Autolock: Why Cache Attacks on ARM Are Harder Than You Think

Total Page:16

File Type:pdf, Size:1020Kb

Autolock: Why Cache Attacks on ARM Are Harder Than You Think AutoLock: Why Cache Attacks on ARM Are Harder Than You Think Marc Green, Worcester Polytechnic Institute; Leandro Rodrigues-Lima and Andreas Zankl, Fraunhofer AISEC; Gorka Irazoqui, Worcester Polytechnic Institute; Johann Heyszl, Fraunhofer AISEC; Thomas Eisenbarth, Worcester Polytechnic Institute https://www.usenix.org/conference/usenixsecurity17/technical-sessions/presentation/green This paper is included in the Proceedings of the 26th USENIX Security Symposium August 16–18, 2017 • Vancouver, BC, Canada ISBN 978-1-931971-40-9 Open access to the Proceedings of the 26th USENIX Security Symposium is sponsored by USENIX AutoLock: Why Cache Attacks on ARM Are Harder Than You Think Marc Green Leandro Rodrigues-Lima Andreas Zankl Worcester Polytechnic Institute Fraunhofer AISEC Fraunhofer AISEC Gorka Irazoqui Johann Heyszl Thomas Eisenbarth Worcester Polytechnic Institute Fraunhofer AISEC Worcester Polytechnic Institute Abstract caches speed up the access to data and instructions, tim- ing measurements allow an adversary to infer the activity Attacks on the microarchitecture of modern processors of other applications and the data processed by them. In have become a practical threat to security and privacy fact, cache attacks have been demonstrated in multiple in desktop and cloud computing. Recently, cache at- scenarios in which our personal data is processed, e.g., tacks have successfully been demonstrated on ARM web browsing [41] or cloud computing [26, 58]. These based mobile devices, suggesting they are as vulner- attacks have severe security implications, as they can re- able as their desktop or server counterparts. In this cover sensitive information such as passwords, crypto- work, we show that previous literature might have left graphic keys, and private user behavior. The majority of an overly pessimistic conclusion of ARM’s security as attacks have been demonstrated on classic desktop and we unveil AutoLock: an internal performance enhance- server hardware [25, 30, 42, 51], and with Intel’s market ment found in inclusive cache levels of ARM proces- share for server processors being over 98% [31], their sors that adversely affects Evict+Time, Prime+Probe, platforms have been targeted most frequently. and Evict+Reload attacks. AutoLock’s presence on system-on-chips (SoCs) is not publicly documented, yet With mobile usage skyrocketing, the feasibility of knowing that it is implemented is vital to correctly as- cache attacks on smartphone and IoT processors – which sess the risk of cache attacks. We therefore provide a de- are predominantly ARM-based – has become a rele- tailed description of the feature and propose three ways vant issue. Attacks that rely on the existence of a to detect its presence on actual SoCs. We illustrate how cache flush instruction, i.e., Flush+Reload [51] and AutoLock impedes cross-core cache evictions, but show Flush+Flush [23], work efficiently across a broad range that its effect can also be compensated in a practical at- of x86 processors, but have limited applicability on ARM tack. Our findings highlight the intricacies of cache at- devices. In general, cache flush instructions serve the tacks on ARM and suggest that a fair and comprehen- legitimate purpose of manually maintaining coherency, sive vulnerability assessment requires an in-depth under- e.g., for memory mapped input-output or self-modifying standing of ARM’s cache architectures and rigorous test- code. On any x86 processor implementing the SSE2 ing across a broad range of ARM based devices. instruction set extension, this flush instruction is avail- able from all privilege levels as clflush. A similar in- struction was introduced for ARM processors only in the 1 Introduction most recent architecture version, ARMv8. In contrast to clflush, it must be specifically enabled to be accessi- The rapid growth of mobile computing illustrates the ble from userspace. This leaves a significant number of continually increasing role of digital services in our daily ARM processors without a cache flush instruction. lives. As more and more information is processed digi- For all processors with a disabled flush instruction or tally, data privacy and security are of utmost importance. an earlier architecture version, e.g., ARMv7, only evic- One of the threats known today aims directly at the fab- tion based cache attacks can be deployed. In particular, ric of digital computing. Attacks on processors and their these attacks are Evict+Time [42], Prime+Probe [42], microarchitecture exploit the very core that handles our and Evict+Reload [24]. On multi-core systems, they data. In particular, processor caches have been exploited target the last-level cache (LLC) to succeed regardless of to retrieve sensitive information across logic boundaries which core a victim process is running on. This requires established by operating systems and hypervisors. As the LLC to be inclusive, i.e., to always contain the con- USENIX Association 26th USENIX Security Symposium 1075 tents of all core-private cache levels. On Intel processors, technical reference manuals (TRMs) nor any other pub- the entire cache hierarchy fulfills the inclusiveness prop- lic product documentation by ARM mention AutoLock. erty and is therefore a viable target for eviction based We therefore provide a detailed description of the fea- attacks. ARM devices, on the contrary, implement inclu- ture and propose three methodologies to test for it: using sive and non-inclusive caches alike. Both properties can a hardware debugging probe, reading the performance co-exist, even in the same cache hierarchy. This renders monitoring unit (PMU), and conducting simple cache- eviction based attacks to be practicable only on a limited timing measurements. Each test strategy has different number of devices, in particular those that implement requirements and reliability; having multiple of them inclusive last-level caches. Yet, our findings show that is vital to test for AutoLock under any circumstances. an internal performance enhancement in inclusive last- With the proposed test suite, we verify AutoLock on level caches, dubbed AutoLock, can still impede evic- ARM Cortex-A7, A15, A53, and A57 processors. As tion based cache attacks. In short, AutoLock prevents AutoLock is likely implemented on a larger number of a processor core from evicting a cache line from an in- ARM processors, we discuss its general implications and clusive last-level cache, if said line is allocated in any how our results relate to previous literature. of the other cores’ private cache levels. This inhibits Despite its adverse effect on eviction based cache at- cross-core LLC evictions, a key requirement for practi- tacks, the impact of AutoLock can be reduced. We dis- cal Evict+Time, Prime+Probe, and Evict+Reload at- cuss generic circumvention strategies and execute the tacks on multi-core systems, and further limits the num- attack by Irazoqui et al. [29] in a practical cross-core ber of ARM based attack targets in practice. Evict+Reload scenario on a Cortex-A15 implementing In literature, Lipp et al. [37], Zhang et al. [54], and AutoLock. We successfully recover the secret key from Zhang et al. [56] confirmed the general feasibility of a table based implementation of AES and show that at- flush and eviction based cache attacks from unprivileged tacks can tolerate AutoLock if multiple cache lines are code on ARM processors. Given the lack of flush in- exploitable. Furthermore, the presented circumvention structions on a large selection of ARM devices and the strategies implicitly facilitate cross-core eviction based deployment of non-inclusive LLCs or inclusive LLCs attacks also on non-inclusive caches. This is because in implementing AutoLock, the authors might have left an the context of cross-core LLC evictions, inclusive last- overly pessimistic conclusion of ARM’s security against level caches with AutoLock behave identically to non- cache attacks. In addition, ARM’s highly flexible licens- inclusive ones. In summary, our main contributions are: ing ecosystem creates a heterogeneous market of system- on-chips (SoCs) that can exhibit significant differences • the disclosure and description of AutoLock, an un- in their microarchitectural implementations. Demme et documented and previously unknown cache imple- al. [17] illustrate that already small changes to the cache mentation feature with adverse impact on practical architecture can have considerable impact on the side- eviction based cache attacks on ARM devices, channel vulnerability of the processor. Consequently, judging the true impact of cache attacks on a broad range • a comprehensive test suite to determine the exis- of ARM based platforms remains to be a challenge. Our tence of AutoLock on actual devices, as its presence work adds another step in this process. It is a contribu- is not documented publicly, tion to an in-depth understanding of microarchitectural • a discussion of AutoLock’s implications and its re- features on ARM in general and an extension to our cur- lation to previous literature demonstrating cache at- rent knowledge of cache implementations in particular. tacks on ARM, and 1.1 Contribution • a set of strategies to circumvent AutoLock to- gether with a practical demonstration of a cross- This work unveils AutoLock, an internal and undocu- core Evict+Reload attack on a multi-core SoC im- mented performance enhancement feature found in in- plementing AutoLock. clusive cache levels on ARM processors. It prevents cross-core evictions of cache lines from inclusive
Recommended publications
  • Data Storage the CPU-Memory
    Data Storage •Disks • Hard disk (HDD) • Solid state drive (SSD) •Random Access Memory • Dynamic RAM (DRAM) • Static RAM (SRAM) •Registers • %rax, %rbx, ... Sean Barker 1 The CPU-Memory Gap 100,000,000.0 10,000,000.0 Disk 1,000,000.0 100,000.0 SSD Disk seek time 10,000.0 SSD access time 1,000.0 DRAM access time Time (ns) Time 100.0 DRAM SRAM access time CPU cycle time 10.0 Effective CPU cycle time 1.0 0.1 CPU 0.0 1985 1990 1995 2000 2003 2005 2010 2015 Year Sean Barker 2 Caching Smaller, faster, more expensive Cache 8 4 9 10 14 3 memory caches a subset of the blocks Data is copied in block-sized 10 4 transfer units Larger, slower, cheaper memory Memory 0 1 2 3 viewed as par@@oned into “blocks” 4 5 6 7 8 9 10 11 12 13 14 15 Sean Barker 3 Cache Hit Request: 14 Cache 8 9 14 3 Memory 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Sean Barker 4 Cache Miss Request: 12 Cache 8 12 9 14 3 12 Request: 12 Memory 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Sean Barker 5 Locality ¢ Temporal locality: ¢ Spa0al locality: Sean Barker 6 Locality Example (1) sum = 0; for (i = 0; i < n; i++) sum += a[i]; return sum; Sean Barker 7 Locality Example (2) int sum_array_rows(int a[M][N]) { int i, j, sum = 0; for (i = 0; i < M; i++) for (j = 0; j < N; j++) sum += a[i][j]; return sum; } Sean Barker 8 Locality Example (3) int sum_array_cols(int a[M][N]) { int i, j, sum = 0; for (j = 0; j < N; j++) for (i = 0; i < M; i++) sum += a[i][j]; return sum; } Sean Barker 9 The Memory Hierarchy The Memory Hierarchy Smaller On 1 cycle to access CPU Chip Registers Faster Storage Costlier instrs can L1, L2 per byte directly Cache(s) ~10’s of cycles to access access (SRAM) Main memory ~100 cycles to access (DRAM) Larger Slower Flash SSD / Local network ~100 M cycles to access Cheaper Local secondary storage (disk) per byte slower Remote secondary storage than local (tapes, Web servers / Internet) disk to access Sean Barker 10.
    [Show full text]
  • GPU Developments 2018
    GPU Developments 2018 2018 GPU Developments 2018 © Copyright Jon Peddie Research 2019. All rights reserved. Reproduction in whole or in part is prohibited without written permission from Jon Peddie Research. This report is the property of Jon Peddie Research (JPR) and made available to a restricted number of clients only upon these terms and conditions. Agreement not to copy or disclose. This report and all future reports or other materials provided by JPR pursuant to this subscription (collectively, “Reports”) are protected by: (i) federal copyright, pursuant to the Copyright Act of 1976; and (ii) the nondisclosure provisions set forth immediately following. License, exclusive use, and agreement not to disclose. Reports are the trade secret property exclusively of JPR and are made available to a restricted number of clients, for their exclusive use and only upon the following terms and conditions. JPR grants site-wide license to read and utilize the information in the Reports, exclusively to the initial subscriber to the Reports, its subsidiaries, divisions, and employees (collectively, “Subscriber”). The Reports shall, at all times, be treated by Subscriber as proprietary and confidential documents, for internal use only. Subscriber agrees that it will not reproduce for or share any of the material in the Reports (“Material”) with any entity or individual other than Subscriber (“Shared Third Party”) (collectively, “Share” or “Sharing”), without the advance written permission of JPR. Subscriber shall be liable for any breach of this agreement and shall be subject to cancellation of its subscription to Reports. Without limiting this liability, Subscriber shall be liable for any damages suffered by JPR as a result of any Sharing of any Material, without advance written permission of JPR.
    [Show full text]
  • Cache-Aware Roofline Model: Upgrading the Loft
    1 Cache-aware Roofline model: Upgrading the loft Aleksandar Ilic, Frederico Pratas, and Leonel Sousa INESC-ID/IST, Technical University of Lisbon, Portugal ilic,fcpp,las @inesc-id.pt f g Abstract—The Roofline model graphically represents the attainable upper bound performance of a computer architecture. This paper analyzes the original Roofline model and proposes a novel approach to provide a more insightful performance modeling of modern architectures by introducing cache-awareness, thus significantly improving the guidelines for application optimization. The proposed model was experimentally verified for different architectures by taking advantage of built-in hardware counters with a curve fitness above 90%. Index Terms—Multicore computer architectures, Performance modeling, Application optimization F 1 INTRODUCTION in Tab. 1. The horizontal part of the Roofline model cor- Driven by the increasing complexity of modern applica- responds to the compute bound region where the Fp can tions, microprocessors provide a huge diversity of compu- be achieved. The slanted part of the model represents the tational characteristics and capabilities. While this diversity memory bound region where the performance is limited is important to fulfill the existing computational needs, it by BD. The ridge point, where the two lines meet, marks imposes challenges to fully exploit architectures’ potential. the minimum I required to achieve Fp. In general, the A model that provides insights into the system performance original Roofline modeling concept [10] ties the Fp and the capabilities is a valuable tool to assist in the development theoretical bandwidth of a single memory level, with the I and optimization of applications and future architectures.
    [Show full text]
  • ARM-Architecture Simulator Archulator
    Embedded ECE Lab Arm Architecture Simulator University of Arizona ARM-Architecture Simulator Archulator Contents Purpose ............................................................................................................................................ 2 Description ...................................................................................................................................... 2 Block Diagram ............................................................................................................................ 2 Component Description .............................................................................................................. 3 Buses ..................................................................................................................................................... 3 µP .......................................................................................................................................................... 3 Caches ................................................................................................................................................... 3 Memory ................................................................................................................................................. 3 CoProcessor .......................................................................................................................................... 3 Using the Archulator ......................................................................................................................
    [Show full text]
  • Multi-Tier Caching Technology™
    Multi-Tier Caching Technology™ Technology Paper Authored by: How Layering an Application’s Cache Improves Performance Modern data storage needs go far beyond just computing. From creative professional environments to desktop systems, Seagate provides solutions for almost any application that requires large volumes of storage. As a leader in NAND, hybrid, SMR and conventional magnetic recording technologies, Seagate® applies different levels of caching and media optimization to benefit performance and capacity. Multi-Tier Caching (MTC) Technology brings the highest performance and areal density to a multitude of applications. Each Seagate product is uniquely tailored to meet the performance requirements of a specific use case with the right memory, NAND, and media type and size. This paper explains how MTC Technology works to optimize hard drive performance. MTC Technology: Key Advantages Capacity requirements can vary greatly from business to business. While the fastest performance can be achieved using Dynamic Random Access Memory (DRAM) cache, the data in the DRAM is not persistent through power cycles and DRAM is very expensive compared to other media. NAND flash data survives through power cycles but it is still very expensive compared to a magnetic storage medium. Magnetic storage media cache offers good performance at a very low cost. On the downside, media cache takes away overall disk drive capacity from PMR or SMR main store media. MTC Technology solves this dilemma by using these diverse media components in combination to offer different levels of performance and capacity at varying price points. By carefully tuning firmware with appropriate cache types and sizes, the end user can experience excellent overall system performance.
    [Show full text]
  • CUDA 11 and A100 - WHAT’S NEW? Markus Hrywniak, 23Rd June 2020 TOPICS for TODAY
    CUDA 11 AND A100 - WHAT’S NEW? Markus Hrywniak, 23rd June 2020 TOPICS FOR TODAY Ampere architecture – A100, powering DGX–A100, HGX-A100... and soon, FZ Jülich‘s JUWELS Booster New CUDA 11 Toolkit release Overview of features Talk next week: Third generation Tensor Cores GTC talks go into much more details. See references! 2 HGX-A100 4-GPU HGX-A100 8-GPU • 4 A100 with NVLINK • 8 A100 with NVSwitch 3 HIERARCHY OF SCALES Multi-System Rack Multi-GPU System Multi-SM GPU Multi-Core SM Unlimited Scale 8 GPUs 108 Multiprocessors 2048 threads 4 AMDAHL’S LAW serial section parallel section serial section Amdahl’s Law parallel section Shortest possible runtime is sum of serial section times Time saved serial section Some Parallelism Increased Parallelism Infinite Parallelism Program time = Parallel sections take less time Parallel sections take no time sum(serial times + parallel times) Serial sections take same time Serial sections take same time 5 OVERCOMING AMDAHL: ASYNCHRONY & LATENCY serial section parallel section serial section Split up serial & parallel components parallel section serial section Some Parallelism Task Parallelism Infinite Parallelism Program time = Parallel sections overlap with serial sections Parallel sections take no time sum(serial times + parallel times) Serial sections take same time 6 OVERCOMING AMDAHL: ASYNCHRONY & LATENCY CUDA Concurrency Mechanisms At Every Scope CUDA Kernel Threads, Warps, Blocks, Barriers Application CUDA Streams, CUDA Graphs Node Multi-Process Service, GPU-Direct System NCCL, CUDA-Aware MPI, NVSHMEM 7 OVERCOMING AMDAHL: ASYNCHRONY & LATENCY Execution Overheads Non-productive latencies (waste) Operation Latency Network latencies Memory read/write File I/O ..
    [Show full text]
  • Sensors and Data Encryption, Two Aspects of Electronics That Used to Be Two Worlds Apart and That Are Now Often Tightly Integrated, One Relying on the Other
    www.eenewseurope.com January 2019 electronics europe News News e-skin beats human touch Swedish startup beats e-Ink on low-power Special Focus: Power Sources european business press November 2011 Electronic Engineering Times Europe1 181231_8-3_Mill_EENE_EU_Snipe.indd 1 12/14/18 3:59 PM 181231_QualR_EENE_EU.indd 1 12/14/18 3:53 PM CONTENTS JANUARY 2019 Dear readers, www.eenewseurope.com January 2019 The Consumer Electronics Show has just closed its doors in Las Vegas, yet the show has opened the mind of many designers, some returning home with new electronics europe News News ideas and possibly new companies to be founded. All the electronic devices unveiled at CES share in common the need for a cheap power source and a lot of research goes into making power sources more sus- tainable. While lithium-ion batteries are commercially mature, their long-term viability is often questioned and new battery chemistries are being investigated for their simpler material sourcing, lower cost and sometime increased energy density. Energy harvesting is another feature that is more and more often inte- e-skin beats human touch grated into wearables but also at grid-level. Our Power Sources feature will give you a market insight and reviews some of the latest findings. Other topics covered in our January edition are Sensors and Data Encryption, two aspects of electronics that used to be two worlds apart and that are now often tightly integrated, one relying on the other. Swedish startup beats e-Ink on low-power With this first edition of 2019, let me wish you all an excellent year and plenty Special Focus: Power Sources of new business opportunities, whether you are designing the future for a european business press startup or working for a well-established company.
    [Show full text]
  • NVIDIA Tegra 4 Family CPU Architecture 4-PLUS-1 Quad Core
    Whitepaper NVIDIA Tegra 4 Family CPU Architecture 4-PLUS-1 Quad core 1 Table of Contents ...................................................................................................................................................................... 1 Introduction .............................................................................................................................................. 3 NVIDIA Tegra 4 Family of Mobile Processors ............................................................................................ 3 Benchmarking CPU Performance .............................................................................................................. 4 Tegra 4 Family CPUs Architected for High Performance and Power Efficiency ......................................... 6 Wider Issue Execution Units for Higher Throughput ............................................................................ 6 Better Memory Level Parallelism from a Larger Instruction Window for Out-of-Order Execution ...... 7 Fast Load-To-Use Logic allows larger L1 Data Cache ............................................................................. 8 Enhanced branch prediction for higher efficiency .............................................................................. 10 Advanced Prefetcher for higher MLP and lower latency .................................................................... 10 Large Unified L2 Cache .......................................................................................................................
    [Show full text]
  • A Characterization of Processor Performance in the VAX-1 L/780
    A Characterization of Processor Performance in the VAX-1 l/780 Joel S. Emer Douglas W. Clark Digital Equipment Corp. Digital Equipment Corp. 77 Reed Road 295 Foster Street Hudson, MA 01749 Littleton, MA 01460 ABSTRACT effect of many architectural and implementation features. This paper reports the results of a study of VAX- llR80 processor performance using a novel hardware Prior related work includes studies of opcode monitoring technique. A micro-PC histogram frequency and other features of instruction- monitor was buiit for these measurements. It kee s a processing [lo. 11,15,161; some studies report timing count of the number of microcode cycles execute z( at Information as well [l, 4,121. each microcode location. Measurement ex eriments were performed on live timesharing wor i loads as After describing our methods and workloads in well as on synthetic workloads of several types. The Section 2, we will re ort the frequencies of various histogram counts allow the calculation of the processor events in 5 ections 3 and 4. Section 5 frequency of various architectural events, such as the resents the complete, detailed timing results, and frequency of different types of opcodes and operand !!Iection 6 concludes the paper. specifiers, as well as the frequency of some im lementation-s ecific events, such as translation bu h er misses. ?phe measurement technique also yields the amount of processing time spent, in various 2. DEFINITIONS AND METHODS activities, such as ordinary microcode computation, memory management, and processor stalls of 2.1 VAX-l l/780 Structure different kinds. This paper reports in detail the amount of time the “average’ VAX instruction The llf780 processor is composed of two major spends in these activities.
    [Show full text]
  • Hardware-Assisted Rootkits: Abusing Performance Counters on the ARM and X86 Architectures
    Hardware-Assisted Rootkits: Abusing Performance Counters on the ARM and x86 Architectures Matt Spisak Endgame, Inc. [email protected] Abstract the OS. With KPP in place, attackers are often forced to move malicious code to less privileged user-mode, to ele- In this paper, a novel hardware-assisted rootkit is intro- vate privileges enabling a hypervisor or TrustZone based duced, which leverages the performance monitoring unit rootkit, or to become more creative in their approach to (PMU) of a CPU. By configuring hardware performance achieving a kernel mode rootkit. counters to count specific architectural events, this re- Early advances in rootkit design focused on low-level search effort proves it is possible to transparently trap hooks to system calls and interrupts within the kernel. system calls and other interrupts driven entirely by the With the introduction of hardware virtualization exten- PMU. This offers an attacker the opportunity to redirect sions, hypervisor based rootkits became a popular area control flow to malicious code without requiring modifi- of study allowing malicious code to run underneath a cations to a kernel image. guest operating system [4, 5]. Another class of OS ag- The approach is demonstrated as a kernel-mode nostic rootkits also emerged that run in System Manage- rootkit on both the ARM and Intel x86-64 architectures ment Mode (SMM) on x86 [6] or within ARM Trust- that is capable of intercepting system calls while evad- Zone [7]. The latter two categories, which leverage vir- ing current kernel patch protection implementations such tualization extensions, SMM on x86, and security exten- as PatchGuard.
    [Show full text]
  • Arxiv:1910.06663V1 [Cs.PF] 15 Oct 2019
    AI Benchmark: All About Deep Learning on Smartphones in 2019 Andrey Ignatov Radu Timofte Andrei Kulik ETH Zurich ETH Zurich Google Research [email protected] [email protected] [email protected] Seungsoo Yang Ke Wang Felix Baum Max Wu Samsung, Inc. Huawei, Inc. Qualcomm, Inc. MediaTek, Inc. [email protected] [email protected] [email protected] [email protected] Lirong Xu Luc Van Gool∗ Unisoc, Inc. ETH Zurich [email protected] [email protected] Abstract compact models as they were running at best on devices with a single-core 600 MHz Arm CPU and 8-128 MB of The performance of mobile AI accelerators has been evolv- RAM. The situation changed after 2010, when mobile de- ing rapidly in the past two years, nearly doubling with each vices started to get multi-core processors, as well as power- new generation of SoCs. The current 4th generation of mo- ful GPUs, DSPs and NPUs, well suitable for machine and bile NPUs is already approaching the results of CUDA- deep learning tasks. At the same time, there was a fast de- compatible Nvidia graphics cards presented not long ago, velopment of the deep learning field, with numerous novel which together with the increased capabilities of mobile approaches and models that were achieving a fundamentally deep learning frameworks makes it possible to run com- new level of performance for many practical tasks, such as plex and deep AI models on mobile devices. In this pa- image classification, photo and speech processing, neural per, we evaluate the performance and compare the results of language understanding, etc.
    [Show full text]
  • A Tour of the ARM Architecture and Its Linux Support
    Linux Conf Australia 2017 A tour of the ARM architecture and its Linux support Thomas Petazzoni Bootlin [email protected] - Kernel, drivers and embedded Linux - Development, consulting, training and support - https://bootlin.com 1/1 I Since 2012: Linux kernel contributor, adding support for Marvell ARM processors I Core contributor to the Buildroot project, an embedded Linux build system I From Toulouse, France Thomas Petazzoni I Thomas Petazzoni I CTO and Embedded Linux engineer at Bootlin I Embedded Linux expertise I Development, consulting and training I Strong open-source focus I Linux kernel contributors, ARM SoC support, kernel maintainers embedded Linux and kernel engineering - Kernel, drivers and embedded Linux - Development, consulting, training and support - https://bootlin.com 2/1 I Core contributor to the Buildroot project, an embedded Linux build system I From Toulouse, France Thomas Petazzoni I Thomas Petazzoni I CTO and Embedded Linux engineer at Bootlin I Embedded Linux expertise I Development, consulting and training I Strong open-source focus I Linux kernel contributors, ARM SoC support, kernel maintainers I Since 2012: Linux kernel contributor, adding support for Marvell ARM processors - Kernel, drivers and embedded Linux - Development, consulting, training and support - https://bootlin.com 2/1 I From Toulouse, France Thomas Petazzoni I Thomas Petazzoni I CTO and Embedded Linux engineer at Bootlin I Embedded Linux expertise I Development, consulting and training I Strong open-source focus I Linux kernel contributors,
    [Show full text]