Hubbry Logo
List of ARM processorsList of ARM processorsMain
Open search
List of ARM processors
Community hub
List of ARM processors
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
List of ARM processors
List of ARM processors
from Wikipedia

This is a list of central processing units based on the ARM family of instruction sets designed by ARM Ltd. and third parties, sorted by version of the ARM instruction set, release and name. In 2005, ARM provided a summary of the numerous vendors who implement ARM cores in their design.[1] Keil also provides a somewhat newer summary of vendors of ARM based processors.[2] ARM further provides a chart[3] displaying an overview of the ARM processor lineup with performance and functionality versus capabilities for the more recent ARM core families.

Processors

[edit]

Designed by ARM

[edit]

Designed by third parties

[edit]

These cores implement the ARM instruction set, and were developed independently by companies with an architectural license from ARM.

Timeline

[edit]

The following table lists each core by the year it was announced.[110][111]

ARM Classic

[edit]
Year Classic cores
ARM1-3 ARM6 ARM7 ARM8 ARM9 ARM10 ARM11
1985 ARM1
1986 ARM2
1989 ARM3
1992 ARM250
1993 ARM60
ARM610
ARM700
1994 ARM710
ARM7DI
ARM7TDMI
1995 ARM710a
1996 ARM810
1997 ARM710T
ARM720T
ARM740T
1998 ARM9TDMI
ARM940T
1999 ARM9E-S
ARM966E-S
2000 ARM920T
ARM922T
ARM946E-S
ARM1020T
2001 ARM7EJ-S
ARM7TDMI-S
ARM9EJ-S
ARM926EJ-S
ARM1020E
ARM1022E
2002 ARM1026EJ-S ARM1136J(F)-S
2003 ARM968E-S ARM1156T2(F)-S
ARM1176JZ(F)-S
2004
2005 ARM11MPCore
2006 ARM996HS

ARM Cortex / Neoverse

[edit]
Year Cortex cores Neoverse cores
Microcontroller
(Cortex-M)
Real-time
(Cortex-R)
Application
(Cortex-A)

(32-bit)
Application
(Cortex-A)

(64-bit)
Application
(Cortex-X)

(64-bit)
Application
(Neoverse)

(64-bit)
2004 Cortex-M3
2005 Cortex-A8
2006
2007 Cortex-M1 Cortex-A9
2008
2009 Cortex-M0 Cortex-A5
2010 Cortex-M4(F) Cortex-A15
2011 Cortex-R4(F)
Cortex-R5(F)
Cortex-R7(F)
Cortex-A7
2012 Cortex-M0+ Cortex-A53
Cortex-A57
2013 Cortex-A12
2014 Cortex-M7(F) Cortex-A17
2015 Cortex-A35
Cortex-A72
2016 Cortex-M23
Cortex-M33(F)
Cortex-R8(F)
Cortex-R52(F)
Cortex-A32 Cortex-A73
2017 Cortex-A55
Cortex-A75
2018 Cortex-M35P(F) Cortex-A65
Cortex-A65AE
Cortex-A76
Cortex-A76AE
2019 Cortex-A34 Cortex-A77 Neoverse E1
Neoverse N1
2020 Cortex-M55(F) Cortex-R82(F) Cortex-A78
Cortex-A78AE
Cortex-A78C
Cortex-X1
[112]
Neoverse V1
[113]
2021 Cortex-A510
Cortex-A710
Cortex-X2 Neoverse E2
Neoverse N2
2022 Cortex-M85(F) Cortex-R52+(F) Cortex-A715 Cortex-X3 Neoverse V2
2023 Cortex-M52(F) Cortex-A520
Cortex-A720
Cortex-X4 Neoverse E3
Neoverse N3
2024 Cortex-R82AE Cortex-A520AE
Cortex-A720AE
Cortex-A725
Cortex-X925 Neoverse V3
Neoverse V3AE
Neoverse VN
2025 Cortex-A320
Cortex-A530
Cortex-A730
Cortex-X930 Neoverse E4
Neoverse N4
Neoverse V4

See also

[edit]

References

[edit]

Further reading

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
The ARM processors comprise a diverse family of central processing unit (CPU) cores based on the ARM architecture, a reduced instruction set computing (RISC) (ISA) originally developed by in the and commercialized through licensing by since 1990. These processors, which implement versions of the ARM ISA from ARMv1 to the current ARMv9, are designed for a wide of applications, including low-power embedded systems, high-performance mobile devices, automotive real-time controllers, and , powering over 325 billion chips shipped worldwide as of 2025. The evolution of ARM processors began with the prototype in 1985, a 32-bit RISC design led by engineers and at , aimed at efficient battery-powered computing for personal devices. This was followed by the ARM2 in 1986, which introduced a hardware multiplier and powered the computer, marking the architecture's entry into commercial products. By 1990, (initially Advanced RISC Machines Ltd.) was formed as a between , Apple, and , adopting an intellectual property (IP) licensing model that allowed third-party manufacturers like , , and to integrate and customize ARM cores into their systems-on-chip (SoCs). This business strategy fueled rapid adoption, with early successes in Apple's Newton PDA (1993) and mobile phones from and in the late 1990s. ARM processors are categorized into several families tailored to specific performance, power, and efficiency needs:
  • Cortex-A series: High-performance application processors for smartphones, tablets, and laptops, supporting advanced features like simultaneous multithreading (SMT) and Armv9's Scalable Vector Extension (SVE2) for AI and machine learning workloads. In 2025, Arm began rebranding its processor lines, moving away from the traditional Cortex nomenclature.
  • Cortex-M series: Ultra-low-power microcontrollers for IoT, wearables, and industrial sensors, emphasizing deterministic execution and energy efficiency in devices like smart home gadgets.
  • Cortex-R series: Real-time processors for automotive, networking, and storage systems, providing low-latency response critical for safety-critical applications such as engine control units.
  • Neoverse series: Infrastructure-grade cores for servers and cloud computing, optimized for high-throughput tasks in data centers, with models like Neoverse V2 delivering scalable performance for hyperscale environments.
Key architectural milestones include the introduction of Thumb mode in ARMv4 (1994) for code density, Jazelle for Java acceleration in ARMv5 (2001), and the shift to the Cortex branding in 2005, which standardized microarchitectures across ARMv6 and beyond for improved scalability and security features like TrustZone. The latest Armv9 (2021) enhances security with and boosts AI capabilities through extensions like Scalable Matrix Extension 2 (SME2), enabling processors to handle complex vector and matrix operations efficiently. Today, ARM's ecosystem dominates —99% of smartphones use ARM-based SoCs—and is expanding into PCs via partnerships with and , as well as automotive via electrification trends. This list catalogs these processors chronologically and by family, highlighting their technical specifications, release dates, and primary use cases to illustrate the architecture's versatility and enduring impact on modern computing.

Instruction Set Architectures

Early Architectures (ARMv1 to ARMv6)

The architecture originated with ARMv1 in , establishing a foundational 32-bit reduced instruction set computing (RISC) design optimized for low power and efficiency in embedded systems. It employed a , where data processing instructions operated only on registers, while loads and stores handled memory access. The processor featured a 3-stage consisting of fetch, decode, and execute stages, enabling simple yet effective instruction throughput. At its core were 16 visible 32-bit general-purpose registers (R0–R12 for general use, R13 as stack pointer, R14 as , and R15 as ), drawn from a total of 37 banked registers to support context switching. Condition flags for negative (N), zero (Z), carry (C), and overflow (V) were maintained in the Current Program Status Register (CPSR). The architecture supported multiple execution modes—User for unprivileged code, for operating system tasks, IRQ for interrupt requests, and FIQ for fast s—to manage privilege levels and exceptions securely. A key innovation was the integrated into (ALU) operations, allowing efficient single-cycle shifts, rotates, or multiplies by powers of two on operands. Both big-endian and little-endian byte ordering were configurable, providing flexibility for diverse system requirements. ARMv2, introduced in 1986, built upon this foundation by adding multiply and multiply-accumulate instructions, which accelerated integer computations essential for early tasks, along with interface support for extending functionality via external units. The register set, , modes, and remained consistent, preserving while enhancing performance. ARMv3, released in 1990, advanced the design to a full 32-bit from the prior 26-bit limitation, improving with added modes including Abort for memory faults, Undefined for unimplemented instructions, and System for privileged operations. It refined load/store multiple (LDM/STM) instructions for efficient block transfers, such as stack operations, and strengthened interworking between modes via banked registers. These changes solidified ARM's suitability for more complex embedded applications without altering the core RISC principles. By 1994, ARMv4 introduced halfword (16-bit) load and store instructions to better handle mixed data sizes, alongside a dedicated kernel privilege mode for low-level system access. The T variant (ARMv4T) marked a significant evolution with instruction set, a 16-bit compressed subset of the 32-bit ARM instructions, aimed at reducing code size by up to 30% in memory-constrained environments like early mobile devices; interworking between ARM and Thumb modes was seamless via branch instructions. Endianness support was formalized with BE-32 for big-endian compatibility. Pipeline designs evolved to 3–5 stages in implementations, balancing performance and complexity. ARMv5, launched in 2001, emphasized (DSP) and embedded with its TE variants: the E extension added DSP instructions like saturated arithmetic, single-instruction multiple-data (SIMD)-like 16-bit multiplies, and packing/unpacking operations for media handling. The J extension introduced DBX, enabling direct execution of Java bytecodes in hardware for faster performance. These built on mode with improved interworking and added instructions like load/store doubleword (LDRD/STRD) for paired accesses. Condition flags expanded in E variants to include a Q flag for saturation overflow. The architecture retained the banked register model and , with pipelines scaling to 5 stages or more in advanced designs. ARMv6, introduced in , further optimized for and real-time embedded use by incorporating SIMD extensions for parallel byte and halfword operations, such as signed/unsigned saturating additions and multiplies, which boosted media processing efficiency without dedicated vector units. It added configurable unaligned access support for words and halfwords, eliminating software workarounds for misaligned data common in packed structures. The Vector Floating Point (VFP) was integrated as an optional extension, providing single- and double-precision floating-point operations with 32 single-precision registers (or 16 double-precision equivalents). Jazelle evolved to RCT (Run-time Compilation and Translation), enhancing dynamic for bytecodes beyond . New exception-handling instructions like Change Processor State (CPS), Store Return State (SRS), and Return From Exception (RFE) streamlined mode switches, while media-specific instructions (e.g., sign/zero extensions and bit reversals) supported embedded audio/video tasks. options expanded to include BE-8 for byte-invariant big-endian. These enhancements, while maintaining 32-bit compatibility, laid groundwork for the multicore and advances in ARMv7.

ARMv7 Architecture

The ARMv7 architecture, introduced in 2006, represents a significant evolution in the ARM instruction set, emphasizing enhanced performance, security, and efficiency for embedded and mobile applications. It defines three distinct profiles tailored to specific use cases: the A-profile for high-performance application processors in devices like smartphones and servers, the R-profile for real-time systems requiring deterministic behavior in automotive and industrial controls, and the M-profile for low-cost, low-power microcontrollers in deeply embedded systems. These profiles share a 32-bit AArch32 execution state while allowing optional extensions for specialized functionality, enabling scalable designs that bridged earlier ARM versions to more advanced paradigms. A key advancement in ARMv7 is the Thumb-2 instruction set, which combines 16-bit and 32-bit instructions to achieve superior code density over the original format, reducing memory footprint by up to 30% in typical workloads without sacrificing performance. Thumb-2 became mandatory in the A-profile and optional in others, facilitating denser binaries for resource-constrained environments. Security features were bolstered with TrustZone, a hardware-enforced partitioning mechanism that divides the system into secure and non-secure worlds, protecting sensitive operations like cryptographic keys from in the normal world. Optional virtualization extensions (denoted as -V) further enable support, allowing multiple guest operating systems to run isolated in the non-secure world while maintaining TrustZone isolation. ARMv7 also introduced (SMP) capabilities, supporting coherent multi-core configurations typically up to four cores with cache coherency protocols for shared memory access. For and , ARMv7 incorporates the advanced SIMD extension and the VFPv3 , providing vectorized operations on 128-bit registers to accelerate tasks like video decoding and image . , optional but widely implemented in A-profile cores, enabled efficient handling of workloads in early smartphones, such as those using the Cortex-A8 processor, by multiple elements in parallel for formats like H.264 video. Power efficiency was improved through hardware support for integer divide instructions (SDIV and UDIV), which execute in a fixed number of cycles to avoid software emulation overhead, alongside advanced sleep modes that allow cores to enter low-power states while retaining context. The variants of ARMv7 reflect its profile-specific optimizations: ARMv7-A targets high-performance scenarios with full Thumb-2, , and support for complex OSes; ARMv7-R emphasizes real-time determinism with enhanced interrupt handling and optional TrustZone for safety-critical applications; and ARMv7-M focuses on low-power, interrupt-driven execution with Thumb-2 and divide instructions but without full virtualization, suiting simple RTOS environments. These features collectively positioned ARMv7 as a foundational architecture for the proliferation of multi-core, secure mobile processors.

ARMv8 Architecture

The ARMv8 architecture, introduced in 2011, marked a significant evolution in the instruction set by incorporating a 64-bit execution state known as alongside with the 32-bit AArch32 mode, enabling seamless transitions for existing software ecosystems. This dual-mode design allowed processors to operate in either state, with providing expanded address spaces up to 64 terabytes and enhanced integer arithmetic capabilities, while AArch32 ensured compatibility with prior ARMv7 applications. The architecture's foundation emphasized performance improvements for mobile devices and emerging server markets, laying the groundwork for widespread adoption in high-efficiency computing. At its core, ARMv8 features 31 general-purpose registers accessible as 64-bit X0-X30 or 32-bit aliases W0-W30, complemented by 32 advanced vector registers for SIMD operations via the enhanced extension, which supports wider data types and fused multiply-add instructions for better parallel processing. Key extensions include optional cryptographic instructions for AES encryption/decryption and SHA-1/SHA-256 hashing, accelerating secure data handling without external coprocessors. support is provided through Exception Levels EL2 (for hypervisors) and EL3 (for secure monitoring), enabling efficient stage-2 address translation and trap handling for multi-tenant environments. ARMv8 defines profiles tailored to specific use cases, with AArch64-A targeting application processors for general computing and AArch64-R focused on real-time systems requiring deterministic performance. Power efficiency is bolstered by instructions like conditional select (CSEL), which avoids branch mispredictions by directly choosing register values based on flags, reducing energy overhead in control-flow intensive code. The architecture supports advanced mechanisms, allowing implementations to minimize pipeline stalls and further optimize power in dynamic workloads. For scalability, ARMv8 theoretically accommodates up to 4096 logical cores through its generic interrupt controller and coherence protocols, paired with for large-scale deployments. A notable application of was Apple's transition to 64-bit processing in the in 2013, which utilized the architecture's first commercial implementation to enhance app performance and pave the way for optimizations. This shift, exemplified in cores like Cortex-A53 and successors, underscored ARMv8's role in enabling efficient 64-bit mobile ecosystems.

ARMv9 Architecture and Extensions

The ARMv9 architecture, introduced in March 2021, represents the latest major evolution of the ARM A-profile instruction set, building upon the 64-bit AArch64 execution state of ARMv8 while introducing enhancements targeted at , , and scalable computing. It maintains full backward compatibility with ARMv8, allowing seamless migration for existing software ecosystems, and emphasizes a "total compute" approach that integrates CPU, GPU, and accelerator optimizations for diverse workloads from edge devices to data centers. A core focus of ARMv9 is advancing AI and capabilities through the Scalable Vector Extension 2 (SVE2), which supports vector lengths ranging from 128 bits to 2048 bits in 128-bit increments, enabling efficient processing of large-scale vector and matrix operations without fixed register widths that limit scalability. SVE2 builds on the original SVE from ARMv8 by adding instructions for , gather-scatter operations, and string processing, making it particularly suited for AI/ML algorithms like training and inference. In workloads, ARMv9 implementations can achieve up to a 30% improvement in () compared to ARMv8 equivalents, driven by these vector enhancements and improved . Security is another pillar of ARMv9, with the Memory Tagging Extension (MTE) providing hardware-assisted detection of errors such as buffer overflows by assigning 4-bit tags to 16-byte memory granules, enabling runtime checks that mitigate common vulnerabilities without significant performance overhead. Complementing MTE, Pointer Authentication Codes (PAC) cryptographically sign pointers to prevent attacks, a feature matured in ARMv9 from its ARMv8.3 origins. The Realm Management Extension (RME), introduced as part of the Confidential Compute Architecture, enables secure enclaves called "Realms" that isolate sensitive code and data from privileged software like hypervisors, supporting dynamic provisioning for in cloud and edge environments. Subsequent branches of ARMv9 have iteratively expanded these foundations. ARMv9.2, released in 2022, introduced the Scalable Matrix Extension (SME) with support for matrix multiply-accumulate operations using up to 256x256 element tiles, optimized for kernels like generalized (GEMM) and enhancing AI performance through bfloat16 and FP16 types. ARMv9.3, announced in 2022, refined branch target identification via an enhanced Branch Record Buffer Extension (BRBE v1p1), which extends control-flow tracing to Exception Level 3 (EL3) for better and in secure scenarios, while also advancing SME with predication and multi-vector support for more flexible AI workloads. The ARMv9.4 branch, announced in 2022, emphasizes and (HPC) profiles, particularly for Neoverse series implementations, by introducing advanced fault handling mechanisms such as Exception-Based Event Profiling (EBEP) and Synchronous Exception-Based Event Profiling (SEBEP), which report performance monitor unit overflows as low-latency exceptions to improve reliability in large-scale systems. Additionally, v9.4 mandates enhancements to SVE2 and SME2, including non-widening bfloat16 arithmetic and 128-bit path support, boosting HPC applications like scientific simulations with up to 512-bit vector operations for decompression and movement. ARMv9.5, announced in 2023, added support for FP8 formats in SME2, SVE2, and to optimize processing, along with checked pointer arithmetic instructions for enhanced security against pointer corruption and features like FEAT_HDBSS for efficient live migration in virtualized environments. ARMv9.6, announced in October 2024, enhanced SME efficiency for AI with support for 2:4 structured sparsity and quarter tile operations, introduced MPAM domains for multi-chiplet systems, and added granular data isolation (GDI) for improved . These extensions underscore ARMv9's role in powering scalable , with brief adoption in recent Cortex-X and Neoverse cores demonstrating its versatility across consumer and server domains.

ARM-Designed Cores

Legacy Cores

The legacy cores represent ARM's foundational 32-bit processor designs, developed from the mid-1980s to the early 2000s, which established the company's emphasis on low-power, RISC-based architectures for embedded systems and early mobile devices. These in-order execution cores, implementing ARMv1 through ARMv6 instruction sets, prioritized efficiency and simplicity, enabling widespread adoption in battery-constrained applications like personal digital assistants and portable gaming. Their evolution introduced features such as pipelining, coprocessors, and compressed instructions, laying the groundwork for ARM's dominance in embedded computing without venturing into out-of-order or multi-core paradigms seen in later designs. The , introduced in 1985, was the inaugural ARM processor, implementing the ARMv2 architecture with a 3-stage and operating at up to 12 MHz on a 3 μm process. Designed initially as a prototype for under a government contract, it focused on high performance-per-watt for the as a co-processor, marking the birth of ARM's RISC without cache or (MMU). Succeeding it, the ARM2 arrived in 1986, refining the ARMv2 architecture while retaining the 3-stage and adding multiply instructions for improved efficiency, with clock speeds reaching 8-18 MHz. The ARM3, released in 1990, advanced to ARMv3 with a 5-stage , integrated 4 KB unified cache, and support for the Floating Point Accelerator (FPA) , achieving up to 25 MHz and enabling more capable personal computing in systems like the . The ARM6 family, launched in 1991, implemented ARMv3 with a 3-stage and full 32-bit addressing, offering variants like the ARM600 with 4 KB cache and optional FPA10 at up to 33 MHz; it powered early mobile phones such as the 6110. Building on this, the ARM7 debuted in 1994 with ARMv3 and evolved to ARMv4T by 1996, featuring a 3-stage , compressed instruction set for code density, and debug support in the popular ARM7TDMI variant. Fabricated on 0.35-0.18 μm processes with clock speeds up to 133 MHz, the ARM7TDMI achieved over 10 billion shipments, dominating early mobile phones, PDAs like the (using ARM610 variant), and handheld games including the . The ARM8, introduced in 1996 as the ARM810, implemented ARMv4 with a , in-order execution, static , and 8 KB unified cache plus MMU, running at up to 72 MHz for enhanced embedded performance without superscalar capabilities. The family, spanning 1997-2001 and supporting ARMv4T to v5TE, featured 5-6 stage pipelines, direct bytecode execution in later variants like the ARM926EJ-S, and DSP extensions for ; the ARM926EJ-S, with variable cache and MMU, reached 200 MHz and was widely used in smartphones and networking gear. Marking a shift toward higher performance, the ARM10 in 2002 implemented ARMv5TE with a 6-stage dual-issue superscalar —the first in ARM's lineup—offering variants like the ARM1020E with 32 KB instruction/data caches and VFP floating-point support for demanding embedded tasks. The family, from 2002-2005, adopted ARMv6 with up to 8-9 stage pipelines, SIMD instructions, Thumb-2 compression, and multi-core support via the ARM11MPCore for (SMP), achieving clocks up to 1 GHz in variants like the ARM1176JZ(F)-S and powering early devices.

Cortex-A and Cortex-X Series

The Cortex-A series comprises ARM's high-performance application processor cores targeted at consumer devices like smartphones and tablets, emphasizing a balance of performance, power efficiency, and support for rich operating systems. Introduced with the ARMv7-A architecture, these cores evolved to incorporate , vector processing via extensions, and multi-core scalability, enabling complex tasks such as rendering and inference. The series progressed to 64-bit ARMv8-A in 2012 and ARMv9-A in 2021, introducing enhancements like Scalable Vector Extensions (SVE) for AI workloads and improved branch prediction for sustained performance. The Cortex-X series, launched in , extends this lineage with custom, ultra-high-performance variants optimized for devices, featuring wider execution units and deeper pipelines to push single-threaded speeds while maintaining compatibility with DynamIQ heterogeneous clustering. Both series support big.LITTLE and DynamIQ technologies for pairing high-performance "big" cores with efficient "LITTLE" ones, optimizing battery life in mobile scenarios. These designs have powered billions of devices, with process nodes scaling from 65 nm for early implementations to 3 nm in recent generations for greater density and efficiency. In September 2025, ARM announced a rebranding for its mobile CPU cores, dropping the "Cortex" name and introducing the Lumex C1 series for next-generation smartphones and devices, focusing on enhanced on-device AI performance and efficiency.
CoreAnnouncement YearKey FeaturesPerformance NotesNotable Uses
Cortex-A82005ARMv7-AFirst out-of-order core, SIMD, up to 1 GHz clockDual-issue superscalar, 2x performance over predecessors (2009)
Cortex-A92007ARMv7-ADual-issue in-order, SMP up to 4 cores, 30-50% faster than A8 in multi-threaded tasks 2
Cortex-A152011ARMv7-A3-wide out-of-order, introduced big.LITTLE pairing with A7Up to 40% IPC uplift over A9, supports 2.5 GHz 5 Octa
Cortex-A72012ARMv7-AEfficient in-order, big.LITTLE companion20% better efficiency than A9 at iso-performancePaired with A15 in early heterogeneous SoCs
Cortex-A572013ARMv8-AHigh-performance out-of-order, 64-bit, 1.9x single-thread perf vs. A15 X1, 810
Cortex-A532012ARMv8-AEfficient in-order, 64-bit, high-density multi-coreBalances perf/watt for background tasksUbiquitous in mid-range Android devices
Cortex-A722015ARMv8-AOut-of-order, big.LITTLE support, improved cache90% perf uplift over A57 at same power 8890
Cortex-A732016ARMv8-APower-optimized out-of-order, 30% efficiency gainSustained perf focus for mobile Kirin 960
Cortex-A752017ARMv8-ADynamIQ compatible, 3-wide decode, out-of-order1.6x faster than A73 in single-threadBroad adoption in 2018 flagships
Cortex-A552018ARMv8-AHigh-efficiency LITTLE, DynamIQ, in-order15% better perf than A53Paired in big.LITTLE configs
Cortex-A762018ARMv8-AWider dispatch (4-wide), DynamIQ, 35% IPC gain over A75 855
Cortex-A772019ARMv8-ADynamIQ, improved load/store, focus20% IPC over A76 865
Cortex-A782020ARMv8.2-ADynamIQ, 20% sustained perf boost, ML enhancements20% IPC gain over A77 at 1W power 1080, foldables
Cortex-X12020ARMv8.2-ACustom high-perf, 6-wide dispatch, DynamIQ30% faster than A78 in single-thread 1080, premium smartphones
Cortex-A7102021ARMv9-AOut-of-order, SVE support, 64-bit only30% energy efficiency vs. A78 2200
Cortex-A7152022ARMv9.2-ABalanced perf/efficiency, matches X1 in some workloads20% power efficiency gain over A710 Dimensity 9200+
Cortex-A7202023ARMv9.2-APremium-efficiency, sustained perf for gamingOptimized for laptops/wearablesExpected in 2024-2025 devices
Cortex-A5202023ARMv9.2-AHigh-efficiency LITTLE, 22% better than A5103x ML perf vs. A55Heterogeneous clusters in mobiles
Cortex-A7252024ARMv9.2-APremium-efficiency, out-of-order, AI-focusedBalances speed and battery life2025 smartphone SoCs
Cortex-X22022ARMv9-AWider execution, improved ML accel30-35% perf over X1 8 Gen 2
Cortex-X32023ARMv9.2-AEnhanced , vector processing11% IPC over X2 Dimensity 9300
Cortex-X42024ARMv9.3-AFlagship perf, SME for matrix math15% faster than X3 in single-threadExpected 2025 flagships
Cortex-X9252024ARMv9.3-AUltimate perf, 15% IPC uplift, SME2 for AI25% single-thread vs. prior gen GB10, 2025 premium devices
Lumex C1-Ultra2025Armv9.3-AFlagship AI-focused, rebranded from Cortex-X, enhanced SME2Up to 25% faster than X925 in single-thread, 45% multi-core upliftExpected in 2026 flagship smartphones
Lumex C1-Pro2025Armv9.3-AHigh-performance balanced core, AI accelerationSuccessor to A725, improved efficiencyPremium mobile SoCs 2026
Lumex C1-Premium2025Armv9.3-AMid-range performance coreBalanced perf/watt for mainstream devices2026 mid-tier smartphones
Lumex C1-Nano2025Armv9.3-AUltra-efficient LITTLE core, rebranded from A520High-density, low-power for clustersEfficient cores in heterogeneous 2026 designs
Performance advancements in the Cortex-A and X series are quantified through instructions per cycle (IPC) improvements, reflecting architectural refinements like increased reorder buffers and better prefetching; for instance, the transition from Cortex-A77 to A78 delivered a 20% IPC gain at constrained power envelopes, enabling longer battery life in 5G-enabled devices. Later v9 cores like the A710 and X925 incorporate Armv9 extensions for enhanced security (e.g., Pointer Authentication) and AI acceleration, with the X925 providing a 15% IPC boost over the X4 through optimized front-end decoding. These metrics, derived from ARM's standardized benchmarks, underscore the series' focus on perf-per-watt scaling, with modern implementations on 3 nm nodes achieving up to 30% ML inference uplift via Scalable Matrix Extension 2 (SME2) in targeted workloads. The 2025 Lumex C1 series further advances this with up to 45% multi-core performance gains and doubled AI capabilities for on-device processing.

Cortex-M Series

The Cortex-M series comprises ARM's microcontroller-oriented processors, tailored for low-power embedded systems, IoT endpoints, and deterministic applications requiring high code density and interrupt responsiveness. These cores implement the M-profile architecture, emphasizing energy efficiency and integration with peripherals like NVIC for low-latency interrupts, making them ideal for devices from sensors to wearables. Over 50 billion Cortex-M based chips have been shipped globally, powering diverse markets including industrial controls and . The series began with foundational cores supporting ARMv6-M, evolving to incorporate advanced features like floating-point units, DSP extensions, and security enhancements in later ARMv7-M and ARMv8-M iterations. Key designs prioritize a balance of performance, area, and power, with metrics such as DMIPS/MHz indicating efficiency— for instance, early cores achieve around 0.9 DMIPS/MHz while later ones exceed 2 DMIPS/MHz without compromising low-power profiles under 1 mW in active states. Cortex-M0, introduced in 2009, is the entry-level core based on ARMv6-M with a 2-stage and Thumb-only instruction set, targeting ultra-low-power applications like smart sensors; it consumes less than 1 mW at typical operating frequencies and supports up to 32 external interrupts via a basic NVIC. Cortex-M0+, released in 2010 and also on ARMv6-M, refines the M0 with enhanced debug capabilities, sleep modes, and fault handling, reducing active power by up to 30% over its predecessor; it is widely integrated in microcontrollers such as the family for cost-sensitive IoT nodes. Cortex-M1, announced in 2007 as an ARMv6-M FPGA soft core, facilitates and ASIC migration with a 2-stage pipeline and AHB-Lite bus interface, optimized for low gate count in programmable logic environments. Shifting to ARMv7-M, Cortex-M3, launched in 2004, introduced a 3-stage pipeline, Thumb-2 instructions for improved code density, and an advanced NVIC supporting up to 240 interrupts with sub-microsecond latency, establishing a benchmark for general-purpose embedded processing at 1.25 DMIPS/MHz. Cortex-M4, from 2010 and based on ARMv7-M, extends the M3 with single-precision FPU and DSP instructions for tasks, delivering 1.25 DMIPS/MHz while enabling efficient vector math in applications like . Cortex-M7, unveiled in 2014 on ARMv7-M, offers the highest performance in the pre-v8 lineup with a 6-stage dual-issue , double-precision FPU, and optional L1 caches up to 64 KB, supporting clock speeds to 1 GHz and 2.14 DMIPS/MHz for demanding DSP and control in automotive and industrial systems. Transitioning to ARMv8-M, Cortex-M33, introduced in 2017, integrates optional TrustZone for secure IoT partitioning, MPU with 16 regions, and DSP support at 1.5 DMIPS/MHz, facilitating isolated execution environments in connected devices. Cortex-M55, announced in 2020 under ARMv8.1-M Mainline, incorporates vector extensions for acceleration, yielding up to 15x ML performance uplift and 15% better efficiency than the M7 at 1.6 DMIPS/MHz, with support for up to 16 MB TCM and AXI bus. Cortex-M85, released in 2021 on ARMv8.1-M, builds on the M55 with enhanced security via Pointer Authentication and Branch Target Identification, plus for superior scalar and ML workloads, targeting secure, high-performance embedded AI at up to 4x ML inference speed over prior generations. More recently, Cortex-M52, launched in 2023 based on ARMv8.2-M, provides a compact -enabled core for cost-optimized AIoT, emphasizing area efficiency and deterministic behavior for industrial endpoints with ML needs, as the smallest such processor in the series.
ProcessorArchitecturePipeline StagesKey FeaturesDMIPS/MHzTypical Power
Cortex-M0ARMv6-M2Thumb-only, basic NVIC0.87<1 mW
Cortex-M0+ARMv6-M2Enhanced debug/sleep0.95<0.7 mW
Cortex-M1ARMv6-M2FPGA soft core0.8Low gate count
Cortex-M3ARMv7-M3Thumb-2, NVIC (240 ints)1.25Low-cost
Cortex-M4ARMv7-M3FPU, DSP1.25Signal control
Cortex-M7ARMv7-M6 (dual-issue)Double FPU, caches/TCM2.14Up to 1 GHz
Cortex-M33ARMv8-M3TrustZone, MPU (16)1.5Secure IoT
Cortex-M55ARMv8.1-M4 ML, AXI/TCM1.615% > M7
Cortex-M85ARMv8.1-M4PAC/BTI security, ~2.0High-security ML
Cortex-M52ARMv8.2-M3Compact ~1.6Area-efficient AI
This table summarizes representative specifications, drawn from ARM's comparison data; actual implementations vary by configuration.

Cortex-R Series

The Cortex-R series comprises ARM-designed processors optimized for real-time applications in safety-critical domains such as automotive systems, industrial control, and storage controllers, emphasizing deterministic , low-latency interrupt handling, and through features like error-correcting code ( support and units (MPUs). These cores implement the R-profile of the , starting with ARMv7-R for 32-bit operation and extending to ARMv8-R for enhanced 64-bit capabilities in later models, enabling predictable execution in environments requiring high reliability. Unlike application or microcontroller-focused series, Cortex-R processors prioritize time-sensitive tasks with configurable tightly coupled memory (TCM) for zero-latency access and support for (SMP) to scale while maintaining real-time guarantees. Key safety mechanisms across the series include dual-core lockstep (DCLS) for fault detection and compliance with up to ASIL-D when configured appropriately, alongside time-triggered interrupt controllers for precise scheduling in automotive and industrial use cases. The Cortex-R4, introduced in 2006, is a foundational 32-bit processor based on the ARMv7-R , featuring an 8-stage in-order dual-issue for efficient real-time processing in embedded systems. It supports ECC on TCM and data caches to enhance reliability, along with a 12-region MPU for memory isolation, and is designed for single-core configurations without hardware coherency, making it suitable for cost-sensitive deterministic applications like . reaches up to 1.67 DMIPS/MHz, with configurable 4-64 KB instruction and data caches, and 0-8 MB TCM for low-latency code execution. Dual-core modes enable for safety-critical deployments. Building on the R4, the Cortex-R5, released in 2008, refines the ARMv7-R design with an 8-stage in-order dual-issue pipeline and improved deterministic performance, achieving up to 1.5 GHz clock speeds in advanced nodes for applications like and systems. It introduces enhanced error management, including 16-region MPU, DCLS for redundant execution, and optional single-thread (STL) for fault detection, alongside ECC support on caches and TCM up to 8 MB. Dual-core configurations with I/O coherency enable scalability for real-time multitasking, delivering 1.67 DMIPS/MHz while prioritizing predictability over raw throughput. This core's focus on safety has made it prevalent in automotive ECUs certified to ASIL-D. The Cortex-R7, announced in 2011, advances the series with an 11-stage out-of-order superscalar pipeline on ARMv7-R, offering 2.5 DMIPS/MHz for higher performance in demanding real-time scenarios like processing. It supports up to four cores in SMP with L1/L2 cache hierarchies (4-64 KB L1, up to 1 MB L2 shared), 16-region MPU, and ECC for , though it lacks native DCLS. Time-triggered interrupts via the Generic Interrupt Controller facilitate precise real-time scheduling, and its design emphasizes for industrial and automotive control systems. Note that the R7 has been discontinued in favor of newer models. Launched in 2016, the Cortex-R8 is a high-end 32-bit ARMv7-R processor with an 11-stage out-of-order superscalar , targeting 2.5 DMIPS/MHz for applications in LTE/ modems, storage controllers, and advanced driver-assistance systems (ADAS). It features a 24-region MPU, ECC on 0-1 MB TCM and 0-64 KB caches, and supports up to four cores in SMP configurations for balanced real-time and throughput needs. The core's deterministic response times and optional Neon SIMD extensions enhance efficiency in data-intensive tasks, with ASIL-D support through and diagnostic features. Dual-core configurations provide redundancy for automotive safety. The Cortex-R82, introduced in , marks the series' shift to 64-bit ARMv8.2-R with an 8-stage in-order triple-issue , delivering 3.4 DMIPS/MHz—up to 2.25 times faster than the R8 in real-world workloads—for high-performance real-time processing in storage and automotive domains. It supports up to eight cores with hardware coherency, 32+32 region MPU/MMUs, ECC on 0.16-1 MB TCM, 16-128 KB L1 instruction/16-64 KB data caches, and optional 0-4 MB L2 cache, enabling up to 1 TB DRAM addressing for computational storage. Safety features include DCLS and STL for ASIL-D compliance, with time-triggered interrupts ensuring predictability in ADAS and autonomous vehicle controls. Optional and machine learning extensions further boost efficiency in 2025-era automotive applications.
ProcessorArchitecturePipelineDMIPS/MHzMax CoresKey Safety FeaturesPrimary Applications
Cortex-R4ARMv7-R8-stage in-order dual-issue1.671ECC, 12-region MPU, dual-core Embedded real-time control
Cortex-R5ARMv7-R8-stage in-order dual-issue1.672 (IO coherency)ECC, 16-region MPU, DCLS, STL, automotive ECUs
Cortex-R7ARMv7-R11-stage OoO superscalar2.54 (SMP)ECC, 16-region MPU, industrial systems (discontinued)
Cortex-R8ARMv7-R11-stage OoO superscalar2.54 (SMP)ECC, 24-region MPU, modems, ADAS, storage
Cortex-R82ARMv8.2-R8-stage in-order triple-issue3.48 (coherency)ECC, 32+32 MPU/MMUs, DCLS, STLComputational storage, autonomous vehicles
This table summarizes core specifications for quick comparison, highlighting the evolution toward higher performance and safety in automotive-focused designs as of 2025.

Neoverse Series

The Neoverse series comprises Arm's family of 64-bit CPU cores tailored for infrastructure workloads, emphasizing scalability, energy efficiency, and performance in data centers, , high-performance computing (HPC), and edge applications. These cores support high core counts—up to hundreds in multi-socket systems—and integrate advanced interconnects like the Coherent Mesh Network (CMN) for cache-coherent scaling across clusters. Unlike mobile-oriented designs, Neoverse prioritizes sustained throughput for server environments, with features such as enhanced branch prediction, larger caches, and vector processing extensions to handle diverse workloads from general-purpose to AI acceleration. The series builds on the Armv8 and Armv9 instruction set architectures, incorporating extensions for and vectorization to enable efficient multi-threaded operation in large-scale deployments. Introduced in 2019, the Neoverse N1 core implements the Armv8.2-A architecture and derives from the Cortex-A76 microarchitecture, featuring a 4-wide pipeline, 64 KB L1 instruction and data caches per core, and up to 1 MB private L2 cache. It supports up to 128 cores in a single coherent domain via a interconnect, enabling high-density configurations for cloud-native applications. The N1 delivers up to 2.5 times the performance of prior-generation Arm infrastructure cores on key cloud benchmarks, with optimizations for and exceeding 175 GB/s in 64-core systems. It powers the AWS Graviton2 processor, which integrates 64 N1 cores on a for cost-effective EC2 instances offering 40% better price-performance over comparable x86 alternatives. The Neoverse V1, announced in 2020, targets HPC and AI/ML workloads with the Armv8.4-A architecture, including the first implementation of Scalable Vector Extension (SVE) for up to 2x 256-bit vector pipelines and double the floating-point throughput of scalar designs. This out-of-order core provides over 50% instructions-per-cycle (IPC) uplift compared to the N1, with a 10-stage pipeline, 1 MB L2 cache per core, and support for up to 64 cores per cluster. It excels in vector-heavy tasks, achieving leadership in HPC benchmarks through SVE's flexible vector lengths. Deployments include Fujitsu's systems for scientific simulations, contributing to Arm-based advancements in exascale computing ecosystems. Launched in 2021, the Neoverse N2 core advances efficiency with the Armv8.6-A architecture, offering a 40% IPC improvement over the through enhancements like a 1.5K-entry micro-op cache, 5-wide dispatch, and 13 execution ports. It features 64 KB L1 caches and configurable 512 KB to 1 MB L2 per core, with support for up to 192 cores at 350 W TDP or low-power 8-core variants at 20 W. The design yields up to 50% better performance-per-watt for scale-out workloads, including improved branch prediction and subsystem latency under 4 cycles for loads. It underpins the AWS Graviton3 processor with 64 N2 cores, delivering 25% faster performance and 60% greater energy efficiency than Graviton2 for EC2 instances focused on web services and databases. The Neoverse V2, released in 2022, builds on Armv9.1-A with SVE2 extensions for enhanced vector processing, providing up to twice the performance of the V1 through wider 512-bit vectors and improved matrix multiply acceleration. This high-performance core supports up to 128 cores (scalable to 256 with CMN-700 mesh), 2 MB L2 cache per core, and a 6-wide decode with 12K-entry branch target buffer for reduced misprediction penalties. It targets HPC and cloud AI, with 1.7x overall performance gains over V1 in floating-point intensive tasks. Notable implementations include the Grace CPU Superchip, featuring 72 V2 cores for AI training and at up to 114 cores in dual-socket configurations. Announced in 2024 and available in 2025, the Neoverse N3 core utilizes Armv9.2-A for cloud and general-purpose , delivering a 30% IPC uplift over the N2 via a refined 5-wide , 2 MB L2 cache options, and advanced ML optimizations like int8 and bfloat16 support. It scales to 64 cores per cluster with 48-bit physical addressing and ECC-protected caches, emphasizing power efficiency for dense server racks. The core achieves nearly triple the ML inference throughput of N2, suitable for hyperscale environments. Early adoptions include custom designs from cloud providers targeting and edge . The Neoverse V3, introduced in 2024 under Armv9.2-A, focuses on maximum performance for HPC, cloud, and ML with via Realm Management Extension (RME) for secure enclaves. It supports up to 128 cores per socket, 4 MB L2 cache per , and PCIe Gen5/CXL 3.0 integration, offering 1.3x the performance of V2 in integer and floating-point workloads. Designed for 2025 server deployments, it includes enhanced SVE2 for AI vectorization and branch prediction with 2x larger history tables. Implementations are slated for next-generation supercomputers and AI clusters, with and others developing V3-based systems. Across the series, Neoverse cores support CCIX and CHI protocols for in multi-socket setups, enabling seamless data sharing up to thousands of cores in disaggregated systems. This infrastructure optimizes Armv9 extensions like pointer authentication and memory tagging for enhanced security in virtualized environments. In November 2025, announced integration of Fusion into Neoverse platforms for improved coherency with GPUs in AI data centers.

Third-Party Designed Cores

Apple Processors

Apple began designing its own -based system-on-chip (SoC) processors in 2010 with the A4, marking a shift from licensing off-the-shelf ARM cores to creating custom microarchitectures optimized for its and later macOS ecosystems. These processors power iPhones, iPads, and other devices, emphasizing power efficiency, integrated graphics, and neural processing units (NPUs) for AI tasks. The A-series targets mobile devices, while the M-series, introduced in 2020, extends to Macs and high-end iPads with unified that shares RAM across CPU, GPU, and other components for reduced latency and higher bandwidth. The A4, released in 2010, was Apple's first custom ARMv7-based SoC, featuring a single-core CPU derived from the design, clocked at up to 1.0 GHz, and integrated with a PowerVR SGX535 GPU; it debuted in the and first-generation . In 2011, the A5 introduced dual cores at up to 1.0 GHz on ARMv7, doubling CPU performance over the A4 while maintaining similar power envelope, and powered the and ; the A5X variant enhanced GPU performance with quad-core graphics for the third-generation . The A6, launched in 2012 for the , refined the microarchitecture on a for 40% better CPU speed and 2x graphics performance compared to the A5, still on ARMv7. Apple's transition to came with the A7 in 2013, the first ARMv8-based processor with a dual-core custom core at 1.3-1.4 GHz, delivering up to 2x faster CPU and GPU than the A6 while introducing 64-bit support for ; it featured in the . The A8 in 2014 for the and 6 Plus used a dual-core core on ARMv8 at up to 1.4 GHz, built on a 20 nm process with a 50% faster CPU and 2.5x GPU uplift over the A7, plus the first dedicated motion M8. Subsequent A9 (2015, Twister cores, ) and A10 Fusion (2016, Hurricane cores, ) on ARMv8 continued scaling with quad-core configurations starting in the A10—two high- and two efficiency cores—inspired by big.LITTLE asymmetry for balanced power and . From the A11 Bionic in 2017 for the , Apple fully embraced custom microarchitectures with high-performance and Mistral efficiency cores in a hexa-core setup on ARMv8, adding the first embedded Neural Engine for acceleration with 600 billion operations per second (OPS). The A12 Bionic (2018, Vortex/Tempest, ) on ARMv8.2 advanced to 7 nm with an upgraded 8-core Neural Engine at 5 trillion OPS, while the A13 Bionic (2019, Lightning/Liberty, ) refined efficiency on 7 nm. The A14 Bionic (2020, Firestorm/Icestorm, ) on 5 nm introduced the first M-series-like cores to mobile, with a 16-core Neural Engine at 11 trillion OPS. The A15 Bionic (2021, ) and A16 Bionic (2022, /Sawtooth, ) on ARMv8.5 scaled to 5 nm and 4 nm processes, enhancing AI with up to 17 trillion OPS and deeper Neural Engine integration. The A17 Pro in 2023 for the marked a leap to ARMv8.6 on a with dual high-performance cores at up to 3.78 GHz, four efficiency cores at 2.11 GHz, and a 16-core Neural at 35 trillion OPS, enabling advanced ray tracing in the GPU. The A18 and A18 Pro in 2024 for the 16 series adopted ARMv9.2-A on , with the A18 featuring a 6-core CPU (2 performance at 4.04 GHz, 4 efficiency at 2.2 GHz) and 30% faster CPU performance than the A17 Pro, alongside a 16-core Neural optimized for Apple Intelligence AI features.
ProcessorYearARM VersionCPU Cores (Perf/Eff)Process NodeKey InnovationDebut Device
A42010v7145 nmFirst custom Swift core
A52011v7245 nmDual-core debut
A62012v7222 nmProcess shrink for efficiency
A72013v8228 nmFirst 64-bit
A82014v8220 nmIntegrated M8
A92015v8214 nmPeak performance focus
A10 Fusion2016v82/216 nmFirst big.LITTLE
A11 Bionic2017v82/410 nmFirst Neural Engine
A12 Bionic2018v8.22/47 nm process, AI uplift
A13 Bionic2019v8.42/4Efficiency refinements
A14 Bionic2020v8.52/45 nm cores
A15 Bionic2021v8.52/45 nmScalable core count
A16 Bionic2022v8.52/44 nmAdvanced media engine
A17 Pro2023v8.62/43 nmRay tracing GPU
A18/A18 Pro2024v9.22/43 nmApple Intelligence NPUiPhone 16
The M-series, licensed under ARMv8 and later, debuted with the M1 in 2020 for the , , and , featuring an 8-core CPU (4 performance cores at up to 3.2 GHz, 4 Icestorm efficiency cores) on 5 nm with unified LPDDR4X memory up to 16 GB at 68 GB/s bandwidth, an 8-core GPU, and 16-core Neural Engine at 11 OPS, achieving up to 3.5x faster CPU performance than i7 equivalents in laptops. The M1 Pro and M1 Max variants scaled to 10 CPU cores and up to 32 GPU cores for pro workflows. The in 2022 for updated Macs introduced / cores on ARMv9 at 5 nm (later 3 nm for some), with up to 8 CPU cores, 10-core GPU, and 16-core Neural Engine at 15.8 OPS, plus hardware-accelerated ray tracing. The M3 family in 2023 for and used ARMv9.2 on 3 nm with up to 8 performance cores at 4.05 GHz, 10-core GPU supporting dynamic caching and mesh shading, and enhanced media engines for decode, delivering 65% faster CPU performance over M1. The M4 in 2024 for and later Macs refined this on second-generation 3 nm with up to 10 CPU cores (including 4 performance, 6 efficiency), a 10-core GPU with 38 trillion OPS Neural Engine, and up to 120 GB/s , focusing on AI with 4x faster NPU than M2. In October 2025, Apple announced the M5 for , , and Vision Pro, built on third-generation with a next-generation 10-core GPU including per-core Neural Accelerators for over 4x peak GPU performance in AI tasks compared to M1, up to 20% faster multithreaded CPU than M4, and up to 153 GB/s, emphasizing tensor processing for on-device AI. These designs post-A11 and across M-series incorporate big.LITTLE-inspired heterogeneous cores, custom wide-vector execution units, and deep integration of the Neural Engine, enabling features like real-time without cloud dependency.

Qualcomm Processors

Qualcomm has developed a range of ARM-based processors primarily under its Snapdragon brand, targeting mobile devices, PCs, and embedded systems, with a focus on custom and licensed core designs to optimize and in Android smartphones and Windows on Arm laptops. The company's evolution from early custom architectures to adopting ARM's Cortex cores and back to proprietary designs reflects adaptations to power constraints and market demands for AI and multitasking capabilities. The Scorpion core, introduced in 2007 as part of the Snapdragon S1 platform, marked Qualcomm's first custom ARM-compatible CPU, implementing the ARMv6 instruction set architecture in a single-core configuration clocked up to 1 GHz. Designed with a 13-stage pipeline for higher clock speeds, Scorpion powered early 3G smartphones like the HTC Dream, emphasizing multimedia processing alongside an integrated Adreno GPU. Its superscalar design delivered improved integer and floating-point performance over standard ARM11 cores, though limited to 32-bit operations. Building on this, the Krait cores, debuted in 2011 with the Snapdragon S4 series, represented Qualcomm's custom implementation of the ARMv7-A architecture, supporting both 32-bit and early 64-bit extensions in dual- and quad-core variants up to 2.3 GHz. Krait, used through 2014 in devices like the HTC One and , featured and SIMD support, achieving up to 40% better efficiency than equivalents while pairing with 320 GPUs for enhanced graphics. This custom allowed Qualcomm to tailor branch prediction and cache hierarchies for mobile workloads, powering the shift to LTE-enabled handsets. The family, launched in 2016 with the Snapdragon 820, introduced Qualcomm's ARMv8-A 64-bit custom cores, initially based on modifications to ARM's Cortex-A57 and A53 designs in a quad-core big.LITTLE configuration up to 2.15 GHz. Evolving to fully custom variants like 385 in the 2017 Snapdragon 835, these octa-core setups on 10 nm processes delivered 30% better CPU performance over predecessors, integrated with 540 GPUs for VR support in flagships like the S8. By 2018, the Snapdragon 845's 385 further refined custom elements for AI acceleration, emphasizing for gaming and camera tasks. From 2018 to 2020, the 4xx and 5xx series shifted to licensed Cortex cores under ARMv8.2-A, with the Snapdragon 855 featuring 485 (based on Cortex-A76 for performance cores at 2.84 GHz and A55 for efficiency at 1.78 GHz) in an octa-core layout on 7 nm. The 2019 Snapdragon 865 advanced this to 585 (Cortex-A77-based primes at 2.84 GHz with A55 efficiency cores), offering 25% CPU uplift and improved power efficiency for devices like the Galaxy S20, while supporting advanced ISP for . These designs prioritized big.LITTLE asymmetry to balance sustained performance in multitasking scenarios. Qualcomm's Oryon cores, a fully custom announced in and implementing ARMv8.6+ extensions (evolving to v8.7 and v9), debuted in production in 2024, marking a return to proprietary designs for superior IPC and efficiency. Featuring wide execution units and large caches, Oryon powers the Snapdragon 8 Gen 1 through Gen 3 platforms (–2024), though early generations like the 8 Gen 1 on 4 nm process used hybrid configs up to 1+3+4 cores (Cortex-based initially transitioning to Oryon influences); by Gen 3, it adopted 1+5+2 layouts with Oryon primes at 3.3 GHz for AI-driven tasks in devices like the Galaxy S24. The 2024 Snapdragon 8 Gen 4 (also branded Elite) on 3 nm fully leverages second-generation Oryon in an all-big-core 2+6 configuration up to 4.32 GHz, delivering 45% faster CPU performance than Gen 3 while enhancing NPU for on-device generative AI. Plans for 2025 include 2 nm iterations with further custom optimizations. In the PC space, the 2024 Snapdragon X Elite integrates 12 Oryon cores under ARMv9.1 on 4 nm, configured in three quad-core clusters with dual-core boost to 4.2 GHz, targeting Windows on Arm laptops for multi-day battery life and native x86 emulation via Prism. It outperforms prior ARM PCs in multi-threaded workloads, with 42 MB cache enabling seamless productivity and gaming. The Snapdragon X Plus variant offers 10- or 8-core options up to 3.4 GHz for mid-range laptops, while 2025's Snapdragon X2 Elite Extreme (enhanced X Plus successor) introduces third-generation Oryon with up to 18 cores on 3 nm nodes for improved AI inference and connectivity. This expansion underscores Qualcomm's 2025 shift to predominantly custom ARM cores across portfolios, reducing reliance on licensed IP for differentiated efficiency.

Samsung Processors

series represents its line of system-on-chip (SoC) processors based on architecture, primarily designed for mobile devices such as smartphones and wearables in the lineup. These processors often feature a mix of licensed Cortex cores and, in earlier generations, custom CPU designs, optimized for performance in premium devices while offering regional variants to complement Qualcomm's Snapdragon chips in global markets. SoCs have evolved from single-core ARMv7 implementations to advanced ARMv9-based deca-core configurations, emphasizing AI capabilities, power efficiency, and integration with ecosystem. The series began with the processor in 2010, an ARMv7-based single-core design derived from the Cortex-A8, clocked at 1 GHz and fabricated on a . It powered the original Galaxy S and smartphones, supporting 1080p video recording and LPDDR2 memory, marking Samsung's entry into in-house mobile SoC development. This was followed by the Exynos 4210 in 2011, a dual-core ARMv7 Cortex-A9 implementation at 1.2 GHz on a , used in devices like the Galaxy S II and Galaxy Note, with Mali-400MP4 GPU for 1080p/30fps playback. Samsung introduced custom CPU cores with the architecture in 2016, starting with the M1 in the 8890 SoC, an ARMv8-based quad-core design paired with four Cortex-A53 efficiency cores on a . The 8890 powered the S7 series and Note 7, featuring a Mali-T880 MP12 GPU and support for 4K/60fps video. Subsequent iterations advanced to the M3 in the 9810 () and M4 in the 9820/9825 (2019), both ARMv8.2 compliant, combining custom cores with Cortex-A55 for devices like the S9, Note 9, S10, and Note 10 series, adding NPU for AI tasks and 8K video support on 10 nm and 8 nm nodes. The 9xx series (9810–990) represented Samsung's peak in custom CPU innovation before shifting to fully licensed ARM designs. Transitioning to ARMv9, the 2100 in 2021 featured a single Cortex-X1 prime core alongside three Cortex-A78 performance cores and four Cortex-A55 efficiency cores on a , integrated with a Mali-G78 MP14 GPU and modem for the S21 series, supporting 200 MP cameras. This was followed by the 2200 in 2022, with a Cortex-X2 prime core, three Cortex-A710 cores, and four Cortex-A510 cores under ARMv9.1, paired with an AMD-designed Xclipse 920 GPU based on RDNA2 for ray tracing, targeted for S22 but ultimately limited to select regions due to yield issues. The 2400, released in 2024, marks a return to all-Cortex configurations under ARMv9.2, with a 10-core setup: one Cortex-X4 at 3.21 GHz, five Cortex-A720 (two at 2.9 GHz, three at 2.6 GHz), and four Cortex-A520 at 1.95 GHz, built on Samsung's 4 nm LPP+ process for improved efficiency. It includes the Xclipse 940 GPU ( RDNA3-based) and powers international variants of the S24 series, emphasizing AI processing with a 14.7x NPU uplift over predecessors. Looking ahead, the 2500, entering mass production in 2025, adopts ARMv9 with a deca-core layout: one Cortex-X925 at 3.3 GHz, seven Cortex-A725 (two at 2.74 GHz, five at 2.36 GHz), and two Cortex-A520 at 1.8 GHz on a 3 nm GAA process, featuring an enhanced NPU and Xclipse 950 GPU for Z Flip 7 devices. For wearables, Samsung's W9xx series includes the W930 (2023, dual Cortex-A55 at 1.4 GHz on 5 nm) for Galaxy Watch6 and the W1000 (2024, one Cortex-A78 at 1.6 GHz plus four Cortex-A55 at 1.5 GHz on 3 nm GAA) for Galaxy Watch7, both with Mali-G68 GPUs to enable always-on displays and GNSS tracking. These processors highlight Samsung's strategy of regional optimization, where variants are deployed in non-U.S. Galaxy models to leverage control and tailored performance.
ProcessorYearARM VersionCore ConfigurationProcess NodeKey DevicesNotable Features
Hummingbird (Exynos 3110)2010v71x Cortex-A8 @1 GHz45 nmGalaxy S, Nexus S1080p video, PowerVR SGX540 GPU
Exynos 42102011v72x Cortex-A9 @1.2 GHz45 nmGalaxy S II, Galaxy NoteMali-400MP4 GPU, 1080p/30fps
Exynos 8890 (Mongoose M1)2016v84x M1 + 4x A5314 nmGalaxy S7, Note 7Custom CPU, 4K/60fps, Mali-T880 MP12
Exynos 9810 (M3)/9820 (M4)2018–2019v8.24x M3/M4 + 4x A5510 nm/8 nmS9/Note9, S10/Note10NPU, 8K video, Mali-G72/G76
Exynos 21002021v91x X1 + 3x A78 + 4x A555 nmGalaxy S21Integrated 5G, 200 MP camera, Mali-G78 MP14
Exynos 22002022v9.11x X2 + 3x A710 + 4x A5104 nmGalaxy S22 (select regions)AMD Xclipse 920 (RDNA2), ray tracing
Exynos 24002024v9.21x X4 + 5x A720 + 4x A5204 nm LPP+Galaxy S24 (international)Xclipse 940 (RDNA3), 14.7x NPU boost
Exynos 25002025v91x X925 + 7x A725 + 2x A5203 nm GAAGalaxy Z Flip 7 (expected)Enhanced NPU, Xclipse 950 (RDNA3)
Exynos W930/W10002023–2024v82x A55 (W930); 1x A78 + 4x A55 (W1000)5 nm/3 nmGalaxy Watch6/7Mali-G68 GPU, GNSS, always-on display

MediaTek Processors

MediaTek has established itself as a key player in the ecosystem by designing affordable system-on-chips (SoCs) optimized for mid-range smartphones, tablets, and emerging personal computing devices, leveraging licensed architectures to deliver balanced performance and power efficiency. These SoCs emphasize cost-effectiveness, integrated connectivity like , and features such as AI processing and advanced imaging, making them prevalent in budget to flagship Android devices across global markets, particularly in developing regions where they hold significant market share. The company's early mobile efforts centered on the MT65xx series during the , which employed ARMv7 instruction set architecture with Cortex-A7 cores for basic multitasking and later Cortex-A17 for improved graphics in entry-level phones. These quad-core SoCs, fabricated on 28nm processes, powered numerous budget Android handsets by prioritizing low power consumption over high-end performance, enabling widespread adoption in feature-rich yet inexpensive devices. Transitioning to , MediaTek introduced the Helio series in 2015 using ARMv8 architecture, targeting mid-range segments with heterogeneous core designs. Flagship-oriented Helio X models, such as the X10 (octa-core Cortex-A53 at 2.0 GHz) and X20 (Cortex-A72/A53 big.LITTLE configuration), supported LTE and enhanced multimedia, while the P and PX series, like the P23 (octa-core A53 at 2.0 GHz), focused on efficient connectivity for everyday tasks in mid-tier phones through 2019. The series incorporated 's CorePilot technology for dynamic resource management, improving battery life in gaming and camera applications. With the shift to 5G, the Dimensity series debuted in 2020 on ARMv8.2, integrating modems and Cortex-A76/A55 cores for cost-effective connectivity; for instance, the Dimensity 700 (octa-core up to 2.2 GHz) and 800 (A76/A55 hybrid) enabled sub-6GHz in mid-range devices, supporting 90Hz displays and 108MP cameras. Advancing to ARMv9, the Dimensity 9000 (2021) and 9200 (2022), built on 4nm processes, featured a Cortex-X2 prime core alongside A78 and A710 cores, delivering flagship-level multitasking and ray-tracing graphics via Mali GPUs. The Dimensity 9300, released in 2023 under ARMv9.1, pioneered an all-big-core CPU layout with four Cortex-A720 cores clocked up to 3.25 GHz, eliminating efficiency cores for superior sustained performance in AI and gaming workloads, complemented by the Imagiq 890 ISP for advanced . Building on this, the 2024 Dimensity 9400 on a 3nm node adopted ARMv9.2 with a single Cortex-X5 core at 3.62 GHz, three A725 performance cores, and four A520 efficiency cores, achieving a 35% CPU uplift and 28% GPU improvement over the 9300 through optimized power delivery. In 2025, the Dimensity 9500 arrived on a 3nm process with architecture, featuring an all-big-core setup including a custom C1-Ultra prime core and integrated Advanced for ultra-low latency in scenarios, with an enhanced NPU for AI enhancements. Beyond mobiles, expanded into personal computing with the MT8195 (also known as Kompanio 1380) in 2022, an ARMv8 octa-core SoC featuring four Cortex-A78 cores on a 6nm process, designed for premium Chromebooks with support for 8K video and decoding. For Windows on Arm, announced an upcoming 8-core ARM-based chip in late 2025, developed in collaboration with , targeting AI PCs with integrated GPU acceleration and expected compatibility with major OEMs like .
SeriesKey ModelsARM ArchitectureCore Configuration ExampleTarget DevicesNotable Features
MT65xxMT6580, MT6595v7Octa-core A53/A17Budget phonesLow-cost , 28nm process
HelioX20, P23v82x A72 + 4x A53Mid-range phonesLTE Cat-6, CorePilot scheduling
Dimensity (Mid)700, 800v8.22x A76 + 6x A55 entry-levelIntegrated , 90Hz display support
Dimensity (Flagship)9300, 9400, 9500v9.xAll-big A720/X5/C1-UltraPremium phones/PCsAI NPU, 3nm efficiency, ray-tracing GPU

Other Third-Party Processors

Other third-party ARM processors encompass a diverse range of implementations developed by vendors beyond the primary mobile-focused designers, often targeting servers, embedded systems, AI acceleration, and niche consumer devices. These cores frequently build on ARM's Cortex-A or Neoverse architectures but include custom modifications for specific workloads, such as or energy-efficient edge processing. Notable examples include processors from , , Amazon, and others, which have expanded ARM's footprint into data centers and specialized applications since the early 2010s. Huawei's Kirin series, introduced in 2012, utilizes custom ARMv8 and v9 cores like the Da Vinci architecture for mobile devices, with the Kirin 9000 (2020) featuring an ARM Cortex-A77-based prime core for enhanced AI and graphics performance in smartphones. For server applications, Huawei's TaiShan processors, such as the TaiShan V120 (2022) based on the Kunpeng 920 (ARMv8), deliver up to 64 cores optimized for and workloads, achieving significant power efficiency in enterprise environments. These designs incorporate Huawei's proprietary extensions for security and acceleration, positioning them as key players in China's domestic ecosystem. NVIDIA's Tegra family, dating back to 2010 with ARMv7 cores, evolved to include custom elements in later iterations; the Tegra X1 (2015) employed quad Cortex-A57 cores for gaming consoles like the , while the Orin SoC (2022) integrates 12 custom Carmel ARMv8.2 cores tailored for AI inference, supporting up to 275 of performance in autonomous vehicles and . The upcoming NVIDIA Blackwell platform (announced 2024, shipping 2025) incorporates ARM-based elements for control and acceleration in AI superchips, enhancing scalability for . Amazon's processors, launched in 2018, leverage foundations for AWS cloud instances; the Graviton3 (2021) uses Neoverse V1 cores to provide up to 25% better price-performance over predecessors in web servers, while the Graviton4 (2024) adopts Neoverse V2 for improved vector processing and up to 30% faster database workloads, emphasizing sustainable computing with lower energy use. Broadcom's BCM2711 (2019), powering the , features a quad Cortex-A72 core at 1.5 GHz for hobbyist and educational computing, enabling GPIO integration and multimedia decoding in compact boards. AMD's MI300X AI accelerator (2023) supports hybrid workloads in data centers with up to 192 GB of HBM3 memory for training. Rockchip's RK3588 (2022) combines eight cores—four Cortex-A76 and four Cortex-A55—for high-end tablets and single-board computers, delivering 6 TOPS of NPU performance for edge AI tasks like video . Allwinner Technology's processors, such as the H616 used in budget tablets, rely on quad Cortex-A53 cores for cost-effective media playback and basic computing in . Ampere Computing's Altra series (2020 onward) employs up to 128 Neoverse cores per socket for cloud servers, offering scalable performance for virtual machines with densities up to 256 cores in dual-socket configurations. Qualcomm's Oryon CPU, initially for client devices, expanded to server applications in 2025 via the Snapdragon X Elite adaptations, targeting efficiency with custom ARMv9 cores that provide up to 45% better power utilization in edge servers compared to x86 alternatives.

Timeline of Releases

1985–2000

The development of ARM processors began in 1985 when introduced the , the world's first commercial RISC processor, designed by and primarily as a second processor for the computer to accelerate and CAD tasks. This 32-bit design emphasized low power consumption and efficiency, setting the foundation for future embedded applications. In 1987, launched the ARM2 processor at 8 MHz, integrated into the series of personal computers, which became the first commercially successful RISC-based home computers and marked ARM's entry into the PC market. The ARM3 followed in 1989, operating at 25 MHz with added cache support, powering upgraded models like the A5000 and enhancing performance for educational and desktop use in the UK market. These early cores shifted focus from general computing toward more efficient designs suitable for battery-powered devices. A pivotal milestone occurred in November 1990 with the formation of (initially Advanced RISC Machines Ltd.) as a joint venture between , Apple Computer, and , transitioning ARM from an in-house Acorn project to a licensable IP model focused on embedded systems. In 1991, the ARM6 core debuted at 20 MHz as the ARM610, specifically tailored for Apple's Newton MessagePad PDA, which launched in 1993 and represented ARM's first major foray into portable . This collaboration helped establish ARM's low-power credentials for mobile applications. Texas Instruments became one of the first major licensees in May 1993, adopting ARM designs for digital signal processing in mobile communications and advising Nokia on GSM phone implementations. Digital Equipment Corporation (DEC) followed as an architectural licensee in the mid-1990s, developing the StrongARM (based on ARM8 architecture) announced in 1996 at speeds up to 200 MHz for high-performance embedded uses. The ARM7TDMI, introduced in 1994, emerged as ARM's first major embedded success with its Thumb instruction set for code density, powering early Nokia GSM phones like the 8110 in 1996. Between 1996 and 1998, the ARM8 (as in the ARM810) and families expanded adoption in set-top boxes for digital TV decoding and PDAs, with the announced in 1997 offering five-stage pipelining for improved multimedia performance. Examples include the ARM7-based phone in 1997, which sold millions and solidified ARM in , and the PDA in 1997 using the ARM710 at 18 MHz for EPOC OS tasks. went public in 1998, listing on the London Stock Exchange and , reflecting growing industry confidence. By 2000, the ARM10 core was introduced in 1998 with initial silicon in 1999, targeting over 400 MIPS for consumer devices and featuring early (SMP) support in variants like the ARM1020 for multi-core experiments in handhelds. This period saw a market shift from PCs to mobile and embedded sectors, driven by licensees like TI and DEC; partner shipments exceeded 180 million units in 1999 alone, reaching hundreds of millions cumulatively by 2000.

2001–2010

The period from 2001 to 2010 marked a pivotal era for processors, driven by the explosive growth of mobile devices and the transition from feature phones to smartphones, which propelled architectures into billions of units shipped annually. In 2001, released the ARMv5TE architecture, enhancing capabilities with support for saturated arithmetic and improved Thumb instruction set interworking, enabling more efficient implementations in embedded systems. This architecture underpinned the widespread adoption of processors in Symbian-based smartphones, such as those from and , where cores like the ARM926EJ-S provided the computational foundation for early mobile operating systems, contributing to Symbian's dominance in the European market with over 100 million devices by mid-decade. By 2004, ARM announced the Cortex-M3 processor, targeting low-cost, high-performance applications with its ARMv7-M , which included Thumb-2 instructions for better code density and interrupt handling, quickly finding use in and industrial controls. The family, introduced in 2002 but gaining traction in portable media players, powered devices like Apple's starting in 2007, where the Samsung-fabricated S5L8900 SoC with an ARM1176JZF-S core delivered 412 MHz performance for tasks. In 2005, ARM unveiled the Cortex-A8, the first out-of-order execution ARM processor based on ARMv7-A, offering up to 2x the performance of at similar power levels through advanced branch prediction and SIMD extensions, setting the stage for high-end . The 2007 launch of the original , featuring a S5L8900 processor with an core fabricated by on a , revolutionized the industry by integrating touch interfaces and app ecosystems, accelerating 's penetration into premium mobile markets and inspiring competitors to adopt similar architectures. Concurrently, the ARM7TDMI core, a staple since the , powered billions of feature phones worldwide, with ARM partners shipping nearly 3 billion processors in 2007 alone, many embedded in basic cellular devices from manufacturers like and . In 2008, the Cortex-R4 processor was introduced for real-time applications, particularly in automotive systems, providing deterministic performance with dual-issue execution and optional floating-point support, enabling advancements in engine control units and safety-critical embedded environments. By 2010, ARM's ecosystem had scaled dramatically, with cumulative shipments exceeding 10 billion cores since inception, fueled by the mobile revolution and the onset of Android's adoption in 2008, which leveraged ARMv7-compatible processors in devices like the . The iPhone 4 introduced Apple's A4 SoC, based on the Cortex-A8 core and fabricated by on a , delivering 1 GHz performance with integrated PowerVR graphics, underscoring the iPhone's outsized impact in driving demand for powerful, power-efficient designs. NVIDIA's platform, featuring ARM cores in early tablets like the 2010 development kits, expanded ARM into multimedia handhelds, while the conceptual groundwork for —later formalized as big.LITTLE in 2011—was laid through explorations of combining high-performance and efficiency cores. This decade's innovations not only solidified ARM's leadership in low-power computing but also transformed , with smartphones alone accounting for over 90% of ARM-based shipments by 2010.

2011–2020

The decade from 2011 to 2020 marked a pivotal shift for processors toward 64-bit architectures, enabling greater performance and efficiency that propelled widespread adoption in mobile devices and initial forays into server infrastructure. In 2011, ARM announced the ARMv8 architecture, introducing 64-bit capabilities while maintaining with 32-bit code to facilitate a smooth transition for developers and manufacturers. Concurrently, Apple's debuted with the A5 processor, featuring a dual-core design that enhanced graphics and multitasking for tablets, underscoring ARM's growing dominance in . By 2013, the 64-bit era arrived in smartphones with Apple's iPhone 5S, powered by the A7 chip—the first 64-bit ARM-based processor in a consumer device—which delivered up to double the CPU and GPU performance of its predecessor while improving energy efficiency. This milestone accelerated the industry's move to , as ARMv8 enabled more complex applications and better handling of large datasets. In 2014, ARM released the Cortex-A53 and Cortex-A57 cores, optimized for the ARMv8-A instruction set, with the A57 providing high-performance capabilities and the A53 focusing on efficiency; these were first implemented in Qualcomm's Snapdragon 810, which employed a big.LITTLE heterogeneous configuration to dynamically balance power and performance in premium smartphones. The mid-2010s saw explosive ecosystem growth, particularly in mobile. In 2016, ARM introduced the Cortex-A73, a more efficient successor to the A57 that improved single-threaded performance by 30% at iso-power, further refining big.LITTLE implementations for sustained battery life in always-on devices. Samsung advanced custom designs with the 8890, incorporating four proprietary M1 cores alongside Cortex-A53 for flagship S7 devices, demonstrating how licensees could tailor ARM IP for competitive edges in AI and processing. That year, ARM-based processors powered over 10 billion shipments cumulatively, reflecting the architecture's near-ubiquity in the global mobile market. Entering the late 2010s, manufacturing advances amplified ARM's scalability. In 2018, the Cortex-A76 debuted as ARM's first core designed specifically for 7nm process nodes, offering 35% higher performance or 40% better compared to the A73, enabling laptop-class in thin smartphones. This period also saw ARM launch the Neoverse N1 platform in 2019 (building on 2018 announcements), a server-oriented core based on the A76 that supported up to 128 cores per socket for workloads, marking ARM's strategic entry into data centers with hyperscalers like AWS adopting it for energy-efficient scaling. ARMv8 architectures achieved dominance, powering the majority of new mobile and embedded designs. The year 2020 highlighted ARM's maturation across sectors amid global challenges. The Cortex-A78 arrived, enhancing integration with improved modem support and AI acceleration, as seen in Qualcomm's Snapdragon 888, which combined it with a dedicated RF system for sub-6GHz and mmWave connectivity in next-generation devices. Apple announced the M1 chip, transitioning Macs to custom ARM-based silicon for the first time, promising superior performance-per-watt for professional workflows. The accelerated demand for ARM PCs by emphasizing remote computing needs, with efficient, low-power designs gaining traction in laptops and edge devices. By 2020, cumulative shipments of ARM-based chips exceeded 160 billion, with integration becoming standard in premium processors to support emerging connectivity ecosystems.

2021–Present

The period from 2021 onward marked a significant evolution in ARM processors, driven by the introduction of the ARMv9 architecture, which enhanced security, capabilities, and scalability for both mobile and server applications. Announced in March 2021, ARMv9 incorporated scalable vector extensions (SVE2) for improved AI and HPC workloads, building on the 64-bit foundation while addressing emerging needs in and branch prediction. In parallel, the Cortex-X1 core debuted in high-end mobile SoCs, offering up to 22% single-threaded performance gains over its predecessor through wider execution units and advanced caching, first appearing in Qualcomm's Snapdragon 888 in late 2020 but gaining prominence in the Snapdragon 8 Gen 1, unveiled in December 2021 with a 4nm process for flagship Android devices. By 2022, Apple's processor advanced the ARM ecosystem in personal computing, released in June for the and , featuring an 8-core CPU with improved performance cores and up to 24 GPU cores on TSMC's second-generation 5nm node, delivering 18% faster CPU performance than the M1. Server-side progress included the Neoverse V2 core, announced in February 2022, which targeted cloud infrastructure with SVE support and up to 50% better integer performance per watt compared to V1, powering platforms like instances. Process node shifts accelerated with MediaTek's Dimensity 9000 in November 2021 (devices in 2022), adopting TSMC's 4nm for efficient and AI processing, signaling the industry's move toward sub-5nm fabrication. In 2023, mobile AI integration surged with dedicated NPUs becoming standard, exemplified by Apple's A17 Pro in the series (September 2023), which included a 16-core Neural Engine capable of 35 trillion operations per second on a 3nm process, enhancing on-device for features like real-time translation. Server advancements continued with Neoverse V3, unveiled in October 2023, offering 50% more than V2 through and extensions, while AWS's Graviton3 processors, based on Neoverse V1 but scaled in 2023 deployments, provided up to 25% better price- for EC2 instances. The AI NPU boom reflected broader adoption, with over 80% of premium smartphones incorporating dedicated accelerators by year-end, boosting inference speeds for generative AI tasks. 2024 saw further maturation in with Apple's M4 chip in May for the , featuring a 10-core CPU on TSMC's 3nm enhanced process and a 16-core Neural Engine at 38 , prioritizing efficiency for slim designs. Qualcomm's Snapdragon 8 , announced in October, shifted to custom Oryon cores with up to 45% faster CPU performance on a 3nm node, targeting AI-driven laptops and phones. Samsung's 2400, released in January for the Galaxy S24, integrated a deca-core design with Xclipse GPU on 4nm, while MediaTek's Dimensity 9400 in October emphasized 3nm efficiency and ray-tracing GPU for gaming. Announcements for 2nm processes proliferated, with and planning EUV-based production in 2025 for next-gen SoCs, promising 15-20% density gains. As of November 2025, the ARM landscape continued to expand with Cortex-X925 and Cortex-A725 cores announced in May 2024, delivering up to 36% IPC uplift via v9.2 extensions for AI and , set for integration in 2025 flagships. Qualcomm's Snapdragon 8 Elite Gen 5, released in September 2025, featured enhanced NPU at 75 on 3nm, powering AI-centric devices. Apple's A18 (iPhone 16, September 2024, but iterated in 2025 devices) and anticipated M5 (expected late 2025 for Macs) emphasized 2nm readiness and 40+ AI, while expanded into Windows on ARM with Dimensity variants for Copilot+ PCs in Q2 2025. ARM reported surpassing 310 billion total shipped cores as of mid-2025, underscoring ecosystem dominance. Key milestones included the revival of Windows on ARM, fueled by Qualcomm's Snapdragon X series and Microsoft's Copilot+ initiative in 2024, achieving over 20 million units shipped by 2025 with native x86 emulation. EUV-enabled 2nm adoption began in production for select designs, enabling denser integrations. ARMv9.3, announced in September 2025, saw rapid uptake in servers for enhanced . AI performance in processors roughly doubled annually, from ~10 in 2021 mobiles to over 50 by 2025, driven by NPU optimizations.

References

  1. https://en.wikichip.org/wiki/qualcomm/kryo
  2. https://en.wikichip.org/wiki/samsung/exynos
  3. https://en.wikichip.org/wiki/mediatek/helio
  4. https://en.wikichip.org/wiki/acorn/microarchitectures/arm3
  5. https://en.wikichip.org/wiki/arm_holdings/microarchitectures/arm6
  6. https://en.wikichip.org/wiki/dec/microarchitectures/strongarm
  7. https://en.wikichip.org/wiki/arm_holdings/microarchitectures/arm8
  8. https://en.wikichip.org/wiki/arm_holdings/microarchitectures/arm9/pr1
  9. https://en.wikichip.org/wiki/arm_holdings/microarchitectures/arm10
Add your contribution
Related Hubs
User Avatar
No comments yet.