14 Sources
14 Sources
[1]
The Top Android Phones of 2026 Could Get Better AI With Arm's New CPUs
Expertise Smartphones | Gaming | Telecom industry | Mobile semiconductors | Mobile gaming On the same day Apple revealed its new iPhone 17 lineup, chip designer Arm is introducing its next generation of processors bundled into its new Arm Lumex platform. These processors will likely make their way into premium Android phones coming in 2026 -- and improve their AI capabilities without draining device batteries faster, Arm says. "It's pretty amazing how there's been kind of this insatiable amount of performance being asked for [by our customers], and a lot of it's around AI, as well as some graphics workloads," said Chris Bergey, senior vice president and general manager of Arm's client line of business. Arm's processors have traditionally been the centerpieces of holistic systems-on-a-chip, which power smartphones. For instance, Arm's previous top-end central processing unit, the Cortex X925, released last year, is featured in Samsung's Exynos 2500 chipset, which powered the Galaxy Z Flip 7 and MediaTek's Dimensity 9400 chipset, found in the Oppo Find X8 Pro. It's likely that Arm's new chips, either on their own or bundled into the company's Arm Lumex platform, will power premium Android phones and other devices next year. But it faces more competition as Qualcomm (which used Arm CPUs in older chips like the Snapdragon 8 Gen 3) has shifted to its internally designed Oryon CPUs in its latest silicon -- which was the subject of its own tech licensing clash over the last few years between the two chip companies. Arm's new CPU range has shifted with a new naming paradigm. The successors to the X925 are two different chips: the highest-end C1-Ultra, which boasts 25% greater performance over its predecessor, and the next-most-powerful C1-Premium (no performance improvement given). The successor to the A725 chip is the C1-Pro, which has 12% greater efficiency. These CPUs also benefit from an evolution of their chip architecture called Scalable Matrix Extension version 2, which enables better AI performance. The new Arm Lumex platform combines these chips with the new Mali G1-Ultra GPU (which Arm says has 20% better performance and twice the ray tracing as its predecessor) for a system that can be plugged into larger chipsets. The end result: up to 5x improvement in AI performance, 4.7x lower latency for speech-based workloads (think live translation) and 2.8x faster audio generation, Arm says. The goal is faster performance without increasing the battery drain when running AI tasks that tax processing power. For example, a yoga tutor demo app running Arm's new chips saw a 2.4x boost in text-to-speech to give faster feedback to users, the company said in a press release. As a supplier of chips and technology, it's ultimately up to the phonemakers using Arm's silicon to decide how much (or little) of the advancements they integrate into their devices. But as demand for generative AI like ChatGPT increases, so too does the drive to get that functionality working on phones as efficiently as possible, rather than relying on slower responses going to and from the cloud. So what will consumers see with devices running C1-Ultra chips and Arm's new technology? "I think what they're gonna see is the ability to run amazing on-device AI and to do so with significant power savings, significant performance increases, and also third-party support," Bergey said. "Not just first-party devices, but also third-party devices."
[2]
Arm bets on CPU-based AI with Lumex chips for smartphones
Arm has lifted the lid on its latest mobile platform, comprising new CPU and GPU designs plus rearchitected interconnect and memory management logic, all optimized with a coming wave of AI-enabled smartphones in mind. The UK-based chip designer has been moving towards more integrated solutions rather than just offering cores for several years, and this year's Lumex compute subsystem (CSS) is the latest evolution of that philosophy. As with each generation, Arm manages to squeeze more performance and power efficiency out of its designs, claiming an average 15 percent step up in CPU and 20 percent in GPU, while saving 15 percent on power. The key focus with Lumex is on Arm's SME2 Scalable Matrix Extensions in the CPU cluster, which the firm is pushing as the preferred route for AI acceleration, and overall system-level optimizations to boost scalability for devices capable of running AI models. Stefan Rosinger, Senior Director for CPUs, said that SME2 gives AI acceleration "an order of magnitude better" than before, and its advantage for a mobile device is that it uses less power and finishes calculations quicker. According to Arm, we can expect to see Lumex implemented in smartphones and other devices later this year or early next year. It has been crafted with 3nm manufacturing in mind, and the firm says it expects chips produced by its licensees will run at upwards of 4 GHz clock speeds. In a rare outbreak of common sense, all the cores in the Lumex CPU cluster are designated C1, with the maximum performance core design labelled the C1-Ultra. Most smartphone chips now feature a mix of core types, with one or two performance cores for the demanding work, coupled with more power-optimized cores to handle other tasks. This started out as big.LITTLE many years ago. With Lumex, Arm has given chip designers no fewer than four core types to choose from, adding C1-Premium as the next step down, followed by C1-Pro and finally C1-Nano as the smallest footprint core with the greatest power efficiency. Chips aimed at flagship phones will likely see two C1-Ultra cores combined with six C1-Pro cores, Arm believes, while "sub-flagship" silicon might use a mix of two C1-Premium and six C1-Pro, while the mainstream could be served by a mix of four Pro and four Nano. In the GPU department, the Lumex platform has the Mali G1, and like the CPU cores, these are graded from Mali G1-Ultra to Mali1 G1-Premium and Mali G1-Pro. The differences between these three tiers are in the number of shader cores, with Pro having one to five, Premium six to nine, and Ultra 10 cores or more. Mali G1-Ultra is also the only tier with Arm's redesigned Ray Tracing Unit (RTU), claimed to offer 40 percent higher performance than last year's Immortalis-G925, plus higher quality for games. The new GPU design is also claimed to accelerate in-game AI with support for half-precision (FP16) matrix multiplication, which Arm claims improves tensor processing while reducing memory bandwidth and lowering power consumption. Arm has already given us a taste of the neural accelerator hardware it is bringing to its phone GPUs next year, but that is not part of the Mali G1-Ultra in Lumex this year. What is included in this complete compute subsystem are a new purpose built System Interconnect (SI) and System Memory Management Unit (SMMU), designed to cope with the demands that running AI models on smartphones are likely to bring. The System Interconnect has been redesigned with what Arm calls a channelized architecture that can give quality of service (QoS) priority to different traffic, while the System MMU has been optimized to reduce latency by up to 75 percent, the firm says. It isn't just about hardware of course, and Arm says it has been working behind the scenes to ensure that the various developer frameworks support the optimizations coming in its latest platforms. Its KleidiAI libraries have been integrated with frameworks such as PyTorch, Llama, LiteRT and ONNX to enable support for SME2 acceleration when running AI workloads. Of course, Arm believes that AI processing should be kept on the CPU, since "it's the only compute unit in the mobile market you can rely on to be present in every mobile phone," according to Geraint North, Arm Fellow in AI and Developer Platforms. "As you start to move to GPUs and NPUs (neural processing units), you end up doing different work for different handsets," he explained, as vendors may have opted for different GPUs and NPUs in their smartphone silicon. This is a logical viewpoint, but it's not one that everyone necessarily agrees with. Analyst Gartner explicitly defines a GenAI smartphone as a device equipped with a built-in neural engine or neural processing unit (NPU) capable of running small language models. It proclaims that premium smartphones and basic smartphones (under $350) will fit this description, with only "utility smartphones" not expected to have NPU capabilities. This will suit Qualcomm, no doubt, which features an integrated NPU in its smartphone chips and showed off a 7 billion parameter large language model running on an Android phone at last year's MWC show. Bob O'Donnell, President and Chief Analyst at TECHnalysis Research, said that Arm's approach makes sense given where the market is right now. "First, because of the wide range of different NPU architectures available and the lack of standardization, few software developers are actually using the NPUs for their AI applications. Instead, they're defaulting to the CPU and GPU, and with the latest SME2 instructions and logic from Arm, they're going to help accelerate those functions," he told The Register. "Second, many of Arm's partners are choosing to differentiate with their own NPU designs and yet another option from Arm could actually make that NPU confusion worse. Hopefully we'll start to see some standardized means for leveraging different NPU architectures soon so that NPUs can start being used more frequently, but I'm concerned that could still be several years away." Arm's silicon licensees will have to place their bets now - will Arm's approach with SME2 in the CPU prove more popular, or will smartphone makers and the buying public want a built-in NPU? ®
[3]
Arm launches new generation of mobile chip designs geared for AI
SAN FRANCISCO, Sept 9 (Reuters) - Arm Holdings said on Tuesday it was launching its next-generation set of chip designs called Lumex that it has optimized for artificial intelligence to run on mobile devices such as smartphones and watches without accessing the internet. Called Lumex, the new generation of Arm mobile designs come in four types, ranging from less powerful but more energy efficient ones designed for watches and other smart wearable devices to a version designed to maximize the computing horsepower available. The peak performance design aims to run software that harnesses the power of large AI models without accessing cloud computing on devices such as high-end smartphones. "AI is becoming pretty fundamental to kind of what's happening, whether it's kind of real-time interactions or some killer use cases like AI translation," said Chris Bergey, a senior vice president and general manager at Arm. "We're just seeing (AI) become kind of this expectation." The Lumex designs announced on Tuesday are part of the company's Compute Subsystems (CSS) business, which aims to offer handset makers and chip designers a more ready-made piece of technology that enables their speedy use in new products. The more complete designs Arm is now making for data centers and mobile phones are part of the company's longer-term plans to grow its smartphone and other revenue through a variety of means. Arm has also said it plans to invest more money into examining making its own chips and has hired several key people in order to do so. The Lumex designs are optimized for so-called 3-nanometer manufacturing nodes, such as the process offered by TSMC (2330.TW), opens new tab. The series of iPhone chips Apple (AAPL.O), opens new tab announced on Tuesday use a TSMC 3-nanometer process. The company is holding a launch event in China on Wednesday to unveil the new designs because outside of Apple and Samsung (005930.KS), opens new tab the leading handset makers are located there, Bergey said. Reporting by Max A. Cherney in San Francisco; Editing by Jamie Freed Our Standards: The Thomson Reuters Trust Principles., opens new tab * Suggested Topics: * Disrupted * Medtech Max A. Cherney Thomson Reuters Max A. Cherney is a correspondent for Reuters based in San Francisco, where he reports on the semiconductor industry and artificial intelligence. He joined Reuters in 2023 and has previously worked for Barron's magazine and its sister publication, MarketWatch. Cherney graduated from Trent University with a degree in history.
[4]
Arm's Lumex platform promises faster AI and better battery life for next-gen devices
Serving tech enthusiasts for over 25 years. TechSpot means tech analysis and advice you can trust. What just happened? Arm says its new Lumex platform marks the sixth consecutive year of double-digit improvements in instructions per cycle, reinforcing its strategy to keep CPUs central to AI innovation. The company is presenting this leap not as a routine hardware update, but as a shift in how consumer devices will handle advanced computing. Unveiled this week, the Lumex Compute Subsystem (CSS) is designed to run AI directly on the device, rather than offload tasks to the cloud. Arm says Lumex delivers major gains in CPU performance, AI acceleration, and graphics capabilities while maintaining power efficiency across a range of devices, from flagship smartphones to ultra-low-power wearables. At the core of the Lumex architecture is support for AI-heavy workloads such as real-time speech recognition, translation, and generative applications. This is enabled by Arm's integration of Scalable Matrix Extension v2 (SME2), an instruction set developed to accelerate the mathematical operations that underpin modern AI models. Instead of relying on separate neural processors, SME2 enables the CPU to perform AI acceleration directly. Arm claims this shift brings up to five times faster AI performance compared with the previous generation, cuts latency for speech tasks by nearly a factor of five, and triples the speed of audio generation. The Lumex CPU family introduces four different core types, each tailored to different performance and efficiency requirements. At the high end, the C1-Ultra offers a 25 percent boost in single-thread performance over Arm's earlier Neoverse design. It is intended for compute-intensive workloads such as large language model inference and computational photography. The C1-Premium features a smaller footprint - approximately 35 percent smaller than the Ultra - while retaining much of its muscle, making it well-suited for mainstream smartphones where cost and space constraints are key. The C1-Pro strikes a balance between efficiency and power, delivering a 16 percent performance boost for sustained workloads, such as video streaming or background assistant apps. At the most efficient tier, the C1-Nano reduces energy use by up to 26 percent, making it a viable option for wearables such as smartwatches and rings. Arm states that a cluster combining two C1-Ultras with six C1-Pros demonstrates a full fivefold increase in AI performance, showcasing the platform's scalability across devices. Software support is central to this effort. Arm has built its KleidiAI libraries directly into leading frameworks such as PyTorch ExecuTorch, Google's LiteRT, Alibaba's MNN, and Microsoft's ONNX Runtime. As such, developers can take advantage of SME2's acceleration without needing to rewrite their code. Arm points to early examples of adoption, including Google Photos, Gmail, and YouTube, which already show improved performance with the framework integration, as well as Alipay, which has demonstrated running large language models entirely on-device using Lumex prototypes. Gaming also received significant focus with the launch of the Mali G1-Ultra GPU. This new graphics processor introduces a second-generation Ray Tracing Unit, which doubles ray tracing performance compared to the prior Immortalis G925 GPU. The 14-core G1-Ultra is said to deliver a 20 percent graphics performance boost while reducing power draw by 9 percent per frame. It also accelerates GPU-based AI inference by 20 percent, highlighting Arm's effort to spread AI workloads across the entire system-on-chip. Games such as Arena Breakout, Fortnite, Genshin Impact, and Honkai Star Rail are expected to see noticeable improvements. For midrange and efficiency-focused devices, scaled-back models - the G1-Premium and G1-Pro - bring similar architectural gains while extending battery life. Several major technology players have already pledged support for Lumex. Samsung, MediaTek, Alibaba, and vivo have been among the first to highlight improvements in responsiveness for applications like voice interactions, summarization, and translation. Google has emphasized SME2's role in making AI more consistent across Android devices and Windows on Arm laptops, while Meta has pointed to performance gains in its applications built on PyTorch ExecuTorch. Nevertheless, entering the US market can be challenging. Qualcomm remains the leading designer of processors for Android phones sold domestically and is currently locked in litigation with Arm over licensing of its technology. That fight may determine how soon Lumex-based devices become widely available outside Asia, leaving open how quickly Arm's latest CPUs will shape the broader market.
[5]
Arm's bid for smarter, AI-powered phones and PCs begins with Lumex
It's likely that Lumex gives us a hint at what Arm's Niva PC line will eventually offer. With any luck, Arm's new Lumex CPU platform may give us a hint at what to expect for upcoming Windows on Arm PCs: four tiers of CPU power, plus an improved ray-tracing engine and graphics upscaling. Arm says that its new Lumex C1-series chips will deliver 25 percent more performance than the Cortex X925 series of processors it launched in May 2024. Like the X925, the latest Lumex C1 cores are being optimized with 3nm process technologies in mind, with physical implementations and foundry collaborations to speed customers to market. (While Arm has hinted at building its own physical cores for years, however, company representatives would say only that it includes "near production-ready physical implementations for partners" and is "not a chip.") The Lumex platform is Arm's brand for smartphones. PC-specific Arm chips will be branded as "Niva" under Arm's new naming scheme, though they will share some common features with the Lumex cores. Qualcomm, which actually makes the Arm processors, uses the "Snapdragon" brand, which will almost certainly continue. Arm chips power the vast majority of the world's smartphones, as Arm licenses its designs to customers who can choose to bring the Arm cores to market or take an architectural license and design compatible but otherwise brand new designs of their own making. That's the approach Apple and Qualcomm have taken -- who, in addition to designing smartphones, have brought Arm into the Mac OS space as well as Windows on Arm PCs. Though Qualcomm and Arm have had their legal differences over licensing Arm's cores -- which have since been settled -- an Arm representative declined to comment when asked if any ongoing legal issues would prevent Qualcomm from taking a license. And you will, apparently, see the Lumex in more than just PCs. "So the Lumex platform is going to power flagship smartphones through to PCs and tablets," said James McNiven, the vice president of product management for Arm, in a press briefing. Smartphones, though, have their own demands: low power, which Arm's RISC architecture was designed for; and maximizing local AI. The Cortex C1 series includes what Arm calls Scalable Matrix Extension 2 (SME2), using the Arm v9.3 instruction set. Arm doesn't use a dedicated NPU. Instead, it uses a technology called KleidiAI that essentially uses software libraries to address AI-specific functions inside the CPU, no matter which version of the Arm architecture is present. In the C1 CPU cluster, Arm says, you'll see a 5X uplift in AI performance. Arm says that will increase performance on apps that always have some form of AI technology running, such as audio generation, camera inferencing, or computer vision. Specifically, Arm is claiming over 4.7 times improvement in latency, speech recognition, and classical large language model tasks, and about 2.8 times faster audio generation. The other thing Arm has been known for is that it basically pioneered the concept of performance and efficiency cores, a strategy known as "Big-Little." But with the Lumex, that's been taken to a new level with the addition of a new "premium" core. Now, there are four different tiers of CPU cores: the C1-Ultra, the C1-Premium, the C1-Pro, and the C1-Nano. Nothing's really changed all that much: The Ultra and Premium cores are simply two tiers of "performance" cores, while the Pro and Nano deliver different levels of efficiency. Arm executives said that the Premium core could stand in for the Ultra cores on non-flagship, cheaper devices, as it offers similar performance to the Ultra, but in a 35 percent smaller area. The C1 Pro "improves" upon the Cortex A75 in terms of performance and efficiency, McNiven said, while the Nano has the "smallest footprint" and will play a role in both flagship and entry-level devices, he said. Overall, Arm executives said that the on-device AI is three times more power efficient than the previous implementations, and the Pro is 12 percent more power efficient at the same frequency. Arm is also introducing a new GPU, the Mali G1-Ultra, which will promise 20 percent better graphics performance, twice the ray-tracing performance, and 20 percent faster inferencing for AI processing than the earlier Immortalis-G925. The Mali G1-Ultra will also consume less power, as the block is on its own power island with less leakage when idle. Specifically, Arm is claiming that frame rates on ray-traced games will be 40 percent higher than its predecessor, part of moving to a "single-ray" model for improved efficiency and more realistic lighting," McNiven said. The new Mali core also includes upscaling -- quickly rendering at a lower resolution for improved frame rate, then upscaling it for better visual quality -- but it does not use the AI-generated frames of some desktop GPUs. "One of the examples that we have been seeing recently was some of the new ray tracing benchmarks like [UL's 3DMark] Solar Bay Extreme, and I think that we see there up to a doubling in performance, because it is so ray tracing heavy. So it really does depend on just the amount of ray-tracing content," McNiven said. One idea behind the Lumex platform, executives said, was to move certain cloud-based AI functions on to the device. Specifically, a large language model in the world of Krafton's Inzoi (a spiritual successor to The Sims) was run on-device at the GDC conference, they said, as well as a "coach" that watched you play in Tencent's Honor of Kings and offered advice. A major online payment provider is also working to put agentic AI on device to handle payment processing during peak times, instead of committing to expensive, back-end cloud servers, said Chris Bergey, the senior vice president of the client line of business at Arm. "If your device is capable of running a large language model, you have an extra means of interacting with the game that augments your experience," McNiven said. "But if you don't have it, the game is still playable."
[6]
Arm unveils new Lumex AI focused smartphone CPUs with some impressive stats
Arm has lifted the wraps off its next-generation Lumex chip designs, optimized to run some local AI workloads on mobile devices. Its architecture allows for four different design types, ranging from energy-efficient cores for wearables to high-performance cores for flagship smartphones. Slating accelerated product cycles, which result in tighter timescales and reduced margin for error, Arm says its integrated platforms combine CPU, GPU and software stacks to speed up time-to-market. Arm described Lumex as its "new purpose-built compute subsystem (CSS) platform to meet the growing demands of on-device AI experiences." The Armv9.3 C1 CPU cluster includes built-in SME2 units for accelerated AI, promising 5x better AI performance and 3x more efficiency compared with the previous generation. Standard benchmarks see performance rise by 30%, with a 15% speed-up in apps and 12% lower power use in daily workloads compared with the prior generation. The four CPUs on offer are C1-Ultra for large-model inferencing, C1-Premium for multitasking, C1-Pro for video playback and C1-Nano for wearables. The Mali G1-Ultra GPU also enables 20% faster AI/ML inferencing than Immortalis-G295, as well as improvements across gaming like 2x better ray tracing performance. Lumex also offers G1-Premium and G1-Pro options - but no G1-Nano. Interestingly, Arm positions CPUs as the universal AI engine given the lack of standardization in NPUs, even though NPUs are starting to earn their place in PC chips. Launching with Lumex is a complete Android 16-ready software stack, SME2-enabled KleidiAI libraries and telemetry to analyze performance and identify bottlenecks, allowing developers to tailor Lumex to each model. "Mobile computing is entering a new era that is defined by how intelligence is built, scaled, and delivered," Senior Director Kinjal Dave explained. Looking ahead, Arm notes that many popular Google apps are already SME2-enabled, meaning that they're prepared to benefit from improved on-device AI features when next-generation hardware becomes available.
[7]
Arm raises stakes in mobile AI with Lumex CSS; debuts C1 CPU and Mali G1-Ultra GPU
Arm SVP & GM of Client Line of Business, Chris Bergey. Credit: Arm Unlocked Shanghai Arm unveiled its Lumex Compute Subsystem (CSS) on September 10 at the Unlocked Summit in Shanghai, signaling its boldest step yet to elevate AI and gaming on flagship smartphones. The platform debuts the C1 CPU cluster based on Armv9.3 and the Mali G1 GPU lineup, headlined by the Mali G1-Ultra with support for second-generation Scalable Matrix Extensions (SME2). Together, they bring edge devices closer to data center-class performance. Chris Bergey, Arm's SVP and GM of the Client Line of Business, said AI has become the "foundation of next-generation mobile and consumer technology." He noted that SME2 will extend across all Arm CPU platforms, projecting that by 2030, SME and SME2 will add over 10 billion TOPS of compute power across more than 3 billion devices, as Sina and ICsmart noted. C1 CPU cluster: Armv9.3 with SME2 for flagship AI performance Arm rolled out four C1 CPU variants, giving OEMs flexibility to balance performance, power, and area: ● C1-Ultra: The flagship super-core with the industry's widest microarchitecture, boosting IPC by 12%, delivering 25% higher single-thread performance and 28% lower power than Cortex-X925. Geekbench 6.3 scores show a 26% uplift over Cortex-X925 at lower energy cost. ● C1-Premium: Arm's first "sub-flagship" shrinks core area by 35% versus C1-Ultra with L2 cache, enabling cost-efficient 8-core designs that maintain comparable performance. ● C1-Pro: A big core optimized for energy efficiency, offering 11% more performance or 26% lower power draw than Cortex-A725, with real-world workloads showing 16% performance gains and 12% lower power use. ● C1-Nano: Built for wearables and compact devices, improving efficiency by 26% and SPECint2017 scores by 5.5%, with only a 2% area increase versus Cortex-A520. At the cluster level, the new C1-DSU (DynamIQ Shared Unit) manages heterogeneous cores with SME2 support, cutting typical power by 11% and RAM wake-up energy by 7% compared with DSU-120. A top-end 2× C1-Ultra + 6× C1-Pro cluster delivers 17 times the performance of a basic 2× C1-Nano setup, though with 25 times the die area. Across workloads, the C1 cluster averages 30% better performance, 15% faster gaming and streaming, and 12% lower power use in daily mobile tasks. SME2 turns the CPU into an AI accelerator. By enabling low-precision data formats and efficient matrix operations, it narrows the gap with GPUs for smaller AI tasks. Arm says SME2 delivers up to 5× higher AI performance and 3× better efficiency than the prior generation. Benchmarks show sharp gains: ● Whisper Base speech recognition latency dropped 4.7×; ● Neural camera denoising hit 120 fps at 1080p or 30 fps at 4K on a single core. SME2 is already being adopted by Alibaba, Ant Group, Samsung System LSI, Tencent, and Vivo, underscoring its role in next-generation AI optimization. Mali G1-Ultra GPU: doubling down on ray tracing and AI The Mali G1-Ultra GPU headlines Arm's fifth-gen GPU architecture, engineered for mobile gaming and AI. With mobile gamers now making up 83% of the global gaming population and logging 390 billion hours annually, the G1-Ultra pushes performance to new heights. Highlights include: ● RTUv2 ray tracing with 2× performance and 40% higher frame rates; ● 20% graphics uplift; ● 20% faster AI inference with MMUL.FP16, 104% better FP32 ML throughput; ● 9% lower per-frame power use for longer gaming sessions. Arm also launched Mali G1-Premium and G1-Pro GPUs for mid-tier markets, scalable from 1 to 24 shader cores. Early testing shows games like Genshin Impact, Honkai: Star Rail, and Fortnite all running with significant improvements on the G1-Ultra. For developers, the GPU adds tile-based hardware counters, Vulkan support, and future RenderDoc integration for deeper profiling. Arm Accuracy Super Resolution (ASR), already in Fortnite Mobile via Unreal Engine 5, enhances visuals while keeping GPU load low. Lumex CSS platform: system IP, 3nm readiness, and developer tools Lumex CSS is more than CPUs and GPUs. It integrates system IP and is validated for 3nm process nodes across leading foundries, giving chipmakers production-ready building blocks. ● SI L1 interconnect: Reduces leakage power by 71%, cuts latency 75% versus CI-7000, and integrates system-level cache with Memory Tagging Extension (MTE) support. ● MMU L1 memory management unit: Lowers TBU latency by up to 83%, enabling secure, scalable virtualization for Android and Windows devices. ● NoC S3 interconnect: Tailored for cost-sensitive, non-coherent mobile SoCs. On the software side, Arm's KleidiAI library -already integrated into PyTorch, Llama.cpp, MediaPipe, and more- delivers SME2-optimized AI performance automatically. With Android 16-ready stacks and observability tools like Vulkan counters, Streamline, and Perfetto, developers can tune workloads before hardware launches. Industry adoption and market impact No early design wins were announced, though Vivo signaled interest on stage. Analysts suggest MediaTek's Dimensity 9500 may adopt the C1 CPU cluster and Mali G1-Ultra GPU, though not the full Lumex CSS. Executives clarified that Lumex CSS is a reference subsystem, not a turnkey SoC. OEMs must still integrate NPUs, basebands, and interface IPs to build commercial chips. Still, as AI shifts to the edge, Lumex CSS offers pre-validated, 3nm-ready IP that helps chipmakers shorten design cycles and lower risk. By combining C1 CPUs with SME2 and Mali G1 GPUs with RTUv2, Lumex CSS delivers desktop-grade gaming, generative AI, and real-time multimedia on mobile. With 3nm readiness, strong ecosystem support, and a full developer toolchain, Arm positions Lumex CSS as a foundation for next-generation flagship devices in the AI-first era.
[8]
Arm debuts AI-optimized Lumex chip lineup for mobile devices - SiliconANGLE
Arm debuts AI-optimized Lumex chip lineup for mobile devices Arm Holdings plc today introduced a new portfolio of chip designs that handset makers can use to develop mobile processors. The product family is known as Lumex. It includes central processing units, graphics processing units and an interconnect module. Smartphone makers can use the latter component to link together Arm's new GPUs and CPUs into a single system-on-chip. Lumex includes four CPU core designs equipped with the company's SME2 technology. It's an instruction set extension, a collection of low-level computing operations that a chip can mix and match to perform complex calculations. The computing operations in SME2 are optimized to run artificial intelligence models. Mobile developers often compress the AI models they use in apps to improve performance and reduce battery usage. According to Arm, SME2 includes a 512-bit register optimized to power compressed neural networks. A register is a speedy memory pool built directly into a CPU. SME2 also includes new computing operations for moving and changing data. Additionally, a so-called predicate-as-counter mechanism reduces the amount of data that must be processed to complete computing operations. That reduction helps boost throughput in some cases. The most capable of the four CPU core designs in the Lumex chip design family is known as the C1-Ultra. According to Arm, the processor provides 25% better single-thread performance than previous-generation silicon. That makes it suitable for hardware-intensive AI use cases such as content generation. The C-1 Ultra is rolling out alongside the C-1 Premium, which is designed to power simpler AI applications such as voice assistants. The C-1 Premium is 26% smaller than the C-1 Ultra. There are also two other C-1 cores that trade off some performance for higher efficiency. The C1-Pro is geared toward apps that prioritize sustained performance, while the C1-Nano has a compact design that allows it to be used in wearables. Handset makers can mix and match Arm's new CPU cores. A company could, for example, build a smartphone processor that includes one C1-Ultra and several C1-Pro cores. Demanding apps can run on the C-1 Ultra while other workloads may be sent to the other cores to save energy. "For generative AI, speech recognition, classic machine learning (ML) and computer vision (CV) workloads, the SME2-enabled Arm C1 CPU cluster achieves up to 5x AI speed-up compared to the previous generation CPU cluster under the same conditions," Stefan Rosinger, Arm's senior director of CPU product management, wrote in a blog post. "Through SME2, the CPU cluster delivers up to 3x improved efficiency." The C1 cores are designed to work with a new GPU series called the Mali G1 that Arm debuted in conjunction. The most advanced GPU in the lineup, the G1-Ultra, can perform inference 20% faster than its predecessor. A new ray tracing module allows it to render lighting, shadows and reflections twice as fast. The G1-Ultra has 10 shaders, which have a similar role as the cores in a CPU. It's rolling out alongside three other G1 GPUs that include up to nine shaders. Additionally, Arm has a new interconnect module called the S1 L1 that can link together C1 and Mali G1 chips with 75% less static latency than earlier hardware.
[9]
Lumex, Arm's new mobile chip designs purpose-built for AI
Chip-designer Arm has announced its next-generation set of chip designs called Lumex, purpose-built for AI on mobile devices like smart phones and watches. UK semiconductor company Arm designs the core architecture of semiconductor processors and instruction set for processors, which are licensed by chip giants the likes of Apple and Nvidia. Yesterday it announced Arm Lumex its "most advanced compute subsystem (CSS) platform purpose-built to accelerate AI experiences on flagship smartphones and next-gen PCs". "AI is no longer a feature, it's the foundation of next-generation mobile and consumer technology," said Arm SVP Chris Bergey who leads the chip designer's client line of business. "Users now expect real-time assistance, seamless communication, or personalised content that is instant, private, and available on device, without compromise. Meeting these expectations requires more than incremental upgrades, it demands a step change that brings performance, privacy and efficiency together in a scalable way." "Partners can choose exactly how they build Lumex into their SoC (system on a chip) - they can take the platform as delivered and leverage cutting-edge physical implementations tailored to their needs, reaping time to market and time to performance benefits," said Bergey. "Alternatively, partners can configure the platform RTL for their targeted tiers and harden the cores themselves." The new Lumex name was mooted back in May when it said it was introducing a simplified product naming architecture "to better communicate these platforms to the outside world", with each compute platform having a clear identity for the respective end market: Arm Neoverse for infrastructure Arm Niva for PC Arm Lumex for mobile Arm Zena for automotive Arm Orbis for IoT The Mali brand to continue for GPUs We will be Chip building The developments come at a time when the majority SoftBank-owned Arm is believed to be planning to build its own chips, with Amazon's artificial intelligence (AI) chip director Rami Sinno the latest recruit to boost this ambition. Sinno, in his role as director of engineering at Amazon-owned Annapurna Labs, played a large part in the development of Amazon's AI chips Trainium and Inferentia. Prior to joining Amazon's homegrown chip operation, Sinno had held leadership positions in Arm's engineering team for more than five years until 2019. The company has had its sights set on its own in-house manufacturing for a while now, with reports indicating as much as far back as January. In February, sources told Reuters that the company had begun hiring from its own customers and competing with them for deals as part of its plan to sell its own chips. Don't miss out on the knowledge you need to succeed. Sign up for the Daily Brief, Silicon Republic's digest of need-to-know sci-tech news.
[10]
Arm's New Lumex Platform is Geared Towards AI Functionality - Phandroid
Arm recently announced the arrival of its Lumex CSS platform, which it says is designed to support next-generation AI features on consumer devices. That in mind, the new platform combines powerful CPUs, GPUs, and system IP with an optimized software stack. READ: Qualcomm Teams up with BMW for a AI-Enabled Automated Driving System One of Lumex's core components is the Scalable Matrix Extension version 2 (SME2), which Arm says enhances AI performance by enabling new SME2-enabled CPUs to achieve up to a 5x increase in AI performance. On paper, this essentially translates to faster real-time tasks like voice translation and audio generation, for example. Chris Bergey, SVP and GM of Arm's Client Line of Business states: AI is no longer a feature, it's the foundation of next-generation mobile and consumer technology. Users now expect real-time assistance, seamless communication, or personalized content that is instant, private, and available on device, without compromise. Meeting these expectations requires more than incremental upgrades, it demands a step change that brings performance, privacy and efficiency together in a scalable way. By moving AI processing from the cloud to the device, Arm says that Lumex is able to provide users a faster, more private, and more reliable experience. It also supports graphics enhancement and gaming with the new Mali G1-Ultra GPU, with as much as a 2x uplift in ray tracing performance and up to a 20% increase in AI inference performance. For developers, the KleidiAI software library is integrated into major mobile operating systems and AI frameworks, allowing them to easily access SME2 acceleration. Arm says that this technology will add over 10 billion TOPS of compute across more than 3 billion devices over the next half-decade.
[11]
Arm launches new generation of mobile chip designs geared for AI - The Economic Times
Arm Holdings said on Tuesday it was launching its next-generation set of chip designs called Lumex that it has optimized for artificial intelligence to run on mobile devices such as smartphones and watches without accessing the internet. Called Lumex, the new generation of Arm mobile designs come in four types, ranging from less powerful but more energy efficient ones designed for watches and other smart wearable devices to a version designed to maximize the computing horsepower available. The peak performance design aims to run software that harnesses the power of large AI models without accessing cloud computing on devices such as high-end smartphones. "AI is becoming pretty fundamental to kind of what's happening, whether it's kind of real-time interactions or some killer use cases like AI translation," said Chris Bergey, a senior vice president and general manager at Arm. "We're just seeing (AI) become kind of this expectation." The Lumex designs announced on Tuesday are part of the company's Compute Subsystems (CSS) business, which aims to offer handset makers and chip designers a more ready-made piece of technology that enables their speedy use in new products. The more complete designs Arm is now making for data centers and mobile phones are part of the company's longer-term plans to grow its smartphone and other revenue through a variety of means. Arm has also said it plans to invest more money into examining making its own chips and has hired several key people in order to do so. The Lumex designs are optimized for so-called 3-nanometer manufacturing nodes, such as the process offered by TSMC. The series of iPhone chips Apple announced on Tuesday use a TSMC 3-nanometer process. The company is holding a launch event in China on Wednesday to unveil the new designs because outside of Apple and Samsung the leading handset makers are located there, Bergey said.
[12]
New Arm CPUs and GPUs Up to 45% Faster Double the Ray-Tracing Speed
What if your next smartphone could process AI tasks five times faster, render hyper-realistic graphics in real-time, and still last longer on a single charge? That's not a distant dream, it's the reality Arm's latest CPUs and GPUs are bringing to the table. With up to 45% faster multi-threaded performance and double the ray-tracing speed, these innovative processors promise to redefine what's possible in mobile and embedded computing. From powering AI-driven workloads to delivering cinematic visuals for gaming, Arm's innovations are setting a new standard for speed, efficiency, and versatility. This isn't just an incremental upgrade; it's a leap forward that could reshape the devices we rely on every day. In this breakdown, Gary explains how Arm's new CPU lineup, led by the Lumix C1 series, and the Mali G1 Ultra GPU are breaking barriers in performance and energy efficiency. You'll discover how features like Scalable Matrix Extensions 2 (SME2) are transforming AI processing directly on the CPU, and how advanced ray-tracing capabilities are bringing immersive graphics to life. Whether you're a developer, a tech enthusiast, or just curious about the future of computing, these advancements offer a glimpse into the next generation of devices. The question isn't just how fast or efficient technology can get, it's how these breakthroughs will transform the way we interact with the world. The Lumix C1 series represents Arm's most advanced CPU family to date, featuring four distinct cores: C1 Ultra, C1 Premium, C1 Pro, and C1 Nano. These cores succeed the Cortex X925, A725, and A520, using the innovative ARM v9.3 architecture to deliver unparalleled performance and efficiency. A key innovation in this series is the integration of Scalable Matrix Extensions 2 (SME2), which enables efficient AI processing directly on the CPU. This reduces reliance on GPUs or NPUs, optimizing silicon usage while conserving energy. The Lumix C1 series achieves a 45% boost in multi-threaded performance, thanks to enhanced Instructions Per Cycle (IPC) and clock speeds reaching up to 4.1 GHz. These improvements make the CPUs ideal for demanding applications, such as AI-driven tasks, high-performance mobile devices, and advanced computing systems. By combining raw power with energy efficiency, the Lumix C1 series sets a new standard for modern processors. SME2 is a defining feature of the ARM v9.3 architecture, designed to transform how AI workloads are processed. By sharing an SME2 unit across CPU clusters, Arm has achieved up to a 5x performance increase for specific AI tasks, all while minimizing energy consumption. This innovation allows devices to handle complex AI computations without requiring additional hardware, offering a cost-effective and energy-efficient solution for manufacturers. The ability to process AI workloads directly on the CPU not only enhances performance but also simplifies system design. This makes SME2 particularly valuable for applications in machine learning, natural language processing, and real-time data analysis, where efficiency and speed are critical. By reducing the need for external accelerators, SME2 paves the way for more compact and versatile devices. Dive deeper into ARM with other articles and guides we have written below. The Mali G1 Ultra GPU introduces significant advancements in graphics performance and energy efficiency, making it a cornerstone of Arm's latest offerings. With a 20% performance improvement and 9% energy savings per frame, this GPU is optimized for next-generation devices and the Android 16 platform. Its enhanced ray-tracing capabilities double the speed of rendering realistic lighting and shadows, delivering a more immersive experience for gaming, multimedia, and virtual reality applications. In addition to its graphical prowess, the Mali G1 Ultra GPU supports advanced features such as variable rate shading and improved texture compression, allowing developers to create visually stunning applications without compromising performance. These enhancements make the GPU an ideal choice for devices ranging from flagship smartphones to high-performance gaming consoles. Arm's latest CPUs and GPUs are engineered to deliver a balance of speed and efficiency, making sure optimal performance for a wide range of applications. Key features include: These advancements ensure that devices powered by Arm's technology can handle intensive tasks such as AI processing, gaming, and multimedia playback while maintaining energy efficiency. One of the standout features of this new generation is its flexibility. Arm's CPUs and GPUs are designed to be highly customizable, allowing manufacturers to tailor CPU clusters to meet the specific needs of various device categories. Whether it's a high-performance smartphone or a compact, energy-efficient system, the designs can be adapted to suit diverse requirements. The compatibility with 3nm fabrication processes ensures that these CPUs and GPUs are ready for the latest manufacturing technologies, allowing manufacturers to produce devices that are both powerful and efficient. This adaptability makes Arm's new lineup a versatile solution for a wide range of industries, from consumer electronics to industrial automation. The potential applications for Arm's latest CPUs and GPUs are vast, spanning multiple industries and use cases. These technologies are designed to power: From gaming and multimedia to AI and IoT, Arm's innovations provide the scalability and performance needed to meet the evolving demands of modern technology. These advancements are not only shaping the future of mobile and embedded computing but also opening new possibilities for developers and manufacturers alike. Arm's latest CPUs and GPUs represent a significant leap forward in computing technology, combining innovative performance with energy efficiency and scalability. With features like SME2 for AI optimization, enhanced ray tracing, and support for advanced fabrication and memory technologies, these designs are tailored to the needs of next-generation devices. Whether you're developing a flagship smartphone, a high-performance gaming console, or an energy-efficient embedded system, Arm's innovations provide the tools to push the boundaries of what's possible in modern computing.
[13]
Arm Holdings' (ARM) Lumex Platform Lifts Smartphone AI
Lumex is primarily focused on mobile devices, but elements of its design will be impacting future chip designs for PCs and other devices as well. The platform represents the company's latest effort to move further up the design and value creation stack for mobile devices. In the never-ending race to improve the AI performance across every device possible, it's always important to keep an eye on Arm Holdings (NASDAQ:ARM). After all, its compute architectures for both CPUs and Bob O'Donnell is the founder and chief analyst of TECHnalysis Research, LLC a technology consulting and market research firm that provides strategic consulting and market research services to the technology industry and professional financial community. You can follow him on Twitter @bobodtech.
[14]
Arm launches new generation of mobile chip designs geared for AI
SAN FRANCISCO (Reuters) -Arm Holdings said on Tuesday it was launching its next-generation set of chip designs called Lumex that it has optimized for artificial intelligence to run on mobile devices such as smartphones and watches without accessing the internet. Called Lumex, the new generation of Arm mobile designs come in four types, ranging from less powerful but more energy efficient ones designed for watches and other smart wearable devices to a version designed to maximize the computing horsepower available. The peak performance design aims to run software that harnesses the power of large AI models without accessing cloud computing on devices such as high-end smartphones. "AI is becoming pretty fundamental to kind of what's happening, whether it's kind of real-time interactions or some killer use cases like AI translation," said Chris Bergey, a senior vice president and general manager at Arm. "We're just seeing (AI) become kind of this expectation." The Lumex designs announced on Tuesday are part of the company's Compute Subsystems (CSS) business, which aims to offer handset makers and chip designers a more ready-made piece of technology that enables their speedy use in new products. The more complete designs Arm is now making for data centers and mobile phones are part of the company's longer-term plans to grow its smartphone and other revenue through a variety of means. Arm has also said it plans to invest more money into examining making its own chips and has hired several key people in order to do so. The Lumex designs are optimized for so-called 3-nanometer manufacturing nodes, such as the process offered by TSMC. The series of iPhone chips Apple announced on Tuesday use a TSMC 3-nanometer process. The company is holding a launch event in China on Wednesday to unveil the new designs because outside of Apple and Samsung the leading handset makers are located there, Bergey said. (Reporting by Max A. Cherney in San Francisco; Editing by Jamie Freed)
Share
Share
Copy Link
Arm unveils its next-generation Lumex platform, promising significant advancements in on-device AI processing, improved performance, and enhanced power efficiency for smartphones and other mobile devices.
Arm Holdings, the UK-based chip designer, has introduced its latest mobile platform, Lumex, aimed at revolutionizing AI capabilities in smartphones and other mobile devices
1
2
3
. This new generation of chip designs promises significant advancements in on-device AI processing, improved performance, and enhanced power efficiency.Source: PCWorld
The Lumex Compute Subsystem (CSS) introduces a range of CPU cores optimized for different performance and efficiency requirements
2
4
. The lineup includes:1
4
.4
.4
.4
.Arm claims that a cluster combining two C1-Ultras with six C1-Pros demonstrates a fivefold increase in AI performance
4
.Source: Geeky Gadgets
Central to Lumex's AI capabilities is the integration of Scalable Matrix Extension v2 (SME2), an instruction set designed to accelerate AI-related mathematical operations
3
4
. This approach enables the CPU to perform AI acceleration directly, resulting in:3
4
.1
.1
.Source: CNET
The new Mali G1-Ultra GPU introduces several improvements
1
4
5
:4
.4
5
.4
.4
.Several major technology players have pledged support for Lumex, including Samsung, MediaTek, Alibaba, and vivo
4
. Google has emphasized SME2's role in making AI more consistent across Android devices and Windows on Arm laptops4
.While Lumex is primarily designed for smartphones, Arm hints at its potential applications in PCs and tablets
5
. The platform's focus on on-device AI processing and improved efficiency could significantly impact the future of mobile computing, enabling more advanced AI applications without compromising battery life or performance1
2
3
.As the mobile industry continues to evolve, Arm's Lumex platform appears poised to play a crucial role in shaping the next generation of AI-powered devices, potentially redefining user experiences across a wide range of applications and use cases.🟡 disadvantaged=🟡### Arm Unveils Next-Generation Lumex Platform
Arm Holdings, the UK-based chip designer, has introduced its latest mobile platform, Lumex, aimed at revolutionizing AI capabilities in smartphones and other mobile devices
1
2
3
. This new generation of chip designs promises significant advancements in on-device AI processing, improved performance, and enhanced power efficiency.The Lumex Compute Subsystem (CSS) introduces a range of CPU cores optimized for different performance and efficiency requirements
2
4
. The lineup includes:1
4
.4
.4
.4
.Arm claims that a cluster combining two C1-Ultras with six C1-Pros demonstrates a fivefold increase in AI performance
4
.Related Stories
Central to Lumex's AI capabilities is the integration of Scalable Matrix Extension v2 (SME2), an instruction set designed to accelerate AI-related mathematical operations
3
4
. This approach enables the CPU to perform AI acceleration directly, resulting in:3
4
.1
.1
.The new Mali G1-Ultra GPU introduces several improvements
1
4
5
:4
.4
5
.4
.4
.Several major technology players have pledged support for Lumex, including Samsung, MediaTek, Alibaba, and vivo
4
. Google has emphasized SME2's role in making AI more consistent across Android devices and Windows on Arm laptops4
.While Lumex is primarily designed for smartphones, Arm hints at its potential applications in PCs and tablets
5
. The platform's focus on on-device AI processing and improved efficiency could significantly impact the future of mobile computing, enabling more advanced AI applications without compromising battery life or performance1
2
3
.As the mobile industry continues to evolve, Arm's Lumex platform appears poised to play a crucial role in shaping the next generation of AI-powered devices, potentially redefining user experiences across a wide range of applications and use cases.
Summarized by
Navi
[2]