2 Sources
2 Sources
[1]
SiFive aims to cram more RISC-V cores into AI chips
Why reinvent the CPU wheel when you can spend your time engineering a way out of your dependence on Nvidia? Every quarter, Nvidia CEO Jensen Huang is asked about the growing number of custom ASICs encroaching on his AI empire, and each time he downplays the threat, arguing that GPUs offer superior programmability in a rapidly changing environment. That hasn't stopped chip designer SiFive from releasing RISC-V-based core designs for use in everything from IoT devices to high-end AI accelerators, including Google's Tensor Processing Units (TPUs) and Tenstorrent's Blackhole accelerators. Last summer, SiFive revealed that its RISC-V-based core designs power chips from five of the "Magnificent 7" companies - those include Alphabet, Amazon, Apple, Meta, Microsoft, Nvidia, and Tesla. And while we're sure many of those don't involve AI, it's understandable why analysts keep asking Huang about custom ASICs. While many of you will be familiar with SiFive for Meta and Microsoft's RISC-V development boards, the company's main business is designing and licensing core IP, similar to Brit chip designer Arm Holdings. These intellectual property (IP) offerings are what we're discussing here. This week at the AI Infra Summit, the RISC-V chip designer revealed its second generation of Intelligence cores, including new designs aimed at edge AI applications like robotics sensing, as well as upgraded versions of its X200 and X300 Series, and XM Series accelerator. All of these designs are based on an eight-stage dual-issue in-order superscalar processor architecture, which is the long way of saying they aren't designed for use in a general-purpose CPU like the one powering whatever device you're reading this on. Instead, SiFive's Intelligence line is intended to serve as an accelerator control unit (ACU) for keeping its customers' tensor cores and matrix multiply-accumulate units (MAC) from starving for data. The idea is that rather than investing resources to design something custom for the job, SiFive customers can license its Intelligence family instead. This is what Google, Tenstorrent and a number of tech titans have apparently done. SiFive's two newest Intelligence cores, the X160 and X180, are aimed at low-power applications like IoT devices, drones, and robotics. The X180 is a 64-bit core, while the X160 is based on the 32-bit RV32I instruction set architecture. Customers can arrange them in clusters of up to four (although a chip could have multiple four-core clusters on it), and they support 128-bit wide vector registers and feature a 64-bit wide data path. This allows them to contend with many modern data types, including INT8 and more commonly BF16, which is kind of important when they're attached to an accelerator designed to run models at those precisions. How these cores communicate with those accelerators has changed a little in SiFive's second generation. In addition to its Vector Coprocessor Interface Extension (VCIX), which provides high-bandwidth access to the CPU core's vector registers, the chip designer's second gen parts now feature SiFive's Scalar Coprocessor Interface (SSCI), which provides custom RISC-V instructions that give accelerators direct access to the CPU's registers. Alongside the new cores, SiFive is also rolling out updated versions of its X280 and X390 family of processor cores. Like the X100-class, both cores feature eight-stage dual-issue in-order execution pipelines and can be arranged in clusters of one, two, or four cores depending on your application. Like last-gen, the X280 and X390 Gen 2 cores boast support for 512-bit and 1024-bit wide vector registers, respectively. As you may recall, Google used SiFive's X280 cores to manage the matrix multiplication units (MXUs) in its Tensor Processing Units back in 2022. One key difference is that, for the second gen of the cores, SiFive has upgraded to the RVA23 instruction set, which adds hardware support for BF16 and OCP's MXFP8 and MXFP4 micro-scaling data types. The latter is gaining considerable attention as of late, as it's how OpenAI opted to release its open weights gpt-oss models. SiFive has also optimized the X280 Gen 2's cache hierarchy, moving from a three-level arrangement - L1, L2, and shared L3 - to a simpler and more customizable one that ditches the L3 for up to 1MB of shared L2 per core cluster. The company says this new arrangement boosts utilization and saves on die area. In terms of performance, SiFive says its larger X390 Gen 2 now boasts 4x the compute and 32x the data throughput of the original X280, allowing for up to 1TB/s of data movement in the four core cluster config. SiFive is positioning the X390 as both a standalone AI accelerator core, presumably leaning heavily on its massive vector registers, as well as an ACU, an application that also benefits from the introduction of the SSCI interface we mentioned earlier. Finally, SiFive has also upgraded the XM family of take-and-bake AI accelerator IP to use its X390 Gen 2 cores. We first looked at SiFive's XM products last summer, which serve as a blueprint for building a scalable AI accelerator. SiFive's Gen 2 XM clusters now combine its in-house matrix math engine with four of its updated X390 cores. According to SiFive, each XM cluster can deliver 64 teraFLOPS of FP8 performance at 2GHz and can be scaled up to support chips with more than 4 petaFLOPS of AI performance. All of these CPU core and accelerator designs are now available to license, with the first customer silicon based on them expected to hit the market sometime in Q2 2026. ®
[2]
SiFive introduces new processor core designs for AI devices - SiliconANGLE
SiFive introduces new processor core designs for AI devices Chip startup SiFive Inc. today debuted four new central processing unit cores optimized to run artificial intelligence models. Santa Clara, California-based SiFive was valued at $2.5 billion following its most recent funding round in 2022. The company develops CPU designs based on the open-source RISC-V instruction set architecture, or ISA. An ISA defines the collection of low-level computing operations that a chip mixes and matches to process data. SiFive's new cores expand upon RISC-V's feature set by adding vector extensions. Those are components that allow a chip to more efficiently process multiple data points at once. The ability to parallelize computations can significantly speed up AI models, which use a large number of relatively simple computations to crunch data. The new lineup is headlined by the X160 and X180. Both designs include vector processing features that speed up convolutions, which are calculations used by computer vision models to process images. The features can also accelerate certain data filtering and transformation tasks. The X160 and X180 can both be configured to run a real-time operating system. This is a type of operating system that ensures calculations are always carried out in a predefined amount of time, which is important for certain systems. A manufacturer, for example, may wish to guarantee that a factory sensor will always generate an alert about equipment failures within three seconds of detecting them. The X160 ships with up to 200 kibibyte of cache and a 2-mebibyte memory. Besides industrial equipment, SiFive also sees the chip finding use in consumer devices such as fitness trackers. Furthermore, the X160 can be installed in systems that contain multiple AI accelerators to manage the chips and block firmware tempering. The X180 is 10% faster than the X160. It can also hold more data thanks to two built-in caches with a combined capacity of more than 4 mebibytes. According to SiFive, the chip lends itself to training AI models and powering certain types of data center equipment. The two other cores that the company debuted today are upgraded versions of existing designs. The X280 is geared toward consumer devices such as augmented reality headsets, while X390 can also power cars and infrastructure systems. The latter core performs vector processing four times faster. According to SiFive, its engineers enhanced the two designs with a new co-processor interface. The technology will make it easier to integrate the cores into systems-on-chip that also include AI accelerators. Additionally, the company has upgraded the memory subsystem in the X280 and X390 to reduce latency.
Share
Share
Copy Link
SiFive introduces new and upgraded RISC-V-based processor core designs optimized for AI applications, aiming to enhance performance in various devices from IoT to data centers.
SiFive, a prominent chip designer, has unveiled its latest generation of RISC-V-based processor core designs, specifically tailored for AI applications. This move signals a significant push to challenge Nvidia's dominance in the AI chip market by offering alternatives to custom ASIC designs
1
.The company has introduced two new core designs, the X160 and X180, aimed at low-power applications such as IoT devices, drones, and robotics. These cores feature:
1
The X160 and X180 are designed to accelerate convolutions for computer vision models and data filtering tasks. They can run real-time operating systems, making them suitable for industrial equipment and consumer devices like fitness trackers
2
.SiFive has also updated its existing X280 and X390 core designs:
1
The X390 Gen 2 now boasts 4x the compute and 32x the data throughput of the original X280, allowing for up to 1TB/s of data movement in a four-core cluster configuration
1
.Related Stories
SiFive has introduced new features to improve integration and performance:
2
SiFive's RISC-V-based core designs have gained traction in the industry:
1
2
As the AI chip market continues to evolve, SiFive's latest offerings present a compelling alternative to custom ASIC designs, potentially challenging Nvidia's position in the AI acceleration space.
Summarized by
Navi
[1]
19 Sept 2024
23 Oct 2024•Technology
25 Sept 2024
1
Business and Economy
2
Business and Economy
3
Policy and Regulation