Raspberry Pi AI HAT+ 2 brings 40 TOPS and 8GB RAM to run gen AI models locally on Pi 5

8 Sources

Share

Raspberry Pi launched the AI HAT+ 2, a $130 add-on board that equips the Raspberry Pi 5 with a 40 TOPS Hailo 10H chip and 8GB of dedicated RAM. The module enables local processing of generative AI models like Llama 3.2 and Qwen2, though early tests show mixed performance results compared to simply upgrading the base Pi's memory.

Raspberry Pi AI HAT+ 2 Targets Local Generative AI Processing

Raspberry Pi has introduced the AI HAT+ 2, an add-on board for Raspberry Pi 5 designed to run generative AI models locally without cloud connectivity. Announced this week, the $130 module represents a shift from its predecessor's focus on image-based tasks to handling Large Language Models and other generative AI applications

1

. The Hardware Attached on Top (HAT) connects via the single-board computer's PCIe interface and GPIO connector, offloading AI-related workloads from the Raspberry Pi 5's Arm CPU

2

.

Source: Geeky Gadgets

Source: Geeky Gadgets

Hailo 10H AI Chip Delivers 40 TOPS Accelerator Performance

At the heart of the Raspberry Pi AI HAT+ 2 sits the Hailo 10H AI chip, a neural network accelerator capable of delivering 40 TOPS of inference performance

3

. This represents a substantial upgrade from the original AI HAT+, which featured either a Hailo-8 with 26 TOPS or Hailo-8L with 13 TOPS

5

. The new module also includes 8GB of RAM dedicated to AI processing, allowing the board to handle models up to 8GB in size without tapping into the host Pi's memory

2

. This architecture means even lower-spec Raspberry Pi 5 models with 1GB, 2GB, or 4GB of RAM can now serve as viable platforms to accelerate AI workloads, potentially reducing overall project costs

2

.

Source: The Register

Source: The Register

Compatible Models Include Llama 3.2, Qwen2, and DeepSeek-R1-Distill

The AI HAT+ 2 supports several generative AI models at launch, including Llama 3.2, DeepSeek-R1-Distill, Qwen2, Qwen2.5-Coder, and Qwen2.5-Instruct

1

. Most of these are 1.5-billion-parameter models, with Llama 3.2 featuring 1 billion parameters

5

. Raspberry Pi demonstrated the board's capabilities through demos showing text-based descriptions of camera streams and language translation from French to English using Qwen2

1

. Users can install models via hailo-ollama and ollama interfaces, with the foundation promising larger models will become available soon after launch

1

.

Performance Questions and Power Limitations Surface

Early testing reveals performance concerns that potential buyers should consider. Tech YouTuber Jeff Geerling found that a standalone Raspberry Pi 5 with 8GB of RAM generally outperformed the AI HAT+ 2 across supported models

1

. The performance gap appears linked to power draw constraints—while the Pi 5 can operate at up to 10 watts, the AI HAT+ 2 is limited to 3W

1

. Geerling noted that the add-on board's 8GB of RAM "is not quite enough to give this HAT an advantage over just paying for the bigger 16GB Pi with more RAM, which will be more flexible and run models faster"

1

. For edge AI applications requiring local LLM workloads and offline operation, however, the NPU's dedicated processing capabilities may justify the investment

3

.

Computer Vision Capabilities Retained from Previous Generation

While the AI HAT+ 2 focuses primarily on generative AI, it maintains computer vision performance roughly equivalent to the 26 TOPS delivered by the original AI HAT+

3

. Testing confirmed that object identification and pose detection worked as expected using the rpicam-hello suite, with smooth image processing performance

2

. For users exclusively focused on vision-based tasks, The Register questions whether the $130 AI HAT+ 2 makes sense compared to the existing AI HAT+ or the $70 AI camera

3

. The module requires a passive heatsink for the HAT itself, which comes included, while the Raspberry Pi 5 needs separate cooling that fits beneath the board

2

.

Source: XDA-Developers

Source: XDA-Developers

Target Use Cases and Future Model Support

The Raspberry Pi Foundation positions the AI HAT+ 2 for developers building cost-effective, low-latency devices that need to run generative AI models locally

5

. While cloud-based LLMs from OpenAI, Meta, and Anthropic range from 500 billion to 2 trillion parameters, the smaller models supported by the HAT can be fine-tuned or retrained with custom datasets for specific applications

5

. Industry use cases requiring both computer vision and local LLM workloads may find the most value in the new board

3

. The module is available now for $120-$130 depending on retailer, with comprehensive documentation and installation guides provided by the foundation

4

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo