3 Sources
[1]
DeepSeek R1 Distilled Models Now Available on AMD Ryzen AI and Radeon
Reasoning models are a new class of large language models (LLMs) designed to tackle highly complex tasks by employing chain-of-thought (CoT) reasoning with the tradeoff of taking longer to respond. The DeepSeek R1 is a recently released frontier "reasoning" model which has been distilled into highly capable smaller models. Deploying these DeepSeek R1 distilled models on AMD Ryzenâ„¢ AI processors and Radeonâ„¢ graphics cards is incredibly easy and available now through LM Studio.
[2]
AMD shows you how to run DeepSeek on your Ryzen AI CPU and Radeon GPU
TL;DR: AMD has introduced instructions for running DeepSeek's R1 reasoning model on AMD Ryzen AI and Radeon products. This model enhances problem-solving by performing extensive reasoning before providing answers, improving analytical capabilities at the cost of longer response times. It is optimized for complex tasks in math and science. AMD has provided some instructions on how to run DeepSeek's exciting new R1 distilled "reasoning" model on AMD Ryzen AI and Radeon products... yeah, you can run a local DeepSeek R1 model on your PC at home. AMD explains: "Reasoning models add a "thinking" stage before the final output - which you can see by expanding the "thinking" window before the model gives its final answer. Unlike conventional LLMs, which one-shot the response, CoT LLMs perform extensive reasoning before answering. The assumptions and self-reflection the LLM performs are visible to the user and this improves the reasoning and analytical capability of the model - albeit at the cost of significantly longer time-to-first-(final output)token". "A reasoning model may first spend thousands of tokens (and you can view this chain of thought!) to analyze the problem before giving a final response. This allows the model to be excellent at complex problem-solving tasks involving math and science and attack a complex problem from all angles before deciding on a response. Depending on your AMD hardware, each of these models will offer state-of-the-art reasoning capability on your AMD Ryzenâ„¢ AI processor or Radeonâ„¢ graphics cards". You can read the entire article with a step-by-step process on running DeepSeek R1 on your AMD Ryzen AI and Radeon-powered products right here.
[3]
AMD released instructions for running DeepSeek on Ryzen AI CPUs and Radeon GPUs
DeepSeek R1 can now be run on AMD's latest consumer-based hardware. AMD has provided instructions on how to run DeepSeek's R1 AI model on AI-accelerated Ryzen AI and Radeon products, making it easy for users to run the new chain-of-thought model on their PCs locally. Several LLMs using R1 are compatible with RX 7000 series desktop GPUs and select Ryzen CPUs with XDNA NPUs. However, they require the optional driver Adrenalin 25.1.1 to run. The guide has everything AMD users need to get DeepSeek R1 running on their local (supported) machine. LM Studio has a one-click installer tailor-made for Ryzen AI, which is the method AMD users will use to install R1. AMD shows how the application needs to be tuned for its hardware, including a list of the maximum supported LLM parameters. DeepSeek R1 allegedly has only recently been distilled into "highly capable" smaller models, small enough to run on consumer-based hardware. The DeepSeek-V3 model was initially trained on a cluster of 2,048 Nvidia H800 GPUs for context. The maximum supported LLM parameters are based on memory capacity. The RX 7600 XT, 7700 XT, 7800 XT, 7900 GRE, and 7900 XT all support up to "DeepSeek-R1-Distill-Qwen-14B". The flagship RX 7900 XTX supports up to "DeepSeek-R1-Distill-Qwen-32B". The RX 7600, with its 8GB of VRAM, supports up to "DeepSeek-R1-Distill-Llama-8B". Similarly, Ryzen 8040 and 7040 series mobile APUs are equipped with 32GB of RAM, and the Ryzen AI HX 370 and 365 with 24GB and 32GB of RAM can support up to "DeepSeek-R1-Distill-Llama-14B". The Ryzen AI Max+ 395 can support up to "DeepSeek-R1-Distill-Llama-70B", but only in 128GB and 64GB memory capacities. The 32GB supports up to "DeepSeek-R1-Distill-Qwen-32B". DeepSeek's new AI model has taken the world by storm, with its 11 times lower computing cost than leading-edge models. Two days ago, it was solely responsible for Nvidia's record-breaking $589 billion market cap loss. The DeepSeek R1 model relies on extreme optimization levels to provide its 11X efficiency uplift, relying on Nvidia's assembly-like Parallel Thread Execution (PTX) programming for most of the performance uplift. Nvidia and AMD GPUs aren't the only GPUs that can run R1; Huawei has already implemented DeepSeek support into its Ascend AI GPUs, enabling performant AI execution on homegrown Chinese hardware.
Share
Copy Link
AMD has released instructions for running DeepSeek's R1 distilled reasoning models on Ryzen AI processors and Radeon graphics cards, bringing advanced AI capabilities to consumer hardware.
AMD has taken a significant step in democratizing AI technology by providing instructions for running DeepSeek's R1 distilled "reasoning" models on AMD Ryzen AI processors and Radeon graphics cards. This move allows users to harness the power of sophisticated AI models on their personal computers, marking a notable advancement in consumer AI capabilities 12.
DeepSeek R1 is a frontier "reasoning" model, recently distilled into smaller, highly capable versions. These models are designed to tackle complex tasks using chain-of-thought (CoT) reasoning. Unlike conventional large language models (LLMs) that provide immediate responses, CoT LLMs perform extensive reasoning before answering, improving their analytical capabilities at the cost of longer response times 12.
The deployment of DeepSeek R1 distilled models on AMD hardware is remarkably straightforward, thanks to LM Studio's one-click installer tailored for Ryzen AI. AMD has provided a comprehensive guide detailing the process, including how to tune the application for optimal performance on their hardware 23.
AMD's latest consumer hardware, including RX 7000 series desktop GPUs and select Ryzen CPUs with XDNA NPUs, are compatible with several LLMs using R1. The maximum supported LLM parameters vary based on memory capacity:
DeepSeek R1 has garnered significant attention in the AI community due to its impressive efficiency. The model boasts 11 times lower computing costs compared to leading-edge models, achieved through extreme optimization levels and the use of Nvidia's Parallel Thread Execution (PTX) programming 3.
The availability of DeepSeek R1 on consumer hardware represents a significant democratization of AI technology. It allows individuals to run sophisticated AI models locally, potentially opening up new avenues for personal and professional applications of AI. This development also highlights the ongoing competition in the AI hardware space, with AMD positioning itself as a strong contender in providing AI capabilities to end-users 123.
Summarized by
Navi
[1]
NVIDIA CEO Jensen Huang confirms the development of the company's most advanced AI architecture, 'Rubin', with six new chips currently in trial production at TSMC.
2 Sources
Technology
17 hrs ago
2 Sources
Technology
17 hrs ago
Databricks, a leading data and AI company, is set to acquire machine learning startup Tecton to bolster its AI agent offerings. This strategic move aims to improve real-time data processing and expand Databricks' suite of AI tools for enterprise customers.
3 Sources
Technology
17 hrs ago
3 Sources
Technology
17 hrs ago
Google is providing free users of its Gemini app temporary access to the Veo 3 AI video generation tool, typically reserved for paying subscribers, for a limited time this weekend.
3 Sources
Technology
9 hrs ago
3 Sources
Technology
9 hrs ago
Broadcom's stock rises as the company capitalizes on the AI boom, driven by massive investments from tech giants in data infrastructure. The chipmaker faces both opportunities and challenges in this rapidly evolving landscape.
2 Sources
Technology
17 hrs ago
2 Sources
Technology
17 hrs ago
Apple is set to introduce new enterprise-focused AI tools, including ChatGPT configuration options and potential support for other AI providers, as part of its upcoming software updates.
2 Sources
Technology
17 hrs ago
2 Sources
Technology
17 hrs ago