12.9 C
New York
Saturday, November 1, 2025

Buy now

DeepSeek’s distilled new R1 AI model can run on a single GPU

DeepSeek’s up to date R1 reasoning AI mannequin is perhaps getting the majority of the AI neighborhood’s consideration this week. However the Chinese language AI lab additionally launched a smaller, “distilled” model of its new R1, DeepSeek-R1-0528-Qwen3-8B, that DeepSeek claims beats comparably sized fashions on sure benchmarks.

The smaller up to date R1, which was constructed utilizing the Qwen3-8B mannequin Alibaba launched in Might as a basis, performs higher than Google’s Gemini 2.5 Flash on AIME 2025, a group of difficult math questions.

DeepSeek-R1-0528-Qwen3-8B additionally almost matches Microsoft’s lately launched Phi 4 reasoning plus mannequin on one other math expertise take a look at, HMMT.

So-called distilled fashions like DeepSeek-R1-0528-Qwen3-8B are usually much less succesful than their full-sized counterparts. On the plus facet, they’re far much less computationally demanding. In line with the cloud platform NodeShift, Qwen3-8B requires a GPU with 40GB-80GB of RAM to run (e.g., an Nvidia H100). The complete-sized new R1 wants round a dozen 80GB GPUs.

DeepSeek skilled DeepSeek-R1-0528-Qwen3-8B by taking textual content generated by the up to date R1 and utilizing it to fine-tune Qwen3-8B. In a devoted internet web page for the mannequin on the AI dev platform Hugging Face, DeepSeek describes DeepSeek-R1-0528-Qwen3-8B as “for each tutorial analysis on reasoning fashions and industrial improvement centered on small-scale fashions.”

DeepSeek-R1-0528-Qwen3-8B is obtainable underneath a permissive MIT license, that means it may be used commercially with out restriction. A number of hosts, together with LM Studio, already supply the mannequin by way of an API.

Supply hyperlink

See also  Authors call on publishers to limit their use of AI

Related Articles

Leave a Reply

Please enter your comment!
Please enter your name here

Latest Articles