Select Region/Language
EN CN
RPI Developer / Product Applications / How is Deepseek R1 on a Raspberry Pi?

How is Deepseek R1 on a Raspberry Pi?

EDATECPublish at 2025-02-12
8530 views

OpenAI, which is only really open about consuming all the world's energy and half a trillion of our taxpayer dollars, just got rattled to its core.

Deepseek, a new AI startup run by a Chinese hedge fund, allegedly created a new open weights model called R1 that beats OpenAI's best model in every metric.

And they did it for $6 million, with GPUs that run at half the memory bandwidth of OpenAI's.

Besides the embarassment of a Chinese startup beating OpenAI using one percent of the resources (according to Deepseek), their model can 'distill' other models to make them run better on slower hardware.

Meaning a Raspberry Pi can run one of the best local Qwen AI models even better now.

OpenAI's entire moat is predicated on people not having access to the insane energy and GPU resources to train and run massive AI models.

But that moat disappears if everyone can buy a GPU and run a model that's good enough, for free, any time they want.

Raspberry Pi AI

But sensationalist headlines aren't telling you the full story.

The Raspberry Pi can technically run Deepseek R1... but it's not the same thing as Deepseek R1 671b, which is a four hundred gigabyte model.

That model (the one that actually beats ChatGPT), still requires a massive amount of GPU compute.

But the big difference is, assuming you have a few 3090s, you could run it at home. You don't have to pay OpenAI for the privilege of running their fancy models.

You can just install Ollama, download Deepseek, and play with it to your heart's content.

And even if you don't have a bunch of GPUs, you could technically still run Deepseek on any computer with enough RAM.

I tested Deepseek R1 671B using Ollama on the AmpereOne 192-core server with 512 GB of RAM, and it ran at just over 4 tokens per second. Which isn't crazy fast, but the AmpereOne won't set you back like $100,000, either!

Even though it's only using a few hundred watts—which is honestly pretty amazing—a noisy rackmount server isn't going to fit in everyone's living room.

A Pi could, though. So let's look at how the smaller 14b model runs on it:

It's... definitely not gonna win any speed records. I got around 1.2 tokens per second.

It runs, but if you want a chatbot for rubber duck debugging, or to give you a few ideas for your next blog post title, this isn't fun.

Raspberry Pi GPU AI

But we can speed things up. A lot. All we need is an external graphics card, because GPUs and the VRAM on them are faster than CPUs and system memory.

I have this setup I've been testing with an AMD W7700 graphics card. It has 16 gigs of speedy VRAM, and as long as it can fit the whole AI model, it should be much faster:

modelsizeparamsbackendngltestt/s
qwen2 14B Q4_K - Medium8.37 GiB14.77 BVulkan99pp512193.31 ± 0.35
qwen2 14B Q4_K - Medium8.37 GiB14.77 BVulkan99pp4096168.01 ± 0.25
qwen2 14B Q4_K - Medium8.37 GiB14.77 BVulkan99tg12824.41 ± 0.24
qwen2 14B Q4_K - Medium8.37 GiB14.77 BVulkan99pp4096+tg12854.26 ± 0.92

llama-bench reports 24 to 54 tokens per second, and this GPU isn't even targeted at LLMs—you can go a lot faster. For full test results, check out my ollama-benchmark repo: Test Deepseek R1 Qwen 14B on Pi 5 with AMD W7700.

https://github.com/geerlingguy/ollama-benchmark/issues/9

Conclusion

AI is still in a massive bubble. Nvidia just lost more than half a trillion dollars in value in one day after Deepseek was launched.

But their stock price is still 8x higher than it was in 2023, and it's not like anyone's hyping up AI any less now.

Original link: https://www.jeffgeerling.com/blog/2025/how-deepseek-r1-on-raspberry-pi

Raspberry Pi OpenAl RaspberryPi CM5 Deepseek

EDATEC

Started in 2017, EDATEC has been certified as one of the global design partners of Raspberry Pi since 2018, providing Raspberry Pi based industrial computers design and manufacturing services for global customers.

  • Building 29, No. 1661, Jialuo Road, Jiading District, Shanghai.
  • CN: +86-18217351262 (Iris Zhou)
    US: +1 859-653-0800 (Randall Restle)
  • support@edatec.cn
Submit Your Message

INFO

By submitting, you agree to EDATEC's policies

Please enter the verification code

captcha
1.142511s