Brianna Wu

@briannawu@mstdn.social

Doing media today. Do we have any evidence that DeepSeek is computationally less expensive beyond china’s claims?

January 29, 2025 at 11:55:13 AM

Well, you can download and run R1 locally, where you could presumably benchmark it.

But what computational power went into the model they are working with?

I think the only evidence it must have been less computing power is (1) their supposedly lower budget, (2) the fact that they aren’t allowed to buy nice NVIDIA chips.

So yeah there’s some “trust us bro” in there for sure.

I literally just watched this. The presenter is a long-time Pi-et-al reviewer. (I have not tried any of it myself, but I used to know him years ago and I trust him to be honest about his findings.)

youtube.com/watch?v=o1sN1lB76E

YouTube

OpenAI's nightmare: Deepseek R1 on a Raspberry Pi

You have to differentiate between training costs and running costs. Running costs can be independently verified since the model is open source, training costs cannot.

Elk Logo

Welcome to Elk!

Elk is a nimble Mastodon web client. You can login to your Mastodon account and use it to interact with the fediverse.

Expect some bugs and missing features here and there. Elk is Open Source and we're actively improving it as a community project. Join us and let's build it together!

If you'd like to report a bug, help us testing, give feedback, or contribute, reach out to us on GitHub and get involved.

To boost development, you can sponsor the Team through GitHub Sponsors. We hope you enjoy Elk!

Daniel RoeTAKAHASHI ShuujiAnthony FuJoaquín SánchezPatak三咲智子 Kevin Deng

The Elk Team