# For Base Model
-
this is deepseek-v3. deepseek-r1 is the model that got all the media hype: https://huggingface.co/deepseek-ai/DeepSeek-R1
-
[email protected]replied to [email protected] last edited by
r1 is lightweight and optimized for local environments on a home PC. It's supposed to be pretty good at programming and logic and kinda awkward at conversation.
v3 is powerful and meant to run on cloud servers. It's supposed to make for some pretty convincing conversations.
-
[email protected]replied to [email protected] last edited by
Base models are general purpose language models, mainly useful for AI researchers and people who want to build on top of them.
Instruct models are chatbots. They are made by fine-tuning base models.
-
[email protected]replied to [email protected] last edited by
Is that website legit?
I've only ever seen https://www.deepseek.com/And I would personally recommend downloading from HuggingFace or Ollama
-
[email protected]replied to [email protected] last edited by
R1 isn’t really runnable with a home rig. You might be able to run a distilled version of the model though!
-
[email protected]replied to [email protected] last edited by
Tell that to my home rig currently running the 671b model...
-
[email protected]replied to [email protected] last edited by
That likely is one of the distilled versions I’m talking about. R1 is 720 GB, and wouldn’t even fit into memory on a normal computer. Heck, even the 1.58-bit quant is 131GB, which is outside the range of a normal desktop PC.
But I’m sure you know what version you’re running better than I do, so I’m not going to bother guessing.
-
[email protected]replied to [email protected] last edited by
It's not
-
[email protected]replied to [email protected] last edited by
You're absolutely right, I wasn't trying to get that in-depth, which is why I said "lightweight and optimized," instead of "when using a distilled version" because that raises more questions than it answers. But I probably overgeneralized by making it a blanket statement like that.
-
[email protected]replied to [email protected] last edited by
You must have a lot of memory, sounds like a lot of fun!
-
[email protected]replied to [email protected] last edited by
I heard people saying they could run the r1 32B model on moderate gaming hardware albeit slowly
-
[email protected]replied to [email protected] last edited by
Yea, comment OP needs to edit links with howany up votes that got.
-
[email protected]replied to [email protected] last edited by
32b is still distilled. The full one is 671b.
-
[email protected]replied to [email protected] last edited by
My legion slim 5 14" can run it not too bad.
-
[email protected]replied to [email protected] last edited by
I know, but the fall off in performance isn't supposed to be severe
-
[email protected]replied to [email protected] last edited by
You are correct. And yes that is kinda the whole point of the distilled models.
-
[email protected]replied to [email protected] last edited by
I know. Lmao