>>384447 >Are there any models that run locally? Yeah, plenty, but what you can run depends heavily on your VRAM.
Text generation models require a lot of it, way more than image generation, for example.
They'll also be inferior to the big models in terms of intelligence, instruction-following, and pony knowledge.
>I heard Deepseek has models that can run locally with 16GB of RAM. Unfortunately, that claim is mostly clickbait.
When people talk about Deepseek, they're usually referring to one of two models:
>Deepseek R1 An open-source "thinking" model. This is the one that made them well-known.
It's technically open-source, but you're not running that locally, unless you're sitting on something absurd.
Even with quantization, you'd probably need 200–300GB of VRAM or RAM to run it.
Not sure about the calcul, but the short version: you're not running it on a normal consumer machine.
>Deepseek V3 Also a large model, but not a "thinking" one.
Same deal, you're not running this locally either.
The "Deepseek" models people say can run locally aren't really Deepseek models.
They're smaller local models that replicated R1's "thinking" approach and core ideas, but they aren't the same thing.
So if you hear someone say they're running "Deepseek" on a Raspberry Pi, it's actually just some janky 1.5B model that mimics the concept but not the performance.
From what I've heard, those versions aren't great anyway, but I haven't tested them myself.
If you want to try the real R1 or V3, you can use free OpenRouter keys, but obviously, that's not local.