DeepSeek's distilled new R1 AI model can run on a single GPU | TechCrunch
-
Save me the search, please. What's deepsite?
Above is what I can do with deepsite by pasting in the first page of your lemmy profile and the prompt:
"This is double_quack, a lemmy user on Lemmy, a new social media platform. Create a cool profile page in a style that they'll like based on the front page of their lemmy account (pasted in a ctrl + a, ctrl + c, ctrl + v of your profile)."
It not perfect by any stretch of the imagination, but like, its not a bad starting point.
if you want to try it: https://huggingface.co/spaces/enzostvs/deepsite
-
Above is what I can do with deepsite by pasting in the first page of your lemmy profile and the prompt:
"This is double_quack, a lemmy user on Lemmy, a new social media platform. Create a cool profile page in a style that they'll like based on the front page of their lemmy account (pasted in a ctrl + a, ctrl + c, ctrl + v of your profile)."
It not perfect by any stretch of the imagination, but like, its not a bad starting point.
if you want to try it: https://huggingface.co/spaces/enzostvs/deepsite
Excuse me... what? Ok, that's something...
-
Excuse me... what? Ok, that's something...
Here I'm DM"ing you something. Its very personal, but I want to share it with you and I made it using Deepsite (in part).
-
7b trash model?
I'm genuinely curious what you do that a 7b model is "trash" to you? Like yeah sure a gippity now tends to beat out a mistral 7b but I'm pretty happy with my mistral most of the time if I ever even need ai at all.
-
This post did not contain any content.
ew probably still censored.
-
ew probably still censored.
You can self host it right??
-
ew probably still censored.
The censorship only exists on the version they host, which is fair enough. If they're running it themselves in China, they can't just break the law.
If you run it yourself, the censorship isn't there.
-
The censorship only exists on the version they host, which is fair enough. If they're running it themselves in China, they can't just break the law.
If you run it yourself, the censorship isn't there.
Yeah, i think the censoring in the LLM data itself would be pretty vulnerable to circumvention.
-
the Chinese AI lab also released a smaller, “distilled” version of its new R1, DeepSeek-R1-0528-Qwen3-8B, that DeepSeek claims beats comparably sized models on certain benchmarks
Most models come in 1B, 7-8B, 12-14B, and 27+B parameter variants. According to the docs, they benchmarked the 8B model using an NVIDIA H20 (96 GB VRAM) and got between 144-1198 tokens/sec. Most consumer GPUs probably aren’t going to be able to keep up with
Depends on the quantization.
7B is small enough to run it in FP8 or a Marlin quant with SGLang/VLLM/TensorRT, so you can probably get very close to the H20 on a 3090 or 4090 (or even a 3060) and you know a little Docker.
-
You can self host it right??
if the model is censored... then what, retraining it? Or doing it from scratch like what open-r1 is doing?
-
The censorship only exists on the version they host, which is fair enough. If they're running it themselves in China, they can't just break the law.
If you run it yourself, the censorship isn't there.
Untrue, I downloaded the vanilla version and it's hardcoded in.
-
You can self host it right??
The self hosted model has hard coded censored content.