Edward Snowden slams Nvidia's RTX 50-series 'F-tier value,' whistleblows on lackluster VRAM capacity
Edward Snowden slams Nvidia's RTX 50-series 'F-tier value,' whistleblows on lackluster VRAM capacity
Blackwell consumer GPUs offer 'F-tier value for S-tier prices, ' moans the naturalized Russian.

You're viewing a single thread.
TeamAssimilation @infosec.pub
Edward Snowden doing GPU reviews? This timeline is becoming weirder every day.
395 0 ReplyWinged_Hussar @lemmy.world
Legitimately thought this was a hard-drive.net post
89 0 ReplyGamingChairModel @lemmy.world "Whistleblows" as if he's some kind of NVIDIA insider.
56 0 Reply0x0 @programming.dev Intel Insider now that would've made for great whistleblowing headlines.
1 0 Reply
Eager Eagle @lemmy.world
I bet he just wants a card to self host models and not give companies his data, but the amount of vram is indeed ridiculous.
48 0 ReplyJeena @piefed.jeena.net
Exactly, I'm in the same situation now and the 8GB in those cheaper cards don't even let you run a 13B model. I'm trying to research if I can run a 13B one on a 3060 with 12 GB.
25 0 ReplyThe Hobbyist @lemmy.zip You can. I'm running a 14B deepseek model on mine. It achieves 28 t/s.
15 0 ReplyJeena @piefed.jeena.net
Oh nice, that's faster than I imagined.
6 0 Replylevzzz @lemmy.world You need a pretty large context window to fit all the reasoning, ollama forces 2048 by default and more uses more memory
4 0 ReplyViri4thus @feddit.org I also have a 3060, can you detail which framework (sglang, ollama, etc) you are using and how you got that speed? i'm having trouble reaching that level of performance. Thx
2 0 ReplyThe Hobbyist @lemmy.zip Ollama, latest version. I have it setup with Open-WebUI (though that shouldn't matter). The 14B is around 9GB, which easily fits in the 12GB.
I'm repeating the 28 t/s from memory, but even if I'm wrong it's easily above 20.
Specifically, I'm running this model: https://ollama.com/library/deepseek-r1:14b-qwen-distill-q4_K_M
Edit: I confirmed I do get 27.9 t/s, using default ollama settings.
4 0 ReplyViri4thus @feddit.org Ty. I'll try ollama with the Q-4-M quantization. I wouldn't expect to see a difference between ollama and SGlang.
2 0 ReplyJeena @piefed.jeena.net
Thanks for the additional information, that helped me to decide to get the 3060 12G instead of the 4060 8G. They have almost the same price but from what I gather when it comes to my use cases the 3060 12G seems to fit better even though it is a generation older. The memory bus is wider and it has more VRAM. Both video editing and the smaller LLMs should be working well enough.
2 0 Reply
manicdave @feddit.uk I'm running deepseek-r1:14b on a 12GB rx6700. It just about fits in memory and is pretty fast.
4 0 Reply
secret300 @lemmy.sdf.org Swear next he's gonna review hentai games
Oh wait... https://www.youtube.com/watch?v=fAf1Syz17JE
11 0 Replynewcockroach @lemmy.world
"Some hentai games are good" -Edward Snowden
8 0 ReplySiegfried @lemmy.world Note that this is from 2003
2 0 Reply
ඞmir @lemmy.ml
I'll keep believing this is a theonion post
9 0 ReplySimulation6 @sopuli.xyz Does he work for Nvidia? Seems out of character for him.
1 0 Reply