☆ Yσɠƚԋσʂ ☆@lemmygrad.ml to Technology@lemmygrad.mlEnglish · 5 days agoDeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAIventurebeat.comexternal-linkmessage-square15fedilinkarrow-up138arrow-down11cross-posted to: [email protected]
arrow-up137arrow-down1external-linkDeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAIventurebeat.com☆ Yσɠƚԋσʂ ☆@lemmygrad.ml to Technology@lemmygrad.mlEnglish · 5 days agomessage-square15fedilinkcross-posted to: [email protected]
minus-squareCriticalResist8@lemmygrad.mllinkfedilinkarrow-up4·4 days agoBut imagine what you’ll be able to run it on in four more months. But yeah, it’s stretching the definition of consumer hardware a bit.
minus-squarepinguinu [any]@lemmygrad.mllinkfedilinkarrow-up1·4 days agoYou can use the smaller models on (beefy) consumer hardware already. That’s something, right? 😅
minus-squareCriticalResist8@lemmygrad.mllinkfedilinkarrow-up3·4 days agoI want the full 1TB model running on my 10 year old linux laptop
minus-squarepinguinu [any]@lemmygrad.mllinkfedilinkarrow-up2·4 days agoJust put your persistent memory as swap. Easy
But imagine what you’ll be able to run it on in four more months. But yeah, it’s stretching the definition of consumer hardware a bit.
You can use the smaller models on (beefy) consumer hardware already. That’s something, right? 😅
I want the full 1TB model running on my 10 year old linux laptop
Just put your persistent memory as swap. Easy