☆ Yσɠƚԋσʂ ☆@lemmygrad.ml to Technology@lemmygrad.mlEnglish · 9 days agoDeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAIventurebeat.comexternal-linkmessage-square15fedilinkarrow-up139arrow-down10cross-posted to: technology@lemmy.ml
arrow-up139arrow-down1external-linkDeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAIventurebeat.com☆ Yσɠƚԋσʂ ☆@lemmygrad.ml to Technology@lemmygrad.mlEnglish · 9 days agomessage-square15fedilinkcross-posted to: technology@lemmy.ml
minus-squarepinguinu [any]@lemmygrad.mllinkfedilinkarrow-up1·8 days agoYou can use the smaller models on (beefy) consumer hardware already. That’s something, right? 😅
minus-squareCriticalResist8@lemmygrad.mllinkfedilinkarrow-up3·8 days agoI want the full 1TB model running on my 10 year old linux laptop
minus-squarepinguinu [any]@lemmygrad.mllinkfedilinkarrow-up2·8 days agoJust put your persistent memory as swap. Easy
You can use the smaller models on (beefy) consumer hardware already. That’s something, right? 😅
I want the full 1TB model running on my 10 year old linux laptop
Just put your persistent memory as swap. Easy