How do you actually use Qwen3 72B Instruct locally? (www.reddit.com)
model roundup
Qwen 3
-
I just got Qwen3 72B Instruct running on a high RAM setup and I’m kinda confused about the proper way to use it. What’s the correct workflow for running it smoothly (like best quant, tools, or runtime)?
-
Going through university right now, and we have massive 100 page pdfs/ppts with soo much fluff that its annoying to go through. until now ive been using chatgpt for it, but realized that the output tokens are HEAVILY limited, and loses a L…
-
If you have low vram - qwen 3 tts is good If you need something unique go for - tada 3b but it need 28gb vram If you want best tts rn + have the commercial use allowed then go for - moss tts 8b its literally the best model out there Litera…
-
-
LLM speed t/s (www.reddit.com)
-
7B showdown on 18GB (benchmark) (www.reddit.com)
Hey r/LocalLLaMA, I've been coding for a while but not in the local AI space and wanted to run some benchmarks on my 18GB M3 Pro. The theme of this one was "specialists vs generalists" at the 7-8B range: qwen2.5-coder:7b, deepseek-r1:7b, m…