To run deepseek v4 flash how much max vram we need? 175 gb or 320gb? (www.reddit.com)
model roundup
DeepSeek 4
-
As far as i know the weight is of 160gb + 9.6gb needed for max 1 million token window + 5 gigs overhead = 175gb vram. But vllm and othere sources said "To use the full 1M context, you need 4x A100 80G" --> thats a 320gb vram ??
-
Show HN: A CLI to use any model in your coding agent (getaivo.dev via hn)
Hi everyone, I've been working on a CLI tool that can help to easily run any model in claude, Codex, Gemini, Pi, and OpenCode. It's also an API keys manager, supports multiple providers or OpenAI/Claude/Gemini accounts.
-
Is Deepseek V4 really out? (www.reddit.com)
Hello Guys, Each time a new local llm is released, there are a ton of new posts , this is it, it's near Opus level...., the abliteration matrix final something at Q2 KXLDND is the best but it's been a day that deepseek was released and i d…
-
Deepseek V4 flash (high) rivals Gemini 3 flash at 1/5th the cost (www.reddit.com)
-
-
-
DeepSeek v4 - Subjective vibes (www.reddit.com)
-
-
🚨 The Chinese beast is BACK… DeepSeek just dropped V4 (www.reddit.com)
-
What is the next SOTA modei you are excited about? (www.reddit.com)
-
Deepseek V4 AGI comfirmed (www.reddit.com)
-
-
-
Ask HN: Why is cache for DeepSeek-v4 cheapest on Vercel AI Gateway? (news.ycombinator.com)
-
-
-
Takeaways & discussion about the DeepSeek V4 architecture (www.reddit.com)
-
Budget to run Deepseek V4 locally at FP4 precision (www.reddit.com)
-
DeepSeek-v4 has a comical 384K max output capability (www.reddit.com)
-
Deepseek v4 people (www.reddit.com)
-
DeepSeek V4 - almost on the frontier, a fraction of the price (simonwillison.net)
-
DeepSeek-V4 Drops: Open-Source Push Toward Cheaper, Long-Context AI. (www.reddit.com)
-
No Multimodality yet in DeepSeek-V4. But I'll wait. (www.reddit.com)
-
-
DeepSeek V4 Benchmarks! (www.reddit.com)
-
DeepSeek V4 is out. the best open-source on coding. here's the breakdown (news.ycombinator.com)
-
DeepSeek-V4: Towards Highly Efficient Million-Token Context Intelligence (huggingface.co via hn)
- DeepSeek-V4: Making 1M token context efficient (firethering.com)
- DeepSeek V4 in vLLM: Efficient Long-Context Attention (vllm-website-pdzeaspbm-inferact-inc.vercel.app)
- DeepSeek-V4: a million-token context that agents can actually use (huggingface.co)
-
DeepSeek-V4 Preview Version is launched (news.ycombinator.com)
-
DeepSeek V4 has released (www.reddit.com)
-
DeepSeek-V4 Technical Report [pdf] (huggingface.co via hn)
-
Deepseek V4 Flash and Non-Flash Out on HuggingFace (www.reddit.com)
-
Recent Open models from last 6 Months - Nov 2025 - Apr 2026 (www.reddit.com)