model roundup

DeepSeek 4

35 items · started 2026-04-22 · ongoing (last activity 2026-04-25)

  1. As far as i know the weight is of 160gb + 9.6gb needed for max 1 million token window + 5 gigs overhead = 175gb vram. But vllm and othere sources said "To use the full 1M context, you need 4x A100 80G" --> thats a 320gb vram ??

  2. Hi everyone, I've been working on a CLI tool that can help to easily run any model in claude, Codex, Gemini, Pi, and OpenCode. It's also an API keys manager, supports multiple providers or OpenAI/Claude/Gemini accounts.

  3. Hello Guys, Each time a new local llm is released, there are a ton of new posts , this is it, it's near Opus level...., the abliteration matrix final something at Q2 KXLDND is the best but it's been a day that deepseek was released and i d…

← all threads