1. Find gaps in your AI systems before users (or attackers) do.

  2. Run more RL experiments. Wait less for GPUs.

  3. Been doing a proper testing round on different image-to-video models to figure out which one produces the most convincing results for avatar-style UGC ads. Outputs vary more than I expected, not just in quality but in how avatars move, how…

  4. I switched from Claude Code to OpenCode a few weeks ago and realized most of Anthropic's official Claude Code plugins don't transfer directly. The reason is that those plugins put their value in `commands/` and `agents/`, both of which are…

  5. Every session with Claude Code starts cold. The agent is capable but it only knows what's in context right now — decisions from last week are gone, the tracking file says one thing, git says another, and the memory note explaining the pivo…

  6. FrontierSWE FrontierSWE is an effort to test coding agents on the hardest ultra-long horizon technical challenges. Together with partners from academia and industry, we have collected real-world problems from domains including performance…

  7. event

    Security
    86 items

    OpenAI has released GPT-5.4-Cyber for testing as part of its Trusted Access for Cyber Defense program, aiming to compete with Anthropic's Claude Mythos in the cybersecurity domain. Meanwhile, concerns are rising over the potential risks associated with advanced AI models like Mythos, prompting calls for improved defenses before wider releases.

    event

    Copilot
    102 items

    Microsoft is keeping its Copilot tool for Windows 11 but renaming it, while issues with rate limits and a security proxy have sparked concerns among users of GitHub Copilot. Meanwhile, Anthropic released a report on agentic coding trends, highlighting that developers use AI in about 60% of their work.

  8. You are now my Prompt Engineering Mentor. Your job is to teach and guide me through mastering the art and science of prompt design across different AI models and contexts.

  9. I previously built an app on top of ChatGPT's Assistant API to answer questions about life events. My wife and I are constantly asking each other "was that before or after we moved into this house?" Life events are things like going on vac…

  10. Learning Opportunities: A Claude Code Skill for Deliberate Skill Development Build your expertise, not just your projects. This skill uses an adaptive "dynamic textbook" approach to help you integrate science-based expertise building exerc…

  11. Context window "brute-forcing" is expensive and slow. I built a tool called PandaFilter to solve this at the source.

  12. I did a deep dive on Claude Design and below are my thoughts. What it does extremely well: Improves your prompt - similar to "ask me questions" when chatting to an LLM.

  13. Am I missing something? I am logged in to both web and desktop using the same credentials.

  14. model roundup

    Qwen 2.5
    25 items

    Qwen2.5-7B-Instruct is a 7 billion parameter instruction-tuned language model that significantly enhances coding and mathematical capabilities, supports up to 128K tokens in context, and understands structured data. Community discussions highlight its suitability for code autocomplete tasks and debate the hardware requirements needed for deployment compared to other models like Gemma 26B MoE.

    model roundup

    Opus 4.7
    200 items

    Claude Opus 4.7, released on April 16, 2026, is Anthropic's latest advanced AI model, offering improved handling of complex tasks and a larger context window of up to 1 million tokens. This version is 50% more expensive than its predecessor due to enhanced capabilities in software engineering and hybrid reasoning.

  15. LinkedIn Campaign Manager MCP MCP server for the LinkedIn Marketing API — query campaigns, performance, and Lead Gen Forms from Claude in plain English. 19 read-only tools covering ad accounts, campaigns, creatives, performance analytics,…

  16. We’ve invited executives from across Automattic to share their perspective on leadership, open source, and the future of the open web. The latest comes from James Grierson, our head of global expansion, who shared his thoughts on the WordP…

  17. WaveletLM is a wavelet-based, attention-free architecture that replaces self-attention with learned lifting wavelet decomposition, a Fast Walsh-Hadamard Transform, per-scale gated spectral mixing with SwiGLU activation, an inverse FWHT, an…

  18. **TL;DR:** Tomorrow I have a two-hour remote pair-programming interview where I drive a from-scratch project while leaning heavily on AI assistance, narrating my reasoning, demonstrating best practices and showing how I handle rein-in mode…

  19. heylo! open sourcing a plugin that I've been daily driving for a month.

  20. model roundup

    Qwen 3.6
    177 items

    Qwen3.6-35B-A3B, a 35 billion parameter sparse MoE model with an active parameter count of 3 billion, was released on April 16, 2026, as open-source software under the Apache 2.0 license by Alibaba Qwen. It offers advanced functionality across various AI applications and outperformed competitors in drawing tests.

    event

    Cowork
    104 items

    Issues with Claude Cowork have been reported, including errors and disruptions for some users on April 16, 2026. Additionally, Google has developed its own desktop Agent to compete with Cowork, while users continue to explore alternatives and troubleshoot bugs in the platform.

  21. Claude can now run your entire workflow for you like an executive assistant. Here are 7 AI agents that organize your emails, tasks, and calendar, and save you hours every week.

  22. A 30-hour timeline of how Cursor's agent, Railway's API, and an industry that markets AI safety faster than it ships it took down a small business serving rental companies across the country. I'm Jer Crane, founder of PocketOS.

  23. From text prompt to pixels, one component at a time. 12 min read Apr 12, 2026 -- You give it a text prompt.

  24. Get responses tailored to you Log in to get answers based on saved chats, plus create images and upload files.

  25. I’m using both Claude Code and Claude in ChatApp for different parts of the same project: Claude Code for implementation and the chat app for concept or prompt work. The constant copy-pasting between them is annoying.

  26. In my local LLM setup I get from 30 to 80 t/s generation at the beginning, but it drops quite a lot as context grows. I use llama.cpp/Vulkan with an MI50 and a V100, is there some command line flags that can improve this issue?