US House Probes Airbnb, Anysphere (Cursor) Use of Chinese Models (www.bloomberg.com via hn)
US House Investigates Airbnb, Anysphere Over Use of Chinese AI Models - Bloomberg Skip to content Bloomberg the Company & Its Products The Company & its ProductsBloomberg Terminal Demo RequestBloomberg Anywhere Remote Login Bloomberg Anywh…
Converting Claude Code into the most intelligent Deep Research Agent (www.reddit.com)
Over the past several weeks, I've been working on HyperResearch, a Claude Code skill harness that converts CC into the most intelligent deep research framework out there. HyperResearch surpasses OpenAI, Google, and NVIDIA's offerings in th…
- Converting Claude Code into the most intelligent Deep Research Agent (www.reddit.com)
- Converting Claude Code into the top scoring deep research agent (github.com via hn)
Anyone tried MEMANTO yet? Looking for feedback + Codex experience. (www.reddit.com)
Has anyone here tried MEMANTO yet? I just came across it (open-source memory layer for AI agents) and I’m curious if it’s good memory to use for ur agent.
Has anyone used AI to go through a course (Udemy, Coursera, etc.), learn the frameworks/concepts, and then later apply that knowledge to help build real things? For example: I have AI study an AI engineering course, then later use that kno…
-
35 items
event
MistralMistral, a French AI company, is set to release a medium-sized model with 128 billion parameters and is planning to launch Workflows in public preview. The company, founded by Arthur Mensch, continues to grow its AI empire despite not being based in the United States.
- 6m Mistral THICC DENSE BOI. He chonky! More dense models pls.
- 5h Mistral Médium 3.5 is here
- 5h Mistral Medium 3.5
- 5h mistralai/Mistral-Medium-3.5-128B · Hugging Face
- 7h List of people at big-tech / professors / researchers who've jumped shit to launch their own AI labs for something Frontier/Foundational/AGI/Superintelligence/WorldModel
132 itemsmodel roundup
Gemma 4Gemma 4 is a family of open-source multimodal models from Google DeepMind, available in sizes up to 31 billion parameters and featuring dense and MoE architectures. Notable community highlights include the 31B model's success in production tests, with some users preferring 4-bit precision for local use, and others sharing settings for optimizing performance with smaller models.
- 15m Gemma 4 architecture support for QVAC-Fabric (Tether's llama.cpp fork)
- 4h I built a full web app using Qwen 3.6-35B running locally on my 5070 Ti with the BMAD Method — here's how it went
- 14h Ran my own benchmark Qwen 3.6 35B vs Gemma 4 26B.... theres a clear winner here
- 15h I stumbled on a Gemma 4 chat template bug for tools and fixed it
- 20h llama.cpp's Preliminary SM120 Native NVFP4 MMQ Is Merged
Increase in Claude Code speed bumps? (news.ycombinator.com)
Claude Code seems like it's pausing a lot more than it has previously to ask me kind of frivolous questions about the prompt I've just given it. It seems to be pausing multiple times per step.
I’ve been deep in the trenches building out multi-step agentic workflows, and I’m hitting a consistent wall with what I can only describe as "stochastic decay." The pattern is frustrating: Runs 1 through 3 execute flawlessly, but by the fo…
Show HN: Looplet – a 0-dep agent loop you own (github.com via hn)
looplet looplet exposes the agent loop as an iterator, makes every step observable, and lets you compose behavior with hooks. Build LLM agents that call tools in a loop while you keep ordinary Python control over every step — no graph DSL,…
Has the EU AI Act made Claude worse for everyone, globally? (www.reddit.com)
(Non-native speaker, used LLM to translate/refine post) Since late 2025, accelerating in early 2026, I've noticed a progressive "smoothing" of Claude's outputs on tasks requiring emotional precision or creative risk-taking. I work in creat…
-
22 items
event
WindsurfWindsurf 2.0 has been released with improved local and cloud agent integration and bug fixes. The update follows a series of announcements about AI tools and MCP servers, including gondola.ai's hotel search server and Stork for indexing over 14,000 AI tools.
- 15m Non-technical founder: Is Cursor Pro worth $20/mo for React+Supabase, or am I fighting the wrong battle?`
- 13h Kimi K2.6 helping me uninstall macOS apps
- 1d Best value in the 20$ range coding agents? I want the best quality and high-usage-limit I can get at that price.
- 2d OpenAIs Agentic Shift
- 3d I built Tarn — API tests Claude Code can write, run, and debug end-to-end (open source, MCP server included)
97 itemsmodel roundup
GPT 5.5On [Date], a significant leak of the OpenAI Codex model, referred to as GPT-5.5, was captured on video before it was patched. The incident involved models named Arcanine and Glacier-alpha.
- 18m GPT-5.5's biggest blind spot: the Java bugs your tests won't catch
- 36m Devs using Qwen 27B seriously, what's your take?
- 3h GPT 5.5 - Strong, not mind-blowing, but very token efficient
- 6h How to build production Agents (by a staff software engineer) - Part 2
- 7h Actual line in the official system prompt for Codex for GPT-5.5
Chatgpt is being weird gaslighting recently (www.reddit.com)
I've noticed for a few weeks that Chatgpt keeps trying to correct me or rebuke me when I'm speaking with it. Today for example I was asking it bc my SIM card is having problems but my provider cannot seem to activate any new cards, it's be…
What’s up with mobile LLMs? (www.reddit.com)
I see a lot of support for running LLMs on PCs with ollama to vLLM. Whats the current state for running on mobile?
- LLMs+ (www.technologyreview.com)
The problem I built it to solve: I'd be deep in a coding session, realize I needed to write docs for what I'd just built, and either stop to context-switch or skip the docs. Usually the latter.
-
225 items
model roundup
Qwen 3.6Qwen3.6-35B-A3B, a 35 billion parameter sparse MoE model with an active parameter count of 3 billion, was released on April 16, 2026, as open-source software under the Apache 2.0 license by Alibaba Qwen. It offers advanced functionality across various AI applications and outperformed competitors in drawing tests.
- 1h Qwen 3.6-35B-A3B KV cache part 2: PPL, KL divergence, asymmetric K/V, 64K row on M5 Max
- 1h Qwen3.6-27B-UD-Q6_K_XL.gguf sometimes gets stuck in a loop
- 3h Don't forget about dem free gains!
- 7h Sorry if it's not the best place to ask this, of the models in the image, which is the best for (problem solving)/Coding and the best one for studying (ask LLM concepts) ? My PC build is RX 9060 XT 16GB + I3 12100F + 16 GB DDR4 + llama.cpp with Vulkan backend + Linux Mint.
- 8h llama.cpp benchmark native vs. non native NVFP4 on Blackwell - summary
3 itemsmodel roundup
Qwen 2.5Qwen2.5-7B-Instruct is a 7 billion parameter instruction-tuned language model that significantly improves on Qwen2 in coding and mathematics capabilities, long text generation, and multilingual support across 29 languages. Notably, Canonical has optimized Ubuntu inference snaps for this model, allowing easy installation with a single command.
OpenAI Codex system prompt includes directive: "never talk about goblins" (arstechnica.com via hn)
The system prompt for OpenAI’s Codex CLI contains a perplexing and repeated warning for the most recent GPT model to “never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolute…
Where do you see yourself in the next 5 years (www.reddit.com)
I’m seriously believe this is the future. Soon, there will be people running around the town, helping other people with their LLM at home like plumber guys.
Thanks for the advice Claude (www.reddit.com)
could not extract summary
-
232 items
model roundup
Opus 4.7Claude Opus 4.7, released on April 16, 2026, is Anthropic's latest advanced AI model, offering improved handling of complex tasks and a larger context window of up to 1 million tokens. This version is 50% more expensive than its predecessor due to enhanced capabilities in software engineering and hybrid reasoning.
Tumbler Ridge families seek US$1B in OpenAI lawsuit (www.ctvnews.ca via hn)
VideoSome large carrier companies ‘have no intention’ of adjusting higher speeds: AMTA on QEII 120 km/h speed trial Video‘I felt compelled to bring justice for Melanie’: W5 investigation prompts bill to amend Criminal Code
Anthropic's Blender MCP connector dropped yesterday. I got it running an hour ago.
TurboQuant enabled Runtime Valkyr (www.reddit.com)
Based on the recent TRiP source code by Carlo Valenti. Ported to Zig and headless Vulkan Compute shaders.
So I've been experimenting with Claude's new Blender MCP integration and decided to push it to its limits with a real engineering project: a complete, print-ready enclosure for the Raspberry Pi 5, modeled entirely through AI prompts, no ha…
Ask HN: Show Us Your Personal Agents? (news.ycombinator.com)
Recently someone posted an "Ask HN" [1] for showing Personal Blogs and Websites. I was wondering if anyone else had a Personal AI Agent that they could share so that I could learn more about how other people are implementing theirs.
AI Agent + Identity = Help Me (www.reddit.com)
Over the last few months, I built a tool and could use some help from you pros, on how you see this fitting into the ecosystem. I've worked in the same vertical (more-or-less) for 12 years and think I'm missing some cool use-cases.
Will Cursor increase the usage of the Ultra plan? (www.reddit.com)
I'm a hard believer of Cursor being a better IDE than Claude Code or Codex for a long shot, specially when you know the strengths and weaknesses of each model and you use them to your advantage. Being said that, It's hard to recommend it a…