Qwen3.6-Plus
Blog post: [https://qwen.ai/blog?id=qwen3.6](https://qwen.ai/blog?id=qwen3.6) From Chujie Zheng on 𝕏: [https://x.com/ChujieZheng/status/2039560126047359394](https://x.com/ChujieZheng/status/2039560126047359394)
They solved AI’s memory problem!
Researchers at the Kimi Team have developed a groundbreaking new AI architecture named Attention Residuals, which is capable of solving the fundamental problem of "AI amnesia" in deep neural networks. While previous large language models have achieved remarkable performance by stacking hundreds of p
The Bonsai 1-bit models are very good
Hey everyone, Tim from [AnythingLLM](https://github.com/Mintplex-Labs/anything-llm/issues) and yesterday I saw the [PrismML Bonsai](https://prismml.com/news/bonsai-8b) post so i had to give it a real shot because 14x smaller models (in size and memory) would actually be a huge game changer for Loca
Claude Code Leak Reveals Always-On ‘Kairos’ Agent
After Anthropic released Claude Code's 2.1.88 update, users quickly discovered that it contained a package with a source map file containing its TypeScript codebase, with one person on X calling attention to the leak and posting a file containing the code. The leaked data reportedly contains more th
Ship Code Faster with Claude Code on Vertex AI
Ship Code Faster with Claude Code on Vertex AI Anthropic
Here's what that Claude Code source leak reveals about Anthropic's plans
A persistent agent, stealth "Undercover" mode, and... a virtual assistant named Buddy?
r/programming bans all discussion of LLM programming
Chinese state media releases episode 2 of their AI generated Iran war animated series
Anthropic took down thousands of GitHub repos trying to yank its leaked source code — a move the company says was an accident
Anthropic executives said it was an accident and retracted the bulk of the takedown notices.
Gemini 4 is coming ??
AI for American-produced cement and concrete
Show HN: Running LLM on smartwatch – found llama.cpp loading model twice in RAM
Gemma time! What are your wishes ?
Gamma 4 drops most likely tomorrow! what will it take to make it a good release for you?
Solar Balconies Take Europe by Storm
ZomboCom stolen by a hacker, sold, now replaced with AI-generated makeover
TurboQuant isn’t just for KV: Qwen3.5-27B at near-Q4_0 quality, about 10% smaller, and finally fitting on my 16GB 5060 Ti
I bought an RTX 5060 Ti 16GB around Christmas and had one goal: get a strong model running locally on my card without paying api fees. I have been testing local ai with open claw. I did not come into this with a quantization background. I only learned about llama, lmstudio and ollama two months ago
Demis Hassabis secretly built a hedge fund inside DeepMind trying to beat Jim Simons. Google shut it down.
Obfuscation is not security – AI can deobfuscate any minified JavaScript code
64Gb ram mac falls right into the local llm dead zone
So I recently bought a Mac (m2 max) with local llm use in mind and I did my research and everywhere everyone was saying go for the larger ram option or I will regret it later... So I did. Time to choose a model: "Okay, - Nice model, Qwen3.5 35b a3b running 8 bit quant, speedy even with full contex