Andrej Karpathy: From Vibe Coding to Agentic Engineering
Andrej Karpathy (co-founder of OpenAI, former head of AI at Tesla, and now founder of Eureka Labs) talks with Sequoia partner Stephanie Zhan at AI Ascent 2026 about what's changed in the year since he coined "vibe coding." He explains why he's never felt more behind as a programmer, why agentic engi
Richard Dawkins and The Claude Delusion
dawkins dropped a piece on unherd yesterday declaring claude conscious after 3 days of talking to it. he calls his instance "claudia". fed it a chunk of the novel he's writing, got eloquent feedback, and wrote: "you may not know you are conscious, but you bloody well are!" i had to read that twice
Richard Dawkins Chats with Claude and Thinks it's Conscious
Thought I'd leave this here since nobody else has done so yet. My personal thoughts? LLMs like to please. The RLFH gets a bit "drifty" and "hallucinatory" after long discussions. It also renders what you want to hear if you don't keep the discussion on a disciplined path. I'd need to see Richar
Let's Buy Spirit Air
Open source models are going to be the future on Cursor, OpenCode etc.
I just wanted to share my experience. At work we have Cursor with the Enterprise tier. Today I burned 10$ with 2 prompts, one on gpt-5.5 and one on claude-opus-4.6-thinking. Last month I burned 80$ in one week with claude-opus-4.7 even with the 50% off they had with the launch. If they continue with
Ilya Sutskever: Accurately predicting the next word leads to real understanding
Source: [https://x.com/vitrupo/status/2050736968041210316](https://x.com/vitrupo/status/2050736968041210316)
One bash permission slipped...
How? It kept getting chained bash commands wrong, with wrong escapes. So it created many bad directories, and tried "fixing" its mistake. It offered to run a large bash command, with `rm -rf` inside, and stupid me missed it. I'm glad I push everything often. But the disruption is massive. FAQ: -
AMD Strix Halo refresh with 192gb!
Looks like the next strix halo, the Gorgon halo 495 max will have more then 128gb! I already bought a strix halo mini forms couple months ago since the 2026 refesh rumors was not interesting. Was not planning on getting another till 2027 with the bigger refresh, and linking them together. But was pl
Got this SVG from A/B test window inside AI Studio. Still can't believe this is an SVG. Most likely the new flash/pro model.
A very basic litmus test for LLMs "ok give me a python program that reads my c: and put names and folders in a sorted list from biggest to small"
Then ask your cloud FOTM api to verify the code it spit. I thought it was an easy question, but my local ones just died on it, with wrong executions, double-reading the sizes of files, putting recursive functions inside recursive functions. I think I got my magic test.
Llama.cpp quantization is broken
Main reason is, that qunatization quality directly affects models performance and stability and this results in real usefullness. Even though GRM-2.6-Plus is in benchmarks better than qwen3.6 27b model from which it derives, it gives worse results than autoround Q2\_K\_mixed quant of qwen3.6 27b whi
Rule suggestion: links to "I made this website" with full disclosure, so we can avoid AI slop.
There's a bunch of posts where people promote their sites related with local LLMs, specially sites for benchmarks. This post for example [https://www.reddit.com/r/LocalLLaMA/comments/1t1m5mn/comment/ojl1vl2/?context=3](https://www.reddit.com/r/LocalLLaMA/comments/1t1m5mn/comment/ojl1vl2/?context=
Friendlier LLMs tell users what they want to hear – even when it is wrong
Are modern ML PhDs becoming too incremental, or is this just what research looks like now? [D]
I’ve been thinking about the current state of machine learning PhDs, including my own work, and I’d like to hear how others see it. My impression is that a large fraction of modern ML PhD work follows a fairly predictable pattern: take an existing idea, connect it to another existing idea, apply i
How much will it cost to host something like qwen3.6 35b a3b in a cloud?
I keep hearing the model is good, I don't have the hardware for it, and I will wait to the end of the year for the hardware to evolve. But, I still need coding, people are saying qwen3.6 35b a3b is good, so the question is now how much will it cost me to host it somwhere until I get new hardware.
What if ChatGPT launched in 1998
The Oscars just banned AI from winning acting and writing awards
Control AI Risk with Pre-Built Frameworks and Ready-to-Run Evaluations
I gave my local LLM a "suffering" meter, and now it won’t stop self-modifying to fix its own stress.
Yesterday I posted about my Agent OS (Hollow) building its own tools. Today, I want to talk about *why* it does it. Most agents sit idle until you prompt them. I wanted something that felt "alive," so I built a **Psychological Stressor Layer**. Each agent has a "suffering" state that worsens over t