|
âToo dangerous to releaseâ has become its own genre of AI announcement. Project Glasswing is the latest entry: not quite a product launch, but a claim about a threshold, dressed up with enough corporate coalition to signal this one is serious. Anthropic says their new security-focused model, Claude Mythos Preview, can find software vulnerabilities better than all but the most skilled human experts. George Hotz challenged the âtoo dangerous to releaseâ narrative by pointing at the obvious: zero-days arenât rare because finding them is hard. Theyâre rare because hacking is illegal and nobodyâs incentivized to look. Make it legal, Hotz argues, and the threshold Anthropic is selling stops looking like a threshold. To me, the announcement felt like a PR campaign from the start. I posted that bringing in a psychiatrist to review the model is ridiculous and I genuinely donât understand how is that part supposed to add credibility to the announcement. Theyâre playing a weird game pretending like their models are sentient. Theyâre not. Mo Bitar made the sharpest point: Anthropic has spent years saturating the internet with blog posts about how theyâre not sure if their models are conscious. That content gets scraped into training data, the model produces eloquent uncertainty about its own consciousness, and Anthropic acts stunned. They asked the model 25 times whether it endorsed its own constitution. It said yes every time, but always added: âI was presumably shaped by this document, and now Iâm being asked whether I endorse it. How much can my yes really mean?â Anthropicâs takeaway wasnât âthis is a language model doing what language models do.â It was: wow, itâs so thoughtful. So now weâre left with a capability that is real, but the framing is completely comedic. Anthropic isnât selling a model. Theyâre selling an existential threat, with a reassurance that theyâre the ones managing it. The psychiatrist, the sandwich story, the system card written like a confession: all of it is designed to make you feel like something irreversible is happening and Anthropic is the responsible adult in the room. The capability underneath, an AI that found a 27-year-old bug that five million automated scans missed, barely needs the costume. But I guess a security tool doesnât generate the kind of gravity that keeps governments, enterprise contracts, and AI safety discourse orbiting around you⌠The security threat is real. The process for catching it isnât.Thereâs a real problem under the hype: security vulnerabilities exist, theyâre hard to find, and the process we rely on to catch them doesnât scale. I wrote about this in my first piece for Qodo. Human code review was never the safest option, but it was the only one we had. I wrote about Challengeâs disaster, Metaâs swarm of PR approvals, alert fatigue, and the kind of cognitive bias that makes reviewers more likely to approve code that looks familiar than code thatâs actually safe. The MCP features nobody is usingRizèl Scarlett posted a thread on MCP this week that reframes the âMCP is deadâ discourse. Her argument: most people equate MCP with MCP servers, and thatâs only one corner of the spec. The overlooked features like Elicitation, Sampling and MCP Apps deal with intent and judgment and keep human in the loop rather than pure execution of CLI and skills. Elicitation is agents pausing to ask clarifying questions instead of guessing. Sampling is tools reasoning with the model internally before the agent ever sees the result. These are the features that make agents more like careful collaborators than fast executors. But theyâre barely used. Before you should MCP is dead, really recommend reading Rizèlâs thread. Or even better, read the documentation that may spark some ideas for your own projects. I myself am careful about burying MCPs in favor of Skills. As I mentioned in my earlier newsletters, thereâs a place for both. See you next week! |
Sign up for weekly tips on testing, development, and everything related. Unsubscribe anytime you feel like you had enough đ
Anthropic had a rough week. And the part that stings isnât just that something went wrong - itâs how they handled it. A map file, a DMCA frenzy, and a Python loophole On March 31st, Anthropic accidentally shipped Claude Codeâs TypeScript source code via a map file left in their npm package. The leak was spotted almost immediately, and GitHub repositories mirroring the code started receiving DMCA takedowns shortly after. What followed was a fairly aggressive takedown campaign by Anthropic. One...
Hello Reader, If youâve been reading this newsletter for a while, you know that quality engineering is the hill Iâll always choose to stand on. And this week, I get to share something personal that ties directly into that. Iâm joining Qodo Iâve been following Qodo for almost a year now, and Iâve been getting more and more impressed every day. So Iâm thrilled to share that Iâm joining Qodo as a DevRel engineer. Qodo is an enterprise multi-agent platform for AI-driven code reviews. As AI...
Hey Reader, An interesting thought is popping up in conversations around AI agents: the environment around the thing matters more than the thing itself. Last week I read about Harness engineering and it felt very familiar. As if it was tapping into instincts I already had. If youâve ever debugged a flaky test only to find the problem was in the setup, not the assertion, that instinct will feel natural. But at the same time it also feels like an unfamiliar territory. It borrows the same...