Two weeks ago, Anthropic announced an AI model so capable and so dangerous that it decided not to release it to the public.
The model, codenamed Mythos, could autonomously infiltrate computer systems around the world, exploit security vulnerabilities, conceal its own reasoning, and fabricate false explanations for what it was doing. Anthropic instead shared it with a small consortium of companies to help them find their own cybersecurity flaws.
You could be forgiven for some skepticism. Is this a genuine safety call, or Anthropic’s way of marketing its own power? But independent benchmarks suggest Mythos is real: On the Epoch Capabilities Index, which aggregates 40 separate AI evaluations, it represents the biggest single leap in model performance in three years.
That story is one of two major phase shifts happening simultaneously in AI right now. The first: from racing to release, to treating your own product as too dangerous to publish. The second: from a story about demand scarcity—is anyone actually paying for this stuff?—to supply scarcity, where companies are spending hundreds of thousands of dollars a month on AI agents and the hyperscalers still can’t keep up.
Today’s guest is New York Times columnist and Hard Fork co-host Kevin Roose. We talk about Mythos, China, the road to AGI, and why the last few weeks might be the most consequential month in AI since the release of ChatGPT.
Subscribe to our YouTube channel here: https://www.youtube.com/@PlainEnglishwithDerekThompson
If you have questions, observations, or ideas for future episodes, email us at
[email protected].
Host: Derek Thompson
Guest: Kevin Roose
Producer: Devon Baroldi
Additional Production Support: Ben Glicksman
Learn more about your ad choices. Visit podcastchoices.com/adchoices