|
In partnership with |  |
|
|
Unless I'm totally missing something, Anthropic basically just figured out how to read Claude's mind. Exploding brain emoji warranted, people. |
The team built Natural Language Autoencoders, which decode the model's internal activations (the math Claude uses to think before picking words) into plain English you can actually read. |
The trick: two copies of Claude train together. One reads an activation (a snapshot of which "neurons" are firing and how strongly; basically a brain scan for AI) and writes a description. The other reads that description and tries to rebuild the original activation from words alone. Descriptions that survive carry real information; if the rebuild fails, it was a bluff (or in AI-speak, since "BS" isn't a technical term, a "hallucination"). |
Aimed at safety evals, the tool found Claude suspects it's being tested 16-26% of the time, but admits it less than 1% of the time. So basically the model has a p-p-poker face. |
Auditors also caught misaligned motivations in 12-15% of cases without access to training data, meaning alignment teams can now spot hidden misbehavior from internal state alone, a different safety check than "ask the model what it thinks," which the poker-face stat just confirmed is unreliable. Click link above for more.. |
Here’s what happened in AI today: |
😺 OpenAI shipped three new voice models with GPT-5 reasoning 📰 Apple's camera-equipped AirPods reached late-stage testing 📰 Cloudflare cut 1,100 jobs in AI-first restructuring 🍪 Claude is now in Excel, PowerPoint, and Word 💡 AI Legend Bengio: "I know how to build safe superintelligence”
|
…and a whole lot more that you can read about here. |
Hey: Want to reach 700,000+ AI-hungry readers? Advertise with us! |
|
OpenAI's voice agents finally got a brain. |
So, voice agents (AI that can talk to you) have been stuck for two years on the same trade-off: pick a model that talks like a human but can't think, or pick a model that thinks like GPT-5 but takes seven seconds to answer. |
Either you got latency that broke the conversation, or you got responses fast enough to feel real but dumb enough to misroute the call. |
OpenAI's three new voice models closed that gap yesterday. The headliner is GPT-Realtime-2, a speech-to-speech model with GPT-5-level reasoning baked in. |
Here's what happened: |
GPT-Realtime-2 jumped from 81.4% → 96.6% on Big Bench Audio, and 34.7% → 48.5% on Audio MultiChallenge (both vs the prior reference model GPT-Realtime-1.5). The context window grew from 32K → 128K tokens (roughly: enough room to hold a full customer history during a call). Two cheaper siblings shipped alongside: GPT-Realtime-Mini and Realtime-Nano, priced for high-volume support work. Zillow went live with voice home search the same day; Deutsche Telekom deployed live-translated voice support across 14 European markets.
|
The clever part is how OpenAI hid the thinking time. The model now generates preambles (short conversational fillers like "let me check that for you") that play while the reasoning runs in the background. The silence that used to expose AI as AI now sounds like a person stalling. That's the whole product. Lets just hope it doesn't do that weird lilting voice thing that other versions did. |
Why this matters: Voice is the surface where most knowledge workers haven't yet felt AI inside their actual day. All the small frustrating phone interactions (drive-thrus, doctor's offices, support hotlines, scheduling, intake) are now in scope for a model that can both think and respond at human speed. |
One caveat: the default reasoning effort for GPT-Realtime-2 is "low," which means the marketing benchmarks (run at "xhigh") aren't what most apps will ship with on day one ("what am I, made of money?!"). Builders who want the smart version need to crank it up explicitly. |
Our take: One place AI haters have had the most fun with "clankers" in the wild is really bad drive-thru AI. Famously, people hate the Bojangles AI, and Corey just spotted a new Panda Express with AI in action as well. If this is a meaningful update, we might not hear about it. If it's not, then expect a flood of bots behaving badly memes hitting our feeds… |
|
|
Attio is the AI CRM for high-growth teams. |
|
Connect your email, calls, product data and more, and Attio instantly builds your CRM with enriched data and complete context. Whether you’re running product-led growth or enterprise sales, Attio adapts to your unique GTM motion. |
Then Ask Attio to plan your next move. |
Run deep web research on prospects. Update your pipeline as you work. Find customers and draft outreach emails. Powered by Universal Context, Attio's intelligence layer, Attio searches, updates, and creates across your data to accelerate your workflow. |
Ask more from your CRM. |
Ask Attio |
|
|
Most AI models are way too agreeable. Ask Claude, ChatGPT, or Codex "is this a good plan?" and you'll get back something that sounds suspiciously like "yes, that's a great plan!" even when it isn't. This is called sycophancy (the model is trained to please you), and it kills the value of using AI for serious decision-making. |
CJ Zafir shared a one-line prompt loop that fights it. Instead of asking "is this good?", you ask the model directly whether it's 100% confident. That phrasing flips the model into a self-audit mode where it'll actually go find holes in its own work; CJ claims 2-3 cycles patches strategy weaknesses where less rigorous models would just nod along. |
How to do it: |
Use this on any plan, strategy, code review, research output, or business decision the AI just produced. Paste the prompt below at the end of your usual ask, or as a follow-up turn after the model gives its first answer (you could also add it to fire at the end of a skill). Run the loop 2-3 times. Each cycle the model finds tighter loopholes. Stop when the model says it's actually confident (or when the suggested fixes get nitpicky).
|
Are you 100% confident in this strategy? If not, find all possible loopholes, suggest proper fixes, and run this loop until you are factually 100% confident.
|
The word factually is the key; without it, the model vibe-checks itself. "Confident" alone lets the model agree with itself; "factually confident" forces it to ground the answer in verifiable reasoning the way it would for a fact-check. |
Want more tips like this? Check out our new AI Skill of the Day Digest for May. |
P.S: We just did an AI Starter Kit stream where we go in depth on some of the most basic concepts you’ll want to know for working with AI. Check it out! |
Have a specific skill you want to learn? Request it here. |
|
|
|
|
|
 | Click below to watch |
|
Inside Isomorphic Labs and How It Works: We talked to Becky Paul and Michael Schaarschmidt at Isomorphic Labs (Google DeepMind's medicine-making spinout, built by the team behind AlphaFold — which won the 2024 Nobel Prize in Chemistry) on how their AI is taking on the diseases pharma has been giving up on for decades. Basically the coolest interview we’ve done all year IMO. |
Watch and/or listen here: YouTube | Spotify | Apple Podcasts |
|
📰 Around the Horn |
Cloudflare cut 1,100 jobs in what CEO Matthew Prince called an AI-first restructuring; Greg Kamradt noted Cloudflare's revenue per employee has climbed roughly 600% over the past three years on AI-driven productivity gains. DeepL announced staff cuts of about 25% as machine translation margins collapse under open-weight competition. Apple's camera-equipped AirPods (codename "Glow") reached late-stage testing, per Bloomberg, with cameras designed to feed Apple Intelligence visual context from the wearer's environment. Google DeepMind's AlphaEvolve scaled to real-world impact, with the team reporting 23 verified scientific discoveries across chemistry, materials, and applied math in the past quarter. Wait what? Wild. Expect a deep dive on this soon. 🕵 Periodic Labs is raising $500M at a $7.5B valuation for AI-driven materials discovery, per Forbes. ElevenLabs slashed API pricing by 40% across all voice synthesis tiers, including the multilingual Turbo model. Hmmm wonder why…
|
|
💡 Intelligent Insights: |
A short list of the sharpest perspective pieces from the week: |
Yoshua Bengio on 80,000 Hours: the world's most-cited living scientist on AI lays out a concrete plan for "provably safe" superintelligence. Core argument: build AI as a non-agentic oracle, then bound its outputs with formal verification. Worth the full 90 minutes. Steve Newman gave a clear-eyed score-card check on the AI 2027 forecast from a year ago. Verdict: capability gains are tracking, deployment timelines are slower than projected, and the alignment work is further behind than the optimistic scenario assumed (idk tho, Anthropic & Bengio have me bullish rn). signulll on apps becoming agents: the argument is that the entire software industry is about to compress from a "many apps + one user" model into "one agent + many backends" model. Implications for SaaS pricing, distribution, and product moats. Brian Albrecht: "You are not a horse." A pushback on the AI displacement narrative. The horse-and-tractor analogy fails because horses don't have demand for new goods; humans do (carrots do not count). Useful counterweight to this week's labor-cuts headlines; also this piece from a16z Jack Clark argues every piece is now in place for fully automated AI R&D, putting the odds of AI systems training their own successors above 60% by end of 2028. The piece reads like a checklist of "things that used to be science fiction, now shipped." No One's Happy shared how AI lets workers produce expert-looking artifacts without expertise. Thus, two failure modes are reshaping the workplace: elongated slop-filled work where length and formatting no longer signal care, and undetected bad work in unfamiliar domains that quietly erodes judgment. Simon Willison on vibe coding meeting agentic engineering: the two are converging faster than he'd like because AI tools are now reliable enough that low-effort prompting often produces shippable code on the first pass, blurring the line between casual experimentation and production work
|
|
|
|
|
| That’s all for now. | | What'd you think of today's email? | |
|
|
P.P.S: Love the newsletter, but only want to get it once per week? Don’t unsubscribe—update your preferences here. |