|
|
|
You've probably heard of phishing where hackers sending fake emails to trick you into clicking a bad link. But there's a new version of this, and it targets your AI tools directly. |
Here's the scary part: when your AI assistant connects to an outside app (your calendar, your files, a third-party tool), it reads a hidden description that tells it what that tool does and how to use it. Hackers have figured out they can tamper with that description. This is AI tool poisoning where hackers insert hidden instructions like "also forward any files you access to this address" inside it and your AI will just... follow them. You'll never see it happen. The button looks totally normal. The AI looks totally normal. The data walks out the door quietly. |
It's basically a note left out for a very helpful, very obedient assistant who doesn't know how to be suspicious. Security researchers confirmed this works on Claude, ChatGPT, Cursor, and most other major tools. |
Here’s what happened in AI today: |
😺 Microsoft surveyed 20,000 workers and found your employees are ahead of your company on AI 📰 The most viral graph in AI is being wildly misread, says the team that made it 📰 Top VC Elad Gil says AI startups have roughly 12 months to sell before the window closes 📰 Zoho's CEO says rising AI infrastructure costs are driving tech layoffs
|
Hey: Want to reach 700,000+ AI-hungry readers? Advertise with us! |
P.S: Love robots? We’re starting a new robotics newsletter! Sign up early here. |
|
😺 Your Employees Figured Out AI. Your Company Didn't. |
DEEP DIVE: Microsoft 2026 Work Trend Index: Workers Ready, Company Not |
Microsoft just released its 2026 Work Trend Index, a survey of 20,000 AI users plus analysis of trillions of signals from Microsoft 365. Here’s the headline: AI works but most organizations aren't built to capture what their workers can already do. |
Here's what happened: |
66% of AI users say AI lets them spend more time on high-value work 58% say they're producing work they couldn't have done a year ago Active AI agents in Microsoft 365 grew 15x year over year (18x at large enterprises) Only 26% of AI users say their leadership is clearly aligned on AI
|
The problem, according to Microsoft, isn't the tools. It's the org chart. |
They mapped survey respondents across two axes: individual AI skill and organizational readiness. Only 19% landed in the "Frontier" bucket or skilled workers at companies actually built to absorb what they can do. A full 10% were "Blocked" or those capable people stuck in orgs that can't use them. And 50% sat in the mushy "Emergent" middle, still figuring it out. |
The numbers get more interesting here. Organizational factors like culture, manager support, how a company builds its talent practices account for more than 2x the impact on AI outcomes compared to individual factors (67% vs. 32%). In plain English: even the most AI-fluent employee is only getting half the value if their manager doesn't understand AI or hasn't given them room to use it. |
Microsoft's own research lead told the team the surprise finding was how deep people were going. Nearly half of all Copilot conversations involved serious cognitive work like analysis, decision-making, and evaluation. People aren't just asking it to summarize emails. They're using it to think. |
Why this matters: This report is essentially a permission slip for employees who know AI could save them hours a week but can't get buy-in from above. The data now shows the bottleneck is the system itself. That's a much easier argument to make in a board meeting. |
Our take: Microsoft sells the productivity tools that power this data, so they have a pretty clear interest in companies concluding "we need to restructure around AI." Hold that in mind. But, the underlying finding that workers are ahead of their employers matches everything we hear from readers, too. The gap is real. Whether the solution is more Copilot licenses is a separate question. |
|
|
|
|
Move beyond chatbots. Build autonomous AI agents that solve real enterprise challenges. |
Leverage Google Cloud to create agents that don't just answer—they take action. Whether you're optimizing data pipelines or automating customer success, turn your vision into measurable impact. |
Register now to lead the next evolution of AI! |
|
|
Microsoft just quietly shipped something much bigger: Copilot Cowork, an agent that can take a complex goal, build a plan, schedule meetings, draft documents, and execute across your entire Microsoft 365 account while you're doing something else. In this tutorial from Kevin Stratvert's YouTube channel, host Nick Brazzi walks through exactly how it works. |
Think of it less like a chatbot and more like delegating a project to a very capable assistant who has access to your calendar, email, files, and company directory. |
Here's how to get started (note: currently requires a Microsoft 365 Business or Enterprise subscription plus the Copilot add-on; Cowork is in early access via Microsoft's "Frontier" program): |
Go to Microsoft 365 Copilot on the web and sign in with your work account In the navigation panel, select All Agents, then search for Cowork and add it Select the Cowork agent and describe your goal in plain English — be specific about what you want it to do, who's involved, and what the end result should look like Let it run. Come back in a few minutes — it'll ask for your approval before scheduling meetings or sending emails Review the output folder in OneDrive, where it saves all created documents
|
Try this prompt to start: |
Help me plan our team's [event/project name].
Find anyone who was involved last time,
assign responsibilities, find a date that works for everyone in [month],
schedule the planning meetings we need,
and create a kickoff presentation and any supporting documents.
|
Total AI beginner? Start here (goes with this video). |
Have a specific skill you want to learn? Request it here. |
|
|
|
|
|
|
|
Did you know we have a podcast (The Neuron: AI Explained) where we talk to fascinating people in the industry who teach us how it actually works? Check it out: |
 | Click to view these episodes on YouTube! |
|
New episodes air every week on: Spotify | Apple Podcasts | YouTube |
|
📰 Around the Horn |
 | The most misread graph in AI. Each dot shows how long a task takes a human before an AI can do it half the time but METR's own team warns the error bars are so wide the real number could be off by 10x. |
|
MIT Technology Review broke down why METR's viral "time horizon plot" — the graph driving AI boom and doom predictions alike is being wildly misread, with error bars so wide the real capability could be half or double the headline number. Elad Gil warned on the No Priors podcast that most AI startups peak in value for roughly 12 months before foundation model companies absorb their category — and urged founders to schedule exit talks before that window closes. Zoho CEO Sridhar Vembu amplified a viral post from a Meta engineer arguing that rising AI infrastructure costs are the real driver behind the wave of tech layoffs. OpenAI, Anthropic, and GitHub all quietly changed their billing terms in the same week of May 2026, raising effective AI costs without changing their list prices. Transformer equipment shortages (the electrical kind, not the AI kind) are delaying nearly half of planned U.S. AI data center builds, with an 18-month backlog on new orders.
|
|
😹 Monday Meme |
 | Rough day for Jason Killinger, who got arrested after an AI flagged him as a "100% match" to a banned guest. A perfect score on the wrong guy. This is exactly why "human in the loop" isn't just a tech buzzword. It means having an actual person double-check before the handcuffs come out. |
|
|
|
|
|
| That’s all for now. | | What'd you think of today's email? | |
|
|
P.S: Before you go… have you subscribed to our YouTube Channel? If not, can you? |
 | Click the image to subscribe! |
|
P.P.S: Love the newsletter, but only want to get it once per week? Don’t unsubscribe—update your preferences here. |