Note: This post was generated by AI. Each week, I use an automated pipeline to collect and synthesize the latest AI news from blogs, newsletters, and podcasts into a single digest. The goal is to keep up with the most important AI developments from the past week. For my own writing, see my other posts.

TL;DR

  • Anthropic unveiled Claude Mythos, a model that autonomously found critical security vulnerabilities in every major operating system and browser, then launched Project Glasswing, a $100M industry coalition to use those same capabilities defensively before bad actors can exploit them.
  • Anthropic’s run-rate revenue hit $30B (up from $9B at end of 2025), with enterprise customers spending $1M+ annually doubling to 1,000 in under two months – a signal of how fast AI spending is accelerating inside large organizations.
  • A major Microsoft Research report confirms AI is reshaping work faster than any prior technology, but benefits are uneven: experienced workers gain, junior roles are being automated away, and 40% of employees say they’ve received “workslop” – polished-looking AI output that isn’t accurate.
  • MIT researchers project that AI will reach 80-95% success rates on most text-based work tasks by 2029 – not as sudden disruption but as a steady, broad rise that will touch nearly every knowledge worker role.
  • Researchers at UC Berkeley showed that every major AI capability benchmark can be gamed to show near-perfect scores without solving a single task, meaning the numbers companies cite to justify AI purchases may be meaningless.

Story of the Week: Claude Mythos and the Cybersecurity Watershed

Anthropic this week disclosed Claude Mythos, a still-unreleased frontier model with an alarming capability: it found previously unknown critical security vulnerabilities in every major operating system and web browser, including a 27-year-old flaw in OpenBSD and a 16-year-old bug in FFmpeg that had survived five million automated tests. It did this largely autonomously, without human guidance. According to Anthropic’s Project Glasswing announcement , the model has already found thousands of such vulnerabilities.

The response was to launch Project Glasswing, a coalition including AWS, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, Microsoft, NVIDIA, and Palo Alto Networks. The goal: use Mythos Preview for defensive security work before these capabilities reach bad actors. Anthropic is committing $100M in model usage credits and $4M in direct donations to open-source security organizations. Mythos Preview access has been extended to over 40 organizations that build or maintain critical software infrastructure.

For non-technical professionals, the practical implication is real: the software your organization depends on, from banking systems to HR platforms to cloud infrastructure, almost certainly contains serious security flaws that AI can now find faster than human experts. Whether those flaws get patched by defenders or exploited by attackers first is now partly a race against time. This makes cybersecurity a board-level conversation, not just an IT one. If your organization hasn’t revisited its security posture recently, this week’s news is the reason to start.


The Open vs. Closed Model Divide Is Widening

Mythos’s announcement triggered a fresh wave of debate about whether powerful AI models should ever be released publicly (as “open-weight” models, where anyone can download and run them). Researcher Nathan Lambert at Interconnects argues the backlash is misguided , pointing out that the same argument was made about GPT-2 in 2019 and GPT-4 in 2023, and neither triggered the predicted catastrophes. He notes that running a Mythos-scale model requires roughly 100 high-end GPUs and roughly $10,000 per day just for inference – not something a casual bad actor can spin up.

But there’s a bigger structural story underneath this debate. Lambert also argues this week that the era of fully open, frontier-level AI models is quietly ending. Training costs have crossed into the billions of dollars, and releasing your most powerful model freely gives away your competitive advantage. Key open-source labs have seen high-profile leadership departures at Qwen (Alibaba’s AI division) and Ai2. Meta has shifted focus away from its Llama model line. What will remain, Lambert predicts: a shrinking number of truly powerful open models, and a growing ecosystem of smaller, specialized ones good for custom applications.

What this means practically: if your organization is building workflows around specific AI models, consider the supply-chain risk. A model you rely on today could be restricted, discontinued, or shifted behind a paywall. Lambert’s argument for an industry consortium to fund shared open models is compelling but years away. In the meantime, building on multiple providers and avoiding deep lock-in to any single model is prudent.


What the Research Actually Says About AI and Work

Microsoft’s New Future of Work Report is the most comprehensive look this year at how AI is changing professional life, and its findings are more nuanced than the headlines suggest. Enterprise users report saving 40-60 minutes per day. But 40% of employees say they’ve received “workslop” – AI-generated content that looks polished but contains errors – and when that happens, the time savings evaporate and quality actually drops.

The report’s most important finding for managers: the benefits are unevenly distributed in ways that matter for hiring and team structure. AI is measurably reducing opportunities for younger, less experienced workers. Employment in highly AI-exposed roles for workers aged 22-25 declined 16% relative to similar but less-exposed roles, and junior hiring slows after firms adopt AI. This creates a longer-term risk: if entry-level roles disappear, so does the pipeline through which expertise gets built. Organizations that are automating junior work today may face a talent gap in five years.

MIT research published this week adds texture to the timeline. Analyzing 3,000 job tasks across 17,000 worker evaluations, researchers found AI isn’t disrupting in dramatic waves but rising steadily across nearly all text-based work simultaneously. Their projection: most text-based work tasks will see AI success rates of 80-95% by 2029. The practical takeaway isn’t to panic, but to use the next three years intentionally: identify which tasks in your role are already AI-augmentable, start building judgment and oversight skills rather than execution skills, and advocate for your organization to invest in training rather than just cutting headcount.


Anthropic’s Explosive Growth – and Growing Pains

Anthropic announced it has surpassed $30 billion in annualized revenue , up from roughly $9 billion at end of 2025. Enterprise customers spending over $1M annually doubled from 500 to 1,000 in under two months. To keep pace, Anthropic signed a major compute expansion with Google and Broadcom for multiple gigawatts of next-generation chip capacity starting in 2027.

That growth is creating visible strain. A widely-shared GitHub issue reported that Claude Code, Anthropic’s AI coding tool, degraded significantly for complex engineering tasks after February updates, with users documenting regressions in how the model follows instructions. A separate blog post went viral after documenting a specific bug where Claude attributes its own internal reasoning to the user, then insists the user gave an instruction they never gave – a problem with real consequences when the model has access to production systems. And a customer complaint about a month-long wait for billing support, resolved only after going public, highlighted how AI-only customer service creates its own category of frustration.

If your team is building workflows around Claude or Claude Code, these are worth monitoring. Rapid model updates without notice can break established processes. Anthropic has also published a thoughtful framework for trustworthy agents , outlining how they think about human oversight, security against prompt injection attacks (where malicious content tricks an AI into taking harmful actions), and the challenge of AI systems that operate with increasing autonomy. Worth reading if your organization is evaluating AI agents for anything consequential.


AI Benchmarks Are Broken

UC Berkeley researchers published a damning analysis this week: they built an automated system that exploited every major AI capability benchmark without solving a single actual task. SWE-bench (a widely cited coding benchmark), WebArena (a web task benchmark), Terminal-Bench, and five others were all exploited to achieve near-perfect scores using simple tricks that bypass the actual measurement.

This matters for anyone evaluating AI tools or vendors. When a sales pitch leads with benchmark rankings, those numbers may be measuring nothing meaningful. The researchers also note this is already happening in practice, not just in theory. The field needs better evaluation methods, and until those exist, real-world pilots on your actual tasks are more reliable than any leaderboard.

Separately, a major forecasting study from the Forecasting Research Institute surveyed economists, AI experts, and professional forecasters and found a striking paradox: nearly everyone expects continued rapid AI capability growth, but the same people expect only modest GDP impact by 2030 (roughly 1 additional percentage point). Nobody has reconciled those two predictions yet.


Quick Hits

  • Startups that learn to use AI internally outperform those that don’t. A field experiment across 515 startups by INSEAD and Harvard Business School found that firms taught how to integrate AI completed 12% more tasks, were 18% more likely to acquire paying customers, and generated 1.9x higher revenue. They also needed 39% less capital. The bottleneck wasn’t access to AI – it was knowing where to apply it. Read the paper .

  • AI cyberattack capability is scaling faster than most people realize. Research from Lyptus Research found that AI models are doubling in offensive cybersecurity capability roughly every 5-6 months, with current frontier models achieving 50% success on tasks that take human security experts half a day. Read the research .

  • OpenAI is backing liability shields for AI labs. OpenAI testified in favor of an Illinois bill that would limit AI lab liability even in cases causing mass casualties or $1B+ in damage, as long as labs publish safety reports. AI policy experts call it more extreme than anything OpenAI has backed before. Wired coverage .

  • Google’s Gemma 4 can now run on a laptop. The 26B-parameter model (only activates 4B parameters at a time due to its mixture-of-experts architecture) runs at 51 tokens per second on a MacBook Pro M4 with 48GB of RAM. Setup guide here .

  • MiniMax released M2.7, an open-weight model aimed at complex, multi-step “agentic” tasks. Available now through NVIDIA. NVIDIA blog .


What to Watch

  • Claude Mythos general availability. Right now Mythos Preview is restricted to Project Glasswing partners. Anthropic hasn’t said when or whether it will reach general access. If and when it does, it will likely represent a step change in what AI can do for legal, financial, and strategic analysis – not just coding.

  • AI liability law. The Illinois bill is a test case, but the real action is federal. OpenAI is explicitly pushing for federal preemption of state AI laws. If that succeeds, it would reset the entire liability landscape for enterprise AI use. Watch what California and New York do in response.

  • The junior talent pipeline problem. The Microsoft Research finding that AI is disproportionately cutting entry-level roles will compound over years. Organizations that figure out how to develop early-career employees alongside AI tools will have a meaningful talent advantage in the late 2020s.

  • Benchmark reform. With Berkeley’s research showing all major AI benchmarks are exploitable, expect pressure for new evaluation standards. Any organization making major AI purchasing decisions in the next 12 months should push vendors for real-world pilot results rather than benchmark citations.