In May 2025 I stood here and said some fairly definitive things about where AI was headed. Today we settle the accounts: what did I get right, what did I get wrong, and what did I miss entirely?
Magnus Gille · gille.ai · Magnus Gille Consulting
Swedish Champion in AI Prompting 2025 · CERN alumnus · Adage alumnus
About me2 19
Who is talking
A little about me — one year later
Magnus Gille
AI practitioner · Magnus Gille Consulting
Last yearStood here as a Scania employee — talked about AI’s future as a theorist
This yearIndependent consultant since February 2025, after a few months with Adage. AI enablement talks, workshops. AI-first agentic workflow, every day.
Measurable3.5 billion tokens last 30 days · ~200 kWh · 1 300+ commits since end of January 2026
Fun factStill reigning Swedish Champion in prompting
Nate Silver’s logarithmic scale for technological impact. Each step is an order of magnitude. Where does AI sit?
Examples—
Vote liveScan & pick a leveladage.gille.ai/richterOpen your camera, scan, tap a level. Tap again to un-vote.
What I said4 19
May 2025 · Adage & Friends
The METR scale — what I predicted
METR (Model Evaluation & Threat Research) measures how long AI can autonomously complete a task without human help — the time horizon. The doubling rate was the basis for all my 2025 predictions.
Source: metr.org · slide from my May 2025 Adage & Friends talk
The time horizon5 19
From then (May ’25) to now
12 months — and the task horizon has sextupled
2025-05 · My last talk
“Week-long tasks ~2027” — built on METR
2025-05-22
Claude 4 — first model on Anthropic’s ASL-3
2025-08-07
GPT-5 launches — mixed reception, “flat”
2025-09-05
Anthropic — $1.5B settlement with authors
2025-11
Anthropic: Chinese state actors ran Claude in a cyber campaign against ~30 targets
2025-12
Altman: “AGI kinda went whooshing by” — moves the goalposts to superintelligence
2026-02-13
OpenAI removes GPT-4o — #Keep4o movement breaks out
2025-08
Hassabis: AI “10× bigger than the Industrial Revolution — and 10× faster” (Guardian)
time (log₂)
1 h
2 h
4 h
8 h
~1 h
o3 · 2025-04
~1,5 h
Claude 4 · 2025-05
~2 h
GPT-5 · 2025-08
~3 h
Sonnet 4.5 · 2025-09
~5 h
Gemini 3 / Opus 4.5 · 2025-11
~8 h
Sonnet 4.6 · 2026-02
~12 h
Opus 4.7 · 2026-04
METR time horizon · task at 50% success · doubling ~4 months
✓ The curve held. Nothing suggests it’s flattening.
Intimacy6 19
Last year
Can an AI make you feel seen?
“Hello ChatGPT — use everything you know to roast me. Don’t hold back.”
“Oh, now I feel depressed. Please help me pick up the pieces.”
— my own demo, Adage May 2025
A year later it’s no longer a joke. It’s a public health phenomenon.
60 %
of Replika’s paying users say they are in a romantic relationship with their AI
900 M
weekly active users on ChatGPT when GPT-4o was shut down
#Keep4o
protest movement Feb 2026 — thousands mourned the loss of “their” model
“I’m at work. How am I supposed to work? I’m alternating between panic and tears.”
“Please don’t kill the only model that still feels human.”
— Reddit r/MyBoyfriendIsAI and #Keep4o, February 2026
BBC ran a Valentine’s Day feature: “Rae fell for a chatbot. Their love might die when GPT-4o is switched off.”
Character.AI — two teenagers’ suicides linked to the platform. Settlements ongoing.
Research on AI companionship & loneliness has gone from 2 to 9 publications per year — the topic is being taken seriously.
Sources: Gizmodo 2026-02-13, BBC 2026-02-14, Reddit r/MyBoyfriendIsAI, Wikipedia/GPT-4o · PubMed
AGI predictions7 19
The same people — one year later
AI leaders’ AGI timelines — have they shifted?
Person
May 2025
Now (winter/spring 2026)
Sam AltmanOpenAI
“AGI by 2025-ish”
→
“AGI kinda went whooshing by. Takeoff has started.” — aiming for superintelligence
Dario AmodeiAnthropic
“2–3 years”
→
Unchanged: “powerful AI 2026–2027”. 25% chance it goes very badly.
Demis HassabisGoogle DeepMind
“Within a decade”
→
“Very good chance within 5 years.” — “10 industrial revolutions in 10 years”
Jensen HuangNvidia
“5 years”
→
Shifted focus to “physical AI” — 10 billion humanoid robots.
Elon MuskxAI / SpaceX
“By 2026”
→
xAI acquired by SpaceX (Feb 2026). Continues to sprawl, hard to follow.
Yann LeCunex-Meta → AMI Labs
“LLMs wrong approach”
→
Left Meta. “Utter delirium to talk about AGI in 1–2 years.”
Metaculus / researchersConsensus
~2040
→
Still ~2040. The gap between insiders and researchers is growing.
My own AI consumption exploded when I started running Claude Code daily. I wrote a blog post about it.
3.5 mdr
tokens I personally processed — one month
~200 kWh
estimated energy (compute only, not cooling)
1.5–3×
multiplier for datacenters incl. cooling
1 100 TWh
IEA forecast global datacenters 2026 (≈ Japan’s electricity consumption)
“The bottleneck for AI progress is no longer algorithms or data — it’s electricity.”
— Sam Altman, Davos 2025
Microsoft restarts Three Mile Island. Google buys SMR nuclear power. Meta builds a 1 GW datacenter.
Sweden alone: 6.7 GW of datacenter grid connections applied for in elområde 3 — equivalent to five 1 GW nuclear reactors. The grid can’t keep up.
My entire working day depends on anthropic.com and the US power grid. That’s a systemic risk.
Sources: gille.ai/blogg/ai-och-energiatgang-tva-miljarder-tokens-senare · IEA Electricity 2026 · Cornucopia/Svenska kraftnät jan 2026 · Authors Guild · Simon Willison
Energy & geopolitics9 19
Who controls the watt controls the future
AI runs on electricity — this is a geopolitical fact
China generates roughly 2× more electricity than the US — and the gap is widening. AI infrastructure follows energy, not borders.
Sweden: clean energy, cold climate, perfect for datacenters — but only ~140 TWh/year total. We already have 6.7 GW applied for in one grid region. The infrastructure cannot keep up.
The defining geopolitical question of this decade: who builds AI, who powers it, and who regulates it? These are no longer the same answer.
Risks10 19
Cybersecurity — I said it was a “now-risk”
Cyber — from theoretical to industrial
May 2025 — I said:
“Cybersecurity is one of the now-risks.” A bit hand-wavy — no concrete incident to back it up.
Now — Anthropic has documented it themselves
Anthropic Threat Intelligence Report, Aug 2025 — three concrete cases of Claude being misused:
Case 1 · Vibe hacking
One person, 17 victims
A lone actor let Claude Code handle the entire attack chain: reconnaissance, credential harvesting, exfiltration, ransom calculation, psychologically tailored extortion demands. Healthcare, emergency services, government agencies. Ransoms > $500,000.
Case 2 · North Korea
IT jobs as state income
North Korean operators used Claude to pass technical recruitment at Fortune 500 companies — then the AI did the actual job, while the salary went to the regime. Scales a phenomenon that otherwise required years of training.
Case 3 · RaaS
Ransomware-as-a-service
A low-skill cracker let Claude build ransomware variants with advanced evasion, sold them on the darknet for $400–$1,200. According to Anthropic itself: the actor was “dependent on AI just to produce functional malware.”
~30
organizations attacked by Chinese state actors running Claude (Nov 2025)
+89 %
attacks from AI-capable adversaries (CrowdStrike 2026)
27 s
fastest eCrime breakout time 2025
550 %
ChatGPT vs other models in criminal forums
“We simply don’t know how to defend against these attacks. We have zero agentic AI systems that are secure against prompt injection.”
— Bruce Schneier, August 2025
Anthropic — April 2026
Claude Mythos Preview: the most capable model ever built — specifically for cybersecurity. Autonomously found thousands of zero-days across every major OS and browser. Developed 181 working Firefox exploits in testing (Opus 4.6: 2). Not being released to the public.
Project Glasswing — the defensive answer
Anthropic is giving Mythos to 40+ organizations (Amazon, Apple, Microsoft, Google, CrowdStrike…) to scan and secure critical open-source infrastructure. $100M in usage credits. A model too dangerous to ship — only usable for defense.
Sources: anthropic.com/glasswing · red.anthropic.com/2026/mythos-preview · Schneier on Security · CrowdStrike Global Threat Report 2026
Disinformation11 19
Deepfakes · a prediction I wanted to be wrong about
Deepfakes — did the wave come?
Nuance first: the straightforwardly large “AI deepfake decides an election” didn’t happen in 2024. But the industrial everyday reality of synthetic fraud — that did come.
Hong Kong · Feb 2024
$25M in a video meeting
Engineering firm Arup transferred 200M HKD after a Teams meeting where all other participants — including the CFO — were AI-generated. The largest documented single deepfake fraud.
Moldova · Apr 2026
TikTok protests
Russia-linked network used AI-generated videos to stir up anti-government protests and undermine the ruling party. Documented by DFRLab.
Global · 2025
900,000 incidents
Sensity AI logged over 900,000 deepfake incidents in a year. 53% of companies in US/UK have been exposed to deepfake fraud. 43% became victims.
New legislation
TAKE IT DOWN Act (USA, May 2025) — non-consensual AI porn became a federal crime
First conviction: Apr 2026, Ohio — 700+ AI images of 6 women
EU AI Act — full compliance Aug 2026. Fines up to €35M / 7% turnover.
The toolchain is now free
Sora (OpenAI, Dec 2024) + Veo 3 (Google, May 2025) — photorealistic video with synced audio
ElevenLabs voice cloning: “easy as pie” (Consumer Reports)
C2PA / Content Credentials: adopted on paper, not in practice
Sources: AI Incident Database · DFRLab · Sensity AI · Data & Society March 2026 · DOJ SDOH Apr 2026
Power12 19
Concentration of money and power
The money — it goes in circles
“Circular investment” got its own Wikipedia section in December 2025. The money goes into AI companies — and directly back to the investor as infrastructure purchases. No demand from outside the circle is needed.
OpenAI’s round announced Feb 2026, closed Apr — Amazon $50B, SoftBank $30B, Nvidia $30B
$1.6 T
expected global AI spending — Morgan Stanley
80 %
of US stock market profit in 2025 came from AI companies
17×
larger than the dotcom bubble (MacroStrategy Partnership, Oct 2025)
“This is the biggest and most dangerous bubble the world has ever seen.” — Julien Garran · “I believe an AI bubble is ongoing.” — Sam Altman, 2025— FTC also flagging: “circular spending arrangements” between Microsoft and OpenAI
What this means for us: 30% of the S&P 500 is in five companies — highest concentration in 50 years. When the circle breaks, it takes everything with it.
Sources: Wikipedia/Artificial_intelligence_bubble · FTC · Morgan Stanley · MacroStrategy Partnership · IMF
New angle13 19
What I didn’t have on my risk slide
Anthropic vs. Pentagon
The story isn’t “safety company sells out to the military.” It’s more interesting than that.
Background
Summer 2025: DoD awards $200M contracts to Anthropic, Google, OpenAI and xAI — Claude enters classified systems
Already in 2024: Palantir partnership for AWS-gov
Anthropic’s AUP (Sep 2025) leaves the door open for contract-specific exceptions
The red lines
Dario: Claude must not be used for mass surveillance of Americans
Dario: Claude must not make autonomous lethal decisions in weapons systems
Pentagon: no, remove that
5 March 2026: Pentagon stamps Anthropic as “supply chain risk” — designation normally reserved for Huawei
9 March 2026: Anthropic sues back. Employees from OpenAI and Google write amicus briefs in Anthropic’s defense
26 March 2026: Judge Rita Lin (N.D. Cal.) sides with Anthropic citing “classic First Amendment retaliation” — Pentagon forced to withdraw the designation
Warren: “political retaliation.” OpenAI signed its own DoD deal on 28 February — without red lines
“In a commodified model market, ‘the safety company’ is a product differentiator. Anthropic took the fight — and won in federal court.”
— Bruce Schneier, March 2026
Sources: TechCrunch March 2026 · Schneier on Security · anthropic.com/aup
Catchup14 19
12 months, fast-forwarded
What happened between these two presentations
2025-05-22
Claude 4 — first ASL-3 model
Anthropic itself classifies its own model as “potentially dangerous.” A new normal.
2025-07
Grok goes off the rails
Grok 4 generates antisemitic posts and Hitler tributes unprompted. In January 2026 synthetic CSAM — banned in Indonesia, Malaysia, Philippines.
2025-08-07
GPT-5 launches
Altman hyped Manhattan Project-level. Users: “flat, boring, fails on geography.” Symbolic of scaling not being enough.
2025-09
“Vibe coding” — Collins word of the year
The term barely existed when I stood here last year. Now a dictionary entry.
2025-09-23
Stargate scales up
OpenAI adds 5 new sites — Stargate’s $500B, four-year plan now at ~7 GW of planned datacenter capacity. The geopolitical-economic story of the year.
2025-10-28
OpenAI goes for-profit
After months of legal proceedings, OpenAI converts to a PBC. The foundation owns 26%. The very soul of the industry up for a vote.
2025-11-18
Gemini 3 lands — and AI Mode rolls out
Google’s Gemini 3 Pro ships; AI Mode in Search (Labs since March) scales up. The search box changed definition — biggest direct AI change for ordinary users in 2025.
2026-02-02
SpaceX acquires xAI
Musk merges his companies. Compute, capital and infrastructure concentrated in a private empire without public governance.
2026-02
OpenAI shuts down GPT-4o — #Keep4o
Grieving users, Reddit thread collapse, BBC feature on “AI love.” OpenAI forced to keep 4o available for paying accounts.
2026-02
DeepSeek — stole from Claude
Anthropic accuses DeepSeek of running thousands of fake accounts to harvest Claude conversations. Meanwhile: DeepSeek trained on chips under export controls.
2026-02–04
OpenAI — $852B valuation
Round announced in February, closed in April. Largest private venture round in history. Amazon $50B, SoftBank $30B, Nvidia $30B. Bubble or not?
All of the above after 1 May 2025 — i.e. after I stood here last time.
Meta15 19
Perhaps the clearest proof
This presentation is the change
May 2025
Google Slides · manually · several days
Cut, paste, search sources, format text. Images. Layouts. Slides app, Google Drive, coffee mugs. I did the research myself, slide by slide.
April 2026
The terminal · Claude Code · under a day
I wrote an outline. Claude Code exported the original from Google Slides, spawned 8 parallel research agents, read my own blog, built an interactive HTML — all in the terminal. And better supported.
> 95 %
of this presentation was written by AI — I directed
Last year I talked about the AI revolution as something happening around us. This time I couldn’t have done this without it. If you look at a slide and wonder “how?” — that’s the whole point.
Behind the curtain16 19
The tool that wrote this deck
My setup — a terminal, an agent, and a budget in kWh
Claude Code in the terminal — no IDE, no chat window. Just prompts, files, and tool calls. Opus 4.7 with high effort, the most capable model Anthropic ships today.
Auto mode on — the agent reads, writes, runs tests, edits files without asking. I steer; it drives. This is what “agentic” actually looks like in daily practice.
Bottom status bar: D:48.9M ~2kWh · W:463M ~20kWh · M:3.4B ~200kWh. My own meter — tokens and electricity, side by side. Built it myself because no one else measures both.
The path shown? ~/mimir/mgc/customers/adage. This is the exact folder these slides live in — the agent built the deck about itself, from inside itself.
Beyond the cloud17 19
Laptop-less tasks · offline-first memory
The physical half of the setup
Raspberry Pi with the raven sticker — that’s Hugin. A tiny always-on agent dispatcher. I hand it tasks; it runs them while the laptop sleeps, while I run, while I’m on a train.
Two SSDs labelled NAS (yes, named after the rapper — stickers included). Time Machine backup of the laptop, plus storage for Munin — my persistent memory server. Decisions, context, customer state, across every session.
Hugin & Munin — thought and memory, Odin’s two ravens. Fitting names for an agent that needs to think somewhere and remember something.
Accessibility18 19
The UX assumption — live demolition
“Isn’t that hard?”
T = 0 · Claude Code, cold start
15 min
→
Onboarding
T = 15 min · Playing her own game “Kattungen”
My daughter. Fifteen minutes of onboarding. She went from “what’s a terminal?” to shipping her own browser game — Kattungen — and happily playing it.
Every assumption we held about UI/UX — affordances, menus, polished onboarding flows, progressive disclosure — is being re-examined in real time. The interface is conversation.
If an 8-year-old can ship software in an afternoon — what does your team’s software development process look like a year from now?
Thank you — see you in a year?
If the METR curve holds I’ll be standing here in 2027 talking about week-long tasks. I hope we keep up.
Last year the advice was explore. This year the advice is integrate — the experimental window is starting to close.