gille.ai x Adage & Friends · Spring 2026

One year of AI

In May 2025 I stood here and said some fairly definitive things about where AI was headed. Today we settle the accounts: what did I get right, what did I get wrong, and what did I miss entirely?

Magnus Gille · gille.ai · Magnus Gille Consulting
Swedish Champion in AI Prompting 2025 · CERN alumnus · Adage alumnus
About me2 19
Who is talking

A little about me — one year later

Magnus Gille
Magnus Gille
AI practitioner · Magnus Gille Consulting
Last yearStood here as a Scania employee — talked about AI’s future as a theorist
This yearIndependent consultant since February 2025, after a few months with Adage.
AI enablement talks, workshops.
AI-first agentic workflow, every day.
Measurable3.5 billion tokens last 30 days · ~200 kWh · 1 300+ commits since end of January 2026
Fun factStill reigning Swedish Champion in prompting
Conceptual primer3 19
Nate Silver · the technological richter scale

How big an innovation is AI ? — click a level

Click a level to explore

Nate Silver’s logarithmic scale for technological impact. Each step is an order of magnitude. Where does AI sit?

Examples
QR code to adage.gille.ai/richter
Vote live Scan & pick a level adage.gille.ai/richter Open your camera, scan, tap a level. Tap again to un-vote.
What I said4 19
May 2025 · Adage & Friends

The METR scale — what I predicted

METR (Model Evaluation & Threat Research) measures how long AI can autonomously complete a task without human help — the time horizon. The doubling rate was the basis for all my 2025 predictions.

METR AI task completion time horizon — May 2025 talk
Source: metr.org · slide from my May 2025 Adage & Friends talk
The time horizon5 19
From then (May ’25) to now

12 months — and the task horizon has sextupled

2025-05 · My last talk
“Week-long tasks ~2027” — built on METR
2025-05-22
Claude 4 — first model on Anthropic’s ASL-3
2025-08-07
GPT-5 launches — mixed reception, “flat”
2025-09-05
Anthropic — $1.5B settlement with authors
2025-11
Anthropic: Chinese state actors ran Claude in a cyber campaign against ~30 targets
2025-12
Altman: “AGI kinda went whooshing by” — moves the goalposts to superintelligence
2026-02-13
OpenAI removes GPT-4o — #Keep4o movement breaks out
2025-08
Hassabis: AI “10× bigger than the Industrial Revolution — and 10× faster” (Guardian)
time (log₂)
1 h
2 h
4 h
8 h
~1 h
o3 · 2025-04
~1,5 h
Claude 4 · 2025-05
~2 h
GPT-5 · 2025-08
~3 h
Sonnet 4.5 · 2025-09
~5 h
Gemini 3 / Opus 4.5 · 2025-11
~8 h
Sonnet 4.6 · 2026-02
~12 h
Opus 4.7 · 2026-04
METR time horizon · task at 50% success · doubling ~4 months
✓ The curve held. Nothing suggests it’s flattening.
Intimacy6 19
Last year

Can an AI make you feel seen?

“Hello ChatGPT — use everything you know to roast me. Don’t hold back.”
“Oh, now I feel depressed. Please help me pick up the pieces.” — my own demo, Adage May 2025

A year later it’s no longer a joke. It’s a public health phenomenon.

60 %
of Replika’s paying users say they are in a romantic relationship with their AI
900 M
weekly active users on ChatGPT when GPT-4o was shut down
#Keep4o
protest movement Feb 2026 — thousands mourned the loss of “their” model
“I’m at work. How am I supposed to work? I’m alternating between panic and tears.”
Please don’t kill the only model that still feels human.— Reddit r/MyBoyfriendIsAI and #Keep4o, February 2026
Sources: Gizmodo 2026-02-13, BBC 2026-02-14, Reddit r/MyBoyfriendIsAI, Wikipedia/GPT-4o · PubMed
AGI predictions7 19
The same people — one year later

AI leaders’ AGI timelines — have they shifted?

Person May 2025 Now (winter/spring 2026)
Sam AltmanOpenAI“AGI by 2025-ish”AGI kinda went whooshing by. Takeoff has started.” — aiming for superintelligence
Dario AmodeiAnthropic“2–3 years”Unchanged: “powerful AI 2026–2027”. 25% chance it goes very badly.
Demis HassabisGoogle DeepMind“Within a decade”Very good chance within 5 years.” — “10 industrial revolutions in 10 years”
Jensen HuangNvidia“5 years”Shifted focus to “physical AI” — 10 billion humanoid robots.
Elon MuskxAI / SpaceX“By 2026”xAI acquired by SpaceX (Feb 2026). Continues to sprawl, hard to follow.
Yann LeCunex-Meta → AMI Labs“LLMs wrong approach”Left Meta. “Utter delirium to talk about AGI in 1–2 years.”
Metaculus / researchersConsensus~2040Still ~2040. The gap between insiders and researchers is growing.
Sources: blog.samaltman.com/the-gentle-singularity · darioamodei.com · the-decoder.com · wikipedia/AGI
Legal, ethics, practice8 19
Copyright · Energy

Legally, ethically, practically — what happened?

Copyright — May 2025

I mentioned Meta’s “no economic value” quote. Assumed it would be a matter for the courts for over a decade.

Now

Bartz v. Anthropic, Sep 2025$1.5 billion settlement. Largest copyright settlement in US history. ~$3,000 per book across ~500,000 pirated titles.

Ruling: training on pirated copies = not fair use. Training on purchased & scanned books = fair use. Anthropic pays — but wins the practice.

Energy — May 2025

I said: xAI Colossus — ~150 MW now, multi-GW planned. Sounded abstract.

Now — it hit me personally

My own AI consumption exploded when I started running Claude Code daily. I wrote a blog post about it.

3.5 mdr
tokens I personally processed — one month
~200 kWh
estimated energy (compute only, not cooling)
1.5–3×
multiplier for datacenters incl. cooling
1 100 TWh
IEA forecast global datacenters 2026 (≈ Japan’s electricity consumption)
“The bottleneck for AI progress is no longer algorithms or data — it’s electricity.” — Sam Altman, Davos 2025
Sources: gille.ai/blogg/ai-och-energiatgang-tva-miljarder-tokens-senare · IEA Electricity 2026 · Cornucopia/Svenska kraftnät jan 2026 · Authors Guild · Simon Willison
Energy & geopolitics9 19
Who controls the watt controls the future

AI runs on electricity — this is a geopolitical fact

Electricity generation — China, USA, Sweden 1985–2025 (Our World in Data)

China generates roughly 2× more electricity than the US — and the gap is widening. AI infrastructure follows energy, not borders.

Sweden: clean energy, cold climate, perfect for datacenters — but only ~140 TWh/year total. We already have 6.7 GW applied for in one grid region. The infrastructure cannot keep up.

The defining geopolitical question of this decade: who builds AI, who powers it, and who regulates it? These are no longer the same answer.

Risks10 19
Cybersecurity — I said it was a “now-risk”

Cyber — from theoretical to industrial

May 2025 — I said:

“Cybersecurity is one of the now-risks.” A bit hand-wavy — no concrete incident to back it up.

Now — Anthropic has documented it themselves

Anthropic Threat Intelligence Report, Aug 2025 — three concrete cases of Claude being misused:

Case 1 · Vibe hacking

One person, 17 victims

A lone actor let Claude Code handle the entire attack chain: reconnaissance, credential harvesting, exfiltration, ransom calculation, psychologically tailored extortion demands. Healthcare, emergency services, government agencies. Ransoms > $500,000.

Case 2 · North Korea

IT jobs as state income

North Korean operators used Claude to pass technical recruitment at Fortune 500 companies — then the AI did the actual job, while the salary went to the regime. Scales a phenomenon that otherwise required years of training.

Case 3 · RaaS

Ransomware-as-a-service

A low-skill cracker let Claude build ransomware variants with advanced evasion, sold them on the darknet for $400–$1,200. According to Anthropic itself: the actor was “dependent on AI just to produce functional malware.”

~30
organizations attacked by Chinese state actors running Claude (Nov 2025)
+89 %
attacks from AI-capable adversaries (CrowdStrike 2026)
27 s
fastest eCrime breakout time 2025
550 %
ChatGPT vs other models in criminal forums
“We simply don’t know how to defend against these attacks. We have zero agentic AI systems that are secure against prompt injection.” — Bruce Schneier, August 2025
Anthropic — April 2026

Claude Mythos Preview: the most capable model ever built — specifically for cybersecurity. Autonomously found thousands of zero-days across every major OS and browser. Developed 181 working Firefox exploits in testing (Opus 4.6: 2). Not being released to the public.

Project Glasswing — the defensive answer

Anthropic is giving Mythos to 40+ organizations (Amazon, Apple, Microsoft, Google, CrowdStrike…) to scan and secure critical open-source infrastructure. $100M in usage credits. A model too dangerous to ship — only usable for defense.

Sources: anthropic.com/glasswing · red.anthropic.com/2026/mythos-preview · Schneier on Security · CrowdStrike Global Threat Report 2026
Disinformation11 19
Deepfakes · a prediction I wanted to be wrong about

Deepfakes — did the wave come?

Nuance first: the straightforwardly large “AI deepfake decides an election” didn’t happen in 2024. But the industrial everyday reality of synthetic fraud — that did come.

Hong Kong · Feb 2024

$25M in a video meeting

Engineering firm Arup transferred 200M HKD after a Teams meeting where all other participants — including the CFO — were AI-generated. The largest documented single deepfake fraud.

Moldova · Apr 2026

TikTok protests

Russia-linked network used AI-generated videos to stir up anti-government protests and undermine the ruling party. Documented by DFRLab.

Global · 2025

900,000 incidents

Sensity AI logged over 900,000 deepfake incidents in a year. 53% of companies in US/UK have been exposed to deepfake fraud. 43% became victims.

New legislation
  • TAKE IT DOWN Act (USA, May 2025) — non-consensual AI porn became a federal crime
  • First conviction: Apr 2026, Ohio — 700+ AI images of 6 women
  • EU AI Act — full compliance Aug 2026. Fines up to €35M / 7% turnover.
The toolchain is now free
  • Sora (OpenAI, Dec 2024) + Veo 3 (Google, May 2025) — photorealistic video with synced audio
  • ElevenLabs voice cloning: “easy as pie” (Consumer Reports)
  • C2PA / Content Credentials: adopted on paper, not in practice
Sources: AI Incident Database · DFRLab · Sensity AI · Data & Society March 2026 · DOJ SDOH Apr 2026
Power12 19
Concentration of money and power

The money — it goes in circles

“Circular investment” got its own Wikipedia section in December 2025. The money goes into AI companies — and directly back to the investor as infrastructure purchases. No demand from outside the circle is needed.

Nvidiachips
$100B investment →
← ~$100B chip contracts
OpenAI$852B valuation
Microsoftcloud
$13B equity →
← $250B Azure commitment
OpenAI$852B valuation
Amazoncloud
$8B equity →
← multi-year AWS commitment
Anthropic$380B valuation
GoogleTPU
$3B equity →
← TPU commitment
Anthropic$380B valuation
Stargate $500B, 4-year infrastructure buildout (OpenAI + Oracle + SoftBank + Nvidia) — ~7 GW planned datacenter capacity. Where the cloud & chip money ultimately flows.
$122 B
OpenAI’s round announced Feb 2026, closed Apr — Amazon $50B, SoftBank $30B, Nvidia $30B
$1.6 T
expected global AI spending — Morgan Stanley
80 %
of US stock market profit in 2025 came from AI companies
17×
larger than the dotcom bubble (MacroStrategy Partnership, Oct 2025)
“This is the biggest and most dangerous bubble the world has ever seen.” — Julien Garran  ·  “I believe an AI bubble is ongoing.” — Sam Altman, 2025 — FTC also flagging: “circular spending arrangements” between Microsoft and OpenAI

What this means for us: 30% of the S&P 500 is in five companies — highest concentration in 50 years. When the circle breaks, it takes everything with it.

Sources: Wikipedia/Artificial_intelligence_bubble · FTC · Morgan Stanley · MacroStrategy Partnership · IMF
New angle13 19
What I didn’t have on my risk slide

Anthropic vs. Pentagon

The story isn’t “safety company sells out to the military.” It’s more interesting than that.

Background
  • Summer 2025: DoD awards $200M contracts to Anthropic, Google, OpenAI and xAI — Claude enters classified systems
  • Already in 2024: Palantir partnership for AWS-gov
  • Anthropic’s AUP (Sep 2025) leaves the door open for contract-specific exceptions
The red lines
  • Dario: Claude must not be used for mass surveillance of Americans
  • Dario: Claude must not make autonomous lethal decisions in weapons systems
  • Pentagon: no, remove that
“In a commodified model market, ‘the safety company’ is a product differentiator. Anthropic took the fight — and won in federal court.” — Bruce Schneier, March 2026
Sources: TechCrunch March 2026 · Schneier on Security · anthropic.com/aup
Catchup14 19
12 months, fast-forwarded

What happened between these two presentations

2025-05-22

Claude 4 — first ASL-3 model

Anthropic itself classifies its own model as “potentially dangerous.” A new normal.

2025-07

Grok goes off the rails

Grok 4 generates antisemitic posts and Hitler tributes unprompted. In January 2026 synthetic CSAM — banned in Indonesia, Malaysia, Philippines.

2025-08-07

GPT-5 launches

Altman hyped Manhattan Project-level. Users: “flat, boring, fails on geography.” Symbolic of scaling not being enough.

2025-09

“Vibe coding” — Collins word of the year

The term barely existed when I stood here last year. Now a dictionary entry.

2025-09-23

Stargate scales up

OpenAI adds 5 new sites — Stargate’s $500B, four-year plan now at ~7 GW of planned datacenter capacity. The geopolitical-economic story of the year.

2025-10-28

OpenAI goes for-profit

After months of legal proceedings, OpenAI converts to a PBC. The foundation owns 26%. The very soul of the industry up for a vote.

2025-11-18

Gemini 3 lands — and AI Mode rolls out

Google’s Gemini 3 Pro ships; AI Mode in Search (Labs since March) scales up. The search box changed definition — biggest direct AI change for ordinary users in 2025.

2026-02-02

SpaceX acquires xAI

Musk merges his companies. Compute, capital and infrastructure concentrated in a private empire without public governance.

2026-02

OpenAI shuts down GPT-4o — #Keep4o

Grieving users, Reddit thread collapse, BBC feature on “AI love.” OpenAI forced to keep 4o available for paying accounts.

2026-02

DeepSeek — stole from Claude

Anthropic accuses DeepSeek of running thousands of fake accounts to harvest Claude conversations. Meanwhile: DeepSeek trained on chips under export controls.

2026-02–04

OpenAI — $852B valuation

Round announced in February, closed in April. Largest private venture round in history. Amazon $50B, SoftBank $30B, Nvidia $30B. Bubble or not?

All of the above after 1 May 2025 — i.e. after I stood here last time.
Meta15 19
Perhaps the clearest proof

This presentation is the change

May 2025

Google Slides · manually · several days

Cut, paste, search sources, format text. Images. Layouts. Slides app, Google Drive, coffee mugs. I did the research myself, slide by slide.

April 2026

The terminal · Claude Code · under a day

I wrote an outline. Claude Code exported the original from Google Slides, spawned 8 parallel research agents, read my own blog, built an interactive HTML — all in the terminal. And better supported.

> 95 %
of this presentation was written by AI — I directed

Last year I talked about the AI revolution as something happening around us. This time I couldn’t have done this without it. If you look at a slide and wonder “how?” — that’s the whole point.

Behind the curtain16 19
The tool that wrote this deck

My setup — a terminal, an agent, and a budget in kWh

Claude Code terminal — Opus 4.7, auto mode, token/kWh meter
Claude Code in the terminal — no IDE, no chat window. Just prompts, files, and tool calls. Opus 4.7 with high effort, the most capable model Anthropic ships today.
Auto mode on — the agent reads, writes, runs tests, edits files without asking. I steer; it drives. This is what “agentic” actually looks like in daily practice.
Bottom status bar: D:48.9M ~2kWh · W:463M ~20kWh · M:3.4B ~200kWh. My own meter — tokens and electricity, side by side. Built it myself because no one else measures both.
The path shown? ~/mimir/mgc/customers/adage. This is the exact folder these slides live in — the agent built the deck about itself, from inside itself.
Beyond the cloud17 19
Laptop-less tasks · offline-first memory

The physical half of the setup

Raspberry Pi (Hugin, bird sticker) and two SSD drives with Nas rap stickers acting as NAS
Raspberry Pi with the raven sticker — that’s Hugin. A tiny always-on agent dispatcher. I hand it tasks; it runs them while the laptop sleeps, while I run, while I’m on a train.
Two SSDs labelled NAS (yes, named after the rapper — stickers included). Time Machine backup of the laptop, plus storage for Munin — my persistent memory server. Decisions, context, customer state, across every session.
Hugin & Munin — thought and memory, Odin’s two ravens. Fitting names for an agent that needs to think somewhere and remember something.
Accessibility18 19
The UX assumption — live demolition

“Isn’t that hard?”

My daughter at the laptop, Claude Code terminal open
T = 0 · Claude Code, cold start
15 min
Onboarding
My daughter playing Kattungen, a browser game she built with Claude Code
T = 15 min · Playing her own game “Kattungen”

My daughter. Fifteen minutes of onboarding. She went from “what’s a terminal?” to shipping her own browser game — Kattungen — and happily playing it.

Every assumption we held about UI/UX — affordances, menus, polished onboarding flows, progressive disclosure — is being re-examined in real time. The interface is conversation.

If an 8-year-old can ship software in an afternoon — what does your team’s software development process look like a year from now?

Thank you — see you in a year?

If the METR curve holds I’ll be standing here in 2027 talking about week-long tasks. I hope we keep up.

Last year the advice was explore. This year the advice is integrate — the experimental window is starting to close.