AGI remains unsolved.
New ideas still needed.

ARC-AGI-3

Interactive Reasoning Benchmark


Measuring interactive human-like intelligence in AI

Coming March 25, 2026

The First Interactive Reasoning Benchmark

ARC-AGI-3 is the first interactive reasoning benchmark designed to measure human-like intelligence in AI. Launching March 25, 2026, it will include 1,000+ levels across 150+ environments that require agents to explore, learn, plan, and adapt. ARC-AGI-3 will display the most authoritative evidence of AI generalization to date.

ARC-AGI-3 uses video-game-like environments where agents must act across multiple steps to achieve long-horizon goals. The games provide no instructions, so players must explore and discover the rules to succeed. Each environment is hand-crafted and novel, so systems cannot memorize their way to success.

Every environment (100%) is human-solvable.

When testing AI, the question isn't whether it solves the environment, it's how efficiently it does so. We measure this through action efficiency: how many actions does it take to complete a goal? This shows how effective a test-taker (human or AI) is at converting environment information into a working strategy.

Humans do this well. AI does not.

Experience ARC-AGI-3

ARC-AGI-3 Developer Toolkit

The ARC-AGI-3 developer toolkit is a series of tools that allows you to play and interact with ARC-AGI-3 environments either locally (up to 2000 FPS), online, or via a hosted API. The toolkit is the best way to get started with research on ARC-AGI-3. See the documentation.

Developer Toolkit code preview
Create an environment to play LS20. Get started in a few lines of code.

Public Environment #1

LS20

Navigating conditional interactions

Planning and memory in an environment governed by latent state.

See LS20 human baseline replays

LS20 gameplay demo

Public Environment #2

VC33

Use budgets and logic to complete a puzzle

Budget gates to complete a goal across the map.

See VC33 human baseline replays

VC33 gameplay demo

Public Environment #3

FT09

Abstract logic and pattern matching with new mechanics

Complete the pattern while dealing with unified goals.

See FT09 human baseline replays

FT09 gameplay demo

Videos

ARC-AGI-3 Preview
ARC-AGI-3 Preview

In July 2025, ARC Prize President Greg Kamradt shares the first look at the future of ARC-AGI-3.
ARC-AGI-3 in Action: Efficiency
ARC-AGI-3 in Action: Efficiency

Greg Kamradt explains how ARC-AGI-3 is scored: measuring how quickly AI completes games compared directly to humans.
François Chollet at YC Startup School
François Chollet at YC Startup School

François Chollet previews ARC-AGI-3 at Y Combinator Startup School in San Francisco.

ARC-AGI-3 Timeline

March 25, 2026
Launch

ARC-AGI-3 Launch

Public release of ARC-AGI-3 benchmark and competition.

January 29, 2026
Release

Developer Toolkit Launch

ARC-AGI-3 Developer Toolkit released. View docs

October 23, 2025
Event

ARC-AGI-3 in Action: Efficiency

Presentation at MIT. Watch video

August 19, 2025
Blog Release

30-Day Learnings

July 17, 2025
Event Release

ARC-AGI-3 Preview Event

June 6, 2025
Announcement

ARC-AGI-3 Announced

Measuring AGI, Interactive Reasoning Benchmarks presentation at AI World's Fair. Watch video

Toggle Animation