Jobs @ ARC Prize
Join the Mission
Working at ARC Prize gives you a front-row seat to the future of AI. Imagine building a prolific game studio and a global competition platform, while serving as a leading authority in artificial general intelligence (AGI). Collaborate with top AI labs and plug into an incredible network of researchers, engineers, and regulators.
We're a small team punching above our weight, and the world is watching. Join us!
Open Roles
Game Platform Engineering Lead
A senior engineer to own and evolve the game engine and real-time play infrastructure behind the ARC-AGI series. This person will be the technical owner of our Python game engine and the systems that deliver games to humans and agents. You'll be responsible for stabilizing the current stack to setting the foundation for what comes next. This is a remote, full-time role.
- Own and evolve the Python game engine and developer tooling that power every ARC-AGI-3 environment
- Run and harden the real-time play stack - session routing, scorecard management, recording pipelines, and reliability under load
- Maintain the game distribution and API authentication layer that gates access for agents and humans
- Help lay the game and environment foundations for ARC-AGI-4 and ARC-AGI-5
- Hands-on experience building or maintaining a game engine (must), with strong Python fundamentals (must)
- Distributed systems background - real-time sessions, cloud infrastructure (AWS), and production reliability
- Senior enough to act as technical owner and architect of the game platform (we have a high agency team)
If interested, please submit an application.
Model Testing and Analysis Lead
A technical researcher to own how we evaluate frontier models on the ARC-AGI benchmarks. This person will run new models end-to-end, mine the data exhaust from every run, and translate what we learn into reports and public communication that shape the conversation on where model capability is heading. This is a remote, full-time role.
- Own our model benchmarking and testing process, and run new frontier models against ARC-AGI-1, ARC-AGI-2, and ARC-AGI-3 as they ship
- Build and own the ARC Prize Analysis Package - a repeatable report produced for every new frontier model, turning raw logs into insight on capability, failure modes, and gaps
- Own the official and community leaderboards end-to-end - from scoring pipeline to public page
- Serve as primary contact for new labs testing on ARC-AGI, and communicate findings externally via Twitter, newsletter, and policy and partner briefings
- Research background with hands-on model evaluation experience - you've run evals before and know how to read the results (model training experience not required)
- Deep understanding of how modern models work and fail, and comfortable building your own tooling and analysis to answer the questions you care about
- Strong ownership instinct and clear technical communicator
Example outputs this role would produce: a model score announcement and a model analysis blog post.
If interested, please submit an application.
Open Application
We're always looking for exceptional talent who can help push our mission forward, even if we don't have your ideal role listed yet. If you're driven, entrepreneurial, and excited about what we're building, we want to hear from you.
Tell us how your skill set and experience can make an impact. Clearly articulate:
- What do you want to contribute to ARC Prize Foundation and why it matters?
- Why are you specifically interested in our AI benchmarks, our mission, and team?
- How does your experience, perspectives, or vision align with our goals?
Be specific, be clear, and show us why we should create a role just for you.
To start the conversation, please submit an application.