
Background
ARC‑AGI was introduced in 2019 as a grid‑based reasoning benchmark (“v1”) designed to test whether AI systems can infer novel rules from a few examples rather than rely on pattern memorization. Open‑source solvers plateaued near 53 % accuracy, while a high‑compute run of OpenAI’s o3‑preview model achieved roughly 75–88 %, indicating that v1 was largely saturated.
To raise the bar, the ARC Prize Foundation unveiled the harder, human‑validated “ARC‑AGI‑2” (v2) on 24 March 2025 and opened a Kaggle contest capped at about US $0.42 of compute per task. The headline rule remains: the first fully open‑source system to reach ≥ 85 % on the private v2 set wins the $1 million Grand Prize.
Resolution Criteria
The market resolves YES if before January 1, 2026 the ARC Prize Foundation publicly announces and awards any portion of the $1 million Grand Prize to one or more teams.
Primary rule: The winning submission must achieve ≥ 85 % accuracy on ARC‑AGI‑2 (or an officially designated successor) during an official competition period.
Future changes: If ARC publishes a new test or alters the accuracy threshold, the operative condition remains “the first public, binding commitment to pay out—or the actual payout of—the prize labelled the ARC Grand Prize.”
@fornever while I doubt this will change, this market still has time to resolve the other way. please ping again in Jan!
This is super interesting:
https://iliao2345.github.io/blog_posts/arc_agi_without_pretraining/arc_agi_without_pretraining.html
@MalachiteEagle Does rhat mean you only get 1 million $ by passing 85 on v2? If so that’s crap!
@AndrewMcKnight it is not an accident. I thought it would be better if people predicted it a year ahead of time.