
If AGI has an okay outcome, will there be an AGI singleton?
Mini
5
Ṁ6482101
25%
chance
1D
1W
1M
ALL
An okay outcome is defined in Eliezer Yudkowsky's market as:
An outcome is "okay" if it gets at least 20% of the maximum attainable cosmopolitan value that could've been attained by a positive Singularity (a la full Coherent Extrapolated Volition done correctly), and existing humans don't suffer death or any other awful fates.
This resolves YES if I can easily point to the single AGI that has an okay outcome, and NO otherwise.
Get
1,000and
1.00
Related questions
Related questions
Will we get AGI before 2026?
5% chance
Will we get AGI before 2048?
90% chance
Will we get AGI before 2030?
65% chance
Will we get AGI before 2029?
50% chance
Will we get AGI before 2041?
82% chance
Will we get AGI before 2031?
62% chance
A multipolar AGI scenario is safer than a singleton AGI scenario
30% chance
Will AGI be a problem before non-G AI?
20% chance
By when will we have AGI?
Will we get AGI before 2026?
11% chance