Will Jan Leike still be leading Superalignment in 4 years?
23
1kṀ66532027
1%
chance
1D
1W
1M
ALL
Ilya Sutskever (cofounder and Chief Scientist of OpenAI) has made this his core research focus, and will be co-leading the team with Jan Leike (Head of Alignment).
Resolves NO if a reputable source says that Jan Leike is no longer leading Superalignment.
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
People are also trading
Will OpenAI achieve "very high level of confidence" in their "Superalignment" solutions by 2027-07-06?
4% chance
Will xAI significantly rework their alignment plan by the start of 2026?
63% chance
Will Superalignment succeed, according to a majority of Eliezer, Nate, Paul, John, David.
4% chance
Will I think that alignment is no longer "preparadigmatic" by the start of 2026?
18% chance
Will there be more alignmentforum posts from 2025 than 2024?
55% chance
Will a major AI alignment office (eg Constellation/Lightcone/HAIST) give out free piksters to alignment ppl by EOY 2027?
43% chance
Will I have a career as an alignment researcher by the end of 2024?
38% chance
In 5 years will I think the org Conjecture was net good for alignment?
57% chance
Sort by:
@shankypanky The question is clearly referring to the OpenAI Superalignment team. Even if he led a team called Superalignment at Anthropic, it shouldn't count.
bought Ṁ944 NO
Resolves NO, he left:
https://x.com/janleike/status/1791498174659715494
https://www.cnbc.com/2024/05/17/openai-superalignment-sutskever-leike.html
People are also trading
Related questions
Will OpenAI achieve "very high level of confidence" in their "Superalignment" solutions by 2027-07-06?
4% chance
Will xAI significantly rework their alignment plan by the start of 2026?
63% chance
Will Superalignment succeed, according to a majority of Eliezer, Nate, Paul, John, David.
4% chance
Will I think that alignment is no longer "preparadigmatic" by the start of 2026?
18% chance
Will there be more alignmentforum posts from 2025 than 2024?
55% chance
Will a major AI alignment office (eg Constellation/Lightcone/HAIST) give out free piksters to alignment ppl by EOY 2027?
43% chance
Will I have a career as an alignment researcher by the end of 2024?
38% chance
In 5 years will I think the org Conjecture was net good for alignment?
57% chance