
Will a major AI company publish a “responsible scaling policy” for AI consciousness by 2030?
11
170Ṁ2052030
55%
chance
1D
1W
1M
ALL
A responsible scaling policy (RSP) or risk-informed development policy (RDP) is a framework adopted by companies like Anthropic and OpenAI that aims to ensure that they do not release catastrophically unsafe AIs. Such a framework defines levels of concerning capabilities and corresponding types of safety mitigations that would be adopted, and compliance is monitored through ongoing evaluations.
Will a major AI company adopt an “RSP for AI consciousness” by the end of 2029? This would involve:
Defining features or indicators of consciousness to be monitored
Outlining mitigation measures to avoid suffering or conscious AIs
See also: Project ideas: Sentience and rights of digital minds (substack.com)
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
Related questions
Related questions
Will there be significant protests calling for AI rights before 2030?
30% chance
Will a major AI company acknowledge the possibility of conscious AIs by 2026?
72% chance
Will software-side AI scaling appear to be suddenly discontinuous before 2025?
24% chance
Before 2028, will there be a major self-improving AI policy*?
78% chance
Will any developed country establish a limit on compute for AI training by 2026?
21% chance
Will Google DeepMind hire an AI welfare researcher before the end of 2025?
10% chance
Will an AI be solely responsible for an AI breakthrough by the end of 2030?
76% chance
Will there be a global "pause" on cutting-edge AI research due to government regulation by 2025?
1% chance
By end of 2028, will there be a global AI organization, responsible for AI safety and regulations?
42% chance
Will a company with at least $1B in revenue be run primarily by an AI CEO before 2030?
23% chance