If Tetra writes [a] Lesswrong post(s) about alignment she is proud of in 2024, which subfield(s) will it be about?
Premium
17
Ṁ1657Jan 1
63%
Re-explaining the basics
48%
Decision theory
46%
Corrigibility
44%
Public communications
40%
Claude
37%
Embedded agency
26%
Policy
26%
Evals
25%
AI control (https://www.lesswrong.com/posts/kcKrE9mzEHrdqtDpE/the-case-for-ensuring-that-powerful-ais-are-controlled)
25%
Agent foundations
25%
Forecasting
22%
Mechanistic interpretability
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Related questions
Related questions
Will Tetra make an alignment-focused LessWrong post that she is proud of by the end of 2024?
15% chance
Will "Alignment Implications of LLM Successes: a De..." make the top fifty posts in LessWrong's 2023 Annual Review?
57% chance
Will "AI alignment researchers don't (seem to) stack
" make the top fifty posts in LessWrong's 2023 Annual Review?
29% chance
Will "Without fundamental advances, misalignment an..." make the top fifty posts in LessWrong's 2024 Annual Review?
46% chance
Will "The self-unalignment problem" make the top fifty posts in LessWrong's 2023 Annual Review?
14% chance
Will "Tips for Empirical Alignment Research" make the top fifty posts in LessWrong's 2024 Annual Review?
24% chance
Will "A Case for the Least Forgiving Take On Alignment" make the top fifty posts in LessWrong's 2023 Annual Review?
11% chance
Will "What I mean by "alignment is in large part ab..." make the top fifty posts in LessWrong's 2023 Annual Review?
14% chance
Will "Why Not Just Outsource Alignment Research To ..." make the top fifty posts in LessWrong's 2023 Annual Review?
15% chance
Will "Alignment Faking in Large Language Models" make the top fifty posts in LessWrong's 2024 Annual Review?
93% chance