Is it possible to align scaffolded LLMs with human values?
Basic
5
Ṁ712100
70%
chance
1D
1W
1M
ALL
This question is managed and resolved by Manifold.
Get
1,000and
3.00
Related questions
Related questions
Do LLMs experience qualia?
25% chance
are LLMs easy to align because unsupervised learning imbues them with an ontology where human values are easy to express
32% chance
By 2025 end, will it be generally agreed upon that LLM produced text/code > human text/code for training LLMs?
11% chance
Will "LLMs for Alignment Research: a safety priority?" make the top fifty posts in LessWrong's 2024 Annual Review?
14% chance
Will an LLM improve its own ability along some important metric well beyond the best trained LLMs before 2026?
50% chance
Will relaxed adversarial training be used in practice for LLM alignment or auditing before 2028?
79% chance
EOY 2025: Will open LLMs perform at least as well as 50 Elo below closed-source LLMs on coding?
30% chance
Will LLMs become a ubiquitous part of everyday life by June 2026?
82% chance
Will LLMs be the best reasoning models on these dates?
Will there be an LLM which scores above what a human can do in 2 hours on METR's eval suite before 2026?
70% chance