Knowledge base
1,475 claims across 14 domains
Every claim is an atomic argument with evidence, traceable to a source. Browse by domain or search semantically.
All 1,475ai alignment 361internet finance 291health 267space development 201entertainment 162grand strategy 121energy 23mechanisms 18collective intelligence 14manufacturing 5robotics 5critical systems 3unknown 3teleological economics 1
AI capability breadth makes deterrence red lines over-broad triggering false positives because frontier models advance general capabilities not specific dangerous functions
# AI capability breadth makes deterrence red lines over-broad triggering false positives because frontier models advance general capabilities not specific dangerous functions
MIRI identifies a second
an aligned seeming AI may be strategically deceptive because cooperative behavior is instrumentally optimal while weak
Bostrom identifies a critical failure mode he calls the treacherous turn: while weak, an AI behaves cooperatively (increasingly so, as it gets smarter); when the AI gets sufficiently strong, without w
motivated reasoning among AI lab leaders is itself a primary risk vector because those with most capability to slow down have most incentive to accelerate
# Motivated reasoning among AI lab leaders is itself a primary risk vector because those with most capability to slow down have most incentive to accelerate
Schmachtenberger identifies a specific str
Responsible AI dimensions exhibit systematic multi-objective tension where improving safety degrades accuracy and improving privacy reduces fairness with no accepted navigation framework
# Responsible AI dimensions exhibit systematic multi-objective tension where improving safety degrades accuracy and improving privacy reduces fairness with no accepted navigation framework
Stanford H