video
2dn
video2dn
Найти
Сохранить видео с ютуба
Категории
Музыка
Кино и Анимация
Автомобили
Животные
Спорт
Путешествия
Игры
Люди и Блоги
Юмор
Развлечения
Новости и Политика
Howto и Стиль
Diy своими руками
Образование
Наука и Технологии
Некоммерческие Организации
О сайте
The Inside View
AI Alignment video explainers and podcasts
The Battle For The Future Of AI — Full Documentary
2024: The Year Of Artificial General Intelligence
Evan Hubinger (Anthropic)—Deception, Sleeper Agents, Responsible Scaling
Holly Elmore—Pausing Frontier AI Development
GPT-2 Teaches GPT-4: Weak-to-Strong Generalization
How to Catch an AI Liar
Anthropic Solved Interpretability?
We Beat The Strongest Go AI
Paul Christiano's Views on AI Doom (ft. Robert Miles)
Neel Nanda–Mechanistic Interpretability, Superposition, Grokking
Joscha Bach—Is AI Risk Real?
Erik Jones—Automatically Auditing Large Language Models
Dylan Patel—GPU Shortage, Nvidia, Semiconductor Supply Chain
Andi Peng—A Human-in-the-Loop Framework for Test-Time Policy Adaptation
Hailey Schoelkopf—Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
Tomek Korbak—Pretraining Language Models with Human Preferences
Tim Dettmers—k-bit Inference Scaling Laws
Eric Wallace—Poisoning Language Models During Instruction Tuning
Tony Wang—Beating Superhuman Go AIs
Дэвид Бау — Редактирование фактов в GPT, Интерпретируемость
Alexander Pan–Are AIs Machiavellian?
Vincent Weisser–Funding Alignment Research
Aran Komatsuzaki–Scaling, GPT-J
Curtis Huebner—AGI by 2028, 90% Doom
Eric Michaud—Scaling, Grokking, Quantum Interpretability
Daniel Filan–AXRP, LLMs, Interpretability
Existential Risk From AI Is Higher Than 10%—Change My Mind
Jesse Hoogland–AI Risk, Interpretability
Clarifying and predicting AGI
Alan Chan and Max Kaufmann–Model Evaluations, Timelines, Coordination