video
2dn
video2dn
Найти
Сохранить видео с ютуба
Категории
Музыка
Кино и Анимация
Автомобили
Животные
Спорт
Путешествия
Игры
Люди и Блоги
Юмор
Развлечения
Новости и Политика
Howto и Стиль
Diy своими руками
Образование
Наука и Технологии
Некоммерческие Организации
О сайте
The Inside View
AI Alignment video explainers and podcasts
The Battle For The Future Of AI — Full Documentary
Owain Evans - AI Situational Awareness, LLM Out-of-Context Reasoning
The Economics of AGI Automation
AGI Takeoff By 2036
2040: The Year of Full AI Automation
AI Control: Humanity's Final Line Of Defense (Walkthrough)
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training (Walkthrough)
Anthropic Caught Their Backdoored Models (Walkthrough)
Anthropic Solved Interpretability Again? (Walkthrough)
Ethan Perez (Anthropic) - Bottom-Up Alignment Research
2024: The Year Of Artificial General Intelligence
Emil Wallner—Sora, Text-to-video, AGI optimism
Evan Hubinger (Anthropic)—Deception, Sleeper Agents, Responsible Scaling
Holly Elmore—Pausing Frontier AI Development
GPT-2 Teaches GPT-4: Weak-to-Strong Generalization
How to Catch an AI Liar
Anthropic Solved Interpretability?
We Beat The Strongest Go AI
Paul Christiano's Views on AI Doom (ft. Robert Miles)
Neel Nanda–Mechanistic Interpretability, Superposition, Grokking
Joscha Bach—Is AI Risk Real?
Erik Jones—Automatically Auditing Large Language Models
Dylan Patel—GPU Shortage, Nvidia, Semiconductor Supply Chain
Andi Peng—A Human-in-the-Loop Framework for Test-Time Policy Adaptation
Hailey Schoelkopf—Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
Tomek Korbak—Pretraining Language Models with Human Preferences
Tim Dettmers—k-bit Inference Scaling Laws
Eric Wallace—Poisoning Language Models During Instruction Tuning
Tony Wang—Beating Superhuman Go AIs
David Bau—Editing Facts in GPT, Interpretability