Quotes from Leopold Aschenbrenner's Situational Awareness Paper - By Zvi Mowshowitz
Is the Intelligence-Explosion Near? A Reality Check.
Situational Awareness by Leopold Aschenbrenner: Part 18
Situational Awareness by Leopold Aschenbrenner: Part 16
Ex-OpenAI Employee Just Revealed it ALL!
Ex-OpenAI Employee Reveals TERRIFYING Future of AI
Ex OpenAI Employee "ASI by 2028" | Sabine Hossenfelder responds...
AI Alignment as a Solvable Problem | Leopold Aschenbrenner & Richard Hanania
P2. AI will probably steal your job sooner than you think, and why it mightn't matter
AI safety, RAG benchmarking and responsible AI at ACM FAccT Conference
AI will probably steal your job sooner than you think, and why it mightn't matter
Yann Lecun: Meta AI, Open Source, Limits of LLMs, AGI & the Future of AI | Lex Fridman Podcast #416
AI Won't Be AGI, Until It Can At Least Do This (plus 6 key ways LLMs are being upgraded)
Francois Chollet - LLMs won’t lead to AGI - $1,000,000 Prize to find true solution
The Threat of AI Regulation with Brian Chau
OpenAI’s huge push to make superintelligence safe | Jan Leike
Open AI's Alignment Problem
The double-edged sword of technology, with Wendell Wallach
Revealing CriticGPT: The Final Work of the Superalignment Team 【Literature Review 07】
Populism, Elites, and Preparing for AGI