
“Foundational Challenges In Assuring Alignment and Safety of LLMs” has been released alongside 35+ co-authors from NLP, ML, and AI Safety communities! This work identifies 18 foundational challenges in assuring the alignment and safety of large...
“Foundational Challenges In Assuring Alignment and Safety of LLMs” has been released alongside 35+ co-authors from NLP, ML, and AI Safety communities! This work identifies 18 foundational challenges in assuring the alignment and safety of large...
🌐 ERA-Krueger Lab (University of Cambridge) AI Safety Internship 2024 Join the Krueger AI Safety Lab (KASL) at the University of Cambridge for a paid Research Internship focusing on technical and governance aspects of AI safety....
“Foundational Challenges In Assuring Alignment and Safety of LLMs” has been released alongside 35+ co-authors from NLP, ML, and AI Safety communities! This work identifies 18 foundational challenges in assuring the alignment and safety of large language models (LLMs). These...
Ethan Caballero, Kshitij Gupta, Irina Rish, David Krueger.
International Conference on Learning Representations (2023)
Ekdeep Singh Lubana, Eric J Bigelow, Robert Dick, David Krueger, Hidenori Tanaka.
International Conference on Machine Learning (2023)
Joar Skalse, Niki Howe, Dmitrii Krasheninnikov, David Krueger.
Neural Information Processing Systems (2022)
Lauro Langosco, Jack Koch, Lee Sharkey, Jacob Pfau, David Krueger.
International Conference on Machine Learning (2022)
We are an artificial intelligence safety research group at Mila – Quebec Artificial Intelligence Institute.
Copyright © KASL 2025