“Foundational Challenges In Assuring Alignment and Safety of LLMs” has been released alongside 35+ co-authors from NLP, ML, and AI Safety communities! This work identifies 18 foundational challenges in assuring the alignment and safety of large…
We are part of the Computational and Biological Learning Lab and the Machine Learning Group.
“Foundational Challenges In Assuring Alignment and Safety of LLMs” has been released alongside 35+ co-authors from NLP, ML, and AI Safety communities! This work identifies 18 foundational challenges in assuring the alignment and safety of large…
🌐 ERA-Krueger Lab (University of Cambridge) AI Safety Internship 2024 Join the Krueger AI Safety Lab (KASL) at the University of Cambridge for a paid Research Internship focusing on technical and governance aspects of AI safety.…
“Foundational Challenges In Assuring Alignment and Safety of LLMs” has been released alongside 35+ co-authors from NLP, ML, and AI Safety communities! This work identifies 18 foundational challenges in assuring the alignment and safety of large language models (LLMs). These…
Ethan Caballero, Kshitij Gupta, Irina Rish, David Krueger.
International Conference on Learning Representations (2023)
Ekdeep Singh Lubana, Eric J Bigelow, Robert Dick, David Krueger, Hidenori Tanaka.
International Conference on Machine Learning (2023)
Joar Skalse, Niki Howe, Dmitrii Krasheninnikov, David Krueger.
Neural Information Processing Systems (2022)
Lauro Langosco, Jack Koch, Lee Sharkey, Jacob Pfau, David Krueger.
International Conference on Machine Learning (2022)
We are an artificial intelligence safety research group at the University of Cambridge’s Department of Engineering. We are part of the Computational and Biological Learning Lab (CBL).
Copyright © KASL 2024