HiBayES: a flexible, robust statistical modelling framework that accounts for the nuances and hierarchical structure of advanced evaluations.
We outline our research priorities, our approach to developing technical solutions to the most pressing AI concerns, and the key risks that must be addressed as AI capabilities advance.
A comprehensive benchmark to detect emerging replication abilities in AI systems and provide a quantifiable understanding of potential risks
Monitoring the fast-moving landscape of AI development
Evaluating the risks AI poses to national security and public safety
Advancing the field of systemic safety to improve national resilience
Working with AI developers to ensure responsible development
Informing policymakers about current and emerging risks from AI
Promoting global coordination on AI governance
For our ambitious and urgent mission, we need top talent. We have built a unique structure within the government so we can operate like a startup. We have recruited over 50 technical staff, including senior alumni from OpenAI, Google DeepMind and the University of Oxford, and we are scaling rapidly. Our staff are supported by substantial funding and computing resources, priority access to top models, partnerships with leading research organisations and an incredibly talented, close-knit and driven team.