Please enable javascript for this website.

About

Our mission is to equip governments with an empirical understanding of the safety of advanced AI systems.

The AI Safety Institute is a research organisation within the UK Government’s Department for Science, Innovation, and Technology.

AISI brand artwork

Organisation

The AI Safety Institute is a research organisation within the Department of Science, Innovation and Technology.

We are working to:

  • Test advanced AI systems and inform policymakers about their risks;
  • Foster collaboration across companies, governments, and the wider research community to mitigate risks and advance publicly beneficial research; and
  • Strengthen AI development practices and policy globally.

To deliver on these ambitious goals, we designed the Institute as a startup in the government, combining the authority of government with the expertise and agility of the private sector.

We have recruited top talent from across the public and private sectors  

  • Our Chair Ian Hogarth brings experience as a tech investor and entrepreneur.  
  • Our Director Oliver Ilott previously led the Prime Minister’s domestic private office.
  • Our advisory board comprises national security and machine learning leaders, such as Yoshua Bengio.  
  • Our Chief Technology Officer Jade Leung previously led the Governance team at OpenAI.  
  • Our Research Directors Geoffrey Irving, Professor Yarin Gal, and Professor Chris Summerfield have led teams at OpenAI, Google DeepMind and the University of Oxford.  
  • We already have >30 technical staff, and we are scaling rapidly.

We are backing our team with the resources they need to move fast

  • £100m in initial funding for the organisation
  • Privileged access to top AI models from leading companies
  • Priority access to over £1.5 billion of compute in the UK’s AI Research Resource and exascale supercomputing programme 
  • Over 20 partnerships with top research organisations
  • Direct reporting to Secretary of State for Science, Innovation and Technology Michelle Donelan

Research

A core driver of our work is the belief that governments need to understand advanced AI to inform policy decisions and to enable public accountability.  

Because of this, we have focused on building in-house capabilities to test the safety of advanced AI systems such as large language model assistants. We aim to conduct rigorous, trustworthy assessments of advanced AI systems before and after they are launched.  

We are currently building and running evaluations for: 

  • Misuse: How much models could assist with dual-use cyber, chemical and biological attacks
  • Societal impacts: How models could affect our social fabric, e.g. by weakening democracy, harming individual welfare and perpetuating unequal outcomes
  • Autonomy: How well models could conduct AI research & development, autonomously make copies of themselves, interact with and manipulate humans and evade human intervention  
  • Safeguards: How effective safety and security features of advanced AI systems are against attempts to circumvent them

By evaluating these risks now, we can help governments assess their significance and get ahead of them.

We have open-sourced our testing framework Inspect so the research community can use and build upon our work.   

We are also pursuing research beyond evaluations to mitigate risks and make AI more publicly beneficial, such as research to make AI systems fundamentally safer and to increase societal resilience to advanced AI.

Setting the global standard

The UK has driven the global conversation on AI governance. We have already shaped how companies and other governments are navigating the technology. To date, we have:

AI Safety Summit
We contributed to the first AI Safety Summit hosted by the UK at Bletchley Park and the follow-up summit hosted by the Republic of Korea in Seoul. The summits are bringing together world leaders, top AI companies and civil society to make unprecedented commitments to mitigate risks.
Partnered with the US AI Safety Institute
We partnered with the US AI Safety Institute to jointly test advanced AI models, share research insights, share model access, and enable expert talent transfers.
International Scientific Report on the Safety of Advanced AI
We commissioned Yoshua Bengio to chair the International Scientific Report on the Safety of Advanced AI, an evidence-based report on the state of the science of advanced AI safety.

We are excited by the amount our team has been able to accomplish — and we are just getting started.