New

Research Manager – Alignment

London, UK

About the AI Security Institute

The AI Security Institute is the largest team in a government dedicated to understanding AI capabilities and risks in the world. 

Our mission is to equip governments with an empirical understanding of the safety of advanced AI systems. We conduct research to understand the capabilities and impacts of advanced AI and develop and test risk mitigations. We focus on risks with security implications, including the potential of AI to assist with the development of chemical and biological weapons, how it can be used to carry out cyber-attacks, enable crimes such as fraud, and the possibility of loss of control. 

The risks from AI are not sci-fi, they are urgent. By combining the agility of a tech start-up with the expertise and mission-driven focus of government, we’re building a unique and innovative organisation to prevent AI’s harms from impeding its potential. 

This role sits outside of the DDaT pay framework given the scope of this role requires in depth technical expertise in frontier AI safety, robustness and advanced AI architectures. 

The deadline for applying this role is Sunday 15 June 2025, end of day, anywhere on Earth.

AISI’s alignment team examines ways to prevent models from autonomously attempting to cause harm. The team’s research is led by Geoffrey Irving, AISI’s Chief Scientist – a research agenda can be found here.

This exciting new team is part of AISI’s Solutions Group, which will also examine ways to prevent misuse risks (our Safeguards team) and ways to prevent models from causing harm, even if they are autonomously attempting to do so (our Control team).   

ROLE SUMMARY

In this role, you will be people managing a team of very exceptional and highly motivated individuals, including complexity theorists, game theorists, learning theorists and ML research scientists. You’ll also be growing your team – getting stuck in with sourcing specific people we want, often world-leading experts in their domain, and then helping run the full process and getting them excited about the role.

Within your team you’re expected to provide fantastic management, including building strong relationships with the other team members, as well as giving regular feedback and coaching. You’ll receive mentorship and coaching both from the workstream lead, as well as the broader group of research directors in AISI (people like Geoffrey Irving and Yarin Gal). In addition, we have a very strong learning culture including paper-reading groups and Friday afternoons dedicated to deep learning.

Your team is given huge amounts of autonomy to chase exciting research directions, including:

  • Conducting foundational research pushing forward the frontier of our understanding about how to make highly advanced AI systems safe alongside AISI expertise and external collaborators 
  • Breaking down the alignment problem by producing safety case sketches for highly-capable AI systems using alignment and scalable oversight techniques 
  • Making funding recommendations, finding the best research projects targeting the most important problems to solve – and supervising this external research

Person Specification 

You may be a good fit if you have some of the following skills, experience and attitudes:

  • Experience people managing strong research or engineering teams, with well thought-out views on management philosophy and style.
  • Experience coaching team members & providing feedback.
  • Strong understanding of large language models (e.g. GPT-4). This includes both a broad understanding of the literature, as well as hands-on experience with either research or research engineering.
  • Broad knowledge of existing approaches to alignment (T-shaped: some deep knowledge, lots of shallow knowledge).
  • Strong written and verbal communication skills.
  • A track record of helping teams achieve exceptional things.
  • Experience working with world-class multi-disciplinary teams (e.g. in a top-3 AI lab).
  • Experience working within a research team that has delivered multiple exceptional scientific breakthroughs, in deep learning (or a related field).
  • Strong track-record of academic excellence (e.g. multiple spotlight papers at top-tier conferences).
  • Acting as a bar raiser for interviews.

Salary & Benefits 

We are hiring individuals at a range of seniority and experience. Your dedicated talent partner will work with you as you move through our assessment process to explain our internal benchmarking process. For this role, we would expect candidates to be L5 and above. The range of salaries are available below, salaries comprise of a base salary, technical allowance plus additional benefits as detailed on this page.

  • Level 5 - Total Package £105,000 - £115,000 inclusive of a base salary £55,805 plus additional technical talent allowance of between £49,195 - £59,195 
  • Level 6 - Total Package £125,000 - £135,000 inclusive of a base salary £68,770 plus additional technical talent allowance of between £56,230 - £66,230 
  • Level 7 - Total Package £145,000 inclusive of a base salary £68,770 plus additional technical talent allowance of £76,230 

There are a range of pension options available which can be found through the Civil Service website.

 


Additional Information

Internal Fraud Database 

The Internal Fraud function of the Fraud, Error, Debt and Grants Function at the Cabinet Office processes details of civil servants who have been dismissed for committing internal fraud, or who would have been dismissed had they not resigned. The Cabinet Office receives the details from participating government organisations of civil servants who have been dismissed, or who would have been dismissed had they not resigned, for internal fraud. In instances such as this, civil servants are then banned for 5 years from further employment in the civil service. The Cabinet Office then processes this data and discloses a limited dataset back to DLUHC as a participating government organisations. DLUHC then carry out the pre employment checks so as to detect instances where known fraudsters are attempting to reapply for roles in the civil service. In this way, the policy is ensured and the repetition of internal fraud is prevented.  For more information please see - Internal Fraud Register.

Security

Successful candidates must undergo a criminal record check and get baseline personnel security standard (BPSS) clearance before they can be appointed. Additionally, there is a strong preference for eligibility for counter-terrorist check (CTC) clearance. Some roles may require higher levels of clearance, and we will state this by exception in the job advertisement. See our vetting charter here.

 

Nationality requirements

We may be able to offer roles to applicant from any nationality or background. As such we encourage you to apply even if you do not meet the standard nationality requirements (opens in a new window).

Working for the Civil Service

The Civil Service Code (opens in a new window) sets out the standards of behaviour expected of civil servants. We recruit by merit on the basis of fair and open competition, as outlined in the Civil Service Commission's recruitment principles (opens in a new window). The Civil Service embraces diversity and promotes equal opportunities. As such, we run a Disability Confident Scheme (DCS) for candidates with disabilities who meet the minimum selection criteria. The Civil Service also offers a Redeployment Interview Scheme to civil servants who are at risk of redundancy, and who meet the minimum requirements for the advertised vacancy.

Diversity and Inclusion

The Civil Service is committed to attract, retain and invest in talent wherever it is found. To learn more please see the Civil Service People Plan (opens in a new window) and the Civil Service Diversity and Inclusion Strategy (opens in a new window).

Apply for this job

*

indicates a required field

Resume/CV*

Accepted file types: pdf, doc, docx, txt, rtf


You are managing a team of 3 research scientists focused on AI safety via debate. They are currently working two projects:
A: some complexity theory proofs about a possible new debate protocol.
B: a study that simulates debate protocols using humans playing the part of AIs.

A research scientist on your team proposes an empirical study attempting to validate whether the stability requirement of a new debate protocol could be violated in practice. The study doesn't require human subjects. The study wold require ~25% of the team's effort for four months.

Assume that the goal of the alignment team is to make as much progress as possible on the alignment problem. Assume you do not have a background in ML theory; focus on strategic reasoning.

Please:
(a) List up to three strategic criteria you would apply to decide whether the project should take priority ongoing work, and briefly explain why each criterion matters.
(b) Then outline, step-by-step, how you would gather evidence to evaluate those criteria.

Select...
Select...
Select...
Select...
Select...

UK Diversity Questions

It's important to us that everyone at AISI feels an included part of the team, whoever they are and whatever their background. These questions will help us to identify the diversity of our applicants. Should you not wish to provide an answer, you will always have the option to not provide a response with a 'I don't wish to answer' option. Your answers will not impact your hiring outcomes whatsoever.

If there are any questions you would like to further discuss or want clarity on, we'd be happy to talk to you about this if you reach out to active.campaigns@dsit.gov.uk

Select...
Select...
Select...
Select...
Select...
Select...
Select...