1h ago

Research Engineer / Research Scientist - Red Team

London, UK
Government AI Research

Description

You will be part of the Alignment Red Team at the AI Security Institute, focusing on detecting and understanding misalignment in frontier AI systems. You will develop novel techniques for finding misaligned behavior, perform pre- and post-deployment evaluations, and share findings with AI companies and governments to inform safety improvements.

Requirements

  • Expertise in AI alignment, red teaming, or related fields
  • Experience with frontier AI model evaluation and testing
  • Ability to conduct independent research and collaborate with cross-functional teams
  • Strong understanding of AI safety risks and mitigation strategies

Responsibilities

  • Develop novel research techniques for detecting misalignment in frontier AI systems
  • Attribute misaligned behavior to fundamental alignment concerns like instrumental convergence
  • Conduct pre- and post-deployment evaluations of frontier AI systems for loss-of-control risks
  • Share evaluation findings with AI companies and governments to inform research and policy-making
  • Work with safety teams at frontier labs to improve model alignment training and monitoring
0 views 0 saves 0 applications