3h ago
Research Engineer / Scientist, Alignment Science
London, UK
full-timesenior HybridArtificial Intelligence
Tech Stack
Description
You will design and run machine learning experiments to understand and steer the behavior of powerful AI systems, contributing to AI safety research in areas like AI control and alignment stress-testing. Your work will involve building tooling, evaluating jailbreaks, and collaborating with teams to mitigate risks from advanced AI.
Requirements
- Significant software, ML, or research engineering experience
- Experience contributing to empirical AI research projects
- Familiarity with technical AI safety research
- Preference for fast-moving collaborative projects
Responsibilities
- Run multi-agent reinforcement learning experiments for AI safety
- Build tooling to evaluate effectiveness of LLM-generated jailbreaks
- Contribute to research papers, blog posts, and talks
- Test robustness of safety techniques by training models to subvert them
- Collaborate on safety-relevant projects with teams like Interpretability and Red Team
0 views 0 saves 0 applications