12h ago
Researcher, Alignment Oversight
San Francisco, CA
$250k-$445k / year
full-time Hybridai-ml
๐ผ About This Role
You'll design and run experiments to improve oversight of increasingly capable AI models as part of the Alignment Oversight team at OpenAI. You'll combine longer-horizon research with hands-on deployment, building practical oversight systems used today. This role offers the chance to produce externally publishable research and directly shape future model behavior.
๐ฏ What You'll Do
- Design and implement alignment experiments for oversight systems
- Deploy systems for action monitoring, red-teaming, and human-in-the-loop control
- Develop evaluations for alignment failure modes of frontier models
- Analyze deployment data to understand model failures and oversight gaps
๐ Requirements
- Hands-on experience training or evaluating large ML models, especially LLMs
- Experience with reinforcement learning or post-training methods
- Strong engineering execution to turn research ideas into reliable systems
- Ability to move quickly from research intuition to working experiments
โจ Nice to Have
- Experience with scalable oversight or model evaluation
- Familiarity with human-in-the-loop control systems
- Published research in alignment or AI safety
๐ Benefits & Perks
- ๐ฐ Competitive salary and equity
- ๐๏ธ Hybrid work model (3 days in office per week)
- ๐ Relocation assistance
- ๐ฅ Comprehensive benefits (implied by equal opportunity employer)
๐จ Hiring Process
Estimated timeline: 2-4 weeks ยท AI estimate
- 1Recruiter callยท 30 min
- 2Technical interviewยท 60 min
- 3On-site interviewsยท Full day
0 0 0