2d ago
Researcher, Misalignment Research
San Francisco
$295k-$445k / year
full-timeseniorai-ml
๐ผ About This Role
You'll join the Safety Systems team at OpenAI, focusing on identifying and mitigating future AGI misalignment risks. You'll design worst-case demonstrations and rigorous evaluations to make these risks concrete and drive improvements. Your work will directly influence OpenAI's product launches and long-term safety roadmap.
๐ฏ What You'll Do
- Design worst-case demonstrations of AGI alignment risks
- Develop adversarial evaluations and system-level stress tests
- Create automated tools for scalable red-teaming
- Conduct research on failure modes of alignment techniques
๐ Requirements
- 4+ years in AI red-teaming or related safety fields
- Strong research track record (publications or high-impact work)
- Fluent in modern ML/AI techniques and large-scale codebases
- Clear communication with technical and non-technical audiences
โจ Nice to Have
- PhD or Master's in CS, ML, security, or related discipline
- Experience publishing influential papers
- Mentoring experience
๐ Benefits & Perks
- ๐ฐ Competitive compensation with equity
๐จ Hiring Process
Estimated timeline: 2-4 weeks ยท AI estimate
- 1Recruiter Callยท 30 min
- 2Technical Interviewยท 60 min
- 3Onsite Interviewsยท Half day
0 0 0