2d ago

Researcher, Misalignment Research

San Francisco

$295k-$445k / year

full-timeseniorai-ml

๐Ÿ’ผ About This Role

You'll join the Safety Systems team at OpenAI, focusing on identifying and mitigating future AGI misalignment risks. You'll design worst-case demonstrations and rigorous evaluations to make these risks concrete and drive improvements. Your work will directly influence OpenAI's product launches and long-term safety roadmap.

๐ŸŽฏ What You'll Do

  • Design worst-case demonstrations of AGI alignment risks
  • Develop adversarial evaluations and system-level stress tests
  • Create automated tools for scalable red-teaming
  • Conduct research on failure modes of alignment techniques

๐Ÿ“‹ Requirements

  • 4+ years in AI red-teaming or related safety fields
  • Strong research track record (publications or high-impact work)
  • Fluent in modern ML/AI techniques and large-scale codebases
  • Clear communication with technical and non-technical audiences

โœจ Nice to Have

  • PhD or Master's in CS, ML, security, or related discipline
  • Experience publishing influential papers
  • Mentoring experience

๐ŸŽ Benefits & Perks

  • ๐Ÿ’ฐ Competitive compensation with equity

๐Ÿ“จ Hiring Process

Estimated timeline: 2-4 weeks ยท AI estimate

  1. 1Recruiter Callยท 30 min
  2. 2Technical Interviewยท 60 min
  3. 3Onsite Interviewsยท Half day
0 0 0