6h ago
Researcher, Misalignment Research
New York City
$380k-$445k / year
full-timesenior Hybridai-ml
💼 About This Role
You'll join Safety Systems to identify and mitigate future AGI misalignment risks by designing worst-case demonstrations and adversarial evaluations. Your work will directly influence product safety and lower the chance of catastrophic outcomes. You'll collaborate with world-class researchers and have access to cutting-edge models.
🎯 What You'll Do
- Design worst-case demonstrations of AGI alignment risks
- Develop adversarial and system-level safety evaluations
- Build automated red-teaming infrastructure and tools
- Conduct research on alignment technique failure modes
📋 Requirements
- 4+ years experience in AI red-teaming, security research, or adversarial ML
- Strong research track record with publications or high-impact internal work
- Fluency in modern ML/AI techniques and large-scale codebases
- Ability to communicate complex findings to technical and non-technical audiences
✨ Nice to Have
- PhD in computer science, machine learning, security, or related field
- Experience with automated red-teaming and evaluation infrastructure
- Cross-functional collaboration skills spanning research, engineering, and policy
🎁 Benefits & Perks
- 🚀 Shape safety practices at the AGI frontier
- 🔬 Access to cutting-edge models, tooling, and compute
- 🤝 Collaborative, mission-driven environment
- 💰 Competitive compensation, equity, and benefits
0 0 0