6h ago
ML/Research Engineer, Safeguards
San Francisco, CA; New York City, NY
$350,000-$500,000 / year
full-timeseniorArtificial Intelligence Visa Sponsor
Tech Stack
Description
You will build systems to detect and mitigate misuse of AI systems, including developing classifiers for harmful behavior, monitoring multi-exchange attacks, improving agentic product safety, and conducting research on red-teaming and adversarial robustness.
Requirements
- 4+ years experience in ML engineering, research engineering, or applied research
- Proficiency in Python and experience building ML systems
- Comfortable working across research-to-deployment pipeline
- Strong communication skills to explain complex technical concepts to non-technical stakeholders
- Bachelor's degree in relevant field or equivalent
Responsibilities
- Develop classifiers to detect misuse and anomalous behavior at scale, including synthetic data pipelines and representative evaluations
- Build systems to monitor harms spanning multiple exchanges (e.g., coordinated cyber attacks, influence operations)
- Evaluate and improve safety of agentic products, including threat models and mitigations for prompt injection
- Conduct research on automated red-teaming and adversarial robustness
0 views 0 saves 0 applications