20h ago
Machine Learning Engineer - Inference / Serving
Remote (US + Canada Only)
$180k-$275k / year
full-timesenior Remoteai-ml
๐ Tech Stack
๐ผ About This Role
You'll design, optimize, and operate systems that bring Behavioral AI models to life in real time. You'll work at the core of production, turning trained models into performant services powering open-web and CTV products. This is an applied ML systems role combining engineering depth, deployment craft, and model intuition.
๐ฏ What You'll Do
- Design and optimize model serving infrastructure for low-latency inference
- Implement versioning, rollouts, rollback, and live experimentation for models
- Monitor model drift, lineage, and observability in production
- Collaborate with researchers to make models more deployable
๐ Requirements
- Deep expertise in model deployment for production ML serving systems
- Low-latency mindset with skills in graph optimization, quantization, caching
- Proficiency in Go, Rust, C++, or Java with Python integration
- Operational maturity in monitoring drift and model lineage
โจ Nice to Have
- Experience with model registries and lightweight orchestration
- Applied ML understanding to reason about model performance trade-offs
๐ Benefits & Perks
- ๐ฐ Competitive Base Salary $180K-$275K
- ๐ Meaningful equity and annual bonus target
- ๐ฅ Health, Dental, Vision with low out-of-pocket
- ๐๏ธ Unlimited PTO
- ๐ฆ 401k with company match
๐จ Hiring Process
Estimated timeline: 2-4 weeks ยท AI estimate
- 1Recruiter Screenยท 30 min
- 2Technical Interviewยท 60 min
- 3Team Fit Interviewยท 45 min
0 0 0