20h ago

Machine Learning Engineer - Inference / Serving

Remote (US + Canada Only)

$180k-$275k / year

full-timesenior Remoteai-ml

๐Ÿ›  Tech Stack

๐Ÿ’ผ About This Role

You'll design, optimize, and operate systems that bring Behavioral AI models to life in real time. You'll work at the core of production, turning trained models into performant services powering open-web and CTV products. This is an applied ML systems role combining engineering depth, deployment craft, and model intuition.

๐ŸŽฏ What You'll Do

  • Design and optimize model serving infrastructure for low-latency inference
  • Implement versioning, rollouts, rollback, and live experimentation for models
  • Monitor model drift, lineage, and observability in production
  • Collaborate with researchers to make models more deployable

๐Ÿ“‹ Requirements

  • Deep expertise in model deployment for production ML serving systems
  • Low-latency mindset with skills in graph optimization, quantization, caching
  • Proficiency in Go, Rust, C++, or Java with Python integration
  • Operational maturity in monitoring drift and model lineage

โœจ Nice to Have

  • Experience with model registries and lightweight orchestration
  • Applied ML understanding to reason about model performance trade-offs

๐ŸŽ Benefits & Perks

  • ๐Ÿ’ฐ Competitive Base Salary $180K-$275K
  • ๐Ÿ“ˆ Meaningful equity and annual bonus target
  • ๐Ÿฅ Health, Dental, Vision with low out-of-pocket
  • ๐Ÿ–๏ธ Unlimited PTO
  • ๐Ÿฆ 401k with company match

๐Ÿ“จ Hiring Process

Estimated timeline: 2-4 weeks ยท AI estimate

  1. 1Recruiter Screenยท 30 min
  2. 2Technical Interviewยท 60 min
  3. 3Team Fit Interviewยท 45 min
0 0 0