3h ago
Staff Backend Engineer, ML Inference Systems
Mountain View, CA
$192.6k-$305.6k / year
full-timeseniorgaming
๐ Tech Stack
๐ผ About This Role
You'll design and operate distributed systems that power billions of daily real-time ML inference decisions, partnering with ML engineers to scale serving infrastructure. You'll drive technical direction on low-latency, high-throughput platforms using GCP, Kubernetes, Prometheus, and Grafana. Join a team influencing how billions of gaming experiences are discovered and monetized.
๐ฏ What You'll Do
- Design and deploy production-grade backend services for large-scale online model inference
- Drive technical direction of the inference platform focusing on low-latency, high-throughput serving
- Partner with ML engineers to ensure serving infrastructure scales with model complexity
- Ensure reliability, scalability, and efficiency using Prometheus, Grafana, GCP, and Kubernetes
๐ Requirements
- 5+ years designing and maintaining distributed systems at scale
- Expertise in Golang for high-performance backend infrastructure
- Hands-on experience with GCP and Kubernetes for workload orchestration
- Strong grounding in monitoring and observability with Prometheus and Grafana
โจ Nice to Have
- Experience with ML inference servers like NVIDIA Triton Inference Server
- Familiarity with auction mechanics or bidding systems in ad tech
- Experience embracing AI as a strategic advantage in engineering
๐ Benefits & Perks
- ๐๏ธ Generous vacation and personal days
- ๐ฅ Comprehensive health, life, and disability insurance
- ๐ฐ Employee stock ownership
- ๐ถ Support for new parents through leave and family-care programs
- ๐ง Mental Health and Wellbeing programs and support
๐จ Hiring Process
Estimated timeline: 3-5 weeks ยท AI estimate
- 1Recruiter Screenยท 30 min
- 2Technical Interviewยท 60 min
- 3System Design Interviewยท 60 min
- 4Hiring Manager Interviewยท 45 min
0 0 0