3h ago

Staff Backend Engineer, ML Inference Systems

Mountain View, CA

$192.6k-$305.6k / year

full-timeseniorgaming

๐Ÿ›  Tech Stack

๐Ÿ’ผ About This Role

You'll design and operate distributed systems that power billions of daily real-time ML inference decisions, partnering with ML engineers to scale serving infrastructure. You'll drive technical direction on low-latency, high-throughput platforms using GCP, Kubernetes, Prometheus, and Grafana. Join a team influencing how billions of gaming experiences are discovered and monetized.

๐ŸŽฏ What You'll Do

  • Design and deploy production-grade backend services for large-scale online model inference
  • Drive technical direction of the inference platform focusing on low-latency, high-throughput serving
  • Partner with ML engineers to ensure serving infrastructure scales with model complexity
  • Ensure reliability, scalability, and efficiency using Prometheus, Grafana, GCP, and Kubernetes

๐Ÿ“‹ Requirements

  • 5+ years designing and maintaining distributed systems at scale
  • Expertise in Golang for high-performance backend infrastructure
  • Hands-on experience with GCP and Kubernetes for workload orchestration
  • Strong grounding in monitoring and observability with Prometheus and Grafana

โœจ Nice to Have

  • Experience with ML inference servers like NVIDIA Triton Inference Server
  • Familiarity with auction mechanics or bidding systems in ad tech
  • Experience embracing AI as a strategic advantage in engineering

๐ŸŽ Benefits & Perks

  • ๐Ÿ–๏ธ Generous vacation and personal days
  • ๐Ÿฅ Comprehensive health, life, and disability insurance
  • ๐Ÿ’ฐ Employee stock ownership
  • ๐Ÿ‘ถ Support for new parents through leave and family-care programs
  • ๐Ÿง  Mental Health and Wellbeing programs and support

๐Ÿ“จ Hiring Process

Estimated timeline: 3-5 weeks ยท AI estimate

  1. 1Recruiter Screenยท 30 min
  2. 2Technical Interviewยท 60 min
  3. 3System Design Interviewยท 60 min
  4. 4Hiring Manager Interviewยท 45 min
0 0 0