20h ago
Principal Machine Learning Engineer, Mobile AI Inference Optimization
Mountain View, CA, USA
โจ $270k-$370k / yearest.
full-timeleadgaming
๐ Tech Stack
๐ผ About This Role
You'll define the inference strategy and drive architectural decisions for deploying multi-modal AI models on mobile hardware. Your work will directly determine the latency, quality, and power profile of AI-driven features affecting billions of mobile game players. This hands-on role includes mentoring a team and pioneering mobile-optimized inference techniques.
๐ฏ What You'll Do
- Set technical vision for deploying multi-modal AI to iOS and Android
- Select and evaluate mobile inference runtimes like CoreML and TFLite
- Own end-to-end optimization pipeline from model export to kernel tuning
- Mentor ML engineers and define engineering best practices
๐ Requirements
- 8+ years in ML engineering with 3+ years on-device inference optimization
- Proven production deployment of transformer-based models on mobile hardware
- Hands-on expertise with CoreML, TFLite, ONNX Runtime, or ExecuTorch
- Expert-level command of INT8/INT4/FP16 quantization and pruning
โจ Nice to Have
- Experience shipping world-model or neural rendering pipelines on mobile
- Contributions to open-source ML inference frameworks
- Familiarity with compiler stacks such as MLIR, TVM, or XLA
๐ Benefits & Perks
- ๐ฅ Comprehensive health, life, and disability insurance
- ๐ Employee stock ownership
- ๐ป Commute subsidy
- โ๏ธ Generous vacation and personal days
- ๐ถ Support for new parents through leave and family-care programs
๐จ Hiring Process
Estimated timeline: 2-4 weeks ยท AI estimate
- 1Recruiter Screenยท 30 min
- 2Technical Interviewยท 60 min
- 3Hiring Manager Interviewยท 45 min
0 0 0