2h ago
Research Scientist, Multimodal AI
San Francisco
$350k-$475k / year
full-timeseniorArtificial Intelligence Visa Sponsor
🛠 Tech Stack
💼 About This Role
You'll own research projects on multimodal AI models and large-scale dataset curation at Thinking Machines Lab. Your work will directly shape how AI understands vision and language together. This role blends fundamental research with practical engineering in a unified team.
🎯 What You'll Do
- Own research projects on training and performance analysis of multimodal models.
- Curate and build large-scale datasets and evaluation benchmarks.
- Collaborate with engineers to create frontier multimodal models and products.
- Publish research and share code, datasets, and insights.
📋 Requirements
- Ability to design, run, and analyze experiments with research judgment and empirical rigor.
- Understanding of machine learning fundamentals, large-scale training, and distributed compute.
- Proficiency in Python and familiarity with a deep learning framework (PyTorch, TensorFlow, or JAX).
- Bachelor's degree or equivalent in Computer Science, Machine Learning, Physics, or Mathematics.
✨ Nice to Have
- Research or engineering contributions in visual reasoning or multimodal architecture design.
- Experience developing evaluation frameworks for multimodal tasks.
- Publications or open-source contributions in vision-language modeling or multimodal AI.
🎁 Benefits & Perks
- 🏥 Health, dental, and vision benefits
- 🏖️ Unlimited PTO
- 👶 Paid parental leave
- 🚚 Relocation support
📨 Hiring Process
We continuously review applications and reach out as new opportunities open; you will hear from us if there is a match.
🚩 Heads Up
- Role blends research and engineering without clear distinction, which may lead to ambiguous expectations.
- Evergreen role may not have an immediate opening, causing potential delay or no response.
- Requires both deep theoretical exploration and hands-on experimentation, which may be demanding.
0 0 0