2h ago

Research Scientist, Multimodal AI

San Francisco

$350k-$475k / year

full-timeseniorArtificial Intelligence Visa Sponsor

🛠 Tech Stack

💼 About This Role

You'll own research projects on multimodal AI models and large-scale dataset curation at Thinking Machines Lab. Your work will directly shape how AI understands vision and language together. This role blends fundamental research with practical engineering in a unified team.

🎯 What You'll Do

  • Own research projects on training and performance analysis of multimodal models.
  • Curate and build large-scale datasets and evaluation benchmarks.
  • Collaborate with engineers to create frontier multimodal models and products.
  • Publish research and share code, datasets, and insights.

📋 Requirements

  • Ability to design, run, and analyze experiments with research judgment and empirical rigor.
  • Understanding of machine learning fundamentals, large-scale training, and distributed compute.
  • Proficiency in Python and familiarity with a deep learning framework (PyTorch, TensorFlow, or JAX).
  • Bachelor's degree or equivalent in Computer Science, Machine Learning, Physics, or Mathematics.

✨ Nice to Have

  • Research or engineering contributions in visual reasoning or multimodal architecture design.
  • Experience developing evaluation frameworks for multimodal tasks.
  • Publications or open-source contributions in vision-language modeling or multimodal AI.

🎁 Benefits & Perks

  • 🏥 Health, dental, and vision benefits
  • 🏖️ Unlimited PTO
  • 👶 Paid parental leave
  • 🚚 Relocation support

📨 Hiring Process

We continuously review applications and reach out as new opportunities open; you will hear from us if there is a match.

🚩 Heads Up

  • Role blends research and engineering without clear distinction, which may lead to ambiguous expectations.
  • Evergreen role may not have an immediate opening, causing potential delay or no response.
  • Requires both deep theoretical exploration and hands-on experimentation, which may be demanding.
0 0 0