Machine Learning Researcher, Multimodal LLMs

New
B
BlandAI, Voice
San Francisco, CA or Remote (US)Full-TimeMiddle
Salary180000 - 260000 USD per year
Apply NowOpens the employer's application page

Job Details

Requirements

  • Experience with LLMs
  • Experience with multimodal models
  • Experience with speech-language systems
  • Deep understanding of prompting techniques
  • Deep understanding of fine-tuning techniques
  • Deep understanding of alignment techniques
  • Familiarity with neural audio codecs
  • Familiarity with modern multimodal LLM techniques
  • Ability to go from idea → dataset → experiment → conclusion in days
  • Strong sense for what makes an interaction feel natural vs robotic
  • Ability to translate abstract modeling ideas into user-facing improvements
  • Take ownership from research through deployment
  • Thrive in ambiguous, fast-moving environments
  • Care about impact, not just elegance
  • Think in systems, not just models
  • Obsess over latency, correctness, and real-world behavior
  • Comfortable discarding ideas quickly when data disagrees
  • Push toward simple abstractions for complex problems

Responsibilities

  • Contribute to the development of our next-generation multimodal LLM stack
  • Combine speech, text, tools, and real-time reasoning into a single unified system
  • Build industry-leading conversational AI models that power Bland's agent
  • Take models from idea to production
  • Define how agents listen, think, and act in real time
  • Integrate streaming audio, tool execution, and dynamic context into a single coherent system
  • Take ideas from research through production systems serving millions of calls per day
View Full Description & ApplyYou'll be redirected to the employer's site
180000 - 260000 USD per year
Apply Now