- Lead the technical development and continuous improvement of Omilia’s proprietary LLM and NLU service portfolio
- Ensure technical correctness, system quality, and delivery monitoring for mission-critical AI services
- Hold final technical authority over all LLM/NLU services, including entity/intent classification, specialized LLMs, and agentic orchestration
- Ensure production stability, performance, and compliance (including PCI/PII) across the LLM/NLU domain
- Commit to delivery dates, drive features from design through deployment, and proactively flag risks
- Resolve technical ambiguity, structure loosely defined requirements, and make architectural decisions independently
- Lead the most complex, ambiguous, or cross-cutting features, including model research, agentic reasoning, and inference server development
- Directly influence the quality and reliability of AI services serving millions of customer interactions in regulated industries
- Guide and mentor mid-level and junior engineers through code reviews, pairing, and knowledge transfer; drive alignment between Product, Architecture, and Engineering
- Lead research and experimentation on new model architectures, training strategies, and evaluation methodologies for LLM/NLU
- Design, develop, fine-tune, and evaluate specialized LLMs for Concierge and Task Agents
- Develop and optimize ML pipelines for training, evaluation, and deployment (AWS SageMaker)
- Architect and maintain inference servers, ensuring low latency and high reliability
- Implement and evolve closed-loop self-learning systems for continuous model improvement
- Drive benchmarking, experiment reproducibility, and documentation quality
- Ensure compliance with data privacy standards throughout the ML lifecycle
- Mentor and support the growth of team members; share expertise via tech talks and guides
PythonPyTorchFastAPI+1 more