Design, build, and maintain production-grade ML systems, from data ingestion and processing to model deployment and monitoring. Develop and fine-tune generative AI models, including LLMs, for specialized tasks. Architect and implement reliable data pipelines and low-latency inference services using our core stack (FastAPI, Docker, Kubeflow, AWS/GCP). Collaborate with senior engineers, researchers, and client stakeholders to translate business problems into technical solutions. Take ownership of key components of our ML platform, ensuring code quality, performance, and scalability.