Deploy, fine-tune, and serve ML models in production environments. Work hands-on with customer data, model architectures, training loops, and inference stacks. Debug performance issues across training, evaluation, latency, cost, and reliability. Adapt the platform to customer-specific workflows and constraints. Build and maintain model-serving pipelines (batch and real-time). Optimize inference performance (throughput, latency, cost). Help productionize evaluation, monitoring, and retraining workflows. Work across cloud infrastructure, GPUs, and ML tooling stacks. Act as the “voice of the customer” to internal product and engineering teams. Identify recurring patterns, edge cases, and gaps in the platform. Contribute to internal tooling, templates, and best practices.