2+ years of experience with transformer architectures and LLM fine-tuning (e.g., Llama, Mistral, GPT, Mixtral, Gemma, Falcon, Claude) Hands-on experience with RLHF/RLAIF, reward modeling, and multi-objective optimization for generative models Deep knowledge of auto-prompting, chain-of-thought evaluation, and self-improving agent loops Experience improving throughput, quantization, and token efficiency on GPUs or specialized inference hardware Knowledge of PII masking, data encryption, and secure model pipelines in production settings Experience with frameworks such as PyTorch, Transformers, Deep Speed, Hugging Face, LangChain, or vLLM