2+ years of experience with transformer architectures and LLM fine-tuning Hands-on experience with RLHF/RLAIF, reward modeling, and multi-objective optimization Deep knowledge of auto-prompting, chain-of-thought evaluation, and self-improving agent loops Experience improving throughput, quantization, and token efficiency on GPUs or specialized inference hardware Knowledge of PII masking, data encryption, and secure model pipelines Experience with frameworks such as PyTorch, Transformers, Deep Speed, Hugging Face, LangChain, or vLLM