Proficiency in Python and related ML frameworks such as Tensorflow, TF-Serving, JAX, and XLA/MLIR Experience using large-scale distributed training strategies Familiarity with autoregressive sequence models, such as Transformers Strong communication and problem-solving skills Demonstrated passion for applied NLP models and products Experience writing kernels for GPUs using CUDA (bonus) Experience training on TPUs (bonus) Papers at top-tier venues (bonus)