4+ years of professional software experience in production environments 4+ years building or maintaining large-scale, production-grade data platforms (including data lakes) Strong expertise with Apache Spark Strong AWS experience (S3, Lambda, SQS, Kinesis) Experience with Databricks, Delta Lake, or similar Proficiency with Infrastructure-as-Code (e.g. Terraform, CloudFormation) Familiarity with data catalogs and discovery tools Experience with streaming data, event-driven architectures, or handling late-arriving data Familiarity with containerization and orchestration (Docker, Kubernetes, Airflow, Dagster, or Prefect)