B.S. in Computer Science or equivalent experience. 8+ years building and operating production data platforms. 4+ years deep, hands-on Databricks/Spark (PySpark + SQL). Proven ownership of a production lakehouse (S3 + Delta Lake) with strict SLAs and compliance requirements. Expertise with Delta Lake (MERGE/CDC, schema evolution, time travel, OPTIMIZE/Z-ORDER, VACUUM) and DLT, Workflows, Auto Loader. Feature Store experience in production. Strong data modeling (dimensional, canonical), SCD Types 1/2, and handling slowly changing entities and schema drift. Track record delivering trustworthy datasets with monitoring, alerting, lineage, and clear documentation. Able to define and maintain metric layers consumed by product and business. Advanced Python and SQL. Testing culture (pytest), CI/CD (GitHub Actions), and Terraform for Databricks. Solid Git practices. AWS foundations: S3, IAM, networking basics; event ingestion. Excellent communication and leadership skills. Able to drive design reviews, write clear technical docs, and mentor engineers in a remote, async environment.