5+ years of experience in data engineering, ETL, data architecture, and cloud-based data platforms. Strong proficiency in AWS (S3, RDS, Redshift, Glue, IAM, DynamoDB). Deep understanding of database design, query optimization, and data modelling. Expertise in distributed data processing frameworks (Spark, Databricks, Presto). Strong Python and PySpark development skills. Experience with CI/CD pipelines (Jenkins) and data workflow automation (Airflow, Databricks Jobs). Proven ability to build and optimize large-scale data infrastructure. Strong problem-solving skills and experience mentoring engineers.