Knowledge in Big Data technologies, solutions, and concepts (Spark, Hive, MapReduce) Experience with Airflow, Spark, AWS and Databricks Strong foundation in software engineering principles, with experience working on data-centric systems Proficiency in Python, or one of the main programming languages Strong knowledge in optimizing SQL query performance Experience in building multidimensional data models (Star and/or Snowflake schema) Understanding of the data lifecycle and related concepts Knowledge in infrastructure areas such as containers and orchestration (Kubernetes, ECS) Excellent communication skills