6+ years of experience in software development. At least 2+ years of experience in building scalable, production-grade data pipelines. Expert-level proficiency in SQL. Expert-level proficiency in distributed data processing tools like Spark, Flink, or Kafka Streams. Strong programming foundations in Python or similar language. Good software engineering design patterns and principles (testing, CI/CD, monitoring). Expert in modern data lakes (e.g., Delta Lake, Iceberg). Familiar with data warehouses (e.g., Snowflake, Redshift, BigQuery). Familiar with production data stores such as relational (e.g., MySQL, PostgreSQL), object (e.g., S3), key-value (e.g., DynamoDB), and message queues (e.g., Kinesis, Kafka). Excellent collaboration and communication skills. Understanding of data modeling and metric design principles. Bachelor’s degree in Computer Science, Computer Engineering, or related field, or equivalent experience.