2+ years of practical experience (professional, internship, and bootcamps count) Solid Python and working SQL proficiency Familiarity with cloud platforms (AWS/GCP/Azure) and modern data tooling (e.g Airflow, dbt, Spark, Snowflake, Databricks, Redshift) Understanding of data modeling concepts (e.g., star schema, normalization) and ETL/ELT design practices Experience reading/writing Parquet Ability to write and run basic Airflow DAGs Docker fundamentals Git fundamentals, agile development and CI/CD practices Demonstrated curiosity Genuine tinkerer energy: you’ve built personal data projects for fun, tried random tools (Polars, DuckDB, Ollama, local LLMs, MotherDuck, etc.), and probably have a messy but awesome docker-compose.yml on your machine