BS degree in a STEM field. 4 years+ experience in a data engineering or python development role. In-depth understanding of backend data engineering, relational database design, ETL and ELT pipelines, building scalable near-real time data streams, batch processing. 2+ years of experience working in Cloud Systems (GCP, AWS, Azure). Experience with Apache Spark, Apache Airflow, Dagster (in Python or Scala). Experience with geospatial data processing tools and databases is a plus. Experience with multiprocessing, multithreading, query optimization, and Cython is desirable. The ability to work effectively in a team environment with remotely located staff in different time zones, demonstrating strong communication and coordination skills. Exposure to energy, oil & gas domain is preferred. API-first design, internal package development, and SQL database optimization. Data pipeline development, ETL processes, and real-time streaming systems. Performance optimization including multiprocessing, multithreading, query optimization. GitHub Actions and advanced CI/CD pipeline development.