Several years of experience developing in a modern programming language, preferably Java and Python.
Significant experience with developing and maintaining distributed big data systems with production quality deployment and monitoring.
Exposure to high-performance data pipelines, preferably with Apache Kafka & Spark.
Experience with scheduling systems such as Airflow and SQL/NoSQL Databases.
Experience with cloud data platforms is a plus.
Exposure to Docker and/or Kubernetes is preferred.
Good command of spoken and written English.
University degree in computer sciences or equivalent professional experience.
Responsibilities:
Develop new data pipelines and maintain our data ecosystem, focusing on fault-tolerant ingestion, storage and data lifecycle, and the computation of metrics, reports and derived information.
Communicate efficiently with your team-mates in order to develop software and creative solutions for our customer needs.
Write high-quality, reusable code, test it, and bring it to production.
Familiarity with applying best practices according to industry standards whilst promoting a culture of agility and excellence.