BSc degree in Computer Science, Information Systems, Engineering, or related technical field or equivalent work experience. 3+ years related work experience. Minimum of 2 years experience building and optimizing ‘big data’ data pipelines, architectures and maintaining data sets. Experienced in Python. Experienced in SQL (PostgreSQL, MS SQL). Experienced in using cloud services: AWS, Azure or GCP. Proficiency in version control, CI/CD and GitHub. Understanding/experience in Glue and PySpark highly desirable. Experience in managing data life cycle. Proficiency in manipulating, processing and architecting large disconnected data sets for analytical requirements. Ability to maintain and optimise processes supporting data transformation, data structures, metadata, dependency and workload management. Good understanding of data management principles - data quality assurance and governance. Strong analytical skills related to working with unstructured datasets. Understanding of message queuing, stream processing, and highly scalable ‘big data’ datastores. Strong attention to detail. Good communication and interpersonal skills.