Apply

Principal Data Engineer

Posted about 9 hours agoViewed

View full description

💎 Seniority level: Principal, 10 years +

📍 Location: United Kingdom

🔍 Industry: Insurance

🏢 Company: external

🗣️ Languages: English

⏳ Experience: 10 years +

🪄 Skills: PythonSQLAgileBusiness IntelligenceETLMicrosoft Power BIApache KafkaAzureData engineeringCI/CDData modelingData analytics

Requirements:
  • Extensive experience of designing and building end to end data solutions (10 years +)
  • Experience of carrying out data engineering design and build activities using agile working practices
  • Experience of Databricks solutions, Databricks administration and pyspark
  • Data Factory/Synapse Workspace – for building data pipelines or synapse analytics pipelines
  • Data Lake – Delta Lake design pattern implementation experience in Azure Data Lake Gen2
  • Synapse Warehouse/Analytics – Experience in Synapse data mappings, external tables, schema creation from SSMS, knowledge on how Synapse pool works behind the scenes
  • Azure Active Directory – for Managed identities creation and usage or for generating service principles for authentication and authorization
  • Version Control – Experience in building Data Ops i.e., CICD pipelines in Azure DevOps with managed identity
  • Unit Testing – Experience in writing unit tests for data pipelines
  • Data Architecture – Knowledge or experience in implementing, Kimball style Data Warehouse
  • Data Quality – Experience in applying Data Quality rules within Azure Data Flow Activities
  • Data Transformation – Extensive hands on with Azure Data Flow Activities for Cleansing, transforming, validation and quality checks
  • Azure Cloud – Knowledge and confidence in effective communication on Azure Cloud Subscriptions
Responsibilities:
  • Create or guide the low-level design of data solutions
  • Be responsible for the quality of the overall data platform(s)
  • Be responsible for coding standards, low level design and ingestion patterns for the data platform(s)
  • Develop high complexity, secure, governed, high quality, efficient data pipelines
  • Set the standards and ensure that data is cleansed, mapped, transformed and optimised for storage
  • Design and build of data observability and data quality by design into all Data pipelines
  • Build solutions that pipe transform data into data lake storage areas, physical database models and reporting structures across data lake, data warehouse, business intelligence systems and analytics applications
  • Build physical data models that are appropriately designed to meet business needs and optimise storage requirements
  • Carry out unit testing of own code, peer testing of others code
  • Ensure that effective, and appropriate documentation that brings transparency and understandability are in place for all content on the data platform(s)
  • Coach and mentor Senior Data Engineers, Data engineers & Associate Data Engineers
  • Create high complexity BI solutions
Apply

Related Jobs

Apply
🔥 Principal Data Engineer (m/f/d)
Posted about 2 months ago

📍 Europe

🧭 Full-Time

🔍 Supply Chain Risk Analytics

🏢 Company: Everstream Analytics👥 251-500💰 $50,000,000 Series B almost 2 years agoProductivity ToolsArtificial Intelligence (AI)LogisticsMachine LearningRisk ManagementAnalyticsSupply Chain ManagementProcurement

  • Deep understanding of Python, including data manipulation and analysis libraries like Pandas and NumPy.
  • Extensive experience in data engineering, including ETL, data warehousing, and data pipelines.
  • Strong knowledge of AWS services, such as RDS, Lake Formation, Glue, Spark, etc.
  • Experience with real-time data processing frameworks like Apache Kafka/MSK.
  • Proficiency in SQL and NoSQL databases, including PostgreSQL, Opensearch, and Athena.
  • Ability to design efficient and scalable data models.
  • Strong analytical skills to identify and solve complex data problems.
  • Excellent communication and collaboration skills to work effectively with cross-functional teams.
  • Manage and grow a remote team of data engineers based in Europe.
  • Collaborate with Platform and Data Architecture teams to deliver robust, scalable, and maintainable data pipelines.
  • Lead and own data engineering projects, including data ingestion, transformation, and storage.
  • Develop and optimize real-time data processing pipelines using technologies like Apache Kafka/MSK or similar.
  • Design and implement data lakehouses and ETL pipelines using AWS services like Glue or similar.
  • Create efficient data models and optimize database queries for optimal performance.
  • Work closely with data scientists, product managers, and engineers to understand data requirements and translate them into technical solutions.
  • Mentor junior data engineers and share your expertise. Establish and promote best practices.

AWSPostgreSQLPythonSQLETLApache KafkaNosqlSparkData modeling

Posted about 2 months ago
Apply