Apply

Senior Data Engineer

Posted about 1 month agoViewed

View full description

💎 Seniority level: Senior, 4+ years

📍 Location: Europe, APAC, Americas

🔍 Industry: Software Development

🏢 Company: Docker👥 251-500💰 $105,000,000 Series C about 3 years agoDeveloper ToolsDeveloper PlatformInformation TechnologySoftware

🗣️ Languages: English

⏳ Experience: 4+ years

🪄 Skills: PythonSQLETLSnowflakeAirflowData engineeringData visualizationData modeling

Requirements:
  • 4+ years of relevant industry experience
  • Experience with data modeling and building scalable pipelines
  • Proficiency with Snowflake or BigQuery
  • Experience with data governance and security controls
  • Experience creating ETL scripts using Python and SQL
  • Familiarity with a cloud ecosystem: AWS/Azure/Google Cloud
  • Experience with Tableau or Looker
Responsibilities:
  • Manage and develop ETL jobs, warehouse, and event collection tools
  • Build and manage the Central Data Model for reporting
  • Integrate emerging methodologies and technologies
  • Build data pipelines for ML and AI projects
  • Contribute to SOC2 compliance across the data platform
  • Document technical architecture
Apply

Related Jobs

Apply

📍 Germany, Italy, Netherlands, Portugal, Romania, Spain, UK

🧭 Full-Time

🔍 Wellness

  • You have a proven track record of designing and building robust, scalable, and maintainable data models and corresponding pipelines from business requirements.
  • You are skilled at engaging with engineering and product teams to elicit requirements.
  • You are comfortable with big data concepts, ensuring data is efficiently ingested, processed, and made available for data scientists, business analysts, and product teams.
  • You are experienced in maintaining data consistency across the entire data ecosystem.
  • You have experience maintaining and debugging data pipelines in production environments with high criticality, ensuring reliability and performance.
  • Develop and maintain efficient and scalable data models and structures to support analytical workloads.
  • Design, develop, and maintain data pipelines that transform and process large volumes of data while embedding business context and semantics.
  • Implement automated data quality checks to ensure consistency, accuracy, and reliability of data.
  • Ensure correct adoption and usage of Wellhub’s data by data practitioners across the company
  • Live the mission: inspire and empower others by genuinely caring for your own wellbeing and your colleagues. Bring wellbeing to the forefront of work, and create a supportive environment where everyone feels comfortable taking care of themselves, taking time off, and finding work-life balance.

SQLApache AirflowKubernetesApache KafkaData engineeringSparkData modeling

Posted 1 day ago
Apply
Apply

📍 Portugal

🧭 Full-Time

🏢 Company: Wellhub

  • Proven track record of designing and building robust, scalable, and maintainable data models and corresponding pipelines from business requirements.
  • Skilled at engaging with engineering and product teams to elicit requirements.
  • Comfortable with big data concepts, ensuring data is efficiently ingested, processed, and made available for data scientists, business analysts, and product teams.
  • Experienced in maintaining data consistency across the entire data ecosystem.
  • Experience maintaining and debugging data pipelines in production environments with high criticality, ensuring reliability and performance.
  • Motivated to contribute to a data-driven culture and take pride in seeing the impact of your work across the company
  • Develop and maintain efficient and scalable data models and structures to support analytical workloads.
  • Design, develop, and maintain data pipelines that transform and process large volumes of data while embedding business context and semantics.
  • Implement automated data quality checks to ensure consistency, accuracy, and reliability of data.
  • Ensure correct adoption and usage of Wellhub’s data by data practitioners across the company
  • Live the mission: inspire and empower others by genuinely caring for your own wellbeing and your colleagues. Bring wellbeing to the forefront of work, and create a supportive environment where everyone feels comfortable taking care of themselves, taking time off, and finding work-life balance.

SQLApache AirflowETLKubernetesApache KafkaData engineeringSparkData visualizationData modelingData analyticsData management

Posted 1 day ago
Apply
Apply
🔥 Senior Data Engineer
Posted 3 days ago

📍 Worldwide

🔍 Hospitality

🏢 Company: Lighthouse

  • 4+ years of professional experience using Python, Java, or Scala for data processing (Python preferred)
  • You stay up-to-date with industry trends, emerging technologies, and best practices in data engineering.
  • Improve, manage, and teach standards for code maintainability and performance in code submitted and reviewed
  • Ship large features independently, generate architecture recommendations and have the ability to implement them
  • Great communication: Regularly achieve consensus amongst teams
  • Familiarity with GCP, Kubernetes (GKE preferred),  CI/CD tools (Gitlab CI preferred), familiarity with the concept of Lambda Architecture.
  • Experience with Apache Beam or Apache Spark for distributed data processing or event sourcing technologies like Apache Kafka.
  • Familiarity with monitoring tools like Grafana & Prometheus.
  • Design and develop scalable, reliable data pipelines using the Google Cloud stack.
  • Optimise data pipelines for performance and scalability.
  • Implement and maintain data governance frameworks, ensuring data accuracy, consistency, and compliance.
  • Monitor and troubleshoot data pipeline issues, implementing proactive measures for reliability and performance.
  • Collaborate with the DevOps team to automate deployments and improve developer experience on the data front.
  • Work with data science and analytics teams to enable them to bring their research to production grade data solutions, using technologies like airflow, dbt or MLflow (but not limited to)
  • As a part of a platform team, you will communicate effectively with teams across the entire engineering organisation, to provide them with reliable foundational data models and data tools.
  • Mentor and provide technical guidance to other engineers working with data.

PythonSQLApache AirflowETLGCPKubernetesApache KafkaData engineeringCI/CDMentoringTerraformScalaData modeling

Posted 3 days ago
Apply
Apply

📍 Worldwide

🧭 Full-Time

NOT STATED
  • Own the design and implementation of cross-domain data models that support key business metrics and use cases.
  • Partner with analysts and data engineers to translate business logic into performant, well-documented dbt models.
  • Champion best practices in testing, documentation, CI/CD, and version control, and guide others in applying them.
  • Act as a technical mentor to other analytics engineers, supporting their development and reviewing their code.
  • Collaborate with central data platform and embedded teams to improve data quality, metric consistency, and lineage tracking.
  • Drive alignment on model architecture across domains—ensuring models are reusable, auditable, and trusted.
  • Identify and lead initiatives to reduce technical debt and modernise legacy reporting pipelines.
  • Contribute to the long-term vision of analytics engineering at Pleo and help shape our roadmap for scalability and impact.

SQLData AnalysisETLData engineeringCI/CDMentoringDocumentationData visualizationData modelingData analyticsData management

Posted 4 days ago
Apply
Apply

📍 Germany, Austria, Italy, Spain, Portugal

🔍 Financial and Real Estate

🏢 Company: PriceHubble👥 101-250💰 Non-equity Assistance over 3 years agoArtificial Intelligence (AI)PropTechBig DataMachine LearningAnalyticsReal Estate

  • 3+ years of experience building and maintaining production data pipelines.
  • Excellent English communication skills, both spoken and written, to effectively collaborate with cross-functional teams and mentor other engineers.
  • Clear writing is key in our remote-first setup.
  • Proficient in working with geospatial data and leveraging geospatial features.
  • Work with backend engineers and data scientists to turn raw data into trusted insights, handling everything from scraping and ingestion to transformation and monitoring.
  • Navigate cost-value trade-offs to make decisions that deliver value to customers at an appropriate cost.
  • Develop solutions that work in over 10 countries, considering local specifics.
  • Lead a project from concept to launch with a temporary team of engineers.
  • Raise the bar and drive the team to deliver high-quality products, services, and processes.
  • Improve the performance, data quality, and cost-efficiency of our data pipelines at scale.
  • Maintain and monitor the data systems your team owns.

AWSDockerLeadershipPostgreSQLPythonSQLApache AirflowCloud ComputingData AnalysisETLGitKubernetesApache KafkaData engineeringData scienceSparkCI/CDProblem SolvingRESTful APIsMentoringLinuxExcellent communication skillsTeamworkCross-functional collaborationData visualizationData modelingData managementEnglish communication

Posted 5 days ago
Apply
Apply

📍 Poland, Ukraine, Cyprus

🧭 Full-Time

🔍 Software Development

🏢 Company: Competera👥 51-100💰 $3,000,000 Seed about 1 year agoArtificial Intelligence (AI)Big DataE-CommerceRetailMachine LearningAnalyticsRetail TechnologyInformation TechnologyEnterprise SoftwareSoftware

  • 5+ years of experience in data engineer role.
  • Strong knowledge of SQL, Spark, Python, Airflow, binary file formats.
  • Contribute to the development of the new data platform.
  • Collaborate with platform and ML teams to create ETL pipelines that efficiently deliver clean and trustworthy data.
  • Engage in architectural decisions regarding the current and future state of the data platform.
  • Design and optimize data models based on business and engineering needs.

PythonSQLETLKafkaAirflowSparkData modeling

Posted 5 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 6 days ago

📍 Costa Rica, Brazil, Argentina, Chile, Mexico

🔍 Insider Risk Management

🏢 Company: Teramind👥 51-100Productivity ToolsSecurityCyber SecurityEnterprise SoftwareSoftware

  • 6+ years of experience in data engineering, with a proven track record of successfully delivering data-driven solutions.
  • Strong expertise in designing and building scalable data pipelines using industry-standard tools and frameworks.
  • Experience with big data technologies and distributed systems, such as Hadoop, Spark, or similar frameworks.
  • Proficient programming skills in languages such as Python, Java, or Scala, alongside a solid understanding of database management systems (SQL and NoSQL).
  • Understanding of data requirements for machine learning applications and how to optimize data for model training.
  • Experience with security data processing and compliance standards is preferred, ensuring that data handling meets industry regulations and best practices.
  • Design and implement robust data architecture tailored for AI-driven features, ensuring it meets the evolving needs of our platform.
  • Build and maintain efficient data pipelines for processing user activity data, ensuring data flows seamlessly throughout our systems.
  • Develop comprehensive systems for data storage, retrieval, and processing, facilitating quick and reliable access to information.
  • Ensure high standards of data quality and availability, enabling machine learning models to produce accurate and actionable insights.
  • Enhance the performance and scalability of our data infrastructure to accommodate growing data demands and user activity.
  • Work closely with data scientists and machine learning engineers to understand their data requirements and ensure data solutions are tailored to their needs.

PythonSQLApache HadoopETLMachine LearningAzureData engineeringNosqlComplianceScalaData visualizationData modelingData management

Posted 6 days ago
Apply
Apply

📍 United States, Canada

🧭 Full-Time

🔍 Software Development

  • Strong hands-on experience with Python and core Python Data Processing tools such as pandas, numpy, scipy, scikit
  • Experience with cloud tools and environments like Docker, Kubernetes, GCP, and/or Azure
  • Experience with Spark/PySpark
  • Experience with Data Lineage and Data Cataloging
  • Relational and non-relational database experience
  • Experience with Data Warehouses and Lakes, such as Bigquery, Databricks, or Snowflake
  • Experience in designing and building data pipelines that scale
  • Strong communication skills, with the ability to convey technical solutions to both technical and non-technical stakeholders
  • Experience working effectively in a fast-paced, agile environment as part of a collaborative team
  • Ability to work independently and as part of a team
  • Willingness and enthusiasm to learn new technologies and tackle challenging problems
  • Experience in Infrastructure as Code tools like Terraform
  • Advanced SQL expertise, including experience with complex queries, query optimization, and working with various database systems
  • Work with business stakeholders to understand their goals, challenges, and decisions
  • Assist with building solutions that standardize their data approach to common problems across the company
  • Incorporate observability and testing best practices into projects
  • Assist in the development of processes to ensure their data is trusted and well-documented
  • Effectively work with data analysts on refining the data model used for reporting and analytical purposes
  • Improve the availability and consistency of data points crucial for analysis
  • Standing up a reporting system in BigQuery from scratch, including data replication, infrastructure setup, dbt model creation, and Integration with reporting endpoints
  • Revamping orchestration and execution to reduce critical data delivery times
  • Database archiving to move data from a live database to cold storage

AWSSQLCloud ComputingData AnalysisETLData engineeringData visualizationData modeling

Posted 11 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 13 days ago

📍 Canada

🧭 Full-Time

🔍 Fintech

🏢 Company: Coinme👥 51-100💰 $772,801 Seed over 2 years agoCryptocurrencyBlockchainBitcoinFinTechVirtual Currency

  • 7+ years of experience with ETL, SQL, PowerBI, Tableau, or similar technologies
  • Strong understanding of data modeling, database design, and SQL
  • Experience working with Apache Kafka or MSK solution
  • Extensive experience delivering solutions on Snowflake or other cloud-based data warehouses
  • Proficiency in Python/R and familiarity with modern data engineering practices
  • Strong analytical and problem-solving skills
  • Experience with machine learning (ML)
  • Design, develop, and maintain scalable data pipelines.
  • Implement data ingestion frameworks.
  • Optimize data pipelines for performance.
  • Develop and deliver data assets.
  • Evaluate and improve existing data solutions.
  • Experience in data quality management.
  • Collaborate with engineers and product managers.
  • Lead the deployment and maintenance of data solutions.
  • Champion best practices in data development.
  • Conduct code reviews and provide mentorship.
  • Create and maintain process documentation.
  • Monitor data pipelines for performance.
  • Implement logging, monitoring, and alerting systems.
  • Drive the team’s Agile process.

PythonSQLAgileETLMachine LearningSnowflakeTableauApache KafkaData engineeringData visualizationData modeling

Posted 13 days ago
Apply
Apply

📍 United Kingdom

🧭 Full-Time

🔍 Cybersecurity

🏢 Company: Immersive

  • Proficient in python programming with experience using Plotly graphing libraries and application development, experience with Flask and SQLAlchemy is a plus
  • Experience maintaining data pipelines, managing infrastructure as code, and implementing data model changes
  • Experience following software engineering best practices like version control and continuous integration
  • Strong proficiency using SQL in cloud data warehouses (e.g. BigQuery, Redshift, Snowflake) and are comfortable with performance optimization, data partitioning, and window functions
  • Experience with dbt for data transformation layer
  • Experience with IaC tooling such as Terraform or CloudFormation
  • Experience with BI tooling such as Power BI or Looker
  • Experience with AWS, Azure or GCP
  • Design, build and maintain high quality python applications for customer facing reporting
  • Maintain and develop data pipelines to ensure data quality and consistency
  • Collaborate closely with analytics engineers to implement data model changes
  • Apply domain knowledge to enable the rest of the business to access the data they need to make informed business decisions

AWSPythonSQLFlaskGitSnowflakeData engineeringCI/CDRESTful APIsTerraformData visualizationData modelingSoftware Engineering

Posted 16 days ago
Apply