Apply

Sr. Data Engineer

Posted 3 days agoViewed

View full description

💎 Seniority level: Senior, 5+ years

💸 Salary: 115000.0 - 145000.0 USD per year

⏳ Experience: 5+ years

Requirements:
  • 5+ years of experience in a data engineering or related role
  • Direct experience designing and building data modeling, ETL/ELT development principles, or data warehousing concepts
  • Strong knowledge of data management fundamentals and data storage principles
  • Deep experience in building data pipelines using Python/SQL
  • Deep experience in Airflow or similar orchestration engines
  • Deep experience in applying CI/CD principles and processes to data engineering solutions.
  • Strong understanding of cloud data engineering design patterns and use cases
  • Experience with Google Cloud Platform (GCP), such as Big Query, Composer, VertexAI, GCS, and other GCP resources
  • Bachelor's degree in Computer Science, Data Science, Statistics, Informatics, Information Systems, Mathematics, Computer Engineering, or quantitative field.
Responsibilities:
  • Collaborate with business leaders, engineers, and product managers to understand data needs.
  • Interface with other technology teams to extract, transform, and load data from a wide variety of data sources using cloud-native data engineering principles
  • Design, build, and scale data pipelines across a variety of source systems and streams (internal, third-party, as well as cloud-based), distributed/elastic environments, and downstream applications and/or self-service solutions
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Implement the appropriate design patterns while optimizing performance, cost, security, and scale and end user experience
  • Participate and lead in development sprints, demos, and retrospectives, as well as release and deployment
  • Build and manage relationships with supporting IT teams in order to effectively deliver work products to production
Apply

Related Jobs

Apply
🔥 Sr. Data Engineer
Posted 1 day ago

🧭 Full-Time

💸 100000.0 - 140000.0 USD per year

🔍 Healthcare IT

  • 7+ years of professional experience as engineer.
  • 3+ years of experience with healthcare data ecosystems (EHR systems, e-prescription workflows, pharmacy data) and HIPAA compliance.
  • Proficiency with AWS cloud services (Redshift, S3, Lambda, Glue, EMR) and data orchestration tools (Airflow, AWS Step Functions).
  • Experience developing ETL pipelines with Python and CI/CD in GitLab or similar platforms.
  • Design, develop, and maintain production-grade ETL pipelines using Python and GitLab CI/CD to process healthcare data from multiple sources.
  • Configure and optimize AWS cloud services including Redshift, S3, Lambda, Glue, and EMR to build scalable data solutions.
  • Extract, transform, and load data from various healthcare systems including e-prescription workflows, and pharmacy fill/claim data sources.
  • Address unique and complex healthcare data challenges through critical thinking, root cause analysis, and collaborative issue resolution.
  • Perform thorough code and data reviews to certify projects as 'production-ready'.
  • Document data pipelines and technical requirements.
  • Identify opportunities to leverage healthcare data in new ways.
Posted 1 day ago
Apply
Apply
🔥 Sr. Data Engineer
Posted 2 days ago

🧭 Full-Time

💸 100000.0 - 200000.0 USD per year

🔍 Software Development

🏢 Company: Risk Labs👥 11-50💰 almost 5 years agoCryptocurrencyFinance

  • 4+ years full-time experience in Data Engineering at a software, financial services, or related technology organization
  • 2+ years of experience with databases and data pipelines
  • Passion for blockchain technology, cryptonomic protocol design, game theory, and decentralized finance
  • Strong communication and collaboration skills, especially written
  • Indexing complex onchain data
  • Design and develop large-scale data pipelines
  • Write code in a highly collaborative open source ecosystem. As much of your day will be spent reviewing and designing code as it will be writing new code. We take pride in writing clean code.
Posted 2 days ago
Apply
Apply

🔍 Software Development

🏢 Company: OfferUp👥 251-500💰 $120,000,000 about 5 years ago🫂 Last layoff over 2 years agoMobile PaymentsMarketplaceE-CommerceE-Commerce PlatformsAppsMobileClassifieds

  • 5+ years of professional software development experience
  • Strong ability in distributed systems for processing large scale data processing
  • Ability to communicate technical information effectively to technical and non-technical audiences
  • Proficiency in SQL and Python
  • Experience leveraging open source data infrastructure projects, such as Apache Spark, Airflow, Kafka, Flink, Samza, Avro, Parquet, Hadoop, Hive, HBase, Phoenix, Presto or Druid
  • Experience building scalable data pipelines and real-time data streams
  • Experience building software in AWS or a similar cloud environment
  • Experience with AWS services like Kinesis, Firehose, Lambda, Sagemaker is a big plus
  • Experience with GCP services like BigQuery, Cloud Functions is a big plus
  • Experience with MLOPS platform like Databricks, VertexAI, Sagemaker is a big plus
  • Computer Science or Engineering degree required, Masters degree preferred
  • Be the technical lead on the team, owning the architecture of the data solutions, and the data platform.
  • Lead the team by example, mentor other engineers and help them with their growth.
  • Drive engineering best practices, set standards and propose larger projects which may require cross-team collaboration.
  • Design and develop applications to process large amounts of critical information in batch and near real-time to power user-facing features.
  • Influence technical direction for the company, leveraging your prior experiences and helping evaluate emerging technologies and approaches.
  • Help bring engineering maturity to a growing team that is at the center of a lot of critical initiatives for the company
Posted 5 days ago
Apply
Apply

💸 115000.0 - 145000.0 USD per year

🔍 Software Development

  • 5+ years of relevant experience
  • 1+ Experience in Data Modeling, Data architecture, Data Quality, Metadata, ETL and Data Warehouse methodologies and technologies.
  • Demonstrated experience as an engineering lead in implementing scalable, distributed, and highly available systems using cloud technologies, specifically Amazon Web Services.
  • Experience in any combination of the following: SQL, AWS EMR/Lambda, Iceberg, Linux, Python, Web technologies, APIs, Spark, Scala.
  • 2+ years’ experience with AWS tech stack.
  • Basic understanding of relational data modeling and be able to clearly articulate the reasons to use non-relational systems in our architecture.
  • Build data pipelines to land, profile, and store multiple internal & external datasets, and build applications that surface this data to support our business partners’ strategic decision making for analytics solution in media industry.
  • Be a Contributor to the overall Data Product roadmap by working closely with our business partners to understand their challenges and develop analytical tools to help drive business decisions.
  • Leverage prototyping methodologies to propose and design creative business solutions that exploit our broad toolset of technologies (AWS tech stack, data lake, Python, Spark, Scala, Iceberg, EMR Serverless, SNS, SQS, AWS Lambda, Step Functions, Event Triggers etc.)
  • Educate and inform business partners on architecture, capabilities, best practices, and solutions to build out future enhancements.
  • Assist in analyzing business requirements, source systems, understand underlying data sources, transformation requirements, data mapping, data model and metadata for reporting solutions.
  • Write easily understood documentation and architecture diagrams and keeping them up to date as code and frameworks change over time.
Posted 5 days ago
Apply
Apply
🔥 Sr. Data Engineer
Posted 9 days ago

🧭 Full-Time

🔍 Healthcare

🏢 Company: Chamber Cardio

  • 5+ years of experience as a data engineer or similar role.
  • Strong expertise in SQL and Python.
  • Hands-on experience with modern data warehousing technologies (e.g., Snowflake, BigQuery, Redshift).
  • Proficient in data pipeline and workflow orchestration tools (e.g., Airflow, Prefect, Dagster).
  • Experience with cloud platforms such as AWS (preferred), GCP, or Azure.
  • Solid understanding of data security principles and familiarity with healthcare data compliance requirements.
  • Demonstrated ability to work with messy, complex data and transform it into actionable insights.
  • Strong communication skills and the ability to work effectively in cross-functional teams.
  • Design, build, and maintain scalable ETL pipelines and data models to support our analytics and product needs.
  • Develop robust data integration workflows, enabling seamless data exchange between internal systems and external data sources.
  • Optimize and scale cloud-hosted databases and data infrastructure to handle complex healthcare datasets efficiently.
  • Ensure data security and compliance with healthcare industry standards (e.g., HIPAA, HITRUST).
  • Collaborate with product, engineering, and clinical teams to define and implement data requirements.
  • Create documentation and implement monitoring solutions to ensure data reliability and transparency.
Posted 9 days ago
Apply
Apply

🏢 Company: RecargaPay👥 501-1000💰 $10,000,000 Debt Financing almost 3 years agoMobile PaymentsFinancial ServicesFinTech

  • Solid hands-on experience with cloud platforms (AWS, GCP, or Azure) in data engineering and architecture roles.
  • Strong background in data modeling with practical experience using dbt, ER modeling tools, and advanced SQL techniques.
  • Advanced knowledge of batch and streaming data processing frameworks (Airflow, Kafka, Spark).
  • Proficient in managing data storage and processing solutions such as Snowflake, Redshift, and Delta Lake.
  • Expertise in orchestration tools and CI/CD practices (Airflow, Databricks, Git).
  • Deep understanding of data security principles, data masking, encryption, access control, and governance frameworks.
  • Excellent problem-solving skills, technical leadership, and the ability to communicate effectively with cross-functional stakeholders.
  • Experience working in regulatory, financial, or compliance-focused environments is a plus.
  • Design and build scalable, secure, and high-performance data architectures in cloud environments (AWS, GCP, or Azure), applying architectural patterns like Lambda Architecture and domain-driven design.
  • Model data across conceptual, logical, and physical levels using dbt and ER modeling tools, ensuring data integrity, consistency, and alignment with architectural standards.
  • Develop optimized data structures to support regulatory reporting, analytics, and operational workloads with a strong focus on reliability and compliance.
  • Design, build, and maintain batch and real-time data pipelines using Airflow, Kafka, and Spark, integrating structured and unstructured data from multiple sources with low latency and high availability.
  • Optimize data ingestion, transformation, and delivery workflows to ensure timely and accurate regulatory reporting.
  • Build and manage scalable data platforms, including Data Lakes, Data Warehouses, and Lakehouses, utilizing technologies such as Snowflake, Redshift, and Delta Lake.
  • Continuously monitor and optimize platforms for performance, scalability, and cost-effectiveness.
  • Orchestrate complex workflows and automate data processes using Airflow, Databricks, or equivalent orchestration tools.
  • Implement best practices for pipeline modularization, versioning (Git), and active monitoring to ensure robustness and maintainability.
  • Implement layered security measures across data platforms, including encryption, RBAC, data masking, and auditing practices, leveraging tools like Ranger, IAM, and cloud-native solutions.
  • Ensure compliance with internal and external governance, security, and regulatory requirements across all data processes.
  • Document data flows, technical standards, architecture decisions, and operational procedures clearly and effectively.
  • Mentor and support other engineers in adopting engineering best practices, ensuring a high level of technical excellence and consistency across the team.
Posted 11 days ago
Apply
Apply
🔥 Sr. Data Engineer
Posted 24 days ago

📍 United States

🧭 Full-Time

💸 140000.0 - 175000.0 USD per year

🔍 Software Development

🏢 Company: InMarket👥 251-500💰 $11,500,000 Debt Financing about 4 years agoDigital MarketingAdvertisingMobile AdvertisingMarketing

  • Strong SQL experience
  • Expert in at least one data pipelining orchestration framework (Airflow, Luigi, etc.)
  • Experience building ETL pipelines
  • Strong software engineering skills in an Object Oriented language such as Java or Python
  • Experience building and optimizing a data warehouse on a major cloud platform (Big Query preferred but not required)
  • Experience and deep understanding of Big Data Technologies such as Hadoop and Spark
  • Experience with building and optimizing large scale and high-performance systems
  • Extensive knowledge of data related tools and architecture concepts in a major cloud platform
  • Knowledge of Kubernetes, Docker, Airflow, Git, and CD/CI best practices
  • Strong collaboration and communication skills within and across teams
  • B.S. or M.S. in Computer Science or a related technical field
  • At least 4 years of software engineering experience, ideally in data engineering
  • Design and implement ETL pipelines in Apache Airflow, Big Query, Python, and Spark to transform various upstream sources data into curated data assets for use across InMarket’s vast array of data driven products
  • Stay up to date with and advocate internally for Data Engineering best practices, technologies, and testing frameworks that will help shape InMarket’s technical future
  • Architect and plan projects involving highly distributed and high performance data platform systems
  • Provide technical guidance to other engineers in order to help promote the growth of their knowledge and technical capabilities
  • Communicate complex concepts and the results of the analyses in a clear and effective manner to Product and Engineering Managers
  • Identify areas for improvement in existing pipelines and processes
  • Understand business requirements and convert them to technical solutions

DockerPythonSQLApache AirflowETLGitHadoopKubernetesData engineeringSparkCI/CD

Posted 24 days ago
Apply
Apply
🔥 Sr. Data Engineer
Posted 2 months ago

📍 Mexico

🔍 Software Development

  • 5+ years of experience as a Data Engineer
  • Snowflake architecture expertise for multi-tenant B2B applications
  • Performance optimization for customer-facing data models and analytics.
  • Advanced SQL skills for complex query optimization
  • Proficiency in Python, Scala, or Go for data pipeline development
  • Experience analyzing source data structures and recommending improvements
  • Ability to collaborate with engineering teams on data design
  • Experience with ETL/ELT pipelines (Airflow, dbt)
  • Integration experience with Power BI, Tableau, and Sigma
  • Mentoring skills for report creation using BI tools
  • Data quality management for customer-facing products
  • Experience with GitHub/source control and CI/CD pipelines (GitHub Actions or Jenkins)
  • Understanding of multi-tenant data security and governance
  • Evaluate and improve data models and schemas in Snowflake to enhance our B2B analytics products.
  • Analyze BI warehouse views and identify structural improvements at the source, collaborating with engineering teams to recommend beneficial changes for downstream BI products.
  • Optimize ETL/ELT pipelines for timely and reliable data synchronization.
  • Improve query performance and storage efficiency while controlling costs.
  • Enhance data quality monitoring and validation processes.
  • Work with product management to translate customer requirements into data solutions.
  • Guide team members and customers in creating effective reports using Power BI, Tableau, and Sigma.
  • Improve data lineage documentation and warehouse architecture.
  • Monitor and resolve pipeline issues to prevent service disruptions.
  • Implement improvements to multi-tenant data architecture for better data isolation and resource utilization.

PythonSQLETLGitSnowflakeTableauAirflowData engineeringREST APICI/CDData modelingDebugging

Posted 2 months ago
Apply
Apply
🔥 Sr. Data Engineer
Posted 2 months ago

📍 United States

🧭 Full-Time

💸 175000.0 - 220000.0 USD per year

🔍 Healthcare

🏢 Company: Motion MSK

  • 4+ years of experience working with big data processing frameworks, such as Apache Spark (Scala or PySpark).
  • Hands-on experience building ETL pipelines for structured and semi-structured healthcare claims data.
  • Proficient in SQL and have experience working with distributed databases and data lakes (e.g., Delta Lake, Redshift, Snowflake).
  • Experience orchestrating workflows using Apache Airflow and managing data processing on AWS Glue or EMR.
  • Comfortable working in a cloud-native environment (AWS preferred) and using infrastructure-as-code tools like Terraform.
  • Design, build, and maintain scalable data pipelines to process claims data efficiently using Spark (Scala/PySpark), AWS Glue, and Airflow.
  • Work closely with analysts, backend engineers, and other stakeholders to ensure high data quality, reliability, and accessibility.
  • Optimize and tune distributed data processing workflows for performance and cost efficiency.
  • Develop and enforce best practices for data governance, schema evolution, and pipeline observability while understanding the nuances of security and compliance.
  • Implement automated testing, monitoring, and alerting for data workflows to ensure operational excellence.
  • Collaborate with stakeholders to define and implement data models that support analytics and business intelligence initiatives.

AWSSQLApache AirflowETLSnowflakeData engineeringTerraformScalaData visualizationData modelingData management

Posted 2 months ago
Apply
Apply

📍 Poland

🧭 Full-Time

🔍 Software Development

🏢 Company: Craft Machine Inc

  • 4+ years of experience in Data Engineering.
  • 4+ years of experience with Python.
  • Experience in developing, maintaining, and ensuring the reliability, scalability, fault tolerance, and observability of data pipelines in a production environment.
  • Strong knowledge of SDLC and solid software engineering practices.
  • Knowledge and experience with Amazon Web Services (AWS) and Databricks (nice to have).
  • Familiar with infrastructure-as-code approach.
  • Have fundamental knowledge of data engineering techniques: ETL/ELT, batch and streaming, DWH, Data Lakes, distributed processing.
  • Building and optimizing data pipelines (batch and streaming).
  • Extracting, analyzing and modeling rich and diverse datasets of structured and unstructured data
  • Designing software that is easily testable and maintainable.
  • Support in setting data strategies and our vision.
  • Keep track of emerging technologies and trends in the Data Engineering world, incorporating modern tooling and best practices at Craft.
  • Work on extendable data processing systems that allows to add and scale pipelines.
  • Applying machine learning techniques such as anomaly detection, clustering, regression classification, and summarization to extract value from our data sets.

AWSDockerPostgreSQLPythonSQLETLMachine LearningAirflowAmazon Web ServicesData engineeringPandasCI/CDTerraformData modelingSoftware Engineering

Posted 2 months ago
Apply

Related Articles

Posted about 1 month ago

How to Overcome Burnout While Working Remotely: Practical Strategies for Recovery

Burnout is a silent epidemic among remote workers. The blurred lines between work and home life, coupled with the pressure to always be “on,” can leave even the most dedicated professionals feeling drained. But burnout doesn’t have to define your remote work experience. With the right strategies, you can recover, recharge, and prevent future episodes. Here’s how.



Posted 6 days ago

Top 10 Skills to Become a Successful Remote Worker by 2025

Remote work is here to stay, and by 2025, the competition for remote jobs will be tougher than ever. To stand out, you need more than just basic skills. Employers want people who can adapt, communicate well, and stay productive without constant supervision. Here’s a simple guide to the top 10 skills that will make you a top candidate for remote jobs in the near future.

Posted 9 months ago

Google is gearing up to expand its remote job listings, promising more opportunities across various departments and regions. Find out how this move can benefit job seekers and impact the market.

Posted 10 months ago

Read about the recent updates in remote work policies by major companies, the latest tools enhancing remote work productivity, and predictive statistics for remote work in 2024.

Posted 10 months ago

In-depth analysis of the tech layoffs in 2024, covering the reasons behind the layoffs, comparisons to previous years, immediate impacts, statistics, and the influence on the remote job market. Discover how startups and large tech companies are adapting, and learn strategies for navigating the new dynamics of the remote job market.