Apply

Senior Data Engineer

Posted 2 days agoViewed

View full description

💎 Seniority level: Senior, 5 years

🔍 Industry: Healthcare IT

🏢 Company: NeueHealth

⏳ Experience: 5 years

Requirements:
  • Bachelor’s degree in Computer Science, Computer Engineering, Information Systems, or equivalent experience required.
  • Around five (5) years of experience in an enterprise data engineering role in an Azure environment.
  • Enterprise development experience coding in Scala and building batch and streaming data pipelines (Scala).
  • Experience with API design.
  • Extensive experience developing data-intensive solutions in an Azure Cloud environment.
  • Extensive experience developing solutions that use event sourcing and/or Big Data architectures.
Responsibilities:
  • Write traditional code and server-less functions using the language best suited for the task, which is primarily Scala.
  • Build APIs, data microservices and ETL pipelines, to share data with internal and external partners and write interfaces to public data sets to enrich our analytics data stores.
  • Develop and optimize processes for fine-tuning large language models (LLMs) and implementing Retrieval-Augmented Generation (RAG) frameworks to enhance AI-driven solutions.
  • Build and support Data Ingestion frameworks deployed in Azure.
  • Participate in building and owning a culture of DevOps and Quality Assurance.
  • Act as tech lead on projects and guide junior engineers.
  • Continuously document your code, framework standards, and team processes.
Apply

Related Jobs

Apply

📍 States of São Paulo and Rio Grande do Sul, cities of Rio de Janeiro, Belo Horizonte, Florianópolis and Fortaleza

🏢 Company: TELUS Digital Brazil

  • 5+ years of relevant development experience writing high-quality code as a Data Engineer
  • Have actively participated in the design and development of data architectures
  • Hands-on experience in developing and optimizing data pipelines
  • Comprehensive understanding of data modeling, ETL processes, and both SQL and NoSQL databases
  • Experience with a general-purpose programming language such as Python or Scala
  • Experience with GCP platforms and services.
  • Experience with containerization technologies such as Docker and Kubernetes
  • Proven track record in implementing and optimizing data warehousing solutions and data lakes
  • Proficiency in DevOps practices and automation tools for continuous integration and deployment of data solutions
  • Experience with machine learning workflows and supporting data scientists in model deployment
  • Solid understanding of data security and compliance requirements in large-scale data environments
  • Strong ability to communicate effectively with teams and stakeholders, providing and receiving feedback to improve product outcomes.
  • Proficient in communicating and writing in English
  • Develop and optimize scalable, high-performing, secure, and reliable data pipelines that address diverse business needs and considerations
  • Identify opportunities to enhance internal processes, implement automation to streamline manual tasks, and contribute to infrastructure redesign
  • Help mentor and coach a product team towards shared goals and outcomes
  • Navigate difficult conversations by providing constructive feedback to teams
  • Identify obstacles to ensure quality, improve our user experience and how we build tests
  • Be self-aware of limitations, yet curious to learn new solutions while being receptive to constructive feedback from teammates
  • Engage in ongoing research and adoption of new technologies, libraries, frameworks, and best practices to enhance the capabilities of the data team

DockerPythonSQLETLGCPHadoopKafkaKubernetesMachine LearningAirflowData engineeringNosqlSparkCI/CDAgile methodologiesRESTful APIsDevOpsScalaData visualizationData modeling

Posted about 9 hours ago
Apply
Apply
🔥 Senior Data Engineer
Posted about 14 hours ago

📍 India

🧭 Full-Time

  • Hands-on experience in implementing, supporting, and administering modern cloud-based data solutions (Google BigQuery, AWS Redshift, Azure Synapse, Snowflake, etc.).
  • Strong programming skills in SQL, Java, and Python.
  • Experience in configuring and managing data pipelines using Apache Airflow, Informatica, Talend, SAP BODS or API-based extraction.
  • Expertise in real-time data processing frameworks.
  • Strong understanding of Git and CI/CD for automated deployment and version control.
  • Experience with Infrastructure-as-Code tools like Terraform for cloud resource management.
  • Good stakeholder management skills to collaborate effectively across teams.
  • Solid understanding of SAP ERP data and processes to integrate enterprise data sources.
  • Exposure to data visualization and front-end tools (Tableau, Looker, etc).
  • Design and Develop Data Pipelines: Create data pipelines to extract data from various sources, transform it into a standardized format, and load it into a centralized data repository.
  • Build and Maintain Data Infrastructure: Design, implement, and manage data warehouses, data lakes, and other data storage solutions.
  • Ensure Data Quality and Integrity: Develop data validation, cleansing, and normalization processes to ensure data accuracy and consistency.
  • Collaborate with Data Analysts and Business Process Owners: Work with data analysts and business process owners to understand their data requirements and provide data support for their projects.
  • Optimize Data Systems for Performance: Continuously monitor and optimize data systems for performance, scalability, and reliability.
  • Develop and Maintain Data Governance Policies: Create and enforce data governance policies to ensure data security, compliance, and regulatory requirements.

AWSPythonSQLApache AirflowCloud ComputingETLGitJavaSAPSnowflakeData engineeringCommunication SkillsCI/CDRESTful APIsTerraformData visualizationStakeholder managementData modelingEnglish communication

Posted about 14 hours ago
Apply
Apply

🧭 Full-Time

🔍 Health & Bioinformatics

🏢 Company: Gradient AI👥 101-250💰 $20,000,000 Series B about 4 years agoArtificial Intelligence (AI)Machine LearningInsurTechInsuranceHealth Care

  • BS in Computer Science, Bioinformatics, or another quantitative discipline with 7+ years working with and interpreting health, medical, and bioinformatics data, including real-world healthcare datasets.
  • Subject matter expertise (SME) in health and bioinformatics data, with a strong grasp of the complexities and challenges of processing medical and biological information.
  • Knowledge of healthcare data standards (e.g., FHIR, HL7) and a solid understanding of healthcare data privacy and security regulations (such as HIPAA) are highly desirable.
  • Proficiency in Python and SQL within a professional environment.
  • Hands-on knowledge of big data tools like Apache Spark (PySpark), DataBricks, Snowflake, or similar platforms.
  • Skilled in using data orchestration frameworks such as Airflow, Dagster, or Prefect.
  • Comfortable working within cloud computing environments, preferably AWS, along with Linux systems.
  • Design, build, and implement data systems to support ML and AI models for our health insurance clients, ensuring strict compliance with healthcare data privacy and security regulations (e.g., HIPAA).
  • Develop tools for extracting, processing, and profiling diverse healthcare data sources, including EHRs, medical claims, pharmacy data, and genomic data.
  • Collaborate with data scientists to transform large volumes of health-related and bioinformatics data into modeling-ready formats, prioritizing data quality, integrity, and reliability in healthcare applications.
  • Build and maintain infrastructure for the extraction, transformation, and loading (ETL) of data from a variety of sources using SQL, AWS, and healthcare-specific big data technologies and analytics platforms.
  • Apply health and bioinformatics subject matter expertise to ensure data pipelines meet the unique requirements of health, medical, and bioinformatics data processing - including translating complex medical and biological concepts into actionable data requirements.
Posted 1 day ago
Apply
Apply

📍 Worldwide

🔍 Algorithmic Trading

  • 7 + years building production‑grade data systems.
  • Familiarity with market data formats (e.g., MDP, ITCH, FIX, proprietary exchange APIs) and market data providers.
  • Expert‑level Python (Go and C++ nice to have).
  • Hands‑on with modern orchestration (Airflow) and event streams (Kafka).
  • Strong SQL proficiency: aggregations, joins, subqueries, window functions (first, last, candle, histogram), indexes, query planning, and optimization.
  • Designing high‑throughput APIs (REST/gRPC) and data access libraries.
  • Strong Linux fundamentals, containers (Docker) and cloud object storage (AWS S3 / GCS).
  • Proven track record of mentoring, code reviews and driving engineering excellence.
  • Architect batch + stream pipelines (Airflow, Kafka, dbt) for diverse structured and unstructured marked data.
  • Implement and tune S3, column‑oriented and time‑series data storage for petabyte‑scale analytics; own partitioning, compression, TTL, versioning and cost optimisation.
  • Develop internal libraries for schema management, data contracts, validation and lineage; contribute to shared libraries and services for internal data consumers for research, backtesting and real-time trading purposes.
  • Embed monitoring, alerting, SLAs, SLOs and CI/CD; champion automated testing, data quality dashboards and incident runbooks.
  • Partner with Data Science, Quant Research, Backend and DevOps to translate requirements into platform capabilities and evangelise best practices.
Posted 1 day ago
Apply
Apply

🧭 Full-Time

🔍 DeFi

🏢 Company: CoW DAO

  • Solid experience designing and maintaining scalable data architectures (data lakes, warehouses, pipelines)
  • Strong coding skills in Python and SQL
  • Comfortable working with blockchain data providers and APIs (e.g. Dune, Etherscan, CoinGecko)
  • Familiar with decoding Ethereum transactions and simulation tools (e.g. Tenderly, Phalcon)
  • Experience with modern data orchestration frameworks (Dagster, Airflow, or Prefect)
  • You write clean, tested, and well-reviewed code and follow best practices for deployment and monitoring
  • You’re a self-starter, collaborative, and not afraid to take ownership
  • Collaborate with teams across the company to understand their data needs and provide actionable insights
  • Design, build, and maintain robust data pipelines and infrastructure for ingesting and processing on-chain and off-chain data
  • Ensure data quality, reliability, and discoverability through testing, monitoring, and documentation
  • Create tools and systems that make it easy to access, query, and visualize data (using tools like Redash, Metabase, or Looker)
  • Keep an eye on the evolving blockchain data ecosystem—experiment with new tools and techniques to stay ahead
  • Own your work end-to-end: from initial idea to deployment and maintenance
Posted 1 day ago
Apply
Apply

📍 United States

🔍 Computer software

🏢 Company: Worth AI👥 11-50💰 $12,000,000 Seed over 1 year agoArtificial Intelligence (AI)Business IntelligenceRisk ManagementFinTech

  • 7+ years of proven experience as a Senior Data Engineer or similar role, preferably in a software or technology-driven company with experience processing several to several hundreds of Gigabytes of data or more
  • In-depth knowledge of data modeling, data warehousing, and database design principles.
  • Strong programming skills in Python, SQL, and other relevant languages.
  • Experience with relational and NoSQL databases, such as PostgreSQL, MySQL, MongoDB
  • Proficiency in data integration and ETL tools, such as Apache Kafka, Apache Airflow, or Informatica.
  • Familiarity with big data processing frameworks, such as Hadoop, Spark, or Flink.
  • Knowledge of cloud platforms, such as AWS, Azure, or GCP, and experience with data storage and processing services in the cloud.
  • Understanding of data governance, data privacy, and data security best practices.
  • Strong problem-solving and troubleshooting skills, with a focus on data quality and system performance.
  • Excellent communication and collaboration skills to work effectively with cross-functional teams.
  • Prior collaborative work with data scientists or machine learning professionals with respect to sourcing, processing and scaling both input and output data
  • Comfortable going through documentation of third-party API’s and identifying best procedures for integrating data from API’s into broader ETL processes
  • Design, build, code and maintain large-scale data processing systems and architectures that support AI initiatives.
  • Develop and implement data pipelines and ETL processes to ingest, transform, and load data from various sources.
  • Design and optimize databases and data storage solutions for high performance and scalability.
  • Collaborate with cross-functional teams to understand data requirements and ensure data quality and integrity.
  • Implement data governance and data security measures to protect sensitive data.
  • Monitor and troubleshoot data infrastructure and pipeline issues in a timely manner.
  • Stay up-to-date with the latest trends and technologies in data engineering and recommend improvements to enhance the company's data capabilities.
Posted 1 day ago
Apply
Apply

🔍 AI Consultancy

  • 5-8 years of data engineering experience with proven expertise in AWS
  • Strong technical leadership capabilities and consulting skills
  • Deep knowledge of modern data engineering principles and best practices
  • Expertise in Python, Java, or Scala with strong system design skills
  • Lead technical implementation of enterprise-scale data solutions using advanced cloud technologies
  • Drive architecture decisions and establish technical standards across client engagements
  • Mentor junior engineers while managing complex technical workstreams
  • Build strong client relationships as a trusted technical advisor
Posted 2 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 2 days ago

📍 Germany, Spain, Portugal, Greece

🏢 Company: WorkMotion👥 101-250💰 $10,000,000 Series B almost 3 years agoComplianceHuman ResourcesEmployee Benefits

  • 3-5 years of professional experience in Data Engineering or Software Development with a focus on data
  • Strong knowledge of Python and SQL; and PySpark
  • Hands-on experience with AWS services (Glue, S3, Athena, EC2)
  • Experience with Apache Airflow, preferably in a Dockerized/cloud-native environment
  • Familiarity with Delta Lake or similar data lake frameworks
  • Proficiency with source control (GitHub) and CI/CD workflows
  • Strong understanding of data modeling, ETL best practices, and data pipeline performance optimization
  • Design, build, and maintain scalable ETL pipelines using Apache Airflow and AWS Glue (Spark)
  • Work with a range of data sources including Salesforce, NetSuite, PostgreSQL, and MongoDB
  • Develop and optimize PySpark jobs for large-scale data transformation and analytics
  • Manage data lake infrastructure using Delta Lake on S3 with Athena as the query layer
  • Ensure data quality, performance, and reliability through monitoring, testing, and documentation
  • Collaborate with analytics, product, and engineering teams to define data requirements
  • Contribute to CI/CD workflows with GitHub and deployment automation
  • Participate in architectural discussions and advocate for best practices in data engineering

AWSDockerPythonSoftware DevelopmentSQLApache AirflowETLGitData engineeringSparkCI/CDData modeling

Posted 2 days ago
Apply
Apply

📍 Colombia

🧭 Full-Time

🔍 Software Development

  • 5+ years of experience in developing scalable data pipeline infrastructure, preferably for sales organizations
  • Proven track record of delivering large-scale data projects and working with business partners
  • Experience with big data processing frameworks such as Apache Spark
  • Experience with data orchestration tools like Airflow or Dagster
  • Experience with infrastructure-as-code tools (e.g., Terraform) and modern CI/CD pipelines
  • Collaborate with other engineers, business partners, and data scientists to build best-in-class data infrastructure that meets evolving needs
  • Design and implement scalable data pipelines that integrate Salesforce and other sales systems data into our enterprise data lake
  • Build automated solutions for sales data quality, enrichment, and standardization
  • Create and maintain data models that power sales analytics, forecasting, and reporting systems
  • Design and manage reverse ETL pipelines to power sales operations and marketing automation
  • Partner with AI/ML engineers to develop Sales predictive and generative models
  • Architect solutions for real-time sales data synchronization and processing
  • Optimize data flows between Salesforce, Snowflake, AWS Athena, and other enterprise systems
  • Build robust monitoring and alerting systems for sales data pipelines
  • Collaborate with Sales Operations to automate manual processes and improve data accuracy
  • Create documentation and enable self-service capabilities for sales teams

AWSPythonApache AirflowSalesforceSnowflakeData engineeringCI/CDTerraformData modeling

Posted 3 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 3 days ago

🧭 Full-Time

🔍 AI

🏢 Company: Zencoder

  • 5+ years of experience in data warehouse (DWH) development and data transformation workflows.
  • Strong SQL skills, including the ability to write and optimize complex queries.
  • A solid understanding of Data Governance and Data Quality best practices.
  • Experience building and maintaining data transformation workflows, with knowledge of ETL/ELT processes, batch, and streaming data processing.
  • Proficiency in Python for data-related tasks.
  • A passion for data-driven decision making and the ability to extract insights from complex datasets.
  • Design and run A/B tests that help us move fast without guessing — you'll guide the team in making smart, data-informed choices, and build out best practices for experimentation that everyone can rely on.
  • Develop and implement standardized processes for tracking new analytical events;
  • Turn raw data into real insights — dig into product usage, agent performance, and user behavior to help teams understand what’s working, what’s not, and where to go next.
  • Level up our analytics game by improving our reporting systems and introducing new metrics that reflect how users experience our AI tools.
  • Build trust in our data with robust quality checks and validation processes.
  • Architect DWH data models, create and optimize SQL queries and data models for cross-layer DWH data transformation using DBT.
Posted 3 days ago
Apply

Related Articles

Posted about 1 month ago

How to Overcome Burnout While Working Remotely: Practical Strategies for Recovery

Burnout is a silent epidemic among remote workers. The blurred lines between work and home life, coupled with the pressure to always be “on,” can leave even the most dedicated professionals feeling drained. But burnout doesn’t have to define your remote work experience. With the right strategies, you can recover, recharge, and prevent future episodes. Here’s how.



Posted 7 days ago

Top 10 Skills to Become a Successful Remote Worker by 2025

Remote work is here to stay, and by 2025, the competition for remote jobs will be tougher than ever. To stand out, you need more than just basic skills. Employers want people who can adapt, communicate well, and stay productive without constant supervision. Here’s a simple guide to the top 10 skills that will make you a top candidate for remote jobs in the near future.

Posted 9 months ago

Google is gearing up to expand its remote job listings, promising more opportunities across various departments and regions. Find out how this move can benefit job seekers and impact the market.

Posted 10 months ago

Read about the recent updates in remote work policies by major companies, the latest tools enhancing remote work productivity, and predictive statistics for remote work in 2024.

Posted 10 months ago

In-depth analysis of the tech layoffs in 2024, covering the reasons behind the layoffs, comparisons to previous years, immediate impacts, statistics, and the influence on the remote job market. Discover how startups and large tech companies are adapting, and learn strategies for navigating the new dynamics of the remote job market.