Apply

Data Engineer

Posted about 2 months agoViewed

View full description

💎 Seniority level: Senior, 3+ years

📍 Location: United States

💸 Salary: 110000.0 - 150000.0 USD per year

🔍 Industry: Healthcare

🏢 Company: Oshi Health👥 51-100💰 $60,000,000 Series C 6 months agoMedicalMobileHealth Care

🗣️ Languages: English

⏳ Experience: 3+ years

🪄 Skills: PythonSQLCloud ComputingETLData engineeringCI/CDData modeling

Requirements:
  • Hold a BS/BA degree in Computer Science, Math, Physics, or a related field, or have equivalent experience in a relevant area.
  • An experienced data engineer with experience in startup environments.
  • Have 3+ years of data development experience.
  • You can understand a complex requirement, componentize it, and develop a scalable solution.
  • Advanced SQL chops and knowledge of data warehousing standards.
  • Comfortable in one or more programming languages (e.g. golang/python) and efficient data transport/manipulation.
  • Familiar with dbt (Data Build Tool) for managing warehouse transformations.
  • Knowledge and experience with cloud environments and associated tooling is a plus.
  • Experience with FHIR standards, HL7, or other healthcare data interchange formats is a plus.
  • Knowledge about data security and HIPAA compliance, a plus.
Responsibilities:
  • Contribute to Oshi’s existing data warehouse, supporting product, clinical, and strategy teams in developing underlying data models.
  • Collaborate with marketing and growth teams to build data pipelines and processes that support member marketing, analytic, and outreach efforts.
  • Develop reusable queries, data quality tests, and insights for reporting and backend systems.
  • Design and implement complex data models for various use cases, including real-time analytics, data pipelines, and business intelligence.
  • Work across the data stack, including CI/CD pipelines for data workflows, infrastructure as code, and platform integrations.
  • Support, expand, and help standardize data governance structures for handling sensitive client data.
  • Ensure the integrity and protection of data across Oshi’s systems.
Apply

Related Jobs

Apply
🔥 Staff Data Engineer
Posted about 3 hours ago

📍 United States

🧭 Full-Time

🔍 Healthcare

🏢 Company: Atropos Health

  • At least 5 years of experience working with messy real-world healthcare data (EHR data, claims) and experience with Real World Data / Evidence at least 3 years
  • Expertise working with common data models and terminology systems in healthcare data (data models and standards such as FHIR, OMOP, codesets like ICD 9/10, CPT, ATC, RxNorm, NDC, etc.)
  • Knowledge of common EHR and medical claim / administrative workflows and the basics of the US healthcare system (patient movement through a hospital or clinic, how medical procedures are reimbursed, etc.)
  • SQL (any dialect): at least 5 years, Python at least 4 years (Required)
  • Significant and deep experience as a data professional working with public cloud infrastructure to build data products (Required)
  • Knowledge of EHR data systems and at least 4 years of experience working intensively with EHR data in a data engineering and/or analytics environment
  • High level knowledge of ETL/ELT operations and data modeling
  • Create and maintain data pipelines to integrate, enrich, and map clinical, claims, and other data from multiple sources in the cloud
  • Design systems of data quality checks and assessments on clinical datasets
  • Expert knowledge of how to automate workflows and create reusable infrastructure in multiple public cloud environments (AWS, Azure, GCP)
  • Map data from one source into a common data model using many different frameworks, with a particular focus on PySpark and Spark SQL.
  • Create, maintain, tune, and document data flows in Databricks and Snowflake, Bigquery, and other diverse cloud tools
  • Filter, clean, transform, and integrate clinical data using your domain expertise of healthcare data and code systems and embedded knowledge
  • Work remotely with an interdisciplinary team and build and maintain strong relationships with research, clinical, product, and commercial stakeholders
  • Manage external customer, vendors, and customer relationships and create internal and external communications
  • Travel required to company or team offsites 3-4 weeks per year

AWSPythonSQLCloud ComputingETLSnowflakeData engineeringRDBMSSparkData modeling

Posted about 3 hours ago
Apply
Apply
🔥 Staff Data Engineer
Posted about 5 hours ago

📍 US

🧭 Full-Time

💸 185000.0 - 200000.0 USD per year

🔍 Adtech

  • 8+ years of experience in data engineering.
  • Proven experience building data infrastructure using Spark with Scala.
  • Familiarity with data lakes, cloud warehouses, and storage formats.
  • Strong proficiency in AWS services.
  • Expertise in SQL for data manipulation and extraction.
  • Bachelor's degree in Computer Science or a related field.
  • Design and implement robust data infrastructure using Spark with Scala.
  • Collaborate with our cross-functional teams to design data solutions that meet business needs.
  • Build out our core data pipelines, store data in optimal engines and formats, and feed our machine learning models.
  • Leverage and optimize AWS resources.
  • Collaborate closely with the Data Science team.

AWSSQLCloud ComputingETLMachine LearningData engineeringData scienceSparkScalaData modeling

Posted about 5 hours ago
Apply
Apply

📍 AL, AR, AZ, CA (exempt only), CO, CT, FL, GA, ID, IL, IN, IA, KS, KY, MA, ME, MD, MI, MN, MO, MT, NC, NE, NJ, NM, NV, NY, OH, OK, OR, PA, SC, SD, TN, TX, UT, VT, VA, WA, WI

🧭 Full-Time

🔍 Insurance

🏢 Company: Kin Insurance

  • Depth of experience in modern big data environments.
  • Advanced knowledge and experience with SQL and Python is expected.
  • Insurance domain knowledge.
  • Design and develop data pipelines and modeling raw data for downstream ingestion.
  • Mentor and guide data engineers on your team and across the organization, while collaborating with other engineers, product managers, analysts, and stakeholders.
  • Lead a cross-functional project team with members from AppEng, DataEng, BI, and business stakeholders.

AWSPythonSQLApache AirflowETLCross-functional Team LeadershipData engineeringProblem SolvingMentoringDocumentationComplianceData visualizationData modelingData management

Posted about 17 hours ago
Apply
Apply
🔥 Data Engineer
Posted about 18 hours ago

📍 United States

🧭 Contract

🔍 Biotechnology

🏢 Company: Avomind👥 11-50EmploymentHuman ResourcesRecruiting

  • Strong experience in data engineering and cloud platforms (preferably GCP).
  • Proficiency in programming languages like Python, SQL, and shell scripting.
  • Familiarity with data catalog tools (e.g., DataHub, Apache Atlas) and metadata management.
  • Experience with building and maintaining scalable ETL pipelines using orchestration tools (Dagster, Airflow).
  • Understanding of API development and integration.
  • Knowledge of data governance and data quality principles.
  • Background in biological or scientific data is a plus but not mandatory.
  • Design, build, and maintain ETL/ELT pipelines to process and transform data efficiently.
  • Develop and optimise scalable data architectures in the cloud.
  • Implement and maintain data cataloging solutions to ensure discoverability and governance.
  • Build APIs and integrations for seamless data exchange across systems.
  • Perform data quality checks and implement automated testing frameworks to ensure data accuracy and reliability.
  • Collaborate with teams to build self-service systems and promote data democratisation.
  • Document and maintain data engineering processes and best practices

PythonSQLETLGCPAirflowAPI testingData engineeringData modeling

Posted about 18 hours ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 150363.0 - 180870.0 USD per year

🔍 Software Development

  • At least a Bachelors Degree or foreign equivalent in Computer Science, Computer Engineering, Electrical and Electronics Engineering, or a closely related technical field, and at least five (5) years of post-bachelor’s, progressive experience writing shell scripts; validating data; and engaging in data wrangling.
  • Experience must include at least three (3) years of experience debugging data; transforming data into Microsoft SQL server; developing processes to import data into HDFS using Sqoop; and using Java, UNIX Shell Scripts, and Python.
  • Experience must also include at least one (1) year of experience developing Hive scripts for data transformation on data lake projects; converting Hive scripts to Pyspark applications; automating in Hadoop; and implementing CI/CD pipelines.
  • Design, develop, test, and implement Big Data technical solutions.
  • Recommend the right technologies and solutions for a given use case, from the application layer to infrastructure.
  • Lead the delivery of compiling and installing database systems, integrating data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures.
  • Drive solution architecture and perform deployments of data pipelines and applications.
  • Author DDL and DML SQL spanning technical tacks.
  • Develop data transformation code and highly complex provisioning pipelines.
  • Ingest data from relational databases.
  • Execute automation strategy.

AWSPythonSQLHadoopJavaKafkaSnowflakeData engineeringSparkCI/CDScalaScriptingDebugging

Posted 2 days ago
Apply
Apply
🔥 Data Engineer-I
Posted 2 days ago

📍 USA

🔍 Healthcare

🏢 Company: Innovaccer Inc.

  • SQL knowledge
  • ETL/ELT/Data pipeline knowledge
  • Python knowledge
  • Powershell / Bash knowledge
  • Excellent problem-solving and effective communication skills
  • Self-motivation, integrity and honesty
  • Collaborate with team, management, departments using virtual tools
  • Run Production data pipelines/processes, ensure the integrity of the data, and send out deliverables based on requirement/runbook documentation
  • Coordinate with the various technical teams to resolve issues/bugs/optimize said production processes
  • Coordinate with internal client facing team members to communicate the status of deliverables
  • Help develop/improve technical documentation to guide future software development projects and operations
  • Dedicated time to explore building out tech stack and capabilities where there are applicable use cases
  • Provide critical thinking, technical innovation, and extra attention to detail by serving as a trusted team member and peer code reviewer
  • Assists with external client communications when deliverables or receivables do not meet technical or project requirements, ensuring timely resolution and alignment

PythonSQLBashETLMicrosoft AzurePostgresData modeling

Posted 2 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 5 days ago

📍 Worldwide

🔍 Hospitality

🏢 Company: Lighthouse

  • 4+ years of professional experience using Python, Java, or Scala for data processing (Python preferred)
  • You stay up-to-date with industry trends, emerging technologies, and best practices in data engineering.
  • Improve, manage, and teach standards for code maintainability and performance in code submitted and reviewed
  • Ship large features independently, generate architecture recommendations and have the ability to implement them
  • Great communication: Regularly achieve consensus amongst teams
  • Familiarity with GCP, Kubernetes (GKE preferred),  CI/CD tools (Gitlab CI preferred), familiarity with the concept of Lambda Architecture.
  • Experience with Apache Beam or Apache Spark for distributed data processing or event sourcing technologies like Apache Kafka.
  • Familiarity with monitoring tools like Grafana & Prometheus.
  • Design and develop scalable, reliable data pipelines using the Google Cloud stack.
  • Optimise data pipelines for performance and scalability.
  • Implement and maintain data governance frameworks, ensuring data accuracy, consistency, and compliance.
  • Monitor and troubleshoot data pipeline issues, implementing proactive measures for reliability and performance.
  • Collaborate with the DevOps team to automate deployments and improve developer experience on the data front.
  • Work with data science and analytics teams to enable them to bring their research to production grade data solutions, using technologies like airflow, dbt or MLflow (but not limited to)
  • As a part of a platform team, you will communicate effectively with teams across the entire engineering organisation, to provide them with reliable foundational data models and data tools.
  • Mentor and provide technical guidance to other engineers working with data.

PythonSQLApache AirflowETLGCPKubernetesApache KafkaData engineeringCI/CDMentoringTerraformScalaData modeling

Posted 5 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 6 days ago

📍 United States

🧭 Full-Time

💸 183600.0 - 216000.0 USD per year

🔍 Software Development

  • 6+ years of experience in a data engineering role building products, ideally in a fast-paced environment
  • Good foundations in Python and SQL.
  • Experience with Spark, PySpark, DBT, Snowflake and Airflow
  • Knowledge of visualization tools, such as Metabase, Jupyter Notebooks (Python)
  • Collaborate on the design and improvements of the data infrastructure
  • Partner with product and engineering to advocate best practices and build supporting systems and infrastructure for the various data needs
  • Create data pipelines that stitch together various data sources in order to produce valuable business insights
  • Create real-time data pipelines in collaboration with the Data Science team

PythonSQLSnowflakeAirflowData engineeringSparkData visualizationData modeling

Posted 6 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 6 days ago

📍 United States

🧭 Full-Time

🔍 Healthcare

🏢 Company: Rad AI👥 101-250💰 $60,000,000 Series C 2 months agoArtificial Intelligence (AI)Enterprise SoftwareHealth Care

  • 4+ years relevant experience in data engineering.
  • Expertise in designing and developing distributed data pipelines using big data technologies on large scale data sets.
  • Deep and hands-on experience designing, planning, productionizing, maintaining and documenting reliable and scalable data infrastructure and data products in complex environments.
  • Solid experience with big data processing and analytics on AWS, using services such as Amazon EMR and AWS Batch.
  • Experience in large scale data processing technologies such as Spark.
  • Expertise in orchestrating workflows using tools like Metaflow.
  • Experience with various database technologies including SQL, NoSQL databases (e.g., AWS DynamoDB, ElasticSearch, Postgresql).
  • Hands-on experience with containerization technologies, such as Docker and Kubernetes.
  • Design and implement the data architecture, ensuring scalability, flexibility, and efficiency using pipeline authoring tools like Metaflow and large-scale data processing technologies like Spark.
  • Define and extend our internal standards for style, maintenance, and best practices for a high-scale data platform.
  • Collaborate with researchers and other stakeholders to understand their data needs including model training and production monitoring systems and develop solutions that meet those requirements.
  • Take ownership of key data engineering projects and work independently to design, develop, and maintain high-quality data solutions.
  • Ensure data quality, integrity, and security by implementing robust data validation, monitoring, and access controls.
  • Evaluate and recommend data technologies and tools to improve the efficiency and effectiveness of the data engineering process.
  • Continuously monitor, maintain, and improve the performance and stability of the data infrastructure.

AWSDockerSQLElasticSearchETLKubernetesData engineeringNosqlSparkData modeling

Posted 6 days ago
Apply
Apply

📍 Worldwide

🧭 Full-Time

NOT STATED
  • Own the design and implementation of cross-domain data models that support key business metrics and use cases.
  • Partner with analysts and data engineers to translate business logic into performant, well-documented dbt models.
  • Champion best practices in testing, documentation, CI/CD, and version control, and guide others in applying them.
  • Act as a technical mentor to other analytics engineers, supporting their development and reviewing their code.
  • Collaborate with central data platform and embedded teams to improve data quality, metric consistency, and lineage tracking.
  • Drive alignment on model architecture across domains—ensuring models are reusable, auditable, and trusted.
  • Identify and lead initiatives to reduce technical debt and modernise legacy reporting pipelines.
  • Contribute to the long-term vision of analytics engineering at Pleo and help shape our roadmap for scalability and impact.

SQLData AnalysisETLData engineeringCI/CDMentoringDocumentationData visualizationData modelingData analyticsData management

Posted 6 days ago
Apply