Apply

Senior Data Engineer

Posted about 2 months agoViewed

View full description

๐Ÿ’Ž Seniority level: Senior, 5+ years

๐Ÿ“ Location: United States

๐Ÿ’ธ Salary: 100000.0 - 120000.0 USD per year

๐Ÿ” Industry: Healthcare Technology

๐Ÿข Company: Found๐Ÿ‘ฅ 51-100๐Ÿ’ฐ $45,999,997 Series C 9 months agoFinancial ServicesBankingFinTech

๐Ÿ—ฃ๏ธ Languages: English

โณ Experience: 5+ years

๐Ÿช„ Skills: PythonSQLApache AirflowETLSnowflakePandasSpark

Requirements:
  • Bachelorโ€™s or Masterโ€™s degree in Computer Science or related field
  • 5+ years of experience in data engineering
  • Expertise in SQL and data manipulation languages
  • Proficiency in data pipeline tools (Airflow, AWS Glue, Spark/PySpark, Pandas)
  • Strong programming skills in Python
  • Experience with Snowflake, Redshift, Databricks, Glue Catalog/S3
Responsibilities:
  • Design, implement, and manage data pipelines
  • Develop and maintain data models
  • Design and implement data warehousing solutions
  • Develop and optimize ETL processes
  • Implement data quality checks
  • Monitor and optimize data pipelines and queries
  • Collaborate with data analysts and stakeholders
  • Create and maintain documentation
Apply

Related Jobs

Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ” B2B SaaS

๐Ÿข Company: Sanity

  • 4+ years of experience building data pipelines at scale
  • Deep expertise in SQL, Python, and Node.js/TypeScript
  • Production experience with Airflow and RudderStack
  • Track record of building reliable data infrastructure
  • Design, develop, and maintain scalable ETL/ELT pipelines
  • Collaborate to implement and scale product telemetry
  • Establish best practices for data ingestion and transformation
  • Monitor and optimize data pipeline performance

Node.jsPythonSQLApache AirflowETLTypeScript

Posted 2 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ” E-commerce

  • Bachelor's or Master's degree in Computer Science or related field
  • 5+ years of experience in data engineering
  • Strong proficiency in SQL and database technologies
  • Experience with data pipeline orchestration tools
  • Proficiency in programming languages like Python and Scala
  • Hands-on experience with AWS cloud data services
  • Familiarity with big data frameworks like Apache Spark
  • Knowledge of data modeling and warehousing
  • Experience implementing CI/CD for data pipelines
  • Real-time data processing architectures experience
  • Design, develop, and maintain ETL/ELT pipelines
  • Optimize data architecture and storage solutions
  • Work with AWS for scalable data solutions
  • Ensure data quality, integrity, and security
  • Collaborate with cross-functional teams
  • Monitor and troubleshoot data workflows
  • Create APIs for analytical information

AWSPostgreSQLPythonSQLApache AirflowETLKafkaMySQLSnowflakeCI/CDScala

Posted 3 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ” Software Development

๐Ÿข Company: BioRender๐Ÿ‘ฅ 101-250๐Ÿ’ฐ $15,319,133 Series A almost 2 years agoLife ScienceGraphic DesignSoftware

  • 7+ years of data engineering experience of relevant industry experience
  • Expertise working with Data Warehousing platforms (AWS RedShift or Snowflake preferred) and data lake / lakehouse architectures
  • Experience with Data Streaming platforms (AWS Kinesis / Firehose preferred)
  • Expertise with SQL and programming languages commonly used in data platforms (Python, Spark, etc)
  • Experience with data pipeline orchestration (e.g., Airflow) and data pipeline integrations (e.g. Airbyte, Stitch)
  • Building and maintaining the right architecture and tooling to support our data science, analytics, product, and machine learning initiatives.
  • Solve complex architectural problems
  • Translate deeply technical designs into business appropriate representations as well as analyze business needs and requirements ensuring implementation of data services directly correlates to the strategy and growth of the business

AWSPythonSQLApache AirflowSnowflakeData engineeringSparkData modeling

Posted 6 days ago
Apply
Apply

๐Ÿ“ United States

๐Ÿงญ Full-Time

๐Ÿ’ธ 175000.0 - 205000.0 USD per year

๐Ÿ” Software Development

๐Ÿข Company: CoreWeave๐Ÿ’ฐ $642,000,000 Secondary Market about 1 year agoCloud ComputingMachine LearningInformation TechnologyCloud Infrastructure

  • Hands-on experience applying Kimball Dimensional Data Modeling principles to large datasets.
  • Expertise in working with analytical table/file formats, including Iceberg, Parquet, Avro, and ORC.
  • Proven experience optimizing MPP databases (StarRocks, Snowflake, BigQuery, Redshift).
  • Minimum 5+ years of programming experience in Python or Scala.
  • Advanced SQL skills, with a strong ability to write, optimize, and debug complex queries.
  • Hands-on experience with Airflow for batch orchestration distributed computing frameworks like Spark or Flink.
  • Develop and maintain data models, including star and snowflake schemas, to support analytical needs across the organization.
  • Establish and enforce best practices for dimensional modeling in our Lakehouse.
  • Engineer and optimize data storage using analytical table/file formats (e.g., Iceberg, Parquet, Avro, ORC).
  • Partner with BI, analytics, and data science teams to design datasets that accurately reflect business metrics.
  • Tune and optimize data in MPP databases such as StarRocks, Snowflake, BigQuery, or Redshift.
  • Collaborate on data workflows using Airflow, building and managing pipelines that power our analytical infrastructure.
  • Ensure efficient processing of large datasets through distributed computing frameworks like Spark or Flink.

AWSDockerPythonSQLCloud ComputingETLKubernetesSnowflakeAirflowAlgorithmsApache KafkaData engineeringData StructuresREST APISparkCommunication SkillsAnalytical SkillsCollaborationCI/CDRESTful APIsDevOpsTerraformProblem-solving skillsJSONScalaData visualizationAnsibleData modelingData analyticsDebugging

Posted 12 days ago
Apply
Apply

๐Ÿ“ OR, WA, CA, CO, TX, IL

๐Ÿงญ Contract

๐Ÿ’ธ 65.0 - 75.0 USD per hour

๐Ÿ” Music industry

๐Ÿข Company: Discogs๐Ÿ‘ฅ 51-100๐Ÿ’ฐ $2,500,000 about 7 years agoDatabaseCommunitiesMusic

  • Proficiency in data integration and ETL processes.
  • Knowledge of programming languages such as Python, Java, or Javascript.
  • Familiarity with cloud platforms and services (e.g., AWS, GCP, Azure).
  • Understanding of data warehousing concepts and technologies (e.g., Redshift, BigQuery, Snowflake).
  • Excellent problem-solving skills and attention to detail.
  • Strong communication and collaboration skills to work effectively with cross-functional teams.
  • Experience with marketing automation platforms.
  • Experience with data warehouses in a marketing context.
  • Knowledge of API integration and data exchange formats such as JSON, XML, and CSV.
  • Design, develop, and maintain data pipelines to ingest, process, and store data.
  • Implement data validation and quality checks to maintain the integrity of incoming data.
  • Optimize and automate data workflows to improve efficiency and reduce manual intervention.
  • Work closely with the product, engineering, marketing and analytics teams to support data-driven decision-making.
  • Develop and maintain documentation related to data processes, workflows, and system architecture.
  • Troubleshoot and resolve data-related issues promptly to minimize disruptions.
  • Monitor and enhance the performance of data infrastructure, ensuring scalability and reliability.
  • Stay updated with industry trends and best practices in data engineering to apply improvements.

AWSPythonApache AirflowETLGCPMySQLSnowflakeApache KafkaAzureJSON

Posted 17 days ago
Apply
Apply

๐Ÿ“ United States

๐Ÿ’ธ 104981.0 - 157476.0 USD per year

๐Ÿ” Mental healthcare

๐Ÿข Company: Headspace๐Ÿ‘ฅ 11-50WellnessHealth CareChild Care

  • 7+ years of proven success designing and implementing large-scale enterprise data systems.
  • Deep experience with industry-leading tools such as Databricks, Snowflake, and Redshift.
  • Demonstrated expertise in architectural patterns for building high-volume real-time and batch ETL pipelines.
  • Proven ability to partner effectively with product teams to drive alignment and deliver solutions.
  • Exceptional oral and written communication abilities.
  • Experience in coaching and mentoring team members.
  • Architect and implement robust data pipelines to ingest, aggregate, and index diverse data sources into the organizationโ€™s data lake.
  • Lead the creation of a secure, compliant, and privacy-focused data warehousing solution tailored to healthcare industry requirements.
  • Partner with the data analytics team to deliver a data platform that supports accurate reporting on business metrics.
  • Collaborate with data science and machine learning teams to build tools for rapid experimentation and innovation.
  • Mentor and coach data engineers while promoting a culture valuing data as a strategic asset.

AWSETLSnowflakeData engineeringData modeling

Posted 21 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Regular

๐Ÿ’ธ 125000.0 - 160000.0 USD per year

๐Ÿ” Digital driver assistance services

๐Ÿข Company: Agero๐Ÿ‘ฅ 1001-5000๐Ÿ’ฐ $4,750,000 over 2 years agoAutomotiveInsurTechInformation TechnologyInsurance

  • Bachelor's degree in a technical field and 5+ years or Master's degree with 3+ years of industry experience.
  • Extensive experience with Snowflake or other cloud-based data warehousing solutions.
  • Expertise in ETL/ELT pipelines using tools like Airflow, DBT, Fivetran.
  • Proficiency in Python for data processing and advanced SQL for managing databases.
  • Solid understanding of data modeling techniques and cost management strategies.
  • Experience with data quality frameworks and deploying data solutions in the cloud.
  • Familiarity with version control systems and implementing CI/CD pipelines.
  • Develop and maintain ETL/ELT pipelines to ingest data from diverse sources.
  • Monitor and optimize cloud costs while performing query optimization in Snowflake.
  • Establish modern data architectures including data lakes and warehouses.
  • Apply dimensional modeling techniques and develop transformations using DBT or Spark.
  • Write reusable and efficient code, and develop data-intensive UIs and dashboards.
  • Implement data quality frameworks and observability solutions.
  • Collaborate cross-functionally and document data flows, processes, and architecture.

AWSPythonSQLApache AirflowDynamoDBETLFlaskMongoDBSnowflakeFastAPIPandasCI/CDData modeling

Posted 24 days ago
Apply
Apply

๐Ÿ“ United States of America

๐Ÿงญ Full-Time

๐Ÿ’ธ 110000.0 - 160000.0 USD per year

๐Ÿ” Insurance industry

๐Ÿข Company: Verikai_External

  • Bachelor's degree or above in Computer Science, Data Science, or a related field.
  • At least 5 years of relevant experience.
  • Proficient in SQL, Python, and data processing frameworks such as Spark.
  • Hands-on experience with AWS services including Lambda, Athena, Dynamo, Glue, Kinesis, and Data Wrangler.
  • Expertise in handling large datasets using technologies like Hadoop and Spark.
  • Experience working with PII and PHI under HIPAA constraints.
  • Strong commitment to data security, accuracy, and compliance.
  • Exceptional ability to communicate complex technical concepts to stakeholders.
  • Design, build, and maintain robust ETL processes and data pipelines for large-scale data ingestion and transformation.
  • Manage third-party data sources and customer data to ensure clean and deduplicated datasets.
  • Develop scalable data storage systems using cloud platforms like AWS.
  • Collaborate with data scientists and product teams to support data needs.
  • Implement data validation and quality checks, ensuring accuracy and compliance with regulations.
  • Integrate new data sources to enhance the data ecosystem and document data strategies.
  • Continuously optimize data workflows and research new tools for the data infrastructure.

AWSPythonSQLDynamoDBETLSpark

Posted about 1 month ago
Apply
Apply
๐Ÿ”ฅ Senior Data Engineer
Posted about 1 month ago

๐Ÿ“ United States

๐Ÿ’ธ 229500.0 - 280500.0 USD per year

๐Ÿ” Event analytics

๐Ÿข Company: Mixpanel๐Ÿ‘ฅ 251-500๐Ÿ’ฐ $200,000,000 Series C over 3 years agoWeb AppsSaaSAnalyticsMobile Apps

  • A strong background in both data and software engineering, with at least 5 years of professional experience.
  • Proficiency with at least one programming language (Python, Java, etc.) and SQL.
  • Excellent debugging and technical investigation skills.
  • Excellent technical communication skills, ideally in a remote environment.
  • Familiarity with rETL tools (e.g., Hightouch, Census, Rudderstack).
  • Experience with modern data storage technologies (e.g., BigQuery SQL, Airflow, DBT or similar).
  • Build and maintain software and data pipelines across backend and data orchestration systems.
  • Design and build data architecture spanning a wide range of complex data.
  • Create and maintain foundational datasets to support analytics, modeling, and product/business needs.
  • Collaborate with, teach, and learn from engineers across the organization.
  • Work with Finance and Data Science to ensure relevance and understanding.
  • Participate in team on-call rotation to maintain system health.
  • Build testing and alerting features for data hygiene.
  • Write internal technical documentation for systems designed and maintained.

PythonSQLApache AirflowETLData engineeringDebugging

Posted about 1 month ago
Apply
Apply
๐Ÿ”ฅ Senior Data Engineer
Posted about 1 month ago

๐Ÿ“ South Africa, Mauritius, Kenya, Nigeria

๐Ÿ” Technology, Marketplaces

  • BSc degree in Computer Science, Information Systems, Engineering, or related technical field or equivalent work experience.
  • 3+ years related work experience.
  • Minimum of 2 years experience building and optimizing โ€˜big dataโ€™ data pipelines, architectures and maintaining data sets.
  • Experienced in Python.
  • Experienced in SQL (PostgreSQL, MS SQL).
  • Experienced in using cloud services: AWS, Azure or GCP.
  • Proficiency in version control, CI/CD and GitHub.
  • Understanding/experience in Glue and PySpark highly desirable.
  • Experience in managing data life cycle.
  • Proficiency in manipulating, processing and architecting large disconnected data sets for analytical requirements.
  • Ability to maintain and optimise processes supporting data transformation, data structures, metadata, dependency and workload management.
  • Good understanding of data management principles - data quality assurance and governance.
  • Strong analytical skills related to working with unstructured datasets.
  • Understanding of message queuing, stream processing, and highly scalable โ€˜big dataโ€™ datastores.
  • Strong attention to detail.
  • Good communication and interpersonal skills.
  • Suggest efficiencies and execute on implementation of internal process improvements in automating manual processes.
  • Implement enhancements and new features across data systems.
  • Improve streamline processes within data systems with support from Senior Data Engineer.
  • Test CI/CD process for optimal data pipelines.
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Highly efficient in ETL processes.
  • Develop and conduct unit tests on data pipelines as well as ensuring data consistency.
  • Develop and maintain automated monitoring solutions.
  • Support reporting and analytics infrastructure.
  • Maintain data quality and data governance as well as upkeep of overall maintenance of data infrastructure systems.
  • Maintain data warehouse and data lake metadata, data catalogue, and user documentation for internal business users.
  • Ensure best practice is implemented and maintained on database.

AWSPostgreSQLPythonSQLETLGitCI/CD

Posted about 1 month ago
Apply