Apply

Senior Data Engineer

Posted about 21 hours agoViewed

View full description

๐Ÿ’Ž Seniority level: Senior, 5+ years or 3+ years with a Master's degree

๐Ÿ“ Location: United States, Canada

๐Ÿ’ธ Salary: 125000.0 - 160000.0 USD per year

๐Ÿ” Industry: Digital driver assistance services

๐Ÿข Company: Agero๐Ÿ‘ฅ 1001-5000๐Ÿ’ฐ $4,750,000 over 2 years agoAutomotiveInsurTechInformation TechnologyInsurance

๐Ÿ—ฃ๏ธ Languages: English

โณ Experience: 5+ years or 3+ years with a Master's degree

๐Ÿช„ Skills: AWSPythonSQLApache AirflowDynamoDBETLFlaskMongoDBSnowflakeFastAPIPandasCI/CDData modeling

Requirements:
  • Bachelor's degree in a technical field and 5+ years or Master's degree with 3+ years of industry experience.
  • Extensive experience with Snowflake or other cloud-based data warehousing solutions.
  • Expertise in ETL/ELT pipelines using tools like Airflow, DBT, Fivetran.
  • Proficiency in Python for data processing and advanced SQL for managing databases.
  • Solid understanding of data modeling techniques and cost management strategies.
  • Experience with data quality frameworks and deploying data solutions in the cloud.
  • Familiarity with version control systems and implementing CI/CD pipelines.
Responsibilities:
  • Develop and maintain ETL/ELT pipelines to ingest data from diverse sources.
  • Monitor and optimize cloud costs while performing query optimization in Snowflake.
  • Establish modern data architectures including data lakes and warehouses.
  • Apply dimensional modeling techniques and develop transformations using DBT or Spark.
  • Write reusable and efficient code, and develop data-intensive UIs and dashboards.
  • Implement data quality frameworks and observability solutions.
  • Collaborate cross-functionally and document data flows, processes, and architecture.
Apply

Related Jobs

Apply

๐Ÿ“ United States of America

๐Ÿงญ Full-Time

๐Ÿ’ธ 110000.0 - 160000.0 USD per year

๐Ÿ” Insurance industry

๐Ÿข Company: Verikai_External

  • Bachelor's degree or above in Computer Science, Data Science, or a related field.
  • At least 5 years of relevant experience.
  • Proficient in SQL, Python, and data processing frameworks such as Spark.
  • Hands-on experience with AWS services including Lambda, Athena, Dynamo, Glue, Kinesis, and Data Wrangler.
  • Expertise in handling large datasets using technologies like Hadoop and Spark.
  • Experience working with PII and PHI under HIPAA constraints.
  • Strong commitment to data security, accuracy, and compliance.
  • Exceptional ability to communicate complex technical concepts to stakeholders.
  • Design, build, and maintain robust ETL processes and data pipelines for large-scale data ingestion and transformation.
  • Manage third-party data sources and customer data to ensure clean and deduplicated datasets.
  • Develop scalable data storage systems using cloud platforms like AWS.
  • Collaborate with data scientists and product teams to support data needs.
  • Implement data validation and quality checks, ensuring accuracy and compliance with regulations.
  • Integrate new data sources to enhance the data ecosystem and document data strategies.
  • Continuously optimize data workflows and research new tools for the data infrastructure.

AWSPythonSQLDynamoDBETLSpark

Posted 8 days ago
Apply
Apply

๐Ÿ“ Canada

๐Ÿ” Home service technology

๐Ÿข Company: Jobber๐Ÿ‘ฅ 501-1000๐Ÿ’ฐ $100,000,000 Series D almost 2 years agoSaaSMobileSmall and Medium BusinessesTask Management

  • Excellent ETL pipeline development skills and hands-on experience with Orchestration (Airflow).
  • Experience with CI/CD practices and optimizing data flow within high-volume infrastructures.
  • Experience with data ingestion systems (e.g., Fivetran, Airbyte) and reverse ETL systems (e.g., HighTouch, Census).
  • Expertise in dimensional modeling, star schemas, and warehousing concepts.
  • Experience with message queues (e.g., Kafka) and real-time stream processing.
  • Proficiency in designing and maintaining efficient data pipelines within cloud infrastructure (preferably AWS).
  • Strong SQL skills and knowledge of containerization (ECS orchestration) and AWS Lambda.
  • Ability to explain complex data concepts to technical and non-technical stakeholders.
  • Empower the Team: Develop tools, frameworks, and workflows to enhance data accessibility and enable data-driven decision-making.
  • Build robust alerting and monitoring systems for data quality and reliability.
  • Collaborate with various functions to support analyses that influence business decisions.
  • Work closely with software engineering teams for collaborative data development.
  • Accelerate Business Growth: Synchronize data between Jobber and external systems. Streamline ETL workflows using tools like Airflow and dbt.
  • Build data models for easier integration.
  • Strategize and innovate: Research emerging technologies to strengthen the data stack.
  • Participate in design and code reviews providing mentorship and knowledge sharing.
  • Ensure Data Integrity: Establish best practices for maintaining data quality.

AWSGraphQLSQLApache AirflowETLKafkaData engineeringNosqlCI/CDRESTful APIsData modeling

Posted 8 days ago
Apply
Apply

๐Ÿ“ Canada

๐Ÿงญ Full-Time

๐Ÿ” Smart home technology

  • Proficiency in building data pipelines using Python, SQL, Apache Spark, Apache Kafka, and Apache Airflow.
  • Experience with cloud-based data platforms, GCP preferred.
  • Good understanding of the machine learning lifecycle and supporting data preparation.
  • Strong experience in data modeling and optimization for data warehousing solutions.
  • Excellent communication skills for collaboration and conveying technical concepts.
  • Design, build, and maintain scalable ETL/ELT pipelines for data ingestion and transformation.
  • Integrate diverse data sources into centralized data repositories.
  • Monitor, optimize, and ensure reliability of data workflows.
  • Collaborate with cross-functional teams to understand data requirements.
  • Implement data quality checks and create comprehensive documentation.

PythonSQLApache AirflowETLGCPIoTMachine LearningApache KafkaData modeling

Posted 9 days ago
Apply
Apply

๐Ÿ“ Canada

๐Ÿงญ Full-Time

๐Ÿ” Technology for small businesses

๐Ÿข Company: Jobber๐Ÿ‘ฅ 501-1000๐Ÿ’ฐ $100,000,000 Series D almost 2 years agoSaaSMobileSmall and Medium BusinessesTask Management

  • Proven ability to lead and collaborate in team environments.
  • Strong coding skills in Python and SQL.
  • Expertise in building and maintaining ETL pipelines using tools like Airflow and dbt.
  • Experience with AWS tools such as Redshift, Glue, and Lambda.
  • Familiarity with handling large datasets using tools like Spark.
  • Experience with Terraform for infrastructure management.
  • Knowledge of dimensional modelling, star schemas, and data warehousing.
  • Design, develop, and maintain batch and real-time data pipelines within cloud infrastructure (preferably AWS).
  • Develop tools that automate processes and set up monitoring systems.
  • Collaborate with teams to extract actionable insights from data.
  • Lead initiatives to propose new technologies, participate in design and code reviews, and maintain data integrity.

AWSPythonSQLApache AirflowETLSparkTerraform

Posted 14 days ago
Apply
Apply

๐Ÿ“ South Africa, Mauritius, Kenya, Nigeria

๐Ÿ” Technology, Marketplaces

  • BSc degree in Computer Science, Information Systems, Engineering, or related technical field or equivalent work experience.
  • 3+ years related work experience.
  • Minimum of 2 years experience building and optimizing โ€˜big dataโ€™ data pipelines, architectures and maintaining data sets.
  • Experienced in Python.
  • Experienced in SQL (PostgreSQL, MS SQL).
  • Experienced in using cloud services: AWS, Azure or GCP.
  • Proficiency in version control, CI/CD and GitHub.
  • Understanding/experience in Glue and PySpark highly desirable.
  • Experience in managing data life cycle.
  • Proficiency in manipulating, processing and architecting large disconnected data sets for analytical requirements.
  • Ability to maintain and optimise processes supporting data transformation, data structures, metadata, dependency and workload management.
  • Good understanding of data management principles - data quality assurance and governance.
  • Strong analytical skills related to working with unstructured datasets.
  • Understanding of message queuing, stream processing, and highly scalable โ€˜big dataโ€™ datastores.
  • Strong attention to detail.
  • Good communication and interpersonal skills.
  • Suggest efficiencies and execute on implementation of internal process improvements in automating manual processes.
  • Implement enhancements and new features across data systems.
  • Improve streamline processes within data systems with support from Senior Data Engineer.
  • Test CI/CD process for optimal data pipelines.
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Highly efficient in ETL processes.
  • Develop and conduct unit tests on data pipelines as well as ensuring data consistency.
  • Develop and maintain automated monitoring solutions.
  • Support reporting and analytics infrastructure.
  • Maintain data quality and data governance as well as upkeep of overall maintenance of data infrastructure systems.
  • Maintain data warehouse and data lake metadata, data catalogue, and user documentation for internal business users.
  • Ensure best practice is implemented and maintained on database.

AWSPostgreSQLPythonSQLETLGitCI/CD

Posted 21 days ago
Apply
Apply
๐Ÿ”ฅ Senior Data Engineer
Posted about 1 month ago

๐Ÿ“ US

๐Ÿ’ธ 103200.0 - 128950.0 USD per year

๐Ÿ” Genetics and healthcare

๐Ÿข Company: Natera๐Ÿ‘ฅ 1001-5000๐Ÿ’ฐ $250,000,000 Post-IPO Equity over 1 year ago๐Ÿซ‚ Last layoff almost 2 years agoWomen'sBiotechnologyMedicalGeneticsHealth Diagnostics

  • BS degree in computer science or a comparable program or equivalent experience.
  • 8+ years of overall software development experience, ideally in complex data management applications.
  • Experience with SQL and No-SQL databases including Dynamo, Cassandra, Postgres, Snowflake.
  • Proficiency in data technologies such as Hive, Hbase, Spark, EMR, Glue.
  • Ability to manipulate and extract value from large datasets.
  • Knowledge of data management fundamentals and distributed systems.
  • Work with other engineers and product managers to make design and implementation decisions.
  • Define requirements in collaboration with stakeholders and users to create reliable applications.
  • Implement best practices in development processes.
  • Write specifications, design software components, fix defects, and create unit tests.
  • Review design proposals and perform code reviews.
  • Develop solutions for the Clinicogenomics platform utilizing AWS cloud services.

AWSPythonSQLAgileDynamoDBSnowflakeData engineeringPostgresSparkData modelingData management

Posted about 1 month ago
Apply
Apply
๐Ÿ”ฅ Senior Data Engineer
Posted about 2 months ago

๐Ÿ“ United States, United Kingdom, Spain, Estonia

๐Ÿ” Identity verification

๐Ÿข Company: Veriff๐Ÿ‘ฅ 501-1000๐Ÿ’ฐ $100,000,000 Series C about 3 years ago๐Ÿซ‚ Last layoff over 1 year agoArtificial Intelligence (AI)Fraud DetectionInformation TechnologyCyber SecurityIdentity Management

  • Expert-level knowledge of SQL, particularly with Redshift.
  • Strong experience in data modeling with an understanding of dimensional data modeling best practices.
  • Proficiency in data transformation frameworks like dbt.
  • Solid programming skills in languages used in data engineering, such as Python or R.
  • Familiarity with orchestration frameworks like Apache Airflow or Luigi.
  • Experience with data from diverse sources including RDBMS and APIs.
  • Collaborate with business stakeholders to design, document, and implement robust data models.
  • Build and optimize data pipelines to transform raw data into actionable insights.
  • Fine-tune query performance and ensure efficient use of data warehouse infrastructure.
  • Ensure data reliability and quality through rigorous testing and monitoring.
  • Assist in migrating from batch processing to real-time streaming systems.
  • Expand support for various use cases including business intelligence and analytics.

PythonSQLApache AirflowETLData engineeringJSONData modeling

Posted about 2 months ago
Apply
Apply
๐Ÿ”ฅ Senior Data Engineer
Posted about 2 months ago

๐Ÿ“ United States

๐Ÿงญ Full-Time

๐Ÿข Company: Avalore, LLC

  • Masterโ€™s or PhD in statistics, mathematics, computer science, or related field.
  • 8+ years of experience as a Data Engineer within the IC.
  • Outstanding communication skills, influencing abilities, and client focus.
  • Professional proficiency in English is required.
  • Current, active Top Secret security clearance.
  • Applicants must be currently authorized to work in the United States on a full-time basis.
  • Develops and documents data pipelines for ingest, transformation, and preparation of data for AI applications.
  • Designs scalable technologies such as streaming and transformation, joining disparate data sets for predictive analytics.
  • Develops API interfaces for accessibility.
  • Leads technical efforts and guides development teams.

PythonSQLApache AirflowArtificial IntelligenceETLMachine LearningAPI testingData engineering

Posted about 2 months ago
Apply
Apply
๐Ÿ”ฅ Senior Data Engineer
Posted about 2 months ago

๐Ÿ“ USA

๐Ÿงญ Full-Time

๐Ÿ’ธ 190000.0 - 220000.0 USD per year

๐Ÿ” B2B data / Data as a Service (DaaS)

๐Ÿข Company: People Data Labs๐Ÿ‘ฅ 101-250๐Ÿ’ฐ $45,000,000 Series B about 3 years agoDatabaseArtificial Intelligence (AI)Developer APIsMachine LearningAnalyticsB2BSoftware

  • 5-7+ years industry experience with strategic technical problem-solving.
  • Strong software development fundamentals.
  • Experience with Python.
  • Expertise in Apache Spark (Java, Scala, or Python-based).
  • Proficiency in SQL.
  • Experience building scalable data processing systems.
  • Familiarity with data pipeline orchestration tools (e.g., Airflow, dbt).
  • Knowledge of modern data design and storage patterns.
  • Experience working in Databricks.
  • Familiarity with cloud computing services (e.g., AWS, GCP, Azure).
  • Experience in data warehousing technologies.
  • Understanding of modern data storage formats and tools.
  • Build infrastructure for ingestion, transformation, and loading of data using Spark, SQL, AWS, and Databricks.
  • Create an entity resolution framework for merging billions of entities into clean datasets.
  • Develop CI/CD pipelines and anomaly detection systems to enhance data quality.
  • Provide solutions to undefined data engineering problems.
  • Assist Engineering and Product teams with data-related technical issues.

AWSPythonSQLKafkaAirflowData engineeringPandasCI/CD

Posted about 2 months ago
Apply
Apply

๐Ÿ“ Paris, New York, San Francisco, Sydney, Madrid, London, Berlin

๐Ÿ” Communication technology

  • Passionate about data engineering.
  • Experience in designing and developing data infrastructure.
  • Technical skills to solve complex challenges.
  • Play a crucial role in designing, developing, and maintaining data infrastructure.
  • Collaborate with teams across the company to solve complex challenges.
  • Improve operational efficiency and lead business towards strategic goals.
  • Contribute to engineering efforts that enhance customer journey.

AWSPostgreSQLPythonSQLApache AirflowETLData engineering

Posted 3 months ago
Apply