Apply

Senior Data Engineer

Posted 5 days agoViewed

View full description

💎 Seniority level: Senior, 5 years

📍 Location: Boston, MA; Vancouver, BC; Chicago, IL; and Vancouver, WA

💸 Salary: 160000.0 - 190000.0 USD per year

🔍 Industry: Social Media Marketing

🏢 Company: Later👥 1-10Consumer ElectronicsiOSAppsSoftware

⏳ Experience: 5 years

🪄 Skills: AWSSQLApache AirflowCloud ComputingETLData engineering

Requirements:
  • Minimum of 5 years in data engineering or related fields, with a strong focus on building data infrastructure and pipelines.
  • Bachelor’s degree in Computer Science, Engineering, or a related technical field; advanced degree preferred.
Responsibilities:
  • Design and build a robust data warehouse architecture.
  • Design, build, and maintain scalable data pipelines for both batch and real-time processing, ensuring high availability and reliability.
  • Develop reliable transformation layers and data pipelines from ambiguous business processes using tools like DBT.
  • Establish optimized data architectures using cloud technologies, and implement both batch and streaming data processing systems.
  • Enforce data quality checks and governance practices to maintain data integrity and compliance.
  • Work with data scientists, product managers, and business stakeholders to understand data needs and deliver actionable insights.
  • Analyze and optimize data pipelines for performance and cost-effectiveness.
Apply

Related Jobs

Apply
🔥 Senior Data Engineer
Posted 2 days ago

📍 Worldwide

🧭 Full-Time

🔍 Software Development

🏢 Company: Kit👥 11-50💰 over 1 year agoEducationFinancial ServicesApps

  • Strong command of SQL, including DDL and DML.
  • Proficient in Python
  • Strong understanding of DBMS internals, including an appreciation for platform-specific nuances.
  • A willingness to work with Redshift and deeply understand its nuances.
  • Familiarity with our key tools (Redshift, Segment, dbt, github)
  • 8+ years in data, with at least 3 years specializing in Data Engineering
  • Proven track record managing and optimizing OLAP clusters
  • Experience refactoring problematic data pipelines without disrupting business operations
  • History of implementing data quality frameworks and validation processes
  • Dive into our Redshift warehouse, dbt models, and workflows.
  • Evaluate the CRM data lifecycle, including source extraction, warehouse ingestion, transformation, and reverse ETL.
  • Refine and start implementing your design for source extraction and warehouse ingestion.
  • Complete the implementation of the CRM source extraction/ingestion project and use the learnings to refine your approach in preparation for other, similar initiatives including, but by no means limited to web traffic events and product usage logs.

PythonSQLETLGitData engineeringRDBMSData modelingData management

Posted 2 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 2 days ago

📍 United States of America

🏢 Company: IDEXX

  • Bachelor’s degree in Computer Science, Computer Engineering, Information Systems, Information Systems Engineering or a related field and 5 years of experience or Master’s degree in Computer Science, Computer Engineering, Information Systems, Information Systems Engineering or a related field and 3 years of related professional experience.
  • Advanced SQL knowledge and experience working with relational databases, including Snowflake, Oracle, Redshift.
  • Experience with AWS or Azure cloud platforms
  • Experience with data pipeline and workflow scheduling tools: Apache Airflow, Informatica.
  • Experience with ETL/ELT tools and data processing techniques
  • Experience in database design, development, and modeling
  • 3 years of related professional experience with object-oriented languages: Python, Java, and Scala
  • Design and implement scalable, reliable distributed data processing frameworks and analytical infrastructure
  • Design metadata and schemas for assigned projects based on a logical model
  • Create scripts for physical data layout
  • Write scripts to load test data
  • Validate schema design
  • Develop and implement node cluster models for unstructured data storage and metadata
  • Design advanced level Structured Query Language (SQL), data definition language (DDL) and Python scripts
  • Define, design, and implement data management, storage, backup and recovery solutions
  • Design automated software deployment functionality
  • Monitor structural performance and utilization, identifying problems and implements solutions
  • Lead the creation of standards, best practices and new processes for operational integration of new technology solutions
  • Ensures environments are compliant with defined standards and operational procedures
  • Implement measures to ensure data accuracy and accessibility, constantly monitoring and refining the performance of data management systems

AWSPythonSQLApache AirflowCloud ComputingETLJavaOracleSnowflakeAzureData engineeringScalaData modelingData management

Posted 2 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 3 days ago

📍 United States

🧭 Full-Time

💸 145000.0 - 200000.0 USD per year

🔍 Daily Fantasy Sports

🏢 Company: PrizePicks👥 101-250💰 Corporate about 2 years agoGamingFantasy SportsSports

  • 5+ years of experience in a data Engineering, or data-oriented software engineering role creating and pushing end-to-end data engineering pipelines.
  • 2+ years of experience acting as technical lead and providing mentorship and feedback to junior engineers.
  • Extensive experience building and optimizing cloud-based data streaming pipelines and infrastructure.
  • Extensive experience exposing real-time predictive model outputs to production-grade systems leveraging large-scale distributed data processing and model training.
  • Experience in most of the following: SQL/NoSQL databases/warehouses: Postgres, BigQuery, BigTable, Materialize, AlloyDB, etc Replication/ELT services: Data Stream, Hevo, etc. Data Transformation services: Spark, Dataproc, etc Scripting languages: SQL, Python, Go. Cloud platform services in GCP and analogous systems: Cloud Storage, Cloud Compute Engine, Cloud Functions, Kubernetes Engine etc. Data Processing and Messaging Systems: Kafka, Pulsar, Flink Code version control: Git Data pipeline and workflow tools: Argo, Airflow, Cloud Composer. Monitoring and Observability platforms: Prometheus, Grafana, ELK stack, Datadog Infrastructure as Code platforms: Terraform, Google Cloud Deployment Manager. Other platform tools such as Redis, FastAPI, and Streamlit.
  • Enhance the capabilities of our existing Core Data platforms and develop new integrations with both internal and external APIs within the Data organization.
  • Work closely with DevOps, architects, and engineers to ensure the success of the Core Data platform.
  • Collaborate with Analytics Engineers to enhance data transformation processes, streamline CI/CD pipelines, and optimize team collaboration workflows.
  • Architect and implement Infrastructure as Code (IaC) solutions to automate and streamline the deployment and management of data infrastructure.
  • Develop and manage CI/CD pipelines to automate and streamline the deployment of data solutions.
  • Ensure code is thoroughly tested, effectively integrated, and efficiently deployed, in alignment with industry best practices for version control, automation, and quality assurance.
  • Serve as a Data Engineering thought leader within the broader PrizePicks technology organization by staying current with emerging technologies, implementing innovative solutions, and sharing knowledge and best practices with junior team members and collaborators.
  • Provide on-call support as part of a shared rotation between the Data and Analytics Engineering teams to maintain system reliability and respond to critical issues.

LeadershipPostgreSQLPythonSQLApache AirflowBashCloud ComputingETLGCPGitKafkaKubernetesData engineeringData scienceREST APICI/CDRESTful APIsMentoringTerraformData modeling

Posted 3 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 135000.0 - 160000.0 USD per year

🔍 Healthcare

🏢 Company: Jobgether👥 11-50💰 $1,493,585 Seed over 2 years agoInternet

  • 5+ years of experience in data engineering roles, preferably in fast-paced or data-centric environments
  • Proficient in SQL and experienced with data warehouses such as Snowflake or Redshift
  • Strong experience with cloud platforms (AWS, GCP, or Azure)
  • Familiarity with workflow management tools like Apache Airflow or Luigi
  • Knowledge of data modeling, warehousing architecture, and pipeline automation best practices
  • Degree in Computer Science, Engineering, Mathematics, or related field (Master’s preferred)
  • Familiarity with healthcare data standards like FHIR or HL7 is a plus
  • Strong problem-solving skills and ability to adapt in a dynamic environment
  • Build, optimize, and maintain highly scalable and reliable data pipelines
  • Collaborate with data scientists and analysts to meet data needs across the business
  • Automate data cleansing, validation, transformation, and mining processes
  • Improve internal data workflows and automate manual processes to enhance scalability
  • Troubleshoot data issues, ensure security compliance, and support infrastructure-related inquiries
  • Deliver high-quality data solutions that empower cross-functional teams with actionable insights

AWSSQLApache AirflowETLGCPSnowflakeAzureData engineeringData modeling

Posted 5 days ago
Apply
Apply

📍 Canada

🧭 Full-Time

💸 125800.0 - 170100.0 USD per year

🔍 Software Development

  • Strong coding skills in Python and SQL.
  • Expertise in building and maintaining ETL pipelines using tools like Airflow and dbt.
  • Experience working with AWS data infrastructure, particularly Redshift, Glue, Lambda, and ECS Fargate.
  • Familiarity with handling large datasets using tools like Spark or similar (e.g., Trino).
  • Experience with Terraform for infrastructure management.
  • Experience with dimensional modelling, star schemas, and data warehousing in a cloud environment (preferably AWS Redshift).
  • Knowledge of CI/CD processes, data ingestion, and optimizing data flow across systems.
  • Proficient in working with high-volume, scalable data infrastructure.
  • Ability to collaborate effectively with both technical and non-technical teams, explaining complex data concepts in a clear and concise manner.
  • Build Scalable Data Solutions: Design, develop, and maintain batch and real-time data pipelines within cloud infrastructure (preferably AWS). Leverage Python, SQL, and AWS technologies (Glue, Lambda, ECS Fargate) to ensure smooth data operations. Build scripts, serverless applications, and automated workflows.
  • Empower Internal Teams: Develop tools and frameworks that automate manual processes, set up alerting/monitoring systems, and help teams run data-driven experiments and analyze results. Work closely with cross-functional teams to support their needs and ensure data accessibility.
  • Accelerate Business Growth: Collaborate with data analysts, scientists, and product teams to extract actionable insights from data. Utilize tools like Airflow and dbt to streamline ETL/ELT pipelines and ensure the seamless flow of data.
  • Strategic Planning and Innovation: Lead initiatives to research and propose new technologies and tooling for our data stack, with an emphasis on performance and scalability. Participate in design and code reviews, continuously learning from and mentoring your peers.
  • Data Integrity: Own the integrity of our data and maintain a high level of trust across the organization.

AWSPythonSQLETLAirflowApache KafkaData engineeringSparkCI/CDTerraformData modeling

Posted 6 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 6 days ago

📍 Canada, Spain

🧭 Full-Time

🔍 Software Development

  • Proven experience writing production-grade code in Python, particularly for data ingestion, transformation, and pipeline orchestration.
  • Strong command of SQL for working with large databases, building complex queries, and optimizing for performance in analytical environments.
  • Hands-on experience using DBT to build, test, and maintain data models that support analytics and business decision-making.
  • Experience designing and maintaining ETL/ELT pipelines for large-scale data processing, including querying APIs for data ingestion.
  • Design and implement robust batch and real-time pipelines using Python and SQL across multi-cloud environments.
  • Develop and optimize modular, testable DBT models to transform raw event and operational data into business-ready datasets that power reporting and analysis.
  • Gather requirements from stakeholders and shape them into well-structured datasets and data products that reflect business logic in a multi-cloud environment.

PostgreSQLPythonSQLApache AirflowETLGCPKafkaData engineeringCI/CDProblem SolvingData modelingData analytics

Posted 6 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 11 days ago

📍 United States of America

💸 78750.0 - 133875.0 USD per year

🏢 Company: vspvisioncareers

  • 6+ years’ experience working in development team providing analytical capabilities
  • 6+ years of hands-on experience in the data space, spanning data preparation, SQL, integration tools, ETL/ELT/data pipeline design
  • SQL coding experience
  • Experience working in an agile development environment (Scrum, Kanban) with a focus on Continuous Integration and Delivery
  • Knowledge about various data architectures, patterns, and capabilities such as event-driven architecture, real-time data flows, non-relational repositories, data virtualization, cloud storage, etc
  • Knowledge of and experience with multiple data integration platforms (IBM InfoSphere DataStage, Oracle Data Integrator, Informatica PowerCenter, MS SSIS, AWS Glue, Denodo), and data warehouse MPP platforms such as Snowflake, Netezza, Teradata, Redshift, etc
  • Collaborate within an agile, multi-disciplinary team to deliver optimal data integration and transformation solutions
  • Analyze data requirements (functional and non-functional) to develop and design robust, scalable, automated, fault-tolerant data pipeline solutions for business and technology initiatives
  • Design, build, maintain, and operationalize data pipelines for high volume and complex data using appropriate tools and practices in development, test, and production environments
  • Develop and design data mappings, programs, routines, and SQL to acquire data from legacy, web, cloud, and purchased package environments into the analytics environment
  • Drive automation of data pipeline preparation and integration tasks to minimize manual and error-prone processes and improve productivity using modern data preparation, integration, and AI-enabled metadata management tools and techniques
  • Participate in architecture, governance, and design reviews, identifying opportunities and making recommendations
  • Collaborate with architects to design and model application data structures, storage, and integration in accordance with enterprise-wide architecture standards across legacy, web, cloud, and purchased package environments

AWSSQLAgileETLSnowflakeApache KafkaData engineeringCI/CDRESTful APIsData visualizationData modelingData management

Posted 11 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 12 days ago

📍 United States

🧭 Full-Time

💸 135000.0 - 145000.0 USD per year

🔍 Life Science

🏢 Company: Medispend

  • Hands-on knowledge of data integration platforms
  • Experience with enterprise systems (ERP, CRM, etc.)
  • Substantial programming experience with Python based data orchestration and transformation frameworks (i.e. Airflow, AWS Glue, Prefect, Dagster, Spark, Polars, Databricks, etc.)
  • Strong working knowledge of traditional RDBMS data warehousing as well as other platforms like Snowflake, RedShift
  • Gather requirements and design the integration of a new data source
  • Design and build data transformations
  • Estimate levels of effort for prospective client implementations
  • Evaluate new open source data management tools
  • Determine root cause for a failed integration
  • Conduct peer review for code check-ins
  • Design and build a data migration framework
  • Monitor infrastructure capacity of the data transformation platform
  • Compile and analyze data transformation success/failure rates

AWSProject ManagementPythonSQLApache AirflowCloud ComputingETLSnowflakeJiraAlgorithmsData engineeringData StructuresPostgresRDBMSREST APISparkCommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingAgile methodologiesDevOpsData visualizationData modelingScriptingData analyticsData managementSaaS

Posted 12 days ago
Apply
Apply

📍 Canada

🧭 Full-Time

💸 125800.0 - 170100.0 USD per year

🔍 Software Development

🏢 Company: Jobber👥 501-1000💰 $100,000,000 Series D over 2 years agoSaaSMobileSmall and Medium BusinessesTask Management

  • Strong coding skills in Python and SQL.
  • Expertise in building and maintaining ETL pipelines using tools like Airflow and dbt.
  • Experience working with AWS data infrastructure, particularly Redshift, Glue, Lambda, and ECS Fargate.
  • Familiarity with handling large datasets using tools like Spark or similar (e.g., Trino).
  • Experience with Terraform for infrastructure management.
  • Experience with dimensional modelling, star schemas, and data warehousing in a cloud environment (preferably AWS Redshift).
  • Knowledge of CI/CD processes, data ingestion, and optimizing data flow across systems.
  • Proficient in working with high-volume, scalable data infrastructure.
  • Build Scalable Data Solutions: Design, develop, and maintain batch and real-time data pipelines within cloud infrastructure (preferably AWS). Leverage Python, SQL, and AWS technologies (Glue, Lambda, ECS Fargate) to ensure smooth data operations. Build scripts, serverless applications, and automated workflows.
  • Empower Internal Teams: Develop tools and frameworks that automate manual processes, set up alerting/monitoring systems, and help teams run data-driven experiments and analyze results. Work closely with cross-functional teams to support their needs and ensure data accessibility.
  • Accelerate Business Growth: Collaborate with data analysts, scientists, and product teams to extract actionable insights from data. Utilize tools like Airflow and dbt to streamline ETL/ELT pipelines and ensure the seamless flow of data.
  • Strategic Planning and Innovation: Lead initiatives to research and propose new technologies and tooling for our data stack, with an emphasis on performance and scalability. Participate in design and code reviews, continuously learning from and mentoring your peers.
  • Data Integrity: Own the integrity of our data and maintain a high level of trust across the organization.

AWSPythonSQLAirflowData engineeringSparkCI/CDTerraformData modeling

Posted 12 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 13 days ago

📍 United States

🧭 Contract

  • Experience with Dataiku.
  • Expertise in IDMC (Informatica Data Management Cloud).
  • Strong knowledge of SQL statements.
  • Basic experience with Python.
  • Knowledge of cloud-based data warehousing solutions, like Snowflake
  • Utilize Dataiku for data preparation, analysis, and workflow automation.
  • Deconstruct complex SQL statements to understand data flows and transformation logic.
  • Migrate data to IDMC (Informatica Data Management Cloud) ensuring quality and validation.
  • Use Snowflake to extract data and convert it into formats compatible with IDMC.
  • Work closely with the business team to validate data accuracy and ensure alignment with business requirements.
  • Provide support for 1-2 products, such as VSM and Flash, ensuring data-related needs are met.

PythonSQLETLSnowflakeData engineeringData visualizationData modeling

Posted 13 days ago
Apply