Apply

Senior Data Engineer

Posted 2 days agoViewed

View full description

πŸ’Ž Seniority level: Senior, 5+ years

πŸ“ Location: Canada

πŸ” Industry: Retail Media

🏒 Company: VantageπŸ‘₯ 1001-5000CryptocurrencyFinancial ServicesFinTechTrading Platform

πŸ—£οΈ Languages: English

⏳ Experience: 5+ years

πŸͺ„ Skills: PythonSQLApache AirflowETLKafkaData engineeringData modeling

Requirements:
  • 5+ years of experience in data engineering, big data, or distributed systems.
  • Strong expertise in Python, SQL (or equivalent big data processing frameworks).
  • Proficiency in ETL/ELT pipelines using Apache Airflow, or similar orchestration tools.
  • Experience working with real-time streaming data (Kafka, Kinesis, or Pub/Sub).
  • Strong understanding of data modelling, data warehousing, and distributed systems.
  • Familiarity with privacy-compliant data processing (GDPR, CCPA) for advertising/retail media use cases.
Responsibilities:
  • Design, develop, and optimize data pipelines, ETL/ELT workflows, and data warehouses to support large-scale retail media analytics.
  • Handle real-time and batch processing at scale
  • Work closely with data scientists, analysts, software engineers, and product teams to ensure seamless data integration and access.
  • Implement robust monitoring, validation, and security controls to maintain high data reliability.
Apply

Related Jobs

Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ’Έ 105825.0 - 136950.0 CAD per year

πŸ” Data Engineering

🏒 Company: SamsaraπŸ‘₯ 1001-5000πŸ’° Secondary Market over 4 years agoπŸ«‚ Last layoff almost 5 years agoCloud Data ServicesBusiness IntelligenceInternet of ThingsSaaSSoftware

  • BS degree in Computer Science, Statistics, Engineering, or a related quantitative discipline
  • 6+ years experience in a data engineering and data science-focused role
  • ​​Proficiency in data manipulation and processing in SQL and Python
  • Expertise building data pipelines with new API endpoints from their documentation
  • Proficiency in building ETL pipelines to handle large volumes of data
  • Demonstrated experience in designing data models at scale
  • Build and maintain highly reliable computed tables, incorporating data from various sources, including unstructured and highly sensitive data
  • Access, manipulate, and integrate external datasets with internal data
  • Building analytical and statistical models to identify patterns, anomalies, and root causes
  • Leverage SQL and Python to shape and aggregate data
  • Incorporate generative AI tools (ChatGPT Enterprise) into production data pipelines and automated workflows
  • Collaborate closely with data scientists, data analysts, and Tableau developers to ship top quality analytic products
  • Champion, role model, and embed Samsara’s cultural principles (Focus on Customer Success, Build for the Long Term, Adopt a Growth Mindset, Be Inclusive, Win as a Team) as we scale globally and across new offices

PythonSQLETLTableauAPI testingData engineeringData scienceSparkCommunication SkillsAnalytical SkillsData visualizationData modeling

Posted 4 days ago
Apply
Apply

πŸ“ United States, Canada

πŸ” Software Development

🏒 Company: OverstoryπŸ‘₯ 1-10E-Commerce

  • Approximately 5 years of experience in Data Engineering with at least one experience in a startup environment
  • Product-minded and able to demonstrate significant impact you have had on a business through the application of technology
  • Proven experience of data engineering across the following (or similar) technologies: Python, data orchestration platforms (Airflow, Luigi, Dagster, etc…), data quality frameworks, data lakes/warehouses
  • Ability to design and implement scalable and resilient data systems
  • Excellent communication skills and ability to collaborate effectively in a cross-functional team environment
  • Passion for learning and staying updated with evolving technologies and industry trends
  • Owning day-to-day operational responsibilities of deliveries our analysis to the customers
  • Developing data-driven solutions to customer problems that our products aren’t solving for yet
  • Building new and improving existing technologies such as:
  • Automation of the analysis for all customers, leading to faster implementation of Overstory’s recommendations
  • Metrics to identify what are the time bottlenecks in the current flow of analysis, therefore helping all Overstory teams identify areas of improvements
  • Visualization of status and progress of the analysis for internal use
  • Working on performance & scalability of our pipelines ensuring that our tech can handle our growth

PythonSQLCloud ComputingGCPAmazon Web ServicesData engineeringCommunication SkillsAnalytical SkillsRESTful APIsData visualizationData modeling

Posted 8 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ’Έ 110000.0 - 130000.0 USD per year

πŸ” Software Development

🏒 Company: CerosπŸ‘₯ 101-250πŸ’° $100,000,000 Private over 4 years agoAdvertisingContent CreatorsContent MarketingGraphic DesignSoftware

  • 5+ years of experience in data engineering, focusing on AWS Redshift and ETL pipeline development.
  • Strong expertise in SQL performance tuning, schema management, and query optimization.
  • Experience designing and maintaining ETL pipelines using AWS Glue, Matillion, or similar tools.
  • Proficiency in JavaScript/TypeScript, with experience building custom ETL workflows and integrations.
  • Hands-on experience with Python for data automation and scripting.
  • Strong understanding of data warehousing best practices, ensuring high-quality, scalable data models.
  • Experience with data monitoring and alerting tools such as AWS CloudWatch and New Relic.
  • Ability to work independently in a fast-paced environment, collaborating across teams to support data-driven initiatives.
  • Own and lead the management of AWS Redshift, ensuring optimal performance, disk usage, and cost efficiency.
  • Design and maintain scalable ETL pipelines using AWS Glue, Lambda, and Matillion to integrate data from Mixpanel, CRM platforms, and customer engagement tools.
  • Optimize SQL-based data transformations and Redshift queries to improve performance and reliability.
  • Automate data offloading and partition management, leveraging AWS services like S3 and external schemas.
  • Ensure version control and documentation of all Redshift queries, ETL processes, and AWS configurations through a centralized GitHub repository.
  • Develop monitoring and alerting for data pipelines using CloudWatch and other observability tools to ensure high availability and early issue detection.
  • Implement and maintain data quality checks and governance processes to ensure accuracy and consistency across foundational tables.
  • Collaborate with AI engineers and business stakeholders to enhance data accessibility and reporting for internal teams.
  • Maintain and optimize BI dashboards in Metabase and HubSpot, ensuring accuracy and efficiency of business reporting.
  • Manage key integrations between Redshift and external platforms, including Mixpanel, HubSpot, and Census, optimizing data accessibility and performance.
  • Administer AWS infrastructure supporting Redshift, ensuring efficient resource utilization, IAM security, and cost management.
  • Automate repetitive data tasks using Python and scripting to enhance data processes and improve team efficiency.

AWSPythonSQLETLGitJavascriptTypeScriptAmazon Web ServicesAPI testingData engineeringREST APICI/CDAnsibleData modelingData analyticsData management

Posted 10 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ’Έ 113000.0 - 130000.0 USD per year

πŸ” Software Development

🏒 Company: LaterπŸ‘₯ 1-10Consumer ElectronicsiOSAppsSoftware

  • Minimum of 5 years in data engineering or related fields
  • Bachelor’s degree in Computer Science, Engineering, or related field
  • Strong focus on building data infrastructure and pipelines
  • Design and build a robust data warehouse architecture
  • Design, build, and maintain scalable data pipelines
  • Develop reliable transformation layers and data pipelines
  • Establish optimized data architectures using cloud technologies
  • Enforce data quality checks and governance practices
  • Collaborate with cross-functional teams to deliver insights
  • Analyze and optimize data pipelines for performance

SQLData engineering

Posted 12 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ” B2B SaaS

🏒 Company: SanityπŸ‘₯ 51-200πŸ’° Corporate over 2 years agoSoftware Development

  • 4+ years of experience building data pipelines at scale
  • Deep expertise in SQL, Python, and Node.js/TypeScript
  • Production experience with Airflow and RudderStack
  • Track record of building reliable data infrastructure
  • Design, develop, and maintain scalable ETL/ELT pipelines
  • Collaborate to implement and scale product telemetry
  • Establish best practices for data ingestion and transformation
  • Monitor and optimize data pipeline performance

Node.jsPythonSQLApache AirflowETLTypeScript

Posted 18 days ago
Apply
Apply

πŸ“ Europe, APAC, Americas

🧭 Full-Time

πŸ” Software Development

🏒 Company: DockerπŸ‘₯ 251-500πŸ’° $105,000,000 Series C almost 3 years agoDeveloper ToolsDeveloper PlatformInformation TechnologySoftware

  • 4+ years of relevant industry experience
  • Experience with data modeling and building scalable pipelines
  • Proficiency with Snowflake or BigQuery
  • Experience with data governance and security controls
  • Experience creating ETL scripts using Python and SQL
  • Familiarity with a cloud ecosystem: AWS/Azure/Google Cloud
  • Experience with Tableau or Looker
  • Manage and develop ETL jobs, warehouse, and event collection tools
  • Build and manage the Central Data Model for reporting
  • Integrate emerging methodologies and technologies
  • Build data pipelines for ML and AI projects
  • Contribute to SOC2 compliance across the data platform
  • Document technical architecture

PythonSQLETLSnowflakeAirflowData engineeringData visualizationData modeling

Posted 19 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ” Data Engineering

🏒 Company: BioRenderπŸ‘₯ 101-250πŸ’° $15,319,133 Series A almost 2 years agoLife ScienceGraphic DesignSoftware

  • 7+ years of data engineering experience of relevant industry experience
  • Expertise working with Data Warehousing platforms (AWS RedShift or Snowflake preferred) and data lake / lakehouse architectures
  • Experience with Data Streaming platforms (AWS Kinesis / Firehose preferred)
  • Expertise with SQL and programming languages commonly used in data platforms (Python, Spark, etc)
  • Experience with data pipeline orchestration (e.g., Airflow) and data pipeline integrations (e.g. Airbyte, Stitch)
NOT STATED

AWSPythonSQLApache AirflowSnowflakeData engineeringSparkData modeling

Posted 22 days ago
Apply
Apply

πŸ“ Worldwide

🧭 Full-Time

πŸ” E-commerce

🏒 Company: TopsortπŸ‘₯ 51-100πŸ’° Corporate 8 months agoAdvertising PlatformsMarketplaceE-Commerce PlatformsRetail TechnologyAd Exchange

  • 5+ years of experience in data engineering or a related field.
  • Strong proficiency in SQL and database technologies with (e.g., PostgreSQL, MySQL, Snowflake, BigQuery).
  • Experience with data pipeline orchestration tools (e.g., Apache Airflow, Prefect, Dagster).
  • Proficiency in programming languages such as Python and Scala.
  • Hands-on experience with AWS cloud data services.
  • Familiarity with big data processing frameworks like Apache Spark.
  • Knowledge of data modeling, warehousing concepts, and distributed computing.
  • Experience implementing CI/CD for data pipelines.
  • Real-time data processing and streaming architectures (RisingWave, Kafka, Flink).
  • Database performance tuning and query optimization.
  • Strong problem-solving skills and the ability to work independently and collaboratively.
  • ETL/ELT pipeline development and automation.
  • Cloud computing and infrastructure management on AWS (nice to have).
  • Design, develop, and maintain robust ETL/ELT pipelines to process and transform large datasets efficiently.
  • Optimize data architecture and storage solutions to support analytics, machine learning, and business intelligence.
  • Work with cloud platforms (AWS) to implement scalable data solutions.
  • Ensure data quality, integrity, and security across all data pipelines.
  • Collaborate with data scientists, analysts, and software engineers to support data-driven initiatives.
  • Monitor and troubleshoot data workflows to ensure system performance and reliability.
  • Create APIs to provide analytical information to our clients.

AWSPostgreSQLPythonSQLApache AirflowETLKafkaMySQLSnowflakeCI/CDScala

Posted 23 days ago
Apply
Apply
πŸ”₯ Senior Data Engineer
Posted about 1 month ago

πŸ“ United States, Canada

🧭 Regular

πŸ’Έ 125000.0 - 160000.0 USD per year

πŸ” Digital driver assistance services

🏒 Company: AgeroπŸ‘₯ 1001-5000πŸ’° $4,750,000 over 2 years agoAutomotiveInsurTechInformation TechnologyInsurance

  • Bachelor's degree in a technical field and 5+ years or Master's degree with 3+ years of industry experience.
  • Extensive experience with Snowflake or other cloud-based data warehousing solutions.
  • Expertise in ETL/ELT pipelines using tools like Airflow, DBT, Fivetran.
  • Proficiency in Python for data processing and advanced SQL for managing databases.
  • Solid understanding of data modeling techniques and cost management strategies.
  • Experience with data quality frameworks and deploying data solutions in the cloud.
  • Familiarity with version control systems and implementing CI/CD pipelines.
  • Develop and maintain ETL/ELT pipelines to ingest data from diverse sources.
  • Monitor and optimize cloud costs while performing query optimization in Snowflake.
  • Establish modern data architectures including data lakes and warehouses.
  • Apply dimensional modeling techniques and develop transformations using DBT or Spark.
  • Write reusable and efficient code, and develop data-intensive UIs and dashboards.
  • Implement data quality frameworks and observability solutions.
  • Collaborate cross-functionally and document data flows, processes, and architecture.

AWSPythonSQLApache AirflowDynamoDBETLFlaskMongoDBSnowflakeFastAPIPandasCI/CDData modeling

Posted about 1 month ago
Apply
Apply

πŸ“ Canada

πŸ” Home service technology

🏒 Company: JobberπŸ‘₯ 501-1000πŸ’° $100,000,000 Series D about 2 years agoSaaSMobileSmall and Medium BusinessesTask Management

  • Excellent ETL pipeline development skills and hands-on experience with Orchestration (Airflow).
  • Experience with CI/CD practices and optimizing data flow within high-volume infrastructures.
  • Experience with data ingestion systems (e.g., Fivetran, Airbyte) and reverse ETL systems (e.g., HighTouch, Census).
  • Expertise in dimensional modeling, star schemas, and warehousing concepts.
  • Experience with message queues (e.g., Kafka) and real-time stream processing.
  • Proficiency in designing and maintaining efficient data pipelines within cloud infrastructure (preferably AWS).
  • Strong SQL skills and knowledge of containerization (ECS orchestration) and AWS Lambda.
  • Ability to explain complex data concepts to technical and non-technical stakeholders.
  • Empower the Team: Develop tools, frameworks, and workflows to enhance data accessibility and enable data-driven decision-making.
  • Build robust alerting and monitoring systems for data quality and reliability.
  • Collaborate with various functions to support analyses that influence business decisions.
  • Work closely with software engineering teams for collaborative data development.
  • Accelerate Business Growth: Synchronize data between Jobber and external systems. Streamline ETL workflows using tools like Airflow and dbt.
  • Build data models for easier integration.
  • Strategize and innovate: Research emerging technologies to strengthen the data stack.
  • Participate in design and code reviews providing mentorship and knowledge sharing.
  • Ensure Data Integrity: Establish best practices for maintaining data quality.

AWSGraphQLSQLApache AirflowETLKafkaData engineeringNosqlCI/CDRESTful APIsData modeling

Posted about 2 months ago
Apply