Apply

Data Engineer

Posted 3 days agoViewed

View full description

💎 Seniority level: Senior, 5+ years

📍 Location: Canada

🔍 Industry: FinTech

🏢 Company: KOHO

🗣️ Languages: English

⏳ Experience: 5+ years

🪄 Skills: AWSPythonSQLETLData engineeringData visualizationData modelingFinanceData analytics

Requirements:
  • 5+ years of mastery in data manipulation and analytics architecture
  • Advanced expertise in dbt (incremental modeling, materializations, snapshots, variables, macros, jinja)
  • Strong knowledge of SQL and how to write efficient SQL queries
  • Strong command of SQL, query optimization, and data warehouse design
Responsibilities:
  • Building strong relationships with stakeholders (the finance team), scope and prioritize their analytics requests.
  • Understanding business needs and translating them to requirements.
  • Using dbt (Core for development and Cloud for orchestration) to transform, test, deploy, and document financial data while applying software engineering best practices.
  • Troubleshooting variances in reports, and striving to eliminate them at the source.
  • Building game-changing data products that empower the finance team
  • Architecting solutions that transform complex financial data into actionable insights
  • Monitoring, optimizing and troubleshooting warehouse performance (AWS Redshift).
  • Creating scalable, self-service analytics solutions that democratize data access
  • Occasionally building dashboards and reports in Sigma and Drivetrain.
  • Defining processes, building tools, and offering training to empower all data users in the organization.
Apply

Related Jobs

Apply
🔥 Senior Data Engineer
Posted 3 days ago

📍 Worldwide

🔍 Hospitality

🏢 Company: Lighthouse

  • 4+ years of professional experience using Python, Java, or Scala for data processing (Python preferred)
  • You stay up-to-date with industry trends, emerging technologies, and best practices in data engineering.
  • Improve, manage, and teach standards for code maintainability and performance in code submitted and reviewed
  • Ship large features independently, generate architecture recommendations and have the ability to implement them
  • Great communication: Regularly achieve consensus amongst teams
  • Familiarity with GCP, Kubernetes (GKE preferred),  CI/CD tools (Gitlab CI preferred), familiarity with the concept of Lambda Architecture.
  • Experience with Apache Beam or Apache Spark for distributed data processing or event sourcing technologies like Apache Kafka.
  • Familiarity with monitoring tools like Grafana & Prometheus.
  • Design and develop scalable, reliable data pipelines using the Google Cloud stack.
  • Optimise data pipelines for performance and scalability.
  • Implement and maintain data governance frameworks, ensuring data accuracy, consistency, and compliance.
  • Monitor and troubleshoot data pipeline issues, implementing proactive measures for reliability and performance.
  • Collaborate with the DevOps team to automate deployments and improve developer experience on the data front.
  • Work with data science and analytics teams to enable them to bring their research to production grade data solutions, using technologies like airflow, dbt or MLflow (but not limited to)
  • As a part of a platform team, you will communicate effectively with teams across the entire engineering organisation, to provide them with reliable foundational data models and data tools.
  • Mentor and provide technical guidance to other engineers working with data.

PythonSQLApache AirflowETLGCPKubernetesApache KafkaData engineeringCI/CDMentoringTerraformScalaData modeling

Posted 3 days ago
Apply
Apply

📍 Worldwide

🧭 Full-Time

NOT STATED
  • Own the design and implementation of cross-domain data models that support key business metrics and use cases.
  • Partner with analysts and data engineers to translate business logic into performant, well-documented dbt models.
  • Champion best practices in testing, documentation, CI/CD, and version control, and guide others in applying them.
  • Act as a technical mentor to other analytics engineers, supporting their development and reviewing their code.
  • Collaborate with central data platform and embedded teams to improve data quality, metric consistency, and lineage tracking.
  • Drive alignment on model architecture across domains—ensuring models are reusable, auditable, and trusted.
  • Identify and lead initiatives to reduce technical debt and modernise legacy reporting pipelines.
  • Contribute to the long-term vision of analytics engineering at Pleo and help shape our roadmap for scalability and impact.

SQLData AnalysisETLData engineeringCI/CDMentoringDocumentationData visualizationData modelingData analyticsData management

Posted 4 days ago
Apply
Apply
🔥 Data Engineer
Posted 6 days ago

📍 Worldwide

🧭 Full-Time

💸 145000.0 - 160000.0 USD per year

  • Proficiency in managing MongoDB databases, including performance tuning and maintenance.
  • Experience with cloud-based data warehousing, particularly using BigQuery.
  • Familiarity with DBT for data transformation and modeling.
  • Exposure to tools like Segment for data collection and integration.
  • Basic knowledge of integrating third-party data sources to build a comprehensive data ecosystem.
  • Overseeing our production MongoDB database to ensure optimal performance, reliability, and security.
  • Assisting in the management and optimization of data pipelines into BigQuery, ensuring data is organized and accessible for downstream users.
  • Utilizing DBT to transform raw data into structured formats, making it useful for analysis and reporting.
  • Collaborating on the integration of data from Segment and various third-party sources to create a unified, clean data ecosystem.
  • Working closely with BI, Marketing, and Data Science teams to understand data requirements and ensure our infrastructure meets their needs.
  • Participating in code reviews, learning new tools, and contributing to the refinement of data processes and best practices.

SQLETLMongoDBData engineeringData modeling

Posted 6 days ago
Apply
Apply

📍 United States, Canada

🧭 Full-Time

🔍 Software Development

  • Strong hands-on experience with Python and core Python Data Processing tools such as pandas, numpy, scipy, scikit
  • Experience with cloud tools and environments like Docker, Kubernetes, GCP, and/or Azure
  • Experience with Spark/PySpark
  • Experience with Data Lineage and Data Cataloging
  • Relational and non-relational database experience
  • Experience with Data Warehouses and Lakes, such as Bigquery, Databricks, or Snowflake
  • Experience in designing and building data pipelines that scale
  • Strong communication skills, with the ability to convey technical solutions to both technical and non-technical stakeholders
  • Experience working effectively in a fast-paced, agile environment as part of a collaborative team
  • Ability to work independently and as part of a team
  • Willingness and enthusiasm to learn new technologies and tackle challenging problems
  • Experience in Infrastructure as Code tools like Terraform
  • Advanced SQL expertise, including experience with complex queries, query optimization, and working with various database systems
  • Work with business stakeholders to understand their goals, challenges, and decisions
  • Assist with building solutions that standardize their data approach to common problems across the company
  • Incorporate observability and testing best practices into projects
  • Assist in the development of processes to ensure their data is trusted and well-documented
  • Effectively work with data analysts on refining the data model used for reporting and analytical purposes
  • Improve the availability and consistency of data points crucial for analysis
  • Standing up a reporting system in BigQuery from scratch, including data replication, infrastructure setup, dbt model creation, and Integration with reporting endpoints
  • Revamping orchestration and execution to reduce critical data delivery times
  • Database archiving to move data from a live database to cold storage

AWSSQLCloud ComputingData AnalysisETLData engineeringData visualizationData modeling

Posted 11 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 13 days ago

📍 Canada

🧭 Full-Time

🔍 Fintech

🏢 Company: Coinme👥 51-100💰 $772,801 Seed over 2 years agoCryptocurrencyBlockchainBitcoinFinTechVirtual Currency

  • 7+ years of experience with ETL, SQL, PowerBI, Tableau, or similar technologies
  • Strong understanding of data modeling, database design, and SQL
  • Experience working with Apache Kafka or MSK solution
  • Extensive experience delivering solutions on Snowflake or other cloud-based data warehouses
  • Proficiency in Python/R and familiarity with modern data engineering practices
  • Strong analytical and problem-solving skills
  • Experience with machine learning (ML)
  • Design, develop, and maintain scalable data pipelines.
  • Implement data ingestion frameworks.
  • Optimize data pipelines for performance.
  • Develop and deliver data assets.
  • Evaluate and improve existing data solutions.
  • Experience in data quality management.
  • Collaborate with engineers and product managers.
  • Lead the deployment and maintenance of data solutions.
  • Champion best practices in data development.
  • Conduct code reviews and provide mentorship.
  • Create and maintain process documentation.
  • Monitor data pipelines for performance.
  • Implement logging, monitoring, and alerting systems.
  • Drive the team’s Agile process.

PythonSQLAgileETLMachine LearningSnowflakeTableauApache KafkaData engineeringData visualizationData modeling

Posted 13 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 18 days ago

📍 India, Canada, United Kingdom

🧭 Full-Time

🔍 Software Development

🏢 Company: Loopio Inc.

  • 5+ years of experience in data engineering in a high-growth agile software development environment
  • Strong understanding of database concepts, modeling, SQL, query optimization
  • Ability to learn fast and translate data into actionable results
  • Experience developing in Python and Pyspark
  • Hands-on experience with the AWS services (RDS, S3, Redshift, Glue, Quicksight, Athena, ECS)
  • Strong understanding of relational databases (RDS, MySQL) and NoSQL
  • Experience with ETL & Data warehousing, building fact & dimensional data models
  • Experience with data processing frameworks such as Spark / Databricks
  • Experience in developing Big Data solutions (migration, storage, processing)
  • Experience with CI/CD tools (Jenkins) and pipeline orchestration tools (Databricks Jobs, Airflow)
  • Experience working with data visualization and BI platforms (Quicksight, Tableau, Sisense, etc)
  • Experience working with Clickstream data (Amplitude, Pendo, etc)
  • Experience building and supporting large-scale systems in a production environment
  • Strong communication, collaboration, and analytical skills
  • Demonstrated ability to work with a high degree of ambiguity, and leadership within a team (mentorship, ownership, innovation)
  • Ability to clearly communicate technical roadmap, challenges, and mitigation
  • Be responsible for building, evolving and scaling data platforms and ETL pipelines, with an eye towards the growth of our business and the reliability of our data
  • Promote data-driven decision-making across the organization through data expertise
  • Build advanced automation tooling tooling for data orchestration, evaluation, testing, monitoring, administration, and data operations.
  • Integrate various data sources into our Data lake, including clickstream, relational, and unstructured data
  • Developing and maintaining a feature store for use in analytics & modeling
  • Partner with data scientists to create predictive models to help drive insights and decisions, both in Loopio’s product and internal teams (RevOps, Marketing, CX)
  • Work closely with stakeholders within and across teams to understand the data needs of the business and produce processes that enable a better product and support data-driven decision-making
  • Build scalable data pipelines using Databricks, and AWS (Redshift, S3, RDS), and other cloud technologies
  • Build and support Loopio’s data warehouse (Redshift) and data lake (Databricks delta lake)
  • Orchestrate pipelines using workflow frameworks/tooling

AWSPythonSQLData AnalysisETLJenkinsMachine LearningAirflowData engineeringNosqlSparkCommunication SkillsAnalytical SkillsCollaborationCI/CDData visualizationData modeling

Posted 18 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 25 days ago

📍 Canada

🧭 Full-Time

🔍 Data Engineering

🏢 Company: Vantage👥 1001-5000CryptocurrencyFinancial ServicesFinTechTrading Platform

  • 5+ years of experience in data engineering, big data, or distributed systems.
  • Strong expertise in Python, SQL (or equivalent big data processing frameworks).
  • Proficiency in ETL/ELT pipelines using Apache Airflow, or similar orchestration tools.
  • Experience working with real-time streaming data (Kafka, Kinesis, or Pub/Sub).
  • Strong understanding of data modelling, data warehousing, and distributed systems.
  • Familiarity with privacy-compliant data processing (GDPR, CCPA) for advertising/retail media use cases.
  • Design, develop, and optimize data pipelines, ETL/ELT workflows, and data warehouses to support large-scale retail media analytics.
  • Handle real-time and batch processing at scale
  • Work closely with data scientists, analysts, software engineers, and product teams to ensure seamless data integration and access.
  • Implement robust monitoring, validation, and security controls to maintain high data reliability.

PythonSQLApache AirflowETLKafkaData engineeringData modeling

Posted 25 days ago
Apply
Apply

📍 United States, Canada

🧭 Full-Time

💸 105825.0 - 136950.0 CAD per year

🔍 Data Engineering

🏢 Company: Samsara👥 1001-5000💰 Secondary Market over 4 years ago🫂 Last layoff almost 5 years agoCloud Data ServicesBusiness IntelligenceInternet of ThingsSaaSSoftware

  • BS degree in Computer Science, Statistics, Engineering, or a related quantitative discipline
  • 6+ years experience in a data engineering and data science-focused role
  • ​​Proficiency in data manipulation and processing in SQL and Python
  • Expertise building data pipelines with new API endpoints from their documentation
  • Proficiency in building ETL pipelines to handle large volumes of data
  • Demonstrated experience in designing data models at scale
  • Build and maintain highly reliable computed tables, incorporating data from various sources, including unstructured and highly sensitive data
  • Access, manipulate, and integrate external datasets with internal data
  • Building analytical and statistical models to identify patterns, anomalies, and root causes
  • Leverage SQL and Python to shape and aggregate data
  • Incorporate generative AI tools (ChatGPT Enterprise) into production data pipelines and automated workflows
  • Collaborate closely with data scientists, data analysts, and Tableau developers to ship top quality analytic products
  • Champion, role model, and embed Samsara’s cultural principles (Focus on Customer Success, Build for the Long Term, Adopt a Growth Mindset, Be Inclusive, Win as a Team) as we scale globally and across new offices

PythonSQLETLTableauAPI testingData engineeringData scienceSparkCommunication SkillsAnalytical SkillsData visualizationData modeling

Posted 27 days ago
Apply
Apply

📍 United States, Canada

🔍 Software Development

🏢 Company: Overstory👥 1-10E-Commerce

  • Approximately 5 years of experience in Data Engineering with at least one experience in a startup environment
  • Product-minded and able to demonstrate significant impact you have had on a business through the application of technology
  • Proven experience of data engineering across the following (or similar) technologies: Python, data orchestration platforms (Airflow, Luigi, Dagster, etc…), data quality frameworks, data lakes/warehouses
  • Ability to design and implement scalable and resilient data systems
  • Excellent communication skills and ability to collaborate effectively in a cross-functional team environment
  • Passion for learning and staying updated with evolving technologies and industry trends
  • Owning day-to-day operational responsibilities of deliveries our analysis to the customers
  • Developing data-driven solutions to customer problems that our products aren’t solving for yet
  • Building new and improving existing technologies such as:
  • Automation of the analysis for all customers, leading to faster implementation of Overstory’s recommendations
  • Metrics to identify what are the time bottlenecks in the current flow of analysis, therefore helping all Overstory teams identify areas of improvements
  • Visualization of status and progress of the analysis for internal use
  • Working on performance & scalability of our pipelines ensuring that our tech can handle our growth

PythonSQLCloud ComputingGCPAmazon Web ServicesData engineeringCommunication SkillsAnalytical SkillsRESTful APIsData visualizationData modeling

Posted about 1 month ago
Apply
Apply

📍 USA, CANADA

🧭 Full-Time

🔍 Software Development

🏢 Company: Wrapbook

  • Hands-on experience deploying production-quality code in fast-paced environments
  • Proficiency in Python (preferred), Java, or Scala for data processing and pipeline development
  • Ability to thrive in fast-changing, ambiguous situations, balancing immediate needs with long-term goals
  • Experience with data pipeline tools, such as Airbyte for ingestion and dbt for transformation/modeling
  • Hands-on expertise with container orchestration tools, such as Kubernetes, and cloud-native environments (e.g., AWS)
  • Proficiency with workflow automation and orchestration tools, like Dagster or Apache Airflow
  • Deep familiarity with PostgreSQL, including administration, tuning, and provisioning in cloud platforms (e.g., AWS)
  • Strong experience in ETL/ELT pipelines and data modeling, including raw vs. curated datasets, star schemas, and incremental loads
  • Advanced SQL skills, with expertise in relational databases and data warehouses (especially Snowflake)
  • Knowledge of best practices in data governance and security
  • Excellent problem-solving skills and ability to troubleshoot complex issues
  • Strong communication skills to collaborate with cross-functional teams
  • Own and optimize data pipeline infrastructure to ensure reliable, efficient, and scalable data flows from diverse sources.
  • Contribute to the development of the data engineering roadmap in collaboration with Platform leadership and cross-functional stakeholders.
  • Design, build, and maintain scalable ETL/ELT pipelines to transform raw data into curated datasets within AWS S3 and Snowflake.
  • Implement and standardize data governance practices, ensuring data quality, lineage tracking, schema consistency, and compliance across pipelines.
  • Collaborate with analytics and engineering teams to manage backfills, resolve schema drift, and implement best practices for incremental loads.
  • Lead the design and implementation of a layered data architecture to improve scalability, governance, and self-service analytics.
  • Develop and implement data contracts by collaborating across teams to align business goals with technical needs.
  • Evaluate, plan, and execute new data tools, infrastructure, and system expansions to support company growth and evolving analytics needs.
  • Deliver scalable, efficient, and maintainable code by applying architectural best practices and adhering to data engineering standards.
  • Maintain SLAs for data freshness, accuracy, and availability by defining clear metrics that foster stakeholder trust and ensure consistent, reliable data delivery.
  • Collaborate with the Data Analytics team to facilitate the delivery of strategic initiatives.

AWSDockerPostgreSQLPythonSQLApache AirflowBashCloud ComputingETLGitKubernetesSnowflakeAlgorithmsData engineeringData StructuresREST APICommunication SkillsAnalytical SkillsCI/CDProblem SolvingRESTful APIsData visualizationAnsibleData modelingData management

Posted about 1 month ago
Apply