Apply

Data Engineer

Posted about 1 month agoViewed

View full description

πŸ’Ž Seniority level: Senior, 6+ years

πŸ“ Location: 30+ countries

πŸ” Industry: Multilingual content industry

🏒 Company: SmartcatπŸ‘₯ 101-250πŸ’° $43,000,000 Series C 4 months agoInformation TechnologySoftware

⏳ Experience: 6+ years

πŸͺ„ Skills: PythonETLAirflowData engineeringData modeling

Requirements:
  • 6+ years of experience in data engineering, analytics engineering, or a related field.
  • Strong background in data modeling, warehousing, managing data quality, orchestration and integration, and efficient data pipeline practices.
  • Demonstrated ability to manage projects and drive them to successful completion.
  • Experience in small to medium-sized SaaS organizations with tech-driven products.
  • High proficiency in Databricks, data modeling and architecture (dbt), Airflow, and Python for data engineering.
  • Familiarity with business intelligence tools and data sources.
Responsibilities:
  • Partner with the head of data to architect and enhance the transition to a cloud-based streaming architecture.
  • Lead the introduction and management of data governance and data operationalization across Smartcat.
  • Foster a culture of technical excellence in data engineering and analytics engineering.
  • Collaborate with product, engineering, and business teams to align initiatives with organizational goals.
  • Stay informed about data engineering trends to recommend productivity-enhancing tools and technologies.
Apply

Related Jobs

Apply
πŸ”₯ Data Engineer
Posted about 4 hours ago

πŸ“ India

🧭 Full-Time

πŸ” Experience Management

🏒 Company: Experience.comπŸ‘₯ 101-250πŸ’° $14,575,000 Series A about 6 years agoCustomer ServiceConsumerInformation ServicesConsultingSaaSAnalyticsQuality AssuranceInformation TechnologySoftware

  • 4+ years of experience with PySpark and SQL for building scalable ETL pipelines.
  • Strong proficiency in Python programming.
  • Knowledge of GCP Data Analytics ecosystem (BigQuery, PySpark, SQL, etc.).
  • Experience with Airflow/Composer for workflow orchestration.
  • Experience with in-memory applications, database design, and data integration.
  • Strong analytical thinking and problem-solving abilities.

  • Design, build, and maintain scalable ETL/ELT pipelines using PySpark and SQL.
  • Work on data extraction, transformation, and loading processes from multiple sources into data warehouses such as BigQuery.
  • Leverage GCP data analytics tools (BigQuery, DataProc, Cloud Functions, etc.) to process and analyze data.
  • Optimize data workflows for benchmarking, performance, and tuning to ensure efficiency and reliability.
  • Collaborate with engineering and analytics teams to develop data integration solutions that meet business needs.
  • Ensure the accuracy and quality of data by implementing strong in-memory applications and database designs.
  • Implement monitoring and alerting for pipelines and workflows to ensure data consistency and issue resolution.

PythonSQLElasticSearchETLGCPMongoDBAirflow

Posted about 4 hours ago
Apply
Apply

πŸ“ Brazil

πŸ” Corporate wellness

🏒 Company: Wellhub

  • Bachelor’s degree in computer science or equivalent professional experience.
  • Experience using Python, Java, or similar programming languages (Python preferred).
  • Familiarity with messaging systems like Kafka, Pulsar, RabbitMQ.
  • Understanding of infrastructure as code tools like Terraform, Pulumi.
  • Experience with AWS or GCP cloud architectures.
  • Proficiency in SQL and NoSQL databases.
  • Understanding of distributed systems and data-intensive applications.
  • Genuine interest in big data technologies.
  • Strong communication skills for multidisciplinary collaboration.

  • Work closely with development and data analysts to understand data needs.
  • Collaborate on designing and implementing scalable data infrastructure.
  • Ensure systems' consistency, reliability, and scalability.
  • Support existing systems and enhance functionality.
  • Participate in incident resolution and system troubleshooting.
  • Keep the tech stack up-to-date with industry practices.
  • Maintain high code and infrastructure quality standards.

AWSPythonSQLCloud ComputingKafkaData engineeringNosqlTerraform

Posted about 5 hours ago
Apply
Apply

πŸ“ Hungary

πŸ” Healthcare

  • Bachelor's degree in mathematics, computer science, statistics, engineering, or information science; Master's is a plus.
  • 5+ years of relevant experience, ideally in a multicultural environment.
  • Experience in designing and developing data flows and ETL processes with modern storage solutions.
  • Proficient in applying advanced statistical methods and machine/deep learning algorithms.
  • Experience in at least one programming language: Python, SQL, JavaScript, dbt.
  • Strong team player with interpersonal skills and a collaborative mindset.
  • Excellent communication and networking skills.
  • Strong experience in stakeholder management.
  • Fluent in English; additional languages are a plus.

  • Deliver high quality outcomes by applying fit-for-purpose analytics solutions.
  • Onboard data into Snowflake and implement data loads and analysis, including SQL backend development.
  • Operationalize models in AWS and set up optimal architectures for data science projects.
  • Lead projects independently while partnering with business stakeholders.
  • Collaborate with colleagues globally on initiatives and projects.
  • Build relationships across departments to identify synergies.
  • Create and manage documentation for data science projects.
  • Participate in agile and cross-functional team collaborations.
  • Perform demos and project reviews, communicate ideas, and receive feedback.
  • Engage in exploratory projects to evaluate new technologies.

AWSPythonSQLData AnalysisETLMachine LearningSnowflakeData engineeringData visualizationStakeholder management

Posted about 5 hours ago
Apply
Apply

πŸ“ Canada

🧭 Full-Time

πŸ” Technology for small businesses

🏒 Company: JobberπŸ‘₯ 501-1000πŸ’° $100,000,000 Series D almost 2 years agoSaaSMobileSmall and Medium BusinessesTask Management

  • Proven ability to lead and collaborate in team environments.
  • Strong coding skills in Python and SQL.
  • Expertise in building and maintaining ETL pipelines using tools like Airflow and dbt.
  • Experience with AWS tools such as Redshift, Glue, and Lambda.
  • Familiarity with handling large datasets using tools like Spark.
  • Experience with Terraform for infrastructure management.
  • Knowledge of dimensional modelling, star schemas, and data warehousing.

  • Design, develop, and maintain batch and real-time data pipelines within cloud infrastructure (preferably AWS).
  • Develop tools that automate processes and set up monitoring systems.
  • Collaborate with teams to extract actionable insights from data.
  • Lead initiatives to propose new technologies, participate in design and code reviews, and maintain data integrity.

AWSPythonSQLApache AirflowETLSparkTerraform

Posted 1 day ago
Apply
Apply
πŸ”₯ Data Engineer
Posted 2 days ago

πŸ“ Philippines

🧭 Full-Time

πŸ” ECommerce and Marketing

🏒 Company: Podean

  • 3+ years of experience in a data engineering or similar role, focusing on API integration.
  • Proficiency in Python, Java, or another programming language suitable for API integration and data engineering.
  • Expertise in SQL and experience with data warehouses (e.g., Redshift, Snowflake, BigQuery).
  • Hands-on experience with workflow orchestration tools.
  • Proven track record of building scalable data pipelines and systems.
  • Strong problem-solving abilities and attention to detail.
  • Excellent communication skills and a collaborative mindset.
  • Ability to manage multiple projects in a fast-paced environment.

  • Develop and maintain integrations with marketplace APIs such as Amazon Selling Partner API.
  • Handle API authentication, rate limits, pagination, and error handling.
  • Design, build, and optimize ETL/ELT pipelines for ingesting and processing data from multiple marketplaces.
  • Automate data workflows to ensure reliable and timely updates.
  • Design and implement data models to support analytical and operational use cases.
  • Utilize data storage solutions such as AWS S3, Redshift, DynamoDB, or Google BigQuery.
  • Monitor and optimize API calls for efficient large-scale data operations.
  • Collaborate with data analysts and product teams to deliver actionable insights.
  • Communicate technical concepts to non-technical stakeholders.
  • Manage API keys, tokens, and access credentials securely.

PythonSQLDynamoDBETL

Posted 2 days ago
Apply
Apply

πŸ“ United States

πŸ” Defense and Financial Technology

🏒 Company: 540

  • Bachelor's Degree.
  • 8+ years of related experience.
  • Well-versed in Python.
  • Experience building and managing data pipelines.
  • Proficient in data analytics tools such as Databricks.
  • Experience building dashboards using PowerBI and/or similar tools.
  • Experience working via the terminal / command line.
  • Experience consuming data via APIs.
  • Hands-on experience using Jira and Confluence.

  • Working directly with government leadership managing teams, customers, and data requirements.
  • Assisting Audit teams with monthly data ingestions from Army systems.
  • Management of data initiatives and small projects from start to finish.
  • Working with Army FM&C Lead to prioritize Advana data product requirements.
  • Developing recurring and ad hoc financial datasets.
  • Developing Advana datasets and analytical products to enable the Army reporting on all Financial System data.
  • Reviewing data pipeline code via GitLab to ensure it meets team and code standards.
  • Overseeing overall architecture and technical direction for FM&C data projects.

AWSPython

Posted 3 days ago
Apply
Apply

πŸ“ US & Canada

πŸ” Fintech

🏒 Company: MesaπŸ‘₯ 11-50Product DesignManufacturingProfessional ServicesSoftware

  • 5+ years of software engineering and operationalizing data pipelines with large and complex datasets.
  • Experience with data modeling, ETL, and patterns for efficient data governance.
  • Experience manipulating large-scale structured and unstructured data.
  • Experience working with batch and stream processing.
  • Strong proficiency with Typescript is a must.
  • Strong SQL skills.
  • Experience using dashboarding tools like Mode, Tableau, Looker.
  • Passionate about event-driven architecture, microservices, data reliability, and observability.
  • Ability to thrive in a fast-paced startup environment and handle ambiguity.

  • Lead data engineering at Mesa by developing and operationalizing scalable and reliable data pipelines.
  • Assemble large, complex data sets that meet functional and non-functional requirements.
  • Work with product and cross functional business stakeholders to enable visualization layers for data-driven decision-making.
  • Drive technical delivery, including architectural design, development, and QA.
  • Participate in customer discovery efforts as beta users help refine the product.

PostgreSQLSQLETLTypeScriptData engineeringMicroservicesData modeling

Posted 3 days ago
Apply
Apply

πŸ“ Brazil

🧭 Full-Time

πŸ” Digital Engineering and Modernization

🏒 Company: EncoraπŸ‘₯ 10001-10001πŸ’° $200,000,000 Private over 5 years agoBig DataCloud ComputingSoftware

  • Experience in data modeling.
  • Experience developing and maintaining data pipelines.
  • Proficiency in SQL.
  • Proficiency in Python.
  • Experience with AWS Redshift.
  • Experience with Apache Airflow.
  • Familiarity with BI tools.

  • Develop and maintain efficient and scalable data pipelines.
  • Model and transform data to meet analysis and reporting needs.
  • Collaborate closely with the customer, including BI and software engineering.
  • Lead other BI or DE team members.
  • Create and maintain detailed technical documentation.
  • Develop dashboards in AWS Quicksight with support from a BI Analyst.

PythonSQLApache AirflowBusiness IntelligenceData modeling

Posted 4 days ago
Apply
Apply

πŸ“ Copenhagen, London, Stockholm, Berlin, Madrid, Montreal, Lisbon, 35 other countries

🧭 Full-Time

πŸ” Financial Technology

  • Strong background in building and managing data infrastructure at scale.
  • Expertise in Python, AWS, dbt, Airflow, and Kubernetes.
  • Ability to translate business and product requirements into technical data solutions.
  • Experience in mentoring and fostering collaboration within teams.
  • Curiosity and enthusiasm for experimenting with new technologies to solve complex problems.
  • Hands-on experience with modern data tools and contributing to strategic decision-making.

  • Partnering with product and business teams to develop data strategies that enable new features and improve user experience.
  • Driving key strategic projects across the organisation, dipping in and out as needed to provide leadership and hands-on support.
  • Supporting multiple teams across Pleo in delivering impactful data and analytics solutions.
  • Building data products that directly support Pleo's product roadmap and business goals.
  • Collaborating with the VP of Data and other data leaders to set the vision for Pleo’s data strategy and ensure alignment with company objectives.
  • Enhancing our data infrastructure and pipelines to improve scalability, performance, and data quality.
  • Experimenting with and implementing innovative technologies to keep Pleo’s data stack at the forefront of the industry.
  • Mentoring engineers, analysts, and data scientists to foster growth and build a world-class data team.

AWSPythonApache AirflowKubernetesData engineering

Posted 4 days ago
Apply
Apply

πŸ“ US, Europe

🧭 Full-Time

πŸ’Έ 175000.0 - 205000.0 USD per year

πŸ” Cloud computing and AI services

🏒 Company: CoreWeaveπŸ’° $642,000,000 Secondary Market about 1 year agoCloud ComputingMachine LearningInformation TechnologyCloud Infrastructure

  • 5+ years of experience with Kubernetes and Helm, with a deep understanding of container orchestration.
  • Hands-on experience administering and optimizing clustered computing technologies on Kubernetes, such as Spark, Trino, Flink, Ray, Kafka, StarRocks or similar.
  • 5+ years of programming experience in C++, C#, Java, or Python.
  • 3+ years of experience scripting in Python or Bash for automation and tooling.
  • Strong understanding of data storage technologies, distributed computing, and big data processing pipelines.
  • Proficiency in data security best practices and managing access in complex systems.

  • Architect, deploy, and scale data storage and processing infrastructure to support analytics and data science workloads.
  • Manage and maintain data lake and clustered computing services, ensuring reliability, security, and scalability.
  • Build and optimize frameworks and tools to simplify the usage of big data technologies.
  • Collaborate with cross-functional teams to align data infrastructure with business goals and requirements.
  • Ensure data governance and security best practices across all platforms.
  • Monitor, troubleshoot, and optimize system performance and resource utilization.

PythonBashKubernetesApache Kafka

Posted 6 days ago
Apply