Apply

Senior Data Engineer

Posted 1 day agoViewed

View full description

πŸ’Ž Seniority level: Senior, 4 + years

πŸ’Έ Salary: 700000000.0 - 900000000.0 COP per year

πŸ” Industry: Software Development

🏒 Company: Newrich Network

⏳ Experience: 4 + years

Requirements:
  • 4 + years experience in Data Engineering
  • 3+ years of experience working on Apache Spark applications using Python (PySpark) or Scala
  • Experience creating spark jobs that work on at least 1 billion records
  • Strong knowledge of ETL architecture and standards
  • Software development experience working with Apache Airflow, Spark, MongoDB, MySQL
  • Strong SQL knowledge
  • Strong command of Python
  • Experience creating data pipelines in a production system
  • Proven experience in building/operating/maintaining fault tolerant and scalable data processing integrations using AWS
  • Experience using Docker or Kubernetes is a plus
  • Ability to identify and resolve problems associated with production grade large scale data processing workflows
  • Experience with crafting and maintaining unit tests and continuous integration.
  • Passion for crafting Intelligent data pipelines that teams love to use
  • Strong capacity to handle numerous projects are a must
Responsibilities:
  • Collaborate with Data architects, Enterprise architects, Solution consultants and Product engineering teams to gather customer data integration requirements, conceptualize solutions & build required technology stack
  • Collaborate with enterprise customer's engineering team to identify data sources, profile and quantify quality of data sources, develop tools to prepare data and build data pipelines for integrating customer data sources and third party data sources.
  • Develop new features and improve existing data integrations with customer data ecosystem
  • Encourage the team to think out-of-the-box and overcome engineering obstacles while incorporating new innovative design principles.
  • Collaborate with a Project Manager to bill and forecast time for product owner solutions
  • Building data pipelines
  • Reconciling missed data
  • Acquire datasets that align with business needs
  • Develop algorithms to transform data into useful, actionable information
  • Build, test, and maintain database pipeline architectures
  • Collaborate with management to understand company objectives
  • Create new data validation methods and data analysis protocols
  • Ensure compliance with data governance and security policies
Apply

Related Jobs

Apply

πŸ“ Lithuania

πŸ’Έ 4000.0 - 6000.0 EUR per month

πŸ” Software Development

🏒 Company: Softeta

  • 4+ years of experience as a Data Engineer
  • Experience with Azure (Certifications are a Plus)
  • Experience with Databricks, Azure Data Lake, Data Factory and Apache Airflow
  • CI/CD or infrastructure as code
  • Knowledge of Medallion Architecture or Multihop architecture
  • Experience developing and administering ETL processes in the Cloud (Azure, AWS or GCP) environment
  • Strong programming skills in Python and SQL
  • Strong problem-solving and analytical skills
  • Design, develop, and maintain data pipelines and ETL processes
  • Data modeling, data cleansing
  • Automating data processing workflows using tools such as Airflow or other workflow management tools
  • Optimizing the performance of databases, including designing and implementing data structures and using indexes appropriately
  • Implement data quality and data governance processes
  • Being a data advocate and helping unlock business value by using data

PythonSQLApache AirflowETLAzureData engineeringCI/CDData modeling

Posted about 7 hours ago
Apply
Apply
πŸ”₯ Senior Data Engineer
Posted about 10 hours ago

πŸ“ United States

πŸ’Έ 144000.0 - 180000.0 USD per year

πŸ” Software Development

🏒 Company: HungryrootπŸ‘₯ 101-250πŸ’° $40,000,000 Series C almost 4 years agoArtificial Intelligence (AI)Food and BeverageE-CommerceRetailConsumer GoodsSoftware

  • 5+ years of experience in ETL development and data modeling
  • 5+ years of experience in both Scala and Python
  • 5+ years of experience in Spark
  • Excellent problem-solving skills and the ability to translate business problems into practical solutions
  • 2+ years of experience working with the Databricks Platform
  • Develop pipelines in Spark (Python + Scala) in the Databricks Platform
  • Build cross-functional working relationships with business partners in Food Analytics, Operations, Marketing, and Web/App Development teams to power pipeline development for the business
  • Ensure system reliability and performance
  • Deploy and maintain data pipelines in production
  • Set an example of code quality, data quality, and best practices
  • Work with Analysts and Data Engineers to enable high quality self-service analytics for all of Hungryroot
  • Investigate datasets to answer business questions, ensuring data quality and business assumptions are understood before deploying a pipeline

AWSPythonSQLApache AirflowData MiningETLSnowflakeAlgorithmsAmazon Web ServicesData engineeringData StructuresSparkCI/CDRESTful APIsMicroservicesJSONScalaData visualizationData modelingData analyticsData management

Posted about 10 hours ago
Apply
Apply

πŸ“ United States

πŸ’Έ 135000.0 - 155000.0 USD per year

πŸ” Software Development

🏒 Company: JobgetherπŸ‘₯ 11-50πŸ’° $1,493,585 Seed about 2 years agoInternet

  • 8+ years of experience as a data engineer, with a strong background in data lake systems and cloud technologies.
  • 4+ years of hands-on experience with AWS technologies, including S3, Redshift, EMR, Kafka, and Spark.
  • Proficient in Python or Node.js for developing data pipelines and creating ETLs.
  • Strong experience with data integration and frameworks like Informatica and Python/Scala.
  • Expertise in creating and managing AWS services (EC2, S3, Lambda, etc.) in a production environment.
  • Solid understanding of Agile methodologies and software development practices.
  • Strong analytical and communication skills, with the ability to influence both IT and business teams.
  • Design and develop scalable data pipelines that integrate enterprise systems and third-party data sources.
  • Build and maintain data infrastructure to ensure speed, accuracy, and uptime.
  • Collaborate with data science teams to build feature engineering pipelines and support machine learning initiatives.
  • Work with AWS cloud technologies like S3, Redshift, and Spark to create a world-class data mesh environment.
  • Ensure proper data governance and implement data quality checks and lineage at every stage of the pipeline.
  • Develop and maintain ETL processes using AWS Glue, Lambda, and other AWS services.
  • Integrate third-party data sources and APIs into the data ecosystem.

AWSNode.jsPythonSQLETLKafkaData engineeringSparkAgile methodologiesScalaData modelingData management

Posted about 12 hours ago
Apply
Apply

πŸ“ Thailand, Philippines

πŸ” Fintech

🏒 Company: EnvissoπŸ‘₯ 11-50CreditComplianceTransaction ProcessingFinancial Services

  • 5+ years of work experience in data engineering.
  • Strong skills in SQL and Python.
  • Experience designing, building and maintaining data models and data pipelines.
  • Experience working with cloud based architecture.
  • Great communication skills with a diverse team of varying technical ability.
  • Create and maintain scalable data pipelines to ingest, transform and serve global payments and risk data.
  • Manage and maintain the data platform, including data pipelines and environments.
  • Collaborate with cross-functional teams of data scientists, software engineers, product managers and business leads, to understand requirements and deliver appropriate solutions.
  • Take ownership of a data area, building subject matter expertise and cultivating trust with stakeholders.
  • Mentor junior members, and grow a strong data culture across the team and organisation.

PythonSQLCloud ComputingETLData engineeringCommunication SkillsData modeling

Posted 1 day ago
Apply
Apply

πŸ” Software Development

🏒 Company: Auctane Poland Careers

  • A strong Data Engineering background in a data warehouse or data lake architecture
  • Experience working in AWS/GCP cloud infrastructure.
  • Experience developing and supporting robust, automated, and reliable data pipelines in Python and SQL.
  • Mastered Python and SQL skills.
  • Experience with data processing frameworks like Spark, Athena, or Pandas.
  • Knowledge of Data Orchestration solutions, preferably Airflow.
  • Contribute to the design, build, and operational management of our data pipelines and analytics solution on top of proven AWS data technologies like S3, Athena, Lambda, and Kinesis.
  • Collaborate with the Product Owner and other stakeholders to implement the Data Products strategy.
  • Develop frameworks and solutions that enable us to acquire, process, monitor and extract value from our massive dataset.
  • Drive the design and architecture of Auctane’s data product.
  • Contribute directly to the implementation and operations of our systems.
  • Be an advocate of data quality and observability principles and use state-of-the-art technologies like Airflow, dbt, and Spark to process data and get our datasets just right.
  • Foster engineering excellence by delivering highly reliable software and data pipelines using Software Engineering best practices like automation, version control, continuous integration/continuous delivery, testing, security, etc.
Posted 2 days ago
Apply
Apply

πŸ“ States of SΓ£o Paulo and Rio Grande do Sul, Rio de Janeiro, Belo Horizonte

πŸ” Data Engineering

🏒 Company: TELUS Digital Brazil

  • At least 3 years of experience as Data Engineer
  • Have actively participated in the design and development of data architectures
  • Hands-on experience in developing and optimizing data pipelines
  • Experience working with databases and data modeling projects, as well as practical experience utilizing SQL
  • Effective English communication - able to explain technical and non-technical concepts to different audiences
  • Experience with a general-purpose programming language such as Python or Scala
  • Ability to work well in teams and interact effectively with others
  • Ability to work independently and manage multiple tasks simultaneously while meeting deadlines
  • Develop and optimize scalable, high-performing, secure, and reliable data pipelines that address diverse business needs and considerations
  • Identify opportunities to enhance internal processes, implement automation to streamline manual tasks, and contribute to infrastructure redesign
  • Act as a guide and mentor to junior engineers, supporting their professional growth and fostering an inclusive working environment
  • Collaborate with cross-functional teams to ensure data quality and support data-driven decision-making to strive for greater functionality in our data systems
  • Collaborate with project managers and product owners to assist in prioritizing, estimating, and planning development tasks
  • Provide constructive feedback, and share expertise with fellow team members, fostering mutual growth and learning
  • Engage in ongoing research and adoption of new technologies, libraries, frameworks, and best practices to enhance the capabilities of the data team
  • Demonstrate a commitment to accessibility and ensure that your work considers and positively impacts others

AWSDockerPythonSQLAgileApache AirflowCloud ComputingETLKubernetesData engineeringData scienceCommunication SkillsAnalytical SkillsTeamworkData modelingEnglish communication

Posted 4 days ago
Apply
Apply

πŸ“ Germany, Spain, United Kingdom, Austria

πŸ” Software Development

🏒 Company: LocalStackπŸ‘₯ 11-50πŸ’° $25,000,000 Series A 4 months agoCloud ComputingInformation TechnologySoftware

  • Ability and experience working with non technical stakeholders to gather requirements
  • Ability to define technical initiatives required to satisfy business requirements
  • Excellent knowledge of Python
  • Experience in designing real time data ingestion solutions with massive volumes of data
  • (preferred) Experience with AWS services commonly used in Data Engineering (like S3, ECS, Glue, EMR)
  • Experience with relational databases and data warehouses, data orchestration and ingestion tools, SQL, and BI tools
  • (preferred) Experience in working remotely/ in async settings
  • Experience owning initiatives at the IC level
  • Experience Providing guidance to junior engineers
  • Maintain, monitor, and optimize data ingestion pipelines for our current data platform.
  • Lead the development of our future data platform based on evolving business needs.
  • Shape the data team roadmap and contribute to long-term strategic planning.
  • Take full ownership of data ingestion from external sources, ensuring smooth functionality.
  • Design and implement a robust data modelling and data lake solution architecture.
  • Provide technical leadership and mentorship to the data engineering team.
  • Collaborate with engineering teams to define and refine ingestion pipeline requirements.
  • Work with stakeholders to gather business questions and data needs.

AWSDockerLeadershipPythonSQLApache AirflowETLKafkaData engineeringData StructuresREST APICommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingMentoringWritten communicationData visualizationTeam managementStakeholder managementData modeling

Posted 4 days ago
Apply
Apply

πŸ” Software Development

NOT STATED
  • Architect and build a modern, scalable data platform
  • Partnering directly with the Head of Data
Posted 4 days ago
Apply
Apply

πŸ” Software Development

🏒 Company: CookUnityπŸ‘₯ 501-1000πŸ’° $47,000,000 almost 2 years agoFood DeliveryFood and BeverageConsumer ApplicationsSubscription ServiceOrganic Food

  • 5+ years of experience as a Data Engineer or Backend Engineer with data expertise.
  • Strong proficiency in SQL, capable of writing optimized queries for large-scale analytical and operational databases.
  • Experience with modern data warehouses (Snowflake, BigQuery, Redshift) and data transformation tools (DBT, Airflow, Prefect).
  • Hands-on experience with Python for data pipeline development.
  • Familiarity with AWS data services (S3, Athena, Lambda, Glue, Kinesis) and cloud-based ETL workflows.
  • Experience working with BI tools (Looker, Tableau, Power BI) for operational reporting.
  • Strong understanding of data modeling, pipeline orchestration, and performance optimization.
  • Advanced English communication skills, with the ability to collaborate across teams.
  • Design, build, and maintain scalable data pipelines (batch and real-time) to support operational and analytical use cases.
  • Develop and optimize SQL queries and data models for Snowflake, AWS Athena, and PostgreSQL.
  • Take full ownership of Chef Payout calculations, ensuring accuracy and automation in financial reporting.
  • Improve data accessibility by developing interactive reports using Looker, Tableau, or Metabase.
  • Enhance data observability with monitoring, logging, and alerting to ensure pipeline reliability.
  • Collaborate with engineers, product managers, and analysts to scope data requirements and translate them into scalable solutions.
  • Maintain best practices for CI/CD, version control (GitHub), and testing in data engineering workflows.
Posted 4 days ago
Apply
Apply

πŸ“ Canada

🧭 Full-Time

πŸ” Retail Media

🏒 Company: VantageπŸ‘₯ 1001-5000CryptocurrencyFinancial ServicesFinTechTrading Platform

  • 5+ years of experience in data engineering, big data, or distributed systems.
  • Strong expertise in Python, SQL (or equivalent big data processing frameworks).
  • Proficiency in ETL/ELT pipelines using Apache Airflow, or similar orchestration tools.
  • Experience working with real-time streaming data (Kafka, Kinesis, or Pub/Sub).
  • Strong understanding of data modelling, data warehousing, and distributed systems.
  • Familiarity with privacy-compliant data processing (GDPR, CCPA) for advertising/retail media use cases.
  • Design, develop, and optimize data pipelines, ETL/ELT workflows, and data warehouses to support large-scale retail media analytics.
  • Handle real-time and batch processing at scale
  • Work closely with data scientists, analysts, software engineers, and product teams to ensure seamless data integration and access.
  • Implement robust monitoring, validation, and security controls to maintain high data reliability.

PythonSQLApache AirflowETLKafkaData engineeringData modeling

Posted 5 days ago
Apply

Related Articles

Posted 12 days ago

Why remote work is such a nice opportunity?

Why is remote work so nice? Let's try to see!

Posted 7 months ago

Insights into the evolving landscape of remote work in 2024 reveal the importance of certifications and continuous learning. This article breaks down emerging trends, sought-after certifications, and provides practical solutions for enhancing your employability and expertise. What skills will be essential for remote job seekers, and how can you navigate this dynamic market to secure your dream role?

Posted 7 months ago

Explore the challenges and strategies of maintaining work-life balance while working remotely. Learn about unique aspects of remote work, associated challenges, historical context, and effective strategies to separate work and personal life.

Posted 7 months ago

Google is gearing up to expand its remote job listings, promising more opportunities across various departments and regions. Find out how this move can benefit job seekers and impact the market.

Posted 7 months ago

Learn about the importance of pre-onboarding preparation for remote employees, including checklist creation, documentation, tools and equipment setup, communication plans, and feedback strategies. Discover how proactive pre-onboarding can enhance job performance, increase retention rates, and foster a sense of belonging from day one.