Apply

Senior Data Engineer

Posted 4 months agoViewed

View full description

πŸ’Ž Seniority level: Senior

πŸ“ Location: Worldwide

πŸ” Industry: Event Technology

πŸͺ„ Skills: AWSDockerPostgreSQLPythonSQLApache AirflowCloud ComputingData AnalysisETLKubernetesAlgorithmsApache KafkaData engineeringData StructuresCI/CDRESTful APIsMicroservicesData visualizationData modeling

Requirements:
NOT STATED
Responsibilities:
NOT STATEDApply

Related Jobs

Apply

πŸ“ States of SΓ£o Paulo and Rio Grande do Sul, cities of Rio de Janeiro, Belo Horizonte, FlorianΓ³polis and Fortaleza

🏒 Company: TELUS Digital Brazil

  • 5+ years of relevant development experience writing high-quality code as a Data Engineer
  • Have actively participated in the design and development of data architectures
  • Hands-on experience in developing and optimizing data pipelines
  • Comprehensive understanding of data modeling, ETL processes, and both SQL and NoSQL databases
  • Experience with a general-purpose programming language such as Python or Scala
  • Experience with GCP platforms and services.
  • Experience with containerization technologies such as Docker and Kubernetes
  • Proven track record in implementing and optimizing data warehousing solutions and data lakes
  • Proficiency in DevOps practices and automation tools for continuous integration and deployment of data solutions
  • Experience with machine learning workflows and supporting data scientists in model deployment
  • Solid understanding of data security and compliance requirements in large-scale data environments
  • Strong ability to communicate effectively with teams and stakeholders, providing and receiving feedback to improve product outcomes.
  • Proficient in communicating and writing in English
  • Develop and optimize scalable, high-performing, secure, and reliable data pipelines that address diverse business needs and considerations
  • Identify opportunities to enhance internal processes, implement automation to streamline manual tasks, and contribute to infrastructure redesign
  • Help mentor and coach a product team towards shared goals and outcomes
  • Navigate difficult conversations by providing constructive feedback to teams
  • Identify obstacles to ensure quality, improve our user experience and how we build tests
  • Be self-aware of limitations, yet curious to learn new solutions while being receptive to constructive feedback from teammates
  • Engage in ongoing research and adoption of new technologies, libraries, frameworks, and best practices to enhance the capabilities of the data team

DockerPythonSQLETLGCPHadoopKafkaKubernetesMachine LearningAirflowData engineeringNosqlSparkCI/CDAgile methodologiesRESTful APIsDevOpsScalaData visualizationData modeling

Posted about 15 hours ago
Apply
Apply
πŸ”₯ Senior Data Engineer
Posted about 21 hours ago

πŸ“ India

🧭 Full-Time

  • Hands-on experience in implementing, supporting, and administering modern cloud-based data solutions (Google BigQuery, AWS Redshift, Azure Synapse, Snowflake, etc.).
  • Strong programming skills in SQL, Java, and Python.
  • Experience in configuring and managing data pipelines using Apache Airflow, Informatica, Talend, SAP BODS or API-based extraction.
  • Expertise in real-time data processing frameworks.
  • Strong understanding of Git and CI/CD for automated deployment and version control.
  • Experience with Infrastructure-as-Code tools like Terraform for cloud resource management.
  • Good stakeholder management skills to collaborate effectively across teams.
  • Solid understanding of SAP ERP data and processes to integrate enterprise data sources.
  • Exposure to data visualization and front-end tools (Tableau, Looker, etc).
  • Design and Develop Data Pipelines: Create data pipelines to extract data from various sources, transform it into a standardized format, and load it into a centralized data repository.
  • Build and Maintain Data Infrastructure: Design, implement, and manage data warehouses, data lakes, and other data storage solutions.
  • Ensure Data Quality and Integrity: Develop data validation, cleansing, and normalization processes to ensure data accuracy and consistency.
  • Collaborate with Data Analysts and Business Process Owners: Work with data analysts and business process owners to understand their data requirements and provide data support for their projects.
  • Optimize Data Systems for Performance: Continuously monitor and optimize data systems for performance, scalability, and reliability.
  • Develop and Maintain Data Governance Policies: Create and enforce data governance policies to ensure data security, compliance, and regulatory requirements.

AWSPythonSQLApache AirflowCloud ComputingETLGitJavaSAPSnowflakeData engineeringCommunication SkillsCI/CDRESTful APIsTerraformData visualizationStakeholder managementData modelingEnglish communication

Posted about 21 hours ago
Apply
Apply

πŸ“ Germany, Spain, Portugal, Greece

🏒 Company: WorkMotionπŸ‘₯ 101-250πŸ’° $10,000,000 Series B almost 3 years agoComplianceHuman ResourcesEmployee Benefits

  • 3-5 years of professional experience in Data Engineering or Software Development with a focus on data
  • Strong knowledge of Python and SQL; and PySpark
  • Hands-on experience with AWS services (Glue, S3, Athena, EC2)
  • Experience with Apache Airflow, preferably in a Dockerized/cloud-native environment
  • Familiarity with Delta Lake or similar data lake frameworks
  • Proficiency with source control (GitHub) and CI/CD workflows
  • Strong understanding of data modeling, ETL best practices, and data pipeline performance optimization
  • Design, build, and maintain scalable ETL pipelines using Apache Airflow and AWS Glue (Spark)
  • Work with a range of data sources including Salesforce, NetSuite, PostgreSQL, and MongoDB
  • Develop and optimize PySpark jobs for large-scale data transformation and analytics
  • Manage data lake infrastructure using Delta Lake on S3 with Athena as the query layer
  • Ensure data quality, performance, and reliability through monitoring, testing, and documentation
  • Collaborate with analytics, product, and engineering teams to define data requirements
  • Contribute to CI/CD workflows with GitHub and deployment automation
  • Participate in architectural discussions and advocate for best practices in data engineering

AWSDockerPythonSoftware DevelopmentSQLApache AirflowETLGitData engineeringSparkCI/CDData modeling

Posted 2 days ago
Apply
Apply

πŸ“ Colombia

🧭 Full-Time

πŸ” Software Development

  • 5+ years of experience in developing scalable data pipeline infrastructure, preferably for sales organizations
  • Proven track record of delivering large-scale data projects and working with business partners
  • Experience with big data processing frameworks such as Apache Spark
  • Experience with data orchestration tools like Airflow or Dagster
  • Experience with infrastructure-as-code tools (e.g., Terraform) and modern CI/CD pipelines
  • Collaborate with other engineers, business partners, and data scientists to build best-in-class data infrastructure that meets evolving needs
  • Design and implement scalable data pipelines that integrate Salesforce and other sales systems data into our enterprise data lake
  • Build automated solutions for sales data quality, enrichment, and standardization
  • Create and maintain data models that power sales analytics, forecasting, and reporting systems
  • Design and manage reverse ETL pipelines to power sales operations and marketing automation
  • Partner with AI/ML engineers to develop Sales predictive and generative models
  • Architect solutions for real-time sales data synchronization and processing
  • Optimize data flows between Salesforce, Snowflake, AWS Athena, and other enterprise systems
  • Build robust monitoring and alerting systems for sales data pipelines
  • Collaborate with Sales Operations to automate manual processes and improve data accuracy
  • Create documentation and enable self-service capabilities for sales teams

AWSPythonApache AirflowSalesforceSnowflakeData engineeringCI/CDTerraformData modeling

Posted 3 days ago
Apply
Apply

πŸ“ United States of America

πŸ’Έ 78750.0 - 133875.0 USD per year

🏒 Company: vspvisioncareers

  • 6+ years’ experience working in development team providing analytical capabilities
  • 6+ years of hands-on experience in the data space, spanning data preparation, SQL, integration tools, ETL/ELT/data pipeline design
  • SQL coding experience
  • Experience working in an agile development environment (Scrum, Kanban) with a focus on Continuous Integration and Delivery
  • Knowledge about various data architectures, patterns, and capabilities such as event-driven architecture, real-time data flows, non-relational repositories, data virtualization, cloud storage, etc
  • Knowledge of and experience with multiple data integration platforms (IBM InfoSphere DataStage, Oracle Data Integrator, Informatica PowerCenter, MS SSIS, AWS Glue, Denodo), and data warehouse MPP platforms such as Snowflake, Netezza, Teradata, Redshift, etc
  • Collaborate within an agile, multi-disciplinary team to deliver optimal data integration and transformation solutions
  • Analyze data requirements (functional and non-functional) to develop and design robust, scalable, automated, fault-tolerant data pipeline solutions for business and technology initiatives
  • Design, build, maintain, and operationalize data pipelines for high volume and complex data using appropriate tools and practices in development, test, and production environments
  • Develop and design data mappings, programs, routines, and SQL to acquire data from legacy, web, cloud, and purchased package environments into the analytics environment
  • Drive automation of data pipeline preparation and integration tasks to minimize manual and error-prone processes and improve productivity using modern data preparation, integration, and AI-enabled metadata management tools and techniques
  • Participate in architecture, governance, and design reviews, identifying opportunities and making recommendations
  • Collaborate with architects to design and model application data structures, storage, and integration in accordance with enterprise-wide architecture standards across legacy, web, cloud, and purchased package environments

AWSSQLAgileETLSnowflakeApache KafkaData engineeringCI/CDRESTful APIsData visualizationData modelingData management

Posted 4 days ago
Apply
Apply

πŸ“ Brazil, Argentina, Colombia, Chile, Peru

πŸ” AdTech

🏒 Company: Workana

  • 5+ years working with data engineering, big data, or similar roles.
  • Strong SQL skills and hands-on experience with databases like BigQuery, Spanner, or equivalents.
  • Proficiency with GCP services (Dataflow, Pub/Sub, Cloud Storage).
  • Experience building ETL/ELT pipelines and working on data for analytics or targeting use cases.
  • Experience with container tools like Docker and Kubernetes.
  • Familiarity with event-streaming platforms (Kafka, Pub/Sub).
  • Knowledge of data modeling, query optimization, and performance tuning.
  • Proficient in at least one programming language used in data (e.g., Python, Go, or Java).
  • Design and build reliable data pipelines and ETL/ELT processes to move and transform data at scale.
  • Use GCP tools like BigQuery, Spanner, and Dataflow to manage real-time and batch data.
  • Work on systems that support audience targeting, insights generation, and campaign analytics.
  • Build and maintain APIs to connect data across different tools and teams.
  • Tune databases and queries for high performance.
  • Work with event-streaming tools such as Kafka or Pub/Sub to enable real-time processing.
  • Monitor and troubleshoot data quality, speed, and reliability issues.
  • Collaborate with engineers and analysts to improve how data is used across the company.

DockerPythonSQLCloud ComputingETLGCPJavaKubernetesApache KafkaAPI testingData engineeringGoData modeling

Posted 4 days ago
Apply
Apply

πŸ“ LATAM

πŸ” Software Development

🏒 Company: NearsureπŸ‘₯ 501-1000Staffing AgencyOutsourcingSoftware

  • Bachelor's Degree in Computer Science, Engineering, or a related field.
  • 5+ Years of experience working with cloud native architecture in a customer-facing or support role.
  • Experience with "Big Data" technologies or concepts, especially BigQuery, such as analytics warehousing, data processing, data transformation, data governance, data migrations, ETL, ELT, SQL, NoSQL, and other data concepts.
  • Experience with Machine Learning model development and deployment frameworks for deep learning (e.g., PyTorch, Tensorflow, Jax, Ray, etc.), AI accelerators (e.g., TPUs, GPUs), model architectures (e.g., encoders, decoders, transformers), and using machine learning APIs.
  • Experience working with and presenting to technical stakeholders and executive leaders.
  • Advanced English Level is required for this role as you will work with US clients. Effective communication in English is essential to deliver the best solutions to our clients and expand your horizons.
  • Design and maintain scalable data pipelines and ETL processes.
  • Manage and optimize relational and non-relational databases.
  • Build and maintain data storage solutions like data warehouses and data lakes.
  • Ensure data quality, accuracy, and integrity through validation processes.
  • Collaborate with analysts and data scientists to meet data needs.
  • Monitor and troubleshoot data workflows for efficiency and reliability.
  • Implement security measures to protect sensitive data.
  • Document data processes and architecture for future maintenance.

AWSPythonSQLApache AirflowCloud ComputingData AnalysisETLMachine LearningPyTorchData engineeringNosqlTensorflowData modeling

Posted 5 days ago
Apply
Apply

πŸ“ AL, AZ, AR, CA, CO, CT, DC, FL, GA, IL, KS, KY, MA, MI, MN, MO, NV, NJ, NY, NC, OH, OR, PA, SC, TN, TX, UT, VA, or WA

πŸ’Έ 116500.0 - 196700.0 USD per year

πŸ” Data Analytics

  • 3-5 years of experience in data engineering, data analytics, or business analysis roles.
  • 3 years of experience writing SQL for data warehousing.
  • 2-3 years of experience working on data warehouses to support analytics.
  • 2-3 years of experience building and maintaining enterprise data models, such as fact/dimensional star schemas.
  • Proficiency in cloud data warehouses (Redshift, BigQuery, Snowflake).
  • Strong SQL query writing and data modeling capabilities.
  • Experience with Python for scripting, data manipulation, and API interaction.
  • Familiarity with version control systems (e.g., Git, GitHub, GitLab).
  • Experience with workflow orchestration tools (e.g., Apache Airflow).
  • Knowledge of cloud platforms, particularly AWS.
  • Create and maintain Apache Airflow data pipeline jobs.
  • Monitor and update Fivetran extract and load runs.
  • Create and govern dbt models, testing, and job runs.
  • Monitor and maintain data quality processes and procedures.
  • Own, manage, and govern Snowflake data warehouse, including permission grants.
  • Maintain and govern usage of data sent from Snowflake to source systems ("Reverse ETL").
  • Collaborate with stakeholders to maintain and monitor existing data pipelines.
  • Evaluate automation opportunities within the business, using available data tools.
  • Provide data set and tooling recommendations where appropriate.
  • Provision data sets for analytic, master data, and reporting use cases.

AWSPythonSQLApache AirflowETLSnowflakeData engineeringData visualizationData modelingData analytics

Posted 5 days ago
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ’Έ 135000.0 - 145000.0 USD per year

πŸ” Life Science

🏒 Company: Medispend

  • Hands-on knowledge of data integration platforms
  • Experience with enterprise systems (ERP, CRM, etc.)
  • Substantial programming experience with Python based data orchestration and transformation frameworks (i.e. Airflow, AWS Glue, Prefect, Dagster, Spark, Polars, Databricks, etc.)
  • Strong working knowledge of traditional RDBMS data warehousing as well as other platforms like Snowflake, RedShift
  • Gather requirements and design the integration of a new data source
  • Design and build data transformations
  • Estimate levels of effort for prospective client implementations
  • Evaluate new open source data management tools
  • Determine root cause for a failed integration
  • Conduct peer review for code check-ins
  • Design and build a data migration framework
  • Monitor infrastructure capacity of the data transformation platform
  • Compile and analyze data transformation success/failure rates

AWSProject ManagementPythonSQLApache AirflowCloud ComputingETLSnowflakeJiraAlgorithmsData engineeringData StructuresPostgresRDBMSREST APISparkCommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingAgile methodologiesDevOpsData visualizationData modelingScriptingData analyticsData managementSaaS

Posted 5 days ago
Apply
Apply

πŸ“ Colombia, Brazil, Venezuela, Bolivarian Republic of

🏒 Company: JobgetherπŸ‘₯ 11-50πŸ’° $1,493,585 Seed about 2 years agoInternet

  • 8+ years of experience in data engineering or related backend roles, preferably in Agile environments
  • Proficiency in Scala and distributed data processing tools such as Apache Spark
  • Deep understanding of PostgreSQL, ClickHouse, and scalable NoSQL solutions like ScyllaDB
  • Experience with CI/CD (Jenkins), containerization (Docker), and cloud infrastructure (AWS)
  • Familiarity with Airflow, Hive, Redis, and Kafka for data orchestration and streaming
  • Ability to design secure, maintainable systems and lead large-scale infrastructure projects
  • Strong version control skills with Git and a commitment to testing and code reliability
  • Design and maintain scalable backend systems and data pipelines for batch and real-time processing
  • Own full project lifecycles from architecture through deployment and monitoring
  • Collaborate with product managers, UX designers, and engineering teams to deliver high-quality, data-driven solutions
  • Write clean, testable, and well-documented code following best software engineering practices
  • Review code, define technical direction, and enforce coding standards
  • Mentor fellow engineers and contribute to knowledge-sharing across the team
  • Continuously improve system performance and identify opportunities for technical enhancement

AWSBackend DevelopmentDockerPostgreSQLSQLGitJenkinsKafkaAirflowClickhouseData engineeringRedisNosqlSparkCI/CDScala

Posted 5 days ago
Apply