Apply

Data Engineer

Posted 13 days agoViewed

View full description

πŸ’Ž Seniority level: Middle, 3+ years

πŸ“ Location: Brazil, Argentina, Uruguay, Paraguay, Mexico

πŸ” Industry: Software Development

🏒 Company: Foxbox DigitalπŸ‘₯ 51-100AndroidiOSWeb DevelopmentMobile AppsWeb DesignSoftware

πŸ—£οΈ Languages: English

⏳ Experience: 3+ years

πŸͺ„ Skills: AWSDockerPythonSQLETLGCPHadoopKafkaKubernetesAirflowAzureData engineeringNosqlSparkCI/CDData modeling

Requirements:
  • 3+ years of hands-on experience in data engineering, building ETL/ELT pipelines, and managing data workflows in a production environment.
  • Proficient in Python or Java for data processing.
  • Familiar with SQL and NoSQL databases.
  • Experience with cloud-based data solutions (AWS, Azure, or GCP).
  • Knowledge of distributed data processing (e.g., Spark, Hadoop, Databricks).
  • Understanding of containerization (e.g., Docker) and orchestration (e.g., Kubernetes) is a plus.
  • Experience with Airflow, Kafka, or Azure Data Factory is preferred.
  • Familiarity with CI/CD practices and Git-based workflows is preferred.
  • Exposure to MLOps tools like MLflow or Kubeflow is a plus.
Responsibilities:
  • Build and maintain scalable ETL/ELT pipelines, transforming raw data into ready-to-use datasets for machine learning.
  • Connect to various data sources (e.g., Quickbase, REST APIs, databases) and ensure seamless data ingestion.
  • Implement validation rules, data profiling, and monitoring to maintain high data integrity.
  • Work with data scientists and MLOps engineers to optimize data workflows, improve model performance, and address bottlenecks.
  • Identify and resolve issues around data throughput and latency, ensuring pipelines can handle production-level loads.
  • Enforce data privacy best practices and role-based access control in accordance with project or regulatory requirements.
  • Set up alerts, logs, and metrics to proactively detect and resolve pipeline or data-related issues.
Apply

Related Jobs

Apply

πŸ“ Worldwide

πŸ” Hospitality

🏒 Company: Lighthouse

  • 4+ years of professional experience using Python, Java, or Scala for data processing (Python preferred)
  • You stay up-to-date with industry trends, emerging technologies, and best practices in data engineering.
  • Improve, manage, and teach standards for code maintainability and performance in code submitted and reviewed
  • Ship large features independently, generate architecture recommendations and have the ability to implement them
  • Great communication: Regularly achieve consensus amongst teams
  • Familiarity with GCP, Kubernetes (GKE preferred),Β  CI/CD tools (Gitlab CI preferred), familiarity with the concept of Lambda Architecture.
  • Experience with Apache Beam or Apache Spark for distributed data processing or event sourcing technologies like Apache Kafka.
  • Familiarity with monitoring tools like Grafana & Prometheus.
  • Design and develop scalable, reliable data pipelines using the Google Cloud stack.
  • Optimise data pipelines for performance and scalability.
  • Implement and maintain data governance frameworks, ensuring data accuracy, consistency, and compliance.
  • Monitor and troubleshoot data pipeline issues, implementing proactive measures for reliability and performance.
  • Collaborate with the DevOps team to automate deployments and improve developer experience on the data front.
  • Work with data science and analytics teams to enable them to bring their research to production grade data solutions, using technologies like airflow, dbt or MLflow (but not limited to)
  • As a part of a platform team, you will communicate effectively with teams across the entire engineering organisation, to provide them with reliable foundational data models and data tools.
  • Mentor and provide technical guidance to other engineers working with data.

PythonSQLApache AirflowETLGCPKubernetesApache KafkaData engineeringCI/CDMentoringTerraformScalaData modeling

Posted 4 days ago
Apply
Apply

πŸ“ Argentina

🧭 Full-Time

πŸ” Software Development

🏒 Company: Austin Software

  • 5+ years experience as a Data Engineer
  • 4+ years experience with MySQL
  • Experience with Python
  • Experience with Spark jobs written in Scala
  • Experience with Databricks

AWSPythonMySQLData engineeringSparkScala

Posted 5 days ago
Apply
Apply

πŸ“ Worldwide

🧭 Full-Time

NOT STATED
  • Own the design and implementation of cross-domain data models that support key business metrics and use cases.
  • Partner with analysts and data engineers to translate business logic into performant, well-documented dbt models.
  • Champion best practices in testing, documentation, CI/CD, and version control, and guide others in applying them.
  • Act as a technical mentor to other analytics engineers, supporting their development and reviewing their code.
  • Collaborate with central data platform and embedded teams to improve data quality, metric consistency, and lineage tracking.
  • Drive alignment on model architecture across domainsβ€”ensuring models are reusable, auditable, and trusted.
  • Identify and lead initiatives to reduce technical debt and modernise legacy reporting pipelines.
  • Contribute to the long-term vision of analytics engineering at Pleo and help shape our roadmap for scalability and impact.

SQLData AnalysisETLData engineeringCI/CDMentoringDocumentationData visualizationData modelingData analyticsData management

Posted 5 days ago
Apply
Apply

πŸ“ LatAm

🧭 Full-Time

πŸ” E-Learning

🏒 Company: TruelogicπŸ‘₯ 101-250ConsultingWeb DevelopmentWeb DesignSoftware

  • 1-3 years of experience working with PySpark and Apache Spark in Big Data environments.
  • Experience with SQL and relational and NoSQL databases (PostgreSQL, MySQL, MongoDB, etc.).
  • Knowledge of ETL processes and data processing in distributed environments.
  • Familiarity with Apache Hadoop, Hive, or Delta Lake.
  • Experience with cloud storage (AWS S3, Google Cloud Storage, Azure Blob).
  • Proficiency in Git and version control.
  • Strong problem-solving skills and a proactive attitude.
  • A passion for learning and continuous improvement.
  • Design, develop, and optimize data pipelines using PySpark and Apache Spark.
  • Integrate and process data from multiple sources (databases, APIs, files, streaming).
  • Implement efficient data transformations for Big Data in distributed environments.
  • Optimize code to improve performance, scalability, and efficiency in data processing.
  • Collaborate with Data Science, BI, and DevOps teams to ensure seamless integration.
  • Monitor and debug data processes to ensure quality and reliability.
  • Apply best practices in data engineering and maintain clear documentation.
  • Stay up to date with the latest trends in Big Data and distributed computing.

PostgreSQLSQLApache HadoopCloud ComputingETLGitMongoDBMySQLApache Kafka

Posted 5 days ago
Apply
Apply
πŸ”₯ Data Engineer
Posted 7 days ago

πŸ“ Worldwide

🧭 Full-Time

πŸ’Έ 145000.0 - 160000.0 USD per year

  • Proficiency in managing MongoDB databases, including performance tuning and maintenance.
  • Experience with cloud-based data warehousing, particularly using BigQuery.
  • Familiarity with DBT for data transformation and modeling.
  • Exposure to tools like Segment for data collection and integration.
  • Basic knowledge of integrating third-party data sources to build a comprehensive data ecosystem.
  • Overseeing our production MongoDB database to ensure optimal performance, reliability, and security.
  • Assisting in the management and optimization of data pipelines into BigQuery, ensuring data is organized and accessible for downstream users.
  • Utilizing DBT to transform raw data into structured formats, making it useful for analysis and reporting.
  • Collaborating on the integration of data from Segment and various third-party sources to create a unified, clean data ecosystem.
  • Working closely with BI, Marketing, and Data Science teams to understand data requirements and ensure our infrastructure meets their needs.
  • Participating in code reviews, learning new tools, and contributing to the refinement of data processes and best practices.

SQLETLMongoDBData engineeringData modeling

Posted 7 days ago
Apply
Apply

πŸ“ Costa Rica, Brazil, Argentina, Chile, Mexico

πŸ” Insider Risk Management

🏒 Company: TeramindπŸ‘₯ 51-100Productivity ToolsSecurityCyber SecurityEnterprise SoftwareSoftware

  • 6+ years of experience in data engineering, with a proven track record of successfully delivering data-driven solutions.
  • Strong expertise in designing and building scalable data pipelines using industry-standard tools and frameworks.
  • Experience with big data technologies and distributed systems, such as Hadoop, Spark, or similar frameworks.
  • Proficient programming skills in languages such as Python, Java, or Scala, alongside a solid understanding of database management systems (SQL and NoSQL).
  • Understanding of data requirements for machine learning applications and how to optimize data for model training.
  • Experience with security data processing and compliance standards is preferred, ensuring that data handling meets industry regulations and best practices.
  • Design and implement robust data architecture tailored for AI-driven features, ensuring it meets the evolving needs of our platform.
  • Build and maintain efficient data pipelines for processing user activity data, ensuring data flows seamlessly throughout our systems.
  • Develop comprehensive systems for data storage, retrieval, and processing, facilitating quick and reliable access to information.
  • Ensure high standards of data quality and availability, enabling machine learning models to produce accurate and actionable insights.
  • Enhance the performance and scalability of our data infrastructure to accommodate growing data demands and user activity.
  • Work closely with data scientists and machine learning engineers to understand their data requirements and ensure data solutions are tailored to their needs.

PythonSQLApache HadoopETLMachine LearningAzureData engineeringNosqlComplianceScalaData visualizationData modelingData management

Posted 7 days ago
Apply
Apply

πŸ“ LATAM

🧭 Contract

πŸ” Pharma

  • Minimum 6+ years of experience working as a Data Engineer is a must have.
  • Experience working with data management teams supporting Pharma field operations is required.
  • Experience in data-centric implementations such as SFA/CRM and Business Intelligence Tools.
  • 3+ years of experience with data warehouse, Master Data Management (MDM), and data integration initiatives.
  • 3+ years of experience building ETL/ELT solutions using data integration tools such as Azure Data Factory, Azure Functions, Logic Apps, Databricks & Snowflake.
  • Proficiency in Snowflake data warehousing.
  • Prior knowledge of best practices in SDLC and use of source control management tool preferably Azure DevOps to GitHub integration is highly desired.
  • Strong organizational skills.
  • Experience working in a production environment for multiple clients in a compliance industry.
  • Ability to work in a team through implementing and adopting a data management framework.
  • Ability to translate customer needs into technical requirements.
  • Advanced written and spoken English fluency along with strong verbal and written communication skills.
  • Detail-oriented and analytical with planning and organizational skills.
  • Excellent interpersonal skills.
  • Specific knowledge of client business and technology/data operations in the pharmaceutical industry.
  • Skill in troubleshooting and problem-solving.
  • Ability to work in an environment that is driven by process governance.
  • Works independently to process information by loading / extraction of data.
  • Designs and develops a large-scale data lake managing data flows including but not limited to Veeva/Salesforce and various sources into a common customer data hub.
  • Orchestrates data pipelines to extract, transform, and load CRM data and create templatized data solutions to provide data to downstream teams, such as Sales Analytics, Incentive Compensation, and Data Science teams.
  • Performs quality control functions on all deliverables, including standard deliverables, ad hoc loading and extraction projects.
  • Develops and maintains project documentation, including work instructions for all repeatable tasks.
  • Provides feedback on the design of new integration initiatives.
  • Meets clients to establish, monitor, and review deliverables.
  • Engages with clients to facilitate data management services related to project delivery schedule.

SQLETLSalesforceSnowflakeData engineeringCommunication SkillsAnalytical SkillsCRMData modelingData management

Posted 12 days ago
Apply
Apply

πŸ“ Brazil

🧭 Full-Time

πŸ” Software Development

🏒 Company: EncoraπŸ‘₯ 10001-10001πŸ’° $200,000,000 Private over 5 years agoBig DataCloud ComputingSoftware

  • Experience as a Data Engineer
  • Strong experience with Apache Flink, including knowledge of the DataStream and Table APIs
  • Experience with Kafka and Kafka Connect for data integration Β· Experience with containerization (Docker) and orchestration (Kubernetes)
  • Experience in Cloud Infrastructure platforms such as AWS, GCP
  • Solid SQL skills with specific experience in BigQuery SQL dialect
  • Experience with data modeling and schema design
  • Proficient in at least one programming language (Java or Python)
  • Experience with version control systems (Git) and CI/CD pipelines
  • Design, develop, and optimize data pipelines using Apache Flink for stream and batch processing
  • Implement and maintain Kafka Connect connectors for seamless data integration
  • Build and maintain data infrastructure on Google Cloud Platform (GCP)
  • Design and optimize BigQuery tables, views, and stored procedures
  • Collaborate with product managers and analysts to understand data requirements
  • Ensure data quality, reliability, and proper governance
  • Troubleshoot and resolve data pipeline issues
  • Document data flows, transformations, and architecture

DockerPythonSQLGCPGitJavaKafkaKubernetesData engineeringCI/CDData modeling

Posted 13 days ago
Apply
Apply

πŸ“ Mexico

πŸ” Software Development

  • Enthusiastic about data and continuous improvement.
  • Ability to balance and prioritize multiple projects.
  • Ability to ask insightful questions in an ambiguous environment.
  • Ability to coordinate execution across various stakeholders.
  • A proactive and hands-on approach
  • Support driving JumpCloud’s core reporting by developing data pipelines, curating and modeling data, and designing databases to enable insights and self-service across the analytics team and organization.
  • Collaborate with analytics teams, technical teams, and business stakeholders to develop and maintain data pipelines, curated data models, and database designs that will support business and analytics insights throughout the organization.

AWSPostgreSQLPythonSQLCloud ComputingData AnalysisETLAirflowData engineeringREST APICommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingAgile methodologiesData visualizationData modelingScriptingData analyticsData management

Posted 14 days ago
Apply
Apply

πŸ“ LATAM

🧭 Full-Time

πŸ” Financial Services

🏒 Company: South GeeksπŸ‘₯ 101-250Web DevelopmentSoftware EngineeringEnterprise SoftwareSoftware

  • Proficiency in administering one or more platforms: Snowflake (required), DBT (required), GitHub, Workato (and/or Made), or other tools (preferred).
  • Data engineering experience with a focus on DBT and Snowflake.
  • Strong desire to expand skills across multiple platforms and contribute to both administrative and engineering functions.
  • Comfort with dynamic, multi-role environments that blend administration with engineering work.
  • Collaborative mindset, able to thrive in a team-first, cross-functional setting.
  • Administer one or more platforms, focusing on: Snowflake, DBT, Fivetran, and other tools: GitHub, Workato (and/or Make), and High Touch.
  • Participate in cross-training to administer multiple platforms, ensuring seamless coverage across the team.
  • Collaborate on data engineering projects, using DBT and Snowflake as part of the stack.
  • Take part in development opportunities and engineering work to broaden your expertise and career path.
  • Focus on team-based service delivery, rather than individual responsibilities, ensuring all systems are effectively managed by the group.

SQLSnowflakeData engineering

Posted 21 days ago
Apply