Apply

Senior Data Engineer

Posted 6 days agoViewed

View full description

💎 Seniority level: Senior, 5+ years

📍 Location: Brazil, Argentina, Colombia, Chile, Peru

🔍 Industry: AdTech

🏢 Company: Workana

⏳ Experience: 5+ years

🪄 Skills: DockerPythonSQLCloud ComputingETLGCPJavaKubernetesApache KafkaAPI testingData engineeringGoData modeling

Requirements:
  • 5+ years working with data engineering, big data, or similar roles.
  • Strong SQL skills and hands-on experience with databases like BigQuery, Spanner, or equivalents.
  • Proficiency with GCP services (Dataflow, Pub/Sub, Cloud Storage).
  • Experience building ETL/ELT pipelines and working on data for analytics or targeting use cases.
  • Experience with container tools like Docker and Kubernetes.
  • Familiarity with event-streaming platforms (Kafka, Pub/Sub).
  • Knowledge of data modeling, query optimization, and performance tuning.
  • Proficient in at least one programming language used in data (e.g., Python, Go, or Java).
Responsibilities:
  • Design and build reliable data pipelines and ETL/ELT processes to move and transform data at scale.
  • Use GCP tools like BigQuery, Spanner, and Dataflow to manage real-time and batch data.
  • Work on systems that support audience targeting, insights generation, and campaign analytics.
  • Build and maintain APIs to connect data across different tools and teams.
  • Tune databases and queries for high performance.
  • Work with event-streaming tools such as Kafka or Pub/Sub to enable real-time processing.
  • Monitor and troubleshoot data quality, speed, and reliability issues.
  • Collaborate with engineers and analysts to improve how data is used across the company.
Apply

Related Jobs

Apply

📍 States of São Paulo and Rio Grande do Sul, cities of Rio de Janeiro, Belo Horizonte, Florianópolis and Fortaleza

🏢 Company: TELUS Digital Brazil

  • 5+ years of relevant development experience writing high-quality code as a Data Engineer
  • Have actively participated in the design and development of data architectures
  • Hands-on experience in developing and optimizing data pipelines
  • Comprehensive understanding of data modeling, ETL processes, and both SQL and NoSQL databases
  • Experience with a general-purpose programming language such as Python or Scala
  • Experience with GCP platforms and services.
  • Experience with containerization technologies such as Docker and Kubernetes
  • Proven track record in implementing and optimizing data warehousing solutions and data lakes
  • Proficiency in DevOps practices and automation tools for continuous integration and deployment of data solutions
  • Experience with machine learning workflows and supporting data scientists in model deployment
  • Solid understanding of data security and compliance requirements in large-scale data environments
  • Strong ability to communicate effectively with teams and stakeholders, providing and receiving feedback to improve product outcomes.
  • Proficient in communicating and writing in English
  • Develop and optimize scalable, high-performing, secure, and reliable data pipelines that address diverse business needs and considerations
  • Identify opportunities to enhance internal processes, implement automation to streamline manual tasks, and contribute to infrastructure redesign
  • Help mentor and coach a product team towards shared goals and outcomes
  • Navigate difficult conversations by providing constructive feedback to teams
  • Identify obstacles to ensure quality, improve our user experience and how we build tests
  • Be self-aware of limitations, yet curious to learn new solutions while being receptive to constructive feedback from teammates
  • Engage in ongoing research and adoption of new technologies, libraries, frameworks, and best practices to enhance the capabilities of the data team

DockerPythonSQLETLGCPHadoopKafkaKubernetesMachine LearningAirflowData engineeringNosqlSparkCI/CDAgile methodologiesRESTful APIsDevOpsScalaData visualizationData modeling

Posted 2 days ago
Apply
Apply

📍 Colombia

🧭 Full-Time

🔍 Software Development

  • 5+ years of experience in developing scalable data pipeline infrastructure, preferably for sales organizations
  • Proven track record of delivering large-scale data projects and working with business partners
  • Experience with big data processing frameworks such as Apache Spark
  • Experience with data orchestration tools like Airflow or Dagster
  • Experience with infrastructure-as-code tools (e.g., Terraform) and modern CI/CD pipelines
  • Collaborate with other engineers, business partners, and data scientists to build best-in-class data infrastructure that meets evolving needs
  • Design and implement scalable data pipelines that integrate Salesforce and other sales systems data into our enterprise data lake
  • Build automated solutions for sales data quality, enrichment, and standardization
  • Create and maintain data models that power sales analytics, forecasting, and reporting systems
  • Design and manage reverse ETL pipelines to power sales operations and marketing automation
  • Partner with AI/ML engineers to develop Sales predictive and generative models
  • Architect solutions for real-time sales data synchronization and processing
  • Optimize data flows between Salesforce, Snowflake, AWS Athena, and other enterprise systems
  • Build robust monitoring and alerting systems for sales data pipelines
  • Collaborate with Sales Operations to automate manual processes and improve data accuracy
  • Create documentation and enable self-service capabilities for sales teams

AWSPythonApache AirflowSalesforceSnowflakeData engineeringCI/CDTerraformData modeling

Posted 5 days ago
Apply
Apply

📍 LATAM

🔍 Software Development

🏢 Company: Nearsure👥 501-1000Staffing AgencyOutsourcingSoftware

  • Bachelor's Degree in Computer Science, Engineering, or a related field.
  • 5+ Years of experience working with cloud native architecture in a customer-facing or support role.
  • Experience with "Big Data" technologies or concepts, especially BigQuery, such as analytics warehousing, data processing, data transformation, data governance, data migrations, ETL, ELT, SQL, NoSQL, and other data concepts.
  • Experience with Machine Learning model development and deployment frameworks for deep learning (e.g., PyTorch, Tensorflow, Jax, Ray, etc.), AI accelerators (e.g., TPUs, GPUs), model architectures (e.g., encoders, decoders, transformers), and using machine learning APIs.
  • Experience working with and presenting to technical stakeholders and executive leaders.
  • Advanced English Level is required for this role as you will work with US clients. Effective communication in English is essential to deliver the best solutions to our clients and expand your horizons.
  • Design and maintain scalable data pipelines and ETL processes.
  • Manage and optimize relational and non-relational databases.
  • Build and maintain data storage solutions like data warehouses and data lakes.
  • Ensure data quality, accuracy, and integrity through validation processes.
  • Collaborate with analysts and data scientists to meet data needs.
  • Monitor and troubleshoot data workflows for efficiency and reliability.
  • Implement security measures to protect sensitive data.
  • Document data processes and architecture for future maintenance.

AWSPythonSQLApache AirflowCloud ComputingData AnalysisETLMachine LearningPyTorchData engineeringNosqlTensorflowData modeling

Posted 6 days ago
Apply
Apply

📍 Colombia, Brazil, Venezuela, Bolivarian Republic of

🏢 Company: Jobgether👥 11-50💰 $1,493,585 Seed about 2 years agoInternet

  • 8+ years of experience in data engineering or related backend roles, preferably in Agile environments
  • Proficiency in Scala and distributed data processing tools such as Apache Spark
  • Deep understanding of PostgreSQL, ClickHouse, and scalable NoSQL solutions like ScyllaDB
  • Experience with CI/CD (Jenkins), containerization (Docker), and cloud infrastructure (AWS)
  • Familiarity with Airflow, Hive, Redis, and Kafka for data orchestration and streaming
  • Ability to design secure, maintainable systems and lead large-scale infrastructure projects
  • Strong version control skills with Git and a commitment to testing and code reliability
  • Design and maintain scalable backend systems and data pipelines for batch and real-time processing
  • Own full project lifecycles from architecture through deployment and monitoring
  • Collaborate with product managers, UX designers, and engineering teams to deliver high-quality, data-driven solutions
  • Write clean, testable, and well-documented code following best software engineering practices
  • Review code, define technical direction, and enforce coding standards
  • Mentor fellow engineers and contribute to knowledge-sharing across the team
  • Continuously improve system performance and identify opportunities for technical enhancement

AWSBackend DevelopmentDockerPostgreSQLSQLGitJenkinsKafkaAirflowClickhouseData engineeringRedisNosqlSparkCI/CDScala

Posted 6 days ago
Apply
Apply

📍 Bogotá, Bogota, Colombia, Mexico, Brazil, Peru, Uruguay

🔍 AdTech

🏢 Company: Workana

  • 6+ years of experience in data engineering or related roles, preferably within the AdTech industry.
  • Expertise in SQL and experience with relational databases such as BigQuery and SpannerDB or similar.
  • Experience with GCP services, including Dataflow, Pub/Sub, and Cloud Storage.
  • Experience building and optimizing ETL/ELT pipelines in support of audience segmentation and analytics use cases.
  • Experience with Docker and Kubernetes for containerization and orchestration.
  • Familiarity with message queues or event-streaming tools, such as Kafka or Pub/Sub.
  • Knowledge of data modeling, schema design, and query optimization for performance at scale.
  • Programming experience in languages like Python, Go, or Java for data engineering tasks
  • Build and optimize data pipelines and ETL/ELT processes to support AdTech products: Insights, Activation, and Measurement.
  • Leverage GCP tools like BigQuery, SpannerDB, and Dataflow to process and analyze real-time consumer-permissioned data.
  • Design scalable and robust data solutions to power audience segmentation, targeted advertising, and outcome measurement.
  • Develop and maintain APIs to facilitate data sharing and integration across the platform’s products.
  • Optimize database and query performance to ensure efficient delivery of advertising insights and analytics.
  • Work with event-driven architectures using tools like Pub/Sub or Kafka to ensure seamless data processing across products.
  • Proactively monitor and troubleshoot issues to maintain data accuracy, security, and performance.
  • Drive innovation by identifying opportunities to enhance the platform’s capabilities in audience targeting and measurement.

DockerPythonSQLETLGCPJavaKafkaKubernetesData engineeringGoData modeling

Posted 10 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 11 days ago

📍 Worldwide

🧭 Full-Time

💸 140000.0 - 175000.0 USD per year

🔍 Software Development

🏢 Company: Figment👥 11-50HospitalityTravel AccommodationsArt

  • Extensive experience with data engineering, including building and managing data pipelines and ETL processes.
  • Proficiency in the Python programming language and SQL.
  • Experience developing highly concurrent and performant applications ensuring scalability and efficient resource utilization in distributed or multi-threaded systems.
  • Experience implementing robust microservices following best practices in error handling, logging, and testing for production-grade systems.
  • Experience with using CI/CD pipelines for automated data infrastructure provisioning and application deployment.
  • Experience with the data orchestration tool Dagster or Airflow.
  • Experience designing and orchestrating complex DAGs to manage dependencies, triggers, and retries for data workflows, ensuring reliable and efficient pipeline execution.
  • Experience with the data transformation tool DBT.
  • Experience designing and implementing complex data transformations using advanced DBT models, materializations, and configurations to streamline data workflows and improve performance.
  • Experience optimizing and troubleshoot DBT pipelines for scale, ensuring that transformations run efficiently in production environments, handling large datasets without issues.
  • Experience with cloud data warehousing platforms (e.g. Snowflake)
  • Experience architecting and optimizing Snowflake environments for performance, including designing partitioning strategies, clustering keys, and storage optimizations for cost-effective scaling.
  • Has an understanding of security and governance policies within Snowflake, including data encryption, access control, and audit logging to meet compliance and security best practices.
  • Implement and maintain reliable data pipelines and data storage solutions.
  • Implement data modeling and integrate technologies according to project needs.
  • Manage specific data pipelines and oversees the technical aspects of data operations
  • Ensure data processes are optimized and align with business requirements
  • Identify areas for process improvements and suggests tools and technologies to enhance efficiency
  • Continuously improve data infrastructure automation, ensuring reliable and efficient data processing.
  • Develop and maintain data pipelines and ETL processes using technologies such as Dagster and DBT to ensure efficient data flow and processing.
  • Automate data ingestion, transformation, and loading processes to support blockchain data analytics and reporting.
  • Utilize Snowflake data warehousing solutions to manage and optimize data storage and retrieval.
  • Collaborate with Engineering Leadership and Product teams to articulate data strategies and progress.
  • Promote best practices in data engineering, cloud infrastructure, networking, and security.

PythonSQLCloud ComputingETLSnowflakeData engineeringCI/CDRESTful APIsMicroservicesData modeling

Posted 11 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 13 days ago

📍 Americas

🧭 Full-Time

🔍 Software Development

🏢 Company: Virtasant Inc.

  • Strong proficiency in Python and SQL (non-negotiable).
  • Experience building ETL pipelines in production environments.
  • Familiarity with Databricks and big data tools (preferred).
  • Knowledge of visualization tools like Superset, Dash, Plotly.
  • Strong knowledge of Pandas; PySpark is a plus.
  • Focus on batch data processing (not real-time streaming).
  • Building and maintaining ETL pipelines using Python and SQL.
  • Managing, transforming, and optimizing structured datasets within Databricks.
  • Writing performant SQL queries to support reporting and analytics.
  • Developing dashboards and reporting tools using Superset, Dash, and Plotly.
  • Identifying and resolving inefficiencies in data workflows.
  • Enabling stakeholders to make data-driven decisions through automated reporting.
  • Collaborating closely with Client teams to understand data needs and deliver solutions.

PythonSQLETLPandas

Posted 13 days ago
Apply
Apply

📍 LATAM

🏢 Company: Nearsure👥 501-1000Staffing AgencyOutsourcingSoftware

  • 5+ Years of experience working in data engineering.
  • Advanced skills in Databricks, Delta Lake, Spark, SQL, and Python
  • Proficiency in BI tools: QuickSight, Tableau, or Power BI.
  • Design, build, and maintain robust data pipelines using Databricks.
  • Create and manage gold tables optimized for business analytics.
  • Develop dashboards using AWS QuickSight or similar BI platforms.
  • Ensure high data quality, performance, and traceability.
  • Collaborate with engineering, product, and analytics teams.

AWSPythonSQLData engineeringSpark

Posted 18 days ago
Apply
Apply

📍 Brazil

🧭 Full-Time

🔍 Software Development

🏢 Company: Quorum👥 251-500💰 almost 5 years agoCRMGovernmentPoliticsSaaSData VisualizationSoftware

  • 5+ years in data engineering with a track record of developing and scaling data-driven products
  • Expertise in building data pipelines, architectures, and datasets, with experience in AWS cloud services (EC2, EMR, RDS, Redshift).
  • Proficient in big data tools (Hadoop, Spark, Kafka) and machine learning frameworks (TensorFlow, PyTorch).
  • 3+ years experience with Python
  • Deep knowledge of SQL and NoSQL databases, workflow management tools (Azkaban, Luigi, Airflow), and an understanding of the machine learning model deployment cycle.
  • Top candidates will have experience or working knowledge of vector databases and advanced Retrieval-Augmented Generation (RAG) systems (Langchain, Pinecone, OpenAI/ChatGPT).
  • Architect and implement highly scalable advanced Retrieval-Augmented Generation (RAG) data pipelines to support the aggregation and analysis of legislative bills, social media posts, documents, and testimonial content.
  • Design robust data pipelines that enable real-time processing and analysis of vast datasets to inform public affairs strategies and decision-making.
  • Design and implement data cleansing and transformation pipelines to power the AI product.
  • Lead cloud-based deployments, primarily in AWS, ensuring optimal performance, security, and cost-efficiency.
  • Innovate and iterate on data architecture to support Quorum Copilot's dynamic needs, ensuring the product remains at the cutting edge of AI-driven public affairs tools.
  • Drive build vs buy, tool selection, and solution trade-off analysis using standard engineering principles.

AWSPythonSQLCloud ComputingETLMachine LearningAirflowData engineeringNosqlCI/CDRESTful APIsData modeling

Posted 23 days ago
Apply
Apply

📍 Colombia

🧭 Full-Time

🔍 Data Engineering

🏢 Company: Aimpoint Digital👥 1-50ConsultingAnalyticsAdvice

  • 3+ years working with relational databases and query languages
  • 3+ years building data pipelines in production and ability to work across structured, semi-structured and unstructured data
  • 3+ years data modeling (e.g. star schema, entity-relationship)
  • 3+ years writing clean, maintainable, and robust code in Python, Scala, Java, or similar coding languages
  • Become a trusted advisor working together with our clients, from data owners and analytic users to C-level executives
  • Work independently as part of a small team to solve complex data engineering use-cases across a variety of industries
  • Design and develop the analytical layer, building cloud data warehouses, data lakes, ETL/ELT pipelines, and orchestration jobs
  • Support the deployment of data science and ML projects into production

AWSPythonSQLCloud ComputingData AnalysisETLGCPGitSnowflakeAzureData engineeringRDBMSSparkCommunication SkillsCI/CDProblem SolvingRESTful APIsDevOpsData visualizationData modeling

Posted 26 days ago
Apply