Apply

Data Engineer

Posted 16 days agoViewed

View full description

🔍 Industry: Logistics

🏢 Company: Skydropx👥 251-500💰 Series B over 2 years agoProductivity ToolsLogisticsDeliveryE-CommerceInformation TechnologySoftware

Requirements:
  • Ingeniería, Actuaría, Matemáticas Aplicadas, Ingeniería Industrial o afín.
  • Experiencia sólida en programación, con especial énfasis en Python.
  • Experiencia en la manipulación de datos a gran escala y optimización de consultas SQL.
  • Experiencia en el uso de herramientas como Matillion, SSIS, Oracle para la construcción y gestión de flujos ETL.
  • Experiencia demostrable en el manejo de bases de datos, incluyendo la optimización de consultas SQL.
  • Conocimiento sobre bases de datos relacionales y no relacionales. Asi como de Datawarehouse, datalake, datavault.
Responsibilities:
  • Recolectar datos de diversas fuentes, incluyendo bases de datos, archivos CSV y APIs, utilizando herramientas ETL para realizar limpieza y preprocesamiento.
  • Diseñar y construir pipelines de datos eficientes utilizando herramientas como Matillion, garantizando la automatización en la recolección, transformación y carga (ETL) de datos.
  • Integrar y consumir datos de APIs externas para enriquecer conjuntos de datos y mejorar la calidad de la información.
  • Contribuir al diseño de arquitecturas eficientes para el manejo de grandes volúmenes de datos, utilizando tecnologías como Python y SQL.
Apply

Related Jobs

Apply
🔥 Staff Data Engineer
Posted about 3 hours ago

📍 United States

🧭 Full-Time

🔍 Healthcare

🏢 Company: Atropos Health

  • At least 5 years of experience working with messy real-world healthcare data (EHR data, claims) and experience with Real World Data / Evidence at least 3 years
  • Expertise working with common data models and terminology systems in healthcare data (data models and standards such as FHIR, OMOP, codesets like ICD 9/10, CPT, ATC, RxNorm, NDC, etc.)
  • Knowledge of common EHR and medical claim / administrative workflows and the basics of the US healthcare system (patient movement through a hospital or clinic, how medical procedures are reimbursed, etc.)
  • SQL (any dialect): at least 5 years, Python at least 4 years (Required)
  • Significant and deep experience as a data professional working with public cloud infrastructure to build data products (Required)
  • Knowledge of EHR data systems and at least 4 years of experience working intensively with EHR data in a data engineering and/or analytics environment
  • High level knowledge of ETL/ELT operations and data modeling
  • Create and maintain data pipelines to integrate, enrich, and map clinical, claims, and other data from multiple sources in the cloud
  • Design systems of data quality checks and assessments on clinical datasets
  • Expert knowledge of how to automate workflows and create reusable infrastructure in multiple public cloud environments (AWS, Azure, GCP)
  • Map data from one source into a common data model using many different frameworks, with a particular focus on PySpark and Spark SQL.
  • Create, maintain, tune, and document data flows in Databricks and Snowflake, Bigquery, and other diverse cloud tools
  • Filter, clean, transform, and integrate clinical data using your domain expertise of healthcare data and code systems and embedded knowledge
  • Work remotely with an interdisciplinary team and build and maintain strong relationships with research, clinical, product, and commercial stakeholders
  • Manage external customer, vendors, and customer relationships and create internal and external communications
  • Travel required to company or team offsites 3-4 weeks per year

AWSPythonSQLCloud ComputingETLSnowflakeData engineeringRDBMSSparkData modeling

Posted about 3 hours ago
Apply
Apply
🔥 Data Engineer
Posted about 3 hours ago

📍 Medellin, Colombia

🧭 Full-Time

🔍 Software Development

🏢 Company: GoodUnited👥 11-50💰 $1,148,052 over 1 year agoInternetSaaSInformation TechnologySocial Media

  • Proficiency in Python and SQL.
  • Experience with ETL tools such as Matillion, Apache AirFlow, or equivalent.
  • Experience with data warehouses like Snowflake or Redshift
  • Experience with DataBase Engines such as PostgreSQL.
  • Familiarity with AWS services related to data (S3, RDS, etc.)
  • Solid understanding of data modeling, data transformation, and data pipeline orchestration.
  • Build and maintain robust, scalable, and reliable ETL pipelines using tools like Matillion, Apache AirFlow and Python, under the guidance of the Data Architect.
  • Implement and monitor data validation rules and automated data checks as defined in the data quality framework.
  • Assist in the integration of new data sources and support cross-team data initiatives in collaboration with Engineering, Product, and Compliance teams.
  • Help set up alerts, dashboards, and automated checks to monitor data pipelines and support auditing activities.
  • Maintain clear and up-to-date documentation for data workflows and tools to support team collaboration and continuity.

AWSNode.jsPostgreSQLPythonSQLApache AirflowETLSnowflakeData engineeringCommunication SkillsRESTful APIsData modelingEnglish communication

Posted about 3 hours ago
Apply
Apply
🔥 Staff Data Engineer
Posted about 4 hours ago

🧭 Full-Time

🏢 Company: Kindred👥 11-50💰 $15,000,000 Series A almost 2 years agoTravel AccommodationsTravel AgencyTravel

  • Have 5+ years of experience building and maintaining data infrastructure at scale
  • Are fluent in modern data tools and cloud-native ecosystems (e.g. dbt, Airflow, Snowflake/BigQuery, Kafka, Spark, etc.)
  • Have experience designing dimensional data models and optimizing warehouse performance
  • Have partnered closely with analytics, product, or ML teams to deliver clean, reliable, well-documented data
  • Have built data platforms from the ground up or been a key contributor in a high-growth startup environment
  • Play a key role in shaping the vision, architecture, and implementation of our end-to-end data infrastructure
  • Partner with engineering, product, and business leaders to understand data needs and influence the long-term data architecture
  • Build scalable, reliable data pipelines that support analytics, operations, and product experiences
  • Design and implement efficient batch and real-time data flows to power reporting, insights, and automation
  • Collaborate with software engineers to instrument key events and ensure data is captured cleanly and consistently
  • Develop well-modeled, intuitive data layers that make it easy for others to query and consume data
  • Implement quality checks, monitoring, and alerts to ensure trust in the data
  • Continuously optimize pipelines and workflows for performance, cost, and maintainability
  • Play a key role in growing our data engineering capabilities as we scale — including mentoring others and setting foundational practices
  • Contribute to incident response and support rotational coverage when needed to maintain data uptime and reliability
Posted about 4 hours ago
Apply
Apply
🔥 Staff Data Engineer
Posted about 5 hours ago

📍 US

🧭 Full-Time

💸 185000.0 - 200000.0 USD per year

🔍 Adtech

  • 8+ years of experience in data engineering.
  • Proven experience building data infrastructure using Spark with Scala.
  • Familiarity with data lakes, cloud warehouses, and storage formats.
  • Strong proficiency in AWS services.
  • Expertise in SQL for data manipulation and extraction.
  • Bachelor's degree in Computer Science or a related field.
  • Design and implement robust data infrastructure using Spark with Scala.
  • Collaborate with our cross-functional teams to design data solutions that meet business needs.
  • Build out our core data pipelines, store data in optimal engines and formats, and feed our machine learning models.
  • Leverage and optimize AWS resources.
  • Collaborate closely with the Data Science team.

AWSSQLCloud ComputingETLMachine LearningData engineeringData scienceSparkScalaData modeling

Posted about 5 hours ago
Apply
Apply
🔥 Data Engineer
Posted about 15 hours ago

💸 120000.0 - 165500.0 USD per year

🔍 Sports Analytics

🏢 Company: Swish Analytics👥 1-10💰 $6,909,110 Series B almost 6 years agoBig DataFantasy SportsPredictive AnalyticsMachine LearningAnalyticsSports

  • BS/BA degree in Mathematics, Computer Science, or related STEM field
  • Minimum of 5+ years of demonstrated experience writing production level code (Python)
  • Proficiency in Python and SQL (preferably MySQL); minimum of 5 years of experience
  • Demonstrated experience with Airflow
  • Demonstrated experience with Kubernetes
  • Experience building end-to-end ETL pipelines
  • Experience utilizing REST APIs
  • Experience with version control (git), continuous integration and deployment, shell scripting, and cloud-computing infrastructures (AWS)
  • Experience with web scraping and cleaning unstructured data
  • Knowledge of data science and machine learning concepts
  • Knowledge of sports betting
  • Must have knowledge and understanding of NBA OR NFL and the ability use your knowledge of the sport to inform your work with complex datasets
  • Architect low-latency, real-time analytics systems including raw data collection, feature development and endpoint production
  • Build new sports betting data products and predictions offerings
  • Integrate large and complex real-time datasets into new consumer and enterprise products
  • Develop production-level predictive analytics into enterprise-grade APIs
  • Support production systems and help triage issues during live sporting events
  • Contribute to the design and implementation of new, fully-automated sports data delivery frameworks
Posted about 15 hours ago
Apply
Apply

📍 AL, AR, AZ, CA (exempt only), CO, CT, FL, GA, ID, IL, IN, IA, KS, KY, MA, ME, MD, MI, MN, MO, MT, NC, NE, NJ, NM, NV, NY, OH, OK, OR, PA, SC, SD, TN, TX, UT, VT, VA, WA, WI

🧭 Full-Time

🔍 Insurance

🏢 Company: Kin Insurance

  • Depth of experience in modern big data environments.
  • Advanced knowledge and experience with SQL and Python is expected.
  • Insurance domain knowledge.
  • Design and develop data pipelines and modeling raw data for downstream ingestion.
  • Mentor and guide data engineers on your team and across the organization, while collaborating with other engineers, product managers, analysts, and stakeholders.
  • Lead a cross-functional project team with members from AppEng, DataEng, BI, and business stakeholders.

AWSPythonSQLApache AirflowETLCross-functional Team LeadershipData engineeringProblem SolvingMentoringDocumentationComplianceData visualizationData modelingData management

Posted about 17 hours ago
Apply
Apply
🔥 Data Engineer
Posted about 18 hours ago

📍 United States

🧭 Contract

🔍 Biotechnology

🏢 Company: Avomind👥 11-50EmploymentHuman ResourcesRecruiting

  • Strong experience in data engineering and cloud platforms (preferably GCP).
  • Proficiency in programming languages like Python, SQL, and shell scripting.
  • Familiarity with data catalog tools (e.g., DataHub, Apache Atlas) and metadata management.
  • Experience with building and maintaining scalable ETL pipelines using orchestration tools (Dagster, Airflow).
  • Understanding of API development and integration.
  • Knowledge of data governance and data quality principles.
  • Background in biological or scientific data is a plus but not mandatory.
  • Design, build, and maintain ETL/ELT pipelines to process and transform data efficiently.
  • Develop and optimise scalable data architectures in the cloud.
  • Implement and maintain data cataloging solutions to ensure discoverability and governance.
  • Build APIs and integrations for seamless data exchange across systems.
  • Perform data quality checks and implement automated testing frameworks to ensure data accuracy and reliability.
  • Collaborate with teams to build self-service systems and promote data democratisation.
  • Document and maintain data engineering processes and best practices

PythonSQLETLGCPAirflowAPI testingData engineeringData modeling

Posted about 18 hours ago
Apply
Apply
🔥 Senior Data Engineer
Posted about 20 hours ago

🔍 Software Development

  • Degree in a related field and 7+ years of data engineering experience.
  • Proficiency in tools and languages, such as AWS, dbt, Snowflake, Git, R, Python, SQL, SQL Server, and Snowflake.
  • Strong project management skills and the ability to communicate complex concepts effectively.
  • Strong analytical skills with the ability to collect, organize, analyze, and disseminate significant amounts of information with attention to detail and accuracy.
  • Adept at conveying complex data insights through a robust understanding of data management systems, warehouse methodologies, data quality standards, data modeling techniques, governance protocols, and advanced analytics.
  • Familiarity with agile work environments and scrum ceremonies.
  • Strong business acumen and experience in aligning data initiatives with business objectives.
  • Contribute to the strategic vision for data engineering and participate in the architectural design and development of new and complex data solutions, focusing on scalability, performance, and hands-on implementation.
  • Design and implement new data systems and infrastructure to ensure the reliability and scalability of data systems by actively contributing to day-to-day engineering tasks.
  • Influence key decisions regarding the data technology stack, infrastructure, and tools while actively engaging in hands-on engineering efforts in the creation and deployment of new data architectures and workflows.
  • Set coding standards and best practices for the Data Engineering & Operations team, conducting and participating in code reviews to maintain high-quality, consistent code.
  • Work closely with database developers, software development, product management, and AI/ML developers to align data initiatives with Assent’s organizational goals.
  • Collaborate with team members to monitor progress, adjust priorities, and meet project deadlines and objectives.
  • Identify opportunities for internal process improvements, including automating manual processes and optimizing data delivery.
  • Proactively support peers in continuous learning by providing technical guidance and training on data engineering development, analysis, and execution.
  • Be familiar with corporate security policies and follow the guidance set out by processes and procedures of Assent.
Posted about 20 hours ago
Apply
Apply
🔥 Data Engineer London or Greece
Posted about 22 hours ago

📍 London, Greece

🔍 Data Science

🏢 Company: VML Enterprise Solutions

  • Proficiency with Python and SQL programming languages.
  • Hands-on experience with cloud platforms like AWS, GCP, or Azure, and familiarity with big data technologies such as Hadoop or Spark.
  • Experience working with relational databases and NoSQL databases.
  • Strong knowledge of data structures, data modelling, and database schema design.
  • Experience in supporting data science workloads and working with both structured and unstructured data.
  • Familiarity with containerization technologies, such as Docker or Kubernetes.
  • Experience with data visualization tools, such as Tableau or Power BI is a plus.
  • Collaborate closely with data scientists, architects, and other stakeholders to understand and implement business requirements.
  • Provide data engineering support for AI model development and deployment, ensuring data scientists have access to the data they need in the format they need it.
  • Implement and optimize data transformations and ETL/ELT processes, using appropriate data engineering tools.
  • Work with a variety of databases and data warehousing solutions to store and retrieve data efficiently.
  • Implement monitoring, troubleshooting, and maintenance procedures for data pipelines to ensure the high quality of data and optimize performance.
  • Participate in the creation and ongoing maintenance of documentation, including data dictionaries, data catalogs, data flow diagrams, and process documentation.

AWSDockerPythonSQLCloud ComputingETLGCPHadoopKubernetesAzureData engineeringData scienceData StructuresNosqlSparkData visualizationData modeling

Posted about 22 hours ago
Apply
Apply
🔥 Data Engineer
Posted 1 day ago

🏢 Company: Portless👥 11-50💰 Series A about 2 months agoLogisticsE-CommerceTransportation

  • 5+ years of experience in data engineering, preferably in supply chain or logistics.
  • Expertise in Google Cloud Platform (GCP), especially BigQuery and Dataflow.
  • Proficiency in Python and JavaScript for serverless data processing.
  • Experience with MLOps and deploying machine learning models.
  • Strong knowledge of ETL/ELT processes, data modeling, and orchestration.
  • Excellent problem-solving skills and ability to work in a fast-paced environment.
  • Design, develop, and maintain data pipelines using BigQuery and Dataflow.
  • Build and manage MLOps workflows to support machine learning models.
  • Architect and implement serverless data solutions using JavaScript and Python.
  • Ensure data quality, integrity, and governance across platforms.
  • Collaborate with cross-functional teams to support analytics, reporting, and operational insights.
  • Provide technical leadership and mentor junior engineers.
  • Stay up to date with emerging data engineering and cloud technologies.
Posted 1 day ago
Apply

Related Articles

Posted about 1 month ago

Why remote work is such a nice opportunity?

Why is remote work so nice? Let's try to see!

Posted 8 months ago

Insights into the evolving landscape of remote work in 2024 reveal the importance of certifications and continuous learning. This article breaks down emerging trends, sought-after certifications, and provides practical solutions for enhancing your employability and expertise. What skills will be essential for remote job seekers, and how can you navigate this dynamic market to secure your dream role?

Posted 8 months ago

Explore the challenges and strategies of maintaining work-life balance while working remotely. Learn about unique aspects of remote work, associated challenges, historical context, and effective strategies to separate work and personal life.

Posted 8 months ago

Google is gearing up to expand its remote job listings, promising more opportunities across various departments and regions. Find out how this move can benefit job seekers and impact the market.

Posted 8 months ago

Learn about the importance of pre-onboarding preparation for remote employees, including checklist creation, documentation, tools and equipment setup, communication plans, and feedback strategies. Discover how proactive pre-onboarding can enhance job performance, increase retention rates, and foster a sense of belonging from day one.