Apply

Data Engineer

Posted 16 days agoViewed

View full description

💎 Seniority level: Middle, 3+ years

🔍 Industry: Data and Analytics

🏢 Company: Antenna👥 1-10InternetMusicMusic Streaming

🗣️ Languages: English

⏳ Experience: 3+ years

Requirements:
  • Proficient in Python, YAML, and SQL
  • Experience with Google Cloud Platform
  • Experience with distributed data processing and data modeling concepts
  • Web Application development experience. Experience with Django is a plus
Responsibilities:
  • Monitor, maintain, and improve scalable data pipelines
  • Proactively identify and troubleshoot pipeline failures, and develop corrective action plans when necessary
  • Support our world-class Data Operations by improving and optimizing our data pipelines
  • Collaborate with cross-functional teams to ensure clear communication and efficient delivery of our data
  • Update and manage critical configuration files (YAML)
  • Contribute to the documentation of processes and procedures
  • Integrate our systems with various data sources
  • Continuously improve our data quality through automated validation and anomaly detection.
Apply

Related Jobs

Apply
🔥 Data Engineer
Posted 27 minutes ago

🏢 Company: Portless

  • 5+ years of experience in data engineering, preferably in supply chain or logistics.
  • Expertise in Google Cloud Platform (GCP), especially BigQuery and Dataflow.
  • Proficiency in Python and JavaScript for serverless data processing.
  • Experience with MLOps and deploying machine learning models.
  • Strong knowledge of ETL/ELT processes, data modeling, and orchestration.
  • Excellent problem-solving skills and ability to work in a fast-paced environment.
  • Design, develop, and maintain data pipelines using BigQuery and Dataflow.
  • Build and manage MLOps workflows to support machine learning models.
  • Architect and implement serverless data solutions using JavaScript and Python.
  • Ensure data quality, integrity, and governance across platforms.
  • Collaborate with cross-functional teams to support analytics, reporting, and operational insights.
  • Provide technical leadership and mentor junior engineers.
  • Stay up to date with emerging data engineering and cloud technologies.
Posted 27 minutes ago
Apply
Apply

📍 India

🔍 Software Development

🏢 Company: Jobgether👥 11-50💰 $1,493,585 Seed about 2 years agoInternet

  • Hands-on experience with GIS, location-based data ingestion pipelines, and AWS services such as EC2, S3, and Lambda.
  • Proficiency in Python or Java for orchestration of data pipelines.
  • Strong experience in writing analytical queries using SQL.
  • Familiarity with Airflow, Docker, and version control with Git.
  • Design and maintain data ingestion pipelines that integrate and process large datasets from multiple sources.
  • Build infrastructure for ETL (Extract, Transform, Load) processes, utilizing AWS technologies such as EC2, S3, EMR, and Lambda.
  • Collaborate with Product, Analytics, and Client Services teams to resolve data-related technical issues and ensure data infrastructure needs are met.
  • Write and optimize SQL queries to extract and analyze data effectively.
  • Participate in code reviews, ensure quality control, and test applications before deployment.
  • Contribute to the improvement of the location-based platform by proposing and implementing innovative solutions.

AWSDockerPythonSQLCloud ComputingETLGitJavaAirflowData engineering

Posted about 12 hours ago
Apply
Apply
🔥 GCP Data Engineer
Posted about 14 hours ago

🧭 Contract

🏢 Company: Kaizen Analytix👥 11-50Information ServicesAnalyticsSoftware

  • 4 to 6 years of experience
  • GCP services
  • DevOps work – Building Ci/CD Pipeline with Jenkins & Gitlab.
  • Configuring Google Cloud Platform services
  • Managing data storage and processing
  • Designing and deploying data pipelines using GCP services
  • Developing data ingestion and transformation processes
  • Establishing and managing data storage solutions using GCP services.
  • DevOps work – Building Ci/CD Pipeline with Jenkins & Gitlab.
Posted about 14 hours ago
Apply
Apply
🔥 Sr. Data Engineer (GC25001)
Posted about 20 hours ago

📍 United States

🧭 Full-Time

💸 150363.0 - 180870.0 USD per year

🔍 Software Development

  • At least a Bachelors Degree or foreign equivalent in Computer Science, Computer Engineering, Electrical and Electronics Engineering, or a closely related technical field, and at least five (5) years of post-bachelor’s, progressive experience writing shell scripts; validating data; and engaging in data wrangling.
  • Experience must include at least three (3) years of experience debugging data; transforming data into Microsoft SQL server; developing processes to import data into HDFS using Sqoop; and using Java, UNIX Shell Scripts, and Python.
  • Experience must also include at least one (1) year of experience developing Hive scripts for data transformation on data lake projects; converting Hive scripts to Pyspark applications; automating in Hadoop; and implementing CI/CD pipelines.
  • Design, develop, test, and implement Big Data technical solutions.
  • Recommend the right technologies and solutions for a given use case, from the application layer to infrastructure.
  • Lead the delivery of compiling and installing database systems, integrating data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures.
  • Drive solution architecture and perform deployments of data pipelines and applications.
  • Author DDL and DML SQL spanning technical tacks.
  • Develop data transformation code and highly complex provisioning pipelines.
  • Ingest data from relational databases.
  • Execute automation strategy.

AWSPythonSQLHadoopJavaKafkaSnowflakeData engineeringSparkCI/CDScalaScriptingDebugging

Posted about 20 hours ago
Apply
Apply
🔥 Data Engineer-I
Posted 1 day ago

📍 USA

🔍 Healthcare

🏢 Company: Innovaccer Inc.

  • SQL knowledge
  • ETL/ELT/Data pipeline knowledge
  • Python knowledge
  • Powershell / Bash knowledge
  • Excellent problem-solving and effective communication skills
  • Self-motivation, integrity and honesty
  • Collaborate with team, management, departments using virtual tools
  • Run Production data pipelines/processes, ensure the integrity of the data, and send out deliverables based on requirement/runbook documentation
  • Coordinate with the various technical teams to resolve issues/bugs/optimize said production processes
  • Coordinate with internal client facing team members to communicate the status of deliverables
  • Help develop/improve technical documentation to guide future software development projects and operations
  • Dedicated time to explore building out tech stack and capabilities where there are applicable use cases
  • Provide critical thinking, technical innovation, and extra attention to detail by serving as a trusted team member and peer code reviewer
  • Assists with external client communications when deliverables or receivables do not meet technical or project requirements, ensuring timely resolution and alignment

PythonSQLBashETLMicrosoft AzurePostgresData modeling

Posted 1 day ago
Apply
Apply

🧭 Full-Time

🔍 Consulting

🏢 Company: P3 Adaptive

  • US Citizenship or Green Card (We don’t sponsor work visas)
  • Strong written and spoken English
  • Proven time management skills
  • Proven ability to connect with a diverse range of technical and non-technical stakeholders
  • Experienced in Project Management
  • Intermediate or better knowledge of T-SQL for DDL and DML applications.
  • Experience with Azure Active Directory Security Groups and Role-Based Access Controls
  • Experience with SSIS, SSAS preferred
  • Experience with PowerShell and Python preferred
  • Insatiable curiosity and love of learning
  • Support the execution of Power BI projects, working alongside expert Principal Consultants and Solution Architects.
  • Create Data Storage Solutions with SQL Server and Data Lakes.
  • Develop ETL Pipelines with Azure Data Factory.
  • Provision Azure Subscriptions and Resources.
  • Develop Automation Solutions using languages such as PowerShell and Python
Posted 1 day ago
Apply
Apply

📍 Germany, Italy, Netherlands, Portugal, Romania, Spain, UK

🧭 Full-Time

🔍 Wellness

  • You have a proven track record of designing and building robust, scalable, and maintainable data models and corresponding pipelines from business requirements.
  • You are skilled at engaging with engineering and product teams to elicit requirements.
  • You are comfortable with big data concepts, ensuring data is efficiently ingested, processed, and made available for data scientists, business analysts, and product teams.
  • You are experienced in maintaining data consistency across the entire data ecosystem.
  • You have experience maintaining and debugging data pipelines in production environments with high criticality, ensuring reliability and performance.
  • Develop and maintain efficient and scalable data models and structures to support analytical workloads.
  • Design, develop, and maintain data pipelines that transform and process large volumes of data while embedding business context and semantics.
  • Implement automated data quality checks to ensure consistency, accuracy, and reliability of data.
  • Ensure correct adoption and usage of Wellhub’s data by data practitioners across the company
  • Live the mission: inspire and empower others by genuinely caring for your own wellbeing and your colleagues. Bring wellbeing to the forefront of work, and create a supportive environment where everyone feels comfortable taking care of themselves, taking time off, and finding work-life balance.

SQLApache AirflowKubernetesApache KafkaData engineeringSparkData modeling

Posted 2 days ago
Apply
Apply

📍 Portugal

🧭 Full-Time

🏢 Company: Wellhub

  • Proven track record of designing and building robust, scalable, and maintainable data models and corresponding pipelines from business requirements.
  • Skilled at engaging with engineering and product teams to elicit requirements.
  • Comfortable with big data concepts, ensuring data is efficiently ingested, processed, and made available for data scientists, business analysts, and product teams.
  • Experienced in maintaining data consistency across the entire data ecosystem.
  • Experience maintaining and debugging data pipelines in production environments with high criticality, ensuring reliability and performance.
  • Motivated to contribute to a data-driven culture and take pride in seeing the impact of your work across the company
  • Develop and maintain efficient and scalable data models and structures to support analytical workloads.
  • Design, develop, and maintain data pipelines that transform and process large volumes of data while embedding business context and semantics.
  • Implement automated data quality checks to ensure consistency, accuracy, and reliability of data.
  • Ensure correct adoption and usage of Wellhub’s data by data practitioners across the company
  • Live the mission: inspire and empower others by genuinely caring for your own wellbeing and your colleagues. Bring wellbeing to the forefront of work, and create a supportive environment where everyone feels comfortable taking care of themselves, taking time off, and finding work-life balance.

SQLApache AirflowETLKubernetesApache KafkaData engineeringSparkData visualizationData modelingData analyticsData management

Posted 2 days ago
Apply
Apply

🧭 Full-Time

🔍 E-Learning

🏢 Company: Truelogic👥 101-250ConsultingWeb DevelopmentWeb DesignSoftware

  • 3-5 years of experience working with PySpark and Apache Spark in Big Data environments.
  • Experience with SQL and relational and NoSQL databases (PostgreSQL, MySQL, MongoDB, etc.).
  • Knowledge of ETL processes and data processing in distributed environments.
  • Familiarity with Apache Hadoop, Hive, or Delta Lake.
  • Experience with cloud storage (AWS S3, Google Cloud Storage, Azure Blob).
  • Proficiency in Git and version control.
  • Strong problem-solving skills and a proactive attitude.
  • A passion for learning and continuous improvement.
  • Design, develop, and optimize data pipelines using PySpark and Apache Spark.
  • Integrate and process data from multiple sources (databases, APIs, files, streaming).
  • Implement efficient data transformations for Big Data in distributed environments.
  • Optimize code to improve performance, scalability, and efficiency in data processing.
  • Collaborate with Data Science, BI, and DevOps teams to ensure seamless integration.
  • Monitor and debug data processes to ensure quality and reliability.
  • Apply best practices in data engineering and maintain clear documentation.
  • Stay up to date with the latest trends in Big Data and distributed computing.
Posted 4 days ago
Apply
Apply

🔍 Health & Bioinformatics

🏢 Company: Gradient AI👥 101-250💰 $20,000,000 Series B almost 4 years agoArtificial Intelligence (AI)Machine LearningInsurTechInsuranceHealth Care

  • 5+ years of relevant working experience, with a significant portion focused on healthcare data.
  • Proven experience working with and interpreting health, medical, and bioinformatics data is required, including experience with real-world healthcare datasets.
  • Expertise as a subject matter expert (SME) in health and bioinformatics data, with a deep understanding of the nuances and challenges associated with processing medical and bioinformatics data, and a strong understanding of the healthcare industry.
  • Experience working in Python in a professional environment, ideally in a healthcare or life sciences setting.
  • Desire to learn new skills and tools (e.g., Redshift, Tableau, AWS Lambda, etc.); bonus for experience with healthcare-specific data analysis and visualization tools.
  • Design, build, and implement data systems that fuel our ML and AI models for our health insurance clients, ensuring compliance with healthcare data privacy and security regulations (e.g., HIPAA).
  • Develop tools to extract and process diverse healthcare data sources, including electronic health records (EHRs), medical claims, pharmacy data, and genomic data, and create tools to profile and validate data.
  • Work cross-functionally with data scientists to transform large amounts of health-related and bioinformatics data and store it in a format that facilitates modeling, paying close attention to data quality and integrity in the context of healthcare applications.
  • Contribute to production operations, data pipelines, workflow management, reliability engineering, and more, with an understanding of the critical nature of data reliability in healthcare settings.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a variety of sources using SQL and AWS ‘big data’ technologies, including experience with healthcare-specific data warehousing and analytics platforms.
  • Leverage expertise as a health and bioinformatics SME to ensure that data pipelines align with the specific requirements of health, medical, and bioinformatics data processing, including the ability to translate complex medical and biological concepts into data requirements.
Posted 4 days ago
Apply

Related Articles

Posted about 1 month ago

Why remote work is such a nice opportunity?

Why is remote work so nice? Let's try to see!

Posted 7 months ago

Insights into the evolving landscape of remote work in 2024 reveal the importance of certifications and continuous learning. This article breaks down emerging trends, sought-after certifications, and provides practical solutions for enhancing your employability and expertise. What skills will be essential for remote job seekers, and how can you navigate this dynamic market to secure your dream role?

Posted 8 months ago

Explore the challenges and strategies of maintaining work-life balance while working remotely. Learn about unique aspects of remote work, associated challenges, historical context, and effective strategies to separate work and personal life.

Posted 8 months ago

Google is gearing up to expand its remote job listings, promising more opportunities across various departments and regions. Find out how this move can benefit job seekers and impact the market.

Posted 8 months ago

Learn about the importance of pre-onboarding preparation for remote employees, including checklist creation, documentation, tools and equipment setup, communication plans, and feedback strategies. Discover how proactive pre-onboarding can enhance job performance, increase retention rates, and foster a sense of belonging from day one.