Apply

Data Engineer

Posted about 2 months agoViewed

View full description

💎 Seniority level: Junior, 2 years

📍 Location: United States

🔍 Industry: IT Consulting

🏢 Company: RP Pro Services👥 51-100ConsultingInformation TechnologyWeb Design

🗣️ Languages: English

⏳ Experience: 2 years

Requirements:
  • Proficiency in data management tools such as Excel, Access, or Google Sheets.
  • Experience working with database platforms like SQL or Microsoft SQL Server for querying and data maintenance.
  • Strong organizational skills and attention to detail.
  • Ability to manage multiple data sets and identify inconsistencies.
  • Familiarity with data entry standards, data security practices, and quality control.
  • High school diploma or equivalent; additional technical certifications are a plus.
  • Minimum of 2 years of experience in data entry, data management, or administrative support roles.
  • Strong data entry skills with a proven ability to maintain accuracy and completeness.
  • Must be eligible for a Public Trust Clearance
Responsibilities:
  • Design, develop, and maintain data pipelines to ingest, transform, and load data from various sources.
  • Develop ETL processes to support data integration across platforms and systems.
  • Optimize data pipelines for performance, scalability, and data quality.
  • Collaborate with data scientists, analysts, and stakeholders to understand data needs and ensure efficient data delivery.
  • Implement data validation processes to ensure accuracy, consistency, and completeness.
  • Develop and maintain data models and schemas to support data warehousing solutions.
  • Ensure data security and compliance with VA policies and federal data regulations.
  • Create documentation for data engineering processes, workflows, and best practices.
Apply

Related Jobs

Apply
🔥 Data Engineer Intern
Posted about 4 hours ago

📍 United States

🧭 Internship

💸 30.0 USD per hour

🔍 Streaming Media

🏢 Company: Wurl👥 101-250💰 $11,000,000 Series B over 4 years agoAdvertisingTVVideo StreamingMedia and EntertainmentContent MarketingContent Delivery Network

  • Pursuing or recently completed a Bachelor’s or Master’s degree in Computer Science, Data Science, or a related technical field.
  • Proficiency in Python and SQL.
  • Familiarity with data engineering concepts and tools (e.g., ETL processes, data warehousing).
  • Interest in digital advertising, performance marketing, and streaming media technologies.
  • Strong analytical and problem-solving skills.
  • Excellent communication and collaboration abilities.
  • Collaborate with the Data Engineering team to develop and maintain scalable data pipelines that support performance marketing efforts on CTV.
  • Assist in implementing attribution models that tie CTV viewership to campaigns.
  • Work with cross-functional teams to integrate AI-driven targeting into our data infrastructure.
  • Analyze viewer engagement data to identify opportunities for optimizing return on ad spend (ROAS).
  • Contribute to the development of tools and dashboards that provide insights into campaign performance and audience behavior.

PythonSQLETLData engineering

Posted about 4 hours ago
Apply
Apply
🔥 Data Engineer Intern
Posted about 4 hours ago

📍 United States

🧭 Internship

💸 30.0 USD per hour

🔍 Software Development

🏢 Company: Wurl👥 101-250💰 $11,000,000 Series B over 4 years agoAdvertisingTVVideo StreamingMedia and EntertainmentContent MarketingContent Delivery Network

  • Currently pursuing or recently completed a Bachelor’s or Master’s degree in Computer Science, Data Science, Cybersecurity, or related field.
  • Familiarity with Python and SQL.
  • Understanding of basic cloud concepts (AWS preferred).
  • Strong interest in data security, access control, and automation.
  • Eagerness to learn and contribute to a fast-paced, remote engineering team.
  • Support the implementation and automation of key rotation policies and secure credential storage across our cloud infrastructure.
  • Collaborate with the Data Engineering team to ensure proper access control and encryption practices are enforced in all ETL pipelines and data services.
  • Assist in developing monitoring and alerting systems for key usage and rotation schedules.
  • Contribute to internal documentation and knowledge bases related to key rotation procedures.
  • Learn about and contribute to large-scale, cloud-native data systems.

AWSPythonSQLCloud ComputingCybersecurityETLData engineeringCI/CDTerraform

Posted about 4 hours ago
Apply
Apply

🧭 Contract

🔍 Software Development

🏢 Company: Seamless.AI👥 501-1000💰 $75,000,000 Series A about 4 years agoSales AutomationArtificial Intelligence (AI)Lead GenerationMachine LearningInformation TechnologySoftware

  • Strong proficiency in Python and experience with related libraries and frameworks (e.g., pandas, NumPy, PySpark).
  • Hands-on experience with AWS Glue or similar ETL tools and technologies.
  • Solid understanding of data modeling, data warehousing, and data architecture principles.
  • Expertise in working with large data sets, data lakes, and distributed computing frameworks.
  • Experience developing and training machine learning models.
  • Strong proficiency in SQL.
  • Familiarity with data matching, deduplication, and aggregation methodologies.
  • Experience with data governance, data security, and privacy practices.
  • Strong problem-solving and analytical skills, with the ability to identify and resolve data-related issues.
  • Excellent communication and collaboration skills, with the ability to work effectively independently.
  • Design, develop, and maintain scalable ETL pipelines to acquire, transform, and load data from various sources into the data ecosystem.
  • Work with stakeholders to understand data requirements and propose effective data acquisition and integration strategies.
  • Implement data transformation logic using Python and relevant frameworks, ensuring efficiency and reliability.
  • Utilize AWS Glue or similar tools to create and manage ETL jobs, workflows, and data catalogs.
  • Optimize ETL processes to improve performance and scalability, particularly for large datasets.
  • Apply data matching, deduplication, and aggregation techniques to enhance data accuracy and quality.
  • Ensure compliance with data governance, security, and privacy best practices within the scope of project deliverables.
  • Provide recommendations on emerging technologies and tools that enhance data processing efficiency.
Posted about 4 hours ago
Apply
Apply

🧭 Contract

🔍 Software Development

🏢 Company: Seamless.AI👥 501-1000💰 $75,000,000 Series A about 4 years agoSales AutomationArtificial Intelligence (AI)Lead GenerationMachine LearningInformation TechnologySoftware

  • Strong proficiency in Python and experience with related libraries and frameworks (e.g., pandas, NumPy, PySpark).
  • Hands-on experience with AWS Glue or similar ETL tools and technologies.
  • Solid understanding of data modeling, data warehousing, and data architecture principles.
  • Expertise in working with large data sets, data lakes, and distributed computing frameworks.
  • Experience developing and training machine learning models.
  • Strong proficiency in SQL.
  • Familiarity with data matching, deduplication, and aggregation methodologies.
  • Experience with data governance, data security, and privacy practices.
  • Strong problem-solving and analytical skills, with the ability to identify and resolve data-related issues.
  • Excellent communication and collaboration skills, with the ability to work effectively independently.
  • Highly organized and self-motivated, with the ability to manage multiple projects and priorities simultaneously.
  • Design, develop, and maintain scalable ETL pipelines to acquire, transform, and load data from various sources into the data ecosystem.
  • Work with stakeholders to understand data requirements and propose effective data acquisition and integration strategies.
  • Implement data transformation logic using Python and relevant frameworks, ensuring efficiency and reliability.
  • Utilize AWS Glue or similar tools to create and manage ETL jobs, workflows, and data catalogs.
  • Optimize ETL processes to improve performance and scalability, particularly for large datasets.
  • Apply data matching, deduplication, and aggregation techniques to enhance data accuracy and quality.
  • Ensure compliance with data governance, security, and privacy best practices within the scope of project deliverables.
  • Provide recommendations on emerging technologies and tools that enhance data processing efficiency.
Posted about 4 hours ago
Apply
Apply

🧭 Contract

🔍 Software Development

🏢 Company: Seamless.AI👥 501-1000💰 $75,000,000 Series A about 4 years agoSales AutomationArtificial Intelligence (AI)Lead GenerationMachine LearningInformation TechnologySoftware

  • Strong proficiency in Python and experience with related libraries and frameworks (e.g., pandas, NumPy, PySpark).
  • Hands-on experience with AWS Glue or similar ETL tools and technologies.
  • Solid understanding of data modeling, data warehousing, and data architecture principles.
  • Expertise in working with large data sets, data lakes, and distributed computing frameworks.
  • Experience developing and training machine learning models.
  • Strong proficiency in SQL.
  • Familiarity with data matching, deduplication, and aggregation methodologies.
  • Experience with data governance, data security, and privacy practices.
  • Strong problem-solving and analytical skills, with the ability to identify and resolve data-related issues.
  • Excellent communication and collaboration skills, with the ability to work effectively independently.
  • Highly organized and self-motivated, with the ability to manage multiple projects and priorities simultaneously.
  • Design, develop, and maintain scalable ETL pipelines to acquire, transform, and load data from various sources into the data ecosystem.
  • Work with stakeholders to understand data requirements and propose effective data acquisition and integration strategies.
  • Implement data transformation logic using Python and relevant frameworks, ensuring efficiency and reliability.
  • Utilize AWS Glue or similar tools to create and manage ETL jobs, workflows, and data catalogs.
  • Optimize ETL processes to improve performance and scalability, particularly for large datasets.
  • Apply data matching, deduplication, and aggregation techniques to enhance data accuracy and quality.
  • Ensure compliance with data governance, security, and privacy best practices within the scope of project deliverables.
  • Provide recommendations on emerging technologies and tools that enhance data processing efficiency.
Posted about 4 hours ago
Apply
Apply

🧭 Contract

🔍 Software Development

🏢 Company: Seamless.AI👥 501-1000💰 $75,000,000 Series A about 4 years agoSales AutomationArtificial Intelligence (AI)Lead GenerationMachine LearningInformation TechnologySoftware

  • Strong proficiency in Python and experience with related libraries and frameworks (e.g., pandas, NumPy, PySpark).
  • Hands-on experience with AWS Glue or similar ETL tools and technologies.
  • Solid understanding of data modeling, data warehousing, and data architecture principles.
  • Expertise in working with large data sets, data lakes, and distributed computing frameworks.
  • Experience developing and training machine learning models.
  • Strong proficiency in SQL.
  • Familiarity with data matching, deduplication, and aggregation methodologies.
  • Experience with data governance, data security, and privacy practices.
  • Strong problem-solving and analytical skills, with the ability to identify and resolve data-related issues.
  • Excellent communication and collaboration skills, with the ability to work effectively independently.
  • Highly organized and self-motivated, with the ability to manage multiple projects and priorities simultaneously.
  • Design, develop, and maintain scalable ETL pipelines to acquire, transform, and load data from various sources into the data ecosystem.
  • Work with stakeholders to understand data requirements and propose effective data acquisition and integration strategies.
  • Implement data transformation logic using Python and relevant frameworks, ensuring efficiency and reliability.
  • Utilize AWS Glue or similar tools to create and manage ETL jobs, workflows, and data catalogs.
  • Optimize ETL processes to improve performance and scalability, particularly for large datasets.
  • Apply data matching, deduplication, and aggregation techniques to enhance data accuracy and quality.
  • Ensure compliance with data governance, security, and privacy best practices within the scope of project deliverables.
  • Provide recommendations on emerging technologies and tools that enhance data processing efficiency.
Posted about 4 hours ago
Apply
Apply

🔍 AI Consultancy

  • 5-8 years of data engineering experience with proven expertise in AWS
  • Strong technical leadership capabilities and consulting skills
  • Deep knowledge of modern data engineering principles and best practices
  • Expertise in Python, Java, or Scala with strong system design skills
  • Lead technical implementation of enterprise-scale data solutions using advanced cloud technologies
  • Drive architecture decisions and establish technical standards across client engagements
  • Mentor junior engineers while managing complex technical workstreams
  • Build strong client relationships as a trusted technical advisor
Posted about 4 hours ago
Apply
Apply

📍 Central or South America

🧭 Contract

  • 3+ years of hands-on experience developing and managing data pipelines using PySpark
  • Strong programming skills in Python and SQL
  • Deep understanding of ETL processes and data warehousing fundamentals
  • Design and develop scalable data pipelines
  • Write efficient SQL and Python code
  • Collaborate with machine learning engineers, product managers, and developers to understand data requirements and deliver solutions.
  • Implement and maintain robust ETL processes
  • Ensure data quality, integrity, and reliability
  • Participate in code reviews, troubleshooting, and performance tuning.
  • Work independently and proactively to identify and resolve data-related issues.
  • If applicable, contribute to Azure-based data solutions
  • Support cloud migration initiatives and DevOps practices, if relevant to the role.
  • Provide guidance on best practices and mentor junior team members when needed.

PythonSQLCloud ComputingETLAzureData engineeringCI/CD

Posted about 5 hours ago
Apply
Apply
🔥 Senior Data Engineer
Posted about 5 hours ago

📍 Germany, Spain, Portugal, Greece

🏢 Company: WorkMotion👥 101-250💰 $10,000,000 Series B almost 3 years agoComplianceHuman ResourcesEmployee Benefits

  • 3-5 years of professional experience in Data Engineering or Software Development with a focus on data
  • Strong knowledge of Python and SQL; and PySpark
  • Hands-on experience with AWS services (Glue, S3, Athena, EC2)
  • Experience with Apache Airflow, preferably in a Dockerized/cloud-native environment
  • Familiarity with Delta Lake or similar data lake frameworks
  • Proficiency with source control (GitHub) and CI/CD workflows
  • Strong understanding of data modeling, ETL best practices, and data pipeline performance optimization
  • Design, build, and maintain scalable ETL pipelines using Apache Airflow and AWS Glue (Spark)
  • Work with a range of data sources including Salesforce, NetSuite, PostgreSQL, and MongoDB
  • Develop and optimize PySpark jobs for large-scale data transformation and analytics
  • Manage data lake infrastructure using Delta Lake on S3 with Athena as the query layer
  • Ensure data quality, performance, and reliability through monitoring, testing, and documentation
  • Collaborate with analytics, product, and engineering teams to define data requirements
  • Contribute to CI/CD workflows with GitHub and deployment automation
  • Participate in architectural discussions and advocate for best practices in data engineering

AWSDockerPythonSoftware DevelopmentSQLApache AirflowETLGitData engineeringSparkCI/CDData modeling

Posted about 5 hours ago
Apply
Apply
🔥 Sr. Data Engineer
Posted about 6 hours ago

🧭 Full-Time

💸 100000.0 - 200000.0 USD per year

🔍 Software Development

🏢 Company: Risk Labs👥 11-50💰 almost 5 years agoCryptocurrencyFinance

  • 4+ years full-time experience in Data Engineering at a software, financial services, or related technology organization
  • 2+ years of experience with databases and data pipelines
  • Passion for blockchain technology, cryptonomic protocol design, game theory, and decentralized finance
  • Strong communication and collaboration skills, especially written
  • Indexing complex onchain data
  • Design and develop large-scale data pipelines
  • Write code in a highly collaborative open source ecosystem. As much of your day will be spent reviewing and designing code as it will be writing new code. We take pride in writing clean code.
Posted about 6 hours ago
Apply

Related Articles

Posted about 1 month ago

How to Overcome Burnout While Working Remotely: Practical Strategies for Recovery

Burnout is a silent epidemic among remote workers. The blurred lines between work and home life, coupled with the pressure to always be “on,” can leave even the most dedicated professionals feeling drained. But burnout doesn’t have to define your remote work experience. With the right strategies, you can recover, recharge, and prevent future episodes. Here’s how.



Posted 5 days ago

Top 10 Skills to Become a Successful Remote Worker by 2025

Remote work is here to stay, and by 2025, the competition for remote jobs will be tougher than ever. To stand out, you need more than just basic skills. Employers want people who can adapt, communicate well, and stay productive without constant supervision. Here’s a simple guide to the top 10 skills that will make you a top candidate for remote jobs in the near future.

Posted 9 months ago

Google is gearing up to expand its remote job listings, promising more opportunities across various departments and regions. Find out how this move can benefit job seekers and impact the market.

Posted 10 months ago

Read about the recent updates in remote work policies by major companies, the latest tools enhancing remote work productivity, and predictive statistics for remote work in 2024.

Posted 10 months ago

In-depth analysis of the tech layoffs in 2024, covering the reasons behind the layoffs, comparisons to previous years, immediate impacts, statistics, and the influence on the remote job market. Discover how startups and large tech companies are adapting, and learn strategies for navigating the new dynamics of the remote job market.