Hadoop Jobs

Find remote positions requiring Hadoop skills. Browse through opportunities where you can utilize your expertise and grow your career.

Hadoop
81 jobs found. to receive daily emails with new job openings that match your preferences.
81 jobs found.

Set alerts to receive daily emails with new job openings that match your preferences.

Apply

📍 Germany, India

🧭 Full-Time

💸 195600.0 - 293400.0 USD per year

🔍 Customer relationship management (CRM)

🏢 Company: HubSpot👥 1001-5000💰 $35,000,000 Series E about 12 years ago🫂 Last layoff 12 months agoSaaSAnalyticsMarketingCopywritingSocial Media

  • Strong engineers interested in writing in the Java ecosystem.
  • Understanding that the job is not about implementing pre-written requirements, but collaboratively figuring out needed solutions.
  • Consistently deliver reliable, performant, well-tested code.
  • Technical challenges with big data and distributed systems, leveraging technologies such as HBase, Kafka, Hadoop, Elasticsearch, GraphQL, and Dropwizard.
  • Demonstrate pragmatic decision-making and problem-solving abilities.
  • Have a customer-centric attitude and develop opinions about the HubSpot product.
  • Understand the value of building a diverse team and champion Diversity, Inclusion, and Belonging initiatives.

  • Design, scope, and deliver solutions to help our customers manage and grow their businesses.
  • Manage a small team of engineers while still making significant individual contributions.
  • Foster an environment that encourages autonomy and psychological safety.
  • Enjoy mentoring, coaching, and delivering feedback to other team members in a psychologically safe way.

Backend DevelopmentGraphQLElasticSearchHadoopJavaKafkaMicroservices

Posted 3 days ago
Apply
Apply
🔥 Senior Data Scientist
Posted 11 days ago

📍 United States

🧭 Full-Time

🔍 Consulting

🏢 Company: ICA.ai

  • Education: Bachelor’s or Master’s degree in Statistics, Mathematics, Computer Science, Data Science, or a related quantitative field.
  • Experience: 7 to 8 years of overall professional experience in data science, analytics, or a related field; at least 2–3 years specifically focused on LLMs.
  • Technical Proficiency: Coding experience with Python, C, R, Java, or Scala; statistical computer languages for data manipulation and model development; knowledge of ML techniques.
  • Familiarity with AWS services for scalable model development.
  • Preferred: Exposure to MLOps tools, big data technologies, or other cloud services.
  • Soft Skills: Excellent written and verbal communication skills; strong problem-solver; collaborative mindset.

  • Data Analysis & Modeling: Design advanced analytical models and establish best practices.
  • Client-Facing Presentations: Present complex data insights to diverse audiences, including executive stakeholders.
  • Predictive Modeling & Optimization: Develop models for predictive analytics and optimization.
  • Model Monitoring: Monitor the performance of models to ensure effectiveness.
  • Identify Opportunities: Identify and propose opportunities for innovative solutions.
  • Innovation & Strategy: Drive organizational transformation through advanced analytics.
  • LLM-Focused Development: Develop solutions utilizing large language models and related techniques.

AWSPythonSQLData AnalysisData MiningHadoopMachine LearningSpark

Posted 11 days ago
Apply
Apply

📍 US

🧭 Full-Time

🔍 Creator commerce

  • Proven leadership skills including mentoring and guiding engineers.
  • Project management skills with the ability to lead projects from conception to completion.
  • Commitment to continuous learning and staying updated with industry trends.
  • Technical skills with more than 7 years of experience in distributed systems and software development.
  • A bachelor's or master's degree in Computer Science, Software Engineering, or a related field preferred.
  • Expertise in Golang, Python, Vector Databases, and related technologies.

  • Lead the Building and Scaling of a Product Catalog Service.
  • Lead the Building and Scaling of an Exact Product Matching Service.
  • Lead the Building and Scaling of a Product Search Service.
  • Provide inputs on engineering excellence and complete critical engineering projects.
  • Innovate solutions to complex technical challenges and influence technical direction.
  • Create and maintain technical documentation and project plans.
  • Take ownership of complex issues and drive problem resolution.

AWSDockerPythonETLHadoopKubernetesApache KafkaData modeling

Posted 13 days ago
Apply
Apply

📍 Colombia, Spain, Ecuador, Venezuela, Argentina

🔍 HR Tech

🏢 Company: Jobgether👥 11-50💰 $1,493,585 Seed almost 2 years agoInternet

  • Bachelor's or Master's degree in Computer Science, Engineering, or a related technical field.
  • Minimum of 4 years of experience in data engineering.
  • Proficiency in Python with at least 4 years of experience.
  • Hands-on experience with big data technologies like Hadoop, Spark, or Kafka.
  • Proficiency in relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases like MongoDB.
  • Experience with AWS cloud platforms.
  • Strong understanding of data modeling, schema design, and data warehousing concepts.
  • Excellent analytical and troubleshooting skills.
  • Fluency in English and Spanish.

  • Design, build, and maintain scalable data pipelines and ETL processes for large volumes of data.
  • Develop and optimize data scraping solutions for efficient job data extraction.
  • Collaborate with data scientists on AI-driven matching algorithms.
  • Ensure data integrity and reliability through validation mechanisms.
  • Analyze and optimize system performance, addressing challenges.
  • Work with teams to efficiently deploy machine learning models.
  • Stay updated on emerging technologies and best practices.
  • Collaborate with various teams to understand and implement data requirements.
  • Maintain documentation for systems and processes, ensuring compliance.

AWSDockerPostgreSQLPythonETLHadoopKafkaKubernetesMachine LearningMongoDBMySQLSparkData modeling

Posted 17 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 18 days ago

📍 Brazil

🧭 Full-Time

🔍 Technology in government affairs

  • Bachelor's or Master's degree in Computer Science, Engineering, or a related field.
  • 5+ years in data engineering and development of data-driven products.
  • Expertise in building data pipelines and architectures; AWS services including EC2, EMR, RDS, Redshift.
  • Proficient in big data tools (Hadoop, Spark, Kafka) and machine learning frameworks (TensorFlow, PyTorch).
  • 3+ years experience with Python.
  • Deep knowledge of SQL and NoSQL databases, and workflow management tools.

  • Architect and implement scalable RAG data pipelines for legislative bills, social media, documents, and testimonies.
  • Design data pipelines for real-time processing and analysis.
  • Develop data cleansing and transformation processes for AI products.
  • Oversee cloud-based deployments in AWS, focusing on performance, security, and cost-efficiency.
  • Innovate data architecture to meet dynamic product needs.

AWSPythonSQLApache AirflowETLHadoopKafkaMachine LearningPyTorchNosqlSparkTensorflow

Posted 18 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 164621.0 - 205776.0 USD per year

🔍 Data movement and analytics technology

🏢 Company: Fivetran

  • Deep understanding of GSI and Reseller pre-sales motions and the related technical teams.
  • Experience with GSI and Reseller sales methodologies and complex sales cycles.
  • Knowledge of technical enablement approaches, training programs, and maintaining technical excellence.
  • Hands-on experience with cloud data platforms, data lakes, data movement technologies, and applications.
  • Familiarity with programming languages like Python and SQL.

  • Create, develop, and build ongoing relationships with GSI and Reseller pre-sales teams.
  • Interact with top GSI and Reseller partners, tracking KPIs related to partner enablement and technical content creation.
  • Represent Fivetran as the technical liaison and collaborate with sales managers to clarify Fivetran's value proposition.
  • Assist sales managers in communicating Fivetran’s value to existing and prospective partners.
  • Engage in hands-on activities involving Fivetran and the broader data ecosystem.

PythonSQLCloud ComputingETLHadoopMachine LearningMongoDBOracleSnowflakeData engineeringREST API

Posted 23 days ago
Apply
Apply

📍 US

🧭 Full-Time

💸 164621.0 - 205776.0 USD per year

🔍 Data Movement and Management

  • Deep understanding of GSI and Reseller pre-sales processes.
  • Experience with sales methodologies and managing complex sales cycles.
  • Knowledge of technical enablement for GSI and Reseller efficiency.
  • Hands-on experience with cloud data platforms and data lakes.
  • Familiarity with data movement technologies and databases.
  • Technical skills in programming and data transformation.

  • Create, develop, build, and take ownership of ongoing relationships with GSI and Reseller pre-sales technical team members.
  • Interact daily with Fivetran’s top GSI and Reseller partners and track key performance indicators.
  • Represent Fivetran as the partner technical liaison and collaborate with sales managers.
  • Assist in conveying Fivetran’s value proposition to partners.
  • Participate in hands-on technical demonstrations and labs.

PythonSQLETLHadoopMicrosoft SQL ServerMongoDBOracleSnowflakeData engineeringREST API

Posted 24 days ago
Apply
Apply

📍 Armenia

🧭 Full-Time

🔍 Data Engineering

  • Proven enterprise experience in Python.
  • Proficiency in cloud platforms such as Azure, AWS, or GCP.
  • Advanced command of Linux with skills in service configuration and scripting.
  • Fundamental understanding of containerization technologies like Docker and Kubernetes.
  • Solid database skills, particularly in SQL and relational algebra.
  • Basic networking knowledge.

  • Develop and maintain scalable data pipelines and API integrations to manage data volume and complexity.
  • Collaborate with analytics and business teams to enhance data models for business intelligence tools.
  • Implement systems to monitor data quality, ensuring accuracy for stakeholders.
  • Design architecture for data storage solutions focused on retrieval, security, and performance.
  • Write high-quality documented code for data infrastructure components.
  • Act as a mentor to junior data engineers.

AWSDockerPythonSQLETLGCPHadoopKafkaKubernetesAzureNosqlSparkLinuxTerraform

Posted 25 days ago
Apply
Apply

📍 Argentina, Uruguay, Peru, Paraguay, Colombia

🧭 Full-Time

🔍 Interactive media strategy and development

🏢 Company: Halo Media👥 11-50InternetConsultingWeb DevelopmentAppsMarketingMobileWeb DesignSoftware

  • Minimum of 5 years of professional experience in R development.
  • Bachelor’s or Master’s degree in Computer Science, Data Science, Statistics, or a related field.
  • Proficiency in R programming language and its core packages (dplyr, ggplot2, tidyr, etc.).
  • Strong understanding of statistical methods and data analysis techniques.
  • Experience with version control systems (e.g., Git).
  • Familiarity with database management and SQL.
  • Knowledge of other programming languages (e.g., Python) is a plus.
  • Excellent problem-solving skills with the ability to interpret complex data and provide actionable insights.
  • Strong verbal and written communication skills.
  • Comfortable exploring and adopting new libraries or frameworks as needs demand.
  • Eager to participate in innovation towards new solutions to existing problems.
  • Ability to work in a fast-paced environment and manage multiple projects simultaneously.

  • Manage and execute advanced data analysis and software development projects using R, ensuring timely delivery and high-quality results.
  • Conduct sophisticated statistical analyses and modeling to extract insights from complex data sets.
  • Develop and maintain data processing pipelines, ensuring data integrity and accuracy.
  • Write, test, and deploy efficient R code for data analysis, visualization, and reporting.
  • Provide guidance and mentorship to junior R developers, promoting best practices and fostering a collaborative team environment.
  • Work closely with data scientists, analysts, and other stakeholders to understand their needs and develop appropriate solutions.
  • Conduct code reviews to ensure code quality, performance, and adherence to best practices.
  • Create and maintain comprehensive documentation for developed code and methodologies.
  • Stay updated with the latest developments in R programming and data science to bring innovative solutions to the team.

SQLGitHadoopMachine LearningSparkData visualizationData management

Posted 26 days ago
Apply
Apply
🔥 Head of Data Product
Posted about 1 month ago

📍 USA

🔍 Deal sourcing platform

🏢 Company: Sourcescrub

  • 5+ years of experience in Product Management for data-related products.
  • Direct people and process management experience.
  • Extensive knowledge of the investing/private equity industry.
  • Experience designing research and data collection processes.
  • Strong Excel skills, experience with SQL and data modeling.
  • Knowledge of ontology development and data governance frameworks.
  • Familiarity with big data technologies like Hadoop and Spark.
  • Experience with cloud platforms such as AWS, Azure, or Google Cloud.
  • Understanding of data privacy regulations like GDPR and CCPA.

  • Own the strategy, development, execution, and budgeting for global data products.
  • Lead the Product Owner role for Data Labeling, Data Platform, and Data Automation Ingestion.
  • Engage in the development and optimization of raw and derived data products.
  • Collaborate with Go-to-Market teams on data accuracy, coverage, and freshness.
  • Oversee data acquisition from diverse sources.
  • Design innovative research methodologies ensuring accuracy and efficiency.
  • Identify patterns in data to support strategic planning.
  • Implement data quality measurement practices and conduct regular audits.
  • Document research methodologies to facilitate future initiatives.
  • Work with engineering teams on the data product roadmap.
  • Enhance research operations with new tools and explore new data sources.
  • Anticipate client needs with research solutions.
  • Contribute to the overall research strategy for SourceScrub.

AWSSQLData AnalysisHadoopProduct ManagementSparkData modeling

Posted about 1 month ago
Apply
Shown 10 out of 81