Hadoop Jobs

Find remote positions requiring Hadoop skills. Browse through opportunities where you can utilize your expertise and grow your career.

Hadoop
78 jobs found. to receive daily emails with new job openings that match your preferences.
78 jobs found.

Set alerts to receive daily emails with new job openings that match your preferences.

Apply

📍 Germany, India

🧭 Full-Time

💸 195600.0 - 293400.0 USD per year

🔍 Customer relationship management (CRM)

🏢 Company: HubSpot👥 1001-5000💰 $35,000,000 Series E about 12 years ago🫂 Last layoff 12 months agoSaaSAnalyticsMarketingCopywritingSocial Media

  • Strong engineers interested in writing in the Java ecosystem.
  • Understanding that the job is not about implementing pre-written requirements, but collaboratively figuring out needed solutions.
  • Consistently deliver reliable, performant, well-tested code.
  • Technical challenges with big data and distributed systems, leveraging technologies such as HBase, Kafka, Hadoop, Elasticsearch, GraphQL, and Dropwizard.
  • Demonstrate pragmatic decision-making and problem-solving abilities.
  • Have a customer-centric attitude and develop opinions about the HubSpot product.
  • Understand the value of building a diverse team and champion Diversity, Inclusion, and Belonging initiatives.

  • Design, scope, and deliver solutions to help our customers manage and grow their businesses.
  • Manage a small team of engineers while still making significant individual contributions.
  • Foster an environment that encourages autonomy and psychological safety.
  • Enjoy mentoring, coaching, and delivering feedback to other team members in a psychologically safe way.

Backend DevelopmentGraphQLElasticSearchHadoopJavaKafkaMicroservices

Posted 7 days ago
Apply
Apply
🔥 Senior Data Scientist
Posted 15 days ago

📍 United States

🧭 Full-Time

🔍 Consulting

🏢 Company: ICA.ai

  • Education: Bachelor’s or Master’s degree in Statistics, Mathematics, Computer Science, Data Science, or a related quantitative field.
  • Experience: 7 to 8 years of overall professional experience in data science, analytics, or a related field; at least 2–3 years specifically focused on LLMs.
  • Technical Proficiency: Coding experience with Python, C, R, Java, or Scala; statistical computer languages for data manipulation and model development; knowledge of ML techniques.
  • Familiarity with AWS services for scalable model development.
  • Preferred: Exposure to MLOps tools, big data technologies, or other cloud services.
  • Soft Skills: Excellent written and verbal communication skills; strong problem-solver; collaborative mindset.

  • Data Analysis & Modeling: Design advanced analytical models and establish best practices.
  • Client-Facing Presentations: Present complex data insights to diverse audiences, including executive stakeholders.
  • Predictive Modeling & Optimization: Develop models for predictive analytics and optimization.
  • Model Monitoring: Monitor the performance of models to ensure effectiveness.
  • Identify Opportunities: Identify and propose opportunities for innovative solutions.
  • Innovation & Strategy: Drive organizational transformation through advanced analytics.
  • LLM-Focused Development: Develop solutions utilizing large language models and related techniques.

AWSPythonSQLData AnalysisData MiningHadoopMachine LearningSpark

Posted 15 days ago
Apply
Apply

📍 US

🧭 Full-Time

🔍 Creator commerce

  • Proven leadership skills including mentoring and guiding engineers.
  • Project management skills with the ability to lead projects from conception to completion.
  • Commitment to continuous learning and staying updated with industry trends.
  • Technical skills with more than 7 years of experience in distributed systems and software development.
  • A bachelor's or master's degree in Computer Science, Software Engineering, or a related field preferred.
  • Expertise in Golang, Python, Vector Databases, and related technologies.

  • Lead the Building and Scaling of a Product Catalog Service.
  • Lead the Building and Scaling of an Exact Product Matching Service.
  • Lead the Building and Scaling of a Product Search Service.
  • Provide inputs on engineering excellence and complete critical engineering projects.
  • Innovate solutions to complex technical challenges and influence technical direction.
  • Create and maintain technical documentation and project plans.
  • Take ownership of complex issues and drive problem resolution.

AWSDockerPythonETLHadoopKubernetesApache KafkaData modeling

Posted 18 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 22 days ago

📍 Brazil

🧭 Full-Time

🔍 Government affairs technology

  • Bachelor's or Master's degree in Computer Science, Engineering, or related field.
  • 5+ years in data engineering with a proven track record.
  • Expertise in building data pipelines and architectures.
  • Experience in AWS cloud services (EC2, EMR, RDS, Redshift).
  • Proficient in big data tools (Hadoop, Spark, Kafka) and machine learning frameworks (TensorFlow, PyTorch).
  • 3+ years experience with Python.
  • Deep knowledge of SQL and NoSQL databases, workflow management tools (Azkaban, Luigi, Airflow).
  • Understanding of the machine learning model deployment cycle.
  • Experience with vector databases and RAG systems (Langchain, Pinecone, OpenAI/ChatGPT) is a plus.

  • Architect and implement highly scalable advanced Retrieval-Augmented Generation (RAG) data pipelines.
  • Design robust data pipelines for real-time processing and analysis of vast datasets.
  • Design and implement data cleansing and transformation pipelines.
  • Lead cloud-based deployments in AWS ensuring performance and security.
  • Innovate on data architecture for Quorum Copilot's evolving needs.
  • Drive build vs buy, tool selection, and analysis using engineering principles.

AWSPythonSQLApache AirflowETLHadoopKafkaMachine LearningPyTorchNosqlSparkTensorflow

Posted 22 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 164621.0 - 205776.0 USD per year

🔍 Data movement and analytics technology

🏢 Company: Fivetran

  • Deep understanding of GSI and Reseller pre-sales motions and the related technical teams.
  • Experience with GSI and Reseller sales methodologies and complex sales cycles.
  • Knowledge of technical enablement approaches, training programs, and maintaining technical excellence.
  • Hands-on experience with cloud data platforms, data lakes, data movement technologies, and applications.
  • Familiarity with programming languages like Python and SQL.

  • Create, develop, and build ongoing relationships with GSI and Reseller pre-sales teams.
  • Interact with top GSI and Reseller partners, tracking KPIs related to partner enablement and technical content creation.
  • Represent Fivetran as the technical liaison and collaborate with sales managers to clarify Fivetran's value proposition.
  • Assist sales managers in communicating Fivetran’s value to existing and prospective partners.
  • Engage in hands-on activities involving Fivetran and the broader data ecosystem.

PythonSQLCloud ComputingETLHadoopMachine LearningMongoDBOracleSnowflakeData engineeringREST API

Posted 28 days ago
Apply
Apply

📍 US

🧭 Full-Time

💸 164621.0 - 205776.0 USD per year

🔍 Data Movement and Management

  • Deep understanding of GSI and Reseller pre-sales processes.
  • Experience with sales methodologies and managing complex sales cycles.
  • Knowledge of technical enablement for GSI and Reseller efficiency.
  • Hands-on experience with cloud data platforms and data lakes.
  • Familiarity with data movement technologies and databases.
  • Technical skills in programming and data transformation.

  • Create, develop, build, and take ownership of ongoing relationships with GSI and Reseller pre-sales technical team members.
  • Interact daily with Fivetran’s top GSI and Reseller partners and track key performance indicators.
  • Represent Fivetran as the partner technical liaison and collaborate with sales managers.
  • Assist in conveying Fivetran’s value proposition to partners.
  • Participate in hands-on technical demonstrations and labs.

PythonSQLETLHadoopMicrosoft SQL ServerMongoDBOracleSnowflakeData engineeringREST API

Posted 29 days ago
Apply
Apply

📍 Armenia

🧭 Full-Time

🔍 Data Engineering

  • Proven enterprise experience in Python.
  • Proficiency in cloud platforms such as Azure, AWS, or GCP.
  • Advanced command of Linux with skills in service configuration and scripting.
  • Fundamental understanding of containerization technologies like Docker and Kubernetes.
  • Solid database skills, particularly in SQL and relational algebra.
  • Basic networking knowledge.

  • Develop and maintain scalable data pipelines and API integrations to manage data volume and complexity.
  • Collaborate with analytics and business teams to enhance data models for business intelligence tools.
  • Implement systems to monitor data quality, ensuring accuracy for stakeholders.
  • Design architecture for data storage solutions focused on retrieval, security, and performance.
  • Write high-quality documented code for data infrastructure components.
  • Act as a mentor to junior data engineers.

AWSDockerPythonSQLETLGCPHadoopKafkaKubernetesAzureNosqlSparkLinuxTerraform

Posted 29 days ago
Apply
Apply

📍 Argentina, Uruguay, Peru, Paraguay, Colombia

🧭 Full-Time

🔍 Interactive media strategy and development

🏢 Company: Halo Media👥 11-50InternetConsultingWeb DevelopmentAppsMarketingMobileWeb DesignSoftware

  • Minimum of 5 years of professional experience in R development.
  • Bachelor’s or Master’s degree in Computer Science, Data Science, Statistics, or a related field.
  • Proficiency in R programming language and its core packages (dplyr, ggplot2, tidyr, etc.).
  • Strong understanding of statistical methods and data analysis techniques.
  • Experience with version control systems (e.g., Git).
  • Familiarity with database management and SQL.
  • Knowledge of other programming languages (e.g., Python) is a plus.
  • Excellent problem-solving skills with the ability to interpret complex data and provide actionable insights.
  • Strong verbal and written communication skills.
  • Comfortable exploring and adopting new libraries or frameworks as needs demand.
  • Eager to participate in innovation towards new solutions to existing problems.
  • Ability to work in a fast-paced environment and manage multiple projects simultaneously.

  • Manage and execute advanced data analysis and software development projects using R, ensuring timely delivery and high-quality results.
  • Conduct sophisticated statistical analyses and modeling to extract insights from complex data sets.
  • Develop and maintain data processing pipelines, ensuring data integrity and accuracy.
  • Write, test, and deploy efficient R code for data analysis, visualization, and reporting.
  • Provide guidance and mentorship to junior R developers, promoting best practices and fostering a collaborative team environment.
  • Work closely with data scientists, analysts, and other stakeholders to understand their needs and develop appropriate solutions.
  • Conduct code reviews to ensure code quality, performance, and adherence to best practices.
  • Create and maintain comprehensive documentation for developed code and methodologies.
  • Stay updated with the latest developments in R programming and data science to bring innovative solutions to the team.

SQLGitHadoopMachine LearningSparkData visualizationData management

Posted about 1 month ago
Apply
Apply
🔥 Head of Data Product
Posted about 1 month ago

📍 USA

🔍 Deal sourcing platform

🏢 Company: Sourcescrub

  • 5+ years of experience in Product Management for data-related products.
  • Direct people and process management experience.
  • Extensive knowledge of the investing/private equity industry.
  • Experience designing research and data collection processes.
  • Strong Excel skills, experience with SQL and data modeling.
  • Knowledge of ontology development and data governance frameworks.
  • Familiarity with big data technologies like Hadoop and Spark.
  • Experience with cloud platforms such as AWS, Azure, or Google Cloud.
  • Understanding of data privacy regulations like GDPR and CCPA.

  • Own the strategy, development, execution, and budgeting for global data products.
  • Lead the Product Owner role for Data Labeling, Data Platform, and Data Automation Ingestion.
  • Engage in the development and optimization of raw and derived data products.
  • Collaborate with Go-to-Market teams on data accuracy, coverage, and freshness.
  • Oversee data acquisition from diverse sources.
  • Design innovative research methodologies ensuring accuracy and efficiency.
  • Identify patterns in data to support strategic planning.
  • Implement data quality measurement practices and conduct regular audits.
  • Document research methodologies to facilitate future initiatives.
  • Work with engineering teams on the data product roadmap.
  • Enhance research operations with new tools and explore new data sources.
  • Anticipate client needs with research solutions.
  • Contribute to the overall research strategy for SourceScrub.

AWSSQLData AnalysisHadoopProduct ManagementSparkData modeling

Posted about 1 month ago
Apply
Apply

📍 US

🧭 Full-Time

💸 163000.0 - 189000.0 USD per year

🔍 Software development for data science and AI solutions

🏢 Company: Domino Data Lab👥 251-500💰 Series F over 2 years agoArtificial Intelligence (AI)Big DataMachine LearningAnalyticsEnterprise ApplicationsData MiningEnterprise SoftwareSoftware

  • 5+ years of experience in backend development, especially in distributed computing with Hadoop, Spark, Kafka.
  • Experience in API design and development, e.g., RESTful APIs, GraphQL with a focus on integrating backend and front-end systems.
  • Strong proficiency in Java, Scala, Go, or Python, dedicated to building high-performance backend solutions.
  • Skilled in enhancing backend performance in cloud environments, well-versed in Docker, Kubernetes, and CI/CD practices.
  • Experience building production-grade software, preferably in Scala or Go.
  • Comfortable in dynamic, fast-paced settings with strong problem-solving skills and a comprehensive understanding of system design and architecture.
  • Strong familiarity with service-oriented architecture (SOA), adept at designing and developing modular, maintainable services for seamless system integration.
  • Ability to work independently and be a reliable teammate.

  • Design and develop software solutions that improve Domino’s user experience with regards to the new, highly-sought Governance feature.
  • Extend and contribute enhancements to the open-source software powering Domino.
  • Create technical designs and clearly communicate them to cross-functional stakeholders.
  • Enable fellow engineers to achieve high quality through design and code reviews.
  • Work with product managers to ensure solutions are well planned and delivered on-time.

Backend DevelopmentDockerGraphQLPythonHadoopJavaKafkaKubernetesGoSparkCI/CDRESTful APIsScala

Posted about 1 month ago
Apply
Shown 10 out of 78