Hadoop Jobs

Find remote positions requiring Hadoop skills. Browse through opportunities where you can utilize your expertise and grow your career.

Hadoop
125 jobs found. to receive daily emails with new job openings that match your preferences.
125 jobs found.

Set alerts to receive daily emails with new job openings that match your preferences.

Apply

πŸ“ US

🧭 Full-Time

πŸ’Έ 165000 - 185000 USD per year

πŸ” Healthcare technology

  • Bachelor’s degree in Computer Science or related field, or equivalent work experience.
  • 7+ years of professional experience as a software developer or data engineer.
  • Experience with healthcare data such as health payments and electronic health records.
  • Knowledge of interoperability technologies and standards like X12, NCPDP, and FHIR.
  • Expertise in Big Data processing, Analytics, and relational databases.
  • Experience with modern big data technologies such as Databricks, Hadoop, and Kafka.
  • Familiarity with object-oriented languages, primarily Java and Scala.
  • Experience in cloud-based microservice architecture and CI/CD frameworks.

  • Performs advanced programming tasks to satisfy business requirements.
  • Plans, designs, develops, tests software system features, enhancements, or bugs at an advanced level.
  • Takes technical leadership on design and development of new features.
  • Works with team to break complex problems into actionable tasks.
  • Independently modifies existing functionality and fixes bugs.
  • Maintains unit/integration tests for all completed coding.
  • Mentors junior team members and contributes to a positive work culture.

AWSDockerLeadershipAgileBusiness IntelligenceGCPHadoopJavaKafkaProduct ManagementAzureCollaborationCI/CDWritten communication

Posted 2024-11-21
Apply
Apply

πŸ“ LATAM

πŸ” Data Science Consultancy

  • MS degree in Statistics, Math, Data Analytics, or a related quantitative field.
  • 3+ years of professional experience in Advanced Data Science, including predictive modeling, statistical analysis, machine learning, text mining, geospatial analytics, and time series forecasting.
  • Experience with Advanced Data Science software languages, such as Python, R, or SAS.
  • Proven ability to deploy machine learning models.
  • Experience with SQL, relational databases, and query authoring and tuning.
  • Familiarity with Hadoop/Hive, Spark, and data-frames in PySpark or Scala.
  • Strong problem-solving skills and comfortable with cloud-based platforms (AWS, Azure, Google).
  • Experience with Google Analytics, Adobe Analytics, and/or Optimizely are a plus.

  • Working with practice leaders and clients to understand business problems, industry context, data sources, risks, and constraints.
  • Solving challenges through diverse Data Science solutions and aligning stakeholders on strategies, deliverables, and roadmaps.
  • Developing project plans with milestones, dates, owners, and risk contingency.
  • Creating and maintaining efficient data pipelines within clients' architecture using SQL, Spark, and Cloud technologies.
  • Assembling large, complex datasets that meet business requirements and constructing analytics tools to provide actionable insights.
  • Performing data cleaning, QC, and integration, conducting statistical analysis for decision-making, and documenting results for client deliverables.

PythonSQLHadoopMachine LearningData scienceSpark

Posted 2024-11-21
Apply
Apply

πŸ“ Colombia

🧭 Full-Time

πŸ” Fintech

  • Experience working on an agile team.
  • Strong communication and collaboration skills.
  • Minimum 4 years creating distributed, data-intensive, and highly scalable applications with Java 8/11, SpringBoot, and AWS.
  • Experience with Rest API/Microservices and concepts like domain-driven design.
  • Proven experience with AWS services like SQS, SNS, DynamoDB, and Lambda.
  • Strong SQL knowledge and experience with data transformation processes.

  • Leading and delivering development tasks through all SDLC phases.
  • Completing research and proof of concepts that can be converted into MVPs.
  • Collaborating with other teams in Caseware, including cloudops, devops, and product.
  • Participating in code reviews and understanding existing architecture.
  • Taking ownership of modules in the DA architecture and providing technical guidance.

AWSSQLAgileDynamoDBETLHadoopJavaJavascriptJavaScriptAngularREST APISparkCollaborationMentoringMicroservices

Posted 2024-11-21
Apply
Apply

πŸ“ Colombia

πŸ” Communications

  • Typically 6+ years of proven experience in data science, focusing on LLMs and supervised ML models.
  • Proficiency in programming languages such as Python or R, with experience in SciKit-Learn, XGBoost, Keras.
  • Strong understanding of data processing and transformation techniques, including SQL and big data technologies.
  • Experience developing, training, and deploying ML models and Python applications in production environments.
  • Familiarity with cloud platforms like AWS Sagemaker and containerization tools like Docker and Kubernetes.
  • Excellent problem-solving skills and strong communication skills.

  • Build and deploy machine learning models, including propensity models and GenAI applications.
  • Collaborate closely with product, program, analytics, and engineering teams to refine models.
  • Leverage technical stack including Python, SQL, R, AWS, and data science libraries for AI/ML solutions.
  • Utilize enterprise data sources like Salesforce and Zendesk for model development.
  • Apply knowledge of LLMs to develop innovative solutions.

AWSDockerPythonSQLHadoopKerasKubernetesMachine LearningMySQLNumpySalesforceData sciencePandasSparkCommunication Skills

Posted 2024-11-20
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ” Consulting/Biomedical Services

🏒 Company: ICA.ai

  • Master's or PhD degree in Statistics, Mathematics, Computer Science, Data Science, or a related quantitative field.
  • 12+ years of experience in data science or a related field.
  • At least 5 years of management experience leading teams of 10 or more data scientists, ML engineers, solution architects, and data engineers.
  • Deep expertise in statistical modeling, machine learning algorithms, and AI techniques.
  • Proficiency with programming languages such as Python and R, and experience with data manipulation libraries (e.g., pandas, NumPy).
  • Extensive experience with AI/ML frameworks like TensorFlow, PyTorch, scikit-learn.
  • Strong knowledge of big data technologies (e.g., Hadoop, Spark) and data warehousing solutions.
  • Advanced knowledge of AWS services (Athena, S3, Glue, SageMaker, Comprehend).
  • Familiarity with MLOps practices and tools (e.g., MLflow, Kubeflow).

  • Leverage technical expertise and business acumen to collaborate with senior stakeholders.
  • Ensure the Data Science team has necessary skills and resources to solve client data challenges.
  • Oversee a team of full-time employees and consultants supporting government clients.
  • Plan and direct operational policies, objectives, and initiatives.
  • Develop strategies and drive actions to meet financial and operational goals.
  • Monitor financial and operational metrics to achieve objectives.

AWSLeadershipProject ManagementPythonAgileArtificial IntelligenceCloud ComputingData AnalysisHadoopKubeflowMachine LearningMLFlowNumpyPyTorchSCRUMAlgorithmsData analysisData sciencePandasSparkTensorflowCommunication SkillsCollaboration

Posted 2024-11-17
Apply
Apply

πŸ“ Canada

πŸ” Cybersecurity

🏒 Company: Abnormal Security

  • 5+ years of experience as a data engineer or in a similar role, with hands-on experience in building data-focused solutions.
  • Expertise in ETL, data pipeline design, and data engineering tools and technologies (e.g., Apache Spark, Hadoop, Airflow, Kafka).
  • Experience with maintaining real-time and near real-time data pipelines or streaming services at high scale.
  • Experience with maintaining large scale distributed systems on cloud platforms such as AWS, GCP, or Azure.
  • Strong grasp of best practices in cloud-based data engineering and data quality frameworks.

  • Architect, design, build, and deploy backend ETL jobs and infrastructure that support a world-class Detection Engine.
  • Ownership of projects that enable scaling of Detection’s Data Pipeline by 10x.
  • Manage real-time and near real-time streaming pipelines and online feature serving services.
  • Collaborate closely with MLE and Data Science teams, executing based on feedback.
  • Coach and mentor junior engineers through various methods.

AWSETLGCPHadoopKafkaStrategyAirflowAzureData engineeringSpark

Posted 2024-11-16
Apply
Apply

πŸ“ India

🧭 Full-Time

πŸ’Έ 3500000 - 3700000 INR per year

🏒 Company: CloudHire

  • 7+ years of experience with AWS, GCP, or Azure.
  • 10+ years experience in designing and supporting data pipelines and analytical solutions with technologies like Postgres, Redshift, BigQuery, Hadoop, Teradata, MS SQL Server, Talend, Informatica, PowerBI, and/or SSIS.
  • 10+ years with object-oriented languages (.Net, Java, Python) for streaming analytics.
  • Experience in agile teams delivering data solutions.

  • Integrate multiple databases together including Snowflake schema, Star schema, and Network model.
  • Work with message buses such as Kafka and IBM MQ to targets like Redshift and Postgres.
  • Define future-state business capabilities and translate them into data architectures.
  • Analyze current technologies to find deficiencies and recommend improvements.
  • Implement real-time data pipelines using existing and emerging technologies.
  • Develop continuous integration and deployment for data pipelines.
  • Mentor and support the team to achieve organizational goals.

AWSPythonSQLAgileErwinGCPHadoopJavaKafkaMongoDBSnowflakeAirflowAzureData engineering.NETPostgresMentoringNegotiation

Posted 2024-11-13
Apply
Apply

πŸ“ India

🧭 Contract

πŸ” Consulting and professional services, specializing in AI and data solutions

🏒 Company: 66degrees

  • In-depth knowledge of GCP data services such as BigQuery, Cloud Dataflow, Cloud Pub/Sub, Cloud Storage, and more.
  • Expertise in data modeling techniques.
  • Hands-on experience with big data technologies like Spark and Hadoop.
  • Proven ability to design reliable and cost-effective cloud architectures.
  • Understanding of data quality, security, compliance, and metadata management.
  • Proficiency in SQL, Python, and DBT.
  • Strong analytical and problem-solving skills.
  • Excellent written and verbal communication skills.
  • Bachelor’s degree in Computer Science, Computer Engineering, Data or related field.

  • Design, implement, and manage cloud-based data architectures on GCP.
  • Develop and maintain data models to support business needs.
  • Design and implement solutions for processing large datasets.
  • Establish and enforce data governance policies.
  • Build and optimize data pipelines for efficient data management.
  • Monitor and tune data systems for high performance.
  • Collaborate with data engineers, data scientists, and stakeholders.

PythonSQLCloud ComputingGCPHadoopOracleData engineeringSparkCommunication SkillsCollaborationCompliance

Posted 2024-11-12
Apply
Apply

πŸ“ Slovakia

🧭 Full-Time

πŸ” Cybersecurity

🏒 Company: SentinelOne - Czech Republic

  • Multiple years of experience in (big) data engineering and architecture.
  • Proficiency in programming languages such as Python, Java, or Scala.
  • Experience with cloud-based technologies, particularly AWS.
  • Familiarity with big data technologies like Apache Spark, Hadoop, Hive, IceBerg, or Trino.
  • Knowledge of data warehousing concepts and technologies.
  • Experience with data visualization tools such as Tableau, PowerBI, or Grafana.
  • Strong problem-solving skills and excellent communication abilities.
  • A BS or higher degree in Computer Science, Statistics, Mathematics, or a related field.

  • Play a vital role in designing, building, and maintaining data infrastructure for cyber threat detection research.
  • Collaborate closely with data scientists, cybersecurity researchers, product managers, and other production teams.
  • Deliver innovative big data solutions aimed at developing advanced threat detection systems.

AWSPythonHadoopJavaTableauData engineeringGrafanaSparkCommunication SkillsCollaboration

Posted 2024-11-11
Apply
Apply

πŸ“ Canada

🧭 Full-Time

πŸ” Software, Open-source

🏒 Company: Canonical - Jobs

  • Experience in AI/ML and MLOps.
  • Practical knowledge of Linux, virtualisation, containers, and networking.
  • Sound knowledge of cloud computing concepts and technologies such as Kubernetes, AWS, GCP, Azure.
  • Intermediate level of Python programming skills.
  • Excellent communication skills in English and Spanish.
  • Demonstrated drive for continual learning and good organizational skills.
  • University degree in Computer Science or related software engineering expertise.

  • Work across the entire Linux stack, from kernel to applications.
  • Design and deploy clouds and applications for external customers.
  • Architect cloud infrastructure solutions like Kubernetes, Kubeflow, OpenStack, Ceph, Hadoop, and Spark.
  • Grow a collaborative engineering culture aligned with company values.
  • Travel up to 25% of the time for internal and external events.

AWSPythonArtificial IntelligenceCloud ComputingGCPHadoopKubeflowKubernetesMachine LearningAzureSparkLinuxPresentation skills

Posted 2024-11-11
Apply
Shown 10 out of 125