Apache Hadoop Jobs

Find remote positions requiring Apache Hadoop skills. Browse through opportunities where you can utilize your expertise and grow your career.

Apache Hadoop
16 jobs found. to receive daily emails with new job openings that match your preferences.
16 jobs found.

Set alerts to receive daily emails with new job openings that match your preferences.

Apply

📍 San Francisco Bay Area

🔍 Financial planning and decision-making software

  • Confidence with a chosen programming language.
  • Ability to quickly learn new technologies.
  • Strong software engineering and computer science fundamentals.
  • Extensive experience with common big data workflow frameworks and solutions.

  • Laying the foundation of an exceptional data engineering practice.
  • Collaborating with the team to enhance big data workflow frameworks and solutions.

Backend DevelopmentPythonSoftware DevelopmentSQLApache AirflowApache HadoopData AnalysisElasticSearchGitHadoopAirflowData analysisData engineeringElasticsearchREST APIRESTful APIs

Posted 2024-11-10
Apply
Apply

📍 Canada

  • Extensive experience in the Apache Hadoop ecosystem.
  • Proficiency in Java programming.
  • Experience with designing and implementing Machine Learning pipelines.

  • Join a dynamic team as a Senior Big Data Engineer.
  • Develop batch processing systems for data collection and batch inference.
  • Work extensively with the Apache Hadoop ecosystem tools such as Map Reduce, Oozie, Hive, Pig, HBase, and Storm.

Apache HadoopHadoopJavaMachine LearningCommunication Skills

Posted 2024-11-07
Apply
Apply

📍 US

  • Extensive experience in the Apache Hadoop ecosystem such as Map Reduce, Oozie, Hive, Pig, HBase, and Storm.
  • Proficiency in Java programming language.

  • Develop batch processing systems for data collection and batch inference.
  • Work within the Apache Hadoop ecosystem including technologies like Map Reduce, Oozie, Hive, Pig, HBase, and Storm.
  • Contribute to the development of Machine Learning pipelines.

Apache HadoopHadoopJavaMachine LearningCommunication Skills

Posted 2024-11-07
Apply
Apply

📍 United States

🔍 Data Architecture and Management

🏢 Company: Unreal Gigs

  • Bachelor’s or Master’s degree in Computer Science, Data Management, Information Technology, or a related field.
  • 5+ years of experience in data architecture or a related field.
  • Extensive experience in designing and implementing data architectures, including data modeling and data warehousing.
  • Proficiency with cloud data services such as AWS, GCP, or Azure.
  • Strong understanding of ETL processes and data integration tools like Apache NiFi or Talend.
  • Expertise in data governance frameworks and data security policies.
  • Certifications in data architecture or cloud platforms are highly desirable.

  • Architect and implement a scalable, high-performance data infrastructure to support business analytics and operational reporting.
  • Create data governance frameworks and enforce data security policies.
  • Work closely with leadership and data teams to align architecture with strategic goals.
  • Design and manage data warehousing solutions for efficient storage and retrieval.
  • Develop and oversee ETL pipelines for accurate data integration.
  • Enable advanced analytics by building architecture for machine learning.
  • Create comprehensive documentation for data architectures and best practices.

AWSSQLApache AirflowApache HadoopETLGCPHadoopMachine LearningSnowflakeStrategyAirflowAzureData scienceSparkCollaboration

Posted 2024-11-07
Apply
Apply

📍 United States

🔍 Data Architecture

🏢 Company: Unreal Gigs

  • Extensive experience in designing and implementing data architectures, including data modeling, data warehousing, and data lakes.
  • Proficiency with cloud data services such as AWS, GCP, or Azure and hands-on experience with cloud-native storage solutions like BigQuery, Redshift, or Azure SQL Data Warehouse.
  • Strong understanding of ETL processes, data integration, and data transformation best practices.
  • Expertise in setting up data governance frameworks, data quality standards, and data security policies.
  • Proven ability to work cross-functionally with data engineers, analysts, and business stakeholders.
  • Bachelor’s or Master’s degree in Computer Science, Data Management, Information Technology, or a related field.
  • 5+ years of experience in data architecture or a related field.

  • Architect and implement a scalable, high-performance data infrastructure to support business analytics, data science, and operational reporting.
  • Create data governance frameworks and enforce data security policies to ensure data accuracy, privacy, and compliance.
  • Work closely with executive leadership, data engineers, and analysts to align the data architecture with the company’s strategic goals.
  • Design and manage data warehousing solutions such as Snowflake, BigQuery, or Redshift.
  • Develop and oversee ETL pipelines to ensure seamless integration of data from multiple sources.
  • Enable advanced analytics by building architecture that supports machine learning and predictive analytics.
  • Create comprehensive documentation of data architectures, schemas, and best practices.

AWSSQLApache AirflowApache HadoopETLGCPHadoopMachine LearningSnowflakeStrategyAirflowAzureData scienceSparkCollaboration

Posted 2024-11-07
Apply
Apply

📍 United States

🧭 Full-Time

💸 164200 - 229900 USD per year

🔍 Social Media

🏢 Company: Reddit

  • PhD in Statistics, Experimental Economics/Econometrics, Statistical Physics, Epidemiology, Political Science, Computational Linguistics, Computational Neuroscience, or similar quantitative discipline.
  • Expertise in experimental and quasi-experimental design and analysis.
  • Deep knowledge of advanced statistical methods, particularly causal inference.
  • Familiarity with qualitative research methods.
  • Proficiency with big data technologies (SQL, Hadoop, Big Query).
  • Strong background in the execution, analysis, and presentation of quantitative research.
  • Ability to drive innovative research and work through ambiguity.

  • Core responsibility is on research and development of various measurement methodologies, algorithms, and products that support sales and measurement efforts.
  • Utilize quantitative and qualitative methods to enhance measurement and experimentation capabilities.
  • Work with measurement partners to assess their methodologies for measuring ad impact.
  • Advise go-to-market teams with deep measurement expertise.
  • Support execution of primary research that enhances value proposition to advertisers.
  • Consult with advertisers' analytics teams on methodology and solutions.
  • Represent Reddit to the research and analytics industry through speaking engagements and publications.

SQLApache HadoopData AnalysisHadoopData analysisAnalytical SkillsProblem Solving

Posted 2024-11-07
Apply
Apply
🔥 Data Architect
Posted 2024-10-23

📍 Mexico

🧭 Full-Time

🔍 AI consulting and custom software development

🏢 Company: Creai

  • Proven experience as a Data Architect with expertise in Microsoft Azure data services.
  • Strong understanding of data management principles, data modeling, and data integration techniques.
  • Hands-on experience with Azure Data Factory, Azure SQL, Cosmos DB, Azure Synapse Analytics, and other Azure cloud-based data tools.
  • Proficiency in building and maintaining data pipelines, data lakes, and data warehouses.
  • Experience with ETL processes and tools to automate and optimize data flows.
  • Strong knowledge of SQL, as well as experience with NoSQL databases.
  • Familiarity with data governance, security standards, and best practices for data architecture in cloud environments.
  • Excellent problem-solving and analytical skills, with the ability to work in a fast-paced environment.
  • Bachelor’s degree in Computer Science, Data Science, Information Technology, or a related field, or equivalent experience.

  • Design and implement data architectures that support AI and machine learning solutions, ensuring scalability, reliability, and performance.
  • Lead data integration efforts, including data pipelines, data warehouses, and data lakes, using Microsoft Azure services.
  • Work closely with cross-functional teams to ensure that data architecture meets business and technical requirements.
  • Optimize database performance, troubleshoot issues, and ensure data security and governance in compliance with industry standards.
  • Implement ETL processes, and manage data storage solutions such as Azure SQL, Cosmos DB, or Data Lake Storage.
  • Leverage Azure services like Azure Data Factory, Azure Synapse Analytics, and Azure Databricks to build and maintain robust data pipelines.
  • Maintain and document data models, architectural guidelines, and best practices to ensure consistent data architecture across projects.
  • Monitor and optimize data architectures to improve efficiency and cost-effectiveness in Azure environments.
  • Stay updated on new Azure data services and industry best practices.

SQLApache AirflowApache HadoopArtificial IntelligenceETLHadoopMachine LearningMicrosoft AzureAirflowAzureData scienceNosqlSparkAnalytical Skills

Posted 2024-10-23
Apply
Apply

📍 United States

🧭 Full-Time

🔍 Cyber-security

🏢 Company: Shuvel Digital

  • Experience in machine learning and data science.
  • Knowledge of cyber-security data and statistical analysis.
  • Understanding of ETL data hygiene methods.
  • Familiarity with machine learning algorithms and frameworks.
  • Hands-on experience with Python, SQL, and data analysis tools.

  • Research, develop, architect, and integrate ML models and algorithms.
  • Collaborate with data scientists and other teams to address specific problems.
  • Design and implement data processing and ETL algorithms.
  • Analyze structured cyber-security data for insights.
  • Deploy ML solutions into production environments.

AWSDockerPythonSQLApache HadoopCloud ComputingData AnalysisETLHadoopJavaKafkaKerasKubernetesMachine LearningNumpyPyTorchAlgorithmsAzureData analysisData scienceGoRustSparkTensorflowAnalytical SkillsCollaborationProblem Solving

Posted 2024-10-23
Apply
Apply

📍 Dubai, London

🔍 Data Infrastructure

🏢 Company: Eqvilent

  • 3+ years in a similar role.
  • Proven experience with AWS or other cloud providers.
  • Experience with distributed systems (e.g. Apache Kafka, Apache Airflow, Apache Hadoop).
  • Proficiency with Terraform.
  • Extensive experience with Docker and Kubernetes, including cluster setup, node pools, and Helm charts.
  • Experience with CI/CD tools (e.g. GitLab CI, Jenkins).
  • Familiarity with observability tools such as Prometheus, Grafana, ELK stack.
  • Solid understanding of networking, security, and system architecture.
  • Strong scripting skills (e.g., Python, Bash).
  • Excellent problem-solving skills, communication, and collaboration abilities.

  • Design, implement, and maintain both cloud and on-premise compute and storage infrastructure.
  • Set up and manage Kubernetes clusters, implement Helm charts, ensuring high availability and performance.
  • Set up, maintain, and scale distributed systems (e.g. Apache Kafka, Apache Airflow) ensuring data integrity and security.
  • Automate code delivery processes and implement CI/CD, monitoring, logging, and alerting solutions.
  • Collaborate with development and operations teams, provide production support, and participate in on-call rotations.

AWSDockerPythonApache AirflowApache HadoopBashHadoopJenkinsKafkaKubernetesAirflowApache KafkaGrafanaPrometheusCollaborationCI/CDTerraform

Posted 2024-10-21
Apply
Apply

📍 Japan

🔍 Telecommunications, IoT

🏢 Company: SORACOM

  • Javaを使った業務ロジックの実装経験
  • ビジネスに必要な機能を把握し、ビジネスサイドと共に適切な仕様を定義し、実装・テストする能力
  • 3年以上の業務システム設計/実装と運用経験
  • 自らの手を動かして試す・体験することが出来る姿勢
  • システム全体および利用するライブラリ・フレームワークを理解した上で実装する姿勢
  • 社内のグローバルチームと仕事を進めるための日本語力/英語力

  • 日本およびグローバルで提供するソラコム各サービスの課金・決済システムの設計・構築・運用
  • デバイスや通信機器の直販/出荷システムの設計・構築・運用
  • 社内利用の業務システムの設計・構築・運用
  • 利用者のフィードバックを元に、継続的にシステムの改善とリリースを行う
  • ソラコムのビジネス拡大をシステムで支える

Backend DevelopmentSoftware DevelopmentApache HadoopHadoopJavaSpringSpring BootSpark

Posted 2024-09-20
Apply
Shown 10 out of 16