Apply

Senior Data Engineer

Posted 2024-10-12

View full description

💎 Seniority level: Senior, Minimum of 3 years in a data role and at least 2 years in software development

📍 Location: Poland

🔍 Industry: Enterprise security products

🏢 Company: Intuition Machines, Inc.

⏳ Experience: Minimum of 3 years in a data role and at least 2 years in software development

🪄 Skills: AWSPythonSoftware DevelopmentSQLKafkaKubernetesStrategyAzureClickhouseData engineeringNosqlCI/CD

Requirements:
  • Thoughtful, conscientious, and self-directed.
  • Experience working with data engineering services on major cloud providers.
  • Minimum of 3 years of experience in a data role involving designing and building data stores, feature engineering, and building reliable data pipelines that handle high loads.
  • Proven ability to make independent decisions regarding data processing strategy and architecture.
  • At least 2 years of professional software development experience in a role other than data engineering.
  • Significant experience coding and developing in Python.
  • Experience in building and maintaining distributed data pipelines.
  • Experience working with Kafka infrastructure and applications.
  • Deep understanding of SQL and NoSQL databases (preferably Clickhouse).
  • Familiarity with public cloud providers (AWS or Azure).
  • Experience with CI/CD and orchestration platforms: Kubernetes, containerization, and microservice design.
  • Familiarity with distributed systems and architectures.
Responsibilities:
  • Maintain, extend, and improve existing data/ML workflows, and implement new ones to handle high-velocity data.
  • Provide interfaces and systems that enable ML engineers and researchers to build datasets on demand.
  • Influence data storage and processing strategies.
  • Collaborate with the ML team, as well as frontend and backend teams, to build out our data platform.
  • Reduce time-to-deployment for dashboards and ML models.
  • Establish best practices and develop pipelines that enable ML engineers and researchers to efficiently build and use datasets.
  • Work with large datasets under performance constraints comparable to those at the largest companies.
  • Iterate quickly, focusing on shipping early and often, ensuring that new products or features can be deployed to millions of users.
Apply

Related Jobs

Apply
🔥 Senior Data Engineer
Posted 2024-11-21

📍 Poland

🧭 Full-Time

🔍 Software development

🏢 Company: Sunscrapers sp. z o.o.

  • At least 5 years of professional experience as a data engineer.
  • Undergraduate or graduate degree in Computer Science, Engineering, Mathematics, or similar.
  • Excellent command in spoken and written English, at least C1.
  • Strong professional experience with Python and SQL.
  • Hands-on experience with DBT and Snowflake.
  • Experience in building data pipelines with Airflow or alternative solutions.
  • Strong understanding of various data modeling techniques like Kimball Star Schema.
  • Great analytical skills and attention to detail.
  • Creative problem-solving skills.
  • Great customer service and troubleshooting skills.

  • Modeling datasets and schemes for consistency and easy access.
  • Design and implement data transformations and data marts.
  • Integrating third-party systems and external data sources into data warehouse.
  • Building data flows for fetching, aggregation and data modeling using batch pipelines.

PythonSQLSnowflakeAirflowAnalytical SkillsCustomer serviceDevOpsAttention to detail

Posted 2024-11-21
Apply
Apply
🔥 Senior Data Engineer
Posted 2024-11-07

📍 Any European country

🧭 Full-Time

🔍 Software development

🏢 Company: Janea Systems

  • Proven experience as a data engineer, preferably with at least 3 or more years of relevant experience.
  • Experience designing cloud native solutions and implementations with Kubernetes.
  • Experience with Airflow or similar pipeline orchestration tools.
  • Strong Python programming skills.
  • Experience collaborating with Data Science and Engineering teams in production environments.
  • Solid understanding of SQL and relational data modeling schemas.
  • Preference for experience with Databricks or Spark.
  • Familiarity with modern data stack design and data lifecycle management.
  • Experience with distributed systems, microservices architecture, and cloud platforms like AWS, Azure, Google Cloud.
  • Excellent problem-solving skills and strong communication skills.

  • Develop and maintain data pipelines using Databricks, Airflow, or similar orchestration systems.
  • Design and implement cloud-native solutions using Kubernetes for high availability.
  • Gather product data requirements and implement solutions to ingest and process data for applications.
  • Collaborate with Data Science and Engineering teams to optimize production-ready applications.
  • Cultivate data from various sources for data scientists and maintain documentation.
  • Design modern data stack for data scientists and ML engineers.

AWSPythonSoftware DevelopmentSQLKubernetesAirflowAzureData scienceSparkCollaboration

Posted 2024-11-07
Apply
Apply

📍 Poland

🔍 Financial services industry

🏢 Company: Capco

  • Extensive experience with Databricks, including ETL processes and data migration.
  • Experience with additional cloud platforms like AWS, Azure, or GCP.
  • Strong knowledge of data warehousing concepts, data modeling, and SQL.
  • Proficiency in programming languages such as Python, SQL, and scripting languages.
  • Knowledge of data governance frameworks and data security principles.
  • Familiarity with containerization technologies such as Docker and orchestration tools like Kubernetes.
  • Bachelor or Master Degree in Computer Science or related field.

  • Design, develop, and implement robust data architecture solutions utilizing modern data platforms like Databricks.
  • Ensure scalable, reliable, and secure data environments that meet business requirements and support advanced analytics.
  • Lead the migration of data from traditional RDBMS systems to Databricks environments.
  • Architect and design scalable data pipelines and infrastructure to support the organization's data needs.
  • Develop and manage ETL processes using Databricks to ensure efficient data extraction, transformation, and loading.
  • Optimize ETL workflows to enhance performance and maintain data integrity.
  • Monitor and optimize performance of data systems to ensure reliability, scalability, and cost-effectiveness.
  • Collaborate with cross-functional teams to understand data requirements and deliver solutions.
  • Define best practices for data engineering and ensure adherence to them.
  • Evaluate and implement new technologies to improve data pipeline efficiency.

AWSDockerLeadershipPythonSQLETLGCPKubernetesAzureData engineeringRDBMSAnalytical Skills

Posted 2024-11-07
Apply
Apply

📍 UK, EU

🔍 Consultancy

🏢 Company: The Dot Collective

  • Advanced knowledge of distributed computing with Spark.
  • Extensive experience with AWS data offerings such as S3, Glue, Lambda.
  • Ability to build CI/CD processes including Infrastructure as Code (e.g. terraform).
  • Expert Python and SQL skills.
  • Agile ways of working.

  • Leading a team of data engineers.
  • Designing and implementing cloud-native data platforms.
  • Owning and managing technical roadmap.
  • Engineering well-tested, scalable, and reliable data pipelines.

AWSPythonSQLAgileSCRUMSparkCollaborationAgile methodologies

Posted 2024-11-07
Apply
Apply

📍 Central EU or Americas

🧭 Full-Time

🔍 Real estate investment

🏢 Company: Roofstock👥 501-1000💰 $240.0m Series E on 2022-03-10🫂 on 2023-03-22Rental PropertyPropTechMarketplaceReal EstateFinTech

  • BS or MS in a technical field: computer science, engineering or similar.
  • 8+ years technical experience working with data.
  • 5+ years strong experience building scalable data services and applications using SQL, Python, Java/Kotlin.
  • Deep understanding of microservices architecture and RESTful API development.
  • Experience with AWS services including messaging and familiarity with real-time data processing frameworks.
  • Significant experience building and deploying data-related infrastructure and robust data pipelines.
  • Strong understanding of data architecture and related challenges.
  • Experience with complex problems and distributed systems focusing on scalability and performance.
  • Strong communication and interpersonal skills.
  • Independent worker able to collaborate with cross-functional teams.

  • Improve and maintain the data services platform.
  • Deliver high-quality data services promptly, ensuring data governance and integrity while meeting objectives and maintaining SLAs.
  • Develop effective architectures and produce key code components contributing to technical solutions.
  • Integrate a diverse network of third-party tools into a cohesive, scalable platform.
  • Continuously enhance system performance and reliability by diagnosing and resolving operational issues.
  • Ensure rigorous testing of the team's work through automated methods.
  • Support data infrastructure and collaborate with the data team on scalable data pipelines.
  • Work within an Agile/Scrum framework with cross-functional teams to deliver value.
  • Influence the enterprise data platform architecture and standards.

AWSDockerPythonSQLAgileETLSCRUMSnowflakeAirflowData engineeringgRPCRESTful APIsMicroservices

Posted 2024-08-10
Apply