Apply

Senior Data Engineer

Posted 2024-11-21

View full description

πŸ’Ž Seniority level: Senior, At least 5 years

πŸ“ Location: Poland

πŸ” Industry: Software development

🏒 Company: Sunscrapers sp. z o.o.

πŸ—£οΈ Languages: English

⏳ Experience: At least 5 years

πŸͺ„ Skills: PythonSQLSnowflakeAirflowAnalytical SkillsCustomer serviceDevOpsAttention to detail

Requirements:
  • At least 5 years of professional experience as a data engineer.
  • Undergraduate or graduate degree in Computer Science, Engineering, Mathematics, or similar.
  • Excellent command in spoken and written English, at least C1.
  • Strong professional experience with Python and SQL.
  • Hands-on experience on DBT and Snowflake.
  • Experience in building data pipelines with Airflow or alternative solutions.
  • Strong understanding of data modeling techniques like Kimball Star Schema.
  • Great analytical skills and attention to detail.
  • Creative problem-solving skills.
  • Great customer service and troubleshooting skills.
Responsibilities:
  • Modeling datasets and schemes for consistency and easy access.
  • Design and implement data transformations and data marts.
  • Integrating third-party systems and external data sources into data warehouse.
  • Building data flows for fetching, aggregation, and data modeling using batch pipelines.
Apply

Related Jobs

Apply

πŸ“ Any European country

🧭 Full-Time

πŸ” Software development

🏒 Company: Janea Systems

  • Proven experience as a data engineer, preferably with at least 3 or more years of relevant experience.
  • Experience designing cloud native solutions and implementations with Kubernetes.
  • Experience with Airflow or similar pipeline orchestration tools.
  • Strong Python programming skills.
  • Experience collaborating with Data Science and Engineering teams in production environments.
  • Solid understanding of SQL and relational data modeling schemas.
  • Preference for experience with Databricks or Spark.
  • Familiarity with modern data stack design and data lifecycle management.
  • Experience with distributed systems, microservices architecture, and cloud platforms like AWS, Azure, Google Cloud.
  • Excellent problem-solving skills and strong communication skills.

  • Develop and maintain data pipelines using Databricks, Airflow, or similar orchestration systems.
  • Design and implement cloud-native solutions using Kubernetes for high availability.
  • Gather product data requirements and implement solutions to ingest and process data for applications.
  • Collaborate with Data Science and Engineering teams to optimize production-ready applications.
  • Cultivate data from various sources for data scientists and maintain documentation.
  • Design modern data stack for data scientists and ML engineers.

AWSPythonSoftware DevelopmentSQLKubernetesAirflowAzureData scienceSparkCollaboration

Posted 2024-11-07
Apply
Apply

πŸ“ UK, EU

πŸ” Consultancy

🏒 Company: The Dot Collective

  • Advanced knowledge of distributed computing with Spark.
  • Extensive experience with AWS data offerings such as S3, Glue, Lambda.
  • Ability to build CI/CD processes including Infrastructure as Code (e.g. terraform).
  • Expert Python and SQL skills.
  • Agile ways of working.

  • Leading a team of data engineers.
  • Designing and implementing cloud-native data platforms.
  • Owning and managing technical roadmap.
  • Engineering well-tested, scalable, and reliable data pipelines.

AWSPythonSQLAgileSCRUMSparkCollaborationAgile methodologies

Posted 2024-11-07
Apply
Apply

πŸ“ Poland

🧭 Full-Time

πŸ” Cyber security, user privacy, and machine learning

🏒 Company: Intuition Machines, Inc.πŸ‘₯ 51-100InternetEducationInternet of ThingsMachine LearningSoftware

  • Thoughtful, conscientious, and self-directed.
  • Minimum of 3 years in a data role involving designing and building data stores, feature engineering, and building reliable data pipelines that handle high loads.
  • At least 2 years of professional software development experience in a role other than data engineering.
  • Significant experience coding and developing in Python.
  • Experience in building and maintaining distributed data pipelines.
  • Experience working with Kafka infrastructure and applications.
  • Deep understanding of SQL and NoSQL databases (preferably Clickhouse).
  • Familiarity with public cloud providers (AWS or Azure).
  • Experience with CI/CD and orchestration platforms: Kubernetes, containerization, and microservice design.
  • Familiarity with distributed systems and architectures.

  • Maintain, extend, and improve existing data/ML workflows and implement new ones to handle high-velocity data.
  • Provide interfaces and systems that enable ML engineers and researchers to build datasets on demand.
  • Influence data storage and processing strategies.
  • Collaborate with the ML team, as well as frontend and backend teams, to build out our data platform.
  • Reduce time-to-deployment for dashboards and ML models.
  • Establish best practices and develop pipelines that enable ML engineers and researchers to efficiently build and use datasets.
  • Work with large datasets under performance constraints comparable to those at the largest companies.
  • Iterate quickly, focusing on early and frequent shipping to deploy new products or features to millions of users.

AWSPythonSoftware DevelopmentSQLKafkaKubernetesStrategyAzureClickhouseData engineeringNosqlCI/CD

Posted 2024-10-12
Apply
Apply

πŸ“ Central EU or Americas

🧭 Full-Time

πŸ” Real estate investment

🏒 Company: RoofstockπŸ‘₯ 501-1000πŸ’° $240.0m Series E on 2022-03-10πŸ«‚ on 2023-03-22Rental PropertyPropTechMarketplaceReal EstateFinTech

  • BS or MS in a technical field: computer science, engineering or similar.
  • 8+ years technical experience working with data.
  • 5+ years strong experience building scalable data services and applications using SQL, Python, Java/Kotlin.
  • Deep understanding of microservices architecture and RESTful API development.
  • Experience with AWS services including messaging and familiarity with real-time data processing frameworks.
  • Significant experience building and deploying data-related infrastructure and robust data pipelines.
  • Strong understanding of data architecture and related challenges.
  • Experience with complex problems and distributed systems focusing on scalability and performance.
  • Strong communication and interpersonal skills.
  • Independent worker able to collaborate with cross-functional teams.

  • Improve and maintain the data services platform.
  • Deliver high-quality data services promptly, ensuring data governance and integrity while meeting objectives and maintaining SLAs.
  • Develop effective architectures and produce key code components contributing to technical solutions.
  • Integrate a diverse network of third-party tools into a cohesive, scalable platform.
  • Continuously enhance system performance and reliability by diagnosing and resolving operational issues.
  • Ensure rigorous testing of the team's work through automated methods.
  • Support data infrastructure and collaborate with the data team on scalable data pipelines.
  • Work within an Agile/Scrum framework with cross-functional teams to deliver value.
  • Influence the enterprise data platform architecture and standards.

AWSDockerPythonSQLAgileETLSCRUMSnowflakeAirflowData engineeringgRPCRESTful APIsMicroservices

Posted 2024-08-10
Apply