Apply

Senior Data Engineer

Posted 8 months agoViewed

View full description

πŸ’Ž Seniority level: Senior, 7+ years

πŸ” Industry: Construction technology

🏒 Company: EquipmentShareπŸ‘₯ 1001-5000πŸ’° $400,000,000 Debt Financing over 1 year agoConstruction

πŸ—£οΈ Languages: English

⏳ Experience: 7+ years

πŸͺ„ Skills: AWSPostgreSQLPythonSQLAgileApache AirflowKafkaMachine LearningMLFlowSnowflakeAirflowSpark

Requirements:
  • The ideal candidate should have 7+ years of relevant data platform development experience, proficiency in SQL and Python, experience with designing and building distributed data architecture, managing data pipelines using tools like Airflow and building and managing data platforms using distributed systems like Kafka and Spark
  • Additionally, familiarity with event data streaming at scale, proven track record of learning new technologies quickly, and experience with building observability and monitoring into data products are required.
Responsibilities:
  • The Senior Data Engineer will collaborate with cross-functional teams to design, build, and maintain a data platform that supports analytics and machine learning products for the construction industry
  • Responsibilities include working closely with various team members to bring ideas to production, develop data product frameworks, and mentor peers.
Apply

Related Jobs

Apply

πŸ“ Lithuania

πŸ’Έ 4000.0 - 6000.0 EUR per month

πŸ” Software Development

🏒 Company: Softeta

  • 4+ years of experience as a Data Engineer
  • Experience with Azure (Certifications are a Plus)
  • Experience with Databricks, Azure Data Lake, Data Factory and Apache Airflow
  • CI/CD or infrastructure as code
  • Knowledge of Medallion Architecture or Multihop architecture
  • Experience developing and administering ETL processes in the Cloud (Azure, AWS or GCP) environment
  • Strong programming skills in Python and SQL
  • Strong problem-solving and analytical skills
  • Design, develop, and maintain data pipelines and ETL processes
  • Data modeling, data cleansing
  • Automating data processing workflows using tools such as Airflow or other workflow management tools
  • Optimizing the performance of databases, including designing and implementing data structures and using indexes appropriately
  • Implement data quality and data governance processes
  • Being a data advocate and helping unlock business value by using data

PythonSQLApache AirflowETLAzureData engineeringCI/CDData modeling

Posted 1 day ago
Apply
Apply

πŸ“ Thailand, Philippines

πŸ” Fintech

🏒 Company: EnvissoπŸ‘₯ 11-50CreditComplianceTransaction ProcessingFinancial Services

  • 5+ years of work experience in data engineering.
  • Strong skills in SQL and Python.
  • Experience designing, building and maintaining data models and data pipelines.
  • Experience working with cloud based architecture.
  • Great communication skills with a diverse team of varying technical ability.
  • Create and maintain scalable data pipelines to ingest, transform and serve global payments and risk data.
  • Manage and maintain the data platform, including data pipelines and environments.
  • Collaborate with cross-functional teams of data scientists, software engineers, product managers and business leads, to understand requirements and deliver appropriate solutions.
  • Take ownership of a data area, building subject matter expertise and cultivating trust with stakeholders.
  • Mentor junior members, and grow a strong data culture across the team and organisation.

PythonSQLCloud ComputingETLData engineeringCommunication SkillsData modeling

Posted 3 days ago
Apply
Apply

πŸ“ States of SΓ£o Paulo and Rio Grande do Sul, Rio de Janeiro, Belo Horizonte

πŸ” Data Engineering

🏒 Company: TELUS Digital Brazil

  • At least 3 years of experience as Data Engineer
  • Have actively participated in the design and development of data architectures
  • Hands-on experience in developing and optimizing data pipelines
  • Experience working with databases and data modeling projects, as well as practical experience utilizing SQL
  • Effective English communication - able to explain technical and non-technical concepts to different audiences
  • Experience with a general-purpose programming language such as Python or Scala
  • Ability to work well in teams and interact effectively with others
  • Ability to work independently and manage multiple tasks simultaneously while meeting deadlines
  • Develop and optimize scalable, high-performing, secure, and reliable data pipelines that address diverse business needs and considerations
  • Identify opportunities to enhance internal processes, implement automation to streamline manual tasks, and contribute to infrastructure redesign
  • Act as a guide and mentor to junior engineers, supporting their professional growth and fostering an inclusive working environment
  • Collaborate with cross-functional teams to ensure data quality and support data-driven decision-making to strive for greater functionality in our data systems
  • Collaborate with project managers and product owners to assist in prioritizing, estimating, and planning development tasks
  • Provide constructive feedback, and share expertise with fellow team members, fostering mutual growth and learning
  • Engage in ongoing research and adoption of new technologies, libraries, frameworks, and best practices to enhance the capabilities of the data team
  • Demonstrate a commitment to accessibility and ensure that your work considers and positively impacts others

AWSDockerPythonSQLAgileApache AirflowCloud ComputingETLKubernetesData engineeringData scienceCommunication SkillsAnalytical SkillsTeamworkData modelingEnglish communication

Posted 6 days ago
Apply
Apply

πŸ“ Germany, Spain, United Kingdom, Austria

πŸ” Software Development

🏒 Company: LocalStackπŸ‘₯ 11-50πŸ’° $25,000,000 Series A 4 months agoCloud ComputingInformation TechnologySoftware

  • Ability and experience working with non technical stakeholders to gather requirements
  • Ability to define technical initiatives required to satisfy business requirements
  • Excellent knowledge of Python
  • Experience in designing real time data ingestion solutions with massive volumes of data
  • (preferred) Experience with AWS services commonly used in Data Engineering (like S3, ECS, Glue, EMR)
  • Experience with relational databases and data warehouses, data orchestration and ingestion tools, SQL, and BI tools
  • (preferred) Experience in working remotely/ in async settings
  • Experience owning initiatives at the IC level
  • Experience Providing guidance to junior engineers
  • Maintain, monitor, and optimize data ingestion pipelines for our current data platform.
  • Lead the development of our future data platform based on evolving business needs.
  • Shape the data team roadmap and contribute to long-term strategic planning.
  • Take full ownership of data ingestion from external sources, ensuring smooth functionality.
  • Design and implement a robust data modelling and data lake solution architecture.
  • Provide technical leadership and mentorship to the data engineering team.
  • Collaborate with engineering teams to define and refine ingestion pipeline requirements.
  • Work with stakeholders to gather business questions and data needs.

AWSDockerLeadershipPythonSQLApache AirflowETLKafkaData engineeringData StructuresREST APICommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingMentoringWritten communicationData visualizationTeam managementStakeholder managementData modeling

Posted 6 days ago
Apply
Apply

πŸ“ United Kingdom

🏒 Company: Pharma Universe

  • 5+ years in Data Warehousing (Kimball methodology)
  • Expertise in Azure Data Factory, Azure SQL Server, DevOps
  • Strong SQL skills for database management & reporting
  • Experience with data modeling, integration, and performance optimization
  • Agile & DevOps experience, with a problem-solving mindset
  • Architect and implement robust data solutions on Azure.
  • Develop and optimize ETL processes for seamless data integration.
  • Maintain data governance, ensuring accuracy, reliability, and compliance.
  • Collaborate with analysts to support Power BI reporting.
  • Own and maintain data documentation and models.

SQLETLAzureDevOpsData modelingData management

Posted 10 days ago
Apply
Apply

πŸ“ Madrid, Barcelona

πŸ” Software Development

🏒 Company: Clarity AI

  • 5+ years in data architecture, data engineering, or a related role, with hands-on experience in data modeling, schema design, and cloud-based data systems
  • Strong Python and SQL skills and proficiency with distributed data stores
  • Proficiency in schema design, data modeling, and building data products
  • Proficient in at least one major programming language, preferably Python, with a software engineering mindset for writing clean, maintainable code
  • Deep understanding of architectural principles in microservices, distributed systems, and data pipeline design
  • Familiarity with containerized environments, public/private API integrations, and security best practices
  • Strong communication and interpersonal skills, with experience working cross-functionally
  • Proven ability to guide teams and drive complex projects to successful completion
  • Self-starter, able to take ownership and initiative, with high energy and stamina
  • Decisive and action-oriented, able to make rapid decisions even when they are short of information
  • Highly motivated, independent and deeply passionate about sustainability and impact
  • Excellent oral and written English communication skills (minimum C1 level-proficient user)
  • Transforming raw data into intuitive, high-quality data models that support analytical and reporting needs across the organization. Applying scalable design patterns but avoid over-engineering, ensuring solutions remain efficient and fit-for-purpose
  • Implementing and monitor data quality checks to ensure the accuracy, completeness, and reliability of data across all systems
  • Collaborating across functions
  • Working closely with engineering and product teams to understand business requirements, translating them into scalable data solutions
  • Acting as a bridge between technical and non-technical stakeholders, ensuring alignment with strategic goals and effective communication of technical designs
  • Leading initiatives to improve data practices, from schema design to data governance, ensuring data quality, consistency, and security
  • Guiding the team in experimenting with new tools and technologies thoughtfully, focusing on understanding both the benefits and limitations of each option
  • Continuously evolving the data architecture for optimal performance, balancing scalability with cost-efficiency and reliability
  • Applying a pragmatic approach to performance metrics and scaling decisions, ensuring that the system remains performant without unnecessary complexity
  • Implementing performance metrics to monitor system health, proposing improvements where necessary
  • Maintaining comprehensive documentation of data systems, processes, and best practices to facilitate knowledge sharing and compliance

DockerPythonSQLApache AirflowCloud ComputingKubernetesAlgorithmsData engineeringData StructuresPostgresREST APICommunication SkillsAnalytical SkillsCI/CDProblem SolvingAgile methodologiesMicroservicesData visualizationData modelingData analyticsData management

Posted 20 days ago
Apply
Apply

πŸ“ Europe, APAC, Americas

🧭 Full-Time

πŸ” Software Development

🏒 Company: DockerπŸ‘₯ 251-500πŸ’° $105,000,000 Series C almost 3 years agoDeveloper ToolsDeveloper PlatformInformation TechnologySoftware

  • 4+ years of relevant industry experience
  • Experience with data modeling and building scalable pipelines
  • Proficiency with Snowflake or BigQuery
  • Experience with data governance and security controls
  • Experience creating ETL scripts using Python and SQL
  • Familiarity with a cloud ecosystem: AWS/Azure/Google Cloud
  • Experience with Tableau or Looker
  • Manage and develop ETL jobs, warehouse, and event collection tools
  • Build and manage the Central Data Model for reporting
  • Integrate emerging methodologies and technologies
  • Build data pipelines for ML and AI projects
  • Contribute to SOC2 compliance across the data platform
  • Document technical architecture

PythonSQLETLSnowflakeAirflowData engineeringData visualizationData modeling

Posted 23 days ago
Apply
Apply

πŸ“ Poland

🧭 Full-Time

πŸ” Software Development

🏒 Company: N-iXπŸ‘₯ 1001-5000IT Services and IT Consulting

  • Minimum of 3-4 years as data engineer, or in a relevant field
  • Advanced experience in Python, particularly in delivering production-grade data pipelines and troubleshooting code-based bugs
  • Structured approach to data insights
  • Familiarity with cloud platforms (preferably Azure)
  • Experience with Databricks, Snowflake, or similar data platforms
  • Knowledge of relational databases, with proficiency in SQL
  • Experience using Apache Spark
  • Experience in creating and maintaining structured documentation
  • Proficiency in utilizing testing frameworks to ensure code reliability and maintainability
  • Experience with Gitlab or equivalent tools
  • B2 level or higher English Proficiency
  • Strong collaboration abilities, experience in an international team environment, willing to learn new skills and tools, adaptive and exploring mindset
  • Design, build, and maintain data pipelines using Python
  • Collaborate with an international team to develop scalable data solutions
  • Conduct in-depth analysis and debugging of system bugs (Tier 2)
  • Develop and maintain smart documentation for process consistency, including the creation and refinement of checklists and workflows
  • Set up and configure new tenants, collaborating closely with team members to ensure smooth onboarding
  • Write integration tests to ensure the quality and reliability of data services
  • Work with Gitlab to manage code and collaborate with team members
  • Utilize Databricks for data processing and management

DockerPythonSQLCloud ComputingData AnalysisETLGitKubernetesSnowflakeApache KafkaAzureData engineeringRDBMSREST APIPandasCI/CDDocumentationMicroservicesDebugging

Posted 25 days ago
Apply
Apply

πŸ“ Berlin, Heidelberg

🧭 Full-Time

πŸ” Insurtech

🏒 Company: Getsafe

  • 4+ years of experience in creating data pipelines - especially ingestion - using SQL / Python / Airflow is required
  • Experience with designing Data Mart, Data Warehouse and database objects within relational databases
  • Experience running software in the cloud - including infrastructure as code via e.g. Terraform
  • Strong proficiency in Python and SQL
  • Good Problem-Solving, Verbal and Written communication skills
  • Excellent analytical, organizational skills and ability to work under pressure/deliver on tight deadlines is a must
  • Work in a time constrained environment to analyze, design, develop and deliver Data Warehouse solutions
  • Create ETL/ELT pipelines using Python and Airflow
  • Design, develop, maintain and support our Data Warehouse & BI platform within Getsafe using various data & BI tools
  • Build ad-hoc applications as needed to support more curious data users and to provide automation as possible
  • Develop collaborative relationships and work with key business sponsors, cloud resources to gather requirements and for the efficient resolution of requests.
  • Communicate technical and business topics, as appropriate, in a 360 degree fashion, when required; communicate using written, verbal and/or presentation materials as necessary.

PythonSQLApache AirflowETLTerraform

Posted 27 days ago
Apply
Apply

πŸ“ LATAM

🧭 Full-Time

πŸ” Fintech

  • 7+ years of experience with ETL, SQL, PowerBI, Tableau, or similar technologies
  • Strong understanding of data modeling, database design, and SQL
  • Experience working with Apache Kafka or MSK solution
  • Extensive experience delivering solutions on Snowflake or other cloud-based data warehouses, and generally an understanding of data warehousing technologies and event-driven architectures
  • Proficiency in Python/R and familiarity with modern data engineering practices
  • Strong analytical and problem-solving skills with a focus on delivering high-quality solutions
  • Experience with machine learning (ML) and building natural language interfaces for business data
  • Proven track record in a fast-paced Agile development environment
  • Ability to work autonomously while effectively engaging with multiple business teams and stakeholders
  • Design, develop, and maintain scalable data pipelines for ingesting, processing, and transforming large volumes of data from various sources
  • Implement data ingestion frameworks to efficiently collect data from internal and external sources
  • Optimize data pipelines for performance, reliability, and scalability
  • Develop and deliver scalable, unit-tested data assets and products that empower analysts and drive business workflows
  • Evaluate and continuously improve existing data products and solutions for performance, scalability and security
  • Experience in data quality management, including software implementation for data correction, reconciliation, and validation of data workflows to ensure accuracy and integrity in the data warehouse
  • Collaborate with engineers, data scientists, and product managers to analyze edge cases and plan for architectural scalability
  • Lead the deployment and maintenance of multiple data solutions such as business dashboards and machine learning models
  • Champion best practices in data development, design, and architecture
  • Conduct comprehensive code reviews, providing mentorship and meaningful feedback to junior team members
  • Collaborate with other team members to create and maintain process documentation, data flows, and ETL diagrams for both new and existing data pipelines and processes
  • Monitor data pipelines for performance, reliability, and security issues
  • Implement logging, monitoring, and alerting systems to detect and respond to data-related issues proactively
  • Drive the team's Agile process, ensuring high standards of productivity and collaboration

AWSPythonSQLAgileCloud ComputingETLMachine LearningSnowflakeTableauApache KafkaData engineeringREST APIData visualizationData modeling

Posted 28 days ago
Apply