Apply

Lead Data Engineer

Posted 2024-10-16

View full description

💎 Seniority level: Lead, 8 years

📍 Location: India

🔍 Industry: Digital engineering and modernization

🏢 Company: Encora

🗣️ Languages: English

⏳ Experience: 8 years

🪄 Skills: AWSLeadershipPythonSQLData AnalysisETLHadoopJenkinsAirflowData analysisData science

Requirements:
  • 7-10 years of strong development experience performing ETL and/or data pipeline implementations.
  • Expert in programming languages, preferably Python.
  • Expert in delivering end-to-end analytic solutions using AWS services (EMR, Airflow, S3, Athena, Kinesis, Redshift).
  • Experience in batch technologies like Hadoop, Hive, Athena, Presto.
  • Strong SQL skills, including query optimization, schema design, complex analytics.
  • Expert in data modeling and metadata management like Glue Catalog etc.
  • Experience in deployment tools like GitHub actions, Jenkins, AWS Code Pipeline etc.
  • Experience in data quality tools like Deque or Great Expectations is Nice To Have.
Responsibilities:
  • Collaborate and partner with Business Analyst teams located in US and EMEA regions.
  • Interface across our Business Analyst and Data Science teams.
  • Play a key role in integrating new data sources into our data & analytical ecosystem over AWS cloud.
  • Implement data lake solutions while addressing common data concerns, such as data quality, data governance.
  • Set the standard for technical excellence as we move / build our data ecosystem into the cloud.
  • Understand their common data problems and deliver scalable solutions.
Apply

Related Jobs

Apply
🔥 Lead Data Engineer
Posted 2024-10-13

📍 India

🧭 Full-Time

🔍 Healthcare technology

🏢 Company: Reveal Health Tech

  • Bachelor’s degree in Computer Science, Engineering, or related field.
  • Proficiency in data modeling and using AWS services.
  • Proven experience as a Data Engineer or similar role.
  • Strong knowledge of AWS services related to data engineering.
  • Experience designing and building data pipelines.
  • Understanding of data lake and warehousing best practices.
  • Proficiency in programming languages for data manipulation (e.g., Python, SQL).
  • Experience with relational databases like PostgreSQL.
  • Understanding of database systems and performance optimization.
  • Experience implementing data quality checks.
  • Familiarity with Cloudformation and CI/CD for deployment.
  • Experience with data quality validation tools.
  • Ability to work independently and collaboratively.

  • Designing and implementing scalable, robust, and maintainable data pipelines using AWS services.
  • Implementing data quality checks and validation processes.
  • Designing data models and architectures for data processing workflows.
  • Utilizing validation tools to automate processes.
  • Collaborating with cross-functional teams to meet data requirements.
  • Monitoring and optimizing data pipeline performance.
  • Debugging and resolving data-related issues.
  • Documenting data pipeline architectures and processes.
  • Staying updated with new technologies in data engineering.

AWSPostgreSQLPythonSQLDynamoDBMachine LearningData engineeringServerlessCI/CD

Posted 2024-10-13
Apply