Senior Data Engineer - AI-enabled Data Platforms (PySpark)

New
Remotely within PolandSenior
Salary not disclosed
Apply NowOpens the employer's application page

Job Details

Languages
English (B2+)
Experience
5+ years
Required Skills
PythonAgileMicrosoft Power BISCRUMAzureNosqlCI/CDDevOpsAzure DevOpsPySpark

Requirements

  • 5+ years of experience in Data Software Engineering or related engineering roles
  • Solid experience with PySpark and SparkSQL
  • Experience with Cosmos DB (NoSQL API)
  • Hands-on experience with OneLake or Delta Lake and OpenLake concepts
  • Knowledge of DF Gen2 and M-code
  • Experience with CI/CD pipelines using Azure DevOps or equivalent
  • Good understanding of Azure services
  • Experience integrating data solutions with Power BI
  • Experience with Azure Fabric would be an asset
  • Strong problem-solving and analytical skills
  • Ability to work independently on complex tasks
  • Experience working in Agile or Scrum environments
  • Upper-intermediate proficiency in English (B2+)

Responsibilities

  • Implement data processing and transformation with Python, PySpark and SparkSQL
  • Work with OneLake (Delta / OpenLake) for efficient data storage and analytics
  • Develop and support solutions using Cosmos DB (NoSQL API)
  • Contribute to Fabric workloads including Data Engineering, Data Factory Gen2 and Lakehouse
  • Design, develop and maintain scalable data pipelines using Azure Fabric
  • Implement and maintain CI/CD pipelines and follow DevOps best practices
  • Integrate data solutions with Power BI for reporting and analytics
  • Collaborate with AI, data science and product teams to support AI-driven use cases
  • Ensure data quality, performance, security and reliability
  • Participate in Agile ceremonies and contribute to sprint delivery
  • Support production issues and continuous improvements
View Full Description & ApplyYou'll be redirected to the employer's site
View details
Apply Now