Apply

Data Engineer

Posted about 23 hours agoViewed

View full description

πŸ’Ž Seniority level: Junior, 2-3 years

πŸ“ Location: Ukraine, Poland

πŸ” Industry: Green Energy

🏒 Company: Globaldev GroupπŸ‘₯ 251-500πŸ’° Seed almost 4 years agoDevOpsInformation TechnologySoftware

πŸ—£οΈ Languages: English

⏳ Experience: 2-3 years

πŸͺ„ Skills: PythonSQLETLTableauData engineeringREST APIData visualizationData modeling

Requirements:
  • Proven experience with REST APIs, including working with custom APIs and data extraction
  • Expertise in Databricks, Python, SQL, and Data Factory
  • Familiarity with Tableau is a plus
  • 2-3 years of hands-on experience in data engineering or related fields
  • Strong proficiency in English communication skills (written and spoken)
Responsibilities:
  • Work with REST APIs to integrate and pull data from various custom APIs
  • Design, develop, and maintain data pipelines using Databricks, Python, SQL, and Data Factory
  • Collaborate with cross-functional teams to understand data requirements and provide data solutions
  • Ensure high-quality, scalable, and efficient data models and processes
  • Support the integration of Tableau for data visualization and reporting (if applicable)
  • Troubleshoot and optimize data processing and storage systems
  • Ensure data security, accuracy, and reliability across all pipelines and systems
Apply

Related Jobs

Apply

πŸ“ Poland

🧭 Full-Time

πŸ” Software Development

🏒 Company: Craft Machine Inc

  • 2+ years of experience in Data Engineering.
  • 2+ years of experience with Python.
  • Experience in developing, maintaining, and ensuring the reliability, scalability, fault tolerance and observability of data pipelines in a production environment.
  • Strong knowledge of SDLC and solid software engineering practices.
  • Knowledge of and experience with Amazon Web Services (AWS) and Databricks.
  • Demonstrated curiosity through asking questions, digging into new technologies, and always trying to grow.
  • Strong problem solving and the ability to communicate ideas effectively.
  • Familiar with infrastructure-as-code approach.
  • Self-starter, independent, likes to take initiative.
  • Have fundamental knowledge of data engineering techniques: ETL/ELT, batch and streaming, DWH, Data Lakes, distributed processing.
  • Familiarity with at least some technologies in our current tech stack: Python, PySpark, Pandas, SQL (PostgreSQL), Airflow, Docker, Databricks & AWS (S3, Batch, Athena, RDS, DynamoDB, Glue, ECS), CircleCI, GitHub, Terraform
  • Building and optimizing data pipelines (batch and streaming).
  • Extracting, analyzing and modeling of rich and diverse datasets.
  • Designing software that is easily testable and maintainable.
  • Support in setting data strategies and our vision.
  • Keep track of emerging technologies and trends in the Data Engineering world, incorporating modern tooling and best practices at Craft.
  • Work on extendable data processing systems that allows to add and scale pipelines.
  • Applying machine learning techniques such as anomaly detection, clustering, regression classification, summarization to extract value from our data sets.

AWSDockerPostgreSQLPythonSQLETLMachine LearningAirflowAmazon Web ServicesData engineeringPandasCI/CDTerraformData modelingSoftware Engineering

Posted 4 days ago
Apply
Apply

πŸ“ Worldwide

🧭 Full-Time

πŸ” AI

🏒 Company: ElevenLabsπŸ‘₯ 101-250πŸ’° $180,000,000 Series C about 2 months agoArtificial Intelligence (AI)Developer APIsContent CreatorsGenerative AI

  • A track record of partnering with RevOps and Finance teams to translate business challenges into data-driven solutions, ensuring alignment on key performance metrics.
  • Proficiency with tools across the modern data stack (python, SQL, BI tools, dbt)
  • familiarity with Salesforce, Gong, Stripe, Netsuite APIs
  • Develop robust ETL processes that integrate data from various sources (CRM, ERP, marketing platforms, financial systems) to ensure that RevOps and Finance have reliable, timely data.
  • Implement automated data validation and cleansing processes to maintain high-quality datasets, reducing errors that could impact financial reporting or revenue forecasting.
  • Create and maintain data models that drive key performance indicators (KPIs) for revenue operations and finance.
  • Streamline regular reporting tasks by automating data extractions and report generation, ensuring stakeholders have access to real-time insights.
  • Maintain thorough documentation of data pipelines, models, and analytical methodologies to facilitate transparency and ensure consistency across teams.

PythonSQLApache AirflowData AnalysisETLData engineeringRESTful APIsData visualizationCRMData modelingFinanceData management

Posted 17 days ago
Apply
Apply

πŸ“ Poland

πŸ” Software Development

  • Experience working with SQL and Python
  • Familiarity with modern data stack tools such as Snowflake, dbt, ThoughtSpot, and AWS
Design, implement, and optimize data workflows

AWSPythonSQLSnowflakeData engineering

Posted 22 days ago
Apply
Apply

πŸ“ Poland

🧭 Full-Time

πŸ” Banking

🏒 Company: CapcoπŸ‘₯ 101-250Electric VehicleProduct DesignMechanical EngineeringManufacturing

  • 5+ years of experience with Python
  • 2+ years of experience with Spark and Scala
  • Experience with agile continuous integration/ DevOps tools (Git, Jenkins, Jira)
  • Develop end-to-end ETL processes with Spark/Scala
  • Use Scrum methodology as part of a high-performance team
  • Document solutions in tools like JIRA and Confluence

DockerPostgreSQLPythonSQLAgileApache AirflowETLGitJenkinsKubernetesData engineeringSparkScala

Posted 23 days ago
Apply
Apply

πŸ“ Worldwide

🧭 Full-Time

πŸ’Έ 167471.0 USD per year

πŸ” Software Development

🏒 Company: Float.com

  • Expertise in ML, expert systems, and advanced algorithms (e.g., pattern matching, optimization) with applied experience in Scheduling, Recommendations, or Personalization.
  • Proficient in Python or Java and comfortable with SQL and Javascript/Typescript.
  • Experience with large-scale data pipelines and stream processing (e.g., Kafka, Debezium, Flink).
  • Skilled in data integration, cleaning, and validation.
  • Familiar with vector and graph databases (e.g., Neo4j).
  • Lead technical viability discussions
  • Develop and test proof-of-concepts for this project.
  • Conduct a comprehensive analysis of existing data to uncover patterns, identify optimization opportunities, and support the squad’s next deliveries.
  • Evaluate our data streaming pipeline: Y
  • Lead technical discussions related to optimization, pattern detection, and AI, serving as the primary point of contact for these areas within Float.
  • Develop and implement advanced algorithms to enhance the Resource Recommendation Engine and other product features, initially focused on pattern detection and optimization.
  • Design, implement, and maintain our streaming data architecture to support real-time data processing and analytics, ensuring data integrity and reliability.
  • Establish best practices and standards for optimization, AI, and data engineering development within the organization.
  • Mentor and train team members on optimization, AI, and data engineering concepts and techniques, fostering a culture of continuous learning and innovation.
  • Stay updated with the latest trends and related technologies, and proactively identify opportunities to incorporate them into Float's solutions.

PythonSQLKafkaMachine LearningAlgorithmsData engineering

Posted 23 days ago
Apply
Apply

πŸ“ Europe, APAC, Americas

🧭 Full-Time

πŸ” Software Development

🏒 Company: DockerπŸ‘₯ 251-500πŸ’° $105,000,000 Series C almost 3 years agoDeveloper ToolsDeveloper PlatformInformation TechnologySoftware

  • 4+ years of relevant industry experience
  • Experience with data modeling and building scalable pipelines
  • Proficiency with Snowflake or BigQuery
  • Experience with data governance and security controls
  • Experience creating ETL scripts using Python and SQL
  • Familiarity with a cloud ecosystem: AWS/Azure/Google Cloud
  • Experience with Tableau or Looker
  • Manage and develop ETL jobs, warehouse, and event collection tools
  • Build and manage the Central Data Model for reporting
  • Integrate emerging methodologies and technologies
  • Build data pipelines for ML and AI projects
  • Contribute to SOC2 compliance across the data platform
  • Document technical architecture

PythonSQLETLSnowflakeAirflowData engineeringData visualizationData modeling

Posted 26 days ago
Apply
Apply

πŸ“ Argentina, Brazil, Bulgaria, Colombia, Poland, Romania

🧭 Contract

πŸ” Software Development

🏒 Company: N-iXπŸ‘₯ 1001-5000IT Services and IT Consulting

  • Bachelor’s or Master’s degree in Computer Science, Software Engineering, or related fields.
  • 3+ years of experience in developing and deploying data services for HPC or similar systems.
  • Proven expertise with HDF5 or similar, in parallel I/O operations. Equivalent experience in distributed systems is also applicable.
  • Programming: Strong proficiency in (at least one): C++, Python, GoLang, or Fortran.
  • HDF5 Expertise: In-depth knowledge of HDF5 APIs and advanced features like parallel HDF5.
  • Parallel I/O: Experience with MPI I/O, POSIX I/O, or similar frameworks for concurrent/parallel data access.
  • Performance Optimization: Skills in profiling and optimizing I/O operations for large datasets.
  • Proficiency in SQL and experience with any RDMS
  • Design and implement the data service module using HDF5 for efficient data storage and retrieval.
  • Develop parallel and concurrent I/O mechanisms to optimize performance for large-scale datasets.
  • Ensure the module is tightly integrated with HPC and visualization workflows.
  • Optimize I/O operations for CPU/GPU-based workflows to minimize bottlenecks.
  • Implement caching, compression, and other strategies to enhance performance.
  • Design data structures and schemas suitable for storing 3D grid data and other simulation outputs.
  • Ensure data integrity and consistency during concurrent read/write operations.
  • Develop and execute test cases to validate module performance and reliability under various load conditions.
  • Conduct benchmarking to ensure scalability across different hardware configurations.
  • Document the architecture, APIs, and usage guidelines for the data service module.
  • Provide technical support to the development and visualization teams for data integration.

PythonSQLFortranC++Data engineeringData Structures

Posted 28 days ago
Apply
Apply

πŸ“ Ukraine

🧭 Contract

πŸ” Data Engineering

🏒 Company: N-iXπŸ‘₯ 1001-5000IT Services and IT Consulting

  • 2+ years experience with SQL
  • Good Python skills
  • Hands-on experience with PySpark (or Spark)
  • Good knowledge of Data Warehouse concepts
  • Intermediate+ English level
  • Ensure that all data is getting refreshed automatically daily as per SLA and data quality is being permanently monitored
  • Troubleshoot and resolve complex issues of Foundry users within SLAs
  • Solve small change requests and bug fixes self-sufficiently
  • Provide a standard and ad-hoc reports about pipeline monitoring, issues, health checks and system usage
  • Create new/missing documentation based on the activities done to share newly gained knowledge within the team, as well as keep up-to-date the existing documentation
  • Ensure immediate communications in case of Foundry outages and delays, but also for upcoming changes and other support information to end users.
  • Implement continuous improvement of existing processes

PythonSQL

Posted 28 days ago
Apply
Apply

πŸ“ Poland

🧭 Full-Time

πŸ” Software Development

🏒 Company: N-iXπŸ‘₯ 1001-5000IT Services and IT Consulting

  • Minimum of 3-4 years as data engineer, or in a relevant field
  • Advanced experience in Python, particularly in delivering production-grade data pipelines and troubleshooting code-based bugs
  • Structured approach to data insights
  • Familiarity with cloud platforms (preferably Azure)
  • Experience with Databricks, Snowflake, or similar data platforms
  • Knowledge of relational databases, with proficiency in SQL
  • Experience using Apache Spark
  • Experience in creating and maintaining structured documentation
  • Proficiency in utilizing testing frameworks to ensure code reliability and maintainability
  • Experience with Gitlab or equivalent tools
  • B2 level or higher English Proficiency
  • Strong collaboration abilities, experience in an international team environment, willing to learn new skills and tools, adaptive and exploring mindset
  • Design, build, and maintain data pipelines using Python
  • Collaborate with an international team to develop scalable data solutions
  • Conduct in-depth analysis and debugging of system bugs (Tier 2)
  • Develop and maintain smart documentation for process consistency, including the creation and refinement of checklists and workflows
  • Set up and configure new tenants, collaborating closely with team members to ensure smooth onboarding
  • Write integration tests to ensure the quality and reliability of data services
  • Work with Gitlab to manage code and collaborate with team members
  • Utilize Databricks for data processing and management

DockerPythonSQLCloud ComputingData AnalysisETLGitKubernetesSnowflakeApache KafkaAzureData engineeringRDBMSREST APIPandasCI/CDDocumentationMicroservicesDebugging

Posted 28 days ago
Apply
Apply

πŸ“ Ukraine, Canada

🧭 Full-Time

πŸ” Software Development

🏒 Company: N-iXπŸ‘₯ 1001-5000IT Services and IT Consulting

  • 8+ years of experience in software engineering
  • Proven experience leading complex data projects, organizing teams, and architecting and designing new solutions
  • Extensive and proven expertise in AWS services, especially S3, Glue, Athena
  • Proficiency in Python, PySpark, GCP
  • Experience with SQL, Airflow, data streaming frameworks (Kafka or Firehose)
  • Experience in building and maintaining data pipelines
  • Experience in building data platforms, not just using them
  • Proficiency in data modeling techniques and best practices
  • Simplify existing tools, make them more modular, work with technical debt
  • Propose improvements, be actively involved in team discussions and decision-making, drive initiatives
  • Work closely with a team of Senior Engineers, Product Manager and Engineering Manager

AWSLeadershipPythonSQLApache AirflowGCPApache KafkaData modelingSoftware Engineering

Posted 28 days ago
Apply