Apply

Staff Data Engineer

Posted 2024-08-17

View full description

πŸ’Ž Seniority level: Staff, 6+ years

πŸ“ Location: US

πŸ’Έ Salary: $176,500 - $263,000 per year

πŸ” Industry: Financial technology

🏒 Company: EarnIn

πŸ—£οΈ Languages: English

⏳ Experience: 6+ years

πŸͺ„ Skills: AWSLeadershipPythonSQLApache HadoopETLGCPGitHadoopKafkaAzureData engineeringNosqlSparkCI/CDTerraform

Requirements:
  • 6+ years in designing, building, and maintaining Data infrastructure and the ability to lead complex projects and teams.
  • Bachelor's, Master's or PhD degree in computer science, computer engineering, or a related technical discipline or equivalent industry experience.
  • Deep knowledge of Kafka.
  • Proficiency in programming languages like Python and Scala.
  • Strong knowledge of distributed computing frameworks such as Apache Hadoop and Apache Spark with cloud platforms such as AWS, Azure, or GCP.
  • Working experience with Databricks would be nice to have.
  • Deep understanding of database design, SQL, and NoSQL databases.
  • Experience in managing large datasets and optimizing database performance.
  • Proficiency in Git and Terraform and experience in deploying continuous integration and continuous deployment (CI/CD) practices.
  • Experience in managing event-driven systems, particularly with Kafka in cloud environments.
  • Expertise in developing and implementing data governance frameworks, policies, and procedures to ensure data quality, compliance, and effective data management practices.
  • Deep understanding of data security principles, including encryption, decryption, and secure data storage and transfer protocol.
Responsibilities:
  • Lead the enhancement of internal processes, focusing on scaling infrastructure, streamlining data delivery, and implementing automation to replace manual operations.
  • Design and implement advanced infrastructure for efficient data extraction, transformation, and loading (ETL) using cutting-edge AWS and SQL technologies.
  • Develop sophisticated analytical tools that tap into the data pipeline, offering deep insights into crucial business metrics like customer growth and operational efficiency.
  • Architect and manage extensive, sophisticated data sets to meet complex business needs and requirements.
  • Engage closely with a wide array of stakeholders, including executive, product, data, and design teams, providing high-level support for data infrastructure challenges and advising on technical data issues.
  • Collaborate and mentor other senior engineers while providing thoughtful guidance using code, design, and architecture reviews.
  • Contribute to defining technical direction, planning the roadmap, escalating issues, and synthesizing feedback to ensure team success.
  • Estimate and manage team project timelines and risks.
  • Participate in hiring and onboarding for new team members.
  • Lead cross-team engineering initiatives.
  • Constantly learning about new technologies and industry standards in data engineering.
Apply

Related Jobs

Apply

πŸ“ US

🧭 Full-Time

πŸ’Έ 206700 - 289400 USD per year

πŸ” Social media / Online community

  • MS or PhD in a quantitative discipline: engineering, statistics, operations research, computer science, informatics, applied mathematics, economics, etc.
  • 7+ years of experience with large-scale ETL systems, building clean, maintainable, object-oriented code (Python preferred).
  • Strong programming proficiency in Python, SQL, Spark, Scala.
  • Experience with data modeling, ETL concepts, and manipulating large structured and unstructured data.
  • Experience with data workflows (e.g., Airflow) and data visualization tools (e.g., Looker, Tableau).
  • Deep understanding of technical and functional designs for relational and MPP databases.
  • Proven track record of collaboration and excellent communication skills.
  • Experience in mentoring junior data scientists and analytics engineers.

  • Act as the analytics engineering lead within Ads DS team and contribute to data science data quality and automation initiatives.
  • Ensure high-quality data through ETLs, reporting dashboards, and data aggregations for business tracking and ML model development.
  • Develop and maintain robust data pipelines and workflows for data ingestion, processing, and transformation.
  • Create user-friendly tools for internal use across Data Science and cross-functional teams.
  • Lead efforts to build a data-driven culture by enabling data self-service.
  • Provide mentorship and coaching to data analysts and act as a thought partner for data teams.

LeadershipPythonSQLData AnalysisETLTableauStrategyAirflowData analysisData engineeringData scienceSparkCommunication SkillsCollaborationMentoringCoaching

Posted 2024-11-21
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ’Έ 120000 - 134999 USD per year

πŸ” Nonprofit/Movement Organizing

🏒 Company: The Movement Cooperative

  • 5+ years of relevant software and/or data engineering experience.
  • 2+ years leading technical projects.
  • Deep familiarity with Python and SQL.
  • Experience architecting and maintaining complex environments with both custom-built and third-party applications.
  • Experience with implementing development best practices, including CI/CD, testing, and observability.
  • Comfort with data storage and managing containerized orchestration (e.g. S3, Redshift, BigQuery, EC2, Kubernetes) on AWS, Azure, or GCP.
  • Ability to coach and mentor engineers and influence organizational leadership.

  • Take a lead role in architecting data pipelines, with an emphasis on interoperability of tool data.
  • Architect new systems, own tool selection, and lead technical aspects of the team vision.
  • Create design patterns for other teams to deliver member-specific work.
  • Work with external partners toward common data best practices.
  • Drive best practices for software development and dev ops.
  • Implement CI/CD best practices and develop an observability strategy.
  • Lead the design and maintenance of the data warehouse, ensuring cost savings and performance.
  • Mentor and support other members of the team, increasing the overall depth of engineering knowledge.

AWSLeadershipPythonSoftware DevelopmentSQLDesign PatternsGCPKubernetesStrategyAzureData engineeringCI/CD

Posted 2024-11-08
Apply
Apply

πŸ“ Canada, United States, United Kingdom

πŸ” Smart home technology

🏒 Company: ecobee

  • 10+ years of experience in data/software engineering with proven track record.
  • Extensive experience in building and maintaining scalable data pipelines with tools like Apache Spark, Kafka, and Airflow.
  • Expertise in cloud data platforms (AWS, GCP, or Azure), focusing on distributed systems.
  • Solid understanding of end-to-end data systems and machine learning deployment.
  • Knowledge in data security, governance, and compliance.
  • Experience in data architecture and engineering methodologies across industries.
  • Experience with real-time data processing and analytics platforms.
  • Proven ability to mentor and guide engineers across teams.

  • Lead the design and implementation of scalable data pipelines and systems for complex problems.
  • Contribute to ecobee’s system architecture with impactful designs.
  • Take end-to-end ownership of components within your domain.
  • Define and track SLAs for components to ensure reliability.
  • Mentor engineers and drive knowledge-sharing initiatives.
  • Collaborate across squads to align technical discussions with strategic goals.
  • Anticipate future data challenges and propose strategies.
  • Evaluate and recommend new technologies for data engineering.

AWSLeadershipGCPIoTKafkaMachine LearningStrategyAirflowAzureData engineeringSparkCollaboration

Posted 2024-11-07
Apply
Apply

πŸ“ ANY COUNTRY WITH A PHYSICAL PRESENCE

🧭 Full-Time

πŸ’Έ 206700 - 289400 USD per year

πŸ” Internet

🏒 Company: Reddit

  • MS or PhD in a quantitative discipline: engineering, statistics, operations research, computer science, informatics, applied mathematics, economics, etc.
  • 7+ years of experience with large-scale ETL systems, building clean and maintainable code (Python preferred).
  • Strong programming proficiency in Python, SQL, Spark, Scala, etc.
  • Experience with data modeling, ETL concepts, and handling structured and unstructured data.
  • Experience with data workflows (Airflow), front-end or back-end engineering, and data visualization tools.
  • Understanding of relational and MPP databases and proven track record of cross-functional collaboration.
  • Experience mentoring junior data scientists and analytics engineers; self-starter able to work autonomously and in teams.

  • Act as the analytics engineering lead within Ads DS team and a key contributor to data quality and automation initiatives.
  • Work on ETLs, reporting dashboards, and data aggregations for business tracking and ML model development.
  • Develop and maintain robust data pipelines and workflows for data ingestion and transformation.
  • Create user-friendly tools and applications for internal use, streamlining data analysis processes.
  • Lead efforts to enable data self-service and build a data-driven culture at Reddit.
  • Provide technical guidance and mentorship to data analysts and serve as a thought partner for various teams.

LeadershipPythonSQLData AnalysisETLTableauStrategyAirflowData analysisData engineeringData scienceSparkCommunication SkillsCollaborationMentoringCoaching

Posted 2024-11-07
Apply
Apply

πŸ“ AR, CA, CO, FL, GA, IL, KY, MA, MI, MT, MO, NV, NJ, NY, NC, OR, PA, TX, WA, WI

πŸ” Food waste reduction and grocery technology

🏒 Company: Afresh

  • 6+ years of experience as a data engineer, analytics engineer, or similar role.
  • Strong understanding of advanced SQL concepts.
  • Exceptional communication and leadership skills.
  • 1+ years of experience with SQL-driven transform libraries supporting ELT, including CI/CD pipelines.
  • Expert knowledge of OLTP and OLAP database design.
  • Familiarity with data engineering concepts like Data Mesh, Data Lake, Data Warehouse.
  • Experience with semantic layer setup defined with code (LookML, Cube.dev, etc.).
  • Technologies: SQL, Python, Airflow, dbt, Snowflake/Databricks/BigQuery, Spark.

  • Improve and extend data analytics architecture for reliable data across use cases.
  • Collaborate with engineers, product managers, and data scientists to understand data needs.
  • Build dimensional models and metrics for consistent insights.
  • Evolve existing data quality and governance processes.
  • Mentor and up-skill other engineers.

LeadershipPythonSQLSnowflakeAirflowData engineeringSparkCollaborationCI/CD

Posted 2024-10-21
Apply