Apply

Senior Software Engineer (Data Platform)

Posted 3 months agoViewed

View full description

💎 Seniority level: Senior, 5+ years

📍 Location: USA

💸 Salary: $169,000 - $240,000 per year

🔍 Industry: Consumer banking

🗣️ Languages: Python, Kotlin

⏳ Experience: 5+ years

🪄 Skills: AWSPythonSoftware DevelopmentKafkaKotlinKubernetesStrategyAirflowNosqlSpark

Requirements:
  • 5+ years experience in software development or related fields.
  • Experience building and owning large-scale, geographically distributed compute and data processing systems.
  • Experience building and owning data lake solutions like Iceberg, Hudi, Delta, etc.
  • Experience building and managing Workflow Orchestration frameworks like Airflow, Flyte, Prefect, Temporal, Luigi, etc.
  • Experience with or working knowledge for efficiently scaling frameworks like Spark/Flink for extremely large scale datasets on Kubernetes.
  • Skilled at developing and debugging in Python/Kotlin or the ability to learn them quickly.
  • Working knowledge of Relational and NoSQL databases.
  • Experience with AWS and/or other cloud providers.
  • Ability to collaborate with your team in a fast-paced environment.
  • Passion and drive to change consumer banking for the better!
Responsibilities:
  • Help support the Data Platform that forms the backbone for several thousands of offline workloads at Affirm.
  • Design and build data infrastructure systems, services and tools to handle new Affirm products and business requirements that securely scale over millions of users and their transactions.
  • Build frameworks and services which will be used by other engineering teams at Affirm to manage billions of dollars in loans and power customer experiences.
  • Improve the reliability and efficiency of the Data Platform at scale and high reliability.
  • Engage other teams at Affirm about their use of the Data platform to ensure we are always building the right thing.
Apply

Related Jobs

Apply

📍 United States

🧭 Full-Time

💸 165000.0 - 210000.0 USD per year

🔍 Software Development / Data Platform

🏢 Company: Temporal Technologies👥 51-200💰 $75,000,000 Series B almost 2 years agoSoftware Development

  • At minimum 7 years of industry (post graduate) experience across the data stack (ingest, storage, compute, data modeling, visualization).
  • 0-1 experience designing and architecting a data platform for a SaaS business.
  • Experience owning key components of SLA bearing production data platform.
  • Experience building out data lake architectures.
  • Deep expertise within at least one of the major cloud providers (AWS, GCP, Azure).
  • Experience working with a wide range of data sources (APIs, logs, event stores, etc.).
  • Extreme proficiency in Python and SQL, additional languages a plus.
  • Experience with multiple data processing and query engines (Spark, Presto/Trino, Athena, BigQuery, etc.).
  • Significant experience with both object stores and relational databases such as S3 and Redshift.
  • Ability to quickly gain proficiency in new tools and technologies.
  • Strong desire to continue to learn and experiment.
  • Strong communicator and collaborator with business impact focus.

  • Designing and architecting the data platform for scalability and near-term value.
  • Building data pipelines for event-driven and batch workloads.
  • Implementing data quality checks in existing pipelines.
  • Modeling data for OLAP purposes.
  • Evaluating and recommending data processing tools.
  • Owning projects through collaboration with engineers and business stakeholders.
  • Creating guidelines for data access and ingestion.
  • Monitoring performance of data operations.
  • Contributing to the data platform roadmap.
  • Creating actionable dashboards.
  • Training stakeholders on new data products.
  • Mentoring junior engineers.

AWSPythonSQLETLGCPData engineeringSparkData modeling

Posted 16 days ago
Apply
Apply

📍 San Francisco Bay Area, Seattle, India, UK

💸 150000.0 - 180000.0 USD per year

🔍 B2B technology

  • Four-year degree in Computer Science, or related field OR equivalent experience.
  • Understanding of data strategies, data analysis, and data model design.
  • Experience designing and building low latency analytics APIs.
  • Proficiency in at least one JVM language (Java, Scala, Kotlin, etc.).
  • Familiarity with the Spark Ecosystem (YARN, Executors, Livy, etc.).
  • Data orchestration frameworks, particularly Airflow or similar.
  • Experience with columnar data stores, particularly Parquet and StarRocks.
  • Strong SDLC principles (CI/CD, Unit Testing, git, etc.).
  • General understanding of AWS EMR, EC2, S3.

  • Design and build the next generation of Demandbase’s Unified Data Platform.
  • Develop data pipelines for ingestion, normalization, and analysis.
  • Integrate 3rd party and open source tools into the data platform.
  • Build DAGs in Airflow for orchestration and monitoring of data pipelines.

AWSApache AirflowJavaKafkaKotlinSparkTerraformScalaData modeling

Posted 18 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 190800.0 - 267100.0 USD per year

🔍 Technology / Social Media

🏢 Company: Reddit👥 1001-5000💰 $410,000,000 Series F over 3 years ago🫂 Last layoff over 1 year agoNewsContentSocial NetworkSocial Media

  • 4+ years of software engineering experience in a production setting writing clean, maintainable, and well-tested code.
  • Proficient in object-oriented programming languages like Python, Scala, and experienced in Go.
  • Expertise in SQL languages such as BigQuery, SparkSQL, or Postgres.
  • Experience in designing and implementing large-scale systems and proactive leadership.
  • Familiarity with cloud services, GCP products, terraform, airflow, Kubernetes, CI/CD, and modern cloud infrastructure.
  • Excellent communication skills for collaboration within a service-oriented team and broader organizational context.

  • Collaborate effectively with a team of proficient software engineers to develop and maintain the fundamental platform that powers Reddit's data warehouse infrastructure.
  • Engage in the complete data lifecycle at Reddit, participating in the development process and working with extensive datasets.
  • Design, build, and deliver end-to-end data solutions to improve the reliability, scalability, latency, and efficiency of Reddit’s Data Platform.
  • Implement automation for key elements of the development process, including data quality, managing alerts, and handling critical infrastructure operations.
  • Collaborate and share on-call responsibilities, including incident management, with the Data Warehouse team.
  • Guide and support fellow engineers by mentoring and contributing to knowledge sharing through training sessions and documentation.

PythonSQLApache AirflowGCPKubernetesGoCI/CDTerraformScala

Posted 29 days ago
Apply
Apply

📍 United States

🔍 Life sciences

  • Applicants must have the unrestricted right to work in the United States.
  • Veeva will not provide sponsorship at this time.

  • Spearhead the development of new architecture for the Data platform from the ground up.
  • Design and build a resilient, scalable cloud-based platform along with its accompanying tools.
  • Empower Opendata teams to efficiently create and distribute valuable data assets.
  • Exercise end-to-end ownership for the project.

Backend DevelopmentLeadershipSoftware DevelopmentCross-functional Team LeadershipCommunication SkillsAnalytical SkillsCollaboration

Posted about 2 months ago
Apply
Apply

📍 USA

🧭 Full-Time

💸 169000 - 240000 USD per year

🔍 Financial services

  • 5+ years of industry experience in building large scale production systems.
  • Experience building and owning large-scale stream processing systems.
  • Experience building and operating robust and highly available infrastructure.
  • Working knowledge of Relational and NoSQL databases.
  • Experience working with Data Warehouse solutions.
  • Experience with industry standard stream processing frameworks like Spark, Samza, Flink, Beam etc.
  • Experience leading technical projects and mentoring junior engineers.
  • Exceptionally collaborative with a history of delivering complex technical projects and working closely with stakeholders.
  • This position requires either equivalent practical experience or a Bachelor’s degree in a related field.

  • Help support the Data Platform that forms the backbone for several thousands of offline workloads at Affirm.
  • Design and build data infrastructure systems, services, and tools to handle new Affirm products and business requirements that securely scale over millions of users and their transactions.
  • Build frameworks and services which will be used by other engineering teams at Affirm to manage billions of dollars in loans and power customer experiences.
  • Improve the reliability and efficiency of the Data Platform at scale and high reliability.
  • Engage other teams at Affirm about their use of the Data platform to ensure we are always building the right thing.

Backend DevelopmentLeadershipSoftware DevelopmentSQLData AnalysisElasticSearchKafkaCross-functional Team LeadershipApache KafkaSparkCollaboration

Posted 2 months ago
Apply