Apply

Software Engineer, Data

Posted 1 day agoViewed

View full description

💎 Seniority level: Senior, 5+ years

📍 Location: United States

💸 Salary: 181000.0 - 209000.0 USD per year

🔍 Industry: Fintech

🏢 Company: Found

⏳ Experience: 5+ years

🪄 Skills: AWSPostgreSQLPythonSQLData AnalysisETLGCPMachine LearningAirflowData engineering

Requirements:
  • 5+ years experience in data infrastructure, data engineering, or analytics engineering roles.
  • Experience with relational databases (Postgres/MySQL), dbt, and Airflow
  • Advocated for and rolled out new technologies or open-source frameworks that empower data and product organizations to make data driven decisions.
  • Experience working with cloud-native big data infrastructure on the public cloud (GCP/AWS/Azure, BigQuery/Redshift/Synapse).
  • Ability to write clean and maintainable code (primarily Python).
Responsibilities:
  • Design, build, and operate large scale data infrastructure systems across multiple environments to store, aggregate, and progress large amounts of data.
  • Build a data platform-as-a-service for internal customers, ensuring data integrity, sanity, tagging, and discoverability.
  • Bridge the gap between engineering and analytics, helping inform the roadmap for data infrastructure for the company.
  • Implement various ETL infrastructures for the company’s topline metrics, reporting, and product functionality.
Apply

Related Jobs

Apply

📍 Worldwide

🧭 Full-Time

🔍 Software Development

🏢 Company: Mysten Labs

  • 5+ years of experience in systems and/or network programming, ideally in Rust, C or C++.
  • Experience designing and developing systems in one or more of the following areas: Distributed systems, storage/database systems, high performance systems, or networking protocols.
  • Production experience: operating and managing production systems, proficiency in system monitoring and alerting, with strong focus on reliability.
  • Bachelor’s degree in Computer Science, Engineering, or a related field, or equivalent practical experience.
  • Expertise in performance tuning and optimizing both system architecture and low-level services.
  • Proven ability to drive complex projects to completion.
  • Optimize Sui’s data infrastructure from end-to-end: write performance, storage footprint, read performance, scaling, reliability, and costs.
  • Build and maintain robust, high-performance RPC services to facilitate reading of blockchain data.
  • Design and build low-latency, and reliable subscription services
  • Collaborate with cross-functional teams and external partners to ensure seamless integration of data platform solutions with first-party applications and the ecosystem at large.

DockerBlockchainKubernetesSoftware ArchitectureC++AlgorithmsApache KafkaData engineeringData StructuresgRPCPostgresRustCI/CDRESTful APIsLinuxDevOpsMicroservicesJSONData modelingData management

Posted about 5 hours ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 205000.0 - 235000.0 USD per year

🔍 Fintech

🏢 Company: Found

  • 8+ years experience in data infrastructure, data engineering, or analytics engineering roles.
  • Experience with relational databases (Postgres/MySQL), dbt, and Airflow
  • Advocated for and rolled out new technologies or open-source frameworks that empower data and product organizations to make data driven decisions.
  • Experience working with cloud-native big data infrastructure on the public cloud (GCP/AWS/Azure, BigQuery/Redshift/Synapse).
  • Ability to write clean and maintainable code (primarily Python).
  • Strong communication and collaboration skills, with the ability to work effectively with cross-functional teams.
  • Design, build, and operate large scale data infrastructure systems across multiple environments to store, aggregate, and progress large amounts of data.
  • Build a data platform-as-a-service for internal customers, ensuring data integrity, sanity, tagging, and discoverability.
  • Bridge the gap between engineering and analytics, helping inform the roadmap for data infrastructure for the company.
  • Implement various ETL infrastructures for the company’s topline metrics, reporting, and product functionality.
  • Contribute to the development of best practices, standards, and frameworks for data engineering at Found.
  • Provide mentorship and guidance to help grow and develop the skills of the broader data team, fostering a culture of continuous learning and excellence.

AWSPostgreSQLPythonSQLCloud ComputingData AnalysisETLGCPMachine LearningMySQLAirflowAzureData engineeringCommunication SkillsCollaborationMentoringData visualizationData modeling

Posted 1 day ago
Apply
Apply

📍 United States

🧭 Full-Time

🔍 Software Development

🏢 Company: Anrok👥 51-100💰 $30,000,000 Series B 11 months agoSaaSSoftwareTax Preparation

  • Product-minded software engineer with 5+ years of experience
  • Experience talking to users and synthesizing needs
  • Deep understanding of data platforms and building for scale
  • Design, build, and scale systems that let customers reliably sync their Anrok financial data to their data warehouses (e.g. Snowflake, BigQuery), allowing them to reconcile invoicing, payment, and bank account data.
  • Interface directly with customers to ensure the formats and mechanisms are tailored to the domain, but still flexible enough to accommodate different accounting practices.
  • Increase the flexibility and scale of Anrok's internal data infrastructure, allowing us to more effectively leverage data in our decision making.

AWSBackend DevelopmentPostgreSQLPythonSQLCloud ComputingData AnalysisETLGitSnowflakeAlgorithmsData engineeringData StructuresREST APICI/CDMicroservicesData visualizationData modelingNodeJSSoftware EngineeringData analyticsData managementDebuggingCustomer Success

Posted 3 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 119200.0 - 280000.0 USD per year

🔍 Software Development

🏢 Company: Figma

  • 6+ years of experience designing and building distributed data infrastructure at scale.
  • Strong expertise in batch and streaming data processing technologies such as Spark, Flink, Kafka, or Airflow/Dagster.
  • A proven track record of impact-driven problem-solving in a fast-paced environment.
  • A strong sense of engineering excellence, with a focus on high-quality, reliable, and performant systems.
  • Excellent technical communication skills, with experience working across both technical and non-technical counterparts.
  • The ability to navigate ambiguity, take ownership, and drive projects from inception to execution.
  • Experience mentoring and supporting engineers, fostering a culture of learning and technical excellence.
  • Design and build large-scale distributed data systems that power analytics, AI/ML, and business intelligence.
  • Develop batch and streaming solutions to ensure data is reliable, efficient, and scalable across the company.
  • Manage data ingestion, movement, and processing through core platforms like Snowflake, our ML Datalake, and real-time streaming systems.
  • Improve data reliability, consistency, and performance, ensuring high-quality data for engineering, research, and business stakeholders.
  • Collaborate with AI researchers, data scientists, product engineers, and business teams to understand data needs and build scalable solutions.
  • Drive technical decisions and best practices for data ingestion, orchestration, processing, and storage.

PythonSQLKafkaMachine LearningSnowflakeAirflowData engineeringGoSparkData modelingData management

Posted 4 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 240000.0 - 265000.0 USD per year

🔍 Software Development

🏢 Company: TRM Labs👥 101-250💰 $70,000,000 Series B over 2 years agoCryptocurrencyComplianceBlockchainBig Data

  • 7+ years of hands-on experience in architecting distributed system architecture, guiding projects from initial ideation through to successful production deployment.
  • Exceptional programming skills in Python, as well as adeptness in SQL or SparkSQL.
  • In-depth experience with data stores such as Icerberg, Trino, BigQuery, and StarRocks, and Citus.
  • Proficiency in data pipeline and workflow orchestration tools like Airflow, DBT, etc.
  • Expertise in data processing technologies and streaming workflows including Spark, Kafka, and Flink.
  • Competence in deploying and monitoring infrastructure within public cloud platforms, utilizing tools such as Docker, Terraform, Kubernetes, and Datadog.
  • Proven ability in loading, querying, and transforming extensive datasets.
  • Build highly reliable data services to integrate with dozens of blockchains.
  • Develop complex ETL pipelines that transform and process petabytes of structured and unstructured data in real-time.
  • Design and architect intricate data models for optimal storage and retrieval to support sub-second latency for querying blockchain data.
  • Oversee the deployment and monitoring of large database clusters with an unwavering focus on performance and high availability.
  • Collaborate across departments, partnering with data scientists, backend engineers, and product managers to design and implement novel data models that enhance TRM’s products.

AWSDockerPythonSQLCloud ComputingETLKafkaKubernetesAirflowData engineeringPostgresSparkTerraformData modeling

Posted 5 days ago
Apply
Apply

📍 United States

🧭 Full-Time

💸 177000.0 - 213000.0 USD per year

🔍 FinTech

🏢 Company: Flex

  • A minimum of 6 years of industry experience in the data infrastructure/data engineering domain.
  • A minimum of 6 years of experience with Python and SQL.
  • A minimum of 3 years of industry experience using DBT.
  • A minimum of 3 years of industry experience using Snowflake and its basic features.
  • Familiarity with AWS services, with industry experience using Lambda, Step Functions, Glue, RDS, EKS, DMS, EMR, etc.
  • Industry experience with different big data platforms and tools such as Snowflake, Kafka, Hadoop, Hive, Spark, Cassandra, Airflow, etc.
  • Industry experience working with relational and NoSQL databases in a production environment.
  • Strong fundamentals in data structures, algorithms, and design patterns.
  • Design, implement, and maintain high-quality data infrastructure services, including but not limited to Data Lake, Kafka, Amazon Kinesis, and data access layers.
  • Develop robust and efficient DBT models and jobs to support analytics reporting and machine learning modeling.
  • Closely collaborating with the Analytics team for data modeling, reporting, and data ingestion.
  • Create scalable real-time streaming pipelines and offline ETL pipelines.
  • Design, implement, and manage a data warehouse that provides secure access to large datasets.
  • Continuously improve data operations by automating manual processes, optimizing data delivery, and redesigning infrastructure for greater scalability.
  • Create engineering documentation for design, runbooks, and best practices.

AWSPythonSQLBashDesign PatternsETLHadoopJavaKafkaSnowflakeAirflowAlgorithmsCassandraData engineeringData StructuresNosqlSparkCommunication SkillsCI/CDRESTful APIsTerraformWritten communicationDocumentationData modelingDebugging

Posted 10 days ago
Apply
Apply

📍 United States, Canada

🧭 Full-Time

🔍 Health-tech

🏢 Company: PHIL

  • 10+ years of progressive experience in software development, with focus on data. Experience within the health-tech sector is a plus.
  • Experience building and maintaining secure, end-to-end systems and services. Experience in Golang, PostgresDB, and MongoDB is preferred.
  • Should enjoy working with data and understanding the relationship with the business. Comfortable writing standard, maintainable SQL, from the basics to window functions
  • Deep understanding of algorithms and efficient data structures
  • Experience working with all phases of the Software Development Lifecycle
  • Strong critical thinking, debugging and problem solving skills
  • A sense of ownership: you would develop, manage, and evolve the core logic of a mission-critical and consequential consumer experience
  • Experience with Looker, Snowflake and dbt is a plus
  • Experience building systems and services for health-tech company is preferred
  • Design, implement, test, and launch PHIL’s backend data services and API platform for multiple client facing apps
  • Collaborate with product and analytics team to improve data quality and governance
  • Design and maintain data pipelines in Golang and Python
  • Optimize performance for data ingestion, transformation, and loading processes
  • Maintain comprehensive documentation of data systems, processes, and workflows
  • Integrate data sources from Phil’s core application, Sales CRM, Inventory tracking, and more
  • Work collaboratively with remote Dev and QA teams reviewing implementation and answering queries about the platform
  • Deeply understand complex business processes and end user needs (internal and external, technical and nontechnical)
  • Make data at Phil easy to understand and work with - whether you're an experienced analyst or a pharmacist
  • Design and test validation frameworks, monitor the results and improve upon them
  • Deliver performant, reliable, scalable, testable, and secure code

PythonSQLETLMongoDBAlgorithmsData engineeringData StructuresREST API

Posted 23 days ago
Apply
Apply

📍 United States, Canada

🧭 Full-Time

💸 140000.0 - 160000.0 USD per year

🔍 Fraud Prevention and AML Compliance

🏢 Company: Sardine👥 101-250💰 $70,000,000 Series C about 1 month agoCryptocurrencyFraud DetectionFinTechSoftware

  • 5+ years of experience in backend or data engineering roles
  • Strong knowledge of database systems (SQL and NoSQL)
  • Expertise in a modern programming language (Go, Python, Java)
  • Familiarity with cloud platforms (AWS, GCP, Azure)
  • Experience with containerization (Docker, Kubernetes)
  • Design and implement ETL pipelines for large datasets
  • Develop and optimize APIs for data retrieval
  • Architect and manage scalable storage solutions
  • Collaborate on data product development
  • Perform data analysis for client value
  • Document processes and mentor junior engineers

AWSDockerPythonSQLDynamoDBElasticSearchETLGCPKubernetesNosqlCI/CD

Posted about 1 month ago
Apply
Apply

📍 United States, Canada

🧭 Full-Time

🔍 Software Development

  • 7+ years of software development experience
  • Experience with Java and Python applications
  • Current cloud technology experience with AWS and Kubernetes
  • Develop core functionality using cloud-native Java
  • Work with Data Science teams on machine learning solutions
  • Ensure secure, efficient solutions in a determined timeframe

AWSDockerPostgreSQLPythonETLJavaKafkaMachine LearningSpringCI/CD

Posted about 1 month ago
Apply
Apply

📍 California, Colorado, Hawaii, Illinois, Maryland, Massachusetts, New York, Oregon, Texas, Washington

🧭 Full-Time

💸 150000.0 - 170000.0 USD per year

🔍 Recruiting and feedback tools

🏢 Company: Textio👥 51-100💰 $999,972 about 3 years ago🫂 Last layoff about 1 year agoArtificial Intelligence (AI)Human ResourcesMachine LearningEnterprise SoftwareNatural Language ProcessingSoftware

  • Hands-on experience shipping customer-facing features including reports and analytics
  • Solid experience with data warehouse software and cloud services (SQL, Redshift, AWS CDK, Meltano)
  • Track record of writing complex SQL queries for performance and efficiency
  • Ability to work with design and product to build user-friendly features
  • Collaboration skills in a diverse and inclusive environment
  • Fast-paced startup experience is a plus
  • Maintain a strong, user-centric approach to feature development
  • Work on ambiguous problems and advocate for clear solutions
  • Enhance data warehouse to meet business needs
  • Collaborate with diverse teams using AI/LLM technologies
  • Enable advanced analytics through real-time and batch processing
  • Improve scalability and performance of data pipelines
  • Implement monitoring, alerting, and self-healing mechanisms for data systems

AWSSQLETL

Posted about 2 months ago
Apply