Apply

Staff Software Engineer, Data

Posted 3 months agoViewed

View full description

πŸ’Ž Seniority level: Staff, 11+ years

πŸ“ Location: U.S.

πŸ’Έ Salary: 190000.0 - 244000.0 USD per year

πŸ” Industry: Commercial aviation

🏒 Company: AcubedπŸ‘₯ 51-100Innovation ManagementAerospaceManufacturing

πŸ—£οΈ Languages: English

⏳ Experience: 11+ years

πŸͺ„ Skills: PythonSQLETLMachine LearningData engineeringPandasData modeling

Requirements:
  • Bachelor's or Master's degree in Computer Science, Software Engineering, or related field.
  • Strong coding skills in Python (Pandas, PySpark) and experience with SQL and analytics databases.
  • 11+ years of experience in software/data engineering, focusing on data quality and ownership.
  • Deep understanding of data pipelines with experience in building and optimizing them.
  • Familiarity with sensor data processing (visual, IR cameras, LiDAR) and embeddings.
  • Proficiency in machine learning-based labeling techniques.
  • Experience developing metrics and dashboards for data performance tracking.
  • Knowledge of data management technologies (e.g., ETL, data lakes).
  • Excellent problem-solving skills and attention to detail in safety-critical environments.
Responsibilities:
  • Identify and address gaps in data coverage for robust ML model performance.
  • Develop and implement metrics to monitor and improve data quality.
  • Collaborate with data scientists and ML engineers to refine datasets.
  • Implement machine learning-based labeling techniques for sensor data accuracy.
  • Design and maintain data pipelines for high-quality data ingestion.
  • Partner with cross-functional teams to enhance data availability.
Apply

Related Jobs

Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ” Software Development

  • 7+ years of software development experience
  • Experience with Java and Python applications
  • Current cloud technology experience with AWS and Kubernetes
  • Develop core functionality using cloud-native Java
  • Work with Data Science teams on machine learning solutions
  • Ensure secure, efficient solutions in a determined timeframe

AWSDockerPostgreSQLPythonETLJavaKafkaMachine LearningSpringCI/CD

Posted 16 days ago
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ” Blockchain intelligence and financial technology

🏒 Company: TRM LabsπŸ‘₯ 101-250πŸ’° $70,000,000 Series B over 2 years agoCryptocurrencyComplianceBlockchainBig Data

  • Bachelor's degree (or equivalent) in Computer Science or a related field.
  • 5+ years of hands-on experience in architecting scalable API development and distributed system architecture.
  • Exceptional programming skills in Python and proficiency in SQL or SparkSQL.
  • In-depth experience with data stores such as BigQuery and Postgres.
  • Proficiency in data pipeline tools like Airflow and DBT.
  • Expertise in data processing technologies including Dataflow, Spark, Kafka, and Flink.
  • Competence in deploying and monitoring infrastructure using tools like Docker, Terraform, Kubernetes, and Datadog.
  • Proven ability in loading, querying, and transforming extensive datasets.
  • Build highly scalable features integrating with multiple blockchains.
  • Design intricate data models for optimal storage and retrieval supporting sub-second latency for querying blockchain data.
  • Collaborate across departments with data scientists, backend engineers, and product managers to enhance TRM’s products.

DockerPythonSQLApache AirflowKafkaKubernetesPostgresSparkTerraform

Posted 3 months ago
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ’Έ $240,000 - $270,000 per year

πŸ” Blockchain intelligence data platform

  • Bachelor's degree (or equivalent) in Computer Science or a related field.
  • 5+ years of experience in building distributed system architecture, with a particular focus on incremental updates from inception to production.
  • Strong programming skills in Python and SQL.
  • Deep technical expertise in advanced data structures and algorithms for incremental updating of data stores (e.g., Graphs, Trees, Hash Maps).
  • Comprehensive knowledge across all facets of data engineering, including implementing and managing incremental updates in data stores like BigQuery, Snowflake, RedShift, Athena, Hive, and Postgres.
  • Orchestrating data pipelines and workflows focused on incremental processing using tools such as Airflow, DBT, Luigi, Azkaban, and Storm.
  • Developing and optimizing data processing technologies and streaming workflows for incremental updates (e.g., Spark, Kafka, Flink).
  • Deploying and monitoring scalable, incremental update systems in public cloud environments (e.g., Docker, Terraform, Kubernetes, Datadog).
  • Expertise in loading, querying, and transforming large datasets with a focus on efficiency and incremental growth.
  • Design and build our Cloud Data Warehouse with a focus on incremental updates to improve cost efficiency and scalability.
  • Research innovative methods to incrementally optimize data processing, storage, and retrieval to support efficient data analytics and insights.
  • Develop and maintain ETL pipelines that transform and incrementally process petabytes of structured and unstructured data to enable data-driven decision-making.
  • Collaborate with cross-functional teams to design and implement new data models and tools focused on accelerating innovation through incremental updates.
  • Continuously monitor and optimize the Data Platform's performance, focusing on enhancing cost efficiency, scalability, and reliability.

DockerPythonSQLETLKafkaKubernetesMachine LearningSnowflakeAirflowAlgorithmsData engineeringData scienceData StructuresPostgresSparkCollaborationTerraformData analytics

Posted 5 months ago
Apply
Apply

πŸ“ United States

πŸ’Έ 200000 - 255000 USD per year

πŸ” Blockchain intelligence and financial services

🏒 Company: TRM LabsπŸ‘₯ 101-250πŸ’° $70,000,000 Series B over 2 years agoCryptocurrencyComplianceBlockchainBig Data

  • A Bachelor's degree (or equivalent) in Computer Science or a related field.
  • 8+ years of hands-on experience in architecting distributed systems.
  • Exceptional programming skills in Python and proficiency in SQL or SparkSQL.
  • In-depth experience with data stores such as Iceberg, Trino, BigQuery, and StarRocks.
  • Proficiency in tools like Airflow and DBT for data pipeline orchestration.
  • Expertise in technologies like Spark, Kafka, and Flink.
  • Competence in deploying and monitoring cloud infrastructure with tools like Docker and Terraform.
  • Proven ability in managing extensive datasets.
  • Build highly reliable data services to integrate with blockchains.
  • Develop complex ETL pipelines for real-time data processing.
  • Design intricate data models for optimal storage and retrieval.
  • Oversee deployment and monitoring of large database clusters.
  • Collaborate with data scientists and engineers to enhance products.

DockerPythonSQLBlockchainETLKafkaKubernetesAirflowData engineeringSparkCollaborationTerraform

Posted 6 months ago
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ’Έ 240000 - 270000 USD per year

πŸ” Blockchain intelligence and financial services

🏒 Company: TRM LabsπŸ‘₯ 101-250πŸ’° $70,000,000 Series B over 2 years agoCryptocurrencyComplianceBlockchainBig Data

  • Bachelor's degree (or equivalent) in Computer Science or related field.
  • 5+ years of experience in building distributed system architecture, focusing on incremental updates.
  • Strong programming skills in Python and SQL.
  • Deep technical expertise in advanced data structures and algorithms for incremental updating.
  • Knowledge of data stores like BigQuery, Snowflake, RedShift, and more.
  • Experience orchestrating data pipelines using Airflow, DBT, Luigi, etc.
  • Proficiency in data processing technologies like Spark, Kafka, Flink.
  • Ability to deploy and monitor systems in public cloud environments.
  • Design and build our Cloud Data Warehouse focusing on incremental updates for cost efficiency.
  • Research methods to optimize data processing, storage, and retrieval.
  • Develop and maintain ETL pipelines for structured and unstructured data.
  • Collaborate with teams on new data models and tools for innovation.
  • Continuously monitor and optimize performance for cost, scalability, and reliability.

DockerPythonSQLETLKafkaKubernetesSnowflakeAirflowAlgorithmsData engineeringData StructuresPostgresSparkCollaborationTerraform

Posted 7 months ago
Apply