Apply

Senior Data Engineer

Posted 4 days agoViewed

View full description

๐Ÿ’Ž Seniority level: Senior, 5+ years

๐Ÿ“ Location: United States

๐Ÿ’ธ Salary: 136000.0 - 190000.0 USD per year

๐Ÿ” Industry: Crypto

๐Ÿข Company: Gemini๐Ÿ‘ฅ 501-1000๐Ÿ’ฐ $1,000,000 Secondary Market over 2 years ago๐Ÿซ‚ Last layoff about 2 years agoCryptocurrencyWeb3Financial ServicesFinanceFinTech

๐Ÿ—ฃ๏ธ Languages: English

โณ Experience: 5+ years

๐Ÿช„ Skills: AWSPythonSQLApache AirflowCloud ComputingETLKafkaSnowflakeAlgorithmsData engineeringData StructuresCI/CDData modeling

Requirements:
  • 5+ years experience in data engineering with data warehouse technologies
  • 5+ years experience in custom ETL design, implementation and maintenance
  • 5+ years experience with schema design and dimensional data modeling
  • Experience building real-time data solutions and processes
  • Advanced skills with Python and SQL are a must
  • Experience with one or more MPP databases(Redshift, Bigquery, Snowflake, etc)
  • Experience with one or more ETL tools(Informatica, Pentaho, SSIS, Alooma, etc)
  • Strong computer science fundamentals including data structures and algorithms
  • Strong software engineering skills in any server side language, preferable Python
  • Experienced in working collaboratively across different teams and departments
  • Strong technical and business communication
Responsibilities:
  • Design, architect and implement best-in-class Data Warehousing and reporting solutions
  • Lead and participate in design discussions and meetings
  • Mentor data engineers and analysts
  • Design, automate, build, and launch scalable, efficient and reliable data pipelines into production using Python
  • Build real-time data and reporting solutions
  • Design, build and enhance dimensional models for Data Warehouse and BI solutions
  • Research new tools and technologies to improve existing processes
  • Develop new systems and tools to enable the teams to consume and understand data more intuitively
  • Partner with engineers, project managers, and analysts to deliver insights to the business
  • Perform root cause analysis and resolve production and data issues
  • Create test plans, test scripts and perform data validation
  • Tune SQL queries, reports and ETL pipelines
  • Build and maintain data dictionary and process documentation
Apply

Related Jobs

Apply

๐Ÿ“ United States, Canada

๐Ÿ” Software Development

๐Ÿข Company: Overstory๐Ÿ‘ฅ 1-10E-Commerce

  • Approximately 5 years of experience in Data Engineering with at least one experience in a startup environment
  • Product-minded and able to demonstrate significant impact you have had on a business through the application of technology
  • Proven experience of data engineering across the following (or similar) technologies: Python, data orchestration platforms (Airflow, Luigi, Dagster, etcโ€ฆ), data quality frameworks, data lakes/warehouses
  • Ability to design and implement scalable and resilient data systems
  • Excellent communication skills and ability to collaborate effectively in a cross-functional team environment
  • Passion for learning and staying updated with evolving technologies and industry trends
  • Owning day-to-day operational responsibilities of deliveries our analysis to the customers
  • Developing data-driven solutions to customer problems that our products arenโ€™t solving for yet
  • Building new and improving existing technologies such as:
  • Automation of the analysis for all customers, leading to faster implementation of Overstoryโ€™s recommendations
  • Metrics to identify what are the time bottlenecks in the current flow of analysis, therefore helping all Overstory teams identify areas of improvements
  • Visualization of status and progress of the analysis for internal use
  • Working on performance & scalability of our pipelines ensuring that our tech can handle our growth

PythonSQLCloud ComputingGCPAmazon Web ServicesData engineeringCommunication SkillsAnalytical SkillsRESTful APIsData visualizationData modeling

Posted 3 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ” SaaS

๐Ÿข Company: Mangomint๐Ÿ‘ฅ 51-100๐Ÿ’ฐ $35,000,000 Series B 6 months agoManagement Information SystemsBeautySoftware

  • 3+ years of experience in data engineering or a related role
  • Proficiency in SQL and Python for data pipelines and automation
  • Experience with dbt (or similar data modeling tools)
  • Familiarity with Snowflake (or other cloud data warehouses)
  • Knowledge of APIs and experience integrating various data sources
  • Experience with CRM and business systems (Salesforce, Outreach, Stripe, etc.)
  • Strong problem-solving skills and ability to take ownership of projects
  • Ability to work independently in a small, fast-paced startup environment
  • Effective communication skills to translate business needs into technical solutions
  • Design, develop, and maintain ETL/ELT data pipelines using Snowflake, dbt, Prefect, and other modern data tools
  • Automate data workflows to improve efficiency and reliability
  • Integrate CRM and other business systems to support cross-functional needs
  • Develop data enrichment pipelines to power our sales process
  • Build internal data tools to drive data-driven decision making
  • Work directly with stakeholders to define requirements and implement data solutions that support business objectives
  • Ensure data integrity, governance, and security best practices are upheld
  • Support analytics and reporting efforts by building dashboards and data models in dbt and Sigma

AWSPythonSQLETLSnowflakeCRMData modeling

Posted 4 days ago
Apply
Apply

๐Ÿ“ USA

๐Ÿงญ Full-Time

๐Ÿ’ธ 160000.0 - 182000.0 USD per year

๐Ÿ” Adtech

๐Ÿข Company: tvScientific๐Ÿ‘ฅ 11-50๐Ÿ’ฐ $9,400,000 Convertible Note about 1 year agoInternetAdvertising

  • 7+ years of experience in data engineering.
  • Proven experience building data infrastructure using Spark with Scala.
  • Familiarity with data lakes, cloud warehouses, and storage formats.
  • Strong proficiency in AWS services.
  • Expertise in SQL for data manipulation and extraction.
  • Bachelor's degree in Computer Science or a related field.
  • Design and implement robust data infrastructure using Spark with Scala.
  • Collaborate with our cross-functional teams to design data solutions that meet business needs.
  • Build out our core data pipelines, store data in optimal engines and formats, and feed our machine learning models.
  • Leverage and optimize AWS resources.
  • Collaborate closely with the Data Science team.

AWSSQLCloud ComputingETLMachine LearningData engineeringSparkScalaData modeling

Posted 4 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ’ธ 110000.0 - 130000.0 USD per year

๐Ÿ” Software Development

๐Ÿข Company: Ceros๐Ÿ‘ฅ 101-250๐Ÿ’ฐ $100,000,000 Private over 4 years agoAdvertisingContent CreatorsContent MarketingGraphic DesignSoftware

  • 5+ years of experience in data engineering, focusing on AWS Redshift and ETL pipeline development.
  • Strong expertise in SQL performance tuning, schema management, and query optimization.
  • Experience designing and maintaining ETL pipelines using AWS Glue, Matillion, or similar tools.
  • Proficiency in JavaScript/TypeScript, with experience building custom ETL workflows and integrations.
  • Hands-on experience with Python for data automation and scripting.
  • Strong understanding of data warehousing best practices, ensuring high-quality, scalable data models.
  • Experience with data monitoring and alerting tools such as AWS CloudWatch and New Relic.
  • Ability to work independently in a fast-paced environment, collaborating across teams to support data-driven initiatives.
  • Own and lead the management of AWS Redshift, ensuring optimal performance, disk usage, and cost efficiency.
  • Design and maintain scalable ETL pipelines using AWS Glue, Lambda, and Matillion to integrate data from Mixpanel, CRM platforms, and customer engagement tools.
  • Optimize SQL-based data transformations and Redshift queries to improve performance and reliability.
  • Automate data offloading and partition management, leveraging AWS services like S3 and external schemas.
  • Ensure version control and documentation of all Redshift queries, ETL processes, and AWS configurations through a centralized GitHub repository.
  • Develop monitoring and alerting for data pipelines using CloudWatch and other observability tools to ensure high availability and early issue detection.
  • Implement and maintain data quality checks and governance processes to ensure accuracy and consistency across foundational tables.
  • Collaborate with AI engineers and business stakeholders to enhance data accessibility and reporting for internal teams.
  • Maintain and optimize BI dashboards in Metabase and HubSpot, ensuring accuracy and efficiency of business reporting.
  • Manage key integrations between Redshift and external platforms, including Mixpanel, HubSpot, and Census, optimizing data accessibility and performance.
  • Administer AWS infrastructure supporting Redshift, ensuring efficient resource utilization, IAM security, and cost management.
  • Automate repetitive data tasks using Python and scripting to enhance data processes and improve team efficiency.

AWSPythonSQLETLGitJavascriptTypeScriptAmazon Web ServicesAPI testingData engineeringREST APICI/CDAnsibleData modelingData analyticsData management

Posted 4 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ’ธ 113000.0 - 130000.0 USD per year

๐Ÿ” Software Development

๐Ÿข Company: Later๐Ÿ‘ฅ 1-10Consumer ElectronicsiOSAppsSoftware

  • Minimum of 5 years in data engineering or related fields
  • Bachelorโ€™s degree in Computer Science, Engineering, or related field
  • Strong focus on building data infrastructure and pipelines
  • Design and build a robust data warehouse architecture
  • Design, build, and maintain scalable data pipelines
  • Develop reliable transformation layers and data pipelines
  • Establish optimized data architectures using cloud technologies
  • Enforce data quality checks and governance practices
  • Collaborate with cross-functional teams to deliver insights
  • Analyze and optimize data pipelines for performance

SQLData engineering

Posted 7 days ago
Apply
Apply

๐Ÿ“ United States

๐Ÿงญ Full-Time

๐Ÿ’ธ 150000.0 - 190000.0 USD per year

๐Ÿ” Game Development

๐Ÿข Company: Second Dinner

  • Demonstrated experience in large-scale distributed data systems such as Spark and Flink
  • Deep expertise in analytical database technologies, including SQL and NoSQL
  • Experience with database technologies and ETL/ELT processes
  • Experience with orchestration and automation tools like Airflow and Beam
  • Experience with Databricks and AWS-based data/analytics solutions
  • Develop and operate data infrastructure and pipelines for analytics and reporting
  • Empower Marketing team with high-quality data for user acquisition and retention
  • Collaborate with teams to gain insights from analytics

AWSPythonSQLApache AirflowETLData engineeringNosqlSpark

Posted 8 days ago
Apply
Apply

๐Ÿ“ United States

๐Ÿงญ Full-Time

๐Ÿ’ธ 145000.0 - 160000.0 USD per year

๐Ÿ” EdTech

๐Ÿข Company: Magpie Literacy

  • 8+ years of experience in data engineering or related fields
  • Advanced SQL expertise
  • Proficiency in Python
  • Experience with BI tools and dashboard creation
  • Understanding of data warehouse concepts
  • Lead the transformation of reporting needs into automated data pipelines
  • Develop reusable reporting components
  • Design and implement modular SQL and dbt models
  • Oversee automated reporting workflows
  • Collaborate with stakeholders on reporting opportunities

AWSPythonSQLETLGit

Posted 8 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Full-Time

๐Ÿ” B2B SaaS

๐Ÿข Company: Sanity๐Ÿ‘ฅ 51-200๐Ÿ’ฐ Corporate over 2 years agoSoftware Development

  • 4+ years of experience building data pipelines at scale
  • Deep expertise in SQL, Python, and Node.js/TypeScript
  • Production experience with Airflow and RudderStack
  • Track record of building reliable data infrastructure
  • Design, develop, and maintain scalable ETL/ELT pipelines
  • Collaborate to implement and scale product telemetry
  • Establish best practices for data ingestion and transformation
  • Monitor and optimize data pipeline performance

Node.jsPythonSQLApache AirflowETLTypeScript

Posted 13 days ago
Apply
Apply

๐Ÿ“ United States, Canada

๐Ÿงญ Fulltime

๐Ÿ” Software Development

๐Ÿข Company: BioRender๐Ÿ‘ฅ 101-250๐Ÿ’ฐ $15,319,133 Series A almost 2 years agoLife ScienceGraphic DesignSoftware

  • 7+ years of data engineering experience of relevant industry experience
  • Expertise working with Data Warehousing platforms (AWS RedShift or Snowflake preferred) and data lake / lakehouse architectures
  • Experience with Data Streaming platforms (AWS Kinesis / Firehose preferred)
  • Expertise with SQL and programming languages commonly used in data platforms (Python, Spark, etc)
  • Experience with data pipeline orchestration (e.g., Airflow) and data pipeline integrations (e.g. Airbyte, Stitch)
NOT STATED

AWSPythonSQLApache AirflowSnowflakeData engineeringSparkData modeling

Posted 17 days ago
Apply
Apply

๐Ÿ“ United States

๐Ÿงญ Full-Time

๐Ÿ’ธ 175000.0 - 205000.0 USD per year

๐Ÿ” Software Development

๐Ÿข Company: CoreWeave๐Ÿ’ฐ $642,000,000 Secondary Market over 1 year agoCloud ComputingMachine LearningInformation TechnologyCloud Infrastructure

  • Hands-on experience applying Kimball Dimensional Data Modeling principles to large datasets.
  • Expertise in working with analytical table/file formats, including Iceberg, Parquet, Avro, and ORC.
  • Proven experience optimizing MPP databases (StarRocks, Snowflake, BigQuery, Redshift).
  • Minimum 5+ years of programming experience in Python or Scala.
  • Advanced SQL skills, with a strong ability to write, optimize, and debug complex queries.
  • Hands-on experience with Airflow for batch orchestration distributed computing frameworks like Spark or Flink.
  • Develop and maintain data models, including star and snowflake schemas, to support analytical needs across the organization.
  • Establish and enforce best practices for dimensional modeling in our Lakehouse.
  • Engineer and optimize data storage using analytical table/file formats (e.g., Iceberg, Parquet, Avro, ORC).
  • Partner with BI, analytics, and data science teams to design datasets that accurately reflect business metrics.
  • Tune and optimize data in MPP databases such as StarRocks, Snowflake, BigQuery, or Redshift.
  • Collaborate on data workflows using Airflow, building and managing pipelines that power our analytical infrastructure.
  • Ensure efficient processing of large datasets through distributed computing frameworks like Spark or Flink.

AWSDockerPythonSQLCloud ComputingETLKubernetesSnowflakeAirflowAlgorithmsApache KafkaData engineeringData StructuresREST APISparkCommunication SkillsAnalytical SkillsCollaborationCI/CDRESTful APIsDevOpsTerraformProblem-solving skillsJSONScalaData visualizationAnsibleData modelingData analyticsDebugging

Posted 22 days ago
Apply