Apply

Senior Data Engineer

Posted 4 days agoViewed

View full description

πŸ’Ž Seniority level: Senior, 5+ years

πŸ“ Location: United States, Canada, EST

πŸ’Έ Salary: 110000.0 - 130000.0 USD per year

πŸ” Industry: Software Development

🏒 Company: CerosπŸ‘₯ 101-250πŸ’° $100,000,000 Private over 4 years agoAdvertisingContent CreatorsContent MarketingGraphic DesignSoftware

πŸ—£οΈ Languages: English

⏳ Experience: 5+ years

πŸͺ„ Skills: AWSPythonSQLETLGitJavascriptTypeScriptAmazon Web ServicesAPI testingData engineeringREST APICI/CDAnsibleData modelingData analyticsData management

Requirements:
  • 5+ years of experience in data engineering, focusing on AWS Redshift and ETL pipeline development.
  • Strong expertise in SQL performance tuning, schema management, and query optimization.
  • Experience designing and maintaining ETL pipelines using AWS Glue, Matillion, or similar tools.
  • Proficiency in JavaScript/TypeScript, with experience building custom ETL workflows and integrations.
  • Hands-on experience with Python for data automation and scripting.
  • Strong understanding of data warehousing best practices, ensuring high-quality, scalable data models.
  • Experience with data monitoring and alerting tools such as AWS CloudWatch and New Relic.
  • Ability to work independently in a fast-paced environment, collaborating across teams to support data-driven initiatives.
Responsibilities:
  • Own and lead the management of AWS Redshift, ensuring optimal performance, disk usage, and cost efficiency.
  • Design and maintain scalable ETL pipelines using AWS Glue, Lambda, and Matillion to integrate data from Mixpanel, CRM platforms, and customer engagement tools.
  • Optimize SQL-based data transformations and Redshift queries to improve performance and reliability.
  • Automate data offloading and partition management, leveraging AWS services like S3 and external schemas.
  • Ensure version control and documentation of all Redshift queries, ETL processes, and AWS configurations through a centralized GitHub repository.
  • Develop monitoring and alerting for data pipelines using CloudWatch and other observability tools to ensure high availability and early issue detection.
  • Implement and maintain data quality checks and governance processes to ensure accuracy and consistency across foundational tables.
  • Collaborate with AI engineers and business stakeholders to enhance data accessibility and reporting for internal teams.
  • Maintain and optimize BI dashboards in Metabase and HubSpot, ensuring accuracy and efficiency of business reporting.
  • Manage key integrations between Redshift and external platforms, including Mixpanel, HubSpot, and Census, optimizing data accessibility and performance.
  • Administer AWS infrastructure supporting Redshift, ensuring efficient resource utilization, IAM security, and cost management.
  • Automate repetitive data tasks using Python and scripting to enhance data processes and improve team efficiency.
Apply

Related Jobs

Apply

πŸ“ United States, Canada

πŸ” Software Development

🏒 Company: OverstoryπŸ‘₯ 1-10E-Commerce

  • Approximately 5 years of experience in Data Engineering with at least one experience in a startup environment
  • Product-minded and able to demonstrate significant impact you have had on a business through the application of technology
  • Proven experience of data engineering across the following (or similar) technologies: Python, data orchestration platforms (Airflow, Luigi, Dagster, etc…), data quality frameworks, data lakes/warehouses
  • Ability to design and implement scalable and resilient data systems
  • Excellent communication skills and ability to collaborate effectively in a cross-functional team environment
  • Passion for learning and staying updated with evolving technologies and industry trends
  • Owning day-to-day operational responsibilities of deliveries our analysis to the customers
  • Developing data-driven solutions to customer problems that our products aren’t solving for yet
  • Building new and improving existing technologies such as:
  • Automation of the analysis for all customers, leading to faster implementation of Overstory’s recommendations
  • Metrics to identify what are the time bottlenecks in the current flow of analysis, therefore helping all Overstory teams identify areas of improvements
  • Visualization of status and progress of the analysis for internal use
  • Working on performance & scalability of our pipelines ensuring that our tech can handle our growth

PythonSQLCloud ComputingGCPAmazon Web ServicesData engineeringCommunication SkillsAnalytical SkillsRESTful APIsData visualizationData modeling

Posted 3 days ago
Apply
Apply

πŸ“ United States

πŸ’Έ 136000.0 - 190000.0 USD per year

πŸ” Crypto

🏒 Company: GeminiπŸ‘₯ 501-1000πŸ’° $1,000,000 Secondary Market over 2 years agoπŸ«‚ Last layoff about 2 years agoCryptocurrencyWeb3Financial ServicesFinanceFinTech

  • 5+ years experience in data engineering with data warehouse technologies
  • 5+ years experience in custom ETL design, implementation and maintenance
  • 5+ years experience with schema design and dimensional data modeling
  • Experience building real-time data solutions and processes
  • Advanced skills with Python and SQL are a must
  • Experience with one or more MPP databases(Redshift, Bigquery, Snowflake, etc)
  • Experience with one or more ETL tools(Informatica, Pentaho, SSIS, Alooma, etc)
  • Strong computer science fundamentals including data structures and algorithms
  • Strong software engineering skills in any server side language, preferable Python
  • Experienced in working collaboratively across different teams and departments
  • Strong technical and business communication
  • Design, architect and implement best-in-class Data Warehousing and reporting solutions
  • Lead and participate in design discussions and meetings
  • Mentor data engineers and analysts
  • Design, automate, build, and launch scalable, efficient and reliable data pipelines into production using Python
  • Build real-time data and reporting solutions
  • Design, build and enhance dimensional models for Data Warehouse and BI solutions
  • Research new tools and technologies to improve existing processes
  • Develop new systems and tools to enable the teams to consume and understand data more intuitively
  • Partner with engineers, project managers, and analysts to deliver insights to the business
  • Perform root cause analysis and resolve production and data issues
  • Create test plans, test scripts and perform data validation
  • Tune SQL queries, reports and ETL pipelines
  • Build and maintain data dictionary and process documentation

AWSPythonSQLApache AirflowCloud ComputingETLKafkaSnowflakeAlgorithmsData engineeringData StructuresCI/CDData modeling

Posted 4 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ” SaaS

🏒 Company: MangomintπŸ‘₯ 51-100πŸ’° $35,000,000 Series B 6 months agoManagement Information SystemsBeautySoftware

  • 3+ years of experience in data engineering or a related role
  • Proficiency in SQL and Python for data pipelines and automation
  • Experience with dbt (or similar data modeling tools)
  • Familiarity with Snowflake (or other cloud data warehouses)
  • Knowledge of APIs and experience integrating various data sources
  • Experience with CRM and business systems (Salesforce, Outreach, Stripe, etc.)
  • Strong problem-solving skills and ability to take ownership of projects
  • Ability to work independently in a small, fast-paced startup environment
  • Effective communication skills to translate business needs into technical solutions
  • Design, develop, and maintain ETL/ELT data pipelines using Snowflake, dbt, Prefect, and other modern data tools
  • Automate data workflows to improve efficiency and reliability
  • Integrate CRM and other business systems to support cross-functional needs
  • Develop data enrichment pipelines to power our sales process
  • Build internal data tools to drive data-driven decision making
  • Work directly with stakeholders to define requirements and implement data solutions that support business objectives
  • Ensure data integrity, governance, and security best practices are upheld
  • Support analytics and reporting efforts by building dashboards and data models in dbt and Sigma

AWSPythonSQLETLSnowflakeCRMData modeling

Posted 4 days ago
Apply
Apply

πŸ“ USA

🧭 Full-Time

πŸ’Έ 160000.0 - 182000.0 USD per year

πŸ” Adtech

🏒 Company: tvScientificπŸ‘₯ 11-50πŸ’° $9,400,000 Convertible Note about 1 year agoInternetAdvertising

  • 7+ years of experience in data engineering.
  • Proven experience building data infrastructure using Spark with Scala.
  • Familiarity with data lakes, cloud warehouses, and storage formats.
  • Strong proficiency in AWS services.
  • Expertise in SQL for data manipulation and extraction.
  • Bachelor's degree in Computer Science or a related field.
  • Design and implement robust data infrastructure using Spark with Scala.
  • Collaborate with our cross-functional teams to design data solutions that meet business needs.
  • Build out our core data pipelines, store data in optimal engines and formats, and feed our machine learning models.
  • Leverage and optimize AWS resources.
  • Collaborate closely with the Data Science team.

AWSSQLCloud ComputingETLMachine LearningData engineeringSparkScalaData modeling

Posted 4 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ’Έ 113000.0 - 130000.0 USD per year

πŸ” Software Development

🏒 Company: LaterπŸ‘₯ 1-10Consumer ElectronicsiOSAppsSoftware

  • Minimum of 5 years in data engineering or related fields
  • Bachelor’s degree in Computer Science, Engineering, or related field
  • Strong focus on building data infrastructure and pipelines
  • Design and build a robust data warehouse architecture
  • Design, build, and maintain scalable data pipelines
  • Develop reliable transformation layers and data pipelines
  • Establish optimized data architectures using cloud technologies
  • Enforce data quality checks and governance practices
  • Collaborate with cross-functional teams to deliver insights
  • Analyze and optimize data pipelines for performance

SQLData engineering

Posted 7 days ago
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ’Έ 150000.0 - 190000.0 USD per year

πŸ” Game Development

🏒 Company: Second Dinner

  • Demonstrated experience in large-scale distributed data systems such as Spark and Flink
  • Deep expertise in analytical database technologies, including SQL and NoSQL
  • Experience with database technologies and ETL/ELT processes
  • Experience with orchestration and automation tools like Airflow and Beam
  • Experience with Databricks and AWS-based data/analytics solutions
  • Develop and operate data infrastructure and pipelines for analytics and reporting
  • Empower Marketing team with high-quality data for user acquisition and retention
  • Collaborate with teams to gain insights from analytics

AWSPythonSQLApache AirflowETLData engineeringNosqlSpark

Posted 8 days ago
Apply
Apply

πŸ“ United States

🧭 Full-Time

πŸ’Έ 145000.0 - 160000.0 USD per year

πŸ” EdTech

🏒 Company: Magpie Literacy

  • 8+ years of experience in data engineering or related fields
  • Advanced SQL expertise
  • Proficiency in Python
  • Experience with BI tools and dashboard creation
  • Understanding of data warehouse concepts
  • Lead the transformation of reporting needs into automated data pipelines
  • Develop reusable reporting components
  • Design and implement modular SQL and dbt models
  • Oversee automated reporting workflows
  • Collaborate with stakeholders on reporting opportunities

AWSPythonSQLETLGit

Posted 8 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Full-Time

πŸ” B2B SaaS

🏒 Company: SanityπŸ‘₯ 51-200πŸ’° Corporate over 2 years agoSoftware Development

  • 4+ years of experience building data pipelines at scale
  • Deep expertise in SQL, Python, and Node.js/TypeScript
  • Production experience with Airflow and RudderStack
  • Track record of building reliable data infrastructure
  • Design, develop, and maintain scalable ETL/ELT pipelines
  • Collaborate to implement and scale product telemetry
  • Establish best practices for data ingestion and transformation
  • Monitor and optimize data pipeline performance

Node.jsPythonSQLApache AirflowETLTypeScript

Posted 13 days ago
Apply
Apply

πŸ“ Europe, APAC, Americas

🧭 Full-Time

πŸ” Software Development

🏒 Company: DockerπŸ‘₯ 251-500πŸ’° $105,000,000 Series C almost 3 years agoDeveloper ToolsDeveloper PlatformInformation TechnologySoftware

  • 4+ years of relevant industry experience
  • Experience with data modeling and building scalable pipelines
  • Proficiency with Snowflake or BigQuery
  • Experience with data governance and security controls
  • Experience creating ETL scripts using Python and SQL
  • Familiarity with a cloud ecosystem: AWS/Azure/Google Cloud
  • Experience with Tableau or Looker
  • Manage and develop ETL jobs, warehouse, and event collection tools
  • Build and manage the Central Data Model for reporting
  • Integrate emerging methodologies and technologies
  • Build data pipelines for ML and AI projects
  • Contribute to SOC2 compliance across the data platform
  • Document technical architecture

PythonSQLETLSnowflakeAirflowData engineeringData visualizationData modeling

Posted 14 days ago
Apply
Apply

πŸ“ United States, Canada

🧭 Fulltime

πŸ” Software Development

🏒 Company: BioRenderπŸ‘₯ 101-250πŸ’° $15,319,133 Series A almost 2 years agoLife ScienceGraphic DesignSoftware

  • 7+ years of data engineering experience of relevant industry experience
  • Expertise working with Data Warehousing platforms (AWS RedShift or Snowflake preferred) and data lake / lakehouse architectures
  • Experience with Data Streaming platforms (AWS Kinesis / Firehose preferred)
  • Expertise with SQL and programming languages commonly used in data platforms (Python, Spark, etc)
  • Experience with data pipeline orchestration (e.g., Airflow) and data pipeline integrations (e.g. Airbyte, Stitch)
NOT STATED

AWSPythonSQLApache AirflowSnowflakeData engineeringSparkData modeling

Posted 17 days ago
Apply