Senior AI Data Engineer

New
CA, TX, NC, WA, ID, NV, AZ, CO, KS, AR, LA, AL, GA, FL, SC, TN, VA, MD, NJ, DE, IL, WI, MI, OH, MA, PA, NH, CTFull-TimeSenior
Salary160,000 - 200,000 USD per year
Apply NowOpens the employer's application page

Job Details

Experience
5+ years of hands-on data engineering experience, including building and operating production data pipelines. 2+ years’ experience working with AI first development workflows. 4+ years’ experience with the following AWS (S3, Glue, Lambda, Redshift), and/or Azure big data services. 1+ year of experience with Snowflake. 2+ years of experience with orchestration frameworks. 2+ years of Salesforce experience with Apex and configurations.
Required Skills
GraphQLNode.jsPythonSQLFlaskKafkaMicrosoft Power BISalesforceSnowflakeAzureRESTful APIsAWS Lambda

Requirements

  • 5+ years of hands-on data engineering experience, including building and operating production data pipelines.
  • Expert-level Python skills for ETL, pipeline orchestration, and automation.
  • Deep SQL proficiency — query optimization, data modeling, stored procedures.
  • 2+ years’ experience working with AI first development workflows.
  • 4+ years’ experience with AWS (S3, Glue, Lambda, Redshift), and/or Azure big data services.
  • 1+ year of experience with Snowflake.
  • 2+ years of experience with orchestration frameworks.
  • 2+ years of Salesforce experience with Apex and configurations.
  • Experienced with Kimball dimensional modeling — you've built star schemas and conformed dimensions in production.
  • Power BI (or equivalent BI tool) experience — data model design and report development.
  • API integration experience — REST, GraphQL, event streaming (Kafka, Kinesis, or similar).
  • Application development literacy — comfortable building lightweight web tooling (Python/Flask, Node, or similar) to complement data products.

Responsibilities

  • Engage stakeholders directly to gather, clarify, and document project requirements.
  • Translate requirements into architected data solutions: choose the right storage, pipeline, modeling, and delivery approach for each problem.
  • Own testing end-to-end — unit tests, data quality checks, reconciliation, and integration tests before anything reaches production.
  • Deploy solutions to production and monitor post-deployment health, iterating rapidly based on real-world feedback.
  • Run parallel AI coding sessions (Claude Code, Cursor, Codex) across different facets of a pipeline simultaneously — orchestrate, verify, and integrate the outputs.
  • Build and maintain context files (CLAUDE.md equivalents) for data projects that encode schema conventions, pipeline patterns, and institutional knowledge — making every future AI session smarter.
  • Design verification loops: automated data quality checks, dbt tests, CI hooks, and pipeline monitors that give AI agents concrete feedback on correctness.
  • Build MCP (Model Context Protocol) or equivalent integrations to connect AI agents directly to Snowflake, Amazon Athena, Postgresql, MySql, Power BI APIs, Salesforce, and internal tooling.
  • Design and build complex, reliable data pipelines ingesting from AWS, Azure, Salesforce, MuleSoft, and multiple third-party APIs into our AWS Data Lake and Snowflake warehouse.
  • Implement and evolve data models using Kimball methodology to support financial, operational, and clinical analytics.
View Full Description & ApplyYou'll be redirected to the employer's site
160,000 - 200,000 USD per year
Apply Now