Data Engineering Intern (AI & Automation)

New
R
Rocket LawyerLegal Services
Remote, but you must be located in CA, AZ, CO, NC, or UT during the internship.InternshipEntry
Salary40 - 50 USD per hour
Apply NowOpens the employer's application page

Job Details

Required Skills
DockerPythonSQLArtificial IntelligenceKubernetesSnowflakeTableauAirflow

Requirements

  • Currently pursuing an undergraduate degree with a targeted graduation date in 2026 or early 2027
  • Pursuing a degree in Computer Science, Data Science, or a related quantitative field
  • Expertise in Python
  • Expertise in SQL
  • Strong understanding of Data Warehousing (Snowflake)
  • Strong understanding of ETL orchestration (Airflow)
  • Familiarity with CLI
  • Familiarity with Docker
  • Familiarity with Kubernetes for managing cloud-based environments
  • Experience with Jupyter Notebooks
  • Experience with Tableau
  • Experience with Streamlit
  • A proactive approach to using AI/LLMs to automate repetitive tasks and improve system reliability

Responsibilities

  • Design, develop, and maintain ETL pipelines to ingest data into our Snowflake warehouse using Python, SQL, and Airflow
  • Implement AI-powered solutions to streamline engineering tasks, including automating code generation and documentation
  • Build AI-driven data quality checks and anomaly detection
  • Develop "self-healing" pipelines that can identify and alert on ingestion errors
  • Use Jupyter Notebooks and Streamlit to analyze data and build internal tools for data-driven decisions
  • Create high-impact dashboards in Tableau that translate complex data into a clear narrative for stakeholders
  • Participate in daily Scrum huddles, manage tasks via Jira
  • Work closely with product owners and QA to promote code to production
  • Interact with cloud services via CLI and manage containerized environments using Docker and Kubernetes
View Full Description & ApplyYou'll be redirected to the employer's site
40 - 50 USD per hour
Apply Now