Location:Arizona, California, Colorado, Connecticut, Florida, Georgia, Hawaii, Illinois, Maryland, Massachusetts, Michigan, Minnesota, Missouri, New Hampshire, New Jersey, New York, North Carolina, North Dakota, Ohio, Oregon, Pennsylvania, Rhode Island, South Carolina, Texas, Utah, Vermont, Virginia, Washington, Wisconsin, Washington D.C., EST, PST
5+ years of relevant professional experience in data engineering or backend development with a strong focus on Python. Expertise in writing clean, modular, tested, and production-ready Python code. Strong understanding of data architecture, distributed systems, and security best practices. Experience deploying and supporting production ML workflows (e.g., SageMaker, Vertex AI, or equivalent). Familiarity with ELT tools such as Fivetran and data modeling frameworks like DBT. Solid command of SQL and experience working with large analytical databases (e.g., Redshift, PostgreSQL). Experience with monitoring and observability using Datadog or similar tools.
Responsibilities:
Design, build, and maintain scalable, reliable, and secure data pipelines using Python. Develop reusable data services and frameworks. Collaborate with data scientists and ML engineers to productionize machine learning workflows. Implement monitoring, testing, and CI/CD automation for data pipelines and ML services. Own and evolve real-time and batch data integrations. Develop, optimize and support reverse ETL workflows. Participate in code reviews, mentor junior engineers, and help foster a high-trust engineering culture. Demonstrate technical leadership through writing documentation, establishing effective monitoring, and fostering clear communication.