Apply

Data Engineer

Posted 5 days agoViewed

View full description

💎 Seniority level: Middle, 3-5 years

📍 Location: United States

💸 Salary: 85000.0 - 100000.0 USD per year

🏢 Company: Middle Seat

🗣️ Languages: English

⏳ Experience: 3-5 years

🪄 Skills: AWSPythonSQLETLData engineeringData modelingData analytics

Requirements:
  • Proficiency in SQL, Redshift, and Python.
  • Demonstrated experience in deploying and managing data pipelines.
  • Strong analytical skills with a passion for quantitative problem-solving, meticulous attention to detail, and effective multitasking abilities.
  • Proven ability to prioritize projects, with a capacity for both independent direction and delegating tasks as needed.
  • Eagerness to excel in a fast-paced environment, possessing a proactive attitude toward tackling unique challenges and devising creative solutions.
  • Proficient in Excel/Google Sheets, with experience utilizing advanced formulas.
  • Familiarity with dbt for data build tooling.
  • Familiarity with the current technology stack: AWS (S3), SQL, Redshift, Python, GitHub, and Google Looker Studio.
  • Experience with the Democratic Party's technology stack, including ActBlue, ActionKit, Action Network, Civis, NGPVAN, and EveryAction.
  • Experience in designing or developing analytical models.
Responsibilities:
  • Develop, manage, and optimize data systems for storing, transforming, and analyzing large datasets for Middle Seat and its clients, ensuring high performance, security, and reliability.
  • Utilize Python, Redshift, SQL, and additional programming languages as required to analyze, summarize, and report on critical data insights.
  • Assemble, manage, and optimize large and complex data sets to meet both functional and non-functional business requirements, enhancing data accessibility and utility.
  • Engineer robust database systems to guarantee database integrity, security, and performance, aiding analysts and other stakeholders in optimizing data pipelines and code for efficiency.
  • Design and implement efficient data movement strategies between databases, multiple data sources, and analytical tools, ensuring seamless data integration and accessibility.
  • Develop and maintain infrastructure for effective data extraction, transformation, and loading (ETL) utilizing AWS and SQL technologies, streamlining data processing workflows.
  • Collaborate extensively with stakeholders across Executive, Product, Data, and Design teams to address data infrastructure needs, troubleshoot technical issues, and support cross-functional data initiatives.
  • Drive process improvements by optimizing data delivery, automating manual tasks, and enhancing data management practices to boost efficiency and data quality.
  • Uphold the highest standards of data quality and reliability, proactively identifying opportunities for data acquisition and integration to support organizational objectives.
  • Build, maintain, and support analytical tools, providing technical assistance to team members to facilitate data-driven decision-making.
  • Act as the primary liaison and account manager for external data vendors and platforms, ensuring effective partnership and data integration.
  • Provide agency-wide leadership in data engineering, offering support and guidance for utilizing Democratic political CRMs such as ActionKit, ActionNetwork, and NGP/EveryAction.
  • Perform additional responsibilities as assigned by management, adapting to evolving business needs and contributing to the team’s success.
Apply

Related Jobs

Apply

📍 United States

🧭 Internship

💸 30.0 USD per hour

🔍 Streaming Media

🏢 Company: Wurl👥 101-250💰 $11,000,000 Series B over 4 years agoAdvertisingTVVideo StreamingMedia and EntertainmentContent MarketingContent Delivery Network

  • Pursuing or recently completed a Bachelor’s or Master’s degree in Computer Science, Data Science, or a related technical field.
  • Proficiency in Python and SQL.
  • Familiarity with data engineering concepts and tools (e.g., ETL processes, data warehousing).
  • Interest in digital advertising, performance marketing, and streaming media technologies.
  • Strong analytical and problem-solving skills.
  • Excellent communication and collaboration abilities.
  • Collaborate with the Data Engineering team to develop and maintain scalable data pipelines that support performance marketing efforts on CTV.
  • Assist in implementing attribution models that tie CTV viewership to campaigns.
  • Work with cross-functional teams to integrate AI-driven targeting into our data infrastructure.
  • Analyze viewer engagement data to identify opportunities for optimizing return on ad spend (ROAS).
  • Contribute to the development of tools and dashboards that provide insights into campaign performance and audience behavior.

PythonSQLETLData engineering

Posted 1 day ago
Apply
Apply
🔥 Data Engineer
Posted 1 day ago

📍 United States

🧭 Full-Time

🔍 Pharmaceuticals

🏢 Company: Azurity Pharmaceuticals - US

Bachelor’s degree or foreign equivalent in computer science, computer information systems, mathematics, or a related field, plus one year of experience in the job offered or as a data engineer creating data models and data flow diagrams; performing data mining; using segmentation techniques, Python Scripting and SQL; building ETL data pipelines from scratch; building an ETL testing pipeline.
  • Serve as the principal data engineer for data warehouse initiatives.
  • Partner with product owners and business stakeholders to analyze corporate needs.
  • Provide data architecture and solutions that meet business data insights and reporting requirements.
  • Design, build, and maintain data pipelines, workflows, data modeling, and database structures.
  • Implement data storage solutions and ensure data quality and consistency.
  • Optimize data flow and collection for efficiency.
  • Collaborate with data scientists, software developers, and commercial operations colleagues to create coherent solutions.
  • Find and recommend process improvements, particularly transitioning from manual processes to automation.
  • Ensure compliance with information security and privacy policies in data architecture and data warehouse capabilities.
  • Provide scripting and software development support as needed (P-SQL, T-SQL, Python).
  • Create and maintain documentation, including sources to target data mapping and systems architecture diagrams.
  • Monitor, support, and troubleshoot data processing systems to improve data analytics platforms.

PythonSQLData MiningETLData engineeringData modelingScriptingData analytics

Posted 1 day ago
Apply
Apply

📍 US

🧭 Full-Time

🔍 Software Development

🏢 Company: Common Room

  • 3+ years of experience building data or/or ML-based models leveraging industry best practices.
  • Exceptional SQL skills and experience with large-scale data transformation.
  • Build data or/or ML-based models leveraging industry best practices.
  • Shape the future of Common Room’s data capabilities by integrating data science into our product development process.
  • Partner with customers to solve unique data challenges to optimize their workflows, and validate methodologies to develop pragmatic, scalable solutions for customers.
  • Partner with the product team to convert data insights into product improvements

AWSNode.jsPythonSQLKafkaMachine LearningSnowflakeTypeScriptClickhouseData engineeringData sciencePostgresReactData modeling

Posted 1 day ago
Apply
Apply

📍 US

🧭 Full-Time

💸 125000.0 - 200000.0 USD per year

🔍 Software Development

🏢 Company: PredictionHealth

  • Reason around messy, clinical datasets and conduct analysis to communicate out findings
  • Learn and adapt quickly, collaborate with a team.
  • Reason around messy, clinical datasets
  • Identify patterns in datasets and conduct analysis to communicate out findings
  • Independently generate your own hypothesis and work to answer them with data
  • Scope, plan, and deliver on complex projects with minimal direction
  • Learn and adapt quickly
  • Collaborate with a team of individuals with diverse skill sets in a fast-paced environment

AWSPythonSQLCloud ComputingData AnalysisETLData engineeringData visualizationData modelingData analytics

Posted 2 days ago
Apply
Apply
🔥 Data Engineer
Posted 2 days ago

📍 United States

💸 97000.0 - 153595.0 USD per year

🔍 Healthcare

🏢 Company: healthfirst

  • Work experience in a data engineering
  • Work experience in data programing languages such as Java or Python
  • Work experience in a Big Data ecosystem processing data including file systems, data structures/databases, automation, security, messaging, movement, etc.
  • Work experience in a production cloud infrastructure
  • Bachelor’s Degree in computer engineering or related field (preferred)
  • Hands on experience in leading healthcare data transformation initiatives from on-premise to cloud deployment (preferred)
  • Demonstrated experience working in an Agile environment as a Data Engineer (preferred)
  • Hands on work with Amazon Web Services, including creating Redshift data structures, accessing them with Spectrum and storing data in S3 (preferred)
  • Proven results using an analytical perspective to identify engineering patterns within complex strategies and ideas, and break them down into engineered code components (preferred)
  • Knowledge of provider-sponsored health insurance systems/processes and the Healthcare industry (preferred)
  • Experience developing, prototyping, and testing engineered processes, products, or services (preferred)
  • Proficiency with relational, graph and noSQL databases; expertise in SQL (preferred)
  • Demonstrates critical thinking skills with ability to problem solve (preferred)
  • Excellent interpersonal skills with proven ability to influence with impact across functions and disciplines (preferred)
  • Skilled in Microsoft Office including Project, PowerPoint, Word, Excel and Visio (preferred)
  • Finds trends in datasets and develops workflows and algorithms to make raw data useful to the enterprise
  • Designs and implements standardized data management procedures around data staging, data ingestion, data preparation, data provisioning, and data destruction (scripts, programs, automation, assisted by automation, etc.)
  • Ensures quality of technical solutions as data moves across Healthfirst environments
  • Provides insight into the changing data environment, data processing, data storage and utilization requirements for the company and offer suggestions for solutions
  • Ensures managed analytic assets support Healthfirst’s strategic goals by creating and verifying data acquisition requirements and strategy
  • Develops, constructs, tests and maintains architectures
  • Aligns architecture with business requirements and uses programming language and tools
  • Identifies ways to improve data reliability, efficiency, and quality
  • Conducts research for industry and business questions
  • Deploys sophisticated analytics programs, machine learning and statistical methods
  • Prepares data for predictive and prescriptive modeling and finds hidden patterns using data
  • Uses data to discover tasks that can be automated
  • Creates data monitoring capabilities for each business process and works with data consumers on updates
  • Aligns data architecture to Healthfirst solution architecture; contributes to overall solution architecture
  • Helps maintain the integrity and security of the company data
  • Additional duties as assigned or required

AWSPythonSQLAgileApache AirflowCloud ComputingETLJavaJava EEAmazon Web ServicesData engineeringData StructuresNosqlData visualizationData modelingScriptingData analyticsData management

Posted 2 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 2 days ago

📍 United States of America

💸 78750.0 - 133875.0 USD per year

🏢 Company: vspvisioncareers

  • 6+ years’ experience working in development team providing analytical capabilities
  • 6+ years of hands-on experience in the data space, spanning data preparation, SQL, integration tools, ETL/ELT/data pipeline design
  • SQL coding experience
  • Experience working in an agile development environment (Scrum, Kanban) with a focus on Continuous Integration and Delivery
  • Knowledge about various data architectures, patterns, and capabilities such as event-driven architecture, real-time data flows, non-relational repositories, data virtualization, cloud storage, etc
  • Knowledge of and experience with multiple data integration platforms (IBM InfoSphere DataStage, Oracle Data Integrator, Informatica PowerCenter, MS SSIS, AWS Glue, Denodo), and data warehouse MPP platforms such as Snowflake, Netezza, Teradata, Redshift, etc
  • Collaborate within an agile, multi-disciplinary team to deliver optimal data integration and transformation solutions
  • Analyze data requirements (functional and non-functional) to develop and design robust, scalable, automated, fault-tolerant data pipeline solutions for business and technology initiatives
  • Design, build, maintain, and operationalize data pipelines for high volume and complex data using appropriate tools and practices in development, test, and production environments
  • Develop and design data mappings, programs, routines, and SQL to acquire data from legacy, web, cloud, and purchased package environments into the analytics environment
  • Drive automation of data pipeline preparation and integration tasks to minimize manual and error-prone processes and improve productivity using modern data preparation, integration, and AI-enabled metadata management tools and techniques
  • Participate in architecture, governance, and design reviews, identifying opportunities and making recommendations
  • Collaborate with architects to design and model application data structures, storage, and integration in accordance with enterprise-wide architecture standards across legacy, web, cloud, and purchased package environments

AWSSQLAgileETLSnowflakeApache KafkaData engineeringCI/CDRESTful APIsData visualizationData modelingData management

Posted 2 days ago
Apply
Apply
🔥 Staff Data Engineer
Posted 3 days ago

📍 United States

🧭 Full-Time

💸 160000.0 - 230000.0 USD per year

🔍 Daily Fantasy Sports

🏢 Company: PrizePicks👥 101-250💰 Corporate about 2 years agoGamingFantasy SportsSports

  • 7+ years of experience in a data Engineering, or data-oriented software engineering role creating and pushing end-to-end data engineering pipelines.
  • 3+ years of experience acting as technical lead and providing mentorship and feedback to junior engineers.
  • Extensive experience building and optimizing cloud-based data streaming pipelines and infrastructure.
  • Extensive experience exposing real-time predictive model outputs to production-grade systems leveraging large-scale distributed data processing and model training.
  • Experience in most of the following:
  • Excellent organizational, communication, presentation, and collaboration experience with organizational technical and non-technical teams
  • Graduate degree in Computer Science, Mathematics, Informatics, Information Systems or other quantitative field
  • Enhance the capabilities of our existing Core Data Platform and develop new integrations with both internal and external APIs within the Data organization.
  • Develop and maintain advanced data pipelines and transformation logic using Python and Go, ensuring efficient and reliable data processing.
  • Collaborate with Data Scientists and Data Science Engineers to support the needs of advanced ML development.
  • Collaborate with Analytics Engineers to enhance data transformation processes, streamline CI/CD pipelines, and optimize team collaboration workflows Using DBT.
  • Work closely with DevOps and Infrastructure teams to ensure the maturity and success of the Core Data platform.
  • Guide teams in implementing and maintaining comprehensive monitoring, alerting, and documentation practices, and coordinate with Engineering teams to ensure continuous feature availability.
  • Design and implement Infrastructure as Code (IaC) solutions to automate and streamline data infrastructure deployment, ensuring scalable, consistent configurations aligned with data engineering best practices.
  • Build and maintain CI/CD pipelines to automate the deployment of data solutions, ensuring robust testing, seamless integration, and adherence to best practices in version control, automation, and quality assurance.
  • Experienced in designing and automating data governance workflows and tool integrations across complex environments, ensuring data integrity and protection throughout the data lifecycle
  • Serve as a Staff Engineer within the broader PrizePicks technology organization by staying current with emerging technologies, implementing innovative solutions, and sharing knowledge and best practices with junior team members and collaborators.
  • Ensure code is thoroughly tested, effectively integrated, and efficiently deployed, in alignment with industry best practices for version control, automation, and quality assurance.
  • Mentor and support junior engineers by providing guidance, coaching and educational opportunities
  • Provide on-call support as part of a shared rotation between the Data and Analytics Engineering teams to maintain system reliability and respond to critical issues.

AWSBackend DevelopmentDockerLeadershipPythonSQLApache AirflowCloud ComputingETLGitKafkaKubernetesRabbitmqAlgorithmsApache KafkaData engineeringData StructuresGoPostgresSparkCommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingRESTful APIsMentoringLinuxDevOpsTerraformExcellent communication skillsStrong communication skillsData visualizationData modelingScriptingSoftware EngineeringData analyticsData management

Posted 3 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 3 days ago

📍 AL, AZ, AR, CA, CO, CT, DC, FL, GA, IL, KS, KY, MA, MI, MN, MO, NV, NJ, NY, NC, OH, OR, PA, SC, TN, TX, UT, VA, or WA

💸 116500.0 - 196700.0 USD per year

🔍 Data Analytics

  • 3-5 years of experience in data engineering, data analytics, or business analysis roles.
  • 3 years of experience writing SQL for data warehousing.
  • 2-3 years of experience working on data warehouses to support analytics.
  • 2-3 years of experience building and maintaining enterprise data models, such as fact/dimensional star schemas.
  • Proficiency in cloud data warehouses (Redshift, BigQuery, Snowflake).
  • Strong SQL query writing and data modeling capabilities.
  • Experience with Python for scripting, data manipulation, and API interaction.
  • Familiarity with version control systems (e.g., Git, GitHub, GitLab).
  • Experience with workflow orchestration tools (e.g., Apache Airflow).
  • Knowledge of cloud platforms, particularly AWS.
  • Create and maintain Apache Airflow data pipeline jobs.
  • Monitor and update Fivetran extract and load runs.
  • Create and govern dbt models, testing, and job runs.
  • Monitor and maintain data quality processes and procedures.
  • Own, manage, and govern Snowflake data warehouse, including permission grants.
  • Maintain and govern usage of data sent from Snowflake to source systems ("Reverse ETL").
  • Collaborate with stakeholders to maintain and monitor existing data pipelines.
  • Evaluate automation opportunities within the business, using available data tools.
  • Provide data set and tooling recommendations where appropriate.
  • Provision data sets for analytic, master data, and reporting use cases.

AWSPythonSQLApache AirflowETLSnowflakeData engineeringData visualizationData modelingData analytics

Posted 3 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 3 days ago

📍 United States

🧭 Full-Time

💸 135000.0 - 145000.0 USD per year

🔍 Life Science

🏢 Company: Medispend

  • Hands-on knowledge of data integration platforms
  • Experience with enterprise systems (ERP, CRM, etc.)
  • Substantial programming experience with Python based data orchestration and transformation frameworks (i.e. Airflow, AWS Glue, Prefect, Dagster, Spark, Polars, Databricks, etc.)
  • Strong working knowledge of traditional RDBMS data warehousing as well as other platforms like Snowflake, RedShift
  • Gather requirements and design the integration of a new data source
  • Design and build data transformations
  • Estimate levels of effort for prospective client implementations
  • Evaluate new open source data management tools
  • Determine root cause for a failed integration
  • Conduct peer review for code check-ins
  • Design and build a data migration framework
  • Monitor infrastructure capacity of the data transformation platform
  • Compile and analyze data transformation success/failure rates

AWSProject ManagementPythonSQLApache AirflowCloud ComputingETLSnowflakeJiraAlgorithmsData engineeringData StructuresPostgresRDBMSREST APISparkCommunication SkillsAnalytical SkillsCollaborationCI/CDProblem SolvingAgile methodologiesDevOpsData visualizationData modelingScriptingData analyticsData managementSaaS

Posted 3 days ago
Apply
Apply
🔥 Senior Data Engineer
Posted 4 days ago

📍 United States

🧭 Contract

  • Experience with Dataiku.
  • Expertise in IDMC (Informatica Data Management Cloud).
  • Strong knowledge of SQL statements.
  • Basic experience with Python.
  • Knowledge of cloud-based data warehousing solutions, like Snowflake
  • Utilize Dataiku for data preparation, analysis, and workflow automation.
  • Deconstruct complex SQL statements to understand data flows and transformation logic.
  • Migrate data to IDMC (Informatica Data Management Cloud) ensuring quality and validation.
  • Use Snowflake to extract data and convert it into formats compatible with IDMC.
  • Work closely with the business team to validate data accuracy and ensure alignment with business requirements.
  • Provide support for 1-2 products, such as VSM and Flash, ensuring data-related needs are met.

PythonSQLETLSnowflakeData engineeringData visualizationData modeling

Posted 4 days ago
Apply