Apply

Software Engineer (Data)

Posted 4 months ago

View full description

💎 Seniority level: Middle, 3+ years of experience

📍 Location: India

🔍 Industry: Weather

🗣️ Languages: English

⏳ Experience: 3+ years of experience

🪄 Skills: AWSPostgreSQLPythonSQLApache AirflowDynamoDBETLHadoopAirflowData engineeringData scienceC (Programming language)

Requirements:
  • 4 year degree in Computer Science or related field
  • 3+ years of experience in Data Engineering or Data Science
  • 3+ years of experience with Python, SQL, AWS Technologies, Data Technologies, Looker, data formats
  • Ability to troubleshoot complex problems and communicate with cross-functional teams
Responsibilities:
  • Develop ETL processes using SQL and Python scripts
  • Code visualizations in Google Looker with supporting backend components
  • Participate in code reviews to improve code quality
  • Provide operational support for deployed code
Apply

Related Jobs

Apply

📍 United States, India, United Kingdom

🔍 B2B Technology

🏢 Company: Demandbase👥 501-1000💰 $175,000,000 Debt Financing almost 2 years agoSales AutomationAdvertisingBig DataSaaSAnalyticsB2BMarketingMarketing AutomationSoftware

  • Bachelor’s or master’s degree in computer science, Mathematics, or Statistics from a top engineering institution.
  • 4+ years of data engineering experience in building enterprise data/analytics solutions.
  • Practical experience with complex analytics projects and advanced SQL for data analysis.
  • Strong practical experience in databases, Advanced SQL, and Python/R.
  • Good understanding of data strategies and data model design.

  • Design, model, and implement data analysis and analytics solutions.
  • Contribute hand-on to data projects involving high-level design, analysis, experiments, data architecture, and data modeling.
  • Support ETL pipeline modules through effective data transformation, data cleaning, reporting, and statistical analysis.
  • Apply analysis techniques such as segmentation, regression, clustering, and data profiling to analyze trends and report KPIs.
  • Collaborate with cross-functional teams in an Agile setting to build a scalable, high-availability data analytics platform.

PythonSQLAgileData AnalysisETLJavaJavascriptProduct DevelopmentData engineeringSparkCommunication SkillsProblem SolvingData modeling

Posted 8 days ago
Apply
Apply

📍 USA, India, UK

🧭 Full-Time

🔍 B2B technology

🏢 Company: Demandbase👥 501-1000💰 $175,000,000 Debt Financing almost 2 years agoSales AutomationAdvertisingBig DataSaaSAnalyticsB2BMarketingMarketing AutomationSoftware

  • Bachelor’s or master’s degree in computer science, Mathematics, Statistics from a top engineering institution.
  • 4+ years of Data Engineering experience in building enterprise data/analytics solutions.
  • Strong practical experience in Databases, Advanced SQL, and Python/R.
  • Good to have experience in designing/implementing ETL data pipelines using open-source platforms.
  • Familiarity with big data technologies like Hive, Red Shift, Hbase, and Apache Spark.

  • Design, Model and Implement data analysis and analytics solutions.
  • Be a hands-on individual contributor for data projects in high-level design, analysis, experiments, data architecture, and data modeling.
  • Support ETL pipeline modules by designing state-of-the-art transformations, data cleaning, matching, reports dashboards, and statistical analysis.
  • Work closely with cross-functional teams in an Agile environment.

PythonSQLAgileData AnalysisETLData engineeringCommunication SkillsProblem SolvingData visualizationData modeling

Posted 22 days ago
Apply
Apply

📍 San Francisco Bay Area, Seattle, India, UK

🧭 Full-Time

💸 150000.0 - 180000.0 USD per year

🔍 B2B technology

  • Four-year degree in Computer Science, or related field OR equivalent experience.
  • Understanding of data strategies, articulated data analysis & data model design.
  • Experience in designing and building low latency analytics APIs at scale.
  • Proficiency in at least one JVM language (Java, Scala, Kotlin, etc.).
  • Familiarity with the Spark Ecosystem (YARN, Executors, Livy, etc.).
  • Knowledge of data orchestration frameworks, particularly Airflow or similar.
  • Experience with columnar data stores, particularly Parquet and StarRocks.
  • Strong SDLC principles (CI/CD, Unit Testing, git, etc.).
  • General understanding of AWS EMR, EC2, S3.

  • Design and build the next generation of Demandbase’s Unified Data Platform through a combination of data pipelines, APIs, internal tools and 3rd party/open source tooling.
  • Work across the data stack to build and productionalize data pipelines for massive amounts of data.
  • Integrate 3rd party and open source tools into our data platform.
  • Build DAGs in Airflow for orchestration and monitoring of data pipelines.

AWSData AnalysisGitJavaJVMKafkaKotlinYarnAirflowSparkCI/CD

Posted 3 months ago
Apply
Apply

📍 San Francisco Bay Area, Seattle, India, UK

🧭 Full-Time

💸 150000.0 - 180000.0 USD per year

🔍 B2B technology

  • Four-year degree in Computer Science or related field OR equivalent experience.
  • Progressive experience in designing frameworks and writing efficient data pipelines, including batches and real-time streams.
  • Understanding of data strategies, articulate data analysis & data model design and evolve data products according to business requirements.
  • Experience with the Spark Ecosystem (YARN, Executors, Livy, etc).
  • Experience in large scale data streaming, particularly Kafka or similar technologies (Pulsar, Kinesis, etc).
  • Experience with data orchestration frameworks, particularly Airflow or similar.
  • Experience with columnar data stores, particularly Parquet and Clickhouse.
  • Strong SDLC principles (CI/CD, Unit Testing, git, etc).
  • General understanding of AWS EMR, EC2, S3.

  • Build out all aspects of the Demandbase Data ecosystem and move products from R&D into production scale.
  • Design and build data pipelines to create the next generation of Demandbase’s Unified Data Platform.
  • Work across the data stack to build and productionalize data pipelines for massive amounts of data.
  • Build DAGs in Airflow for orchestration and monitoring of data pipelines.

Data AnalysisGitKafkaYarnAirflowClickhouseSpark

Posted 5 months ago
Apply