Data Engineer

Company Description

BumbleB is an enterprise software company that specializes in building software as a service (SaaS) and platform as a service (PaaS) solutions in the field of data insights and artificial intelligence (AI). Our goal is to enable businesses to leverage the power of AI and data to drive innovation and make intelligent decisions.

Role Description

This is a full-time role (location Bengaluru) for a Data Engineer at BumbleB. You will be responsible for designing, building, and maintaining scalable data pipelines and infrastructure. You will work closely with the founding team to ensure reliable data ingestion, transformation, and storage that powers our AI and analytics products.

What You’ll Do

  • Design and build robust data pipelines for ingestion, transformation, and delivery
  • Write and optimize complex SQL for large-scale data processing and analytics
  • Model data for analytical and operational workloads
  • Build and maintain data infrastructure on cloud platforms
  • Collaborate with AI and product teams to deliver clean, reliable datasets

Qualifications

  • Strong foundation in SQL — ability to write performant, complex queries and a thorough understanding of query execution and optimization
  • Solid grasp of relational theory — normalization, set operations, joins, indexing strategies, and data integrity constraints
  • Strong programming skills in Python and/or Scala
  • Experience building and maintaining ETL/ELT pipelines at scale
  • Familiarity with data warehousing solutions (BigQuery, Snowflake, Redshift)
  • Understanding of data modeling techniques (dimensional modeling, star/snowflake schemas)
  • Experience with streaming and event-driven systems (Kafka, Spark Streaming, Flink, or Pulsar)
  • Familiarity with cloud platforms (AWS/GCP) and infrastructure as code
  • 5+ years of industry experience in data engineering or related roles
  • Ability to work independently
  • Excellent problem-solving and analytical thinking skills
  • Effective written and verbal communication skills
  • Bachelor’s or Master’s degree in Computer Science, Data Engineering, or a related field

Nice to Have

  • Experience with database internals — storage engines, query planners, or SQL engine architecture
  • Familiarity with data orchestration tools (Airflow, Dagster, Prefect)
  • Experience with Kafka internals — partitioning, consumer groups, exactly-once semantics, schema registry
  • Experience with data quality frameworks and observability tooling
Click here to apply.