Software Engineer - Data Platform

  • Engineering
  • Gurgaon, India

Software Engineer - Data Platform

Job description

At Zomato, Data is central to decision making across Product Development, Engineering and Business. We capture a huge amount of data from across our products ranging across Ad performance, application performance, user journey and beyond. We cover the widest possible gamut of services in Restaurants/Food in the world, from Search & Discovery, to Online Ordering, Table Reservations on the consumer side to multitude of B2B products. All this brings in a uniqe set of data which feeds back into our current and future products.

 

The Data Infrastructure team, builds distributed components, systems, and tools that enable capturing, processing and analysis of this data to distill into usable insights. We work with Open Source technologies like Apache Kafka, Hadoop, Presto, Spark and also write some of our own.

 

Here's what you'll do everyday:

  • Build data architecture / data models in support of Data warehouse, Big data and Analytics capabilities, and business requirements
  • Evolve the Data Pipeline and Architecture to allow (near) real time access to data
  • Build tools and systems to make data access easier and friendler to everyone within the organization
  • Improve system efficiency to bring down cost per unit of data stored
  • Work with various teams to understand their data needs and provide solutions that scale
  • Improve data consistency and quality by defining and enforcing better guidelines


You'll also get:

  • To work in our state of the art office with Macbooks, a big screen for debugging, designing or whatever you're into, and high speed internet
  • Open workspaces where the glass walls bear the finest ideas and cryptic musings (quite literally)
  • A cup of coffee (or many) while your code compiles at our own cafeteria which also serves breakfast, lunch, and dinner

Requirements

You must have:

  • You have 3+ years of professional experience working with Big Data technologies
  • Bachelor’s degree or higher in Computer Science or related field or equivalent experience


Here's what we're looking for:

  • Proficiency with technologies like Kafka, Zookeeper, Hadoop, Hive, Yarn, Presto, Spark, Flink, Parquet and ORC
  • Experience with one of these messaging systems (Kafka, Kinesis) and serialization formats (JSON, Avro, Protobuf etc)
  • Experience with AWS or other Cloud technology
  • Experience with designing, building & deploying self-service high-volume data pipelines
  • Experience with designing and building dimensional data models to improve accessibility, efficiency, and quality of data
  • Good understanding of SQL engines and ability to perform advanced performance tuning
  • Good knowledge of and programming experience with scripting languages such as Perl or Unix shell
  • Experience with large Operational Data Store, Data Warehouse, Business Intelligence databases
  • You enjoy working in a fast, agile and nimble environment with frequent changes
  • You have excellent problem solving and critical thinking skills