Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 5

AIM:

To study how Flipkart uses Big Data to enhance shopping experience.

INTRODUCTION:
Flipkart Private Limited, d/b/a Flipkart is an Indian e-commerce company based in
Bengaluru, India. It was founded by Sachin Bansal and Binny Bansal in 2007. The company
initially focused on book sales, before expanding into other product categories such as
consumer electronics, fashion, and lifestyle products
The service competes primarily with Amazon's Indian subsidiary, and the domestic rival
Snapdeal.[5][6] As of March 2017, Flipkart held a 39.5% market share of India's e-commerce
industry.[7] Flipkart is significantly dominant in the sale of apparel (a position that was
bolstered by its acquisitions of Myntra and Jabong.com), and was described as being "neck
and neck" with Amazon in the sale of electronics and mobile phones.[8] Flipkart also owns
PhonePe, a mobile payments service based on the Unified Payments Interface (UPI).

OBJECTIVE:
In today’s hyper-connected world, businesses are under enormous pressure to build
relationships with fully engaged consumers who keep coming back for more. In theory,
fostering more intimate consumer relationships becomes easier as new sources of data
emerge, data volumes continue their unprecedented growth, and technology becomes more
sophisticated. These developments should enable businesses to do a much better job of
personalizing marketing campaigns and generating precise content recommendations that
drive engagement, adoption and value for subscribers.

Problem:

1. Data volumes continue their unprecedented growth, and technology becomes more
sophisticated so it’s very difficult to manage the data

2. Personalizing marketing campaigns and generating precise content recommendations that


drive engagement, adoption and value for subscribers is becoming more and more difficult

3. Achieving an advanced understanding of one’s audience is a continuous process of testing


and learning. It demands the ability to quickly gather and reliably analyse thousands,
millions, even billions of events every day found in a variety of data sources, formats and
locations — otherwise known as Big Data

Solution Idea
Flipkart the World’s number one e-commerce platform is using analytics and algorithms to
get better insights into its business during any type of sale or festival season. This report will
explain how the Flipkart is leveraging Big Data Platform for processing big data in streams
and batches. This service-oriented architecture empowers user experience, optimizes
logistics and improves product listings. It will also give an insight into how this ingenious big
data platform is able to process such large amounts of data.

BIG DATA AT FLIPKART


Flipkart Data Platform is a service-oriented architecture that is capable of computing batch
data as well as streaming data. This platform comprises of various micro-services that
promote user experience through efficient product listings, optimization of prices,
maintaining various types of data domains – Redis, HBase, SQL, etc. This FDP is capable of
storing 35 PetaBytes of data and is capable of managing 800+ Hadoop nodes on the server.
This is just a brief of how Big Data is helping Flipkart.

The Architecture of Flipkart Data Platform

To know how Flipkart is using Big Data, you need to understand the flow of data or Flipkart’s data
platform architecture which is explained through the below flow chart-

Methodology

The data and analytics platform methodology:


· Rapidly and reliably handle staggering workloads; support insightful analysis of billions of
transactional events each day — every search, browse, stop and start — in whatever data
format that records the events.

· Work with a variety of analytics approaches, including neural networks, Python, Pig, as well
as varied Business Intelligence tools, like MicroStrategy.

How Big Data is helping Flipkart?

Let’s understand the complete process of how Flipkart works on Big Data. Starting with the
FDP ingestion system –
1. FPD Ingestion System
A Big Data Ingestion System is the first place where all the variables start their journey into
the data system. It is a process that involves the import and storage of data in a database.
This data can either be taken in the form of batches or real-time streams. Simply speaking,
batch consists of a collection of data points that are grouped in a specific time interval. On
the contrary, streaming data has to deal with a continuous flow of data. Batch Data has
greater latency than streaming data which is less than sub-seconds. There are three ways in
which ingestion can be performed –
Specter – This is a Java library that is used for sending the draft to Kafka.
Dart Service – This is a REST service which allows the payload to be sent over HTTP.
File Ingestor – With this, we can make use of the CLI tool to dump data into the HDFS.
Then, the user creates a schema for which the corresponding Kafka topic is created. Using
Specter, data is then ingested into the FDP. The payload in the HDFS file is stored in the
form of HIVE tables.
2. Batch Compute
This part of the big data ecosystem is used for computing and processing data that is
present in batches. Batch Compute is an efficient method for processing large scale data
that is present in the form of transactions that are collected over a period of time. These
batches can be computed at the end of the day when the data is collected in large volumes,
only to be processed once. This is the time you need to explore Big Data as much as
possible.
3. Streaming Platform
The streaming platforms process the data that is generated in sub-seconds. Apache Flink is
one of the most popular real-time streaming platforms that are used to produce fast-paced
analytical results. It provides a distributed, fault-tolerant and scalable data streaming
capabilities that can be used by the industries to process a million transactions at one time
without any latency.
4. Messaging Queue
A messaging queue acts like a buffer or a temporary storage system for messages when the
destination is busy or not connected. The message can be in the form of a plain message, a
byte array consisting of headers or a prompt that commands the messaging queue to
process a task. There are two components in the Messaging Queue Architecture – Producer
and Consumer. A Producer generates the messages and delivers them to the messaging
queue. A Consumer is the end destination of the message where the message is processed.

5. Real-time Serving
After the messages are retrieved from the Messaging Queue, the real-time serving system
acts as a consumer for the messaging queue. With the help of this real-time serving
platform, users can gather real-time insights from the data platform. Furthermore, with the
help of real-time serving, the users can access the data through dynamic pipelines.
6. Data Lake
The core component of this architecture is the data storage platform. This is a Hadoop
platform that stores raw data, journaled data as well as derived data. Using this, the data is
stored in the form of a backup, archive that can be retrieved or purged according to the
requirements. The raw data is mostly used by the data scientists who use the insights from
the original data to make decisions and develop data products. The data is present in the
form of batches or real-time streams. The real-time data is in the form of click streams,
summarized reports of user data, product insights, reviews, etc.
From the data lake, data is transferred to three main routes –
a. Reports
The reports are generally produced from the batch data. These reports allow a
comprehensive insight into website logs, daily website readings and other forms of reports.
With the help of these reports, companies like Flipkart are able to quantify the market
needs as well as
b. Ad hoc Query
Ad hoc Query is designed for some specific purpose or use. The Adhoc Queries that are
generated from the data-lake are handled by the data analysts. These data analysts make
use of various business intelligence tools to discover meaning from the data.
c. Batch Export
This part of the data platform takes the data from data-lake and exports it in various
formats to the further processing platforms. The data is present in huge bulks that are
exported.
7. Knowledge Graphs
Knowledge graphs represent an inter-linked network of real-world entities or objects
through which we can extract information to process it in an efficient manner. This
knowledge graph takes input from the meta-data. This metadata is beneficial for
understanding the underlying semantics which is used for deriving newer facts. The
knowledge graph also makes use of various machine learning tools and libraries to gain
insights and understand the relationships between the objects. One of the most popular
tools that are used for building graph is Apache Spark’s GraphX library.

Advantages received by Flipkart from using data analytics :

The Benefits of Using Big Data in Ecommerce include:

Making better strategic decisions (69%).


Improved control of operational processes (54%).
Better understanding of customers (52%).
Cost reductions (47%).

Conclusion:

We studied how Big Data is helping Flipkart to offer the best services all around the World. We also
looked at the ingenious big data platform that is designed by Flipkart to handle large scale data
transactions. We also understood how Flipkart makes use of various big data components to deliver
dynamic results to the user. We also had a look at how the Big Data Platform is capable of processing
large scale data queries that allow it to produce results.

You might also like