Welcome: It Is My Greatest Pleasure To Welcome One and All Present Here To My Seminar On

You might also like

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 36

Welcome

It is my greatest Pleasure to welcome one and


all present here to my Seminar on

“BIG Data”
Project is running fine in RDBMS
Project is running fine in RDBMS
Question:-Why we need to Migrate from
RDBMS to Hadoop.
Reasons:-

 Can handle Data beyond Petabyte


 Schema is Dynamic, no more effort required

to change schema
 No License costs
 Very less chance of data loss
 Many more...
Architecture of Hadoop
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Interview Question:-
How can we Optimize it
Answer:-By Reducing Traffic
By adding combiner at Each mapper
Code of combiner almost similar to
code of reducer
Can we load, such type of data in
RDBMS
 Answer :- No
Can we load, such type of data in
RDBMS
 Answer :- No
 Error:-
Can we load, such type of data in
RDBMS
 Answer :- No
 Error:-
 Reason:- Check data type at time of data load
How were we able to load data ,with
data type mismatched
How able to load data ,with data
type mismatched
Reason:- Does not check data type at time of
loading
When it check data type
When it check data type
It compare data type at time of read operation.
If data type does not matched, it consider that
value NULL
Where data stored in Hive Table
Where data stored in Hive Table
 Stored in form of file under warehouse
direcory
Where data stored in Hive Table
 Stored in form of file under warehouse
direcory
 Location of datafile
Where data stored in Hive Table
 Stored in form of file under warehouse
direcory
 Location of datafile
 Display data under file
File Under partition table
In older version,No Update operation
 How to get updated records by using row
no.
There are lot of other component of
hadoop
 Pig
 Hive
 Sqoop
 Oozie
 Flume
Sqoop
 Use of sqoop
you have data in RDBMS, not able to process
data in RDBMS. So we do import and export
What were the drawback of
Mapreduce
 Write intermediate result to disk
To overcome drawback of mapreduce
we switched to spark
 Every Result written in Ram
Hadoop Vs Spark Vs Teradata
When we need to shift to cloud
Display Hive and Spark Terminal
Thank You

 For Your Attentention

You might also like