Professional Documents
Culture Documents
Welcome: It Is My Greatest Pleasure To Welcome One and All Present Here To My Seminar On
Welcome: It Is My Greatest Pleasure To Welcome One and All Present Here To My Seminar On
Welcome: It Is My Greatest Pleasure To Welcome One and All Present Here To My Seminar On
“BIG Data”
Project is running fine in RDBMS
Project is running fine in RDBMS
Question:-Why we need to Migrate from
RDBMS to Hadoop.
Reasons:-
to change schema
No License costs
Very less chance of data loss
Many more...
Architecture of Hadoop
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Understanding of Internal Working
using Word count Problem
Interview Question:-
How can we Optimize it
Answer:-By Reducing Traffic
By adding combiner at Each mapper
Code of combiner almost similar to
code of reducer
Can we load, such type of data in
RDBMS
Answer :- No
Can we load, such type of data in
RDBMS
Answer :- No
Error:-
Can we load, such type of data in
RDBMS
Answer :- No
Error:-
Reason:- Check data type at time of data load
How were we able to load data ,with
data type mismatched
How able to load data ,with data
type mismatched
Reason:- Does not check data type at time of
loading
When it check data type
When it check data type
It compare data type at time of read operation.
If data type does not matched, it consider that
value NULL
Where data stored in Hive Table
Where data stored in Hive Table
Stored in form of file under warehouse
direcory
Where data stored in Hive Table
Stored in form of file under warehouse
direcory
Location of datafile
Where data stored in Hive Table
Stored in form of file under warehouse
direcory
Location of datafile
Display data under file
File Under partition table
In older version,No Update operation
How to get updated records by using row
no.
There are lot of other component of
hadoop
Pig
Hive
Sqoop
Oozie
Flume
Sqoop
Use of sqoop
you have data in RDBMS, not able to process
data in RDBMS. So we do import and export
What were the drawback of
Mapreduce
Write intermediate result to disk
To overcome drawback of mapreduce
we switched to spark
Every Result written in Ram
Hadoop Vs Spark Vs Teradata
When we need to shift to cloud
Display Hive and Spark Terminal
Thank You