Professional Documents
Culture Documents
HDFS Notes
HDFS Notes
HDFS Notes
==============COLUMNARY IMPORT============================
sqoop import --connect jdbc:mysql://localhost:3306/vaibhav --columns 'FIRST_NAME'
--table employees --password hr --username root --target-dir /sqoop/sunny/6
*******************************************************************
output : No new rows detected since last import (coz we have given 5 total there
and 5 last value)
output :
18/02/03 02:59:31 INFO tool.ImportTool: Maximal id query for free form incremental
import: SELECT MAX(`user_id`) FROM user_detail
18/02/03 02:59:31 INFO tool.ImportTool: Incremental import based on column
`user_id`
18/02/03 02:59:31 INFO tool.ImportTool: Lower bound value: 6
18/02/03 02:59:31 INFO tool.ImportTool: Upper bound value: 7
*imp - not good practive to give it for strings as it sort and to min max
caluculations
output:
18/02/03 03:14:42 INFO tool.ImportTool: Lower bound value: '2017-02-03 00:00:00'
18/02/03 03:14:42 INFO tool.ImportTool: Upper bound value: '2018-02-03 03:14:42.0'
##query 2
??why its takign reg >= what i gave and < system time?????
in production date- generally future date wont be dere
but what about columns where future date is there?
l
output:
18/02/03 07:47:17 INFO tool.ImportTool: Lower bound value: '2018-02-02 00:00:00'
18/02/03 07:47:17 INFO tool.ImportTool: Upper bound value: '2018-02-03 07:47:17.0'
1,A1,30,2018-02-03 00:00:00.0
2,A2,31,2018-02-03 00:00:00.0
3,A3,32,2018-02-03 00:00:00.0
4,A4,33,2018-02-03 00:00:00.0
5,A5,34,2018-02-03 00:00:00.0
6,A6,35,2018-02-03 00:00:00.0
7,A7,36,2018-02-03 00:00:00.0
9,A9,38,2018-02-02 10:00:00.0
Note:
sqoop job --create <jobname> -- <mandatory_space> <import/export> <query>
output : no specific output, check in all job list
================== LIST All JOB=================================
sqoop job --list
output:
Available jobs:
myjob1
output: (it asks for password and print all the paramters)
Enter password:
Job: myjob1
Tool: import
Options:
----------------------------
verbose = false
db.connect.string = jdbc:mysql://localhost:3306/vaibhav
codegen.output.delimiters.escape = 0
codegen.output.delimiters.enclose.required = false
codegen.input.delimiters.field = 0
hbase.create.table = false
db.require.password = true
hdfs.append.dir = false
import.fetch.size = null
codegen.input.delimiters.escape = 0
codegen.input.delimiters.enclose.required = false
db.username = root
codegen.output.delimiters.record = 10
import.max.inline.lob.size = 16777216
hcatalog.create.table = false
db.clear.staging.table = false
codegen.input.delimiters.record = 0
enable.compression = false
hive.overwrite.table = false
hive.import = false
codegen.input.delimiters.enclose = 0
hive.drop.delims = false
codegen.output.delimiters.enclose = 0
hdfs.delete-target.dir = true
codegen.output.dir = .
codegen.auto.compile.dir = true
db.query = SELECT * FROM employees WHERE 1>0 AND $CONDITIONS
mapreduce.num.mappers = 4
import.direct.split.size = 0
export.new.update = UpdateOnly
codegen.output.delimiters.field = 44
incremental.mode = None
hdfs.file.format = TextFile
codegen.compile.dir = /tmp/sqoop-
itelligence/compile/c43e0f1df78448d09edc5b36e35ee5ea
direct.import = false
db.split.column = EMPLOYEE_ID
hdfs.target.dir = /sqoop/sunny/172
hive.fail.table.exists = false
db.batch = false
**IMP - You can not edit the job, Delete and create new one