sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Chalcy <cha...@gmail.com>
Subject Re: Data mismatch when importing data from Oracle to Hive through Sqoop without an error
Date Sun, 17 Mar 2013 12:55:02 GMT
When you create the table copy table, use, "create table copytable like
origtable" instead of creating hive copy table using script like you are
doing.

Cheers,
Chalcy

On Wed, Mar 13, 2013 at 4:44 AM, Ajit Kumar Shreevastava <
Ajit.Shreevastava@hcl.com> wrote:

>  Thanks Chalcy and Jarek****
>
> ** **
>
> Now, I am facing one new problem with Sqoop export utility.****
>
> ** **
>
> I have implemented what both of you suggested in chain mail and my problem
> is solved for importing the data to Hive.****
>
> ** **
>
> Now I have truncate the table in oracle and try to export the data from
> just populated table in Hive is working fine and no record mismatch was
> found.****
>
> Command for export to oracle is :à****
>
> *[hadoop@NHCLT-PC44-2 sqoop-oper]$ *sqoop export --connect
> jdbc:oracle:thin:@10.99.42.11:1521/clouddb --username HDFSUSER  --table
> BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp_test
> -P -m 1  --input-fields-terminated-by '\0001' –verbose****
>
> ** **
>
> Now I want to create a table in hive as a copy of bttn_bkp_test.****
>
> I described the bttn_bkp_test table and created the table column described
> in bttn_bkp_test like :à****
>
> ** **
>
> *hive>* CREATE TABLE IF NOT EXISTS BTTN_BKP(****
>
>     > bttn_id                 double,****
>
>     > data_inst_id            double,****
>
>     > scr_id                  double,****
>
>     > bttn_nu                 double,****
>
>     > cat                     double,****
>
>     > wdth                    double,****
>
>     > hght                    double,****
>
>     > key_scan                double,****
>
>     > key_shft                double,****
>
>     > frgnd_cptn_colr         string,****
>
>     > frgnd_cptn_colr_prsd    string,****
>
>     > bkgd_cptn_colr          string,****
>
>     > bkgd_cptn_colr_prsd     string,****
>
>     > blm_fl                  double,****
>
>     > lclz_fl                 double,****
>
>     > menu_item_nu            double,****
>
>     > bttn_asgn_lvl_id        double,****
>
>     > on_atvt                 double,****
>
>     > on_clik                 double,****
>
>     > enbl_fl                 double,****
>
>     > blm_set_id              double,****
>
>     > bttn_asgn_lvl_name      string,****
>
>     > mkt_id                  double,****
>
>     > crte_ts                 string,****
>
>     > crte_user_id            string,****
>
>     > updt_ts                 string,****
>
>     > updt_user_id            string,****
>
>     > del_ts                  string,****
>
>     > del_user_id             string,****
>
>     > dltd_fl                 double,****
>
>     > menu_item_na            string,****
>
>     > prd_cd                  double,****
>
>     > blm_set_na              string,****
>
>     > sound_file_id           double,****
>
>     > is_dynmc_bttn           double,****
>
>     > frgnd_cptn_colr_id      double,****
>
>     > frgnd_cptn_colr_prsd_id double,****
>
>     > bkgd_cptn_colr_id       double,****
>
>     > bkgd_cptn_colr_prsd_id  double);****
>
> ** **
>
> Then I populate this table by following command:à****
>
> *hive>* insert OVERWRITE table bttn_bkp****
>
>    > select * from BTTN_BKP_TEST;****
>
> ** **
>
> And its insert the same data in bttn_bkp without any error.****
>
> ****
>
> Now I have truncated the oracle table bttn_bkp_test and try to populate
> the table by bttn_bkp table data of Hive through below Sqoop command :à **
> **
>
> ** **
>
> *[hadoop@NHCLT-PC44-2 sqoop-oper]$* sqoop export --connect
> jdbc:oracle:thin:@10.99.42.11:1521/clouddb --username HDFSUSER  --table
> BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp_test
> -P -m 1  --input-fields-terminated-by '\0001' –verbose****
>
> ** **
>
> And I got below error error:à****
>
> ** **
>
> Warning: /usr/lib/hbase does not exist! HBase imports will fail.****
>
> Please set $HBASE_HOME to the root of your HBase installation.****
>
> 13/03/13 14:00:04 DEBUG tool.BaseSqoopTool: Enabled debug logging.****
>
> Enter password:****
>
> 13/03/13 14:00:09 DEBUG sqoop.ConnFactory: Loaded manager factory:
> com.cloudera.sqoop.manager.DefaultManagerFactory****
>
> 13/03/13 14:00:09 DEBUG sqoop.ConnFactory: Trying ManagerFactory:
> com.cloudera.sqoop.manager.DefaultManagerFactory****
>
> 13/03/13 14:00:09 DEBUG manager.DefaultManagerFactory: Trying with scheme:
> jdbc:oracle:thin:@10.99.42.11****
>
> 13/03/13 14:00:09 DEBUG manager.OracleManager$ConnCache: Instantiated new
> connection cache.****
>
> 13/03/13 14:00:09 INFO manager.SqlManager: Using default fetchSize of 1000
> ****
>
> 13/03/13 14:00:09 DEBUG sqoop.ConnFactory: Instantiated ConnManager
> org.apache.sqoop.manager.OracleManager@5557c2bd****
>
> 13/03/13 14:00:09 INFO tool.CodeGenTool: Beginning code generation****
>
> 13/03/13 14:00:09 DEBUG manager.OracleManager: Using column names query:
> SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0****
>
> 13/03/13 14:00:09 DEBUG manager.OracleManager: Creating a new connection
> for jdbc:oracle:thin:@10.99.42.11:1521/clouddb, using username: HDFSUSER**
> **
>
> 13/03/13 14:00:09 DEBUG manager.OracleManager: No connection paramenters
> specified. Using regular API for making connection.****
>
> 13/03/13 14:00:10 INFO manager.OracleManager: Time zone has been set to GMT
> ****
>
> 13/03/13 14:00:10 DEBUG manager.SqlManager: Using fetchSize for next
> query: 1000****
>
> 13/03/13 14:00:10 INFO manager.SqlManager: Executing SQL statement: SELECT
> t.* FROM BTTN_BKP_TEST t WHERE 1=0****
>
> 13/03/13 14:00:10 DEBUG manager.OracleManager$ConnCache: Caching released
> connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: selected columns:****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DATA_INST_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   SCR_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_NU****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   CAT****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   WDTH****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   HGHT****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   KEY_SCAN****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   KEY_SHFT****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BLM_FL****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   LCLZ_FL****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   MENU_ITEM_NU****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   ON_ATVT****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   ON_CLIK****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   ENBL_FL****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BLM_SET_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   MKT_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   CRTE_TS****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   CRTE_USER_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   UPDT_TS****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   UPDT_USER_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DEL_TS****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DEL_USER_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   DLTD_FL****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   MENU_ITEM_NA****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   PRD_CD****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BLM_SET_NA****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   SOUND_FILE_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: Writing source file:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.java
> ****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: Table name: BTTN_BKP_TEST****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: Columns: BTTN_ID:2,
> DATA_INST_ID:2, SCR_ID:2, BTTN_NU:2, CAT:2, WDTH:2, HGHT:2, KEY_SCAN:2,
> KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12, BKGD_CPTN_COLR:12,
> BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2, MENU_ITEM_NU:2,
> BTTN_ASGN_LVL_ID:2, ON_ATVT:2, ON_CLIK:2, ENBL_FL:2, BLM_SET_ID:2,
> BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93, CRTE_USER_ID:12, UPDT_TS:93,
> UPDT_USER_ID:12, DEL_TS:93, DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12,
> PRD_CD:2, BLM_SET_NA:12, SOUND_FILE_ID:2, IS_DYNMC_BTTN:2,
> FRGND_CPTN_COLR_ID:2, FRGND_CPTN_COLR_PRSD_ID:2, BKGD_CPTN_COLR_ID:2,
> BKGD_CPTN_COLR_PRSD_ID:2,****
>
> 13/03/13 14:00:10 DEBUG orm.ClassWriter: sourceFilename is
> BTTN_BKP_TEST.java****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager: Found existing
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/****
>
> 13/03/13 14:00:10 INFO orm.CompilationManager: HADOOP_HOME is
> /home/hadoop/hadoop-1.0.3/libexec/..****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager: Adding source file:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.java
> ****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager: Invoking javac with args:*
> ***
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   -sourcepath****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   -d****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:   -classpath****
>
> 13/03/13 14:00:10 DEBUG orm.CompilationManager:
> /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar
> ****
>
> Note:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.java
> uses or overrides a deprecated API.****
>
> Note: Recompile with -Xlint:deprecation for details.****
>
> 13/03/13 14:00:11 INFO orm.CompilationManager: Writing jar file:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.jar
> ****
>
> 13/03/13 14:00:11 DEBUG orm.CompilationManager: Scanning for .class files
> in directory: /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665**
> **
>
> 13/03/13 14:00:11 DEBUG orm.CompilationManager: Got classfile:
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.class
> -> BTTN_BKP_TEST.class****
>
> 13/03/13 14:00:11 DEBUG orm.CompilationManager: Finished writing jar file
> /tmp/sqoop-hadoop/compile/70e185f5d7f3490053e54864e50fc665/BTTN_BKP_TEST.jar
> ****
>
> 13/03/13 14:00:11 INFO mapreduce.ExportJobBase: Beginning export of
> BTTN_BKP_TEST****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Using InputFormat: class
> org.apache.sqoop.mapreduce.ExportInputFormat****
>
> 13/03/13 14:00:11 DEBUG manager.OracleManager$ConnCache: Got cached
> connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
>
> 13/03/13 14:00:11 INFO manager.OracleManager: Time zone has been set to GMT
> ****
>
> 13/03/13 14:00:11 DEBUG manager.OracleManager$ConnCache: Caching released
> connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/sqoop-1.4.2.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/ojdbc6.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/sqoop-1.4.2.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/sqoop-1.4.2.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/ojdbc6.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/paranamer-2.3.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/avro-1.5.3.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/commons-io-1.4.jar****
>
> 13/03/13 14:00:11 DEBUG mapreduce.JobBase: Adding to job classpath:
> file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar****
>
> 13/03/13 14:00:12 INFO input.FileInputFormat: Total input paths to process
> : 1****
>
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: Target numMapTasks=1*
> ***
>
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: Total input
> bytes=172704981****
>
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: maxSplitSize=172704981
> ****
>
> 13/03/13 14:00:12 INFO input.FileInputFormat: Total input paths to process
> : 1****
>
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat: Generated splits:****
>
> 13/03/13 14:00:12 DEBUG mapreduce.ExportInputFormat:
> Paths:/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:0+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:67108864+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:134217728+38487253
> Locations:NHCLT-PC44-2:;****
>
> 13/03/13 14:00:12 INFO mapred.JobClient: Running job: job_201303121648_0015
> ****
>
> 13/03/13 14:00:13 INFO mapred.JobClient:  map 0% reduce 0%****
>
> 13/03/13 14:00:27 INFO mapred.JobClient: Task Id :
> attempt_201303121648_0015_m_000000_0, Status : FAILED****
>
> java.lang.NumberFormatException****
>
>         at java.math.BigDecimal.<init>(BigDecimal.java:459)****
>
>         at java.math.BigDecimal.<init>(BigDecimal.java:728)****
>
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1271)****
>
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)****
>
>         at
> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)*
> ***
>
>         at
> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)*
> ***
>
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)****
>
>         at
> org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
> ****
>
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
> ****
>
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)****
>
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)****
>
>         at java.security.AccessController.doPrivileged(Native Method)****
>
>         at javax.security.auth.Subject.doAs(Subject.java:396)****
>
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
> ****
>
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)****
>
> ** **
>
> 13/03/13 14:00:33 INFO mapred.JobClient: Task Id :
> attempt_201303121648_0015_m_000000_1, Status : FAILED****
>
> java.lang.NumberFormatException****
>
>         at java.math.BigDecimal.<init>(BigDecimal.java:459)****
>
>         at java.math.BigDecimal.<init>(BigDecimal.java:728)****
>
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1271)****
>
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)****
>
>         at
> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)*
> ***
>
>         at
> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)*
> ***
>
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)****
>
>         at
> org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
> ****
>
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
> ****
>
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)****
>
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)****
>
>         at java.security.AccessController.doPrivileged(Native Method)****
>
>         at javax.security.auth.Subject.doAs(Subject.java:396)****
>
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
> ****
>
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)****
>
> ** **
>
> 13/03/13 14:00:39 INFO mapred.JobClient: Task Id :
> attempt_201303121648_0015_m_000000_2, Status : FAILED****
>
> java.lang.NumberFormatException****
>
>         at java.math.BigDecimal.<init>(BigDecimal.java:459)****
>
>         at java.math.BigDecimal.<init>(BigDecimal.java:728)****
>
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1271)****
>
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)****
>
>         at
> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)*
> ***
>
>         at
> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)*
> ***
>
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)****
>
>         at
> org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
> ****
>
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
> ****
>
>        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)****
>
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)****
>
>         at java.security.AccessController.doPrivileged(Native Method)****
>
>         at javax.security.auth.Subject.doAs(Subject.java:396)****
>
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
> ****
>
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)****
>
> ** **
>
> 13/03/13 14:00:50 INFO mapred.JobClient: Job complete:
> job_201303121648_0015****
>
> 13/03/13 14:00:50 INFO mapred.JobClient: Counters: 7****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:   Job Counters****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=25370****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     Total time spent by all
> reduces waiting after reserving slots (ms)=0****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     Total time spent by all maps
> waiting after reserving slots (ms)=0****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     Launched map tasks=4****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     Data-local map tasks=4****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=0****
>
> 13/03/13 14:00:50 INFO mapred.JobClient:     Failed map tasks=1****
>
> 13/03/13 14:00:50 INFO mapreduce.ExportJobBase: Transferred 0 bytes in
> 39.0699 seconds (0 bytes/sec)****
>
> 13/03/13 14:00:50 INFO mapreduce.ExportJobBase: Exported 0 records.****
>
> 13/03/13 14:00:50 ERROR tool.ExportTool: Error during export: Export job
> failed!****
>
> ** **
>
> ** **
>
> ** **
>
> *From:* Chalcy [mailto:chalcy@gmail.com]
> *Sent:* Friday, March 08, 2013 11:15 PM
> *To:* user@sqoop.apache.org
> *Cc:* user@hive.apache.org
> *Subject:* Re: Data mismatch when importing data from Oracle to Hive
> through Sqoop without an error****
>
> ** **
>
> Yes, This is newline char in one of the text fields.****
>
> ** **
>
> The data is split and forms two rows if the original field has one new
> line char.  -hive-drop-import-delims is the helper.****
>
> ** **
>
> Thanks,****
>
> Chalcy****
>
> On Wed, Mar 6, 2013 at 12:35 PM, Jarek Jarcec Cecho <jarcec@apache.org>
> wrote:****
>
> Hi Ajit,
> I've seen similar issue many times. Does your table have textual data? If
> so, can it happen that your textual data contains hive delimiters like new
> line characters? Because if so then Sqoop might create two lines in for one
> single row in the table that will be consequently seen as two rows in Hive.
> As Hive will implicitly convert any invalid values into NULL, it would also
> explain your NULL values.
>
> Sqoop offers arguments --hive-drop-import-delims and
> --hive-delims-replacement to deal with this problem. More information can
> be found in Sqoop user guide [1].
>
> Jarcec
>
> Links:
> 1:
> http://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_importing_data_into_hive
> ****
>
>
> On Wed, Mar 06, 2013 at 10:50:26PM +0530, abhijeet gaikwad wrote:
> > If you see Hadoop job counters in your logs - numbers are correct till
> that
> > point. So normal import in HDFS is working fine. Only reasonable
> > explanation is that there is an issue inserting data in Hive.
> >
> > Have a look at the file on HDFS it should contain the correct number of
> > records with correct data. But if you do "select * ..." from Hive console
> > it will give you NULL values for invalid data, as in the data that cannot
> > be converted to the respective data type of the column of Hive table.
> This
> > is because Hive imposes data type constraints on the data in files on
> HDFS.
> > Those extra records is a mystry for me too. I don't think actual file on
> > HDFS has extra records.
> >****
>
> > Sqoop is creating Hive table incorrectly which **may** have caused this*
> ***
>
> > issue. Create table manually with correct data types if equivalent
> > available or as string and try loading data. If this works fine we know
> > where the issue is!
> >
> > Thanks,
> > Abhijeet
> > On Wed, Mar 6, 2013 at 2:05 PM, Ajit Kumar Shreevastava <
> > Ajit.Shreevastava@hcl.com> wrote:
> >
> >****
>
> > >  Hi Abhijeet,****
> > >
> > > ** **
> > >
> > > Thanks for your response.********
>
> > >
> > > If values that don’t fit in double must be getting inserted as Null is
> the****
>
> > > case then count should not be mis-match in both the case.********
>
> > >
> > > Here the null value inserted are extra value apart from the other value
> ****
>
> > > which is already present in both Oracle Table and Hive table.****
> > >
> > > ** **
> > >
> > > Correct me if I am wrong in interpretation.****
> > >
> > > ** **
> > >
> > > Thanks and Regards,****
> > >
> > > Ajit Kumar Shreevastava****
> > >
> > > ** **
> > >
> > > *From:* abhijeet gaikwad [mailto:abygaikwad17@gmail.com]
> > > *Sent:* Wednesday, March 06, 2013 1:46 PM
> > > *To:* user@hive.apache.org
> > > *Cc:* user@sqoop.apache.org
> > > *Subject:* Re: Data mismatch when importing data from Oracle to Hive
> > > through Sqoop without an error****
> > >
> > >   ** ******
>
> > >
> > > Sqoop maps numeric and decimal types (RDBMS) to double (Hive). I think
> the
> > > values that don't fit in double must be getting inserted as NULL.
> > > You can see this warning in your logs.
> > >
> > > Thanks,****
>
> > > Abhijeet********
>
> > >
> > > On Wed, Mar 6, 2013 at 1:32 PM, Ajit Kumar Shreevastava <****
>
> > > Ajit.Shreevastava@hcl.com> wrote:****
> > >
> > > Hi all,****
> > >
> > > I have notice one interesting thing in the below result-set.********
>
> > >
> > > I have fired one query in both Oracle and Hive shell and found the****
>
> > > following result set:à****
> > >
> > >  ****
> > >
> > > *SQL> select count(1) from bttn*****
> > >
> > > *  2  where bttn_id is null or data_inst_id is null or scr_id is
> null;****
> > > *
> > >
> > > * *****
> > >
> > > *  COUNT(1)*****
> > >
> > > *----------*****
> > >
> > > *         0*****
> > >
> > > *hive> select count(1) from bttn*****
> > >
> > > *    > where bttn_id is null or data_inst_id is null or scr_id is
> null;***
> > > **
> > >
> > > Total MapReduce jobs = 1****
> > >
> > > Launching Job 1 out of 1****
> > >
> > > Number of reduce tasks determined at compile time: 1****
> > >
> > > In order to change the average load for a reducer (in bytes):****
> > >
> > >   set hive.exec.reducers.bytes.per.reducer=<number>****
> > >
> > > In order to limit the maximum number of reducers:****
> > >
> > >   set hive.exec.reducers.max=<number>****
> > >
> > > In order to set a constant number of reducers:****
> > >
> > >   set mapred.reduce.tasks=<number>********
>
> > >
> > > Starting Job = job_201303051835_0020, Tracking URL =****
>
> > > http://NHCLT-PC44-2:50030/jobdetails.jsp?jobid=job_201303051835_0020<
> http://nhclt-pc44-2:50030/jobdetails.jsp?jobid=job_201303051835_0020>
> > > ********
>
> > >
> > > Kill Command = /home/hadoop/hadoop-1.0.3/bin/hadoop job  -kill****
>
> > > job_201303051835_0020********
>
> > >
> > > Hadoop job information for Stage-1: number of mappers: 1; number of***
> *
>
> > > reducers: 1****
> > >
> > > 2013-03-06 13:22:56,908 Stage-1 map = 0%,  reduce = 0%********
>
> > >
> > > 2013-03-06 13:23:05,928 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:06,931 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:07,934 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:08,938 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:09,941 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:10,944 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:11,947 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:12,956 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:13,959 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:14,962 Stage-1 map = 100%,  reduce = 33%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:15,965 Stage-1 map = 100%,  reduce = 33%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:16,969 Stage-1 map = 100%,  reduce = 33%, Cumulative
> CPU****
>
> > > 5.2 sec********
>
> > >
> > > 2013-03-06 13:23:17,974 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec********
>
> > >
> > > 2013-03-06 13:23:18,977 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec********
>
> > >
> > > 2013-03-06 13:23:19,981 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec********
>
> > >
> > > 2013-03-06 13:23:20,985 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec********
>
> > >
> > > 2013-03-06 13:23:21,988 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec********
>
> > >
> > > 2013-03-06 13:23:22,995 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec********
>
> > >
> > > 2013-03-06 13:23:23,998 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 6.95 sec****
> > >
> > > MapReduce Total cumulative CPU time: 6 seconds 950 msec****
> > >
> > > Ended Job = job_201303051835_0020****
> > >
> > > MapReduce Jobs Launched:********
>
> > >
> > > Job 0: Map: 1  Reduce: 1   Cumulative CPU: 6.95 sec   HDFS Read:
> 184270926****
>
> > > HDFS Write: 4 SUCCESS****
> > >
> > > Total MapReduce CPU Time Spent: 6 seconds 950 msec****
> > >
> > > OK****
> > >
> > > *986*****
> > >
> > > Time taken: 35.983 seconds****
> > >
> > > *hive>*****
> > >
> > >  ****
> > >
> > > and  *739169 – 738183=986*****
> > >
> > > * *****
> > >
> > > *can anyone tell me why this happened as BTTN_ID  ,  DATA_INST_ID,****
>
> > > SCR_ID   are not null constrains of BTTN table and also composite
> Primary****
>
> > > Key of the table?*****
> > >
> > > *Also tell me how can I prevent this unnecessary data generation in
> HIVE
> > > table.*****
> > >
> > > * *****
> > >
> > > *Regards*****
> > >
> > > *Ajit Kumar Shreevastava*****
> > >
> > > * *****
> > >
> > > *From:* Ajit Kumar Shreevastava
> > > *Sent:* Wednesday, March 06, 2013 12:40 PM
> > > *To:* 'user@sqoop.apache.org'
> > > *Cc:* user@hive.apache.org
> > > *Subject:* Data mismatch when importing data from Oracle to Hive
> through
> > > Sqoop without an error
> > > ****
> > >
> > >    ****
> > >
> > > HI,****
> > >
> > >  ****
> > >
> > > I have a table BTTN in Oracle database having *738183* records:à****
> > >
> > >  ****
> > >
> > > *SQL> select count(1) from bttn;*****
> > >
> > >  ****
> > >
> > > *  COUNT(1)*****
> > >
> > > *----------*****
> > >
> > > *    738183*****
> > >
> > > * *********
>
> > >
> > > Now I want to import this table to HIVE  and I have fired the following
> ****
>
> > > command at command prompt:à****
> > >
> > >  ****
> > >
> > > *[hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop import --connect****
>
> > > jdbc:oracle:thin:@10.99.42.11:1521/clouddb --username HDFSUSER
>  --table
> > > BTTN --verbose -P --hive-table bttn --create-hive-table --hive-import*
> ***
>
> > > --hive-home /home/hadoop/user/hive/warehouse*****
> > >
> > > *The output is as follows:**à*****
> > >
> > > Warning: /usr/lib/hbase does not exist! HBase imports will fail.****
> > >
> > > Please set $HBASE_HOME to the root of your HBase installation.****
> > >
> > > 13/03/06 12:16:25 DEBUG tool.BaseSqoopTool: Enabled debug logging.****
> > >
> > > Enter password:********
>
> > >
> > > 13/03/06 12:16:32 INFO tool.BaseSqoopTool: Using Hive-specific
> delimiters****
>
> > > for output. You can override********
>
> > >
> > > 13/03/06 12:16:32 INFO tool.BaseSqoopTool: delimiters with****
>
> > > --fields-terminated-by, etc.********
>
> > >
> > > 13/03/06 12:16:32 DEBUG sqoop.ConnFactory: Loaded manager factory:****
>
> > > com.cloudera.sqoop.manager.DefaultManagerFactory********
>
> > >
> > > 13/03/06 12:16:32 DEBUG sqoop.ConnFactory: Trying ManagerFactory:****
>
> > > com.cloudera.sqoop.manager.DefaultManagerFactory********
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.DefaultManagerFactory: Trying with
> scheme:****
>
> > > jdbc:oracle:thin:@10.99.42.11********
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager$ConnCache: Instantiated
> new****
>
> > > connection cache.********
>
> > >
> > > 13/03/06 12:16:32 INFO manager.SqlManager: Using default fetchSize of
> 1000****
>
> > > ********
>
> > >
> > > 13/03/06 12:16:32 DEBUG sqoop.ConnFactory: Instantiated ConnManager***
> *
>
> > > org.apache.sqoop.manager.OracleManager@2393385d****
> > >
> > > 13/03/06 12:16:32 INFO tool.CodeGenTool: Beginning code generation****
> ****
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager: Using column names
> query:****
>
> > > SELECT t.* FROM BTTN t WHERE 1=0********
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager: Creating a new
> connection****
>
> > > for jdbc:oracle:thin:@10.99.42.11:1521/clouddb, using username:
> HDFSUSER**
> > > ******
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager: No connection
> paramenters****
>
> > > specified. Using regular API for making connection.********
>
> > >
> > > 13/03/06 12:16:32 INFO manager.OracleManager: Time zone has been set
> to GMT****
>
> > > ********
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.SqlManager: Using fetchSize for next**
> **
>
> > > query: 1000********
>
> > >
> > > 13/03/06 12:16:32 INFO manager.SqlManager: Executing SQL statement:
> SELECT****
>
> > > t.* FROM BTTN t WHERE 1=0********
>
> > >
> > > 13/03/06 12:16:32 DEBUG manager.OracleManager$ConnCache: Caching
> released****
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: selected columns:****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DATA_INST_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   SCR_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_NU****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   CAT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   WDTH****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   HGHT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   KEY_SCAN****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   KEY_SHFT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BLM_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   LCLZ_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   MENU_ITEM_NU****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   ON_ATVT****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   ON_CLIK****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   ENBL_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BLM_SET_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   MKT_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   CRTE_TS****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   CRTE_USER_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   UPDT_TS****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   UPDT_USER_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DEL_TS****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DEL_USER_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   DLTD_FL****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   MENU_ITEM_NA****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   PRD_CD****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BLM_SET_NA****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   SOUND_FILE_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID*****
> ***
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: Writing source file:****
>
> > >
> /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.java****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: Table name: BTTN********
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: Columns: BTTN_ID:2,
> > > DATA_INST_ID:2, SCR_ID:2, BTTN_NU:2, CAT:2, WDTH:2, HGHT:2, KEY_SCAN:2,
> > > KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12,
> BKGD_CPTN_COLR:12,
> > > BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2, MENU_ITEM_NU:2,
> > > BTTN_ASGN_LVL_ID:2, ON_ATVT:2, ON_CLIK:2, ENBL_FL:2, BLM_SET_ID:2,
> > > BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93, CRTE_USER_ID:12,
> UPDT_TS:93,
> > > UPDT_USER_ID:12, DEL_TS:93, DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12,
> > > PRD_CD:2, BLM_SET_NA:12, SOUND_FILE_ID:2, IS_DYNMC_BTTN:2,
> > > FRGND_CPTN_COLR_ID:2, FRGND_CPTN_COLR_PRSD_ID:2, BKGD_CPTN_COLR_ID:2,*
> ***
>
> > > BKGD_CPTN_COLR_PRSD_ID:2,****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.ClassWriter: sourceFilename is
> BTTN.java********
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager: Found existing****
>
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/********
>
> > >
> > > 13/03/06 12:16:32 INFO orm.CompilationManager: HADOOP_HOME is****
>
> > > /home/hadoop/hadoop-1.0.3/libexec/..********
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager: Adding source file:***
> *
>
> > >
> /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.java****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager: Invoking javac with
> args:*
> > > ***
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:   -sourcepath********
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:****
>
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:   -d********
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:****
>
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/****
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:   -classpath********
>
> > >
> > > 13/03/06 12:16:32 DEBUG orm.CompilationManager:
> > >
> /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar
> ****
>
> > > ****
> > >
> > > Note:
> /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.java
> > > uses or overrides a deprecated API.****
> > >
> > > Note: Recompile with -Xlint:deprecation for details.********
>
> > >
> > > 13/03/06 12:16:34 INFO orm.CompilationManager: Writing jar file:****
>
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.jar****
> ****
>
> > >
> > > 13/03/06 12:16:34 DEBUG orm.CompilationManager: Scanning for .class
> files****
>
> > > in directory:
> /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e**
> > > ******
>
> > >
> > > 13/03/06 12:16:34 DEBUG orm.CompilationManager: Got classfile:
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.class
> ->****
>
> > > BTTN.class********
>
> > >
> > > 13/03/06 12:16:34 DEBUG orm.CompilationManager: Finished writing jar
> file****
>
> > > /tmp/sqoop-hadoop/compile/6aa39c498a3c94ffa01a5352f368712e/BTTN.jar****
> ****
>
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Got cached***
> *
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:16:34 INFO manager.OracleManager: Time zone has been set
> to GMT****
>
> > > ********
>
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Caching
> released****
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:16:34 WARN manager.OracleManager: The table BTTN contains a
> > > multi-column primary key. Sqoop will default to the column BTTN_ID
> only for****
>
> > > this job.********
>
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Got cached***
> *
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:16:34 INFO manager.OracleManager: Time zone has been set
> to GMT****
>
> > > ********
>
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Caching
> released****
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:16:34 WARN manager.OracleManager: The table BTTN contains a
> > > multi-column primary key. Sqoop will default to the column BTTN_ID
> only for****
>
> > > this job.****
> > >
> > > 13/03/06 12:16:34 INFO mapreduce.ImportJobBase: Beginning import of
> BTTN**
> > > ******
>
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Got cached***
> *
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:16:34 INFO manager.OracleManager: Time zone has been set
> to GMT****
>
> > > ********
>
> > >
> > > 13/03/06 12:16:34 DEBUG manager.OracleManager$ConnCache: Caching
> released****
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.DataDrivenImportJob: Using table
> class:****
>
> > > BTTN********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.DataDrivenImportJob: Using
> InputFormat:****
>
> > > class com.cloudera.sqoop.mapreduce.db.OracleDataDrivenDBInputFormat****
> ****
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/sqoop-1.4.2.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/ojdbc6.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/sqoop-1.4.2.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/sqoop-1.4.2.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/ojdbc6.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/paranamer-2.3.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/avro-1.5.3.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/commons-io-1.4.jar********
>
> > >
> > > 13/03/06 12:16:34 DEBUG mapreduce.JobBase: Adding to job classpath:***
> *
>
> > > file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar********
>
> > >
> > > 13/03/06 12:16:35 INFO db.DataDrivenDBInputFormat: BoundingValsQuery:*
> ***
>
> > > SELECT MIN(BTTN_ID), MAX(BTTN_ID) FROM BTTN********
>
> > >
> > > 13/03/06 12:16:35 INFO mapred.JobClient: Running job:
> job_201303051835_0014****
>
> > > ****
> > >
> > > 13/03/06 12:16:36 INFO mapred.JobClient:  map 0% reduce 0%****
> > >
> > > 13/03/06 12:16:52 INFO mapred.JobClient:  map 25% reduce 0%****
> > >
> > > 13/03/06 12:17:01 INFO mapred.JobClient:  map 75% reduce 0%****
> > >
> > > 13/03/06 12:17:13 INFO mapred.JobClient:  map 100% reduce 0%********
>
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient: Job complete:****
>
> > > job_201303051835_0014****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient: Counters: 18****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   Job Counters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:
> SLOTS_MILLIS_MAPS=71591********
>
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Total time spent by all**
> **
>
> > > reduces waiting after reserving slots (ms)=0********
>
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Total time spent by all
> maps****
>
> > > waiting after reserving slots (ms)=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Launched map tasks=4****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   File Output Format
> Counters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Bytes
> Written=184266237****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   FileSystemCounters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     HDFS_BYTES_READ=454****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:
> FILE_BYTES_WRITTEN=117870****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:
> HDFS_BYTES_WRITTEN=184266237*
> > > ***
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   File Input Format
> Counters****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Bytes Read=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:   Map-Reduce Framework****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Map input
> records=738183********
>
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Physical memory (bytes)**
> **
>
> > > snapshot=735293440****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Spilled Records=0****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     CPU time spent
> (ms)=28950********
>
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Total committed heap usage
> ****
>
> > > (bytes)=581435392********
>
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Virtual memory (bytes)***
> *
>
> > > snapshot=6086893568****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     Map output
> records=738183****
> > >
> > > 13/03/06 12:17:19 INFO mapred.JobClient:     SPLIT_RAW_BYTES=454******
> **
>
> > >
> > > 13/03/06 12:17:19 INFO mapreduce.ImportJobBase: Transferred 175.73 MB
> in****
>
> > > 45.4959 seconds (3.8625 MB/sec)****
> > >
> > > 13/03/06 12:17:19 INFO mapreduce.ImportJobBase: Retrieved 738183
> records.*
> > > ***
> > >
> > > 13/03/06 12:17:19 DEBUG hive.HiveImport: Hive.inputTable: BTTN****
> > >
> > > 13/03/06 12:17:19 DEBUG hive.HiveImport: Hive.outputTable: bttn*******
> *
>
> > >
> > > 13/03/06 12:17:19 DEBUG manager.OracleManager: Using column names
> query:****
>
> > > SELECT t.* FROM BTTN t WHERE 1=0********
>
> > >
> > > 13/03/06 12:17:19 DEBUG manager.OracleManager$ConnCache: Got cached***
> *
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:17:19 INFO manager.OracleManager: Time zone has been set
> to GMT****
>
> > > ********
>
> > >
> > > 13/03/06 12:17:19 DEBUG manager.SqlManager: Using fetchSize for next**
> **
>
> > > query: 1000********
>
> > >
> > > 13/03/06 12:17:19 INFO manager.SqlManager: Executing SQL statement:
> SELECT****
>
> > > t.* FROM BTTN t WHERE 1=0********
>
> > >
> > > 13/03/06 12:17:19 DEBUG manager.OracleManager$ConnCache: Caching
> released****
>
> > > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER****
> ****
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BTTN_ID had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column DATA_INST_ID had to
> be****
>
> > > cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column SCR_ID had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BTTN_NU had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column CAT had to be cast
> to a****
>
> > > less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column WDTH had to be cast
> to****
>
> > > a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column HGHT had to be cast
> to****
>
> > > a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column KEY_SCAN had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column KEY_SHFT had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BLM_FL had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column LCLZ_FL had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column MENU_ITEM_NU had to
> be****
>
> > > cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BTTN_ASGN_LVL_ID
> had to****
>
> > > be cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column ON_ATVT had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column ON_CLIK had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column ENBL_FL had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BLM_SET_ID had to be
> ****
>
> > > cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column MKT_ID had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column CRTE_TS had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column UPDT_TS had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column DEL_TS had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column DLTD_FL had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column PRD_CD had to be
> cast****
>
> > > to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column SOUND_FILE_ID had
> to be****
>
> > > cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column IS_DYNMC_BTTN had
> to be****
>
> > > cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column FRGND_CPTN_COLR_ID
> had****
>
> > > to be cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column
> FRGND_CPTN_COLR_PRSD_ID****
>
> > > had to be cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column BKGD_CPTN_COLR_ID
> had****
>
> > > to be cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 WARN hive.TableDefWriter: Column
> BKGD_CPTN_COLR_PRSD_ID****
>
> > > had to be cast to a less precise type in Hive********
>
> > >
> > > 13/03/06 12:17:19 DEBUG hive.TableDefWriter: Create statement: CREATE
> > > TABLE `bttn` ( `BTTN_ID` DOUBLE, `DATA_INST_ID` DOUBLE, `SCR_ID`
> DOUBLE,
> > > `BTTN_NU` DOUBLE, `CAT` DOUBLE, `WDTH` DOUBLE, `HGHT` DOUBLE,
> `KEY_SCAN`
> > > DOUBLE, `KEY_SHFT` DOUBLE, `FRGND_CPTN_COLR` STRING,
> `FRGND_CPTN_COLR_PRSD`
> > > STRING, `BKGD_CPTN_COLR` STRING, `BKGD_CPTN_COLR_PRSD` STRING, `BLM_FL`
> > > DOUBLE, `LCLZ_FL` DOUBLE, `MENU_ITEM_NU` DOUBLE, `BTTN_ASGN_LVL_ID`
> DOUBLE,
> > > `ON_ATVT` DOUBLE, `ON_CLIK` DOUBLE, `ENBL_FL` DOUBLE, `BLM_SET_ID`
> DOUBLE,
> > > `BTTN_ASGN_LVL_NAME` STRING, `MKT_ID` DOUBLE, `CRTE_TS` STRING,
> > > `CRTE_USER_ID` STRING, `UPDT_TS` STRING, `UPDT_USER_ID` STRING,
> `DEL_TS`
> > > STRING, `DEL_USER_ID` STRING, `DLTD_FL` DOUBLE, `MENU_ITEM_NA` STRING,
> > > `PRD_CD` DOUBLE, `BLM_SET_NA` STRING, `SOUND_FILE_ID` DOUBLE,
> > > `IS_DYNMC_BTTN` DOUBLE, `FRGND_CPTN_COLR_ID` DOUBLE,
> > > `FRGND_CPTN_COLR_PRSD_ID` DOUBLE, `BKGD_CPTN_COLR_ID` DOUBLE,
> > > `BKGD_CPTN_COLR_PRSD_ID` DOUBLE) COMMENT 'Imported by sqoop on
> 2013/03/06
> > > 12:17:19' ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' LINES
> TERMINATED****
>
> > > BY '\012' STORED AS TEXTFILE********
>
> > >
> > > 13/03/06 12:17:19 DEBUG hive.TableDefWriter: Load statement: LOAD DATA
> ****
>
> > > INPATH 'hdfs://NHCLT-PC44-2:8020/user/hadoop/BTTN' INTO TABLE
> `bttn`********
>
> > >
> > > 13/03/06 12:17:19 INFO hive.HiveImport: Removing temporary files from*
> ***
>
> > > import process: hdfs://NHCLT-PC44-2:8020/user/hadoop/BTTN/_logs****
> > >
> > > 13/03/06 12:17:19 INFO hive.HiveImport: Loading uploaded data into
> Hive***
> > > *
> > >
> > > 13/03/06 12:17:19 DEBUG hive.HiveImport: Using external Hive
> process.********
>
> > >
> > > 13/03/06 12:17:21 INFO hive.HiveImport: Logging initialized using****
>
> > > configuration in file:/home/hadoop/Hive/conf/hive-log4j.properties****
> ****
>
> > >
> > > 13/03/06 12:17:21 INFO hive.HiveImport: Hive history
> > >
> file=/home/hadoop/tmp/hadoop/hive_job_log_hadoop_201303061217_790206596.txt
> ****
>
> > > ****
> > >
> > > 13/03/06 12:17:32 INFO hive.HiveImport: OK****
> > >
> > > 13/03/06 12:17:32 INFO hive.HiveImport: Time taken: 11.345 seconds****
> ****
>
> > >
> > > 13/03/06 12:17:32 INFO hive.HiveImport: Loading data to table
> default.bttn****
>
> > > ********
>
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Table default.bttn stats:
> > > [num_partitions: 0, num_files: 5, num_rows: 0, total_size: 184266237,*
> ***
>
> > > raw_data_size: 0]****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: OK****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Time taken: 0.36 seconds****
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Hive import complete.********
>
> > >
> > > 13/03/06 12:17:33 INFO hive.HiveImport: Export directory is empty,****
>
> > > removing it.****
> > >
> > > *[hadoop@NHCLT-PC44-2 sqoop-oper]$*****
> > >
> > > * *****
> > >
> > > Now I have fired the following query at hive prompt:à****
> > >
> > >  ****
> > >
> > > *hive>* *select count(1) from bttn;*****
> > >
> > > Total MapReduce jobs = 1****
> > >
> > > Launching Job 1 out of 1****
> > >
> > > Number of reduce tasks determined at compile time: 1****
> > >
> > > In order to change the average load for a reducer (in bytes):****
> > >
> > >   set hive.exec.reducers.bytes.per.reducer=<number>****
> > >
> > > In order to limit the maximum number of reducers:****
> > >
> > >   set hive.exec.reducers.max=<number>****
> > >
> > > In order to set a constant number of reducers:****
> > >
> > >   set mapred.reduce.tasks=<number>********
>
> > >
> > > Starting Job = job_201303051835_0016, Tracking URL =****
>
> > > http://NHCLT-PC44-2:50030/jobdetails.jsp?jobid=job_201303051835_0016<
> http://nhclt-pc44-2:50030/jobdetails.jsp?jobid=job_201303051835_0016>
> > > ********
>
> > >
> > > Kill Command = /home/hadoop/hadoop-1.0.3/bin/hadoop job  -kill****
>
> > > job_201303051835_0016********
>
> > >
> > > Hadoop job information for Stage-1: number of mappers: 1; number of***
> *
>
> > > reducers: 1****
> > >
> > > 2013-03-06 12:21:45,350 Stage-1 map = 0%,  reduce = 0%****
> > >
> > > 2013-03-06 12:21:54,370 Stage-1 map = 76%,  reduce = 0%********
>
> > >
> > > 2013-03-06 12:21:57,378 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:21:58,381 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:21:59,385 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:00,388 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:01,391 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:02,394 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:03,397 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:04,401 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:05,404 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:06,407 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:07,410 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:08,415 Stage-1 map = 100%,  reduce = 0%, Cumulative
> CPU****
>
> > > 3.68 sec********
>
> > >
> > > 2013-03-06 12:22:09,418 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec********
>
> > >
> > > 2013-03-06 12:22:10,421 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec********
>
> > >
> > > 2013-03-06 12:22:11,425 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec********
>
> > >
> > > 2013-03-06 12:22:12,428 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec********
>
> > >
> > > 2013-03-06 12:22:13,432 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec********
>
> > >
> > > 2013-03-06 12:22:14,435 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec********
>
> > >
> > > 2013-03-06 12:22:15,439 Stage-1 map = 100%,  reduce = 100%, Cumulative
> CPU****
>
> > > 7.37 sec****
> > >
> > > MapReduce Total cumulative CPU time: 7 seconds 370 msec****
> > >
> > > Ended Job = job_201303051835_0016****
> > >
> > > MapReduce Jobs Launched:********
>
> > >
> > > Job 0: Map: 1  Reduce: 1   Cumulative CPU: 7.37 sec   HDFS Read:
> 184270926****
>
> > > HDFS Write: 7 SUCCESS****
> > >
> > > Total MapReduce CPU Time Spent: 7 seconds 370 msec****
> > >
> > > OK****
> > >
> > > *739169*****
> > >
> > > Time taken: 37.639 seconds****
> > >
> > > *hive>*****
> > >
> > >  ****
> > >
> > > The data present in *oracle database* is *738183 *but imported into*
> HIVE
> > > *is* 739169.*****
> > >
> > > Can you please let me know the issue.****
> > >
> > > I have already pasted all my screen log hare.****
> > >
> > > * *****
> > >
> > > *Thanks and Regards,*****
> > >
> > > *Ajit Kumar Shreevastava*****
> > >
> > >
> > > ::DISCLAIMER::
> > >
> > >
> > >
> ----------------------------------------------------------------------------------------------------------------------------------------------------
> > > ********
>
> > >
> > > The contents of this e-mail and any attachment(s) are confidential and
> > > intended for the named recipient(s) only.
> > > E-mail transmission is not guaranteed to be secure or error-free as
> > > information could be intercepted, corrupted,
> > > lost, destroyed, arrive late or incomplete, or may contain viruses in
> > > transmission. The e mail and its contents
> > > (with or without referred errors) shall therefore not attach any
> liability
> > > on the originator or HCL or its affiliates.
> > > Views or opinions, if any, presented in this email are solely those of
> the
> > > author and may not necessarily reflect the
> > > views or opinions of HCL or its affiliates. Any form of reproduction,
> > > dissemination, copying, disclosure, modification,
> > > distribution and / or publication of this message without the prior
> > > written consent of authorized representative of
> > > HCL is strictly prohibited. If you have received this email in error
> > > please delete it and notify the sender immediately.
> > > Before opening any email and/or attachments, please check them for
> viruses
> > > and other defects.****
>
> > > ****
> > >
> > >
> > >
> ----------------------------------------------------------------------------------------------------------------------------------------------------
> > > ****
> > >
> > >  ** ******
>
> ** **
>

Mime
View raw message