sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jarek Jarcec Cecho <jar...@apache.org>
Subject Re: Exporting hive table data into oracle give date format error
Date Tue, 19 Mar 2013 21:25:58 GMT
Hi Ajit,
thank you for sharing the additional data. I've noticed in your data that some of the columns
are using \N to denote the NULL value, however some other columns are using string constant
"null" (that do not denote NULL in Hive). This also seems to be the case for column DEL_TS.
My guess is that Sqoop is trying to decode the "null" string as the timestamp and failing
on the "Timestamp format must be..." exception. I would recommend to unify the null representation
tokens and run Sqoop export with appropriate one.

Jarcec

On Tue, Mar 19, 2013 at 08:13:01AM +0000, Ajit Kumar Shreevastava wrote:
> Hi Jercec,
> 
> 
> 
> Thank you for your valuable suggestions.
> 
> 
> 
> I have applied the below suggestion and re-do all the process again with the SQOOP1.4.3
(sqoop-1.4.3.bin__hadoop-1.0.0.tar.gz) but I have face same below error again. Please suggest
me.
> 
> 
> 
> Here I have created table in hive as  suggested by you.
> 
> 
> 
> hive> create table bttn_bkp_testing like bttn_bkp_test;
> 
> hive> insert OVERWRITE table bttn_bkp_testing
> 
>         > select * from bttn_bkp_test;
> 
> 
> 
> I am also attaching the error file generated by task tracker for your analysis.
> 
> It fails for bttn_id = 194628
> 
> 
> 
> I have queried both the table and records are like
> 
> 
> 
> hive> select * from bttn_bkp_testing
> 
>     > where bttn_id=194628;
> 
> 
> 
> 194628.0        577019.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-04-19 23:25:48.78  ei009724        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        706360.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-05-21 01:01:53.629 ei103215        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        1620395.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-08-10 04:34:00.203 ei103215        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        1694103.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-11-08 01:09:15.136 ei103215        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        1831767.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-12-19 23:44:44.241 e0025129        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 
> 
> And
> 
> hive> select * from bttn_bkp_test_new
> 
>     > where bttn_id=194628;
> 
> 
> 
> 194628.0        577019.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-04-19 23:25:48.78  ei009724        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        706360.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-05-21 01:01:53.629 ei103215        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        1620395.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-08-10 04:34:00.203 ei103215        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        1694103.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-11-08 01:09:15.136 ei103215        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 194628.0        1831767.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0 
   #FFFFFF #FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0    
 NULL    null    20.0    2012-12-19 23:44:44.241 e0025129        2013-01-18 09:11:30.245 DP_CQ4540
      null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0
61230.0 61230.0 61259.0
> 
> 
> 
> Regards,
> 
> Ajit Kumar Shreevastava
> 
> 
> 
> -----Original Message-----
> From: Jarek Jarcec Cecho [mailto:jarcec@apache.org]
> Sent: Sunday, March 17, 2013 4:29 AM
> To: user@sqoop.apache.org
> Subject: Re: Exporting hive table data into oracle give date format error
> 
> 
> 
> [-CC hive@user.apache.org<mailto:hive@user.apache.org>]
> 
> 
> 
> Hi Ajit,
> 
> would you mind upgrading to Sqoop 1.4.3? We've improved the logging for this particular
exception, so it should significantly help in triangulating your issue.
> 
> 
> 
> Jarcec
> 
> 
> 
> On Wed, Mar 13, 2013 at 01:43:11PM +0000, Ajit Kumar Shreevastava wrote:
> 
> > Hi All,
> 
> >
> 
> > Can you please let me know how can I bypass this error. I am currently using Apache
 SQOOP version 1.4.2.
> 
> >
> 
> >
> 
> > [hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop export --connect jdbc:oracle:thin:@10.99.42.11:1521/clouddb
--username HDFSUSER  --table BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp
-P -m 1  --input-fields-terminated-by '\0001' --verbose --input-null-string '\\N' --input-null-non-string
'\\N'
> 
> >
> 
> > Please set $HBASE_HOME to the root of your HBase installation.
> 
> > 13/03/13 18:20:42 DEBUG tool.BaseSqoopTool: Enabled debug logging.
> 
> > Enter password:
> 
> > 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Loaded manager factory:
> 
> > com.cloudera.sqoop.manager.DefaultManagerFactory
> 
> > 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Trying ManagerFactory:
> 
> > com.cloudera.sqoop.manager.DefaultManagerFactory
> 
> > 13/03/13 18:20:47 DEBUG manager.DefaultManagerFactory: Trying with
> 
> > scheme: jdbc:oracle:thin:@10.99.42.11
> 
> > 13/03/13 18:20:47 DEBUG manager.OracleManager$ConnCache: Instantiated new connection
cache.
> 
> > 13/03/13 18:20:47 INFO manager.SqlManager: Using default fetchSize of
> 
> > 1000
> 
> > 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Instantiated ConnManager
> 
> > org.apache.sqoop.manager.OracleManager@74b23210<mailto:org.apache.sqoop.manager.OracleManager@74b23210>
> 
> > 13/03/13 18:20:47 INFO tool.CodeGenTool: Beginning code generation
> 
> > 13/03/13 18:20:47 DEBUG manager.OracleManager: Using column names
> 
> > query: SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0
> 
> > 13/03/13 18:20:47 DEBUG manager.OracleManager: Creating a new
> 
> > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb, using
> 
> > username: HDFSUSER
> 
> > 13/03/13 18:20:47 DEBUG manager.OracleManager: No connection paramenters specified.
Using regular API for making connection.
> 
> > 13/03/13 18:20:47 INFO manager.OracleManager: Time zone has been set
> 
> > to GMT
> 
> > 13/03/13 18:20:47 DEBUG manager.SqlManager: Using fetchSize for next
> 
> > query: 1000
> 
> > 13/03/13 18:20:47 INFO manager.SqlManager: Executing SQL statement:
> 
> > SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0
> 
> > 13/03/13 18:20:47 DEBUG manager.OracleManager$ConnCache: Caching
> 
> > released connection for
> 
> > jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter: selected columns:
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DATA_INST_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   SCR_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_NU
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CAT
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   WDTH
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   HGHT
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   KEY_SCAN
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   KEY_SHFT
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_FL
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   LCLZ_FL
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MENU_ITEM_NU
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ON_ATVT
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ON_CLIK
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ENBL_FL
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_SET_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MKT_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CRTE_TS
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CRTE_USER_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   UPDT_TS
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   UPDT_USER_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DEL_TS
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DEL_USER_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DLTD_FL
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MENU_ITEM_NA
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   PRD_CD
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_SET_NA
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   SOUND_FILE_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter: Writing source file:
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE
> 
> > ST.java
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter: Table name: BTTN_BKP_TEST
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter: Columns: BTTN_ID:2,
> 
> > DATA_INST_ID:2, SCR_ID:2, BTTN_NU:2, CAT:2, WDTH:2, HGHT:2,
> 
> > KEY_SCAN:2, KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12,
> 
> > BKGD_CPTN_COLR:12, BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2,
> 
> > MENU_ITEM_NU:2, BTTN_ASGN_LVL_ID:2, ON_ATVT:2, ON_CLIK:2, ENBL_FL:2,
> 
> > BLM_SET_ID:2, BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93,
> 
> > CRTE_USER_ID:12, UPDT_TS:93, UPDT_USER_ID:12, DEL_TS:93,
> 
> > DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12, PRD_CD:2, BLM_SET_NA:12,
> 
> > SOUND_FILE_ID:2, IS_DYNMC_BTTN:2, FRGND_CPTN_COLR_ID:2,
> 
> > FRGND_CPTN_COLR_PRSD_ID:2, BKGD_CPTN_COLR_ID:2,
> 
> > BKGD_CPTN_COLR_PRSD_ID:2,
> 
> > 13/03/13 18:20:47 DEBUG orm.ClassWriter: sourceFilename is
> 
> > BTTN_BKP_TEST.java
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager: Found existing
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/
> 
> > 13/03/13 18:20:47 INFO orm.CompilationManager: HADOOP_HOME is /home/hadoop/hadoop-1.0.3/libexec/..
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager: Adding source file:
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE
> 
> > ST.java
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager: Invoking javac with args:
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -sourcepath
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -d
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -classpath
> 
> > 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 
> > Note: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.java
uses or overrides a deprecated API.
> 
> > Note: Recompile with -Xlint:deprecation for details.
> 
> > 13/03/13 18:20:48 DEBUG orm.CompilationManager: Could not rename
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE
> 
> > ST.java to /home/hadoop/sqoop-oper/./BTTN_BKP_TEST.java
> 
> > org.apache.commons.io.FileExistsException: Destination '/home/hadoop/sqoop-oper/./BTTN_BKP_TEST.java'
already exists
> 
> >         at org.apache.commons.io.FileUtils.moveFile(FileUtils.java:2378)
> 
> >         at org.apache.sqoop.orm.CompilationManager.compile(CompilationManager.java:227)
> 
> >         at org.apache.sqoop.tool.CodeGenTool.generateORM(CodeGenTool.java:83)
> 
> >         at org.apache.sqoop.tool.ExportTool.exportTable(ExportTool.java:64)
> 
> >         at org.apache.sqoop.tool.ExportTool.run(ExportTool.java:97)
> 
> >         at org.apache.sqoop.Sqoop.run(Sqoop.java:145)
> 
> >         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> 
> >         at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:181)
> 
> >         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:220)
> 
> >         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:229)
> 
> >         at org.apache.sqoop.Sqoop.main(Sqoop.java:238)
> 
> >         at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)
> 
> > 13/03/13 18:20:48 INFO orm.CompilationManager: Writing jar file:
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE
> 
> > ST.jar
> 
> > 13/03/13 18:20:48 DEBUG orm.CompilationManager: Scanning for .class
> 
> > files in directory:
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531
> 
> > 13/03/13 18:20:48 DEBUG orm.CompilationManager: Got classfile:
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE
> 
> > ST.class -> BTTN_BKP_TEST.class
> 
> > 13/03/13 18:20:48 DEBUG orm.CompilationManager: Finished writing jar
> 
> > file
> 
> > /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE
> 
> > ST.jar
> 
> > 13/03/13 18:20:48 INFO mapreduce.ExportJobBase: Beginning export of
> 
> > BTTN_BKP_TEST
> 
> > 13/03/13 18:20:48 DEBUG mapreduce.JobBase: Using InputFormat: class
> 
> > org.apache.sqoop.mapreduce.ExportInputFormat
> 
> > 13/03/13 18:20:49 DEBUG manager.OracleManager$ConnCache: Got cached
> 
> > connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 
> > 13/03/13 18:20:49 INFO manager.OracleManager: Time zone has been set
> 
> > to GMT
> 
> > 13/03/13 18:20:49 DEBUG manager.OracleManager$ConnCache: Caching
> 
> > released connection for
> 
> > jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/ojdbc6.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/ojdbc6.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/paranamer-2.3.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/avro-1.5.3.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/commons-io-1.4.jar
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:
> 
> > file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar
> 
> > 13/03/13 18:20:49 INFO input.FileInputFormat: Total input paths to
> 
> > process : 1
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Target
> 
> > numMapTasks=1
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Total input
> 
> > bytes=172704981
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat:
> 
> > maxSplitSize=172704981
> 
> > 13/03/13 18:20:49 INFO input.FileInputFormat: Total input paths to
> 
> > process : 1
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Generated splits:
> 
> > 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat:   Paths:/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:0+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:67108864+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:134217728+38487253
Locations:NHCLT-PC44-2:;
> 
> > 13/03/13 18:20:49 INFO mapred.JobClient: Running job:
> 
> > job_201303121648_0018
> 
> > 13/03/13 18:20:50 INFO mapred.JobClient:  map 0% reduce 0%
> 
> > 13/03/13 18:21:06 INFO mapred.JobClient:  map 8% reduce 0%
> 
> > 13/03/13 18:21:09 INFO mapred.JobClient:  map 13% reduce 0%
> 
> > 13/03/13 18:21:12 INFO mapred.JobClient:  map 17% reduce 0%
> 
> > 13/03/13 18:21:15 INFO mapred.JobClient:  map 21% reduce 0%
> 
> > 13/03/13 18:21:18 INFO mapred.JobClient:  map 26% reduce 0%
> 
> > 13/03/13 18:21:21 INFO mapred.JobClient:  map 30% reduce 0%
> 
> > 13/03/13 18:21:24 INFO mapred.JobClient:  map 35% reduce 0%
> 
> > 13/03/13 18:21:27 INFO mapred.JobClient:  map 40% reduce 0%
> 
> > 13/03/13 18:21:30 INFO mapred.JobClient:  map 45% reduce 0%
> 
> > 13/03/13 18:21:33 INFO mapred.JobClient:  map 50% reduce 0%
> 
> > 13/03/13 18:21:36 INFO mapred.JobClient:  map 53% reduce 0%
> 
> > 13/03/13 18:21:39 INFO mapred.JobClient:  map 58% reduce 0%
> 
> > 13/03/13 18:21:42 INFO mapred.JobClient:  map 62% reduce 0%
> 
> > 13/03/13 18:21:45 INFO mapred.JobClient:  map 65% reduce 0%
> 
> > 13/03/13 18:21:47 INFO mapred.JobClient: Task Id :
> 
> > attempt_201303121648_0018_m_000000_0, Status : FAILED
> 
> > java.lang.IllegalArgumentException: Timestamp format must be yyyy-mm-dd hh:mm:ss[.fffffffff]
> 
> >         at java.sql.Timestamp.valueOf(Timestamp.java:185)
> 
> >         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1331)
> 
> >         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)
> 
> >         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
> 
> >         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
> 
> >         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
> 
> >        at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
> 
> >         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
> 
> >         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
> 
> >         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
> 
> >         at java.security.AccessController.doPrivileged(Native Method)
> 
> >         at javax.security.auth.Subject.doAs(Subject.java:396)
> 
> >         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
> 
> >         at org.apache.hadoop.mapred.Child.main(Child.java:249)
> 
> >
> 
> >
> 
> > ::DISCLAIMER::
> 
> > ----------------------------------------------------------------------
> 
> > ----------------------------------------------------------------------
> 
> > --------
> 
> >
> 
> > The contents of this e-mail and any attachment(s) are confidential and intended
for the named recipient(s) only.
> 
> > E-mail transmission is not guaranteed to be secure or error-free as
> 
> > information could be intercepted, corrupted, lost, destroyed, arrive
> 
> > late or incomplete, or may contain viruses in transmission. The e mail and its contents
(with or without referred errors) shall therefore not attach any liability on the originator
or HCL or its affiliates.
> 
> > Views or opinions, if any, presented in this email are solely those of
> 
> > the author and may not necessarily reflect the views or opinions of
> 
> > HCL or its affiliates. Any form of reproduction, dissemination,
> 
> > copying, disclosure, modification, distribution and / or publication of this message
without the prior written consent of authorized representative of HCL is strictly prohibited.
If you have received this email in error please delete it and notify the sender immediately.
> 
> > Before opening any email and/or attachments, please check them for viruses and other
defects.
> 
> >
> 
> > ----------------------------------------------------------------------
> 
> > ----------------------------------------------------------------------
> 
> > --------



Mime
View raw message