sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ajit Kumar Shreevastava <Ajit.Shreevast...@hcl.com>
Subject RE: Exporting hive table data into oracle give date format error
Date Tue, 19 Mar 2013 08:13:01 GMT
Hi Jercec,



Thank you for your valuable suggestions.



I have applied the below suggestion and re-do all the process again with the SQOOP1.4.3 (sqoop-1.4.3.bin__hadoop-1.0.0.tar.gz)
but I have face same below error again. Please suggest me.



Here I have created table in hive as  suggested by you.



hive> create table bttn_bkp_testing like bttn_bkp_test;

hive> insert OVERWRITE table bttn_bkp_testing

        > select * from bttn_bkp_test;



I am also attaching the error file generated by task tracker for your analysis.

It fails for bttn_id = 194628



I have queried both the table and records are like



hive> select * from bttn_bkp_testing

    > where bttn_id=194628;



194628.0        577019.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-04-19 23:25:48.78  ei009724        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        706360.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-05-21 01:01:53.629 ei103215        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        1620395.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-08-10 04:34:00.203 ei103215        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        1694103.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-11-08 01:09:15.136 ei103215        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        1831767.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-12-19 23:44:44.241 e0025129        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0



And

hive> select * from bttn_bkp_test_new

    > where bttn_id=194628;



194628.0        577019.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-04-19 23:25:48.78  ei009724        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        706360.0        8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-05-21 01:01:53.629 ei103215        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        1620395.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-08-10 04:34:00.203 ei103215        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        1694103.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-11-08 01:09:15.136 ei103215        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0

194628.0        1831767.0       8910.0  19.0    1.0     1.0     1.0     0.0     0.0     #FFFFFF
#FF0000 #FF0000 #FFFFFF 0.0     0.0     1646.0  NULL    NULL    NULL   1.0      NULL    null
   20.0    2012-12-19 23:44:44.241 e0025129        2013-01-18 09:11:30.245 DP_CQ4540     
 null    null    0.0     BLUEBERRY MUFFIN        7836.0 null     NULL    0.0     61259.0 61230.0
61230.0 61259.0



Regards,

Ajit Kumar Shreevastava



-----Original Message-----
From: Jarek Jarcec Cecho [mailto:jarcec@apache.org]
Sent: Sunday, March 17, 2013 4:29 AM
To: user@sqoop.apache.org
Subject: Re: Exporting hive table data into oracle give date format error



[-CC hive@user.apache.org<mailto:hive@user.apache.org>]



Hi Ajit,

would you mind upgrading to Sqoop 1.4.3? We've improved the logging for this particular exception,
so it should significantly help in triangulating your issue.



Jarcec



On Wed, Mar 13, 2013 at 01:43:11PM +0000, Ajit Kumar Shreevastava wrote:

> Hi All,

>

> Can you please let me know how can I bypass this error. I am currently using Apache 
SQOOP version 1.4.2.

>

>

> [hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop export --connect jdbc:oracle:thin:@10.99.42.11:1521/clouddb
--username HDFSUSER  --table BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp
-P -m 1  --input-fields-terminated-by '\0001' --verbose --input-null-string '\\N' --input-null-non-string
'\\N'

>

> Please set $HBASE_HOME to the root of your HBase installation.

> 13/03/13 18:20:42 DEBUG tool.BaseSqoopTool: Enabled debug logging.

> Enter password:

> 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Loaded manager factory:

> com.cloudera.sqoop.manager.DefaultManagerFactory

> 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Trying ManagerFactory:

> com.cloudera.sqoop.manager.DefaultManagerFactory

> 13/03/13 18:20:47 DEBUG manager.DefaultManagerFactory: Trying with

> scheme: jdbc:oracle:thin:@10.99.42.11

> 13/03/13 18:20:47 DEBUG manager.OracleManager$ConnCache: Instantiated new connection
cache.

> 13/03/13 18:20:47 INFO manager.SqlManager: Using default fetchSize of

> 1000

> 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Instantiated ConnManager

> org.apache.sqoop.manager.OracleManager@74b23210<mailto:org.apache.sqoop.manager.OracleManager@74b23210>

> 13/03/13 18:20:47 INFO tool.CodeGenTool: Beginning code generation

> 13/03/13 18:20:47 DEBUG manager.OracleManager: Using column names

> query: SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0

> 13/03/13 18:20:47 DEBUG manager.OracleManager: Creating a new

> connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb, using

> username: HDFSUSER

> 13/03/13 18:20:47 DEBUG manager.OracleManager: No connection paramenters specified. Using
regular API for making connection.

> 13/03/13 18:20:47 INFO manager.OracleManager: Time zone has been set

> to GMT

> 13/03/13 18:20:47 DEBUG manager.SqlManager: Using fetchSize for next

> query: 1000

> 13/03/13 18:20:47 INFO manager.SqlManager: Executing SQL statement:

> SELECT t.* FROM BTTN_BKP_TEST t WHERE 1=0

> 13/03/13 18:20:47 DEBUG manager.OracleManager$ConnCache: Caching

> released connection for

> jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER

> 13/03/13 18:20:47 DEBUG orm.ClassWriter: selected columns:

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DATA_INST_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   SCR_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_NU

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CAT

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   WDTH

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   HGHT

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   KEY_SCAN

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   KEY_SHFT

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_FL

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   LCLZ_FL

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MENU_ITEM_NU

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ON_ATVT

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ON_CLIK

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ENBL_FL

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_SET_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MKT_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CRTE_TS

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CRTE_USER_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   UPDT_TS

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   UPDT_USER_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DEL_TS

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DEL_USER_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DLTD_FL

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MENU_ITEM_NA

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   PRD_CD

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_SET_NA

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   SOUND_FILE_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID

> 13/03/13 18:20:47 DEBUG orm.ClassWriter: Writing source file:

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE

> ST.java

> 13/03/13 18:20:47 DEBUG orm.ClassWriter: Table name: BTTN_BKP_TEST

> 13/03/13 18:20:47 DEBUG orm.ClassWriter: Columns: BTTN_ID:2,

> DATA_INST_ID:2, SCR_ID:2, BTTN_NU:2, CAT:2, WDTH:2, HGHT:2,

> KEY_SCAN:2, KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12,

> BKGD_CPTN_COLR:12, BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2,

> MENU_ITEM_NU:2, BTTN_ASGN_LVL_ID:2, ON_ATVT:2, ON_CLIK:2, ENBL_FL:2,

> BLM_SET_ID:2, BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93,

> CRTE_USER_ID:12, UPDT_TS:93, UPDT_USER_ID:12, DEL_TS:93,

> DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12, PRD_CD:2, BLM_SET_NA:12,

> SOUND_FILE_ID:2, IS_DYNMC_BTTN:2, FRGND_CPTN_COLR_ID:2,

> FRGND_CPTN_COLR_PRSD_ID:2, BKGD_CPTN_COLR_ID:2,

> BKGD_CPTN_COLR_PRSD_ID:2,

> 13/03/13 18:20:47 DEBUG orm.ClassWriter: sourceFilename is

> BTTN_BKP_TEST.java

> 13/03/13 18:20:47 DEBUG orm.CompilationManager: Found existing

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/

> 13/03/13 18:20:47 INFO orm.CompilationManager: HADOOP_HOME is /home/hadoop/hadoop-1.0.3/libexec/..

> 13/03/13 18:20:47 DEBUG orm.CompilationManager: Adding source file:

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE

> ST.java

> 13/03/13 18:20:47 DEBUG orm.CompilationManager: Invoking javac with args:

> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -sourcepath

> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/

> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -d

> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/

> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -classpath

> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar

> Note: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.java uses
or overrides a deprecated API.

> Note: Recompile with -Xlint:deprecation for details.

> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Could not rename

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE

> ST.java to /home/hadoop/sqoop-oper/./BTTN_BKP_TEST.java

> org.apache.commons.io.FileExistsException: Destination '/home/hadoop/sqoop-oper/./BTTN_BKP_TEST.java'
already exists

>         at org.apache.commons.io.FileUtils.moveFile(FileUtils.java:2378)

>         at org.apache.sqoop.orm.CompilationManager.compile(CompilationManager.java:227)

>         at org.apache.sqoop.tool.CodeGenTool.generateORM(CodeGenTool.java:83)

>         at org.apache.sqoop.tool.ExportTool.exportTable(ExportTool.java:64)

>         at org.apache.sqoop.tool.ExportTool.run(ExportTool.java:97)

>         at org.apache.sqoop.Sqoop.run(Sqoop.java:145)

>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)

>         at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:181)

>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:220)

>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:229)

>         at org.apache.sqoop.Sqoop.main(Sqoop.java:238)

>         at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)

> 13/03/13 18:20:48 INFO orm.CompilationManager: Writing jar file:

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE

> ST.jar

> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Scanning for .class

> files in directory:

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531

> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Got classfile:

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE

> ST.class -> BTTN_BKP_TEST.class

> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Finished writing jar

> file

> /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TE

> ST.jar

> 13/03/13 18:20:48 INFO mapreduce.ExportJobBase: Beginning export of

> BTTN_BKP_TEST

> 13/03/13 18:20:48 DEBUG mapreduce.JobBase: Using InputFormat: class

> org.apache.sqoop.mapreduce.ExportInputFormat

> 13/03/13 18:20:49 DEBUG manager.OracleManager$ConnCache: Got cached

> connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER

> 13/03/13 18:20:49 INFO manager.OracleManager: Time zone has been set

> to GMT

> 13/03/13 18:20:49 DEBUG manager.OracleManager$ConnCache: Caching

> released connection for

> jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/sqoop-1.4.2.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/ojdbc6.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/sqoop-1.4.2.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/sqoop-1.4.2.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/ojdbc6.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/paranamer-2.3.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/avro-1.5.3.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/commons-io-1.4.jar

> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath:

> file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar

> 13/03/13 18:20:49 INFO input.FileInputFormat: Total input paths to

> process : 1

> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Target

> numMapTasks=1

> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Total input

> bytes=172704981

> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat:

> maxSplitSize=172704981

> 13/03/13 18:20:49 INFO input.FileInputFormat: Total input paths to

> process : 1

> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Generated splits:

> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat:   Paths:/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:0+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:67108864+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:134217728+38487253
Locations:NHCLT-PC44-2:;

> 13/03/13 18:20:49 INFO mapred.JobClient: Running job:

> job_201303121648_0018

> 13/03/13 18:20:50 INFO mapred.JobClient:  map 0% reduce 0%

> 13/03/13 18:21:06 INFO mapred.JobClient:  map 8% reduce 0%

> 13/03/13 18:21:09 INFO mapred.JobClient:  map 13% reduce 0%

> 13/03/13 18:21:12 INFO mapred.JobClient:  map 17% reduce 0%

> 13/03/13 18:21:15 INFO mapred.JobClient:  map 21% reduce 0%

> 13/03/13 18:21:18 INFO mapred.JobClient:  map 26% reduce 0%

> 13/03/13 18:21:21 INFO mapred.JobClient:  map 30% reduce 0%

> 13/03/13 18:21:24 INFO mapred.JobClient:  map 35% reduce 0%

> 13/03/13 18:21:27 INFO mapred.JobClient:  map 40% reduce 0%

> 13/03/13 18:21:30 INFO mapred.JobClient:  map 45% reduce 0%

> 13/03/13 18:21:33 INFO mapred.JobClient:  map 50% reduce 0%

> 13/03/13 18:21:36 INFO mapred.JobClient:  map 53% reduce 0%

> 13/03/13 18:21:39 INFO mapred.JobClient:  map 58% reduce 0%

> 13/03/13 18:21:42 INFO mapred.JobClient:  map 62% reduce 0%

> 13/03/13 18:21:45 INFO mapred.JobClient:  map 65% reduce 0%

> 13/03/13 18:21:47 INFO mapred.JobClient: Task Id :

> attempt_201303121648_0018_m_000000_0, Status : FAILED

> java.lang.IllegalArgumentException: Timestamp format must be yyyy-mm-dd hh:mm:ss[.fffffffff]

>         at java.sql.Timestamp.valueOf(Timestamp.java:185)

>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1331)

>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)

>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)

>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)

>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)

>        at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)

>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)

>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)

>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)

>         at java.security.AccessController.doPrivileged(Native Method)

>         at javax.security.auth.Subject.doAs(Subject.java:396)

>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)

>         at org.apache.hadoop.mapred.Child.main(Child.java:249)

>

>

> ::DISCLAIMER::

> ----------------------------------------------------------------------

> ----------------------------------------------------------------------

> --------

>

> The contents of this e-mail and any attachment(s) are confidential and intended for the
named recipient(s) only.

> E-mail transmission is not guaranteed to be secure or error-free as

> information could be intercepted, corrupted, lost, destroyed, arrive

> late or incomplete, or may contain viruses in transmission. The e mail and its contents
(with or without referred errors) shall therefore not attach any liability on the originator
or HCL or its affiliates.

> Views or opinions, if any, presented in this email are solely those of

> the author and may not necessarily reflect the views or opinions of

> HCL or its affiliates. Any form of reproduction, dissemination,

> copying, disclosure, modification, distribution and / or publication of this message
without the prior written consent of authorized representative of HCL is strictly prohibited.
If you have received this email in error please delete it and notify the sender immediately.

> Before opening any email and/or attachments, please check them for viruses and other
defects.

>

> ----------------------------------------------------------------------

> ----------------------------------------------------------------------

> --------

Mime
View raw message