sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jarek Jarcec Cecho <jar...@apache.org>
Subject Re: Exporting hive table data into oracle give date format error
Date Sat, 16 Mar 2013 22:58:37 GMT
[-CC hive@user.apache.org]

Hi Ajit,
would you mind upgrading to Sqoop 1.4.3? We've improved the logging for this particular exception,
so it should significantly help in triangulating your issue.

Jarcec

On Wed, Mar 13, 2013 at 01:43:11PM +0000, Ajit Kumar Shreevastava wrote:
> Hi All,
> 
> Can you please let me know how can I bypass this error. I am currently using Apache 
SQOOP version 1.4.2.
> 
> 
> [hadoop@NHCLT-PC44-2 sqoop-oper]$ sqoop export --connect jdbc:oracle:thin:@10.99.42.11:1521/clouddb
--username HDFSUSER  --table BTTN_BKP_TEST --export-dir  /home/hadoop/user/hive/warehouse/bttn_bkp
-P -m 1  --input-fields-terminated-by '\0001' --verbose --input-null-string '\\N' --input-null-non-string
'\\N'
> 
> Please set $HBASE_HOME to the root of your HBase installation.
> 13/03/13 18:20:42 DEBUG tool.BaseSqoopTool: Enabled debug logging.
> Enter password:
> 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Loaded manager factory: com.cloudera.sqoop.manager.DefaultManagerFactory
> 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Trying ManagerFactory: com.cloudera.sqoop.manager.DefaultManagerFactory
> 13/03/13 18:20:47 DEBUG manager.DefaultManagerFactory: Trying with scheme: jdbc:oracle:thin:@10.99.42.11
> 13/03/13 18:20:47 DEBUG manager.OracleManager$ConnCache: Instantiated new connection
cache.
> 13/03/13 18:20:47 INFO manager.SqlManager: Using default fetchSize of 1000
> 13/03/13 18:20:47 DEBUG sqoop.ConnFactory: Instantiated ConnManager org.apache.sqoop.manager.OracleManager@74b23210
> 13/03/13 18:20:47 INFO tool.CodeGenTool: Beginning code generation
> 13/03/13 18:20:47 DEBUG manager.OracleManager: Using column names query: SELECT t.* FROM
BTTN_BKP_TEST t WHERE 1=0
> 13/03/13 18:20:47 DEBUG manager.OracleManager: Creating a new connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb,
using username: HDFSUSER
> 13/03/13 18:20:47 DEBUG manager.OracleManager: No connection paramenters specified. Using
regular API for making connection.
> 13/03/13 18:20:47 INFO manager.OracleManager: Time zone has been set to GMT
> 13/03/13 18:20:47 DEBUG manager.SqlManager: Using fetchSize for next query: 1000
> 13/03/13 18:20:47 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM BTTN_BKP_TEST
t WHERE 1=0
> 13/03/13 18:20:47 DEBUG manager.OracleManager$ConnCache: Caching released connection
for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 13/03/13 18:20:47 DEBUG orm.ClassWriter: selected columns:
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DATA_INST_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   SCR_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_NU
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CAT
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   WDTH
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   HGHT
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   KEY_SCAN
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   KEY_SHFT
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_FL
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   LCLZ_FL
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MENU_ITEM_NU
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ON_ATVT
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ON_CLIK
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   ENBL_FL
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_SET_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BTTN_ASGN_LVL_NAME
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MKT_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CRTE_TS
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   CRTE_USER_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   UPDT_TS
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   UPDT_USER_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DEL_TS
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DEL_USER_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   DLTD_FL
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   MENU_ITEM_NA
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   PRD_CD
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BLM_SET_NA
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   SOUND_FILE_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   IS_DYNMC_BTTN
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   FRGND_CPTN_COLR_PRSD_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter:   BKGD_CPTN_COLR_PRSD_ID
> 13/03/13 18:20:47 DEBUG orm.ClassWriter: Writing source file: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.java
> 13/03/13 18:20:47 DEBUG orm.ClassWriter: Table name: BTTN_BKP_TEST
> 13/03/13 18:20:47 DEBUG orm.ClassWriter: Columns: BTTN_ID:2, DATA_INST_ID:2, SCR_ID:2,
BTTN_NU:2, CAT:2, WDTH:2, HGHT:2, KEY_SCAN:2, KEY_SHFT:2, FRGND_CPTN_COLR:12, FRGND_CPTN_COLR_PRSD:12,
BKGD_CPTN_COLR:12, BKGD_CPTN_COLR_PRSD:12, BLM_FL:2, LCLZ_FL:2, MENU_ITEM_NU:2, BTTN_ASGN_LVL_ID:2,
ON_ATVT:2, ON_CLIK:2, ENBL_FL:2, BLM_SET_ID:2, BTTN_ASGN_LVL_NAME:12, MKT_ID:2, CRTE_TS:93,
CRTE_USER_ID:12, UPDT_TS:93, UPDT_USER_ID:12, DEL_TS:93, DEL_USER_ID:12, DLTD_FL:2, MENU_ITEM_NA:12,
PRD_CD:2, BLM_SET_NA:12, SOUND_FILE_ID:2, IS_DYNMC_BTTN:2, FRGND_CPTN_COLR_ID:2, FRGND_CPTN_COLR_PRSD_ID:2,
BKGD_CPTN_COLR_ID:2, BKGD_CPTN_COLR_PRSD_ID:2,
> 13/03/13 18:20:47 DEBUG orm.ClassWriter: sourceFilename is BTTN_BKP_TEST.java
> 13/03/13 18:20:47 DEBUG orm.CompilationManager: Found existing /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/
> 13/03/13 18:20:47 INFO orm.CompilationManager: HADOOP_HOME is /home/hadoop/hadoop-1.0.3/libexec/..
> 13/03/13 18:20:47 DEBUG orm.CompilationManager: Adding source file: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.java
> 13/03/13 18:20:47 DEBUG orm.CompilationManager: Invoking javac with args:
> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -sourcepath
> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/
> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -d
> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/
> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   -classpath
> 13/03/13 18:20:47 DEBUG orm.CompilationManager:   /home/hadoop/hadoop-1.0.3/libexec/../conf:/usr/java/jdk1.6.0_32/lib/tools.jar:/home/hadoop/hadoop-1.0.3/libexec/..:/home/hadoop/hadoop-1.0.3/libexec/../hadoop-core-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/asm-3.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjrt-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/aspectjtools-1.6.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-cli-1.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-codec-1.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-collections-3.2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-configuration-1.6.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-daemon-1.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-digester-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-el-1.0.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-httpclient-3.0.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-io-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-lang-2.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-1.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-logging-api-1.0.4.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-math-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/commons-net-1.4.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/core-3.1.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-core-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-compiler-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jasper-runtime-5.5.12.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jdeb-0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-core-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-json-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jersey-server-1.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jets3t-0.6.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jetty-util-6.1.26.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsch-0.1.42.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/junit-4.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/kfs-0.2.2.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/log4j-1.2.15.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/mockito-all-1.8.5.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/oro-2.0.8.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/servlet-api-2.5-20081211.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-api-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/slf4j-log4j12-1.4.3.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/xmlenc-0.52.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-2.1.jar:/home/hadoop/hadoop-1.0.3/libexec/../lib/jsp-2.1/jsp-api-2.1.jar:/home/hadoop/sqoop/conf::/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop/lib/avro-1.5.3.jar:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop/lib/commons-io-1.4.jar:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar:/home/hadoop/sqoop/lib/ojdbc6.jar:/home/hadoop/sqoop/lib/paranamer-2.3.jar:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar:/home/hadoop/sqoop/sqoop-test-1.4.2.jar::/home/hadoop/hadoop-1.0.3/hadoop-core-1.0.3.jar:/home/hadoop/sqoop/sqoop-1.4.2.jar
> Note: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.java uses
or overrides a deprecated API.
> Note: Recompile with -Xlint:deprecation for details.
> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Could not rename /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.java
to /home/hadoop/sqoop-oper/./BTTN_BKP_TEST.java
> org.apache.commons.io.FileExistsException: Destination '/home/hadoop/sqoop-oper/./BTTN_BKP_TEST.java'
already exists
>         at org.apache.commons.io.FileUtils.moveFile(FileUtils.java:2378)
>         at org.apache.sqoop.orm.CompilationManager.compile(CompilationManager.java:227)
>         at org.apache.sqoop.tool.CodeGenTool.generateORM(CodeGenTool.java:83)
>         at org.apache.sqoop.tool.ExportTool.exportTable(ExportTool.java:64)
>         at org.apache.sqoop.tool.ExportTool.run(ExportTool.java:97)
>         at org.apache.sqoop.Sqoop.run(Sqoop.java:145)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>         at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:181)
>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:220)
>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:229)
>         at org.apache.sqoop.Sqoop.main(Sqoop.java:238)
>         at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)
> 13/03/13 18:20:48 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.jar
> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Scanning for .class files in directory:
/tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531
> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Got classfile: /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.class
-> BTTN_BKP_TEST.class
> 13/03/13 18:20:48 DEBUG orm.CompilationManager: Finished writing jar file /tmp/sqoop-hadoop/compile/69b6a9d2ebb99cebced808e559528531/BTTN_BKP_TEST.jar
> 13/03/13 18:20:48 INFO mapreduce.ExportJobBase: Beginning export of BTTN_BKP_TEST
> 13/03/13 18:20:48 DEBUG mapreduce.JobBase: Using InputFormat: class org.apache.sqoop.mapreduce.ExportInputFormat
> 13/03/13 18:20:49 DEBUG manager.OracleManager$ConnCache: Got cached connection for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 13/03/13 18:20:49 INFO manager.OracleManager: Time zone has been set to GMT
> 13/03/13 18:20:49 DEBUG manager.OracleManager$ConnCache: Caching released connection
for jdbc:oracle:thin:@10.99.42.11:1521/clouddb/HDFSUSER
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ojdbc6.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/sqoop-1.4.2.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/jackson-mapper-asl-1.7.3.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/hsqldb-1.8.0.10.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/avro-ipc-1.5.3.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/jopt-simple-3.2.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ojdbc6.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/jackson-core-asl-1.7.3.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ant-contrib-1.0b3.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/ant-eclipse-1.0-jvm1.2.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/snappy-java-1.0.3.2.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/paranamer-2.3.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/avro-1.5.3.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/commons-io-1.4.jar
> 13/03/13 18:20:49 DEBUG mapreduce.JobBase: Adding to job classpath: file:/home/hadoop/sqoop/lib/avro-mapred-1.5.3.jar
> 13/03/13 18:20:49 INFO input.FileInputFormat: Total input paths to process : 1
> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Target numMapTasks=1
> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Total input bytes=172704981
> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: maxSplitSize=172704981
> 13/03/13 18:20:49 INFO input.FileInputFormat: Total input paths to process : 1
> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat: Generated splits:
> 13/03/13 18:20:49 DEBUG mapreduce.ExportInputFormat:   Paths:/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:0+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:67108864+67108864,/home/hadoop/user/hive/warehouse/bttn_bkp/000000_0:134217728+38487253
Locations:NHCLT-PC44-2:;
> 13/03/13 18:20:49 INFO mapred.JobClient: Running job: job_201303121648_0018
> 13/03/13 18:20:50 INFO mapred.JobClient:  map 0% reduce 0%
> 13/03/13 18:21:06 INFO mapred.JobClient:  map 8% reduce 0%
> 13/03/13 18:21:09 INFO mapred.JobClient:  map 13% reduce 0%
> 13/03/13 18:21:12 INFO mapred.JobClient:  map 17% reduce 0%
> 13/03/13 18:21:15 INFO mapred.JobClient:  map 21% reduce 0%
> 13/03/13 18:21:18 INFO mapred.JobClient:  map 26% reduce 0%
> 13/03/13 18:21:21 INFO mapred.JobClient:  map 30% reduce 0%
> 13/03/13 18:21:24 INFO mapred.JobClient:  map 35% reduce 0%
> 13/03/13 18:21:27 INFO mapred.JobClient:  map 40% reduce 0%
> 13/03/13 18:21:30 INFO mapred.JobClient:  map 45% reduce 0%
> 13/03/13 18:21:33 INFO mapred.JobClient:  map 50% reduce 0%
> 13/03/13 18:21:36 INFO mapred.JobClient:  map 53% reduce 0%
> 13/03/13 18:21:39 INFO mapred.JobClient:  map 58% reduce 0%
> 13/03/13 18:21:42 INFO mapred.JobClient:  map 62% reduce 0%
> 13/03/13 18:21:45 INFO mapred.JobClient:  map 65% reduce 0%
> 13/03/13 18:21:47 INFO mapred.JobClient: Task Id : attempt_201303121648_0018_m_000000_0,
Status : FAILED
> java.lang.IllegalArgumentException: Timestamp format must be yyyy-mm-dd hh:mm:ss[.fffffffff]
>         at java.sql.Timestamp.valueOf(Timestamp.java:185)
>         at BTTN_BKP_TEST.__loadFromFields(BTTN_BKP_TEST.java:1331)
>         at BTTN_BKP_TEST.parse(BTTN_BKP_TEST.java:1148)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
>         at org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
>        at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>         at org.apache.hadoop.mapred.Child.main(Child.java:249)
> 
> 
> ::DISCLAIMER::
> ----------------------------------------------------------------------------------------------------------------------------------------------------
> 
> The contents of this e-mail and any attachment(s) are confidential and intended for the
named recipient(s) only.
> E-mail transmission is not guaranteed to be secure or error-free as information could
be intercepted, corrupted,
> lost, destroyed, arrive late or incomplete, or may contain viruses in transmission. The
e mail and its contents
> (with or without referred errors) shall therefore not attach any liability on the originator
or HCL or its affiliates.
> Views or opinions, if any, presented in this email are solely those of the author and
may not necessarily reflect the
> views or opinions of HCL or its affiliates. Any form of reproduction, dissemination,
copying, disclosure, modification,
> distribution and / or publication of this message without the prior written consent of
authorized representative of
> HCL is strictly prohibited. If you have received this email in error please delete it
and notify the sender immediately.
> Before opening any email and/or attachments, please check them for viruses and other
defects.
> 
> ----------------------------------------------------------------------------------------------------------------------------------------------------

Mime
View raw message