sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Lee S <sle...@gmail.com>
Subject Re: execute sqoop job with exception
Date Mon, 17 Aug 2015 02:40:10 GMT
@Abe, Hive dependeces have been added to the comon.loader at
conf/catalina.properties.
And  this line
jars.add(ClassUtils.jarForClass("org.apache.hadoop.hive.conf.HiveConf"));
can find the hive jar which also contains the not founded
class(hive.ql.outputformat).


On Mon, Aug 17, 2015 at 12:30 AM, Abraham Elmahrek <abe@cloudera.com> wrote:

> Sqoop2 delegates Parquet support to Kite. Each file format might be a
> different code path in Kite.
>
> On Fri, Aug 14, 2015 at 6:52 PM, Lee S <sleefd@gmail.com> wrote:
>
>> Hi Abe:
>>  I'll  try to add hive deps in the codes and rebuild sqoop2.
>>  But I have set the file format to FileFormat.CSV,  why it comes about
>> the parquet.
>>
>> On Sat, Aug 15, 2015 at 1:00 AM, Abraham Elmahrek <abe@cloudera.com>
>> wrote:
>>
>>> Hey man,
>>>
>>> It looks like certain Hive jars are missing from the job for some
>>> reason. Seems like we need to add more jars to
>>> https://github.com/apache/sqoop/blob/sqoop2/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteFromInitializer.java#L71
>>> and
>>> https://github.com/apache/sqoop/blob/sqoop2/connector/connector-kite/src/main/java/org/apache/sqoop/connector/kite/KiteToInitializer.java#L78
>>> .
>>>
>>> I've created https://issues.apache.org/jira/browse/SQOOP-2489 to track
>>> this bug. I've also created
>>> https://issues.apache.org/jira/browse/SQOOP-2490 to provide a facility
>>> to work around these kinds of issues in the future.
>>>
>>> Sqoop2 is a work in progress and still needs some battle testing. With
>>> that in mind, can you use the Avro integration instead? Otherwise, you
>>> might need to rebuild Sqoop2 with the fix to get this working at the
>>> moment. Again, https://issues.apache.org/jira/browse/SQOOP-2490 will
>>> change all of that.
>>>
>>> -Abe
>>>
>>> On Fri, Aug 14, 2015 at 1:10 AM, Lee S <sleefd@gmail.com> wrote:
>>>
>>>> Hi all:
>>>>  I try to import from rdms to hive with kite connector in sqoop-shell.
>>>>  And I submit the job successfully , but when I try to track the job
>>>> staus
>>>> on yarn web-ui, I found there are errors in container log as below.
>>>> And the job continues running ,never stops.
>>>> The log shows *java.lang.NoClassDefFoundError:
>>>> org/apache/hadoop/hive/ql/io/HiveOutputFormat.*
>>>>
>>>> *Anybody can help?*
>>>>
>>>> *Showing 4096 bytes. Click here
>>>> <http://pdm-03:8042/node/containerlogs/container_1439241066552_0044_01_000002/root/stderr/?start=0>
for
>>>> full log*
>>>>
>>>> *utFormatLoadExecutor  - SqoopOutputFormatLoadExecutor consumer thread is
starting
>>>> 2015-08-14 15:39:10,765 [OutputFormatLoader-consumer] INFO  org.apache.sqoop.job.mr.SqoopOutputFormatLoadExecutor
 - Running loader class org.apache.sqoop.connector.kite.KiteLoader
>>>> 2015-08-14 15:39:10,771 [main] INFO  org.apache.sqoop.job.mr.SqoopMapper
 - Starting progress service
>>>> 2015-08-14 15:39:10,772 [main] INFO  org.apache.sqoop.job.mr.SqoopMapper
 - Running extractor class org.apache.sqoop.connector.jdbc.GenericJdbcExtractor
>>>> 2015-08-14 15:39:10,981 [OutputFormatLoader-consumer] INFO  org.apache.sqoop.connector.kite.KiteLoader
 - Constructed temporary dataset URI: dataset:hive:wangjun/temp_5bfec97713e04374b2f2efde2dc5e4de?auth:host=pdm-03&auth:port=9083
>>>> 2015-08-14 15:39:11,093 [main] INFO  org.apache.sqoop.connector.jdbc.GenericJdbcExtractor
 - Using query: SELECT id FROM bcpdm.history WHERE 1368 <= id AND id <= 1546
>>>> 2015-08-14 15:39:11,537 [OutputFormatLoader-consumer] ERROR org.apache.sqoop.job.mr.SqoopOutputFormatLoadExecutor
 - Error while loading data out of MR job.
>>>> java.lang.NoClassDefFoundError: org/apache/hadoop/hive/ql/io/HiveOutputFormat
>>>> 	at java.lang.ClassLoader.defineClass1(Native Method)
>>>> 	at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
>>>> 	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
>>>> 	at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
>>>> 	at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
>>>> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>>>> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>>> 	at java.security.AccessController.doPrivileged(Native Method)
>>>> 	at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>>> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>>> 	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>>> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>>> 	at java.lang.Class.forName0(Native Method)
>>>> 	at java.lang.Class.forName(Class.java:190)
>>>> 	at org.kitesdk.data.spi.hive.HiveUtils.getHiveParquetOutputFormat(HiveUtils.java:446)
>>>> 	at org.kitesdk.data.spi.hive.HiveUtils.<clinit>(HiveUtils.java:91)
>>>> 	at org.kitesdk.data.spi.hive.HiveManagedMetadataProvider.create(HiveManagedMetadataProvider.java:83)
>>>> 	at org.kitesdk.data.spi.hive.HiveManagedDatasetRepository.create(HiveManagedDatasetRepository.java:77)
>>>> 	at org.kitesdk.data.Datasets.create(Datasets.java:239)
>>>> 	at org.kitesdk.data.Datasets.create(Datasets.java:307)
>>>> 	at org.kitesdk.data.Datasets.create(Datasets.java:335)
>>>> 	at org.apache.sqoop.connector.kite.KiteDatasetExecutor.createDataset(KiteDatasetExecutor.java:70)
>>>> 	at org.apache.sqoop.connector.kite.KiteLoader.getExecutor(KiteLoader.java:52)
>>>> 	at org.apache.sqoop.connector.kite.KiteLoader.load(KiteLoader.java:62)
>>>> 	at org.apache.sqoop.connector.kite.KiteLoader.load(KiteLoader.java:36)
>>>> 	at org.apache.sqoop.job.mr.SqoopOutputFormatLoadExecutor$ConsumerThread.run(SqoopOutputFormatLoadExecutor.java:250)
>>>> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>>>> 	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> 	at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.hive.ql.io.HiveOutputFormat
>>>> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>>>> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>>> 	at java.security.AccessController.doPrivileged(Native Method)
>>>> 	at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>>> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>>> 	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>>> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>>> 	... 31 more
>>>> 2015-08-14 15:39:11,540 [main] INFO  org.apache.sqoop.job.mr.SqoopMapper
 - Stopping progress service
>>>> 2015-08-14 15:39:11,540 [main] INFO  org.apache.sqoop.job.mr.SqoopOutputFormatLoadExecutor
 - SqoopOutputFormatLoadExecutor::SqoopRecordWriter is about to be closed*
>>>>
>>>>
>>>> p.s. sqoop version: 1.99.6
>>>> hadoop version 2.6.0
>>>>
>>>
>>>
>>
>

Mime
View raw message