mahout-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Allan Roberto Avendano Sudario <aaven...@fiec.espol.edu.ec>
Subject Re: Creating Vectors from Text
Date Thu, 02 Jul 2009 16:09:38 GMT
Regards,
This is the entire exception message:


java -cp $JAVACLASSPATH org.apache.mahout.utils.vectors.Driver --dir
/home/hadoop/Desktop/<urls>/index  --field content  --dictOut
/home/hadoop/Desktop/dictionary/dict.txt --output
/home/hadoop/Desktop/dictionary/out.txt --max 50 --norm 2


09/07/02 09:35:47 INFO vectors.Driver: Output File:
/home/hadoop/Desktop/dictionary/out.txt
09/07/02 09:35:47 INFO util.NativeCodeLoader: Loaded the native-hadoop
library
09/07/02 09:35:47 INFO zlib.ZlibFactory: Successfully loaded & initialized
native-zlib library
09/07/02 09:35:47 INFO compress.CodecPool: Got brand-new compressor
Exception in thread "main" java.lang.NullPointerException
        at
org.apache.mahout.utils.vectors.lucene.LuceneIteratable$TDIterator.next(LuceneIteratable.java:111)
        at
org.apache.mahout.utils.vectors.lucene.LuceneIteratable$TDIterator.next(LuceneIteratable.java:82)
        at
org.apache.mahout.utils.vectors.io.SequenceFileVectorWriter.write(SequenceFileVectorWriter.java:25)
        at org.apache.mahout.utils.vectors.Driver.main(Driver.java:204)


Well, I used a nutch crawl index, is that correct? mmm... I have change to
contenc field, but nothing happened.
Possibly the nutch crawl doesn´t have Term Vector indexed.

Thanks,


2009/7/1 Grant Ingersoll <gsingers@apache.org>

> Is there any more information around the exception?
>
> How did you create your Lucene index?  Does the body field exist and does
> it have Term Vectors stored?
>
>
>
> On Jul 1, 2009, at 7:20 PM, Allan Roberto Avendano Sudario wrote:
>
>  Regards Community,
>> Someone know how to execute the code that create vector from text? This
>> message show me java, when I try to run this:
>>
>> java -cp $CLASSPATH org.apache.mahout.utils.vectors.Driver --dir
>> ~/Desktop/crawlSite/index  --field body --dictOut ~/Desktop/dict/dict.txt
>> --output ~/Desktop/dict/out.txt --max 50
>>
>> ....
>>
>> WARN util.NativeCodeLoader: Unable to load native-hadoop library for your
>> platform... using builtin-java classes where applicable
>> 09/07/01 18:02:32 INFO compress.CodecPool: Got brand-new compressor
>> Exception in thread "main" java.lang.NullPointerException
>>       at
>>
>> org.apache.mahout.utils.vectors.lucene.LuceneIteratable$TDIterator.next(LuceneIteratable.java:111)
>>       at
>>
>> org.apache.mahout.utils.vectors.lucene.LuceneIteratable$TDIterator.next(LuceneIteratable.java:82)
>>       at
>>
>> org.apache.mahout.utils.vectors.io.SequenceFileVectorWriter.write(SequenceFileVectorWriter.java:25)
>>       at org.apache.mahout.utils.vectors.Driver.main(Driver.java:204)
>>
>>
>> Thanks!
>>
>> --
>> Allan Avendaño S.
>> Home: 04 2 800 692
>> Cell: 09 700 42 48
>>
>
> --------------------------
> Grant Ingersoll
> http://www.lucidimagination.com/
>
> Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using
> Solr/Lucene:
> http://www.lucidimagination.com/search
>
>


-- 
Allan Avendaño S.
Home: 04 2 800 692
Cell: 09 700 42 48

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message