spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Walrus theCat <walrusthe...@gmail.com>
Subject Re: Could not find resource path for Web UI: org/apache/spark/ui/static
Date Thu, 05 Dec 2013 19:00:23 GMT
Update on this... it works when I run the main class with the ./run-example
script, but not in any form of scala myjar.jar .


On Mon, Dec 2, 2013 at 7:14 PM, Walrus theCat <walrusthecat@gmail.com>wrote:

> Anyone have any ideas based on the stack trace?
>
> Thanks
>
>
> On Sun, Dec 1, 2013 at 9:09 PM, Walrus theCat <walrusthecat@gmail.com>wrote:
>
>> Shouldn't?  I imported the new 0.8.0 jars into my build path, and had to
>> update my imports accordingly.  The only way I upload the spark jars myself
>> is that they get packaged into my executable jar.  The cluster should have
>> the right version based on the flag used to launch it (and it does.)
>>
>>
>> On Fri, Nov 29, 2013 at 10:12 PM, Ashish Rangole <arangole@gmail.com>wrote:
>>
>>> I am sure you have already checked this, any chance the classpath has
>>> v 0.7.x jars in it?
>>> On Nov 29, 2013 4:40 PM, "Walrus theCat" <walrusthecat@gmail.com> wrote:
>>>
>>>> The "full context" isn't much -- this is the first thing I do in my
>>>> main method (assign a value to sc), and it throws this error.
>>>>
>>>>
>>>> On Fri, Nov 29, 2013 at 10:38 AM, Walrus theCat <walrusthecat@gmail.com
>>>> > wrote:
>>>>
>>>>> Hi Matei,
>>>>>
>>>>> Good to hear from you.  The stack trace is below.  I launched the
>>>>> instances with --spark-version=0.8.0 and verified that the version was
>>>>> correct by launching spark-shell.  Also verified that the version I've
got
>>>>> in my project is 0.8.0.  Nothing else should have changed, as the scripts
I
>>>>> use to set up the classpath and everything is the exact same as I used
in
>>>>> 0.7.3.
>>>>>
>>>>> Cheers,
>>>>>
>>>>>
>>>>> java.lang.Exception: Could not find resource path for Web UI:
>>>>> org/apache/spark/ui/static
>>>>>     at
>>>>> org.apache.spark.ui.JettyUtils$.createStaticHandler(JettyUtils.scala:89)
>>>>>     at org.apache.spark.ui.SparkUI.<init>(SparkUI.scala:40)
>>>>>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:122)
>>>>>     at walrusthecat.ml.spark.SparkSVM$.main(SparkSVM.scala:16)
>>>>>     at walrusthecat.ml.spark.SparkSVM.main(SparkSVM.scala)
>>>>>     at walrusthecat.ml.spark.Main.main(Main.java:7)
>>>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>     at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>     at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>     at
>>>>> org.eclipse.jdt.internal.jarinjarloader.JarRsrcLoader.main(JarRsrcLoader.java:58)
>>>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>     at
>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>     at
>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>     at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>     at
>>>>> scala.tools.nsc.util.ScalaClassLoader$$anonfun$run$1.apply(ScalaClassLoader.scala:78)
>>>>>     at
>>>>> scala.tools.nsc.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:24)
>>>>>     at
>>>>> scala.tools.nsc.util.ScalaClassLoader$URLClassLoader.asContext(ScalaClassLoader.scala:88)
>>>>>     at
>>>>> scala.tools.nsc.util.ScalaClassLoader$class.run(ScalaClassLoader.scala:78)
>>>>>     at
>>>>> scala.tools.nsc.util.ScalaClassLoader$URLClassLoader.run(ScalaClassLoader.scala:101)
>>>>>     at scala.tools.nsc.ObjectRunner$.run(ObjectRunner.scala:33)
>>>>>     at scala.tools.nsc.ObjectRunner$.runAndCatch(ObjectRunner.scala:40)
>>>>>     at
>>>>> scala.tools.nsc.MainGenericRunner.runTarget$1(MainGenericRunner.scala:60)
>>>>>     at
>>>>> scala.tools.nsc.MainGenericRunner.process(MainGenericRunner.scala:80)
>>>>>     at
>>>>> scala.tools.nsc.MainGenericRunner$.main(MainGenericRunner.scala:89)
>>>>>     at scala.tools.nsc.MainGenericRunner.main(MainGenericRunner.scala)
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Nov 27, 2013 at 6:15 PM, Matei Zaharia <
>>>>> matei.zaharia@gmail.com> wrote:
>>>>>
>>>>>> Sorry, what’s the full context for this? Do you have a stack trace?
>>>>>> My guess is that Spark isn’t on your classpath, or maybe you only
have an
>>>>>> old version of it on there.
>>>>>>
>>>>>> Matei
>>>>>>
>>>>>> On Nov 27, 2013, at 6:04 PM, Walrus theCat <walrusthecat@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> To clarify, I just undid that "var... field.." thing described above,
>>>>>> and it throws the same error.
>>>>>>
>>>>>>
>>>>>> On Wed, Nov 27, 2013 at 5:53 PM, Walrus theCat <
>>>>>> walrusthecat@gmail.com> wrote:
>>>>>>
>>>>>>> Hi all,
>>>>>>>
>>>>>>> This exception gets thrown when I assign a value to the variable
>>>>>>> holding my SparkContext.  I initialize it as a var holding a
null value (so
>>>>>>> it can be a field), and then give it a value in my main method.
 This
>>>>>>> worked with the previous version of Spark, but is not working
on Spark
>>>>>>> 0.8.0.
>>>>>>>
>>>>>>> Dankeschöen,
>>>>>>>
>>>>>>> Walrus theCat
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>
>

Mime
View raw message