spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jacek Laskowski <ja...@japila.pl>
Subject Re: Unable to run a Standalone job
Date Fri, 23 May 2014 21:38:22 GMT
Hi Shrikar,

How did you build Spark 1.0.0-SNAPSHOT on your machine? My
understanding is that `sbt publishLocal` is not enough and you really
need `sbt assembly` instead. Give it a try and report back.

As to your build.sbt, upgrade Scala to 2.10.4 and "org.apache.spark"
%% "spark-streaming" % "1.0.0-SNAPSHOT" only that will pull down
spark-core as a transitive dep. The resolver for Akka Repository is
not needed. Your build.sbt should really look as follows:

name := "Simple Project"

version := "1.0"

scalaVersion := "2.10.4"

libraryDependencies += "org.apache.spark" %% "spark-streaming" %
"1.0.0-SNAPSHOT"

Jacek

On Thu, May 22, 2014 at 11:27 PM, Shrikar archak <shrikar84@gmail.com> wrote:
> Hi All,
>
> I am trying to run the network count example as a seperate standalone job
> and running into some issues.
>
> Environment:
> 1) Mac Mavericks
> 2) Latest spark repo from Github.
>
>
> I have a structure like this
>
> Shrikars-MacBook-Pro:SimpleJob shrikar$ find .
> .
> ./simple.sbt
> ./src
> ./src/main
> ./src/main/scala
> ./src/main/scala/NetworkWordCount.scala
> ./src/main/scala/SimpleApp.scala.bk
>
>
> simple.sbt
> name := "Simple Project"
>
> version := "1.0"
>
> scalaVersion := "2.10.3"
>
> libraryDependencies ++= Seq("org.apache.spark" %% "spark-core" %
> "1.0.0-SNAPSHOT",
>                             "org.apache.spark" %% "spark-streaming" %
> "1.0.0-SNAPSHOT")
>
> resolvers += "Akka Repository" at "http://repo.akka.io/releases/"
>
>
> I am able to run the SimpleApp which is mentioned in the doc but when I try
> to run the NetworkWordCount app I get error like this am I missing
> something?
>
> [info] Running com.shrikar.sparkapps.NetworkWordCount
> 14/05/22 14:26:47 INFO spark.SecurityManager: Changing view acls to: shrikar
> 14/05/22 14:26:47 INFO spark.SecurityManager: SecurityManager:
> authentication disabled; ui acls disabled; users with view permissions:
> Set(shrikar)
> 14/05/22 14:26:48 INFO slf4j.Slf4jLogger: Slf4jLogger started
> 14/05/22 14:26:48 INFO Remoting: Starting remoting
> 14/05/22 14:26:48 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://spark@192.168.10.88:49963]
> 14/05/22 14:26:48 INFO Remoting: Remoting now listens on addresses:
> [akka.tcp://spark@192.168.10.88:49963]
> 14/05/22 14:26:48 INFO spark.SparkEnv: Registering MapOutputTracker
> 14/05/22 14:26:48 INFO spark.SparkEnv: Registering BlockManagerMaster
> 14/05/22 14:26:48 INFO storage.DiskBlockManager: Created local directory at
> /var/folders/r2/mbj08pb55n5d_9p8588xk5b00000gn/T/spark-local-20140522142648-0a14
> 14/05/22 14:26:48 INFO storage.MemoryStore: MemoryStore started with
> capacity 911.6 MB.
> 14/05/22 14:26:48 INFO network.ConnectionManager: Bound socket to port 49964
> with id = ConnectionManagerId(192.168.10.88,49964)
> 14/05/22 14:26:48 INFO storage.BlockManagerMaster: Trying to register
> BlockManager
> 14/05/22 14:26:48 INFO storage.BlockManagerInfo: Registering block manager
> 192.168.10.88:49964 with 911.6 MB RAM
> 14/05/22 14:26:48 INFO storage.BlockManagerMaster: Registered BlockManager
> 14/05/22 14:26:48 INFO spark.HttpServer: Starting HTTP Server
> [error] (run-main) java.lang.NoClassDefFoundError:
> javax/servlet/http/HttpServletResponse
> java.lang.NoClassDefFoundError: javax/servlet/http/HttpServletResponse
> at org.apache.spark.HttpServer.start(HttpServer.scala:54)
> at
> org.apache.spark.broadcast.HttpBroadcast$.createServer(HttpBroadcast.scala:156)
> at
> org.apache.spark.broadcast.HttpBroadcast$.initialize(HttpBroadcast.scala:127)
> at
> org.apache.spark.broadcast.HttpBroadcastFactory.initialize(HttpBroadcastFactory.scala:31)
> at
> org.apache.spark.broadcast.BroadcastManager.initialize(BroadcastManager.scala:48)
> at
> org.apache.spark.broadcast.BroadcastManager.<init>(BroadcastManager.scala:35)
> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:218)
> at org.apache.spark.SparkContext.<init>(SparkContext.scala:202)
> at
> org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:549)
> at
> org.apache.spark.streaming.StreamingContext$.createNewSparkContext(StreamingContext.scala:561)
> at
> org.apache.spark.streaming.StreamingContext.<init>(StreamingContext.scala:91)
> at com.shrikar.sparkapps.NetworkWordCount$.main(NetworkWordCount.scala:39)
> at com.shrikar.sparkapps.NetworkWordCount.main(NetworkWordCount.scala)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
>
>
> Thanks,
> Shrikar
>



-- 
Jacek Laskowski | http://blog.japila.pl
"Never discourage anyone who continually makes progress, no matter how
slow." Plato

Mime
View raw message