spark-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Stephen Boesch <java...@gmail.com>
Subject Re: sbt scala compiler crashes on spark-sql
Date Mon, 03 Nov 2014 03:48:47 GMT
Yes I have seen this same error - and for team members as well - repeatedly
since June. A Patrick and Cheng mentioned, the next step is to do an sbt
clean

2014-11-02 19:37 GMT-08:00 Cheng Lian <lian.cs.zju@gmail.com>:

> I often see this when I first build the whole Spark project with SBT, then
> modify some code and tries to build and debug within IDEA, or vice versa. A
> clean rebuild can always solve this.
>
> On Mon, Nov 3, 2014 at 11:28 AM, Patrick Wendell <pwendell@gmail.com>
> wrote:
>
> > Does this happen if you clean and recompile? I've seen failures on and
> > off, but haven't been able to find one that I could reproduce from a
> > clean build such that we could hand it to the scala team.
> >
> > - Patrick
> >
> > On Sun, Nov 2, 2014 at 7:25 PM, Imran Rashid <imran@therashids.com>
> wrote:
> > > I'm finding the scala compiler crashes when I compile the spark-sql
> > project
> > > in sbt.  This happens in both the 1.1 branch and master (full error
> > > below).  The other projects build fine in sbt, and everything builds
> fine
> > > in maven.  is there some sbt option I'm forgetting?  Any one else
> > > experiencing this?
> > >
> > > Also, are there up-to-date instructions on how to do common dev tasks
> in
> > > both sbt & maven?  I have only found these instructions on building
> with
> > > maven:
> > >
> > > http://spark.apache.org/docs/latest/building-with-maven.html
> > >
> > > and some general info here:
> > >
> > >
> https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark
> > >
> > > but I think this doesn't walk through a lot of the steps of a typical
> dev
> > > cycle, eg, continuous compilation, running one test, running one main
> > > class, etc.  (especially since it seems like people still favor sbt for
> > > dev.)  If it doesn't already exist somewhere, I could try to put
> > together a
> > > brief doc for how to do the basics.  (I'm returning to spark dev after
> a
> > > little hiatus myself, and I'm hitting some stumbling blocks that are
> > > probably common knowledge to everyone still dealing with it all the
> > time.)
> > >
> > > thanks,
> > > Imran
> > >
> > > ------------------------------
> > > full crash info from sbt:
> > >
> > >> project sql
> > > [info] Set current project to spark-sql (in build
> > > file:/Users/imran/spark/spark/)
> > >> compile
> > > [info] Compiling 62 Scala sources to
> > > /Users/imran/spark/spark/sql/catalyst/target/scala-2.10/classes...
> > > [info] Compiling 45 Scala sources and 39 Java sources to
> > > /Users/imran/spark/spark/sql/core/target/scala-2.10/classes...
> > > [error]
> > > [error]      while compiling:
> > >
> >
> /Users/imran/spark/spark/sql/core/src/main/scala/org/apache/spark/sql/types/util/DataTypeConversions.scala
> > > [error]         during phase: jvm
> > > [error]      library version: version 2.10.4
> > > [error]     compiler version: version 2.10.4
> > > [error]   reconstructed args: -classpath
> > >
> >
> /Users/imran/spark/spark/sql/core/target/scala-2.10/classes:/Users/imran/spark/spark/core/target/scala-2.10/classes:/Users/imran/spark/spark/sql/catalyst/target/scala-2.10/classes:/Users/imran/spark/spark/lib_managed/jars/hadoop-client-1.0.4.jar:/Users/imran/spark/spark/lib_managed/jars/hadoop-core-1.0.4.jar:/Users/imran/spark/spark/lib_managed/jars/xmlenc-0.52.jar:/Users/imran/spark/spark/lib_managed/jars/commons-math-2.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-configuration-1.6.jar:/Users/imran/spark/spark/lib_managed/jars/commons-collections-3.2.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-lang-2.4.jar:/Users/imran/spark/spark/lib_managed/jars/commons-logging-1.1.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-digester-1.8.jar:/Users/imran/spark/spark/lib_managed/jars/commons-beanutils-1.7.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-beanutils-core-1.8.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-net-2.2.jar:/Users/imran/spark/spark/lib_managed/jars/commons-el-1.0.jar:/Users/imran/spark/spark/lib_managed/jars/hsqldb-1.8.0.10.jar:/Users/imran/spark/spark/lib_managed/jars/oro-2.0.8.jar:/Users/imran/spark/spark/lib_managed/jars/jets3t-0.7.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-httpclient-3.1.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-recipes-2.4.0.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-framework-2.4.0.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-client-2.4.0.jar:/Users/imran/spark/spark/lib_managed/jars/zookeeper-3.4.5.jar:/Users/imran/spark/spark/lib_managed/jars/slf4j-log4j12-1.7.5.jar:/Users/imran/spark/spark/lib_managed/bundles/log4j-1.2.17.jar:/Users/imran/spark/spark/lib_managed/jars/jline-0.9.94.jar:/Users/imran/spark/spark/lib_managed/bundles/guava-14.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-plus-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.transaction-1.1.1.v201105210645.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-webapp-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-xml-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-util-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-servlet-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-security-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-server-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.servlet-3.0.0.v201112011016.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-continuation-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-http-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-io-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-jndi-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.mail.glassfish-1.4.1.v201005082020.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.activation-1.1.0.v201105071233.jar:/Users/imran/spark/spark/lib_managed/jars/commons-lang3-3.3.2.jar:/Users/imran/spark/spark/lib_managed/jars/jsr305-1.3.9.jar:/Users/imran/spark/spark/lib_managed/jars/slf4j-api-1.7.5.jar:/Users/imran/spark/spark/lib_managed/jars/jul-to-slf4j-1.7.5.jar:/Users/imran/spark/spark/lib_managed/jars/jcl-over-slf4j-1.7.5.jar:/Users/imran/spark/spark/lib_managed/bundles/compress-lzf-1.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/snappy-java-1.0.5.3.jar:/Users/imran/spark/spark/lib_managed/jars/lz4-1.2.0.jar:/Users/imran/spark/spark/lib_managed/jars/chill_2.10-0.3.6.jar:/Users/imran/spark/spark/lib_managed/jars/chill-java-0.3.6.jar:/Users/imran/spark/spark/lib_managed/bundles/kryo-2.21.jar:/Users/imran/spark/spark/lib_managed/jars/reflectasm-1.07-shaded.jar:/Users/imran/spark/spark/lib_managed/jars/minlog-1.2.jar:/Users/imran/spark/spark/lib_managed/jars/objenesis-1.2.jar:/Users/imran/spark/spark/lib_managed/bundles/akka-remote_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/akka-actor_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/bundles/config-1.0.2.jar:/Users/imran/spark/spark/lib_managed/bundles/netty-3.6.6.Final.jar:/Users/imran/spark/spark/lib_managed/jars/protobuf-java-2.4.1-shaded.jar:/Users/imran/spark/spark/lib_managed/jars/uncommons-maths-1.2.2a.jar:/Users/imran/spark/spark/lib_managed/bundles/akka-slf4j_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-jackson_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-core_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-ast_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/paranamer-2.6.jar:/Users/imran/spark/spark/lib_managed/jars/scalap-2.10.0.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-databind-2.3.1.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-annotations-2.3.0.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-core-2.3.1.jar:/Users/imran/spark/spark/lib_managed/jars/colt-1.2.0.jar:/Users/imran/spark/spark/lib_managed/jars/concurrent-1.3.4.jar:/Users/imran/spark/spark/lib_managed/jars/mesos-0.18.1-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/netty-all-4.0.23.Final.jar:/Users/imran/spark/spark/lib_managed/jars/stream-2.7.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-core-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-jvm-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-json-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-graphite-3.0.0.jar:/Users/imran/spark/spark/lib_managed/jars/tachyon-client-0.5.0.jar:/Users/imran/spark/spark/lib_managed/jars/tachyon-0.5.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-io-2.4.jar:/Users/imran/spark/spark/lib_managed/jars/pyrolite-2.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/py4j-0.8.2.1.jar:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-compiler.jar:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-reflect.jar:/Users/imran/spark/spark/lib_managed/jars/quasiquotes_2.10-2.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-column-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-common-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-encoding-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-generator-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/commons-codec-1.5.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-hadoop-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-format-2.0.0.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-jackson-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/jackson-mapper-asl-1.9.11
> > >
> >
> .jar:/Users/imran/spark/spark/lib_managed/jars/jackson-core-asl-1.9.11.jar
> > > -deprecation -feature
> > > -P:genjavadoc:out=/Users/imran/spark/spark/sql/core/target/java
> > >
> >
> -Xplugin:/Users/imran/spark/spark/lib_managed/jars/genjavadoc-plugin_2.10.4-0.7.jar
> > > -bootclasspath
> > >
> >
> /Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/sunrsasign.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/classes:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-library.jar
> > > -unchecked -language:postfixOps
> > > [error]
> > > [error]   last tree to typer:
> > > Literal(Constant(org.apache.spark.sql.catalyst.types.PrimitiveType))
> > > [error]               symbol: null
> > > [error]    symbol definition: null
> > > [error]                  tpe:
> > > Class(classOf[org.apache.spark.sql.catalyst.types.PrimitiveType])
> > > [error]        symbol owners:
> > > [error]       context owners: anonymous class anonfun$asScalaDataType$1
> > ->
> > > package util
> > > [error]
> > > [error] == Enclosing template or block ==
> > > [error]
> > > [error] Template( // val <local $anonfun>: <notype>,
> > > tree.tpe=org.apache.spark.sql.types.util.anonfun$asScalaDataType$1
> > > [error]   "scala.runtime.AbstractFunction1", "scala.Serializable" //
> > parents
> > > [error]   ValDef(
> > > [error]     private
> > > [error]     "_"
> > > [error]     <tpt>
> > > [error]     <empty>
> > > [error]   )
> > > [error]   // 3 statements
> > > [error]   DefDef( // final def apply(javaStructField:
> > > org.apache.spark.sql.api.java.StructField):
> > > org.apache.spark.sql.catalyst.types.StructField
> > > [error]     <method> final <triedcooking>
> > > [error]     "apply"
> > > [error]     []
> > > [error]     // 1 parameter list
> > > [error]     ValDef( // javaStructField:
> > > org.apache.spark.sql.api.java.StructField
> > > [error]       <param> <synthetic> <triedcooking>
> > > [error]       "javaStructField"
> > > [error]       <tpt> //
> tree.tpe=org.apache.spark.sql.api.java.StructField
> > > [error]       <empty>
> > > [error]     )
> > > [error]     <tpt> //
> > > tree.tpe=org.apache.spark.sql.catalyst.types.StructField
> > > [error]     Apply( // def asScalaStructField(javaStructField:
> > > org.apache.spark.sql.api.java.StructField):
> > > org.apache.spark.sql.catalyst.types.StructField in object
> > > DataTypeConversions,
> > > tree.tpe=org.apache.spark.sql.catalyst.types.StructField
> > > [error]       DataTypeConversions.this."asScalaStructField" // def
> > > asScalaStructField(javaStructField:
> > > org.apache.spark.sql.api.java.StructField):
> > > org.apache.spark.sql.catalyst.types.StructField in object
> > > DataTypeConversions, tree.tpe=(javaStructField:
> > >
> >
> org.apache.spark.sql.api.java.StructField)org.apache.spark.sql.catalyst.types.StructField
> > > [error]       "javaStructField" // javaStructField:
> > > org.apache.spark.sql.api.java.StructField,
> > > tree.tpe=org.apache.spark.sql.api.java.StructField
> > > [error]     )
> > > [error]   )
> > > [error]   DefDef( // final def apply(v1: Object): Object
> > > [error]     <method> final <bridge>
> > > [error]     "apply"
> > > [error]     []
> > > [error]     // 1 parameter list
> > > [error]     ValDef( // v1: Object
> > > [error]       <param> <triedcooking>
> > > [error]       "v1"
> > > [error]       <tpt> // tree.tpe=Object
> > > [error]       <empty>
> > > [error]     )
> > > [error]     <tpt> // tree.tpe=Object
> > > [error]     Apply( // final def apply(javaStructField:
> > > org.apache.spark.sql.api.java.StructField):
> > > org.apache.spark.sql.catalyst.types.StructField,
> > > tree.tpe=org.apache.spark.sql.catalyst.types.StructField
> > > [error]
>  DataTypeConversions$$anonfun$asScalaDataType$1.this."apply"
> > > // final def apply(javaStructField:
> > > org.apache.spark.sql.api.java.StructField):
> > > org.apache.spark.sql.catalyst.types.StructField,
> > tree.tpe=(javaStructField:
> > >
> >
> org.apache.spark.sql.api.java.StructField)org.apache.spark.sql.catalyst.types.StructField
> > > [error]       Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in
> > > class Object, tree.tpe=org.apache.spark.sql.api.java.StructField
> > > [error]         TypeApply( // final def $asInstanceOf[T0 >: ? <: ?]():
> T0
> > > in class Object, tree.tpe=()org.apache.spark.sql.api.java.StructField
> > > [error]           "v1"."$asInstanceOf" // final def $asInstanceOf[T0
> >: ?
> > > <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0
> > > [error]           <tpt> //
> > > tree.tpe=org.apache.spark.sql.api.java.StructField
> > > [error]         )
> > > [error]         Nil
> > > [error]       )
> > > [error]     )
> > > [error]   )
> > > [error]   DefDef( // def <init>():
> > > org.apache.spark.sql.types.util.anonfun$asScalaDataType$1
> > > [error]     <method> <triedcooking>
> > > [error]     "<init>"
> > > [error]     []
> > > [error]     List(Nil)
> > > [error]     <tpt> //
> > > tree.tpe=org.apache.spark.sql.types.util.anonfun$asScalaDataType$1
> > > [error]     Block( // tree.tpe=Unit
> > > [error]       Apply( // def <init>(): scala.runtime.AbstractFunction1
> in
> > > class AbstractFunction1, tree.tpe=scala.runtime.AbstractFunction1
> > > [error]
> > > DataTypeConversions$$anonfun$asScalaDataType$1.super."<init>" // def
> > > <init>(): scala.runtime.AbstractFunction1 in class AbstractFunction1,
> > > tree.tpe=()scala.runtime.AbstractFunction1
> > > [error]         Nil
> > > [error]       )
> > > [error]       ()
> > > [error]     )
> > > [error]   )
> > > [error] )
> > > [error]
> > > [error] == Expanded type of tree ==
> > > [error]
> > > [error] ConstantType(
> > > [error]   value =
> > > Constant(org.apache.spark.sql.catalyst.types.PrimitiveType)
> > > [error] )
> > > [error]
> > > [error] uncaught exception during compilation: java.lang.AssertionError
> > > [trace] Stack trace suppressed: run last sql/compile:compile for the
> full
> > > output.
> > > [error] (sql/compile:compile) java.lang.AssertionError: assertion
> failed:
> > > List(object package$DebugNode, object package$DebugNode)
> > > [error] Total time: 23 s, completed Nov 2, 2014 1:00:37 PM
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
> > For additional commands, e-mail: dev-help@spark.apache.org
> >
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message