spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Marco Mistroni <mmistr...@gmail.com>
Subject Re: sbt shenanigans for a Spark-based project
Date Tue, 15 Nov 2016 22:45:05 GMT
Uhm i removed mvn repo and ivy folder as well, sbt seems to kick in but for
some reason it cannot 'see' org.apacke.spark-mllib and therefore my
compilation fails
i have temporarily fixed it by placing the spark-mllib jar in my project
\lib directory,
perhaps i'll try to create a brand new Spark project and see if that makes
any differences

thanks for assistance Don!
kr
 marco

On Mon, Nov 14, 2016 at 11:13 PM, Don Drake <dondrake@gmail.com> wrote:

> I would remove your entire local Maven repo (~/.m2/repo in linux) and try
> again. I'm able to compile sample code with your build.sbt and sbt
> v.0.13.12.
>
> -Don
>
> On Mon, Nov 14, 2016 at 3:11 PM, Marco Mistroni <mmistroni@gmail.com>
> wrote:
>
>> uhm.sorry...... still same issues. this is hte new version
>>
>> name := "SparkExamples"
>> version := "1.0"
>> scalaVersion := "2.11.8"
>> val sparkVersion = "2.0.1"
>>
>> // Add a single dependency
>> libraryDependencies += "junit" % "junit" % "4.8" % "test"
>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5",
>>                             "org.slf4j" % "slf4j-simple" % "1.7.5",
>>                             "org.clapper" %% "grizzled-slf4j" % "1.0.2")
>> libraryDependencies += "org.apache.spark"%%"spark-core"   % sparkVersion
>> libraryDependencies += "org.apache.spark"%%"spark-streaming"   %
>> sparkVersion
>> libraryDependencies += "org.apache.spark"%%"spark-mllib"   %
>> sparkVersion
>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume-sink"
>> % "2.0.1"
>> libraryDependencies += "org.apache.spark"%%"spark-sql"   % sparkVersion
>>
>>
>> resolvers += "softprops-maven" at "http://dl.bintray.com/content
>> /softprops/maven"
>>
>> Still seeing these kinds of errors  which seems to lead to the fact that
>> somehow sbt is getting confused..
>>
>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:2:
>> object mllib is not a member of package org.apache.spark
>> [error] import org.apache.spark.mllib.linalg.{ Vector, Vectors }
>> [error]                         ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:3:
>> object mllib is not a member of package org.apache.spark
>> [error] import org.apache.spark.mllib.regression.LabeledPoint
>> [error]                         ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:4:
>> object classification is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.classification.{
>> RandomForestClassifier, RandomForestClassificationModel }
>> [error]                            ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:6:
>> object feature is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.feature.{ StringIndexer,
>> IndexToString, VectorIndexer, VectorAssembler }
>> [error]                            ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:7:
>> object evaluation is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.evaluation.{ RegressionEvaluator,
>> MulticlassClassificationEvaluator }
>> [error]                            ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:8:
>> object classification is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.classification._
>> [error]                            ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:9:
>> object tuning is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.tuning.{ CrossValidator,
>> ParamGridBuilder }
>> [error]                            ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:10:
>> object tuning is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.tuning.{ ParamGridBuilder,
>> TrainValidationSplit }
>> [error]                            ^
>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:16:
>> object Pipeline is not a member of package org.apache.spark.ml
>> [error] import org.apache.spark.ml.{ Pipeline, PipelineModel }
>>
>> any other hints?
>>
>> thanks and regarsd
>>  marco
>>
>>
>>
>>
>> On Sun, Nov 13, 2016 at 10:52 PM, Don Drake <dondrake@gmail.com> wrote:
>>
>>> I would upgrade your Scala version to 2.11.8 as Spark 2.0 uses Scala
>>> 2.11 by default.
>>>
>>> On Sun, Nov 13, 2016 at 3:01 PM, Marco Mistroni <mmistroni@gmail.com>
>>> wrote:
>>>
>>>> HI all
>>>>  i have a small Spark-based project which at the moment depends on jar
>>>> from Spark 1.6.0
>>>> The project has few Spark examples plus one which depends on Flume
>>>> libraries
>>>>
>>>>
>>>> I am attempting to move to Spark 2.0, but i am having issues with
>>>> my dependencies
>>>> The stetup below works fine when compiled against 1.6.0 dependencies
>>>>
>>>> name := "SparkExamples"
>>>> version := "1.0"
>>>> scalaVersion := "2.10.5"
>>>> val sparkVersion = "1.6.0"
>>>>
>>>>
>>>> // Add a single dependency
>>>> libraryDependencies += "junit" % "junit" % "4.8" % "test"
>>>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5",
>>>>                             "org.slf4j" % "slf4j-simple" % "1.7.5",
>>>>                             "org.clapper" %% "grizzled-slf4j" % "1.0.2")
>>>> libraryDependencies += "org.apache.spark"%%"spark-core"   %
>>>> sparkVersion
>>>> libraryDependencies += "org.apache.spark"%%"spark-streaming"   %
>>>> sparkVersion
>>>> libraryDependencies += "org.apache.spark"%%"spark-mllib"   %
>>>> sparkVersion
>>>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume" %
>>>> "1.3.0"
>>>> libraryDependencies += "org.apache.spark"%%"spark-sql"   %
>>>> sparkVersion
>>>>
>>>>
>>>> resolvers += "softprops-maven" at "http://dl.bintray.com/content
>>>> /softprops/maven"
>>>>
>>>>
>>>>
>>>> This is the build.sbt version for using Spark 2 dependencies
>>>>
>>>> name := "SparkExamples"
>>>> version := "1.0"
>>>> scalaVersion := "2.10.6"
>>>> val sparkVersion = "2.0.1"
>>>>
>>>>
>>>> // Add a single dependency
>>>> libraryDependencies += "junit" % "junit" % "4.8" % "test"
>>>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5",
>>>>                             "org.slf4j" % "slf4j-simple" % "1.7.5",
>>>>                             "org.clapper" %% "grizzled-slf4j" % "1.0.2")
>>>> libraryDependencies += "org.apache.spark"%%"spark-core"   %
>>>> sparkVersion
>>>> libraryDependencies += "org.apache.spark"%%"spark-streaming"   %
>>>> sparkVersion
>>>> libraryDependencies += "org.apache.spark"%%"spark-mllib"   %
>>>> sparkVersion
>>>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume-sink"
>>>> % "2.0.1"
>>>> libraryDependencies += "org.apache.spark"%%"spark-sql"   %
>>>> sparkVersion
>>>> resolvers += "softprops-maven" at "http://dl.bintray.com/content
>>>> /softprops/maven"
>>>>
>>>> but the sbt compile fails miserably...below few of the errors (it
>>>> actually compiles like i forgot all the depencencies as it is complaining
>>>> on all org.apache.spark.ml and mllib packages
>>>>
>>>> [warn] Multiple dependencies with the same organization/name but
>>>> different versions. To avoid conflict, pick one version:
>>>> [warn]  * org.apache.spark:spark-core_2.10:(1.6.1, 2.0.1)
>>>> [warn]  * org.apache.spark:spark-streaming_2.10:(1.6.1, 2.0.1)
>>>> [warn]  * org.apache.spark:spark-sql_2.10:(1.6.1, 2.0.1)
>>>> [warn]  * org.apache.spark:spark-mllib_2.10:(1.6.1, 2.0.1)
>>>> [info] Resolving org.scala-lang#scala-library;2.10.6 ...
>>>> .........
>>>> [warn]     * org.apache.spark:spark-mllib_2.10:1.6.1 -> 2.0.1
>>>> [warn]     * org.apache.spark:spark-sql_2.10:1.6.1 -> 2.0.1
>>>> [warn]     * org.apache.spark:spark-streaming_2.10:1.6.1 -> 2.0.1
>>>> [warn]     * org.apache.spark:spark-core_2.10:1.6.1 -> 2.0.1
>>>> [warn] Run 'evicted' to see detailed eviction warnings
>>>> [info] Compiling 18 Scala sources to C:\Users\marco\SparkExamples\t
>>>> arget\scala-2.10\classes...
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:2:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.linalg.{ Vector, Vectors }
>>>> [error]                         ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:3:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.regression.LabeledPoint
>>>> [error]                         ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:4:
>>>> object classification is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.classification._
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:5:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.tree.DecisionTree
>>>> [error]                         ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:6:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.tree.model.DecisionTreeModel
>>>> [error]                         ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:7:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.util.MLUtils
>>>> [error]                         ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:9:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.stat.{
>>>> MultivariateStatisticalSummary, Statistics }
>>>> [error]                         ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:10:
>>>> object mllib is not a member of package org.apache.spark
>>>> [error] import org.apache.spark.mllib.linalg.Vector
>>>> ........
>>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:4:
>>>> object classification is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.classification.{
>>>> RandomForestClassifier, RandomForestClassificationModel }
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:6:
>>>> object feature is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.feature.{ StringIndexer,
>>>> IndexToString, VectorIndexer, VectorAssembler }
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:7:
>>>> object evaluation is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.evaluation.{ RegressionEvaluator,
>>>> MulticlassClassificationEvaluator }
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:8:
>>>> object classification is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.classification._
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:9:
>>>> object tuning is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.tuning.{ CrossValidator,
>>>> ParamGridBuilder }
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:10:
>>>> object tuning is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.tuning.{ ParamGridBuilder,
>>>> TrainValidationSplit }
>>>> [error]                            ^
>>>> [error] C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:16:
>>>> object Pipeline is not a member of package org.apache.spark.ml
>>>> [error] import org.apache.spark.ml.{ Pipeline, PipelineModel }
>>>> [error]
>>>> ....
>>>> and many more
>>>>
>>>> Could anyone assist or pass me the right dependencies? Or am i using
>>>> wrong resolvers?
>>>>
>>>> kr
>>>>  marco
>>>>
>>>
>>>
>>>
>>> --
>>> Donald Drake
>>> Drake Consulting
>>> http://www.drakeconsulting.com/
>>> https://twitter.com/dondrake <http://www.MailLaunder.com/>
>>> 800-733-2143
>>>
>>
>>
>
>
> --
> Donald Drake
> Drake Consulting
> http://www.drakeconsulting.com/
> https://twitter.com/dondrake <http://www.MailLaunder.com/>
> 800-733-2143
>

Mime
View raw message