spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Cody Koeninger <c...@koeninger.org>
Subject Re: Installing Spark on Mac
Date Tue, 08 Mar 2016 21:50:05 GMT
That's what I'm saying, there is no "installing" necessary for
pre-built packages.  Just unpack it and change directory into it.

What happens when you do

./bin/spark-shell --master local[2]

or

./bin/start-all.sh



On Tue, Mar 8, 2016 at 3:45 PM, Aida Tefera <aida1.tefera@gmail.com> wrote:
> Hi Cody, thanks for your reply
>
> I tried "sbt/sbt clean assembly" in the Terminal; somehow I still end up with errors.
>
> I have looked at the below links, doesn't give much detail on how to install it before
executing "./sbin/start-master.sh"
>
> Thanks,
>
> Aida
> Sent from my iPhone
>
>> On 8 Mar 2016, at 19:02, Cody Koeninger <cody@koeninger.org> wrote:
>>
>> You said you downloaded a prebuilt version.
>>
>> You shouldn't have to mess with maven or building spark at all.  All
>> you need is a jvm, which it looks like you already have installed.
>>
>> You should be able to follow the instructions at
>>
>> http://spark.apache.org/docs/latest/
>>
>> and
>>
>> http://spark.apache.org/docs/latest/spark-standalone.html
>>
>> If you want standalone mode (master and several worker processes on
>> your machine) rather than local mode (single process on your machine),
>> you need to set up passwordless ssh to localhost
>>
>> http://stackoverflow.com/questions/7134535/setup-passphraseless-ssh-to-localhost-on-os-x
>>
>>
>>
>> On Tue, Mar 8, 2016 at 12:45 PM, Eduardo Costa Alfaia
>> <e.costaalfaia@unibs.it> wrote:
>>> Hi Aida,
>>> The installation has detected a maven version 3.0.3. Update to 3.3.3 and try
>>> again.
>>>
>>> Il 08/Mar/2016 14:06, "Aida" <Aida1.Tefera@gmail.com> ha scritto:
>>>>
>>>> Hi all,
>>>>
>>>> Thanks everyone for your responses; really appreciate it.
>>>>
>>>> Eduardo - I tried your suggestions but ran into some issues, please see
>>>> below:
>>>>
>>>> ukdrfs01:Spark aidatefera$ cd spark-1.6.0
>>>> ukdrfs01:spark-1.6.0 aidatefera$ build/mvn -DskipTests clean package
>>>> Using `mvn` from path: /usr/bin/mvn
>>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>>> MaxPermSize=512M;
>>>> support was removed in 8.0
>>>> [INFO] Scanning for projects...
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [INFO] Reactor Build Order:
>>>> [INFO]
>>>> [INFO] Spark Project Parent POM
>>>> [INFO] Spark Project Test Tags
>>>> [INFO] Spark Project Launcher
>>>> [INFO] Spark Project Networking
>>>> [INFO] Spark Project Shuffle Streaming Service
>>>> [INFO] Spark Project Unsafe
>>>> [INFO] Spark Project Core
>>>> [INFO] Spark Project Bagel
>>>> [INFO] Spark Project GraphX
>>>> [INFO] Spark Project Streaming
>>>> [INFO] Spark Project Catalyst
>>>> [INFO] Spark Project SQL
>>>> [INFO] Spark Project ML Library
>>>> [INFO] Spark Project Tools
>>>> [INFO] Spark Project Hive
>>>> [INFO] Spark Project Docker Integration Tests
>>>> [INFO] Spark Project REPL
>>>> [INFO] Spark Project Assembly
>>>> [INFO] Spark Project External Twitter
>>>> [INFO] Spark Project External Flume Sink
>>>> [INFO] Spark Project External Flume
>>>> [INFO] Spark Project External Flume Assembly
>>>> [INFO] Spark Project External MQTT
>>>> [INFO] Spark Project External MQTT Assembly
>>>> [INFO] Spark Project External ZeroMQ
>>>> [INFO] Spark Project External Kafka
>>>> [INFO] Spark Project Examples
>>>> [INFO] Spark Project External Kafka Assembly
>>>> [INFO]
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [INFO] Building Spark Project Parent POM 1.6.0
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [INFO]
>>>> [INFO] --- maven-clean-plugin:2.6.1:clean (default-clean) @
>>>> spark-parent_2.10 ---
>>>> [INFO]
>>>> [INFO] --- maven-enforcer-plugin:1.4:enforce (enforce-versions) @
>>>> spark-parent_2.10 ---
>>>> [WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion
>>>> failed with message:
>>>> Detected Maven Version: 3.0.3 is not in the allowed range 3.3.3.
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [INFO] Reactor Summary:
>>>> [INFO]
>>>> [INFO] Spark Project Parent POM .......................... FAILURE
>>>> [0.821s]
>>>> [INFO] Spark Project Test Tags ........................... SKIPPED
>>>> [INFO] Spark Project Launcher ............................ SKIPPED
>>>> [INFO] Spark Project Networking .......................... SKIPPED
>>>> [INFO] Spark Project Shuffle Streaming Service ........... SKIPPED
>>>> [INFO] Spark Project Unsafe .............................. SKIPPED
>>>> [INFO] Spark Project Core ................................ SKIPPED
>>>> [INFO] Spark Project Bagel ............................... SKIPPED
>>>> [INFO] Spark Project GraphX .............................. SKIPPED
>>>> [INFO] Spark Project Streaming ........................... SKIPPED
>>>> [INFO] Spark Project Catalyst ............................ SKIPPED
>>>> [INFO] Spark Project SQL ................................. SKIPPED
>>>> [INFO] Spark Project ML Library .......................... SKIPPED
>>>> [INFO] Spark Project Tools ............................... SKIPPED
>>>> [INFO] Spark Project Hive ................................ SKIPPED
>>>> [INFO] Spark Project Docker Integration Tests ............ SKIPPED
>>>> [INFO] Spark Project REPL ................................ SKIPPED
>>>> [INFO] Spark Project Assembly ............................ SKIPPED
>>>> [INFO] Spark Project External Twitter .................... SKIPPED
>>>> [INFO] Spark Project External Flume Sink ................. SKIPPED
>>>> [INFO] Spark Project External Flume ...................... SKIPPED
>>>> [INFO] Spark Project External Flume Assembly ............. SKIPPED
>>>> [INFO] Spark Project External MQTT ....................... SKIPPED
>>>> [INFO] Spark Project External MQTT Assembly .............. SKIPPED
>>>> [INFO] Spark Project External ZeroMQ ..................... SKIPPED
>>>> [INFO] Spark Project External Kafka ...................... SKIPPED
>>>> [INFO] Spark Project Examples ............................ SKIPPED
>>>> [INFO] Spark Project External Kafka Assembly ............. SKIPPED
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [INFO] BUILD FAILURE
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [INFO] Total time: 1.745s
>>>> [INFO] Finished at: Tue Mar 08 18:01:48 GMT 2016
>>>> [INFO] Final Memory: 19M/183M
>>>> [INFO]
>>>> ------------------------------------------------------------------------
>>>> [ERROR] Failed to execute goal
>>>> org.apache.maven.plugins:maven-enforcer-plugin:1.4:enforce
>>>> (enforce-versions) on project spark-parent_2.10: Some Enforcer rules have
>>>> failed. Look above for specific messages explaining why the rule failed.
>>>> ->
>>>> [Help 1]
>>>> [ERROR]
>>>> [ERROR] To see the full stack trace of the errors, re-run Maven with the
>>>> -e
>>>> switch.
>>>> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
>>>> [ERROR]
>>>> [ERROR] For more information about the errors and possible solutions,
>>>> please
>>>> read the following articles:
>>>> [ERROR] [Help 1]
>>>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
>>>> ukdrfs01:spark-1.6.0 aidatefera$
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> View this message in context:
>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Installing-Spark-on-Mac-tp26397p26431.html
>>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
>>>> For additional commands, e-mail: user-help@spark.apache.org
>>>
>>> Informativa sulla Privacy: http://www.unibs.it/node/8155

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org


Mime
View raw message