spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Zahid Rahman <zahidr1...@gmail.com>
Subject Re: Hostname :BUG
Date Thu, 05 Mar 2020 07:01:29 GMT
Please explain why you think that if there is a different reason from this
: -

If you think that, because the header of /etc/hostname says hosts then that
is because I copied the file header from /etc/hosts to  /etc/hostname.




On Wed, 4 Mar 2020, 21:14 Andrew Melo, <andrew.melo@gmail.com> wrote:

> Hello Zabid,
>
> On Wed, Mar 4, 2020 at 1:47 PM Zahid Rahman <zahidr1000@gmail.com> wrote:
>
>> Hi,
>>
>> I found the problem was because on my  Linux   Operating System the
>> /etc/hostname was blank.
>>
>> *STEP 1*
>> I searched  on google the error message and there was an answer suggesting
>> I should add to /etc/hostname
>>
>> 127.0.0.1  [hostname] localhost.
>>
>
> I believe you've confused /etc/hostname and /etc/hosts --
>
>
>>
>> I did that but there was still  an error,  this time the spark  log in
>> standard output was concatenating the text content
>> of etc/hostname  like so ,   127.0.0.1[hostname]localhost.
>>
>> *STEP 2*
>> My second attempt was to change the /etc/hostname to 127.0.0.1
>> This time I was getting a warning with information about "using loop
>> back"  rather than an error.
>>
>> *STEP 3*
>> I wasn't happy with that so then I changed the /etc/hostname to (see
>> below) ,
>> then the warning message disappeared. my guess is that it is the act of
>> creating spark session as to the cause of error,
>> in SparkConf() API.
>>
>>      SparkConf sparkConf = new SparkConf()
>>              .setAppName("Simple Application")
>>              .setMaster("local")
>>              .set("spark.executor.memory","2g");
>>
>> $ cat /etc/hostname
>> # hosts         This file describes a number of hostname-to-address
>> #               mappings for the TCP/IP subsystem.  It is mostly
>> #               used at boot time, when no name servers are running.
>> #               On small systems, this file can be used instead of a
>> #               "named" name server.
>> # Syntax:
>> #
>> # IP-Address  Full-Qualified-Hostname  Short-Hostname
>> #
>>
>> 192.168.0.42
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> zahid@localhost
>> :~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek>
>> mvn exec:java -Dexec.mainClass=com.javacodegeek.examples.SparkExampleRDD
>> -Dexec.args="input.txt"
>> [INFO] Scanning for projects...
>> [WARNING]
>> [WARNING] Some problems were encountered while building the effective
>> model for javacodegeek:examples:jar:1.0-SNAPSHOT
>> [WARNING] 'build.plugins.plugin.version' for
>> org.apache.maven.plugins:maven-compiler-plugin is missing. @ line 12,
>> column 21
>> [WARNING]
>> [WARNING] It is highly recommended to fix these problems because they
>> threaten the stability of your build.
>> [WARNING]
>> [WARNING] For this reason, future Maven versions might no longer support
>> building such malformed projects.
>> [WARNING]
>> [INFO]
>> [INFO] -----------------------< javacodegeek:examples
>> >------------------------
>> [INFO] Building examples 1.0-SNAPSHOT
>> [INFO] --------------------------------[ jar
>> ]---------------------------------
>> [INFO]
>> [INFO] --- exec-maven-plugin:1.6.0:java (default-cli) @ examples ---
>> WARNING: An illegal reflective access operation has occurred
>> WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform
>> (file:/home/zahid/.m2/repository/org/apache/spark/spark-unsafe_2.12/2.4.5/spark-unsafe_2.12-2.4.5.jar)
>> to method java.nio.Bits.unaligned()
>> WARNING: Please consider reporting this to the maintainers of
>> org.apache.spark.unsafe.Platform
>> WARNING: Use --illegal-access=warn to enable warnings of further illegal
>> reflective access operations
>> WARNING: All illegal access operations will be denied in a future release
>> Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>> 20/02/29 17:20:40 INFO SparkContext: Running Spark version 2.4.5
>> 20/02/29 17:20:40 WARN NativeCodeLoader: Unable to load native-hadoop
>> library for your platform... using builtin-java classes where applicable
>> 20/02/29 17:20:41 INFO SparkContext: Submitted application: Word Count
>> 20/02/29 17:20:41 INFO SecurityManager: Changing view acls to: zahid
>> 20/02/29 17:20:41 INFO SecurityManager: Changing modify acls to: zahid
>> 20/02/29 17:20:41 INFO SecurityManager: Changing view acls groups to:
>> 20/02/29 17:20:41 INFO SecurityManager: Changing modify acls groups to:
>> 20/02/29 17:20:41 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users  with view permissions: Set(zahid);
>> groups with view permissions: Set(); users  with modify permissions:
>> Set(zahid); groups with modify permissions: Set()
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:20:41 ERROR SparkContext: Error initializing SparkContext.
>> java.net.BindException: Cannot assign requested address: Service
>> 'sparkDriver' failed after 16 retries (on a random free port)! Consider
>> explicitly setting the appropriate binding address for the service
>> 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the
>> correct binding address.
>>         at java.base/sun.nio.ch.Net.bind0(Native Method)
>>         at java.base/sun.nio.ch.Net.bind(Net.java:469)
>>         at java.base/sun.nio.ch.Net.bind(Net.java:458)
>>         at
>> java.base/sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:220)
>>         at
>> io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:132)
>>         at
>> io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:551)
>>         at
>> io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1346)
>>         at
>> io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:503)
>>         at
>> io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:488)
>>         at
>> io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:985)
>>         at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:247)
>>         at
>> io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:344)
>>         at
>> io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
>>         at
>> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:510)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:518)
>>         at
>> io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1044)
>>         at
>> io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>>         at
>> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
>>         at java.base/java.lang.Thread.run(Thread.java:830)
>> 20/02/29 17:20:41 INFO SparkContext: Successfully stopped SparkContext
>> [WARNING]
>> java.net.BindException: Cannot assign requested address: Service
>> 'sparkDriver' failed after 16 retries (on a random free port)! Consider
>> explicitly setting the appropriate binding address for the service
>> 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the
>> correct binding address.
>>     at sun.nio.ch.Net.bind0 (Native Method)
>>     at sun.nio.ch.Net.bind (Net.java:469)
>>     at sun.nio.ch.Net.bind (Net.java:458)
>>     at sun.nio.ch.ServerSocketChannelImpl.bind
>> (ServerSocketChannelImpl.java:220)
>>     at io.netty.channel.socket.nio.NioServerSocketChannel.doBind
>> (NioServerSocketChannel.java:132)
>>     at io.netty.channel.AbstractChannel$AbstractUnsafe.bind
>> (AbstractChannel.java:551)
>>     at io.netty.channel.DefaultChannelPipeline$HeadContext.bind
>> (DefaultChannelPipeline.java:1346)
>>     at io.netty.channel.AbstractChannelHandlerContext.invokeBind
>> (AbstractChannelHandlerContext.java:503)
>>     at io.netty.channel.AbstractChannelHandlerContext.bind
>> (AbstractChannelHandlerContext.java:488)
>>     at io.netty.channel.DefaultChannelPipeline.bind
>> (DefaultChannelPipeline.java:985)
>>     at io.netty.channel.AbstractChannel.bind (AbstractChannel.java:247)
>>     at io.netty.bootstrap.AbstractBootstrap$2.run
>> (AbstractBootstrap.java:344)
>>     at io.netty.util.concurrent.AbstractEventExecutor.safeExecute
>> (AbstractEventExecutor.java:163)
>>     at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks
>> (SingleThreadEventExecutor.java:510)
>>     at io.netty.channel.nio.NioEventLoop.run (NioEventLoop.java:518)
>>     at io.netty.util.concurrent.SingleThreadEventExecutor$6.run
>> (SingleThreadEventExecutor.java:1044)
>>     at io.netty.util.internal.ThreadExecutorMap$2.run
>> (ThreadExecutorMap.java:74)
>>     at io.netty.util.concurrent.FastThreadLocalRunnable.run
>> (FastThreadLocalRunnable.java:30)
>>     at java.lang.Thread.run (Thread.java:830)
>> [INFO]
>> ------------------------------------------------------------------------
>> [INFO] BUILD FAILURE
>> [INFO]
>> ------------------------------------------------------------------------
>> [INFO] Total time:  3.977 s
>> [INFO] Finished at: 2020-02-29T17:20:43Z
>> [INFO]
>> ------------------------------------------------------------------------
>> [ERROR] Failed to execute goal
>> org.codehaus.mojo:exec-maven-plugin:1.6.0:java (default-cli) on project
>> examples: An exception occured while executing the Java class. Cannot
>> assign requested address: Service 'sparkDriver' failed after 16 retries (on
>> a random free port)! Consider explicitly setting the appropriate binding
>> address for the service 'sparkDriver' (for example spark.driver.bindAddress
>> for SparkDriver) to the correct binding address. -> [Help 1]
>> [ERROR]
>> [ERROR] To see the full stack trace of the errors, re-run Maven with the
>> -e switch.
>> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
>> [ERROR]
>> [ERROR] For more information about the errors and possible solutions,
>> please read the following articles:
>> [ERROR] [Help 1]
>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
>> zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek>
>> mvn exec:java -Dexec.mainClass=com.javacodegeek.examples.SparkExampleRDD
>> -Dexec.args="input.txt" > log.out
>> WARNING: An illegal reflective access operation has occurred
>> WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform
>> (file:/home/zahid/.m2/repository/org/apache/spark/spark-unsafe_2.12/2.4.5/spark-unsafe_2.12-2.4.5.jar)
>> to method java.nio.Bits.unaligned()
>> WARNING: Please consider reporting this to the maintainers of
>> org.apache.spark.unsafe.Platform
>> WARNING: Use --illegal-access=warn to enable warnings of further illegal
>> reflective access operations
>> WARNING: All illegal access operations will be denied in a future release
>> Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>> 20/02/29 17:21:12 INFO SparkContext: Running Spark version 2.4.5
>> 20/02/29 17:21:12 WARN NativeCodeLoader: Unable to load native-hadoop
>> library for your platform... using builtin-java classes where applicable
>> 20/02/29 17:21:12 INFO SparkContext: Submitted application: Word Count
>> 20/02/29 17:21:12 INFO SecurityManager: Changing view acls to: zahid
>> 20/02/29 17:21:12 INFO SecurityManager: Changing modify acls to: zahid
>> 20/02/29 17:21:12 INFO SecurityManager: Changing view acls groups to:
>> 20/02/29 17:21:12 INFO SecurityManager: Changing modify acls groups to:
>> 20/02/29 17:21:12 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users  with view permissions: Set(zahid);
>> groups with view permissions: Set(); users  with modify permissions:
>> Set(zahid); groups with modify permissions: Set()
>> 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a
>> random free port. You may check whether configuring an appropriate binding
>> address.
>> 20/02/29 17:21:13 ERROR SparkContext: Error initializing SparkContext.
>> java.net.BindException: Cannot assign requested address: Service
>> 'sparkDriver' failed after 16 retries (on a random free port)! Consider
>> explicitly setting the appropriate binding address for the service
>> 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the
>> correct binding address.
>>         at java.base/sun.nio.ch.Net.bind0(Native Method)
>>         at java.base/sun.nio.ch.Net.bind(Net.java:469)
>>         at java.base/sun.nio.ch.Net.bind(Net.java:458)
>>         at
>> java.base/sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:220)
>>         at
>> io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:132)
>>         at
>> io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:551)
>>         at
>> io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1346)
>>         at
>> io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:503)
>>         at
>> io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:488)
>>         at
>> io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:985)
>>         at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:247)
>>         at
>> io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:344)
>>         at
>> io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
>>         at
>> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:510)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:518)
>>         at
>> io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1044)
>>         at
>> io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
>>         at
>> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
>>         at java.base/java.lang.Thread.run(Thread.java:830)
>> 20/02/29 17:21:13 INFO SparkContext: Successfully stopped SparkContext
>> zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek>
>> ls
>> input.txt  log.out  pom.xml  src  target
>> zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek>
>> cat log.out
>> [INFO] Scanning for projects...
>> [WARNING]
>> [WARNING] Some problems were encountered while building the effective
>> model for javacodegeek:examples:jar:1.0-SNAPSHOT
>> [WARNING] 'build.plugins.plugin.version' for
>> org.apache.maven.plugins:maven-compiler-plugin is missing. @ line 12,
>> column 21
>> [WARNING]
>> [WARNING] It is highly recommended to fix these problems because they
>> threaten the stability of your build.
>> [WARNING]
>> [WARNING] For this reason, future Maven versions might no longer support
>> building such malformed projects.
>> [WARNING]
>> [INFO]
>> [INFO] -----------------------< javacodegeek:examples
>> >------------------------
>> [INFO] Building examples 1.0-SNAPSHOT
>> [INFO] --------------------------------[ jar
>> ]---------------------------------
>> [INFO]
>> [INFO] --- exec-maven-plugin:1.6.0:java (default-cli) @ examples ---
>> [WARNING]
>> java.net.BindException: Cannot assign requested address: Service
>> 'sparkDriver' failed after 16 retries (on a random free port)! Consider
>> explicitly setting the appropriate binding address for the service
>> 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the
>> correct binding address.
>>     at sun.nio.ch.Net.bind0 (Native Method)
>>     at sun.nio.ch.Net.bind (Net.java:469)
>>     at sun.nio.ch.Net.bind (Net.java:458)
>>     at sun.nio.ch.ServerSocketChannelImpl.bind
>> (ServerSocketChannelImpl.java:220)
>>     at io.netty.channel.socket.nio.NioServerSocketChannel.doBind
>> (NioServerSocketChannel.java:132)
>>     at io.netty.channel.AbstractChannel$AbstractUnsafe.bind
>> (AbstractChannel.java:551)
>>     at io.netty.channel.DefaultChannelPipeline$HeadContext.bind
>> (DefaultChannelPipeline.java:1346)
>>     at io.netty.channel.AbstractChannelHandlerContext.invokeBind
>> (AbstractChannelHandlerContext.java:503)
>>     at io.netty.channel.AbstractChannelHandlerContext.bind
>> (AbstractChannelHandlerContext.java:488)
>>     at io.netty.channel.DefaultChannelPipeline.bind
>> (DefaultChannelPipeline.java:985)
>>     at io.netty.channel.AbstractChannel.bind (AbstractChannel.java:247)
>>     at io.netty.bootstrap.AbstractBootstrap$2.run
>> (AbstractBootstrap.java:344)
>>     at io.netty.util.concurrent.AbstractEventExecutor.safeExecute
>> (AbstractEventExecutor.java:163)
>>     at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks
>> (SingleThreadEventExecutor.java:510)
>>     at io.netty.channel.nio.NioEventLoop.run (NioEventLoop.java:518)
>>     at io.netty.util.concurrent.SingleThreadEventExecutor$6.run
>> (SingleThreadEventExecutor.java:1044)
>>     at io.netty.util.internal.ThreadExecutorMap$2.run
>> (ThreadExecutorMap.java:74)
>>     at io.netty.util.concurrent.FastThreadLocalRunnable.run
>> (FastThreadLocalRunnable.java:30)
>>     at java.lang.Thread.run (Thread.java:830)
>> [INFO]
>> ------------------------------------------------------------------------
>> [INFO] BUILD FAILURE
>> [INFO]
>> ------------------------------------------------------------------------
>> [INFO] Total time:  4.949 s
>> [INFO] Finished at: 2020-02-29T17:21:16Z
>> [INFO]
>> ------------------------------------------------------------------------
>> [ERROR] Failed to execute goal
>> org.codehaus.mojo:exec-maven-plugin:1.6.0:java (default-cli) on project
>> examples: An exception occured while executing the Java class. Cannot
>> assign requested address: Service 'sparkDriver' failed after 16 retries (on
>> a random free port)! Consider explicitly setting the appropriate binding
>> address for the service 'sparkDriver' (for example spark.driver.bindAddress
>> for SparkDriver) to the correct binding address. -> [Help 1]
>> [ERROR]
>> [ERROR] To see the full stack trace of the errors, re-run Maven with the
>> -e switch.
>> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
>> [ERROR]
>> [ERROR] For more information about the errors and possible solutions,
>> please read the following articles:
>> [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionExc:
>> BUGeption
>> <http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException>
>> zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek>
>>
>>
>>
>> Backbutton.co.uk
>> ¯\_(ツ)_/¯
>> ♡۶Java♡۶RMI ♡۶
>> Make Use Method {MUM}
>> makeuse.org
>> <http://www.backbutton.co.uk>
>>
>

Mime
View raw message