spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From <spark....@yahoo.com.INVALID>
Subject spark job is not running on yarn clustor mode
Date Tue, 17 May 2016 12:38:53 GMT
Hi friends,
I am running spark streaming job on yarn cluster mode but it is failing. It is working fine in yarn-client mode. and also spark-examples are running good in spark-cluster mode. below is the log file for the spark streaming job on yarn-cluster mode. Can anyone help me on this.

SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/filecache/15/spark-assembly-1.5.2-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
16/05/17 16:17:47 INFO yarn.ApplicationMaster: Registered signal handlers for [TERM, HUP, INT]
16/05/17 16:17:48 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
16/05/17 16:17:48 INFO yarn.ApplicationMaster: ApplicationAttemptId: appattempt_1463479181441_0003_000002
16/05/17 16:17:49 INFO spark.SecurityManager: Changing view acls to: hadoop
16/05/17 16:17:49 INFO spark.SecurityManager: Changing modify acls to: hadoop
16/05/17 16:17:49 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); users with modify permissions: Set(hadoop)
16/05/17 16:17:49 INFO yarn.ApplicationMaster: Starting the user application in a separate Thread
16/05/17 16:17:49 INFO yarn.ApplicationMaster: Waiting for spark context initialization
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: found keyword== userTwitterToken=9ACWejzaHVyxpPDYCHnDsO98U 01safwuyLO8B8S94v5i0p90SzxEPZqUUmCaDkYOj1FKN1dXKZC 702828259411521536-PNoSkM8xNIvuEVvoQ9Pj8fj7D8CkYp1 OntoQStrmwrztnzi1MSlM56sKc23bqUCC2WblbDPiiP8P
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = 9ACWejzaHVyxpPDYCHnDsO98U 01safwuyLO8B8S94v5i0p90SzxEPZqUUmCaDkYOj1FKN1dXKZC 702828259411521536-PNoSkM8xNIvuEVvoQ9Pj8fj7D8CkYp1 OntoQStrmwrztnzi1MSlM56sKc23bqUCC2WblbDPiiP8P
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = 1
16/05/17 16:17:49 INFO yarn.ApplicationMaster: Waiting for spark context initialization ... 
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = 2
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = Tue May 17 00:00:00 IST 2016
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = Tue May 17 00:00:00 IST 2016
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = nokia,samsung,iphone,blackberry
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = All
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = mo
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = en
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: DemoJava called = retweet
16/05/17 16:17:49 INFO spark.SparkTweetStreamingHDFSLoad: Twitter Token...........[Ljava.lang.String;@3ee5e48d
16/05/17 16:17:49 INFO spark.SparkContext: Running Spark version 1.5.2
16/05/17 16:17:49 WARN spark.SparkConf: 
SPARK_JAVA_OPTS was detected (set to '-Dspark.driver.port=53411').
This is deprecated in Spark 1.0+.

Please instead use:
 - ./spark-submit with conf/spark-defaults.conf to set defaults for an application
 - ./spark-submit with --driver-java-options to set -X options for a driver
 - spark.executor.extraJavaOptions to set -X options for executors
 - SPARK_DAEMON_JAVA_OPTS to set java options for standalone daemons (master or worker)
        
16/05/17 16:17:49 WARN spark.SparkConf: Setting 'spark.executor.extraJavaOptions' to '-Dspark.driver.port=53411' as a work-around.
16/05/17 16:17:49 WARN spark.SparkConf: Setting 'spark.driver.extraJavaOptions' to '-Dspark.driver.port=53411' as a work-around.
16/05/17 16:17:49 INFO spark.SecurityManager: Changing view acls to: hadoop
16/05/17 16:17:49 INFO spark.SecurityManager: Changing modify acls to: hadoop
16/05/17 16:17:49 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(hadoop); users with modify permissions: Set(hadoop)
16/05/17 16:17:49 INFO slf4j.Slf4jLogger: Slf4jLogger started
16/05/17 16:17:49 INFO Remoting: Starting remoting
16/05/17 16:17:50 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@172.16.28.195:53411]
16/05/17 16:17:50 INFO util.Utils: Successfully started service 'sparkDriver' on port 53411.
16/05/17 16:17:50 INFO spark.SparkEnv: Registering MapOutputTracker
16/05/17 16:17:50 INFO spark.SparkEnv: Registering BlockManagerMaster
16/05/17 16:17:50 INFO storage.DiskBlockManager: Created local directory at /tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/appcache/application_1463479181441_0003/blockmgr-fe61bf50-b650-4db9-989a-11199df6c1ac
16/05/17 16:17:50 INFO storage.MemoryStore: MemoryStore started with capacity 1966.1 MB
16/05/17 16:17:50 INFO spark.HttpFileServer: HTTP File server directory is /tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/appcache/application_1463479181441_0003/spark-5b36342a-6212-4cea-80da-b1961cab161c/httpd-20144975-e972-4b5a-8592-be94029cd0eb
16/05/17 16:17:50 INFO spark.HttpServer: Starting HTTP Server
16/05/17 16:17:50 INFO server.Server: jetty-8.y.z-SNAPSHOT
16/05/17 16:17:50 INFO server.AbstractConnector: Started SocketConnector@0.0.0.0:47195
16/05/17 16:17:50 INFO util.Utils: Successfully started service 'HTTP file server' on port 47195.
16/05/17 16:17:50 INFO spark.SparkEnv: Registering OutputCommitCoordinator
16/05/17 16:17:50 INFO ui.JettyUtils: Adding filter: org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
16/05/17 16:17:55 INFO server.Server: jetty-8.y.z-SNAPSHOT
16/05/17 16:17:55 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:59320
16/05/17 16:17:55 INFO util.Utils: Successfully started service 'SparkUI' on port 59320.
16/05/17 16:17:55 INFO ui.SparkUI: Started SparkUI at http://172.16.28.195:59320
16/05/17 16:17:55 INFO cluster.YarnClusterScheduler: Created YarnClusterScheduler
16/05/17 16:17:55 WARN metrics.MetricsSystem: Using default name DAGScheduler for source because spark.app.id is not set.
16/05/17 16:17:55 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 57488.
16/05/17 16:17:55 INFO netty.NettyBlockTransferService: Server created on 57488
16/05/17 16:17:55 INFO storage.BlockManagerMaster: Trying to register BlockManager
16/05/17 16:17:55 INFO storage.BlockManagerMasterEndpoint: Registering block manager 172.16.28.195:57488 with 1966.1 MB RAM, BlockManagerId(driver, 172.16.28.195, 57488)
16/05/17 16:17:55 INFO storage.BlockManagerMaster: Registered BlockManager
16/05/17 16:17:56 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster registered as AkkaRpcEndpointRef(Actor[akka://sparkDriver/user/YarnAM#-174037885])
16/05/17 16:17:56 INFO client.RMProxy: Connecting to ResourceManager at namenode/172.16.28.190:8030
16/05/17 16:17:56 INFO yarn.YarnRMClient: Registering the ApplicationMaster
16/05/17 16:17:56 INFO yarn.YarnAllocator: Will request 2 executor containers, each with 1 cores and 1408 MB memory including 384 MB overhead
16/05/17 16:17:56 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1>)
16/05/17 16:17:56 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1>)
16/05/17 16:17:56 INFO yarn.ApplicationMaster: Started progress reporter thread with (heartbeat : 3000, initial allocation : 200) intervals
16/05/17 16:17:56 INFO impl.AMRMClientImpl: Received new token for : node4:58299
16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching container container_1463479181441_0003_02_000002 for on host node4
16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler,  executorHostname: node4
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Starting Executor Container
16/05/17 16:17:56 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 1 of them.
16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Preparing Local resources
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__app__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar" } size: 216515519 timestamp: 1463481955892 type: FILE visibility: PRIVATE, __spark__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar" } size: 183993445 timestamp: 1463481933738 type: FILE visibility: PRIVATE)
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: 
===============================================================================
YARN executor launch context:
  env:
    CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*
    SPARK_LOG_URL_STDERR -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000002/hadoop/stderr?start=-4096
    SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1463479181441_0003
    SPARK_YARN_CACHE_FILES_FILE_SIZES -> 183993445,216515519
    SPARK_USER -> hadoop
    SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE,PRIVATE
    SPARK_YARN_MODE -> true
    SPARK_JAVA_OPTS -> -Dspark.driver.port=53411
    SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1463481933738,1463481955892
    SPARK_LOG_URL_STDOUT -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000002/hadoop/stdout?start=-4096
    SPARK_YARN_CACHE_FILES -> hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar,hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar#__app__.jar

  command:
    {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms1024m -Xmx1024m '-Dspark.driver.port=53411' -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.ui.port=0' '-Dspark.driver.port=53411' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler --executor-id 1 --hostname node4 --cores 1 --app-id application_1463479181441_0003 --user-class-path file:$PWD/__app__.jar 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr
===============================================================================
      
16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: Opening proxy : node4:58299
16/05/17 16:17:56 INFO impl.AMRMClientImpl: Received new token for : node2:52751
16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching container container_1463479181441_0003_02_000003 for on host node2
16/05/17 16:17:56 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler,  executorHostname: node2
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Starting Executor Container
16/05/17 16:17:56 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 1 of them.
16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Preparing Local resources
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__app__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar" } size: 216515519 timestamp: 1463481955892 type: FILE visibility: PRIVATE, __spark__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar" } size: 183993445 timestamp: 1463481933738 type: FILE visibility: PRIVATE)
16/05/17 16:17:56 INFO yarn.ExecutorRunnable: 
===============================================================================
YARN executor launch context:
  env:
    CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*
    SPARK_LOG_URL_STDERR -> http://node2:8042/node/containerlogs/container_1463479181441_0003_02_000003/hadoop/stderr?start=-4096
    SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1463479181441_0003
    SPARK_YARN_CACHE_FILES_FILE_SIZES -> 183993445,216515519
    SPARK_USER -> hadoop
    SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE,PRIVATE
    SPARK_YARN_MODE -> true
    SPARK_JAVA_OPTS -> -Dspark.driver.port=53411
    SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1463481933738,1463481955892
    SPARK_LOG_URL_STDOUT -> http://node2:8042/node/containerlogs/container_1463479181441_0003_02_000003/hadoop/stdout?start=-4096
    SPARK_YARN_CACHE_FILES -> hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar,hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar#__app__.jar

  command:
    {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms1024m -Xmx1024m '-Dspark.driver.port=53411' -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.ui.port=0' '-Dspark.driver.port=53411' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler --executor-id 2 --hostname node2 --cores 1 --app-id application_1463479181441_0003 --user-class-path file:$PWD/__app__.jar 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr
===============================================================================
      
16/05/17 16:17:56 INFO impl.ContainerManagementProtocolProxy: Opening proxy : node2:52751
16/05/17 16:17:59 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:39430
16/05/17 16:17:59 INFO cluster.YarnClusterSchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@node4:50089/user/Executor#1750526367]) with ID 1
16/05/17 16:17:59 INFO storage.BlockManagerMasterEndpoint: Registering block manager node4:47743 with 530.0 MB RAM, BlockManagerId(1, node4, 47743)
16/05/17 16:17:59 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 0 of them.
16/05/17 16:17:59 INFO yarn.YarnAllocator: Completed container container_1463479181441_0003_02_000003 (state: COMPLETE, exit status: -1000)
16/05/17 16:17:59 INFO yarn.YarnAllocator: Container marked as failed: container_1463479181441_0003_02_000003. Exit status: -1000. Diagnostics: File does not exist: hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar
java.io.FileNotFoundException: File does not exist: hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar
	at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1122)
	at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
	at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
	at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251)
	at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61)
	at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359)
	at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
	at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356)
	at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)


16/05/17 16:17:59 INFO cluster.YarnClusterSchedulerBackend: Asked to remove non-existent executor 2
16/05/17 16:18:02 INFO yarn.YarnAllocator: Will request 1 executor containers, each with 1 cores and 1408 MB memory including 384 MB overhead
16/05/17 16:18:02 INFO yarn.YarnAllocator: Container request (host: Any, capability: <memory:1408, vCores:1>)
16/05/17 16:18:03 INFO yarn.YarnAllocator: Launching container container_1463479181441_0003_02_000005 for on host node4
16/05/17 16:18:03 INFO yarn.YarnAllocator: Launching ExecutorRunnable. driverUrl: akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler,  executorHostname: node4
16/05/17 16:18:03 INFO yarn.YarnAllocator: Received 1 containers from YARN, launching executors on 1 of them.
16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Starting Executor Container
16/05/17 16:18:03 INFO impl.ContainerManagementProtocolProxy: yarn.client.max-cached-nodemanagers-proxies : 0
16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Setting up ContainerLaunchContext
16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Preparing Local resources
16/05/17 16:18:03 INFO yarn.ExecutorRunnable: Prepared Local resources Map(__app__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar" } size: 216515519 timestamp: 1463481955892 type: FILE visibility: PRIVATE, __spark__.jar -> resource { scheme: "hdfs" host: "namenode" port: 54310 file: "/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar" } size: 183993445 timestamp: 1463481933738 type: FILE visibility: PRIVATE)
16/05/17 16:18:03 INFO yarn.ExecutorRunnable: 
===============================================================================
YARN executor launch context:
  env:
    CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*
    SPARK_LOG_URL_STDERR -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000005/hadoop/stderr?start=-4096
    SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1463479181441_0003
    SPARK_YARN_CACHE_FILES_FILE_SIZES -> 183993445,216515519
    SPARK_USER -> hadoop
    SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE,PRIVATE
    SPARK_YARN_MODE -> true
    SPARK_JAVA_OPTS -> -Dspark.driver.port=53411
    SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1463481933738,1463481955892
    SPARK_LOG_URL_STDOUT -> http://node4:8042/node/containerlogs/container_1463479181441_0003_02_000005/hadoop/stdout?start=-4096
    SPARK_YARN_CACHE_FILES -> hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar,hdfs://namenode:54310/user/hadoop/.sparkStaging/application_1463479181441_0003/SparkTwittterStreamingJob-0.0.1-SNAPSHOT-jar-with-dependencies.jar#__app__.jar

  command:
    {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms1024m -Xmx1024m '-Dspark.driver.port=53411' -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.ui.port=0' '-Dspark.driver.port=53411' -Dspark.yarn.app.container.log.dir=<LOG_DIR> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url akka.tcp://sparkDriver@172.16.28.195:53411/user/CoarseGrainedScheduler --executor-id 3 --hostname node4 --cores 1 --app-id application_1463479181441_0003 --user-class-path file:$PWD/__app__.jar 1> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr
===============================================================================
      
16/05/17 16:18:03 INFO impl.ContainerManagementProtocolProxy: Opening proxy : node4:58299
16/05/17 16:18:06 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:35884
16/05/17 16:18:06 INFO cluster.YarnClusterSchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@node4:46484/user/Executor#-348284167]) with ID 3
16/05/17 16:18:06 INFO cluster.YarnClusterSchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8
16/05/17 16:18:06 INFO cluster.YarnClusterScheduler: YarnClusterScheduler.postStartHook done
16/05/17 16:18:06 INFO storage.BlockManagerMasterEndpoint: Registering block manager node4:58845 with 530.0 MB RAM, BlockManagerId(3, node4, 58845)
16/05/17 16:18:06 INFO spark.SparkTweetStreamingHDFSLoad: dayOfTheWeek .........[Ljava.lang.String;@42c6ef6d
16/05/17 16:18:07 INFO rate.PIDRateEstimator: Created PIDRateEstimator with proportional = 1.0, integral = 0.2, derivative = 0.0, min rate = 100.0
16/05/17 16:18:07 INFO spark.SparkTweetStreamingHDFSLoad: Terminate DAte............Tue May 17 00:00:00 IST 2016
16/05/17 16:18:07 INFO spark.SparkTweetStreamingHDFSLoad: outputURI--------------hdfs://namenode:54310/spark/TweetData/twitterRawDataTest
16/05/17 16:18:07 INFO spark.SparkTweetStreamingHDFSLoad: outputURI--------------hdfs://namenode:54310/spark/TweetData/twitterSeggDataTest
16/05/17 16:18:07 INFO spark.SparkContext: Starting job: start at SparkTweetStreamingHDFSLoad.java:1743
16/05/17 16:18:07 INFO scheduler.DAGScheduler: Registering RDD 1 (start at SparkTweetStreamingHDFSLoad.java:1743)
16/05/17 16:18:07 INFO scheduler.DAGScheduler: Got job 0 (start at SparkTweetStreamingHDFSLoad.java:1743) with 20 output partitions
16/05/17 16:18:07 INFO scheduler.DAGScheduler: Final stage: ResultStage 1(start at SparkTweetStreamingHDFSLoad.java:1743)
16/05/17 16:18:07 INFO scheduler.DAGScheduler: Parents of final stage: List(ShuffleMapStage 0)
16/05/17 16:18:07 INFO scheduler.DAGScheduler: Missing parents: List(ShuffleMapStage 0)
16/05/17 16:18:07 INFO scheduler.DAGScheduler: Submitting ShuffleMapStage 0 (MapPartitionsRDD[1] at start at SparkTweetStreamingHDFSLoad.java:1743), which has no missing parents
16/05/17 16:18:08 INFO storage.MemoryStore: ensureFreeSpace(2736) called with curMem=0, maxMem=2061647216
16/05/17 16:18:08 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 2.7 KB, free 1966.1 MB)
16/05/17 16:18:08 INFO storage.MemoryStore: ensureFreeSpace(1655) called with curMem=2736, maxMem=2061647216
16/05/17 16:18:08 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 1655.0 B, free 1966.1 MB)
16/05/17 16:18:08 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 172.16.28.195:57488 (size: 1655.0 B, free: 1966.1 MB)
16/05/17 16:18:08 INFO spark.SparkContext: Created broadcast 0 from broadcast at DAGScheduler.scala:861
16/05/17 16:18:08 INFO scheduler.DAGScheduler: Submitting 50 missing tasks from ShuffleMapStage 0 (MapPartitionsRDD[1] at start at SparkTweetStreamingHDFSLoad.java:1743)
16/05/17 16:18:08 INFO cluster.YarnClusterScheduler: Adding task set 0.0 with 50 tasks
16/05/17 16:18:08 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:08 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 0.0 (TID 1, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:12 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on node4:47743 (size: 1655.0 B, free: 530.0 MB)
16/05/17 16:18:12 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on node4:58845 (size: 1655.0 B, free: 530.0 MB)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 2.0 in stage 0.0 (TID 2, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 3.0 in stage 0.0 (TID 3, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 1.0 in stage 0.0 (TID 1) in 4243 ms on node4 (1/50)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 4296 ms on node4 (2/50)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 4.0 in stage 0.0 (TID 4, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Starting task 5.0 in stage 0.0 (TID 5, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 2.0 in stage 0.0 (TID 2) in 149 ms on node4 (3/50)
16/05/17 16:18:12 INFO scheduler.TaskSetManager: Finished task 3.0 in stage 0.0 (TID 3) in 143 ms on node4 (4/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 6.0 in stage 0.0 (TID 6, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 7.0 in stage 0.0 (TID 7, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 4.0 in stage 0.0 (TID 4) in 109 ms on node4 (5/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 5.0 in stage 0.0 (TID 5) in 88 ms on node4 (6/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 8.0 in stage 0.0 (TID 8, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 6.0 in stage 0.0 (TID 6) in 74 ms on node4 (7/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 7.0 in stage 0.0 (TID 7) in 75 ms on node4 (8/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 9.0 in stage 0.0 (TID 9, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 8.0 in stage 0.0 (TID 8) in 83 ms on node4 (9/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 10.0 in stage 0.0 (TID 10, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 9.0 in stage 0.0 (TID 9) in 94 ms on node4 (10/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 11.0 in stage 0.0 (TID 11, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 12.0 in stage 0.0 (TID 12, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 10.0 in stage 0.0 (TID 10) in 70 ms on node4 (11/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 13.0 in stage 0.0 (TID 13, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 11.0 in stage 0.0 (TID 11) in 83 ms on node4 (12/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 14.0 in stage 0.0 (TID 14, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 15.0 in stage 0.0 (TID 15, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 14.0 in stage 0.0 (TID 14) in 64 ms on node4 (13/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 16.0 in stage 0.0 (TID 16, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 13.0 in stage 0.0 (TID 13) in 99 ms on node4 (14/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 12.0 in stage 0.0 (TID 12) in 169 ms on node4 (15/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 17.0 in stage 0.0 (TID 17, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 15.0 in stage 0.0 (TID 15) in 79 ms on node4 (16/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 18.0 in stage 0.0 (TID 18, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 16.0 in stage 0.0 (TID 16) in 112 ms on node4 (17/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 19.0 in stage 0.0 (TID 19, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 17.0 in stage 0.0 (TID 17) in 87 ms on node4 (18/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 20.0 in stage 0.0 (TID 20, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 18.0 in stage 0.0 (TID 18) in 73 ms on node4 (19/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 21.0 in stage 0.0 (TID 21, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 19.0 in stage 0.0 (TID 19) in 89 ms on node4 (20/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 22.0 in stage 0.0 (TID 22, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 20.0 in stage 0.0 (TID 20) in 113 ms on node4 (21/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 21.0 in stage 0.0 (TID 21) in 90 ms on node4 (22/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 23.0 in stage 0.0 (TID 23, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 24.0 in stage 0.0 (TID 24, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 22.0 in stage 0.0 (TID 22) in 85 ms on node4 (23/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 23.0 in stage 0.0 (TID 23) in 71 ms on node4 (24/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 25.0 in stage 0.0 (TID 25, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 26.0 in stage 0.0 (TID 26, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 24.0 in stage 0.0 (TID 24) in 79 ms on node4 (25/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 27.0 in stage 0.0 (TID 27, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 25.0 in stage 0.0 (TID 25) in 77 ms on node4 (26/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 28.0 in stage 0.0 (TID 28, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 26.0 in stage 0.0 (TID 26) in 84 ms on node4 (27/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 29.0 in stage 0.0 (TID 29, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 27.0 in stage 0.0 (TID 27) in 81 ms on node4 (28/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 30.0 in stage 0.0 (TID 30, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 28.0 in stage 0.0 (TID 28) in 70 ms on node4 (29/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 31.0 in stage 0.0 (TID 31, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 29.0 in stage 0.0 (TID 29) in 93 ms on node4 (30/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Finished task 30.0 in stage 0.0 (TID 30) in 74 ms on node4 (31/50)
16/05/17 16:18:13 INFO scheduler.TaskSetManager: Starting task 32.0 in stage 0.0 (TID 32, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 33.0 in stage 0.0 (TID 33, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 32.0 in stage 0.0 (TID 32) in 71 ms on node4 (32/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 31.0 in stage 0.0 (TID 31) in 98 ms on node4 (33/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 34.0 in stage 0.0 (TID 34, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 35.0 in stage 0.0 (TID 35, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 33.0 in stage 0.0 (TID 33) in 85 ms on node4 (34/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 36.0 in stage 0.0 (TID 36, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 34.0 in stage 0.0 (TID 34) in 93 ms on node4 (35/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 37.0 in stage 0.0 (TID 37, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 35.0 in stage 0.0 (TID 35) in 503 ms on node4 (36/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 38.0 in stage 0.0 (TID 38, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 36.0 in stage 0.0 (TID 36) in 496 ms on node4 (37/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 39.0 in stage 0.0 (TID 39, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 37.0 in stage 0.0 (TID 37) in 86 ms on node4 (38/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 40.0 in stage 0.0 (TID 40, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 38.0 in stage 0.0 (TID 38) in 68 ms on node4 (39/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 41.0 in stage 0.0 (TID 41, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 40.0 in stage 0.0 (TID 40) in 62 ms on node4 (40/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 39.0 in stage 0.0 (TID 39) in 87 ms on node4 (41/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 42.0 in stage 0.0 (TID 42, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 43.0 in stage 0.0 (TID 43, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 41.0 in stage 0.0 (TID 41) in 95 ms on node4 (42/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 44.0 in stage 0.0 (TID 44, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 42.0 in stage 0.0 (TID 42) in 110 ms on node4 (43/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 45.0 in stage 0.0 (TID 45, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 43.0 in stage 0.0 (TID 43) in 94 ms on node4 (44/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 46.0 in stage 0.0 (TID 46, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 44.0 in stage 0.0 (TID 44) in 95 ms on node4 (45/50)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Starting task 47.0 in stage 0.0 (TID 47, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:14 INFO scheduler.TaskSetManager: Finished task 45.0 in stage 0.0 (TID 45) in 90 ms on node4 (46/50)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 48.0 in stage 0.0 (TID 48, node4, PROCESS_LOCAL, 1962 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 46.0 in stage 0.0 (TID 46) in 103 ms on node4 (47/50)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 49.0 in stage 0.0 (TID 49, node4, PROCESS_LOCAL, 1929 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 47.0 in stage 0.0 (TID 47) in 93 ms on node4 (48/50)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 48.0 in stage 0.0 (TID 48) in 127 ms on node4 (49/50)
16/05/17 16:18:15 INFO scheduler.DAGScheduler: ShuffleMapStage 0 (start at SparkTweetStreamingHDFSLoad.java:1743) finished in 6.553 s
16/05/17 16:18:15 INFO scheduler.DAGScheduler: looking for newly runnable stages
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 49.0 in stage 0.0 (TID 49) in 94 ms on node4 (50/50)
16/05/17 16:18:15 INFO scheduler.DAGScheduler: running: Set()
16/05/17 16:18:15 INFO scheduler.DAGScheduler: waiting: Set(ResultStage 1)
16/05/17 16:18:15 INFO scheduler.DAGScheduler: failed: Set()
16/05/17 16:18:15 INFO cluster.YarnClusterScheduler: Removed TaskSet 0.0, whose tasks have all completed, from pool 
16/05/17 16:18:15 INFO scheduler.DAGScheduler: Missing parents for ResultStage 1: List()
16/05/17 16:18:15 INFO scheduler.DAGScheduler: Submitting ResultStage 1 (ShuffledRDD[2] at start at SparkTweetStreamingHDFSLoad.java:1743), which is now runnable
16/05/17 16:18:15 INFO storage.MemoryStore: ensureFreeSpace(2344) called with curMem=4391, maxMem=2061647216
16/05/17 16:18:15 INFO storage.MemoryStore: Block broadcast_1 stored as values in memory (estimated size 2.3 KB, free 1966.1 MB)
16/05/17 16:18:15 INFO storage.MemoryStore: ensureFreeSpace(1400) called with curMem=6735, maxMem=2061647216
16/05/17 16:18:15 INFO storage.MemoryStore: Block broadcast_1_piece0 stored as bytes in memory (estimated size 1400.0 B, free 1966.1 MB)
16/05/17 16:18:15 INFO storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on 172.16.28.195:57488 (size: 1400.0 B, free: 1966.1 MB)
16/05/17 16:18:15 INFO spark.SparkContext: Created broadcast 1 from broadcast at DAGScheduler.scala:861
16/05/17 16:18:15 INFO scheduler.DAGScheduler: Submitting 20 missing tasks from ResultStage 1 (ShuffledRDD[2] at start at SparkTweetStreamingHDFSLoad.java:1743)
16/05/17 16:18:15 INFO cluster.YarnClusterScheduler: Adding task set 1.0 with 20 tasks
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 1.0 (TID 50, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 1.0 in stage 1.0 (TID 51, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on node4:58845 (size: 1400.0 B, free: 530.0 MB)
16/05/17 16:18:15 INFO storage.BlockManagerInfo: Added broadcast_1_piece0 in memory on node4:47743 (size: 1400.0 B, free: 530.0 MB)
16/05/17 16:18:15 INFO spark.MapOutputTrackerMasterEndpoint: Asked to send map output locations for shuffle 0 to node4:50089
16/05/17 16:18:15 INFO spark.MapOutputTrackerMaster: Size of output statuses for shuffle 0 is 295 bytes
16/05/17 16:18:15 INFO spark.MapOutputTrackerMasterEndpoint: Asked to send map output locations for shuffle 0 to node4:46484
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 2.0 in stage 1.0 (TID 52, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 3.0 in stage 1.0 (TID 53, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 1.0 in stage 1.0 (TID 51) in 454 ms on node4 (1/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 1.0 (TID 50) in 457 ms on node4 (2/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 4.0 in stage 1.0 (TID 54, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 2.0 in stage 1.0 (TID 52) in 69 ms on node4 (3/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 5.0 in stage 1.0 (TID 55, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 3.0 in stage 1.0 (TID 53) in 86 ms on node4 (4/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 6.0 in stage 1.0 (TID 56, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 4.0 in stage 1.0 (TID 54) in 66 ms on node4 (5/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 7.0 in stage 1.0 (TID 57, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 5.0 in stage 1.0 (TID 55) in 55 ms on node4 (6/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 8.0 in stage 1.0 (TID 58, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 6.0 in stage 1.0 (TID 56) in 77 ms on node4 (7/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 9.0 in stage 1.0 (TID 59, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 7.0 in stage 1.0 (TID 57) in 87 ms on node4 (8/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 10.0 in stage 1.0 (TID 60, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 8.0 in stage 1.0 (TID 58) in 49 ms on node4 (9/20)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Starting task 11.0 in stage 1.0 (TID 61, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:15 INFO scheduler.TaskSetManager: Finished task 9.0 in stage 1.0 (TID 59) in 58 ms on node4 (10/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 12.0 in stage 1.0 (TID 62, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 11.0 in stage 1.0 (TID 61) in 79 ms on node4 (11/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 13.0 in stage 1.0 (TID 63, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 10.0 in stage 1.0 (TID 60) in 107 ms on node4 (12/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 14.0 in stage 1.0 (TID 64, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 12.0 in stage 1.0 (TID 62) in 49 ms on node4 (13/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 15.0 in stage 1.0 (TID 65, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 13.0 in stage 1.0 (TID 63) in 64 ms on node4 (14/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 16.0 in stage 1.0 (TID 66, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 17.0 in stage 1.0 (TID 67, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 15.0 in stage 1.0 (TID 65) in 51 ms on node4 (15/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 14.0 in stage 1.0 (TID 64) in 86 ms on node4 (16/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 18.0 in stage 1.0 (TID 68, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 16.0 in stage 1.0 (TID 66) in 52 ms on node4 (17/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 19.0 in stage 1.0 (TID 69, node4, PROCESS_LOCAL, 1901 bytes)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 17.0 in stage 1.0 (TID 67) in 53 ms on node4 (18/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 19.0 in stage 1.0 (TID 69) in 40 ms on node4 (19/20)
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Finished task 18.0 in stage 1.0 (TID 68) in 67 ms on node4 (20/20)
16/05/17 16:18:16 INFO cluster.YarnClusterScheduler: Removed TaskSet 1.0, whose tasks have all completed, from pool 
16/05/17 16:18:16 INFO scheduler.DAGScheduler: ResultStage 1 (start at SparkTweetStreamingHDFSLoad.java:1743) finished in 1.010 s
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Job 0 finished: start at SparkTweetStreamingHDFSLoad.java:1743, took 8.825568 s
16/05/17 16:18:16 INFO scheduler.ReceiverTracker: Starting 1 receivers
16/05/17 16:18:16 INFO scheduler.ReceiverTracker: ReceiverTracker started
16/05/17 16:18:16 INFO dstream.ForEachDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO dstream.FilteredDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO dstream.MappedDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Initialized and validated org.apache.spark.streaming.twitter.TwitterInputDStream@55861179
16/05/17 16:18:16 INFO dstream.MappedDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO dstream.MappedDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO dstream.MappedDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO dstream.MappedDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO dstream.MappedDStream: Initialized and validated org.apache.spark.streaming.dstream.MappedDStream@6e42c819
16/05/17 16:18:16 INFO dstream.FilteredDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO dstream.FilteredDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO dstream.FilteredDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO dstream.FilteredDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO dstream.FilteredDStream: Initialized and validated org.apache.spark.streaming.dstream.FilteredDStream@479cccce
16/05/17 16:18:16 INFO dstream.ForEachDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO dstream.ForEachDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO dstream.ForEachDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO dstream.ForEachDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO dstream.ForEachDStream: Initialized and validated org.apache.spark.streaming.dstream.ForEachDStream@667afcd2
16/05/17 16:18:16 INFO dstream.ForEachDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO dstream.FilteredDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO dstream.MappedDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: metadataCleanupDelay = -1
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO twitter.TwitterInputDStream: Initialized and validated org.apache.spark.streaming.twitter.TwitterInputDStream@55861179
16/05/17 16:18:16 INFO dstream.MappedDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO dstream.MappedDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO dstream.MappedDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO dstream.MappedDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO dstream.MappedDStream: Initialized and validated org.apache.spark.streaming.dstream.MappedDStream@39234bd
16/05/17 16:18:16 INFO dstream.FilteredDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO dstream.FilteredDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO dstream.FilteredDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO dstream.FilteredDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO dstream.FilteredDStream: Initialized and validated org.apache.spark.streaming.dstream.FilteredDStream@7b6836d6
16/05/17 16:18:16 INFO dstream.ForEachDStream: Slide time = 60000 ms
16/05/17 16:18:16 INFO dstream.ForEachDStream: Storage level = StorageLevel(false, false, false, false, 1)
16/05/17 16:18:16 INFO dstream.ForEachDStream: Checkpoint interval = null
16/05/17 16:18:16 INFO dstream.ForEachDStream: Remember duration = 60000 ms
16/05/17 16:18:16 INFO dstream.ForEachDStream: Initialized and validated org.apache.spark.streaming.dstream.ForEachDStream@5ab36fc9
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Got job 1 (start at SparkTweetStreamingHDFSLoad.java:1743) with 1 output partitions
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Final stage: ResultStage 2(start at SparkTweetStreamingHDFSLoad.java:1743)
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Parents of final stage: List()
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Missing parents: List()
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Submitting ResultStage 2 (Receiver 0 ParallelCollectionRDD[3] at makeRDD at ReceiverTracker.scala:556), which has no missing parents
16/05/17 16:18:16 INFO scheduler.ReceiverTracker: Receiver 0 started
16/05/17 16:18:16 INFO storage.MemoryStore: ensureFreeSpace(62448) called with curMem=8135, maxMem=2061647216
16/05/17 16:18:16 INFO storage.MemoryStore: Block broadcast_2 stored as values in memory (estimated size 61.0 KB, free 1966.1 MB)
16/05/17 16:18:16 INFO storage.MemoryStore: ensureFreeSpace(21083) called with curMem=70583, maxMem=2061647216
16/05/17 16:18:16 INFO storage.MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 20.6 KB, free 1966.1 MB)
16/05/17 16:18:16 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on 172.16.28.195:57488 (size: 20.6 KB, free: 1966.1 MB)
16/05/17 16:18:16 INFO spark.SparkContext: Created broadcast 2 from broadcast at DAGScheduler.scala:861
16/05/17 16:18:16 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 2 (Receiver 0 ParallelCollectionRDD[3] at makeRDD at ReceiverTracker.scala:556)
16/05/17 16:18:16 INFO cluster.YarnClusterScheduler: Adding task set 2.0 with 1 tasks
16/05/17 16:18:16 INFO util.RecurringTimer: Started timer for JobGenerator at time 1463482140000
16/05/17 16:18:16 INFO scheduler.JobGenerator: Started JobGenerator at 1463482140000 ms
16/05/17 16:18:16 INFO scheduler.JobScheduler: Started JobScheduler
16/05/17 16:18:16 INFO streaming.StreamingContext: StreamingContext started
16/05/17 16:18:16 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 2.0 (TID 70, node4, NODE_LOCAL, 3094 bytes)
16/05/17 16:18:17 INFO impl.StdSchedulerFactory: Using default implementation for ThreadExecutor
16/05/17 16:18:17 INFO simpl.SimpleThreadPool: Job execution threads will use class loader of thread: Driver
16/05/17 16:18:17 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on node4:58845 (size: 20.6 KB, free: 530.0 MB)
16/05/17 16:18:17 INFO core.SchedulerSignalerImpl: Initialized Scheduler Signaller of type: class org.quartz.core.SchedulerSignalerImpl
16/05/17 16:18:17 INFO core.QuartzScheduler: Quartz Scheduler v.1.8.6 created.
16/05/17 16:18:17 INFO simpl.RAMJobStore: RAMJobStore initialized.
16/05/17 16:18:17 INFO core.QuartzScheduler: Scheduler meta-data: Quartz Scheduler (v1.8.6) 'DefaultQuartzScheduler' with instanceId 'NON_CLUSTERED'
  Scheduler class: 'org.quartz.core.QuartzScheduler' - running locally.
  NOT STARTED.
  Currently in standby mode.
  Number of jobs executed: 0
  Using thread pool 'org.quartz.simpl.SimpleThreadPool' - with 10 threads.
  Using job-store 'org.quartz.simpl.RAMJobStore' - which does not support persistence. and is not clustered.

16/05/17 16:18:17 INFO impl.StdSchedulerFactory: Quartz scheduler 'DefaultQuartzScheduler' initialized from default resource file in Quartz package: 'quartz.properties'
16/05/17 16:18:17 INFO impl.StdSchedulerFactory: Quartz scheduler version: 1.8.6
16/05/17 16:18:17 INFO core.QuartzScheduler: Scheduler DefaultQuartzScheduler_$_NON_CLUSTERED started.
16/05/17 16:18:17 INFO spark.SparkTweetStreamingHDFSLoad: END {}TwitterTweets
16/05/17 16:18:17 INFO yarn.ApplicationMaster: Final app status: SUCCEEDED, exitCode: 0
16/05/17 16:18:17 INFO streaming.StreamingContext: Invoking stop(stopGracefully=false) from shutdown hook
16/05/17 16:18:17 INFO scheduler.ReceiverTracker: Sent stop signal to all 1 receivers
16/05/17 16:18:17 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 2.0 (TID 70) in 718 ms on node4 (1/1)
16/05/17 16:18:17 INFO scheduler.DAGScheduler: ResultStage 2 (start at SparkTweetStreamingHDFSLoad.java:1743) finished in 0.717 s
16/05/17 16:18:17 INFO cluster.YarnClusterScheduler: Removed TaskSet 2.0, whose tasks have all completed, from pool 
16/05/17 16:18:17 INFO scheduler.ReceiverTracker: All of the receivers have deregistered successfully
16/05/17 16:18:17 INFO scheduler.ReceiverTracker: ReceiverTracker stopped
16/05/17 16:18:17 INFO scheduler.JobGenerator: Stopping JobGenerator immediately
16/05/17 16:18:17 INFO util.RecurringTimer: Stopped timer for JobGenerator after time -1
16/05/17 16:18:17 INFO scheduler.JobGenerator: Stopped JobGenerator
16/05/17 16:18:17 INFO scheduler.JobScheduler: Stopped JobScheduler
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming/batch,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/static/streaming,null}
16/05/17 16:18:17 INFO streaming.StreamingContext: StreamingContext stopped successfully
16/05/17 16:18:17 INFO spark.SparkContext: Invoking stop() from shutdown hook
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming/batch/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/streaming/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/metrics/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/api,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/static,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/json,null}
16/05/17 16:18:17 INFO handler.ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs,null}
16/05/17 16:18:17 INFO ui.SparkUI: Stopped Spark web UI at http://172.16.28.195:59320
16/05/17 16:18:17 INFO scheduler.DAGScheduler: Stopping DAGScheduler
16/05/17 16:18:17 INFO cluster.YarnClusterSchedulerBackend: Shutting down all executors
16/05/17 16:18:17 INFO cluster.YarnClusterSchedulerBackend: Asking each executor to shut down
16/05/17 16:18:17 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:50089
16/05/17 16:18:17 INFO yarn.ApplicationMaster$AMEndpoint: Driver terminated or disconnected! Shutting down. node4:46484
16/05/17 16:18:18 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
16/05/17 16:18:18 INFO storage.MemoryStore: MemoryStore cleared
16/05/17 16:18:18 INFO storage.BlockManager: BlockManager stopped
16/05/17 16:18:18 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
16/05/17 16:18:18 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
16/05/17 16:18:18 INFO spark.SparkContext: Successfully stopped SparkContext
16/05/17 16:18:18 INFO remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon.
16/05/17 16:18:18 INFO yarn.ApplicationMaster: Unregistering ApplicationMaster with SUCCEEDED
16/05/17 16:18:18 INFO remote.RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports.
16/05/17 16:18:18 INFO impl.AMRMClientImpl: Waiting for application to be successfully unregistered.
16/05/17 16:18:18 INFO yarn.ApplicationMaster: Deleting staging directory .sparkStaging/application_1463479181441_0003
16/05/17 16:18:19 INFO util.ShutdownHookManager: Shutdown hook called
16/05/17 16:18:19 INFO util.ShutdownHookManager: Deleting directory /tmp/hadoop-hadoop/nm-local-dir/usercache/hadoop/appcache/application_1463479181441_0003/spark-5b36342a-6212-4cea-80da-b1961cab161c
 

Sent from Yahoo Mail. Get the app
Mime
View raw message