samza-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Chris Riccomini <criccom...@apache.org>
Subject Re: Samza on Yarn
Date Thu, 12 Mar 2015 16:36:35 GMT
Hey Shekar,

You need that samza-yarn file on your RM/NM's classpath, along with scala.
We missed this in the docs, and are tracking the issue here:

  https://issues.apache.org/jira/browse/SAMZA-456

You'll also need samza-core in the classpath, based on the discussion on
SAMZA-456. Sorry about that. If you want to update the tutorial when you
get your cluster working, and submit a patch, that'd be great! :)

Cheers,
Chris

On Wed, Mar 11, 2015 at 9:43 PM, Shekar Tippur <ctippur@gmail.com> wrote:

> Here is the corresponding log:
>
> 2015-03-11 20:43:09,665 INFO  [AsyncDispatcher event handler]
> localizer.LocalizedResource (LocalizedResource.java:handle(203)) - Resource
> http://sprfargas102:8000/hello-samza-0.8.0-dist.tar.gz transitioned from
> INIT to DOWNLOADING
>
> 2015-03-11 20:43:09,665 INFO  [AsyncDispatcher event handler]
> localizer.ResourceLocalizationService
> (ResourceLocalizationService.java:handle(679)) - Created localizer for
> container_1426121400423_2587_01_000001
>
> 2015-03-11 20:43:09,669 INFO  [LocalizerRunner for
> container_1426121400423_2587_01_000001]
> localizer.ResourceLocalizationService
> (ResourceLocalizationService.java:writeCredentials(1107)) - Writing
> credentials to the nmPrivate file
>
> /tmp/hadoop-hadoop/nm-local-dir/nmPrivate/container_1426121400423_2587_01_000001.tokens.
> Credentials list:
>
> 2015-03-11 20:43:09,675 INFO  [DeletionService #0]
> nodemanager.DefaultContainerExecutor
> (DefaultContainerExecutor.java:deleteAsUser(378)) - Deleting path :
> /home/hadoop/hadoop-2.5.2/logs/userlogs/application_1426120927668_0010
>
> 2015-03-11 20:43:09,676 INFO  [LocalizerRunner for
> container_1426121400423_2587_01_000001]
> nodemanager.DefaultContainerExecutor
> (DefaultContainerExecutor.java:createUserCacheDirs(469)) - Initializing
> user root
>
> 2015-03-11 20:43:09,685 INFO  [LocalizerRunner for
> container_1426121400423_2587_01_000001]
> nodemanager.DefaultContainerExecutor
> (DefaultContainerExecutor.java:startLocalizer(103)) - Copying from
>
> /tmp/hadoop-hadoop/nm-local-dir/nmPrivate/container_1426121400423_2587_01_000001.tokens
> to
>
> /tmp/hadoop-hadoop/nm-local-dir/usercache/root/appcache/application_1426121400423_2587/container_1426121400423_2587_01_000001.tokens
>
> *2015-03-11 20:43:09,685 INFO  [LocalizerRunner for
> container_1426121400423_2587_01_000001]
> nodemanager.DefaultContainerExecutor
> (DefaultContainerExecutor.java:startLocalizer(105)) - CWD set to
>
> /tmp/hadoop-hadoop/nm-local-dir/usercache/root/appcache/application_1426121400423_2587
> =
>
> file:/tmp/hadoop-hadoop/nm-local-dir/usercache/root/appcache/application_1426121400423_2587*
>
> *2015-03-11 20:43:09,716 INFO  [IPC Server handler 2 on 8040]
> localizer.ResourceLocalizationService
> (ResourceLocalizationService.java:update(1007)) - DEBUG: FAILED {
> http://sprfargas102:8000/hello-samza-0.8.0-dist.tar.gz
> <http://sprfargas102:8000/hello-samza-0.8.0-dist.tar.gz>, 0, ARCHIVE, null
> }, java.lang.ClassNotFoundException: Class
> org.apache.samza.util.hadoop.HttpFileSystem not found*
>
> *2015-03-11 20:43:09,716 INFO  [IPC Server handler 2 on 8040]
> localizer.LocalizedResource (LocalizedResource.java:handle(203)) - Resource
> http://sprfargas102:8000/hello-samza-0.8.0-dist.tar.gz(-
> <http://sprfargas102:8000/hello-samza-0.8.0-dist.tar.gz(-
> >>/tmp/hadoop-hadoop/nm-local-dir/usercache/root/appcache/application_1426121400423_2587/filecache/10/hello-samza-0.8.0-dist.tar.gz)
> transitioned from DOWNLOADING to FAILED*
>
> 2015-03-11 20:43:09,717 INFO  [AsyncDispatcher event handler]
> container.Container (ContainerImpl.java:handle(918)) - Container
> container_1426121400423_2587_01_000001 transitioned from LOCALIZING to
> LOCALIZATION_FAILED
>
> 2015-03-11 20:43:09,717 INFO  [AsyncDispatcher event handler]
> localizer.LocalResourcesTrackerImpl
> (LocalResourcesTrackerImpl.java:handle(151)) - Container
> container_1426121400423_2587_01_000001 sent RELEASE event on a resource
> request { http://sprfargas102:8000/hello-samza-0.8.0-dist.tar.gz, 0,
> ARCHIVE, null } not present in cache.
>
> 2015-03-11 20:43:09,717 WARN  [AsyncDispatcher event handler]
> nodemanager.NMAuditLogger (NMAuditLogger.java:logFailure(150)) -
> USER=root OPERATION=Container
> Finished - Failed TARGET=ContainerImpl RESULT=FAILURE DESCRIPTION=Container
> failed with state: LOCALIZATION_FAILED APPID=application_1426121400423_2587
> CONTAINERID=container_1426121400423_2587_01_000001
>
> 2015-03-11 20:43:09,717 INFO  [AsyncDispatcher event handler]
> container.Container (ContainerImpl.java:handle(918)) - Container
> container_1426121400423_2587_01_000001 transitioned from
> LOCALIZATION_FAILED to DONE
>
> 2015-03-11 20:43:09,717 INFO  [AsyncDispatcher event handler]
> application.Application (ApplicationImpl.java:transition(340)) - Removing
> container_1426121400423_2587_01_000001 from application
> application_1426121400423_2587
>
> 2015-03-11 20:43:09,717 INFO  [AsyncDispatcher event handler]
> containermanager.AuxServices (AuxServices.java:handle(196)) - Got event
> CONTAINER_STOP for appId application_1426121400423_2587
>
> 2015-03-11 20:43:09,717 INFO  [DeletionService #2]
> nodemanager.DefaultContainerExecutor
> (DefaultContainerExecutor.java:deleteAsUser(369)) - Deleting absolute path
> :
>
> /tmp/hadoop-hadoop/nm-local-dir/usercache/root/appcache/application_1426121400423_2587/container_1426121400423_2587_01_000001
>
> 2015-03-11 20:43:09,717 WARN  [DeletionService #2]
> nodemanager.DefaultContainerExecutor
> (DefaultContainerExecutor.java:deleteAsUser(372)) - delete returned false
> for path:
>
> [/tmp/hadoop-hadoop/nm-local-dir/usercache/root/appcache/application_1426121400423_2587/container_1426121400423_2587_01_000001]
>
> 2015-03-11 20:43:09,718 WARN  [LocalizerRunner for
> container_1426121400423_2587_01_000001] ipc.Client (Client.java:call(1389))
> - interrupted waiting to send rpc request to server
>
> java.lang.InterruptedException
>
> at java.util.concurrent.FutureTask.awaitDone(FutureTask.java:400)
>
> at java.util.concurrent.FutureTask.get(FutureTask.java:187)
>
> at org.apache.hadoop.ipc.Client$Connection.sendRpcRequest(Client.java:1030)
>
> at org.apache.hadoop.ipc.Client.call(Client.java:1384)
>
> at org.apache.hadoop.ipc.Client.call(Client.java:1364)
>
> at
>
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>
> at com.sun.proxy.$Proxy29.heartbeat(Unknown Source)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.api.impl.pb.client.LocalizationProtocolPBClientImpl.heartbeat(LocalizationProtocolPBClientImpl.java:62)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.localizeFiles(ContainerLocalizer.java:255)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.runLocalization(ContainerLocalizer.java:169)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:107)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:1073)
>
> java.io.IOException: java.lang.InterruptedException
>
> at org.apache.hadoop.ipc.Client.call(Client.java:1390)
>
> at org.apache.hadoop.ipc.Client.call(Client.java:1364)
>
> at
>
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>
> at com.sun.proxy.$Proxy29.heartbeat(Unknown Source)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.api.impl.pb.client.LocalizationProtocolPBClientImpl.heartbeat(LocalizationProtocolPBClientImpl.java:62)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.localizeFiles(ContainerLocalizer.java:255)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.runLocalization(ContainerLocalizer.java:169)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:107)
>
> at
>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:1073)
>
> Caused by: java.lang.InterruptedException
>
> at java.util.concurrent.FutureTask.awaitDone(FutureTask.java:400)
>
> at java.util.concurrent.FutureTask.get(FutureTask.java:187)
>
> at org.apache.hadoop.ipc.Client$Connection.sendRpcRequest(Client.java:1030)
>
> at org.apache.hadoop.ipc.Client.call(Client.java:1384)
>
> ... 8 more
>
> On Wed, Mar 11, 2015 at 4:56 PM, Shekar Tippur <ctippur@gmail.com> wrote:
>
> > Hello,
> >
> > Sorry to reopen this topic. I had setup yarn couple of months ago and
> cant
> > seem to replicate this now.
> >
> > I see that I have done everything listed here
> >
> http://samza.apache.org/learn/tutorials/0.7.0/run-in-multi-node-yarn.html
> >
> > I see this error on the application side
> >
> > Application application_1426115467623_0492 failed 2 times due to AM
> > Container for appattempt_1426115467623_0492_000002 exited with exitCode:
> > -1000 due to: java.lang.ClassNotFoundException: Class
> > org.apache.samza.util.hadoop.HttpFileSystem not found
> > .Failing this attempt.. Failing the application.
> >
> > I see that
> >
> > /home/hadoop/hadoop-2.5.2/share/hadoop/hdfs/lib/samza-yarn_2.10-0.8.0.jar
> > has that particular class
> >
> >   1739 Tue Nov 25 10:51:40 PST 2014
> >
> org/apache/samza/util/hadoop/HttpFileSystem$$anonfun$getFileStatus$1.class
> >
> >   1570 Tue Nov 25 10:51:40 PST 2014
> > org/apache/samza/util/hadoop/HttpFileSystem$$anonfun$initialize$1.class
> >
> >   1597 Tue Nov 25 10:51:40 PST 2014
> > org/apache/samza/util/hadoop/HttpFileSystem$$anonfun$open$1.class
> >
> >   1797 Tue Nov 25 10:51:40 PST 2014
> > org/apache/samza/util/hadoop/HttpFileSystem$$anonfun$open$2.class
> >
> >   9549 Tue Nov 25 10:51:40 PST 2014
> > org/apache/samza/util/hadoop/HttpFileSystem.class
> >
> >
> > I see that env is set right:
> >
> >
> > HADOOP_YARN_HOME=/home/hadoop/hadoop-2.5.2
> >
> > HADOOP_CONF_DIR=/home/hadoop/hadoop-2.5.2/conf
> >
> > Wondering if I am missing anything...
> > - Shekar
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message