mahout-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From WangRamon <ramon_w...@hotmail.com>
Subject RE: Cancel running distributed RecommenderJob
Date Tue, 03 Apr 2012 01:08:44 GMT

I think you also need to stop/kill the process which submit the RecommenderJob to hadoop too.
RegardsRamon

> Date: Mon, 2 Apr 2012 19:05:27 +0100
> Subject: Re: Cancel running distributed RecommenderJob
> From: srowen@gmail.com
> To: user@mahout.apache.org
> 
> You can use the Hadoop interface itself (like, the command-line hadoop
> tool) to kill a job by its ID. If you kill one MapReduce job the
> entire process should halt after that.
> 
> On Mon, Apr 2, 2012 at 6:44 PM, Sören Brunk <soren.brunk@deri.org> wrote:
> > Hi,
> >
> > I'm using the distributed RecommenderJob from within a Java program.
> > For that, in a separate thread, I'm creating a RecommenderJob object, call
> > setConf() for the hadoop configuration and then run() with the job
> > parameters.
> > This is working fine for me but now I would like to be able to stop a
> > running job.
> > Not sure if that's possible at all since RecommenderJob encapsulates several
> > Hadoop jobs (or even other Mahout jobs that call Hadoop in turn) and runs
> > them in a blocking way.
> >
> > Would be interesting for other Mahout jobs as well.
> > Any ideas?
> >
> > Thanks,
> >
> > --
> > Sören Brunk
> > Research Assistant
> > Data Intensive Infrastructures Unit (DI2)
> > Digital Enterprise Research Institute
> > National University of Ireland Galway
> >
 		 	   		  
Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message