Process process1 = runtime.exec (cmdToBeExec); 
  InputStream is = process1.getInputStream ();  
 InputStreamReader isr = new InputStreamReader (is); 
 BufferedReader br = new BufferedReader (isr); 
 while ((line = br.readLine ())! = null) { 
 System.out.println ("Output is:" + line); 
} 
 I wrote such a program, and then configure the workflow, the mahout command line execution of an algorithm executed as a variable transmission in the past, 
 this is not enforced. Later plus error output process1.geterrorstream (); seemed to begin execution of the 
 But I jobtracker point of view, has been stuck in a no go down, log in no other errors, is the heartbeat detection 
 This is how it happened? 
 ------ Solution ---------------------------------------- ----  
 ?This post last edited by the tntzbzc on 2013-05-26 14:57:35  job track and task track are not being given the Log one o'clock? 
 with HADOOP JAR run mahout can go?  
 ------ Solution ------------------------------------ -------- 
 mahout in multiple iterations JOB what way? 
 ------ Solution ---------------------------------------- ---- 
 java call linux, execute shell commands, there have been problems. Failed to get running.  
 I write directly jni call to solve.  
 c / python no problem 
 ------ For reference only ---------------------------- -----------   
 at the command line to run no problem, use this procedure when performed on the card 
 no other error log is the heartbeat 
 I delegated this directive continues, in fact, a mahout in a job with a job that 
 will not affect it because of this? After all, you really want to prepare this directive began to run 
 ------ For reference only ----------------------------- ----------   
 estimate is not the reason mahout, 
 I used a bit oozie + sqooop is the same error, map0%  
 ------ For reference only ---------------------------------- -----   
 Now the situation is like this, as long as I submit job with oozie 
 before performing any mapreduce program will be blocked 
 When I manually kill off the job when submitted oozie 
 rest mapreduce program will continue down the implementation of the  
 I find it strange that it can not find a mistake 
 ------ For reference only ----------------------- ----------------  
 ??
 Now the situation is like this, as long as I submit job with oozie ??
 before performing any mapreduce program will be blocked ??
 When I manually kill off the job when submitted oozie ??
 rest mapreduce program will continue down the implementation of the ??
 ??
 I find it strange that it can not find a mistake ?    

 so magical?  
Hard to say what causes the problem.
回复删除