213214215216217218219220221222223
if (ctxCreated) { ctx.clear(); } if (rj != null) { if (returnVal != 0) { rj.killJob(); } HadoopJobExecHelper.runningJobKillURIs.remove(rj.getJobID()); jobID = rj.getID().toString(); } ColumnTruncateMapper.jobClose(outputPath, success, job, console,
244245246247248249250251252253254
if (ctxCreated) { ctx.clear(); } if (rj != null) { if (returnVal != 0) { rj.killJob(); } HadoopJobExecHelper.runningJobKillURIs.remove(rj.getJobID()); jobID = rj.getID().toString(); } } catch (Exception e) {
237238239240241242243244245246247
if (ctxCreated) { ctx.clear(); } if (rj != null) { if (returnVal != 0) { rj.killJob(); } HadoopJobExecHelper.runningJobKillURIs.remove(rj.getJobID()); jobID = rj.getID().toString(); } RCFileMergeMapper.jobClose(outputPath, success, job, console,
140141142143144145146147148149150
running = jc.getJob(jobId); } sucess = running.isSuccessful(); } finally { if (!sucess && (running != null)) { running.killJob(); } jc.close(); } return sucess; }
254255256257258259260261262263264
this.state = Job.SUCCESS; } else { this.state = Job.FAILED; this.message = "Job failed!"; try { running.killJob(); } catch (IOException e1) { } try { this.jc.close();
271272273274275276277278279280281
} catch (IOException ioe) { this.state = Job.FAILED; this.message = StringUtils.stringifyException(ioe); try { if (running != null) running.killJob(); } catch (IOException e1) { } try { this.jc.close();
564565566567568569570571572573574
RunningJob job = jc.getJob(id); if (job == null) System.out.println("Job with id " + jobid + " is not active"); else { job.killJob(); log.info("Kill " + id + " submitted."); } } }
286287288289290291292293294295296
303304305306307308309310311312313
642643644645646647648649650651652