trssc 发表于 2018-10-31 10:50:25

Attempt time threshold of job running in hadoop-maxwell

  Attempt time threshold is a good feature to avoid time out or too slowly running of the job. However, you may need to run some scripts by forking a new sub-process sometimes. You may not get the progress of the process. In hadoop, if you do not report the progress during a time interval (defalut as 600s), it will consider the attemp is timeout, kill the current attemp process, and fork a new attemp.
  In streaming way, you can set the mapred.task.timeout as much as you wish in command line running your job, e.g.,
  hadoop jar test.jar -jobconf mapred.task.timeout=3600000
  To develop a jar, you can set it like this:
  long milliSeconds = 1000*60*60;
  conf.setLong("mapred.task.timeout", milliSeconds);
  NOTICE: The unit of timeout property is millisecond
  Or, if you could, you need to report the progress to the job to aviod time out, e.g.,
  context.progress()

页: [1]
查看完整版本: Attempt time threshold of job running in hadoop-maxwell