OutputFormat抽象类描述输出数据的格式。getRecordWriter获取RecordWriter,RecordWriter的write方法将key/value写入文件。checkOutputSpecs验证输出目录或者其他的有效性。getOutputCommitter是OutputFormat中一些临时目录、结果目录的创建、移动、删除等操作。
package org.apache.hadoop.mapreduce;import java.io.IOException;import org.apache.hadoop.classification.InterfaceAudience;import org.apache.hadoop.classification.InterfaceStability;import org.apache.hadoop.fs.FileSystem;/** * OutputFormat describes the output-specification for a * Map-Reduce job. * * The Map-Reduce framework relies on the OutputFormat of the * job to: * * - * Validate the output-specification of the job. For e.g. check that the * output directory doesn't already exist. *
- * Provide the {@link RecordWriter} implementation to be used to write out * the output files of the job. Output files are stored in a * {@link FileSystem}. *
* * * @see RecordWriter */@InterfaceAudience.Public@InterfaceStability.Stablepublic abstract class OutputFormat { /** * Get the {@link RecordWriter} for the given task. * * @param context the information about the current task. * @return a {@link RecordWriter} to write the output for the job. * @throws IOException */ public abstract RecordWriter getRecordWriter(TaskAttemptContext context ) throws IOException, InterruptedException; /** * Check for validity of the output-specification for the job. * * This is to validate the output specification for the job when it is * a job is submitted. Typically checks that it does not already exist, * throwing an exception when it already exists, so that output is not * overwritten. * * @param context information about the job * @throws IOException when output should not be attempted */ public abstract void checkOutputSpecs(JobContext context ) throws IOException, InterruptedException; /** * Get the output committer for this output format. This is responsible * for ensuring the output is committed correctly. * @param context the task context * @return an output committer * @throws IOException * @throws InterruptedException */ public abstract OutputCommitter getOutputCommitter(TaskAttemptContext context ) throws IOException, InterruptedException;} |
OutputFormat的子类
RecordWriter抽象类
package org.apache.hadoop.mapreduce;import java.io.IOException;import org.apache.hadoop.classification.InterfaceAudience;import org.apache.hadoop.classification.InterfaceStability;import org.apache.hadoop.fs.FileSystem;/** * RecordWriter writes the output <key, value> pairs * to an output file. * RecordWriter implementations write the job outputs to the * {@link FileSystem}. * * @see OutputFormat */@InterfaceAudience.Public@InterfaceStability.Stablepublic abstract class RecordWriter { /** * Writes a key/value pair. * * @param key the key to write. * @param value the value to write. * @throws IOException */ public abstract void write(K key, V value ) throws IOException, InterruptedException; /** * Close this RecordWriter to future operations. * * @param context the context of the task * @throws IOException */ public abstract void close(TaskAttemptContext context ) throws IOException, InterruptedException;} |
OutputCommitter抽象类
package org.apache.hadoop.mapreduce;import java.io.IOException;import org.apache.hadoop.classification.InterfaceAudience;import org.apache.hadoop.classification.InterfaceStability;/** * OutputCommitter describes the commit of task output for a * Map-Reduce job. * * The Map-Reduce framework relies on the OutputCommitter of * the job to: * * - * Setup the job during initialization. For example, create the temporary * output directory for the job during the initialization of the job. *
* - * Cleanup the job after the job completion. For example, remove the * temporary output directory after the job completion. *
* - * Setup the task temporary output. *
* - * Check whether a task needs a commit. This is to avoid the commit * procedure if a task does not need commit. *
* - * Commit of the task output. *
* - * Discard the task commit. *
* * The methods in this class can be called from several different processes and * from several different contexts. It is important to know which process and * which context each is called from. Each method should be marked accordingly * in its documentation. It is also important to note that not all methods are * guaranteed to be called once and only once. If a method is not guaranteed to * have this property the output committer needs to handle this appropriately. * Also note it will only be in rare situations where they may be called * multiple times for the same task. * * @see org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter * @see JobContext * @see TaskAttemptContext */@InterfaceAudience.Public@InterfaceStability.Stablepublic abstract class OutputCommitter { /** * For the framework to setup the job output during initialization. This is * called from the application master process for the entire job. This will be * called multiple times, once per job attempt. * * @param jobContext Context of the job whose output is being written. * @throws IOException if temporary output could not be created */ public abstract void setupJob(JobContext jobContext) throws IOException; /** * For cleaning up the job's output after job completion. This is called * from the application master process for the entire job. This may be called * multiple times. * * @param jobContext Context of the job whose output is being written. * @throws IOException * @deprecated Use {@link #commitJob(JobContext)} and * {@link #abortJob(JobContext, JobStatus.State)} instead. */ @Deprecated public void cleanupJob(JobContext jobContext) throws IOException { } /** * For committing job's output after successful job completion. Note that this * is invoked for jobs with final runstate as SUCCESSFUL. This is called * from the application master process for the entire job. This is guaranteed * to only be called once. If it throws an exception the entire job will * fail. * * @param jobContext Context of the job whose output is being written. * @throws IOException */ public void commitJob(JobContext jobContext) throws IOException { cleanupJob(jobContext); } /** * For aborting an unsuccessful job's output. Note that this is invoked for * jobs with final runstate as {@link JobStatus.State#FAILED} or * {@link JobStatus.State#KILLED}. This is called from the application * master process for the entire job. This may be called multiple times. * * @param jobContext Context of the job whose output is being written. * @param state final runstate of the job * @throws IOException */ public void abortJob(JobContext jobContext, JobStatus.State state) throws IOException { cleanupJob(jobContext); } /** * Sets up output for the task. This is called from each individual task's * process that will output to HDFS, and it is called just for that task. This * may be called multiple times for the same task, but for different task * attempts. * * @param taskContext Context of the task whose output is being written. * @throws IOException */ public abstract void setupTask(TaskAttemptContext taskContext) throws IOException; /** * Check whether task needs a commit. This is called from each individual * task's process that will output to HDFS, and it is called just for that * task. * * @param taskContext * @return true/false * @throws IOException */ public abstract boolean needsTaskCommit(TaskAttemptContext taskContext) throws IOException; /** * To promote the task's temporary output to final output location. * If {@link #needsTaskCommit(TaskAttemptContext)} returns true and this * task is the task that the AM determines finished first, this method * is called to commit an individual task's output. This is to mark * that tasks output as complete, as {@link #commitJob(JobContext)} will * also be called later on if the entire job finished successfully. This * is called from a task's process. This may be called multiple times for the * same task, but different task attempts. It should be very rare for this to * be called multiple times and requires odd networking failures to make this * happen. In the future the Hadoop framework may eliminate this race. * * @param taskContext Context of the task whose output is being written. * @throws IOException if commit is not successful. */ public abstract void commitTask(TaskAttemptContext taskContext) throws IOException; /** * Discard the task output. This is called from a task's process to clean * up a single task's output that can not yet been committed. This may be * called multiple times for the same task, but for different task attempts. * * @param taskContext * @throws IOException */ public abstract void abortTask(TaskAttemptContext taskContext) throws IOException; /** * Is task output recovery supported for restarting jobs? * * If task output recovery is supported, job restart can be done more * efficiently. * * @return true if task output recovery is supported, * false otherwise * @see #recoverTask(TaskAttemptContext) * @deprecated Use {@link #isRecoverySupported(JobContext)} instead. */ @Deprecated public boolean isRecoverySupported() { return false; } /** * Is task output recovery supported for restarting jobs? * * If task output recovery is supported, job restart can be done more * efficiently. * * @param jobContext * Context of the job whose output is being written. * @return true if task output recovery is supported, * false otherwise * @throws IOException * @see #recoverTask(TaskAttemptContext) */ public boolean isRecoverySupported(JobContext jobContext) throws IOException { return isRecoverySupported(); } /** * Recover the task output. * * The retry-count for the job will be passed via the * {@link MRJobConfig#APPLICATION_ATTEMPT_ID} key in * {@link TaskAttemptContext#getConfiguration()} for the * OutputCommitter . This is called from the application master * process, but it is called individually for each task. * * If an exception is thrown the task will be attempted again. * * This may be called multiple times for the same task. But from different * application attempts. * * @param taskContext Context of the task whose output is being recovered * @throws IOException */ public void recoverTask(TaskAttemptContext taskContext) throws IOException {}} |