hadoop PipesNonJavaInputFormat 源码
haddop PipesNonJavaInputFormat 代码
文件路径:/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/pipes/PipesNonJavaInputFormat.java
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.mapred.pipes;
import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.io.FloatWritable;
import org.apache.hadoop.io.NullWritable;
import org.apache.hadoop.mapred.InputFormat;
import org.apache.hadoop.mapred.InputSplit;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.RecordReader;
import org.apache.hadoop.mapred.Reporter;
import org.apache.hadoop.mapred.TextInputFormat;
import org.apache.hadoop.util.ReflectionUtils;
/**
* Dummy input format used when non-Java a {@link RecordReader} is used by
* the Pipes' application.
*
* The only useful thing this does is set up the Map-Reduce job to get the
* {@link PipesDummyRecordReader}, everything else left for the 'actual'
* InputFormat specified by the user which is given by
* <i>mapreduce.pipes.inputformat</i>.
*/
class PipesNonJavaInputFormat
implements InputFormat<FloatWritable, NullWritable> {
public RecordReader<FloatWritable, NullWritable> getRecordReader(
InputSplit genericSplit, JobConf job, Reporter reporter)
throws IOException {
return new PipesDummyRecordReader(job, genericSplit);
}
public InputSplit[] getSplits(JobConf job, int numSplits) throws IOException {
// Delegate the generation of input splits to the 'original' InputFormat
return ReflectionUtils.newInstance(
job.getClass(Submitter.INPUT_FORMAT,
TextInputFormat.class,
InputFormat.class), job).getSplits(job, numSplits);
}
/**
* A dummy {@link org.apache.hadoop.mapred.RecordReader} to help track the
* progress of Hadoop Pipes' applications when they are using a non-Java
* <code>RecordReader</code>.
*
* The <code>PipesDummyRecordReader</code> is informed of the 'progress' of
* the task by the {@link OutputHandler#progress(float)} which calls the
* {@link #next(FloatWritable, NullWritable)} with the progress as the
* <code>key</code>.
*/
static class PipesDummyRecordReader implements RecordReader<FloatWritable, NullWritable> {
float progress = 0.0f;
public PipesDummyRecordReader(Configuration job, InputSplit split)
throws IOException{
}
public FloatWritable createKey() {
return null;
}
public NullWritable createValue() {
return null;
}
public synchronized void close() throws IOException {}
public synchronized long getPos() throws IOException {
return 0;
}
public float getProgress() {
return progress;
}
public synchronized boolean next(FloatWritable key, NullWritable value)
throws IOException {
progress = key.get();
return true;
}
}
}
相关信息
相关文章
0
赞
热门推荐
-
2、 - 优质文章
-
3、 gate.io
-
8、 golang
-
9、 openharmony
-
10、 Vue中input框自动聚焦