org.apache.hadoop.mapred.LocalJobRunner Java Examples

The following examples show how to use org.apache.hadoop.mapred.LocalJobRunner. You can vote up the ones you like or vote down the ones you don't like, and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: TestLocalRunner.java    From hadoop with Apache License 2.0 6 votes vote down vote up
/**
 * Run a test with a misconfigured number of mappers.
 * Expect failure.
 */
@Test
public void testInvalidMultiMapParallelism() throws Exception {
  Job job = Job.getInstance();

  Path inputPath = createMultiMapsInput();
  Path outputPath = getOutputPath();

  Configuration conf = new Configuration();
  FileSystem fs = FileSystem.getLocal(conf);

  if (fs.exists(outputPath)) {
    fs.delete(outputPath, true);
  }

  job.setMapperClass(StressMapper.class);
  job.setReducerClass(CountingReducer.class);
  job.setNumReduceTasks(1);
  LocalJobRunner.setLocalMaxRunningMaps(job, -6);
  FileInputFormat.addInputPath(job, inputPath);
  FileOutputFormat.setOutputPath(job, outputPath);

  boolean success = job.waitForCompletion(true);
  assertFalse("Job succeeded somehow", success);
}
 
Example #2
Source File: TestLocalRunner.java    From big-c with Apache License 2.0 6 votes vote down vote up
/**
 * Run a test with a misconfigured number of mappers.
 * Expect failure.
 */
@Test
public void testInvalidMultiMapParallelism() throws Exception {
  Job job = Job.getInstance();

  Path inputPath = createMultiMapsInput();
  Path outputPath = getOutputPath();

  Configuration conf = new Configuration();
  FileSystem fs = FileSystem.getLocal(conf);

  if (fs.exists(outputPath)) {
    fs.delete(outputPath, true);
  }

  job.setMapperClass(StressMapper.class);
  job.setReducerClass(CountingReducer.class);
  job.setNumReduceTasks(1);
  LocalJobRunner.setLocalMaxRunningMaps(job, -6);
  FileInputFormat.addInputPath(job, inputPath);
  FileOutputFormat.setOutputPath(job, outputPath);

  boolean success = job.waitForCompletion(true);
  assertFalse("Job succeeded somehow", success);
}
 
Example #3
Source File: TestCompactCommandCluster.java    From kite with Apache License 2.0 6 votes vote down vote up
private boolean setLocalReducerMax(Configuration conf, int max) {
  try {
    Job job = Hadoop.Job.newInstance.invoke(new Configuration(false));
    DynMethods.StaticMethod setReducerMax = new DynMethods
        .Builder("setLocalMaxRunningReduces")
        .impl(LocalJobRunner.class,
            org.apache.hadoop.mapreduce.JobContext.class, Integer.TYPE)
        .buildStaticChecked();
    setReducerMax.invoke(job, max);
    // copy the setting into the passed configuration
    Configuration jobConf = Hadoop.JobContext.getConfiguration.invoke(job);
    for (Map.Entry<String, String> entry : jobConf) {
      conf.set(entry.getKey(), entry.getValue());
    }
    return true;
  } catch (NoSuchMethodException e) {
    return false;
  }
}
 
Example #4
Source File: TestCopyCommandCluster.java    From kite with Apache License 2.0 6 votes vote down vote up
private boolean setLocalReducerMax(Configuration conf, int max) {
  try {
    Job job = Hadoop.Job.newInstance.invoke(new Configuration(false));
    DynMethods.StaticMethod setReducerMax = new DynMethods
        .Builder("setLocalMaxRunningReduces")
        .impl(LocalJobRunner.class,
            org.apache.hadoop.mapreduce.JobContext.class, Integer.TYPE)
        .buildStaticChecked();
    setReducerMax.invoke(job, max);
    // copy the setting into the passed configuration
    Configuration jobConf = Hadoop.JobContext.getConfiguration.invoke(job);
    for (Map.Entry<String, String> entry : jobConf) {
      conf.set(entry.getKey(), entry.getValue());
    }
    return true;
  } catch (NoSuchMethodException e) {
    return false;
  }
}
 
Example #5
Source File: TestTransformCommandCluster.java    From kite with Apache License 2.0 6 votes vote down vote up
private boolean setLocalReducerMax(Configuration conf, int max) {
  try {
    Job job = Hadoop.Job.newInstance.invoke(new Configuration(false));
    DynMethods.StaticMethod setReducerMax = new DynMethods
        .Builder("setLocalMaxRunningReduces")
        .impl(LocalJobRunner.class,
            org.apache.hadoop.mapreduce.JobContext.class, Integer.TYPE)
        .buildStaticChecked();
    setReducerMax.invoke(job, max);
    // copy the setting into the passed configuration
    Configuration jobConf = Hadoop.JobContext.getConfiguration.invoke(job);
    for (Map.Entry<String, String> entry : jobConf) {
      conf.set(entry.getKey(), entry.getValue());
    }
    return true;
  } catch (NoSuchMethodException e) {
    return false;
  }
}
 
Example #6
Source File: TestClientProtocolProviderImpls.java    From hadoop with Apache License 2.0 5 votes vote down vote up
@Test
public void testClusterWithLocalClientProvider() throws Exception {
  Configuration conf = new Configuration();
  conf.set(MRConfig.FRAMEWORK_NAME, "local");
  Cluster cluster = new Cluster(conf);
  assertTrue(cluster.getClient() instanceof LocalJobRunner);
  cluster.close();
}
 
Example #7
Source File: TestLocalRunner.java    From hadoop with Apache License 2.0 5 votes vote down vote up
/**
 * Run a test which creates a SequenceMapper / IdentityReducer
 * job over a set of generated number files.
 */
private void doMultiReducerTest(int numMaps, int numReduces,
    int parallelMaps, int parallelReduces) throws Exception {

  Path in = getNumberDirPath();
  Path out = getOutputPath();

  // Clear data from any previous tests.
  Configuration conf = new Configuration();
  FileSystem fs = FileSystem.getLocal(conf);
  if (fs.exists(out)) {
    fs.delete(out, true);
  }

  if (fs.exists(in)) {
    fs.delete(in, true);
  }

  for (int i = 0; i < numMaps; i++) {
    makeNumberFile(i, 100);
  }

  Job job = Job.getInstance();
  job.setNumReduceTasks(numReduces);

  job.setMapperClass(SequenceMapper.class);
  job.setOutputKeyClass(Text.class);
  job.setOutputValueClass(NullWritable.class);
  FileInputFormat.addInputPath(job, in);
  FileOutputFormat.setOutputPath(job, out);

  LocalJobRunner.setLocalMaxRunningMaps(job, parallelMaps);
  LocalJobRunner.setLocalMaxRunningReduces(job, parallelReduces);

  boolean result = job.waitForCompletion(true);
  assertTrue("Job failed!!", result);

  verifyNumberJob(numMaps);
}
 
Example #8
Source File: TestClientProtocolProviderImpls.java    From big-c with Apache License 2.0 5 votes vote down vote up
@Test
public void testClusterWithLocalClientProvider() throws Exception {
  Configuration conf = new Configuration();
  conf.set(MRConfig.FRAMEWORK_NAME, "local");
  Cluster cluster = new Cluster(conf);
  assertTrue(cluster.getClient() instanceof LocalJobRunner);
  cluster.close();
}
 
Example #9
Source File: TestLocalRunner.java    From big-c with Apache License 2.0 5 votes vote down vote up
/**
 * Run a test which creates a SequenceMapper / IdentityReducer
 * job over a set of generated number files.
 */
private void doMultiReducerTest(int numMaps, int numReduces,
    int parallelMaps, int parallelReduces) throws Exception {

  Path in = getNumberDirPath();
  Path out = getOutputPath();

  // Clear data from any previous tests.
  Configuration conf = new Configuration();
  FileSystem fs = FileSystem.getLocal(conf);
  if (fs.exists(out)) {
    fs.delete(out, true);
  }

  if (fs.exists(in)) {
    fs.delete(in, true);
  }

  for (int i = 0; i < numMaps; i++) {
    makeNumberFile(i, 100);
  }

  Job job = Job.getInstance();
  job.setNumReduceTasks(numReduces);

  job.setMapperClass(SequenceMapper.class);
  job.setOutputKeyClass(Text.class);
  job.setOutputValueClass(NullWritable.class);
  FileInputFormat.addInputPath(job, in);
  FileOutputFormat.setOutputPath(job, out);

  LocalJobRunner.setLocalMaxRunningMaps(job, parallelMaps);
  LocalJobRunner.setLocalMaxRunningReduces(job, parallelReduces);

  boolean result = job.waitForCompletion(true);
  assertTrue("Job failed!!", result);

  verifyNumberJob(numMaps);
}
 
Example #10
Source File: TestLocalRunner.java    From RDFS with Apache License 2.0 5 votes vote down vote up
/**
 * Run a test with several mappers in parallel, operating at different
 * speeds. Verify that the correct amount of output is created.
 */
@Test
public void testMultiMaps() throws Exception {
  Path inputPath = createMultiMapsInput();
  Path outputPath = getOutputPath();

  Configuration conf = new Configuration();
  conf.setBoolean("mapred.localrunner.sequential", false);
  conf.setBoolean("mapred.localrunner.debug", true);
  conf.setInt(LocalJobRunner.LOCAL_RUNNER_SLOTS, 6);
  conf.set(JobConf.MAPRED_TASK_JAVA_OPTS, "-DtestProperty=testValue");
  Job job = new Job(conf);
  job.setMapperClass(StressMapper.class);
  job.setReducerClass(CountingReducer.class);
  job.setNumReduceTasks(1);
  job.getConfiguration().set("io.sort.record.pct", "0.50");
  job.getConfiguration().set("io.sort.mb", "25");
  FileInputFormat.addInputPath(job, inputPath);
  FileOutputFormat.setOutputPath(job, outputPath);

  FileSystem fs = FileSystem.getLocal(conf);
  if (fs.exists(outputPath)) {
    fs.delete(outputPath, true);
  }

  job.waitForCompletion(true);

  verifyOutput(outputPath);
}