Java Code Examples for org.apache.hadoop.io.SequenceFile#CompressionType

The following examples show how to use org.apache.hadoop.io.SequenceFile#CompressionType . You can vote up the ones you like or vote down the ones you don't like, and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: TestMapRed.java    From RDFS with Apache License 2.0 5 votes vote down vote up
public void testCompression() throws Exception {
  EnumSet<SequenceFile.CompressionType> seq =
    EnumSet.allOf(SequenceFile.CompressionType.class);
  for (CompressionType redCompression : seq) {
    for(int combine=0; combine < 2; ++combine) {
      checkCompression(false, redCompression, combine == 1);
      checkCompression(true, redCompression, combine == 1);
    }
  }
}
 
Example 2
Source File: Export.java    From hbase with Apache License 2.0 5 votes vote down vote up
private static SequenceFile.CompressionType getCompressionType(
    final ExportProtos.ExportRequest request) {
  if (request.hasCompressType()) {
    return SequenceFile.CompressionType.valueOf(request.getCompressType());
  } else {
    return DEFAULT_TYPE;
  }
}
 
Example 3
Source File: TestMapRed.java    From hadoop-gpu with Apache License 2.0 5 votes vote down vote up
public void testCompression() throws Exception {
  EnumSet<SequenceFile.CompressionType> seq =
    EnumSet.allOf(SequenceFile.CompressionType.class);
  for (CompressionType redCompression : seq) {
    for(int combine=0; combine < 2; ++combine) {
      checkCompression(false, redCompression, combine == 1);
      checkCompression(true, redCompression, combine == 1);
    }
  }
}
 
Example 4
Source File: TestMapRed.java    From hadoop with Apache License 2.0 5 votes vote down vote up
@Test 
public void testCompression() throws Exception {
  EnumSet<SequenceFile.CompressionType> seq =
    EnumSet.allOf(SequenceFile.CompressionType.class);
  for (CompressionType redCompression : seq) {
    for(int combine=0; combine < 2; ++combine) {
      checkCompression(false, redCompression, combine == 1);
      checkCompression(true, redCompression, combine == 1);
    }
  }
}
 
Example 5
Source File: TestMapRed.java    From big-c with Apache License 2.0 5 votes vote down vote up
@Test 
public void testCompression() throws Exception {
  EnumSet<SequenceFile.CompressionType> seq =
    EnumSet.allOf(SequenceFile.CompressionType.class);
  for (CompressionType redCompression : seq) {
    for(int combine=0; combine < 2; ++combine) {
      checkCompression(false, redCompression, combine == 1);
      checkCompression(true, redCompression, combine == 1);
    }
  }
}
 
Example 6
Source File: TestRecordWriterManager.java    From datacollector with Apache License 2.0 5 votes vote down vote up
private void testSeqFile(CompressionCodec compressionCodec, SequenceFile.CompressionType compressionType)
    throws Exception {
  RecordWriterManager mgr = managerBuilder()
    .dirPathTemplate(getTestDir().toString() + "/${YYYY()}")
    .compressionCodec(compressionCodec)
    .compressionType(compressionType)
    .fileType(HdfsFileType.SEQUENCE_FILE)
    .build();

  FileSystem fs = FileSystem.get(uri, hdfsConf);
  Path file = new Path(getTestDir(), UUID.randomUUID().toString());
  long expires = System.currentTimeMillis() + 50000;
  RecordWriter writer = mgr.createWriter(fs, file, 50000);
  Assert.assertTrue(expires <= writer.getExpiresOn());
  Assert.assertFalse(writer.isTextFile());
  Assert.assertTrue(writer.isSeqFile());
  Record record = RecordCreator.create();
  record.set(Field.create("a"));
  writer.write(record);
  writer.close();

  SequenceFile.Reader reader = new SequenceFile.Reader(fs, file, new HdfsConfiguration());
  Text key = new Text();
  Text value = new Text();
  Assert.assertTrue(reader.next(key, value));
  Assert.assertNotNull(UUID.fromString(key.toString()));
  Assert.assertEquals("a", value.toString().trim());
  Assert.assertFalse(reader.next(key, value));
  reader.close();
}
 
Example 7
Source File: HdfsState.java    From jstorm with Apache License 2.0 4 votes vote down vote up
public SequenceFileOptions withCompressionType(SequenceFile.CompressionType compressionType) {
    this.compressionType = compressionType;
    return this;
}
 
Example 8
Source File: SequenceFileBolt.java    From storm-hdfs with Apache License 2.0 4 votes vote down vote up
public SequenceFileBolt withCompressionType(SequenceFile.CompressionType compressionType){
    this.compressionType = compressionType;
    return this;
}
 
Example 9
Source File: HdfsState.java    From storm-hdfs with Apache License 2.0 4 votes vote down vote up
public SequenceFileOptions withCompressionType(SequenceFile.CompressionType compressionType){
    this.compressionType = compressionType;
    return this;
}
 
Example 10
Source File: SequenceFileWriter.java    From flink with Apache License 2.0 4 votes vote down vote up
SequenceFile.CompressionType getCompressionType() {
	return compressionType;
}
 
Example 11
Source File: RecordWriterManager.java    From datacollector with Apache License 2.0 4 votes vote down vote up
public RecordWriterManager(
    FileSystem fs,
    Configuration hdfsConf,
    String uniquePrefix,
    String fileNameSuffix,
    boolean dirPathTemplateInHeader,
    String dirPathTemplate,
    TimeZone timeZone,
    long cutOffSecs,
    long cutOffSizeBytes,
    long cutOffRecords,
    HdfsFileType fileType,
    CompressionCodec compressionCodec,
    SequenceFile.CompressionType compressionType,
    String keyEL,
    boolean rollIfHeader,
    String rollHeaderName,
    String fileNameEL,
    WholeFileExistsAction wholeFileAlreadyExistsAction,
    String permissionEL,
    DataGeneratorFactory generatorFactory,
    Target.Context context,
    String config
) {
  this.fs = fs;
  this.hdfsConf = hdfsConf;
  this.uniquePrefix = uniquePrefix;
  this.fileNameSuffix = fileNameSuffix;
  this.dirPathTemplateInHeader = dirPathTemplateInHeader;
  this.dirPathTemplate = dirPathTemplate;
  this.timeZone = timeZone;
  this.cutOffMillis = preventOverflow(cutOffSecs * 1000);
  this.cutOffSize = cutOffSizeBytes;
  this.cutOffRecords = cutOffRecords;
  this.fileType = fileType;
  this.compressionCodec = compressionCodec;
  this.compressionType = compressionType;
  this.keyEL = keyEL;
  this.generatorFactory = generatorFactory;
  this.context = context;
  this.rollIfHeader = rollIfHeader;
  this.rollHeaderName = rollHeaderName;
  closedPaths = new ConcurrentLinkedQueue<>();
  pathResolver = new PathResolver(context, config, dirPathTemplate, timeZone);
  fsHelper = getFsHelper(context, fileNameEL, wholeFileAlreadyExistsAction, permissionEL);
}
 
Example 12
Source File: CreateHadoopSequenceFile.java    From localization_nifi with Apache License 2.0 4 votes vote down vote up
@Override
public void onTrigger(ProcessContext context, ProcessSession session) throws ProcessException {

    FlowFile flowFile = session.get();
    if (flowFile == null) {
        return;
    }

    String mimeType = flowFile.getAttribute(CoreAttributes.MIME_TYPE.key());
    String packagingFormat = NOT_PACKAGED;
    if (null != mimeType) {
        switch (mimeType.toLowerCase()) {
            case "application/tar":
                packagingFormat = TAR_FORMAT;
                break;
            case "application/zip":
                packagingFormat = ZIP_FORMAT;
                break;
            case "application/flowfile-v3":
                packagingFormat = FLOWFILE_STREAM_FORMAT_V3;
                break;
            default:
                getLogger().warn(
                        "Cannot unpack {} because its mime.type attribute is set to '{}', which is not a format that can be unpacked",
                        new Object[]{flowFile, mimeType});
        }
    }
    final SequenceFileWriter sequenceFileWriter;
    switch (packagingFormat) {
        case TAR_FORMAT:
            sequenceFileWriter = new TarUnpackerSequenceFileWriter();
            break;
        case ZIP_FORMAT:
            sequenceFileWriter = new ZipUnpackerSequenceFileWriter();
            break;
        case FLOWFILE_STREAM_FORMAT_V3:
            sequenceFileWriter = new FlowFileStreamUnpackerSequenceFileWriter();
            break;
        default:
            sequenceFileWriter = new SequenceFileWriterImpl();
    }

    final Configuration configuration = getConfiguration();
    if (configuration == null) {
        getLogger().error("HDFS not configured properly");
        session.transfer(flowFile, RELATIONSHIP_FAILURE);
        context.yield();
        return;
    }

    final CompressionCodec codec = getCompressionCodec(context, configuration);

    final String value = context.getProperty(COMPRESSION_TYPE).getValue();
    final SequenceFile.CompressionType compressionType = value == null
        ? SequenceFile.CompressionType.valueOf(DEFAULT_COMPRESSION_TYPE) : SequenceFile.CompressionType.valueOf(value);

    final String fileName = flowFile.getAttribute(CoreAttributes.FILENAME.key()) + ".sf";
    flowFile = session.putAttribute(flowFile, CoreAttributes.FILENAME.key(), fileName);

    try {
        StopWatch stopWatch = new StopWatch(true);
        flowFile = sequenceFileWriter.writeSequenceFile(flowFile, session, configuration, compressionType, codec);
        session.getProvenanceReporter().modifyContent(flowFile, stopWatch.getElapsed(TimeUnit.MILLISECONDS));
        session.transfer(flowFile, RELATIONSHIP_SUCCESS);
        getLogger().info("Transferred flowfile {} to {}", new Object[]{flowFile, RELATIONSHIP_SUCCESS});
    } catch (ProcessException e) {
        getLogger().error("Failed to create Sequence File. Transferring {} to 'failure'", new Object[]{flowFile}, e);
        session.transfer(flowFile, RELATIONSHIP_FAILURE);
    }

}
 
Example 13
Source File: SequenceFileBolt.java    From jstorm with Apache License 2.0 4 votes vote down vote up
public SequenceFileBolt withCompressionType(SequenceFile.CompressionType compressionType){
    this.compressionType = compressionType;
    return this;
}
 
Example 14
Source File: SequenceFileWriter.java    From flink with Apache License 2.0 4 votes vote down vote up
SequenceFile.CompressionType getCompressionType() {
	return compressionType;
}
 
Example 15
Source File: SequenceFileWriter.java    From Flink-CEPplus with Apache License 2.0 4 votes vote down vote up
SequenceFile.CompressionType getCompressionType() {
	return compressionType;
}
 
Example 16
Source File: SequenceFileWriterFactory.java    From flink with Apache License 2.0 3 votes vote down vote up
/**
 * Creates a new SequenceFileWriterFactory using the given builder to assemble the
 * SequenceFileWriter.
 *
 * @param hadoopConf           The Hadoop configuration for Sequence File Writer.
 * @param keyClass             The class of key to write.
 * @param valueClass           The class of value to write.
 * @param compressionCodecName The name of compression codec.
 * @param compressionType      The type of compression level.
 */
public SequenceFileWriterFactory(Configuration hadoopConf, Class<K> keyClass, Class<V> valueClass, String compressionCodecName, SequenceFile.CompressionType compressionType) {
	this.serializableHadoopConfig = new SerializableHadoopConfiguration(checkNotNull(hadoopConf));
	this.keyClass = checkNotNull(keyClass);
	this.valueClass = checkNotNull(valueClass);
	this.compressionCodecName = checkNotNull(compressionCodecName);
	this.compressionType = checkNotNull(compressionType);
}
 
Example 17
Source File: SequenceFileWriterFactory.java    From Flink-CEPplus with Apache License 2.0 3 votes vote down vote up
/**
 * Creates a new SequenceFileWriterFactory using the given builder to assemble the
 * SequenceFileWriter.
 *
 * @param hadoopConf           The Hadoop configuration for Sequence File Writer.
 * @param keyClass             The class of key to write.
 * @param valueClass           The class of value to write.
 * @param compressionCodecName The name of compression codec.
 * @param compressionType      The type of compression level.
 */
public SequenceFileWriterFactory(Configuration hadoopConf, Class<K> keyClass, Class<V> valueClass, String compressionCodecName, SequenceFile.CompressionType compressionType) {
	this.serializableHadoopConfig = new SerializableHadoopConfiguration(checkNotNull(hadoopConf));
	this.keyClass = checkNotNull(keyClass);
	this.valueClass = checkNotNull(valueClass);
	this.compressionCodecName = checkNotNull(compressionCodecName);
	this.compressionType = checkNotNull(compressionType);
}
 
Example 18
Source File: SequenceFileWriter.java    From flink with Apache License 2.0 2 votes vote down vote up
/**
 * Creates a new {@code SequenceFileWriter} that writes sequence with the given
 * compression codec and compression type.
 *
 * @param compressionCodecName Name of a Hadoop Compression Codec.
 * @param compressionType The compression type to use.
 */
public SequenceFileWriter(String compressionCodecName,
		SequenceFile.CompressionType compressionType) {
	this.compressionCodecName = compressionCodecName;
	this.compressionType = compressionType;
}
 
Example 19
Source File: SequenceFileWriter.java    From flink with Apache License 2.0 2 votes vote down vote up
/**
 * Creates a new {@code SequenceFileWriter} that writes sequence with the given
 * compression codec and compression type.
 *
 * @param compressionCodecName Name of a Hadoop Compression Codec.
 * @param compressionType The compression type to use.
 */
public SequenceFileWriter(String compressionCodecName,
		SequenceFile.CompressionType compressionType) {
	this.compressionCodecName = compressionCodecName;
	this.compressionType = compressionType;
}
 
Example 20
Source File: SequenceFileWriter.java    From Flink-CEPplus with Apache License 2.0 2 votes vote down vote up
/**
 * Creates a new {@code SequenceFileWriter} that writes sequence with the given
 * compression codec and compression type.
 *
 * @param compressionCodecName Name of a Hadoop Compression Codec.
 * @param compressionType The compression type to use.
 */
public SequenceFileWriter(String compressionCodecName,
		SequenceFile.CompressionType compressionType) {
	this.compressionCodecName = compressionCodecName;
	this.compressionType = compressionType;
}