org.apache.hadoop.io.BytesWritable Java Examples
The following examples show how to use
org.apache.hadoop.io.BytesWritable.
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example #1
Source File: CommonStub.java From big-c with Apache License 2.0 | 6 votes |
protected void readObject(Writable obj, DataInputStream inStream) throws IOException { int numBytes = WritableUtils.readVInt(inStream); byte[] buffer; // For BytesWritable and Text, use the specified length to set the length // this causes the "obvious" translations to work. So that if you emit // a string "abc" from C++, it shows up as "abc". if (obj instanceof BytesWritable) { buffer = new byte[numBytes]; inStream.readFully(buffer); ((BytesWritable) obj).set(buffer, 0, numBytes); } else if (obj instanceof Text) { buffer = new byte[numBytes]; inStream.readFully(buffer); ((Text) obj).set(buffer); } else { obj.readFields(inStream); } }
Example #2
Source File: UtilsForTests.java From hadoop-gpu with Apache License 2.0 | 6 votes |
/** * Configure a waiting job */ static void configureWaitingJobConf(JobConf jobConf, Path inDir, Path outputPath, int numMaps, int numRed, String jobName, String mapSignalFilename, String redSignalFilename) throws IOException { jobConf.setJobName(jobName); jobConf.setInputFormat(NonSplitableSequenceFileInputFormat.class); jobConf.setOutputFormat(SequenceFileOutputFormat.class); FileInputFormat.setInputPaths(jobConf, inDir); FileOutputFormat.setOutputPath(jobConf, outputPath); jobConf.setMapperClass(UtilsForTests.HalfWaitingMapper.class); jobConf.setReducerClass(IdentityReducer.class); jobConf.setOutputKeyClass(BytesWritable.class); jobConf.setOutputValueClass(BytesWritable.class); jobConf.setInputFormat(RandomInputFormat.class); jobConf.setNumMapTasks(numMaps); jobConf.setNumReduceTasks(numRed); jobConf.setJar("build/test/testjar/testjob.jar"); jobConf.set(getTaskSignalParameter(true), mapSignalFilename); jobConf.set(getTaskSignalParameter(false), redSignalFilename); }
Example #3
Source File: NamecoinUDFTest.java From hadoopcryptoledger with Apache License 2.0 | 6 votes |
@Test public void extractNamecoinFieldFirstUpdate() throws HiveException { String firstUpdateScript ="520A642F666C6173687570641460C7B068EDEA60281DAF424C38D8DAB87C96CF993D7B226970223A223134352E3234392E3130362E323238222C226D6170223A7B222A223A7B226970223A223134352E3234392E3130362E323238227D7D7D6D6D76A91451B4FC93AAB8CBDBD0AC9BC8EAF824643FC1E29B88AC"; byte[] firstUpdateScriptBytes = BitcoinUtil.convertHexStringToByteArray(firstUpdateScript); NamecoinExtractFieldUDF nefu = new NamecoinExtractFieldUDF(); ObjectInspector[] arguments = new ObjectInspector[1]; arguments[0] = PrimitiveObjectInspectorFactory.writableBinaryObjectInspector;; nefu.initialize(arguments); GenericUDF.DeferredObject[] doa = new GenericUDF.DeferredObject[1]; doa[0]=new GenericUDF.DeferredJavaObject(new BytesWritable(firstUpdateScriptBytes)); List<Text> resultList = (List<Text>) nefu.evaluate(doa); Text[] result=resultList.toArray(new Text[resultList.size()]); assertNotNull( result,"Valid result obtained"); // test for domain name assertEquals("d/flashupd",result[0].toString(),"Domain name of first update detected correctly"); // test for domain value assertEquals("{\"ip\":\"145.249.106.228\",\"map\":{\"*\":{\"ip\":\"145.249.106.228\"}}}",result[1].toString(),"Domain value of first update detected correctly"); }
Example #4
Source File: UtilsForTests.java From big-c with Apache License 2.0 | 6 votes |
/** * Configure a waiting job */ static void configureWaitingJobConf(JobConf jobConf, Path inDir, Path outputPath, int numMaps, int numRed, String jobName, String mapSignalFilename, String redSignalFilename) throws IOException { jobConf.setJobName(jobName); jobConf.setInputFormat(NonSplitableSequenceFileInputFormat.class); jobConf.setOutputFormat(SequenceFileOutputFormat.class); FileInputFormat.setInputPaths(jobConf, inDir); FileOutputFormat.setOutputPath(jobConf, outputPath); jobConf.setMapperClass(UtilsForTests.HalfWaitingMapper.class); jobConf.setReducerClass(IdentityReducer.class); jobConf.setOutputKeyClass(BytesWritable.class); jobConf.setOutputValueClass(BytesWritable.class); jobConf.setInputFormat(RandomInputFormat.class); jobConf.setNumMapTasks(numMaps); jobConf.setNumReduceTasks(numRed); jobConf.setJar("build/test/mapred/testjar/testjob.jar"); jobConf.set(getTaskSignalParameter(true), mapSignalFilename); jobConf.set(getTaskSignalParameter(false), redSignalFilename); }
Example #5
Source File: BaileyBorweinPlouffe.java From hadoop with Apache License 2.0 | 6 votes |
/** Compute the (offset+1)th to (offset+length)th digits. */ protected void map(LongWritable offset, IntWritable length, final Context context) throws IOException, InterruptedException { LOG.info("offset=" + offset + ", length=" + length); // compute digits final byte[] bytes = new byte[length.get() >> 1]; long d = offset.get(); for (int i = 0; i < bytes.length; d += 4) { final long digits = hexDigits(d); bytes[i++] = (byte) (digits >> 8); bytes[i++] = (byte) digits; } // output map results context.write(offset, new BytesWritable(bytes)); }
Example #6
Source File: DeduplicationJob.java From nutch-htmlunit with Apache License 2.0 | 6 votes |
@Override public void map(Text key, CrawlDatum value, OutputCollector<BytesWritable, CrawlDatum> output, Reporter reporter) throws IOException { if (value.getStatus() == CrawlDatum.STATUS_DB_FETCHED || value.getStatus() == CrawlDatum.STATUS_DB_NOTMODIFIED) { // || value.getStatus() ==CrawlDatum.STATUS_DB_GONE){ byte[] signature = value.getSignature(); if (signature == null) return; BytesWritable sig = new BytesWritable(signature); // add the URL as a temporary MD value.getMetaData().put(urlKey, key); // reduce on the signature output.collect(sig, value); } }
Example #7
Source File: DataToDoubleSummarySketchUDAFTest.java From incubator-datasketches-hive with Apache License 2.0 | 6 votes |
@Test public void partial1ModeIntKeysDefaultParams() throws Exception { ObjectInspector[] inspectors = new ObjectInspector[] { intInspector, doubleInspector }; GenericUDAFParameterInfo info = new SimpleGenericUDAFParameterInfo(inspectors, false, false, false); try (GenericUDAFEvaluator eval = new DataToDoubleSummarySketchUDAF().getEvaluator(info)) { ObjectInspector resultInspector = eval.init(Mode.PARTIAL1, inspectors); checkIntermediateResultInspector(resultInspector); @SuppressWarnings("unchecked") State<DoubleSummary> state = (State<DoubleSummary>) eval.getNewAggregationBuffer(); eval.iterate(state, new Object[] {new IntWritable(1), new DoubleWritable(1)}); eval.iterate(state, new Object[] {new IntWritable(2), new DoubleWritable(1)}); Object result = eval.terminatePartial(state); Assert.assertNotNull(result); Assert.assertTrue(result instanceof List); List<?> r = (List<?>) result; Assert.assertEquals(r.size(), 2); Assert.assertEquals(((IntWritable) r.get(0)).get(), DEFAULT_NOMINAL_ENTRIES); Sketch<DoubleSummary> resultSketch = Sketches.heapifySketch( BytesWritableHelper.wrapAsMemory((BytesWritable) r.get(1)), new DoubleSummaryDeserializer()); Assert.assertFalse(resultSketch.isEstimationMode()); Assert.assertEquals(resultSketch.getEstimate(), 2.0); } }
Example #8
Source File: ST_GeomFromText.java From spatial-framework-for-hadoop with Apache License 2.0 | 6 votes |
public BytesWritable evaluate(Text wkwrap, int wkid) throws UDFArgumentException { String wkt = wkwrap.toString(); try { SpatialReference spatialReference = null; if (wkid != GeometryUtils.WKID_UNKNOWN) { spatialReference = SpatialReference.create(wkid); } OGCGeometry ogcObj = OGCGeometry.fromText(wkt); ogcObj.setSpatialReference(spatialReference); return GeometryUtils.geometryToEsriShapeBytesWritable(ogcObj); } catch (Exception e) { // IllegalArgumentException, GeometryException LogUtils.Log_InvalidText(LOG, wkt); return null; } }
Example #9
Source File: CommonHadoopShim.java From pentaho-hadoop-shims with Apache License 2.0 | 6 votes |
@Override public Class<? extends Writable> getHadoopWritableCompatibleClass( ValueMetaInterface kettleType ) { if ( kettleType == null ) { return NullWritable.class; } switch ( kettleType.getType() ) { case ValueMetaInterface.TYPE_STRING: case ValueMetaInterface.TYPE_BIGNUMBER: case ValueMetaInterface.TYPE_DATE: return Text.class; case ValueMetaInterface.TYPE_INTEGER: return LongWritable.class; case ValueMetaInterface.TYPE_NUMBER: return DoubleWritable.class; case ValueMetaInterface.TYPE_BOOLEAN: return BooleanWritable.class; case ValueMetaInterface.TYPE_BINARY: return BytesWritable.class; default: return Text.class; } }
Example #10
Source File: ST_Point.java From spatial-framework-for-hadoop with Apache License 2.0 | 6 votes |
public BytesWritable evaluate(Text wkwrap) throws UDFArgumentException { String wkt = wkwrap.toString(); try { OGCGeometry ogcObj = OGCGeometry.fromText(wkt); ogcObj.setSpatialReference(null); if (ogcObj.geometryType().equals("Point")) { return GeometryUtils.geometryToEsriShapeBytesWritable(ogcObj); } else { LogUtils.Log_InvalidType(LOG, GeometryUtils.OGCType.ST_POINT, GeometryUtils.OGCType.UNKNOWN); return null; } } catch (Exception e) { // IllegalArgumentException, GeometryException LogUtils.Log_InvalidText(LOG, wkt); return null; } }
Example #11
Source File: SeqCombiner.java From compiler with Apache License 2.0 | 6 votes |
public static long readAndAppendCommit(Configuration conf, FileSystem fileSystem, MapFile.Writer writer, String fileName, long lastAstKey, long lastCommitKey) throws IOException { long newLastKey = lastCommitKey; SequenceFile.Reader r = new SequenceFile.Reader(fileSystem, new Path(fileName), conf); LongWritable longKey = new LongWritable(); BytesWritable value = new BytesWritable(); try { while (r.next(longKey, value)) { newLastKey = longKey.get() + lastCommitKey; Revision rev = Revision.parseFrom(CodedInputStream.newInstance(value.getBytes(), 0, value.getLength())); Revision.Builder rb = Revision.newBuilder(rev); for (ChangedFile.Builder cfb : rb.getFilesBuilderList()) { long key = cfb.getKey(); if (key > 0) cfb.setKey(lastAstKey + key); } writer.append(new LongWritable(newLastKey), new BytesWritable(rb.build().toByteArray())); } } catch (Exception e) { System.err.println(fileName); e.printStackTrace(); } finally { r.close(); } return newLastKey; }
Example #12
Source File: ST_MinY.java From spatial-framework-for-hadoop with Apache License 2.0 | 6 votes |
public DoubleWritable evaluate(BytesWritable geomref) { if (geomref == null || geomref.getLength() == 0) { LogUtils.Log_ArgumentsNull(LOG); return null; } OGCGeometry ogcGeometry = GeometryUtils.geometryFromEsriShape(geomref); if (ogcGeometry == null) { LogUtils.Log_ArgumentsNull(LOG); return null; } Envelope envBound = new Envelope(); ogcGeometry.getEsriGeometry().queryEnvelope(envBound); resultDouble.set(envBound.getYMin()); return resultDouble; }
Example #13
Source File: TestCombineSequenceFileInputFormat.java From hadoop with Apache License 2.0 | 6 votes |
private static void createFiles(int length, int numFiles, Random random, Job job) throws IOException { Range[] ranges = createRanges(length, numFiles, random); for (int i = 0; i < numFiles; i++) { Path file = new Path(workDir, "test_" + i + ".seq"); // create a file with length entries @SuppressWarnings("deprecation") SequenceFile.Writer writer = SequenceFile.createWriter(localFs, job.getConfiguration(), file, IntWritable.class, BytesWritable.class); Range range = ranges[i]; try { for (int j = range.start; j < range.end; j++) { IntWritable key = new IntWritable(j); byte[] data = new byte[random.nextInt(10)]; random.nextBytes(data); BytesWritable value = new BytesWritable(data); writer.append(key, value); } } finally { writer.close(); } } }
Example #14
Source File: TestIPCServerResponder.java From RDFS with Apache License 2.0 | 6 votes |
@Override public void run() { for (int i = 0; i < count; i++) { try { int byteSize = RANDOM.nextInt(BYTE_COUNT); byte[] bytes = new byte[byteSize]; System.arraycopy(BYTES, 0, bytes, 0, byteSize); Writable param = new BytesWritable(bytes); Writable value = client.call(param, address); Thread.sleep(RANDOM.nextInt(20)); } catch (Exception e) { LOG.fatal("Caught: " + e); failed = true; } } }
Example #15
Source File: GroupbyKeyComparator.java From Eagle with Apache License 2.0 | 6 votes |
@Override public int compare(GroupbyKey key1, GroupbyKey key2){ List<BytesWritable> list1 = key1.getValue(); List<BytesWritable> list2 = key2.getValue(); if(list1 == null || list2 == null || list1.size() != list2.size()) throw new IllegalArgumentException("2 list of groupby fields must be non-null and have the same size"); ListIterator<BytesWritable> e1 = list1.listIterator(); ListIterator<BytesWritable> e2 = list2.listIterator(); while(e1.hasNext() && e2.hasNext()){ int r = Bytes.compareTo(e1.next().copyBytes(), e2.next().copyBytes()); if(r != 0) return r; } return 0; }
Example #16
Source File: GroupbyKeyComparator.java From eagle with Apache License 2.0 | 6 votes |
@Override public int compare(GroupbyKey key1, GroupbyKey key2) { List<BytesWritable> list1 = key1.getValue(); List<BytesWritable> list2 = key2.getValue(); if (list1 == null || list2 == null || list1.size() != list2.size()) { throw new IllegalArgumentException("2 list of groupby fields must be non-null and have the same size"); } ListIterator<BytesWritable> e1 = list1.listIterator(); ListIterator<BytesWritable> e2 = list2.listIterator(); while (e1.hasNext() && e2.hasNext()) { int r = Bytes.compareTo(e1.next().copyBytes(), e2.next().copyBytes()); if (r != 0) { return r; } } return 0; }
Example #17
Source File: TestTupleWritable.java From big-c with Apache License 2.0 | 6 votes |
public void testWritable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); ByteArrayOutputStream out = new ByteArrayOutputStream(); sTuple.write(new DataOutputStream(out)); ByteArrayInputStream in = new ByteArrayInputStream(out.toByteArray()); TupleWritable dTuple = new TupleWritable(); dTuple.readFields(new DataInputStream(in)); assertTrue("Failed to write/read tuple", sTuple.equals(dTuple)); }
Example #18
Source File: TFile.java From hadoop with Apache License 2.0 | 6 votes |
/** * Copy the value into BytesWritable. The input BytesWritable will be * automatically resized to the actual value size. The implementation * directly uses the buffer inside BytesWritable for storing the value. * The call does not require the value length to be known. * * @param value * @throws IOException */ public long getValue(BytesWritable value) throws IOException { DataInputStream dis = getValueStream(); int size = 0; try { int remain; while ((remain = valueBufferInputStream.getRemain()) > 0) { value.setSize(size + remain); dis.readFully(value.getBytes(), size, remain); size += remain; } return value.getLength(); } finally { dis.close(); } }
Example #19
Source File: HadoopCloverConvert.java From CloverETL-Engine with GNU Lesser General Public License v2.1 | 6 votes |
@SuppressWarnings("rawtypes") public static Class cloverType2Hadoop(DataFieldMetadata field) throws IOException{ switch (field.getDataType()){ case BOOLEAN: return BooleanWritable.class; case BYTE: case CBYTE: return BytesWritable.class; case DATE: return LongWritable.class; case INTEGER: return IntWritable.class; case LONG: return LongWritable.class; case NUMBER: return DoubleWritable.class; case STRING: return Text.class; default: throw new IOException(String.format("Unsupported CloverDX data type \"%s\" of field \"%s\" in conversion to Hadoop.",field.getDataType().getName(),field.getName())); } }
Example #20
Source File: TestTupleWritable.java From RDFS with Apache License 2.0 | 6 votes |
public void testNestedIterable() throws Exception { Random r = new Random(); Writable[] writs = { new BooleanWritable(r.nextBoolean()), new FloatWritable(r.nextFloat()), new FloatWritable(r.nextFloat()), new IntWritable(r.nextInt()), new LongWritable(r.nextLong()), new BytesWritable("dingo".getBytes()), new LongWritable(r.nextLong()), new IntWritable(r.nextInt()), new BytesWritable("yak".getBytes()), new IntWritable(r.nextInt()) }; TupleWritable sTuple = makeTuple(writs); assertTrue("Bad count", writs.length == verifIter(writs, sTuple, 0)); }
Example #21
Source File: FlinkBitcoinDataSourceTest.java From hadoopcryptoledger with Apache License 2.0 | 6 votes |
@Test public void parseBitcoinRawBlock() throws HadoopCryptoLedgerConfigurationException, IOException { ClassLoader classLoader = getClass().getClassLoader(); String fileName="genesis.blk"; String fileNameBlock=classLoader.getResource("testdata/"+fileName).getFile(); Path file = new Path(fileNameBlock); FileInputSplit blockInputSplit = new FileInputSplit(0,file,0, -1, null); BitcoinRawBlockFlinkInputFormat inputFormat = new BitcoinRawBlockFlinkInputFormat(1024*1024,"F9BEB4D9",false); inputFormat.open(blockInputSplit); assertFalse(inputFormat.reachedEnd(),"End not reached"); BytesWritable reuse = new BytesWritable(); BytesWritable nextBlock = inputFormat.nextRecord(reuse); assertNotNull(nextBlock,"First Block returned"); assertEquals( 293, nextBlock.getLength(),"First Block must have size of 293"); nextBlock=inputFormat.nextRecord(reuse); assertNull(nextBlock,"No further block"); assertTrue(inputFormat.reachedEnd(),"End reached"); }
Example #22
Source File: MultipleKafkaInputFormat.java From kangaroo with Apache License 2.0 | 5 votes |
/** * {@inheritDoc} */ @Override public RecordReader<LongWritable, BytesWritable> createRecordReader(final InputSplit split, final TaskAttemptContext context) throws IOException, InterruptedException { final TaggedInputSplit taggedInputSplit = (TaggedInputSplit) split; final TaskAttemptContext taskAttemptContextClone = new TaskAttemptContextImpl(taggedInputSplit.getConf(), context.getTaskAttemptID()); taskAttemptContextClone.setStatus(context.getStatus()); return new DelegatingRecordReader<LongWritable, BytesWritable>(split, taskAttemptContextClone); }
Example #23
Source File: RawBytesOutputReader.java From hadoop with Apache License 2.0 | 5 votes |
@Override public void initialize(PipeMapRed pipeMapRed) throws IOException { super.initialize(pipeMapRed); clientIn = pipeMapRed.getClientInput(); key = new BytesWritable(); value = new BytesWritable(); }
Example #24
Source File: DataToDoubleSummarySketchUDAFTest.java From incubator-datasketches-hive with Apache License 2.0 | 5 votes |
@Test public void completeModeDoubleKeysExplicitParams() throws Exception { ObjectInspector[] inspectors = new ObjectInspector[] { doubleInspector, doubleInspector, intInspector, floatInspector }; GenericUDAFParameterInfo info = new SimpleGenericUDAFParameterInfo(inspectors, false, false, false); try (GenericUDAFEvaluator eval = new DataToDoubleSummarySketchUDAF().getEvaluator(info)) { ObjectInspector resultInspector = eval.init(Mode.COMPLETE, inspectors); checkFinalResultInspector(resultInspector); @SuppressWarnings("unchecked") State<DoubleSummary> state = (State<DoubleSummary>) eval.getNewAggregationBuffer(); eval.iterate(state, new Object[] {new DoubleWritable(1), new DoubleWritable(1), new IntWritable(32), new FloatWritable(0.99f)}); eval.iterate(state, new Object[] {new DoubleWritable(2), new DoubleWritable(1), new IntWritable(32), new FloatWritable(0.99f)}); Object result = eval.terminate(state); Assert.assertNotNull(result); Assert.assertTrue(result instanceof BytesWritable); Sketch<DoubleSummary> resultSketch = Sketches.heapifySketch( BytesWritableHelper.wrapAsMemory((BytesWritable) result), new DoubleSummaryDeserializer()); // because of sampling probability < 1 Assert.assertTrue(resultSketch.isEstimationMode()); Assert.assertEquals(resultSketch.getEstimate(), 2.0, 0.05); eval.reset(state); result = eval.terminate(state); Assert.assertNull(result); } }
Example #25
Source File: SortValidator.java From big-c with Apache License 2.0 | 5 votes |
public void map(BytesWritable key, BytesWritable value, OutputCollector<BytesWritable, IntWritable> output, Reporter reporter) throws IOException { // newKey = (key, value) BytesWritable keyValue = new BytesWritable(pair(key, value)); // output (newKey, value) output.collect(keyValue, this.value); }
Example #26
Source File: KVGenerator.java From hbase with Apache License 2.0 | 5 votes |
public void next(BytesWritable key, BytesWritable value, boolean dupKey) { if (dupKey) { key.set(lastKey); } else { fillKey(key); } fillValue(value); }
Example #27
Source File: TestTFileSeqFileComparison.java From attic-apex-malhar with Apache License 2.0 | 5 votes |
private void timeWrite(Path path, KVAppendable appendable, int baseKlen, int baseVlen, long fileSize) throws IOException { int maxKlen = baseKlen * 2; int maxVlen = baseVlen * 2; BytesWritable key = new BytesWritable(); BytesWritable value = new BytesWritable(); byte[] keyBuffer = new byte[maxKlen]; byte[] valueBuffer = new byte[maxVlen]; Random rng = new Random(options.seed); long totalBytes = 0; printlnWithTimestamp("Start writing: " + path.getName() + "..."); startTime(); for (long i = 0; true; ++i) { if (i % 1000 == 0) { // test the size for every 1000 rows. if (fs.getFileStatus(path).getLen() >= fileSize) { break; } } int klen = rng.nextInt(baseKlen) + baseKlen; int vlen = rng.nextInt(baseVlen) + baseVlen; fillBuffer(rng, key, keyBuffer, klen); fillBuffer(rng, value, valueBuffer, vlen); key.set(keyBuffer, 0, klen); value.set(valueBuffer, 0, vlen); appendable.append(key, value); totalBytes += klen; totalBytes += vlen; } stopTime(); appendable.close(); reportStats(path, totalBytes); }
Example #28
Source File: BitcoinUDFTest.java From hadoopcryptoledger with Apache License 2.0 | 5 votes |
@Test public void BitcoinScriptPaymentPatternAnalyzerUDFNotNull() { BitcoinScriptPaymentPatternAnalyzerUDF bsppaUDF = new BitcoinScriptPaymentPatternAnalyzerUDF(); byte[] txOutScriptGenesis= new byte[]{(byte)0x41,(byte)0x04,(byte)0x67,(byte)0x8A,(byte)0xFD,(byte)0xB0,(byte)0xFE,(byte)0x55,(byte)0x48,(byte)0x27,(byte)0x19,(byte)0x67,(byte)0xF1,(byte)0xA6,(byte)0x71,(byte)0x30,(byte)0xB7,(byte)0x10,(byte)0x5C,(byte)0xD6,(byte)0xA8,(byte)0x28,(byte)0xE0,(byte)0x39,(byte)0x09,(byte)0xA6,(byte)0x79,(byte)0x62,(byte)0xE0,(byte)0xEA,(byte)0x1F,(byte)0x61,(byte)0xDE,(byte)0xB6,(byte)0x49,(byte)0xF6,(byte)0xBC,(byte)0x3F,(byte)0x4C,(byte)0xEF,(byte)0x38,(byte)0xC4,(byte)0xF3,(byte)0x55,(byte)0x04,(byte)0xE5,(byte)0x1E,(byte)0xC1,(byte)0x12,(byte)0xDE,(byte)0x5C,(byte)0x38,(byte)0x4D,(byte)0xF7,(byte)0xBA,(byte)0x0B,(byte)0x8D,(byte)0x57,(byte)0x8A,(byte)0x4C,(byte)0x70,(byte)0x2B,(byte)0x6B,(byte)0xF1,(byte)0x1D,(byte)0x5F,(byte)0xAC}; BytesWritable evalObj = new BytesWritable(txOutScriptGenesis); String result = bsppaUDF.evaluate(evalObj).toString(); String comparatorText = "bitcoinpubkey_4104678AFDB0FE5548271967F1A67130B7105CD6A828E03909A67962E0EA1F61DEB649F6BC3F4CEF38C4F35504E51EC112DE5C384DF7BA0B8D578A4C702B6BF11D5F"; assertEquals(comparatorText,result,"TxOutScript from Genesis should be payment to a pubkey address"); }
Example #29
Source File: TestTFileSeek.java From attic-apex-malhar with Apache License 2.0 | 5 votes |
public void seekTFile() throws IOException { int miss = 0; long totalBytes = 0; FSDataInputStream fsdis = fs.open(path); Reader reader = new Reader(fsdis, fs.getFileStatus(path).getLen(), conf); KeySampler kSampler = new KeySampler(rng, reader.getFirstKey(), reader.getLastKey(), keyLenGen); Scanner scanner = reader.createScanner(); BytesWritable key = new BytesWritable(); BytesWritable val = new BytesWritable(); timer.reset(); timer.start(); for (int i = 0; i < options.seekCount; ++i) { kSampler.next(key); scanner.lowerBound(key.get(), 0, key.getSize()); if (!scanner.atEnd()) { scanner.entry().get(key, val); totalBytes += key.getSize(); totalBytes += val.getSize(); } else { ++miss; } } timer.stop(); double duration = (double) timer.read() / 1000; // in us. System.out.printf( "time: %s...avg seek: %s...%d hit...%d miss...avg I/O size: %.2fKB\n", timer.toString(), NanoTimer.nanoTimeToString(timer.read() / options.seekCount), options.seekCount - miss, miss, (double) totalBytes / 1024 / (options.seekCount - miss)); }
Example #30
Source File: Warp10InputFormat.java From warp10-platform with Apache License 2.0 | 5 votes |
@Override public RecordReader<Text, BytesWritable> createRecordReader(InputSplit split, TaskAttemptContext context) throws IOException { if (!(split instanceof Warp10InputSplit)) { throw new IOException("Invalid split type."); } return new Warp10RecordReader(this.suffix); }