/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.flink.formats.parquet.avro; import org.apache.flink.api.common.typeinfo.TypeInformation; import org.apache.flink.api.java.tuple.Tuple2; import org.apache.flink.core.fs.Path; import org.apache.flink.formats.avro.typeutils.GenericRecordAvroTypeInfo; import org.apache.flink.formats.parquet.generated.Address; import org.apache.flink.streaming.api.datastream.DataStream; import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; import org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink; import org.apache.flink.streaming.util.FiniteTestSource; import org.apache.flink.test.util.AbstractTestBase; import org.apache.avro.Schema; import org.apache.avro.generic.GenericData; import org.apache.avro.generic.GenericRecord; import org.apache.avro.reflect.ReflectData; import org.apache.avro.specific.SpecificData; import org.apache.hadoop.conf.Configuration; import org.apache.parquet.avro.AvroParquetReader; import org.apache.parquet.hadoop.ParquetReader; import org.apache.parquet.hadoop.util.HadoopInputFile; import org.apache.parquet.io.InputFile; import org.junit.Rule; import org.junit.Test; import org.junit.rules.Timeout; import java.io.File; import java.io.IOException; import java.io.Serializable; import java.util.AbstractCollection; import java.util.ArrayList; import java.util.Arrays; import java.util.Collection; import java.util.Iterator; import java.util.List; import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertTrue; /** * Simple integration test case for writing bulk encoded files with the * {@link StreamingFileSink} with Parquet. */ @SuppressWarnings("serial") public class ParquetStreamingFileSinkITCase extends AbstractTestBase { @Rule public final Timeout timeoutPerTest = Timeout.seconds(20); @Test public void testWriteParquetAvroSpecific() throws Exception { final File folder = TEMPORARY_FOLDER.newFolder(); final List<Address> data = Arrays.asList( new Address(1, "a", "b", "c", "12345"), new Address(2, "p", "q", "r", "12345"), new Address(3, "x", "y", "z", "12345") ); final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); env.setParallelism(1); env.enableCheckpointing(100); DataStream<Address> stream = env.addSource( new FiniteTestSource<>(data), TypeInformation.of(Address.class)); stream.addSink( StreamingFileSink.forBulkFormat( Path.fromLocalFile(folder), ParquetAvroWriters.forSpecificRecord(Address.class)) .build()); env.execute(); validateResults(folder, SpecificData.get(), data); } @Test public void testWriteParquetAvroGeneric() throws Exception { final File folder = TEMPORARY_FOLDER.newFolder(); final Schema schema = Address.getClassSchema(); final Collection<GenericRecord> data = new GenericTestDataCollection(); final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); env.setParallelism(1); env.enableCheckpointing(100); DataStream<GenericRecord> stream = env.addSource( new FiniteTestSource<>(data), new GenericRecordAvroTypeInfo(schema)); stream.addSink( StreamingFileSink.forBulkFormat( Path.fromLocalFile(folder), ParquetAvroWriters.forGenericRecord(schema)) .build()); env.execute(); List<Address> expected = Arrays.asList( new Address(1, "a", "b", "c", "12345"), new Address(2, "x", "y", "z", "98765")); validateResults(folder, SpecificData.get(), expected); } @Test public void testWriteParquetAvroReflect() throws Exception { final File folder = TEMPORARY_FOLDER.newFolder(); final List<Datum> data = Arrays.asList( new Datum("a", 1), new Datum("b", 2), new Datum("c", 3)); final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); env.setParallelism(1); env.enableCheckpointing(100); DataStream<Datum> stream = env.addSource( new FiniteTestSource<>(data), TypeInformation.of(Datum.class)); stream.addSink( StreamingFileSink.forBulkFormat( Path.fromLocalFile(folder), ParquetAvroWriters.forReflectRecord(Datum.class)) .build()); env.execute(); validateResults(folder, ReflectData.get(), data); } // ------------------------------------------------------------------------ private static <T> void validateResults(File folder, GenericData dataModel, List<T> expected) throws Exception { File[] buckets = folder.listFiles(); assertNotNull(buckets); assertEquals(1, buckets.length); File[] partFiles = buckets[0].listFiles(); assertNotNull(partFiles); assertEquals(2, partFiles.length); for (File partFile : partFiles) { assertTrue(partFile.length() > 0); final List<Tuple2<Long, String>> fileContent = readParquetFile(partFile, dataModel); assertEquals(expected, fileContent); } } private static <T> List<T> readParquetFile(File file, GenericData dataModel) throws IOException { InputFile inFile = HadoopInputFile.fromPath(new org.apache.hadoop.fs.Path(file.toURI()), new Configuration()); ArrayList<T> results = new ArrayList<>(); try (ParquetReader<T> reader = AvroParquetReader.<T>builder(inFile).withDataModel(dataModel).build()) { T next; while ((next = reader.read()) != null) { results.add(next); } } return results; } private static class GenericTestDataCollection extends AbstractCollection<GenericRecord> implements Serializable { @Override public Iterator<GenericRecord> iterator() { final GenericRecord rec1 = new GenericData.Record(Address.getClassSchema()); rec1.put(0, 1); rec1.put(1, "a"); rec1.put(2, "b"); rec1.put(3, "c"); rec1.put(4, "12345"); final GenericRecord rec2 = new GenericData.Record(Address.getClassSchema()); rec2.put(0, 2); rec2.put(1, "x"); rec2.put(2, "y"); rec2.put(3, "z"); rec2.put(4, "98765"); return Arrays.asList(rec1, rec2).iterator(); } @Override public int size() { return 2; } } // ------------------------------------------------------------------------ /** Test datum. */ public static class Datum implements Serializable { public String a; public int b; public Datum() {} public Datum(String a, int b) { this.a = a; this.b = b; } @Override public boolean equals(Object o) { if (this == o) { return true; } if (o == null || getClass() != o.getClass()) { return false; } Datum datum = (Datum) o; return b == datum.b && (a != null ? a.equals(datum.a) : datum.a == null); } @Override public int hashCode() { int result = a != null ? a.hashCode() : 0; result = 31 * result + b; return result; } } }