Skip to content

Commit 83f6075

Browse files
authored
DRILL-8521: Upgrade parquet-avro to 1.15.1, Avro to 1.12.0 (#2986)
* Upgrade parquet-avro to 1.15.1 * Upgrade Avro to 1.12.0 * Upgrade kafka-avro-serializer to 7.9.0
1 parent 1530937 commit 83f6075

9 files changed

Lines changed: 2154 additions & 2078 deletions

File tree

contrib/storage-hive/hive-exec-shade/pom.xml

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@
3232
<name>Drill : Contrib : Storage : Hive : Exec Shaded</name>
3333

3434
<properties>
35-
<hive.parquet.version>1.8.3</hive.parquet.version>
35+
<hive.parquet.version>1.15.1</hive.parquet.version>
3636
</properties>
3737

3838
<dependencyManagement>
@@ -219,6 +219,12 @@
219219
<exclude>META-INF/versions/22/**</exclude>
220220
</excludes>
221221
</filter>
222+
<filter>
223+
<artifact>org.apache.parquet:parquet-hadoop-bundle</artifact>
224+
<excludes>
225+
<exclude>META-INF/versions/22/**</exclude>
226+
</excludes>
227+
</filter>
222228
</filters>
223229
</configuration>
224230
</plugin>

contrib/storage-kafka/pom.xml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -97,7 +97,7 @@
9797
<dependency>
9898
<groupId>io.confluent</groupId>
9999
<artifactId>kafka-avro-serializer</artifactId>
100-
<version>6.1.1</version>
100+
<version>7.9.0</version>
101101
</dependency>
102102

103103
<!-- Test dependencies -->

exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ColumnDataReader.java

Lines changed: 0 additions & 103 deletions
This file was deleted.

exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetRecordWriter.java

Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@
6969
import org.apache.parquet.column.values.factory.DefaultV2ValuesWriterFactory;
7070
import org.apache.parquet.column.values.factory.ValuesWriterFactory;
7171
import org.apache.parquet.compression.CompressionCodecFactory;
72-
import org.apache.parquet.hadoop.ParquetColumnChunkPageWriteStore;
72+
import org.apache.parquet.hadoop.ColumnChunkPageWriteStore;
7373
import org.apache.parquet.hadoop.ParquetFileWriter;
7474
import org.apache.parquet.hadoop.metadata.CompressionCodecName;
7575
import org.apache.parquet.io.ColumnIOFactory;
@@ -120,7 +120,7 @@ public class ParquetRecordWriter extends ParquetOutputRecordWriter {
120120
private long recordCountForNextMemCheck = MINIMUM_RECORD_COUNT_FOR_CHECK;
121121

122122
private ColumnWriteStore store;
123-
private ParquetColumnChunkPageWriteStore pageStore;
123+
private ColumnChunkPageWriteStore pageStore;
124124

125125
private RecordConsumer consumer;
126126
private BatchSchema batchSchema;
@@ -285,10 +285,7 @@ private void newSchema() {
285285
.withWriterVersion(writerVersion)
286286
.build();
287287

288-
// TODO: Replace ParquetColumnChunkPageWriteStore with ColumnChunkPageWriteStore from parquet library
289-
// once DRILL-7906 (PARQUET-1006) will be resolved
290-
pageStore = new ParquetColumnChunkPageWriteStore(codecFactory.getCompressor(codec), schema,
291-
parquetProperties.getInitialSlabSize(), pageSize, parquetProperties.getAllocator(),
288+
pageStore = new ColumnChunkPageWriteStore(codecFactory.getCompressor(codec), schema, parquetProperties.getAllocator(),
292289
parquetProperties.getColumnIndexTruncateLength(), parquetProperties.getPageWriteChecksumEnabled());
293290

294291
store = writerVersion == WriterVersion.PARQUET_1_0

exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet2/DrillParquetReader.java

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,6 @@
5555
import org.slf4j.Logger;
5656
import org.slf4j.LoggerFactory;
5757

58-
import java.io.IOException;
5958
import java.util.ArrayList;
6059
import java.util.Arrays;
6160
import java.util.Collection;
@@ -417,13 +416,9 @@ public void close() {
417416
recordReader = null;
418417
recordMaterializer = null;
419418
nullFilledVectors = null;
420-
try {
421-
if (pageReadStore != null) {
422-
pageReadStore.close();
423-
pageReadStore = null;
424-
}
425-
} catch (IOException e) {
426-
logger.warn("Failure while closing PageReadStore", e);
419+
if (pageReadStore != null) {
420+
pageReadStore.close();
421+
pageReadStore = null;
427422
}
428423
}
429424

exec/java-exec/src/main/java/org/apache/parquet/hadoop/ColumnChunkIncReadStore.java

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -295,9 +295,13 @@ public void addColumn(ColumnDescriptor descriptor, ColumnChunkMetaData metaData)
295295
columns.put(descriptor, reader);
296296
}
297297

298-
public void close() throws IOException {
298+
public void close() {
299299
for (FSDataInputStream stream : streams) {
300-
stream.close();
300+
try {
301+
stream.close();
302+
} catch (IOException e) {
303+
logger.warn("Error closing stream: {}", e.getMessage(), e);
304+
}
301305
}
302306
for (ColumnChunkIncPageReader reader : columns.values()) {
303307
reader.close();

0 commit comments

Comments
 (0)