Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
136 changes: 136 additions & 0 deletions parquet/tests/arrow_reader/bad_data.rs
Original file line number Diff line number Diff line change
@@ -0,0 +1,136 @@
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.

//! Tests that reading invalid parquet files returns an error

use arrow::util::test_util::parquet_test_data;
use parquet::arrow::arrow_reader::ArrowReaderBuilder;
use parquet::errors::ParquetError;
use std::collections::HashSet;
use std::path::PathBuf;

static KNOWN_FILES: &[&str] = &[
"PARQUET-1481.parquet",
"ARROW-GH-41317.parquet",
"ARROW-GH-41321.parquet",
"ARROW-RS-GH-6229-DICTHEADER.parquet",
"ARROW-RS-GH-6229-LEVELS.parquet",
"README.md",
];

/// Returns the path to 'parquet-testing/bad_data'
fn bad_data_dir() -> PathBuf {
// points to parquet-testing/data
let parquet_testing_data = parquet_test_data();
PathBuf::from(parquet_testing_data)
.parent()
.expect("was in parquet-testing/data")
.join("bad_data")
}

#[test]
// Ensure that if we add a new test the files are added to the tests.
fn test_invalid_files() {
let known_files: HashSet<_> = KNOWN_FILES.iter().cloned().collect();
let mut seen_files = HashSet::new();

let files = std::fs::read_dir(bad_data_dir()).unwrap();

for file in files {
let file_name = file
.unwrap()
.path()
.file_name()
.unwrap()
.to_str()
.unwrap()
.to_string();

// If you see this error, please add a test for the new file following the model below
assert!(
known_files.contains(file_name.as_str()),
"Found new file in bad_data, please add test: {file_name}"
);
seen_files.insert(file_name);
}
for expected_file in known_files {
assert!(
seen_files.contains(expected_file),
"Expected file not found in bad_data directory: {expected_file}"
);
}
}

#[test]
fn test_parquet_1481() {
let err = read_file("PARQUET-1481.parquet").unwrap_err();
assert_eq!(
err.to_string(),
"Parquet error: unexpected parquet type: -7"
);
}

#[test]
#[should_panic(expected = "assertion failed: self.current_value.is_some()")]
fn test_arrow_gh_41321() {
let err = read_file("ARROW-GH-41321.parquet").unwrap_err();
assert_eq!(err.to_string(), "TBD (currently panics)");
}

#[test]
fn test_arrow_gh_41317() {
let err = read_file("ARROW-GH-41317.parquet").unwrap_err();
assert_eq!(
err.to_string(),
"External: Parquet argument error: External: bad data"
);
}

#[test]
fn test_arrow_rs_gh_6229_dict_header() {
let err = read_file("ARROW-RS-GH-6229-DICTHEADER.parquet").unwrap_err();
assert_eq!(
err.to_string(),
"External: Parquet argument error: EOF: eof decoding byte array"
);
}

#[test]
fn test_arrow_rs_gh_6229_dict_levels() {
let err = read_file("ARROW-RS-GH-6229-LEVELS.parquet").unwrap_err();
assert_eq!(
err.to_string(),
"External: Parquet argument error: Parquet error: Insufficient repetition levels read from column"
);
}

/// Reads the file and tries to return the total row count
/// Returns an error if the file is invalid
fn read_file(name: &str) -> Result<usize, ParquetError> {
let path = bad_data_dir().join(name);
println!("Reading file: {:?}", path);

let file = std::fs::File::open(&path).unwrap();
let reader = ArrowReaderBuilder::try_new(file)?.build()?;

let mut num_rows = 0;
for batch in reader {
let batch = batch?;
num_rows += batch.num_rows();
}
Ok(num_rows)
}
1 change: 1 addition & 0 deletions parquet/tests/arrow_reader/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -35,6 +35,7 @@ use parquet::file::properties::{EnabledStatistics, WriterProperties};
use std::sync::Arc;
use tempfile::NamedTempFile;

mod bad_data;
mod statistics;

// returns a struct array with columns "int32_col", "float32_col" and "float64_col" with the specified values
Expand Down