@@ -167,7 +167,7 @@ fn collect_small_file_parts(
167
167
}
168
168
if max_compressed_size == 0 {
169
169
// there are no large files, so we choose a default value.
170
- max_compressed_size = ( ( 128usize << 20 ) as f64 / max_compression_ratio ) as u64 ;
170
+ max_compressed_size = 128u64 << 20 ;
171
171
}
172
172
let mut num_small_files = small_files. len ( ) ;
173
173
stats. read_rows += num_small_files;
@@ -195,7 +195,7 @@ fn collect_small_file_parts(
195
195
. map ( |( path, size, dedup_key) | ParquetFilePart {
196
196
file : path,
197
197
compressed_size : size,
198
- estimated_uncompressed_size : ( size as f64 / max_compression_ratio) as u64 ,
198
+ estimated_uncompressed_size : ( size as f64 * max_compression_ratio) as u64 ,
199
199
dedup_key,
200
200
bucket_option : None ,
201
201
} )
@@ -284,18 +284,15 @@ pub(crate) fn collect_parts(
284
284
) ;
285
285
286
286
if !small_files. is_empty ( ) {
287
- let mut max_compression_ratio = compression_ratio;
288
287
let mut max_compressed_size = 0u64 ;
289
288
for part in partitions. partitions . iter ( ) {
290
289
let p = part. as_any ( ) . downcast_ref :: < ParquetPart > ( ) . unwrap ( ) ;
291
- max_compression_ratio = max_compression_ratio
292
- . max ( p. uncompressed_size ( ) as f64 / p. compressed_size ( ) as f64 ) ;
293
290
max_compressed_size = max_compressed_size. max ( p. compressed_size ( ) ) ;
294
291
}
295
292
296
293
collect_small_file_parts (
297
294
small_files,
298
- max_compression_ratio ,
295
+ compression_ratio ,
299
296
max_compressed_size,
300
297
& mut partitions,
301
298
& mut stats,
0 commit comments