@@ -395,7 +395,8 @@ mod tests {
395395 let location = Path :: from ( url. path ( ) ) ;
396396 let meta = store. head ( & location) . await . unwrap ( ) ;
397397
398- let reader = ParquetObjectReader :: new ( store. clone ( ) , meta. clone ( ) ) ;
398+ let reader = ParquetObjectReader :: new ( store. clone ( ) , meta. location )
399+ . with_file_size ( meta. size ) ;
399400 let physical_schema = ParquetRecordBatchStreamBuilder :: new ( reader)
400401 . await
401402 . unwrap ( )
@@ -405,7 +406,7 @@ mod tests {
405406 let files = & [ FileMeta {
406407 location : url. clone ( ) ,
407408 last_modified : meta. last_modified . timestamp ( ) ,
408- size : meta. size ,
409+ size : meta. size as usize ,
409410 } ] ;
410411
411412 let handler = DefaultParquetHandler :: new ( store, Arc :: new ( TokioBackgroundExecutor :: new ( ) ) ) ;
@@ -513,13 +514,14 @@ mod tests {
513514
514515 let filename = location. path ( ) . split ( '/' ) . next_back ( ) . unwrap ( ) ;
515516 assert_eq ! ( & expected_location. join( filename) . unwrap( ) , location) ;
516- assert_eq ! ( expected_size, size) ;
517+ assert_eq ! ( expected_size, size as u64 ) ;
517518 assert ! ( now - last_modified < 10_000 ) ;
518519
519520 // check we can read back
520521 let path = Path :: from ( location. path ( ) ) ;
521522 let meta = store. head ( & path) . await . unwrap ( ) ;
522- let reader = ParquetObjectReader :: new ( store. clone ( ) , meta. clone ( ) ) ;
523+ let reader = ParquetObjectReader :: new ( store. clone ( ) , meta. location )
524+ . with_file_size ( meta. size ) ;
523525 let physical_schema = ParquetRecordBatchStreamBuilder :: new ( reader)
524526 . await
525527 . unwrap ( )
0 commit comments