@@ -27,12 +27,12 @@ use arrow::util::display::{ArrayFormatter, FormatOptions};
2727use datafusion:: arrow:: datatypes:: Schema ;
2828use datafusion:: arrow:: pyarrow:: { PyArrowType , ToPyArrow } ;
2929use datafusion:: arrow:: util:: pretty;
30+ use datafusion:: common:: UnnestOptions ;
3031use datafusion:: config:: { CsvOptions , TableParquetOptions } ;
3132use datafusion:: dataframe:: { DataFrame , DataFrameWriteOptions } ;
3233use datafusion:: execution:: SendableRecordBatchStream ;
3334use datafusion:: parquet:: basic:: { BrotliLevel , Compression , GzipLevel , ZstdLevel } ;
3435use datafusion:: prelude:: * ;
35- use datafusion_common:: UnnestOptions ;
3636use pyo3:: exceptions:: { PyTypeError , PyValueError } ;
3737use pyo3:: prelude:: * ;
3838use pyo3:: pybacked:: PyBackedStr ;
@@ -541,7 +541,7 @@ impl PyDataFrame {
541541 // create a Tokio runtime to run the async code
542542 let rt = & get_tokio_runtime ( py) . 0 ;
543543 let df = self . df . as_ref ( ) . clone ( ) ;
544- let fut: JoinHandle < datafusion_common :: Result < SendableRecordBatchStream > > =
544+ let fut: JoinHandle < datafusion :: common :: Result < SendableRecordBatchStream > > =
545545 rt. spawn ( async move { df. execute_stream ( ) . await } ) ;
546546 let stream = wait_for_future ( py, fut) . map_err ( py_datafusion_err) ?;
547547 Ok ( PyRecordBatchStream :: new ( stream?) )
@@ -551,7 +551,7 @@ impl PyDataFrame {
551551 // create a Tokio runtime to run the async code
552552 let rt = & get_tokio_runtime ( py) . 0 ;
553553 let df = self . df . as_ref ( ) . clone ( ) ;
554- let fut: JoinHandle < datafusion_common :: Result < Vec < SendableRecordBatchStream > > > =
554+ let fut: JoinHandle < datafusion :: common :: Result < Vec < SendableRecordBatchStream > > > =
555555 rt. spawn ( async move { df. execute_stream_partitioned ( ) . await } ) ;
556556 let stream = wait_for_future ( py, fut) . map_err ( py_datafusion_err) ?;
557557
0 commit comments