@@ -23,56 +23,53 @@ use std::sync::Arc;
2323use arrow:: array:: RecordBatchReader ;
2424use arrow:: ffi_stream:: ArrowArrayStreamReader ;
2525use arrow:: pyarrow:: FromPyArrow ;
26- use datafusion:: execution:: session_state:: SessionStateBuilder ;
27- use object_store:: ObjectStore ;
28- use url:: Url ;
29- use uuid:: Uuid ;
30-
31- use pyo3:: exceptions:: { PyKeyError , PyValueError } ;
32- use pyo3:: prelude:: * ;
33-
34- use crate :: catalog:: { PyCatalog , RustWrappedPyCatalogProvider } ;
35- use crate :: dataframe:: PyDataFrame ;
36- use crate :: dataset:: Dataset ;
37- use crate :: errors:: { py_datafusion_err, PyDataFusionResult } ;
38- use crate :: expr:: sort_expr:: PySortExpr ;
39- use crate :: physical_plan:: PyExecutionPlan ;
40- use crate :: record_batch:: PyRecordBatchStream ;
41- use crate :: sql:: exceptions:: py_value_err;
42- use crate :: sql:: logical:: PyLogicalPlan ;
43- use crate :: store:: StorageContexts ;
44- use crate :: table:: PyTable ;
45- use crate :: udaf:: PyAggregateUDF ;
46- use crate :: udf:: PyScalarUDF ;
47- use crate :: udtf:: PyTableFunction ;
48- use crate :: udwf:: PyWindowUDF ;
49- use crate :: utils:: { get_global_ctx, spawn_future, validate_pycapsule, wait_for_future} ;
5026use datafusion:: arrow:: datatypes:: { DataType , Schema , SchemaRef } ;
5127use datafusion:: arrow:: pyarrow:: PyArrowType ;
5228use datafusion:: arrow:: record_batch:: RecordBatch ;
5329use datafusion:: catalog:: CatalogProvider ;
54- use datafusion:: common:: TableReference ;
55- use datafusion:: common:: { exec_err, ScalarValue } ;
30+ use datafusion:: common:: { exec_err, ScalarValue , TableReference } ;
5631use datafusion:: datasource:: file_format:: file_compression_type:: FileCompressionType ;
5732use datafusion:: datasource:: file_format:: parquet:: ParquetFormat ;
5833use datafusion:: datasource:: listing:: {
5934 ListingOptions , ListingTable , ListingTableConfig , ListingTableUrl ,
6035} ;
61- use datafusion:: datasource:: MemTable ;
62- use datafusion:: datasource:: TableProvider ;
36+ use datafusion:: datasource:: { MemTable , TableProvider } ;
6337use datafusion:: execution:: context:: {
6438 DataFilePaths , SQLOptions , SessionConfig , SessionContext , TaskContext ,
6539} ;
6640use datafusion:: execution:: disk_manager:: DiskManagerMode ;
6741use datafusion:: execution:: memory_pool:: { FairSpillPool , GreedyMemoryPool , UnboundedMemoryPool } ;
6842use datafusion:: execution:: options:: ReadOptions ;
6943use datafusion:: execution:: runtime_env:: RuntimeEnvBuilder ;
44+ use datafusion:: execution:: session_state:: SessionStateBuilder ;
7045use datafusion:: prelude:: {
7146 AvroReadOptions , CsvReadOptions , DataFrame , NdJsonReadOptions , ParquetReadOptions ,
7247} ;
7348use datafusion_ffi:: catalog_provider:: { FFI_CatalogProvider , ForeignCatalogProvider } ;
49+ use object_store:: ObjectStore ;
50+ use pyo3:: exceptions:: { PyKeyError , PyValueError } ;
51+ use pyo3:: prelude:: * ;
7452use pyo3:: types:: { PyCapsule , PyDict , PyList , PyTuple , PyType } ;
7553use pyo3:: IntoPyObjectExt ;
54+ use url:: Url ;
55+ use uuid:: Uuid ;
56+
57+ use crate :: catalog:: { PyCatalog , RustWrappedPyCatalogProvider } ;
58+ use crate :: dataframe:: PyDataFrame ;
59+ use crate :: dataset:: Dataset ;
60+ use crate :: errors:: { py_datafusion_err, PyDataFusionResult } ;
61+ use crate :: expr:: sort_expr:: PySortExpr ;
62+ use crate :: physical_plan:: PyExecutionPlan ;
63+ use crate :: record_batch:: PyRecordBatchStream ;
64+ use crate :: sql:: exceptions:: py_value_err;
65+ use crate :: sql:: logical:: PyLogicalPlan ;
66+ use crate :: store:: StorageContexts ;
67+ use crate :: table:: PyTable ;
68+ use crate :: udaf:: PyAggregateUDF ;
69+ use crate :: udf:: PyScalarUDF ;
70+ use crate :: udtf:: PyTableFunction ;
71+ use crate :: udwf:: PyWindowUDF ;
72+ use crate :: utils:: { get_global_ctx, spawn_future, validate_pycapsule, wait_for_future} ;
7673
7774/// Configuration options for a SessionContext
7875#[ pyclass( frozen, name = "SessionConfig" , module = "datafusion" , subclass) ]
0 commit comments