@@ -976,6 +976,7 @@ def merge(
976
976
error_on_type_mismatch : bool = True ,
977
977
writer_properties : Optional [WriterProperties ] = None ,
978
978
large_dtypes : Optional [bool ] = None ,
979
+ streaming : bool = False ,
979
980
custom_metadata : Optional [Dict [str , str ]] = None ,
980
981
post_commithook_properties : Optional [PostCommitHookProperties ] = None ,
981
982
commit_properties : Optional [CommitProperties ] = None ,
@@ -993,6 +994,7 @@ def merge(
993
994
error_on_type_mismatch: specify if merge will return error if data types are mismatching :default = True
994
995
writer_properties: Pass writer properties to the Rust parquet writer
995
996
large_dtypes: Deprecated, will be removed in 1.0
997
+ streaming: Will execute MERGE using a LazyMemoryExec plan
996
998
arrow_schema_conversion_mode: Large converts all types of data schema into Large Arrow types, passthrough keeps string/binary/list types untouched
997
999
custom_metadata: Deprecated and will be removed in future versions. Use commit_properties instead.
998
1000
post_commithook_properties: properties for the post commit hook. If None, default values are used.
@@ -1031,17 +1033,14 @@ def merge(
1031
1033
convert_pyarrow_table ,
1032
1034
)
1033
1035
1034
- streaming = False
1035
1036
if isinstance (source , pyarrow .RecordBatchReader ):
1036
1037
source = convert_pyarrow_recordbatchreader (source , conversion_mode )
1037
- streaming = True
1038
1038
elif isinstance (source , pyarrow .RecordBatch ):
1039
1039
source = convert_pyarrow_recordbatch (source , conversion_mode )
1040
1040
elif isinstance (source , pyarrow .Table ):
1041
1041
source = convert_pyarrow_table (source , conversion_mode )
1042
1042
elif isinstance (source , ds .Dataset ):
1043
1043
source = convert_pyarrow_dataset (source , conversion_mode )
1044
- streaming = True
1045
1044
elif _has_pandas and isinstance (source , pd .DataFrame ):
1046
1045
source = convert_pyarrow_table (
1047
1046
pyarrow .Table .from_pandas (source ), conversion_mode
0 commit comments