in awswrangler/s3/_write.py [0:0]
def write( # noqa: PLR0913
self,
df: pd.DataFrame,
path: str | None,
index: bool,
compression: str | None,
pyarrow_additional_kwargs: dict[str, Any],
max_rows_by_file: int | None,
use_threads: bool | int,
boto3_session: boto3.Session | None,
s3_additional_kwargs: dict[str, Any] | None,
sanitize_columns: bool,
dataset: bool,
filename_prefix: str | None,
partition_cols: list[str] | None,
bucketing_info: typing.BucketingInfoTuple | None,
concurrent_partitioning: bool,
mode: str | None,
catalog_versioning: bool,
schema_evolution: bool,
database: str | None,
table: str | None,
description: str | None,
parameters: dict[str, str] | None,
columns_comments: dict[str, str] | None,
columns_parameters: dict[str, dict[str, str]] | None,
regular_partitions: bool,
table_type: str | None,
dtype: dict[str, str] | None,
athena_partition_projection_settings: typing.AthenaPartitionProjectionSettings | None,
catalog_id: str | None,
compression_ext: str,
encryption_configuration: typing.ArrowEncryptionConfiguration | None,