in awswrangler/redshift/_write.py [0:0]
def copy_from_files( # noqa: PLR0913
path: str,
con: "redshift_connector.Connection",
table: str,
schema: str,
iam_role: str | None = None,
aws_access_key_id: str | None = None,
aws_secret_access_key: str | None = None,
aws_session_token: str | None = None,
data_format: _CopyFromFilesDataFormatLiteral = "parquet",
redshift_column_types: dict[str, str] | None = None,
parquet_infer_sampling: float = 1.0,
mode: _ToSqlModeLiteral = "append",
overwrite_method: _ToSqlOverwriteModeLiteral = "drop",
diststyle: _ToSqlDistStyleLiteral = "AUTO",
distkey: str | None = None,
sortstyle: _ToSqlSortStyleLiteral = "COMPOUND",
sortkey: list[str] | None = None,
primary_keys: list[str] | None = None,
varchar_lengths_default: int = 256,
varchar_lengths: dict[str, int] | None = None,
serialize_to_json: bool = False,
path_suffix: str | None = None,
path_ignore_suffix: str | list[str] | None = None,
use_threads: bool | int = True,
lock: bool = False,
commit_transaction: bool = True,
manifest: bool | None = False,
sql_copy_extra_params: list[str] | None = None,
boto3_session: boto3.Session | None = None,
s3_additional_kwargs: dict[str, str] | None = None,
precombine_key: str | None = None,
column_names: list[str] | None = None,
add_new_columns: bool = False,