awswrangler/s3/_read_text.py (280 lines of code) (raw):
"""Amazon S3 Read Module (PRIVATE)."""
from __future__ import annotations
import datetime
import itertools
import logging
import pprint
from typing import TYPE_CHECKING, Any, Callable, Iterator
import boto3
import pandas as pd
from typing_extensions import Literal
from awswrangler import _utils, exceptions
from awswrangler._distributed import engine
from awswrangler._executor import _BaseExecutor, _get_executor
from awswrangler.s3._list import _path2list
from awswrangler.s3._read import (
_apply_partition_filter,
_check_version_id,
_concat_union_categoricals,
_get_num_output_blocks,
_get_path_ignore_suffix,
_get_path_root,
)
from awswrangler.s3._read_text_core import _read_text_file, _read_text_files_chunked
from awswrangler.typing import RaySettings
if TYPE_CHECKING:
from mypy_boto3_s3 import S3Client
_logger: logging.Logger = logging.getLogger(__name__)
def _resolve_format(read_format: str) -> Any:
if read_format == "csv":
return pd.read_csv
if read_format == "fwf":
return pd.read_fwf
if read_format == "json":
return pd.read_json
raise exceptions.UnsupportedType("Unsupported read format")
@engine.dispatch_on_engine
def _read_text(
read_format: str,
paths: list[str],
path_root: str | None,
use_threads: bool | int,
s3_client: "S3Client",
s3_additional_kwargs: dict[str, str] | None,
dataset: bool,
ignore_index: bool,
override_num_blocks: int,
version_ids: dict[str, str] | None,
pandas_kwargs: dict[str, Any],
) -> pd.DataFrame:
parser_func = _resolve_format(read_format)
executor: _BaseExecutor = _get_executor(use_threads=use_threads)
tables = executor.map(
_read_text_file,
s3_client,
paths,
[version_ids.get(p) if isinstance(version_ids, dict) else None for p in paths],
itertools.repeat(parser_func),
itertools.repeat(path_root),
itertools.repeat(pandas_kwargs),
itertools.repeat(s3_additional_kwargs),
itertools.repeat(dataset),
)
return _concat_union_categoricals(dfs=tables, ignore_index=ignore_index)
def _read_text_format(
read_format: str,
path: str | list[str],
path_suffix: str | list[str] | None,
path_ignore_suffix: str | list[str] | None,
ignore_empty: bool,
use_threads: bool | int,
last_modified_begin: datetime.datetime | None,
last_modified_end: datetime.datetime | None,
s3_client: "S3Client",
s3_additional_kwargs: dict[str, str] | None,
chunksize: int | None,
dataset: bool,
partition_filter: Callable[[dict[str, str]], bool] | None,
ignore_index: bool,
ray_args: RaySettings | None,
version_id: str | dict[str, str] | None = None,
**pandas_kwargs: Any,
) -> pd.DataFrame | Iterator[pd.DataFrame]:
if "iterator" in pandas_kwargs:
raise exceptions.InvalidArgument("Please, use the chunksize argument instead of iterator.")
paths: list[str] = _path2list(
path=path,
s3_client=s3_client,
suffix=path_suffix,
ignore_suffix=_get_path_ignore_suffix(path_ignore_suffix=path_ignore_suffix),
ignore_empty=ignore_empty,
last_modified_begin=last_modified_begin,
last_modified_end=last_modified_end,
s3_additional_kwargs=s3_additional_kwargs,
)
path_root: str | None = _get_path_root(path=path, dataset=dataset)
if path_root is not None:
paths = _apply_partition_filter(path_root=path_root, paths=paths, filter_func=partition_filter)
if len(paths) < 1:
raise exceptions.NoFilesFound(f"No files Found on: {path}.")
version_ids = _check_version_id(paths=paths, version_id=version_id)
args: dict[str, Any] = {
"parser_func": _resolve_format(read_format),
"s3_client": s3_client,
"dataset": dataset,
"path_root": path_root,
"pandas_kwargs": pandas_kwargs,
"s3_additional_kwargs": s3_additional_kwargs,
"use_threads": use_threads,
}
_logger.debug("Read args:\n%s", pprint.pformat(args))
if chunksize is not None:
return _read_text_files_chunked(
paths=paths,
version_ids=version_ids,
chunksize=chunksize,
**args,
)
return _read_text(
read_format,
paths=paths,
path_root=path_root,
use_threads=use_threads,
s3_client=s3_client,
s3_additional_kwargs=s3_additional_kwargs,
dataset=dataset,
ignore_index=ignore_index,
override_num_blocks=_get_num_output_blocks(ray_args),
version_ids=version_ids,
pandas_kwargs=pandas_kwargs,
)
@_utils.validate_distributed_kwargs(
unsupported_kwargs=["boto3_session"],
)
def read_csv(
path: str | list[str],
path_suffix: str | list[str] | None = None,
path_ignore_suffix: str | list[str] | None = None,
version_id: str | dict[str, str] | None = None,
ignore_empty: bool = True,
use_threads: bool | int = True,
last_modified_begin: datetime.datetime | None = None,
last_modified_end: datetime.datetime | None = None,
boto3_session: boto3.Session | None = None,
s3_additional_kwargs: dict[str, Any] | None = None,
dtype_backend: Literal["numpy_nullable", "pyarrow"] = "numpy_nullable",
chunksize: int | None = None,
dataset: bool = False,
partition_filter: Callable[[dict[str, str]], bool] | None = None,
ray_args: RaySettings | None = None,
**pandas_kwargs: Any,
) -> pd.DataFrame | Iterator[pd.DataFrame]:
"""Read CSV file(s) from a received S3 prefix or list of S3 objects paths.
This function accepts Unix shell-style wildcards in the path argument.
* (matches everything), ? (matches any single character),
[seq] (matches any character in seq), [!seq] (matches any character not in seq).
If you want to use a path which includes Unix shell-style wildcard characters (`*, ?, []`),
you can use `glob.escape(path)` before passing the path to this function.
Note
----
For partial and gradual reading use the argument ``chunksize`` instead of ``iterator``.
Note
----
In case of `use_threads=True` the number of threads
that will be spawned will be gotten from os.cpu_count().
Note
----
The filter by last_modified begin last_modified end is applied after list all S3 files
Parameters
----------
path
S3 prefix (accepts Unix shell-style wildcards)
(e.g. s3://bucket/prefix) or list of S3 objects paths (e.g. ``[s3://bucket/key0, s3://bucket/key1]``).
path_suffix
Suffix or List of suffixes to be read (e.g. [".csv"]).
If None, will try to read all files. (default)
path_ignore_suffix
Suffix or List of suffixes for S3 keys to be ignored.(e.g. ["_SUCCESS"]).
If None, will try to read all files. (default)
version_id
Version id of the object or mapping of object path to version id.
(e.g. {'s3://bucket/key0': '121212', 's3://bucket/key1': '343434'})
ignore_empty
Ignore files with 0 bytes.
use_threads
True to enable concurrent requests, False to disable multiple threads.
If enabled os.cpu_count() will be used as the max number of threads.
If integer is provided, specified number is used.
last_modified_begin
Filter the s3 files by the Last modified date of the object.
The filter is applied only after list all s3 files.
last_modified_end
Filter the s3 files by the Last modified date of the object.
The filter is applied only after list all s3 files.
boto3_session
Boto3 Session. The default boto3 session will be used if boto3_session receive None.
pyarrow_additional_kwargs
Forward to botocore requests, only "SSECustomerAlgorithm" and "SSECustomerKey" arguments will be considered.
dtype_backend
Which dtype_backend to use, e.g. whether a DataFrame should have NumPy arrays,
nullable dtypes are used for all dtypes that have a nullable implementation when
“numpy_nullable” is set, pyarrow is used for all dtypes if “pyarrow” is set.
The dtype_backends are still experimential. The "pyarrow" backend is only supported with Pandas 2.0 or above.
chunksize
If specified, return an generator where chunksize is the number of rows to include in each chunk.
dataset
If `True` read a CSV dataset instead of simple file(s) loading all the related partitions as columns.
partition_filter
Callback Function filters to apply on PARTITION columns (PUSH-DOWN filter).
This function MUST receive a single argument (Dict[str, str]) where keys are partitions
names and values are partitions values. Partitions values will be always strings extracted from S3.
This function MUST return a bool, True to read the partition or False to ignore it.
Ignored if `dataset=False`.
E.g ``lambda x: True if x["year"] == "2020" and x["month"] == "1" else False``
https://aws-sdk-pandas.readthedocs.io/en/3.11.0/tutorials/023%20-%20Flexible%20Partitions%20Filter.html
s3_additional_kwargs
Forwarded to botocore requests.
ray_args
Parameters of the Ray Modin settings. Only used when distributed computing is used with Ray and Modin installed.
pandas_kwargs
KEYWORD arguments forwarded to pandas.read_csv(). You can NOT pass `pandas_kwargs` explicitly, just add valid
Pandas arguments in the function call and awswrangler will accept it.
e.g. wr.s3.read_csv('s3://bucket/prefix/', sep='|', na_values=['null', 'none'], skip_blank_lines=True)
https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.read_csv.html
Returns
-------
Pandas DataFrame or a Generator in case of `chunksize != None`.
Examples
--------
Reading all CSV files under a prefix
>>> import awswrangler as wr
>>> df = wr.s3.read_csv(path='s3://bucket/prefix/')
Reading all CSV files under a prefix and using pandas_kwargs
>>> import awswrangler as wr
>>> df = wr.s3.read_csv('s3://bucket/prefix/', sep='|', na_values=['null', 'none'], skip_blank_lines=True)
Reading all CSV files from a list
>>> import awswrangler as wr
>>> df = wr.s3.read_csv(path=['s3://bucket/filename0.csv', 's3://bucket/filename1.csv'])
Reading in chunks of 100 lines
>>> import awswrangler as wr
>>> dfs = wr.s3.read_csv(path=['s3://bucket/filename0.csv', 's3://bucket/filename1.csv'], chunksize=100)
>>> for df in dfs:
>>> print(df) # 100 lines Pandas DataFrame
Reading CSV Dataset with PUSH-DOWN filter over partitions
>>> import awswrangler as wr
>>> my_filter = lambda x: True if x["city"].startswith("new") else False
>>> df = wr.s3.read_csv(path, dataset=True, partition_filter=my_filter)
"""
if "pandas_kwargs" in pandas_kwargs:
raise exceptions.InvalidArgument(
"You can NOT pass `pandas_kwargs` explicitly, just add valid "
"Pandas arguments in the function call and awswrangler will accept it."
"e.g. wr.s3.read_csv('s3://bucket/prefix/', sep='|', skip_blank_lines=True)"
)
if dtype_backend != "numpy_nullable":
pandas_kwargs["dtype_backend"] = dtype_backend
s3_client = _utils.client(service_name="s3", session=boto3_session)
ignore_index: bool = "index_col" not in pandas_kwargs
return _read_text_format(
read_format="csv",
path=path,
path_suffix=path_suffix,
path_ignore_suffix=path_ignore_suffix,
version_id=version_id,
ignore_empty=ignore_empty,
use_threads=use_threads,
s3_client=s3_client,
s3_additional_kwargs=s3_additional_kwargs,
chunksize=chunksize,
dataset=dataset,
partition_filter=partition_filter,
last_modified_begin=last_modified_begin,
last_modified_end=last_modified_end,
ignore_index=ignore_index,
ray_args=ray_args,
**pandas_kwargs,
)
@_utils.validate_distributed_kwargs(
unsupported_kwargs=["boto3_session"],
)
def read_fwf(
path: str | list[str],
path_suffix: str | list[str] | None = None,
path_ignore_suffix: str | list[str] | None = None,
version_id: str | dict[str, str] | None = None,
ignore_empty: bool = True,
use_threads: bool | int = True,
last_modified_begin: datetime.datetime | None = None,
last_modified_end: datetime.datetime | None = None,
boto3_session: boto3.Session | None = None,
s3_additional_kwargs: dict[str, Any] | None = None,
chunksize: int | None = None,
dataset: bool = False,
partition_filter: Callable[[dict[str, str]], bool] | None = None,
ray_args: RaySettings | None = None,
**pandas_kwargs: Any,
) -> pd.DataFrame | Iterator[pd.DataFrame]:
"""Read fixed-width formatted file(s) from a received S3 prefix or list of S3 objects paths.
This function accepts Unix shell-style wildcards in the path argument.
* (matches everything), ? (matches any single character),
[seq] (matches any character in seq), [!seq] (matches any character not in seq).
If you want to use a path which includes Unix shell-style wildcard characters (`*, ?, []`),
you can use `glob.escape(path)` before passing the path to this function.
Note
----
For partial and gradual reading use the argument ``chunksize`` instead of ``iterator``.
Note
----
In case of `use_threads=True` the number of threads
that will be spawned will be gotten from os.cpu_count().
Note
----
The filter by last_modified begin last_modified end is applied after list all S3 files
Parameters
----------
path
S3 prefix (accepts Unix shell-style wildcards)
(e.g. s3://bucket/prefix) or list of S3 objects paths (e.g. ``[s3://bucket/key0, s3://bucket/key1]``).
path_suffix
Suffix or List of suffixes to be read (e.g. [".txt"]).
If None, will try to read all files. (default)
path_ignore_suffix
Suffix or List of suffixes for S3 keys to be ignored.(e.g. ["_SUCCESS"]).
If None, will try to read all files. (default)
version_id
Version id of the object or mapping of object path to version id.
(e.g. {'s3://bucket/key0': '121212', 's3://bucket/key1': '343434'})
ignore_empty
Ignore files with 0 bytes.
use_threads
True to enable concurrent requests, False to disable multiple threads.
If enabled os.cpu_count() will be used as the max number of threads.
If integer is provided, specified number is used.
last_modified_begin
Filter the s3 files by the Last modified date of the object.
The filter is applied only after list all s3 files.
last_modified_end
Filter the s3 files by the Last modified date of the object.
The filter is applied only after list all s3 files.
boto3_session
Boto3 Session. The default boto3 session will be used if boto3_session receive None.
pyarrow_additional_kwargs
Forward to botocore requests, only "SSECustomerAlgorithm" and "SSECustomerKey" arguments will be considered.
chunksize
If specified, return an generator where chunksize is the number of rows to include in each chunk.
dataset
If `True` read a FWF dataset instead of simple file(s) loading all the related partitions as columns.
partition_filter
Callback Function filters to apply on PARTITION columns (PUSH-DOWN filter).
This function MUST receive a single argument (Dict[str, str]) where keys are partitions
names and values are partitions values. Partitions values will be always strings extracted from S3.
This function MUST return a bool, True to read the partition or False to ignore it.
Ignored if `dataset=False`.
E.g ``lambda x: True if x["year"] == "2020" and x["month"] == "1" else False``
https://aws-sdk-pandas.readthedocs.io/en/3.11.0/tutorials/023%20-%20Flexible%20Partitions%20Filter.html
s3_additional_kwargs
Forwarded to botocore requests.
ray_args
Parameters of the Ray Modin settings. Only used when distributed computing is used with Ray and Modin installed.
pandas_kwargs:
KEYWORD arguments forwarded to pandas.read_fwf(). You can NOT pass `pandas_kwargs` explicit, just add valid
Pandas arguments in the function call and awswrangler will accept it.
e.g. wr.s3.read_fwf(path='s3://bucket/prefix/', widths=[1, 3], names=["c0", "c1"])
https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.read_fwf.html
Returns
-------
Pandas DataFrame or a Generator in case of `chunksize != None`.
Examples
--------
Reading all fixed-width formatted (FWF) files under a prefix
>>> import awswrangler as wr
>>> df = wr.s3.read_fwf(path='s3://bucket/prefix/', widths=[1, 3], names=['c0', 'c1'])
Reading all fixed-width formatted (FWF) files from a list
>>> import awswrangler as wr
>>> df = wr.s3.read_fwf(path=['s3://bucket/0.txt', 's3://bucket/1.txt'], widths=[1, 3], names=['c0', 'c1'])
Reading in chunks of 100 lines
>>> import awswrangler as wr
>>> dfs = wr.s3.read_fwf(
... path=['s3://bucket/0.txt', 's3://bucket/1.txt'],
... chunksize=100,
... widths=[1, 3],
... names=["c0", "c1"]
... )
>>> for df in dfs:
>>> print(df) # 100 lines Pandas DataFrame
Reading FWF Dataset with PUSH-DOWN filter over partitions
>>> import awswrangler as wr
>>> my_filter = lambda x: True if x["city"].startswith("new") else False
>>> df = wr.s3.read_fwf(path, dataset=True, partition_filter=my_filter, widths=[1, 3], names=["c0", "c1"])
"""
if "pandas_kwargs" in pandas_kwargs:
raise exceptions.InvalidArgument(
"You can NOT pass `pandas_kwargs` explicit, just add valid "
"Pandas arguments in the function call and awswrangler will accept it."
"e.g. wr.s3.read_fwf(path, widths=[1, 3], names=['c0', 'c1'])"
)
s3_client = _utils.client(service_name="s3", session=boto3_session)
return _read_text_format(
read_format="fwf",
path=path,
path_suffix=path_suffix,
path_ignore_suffix=path_ignore_suffix,
version_id=version_id,
ignore_empty=ignore_empty,
use_threads=use_threads,
s3_client=s3_client,
s3_additional_kwargs=s3_additional_kwargs,
chunksize=chunksize,
dataset=dataset,
partition_filter=partition_filter,
last_modified_begin=last_modified_begin,
last_modified_end=last_modified_end,
ignore_index=True,
sort_index=False,
ray_args=ray_args,
**pandas_kwargs,
)
@_utils.validate_distributed_kwargs(
unsupported_kwargs=["boto3_session"],
)
def read_json(
path: str | list[str],
path_suffix: str | list[str] | None = None,
path_ignore_suffix: str | list[str] | None = None,
version_id: str | dict[str, str] | None = None,
ignore_empty: bool = True,
orient: str = "columns",
use_threads: bool | int = True,
last_modified_begin: datetime.datetime | None = None,
last_modified_end: datetime.datetime | None = None,
boto3_session: boto3.Session | None = None,
s3_additional_kwargs: dict[str, Any] | None = None,
dtype_backend: Literal["numpy_nullable", "pyarrow"] = "numpy_nullable",
chunksize: int | None = None,
dataset: bool = False,
partition_filter: Callable[[dict[str, str]], bool] | None = None,
ray_args: RaySettings | None = None,
**pandas_kwargs: Any,
) -> pd.DataFrame | Iterator[pd.DataFrame]:
"""Read JSON file(s) from a received S3 prefix or list of S3 objects paths.
This function accepts Unix shell-style wildcards in the path argument.
* (matches everything), ? (matches any single character),
[seq] (matches any character in seq), [!seq] (matches any character not in seq).
If you want to use a path which includes Unix shell-style wildcard characters (`*, ?, []`),
you can use `glob.escape(path)` before passing the path to this function.
Note
----
For partial and gradual reading use the argument ``chunksize`` instead of ``iterator``.
Note
----
In case of `use_threads=True` the number of threads
that will be spawned will be gotten from os.cpu_count().
Note
----
The filter by last_modified begin last_modified end is applied after list all S3 files
Parameters
----------
path
S3 prefix (accepts Unix shell-style wildcards)
(e.g. s3://bucket/prefix) or list of S3 objects paths (e.g. ``[s3://bucket/key0, s3://bucket/key1]``).
path_suffix
Suffix or List of suffixes to be read (e.g. [".json"]).
If None, will try to read all files. (default)
path_ignore_suffix
Suffix or List of suffixes for S3 keys to be ignored.(e.g. ["_SUCCESS"]).
If None, will try to read all files. (default)
version_id
Version id of the object or mapping of object path to version id.
(e.g. {'s3://bucket/key0': '121212', 's3://bucket/key1': '343434'})
ignore_empty
Ignore files with 0 bytes.
orient
Same as Pandas: https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.read_json.html
use_threads
True to enable concurrent requests, False to disable multiple threads.
If enabled os.cpu_count() will be used as the max number of threads.
If integer is provided, specified number is used.
last_modified_begin
Filter the s3 files by the Last modified date of the object.
The filter is applied only after list all s3 files.
last_modified_end
Filter the s3 files by the Last modified date of the object.
The filter is applied only after list all s3 files.
boto3_session
Boto3 Session. The default boto3 session will be used if boto3_session receive None.
pyarrow_additional_kwargs
Forward to botocore requests, only "SSECustomerAlgorithm" and "SSECustomerKey" arguments will be considered.
dtype_backend
Which dtype_backend to use, e.g. whether a DataFrame should have NumPy arrays,
nullable dtypes are used for all dtypes that have a nullable implementation when
“numpy_nullable” is set, pyarrow is used for all dtypes if “pyarrow” is set.
The dtype_backends are still experimential. The "pyarrow" backend is only supported with Pandas 2.0 or above.
chunksize
If specified, return an generator where chunksize is the number of rows to include in each chunk.
dataset
If `True` read a JSON dataset instead of simple file(s) loading all the related partitions as columns.
If `True`, the `lines=True` will be assumed by default.
partition_filter
Callback Function filters to apply on PARTITION columns (PUSH-DOWN filter).
This function MUST receive a single argument (Dict[str, str]) where keys are partitions
names and values are partitions values. Partitions values will be always strings extracted from S3.
This function MUST return a bool, True to read the partition or False to ignore it.
Ignored if `dataset=False`.
E.g ``lambda x: True if x["year"] == "2020" and x["month"] == "1" else False``
https://aws-sdk-pandas.readthedocs.io/en/3.11.0/tutorials/023%20-%20Flexible%20Partitions%20Filter.html
s3_additional_kwargs
Forwarded to botocore requests.
ray_args
Parameters of the Ray Modin settings. Only used when distributed computing is used with Ray and Modin installed.
pandas_kwargs:
KEYWORD arguments forwarded to pandas.read_json(). You can NOT pass `pandas_kwargs` explicit, just add valid
Pandas arguments in the function call and awswrangler will accept it.
e.g. wr.s3.read_json('s3://bucket/prefix/', lines=True, keep_default_dates=True)
https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.read_json.html
Returns
-------
Pandas DataFrame or a Generator in case of `chunksize != None`.
Examples
--------
Reading all JSON files under a prefix
>>> import awswrangler as wr
>>> df = wr.s3.read_json(path='s3://bucket/prefix/')
Reading all CSV files under a prefix and using pandas_kwargs
>>> import awswrangler as wr
>>> df = wr.s3.read_json('s3://bucket/prefix/', lines=True, keep_default_dates=True)
Reading all JSON files from a list
>>> import awswrangler as wr
>>> df = wr.s3.read_json(path=['s3://bucket/filename0.json', 's3://bucket/filename1.json'])
Reading in chunks of 100 lines
>>> import awswrangler as wr
>>> dfs = wr.s3.read_json(path=['s3://bucket/0.json', 's3://bucket/1.json'], chunksize=100, lines=True)
>>> for df in dfs:
>>> print(df) # 100 lines Pandas DataFrame
Reading JSON Dataset with PUSH-DOWN filter over partitions
>>> import awswrangler as wr
>>> my_filter = lambda x: True if x["city"].startswith("new") else False
>>> df = wr.s3.read_json(path, dataset=True, partition_filter=my_filter)
"""
if "pandas_kwargs" in pandas_kwargs:
raise exceptions.InvalidArgument(
"You can NOT pass `pandas_kwargs` explicit, just add valid "
"Pandas arguments in the function call and awswrangler will accept it."
"e.g. wr.s3.read_json(path, lines=True, keep_default_dates=True)"
)
if dtype_backend != "numpy_nullable":
pandas_kwargs["dtype_backend"] = dtype_backend
s3_client = _utils.client(service_name="s3", session=boto3_session)
if (dataset is True) and ("lines" not in pandas_kwargs):
pandas_kwargs["lines"] = True
pandas_kwargs["orient"] = orient
ignore_index: bool = orient not in ("split", "index", "columns")
return _read_text_format(
read_format="json",
path=path,
path_suffix=path_suffix,
path_ignore_suffix=path_ignore_suffix,
version_id=version_id,
ignore_empty=ignore_empty,
use_threads=use_threads,
s3_client=s3_client,
s3_additional_kwargs=s3_additional_kwargs,
chunksize=chunksize,
dataset=dataset,
partition_filter=partition_filter,
last_modified_begin=last_modified_begin,
last_modified_end=last_modified_end,
ignore_index=ignore_index,
ray_args=ray_args,
**pandas_kwargs,
)