python/datafusion/input/location.py (44 lines of code) (raw):
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
"""The default input source for DataFusion."""
import glob
from pathlib import Path
from typing import Any
from datafusion.common import DataTypeMap, SqlTable
from datafusion.input.base import BaseInputSource
class LocationInputPlugin(BaseInputSource):
"""Input Plugin for everything.
This can be read in from a file (on disk, remote etc.).
"""
def is_correct_input(self, input_item: Any, table_name: str, **kwargs: Any) -> bool: # noqa: ARG002
"""Returns `True` if the input is valid."""
return isinstance(input_item, str)
def build_table(
self,
input_item: str,
table_name: str,
**kwargs: Any, # noqa: ARG002
) -> SqlTable: # type: ignore[invalid-type-form]
"""Create a table from the input source."""
extension = Path(input_item).suffix
file_format = extension.lstrip(".").lower()
num_rows = 0 # Total number of rows in the file. Used for statistics
columns = []
if file_format == "parquet":
import pyarrow.parquet as pq
# Read the Parquet metadata
metadata = pq.read_metadata(input_item)
num_rows = metadata.num_rows
# Iterate through the schema and build the SqlTable
columns = [
(
col.name,
DataTypeMap.from_parquet_type_str(col.physical_type),
)
for col in metadata.schema
]
elif format == "csv":
import csv
# Consume header row and count number of rows for statistics.
# TODO: Possibly makes sense to have the eager number of rows
# calculated as a configuration since you must read the entire file
# to get that information. However, this should only be occurring
# at table creation time and therefore shouldn't
# slow down query performance.
with Path(input_item).open() as file:
reader = csv.reader(file)
_header_row = next(reader)
for _ in reader:
num_rows += 1
# TODO: Need to actually consume this row into reasonable columns
msg = "TODO: Currently unable to support CSV input files."
raise RuntimeError(msg)
else:
msg = f"Input of format: `{format}` is currently not supported.\
Only Parquet and CSV."
raise RuntimeError(msg)
# Input could possibly be multiple files. Create a list if so
input_files = glob.glob(input_item)
return SqlTable(table_name, columns, num_rows, input_files)