pypaimon/py4j/java_implementation.py (278 lines of code) (raw):
################################################################################
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
################################################################################
# pypaimon.api implementation based on Java code & py4j lib
import pandas as pd
import pyarrow as pa
from pypaimon.py4j.java_gateway import get_gateway
from pypaimon.py4j.util import java_utils, constants
from pypaimon.py4j.util.java_utils import serialize_java_object, deserialize_java_object
from pypaimon.api import \
(catalog, table, read_builder, table_scan, split, row_type,
table_read, write_builder, table_write, commit_message,
table_commit, Schema, predicate)
from typing import List, Iterator, Optional, Any, TYPE_CHECKING
if TYPE_CHECKING:
import ray
from duckdb.duckdb import DuckDBPyConnection
class Catalog(catalog.Catalog):
def __init__(self, j_catalog, catalog_options: dict):
self._j_catalog = j_catalog
self._catalog_options = catalog_options
@staticmethod
def create(catalog_options: dict) -> 'Catalog':
j_catalog_context = java_utils.to_j_catalog_context(catalog_options)
gateway = get_gateway()
j_catalog = gateway.jvm.CatalogFactory.createCatalog(j_catalog_context)
return Catalog(j_catalog, catalog_options)
def get_table(self, identifier: str) -> 'Table':
j_identifier = java_utils.to_j_identifier(identifier)
j_table = self._j_catalog.getTable(j_identifier)
return Table(j_table, self._catalog_options)
def create_database(self, name: str, ignore_if_exists: bool, properties: Optional[dict] = None):
if properties is None:
properties = {}
self._j_catalog.createDatabase(name, ignore_if_exists, properties)
def create_table(self, identifier: str, schema: Schema, ignore_if_exists: bool):
j_identifier = java_utils.to_j_identifier(identifier)
j_schema = java_utils.to_paimon_schema(schema)
self._j_catalog.createTable(j_identifier, j_schema, ignore_if_exists)
class Table(table.Table):
def __init__(self, j_table, catalog_options: dict):
self._j_table = j_table
self._catalog_options = catalog_options
def new_read_builder(self) -> 'ReadBuilder':
j_read_builder = get_gateway().jvm.InvocationUtil.getReadBuilder(self._j_table)
return ReadBuilder(j_read_builder, self._j_table.rowType(), self._catalog_options)
def new_batch_write_builder(self) -> 'BatchWriteBuilder':
java_utils.check_batch_write(self._j_table)
j_batch_write_builder = get_gateway().jvm.InvocationUtil.getBatchWriteBuilder(self._j_table)
return BatchWriteBuilder(j_batch_write_builder)
class ReadBuilder(read_builder.ReadBuilder):
def __init__(self, j_read_builder, j_row_type, catalog_options: dict):
self._j_read_builder = j_read_builder
self._j_row_type = j_row_type
self._catalog_options = catalog_options
def with_filter(self, predicate: 'Predicate'):
self._j_read_builder.withFilter(predicate.to_j_predicate())
return self
def with_projection(self, projection: List[str]) -> 'ReadBuilder':
field_names = list(map(lambda field: field.name(), self._j_row_type.getFields()))
int_projection = list(map(lambda p: field_names.index(p), projection))
gateway = get_gateway()
int_projection_arr = gateway.new_array(gateway.jvm.int, len(projection))
for i in range(len(projection)):
int_projection_arr[i] = int_projection[i]
self._j_read_builder.withProjection(int_projection_arr)
return self
def with_limit(self, limit: int) -> 'ReadBuilder':
self._j_read_builder.withLimit(limit)
return self
def new_scan(self) -> 'TableScan':
j_table_scan = self._j_read_builder.newScan()
return TableScan(j_table_scan)
def new_read(self) -> 'TableRead':
j_table_read = self._j_read_builder.newRead().executeFilter()
return TableRead(j_table_read, self._j_read_builder.readType(), self._catalog_options)
def new_predicate_builder(self) -> 'PredicateBuilder':
return PredicateBuilder(self._j_row_type)
def read_type(self) -> 'RowType':
return RowType(self._j_read_builder.readType())
class RowType(row_type.RowType):
def __init__(self, j_row_type):
self._j_row_type = j_row_type
def as_arrow(self) -> "pa.Schema":
return java_utils.to_arrow_schema(self._j_row_type)
class TableScan(table_scan.TableScan):
def __init__(self, j_table_scan):
self._j_table_scan = j_table_scan
def plan(self) -> 'Plan':
j_plan = self._j_table_scan.plan()
j_splits = j_plan.splits()
return Plan(j_splits)
class Plan(table_scan.Plan):
def __init__(self, j_splits):
self._j_splits = j_splits
def splits(self) -> List['Split']:
return list(map(lambda s: self._build_single_split(s), self._j_splits))
def _build_single_split(self, j_split) -> 'Split':
j_split_bytes = serialize_java_object(j_split)
row_count = j_split.rowCount()
files_optional = j_split.convertToRawFiles()
if not files_optional.isPresent():
file_size = 0
file_paths = []
else:
files = files_optional.get()
file_size = sum(file.length() for file in files)
file_paths = [file.path() for file in files]
return Split(j_split_bytes, row_count, file_size, file_paths)
class Split(split.Split):
def __init__(self, j_split_bytes, row_count: int, file_size: int, file_paths: List[str]):
self._j_split_bytes = j_split_bytes
self._row_count = row_count
self._file_size = file_size
self._file_paths = file_paths
def to_j_split(self):
return deserialize_java_object(self._j_split_bytes)
def row_count(self) -> int:
return self._row_count
def file_size(self) -> int:
return self._file_size
def file_paths(self) -> List[str]:
return self._file_paths
class TableRead(table_read.TableRead):
def __init__(self, j_table_read, j_read_type, catalog_options):
self._arrow_schema = java_utils.to_arrow_schema(j_read_type)
self._j_bytes_reader = get_gateway().jvm.InvocationUtil.createParallelBytesReader(
j_table_read, j_read_type, TableRead._get_max_workers(catalog_options))
def to_arrow(self, splits):
record_batch_reader = self.to_arrow_batch_reader(splits)
return pa.Table.from_batches(record_batch_reader, schema=self._arrow_schema)
def to_arrow_batch_reader(self, splits):
j_splits = list(map(lambda s: s.to_j_split(), splits))
self._j_bytes_reader.setSplits(j_splits)
batch_iterator = self._batch_generator()
return pa.RecordBatchReader.from_batches(self._arrow_schema, batch_iterator)
def to_pandas(self, splits: List[Split]) -> pd.DataFrame:
return self.to_arrow(splits).to_pandas()
def to_duckdb(
self,
splits: List[Split],
table_name: str,
connection: Optional["DuckDBPyConnection"] = None) -> "DuckDBPyConnection":
import duckdb
con = connection or duckdb.connect(database=":memory:")
con.register(table_name, self.to_arrow(splits))
return con
def to_ray(self, splits: List[Split]) -> "ray.data.dataset.Dataset":
import ray
return ray.data.from_arrow(self.to_arrow(splits))
@staticmethod
def _get_max_workers(catalog_options):
# default is sequential
max_workers = int(catalog_options.get(constants.MAX_WORKERS, 1))
if max_workers <= 0:
raise ValueError("max_workers must be greater than 0")
return max_workers
def _batch_generator(self) -> Iterator[pa.RecordBatch]:
while True:
next_bytes = self._j_bytes_reader.next()
if next_bytes is None:
break
else:
stream_reader = pa.RecordBatchStreamReader(pa.BufferReader(next_bytes))
yield from stream_reader
class BatchWriteBuilder(write_builder.BatchWriteBuilder):
def __init__(self, j_batch_write_builder):
self._j_batch_write_builder = j_batch_write_builder
def overwrite(self, static_partition: Optional[dict] = None) -> 'BatchWriteBuilder':
if static_partition is None:
static_partition = {}
self._j_batch_write_builder.withOverwrite(static_partition)
return self
def new_write(self) -> 'BatchTableWrite':
j_batch_table_write = self._j_batch_write_builder.newWrite()
return BatchTableWrite(j_batch_table_write, self._j_batch_write_builder.rowType())
def new_commit(self) -> 'BatchTableCommit':
j_batch_table_commit = self._j_batch_write_builder.newCommit()
return BatchTableCommit(j_batch_table_commit)
class BatchTableWrite(table_write.BatchTableWrite):
def __init__(self, j_batch_table_write, j_row_type):
self._j_batch_table_write = j_batch_table_write
self._j_bytes_writer = get_gateway().jvm.InvocationUtil.createBytesWriter(
j_batch_table_write, j_row_type)
self._arrow_schema = java_utils.to_arrow_schema(j_row_type)
def write_arrow(self, table):
for record_batch in table.to_reader():
# TODO: can we use a reusable stream in #_write_arrow_batch ?
self._write_arrow_batch(record_batch)
def write_arrow_batch(self, record_batch):
self._write_arrow_batch(record_batch)
def write_pandas(self, dataframe: pd.DataFrame):
record_batch = pa.RecordBatch.from_pandas(dataframe, schema=self._arrow_schema)
self._write_arrow_batch(record_batch)
def _write_arrow_batch(self, record_batch):
stream = pa.BufferOutputStream()
with pa.RecordBatchStreamWriter(stream, record_batch.schema) as writer:
writer.write(record_batch)
arrow_bytes = stream.getvalue().to_pybytes()
self._j_bytes_writer.write(arrow_bytes)
def prepare_commit(self) -> List['CommitMessage']:
j_commit_messages = self._j_batch_table_write.prepareCommit()
return list(map(lambda cm: CommitMessage(cm), j_commit_messages))
def close(self):
self._j_batch_table_write.close()
self._j_bytes_writer.close()
class CommitMessage(commit_message.CommitMessage):
def __init__(self, j_commit_message):
self._j_commit_message = j_commit_message
def to_j_commit_message(self):
return self._j_commit_message
class BatchTableCommit(table_commit.BatchTableCommit):
def __init__(self, j_batch_table_commit):
self._j_batch_table_commit = j_batch_table_commit
def commit(self, commit_messages: List[CommitMessage]):
j_commit_messages = list(map(lambda cm: cm.to_j_commit_message(), commit_messages))
self._j_batch_table_commit.commit(j_commit_messages)
def close(self):
self._j_batch_table_commit.close()
class Predicate(predicate.Predicate):
def __init__(self, j_predicate_bytes):
self._j_predicate_bytes = j_predicate_bytes
def to_j_predicate(self):
return deserialize_java_object(self._j_predicate_bytes)
class PredicateBuilder(predicate.PredicateBuilder):
def __init__(self, j_row_type):
self._field_names = j_row_type.getFieldNames()
self._j_row_type = j_row_type
self._j_predicate_builder = get_gateway().jvm.PredicateBuilder(j_row_type)
def _build(self, method: str, field: str, literals: Optional[List[Any]] = None):
error = ValueError(f'The field {field} is not in field list {self._field_names}.')
try:
index = self._field_names.index(field)
if index == -1:
raise error
except ValueError:
raise error
if literals is None:
literals = []
j_predicate = get_gateway().jvm.PredicationUtil.build(
self._j_row_type,
self._j_predicate_builder,
method,
index,
literals
)
return Predicate(serialize_java_object(j_predicate))
def equal(self, field: str, literal: Any) -> Predicate:
return self._build('equal', field, [literal])
def not_equal(self, field: str, literal: Any) -> Predicate:
return self._build('notEqual', field, [literal])
def less_than(self, field: str, literal: Any) -> Predicate:
return self._build('lessThan', field, [literal])
def less_or_equal(self, field: str, literal: Any) -> Predicate:
return self._build('lessOrEqual', field, [literal])
def greater_than(self, field: str, literal: Any) -> Predicate:
return self._build('greaterThan', field, [literal])
def greater_or_equal(self, field: str, literal: Any) -> Predicate:
return self._build('greaterOrEqual', field, [literal])
def is_null(self, field: str) -> Predicate:
return self._build('isNull', field)
def is_not_null(self, field: str) -> Predicate:
return self._build('isNotNull', field)
def startswith(self, field: str, pattern_literal: Any) -> Predicate:
return self._build('startsWith', field, [pattern_literal])
def endswith(self, field: str, pattern_literal: Any) -> Predicate:
return self._build('endsWith', field, [pattern_literal])
def contains(self, field: str, pattern_literal: Any) -> Predicate:
return self._build('contains', field, [pattern_literal])
def is_in(self, field: str, literals: List[Any]) -> Predicate:
return self._build('in', field, literals)
def is_not_in(self, field: str, literals: List[Any]) -> Predicate:
return self._build('notIn', field, literals)
def between(self, field: str, included_lower_bound: Any, included_upper_bound: Any) \
-> Predicate:
return self._build('between', field, [included_lower_bound, included_upper_bound])
def and_predicates(self, predicates: List[Predicate]) -> Predicate:
predicates = list(map(lambda p: p.to_j_predicate(), predicates))
j_predicate = get_gateway().jvm.PredicationUtil.buildAnd(predicates)
return Predicate(serialize_java_object(j_predicate))
def or_predicates(self, predicates: List[Predicate]) -> Predicate:
predicates = list(map(lambda p: p.to_j_predicate(), predicates))
j_predicate = get_gateway().jvm.PredicationUtil.buildOr(predicates)
return Predicate(serialize_java_object(j_predicate))