python/pyspark/sql/connect/proto/catalog_pb2.pyi (913 lines of code) (raw):

# # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """ @generated by mypy-protobuf. Do not edit manually! isort:skip_file Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import builtins import collections.abc import google.protobuf.descriptor import google.protobuf.internal.containers import google.protobuf.message import pyspark.sql.connect.proto.common_pb2 import pyspark.sql.connect.proto.types_pb2 import sys import typing if sys.version_info >= (3, 8): import typing as typing_extensions else: import typing_extensions DESCRIPTOR: google.protobuf.descriptor.FileDescriptor class Catalog(google.protobuf.message.Message): """Catalog messages are marked as unstable.""" DESCRIPTOR: google.protobuf.descriptor.Descriptor CURRENT_DATABASE_FIELD_NUMBER: builtins.int SET_CURRENT_DATABASE_FIELD_NUMBER: builtins.int LIST_DATABASES_FIELD_NUMBER: builtins.int LIST_TABLES_FIELD_NUMBER: builtins.int LIST_FUNCTIONS_FIELD_NUMBER: builtins.int LIST_COLUMNS_FIELD_NUMBER: builtins.int GET_DATABASE_FIELD_NUMBER: builtins.int GET_TABLE_FIELD_NUMBER: builtins.int GET_FUNCTION_FIELD_NUMBER: builtins.int DATABASE_EXISTS_FIELD_NUMBER: builtins.int TABLE_EXISTS_FIELD_NUMBER: builtins.int FUNCTION_EXISTS_FIELD_NUMBER: builtins.int CREATE_EXTERNAL_TABLE_FIELD_NUMBER: builtins.int CREATE_TABLE_FIELD_NUMBER: builtins.int DROP_TEMP_VIEW_FIELD_NUMBER: builtins.int DROP_GLOBAL_TEMP_VIEW_FIELD_NUMBER: builtins.int RECOVER_PARTITIONS_FIELD_NUMBER: builtins.int IS_CACHED_FIELD_NUMBER: builtins.int CACHE_TABLE_FIELD_NUMBER: builtins.int UNCACHE_TABLE_FIELD_NUMBER: builtins.int CLEAR_CACHE_FIELD_NUMBER: builtins.int REFRESH_TABLE_FIELD_NUMBER: builtins.int REFRESH_BY_PATH_FIELD_NUMBER: builtins.int CURRENT_CATALOG_FIELD_NUMBER: builtins.int SET_CURRENT_CATALOG_FIELD_NUMBER: builtins.int LIST_CATALOGS_FIELD_NUMBER: builtins.int @property def current_database(self) -> global___CurrentDatabase: ... @property def set_current_database(self) -> global___SetCurrentDatabase: ... @property def list_databases(self) -> global___ListDatabases: ... @property def list_tables(self) -> global___ListTables: ... @property def list_functions(self) -> global___ListFunctions: ... @property def list_columns(self) -> global___ListColumns: ... @property def get_database(self) -> global___GetDatabase: ... @property def get_table(self) -> global___GetTable: ... @property def get_function(self) -> global___GetFunction: ... @property def database_exists(self) -> global___DatabaseExists: ... @property def table_exists(self) -> global___TableExists: ... @property def function_exists(self) -> global___FunctionExists: ... @property def create_external_table(self) -> global___CreateExternalTable: ... @property def create_table(self) -> global___CreateTable: ... @property def drop_temp_view(self) -> global___DropTempView: ... @property def drop_global_temp_view(self) -> global___DropGlobalTempView: ... @property def recover_partitions(self) -> global___RecoverPartitions: ... @property def is_cached(self) -> global___IsCached: ... @property def cache_table(self) -> global___CacheTable: ... @property def uncache_table(self) -> global___UncacheTable: ... @property def clear_cache(self) -> global___ClearCache: ... @property def refresh_table(self) -> global___RefreshTable: ... @property def refresh_by_path(self) -> global___RefreshByPath: ... @property def current_catalog(self) -> global___CurrentCatalog: ... @property def set_current_catalog(self) -> global___SetCurrentCatalog: ... @property def list_catalogs(self) -> global___ListCatalogs: ... def __init__( self, *, current_database: global___CurrentDatabase | None = ..., set_current_database: global___SetCurrentDatabase | None = ..., list_databases: global___ListDatabases | None = ..., list_tables: global___ListTables | None = ..., list_functions: global___ListFunctions | None = ..., list_columns: global___ListColumns | None = ..., get_database: global___GetDatabase | None = ..., get_table: global___GetTable | None = ..., get_function: global___GetFunction | None = ..., database_exists: global___DatabaseExists | None = ..., table_exists: global___TableExists | None = ..., function_exists: global___FunctionExists | None = ..., create_external_table: global___CreateExternalTable | None = ..., create_table: global___CreateTable | None = ..., drop_temp_view: global___DropTempView | None = ..., drop_global_temp_view: global___DropGlobalTempView | None = ..., recover_partitions: global___RecoverPartitions | None = ..., is_cached: global___IsCached | None = ..., cache_table: global___CacheTable | None = ..., uncache_table: global___UncacheTable | None = ..., clear_cache: global___ClearCache | None = ..., refresh_table: global___RefreshTable | None = ..., refresh_by_path: global___RefreshByPath | None = ..., current_catalog: global___CurrentCatalog | None = ..., set_current_catalog: global___SetCurrentCatalog | None = ..., list_catalogs: global___ListCatalogs | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal[ "cache_table", b"cache_table", "cat_type", b"cat_type", "clear_cache", b"clear_cache", "create_external_table", b"create_external_table", "create_table", b"create_table", "current_catalog", b"current_catalog", "current_database", b"current_database", "database_exists", b"database_exists", "drop_global_temp_view", b"drop_global_temp_view", "drop_temp_view", b"drop_temp_view", "function_exists", b"function_exists", "get_database", b"get_database", "get_function", b"get_function", "get_table", b"get_table", "is_cached", b"is_cached", "list_catalogs", b"list_catalogs", "list_columns", b"list_columns", "list_databases", b"list_databases", "list_functions", b"list_functions", "list_tables", b"list_tables", "recover_partitions", b"recover_partitions", "refresh_by_path", b"refresh_by_path", "refresh_table", b"refresh_table", "set_current_catalog", b"set_current_catalog", "set_current_database", b"set_current_database", "table_exists", b"table_exists", "uncache_table", b"uncache_table", ], ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "cache_table", b"cache_table", "cat_type", b"cat_type", "clear_cache", b"clear_cache", "create_external_table", b"create_external_table", "create_table", b"create_table", "current_catalog", b"current_catalog", "current_database", b"current_database", "database_exists", b"database_exists", "drop_global_temp_view", b"drop_global_temp_view", "drop_temp_view", b"drop_temp_view", "function_exists", b"function_exists", "get_database", b"get_database", "get_function", b"get_function", "get_table", b"get_table", "is_cached", b"is_cached", "list_catalogs", b"list_catalogs", "list_columns", b"list_columns", "list_databases", b"list_databases", "list_functions", b"list_functions", "list_tables", b"list_tables", "recover_partitions", b"recover_partitions", "refresh_by_path", b"refresh_by_path", "refresh_table", b"refresh_table", "set_current_catalog", b"set_current_catalog", "set_current_database", b"set_current_database", "table_exists", b"table_exists", "uncache_table", b"uncache_table", ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["cat_type", b"cat_type"] ) -> ( typing_extensions.Literal[ "current_database", "set_current_database", "list_databases", "list_tables", "list_functions", "list_columns", "get_database", "get_table", "get_function", "database_exists", "table_exists", "function_exists", "create_external_table", "create_table", "drop_temp_view", "drop_global_temp_view", "recover_partitions", "is_cached", "cache_table", "uncache_table", "clear_cache", "refresh_table", "refresh_by_path", "current_catalog", "set_current_catalog", "list_catalogs", ] | None ): ... global___Catalog = Catalog class CurrentDatabase(google.protobuf.message.Message): """See `spark.catalog.currentDatabase`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor def __init__( self, ) -> None: ... global___CurrentDatabase = CurrentDatabase class SetCurrentDatabase(google.protobuf.message.Message): """See `spark.catalog.setCurrentDatabase`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor DB_NAME_FIELD_NUMBER: builtins.int db_name: builtins.str """(Required)""" def __init__( self, *, db_name: builtins.str = ..., ) -> None: ... def ClearField(self, field_name: typing_extensions.Literal["db_name", b"db_name"]) -> None: ... global___SetCurrentDatabase = SetCurrentDatabase class ListDatabases(google.protobuf.message.Message): """See `spark.catalog.listDatabases`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor PATTERN_FIELD_NUMBER: builtins.int pattern: builtins.str """(Optional) The pattern that the database name needs to match""" def __init__( self, *, pattern: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_pattern", b"_pattern", "pattern", b"pattern"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal["_pattern", b"_pattern", "pattern", b"pattern"] ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_pattern", b"_pattern"] ) -> typing_extensions.Literal["pattern"] | None: ... global___ListDatabases = ListDatabases class ListTables(google.protobuf.message.Message): """See `spark.catalog.listTables`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor DB_NAME_FIELD_NUMBER: builtins.int PATTERN_FIELD_NUMBER: builtins.int db_name: builtins.str """(Optional)""" pattern: builtins.str """(Optional) The pattern that the table name needs to match""" def __init__( self, *, db_name: builtins.str | None = ..., pattern: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "_pattern", b"_pattern", "db_name", b"db_name", "pattern", b"pattern", ], ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "_pattern", b"_pattern", "db_name", b"db_name", "pattern", b"pattern", ], ) -> None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_pattern", b"_pattern"] ) -> typing_extensions.Literal["pattern"] | None: ... global___ListTables = ListTables class ListFunctions(google.protobuf.message.Message): """See `spark.catalog.listFunctions`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor DB_NAME_FIELD_NUMBER: builtins.int PATTERN_FIELD_NUMBER: builtins.int db_name: builtins.str """(Optional)""" pattern: builtins.str """(Optional) The pattern that the function name needs to match""" def __init__( self, *, db_name: builtins.str | None = ..., pattern: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "_pattern", b"_pattern", "db_name", b"db_name", "pattern", b"pattern", ], ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "_pattern", b"_pattern", "db_name", b"db_name", "pattern", b"pattern", ], ) -> None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_pattern", b"_pattern"] ) -> typing_extensions.Literal["pattern"] | None: ... global___ListFunctions = ListFunctions class ListColumns(google.protobuf.message.Message): """See `spark.catalog.listColumns`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int DB_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" db_name: builtins.str """(Optional)""" def __init__( self, *, table_name: builtins.str = ..., db_name: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_db_name", b"_db_name", "db_name", b"db_name"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "db_name", b"db_name", "table_name", b"table_name" ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... global___ListColumns = ListColumns class GetDatabase(google.protobuf.message.Message): """See `spark.catalog.getDatabase`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor DB_NAME_FIELD_NUMBER: builtins.int db_name: builtins.str """(Required)""" def __init__( self, *, db_name: builtins.str = ..., ) -> None: ... def ClearField(self, field_name: typing_extensions.Literal["db_name", b"db_name"]) -> None: ... global___GetDatabase = GetDatabase class GetTable(google.protobuf.message.Message): """See `spark.catalog.getTable`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int DB_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" db_name: builtins.str """(Optional)""" def __init__( self, *, table_name: builtins.str = ..., db_name: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_db_name", b"_db_name", "db_name", b"db_name"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "db_name", b"db_name", "table_name", b"table_name" ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... global___GetTable = GetTable class GetFunction(google.protobuf.message.Message): """See `spark.catalog.getFunction`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor FUNCTION_NAME_FIELD_NUMBER: builtins.int DB_NAME_FIELD_NUMBER: builtins.int function_name: builtins.str """(Required)""" db_name: builtins.str """(Optional)""" def __init__( self, *, function_name: builtins.str = ..., db_name: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_db_name", b"_db_name", "db_name", b"db_name"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "db_name", b"db_name", "function_name", b"function_name" ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... global___GetFunction = GetFunction class DatabaseExists(google.protobuf.message.Message): """See `spark.catalog.databaseExists`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor DB_NAME_FIELD_NUMBER: builtins.int db_name: builtins.str """(Required)""" def __init__( self, *, db_name: builtins.str = ..., ) -> None: ... def ClearField(self, field_name: typing_extensions.Literal["db_name", b"db_name"]) -> None: ... global___DatabaseExists = DatabaseExists class TableExists(google.protobuf.message.Message): """See `spark.catalog.tableExists`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int DB_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" db_name: builtins.str """(Optional)""" def __init__( self, *, table_name: builtins.str = ..., db_name: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_db_name", b"_db_name", "db_name", b"db_name"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "db_name", b"db_name", "table_name", b"table_name" ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... global___TableExists = TableExists class FunctionExists(google.protobuf.message.Message): """See `spark.catalog.functionExists`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor FUNCTION_NAME_FIELD_NUMBER: builtins.int DB_NAME_FIELD_NUMBER: builtins.int function_name: builtins.str """(Required)""" db_name: builtins.str """(Optional)""" def __init__( self, *, function_name: builtins.str = ..., db_name: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_db_name", b"_db_name", "db_name", b"db_name"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_db_name", b"_db_name", "db_name", b"db_name", "function_name", b"function_name" ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_db_name", b"_db_name"] ) -> typing_extensions.Literal["db_name"] | None: ... global___FunctionExists = FunctionExists class CreateExternalTable(google.protobuf.message.Message): """See `spark.catalog.createExternalTable`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor class OptionsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int key: builtins.str value: builtins.str def __init__( self, *, key: builtins.str = ..., value: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["key", b"key", "value", b"value"] ) -> None: ... TABLE_NAME_FIELD_NUMBER: builtins.int PATH_FIELD_NUMBER: builtins.int SOURCE_FIELD_NUMBER: builtins.int SCHEMA_FIELD_NUMBER: builtins.int OPTIONS_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" path: builtins.str """(Optional)""" source: builtins.str """(Optional)""" @property def schema(self) -> pyspark.sql.connect.proto.types_pb2.DataType: """(Optional)""" @property def options(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.str]: """Options could be empty for valid data source format. The map key is case insensitive. """ def __init__( self, *, table_name: builtins.str = ..., path: builtins.str | None = ..., source: builtins.str | None = ..., schema: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., options: collections.abc.Mapping[builtins.str, builtins.str] | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal[ "_path", b"_path", "_schema", b"_schema", "_source", b"_source", "path", b"path", "schema", b"schema", "source", b"source", ], ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_path", b"_path", "_schema", b"_schema", "_source", b"_source", "options", b"options", "path", b"path", "schema", b"schema", "source", b"source", "table_name", b"table_name", ], ) -> None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_path", b"_path"] ) -> typing_extensions.Literal["path"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_schema", b"_schema"] ) -> typing_extensions.Literal["schema"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_source", b"_source"] ) -> typing_extensions.Literal["source"] | None: ... global___CreateExternalTable = CreateExternalTable class CreateTable(google.protobuf.message.Message): """See `spark.catalog.createTable`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor class OptionsEntry(google.protobuf.message.Message): DESCRIPTOR: google.protobuf.descriptor.Descriptor KEY_FIELD_NUMBER: builtins.int VALUE_FIELD_NUMBER: builtins.int key: builtins.str value: builtins.str def __init__( self, *, key: builtins.str = ..., value: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["key", b"key", "value", b"value"] ) -> None: ... TABLE_NAME_FIELD_NUMBER: builtins.int PATH_FIELD_NUMBER: builtins.int SOURCE_FIELD_NUMBER: builtins.int DESCRIPTION_FIELD_NUMBER: builtins.int SCHEMA_FIELD_NUMBER: builtins.int OPTIONS_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" path: builtins.str """(Optional)""" source: builtins.str """(Optional)""" description: builtins.str """(Optional)""" @property def schema(self) -> pyspark.sql.connect.proto.types_pb2.DataType: """(Optional)""" @property def options(self) -> google.protobuf.internal.containers.ScalarMap[builtins.str, builtins.str]: """Options could be empty for valid data source format. The map key is case insensitive. """ def __init__( self, *, table_name: builtins.str = ..., path: builtins.str | None = ..., source: builtins.str | None = ..., description: builtins.str | None = ..., schema: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., options: collections.abc.Mapping[builtins.str, builtins.str] | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal[ "_description", b"_description", "_path", b"_path", "_schema", b"_schema", "_source", b"_source", "description", b"description", "path", b"path", "schema", b"schema", "source", b"source", ], ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_description", b"_description", "_path", b"_path", "_schema", b"_schema", "_source", b"_source", "description", b"description", "options", b"options", "path", b"path", "schema", b"schema", "source", b"source", "table_name", b"table_name", ], ) -> None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_description", b"_description"] ) -> typing_extensions.Literal["description"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_path", b"_path"] ) -> typing_extensions.Literal["path"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_schema", b"_schema"] ) -> typing_extensions.Literal["schema"] | None: ... @typing.overload def WhichOneof( self, oneof_group: typing_extensions.Literal["_source", b"_source"] ) -> typing_extensions.Literal["source"] | None: ... global___CreateTable = CreateTable class DropTempView(google.protobuf.message.Message): """See `spark.catalog.dropTempView`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor VIEW_NAME_FIELD_NUMBER: builtins.int view_name: builtins.str """(Required)""" def __init__( self, *, view_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["view_name", b"view_name"] ) -> None: ... global___DropTempView = DropTempView class DropGlobalTempView(google.protobuf.message.Message): """See `spark.catalog.dropGlobalTempView`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor VIEW_NAME_FIELD_NUMBER: builtins.int view_name: builtins.str """(Required)""" def __init__( self, *, view_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["view_name", b"view_name"] ) -> None: ... global___DropGlobalTempView = DropGlobalTempView class RecoverPartitions(google.protobuf.message.Message): """See `spark.catalog.recoverPartitions`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" def __init__( self, *, table_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["table_name", b"table_name"] ) -> None: ... global___RecoverPartitions = RecoverPartitions class IsCached(google.protobuf.message.Message): """See `spark.catalog.isCached`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" def __init__( self, *, table_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["table_name", b"table_name"] ) -> None: ... global___IsCached = IsCached class CacheTable(google.protobuf.message.Message): """See `spark.catalog.cacheTable`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int STORAGE_LEVEL_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" @property def storage_level(self) -> pyspark.sql.connect.proto.common_pb2.StorageLevel: """(Optional)""" def __init__( self, *, table_name: builtins.str = ..., storage_level: pyspark.sql.connect.proto.common_pb2.StorageLevel | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal[ "_storage_level", b"_storage_level", "storage_level", b"storage_level" ], ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal[ "_storage_level", b"_storage_level", "storage_level", b"storage_level", "table_name", b"table_name", ], ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_storage_level", b"_storage_level"] ) -> typing_extensions.Literal["storage_level"] | None: ... global___CacheTable = CacheTable class UncacheTable(google.protobuf.message.Message): """See `spark.catalog.uncacheTable`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" def __init__( self, *, table_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["table_name", b"table_name"] ) -> None: ... global___UncacheTable = UncacheTable class ClearCache(google.protobuf.message.Message): """See `spark.catalog.clearCache`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor def __init__( self, ) -> None: ... global___ClearCache = ClearCache class RefreshTable(google.protobuf.message.Message): """See `spark.catalog.refreshTable`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor TABLE_NAME_FIELD_NUMBER: builtins.int table_name: builtins.str """(Required)""" def __init__( self, *, table_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["table_name", b"table_name"] ) -> None: ... global___RefreshTable = RefreshTable class RefreshByPath(google.protobuf.message.Message): """See `spark.catalog.refreshByPath`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor PATH_FIELD_NUMBER: builtins.int path: builtins.str """(Required)""" def __init__( self, *, path: builtins.str = ..., ) -> None: ... def ClearField(self, field_name: typing_extensions.Literal["path", b"path"]) -> None: ... global___RefreshByPath = RefreshByPath class CurrentCatalog(google.protobuf.message.Message): """See `spark.catalog.currentCatalog`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor def __init__( self, ) -> None: ... global___CurrentCatalog = CurrentCatalog class SetCurrentCatalog(google.protobuf.message.Message): """See `spark.catalog.setCurrentCatalog`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor CATALOG_NAME_FIELD_NUMBER: builtins.int catalog_name: builtins.str """(Required)""" def __init__( self, *, catalog_name: builtins.str = ..., ) -> None: ... def ClearField( self, field_name: typing_extensions.Literal["catalog_name", b"catalog_name"] ) -> None: ... global___SetCurrentCatalog = SetCurrentCatalog class ListCatalogs(google.protobuf.message.Message): """See `spark.catalog.listCatalogs`""" DESCRIPTOR: google.protobuf.descriptor.Descriptor PATTERN_FIELD_NUMBER: builtins.int pattern: builtins.str """(Optional) The pattern that the catalog name needs to match""" def __init__( self, *, pattern: builtins.str | None = ..., ) -> None: ... def HasField( self, field_name: typing_extensions.Literal["_pattern", b"_pattern", "pattern", b"pattern"] ) -> builtins.bool: ... def ClearField( self, field_name: typing_extensions.Literal["_pattern", b"_pattern", "pattern", b"pattern"] ) -> None: ... def WhichOneof( self, oneof_group: typing_extensions.Literal["_pattern", b"_pattern"] ) -> typing_extensions.Literal["pattern"] | None: ... global___ListCatalogs = ListCatalogs