common/thrift/ImpalaService.thrift (306 lines of code) (raw):

// Licensed to the Apache Software Foundation (ASF) under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, // software distributed under the License is distributed on an // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. See the License for the // specific language governing permissions and limitations // under the License. namespace py impala_thrift_gen.ImpalaService namespace cpp impala namespace java org.apache.impala.thrift include "ExecStats.thrift" include "Status.thrift" include "Types.thrift" include "beeswax.thrift" include "TCLIService.thrift" include "RuntimeProfile.thrift" include "Frontend.thrift" include "BackendGflags.thrift" include "Query.thrift" // ImpalaService accepts query execution options through beeswax.Query.configuration in // key:value form. For example, the list of strings could be: // "num_nodes:1", "abort_on_error:false" // The valid keys are listed in this enum. They map to TQueryOptions. // Note: If you add an option or change the default, you also need to update: // - ImpalaInternalService.thrift: TQueryOptions // - SetQueryOption(), SetQueryOptions() // - TQueryOptionsToMap() enum TImpalaQueryOptions { // if true, abort execution on the first error ABORT_ON_ERROR = 0 // maximum # of errors to be reported; Unspecified or 0 indicates backend default MAX_ERRORS = 1 // if true disable llvm codegen DISABLE_CODEGEN = 2 // batch size to be used by backend; Unspecified or a size of 0 indicates backend // default BATCH_SIZE = 3 // a per-machine approximate limit on the memory consumption of this query; // unspecified or a limit of 0 means no limit; // otherwise specified either as: // a) an int (= number of bytes); // b) a float followed by "M" (MB) or "G" (GB) MEM_LIMIT = 4 // specifies the degree of parallelism with which to execute the query; // 1: single-node execution // NUM_NODES_ALL: executes on all nodes that contain relevant data // NUM_NODES_ALL_RACKS: executes on one node per rack that holds relevant data // > 1: executes on at most that many nodes at any point in time (ie, there can be // more nodes than numNodes with plan fragments for this query, but at most // numNodes would be active at any point in time) // Constants (NUM_NODES_ALL, NUM_NODES_ALL_RACKS) are defined in JavaConstants.thrift. NUM_NODES = 5 // maximum length of the scan range; only applicable to HDFS scan range; Unspecified or // a length of 0 indicates backend default; MAX_SCAN_RANGE_LENGTH = 6 MAX_IO_BUFFERS = 7 // Removed // Number of scanner threads. NUM_SCANNER_THREADS = 8 ALLOW_UNSUPPORTED_FORMATS = 9 // Removed DEFAULT_ORDER_BY_LIMIT = 10 // Removed // DEBUG ONLY: // Accepted formats: // 1. ExecNode actions // "[<instance idx>:]<node id>:<TExecNodePhase>:<TDebugAction>", // the exec node with the given id will perform the specified action in the given // phase. If the optional backend number (starting from 0) is specified, only that // backend instance will perform the debug action, otherwise all backends will behave // in that way. // If the string doesn't have the required format or if any of its components is // invalid, the option is ignored. // // 2. Global actions // "<global label>:<arg0>:...:<argN>:<command>@<param0>@<param1>@...<paramN>", // Used with the DebugAction() call, the action will be performed if the label and // optional arguments all match. The arguments can be used to make the debug action // context dependent, for example to only fail rpcs when a particular hostname matches. // Note that some debug actions must be specified as a query option while others must // be passed in with the startup flag. // Available global actions: // - SLEEP@<ms> will sleep for the 'ms' milliseconds. // - JITTER@<ms>[@<probability>] will sleep for a random amount of time between 0 // and 'ms' milliseconds with the given probability. If <probability> is omitted, // it is 1.0. // - FAIL[@<probability>][@<error>] returns an INTERNAL_ERROR status with the given // probability and error. If <probability> is omitted, it is 1.0. If 'error' is // omitted, a generic error of the form: 'Debug Action: <label>:<action>' is used. // The code executing the debug action may respond to different error messages by // exercising different error paths. // Examples: // - AC_BEFORE_ADMISSION:SLEEP@1000 // Causes a 1 second sleep before queries are submitted to the admission controller. // - IMPALA_SERVICE_POOL:127.0.0.1:27002:TransmitData:FAIL@0.1@REJECT_TOO_BUSY // Causes TransmitData rpcs to the third minicluster impalad to fail 10% of the time // with a "service too busy" error. // // Only a single ExecNode action is allowed, but multiple global actions can be // specified. To specify multiple actions, separate them with "|". DEBUG_ACTION = 11 ABORT_ON_DEFAULT_LIMIT_EXCEEDED = 12 // Removed // Compression codec when inserting into tables. // Valid values are "snappy", "gzip", "bzip2" and "none" // Leave blank to use default. COMPRESSION_CODEC = 13 SEQ_COMPRESSION_MODE = 14 // Removed // HBase scan query option. If set and > 0, HBASE_CACHING is the value for // "hbase.client.Scan.setCaching()" when querying HBase table. Otherwise, use backend // default. // If the value is too high, then the hbase region server will have a hard time (GC // pressure and long response times). If the value is too small, then there will be // extra trips to the hbase region server. HBASE_CACHING = 15 // HBase scan query option. If set, HBase scan will always set // "hbase.client.setCacheBlocks" to CACHE_BLOCKS. Default is false. // If the table is large and the query is doing big scan, set it to false to // avoid polluting the cache in the hbase region server. // If the table is small and the table is used several time, set it to true to improve // performance. HBASE_CACHE_BLOCKS = 16 // Target file size for inserts into parquet tables. 0 uses the default. PARQUET_FILE_SIZE = 17 // Level of detail for explain output (NORMAL, VERBOSE). EXPLAIN_LEVEL = 18 // If true, waits for the result of all catalog operations to be processed by all // active impalad in the cluster before completing. SYNC_DDL = 19 // Request pool this request should be submitted to. If not set // the pool is determined based on the user. REQUEST_POOL = 20 V_CPU_CORES = 21 // Removed RESERVATION_REQUEST_TIMEOUT = 22 // Removed // if true, disables cached reads. This option has no effect if REPLICA_PREFERENCE is // configured. // TODO: IMPALA-4306: retire at compatibility-breaking version DISABLE_CACHED_READS = 23 // Temporary testing flag DISABLE_OUTERMOST_TOPN = 24 RM_INITIAL_MEM = 25 // Removed // Time, in s, before a query will be timed out if it is inactive. May not exceed // --idle_query_timeout if that flag > 0. If 0, falls back to --idle_query_timeout. QUERY_TIMEOUT_S = 26 // Test hook for spill to disk operators BUFFER_POOL_LIMIT = 27 // Transforms all count(distinct) aggregations into NDV() APPX_COUNT_DISTINCT = 28 // If true, allows Impala to internally disable spilling for potentially // disastrous query plans. Impala will excercise this option if a query // has no plan hints, and at least one table is missing relevant stats. DISABLE_UNSAFE_SPILLS = 29 // If the number of rows that are processed for a single query is below the // threshold, it will be executed on the coordinator only with codegen disabled EXEC_SINGLE_NODE_ROWS_THRESHOLD = 30 // If true, use the table's metadata to produce the partition columns instead of table // scans whenever possible. This option is opt-in by default as this optimization may // produce different results than the scan based approach in some edge cases. OPTIMIZE_PARTITION_KEY_SCANS = 31 // Prefered memory distance of replicas. This parameter determines the pool of replicas // among which scans will be scheduled in terms of the distance of the replica storage // from the impalad. REPLICA_PREFERENCE = 32 // Enables random backend selection during scheduling. SCHEDULE_RANDOM_REPLICA = 33 SCAN_NODE_CODEGEN_THRESHOLD = 34 // Removed // If true, the planner will not generate plans with streaming preaggregations. DISABLE_STREAMING_PREAGGREGATIONS = 35 // Select between off, local, or global runtime filters. RUNTIME_FILTER_MODE = 36 // Size (in bytes) of a runtime Bloom Filter. Will be rounded up to nearest power of // two. RUNTIME_BLOOM_FILTER_SIZE = 37 // Time (in ms) to wait in scans for runtime filters to arrive. RUNTIME_FILTER_WAIT_TIME_MS = 38 // If true, disable application of runtime filters to individual rows. DISABLE_ROW_RUNTIME_FILTERING = 39 // Maximum number of bloom runtime filters allowed per query. MAX_NUM_RUNTIME_FILTERS = 40 // If true, use UTF-8 annotation for string columns. Note that char and varchar columns // always use the annotation. PARQUET_ANNOTATE_STRINGS_UTF8 = 41 // Determines how to resolve Parquet files' schemas in the absence of field IDs (which // is always, since fields IDs are NYI). Valid values are "position" and "name". PARQUET_FALLBACK_SCHEMA_RESOLUTION = 42 // Multi-threaded execution: degree of parallelism = number of active threads per // backend MT_DOP = 43 // If true, INSERT writes to S3 go directly to their final location rather than being // copied there by the coordinator. We cannot do this for INSERT OVERWRITES because for // those queries, the coordinator deletes all files in the final location before copying // the files there. // TODO: Find a way to get this working for INSERT OVERWRITEs too. S3_SKIP_INSERT_STAGING = 44 // Maximum runtime bloom filter size, in bytes. RUNTIME_FILTER_MAX_SIZE = 45 // Minimum runtime bloom filter size, in bytes. RUNTIME_FILTER_MIN_SIZE = 46 // Prefetching behavior during hash tables' building and probing. PREFETCH_MODE = 47 // Additional strict handling of invalid data parsing and type conversions. STRICT_MODE = 48 // A limit on the amount of scratch directory space that can be used; // Unspecified or a limit of -1 means no limit; // Otherwise specified in the same way as MEM_LIMIT. SCRATCH_LIMIT = 49 // Indicates whether the FE should rewrite Exprs for optimization purposes. // It's sometimes useful to disable rewrites for testing, e.g., expr-test.cc. ENABLE_EXPR_REWRITES = 50 // Indicates whether to use the new decimal semantics, which includes better // rounding and output types for multiply / divide DECIMAL_V2 = 51 // Indicates whether to use dictionary filtering for Parquet files PARQUET_DICTIONARY_FILTERING = 52 // Policy for resolving nested array fields in Parquet files. // An Impala array type can have several different representations in // a Parquet schema (three, two, or one level). There is fundamental ambiguity // between the two and three level encodings with index-based field resolution. // The ambiguity can manually be resolved using this query option, or by using // PARQUET_FALLBACK_SCHEMA_RESOLUTION=name. PARQUET_ARRAY_RESOLUTION = 53 // Indicates whether to read statistics from Parquet files and use them during query // processing. This includes skipping data based on the statistics and computing query // results like "select min()". PARQUET_READ_STATISTICS = 54 // Join distribution mode that is used when the join inputs have an unknown // cardinality, e.g., because of missing table statistics. DEFAULT_JOIN_DISTRIBUTION_MODE = 55 // If the number of rows processed per node is below the threshold and disable_codegen // is unset, codegen will be automatically be disabled by the planner. DISABLE_CODEGEN_ROWS_THRESHOLD = 56 // The default spillable buffer size, in bytes. DEFAULT_SPILLABLE_BUFFER_SIZE = 57 // The minimum spillable buffer size, in bytes. MIN_SPILLABLE_BUFFER_SIZE = 58 // The maximum row size that memory is reserved for, in bytes. MAX_ROW_SIZE = 59 // The time, in seconds, that a session may be idle for before it is closed (and all // running queries cancelled) by Impala. If 0, idle sessions never expire. IDLE_SESSION_TIMEOUT = 60 // Minimum number of bytes that will be scanned in COMPUTE STATS TABLESAMPLE, // regardless of the user-supplied sampling percent. COMPUTE_STATS_MIN_SAMPLE_SIZE = 61 // Time limit, in s, before a query will be timed out after it starts executing. Does // not include time spent in planning, scheduling or admission control. A value of 0 // means no time limit. EXEC_TIME_LIMIT_S = 62 // When a query has both grouping and distinct exprs, impala can optionally include the // distinct exprs in the hash exchange of the first aggregation phase to spread the data // among more nodes. However, this plan requires another hash exchange on the grouping // exprs in the second phase which is not required when omitting the distinct exprs in // the first phase. Shuffling by both is better if the grouping exprs have low NDVs. SHUFFLE_DISTINCT_EXPRS = 63 // This only has an effect if memory-estimate-based admission control is enabled, i.e. // max_mem_resources is set for the pool and, *contrary to best practices*, MEM_LIMIT // is not set. In that case, then min(MAX_MEM_ESTIMATE_FOR_ADMISSION, // planner memory estimate) is used for admission control purposes. This provides a // workaround if the planner's memory estimate is too high and prevents a runnable // query from being admitted. 0 or -1 means this has no effect. Defaults to 0. MAX_MEM_ESTIMATE_FOR_ADMISSION = 64 // Admission control will reject queries when the number of reserved threads per backend // for the query exceeds this number. 0 or -1 means this has no effect. THREAD_RESERVATION_LIMIT = 65 // Admission control will reject queries when the total number of reserved threads // across all backends for the query exceeds this number. 0 or -1 means this has no // effect. THREAD_RESERVATION_AGGREGATE_LIMIT = 66 // Overrides the -kudu_read_mode flag to set the consistency level for Kudu scans. // Possible values are DEFAULT, READ_LATEST, and READ_AT_SNAPSHOT. KUDU_READ_MODE = 67 // Allow reading of erasure coded files. ALLOW_ERASURE_CODED_FILES = 68 // The timezone used in UTC<->localtime conversions. The default is the OS's timezone // at the coordinator, which can be overridden by environment variable $TZ. TIMEZONE = 69 // Scan bytes limit, after which a query will be terminated with an error. SCAN_BYTES_LIMIT = 70 // CPU time limit in seconds, after which a query will be terminated with an error. // Note that until IMPALA-7318 is fixed, CPU usage can be very stale and this may not // terminate queries soon enough. CPU_LIMIT_S = 71 // The max number of estimated bytes a TopN operator is allowed to materialize, if the // planner thinks a TopN operator will exceed this limit, it falls back to a TotalSort // operator which is capable of spilling to disk (unlike the TopN operator which keeps // everything in memory). 0 or -1 means this has no effect. TOPN_BYTES_LIMIT = 72 // An opaque string, not used by Impala itself, that can be used to identify // the client, like a User-Agent in HTTP. Drivers should set this to // their version number. May also be used by tests to help identify queries. CLIENT_IDENTIFIER = 73 // Probability to enable tracing of resource usage consumption on all fragment instance // executors of a query. Must be between 0 and 1 inclusive, 0 means no query will be // traced, 1 means all queries will be traced. RESOURCE_TRACE_RATIO = 74 // The maximum number of executor candidates to consider when scheduling remote // HDFS ranges. When non-zero, the scheduler generates a consistent set of // executor candidates based on the filename and offset. This algorithm is designed // to avoid changing file to node mappings when nodes come and go. It then picks from // among the candidates by the same method used for local scan ranges. Limiting the // number of nodes that can read a single file provides a type of simulated locality. // This increases the efficiency of file-related caches (e.g. the HDFS file handle // cache). If set to 0, the number of executor candidates is unlimited, and remote // ranges will be scheduled across all executors. NUM_REMOTE_EXECUTOR_CANDIDATES = 75 // A limit on the number of rows produced by the query. The query will be // canceled if the query is still executing after this limit is hit. A value // of 0 means there is no limit on the number of rows produced. NUM_ROWS_PRODUCED_LIMIT = 76 // Set when attempting to load a planner testcase. Typically used by developers for // debugging a testcase. Should not be set in user clusters. If set, a warning // is emitted in the query runtime profile. PLANNER_TESTCASE_MODE = 77 // Specifies the default table file format. DEFAULT_FILE_FORMAT = 78 // The physical type and unit used when writing timestamps in Parquet. // Valid values: INT96_NANOS, INT64_MILLIS, INT64_MICROS, INT64_NANOS // Default: INT96_NANOS PARQUET_TIMESTAMP_TYPE = 79 // Enable using the Parquet page index during scans. The page index contains min/max // statistics at page-level granularity. It can be used to skip pages and rows during // scanning. PARQUET_READ_PAGE_INDEX = 80 // Enable writing the Parquet page index. PARQUET_WRITE_PAGE_INDEX = 81 // Maximum number of rows written in a single Parquet data page. PARQUET_PAGE_ROW_COUNT_LIMIT = 82 // Disable the attempt to compute an estimated number of rows in an // hdfs table. DISABLE_HDFS_NUM_ROWS_ESTIMATE = 83 // Default hints for insert statement. Will be overridden by hints in the INSERT // statement, if any. DEFAULT_HINTS_INSERT_STATEMENT = 84 // Enable spooling of query results. If true, query results will be spooled in // memory up to a specified memory limit. If the memory limit is hit, the // coordinator fragment will block until the client has consumed enough rows to free // up more memory. If false, client consumption driven back-pressure controls the rate // at which rows are materialized by the execution tree. SPOOL_QUERY_RESULTS = 85 // Speficies the default transactional type for new HDFS tables. // Valid values: none, insert_only DEFAULT_TRANSACTIONAL_TYPE = 86 // Limit on the total number of expressions in the statement. Statements that exceed // the limit will get an error during analysis. This is intended to set an upper // bound on the complexity of statements to avoid resource impacts such as excessive // time in analysis or codegen. This is enforced only for the first pass of analysis // before any rewrites are applied. STATEMENT_EXPRESSION_LIMIT = 87 // Limit on the total length of a SQL statement. Statements that exceed the maximum // length will get an error before parsing/analysis. This is complementary to the // statement expression limit, because statements of a certain size are highly // likely to violate the statement expression limit. Rejecting them early avoids // the cost of parsing/analysis. MAX_STATEMENT_LENGTH_BYTES = 88 // Disable the data cache. DISABLE_DATA_CACHE = 89 // The maximum amount of memory used when spooling query results. If this value is // exceeded when spooling results, all memory will be unpinned and most likely spilled // to disk. Set to 100 MB by default. Only applicable if SPOOL_QUERY_RESULTS // is true. Setting this to 0 or -1 means the memory is unbounded. Cannot be set to // values below -1. MAX_RESULT_SPOOLING_MEM = 90 // The maximum amount of memory that can be spilled when spooling query results. Must be // greater than or equal to MAX_RESULT_SPOOLING_MEM to allow unpinning all pinned memory // if the amount of spooled results exceeds MAX_RESULT_SPOOLING_MEM. If this value is // exceeded, the coordinator fragment will block until the client has consumed enough // rows to free up more memory. Set to 1 GB by default. Only applicable if // SPOOL_QUERY_RESULTS is true. Setting this to 0 or -1 means the memory is unbounded. // Cannot be set to values below -1. MAX_SPILLED_RESULT_SPOOLING_MEM = 91 // Disable the normal key sampling of HBase tables in row count and row size estimation. // Set this to true will force the use of HMS table stats. DISABLE_HBASE_NUM_ROWS_ESTIMATE = 92 // The maximum amount of time, in milliseconds, a fetch rows request (TFetchResultsReq) // from the client should spend fetching results (including waiting for results to // become available and materialize). When result spooling is enabled, a fetch request // may read multiple RowBatches, in which case, the timeout controls how long the // client waits for all returned RowBatches to be produced. If the timeout is hit, the // client returns whatever rows it has already read. Defaults to 10000 milliseconds. A // value of 0 causes fetch requests to wait indefinitely. FETCH_ROWS_TIMEOUT_MS = 93 // For testing purposes only. This can provide a datetime string to use as now() for // tests. NOW_STRING = 94 // The split size of Parquet files when scanning non-block-based storage systems (e.g. // S3, ADLS, etc.). When reading from block-based storage systems (e.g. HDFS), Impala // sets the split size for Parquet files to the size of the blocks. This is done // because Impala assumes Parquet files have a single row group per block (which is // the recommended way Parquet files should be written). However, since non-block-based // storage systems have no concept of blocks, there is no way to derive a good default // value for Parquet split sizes. Defaults to 256 MB, which is the default size of // Parquet files written by Impala (Impala writes Parquet files with a single row // group per file). Must be >= 1 MB. PARQUET_OBJECT_STORE_SPLIT_SIZE = 95 // For testing purposes only. A per executor approximate limit on the memory consumption // of this query. Only applied if MEM_LIMIT is not specified. // unspecified or a limit of 0 means no limit; // Otherwise specified either as: // a) an int (= number of bytes); // b) a float followed by "M" (MB) or "G" (GB) MEM_LIMIT_EXECUTORS = 96 // The max number of estimated bytes eligible for a Broadcast operation during a join. // If the planner thinks the total bytes sent to all destinations of a broadcast // exchange will exceed this limit, it will not consider a broadcast and instead // fall back on a hash partition exchange. 0 or -1 means this has no effect. BROADCAST_BYTES_LIMIT = 97 // The max reservation that each grouping class in a preaggregation will use. // 0 or -1 means this has no effect. PREAGG_BYTES_LIMIT = 98 // Indicates whether the FE should rewrite disjunctive predicates to conjunctive // normal form (CNF) for optimization purposes. Default is true. ENABLE_CNF_REWRITES = 99 // The max number of conjunctive normal form (CNF) exprs to create when converting // a disjunctive expression to CNF. Each AND counts as 1 expression. A value of // -1 or 0 means no limit. Default is 200. MAX_CNF_EXPRS = 100 // Set the timestamp for Kudu snapshot reads in Unix time micros. Only valid if // KUDU_READ_MODE is set to READ_AT_SNAPSHOT. KUDU_SNAPSHOT_READ_TIMESTAMP_MICROS = 101 // Transparently retry queries that fail due to cluster membership changes. A cluster // membership change includes blacklisting a node and the statestore detecting that a // node has been removed from the cluster membership. From Impala's perspective, a // retried query is a brand new query. From the client perspective, requests for the // failed query are transparently re-routed to the new query. RETRY_FAILED_QUERIES = 102 // Enabled runtime filter types to be applied to scanner. // This option only apply to Hdfs scan node and Kudu scan node. // Specify the enabled types by a comma-separated list or enable all types by "ALL". // BLOOM - apply bloom filter only, // MIN_MAX - apply min-max filter only. // IN_LIST - apply in-list filter only. // ALL - apply all types of runtime filters. // Default is [BLOOM, MIN_MAX]. // Depending on the scan node type, Planner can schedule compatible runtime filter type // as follows: // Kudu scan: BLOOM, MIN_MAX // Hdfs scan on Parquet file: BLOOM, MIN_MAX // Hdfs scan on ORC file: BLOOM, IN_LIST // Hdfs scan on other kind of file: BLOOM ENABLED_RUNTIME_FILTER_TYPES = 103 // Enable asynchronous codegen. ASYNC_CODEGEN = 104 // If true, the planner will consider adding a distinct aggregation to SEMI JOIN // operations. If false, disables the optimization (i.e. falls back to pre-Impala-4.0 // behaviour). ENABLE_DISTINCT_SEMI_JOIN_OPTIMIZATION = 105 // The max reservation that sorter will use for intermediate sort runs. // 0 or -1 means this has no effect. SORT_RUN_BYTES_LIMIT = 106 // Sets an upper limit on the number of fs writer instances to be scheduled during // insert. Currently this limit only applies to HDFS inserts. MAX_FS_WRITERS = 107 // When this query option is set, a refresh table statement will detect existing // partitions which have been changed in metastore and refresh them. By default, this // option is disabled since there is additional performance hit to fetch all the // partitions and detect if they are not same as ones in the catalogd. Currently, this // option is only applicable for refresh table statement. REFRESH_UPDATED_HMS_PARTITIONS = 108 // If RETRY_FAILED_QUERIES and SPOOL_QUERY_RESULTS are enabled and this is true, // retryable queries will try to spool all results before returning any to the client. // If the result set is too large to fit into the spooling memory (including the spill // mem), results will be returned and the query will not be retryable. This may have // some performance impact. Set it to false then clients can fetch results immediately // when any of them are ready. Note that in this case, query retry will be skipped if // the client has fetched some results. SPOOL_ALL_RESULTS_FOR_RETRIES = 109 // A value (0.0, 1.0) that is the target false positive probability for runtime bloom // filters. If not set, falls back to max_filter_error_rate. RUNTIME_FILTER_ERROR_RATE = 110 // When true, TIMESTAMPs are interpreted in the local time zone (set in query option // TIMEZONE) when converting to and from Unix times. // When false, TIMESTAMPs are interpreted in the UTC time zone. USE_LOCAL_TZ_FOR_UNIX_TIMESTAMP_CONVERSIONS = 111 // When true, TIMESTAMPs read from files written by Parquet-MR (used by Hive) will // be converted from UTC to local time. Writes are unaffected. CONVERT_LEGACY_HIVE_PARQUET_UTC_TIMESTAMPS = 112 // Indicates whether the FE should attempt to transform outer joins into inner joins. ENABLE_OUTER_JOIN_TO_INNER_TRANSFORMATION = 113 // Set the target scan range length for scanning kudu tables (in bytes). This is // used to split kudu scan tokens and is treated as a hint by kudu. Therefore, // does not guarantee a limit on the size of the scan range. If unspecified or // set to 0 disables this feature. TARGETED_KUDU_SCAN_RANGE_LENGTH = 114 // Enable (>=0) or disable(<0) reporting of skews for a query in runtime profile. // When enabled, used as the CoV threshold value in the skew detection formula. REPORT_SKEW_LIMIT = 115 // If true, for simple limit queries (limit with no order-by, group-by, aggregates, // joins, analytic functions but does allow where predicates) optimize the planning // time by only considering a small number of partitions. The number of files within // a partition is used as an approximation to number of rows (1 row per file). // This option is opt-in by default as this optimization may in some cases produce // fewer (but correct) rows than the limit value in the query. OPTIMIZE_SIMPLE_LIMIT = 116 // When true, the degree of parallelism (if > 1) is used in costing decision // of a broadcast vs partition distribution. USE_DOP_FOR_COSTING = 117 // A multiplying factor between 0 to 1000 that is applied to the costing decision of // a broadcast vs partition distribution. Fractional values between 0 to 1 favor // broadcast by reducing the build side cost of a broadcast join. Values above 1.0 // favor partition distribution. BROADCAST_TO_PARTITION_FACTOR = 118 // A limit on the number of join rows produced by the query. The query will be // canceled if the query is still executing after this limit is hit. A value // of 0 means there is no limit on the number of join rows produced. JOIN_ROWS_PRODUCED_LIMIT = 119 // If true, strings are processed in a UTF-8 aware way, e.g. counting lengths by UTF-8 // characters instead of bytes. UTF8_MODE = 120 // If > 0, the rank()/row_number() pushdown into pre-analytic sorts is enabled // if the limit would be less than or equal to the threshold. // If 0 or -1, disables the optimization (i.e. falls back to pre-Impala-4.0 // behaviour). // Default is 1000. Setting it higher increases the max size of the in-memory heaps // used in the top-n operation. The larger the heaps, the less beneficial the // optimization is compared to a full sort and the more potential for perf regressions. ANALYTIC_RANK_PUSHDOWN_THRESHOLD = 121 // Specifiy a threshold between 0.0 and 1.0 to control the operation of overlap // predicates (via min/max filters) for equi-hash joins at the Parquet table scan // operator. Set to 0.0 to disable the feature. Set to 1.0 to evaluate all // overlap predicates. Set to a value between 0.0 and 1.0 to evaluate only those // with an overlap ratio less than the threshold. MINMAX_FILTER_THRESHOLD = 122 // Minmax filtering level to be applied to Parquet scanners. // ROW_GROUP - apply to row groups only, // PAGE - apply to row groups and pages only. // ROW - apply to row groups, pages and rows. MINMAX_FILTERING_LEVEL = 123 // If true, compute the column min and max stats during compute stats. COMPUTE_COLUMN_MINMAX_STATS = 124 // If true, show the min and max stats during show column stats. SHOW_COLUMN_MINMAX_STATS = 125 // Default NDV scale settings, make it easier to change scale in SQL function // NDV(<expr>). DEFAULT_NDV_SCALE = 126 // Policy with which to choose amongst multiple Kudu replicas. // LEADER_ONLY - Select the LEADER replica. // CLOSEST_REPLICA - Select the closest replica to the client (default). KUDU_REPLICA_SELECTION = 127 // If false, a truncate DDL operation will not delete table stats. // the default value of this is true as set in Query.thrift DELETE_STATS_IN_TRUNCATE = 128 // Indicates whether to use Bloom filtering for Parquet files PARQUET_BLOOM_FILTERING = 129 // If true, generate min/max filters when join into sorted columns. MINMAX_FILTER_SORTED_COLUMNS = 130 // Control setting for taking the fast code path when min/max filtering sorted columns. // OFF - Take the regular code path, // ON - Take the fast code path, // VERIFICATION - Take both code paths and verify that the results from both are // the same. MINMAX_FILTER_FAST_CODE_PATH = 131 // If true, Kudu's multi-row transaction is enabled. ENABLE_KUDU_TRANSACTION = 132 // Indicates whether to use min/max filtering on partition columns MINMAX_FILTER_PARTITION_COLUMNS = 133 // Controls when to write Parquet Bloom filters. // NEVER - never write Parquet Bloom filters // TBL_PROPS - write Parquet Bloom filters as set in table properties // IF_NO_DICT - write Parquet Bloom filters if the row group is not fully // dictionary encoded // ALWAYS - always write Parquet Bloom filters, even if the row group is fully // dictionary encoded PARQUET_BLOOM_FILTER_WRITE = 134 // Indicates whether to use ORC's search argument to push down predicates. ORC_READ_STATISTICS = 135 // Indicates whether to run most of ddl requests in async mode. ENABLE_ASYNC_DDL_EXECUTION = 136 // Indicates whether to run load data requests in async mode. ENABLE_ASYNC_LOAD_DATA_EXECUTION = 137 // Number of minimum consecutive rows when filtered out, will avoid materialization // of columns in parquet. Set it to -1 to turn off late materialization feature. PARQUET_LATE_MATERIALIZATION_THRESHOLD = 138; // Max entries in the dictionary before skipping runtime filter evaluation for row // groups. If a dictionary has many entries, then runtime filter evaluation is more // likely to give false positive results, which means that the row groups won't be // rejected. Set it to 0 to disable runtime filter dictionary filtering, above 0 will // enable runtime filtering on the row group. For example, 2 means that runtime filter // will be evaluated when the dictionary size is smaller or equal to 2. PARQUET_DICTIONARY_RUNTIME_FILTER_ENTRY_LIMIT = 139; // Abort the Java UDF if an exception is thrown. Default is that only a // warning will be logged if the Java UDF throws an exception. ABORT_JAVA_UDF_ON_EXCEPTION = 140; // Indicates whether to use ORC's async read. ORC_ASYNC_READ = 141 // Maximum number of distinct entries in a runtime in-list filter. RUNTIME_IN_LIST_FILTER_ENTRY_LIMIT = 142; // If true, replanning is enabled. ENABLE_REPLAN = 143; // If true, test replan by imposing artificial two executor groups in FE and always // compute ProcessingCost. The degree of parallelism adjustment, however, still require // COMPUTE_PROCESSING_COST option set to true. TEST_REPLAN = 144; // Maximum wait time on HMS ACID lock in seconds. LOCK_MAX_WAIT_TIME_S = 145 // Determines how to resolve ORC files' schemas. Valid values are "position" and "name". ORC_SCHEMA_RESOLUTION = 146; // Expands complex types in star queries EXPAND_COMPLEX_TYPES = 147; // Specify the database name which stores global udf FALLBACK_DB_FOR_FUNCTIONS = 148; // Specify whether to use codegen cache. DISABLE_CODEGEN_CACHE = 149; // Specify how the entry stores to the codegen cache, would affect the entry size. // Possible values are NORMAL, OPTIMAL, NORMAL_DEBUG and OPTIMAL_DEBUG. // The normal mode will use a full key for the cache, while the optimal mode uses // a hashcode of 128 bits for the key to save the memory consumption. // The debug mode of each of them allows more logs, would be helpful to target // an issue. // Only valid if DISABLE_CODEGEN_CACHE is set to false. CODEGEN_CACHE_MODE = 150; // Convert non-string map keys to string to produce valid JSON. STRINGIFY_MAP_KEYS = 151 // Enable immediate admission for trivial queries. ENABLE_TRIVIAL_QUERY_FOR_ADMISSION = 152 // Control whether to consider CPU processing cost during query planning. COMPUTE_PROCESSING_COST = 153; // Minimum number of threads of a query fragment per node in processing // cost algorithm. It is recommend to not set it with value more than number of // physical cores in executor node. Valid values are in [1, 128]. Default to 1. PROCESSING_COST_MIN_THREADS = 154; // When calculating estimated join cardinality of 2 or more conjuncts // e.g t1.a1 = t2.a2 AND t1.b1 = t2.b2, this selectivity correlation factor // provides more control over the join cardinality estimation. The range is a // double value between 0 to 1 inclusive. The default value of 0 preserves the // existing behavior of using the minimum cardinality of the conjucts. Setting // this to a value between 0 to 1 computes the combined selectivity by // taking the product of the selectivities and dividing by this factor. JOIN_SELECTIVITY_CORRELATION_FACTOR = 155; // Maximum number of threads of a query fragment per node in processing // cost algorithm. It is recommend to not set it with value more than number of // physical cores in executor node. This query option may be ignored if selected // executor group has lower max-query-cpu-core-per-node-limit configuration or // PROCESSING_COST_MIN_THREADS option has higher value. // Valid values are in [1, 128]. Default to 128. MAX_FRAGMENT_INSTANCES_PER_NODE = 156 // Configures the in-memory sort algorithm used in the sorter. Determines the maximum // number of pages in an initial in-memory run (fixed + variable length). // Maximizing the sort run size can help mitigate back-pressure in the sorter. It // creates multiple miniruns and merges them in-memory. The run size must be at least 2, // but 10 or more are recommended to avoid high fragmentation of variable length data. // Setting 0 or a negative value disables the run size limitation. // Defaults to 0 (disabled). MAX_SORT_RUN_SIZE = 157; // Allowing implicit casts with loss of precision, adds the capability to use // implicit casts between numeric and string types in set operations and insert // statements. ALLOW_UNSAFE_CASTS = 158; // The maximum number of threads Impala can use for migrating a table to a different // type. E.g. from Hive table to Iceberg table. NUM_THREADS_FOR_TABLE_MIGRATION = 159; // Turns off optimized Iceberg V2 reads, falls back to Hash Join DISABLE_OPTIMIZED_ICEBERG_V2_READ = 160; // In VALUES clauses, if all values in a column are CHARs but they have different // lengths, choose the VARCHAR type of the longest length instead of the corresponding // CHAR type as the common type. This avoids padding and thereby loss of information. // See IMPALA-10753. VALUES_STMT_AVOID_LOSSY_CHAR_PADDING = 161; // Threshold in bytes to determine whether an aggregation node's memory estimate is // deemed large. If an aggregation node's memory estimate is large, an alternative // estimation is used to lower the memory usage estimation for that aggregation node. // The new memory estimate will not be lower than the specified // LARGE_AGG_MEM_THRESHOLD. Unlike PREAGG_BYTES_LIMIT, LARGE_AGG_MEM_THRESHOLD is // evaluated on both preagg and merge agg, and does not cap max memory reservation of // the aggregation node (it may still increase memory allocation beyond the threshold // if it is available). However, if a plan node is a streaming preaggregation node and // PREAGG_BYTES_LIMIT is set, then PREAGG_BYTES_LIMIT will override the value of // LARGE_AGG_MEM_THRESHOLD as a threshold. 0 or -1 means this option has no effect. LARGE_AGG_MEM_THRESHOLD = 162 // Correlation factor that will be used to calculate a lower memory estimation of // aggregation node when the default memory estimation exceeds // LARGE_AGG_MEM_THRESHOLD. The reduction is achieved by calculating a memScale // multiplier (a fraction between 0.0 and 1.0). Given N as number of non-literal // grouping expressions: // // memScale = (1.0 - AGG_MEM_CORRELATION_FACTOR) ^ max(0, N - 1) // // Valid values are in [0.0, 1.0]. Note that high value of AGG_MEM_CORRELATION_FACTOR // value means there is high correlation between grouping expressions / columns, while // low value means there is low correlation between them. High correlation means // aggregation node can be scheduled with lower memory estimation (lower memScale). // Setting value 0.0 will result in an equal memory estimate as the default estimation // (no change). Defaults to 0.5. AGG_MEM_CORRELATION_FACTOR = 163 // A per coordinator approximate limit on the memory consumption // of this query. Only applied if MEM_LIMIT is not specified. // Unspecified or a limit of 0 or negative value means no limit; // Otherwise specified either as: // a) an int (= number of bytes); // b) a float followed by "M" (MB) or "G" (GB) MEM_LIMIT_COORDINATORS = 164 // Enables predicate subsetting for Iceberg plan nodes. If enabled, expressions // evaluated by Iceberg are not pushed down the scanner node. ICEBERG_PREDICATE_PUSHDOWN_SUBSETTING = 165; // Amount of memory that we approximate a scanner thread will use not including I/O // buffers. The memory used does not vary considerably between file formats (just a // couple of MBs). This amount of memory is not reserved by the planner and only // considered in the old multi-threaded scanner mode (non-MT_DOP) for 2nd and // subsequent additional scanner threads. If this option is not set to a positive // value, the value of flag hdfs_scanner_thread_max_estimated_bytes will be used // (which defaults to 32MB). The default value of this option is -1 (not set). HDFS_SCANNER_NON_RESERVED_BYTES = 166 // Select codegen optimization level from O0, O1, Os, O2, or O3. Higher levels will // overwrite existing codegen cache entries. CODEGEN_OPT_LEVEL = 167 // The reservation time (in seconds) for deleted impala-managed Kudu tables. // During this time deleted Kudu tables can be recovered by Kudu's 'recall table' API. // See KUDU-3326 for details. KUDU_TABLE_RESERVE_SECONDS = 168 // When true, UNIXTIME_MICRO columns read from Kudu will be interpreted as UTC and // and UTC->local timezone conversion is applied when converting to Impala TIMESTAMP. // Writes are unaffected (see WRITE_KUDU_UTC_TIMESTAMPS). CONVERT_KUDU_UTC_TIMESTAMPS = 169 // This only makes sense when 'CONVERT_KUDU_UTC_TIMESTAMPS' is true. When true, it // disables the bloom filter for Kudu's timestamp type, because using local timestamp in // Kudu bloom filter may cause missing rows. // Local timestamp convert to UTC could be ambiguous in the case of DST change. // We can only put one of the two possible UTC timestamps in the bloom filter // for now, which may cause missing rows that have the other UTC timestamp. // For those regions that do not observe DST, could set this flag to false // to re-enable kudu local timestamp bloom filter. DISABLE_KUDU_LOCAL_TIMESTAMP_BLOOM_FILTER = 170 // A range of [0.0..1.0] that controls the cardinality reduction scale from runtime // filter analysis. This is a linear scale with 0.0 meaning no cardinality estimate // reduction should be applied and 1.0 meaning maximum cardinality estimate reduction // should be applied. For example, if a table has 1M rows and runtime filters are // estimated to reduce cardinality to 500K, setting value 0.25 will result in an 875K // cardinality estimate. Default to 1.0. RUNTIME_FILTER_CARDINALITY_REDUCTION_SCALE = 171 // Maximum number of backend executor that can send bloom runtime filter updates to // one intermediate aggregator. Given N as number of backend executor excluding // coordinator, the selected number of designated intermediate aggregator is // ceil(N / MAX_NUM_FILTERS_AGGREGATED_PER_HOST). Setting 1, 0, or negative value // will disable the intermediate aggregator feature. Default to -1 (disabled). MAX_NUM_FILTERS_AGGREGATED_PER_HOST = 172 // Divide the CPU requirement of a query to fit the total available CPU in // the executor group. For example, setting value 2 will fit the query with CPU // requirement 2X to an executor group with total available CPU X. Note that setting // with a fractional value less than 1 effectively multiplies the query CPU // requirement. A valid value is > 0.0. // If this query option is not set, value of backend flag --query_cpu_count_divisor // (default to 1.0) will be picked up instead. QUERY_CPU_COUNT_DIVISOR = 173 // Enables intermediate result caching. The frontend will determine eligibility and // potentially insert tuple cache nodes into the plan. This can only be set if the // allow_tuple_caching feature startup flag is set to true. ENABLE_TUPLE_CACHE = 174 // Disables statistic-based count(*)-optimization for Iceberg tables. ICEBERG_DISABLE_COUNT_STAR_OPTIMIZATION = 175 // List of runtime filter id to skip if it exists in query plan. // If using JDBC client, use double quote to wrap multiple ids, like: // RUNTIME_FILTER_IDS_TO_SKIP="1,2,3" // If using impala-shell client, double quote is not required. RUNTIME_FILTER_IDS_TO_SKIP = 176 // Decide what strategy to use to compute number of slot per node to run a query. // Default to number of instances of largest query fragment (LARGEST_FRAGMENT). // See TSlotCountStrategy in Query.thrift for documentation of its possible values. SLOT_COUNT_STRATEGY = 177 // Indicate if external JDBC table handler should clean DBCP DataSource object from // cache when its reference count equals 0. By caching DBCP DataSource objects, we can // avoid to reload JDBC driver. CLEAN_DBCP_DS_CACHE = 178 // Enables cache for isTrueWithNullSlots, which can be expensive when evaluating lots // of expressions. The cache helps with generated expressions, which often contain lots // of repeated patterns. USE_NULL_SLOTS_CACHE = 179 // When true, Impala TIMESTAMPs are converted from local timezone to UTC before being // written to Kudu as UNIXTIME_MICRO. // Reads are unaffected (see CONVERT_KUDU_UTC_TIMESTAMPS). WRITE_KUDU_UTC_TIMESTAMPS = 180 // Turns off optimized JSON count star (zero slots) scan, falls back to rapidjson parse. DISABLE_OPTIMIZED_JSON_COUNT_STAR = 181 // How long to wait for statement completion for ExecuteStatement/executeAndWait and // GetOperationStatus/get_state RPCs. Waiting on the server side allows for immediate // notification when the query completes and avoid added latency from waiting on the // client side. This defaults to off (0ms). LONG_POLLING_TIME_MS = 182 // Enables the verification process for intermediate result caching. // Tuple cache verification is performed only when the startup flag // tuple_cache_debug_dump_dir is specified and enable_tuple_cache_verification is set // to true. ENABLE_TUPLE_CACHE_VERIFICATION = 183 // If True, enable tuple analysis for both preaggregation and final aggregation node. // Enabling this feature can lower cardinality estimate of multi-column grouping. ENABLE_TUPLE_ANALYSIS_IN_AGGREGATE = 184 // If True, account for probability of having duplicate grouping key exist in multiple // nodes during preaggreation. ESTIMATE_DUPLICATE_IN_PREAGG = 185 // When true and CONVERT_LEGACY_HIVE_PARQUET_UTC_TIMESTAMPS is also enabled, TIMESTAMP // conversion to local time will fallback to the timestamp conversion method from Hive // 3.0 and earlier if not specified in the file. This matches the Hive option // 'hive.parquet.timestamp.legacy.conversion.enabled', which defaults to true. Impala // defaults to false because conversion is ~50x slower than Impala's default conversion // method and they produce the same results for modern time periods (post 1970, and in // most instances before that). USE_LEGACY_HIVE_TIMESTAMP_CONVERSION = 186 // Maximum time in seconds to wait for catalogd catching up HMS events before query // planning. Only events generated before the query is submitted will be waited for. // Defaults to 0 which disables waiting. Please only use this on queries that depend on // external modifications. Don't set it cluster wide since it impacts performance. SYNC_HMS_EVENTS_WAIT_TIME_S = 187 // Whether to fail the query if coordinator fails to wait for HMS events to be synced // in catalogd, e.g. when event-processor is in ERROR state or timed out waiting for // SYNC_HMS_EVENTS_WAIT_TIME_S seconds. Defaults to false, which means coordinator will // start query planning regardless of the failure. SYNC_HMS_EVENTS_STRICT_MODE = 188 // Turns on an optimization in UPDATE operation that skips updating rows which already // have the desired value. The filtering is on by default, skipping unnecessary updates // if there are up to 10 items in the SET list. If more columns are updated than the // given value, then no filtering is done. // Set it non-positive to turn this feature off completely. SKIP_UNNEEDED_UPDATES_COL_LIMIT = 189 // A scale between (0.0, 1.0] (not including 0.0) to control estimate peak memory of // query operators that have spill-to-disk capabilities. Setting value closer to 1.0 // will make Planner bias towards keeping as much rows as possible in memory, while // setting value closer to 0.0 will make Planner bias towards spilling rows to disk // under memory pressure. // Resulting memory estimate will never be less than minimum memory required for // a query operator to start execution. MEM_LIMIT, MEM_LIMIT_COORDINATORS, // MEM_LIMIT_EXECUTORS, per-pool memory clamping configuration, and // MAX_MEM_ESTIMATE_FOR_ADMISSION will be considered into peak memory estimation if any // of them is set. // Setting to 0.0 or unsetting scratch_dirs flag will revert Planner to old behavior // where it assumes that all rows must be held in memory, regardless of any memory // limit configurations mentioned above. It is recommended to set this query option // to either 1.0 (enabled) or 0.0 (disabled) and only set value in between for // experimental purpose. MEM_ESTIMATE_SCALE_FOR_SPILLING_OPERATOR = 190 // If True, use the Calcite planner for compilation USE_CALCITE_PLANNER = 191 // The default format for reading JSON binary columns, can be overridden by table // property 'json.binary.format' (if set). The valid values are: // NONE - default value, means unspecified format, depends on the table property. // BASE64 - the json binary data is read as base64 encoded string. // RAWSTRING - the json binary data is read as raw string. JSON_BINARY_FORMAT = 192 } // The summary of a DML statement. struct TDmlResult { // Number of modified rows per partition. Only applies to HDFS and Kudu tables. // The keys represent partitions to create, coded as k1=v1/k2=v2/k3=v3..., with // the root in an unpartitioned table being the empty string. 1: required map<string, i64> rows_modified 3: optional map<string, i64> rows_deleted // Number of row operations attempted but not completed due to non-fatal errors // reported by the storage engine that Impala treats as warnings. Only applies to Kudu // tables. This includes errors due to duplicate/missing primary keys, nullability // constraint violations, and primary keys in uncovered partition ranges. // TODO: Provide a detailed breakdown of these counts by error. IMPALA-4416. 2: optional i64 num_row_errors } // Response from a call to PingImpalaService struct TPingImpalaServiceResp { // The Impala service's version string. 1: string version // The Impalad's webserver address. 2: string webserver_address } // Parameters for a ResetTable request which will invalidate a table's metadata. // DEPRECATED. struct TResetTableReq { // Name of the table's parent database. 1: required string db_name // Name of the table. 2: required string table_name } // PingImpalaHS2Service() - ImpalaHiveServer2Service version. // Pings the Impala server to confirm that the server is alive and the session identified // by 'sessionHandle' is open. Returns metadata about the server. This exists separate // from the base HS2 GetInfo() methods because not all relevant metadata is accessible // through GetInfo(). struct TPingImpalaHS2ServiceReq { 1: required TCLIService.TSessionHandle sessionHandle } struct TPingImpalaHS2ServiceResp { 1: required TCLIService.TStatus status // The Impala service's version string. 2: optional string version // The Impalad's webserver address. 3: optional string webserver_address // The Impalad's local monotonic time 4: optional i64 timestamp } // CloseImpalaOperation() // // Extended version of CloseOperation() that, if the operation was a DML // operation, returns statistics about the operation. struct TCloseImpalaOperationReq { 1: required TCLIService.TOperationHandle operationHandle } struct TCloseImpalaOperationResp { 1: required TCLIService.TStatus status // Populated if the operation was a DML operation. 2: optional TDmlResult dml_result } // For all rpc that return a TStatus as part of their result type, // if the status_code field is set to anything other than OK, the contents // of the remainder of the result type is undefined (typically not set) service ImpalaService extends beeswax.BeeswaxService { // Cancel execution of query. Returns RUNTIME_ERROR if query_id // unknown. // This terminates all threads running on behalf of this query at // all nodes that were involved in the execution. // Throws BeeswaxException if the query handle is invalid (this doesn't // necessarily indicate an error: the query might have finished). Status.TStatus Cancel(1:beeswax.QueryHandle query_id) throws(1:beeswax.BeeswaxException error); // Invalidates all catalog metadata, forcing a reload // DEPRECATED; execute query "invalidate metadata" to refresh metadata Status.TStatus ResetCatalog(); // Invalidates a specific table's catalog metadata, forcing a reload on the next access // DEPRECATED; execute query "refresh <table>" to refresh metadata Status.TStatus ResetTable(1:TResetTableReq request) // Returns the runtime profile string for the given query handle. string GetRuntimeProfile(1:beeswax.QueryHandle query_id) throws(1:beeswax.BeeswaxException error); // Closes the query handle and return the result summary of the insert. TDmlResult CloseInsert(1:beeswax.QueryHandle handle) throws(1:beeswax.QueryNotFoundException error, 2:beeswax.BeeswaxException error2); // Client calls this RPC to verify that the server is an ImpalaService. Returns the // server version. TPingImpalaServiceResp PingImpalaService(); // Returns the summary of the current execution. ExecStats.TExecSummary GetExecSummary(1:beeswax.QueryHandle handle) throws(1:beeswax.QueryNotFoundException error, 2:beeswax.BeeswaxException error2); } // Impala HiveServer2 service struct TGetExecSummaryReq { 1: optional TCLIService.TOperationHandle operationHandle 2: optional TCLIService.TSessionHandle sessionHandle // If true, returns the summaries of all query attempts. A TGetExecSummaryResp // always returns the profile for the most recent query attempt, regardless of the // query id specified. Clients should set this to true if they want to retrieve the // summaries of all query attempts (including the failed ones). 3: optional bool include_query_attempts = false } struct TGetExecSummaryResp { 1: required TCLIService.TStatus status 2: optional ExecStats.TExecSummary summary // A list of all summaries of the failed query attempts. 3: optional list<ExecStats.TExecSummary> failed_summaries } struct TGetRuntimeProfileReq { 1: optional TCLIService.TOperationHandle operationHandle 2: optional TCLIService.TSessionHandle sessionHandle 3: optional RuntimeProfile.TRuntimeProfileFormat format = RuntimeProfile.TRuntimeProfileFormat.STRING // If true, returns the profiles of all query attempts. A TGetRuntimeProfileResp // always returns the profile for the most recent query attempt, regardless of the // query id specified. Clients should set this to true if they want to retrieve the // profiles of all query attempts (including the failed ones). 4: optional bool include_query_attempts = false } struct TGetRuntimeProfileResp { 1: required TCLIService.TStatus status // Will be set on success if TGetRuntimeProfileReq.format // was STRING, BASE64 or JSON. 2: optional string profile // Will be set on success if TGetRuntimeProfileReq.format was THRIFT. 3: optional RuntimeProfile.TRuntimeProfileTree thrift_profile // A list of all the failed query attempts in either STRING, BASE64 or JSON format. 4: optional list<string> failed_profiles // A list of all the failed query attempts in THRIFT format. 5: optional list<RuntimeProfile.TRuntimeProfileTree> failed_thrift_profiles } // ExecutePlannedStatement() // // Execute a statement where the ExecRequest has been externally supplied. // The returned OperationHandle can be used to check on the // status of the plan, and to fetch results once the // plan has finished executing. struct TExecutePlannedStatementReq { 1: required TCLIService.TExecuteStatementReq statementReq // The plan to be executed 2: required Frontend.TExecRequest plan } struct TGetBackendConfigReq { 1: required TCLIService.TSessionHandle sessionHandle } struct TGetBackendConfigResp { 1: required TCLIService.TStatus status 2: required BackendGflags.TBackendGflags backend_config } struct TGetExecutorMembershipReq { 1: required TCLIService.TSessionHandle sessionHandle } struct TGetExecutorMembershipResp { 1: required TCLIService.TStatus status 2: required Frontend.TUpdateExecutorMembershipRequest executor_membership } struct TInitQueryContextResp { 1: required TCLIService.TStatus status 2: required Query.TQueryCtx query_ctx } service ImpalaHiveServer2Service extends TCLIService.TCLIService { // Returns the exec summary for the given query. The exec summary is only valid for // queries that execute with Impala's backend, i.e. QUERY, DML and COMPUTE_STATS // queries. Otherwise a default-initialized TExecSummary is returned for // backwards-compatibility with impala-shell - see IMPALA-9729. TGetExecSummaryResp GetExecSummary(1:TGetExecSummaryReq req); // Returns the runtime profile string for the given query TGetRuntimeProfileResp GetRuntimeProfile(1:TGetRuntimeProfileReq req); // Client calls this RPC to verify that the server is an ImpalaService. Returns the // server version. TPingImpalaHS2ServiceResp PingImpalaHS2Service(1:TPingImpalaHS2ServiceReq req); // Same as HS2 CloseOperation but can return additional information. TCloseImpalaOperationResp CloseImpalaOperation(1:TCloseImpalaOperationReq req); // Returns an initialized TQueryCtx. Only supported for the "external fe" service. TInitQueryContextResp InitQueryContext(); // Execute statement with supplied ExecRequest TCLIService.TExecuteStatementResp ExecutePlannedStatement( 1:TExecutePlannedStatementReq req); // Returns the current TBackendGflags. Only supported for the "external fe" service. TGetBackendConfigResp GetBackendConfig(1:TGetBackendConfigReq req); // Returns the executor membership information. Only supported for the "external fe" // service. TGetExecutorMembershipResp GetExecutorMembership(1:TGetExecutorMembershipReq req); }