diff --git a/docs/reference.asciidoc b/docs/reference.asciidoc index 3cf8ba8a8..5d85a21cb 100644 --- a/docs/reference.asciidoc +++ b/docs/reference.asciidoc @@ -324,10 +324,7 @@ client.create({ id, index }) ** *`id` (string)*: A unique identifier for the document. To automatically generate a document ID, use the `POST //_doc/` request format. ** *`index` (string)*: The name of the data stream or index to target. If the target doesn't exist and matches the name or wildcard (`*`) pattern of an index template with a `data_stream` definition, this request creates the data stream. If the target doesn't exist and doesn’t match a data stream template, this request creates the index. ** *`document` (Optional, object)*: A document. -** *`if_primary_term` (Optional, number)*: Only perform the operation if the document has this primary term. -** *`if_seq_no` (Optional, number)*: Only perform the operation if the document has this sequence number. ** *`include_source_on_error` (Optional, boolean)*: True or false if to include the document source in the error message in case of parsing errors. -** *`op_type` (Optional, Enum("index" | "create"))*: Set to `create` to only index the document if it does not already exist (put if absent). If a document with the specified `_id` already exists, the indexing operation will fail. The behavior is the same as using the `/_create` endpoint. If a document ID is specified, this paramater defaults to `index`. Otherwise, it defaults to `create`. If the request targets a data stream, an `op_type` of `create` is required. ** *`pipeline` (Optional, string)*: The ID of the pipeline to use to preprocess incoming documents. If the index has a default ingest pipeline specified, setting the value to `_none` turns off the default ingest pipeline for this request. If a final pipeline is configured, it will always run regardless of the value of this parameter. ** *`refresh` (Optional, Enum(true | false | "wait_for"))*: If `true`, Elasticsearch refreshes the affected shards to make this operation visible to search. If `wait_for`, it waits for a refresh to make this operation visible to search. If `false`, it does nothing with refreshes. ** *`require_alias` (Optional, boolean)*: If `true`, the destination must be an index alias. @@ -2824,7 +2821,7 @@ client.cat.indices({ ... }) Supports wildcards (`*`). To target all data streams and indices, omit this parameter or use `*` or `_all`. ** *`bytes` (Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb"))*: The unit used to display byte values. ** *`expand_wildcards` (Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])*: The type of index that wildcard patterns can match. -** *`health` (Optional, Enum("green" | "yellow" | "red"))*: The health status used to limit returned indices. By default, the response includes indices of any health status. +** *`health` (Optional, Enum("green" | "yellow" | "red" | "unknown" | "unavailable"))*: The health status used to limit returned indices. By default, the response includes indices of any health status. ** *`include_unloaded_segments` (Optional, boolean)*: If true, the response includes information from segments that are not loaded into memory. ** *`pri` (Optional, boolean)*: If true, the response only includes information from primary shards. ** *`time` (Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d"))*: The unit used to display time values. @@ -3250,7 +3247,7 @@ Accepts wildcard expressions. `_all` returns all repositories. If any repository fails during the request, Elasticsearch returns an error. ** *`ignore_unavailable` (Optional, boolean)*: If `true`, the response does not include information from unavailable snapshots. -** *`h` (Optional, Enum("id" | "repository" | "status" | "start_epoch" | "start_time" | "end_epoch" | "end_time" | "duration" | "indices" | "successful_shards" | "failed_shards" | "total_shards" | "reason") | Enum("build" | "completion.size" | "cpu" | "disk.avail" | "disk.total" | "disk.used" | "disk.used_percent" | "fielddata.evictions" | "fielddata.memory_size" | "file_desc.current" | "file_desc.max" | "file_desc.percent" | "flush.total" | "flush.total_time" | "get.current" | "get.exists_time" | "get.exists_total" | "get.missing_time" | "get.missing_total" | "get.time" | "get.total" | "heap.current" | "heap.max" | "heap.percent" | "http_address" | "id" | "indexing.delete_current" | "indexing.delete_time" | "indexing.delete_total" | "indexing.index_current" | "indexing.index_failed" | "indexing.index_failed_due_to_version_conflict" | "indexing.index_time" | "indexing.index_total" | "ip" | "jdk" | "load_1m" | "load_5m" | "load_15m" | "mappings.total_count" | "mappings.total_estimated_overhead_in_bytes" | "master" | "merges.current" | "merges.current_docs" | "merges.current_size" | "merges.total" | "merges.total_docs" | "merges.total_size" | "merges.total_time" | "name" | "node.role" | "pid" | "port" | "query_cache.memory_size" | "query_cache.evictions" | "query_cache.hit_count" | "query_cache.miss_count" | "ram.current" | "ram.max" | "ram.percent" | "refresh.total" | "refresh.time" | "request_cache.memory_size" | "request_cache.evictions" | "request_cache.hit_count" | "request_cache.miss_count" | "script.compilations" | "script.cache_evictions" | "search.fetch_current" | "search.fetch_time" | "search.fetch_total" | "search.open_contexts" | "search.query_current" | "search.query_time" | "search.query_total" | "search.scroll_current" | "search.scroll_time" | "search.scroll_total" | "segments.count" | "segments.fixed_bitset_memory" | "segments.index_writer_memory" | "segments.memory" | "segments.version_map_memory" | "shard_stats.total_count" | "suggest.current" | "suggest.time" | "suggest.total" | "uptime" | "version")[])*: A list of columns names to display. +** *`h` (Optional, Enum("id" | "repository" | "status" | "start_epoch" | "start_time" | "end_epoch" | "end_time" | "duration" | "indices" | "successful_shards" | "failed_shards" | "total_shards" | "reason") | Enum("id" | "repository" | "status" | "start_epoch" | "start_time" | "end_epoch" | "end_time" | "duration" | "indices" | "successful_shards" | "failed_shards" | "total_shards" | "reason")[])*: A list of columns names to display. It supports simple wildcards. ** *`s` (Optional, string | string[])*: List of columns that determine how the table should be sorted. Sorting defaults to ascending and can be changed by setting `:asc` @@ -3909,7 +3906,7 @@ client.cluster.health({ ... }) ** *`wait_for_nodes` (Optional, string | number)*: The request waits until the specified number N of nodes is available. It also accepts >=N, <=N, >N and yellow > red. By default, will not wait for any status. +** *`wait_for_status` (Optional, Enum("green" | "yellow" | "red" | "unknown" | "unavailable"))*: One of green, yellow or red. Will wait (until the timeout provided) until the status of the cluster changes to the one provided or better, i.e. green > yellow > red. By default, will not wait for any status. [discrete] ==== info @@ -5056,8 +5053,9 @@ If false, the sequence query will return successfully, but will always have empt ** *`max_samples_per_key` (Optional, number)*: By default, the response of a sample query contains up to `10` samples, with one sample per unique set of join keys. Use the `size` parameter to get a smaller or larger set of samples. To retrieve more than one sample per set of join keys, use the `max_samples_per_key` parameter. Pipes are not supported for sample queries. -** *`allow_no_indices` (Optional, boolean)* -** *`expand_wildcards` (Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])* +** *`allow_no_indices` (Optional, boolean)*: Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes `_all` string or when no indices have been specified) +** *`expand_wildcards` (Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])*: Whether to expand wildcard expression to concrete indices that are open, closed or both. +** *`ccs_minimize_roundtrips` (Optional, boolean)*: Indicates whether network round-trips should be minimized as part of cross-cluster search requests execution ** *`ignore_unavailable` (Optional, boolean)*: If true, missing or closed indices are not included in the response. [discrete] @@ -7424,6 +7422,12 @@ Supports wildcards (`*`). To target all data streams and indices, omit this parameter or use `*` or `_all`. ** *`active_only` (Optional, boolean)*: If `true`, the response only includes ongoing shard recoveries. ** *`detailed` (Optional, boolean)*: If `true`, the response includes detailed information about shard recoveries. +** *`allow_no_indices` (Optional, boolean)*: If `false`, the request returns an error if any wildcard expression, index alias, or `_all` value targets only missing or closed indices. +This behavior applies even if the request targets other open indices. +** *`expand_wildcards` (Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])*: Type of index that wildcard patterns can match. +If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. +Supports a list of values, such as `open,hidden`. +** *`ignore_unavailable` (Optional, boolean)*: If `false`, the request returns an error if it targets a missing or closed index. [discrete] ==== refresh @@ -8224,6 +8228,7 @@ client.inference.put({ inference_id }) ** *`inference_id` (string)*: The inference Id ** *`task_type` (Optional, Enum("sparse_embedding" | "text_embedding" | "rerank" | "completion" | "chat_completion"))*: The task type. Refer to the integration list in the API description for the available task types. ** *`inference_config` (Optional, { chunking_settings, service, service_settings, task_settings })* +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_alibabacloud @@ -8248,6 +8253,7 @@ client.inference.putAlibabacloud({ task_type, alibabacloud_inference_id, service ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { input_type, return_token })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_amazonbedrock @@ -8275,6 +8281,7 @@ client.inference.putAmazonbedrock({ task_type, amazonbedrock_inference_id, servi ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { max_new_tokens, temperature, top_k, top_p })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_amazonsagemaker @@ -8309,6 +8316,7 @@ The only valid task type for the model to perform is `completion`. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { max_tokens, temperature, top_k, top_p })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_azureaistudio @@ -8333,6 +8341,7 @@ client.inference.putAzureaistudio({ task_type, azureaistudio_inference_id, servi ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { do_sample, max_new_tokens, temperature, top_p, user })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_azureopenai @@ -8365,6 +8374,7 @@ NOTE: The `chat_completion` task type only supports streaming and only through t ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { user })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_cohere @@ -8390,6 +8400,7 @@ These settings are specific to the `cohere` service. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { input_type, return_documents, top_n, truncate })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_deepseek @@ -8411,6 +8422,7 @@ client.inference.putDeepseek({ task_type, deepseek_inference_id, service, servic ** *`service_settings` ({ api_key, model_id, url })*: Settings used to install the inference model. These settings are specific to the `deepseek` service. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_elasticsearch @@ -8449,6 +8461,7 @@ The must not match the `model_id`. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { return_documents })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_elser @@ -8485,6 +8498,7 @@ client.inference.putElser({ task_type, elser_inference_id, service, service_sett ** *`service` (Enum("elser"))*: The type of service supported for the specified task type. In this case, `elser`. ** *`service_settings` ({ adaptive_allocations, num_allocations, num_threads })*: Settings used to install the inference model. These settings are specific to the `elser` service. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_googleaistudio @@ -8507,6 +8521,7 @@ client.inference.putGoogleaistudio({ task_type, googleaistudio_inference_id, ser ** *`service` (Enum("googleaistudio"))*: The type of service supported for the specified task type. In this case, `googleaistudio`. ** *`service_settings` ({ api_key, model_id, rate_limit })*: Settings used to install the inference model. These settings are specific to the `googleaistudio` service. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_googlevertexai @@ -8531,6 +8546,7 @@ client.inference.putGooglevertexai({ task_type, googlevertexai_inference_id, ser ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { auto_truncate, top_n })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_hugging_face @@ -8590,6 +8606,7 @@ client.inference.putHuggingFace({ task_type, huggingface_inference_id, service, ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { return_documents, top_n })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_jinaai @@ -8617,6 +8634,7 @@ client.inference.putJinaai({ task_type, jinaai_inference_id, service, service_se ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { return_documents, task, top_n })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_mistral @@ -8639,6 +8657,7 @@ client.inference.putMistral({ task_type, mistral_inference_id, service, service_ ** *`service` (Enum("mistral"))*: The type of service supported for the specified task type. In this case, `mistral`. ** *`service_settings` ({ api_key, max_input_tokens, model, rate_limit })*: Settings used to install the inference model. These settings are specific to the `mistral` service. ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_openai @@ -8664,6 +8683,7 @@ NOTE: The `chat_completion` task type only supports streaming and only through t ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { user })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_voyageai @@ -8690,6 +8710,7 @@ client.inference.putVoyageai({ task_type, voyageai_inference_id, service, servic ** *`chunking_settings` (Optional, { max_chunk_size, overlap, sentence_overlap, strategy })*: The chunking configuration object. ** *`task_settings` (Optional, { input_type, return_documents, top_k, truncation })*: Settings to configure the inference task. These settings are specific to the task type you specified. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== put_watsonx @@ -8714,6 +8735,7 @@ The only valid task type for the model to perform is `text_embedding`. ** *`watsonx_inference_id` (string)*: The unique identifier of the inference endpoint. ** *`service` (Enum("watsonxai"))*: The type of service supported for the specified task type. In this case, `watsonxai`. ** *`service_settings` ({ api_key, api_version, model_id, project_id, rate_limit, url })*: Settings used to install the inference model. These settings are specific to the `watsonxai` service. +** *`timeout` (Optional, string | -1 | 0)*: Specifies the amount of time to wait for the inference endpoint to be created. [discrete] ==== rerank @@ -8786,6 +8808,7 @@ It can be a single string or an array. NOTE: Inference endpoints for the completion task type currently only support a single string as input. ** *`task_settings` (Optional, User-defined value)*: Optional task settings +** *`timeout` (Optional, string | -1 | 0)*: The amount of time to wait for the inference request to complete. [discrete] ==== text_embedding diff --git a/src/api/types.ts b/src/api/types.ts index 270373d22..ed1f41e58 100644 --- a/src/api/types.ts +++ b/src/api/types.ts @@ -163,10 +163,7 @@ export interface CountResponse { export interface CreateRequest extends RequestBase { id: Id index: IndexName - if_primary_term?: long - if_seq_no?: SequenceNumber include_source_on_error?: boolean - op_type?: OpType pipeline?: string refresh?: Refresh require_alias?: boolean @@ -528,7 +525,7 @@ export interface HealthReportImpact { export type HealthReportImpactArea = 'search' | 'ingest' | 'backup' | 'deployment_management' -export type HealthReportIndicatorHealthStatus = 'green' | 'yellow' | 'red' | 'unknown' +export type HealthReportIndicatorHealthStatus = 'green' | 'yellow' | 'red' | 'unknown' | 'unavailable' export interface HealthReportIndicatorNode { name: string | null @@ -2210,6 +2207,8 @@ export type DistanceUnit = 'in' | 'ft' | 'yd' | 'mi' | 'nmi' | 'km' | 'm' | 'cm' export interface DocStats { count: long deleted?: long + total_size_in_bytes: long + total_size?: ByteSize } export type Duration = string | -1 | 0 @@ -2293,6 +2292,7 @@ export interface FielddataStats { memory_size?: ByteSize memory_size_in_bytes: long fields?: Record + global_ordinals: GlobalOrdinalsStats } export type Fields = Field | Field[] @@ -2359,9 +2359,21 @@ export interface GetStats { total: long } +export interface GlobalOrdinalFieldStats { + build_time_in_millis: UnitMillis + build_time?: string + shard_max_value_count: long +} + +export interface GlobalOrdinalsStats { + build_time_in_millis: UnitMillis + build_time?: string + fields?: Record +} + export type GrokPattern = string -export type HealthStatus = 'green' | 'GREEN' | 'yellow' | 'YELLOW' | 'red' | 'RED' +export type HealthStatus = 'green' | 'GREEN' | 'yellow' | 'YELLOW' | 'red' | 'RED' | 'unknown' | 'unavailable' export type Host = string @@ -2479,6 +2491,9 @@ export type LifecycleOperationMode = 'RUNNING' | 'STOPPING' | 'STOPPED' export interface LinearRetriever extends RetrieverBase { retrievers?: InnerRetriever[] rank_window_size?: integer + query?: string + fields?: string[] + normalizer?: ScoreNormalizer } export type MapboxVectorTiles = ArrayBuffer @@ -2613,6 +2628,8 @@ export interface RRFRetriever extends RetrieverBase { retrievers: RetrieverContainer[] rank_constant?: integer rank_window_size?: integer + query?: string + fields?: string[] } export interface RankBase { @@ -2784,7 +2801,6 @@ export interface SegmentsStats { fixed_bit_set?: ByteSize fixed_bit_set_memory_in_bytes: long index_writer_memory?: ByteSize - index_writer_max_memory_in_bytes?: long index_writer_memory_in_bytes: long max_unsafe_auto_id_timestamp: long memory?: ByteSize @@ -2793,11 +2809,11 @@ export interface SegmentsStats { norms_memory_in_bytes: long points_memory?: ByteSize points_memory_in_bytes: long - stored_memory?: ByteSize stored_fields_memory_in_bytes: long + stored_fields_memory?: ByteSize terms_memory_in_bytes: long terms_memory?: ByteSize - term_vectory_memory?: ByteSize + term_vectors_memory?: ByteSize term_vectors_memory_in_bytes: long version_map_memory?: ByteSize version_map_memory_in_bytes: long @@ -7094,7 +7110,7 @@ export type CatCatShardColumns = CatCatShardColumn | CatCatShardColumn[] export type CatCatSnapshotsColumn = 'id' | 'snapshot' | 'repository' | 're' | 'repo' | 'status' | 's' | 'start_epoch' | 'ste' | 'startEpoch' | 'start_time' | 'sti' | 'startTime' | 'end_epoch' | 'ete' | 'endEpoch' | 'end_time' | 'eti' | 'endTime' | 'duration' | 'dur' | 'indices' | 'i' | 'successful_shards' | 'ss' | 'failed_shards' | 'fs' | 'total_shards' | 'ts' | 'reason' | 'r' | string -export type CatCatSnapshotsColumns = CatCatSnapshotsColumn | CatCatNodeColumn[] +export type CatCatSnapshotsColumns = CatCatSnapshotsColumn | CatCatSnapshotsColumn[] export type CatCatThreadPoolColumn = 'active' | 'a' | 'completed' | 'c' | 'core' | 'cr' | 'ephemeral_id' | 'eid' | 'host' | 'h' | 'ip' | 'i' | 'keep_alive' | 'k' | 'largest' | 'l' | 'max' | 'mx' | 'name' | 'node_id' | 'id' | 'node_name' | 'pid' | 'p' | 'pool_size' | 'psz' | 'port' | 'po' | 'queue' | 'q' | 'queue_size' | 'qs' | 'rejected' | 'r' | 'size' | 'sz' | 'type' | 't' | string @@ -9694,26 +9710,44 @@ export interface ClusterStatsCharFilterTypes { char_filter_types: ClusterStatsFieldTypes[] filter_types: ClusterStatsFieldTypes[] tokenizer_types: ClusterStatsFieldTypes[] + synonyms: Record } export interface ClusterStatsClusterFileSystem { - available_in_bytes: long - free_in_bytes: long - total_in_bytes: long + path?: string + mount?: string + type?: string + available_in_bytes?: long + available?: ByteSize + free_in_bytes?: long + free?: ByteSize + total_in_bytes?: long + total?: ByteSize + low_watermark_free_space?: ByteSize + low_watermark_free_space_in_bytes?: long + high_watermark_free_space?: ByteSize + high_watermark_free_space_in_bytes?: long + flood_stage_free_space?: ByteSize + flood_stage_free_space_in_bytes?: long + frozen_flood_stage_free_space?: ByteSize + frozen_flood_stage_free_space_in_bytes?: long } export interface ClusterStatsClusterIndices { - analysis: ClusterStatsCharFilterTypes + analysis?: ClusterStatsCharFilterTypes completion: CompletionStats count: long docs: DocStats fielddata: FielddataStats query_cache: QueryCacheStats + search: ClusterStatsSearchUsageStats segments: SegmentsStats shards: ClusterStatsClusterIndicesShards store: StoreStats - mappings: ClusterStatsFieldTypesMappings + mappings?: ClusterStatsFieldTypesMappings versions?: ClusterStatsIndicesVersions[] + dense_vector: ClusterStatsDenseVectorStats + sparse_vector: ClusterStatsSparseVectorStats } export interface ClusterStatsClusterIndicesShards { @@ -9736,6 +9770,7 @@ export interface ClusterStatsClusterIngest { export interface ClusterStatsClusterJvm { max_uptime_in_millis: DurationValue + max_uptime?: Duration mem: ClusterStatsClusterJvmMemory threads: long versions: ClusterStatsClusterJvmVersion[] @@ -9743,7 +9778,9 @@ export interface ClusterStatsClusterJvm { export interface ClusterStatsClusterJvmMemory { heap_max_in_bytes: long + heap_max?: ByteSize heap_used_in_bytes: long + heap_used?: ByteSize } export interface ClusterStatsClusterJvmVersion { @@ -9762,20 +9799,22 @@ export interface ClusterStatsClusterNetworkTypes { } export interface ClusterStatsClusterNodeCount { - coordinating_only: integer - data: integer - data_cold: integer - data_content: integer - data_frozen?: integer - data_hot: integer - data_warm: integer - ingest: integer - master: integer - ml: integer - remote_cluster_client: integer total: integer - transform: integer - voting_only: integer + coordinating_only?: integer + data?: integer + data_cold?: integer + data_content?: integer + data_frozen?: integer + data_hot?: integer + data_warm?: integer + index?: integer + ingest?: integer + master?: integer + ml?: integer + remote_cluster_client?: integer + search?: integer + transform?: integer + voting_only?: integer } export interface ClusterStatsClusterNodes { @@ -9846,50 +9885,62 @@ export interface ClusterStatsClusterShardMetrics { min: double } +export interface ClusterStatsClusterSnapshotStats { + current_counts: ClusterStatsSnapshotCurrentCounts + repositories: Record +} + +export interface ClusterStatsDenseVectorOffHeapStats { + total_size_bytes: long + total_size?: ByteSize + total_veb_size_bytes: long + total_veb_size?: ByteSize + total_vec_size_bytes: long + total_vec_size?: ByteSize + total_veq_size_bytes: long + total_veq_size?: ByteSize + total_vex_size_bytes: long + total_vex_size?: ByteSize + fielddata?: Record> +} + +export interface ClusterStatsDenseVectorStats { + value_count: long + off_heap?: ClusterStatsDenseVectorOffHeapStats +} + export interface ClusterStatsFieldTypes { name: Name count: integer index_count: integer - indexed_vector_count?: long - indexed_vector_dim_max?: long - indexed_vector_dim_min?: long + indexed_vector_count?: integer + indexed_vector_dim_max?: integer + indexed_vector_dim_min?: integer script_count?: integer + vector_index_type_count?: Record + vector_similarity_type_count?: Record + vector_element_type_count?: Record } export interface ClusterStatsFieldTypesMappings { field_types: ClusterStatsFieldTypes[] - runtime_field_types?: ClusterStatsRuntimeFieldTypes[] - total_field_count?: integer - total_deduplicated_field_count?: integer + runtime_field_types: ClusterStatsRuntimeFieldTypes[] + total_field_count?: long + total_deduplicated_field_count?: long total_deduplicated_mapping_size?: ByteSize total_deduplicated_mapping_size_in_bytes?: long + source_modes: Record } export interface ClusterStatsIndexingPressure { - memory: ClusterStatsIndexingPressureMemory -} - -export interface ClusterStatsIndexingPressureMemory { - current: ClusterStatsIndexingPressureMemorySummary - limit_in_bytes: long - total: ClusterStatsIndexingPressureMemorySummary -} - -export interface ClusterStatsIndexingPressureMemorySummary { - all_in_bytes: long - combined_coordinating_and_primary_in_bytes: long - coordinating_in_bytes: long - coordinating_rejections?: long - primary_in_bytes: long - primary_rejections?: long - replica_in_bytes: long - replica_rejections?: long + memory: NodesIndexingPressureMemory } export interface ClusterStatsIndicesVersions { index_count: integer primary_shard_count: integer total_primary_bytes: long + total_primary_size?: ByteSize version: VersionString } @@ -9901,18 +9952,29 @@ export interface ClusterStatsNodePackagingType { export interface ClusterStatsOperatingSystemMemoryInfo { adjusted_total_in_bytes?: long + adjusted_total?: ByteSize free_in_bytes: long + free?: ByteSize free_percent: integer total_in_bytes: long + total?: ByteSize used_in_bytes: long + used?: ByteSize used_percent: integer } +export interface ClusterStatsPerRepositoryStats { + type: string + oldest_start_time_millis: UnitMillis + oldest_start_time?: DateFormat + current_counts: ClusterStatsRepositoryStatsCurrentCounts +} + export interface ClusterStatsRemoteClusterInfo { cluster_uuid: string mode: string skip_unavailable: boolean - transport_compress: string + 'transport.compress': string status: HealthStatus version: VersionString[] nodes_count: integer @@ -9926,6 +9988,23 @@ export interface ClusterStatsRemoteClusterInfo { mem_total?: string } +export interface ClusterStatsRepositoryStatsCurrentCounts { + snapshots: integer + clones: integer + finalizations: integer + deletions: integer + snapshot_deletions: integer + active_deletions: integer + shards: ClusterStatsRepositoryStatsShards +} + +export interface ClusterStatsRepositoryStatsShards { + total: integer + complete: integer + incomplete: integer + states: Record +} + export interface ClusterStatsRequest extends RequestBase { node_id?: NodeIds include_remotes?: boolean @@ -9951,16 +10030,45 @@ export interface ClusterStatsRuntimeFieldTypes { source_total: integer } +export interface ClusterStatsSearchUsageStats { + total: long + queries: Record + rescorers: Record + sections: Record + retrievers: Record +} + +export type ClusterStatsShardState = 'INIT' | 'SUCCESS' | 'FAILED' | 'ABORTED' | 'MISSING' | 'WAITING' | 'QUEUED' | 'PAUSED_FOR_NODE_REMOVAL' + +export interface ClusterStatsSnapshotCurrentCounts { + snapshots: integer + shard_snapshots: integer + snapshot_deletions: integer + concurrent_operations: integer + cleanups: integer +} + +export interface ClusterStatsSparseVectorStats { + value_count: long +} + export interface ClusterStatsStatsResponseBase extends NodesNodesResponseBase { cluster_name: Name cluster_uuid: Uuid indices: ClusterStatsClusterIndices nodes: ClusterStatsClusterNodes - status: HealthStatus + repositories: Record> + snapshots: ClusterStatsClusterSnapshotStats + status?: HealthStatus timestamp: long ccs: ClusterStatsCCSStats } +export interface ClusterStatsSynonymsStats { + count: integer + index_count: integer +} + export interface ConnectorConnector { api_key_id?: string api_key_secret_id?: string @@ -10671,6 +10779,7 @@ export interface EqlSearchRequest extends RequestBase { index: Indices allow_no_indices?: boolean expand_wildcards?: ExpandWildcards + ccs_minimize_roundtrips?: boolean ignore_unavailable?: boolean query: string case_sensitive?: boolean @@ -12002,6 +12111,10 @@ export interface IndicesDeleteRequest extends RequestBase { export type IndicesDeleteResponse = IndicesResponseBase +export interface IndicesDeleteAliasIndicesAliasesResponseBody extends AcknowledgedResponseBase { + errors?: boolean +} + export interface IndicesDeleteAliasRequest extends RequestBase { index: Indices name: Names @@ -12009,7 +12122,7 @@ export interface IndicesDeleteAliasRequest extends RequestBase { timeout?: Duration } -export type IndicesDeleteAliasResponse = AcknowledgedResponseBase +export type IndicesDeleteAliasResponse = IndicesDeleteAliasIndicesAliasesResponseBody export interface IndicesDeleteDataLifecycleRequest extends RequestBase { name: DataStreamNames @@ -12627,6 +12740,9 @@ export interface IndicesRecoveryRequest extends RequestBase { index?: Indices active_only?: boolean detailed?: boolean + allow_no_indices?: boolean + expand_wildcards?: ExpandWildcards + ignore_unavailable?: boolean } export type IndicesRecoveryResponse = Record @@ -13792,6 +13908,7 @@ export type InferenceInferenceResponse = InferenceInferenceResult export interface InferencePutRequest extends RequestBase { task_type?: InferenceTaskType inference_id: Id + timeout?: Duration inference_config?: InferenceInferenceEndpoint } @@ -13800,6 +13917,7 @@ export type InferencePutResponse = InferenceInferenceEndpointInfo export interface InferencePutAlibabacloudRequest extends RequestBase { task_type: InferenceAlibabaCloudTaskType alibabacloud_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceAlibabaCloudServiceType service_settings: InferenceAlibabaCloudServiceSettings @@ -13811,6 +13929,7 @@ export type InferencePutAlibabacloudResponse = InferenceInferenceEndpointInfoAli export interface InferencePutAmazonbedrockRequest extends RequestBase { task_type: InferenceAmazonBedrockTaskType amazonbedrock_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceAmazonBedrockServiceType service_settings: InferenceAmazonBedrockServiceSettings @@ -13822,6 +13941,7 @@ export type InferencePutAmazonbedrockResponse = InferenceInferenceEndpointInfoAm export interface InferencePutAnthropicRequest extends RequestBase { task_type: InferenceAnthropicTaskType anthropic_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceAnthropicServiceType service_settings: InferenceAnthropicServiceSettings @@ -13833,6 +13953,7 @@ export type InferencePutAnthropicResponse = InferenceInferenceEndpointInfoAnthro export interface InferencePutAzureaistudioRequest extends RequestBase { task_type: InferenceAzureAiStudioTaskType azureaistudio_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceAzureAiStudioServiceType service_settings: InferenceAzureAiStudioServiceSettings @@ -13844,6 +13965,7 @@ export type InferencePutAzureaistudioResponse = InferenceInferenceEndpointInfoAz export interface InferencePutAzureopenaiRequest extends RequestBase { task_type: InferenceAzureOpenAITaskType azureopenai_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceAzureOpenAIServiceType service_settings: InferenceAzureOpenAIServiceSettings @@ -13855,6 +13977,7 @@ export type InferencePutAzureopenaiResponse = InferenceInferenceEndpointInfoAzur export interface InferencePutCohereRequest extends RequestBase { task_type: InferenceCohereTaskType cohere_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceCohereServiceType service_settings: InferenceCohereServiceSettings @@ -13866,6 +13989,7 @@ export type InferencePutCohereResponse = InferenceInferenceEndpointInfoCohere export interface InferencePutDeepseekRequest extends RequestBase { task_type: InferenceTaskTypeDeepSeek deepseek_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceDeepSeekServiceType service_settings: InferenceDeepSeekServiceSettings @@ -13876,6 +14000,7 @@ export type InferencePutDeepseekResponse = InferenceInferenceEndpointInfoDeepSee export interface InferencePutElasticsearchRequest extends RequestBase { task_type: InferenceElasticsearchTaskType elasticsearch_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceElasticsearchServiceType service_settings: InferenceElasticsearchServiceSettings @@ -13887,6 +14012,7 @@ export type InferencePutElasticsearchResponse = InferenceInferenceEndpointInfoEl export interface InferencePutElserRequest extends RequestBase { task_type: InferenceElserTaskType elser_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceElserServiceType service_settings: InferenceElserServiceSettings @@ -13897,6 +14023,7 @@ export type InferencePutElserResponse = InferenceInferenceEndpointInfoELSER export interface InferencePutGoogleaistudioRequest extends RequestBase { task_type: InferenceGoogleAiStudioTaskType googleaistudio_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceGoogleAiServiceType service_settings: InferenceGoogleAiStudioServiceSettings @@ -13907,6 +14034,7 @@ export type InferencePutGoogleaistudioResponse = InferenceInferenceEndpointInfoG export interface InferencePutGooglevertexaiRequest extends RequestBase { task_type: InferenceGoogleVertexAITaskType googlevertexai_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceGoogleVertexAIServiceType service_settings: InferenceGoogleVertexAIServiceSettings @@ -13918,6 +14046,7 @@ export type InferencePutGooglevertexaiResponse = InferenceInferenceEndpointInfoG export interface InferencePutHuggingFaceRequest extends RequestBase { task_type: InferenceHuggingFaceTaskType huggingface_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceHuggingFaceServiceType service_settings: InferenceHuggingFaceServiceSettings @@ -13929,6 +14058,7 @@ export type InferencePutHuggingFaceResponse = InferenceInferenceEndpointInfoHugg export interface InferencePutJinaaiRequest extends RequestBase { task_type: InferenceJinaAITaskType jinaai_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceJinaAIServiceType service_settings: InferenceJinaAIServiceSettings @@ -13940,6 +14070,7 @@ export type InferencePutJinaaiResponse = InferenceInferenceEndpointInfoJinaAi export interface InferencePutMistralRequest extends RequestBase { task_type: InferenceMistralTaskType mistral_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceMistralServiceType service_settings: InferenceMistralServiceSettings @@ -13950,6 +14081,7 @@ export type InferencePutMistralResponse = InferenceInferenceEndpointInfoMistral export interface InferencePutOpenaiRequest extends RequestBase { task_type: InferenceOpenAITaskType openai_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceOpenAIServiceType service_settings: InferenceOpenAIServiceSettings @@ -13961,6 +14093,7 @@ export type InferencePutOpenaiResponse = InferenceInferenceEndpointInfoOpenAI export interface InferencePutVoyageaiRequest extends RequestBase { task_type: InferenceVoyageAITaskType voyageai_inference_id: Id + timeout?: Duration chunking_settings?: InferenceInferenceChunkingSettings service: InferenceVoyageAIServiceType service_settings: InferenceVoyageAIServiceSettings @@ -13972,6 +14105,7 @@ export type InferencePutVoyageaiResponse = InferenceInferenceEndpointInfoVoyageA export interface InferencePutWatsonxRequest extends RequestBase { task_type: InferenceWatsonxTaskType watsonx_inference_id: Id + timeout?: Duration service: InferenceWatsonxServiceType service_settings: InferenceWatsonxServiceSettings } @@ -13999,6 +14133,7 @@ export type InferenceSparseEmbeddingResponse = InferenceSparseEmbeddingInference export interface InferenceStreamCompletionRequest extends RequestBase { inference_id: Id + timeout?: Duration input: string | string[] task_settings?: InferenceTaskSettings } @@ -17734,6 +17869,7 @@ export interface NodesJvmMemoryStats { heap_used_percent?: long heap_committed_in_bytes?: long heap_max_in_bytes?: long + heap_max?: ByteSize non_heap_used_in_bytes?: long non_heap_committed_in_bytes?: long pools?: Record @@ -17808,6 +17944,8 @@ export interface NodesPressureMemory { coordinating_rejections?: long primary_rejections?: long replica_rejections?: long + primary_document_rejections?: long + large_operation_rejections?: long } export interface NodesProcess { diff --git a/src/api/typesWithBodyKey.ts b/src/api/typesWithBodyKey.ts index 6d1430e1a..6aa9c313d 100644 --- a/src/api/typesWithBodyKey.ts +++ b/src/api/typesWithBodyKey.ts @@ -174,10 +174,7 @@ export interface CountResponse { export interface CreateRequest extends RequestBase { id: Id index: IndexName - if_primary_term?: long - if_seq_no?: SequenceNumber include_source_on_error?: boolean - op_type?: OpType pipeline?: string refresh?: Refresh require_alias?: boolean @@ -549,7 +546,7 @@ export interface HealthReportImpact { export type HealthReportImpactArea = 'search' | 'ingest' | 'backup' | 'deployment_management' -export type HealthReportIndicatorHealthStatus = 'green' | 'yellow' | 'red' | 'unknown' +export type HealthReportIndicatorHealthStatus = 'green' | 'yellow' | 'red' | 'unknown' | 'unavailable' export interface HealthReportIndicatorNode { name: string | null @@ -2287,6 +2284,8 @@ export type DistanceUnit = 'in' | 'ft' | 'yd' | 'mi' | 'nmi' | 'km' | 'm' | 'cm' export interface DocStats { count: long deleted?: long + total_size_in_bytes: long + total_size?: ByteSize } export type Duration = string | -1 | 0 @@ -2370,6 +2369,7 @@ export interface FielddataStats { memory_size?: ByteSize memory_size_in_bytes: long fields?: Record + global_ordinals: GlobalOrdinalsStats } export type Fields = Field | Field[] @@ -2436,9 +2436,21 @@ export interface GetStats { total: long } +export interface GlobalOrdinalFieldStats { + build_time_in_millis: UnitMillis + build_time?: string + shard_max_value_count: long +} + +export interface GlobalOrdinalsStats { + build_time_in_millis: UnitMillis + build_time?: string + fields?: Record +} + export type GrokPattern = string -export type HealthStatus = 'green' | 'GREEN' | 'yellow' | 'YELLOW' | 'red' | 'RED' +export type HealthStatus = 'green' | 'GREEN' | 'yellow' | 'YELLOW' | 'red' | 'RED' | 'unknown' | 'unavailable' export type Host = string @@ -2556,6 +2568,9 @@ export type LifecycleOperationMode = 'RUNNING' | 'STOPPING' | 'STOPPED' export interface LinearRetriever extends RetrieverBase { retrievers?: InnerRetriever[] rank_window_size?: integer + query?: string + fields?: string[] + normalizer?: ScoreNormalizer } export type MapboxVectorTiles = ArrayBuffer @@ -2690,6 +2705,8 @@ export interface RRFRetriever extends RetrieverBase { retrievers: RetrieverContainer[] rank_constant?: integer rank_window_size?: integer + query?: string + fields?: string[] } export interface RankBase { @@ -2861,7 +2878,6 @@ export interface SegmentsStats { fixed_bit_set?: ByteSize fixed_bit_set_memory_in_bytes: long index_writer_memory?: ByteSize - index_writer_max_memory_in_bytes?: long index_writer_memory_in_bytes: long max_unsafe_auto_id_timestamp: long memory?: ByteSize @@ -2870,11 +2886,11 @@ export interface SegmentsStats { norms_memory_in_bytes: long points_memory?: ByteSize points_memory_in_bytes: long - stored_memory?: ByteSize stored_fields_memory_in_bytes: long + stored_fields_memory?: ByteSize terms_memory_in_bytes: long terms_memory?: ByteSize - term_vectory_memory?: ByteSize + term_vectors_memory?: ByteSize term_vectors_memory_in_bytes: long version_map_memory?: ByteSize version_map_memory_in_bytes: long @@ -7175,7 +7191,7 @@ export type CatCatShardColumns = CatCatShardColumn | CatCatShardColumn[] export type CatCatSnapshotsColumn = 'id' | 'snapshot' | 'repository' | 're' | 'repo' | 'status' | 's' | 'start_epoch' | 'ste' | 'startEpoch' | 'start_time' | 'sti' | 'startTime' | 'end_epoch' | 'ete' | 'endEpoch' | 'end_time' | 'eti' | 'endTime' | 'duration' | 'dur' | 'indices' | 'i' | 'successful_shards' | 'ss' | 'failed_shards' | 'fs' | 'total_shards' | 'ts' | 'reason' | 'r' | string -export type CatCatSnapshotsColumns = CatCatSnapshotsColumn | CatCatNodeColumn[] +export type CatCatSnapshotsColumns = CatCatSnapshotsColumn | CatCatSnapshotsColumn[] export type CatCatThreadPoolColumn = 'active' | 'a' | 'completed' | 'c' | 'core' | 'cr' | 'ephemeral_id' | 'eid' | 'host' | 'h' | 'ip' | 'i' | 'keep_alive' | 'k' | 'largest' | 'l' | 'max' | 'mx' | 'name' | 'node_id' | 'id' | 'node_name' | 'pid' | 'p' | 'pool_size' | 'psz' | 'port' | 'po' | 'queue' | 'q' | 'queue_size' | 'qs' | 'rejected' | 'r' | 'size' | 'sz' | 'type' | 't' | string @@ -9799,26 +9815,44 @@ export interface ClusterStatsCharFilterTypes { char_filter_types: ClusterStatsFieldTypes[] filter_types: ClusterStatsFieldTypes[] tokenizer_types: ClusterStatsFieldTypes[] + synonyms: Record } export interface ClusterStatsClusterFileSystem { - available_in_bytes: long - free_in_bytes: long - total_in_bytes: long + path?: string + mount?: string + type?: string + available_in_bytes?: long + available?: ByteSize + free_in_bytes?: long + free?: ByteSize + total_in_bytes?: long + total?: ByteSize + low_watermark_free_space?: ByteSize + low_watermark_free_space_in_bytes?: long + high_watermark_free_space?: ByteSize + high_watermark_free_space_in_bytes?: long + flood_stage_free_space?: ByteSize + flood_stage_free_space_in_bytes?: long + frozen_flood_stage_free_space?: ByteSize + frozen_flood_stage_free_space_in_bytes?: long } export interface ClusterStatsClusterIndices { - analysis: ClusterStatsCharFilterTypes + analysis?: ClusterStatsCharFilterTypes completion: CompletionStats count: long docs: DocStats fielddata: FielddataStats query_cache: QueryCacheStats + search: ClusterStatsSearchUsageStats segments: SegmentsStats shards: ClusterStatsClusterIndicesShards store: StoreStats - mappings: ClusterStatsFieldTypesMappings + mappings?: ClusterStatsFieldTypesMappings versions?: ClusterStatsIndicesVersions[] + dense_vector: ClusterStatsDenseVectorStats + sparse_vector: ClusterStatsSparseVectorStats } export interface ClusterStatsClusterIndicesShards { @@ -9841,6 +9875,7 @@ export interface ClusterStatsClusterIngest { export interface ClusterStatsClusterJvm { max_uptime_in_millis: DurationValue + max_uptime?: Duration mem: ClusterStatsClusterJvmMemory threads: long versions: ClusterStatsClusterJvmVersion[] @@ -9848,7 +9883,9 @@ export interface ClusterStatsClusterJvm { export interface ClusterStatsClusterJvmMemory { heap_max_in_bytes: long + heap_max?: ByteSize heap_used_in_bytes: long + heap_used?: ByteSize } export interface ClusterStatsClusterJvmVersion { @@ -9867,20 +9904,22 @@ export interface ClusterStatsClusterNetworkTypes { } export interface ClusterStatsClusterNodeCount { - coordinating_only: integer - data: integer - data_cold: integer - data_content: integer - data_frozen?: integer - data_hot: integer - data_warm: integer - ingest: integer - master: integer - ml: integer - remote_cluster_client: integer total: integer - transform: integer - voting_only: integer + coordinating_only?: integer + data?: integer + data_cold?: integer + data_content?: integer + data_frozen?: integer + data_hot?: integer + data_warm?: integer + index?: integer + ingest?: integer + master?: integer + ml?: integer + remote_cluster_client?: integer + search?: integer + transform?: integer + voting_only?: integer } export interface ClusterStatsClusterNodes { @@ -9951,50 +9990,62 @@ export interface ClusterStatsClusterShardMetrics { min: double } +export interface ClusterStatsClusterSnapshotStats { + current_counts: ClusterStatsSnapshotCurrentCounts + repositories: Record +} + +export interface ClusterStatsDenseVectorOffHeapStats { + total_size_bytes: long + total_size?: ByteSize + total_veb_size_bytes: long + total_veb_size?: ByteSize + total_vec_size_bytes: long + total_vec_size?: ByteSize + total_veq_size_bytes: long + total_veq_size?: ByteSize + total_vex_size_bytes: long + total_vex_size?: ByteSize + fielddata?: Record> +} + +export interface ClusterStatsDenseVectorStats { + value_count: long + off_heap?: ClusterStatsDenseVectorOffHeapStats +} + export interface ClusterStatsFieldTypes { name: Name count: integer index_count: integer - indexed_vector_count?: long - indexed_vector_dim_max?: long - indexed_vector_dim_min?: long + indexed_vector_count?: integer + indexed_vector_dim_max?: integer + indexed_vector_dim_min?: integer script_count?: integer + vector_index_type_count?: Record + vector_similarity_type_count?: Record + vector_element_type_count?: Record } export interface ClusterStatsFieldTypesMappings { field_types: ClusterStatsFieldTypes[] - runtime_field_types?: ClusterStatsRuntimeFieldTypes[] - total_field_count?: integer - total_deduplicated_field_count?: integer + runtime_field_types: ClusterStatsRuntimeFieldTypes[] + total_field_count?: long + total_deduplicated_field_count?: long total_deduplicated_mapping_size?: ByteSize total_deduplicated_mapping_size_in_bytes?: long + source_modes: Record } export interface ClusterStatsIndexingPressure { - memory: ClusterStatsIndexingPressureMemory -} - -export interface ClusterStatsIndexingPressureMemory { - current: ClusterStatsIndexingPressureMemorySummary - limit_in_bytes: long - total: ClusterStatsIndexingPressureMemorySummary -} - -export interface ClusterStatsIndexingPressureMemorySummary { - all_in_bytes: long - combined_coordinating_and_primary_in_bytes: long - coordinating_in_bytes: long - coordinating_rejections?: long - primary_in_bytes: long - primary_rejections?: long - replica_in_bytes: long - replica_rejections?: long + memory: NodesIndexingPressureMemory } export interface ClusterStatsIndicesVersions { index_count: integer primary_shard_count: integer total_primary_bytes: long + total_primary_size?: ByteSize version: VersionString } @@ -10006,18 +10057,29 @@ export interface ClusterStatsNodePackagingType { export interface ClusterStatsOperatingSystemMemoryInfo { adjusted_total_in_bytes?: long + adjusted_total?: ByteSize free_in_bytes: long + free?: ByteSize free_percent: integer total_in_bytes: long + total?: ByteSize used_in_bytes: long + used?: ByteSize used_percent: integer } +export interface ClusterStatsPerRepositoryStats { + type: string + oldest_start_time_millis: UnitMillis + oldest_start_time?: DateFormat + current_counts: ClusterStatsRepositoryStatsCurrentCounts +} + export interface ClusterStatsRemoteClusterInfo { cluster_uuid: string mode: string skip_unavailable: boolean - transport_compress: string + 'transport.compress': string status: HealthStatus version: VersionString[] nodes_count: integer @@ -10031,6 +10093,23 @@ export interface ClusterStatsRemoteClusterInfo { mem_total?: string } +export interface ClusterStatsRepositoryStatsCurrentCounts { + snapshots: integer + clones: integer + finalizations: integer + deletions: integer + snapshot_deletions: integer + active_deletions: integer + shards: ClusterStatsRepositoryStatsShards +} + +export interface ClusterStatsRepositoryStatsShards { + total: integer + complete: integer + incomplete: integer + states: Record +} + export interface ClusterStatsRequest extends RequestBase { node_id?: NodeIds include_remotes?: boolean @@ -10056,16 +10135,45 @@ export interface ClusterStatsRuntimeFieldTypes { source_total: integer } +export interface ClusterStatsSearchUsageStats { + total: long + queries: Record + rescorers: Record + sections: Record + retrievers: Record +} + +export type ClusterStatsShardState = 'INIT' | 'SUCCESS' | 'FAILED' | 'ABORTED' | 'MISSING' | 'WAITING' | 'QUEUED' | 'PAUSED_FOR_NODE_REMOVAL' + +export interface ClusterStatsSnapshotCurrentCounts { + snapshots: integer + shard_snapshots: integer + snapshot_deletions: integer + concurrent_operations: integer + cleanups: integer +} + +export interface ClusterStatsSparseVectorStats { + value_count: long +} + export interface ClusterStatsStatsResponseBase extends NodesNodesResponseBase { cluster_name: Name cluster_uuid: Uuid indices: ClusterStatsClusterIndices nodes: ClusterStatsClusterNodes - status: HealthStatus + repositories: Record> + snapshots: ClusterStatsClusterSnapshotStats + status?: HealthStatus timestamp: long ccs: ClusterStatsCCSStats } +export interface ClusterStatsSynonymsStats { + count: integer + index_count: integer +} + export interface ConnectorConnector { api_key_id?: string api_key_secret_id?: string @@ -10839,6 +10947,7 @@ export interface EqlSearchRequest extends RequestBase { index: Indices allow_no_indices?: boolean expand_wildcards?: ExpandWildcards + ccs_minimize_roundtrips?: boolean ignore_unavailable?: boolean /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { @@ -12205,6 +12314,10 @@ export interface IndicesDeleteRequest extends RequestBase { export type IndicesDeleteResponse = IndicesResponseBase +export interface IndicesDeleteAliasIndicesAliasesResponseBody extends AcknowledgedResponseBase { + errors?: boolean +} + export interface IndicesDeleteAliasRequest extends RequestBase { index: Indices name: Names @@ -12212,7 +12325,7 @@ export interface IndicesDeleteAliasRequest extends RequestBase { timeout?: Duration } -export type IndicesDeleteAliasResponse = AcknowledgedResponseBase +export type IndicesDeleteAliasResponse = IndicesDeleteAliasIndicesAliasesResponseBody export interface IndicesDeleteDataLifecycleRequest extends RequestBase { name: DataStreamNames @@ -12851,6 +12964,9 @@ export interface IndicesRecoveryRequest extends RequestBase { index?: Indices active_only?: boolean detailed?: boolean + allow_no_indices?: boolean + expand_wildcards?: ExpandWildcards + ignore_unavailable?: boolean } export type IndicesRecoveryResponse = Record @@ -14041,6 +14157,7 @@ export type InferenceInferenceResponse = InferenceInferenceResult export interface InferencePutRequest extends RequestBase { task_type?: InferenceTaskType inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, use 'inference_config' instead. */ body?: InferenceInferenceEndpoint } @@ -14050,6 +14167,7 @@ export type InferencePutResponse = InferenceInferenceEndpointInfo export interface InferencePutAlibabacloudRequest extends RequestBase { task_type: InferenceAlibabaCloudTaskType alibabacloud_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14064,6 +14182,7 @@ export type InferencePutAlibabacloudResponse = InferenceInferenceEndpointInfoAli export interface InferencePutAmazonbedrockRequest extends RequestBase { task_type: InferenceAmazonBedrockTaskType amazonbedrock_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14078,6 +14197,7 @@ export type InferencePutAmazonbedrockResponse = InferenceInferenceEndpointInfoAm export interface InferencePutAnthropicRequest extends RequestBase { task_type: InferenceAnthropicTaskType anthropic_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14092,6 +14212,7 @@ export type InferencePutAnthropicResponse = InferenceInferenceEndpointInfoAnthro export interface InferencePutAzureaistudioRequest extends RequestBase { task_type: InferenceAzureAiStudioTaskType azureaistudio_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14106,6 +14227,7 @@ export type InferencePutAzureaistudioResponse = InferenceInferenceEndpointInfoAz export interface InferencePutAzureopenaiRequest extends RequestBase { task_type: InferenceAzureOpenAITaskType azureopenai_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14120,6 +14242,7 @@ export type InferencePutAzureopenaiResponse = InferenceInferenceEndpointInfoAzur export interface InferencePutCohereRequest extends RequestBase { task_type: InferenceCohereTaskType cohere_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14134,6 +14257,7 @@ export type InferencePutCohereResponse = InferenceInferenceEndpointInfoCohere export interface InferencePutDeepseekRequest extends RequestBase { task_type: InferenceTaskTypeDeepSeek deepseek_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14147,6 +14271,7 @@ export type InferencePutDeepseekResponse = InferenceInferenceEndpointInfoDeepSee export interface InferencePutElasticsearchRequest extends RequestBase { task_type: InferenceElasticsearchTaskType elasticsearch_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14161,6 +14286,7 @@ export type InferencePutElasticsearchResponse = InferenceInferenceEndpointInfoEl export interface InferencePutElserRequest extends RequestBase { task_type: InferenceElserTaskType elser_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14174,6 +14300,7 @@ export type InferencePutElserResponse = InferenceInferenceEndpointInfoELSER export interface InferencePutGoogleaistudioRequest extends RequestBase { task_type: InferenceGoogleAiStudioTaskType googleaistudio_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14187,6 +14314,7 @@ export type InferencePutGoogleaistudioResponse = InferenceInferenceEndpointInfoG export interface InferencePutGooglevertexaiRequest extends RequestBase { task_type: InferenceGoogleVertexAITaskType googlevertexai_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14201,6 +14329,7 @@ export type InferencePutGooglevertexaiResponse = InferenceInferenceEndpointInfoG export interface InferencePutHuggingFaceRequest extends RequestBase { task_type: InferenceHuggingFaceTaskType huggingface_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14215,6 +14344,7 @@ export type InferencePutHuggingFaceResponse = InferenceInferenceEndpointInfoHugg export interface InferencePutJinaaiRequest extends RequestBase { task_type: InferenceJinaAITaskType jinaai_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14229,6 +14359,7 @@ export type InferencePutJinaaiResponse = InferenceInferenceEndpointInfoJinaAi export interface InferencePutMistralRequest extends RequestBase { task_type: InferenceMistralTaskType mistral_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14242,6 +14373,7 @@ export type InferencePutMistralResponse = InferenceInferenceEndpointInfoMistral export interface InferencePutOpenaiRequest extends RequestBase { task_type: InferenceOpenAITaskType openai_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14256,6 +14388,7 @@ export type InferencePutOpenaiResponse = InferenceInferenceEndpointInfoOpenAI export interface InferencePutVoyageaiRequest extends RequestBase { task_type: InferenceVoyageAITaskType voyageai_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { chunking_settings?: InferenceInferenceChunkingSettings @@ -14270,6 +14403,7 @@ export type InferencePutVoyageaiResponse = InferenceInferenceEndpointInfoVoyageA export interface InferencePutWatsonxRequest extends RequestBase { task_type: InferenceWatsonxTaskType watsonx_inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { service: InferenceWatsonxServiceType @@ -14306,6 +14440,7 @@ export type InferenceSparseEmbeddingResponse = InferenceSparseEmbeddingInference export interface InferenceStreamCompletionRequest extends RequestBase { inference_id: Id + timeout?: Duration /** @deprecated The use of the 'body' key has been deprecated, move the nested keys to the top level object. */ body?: { input: string | string[] @@ -18180,6 +18315,7 @@ export interface NodesJvmMemoryStats { heap_used_percent?: long heap_committed_in_bytes?: long heap_max_in_bytes?: long + heap_max?: ByteSize non_heap_used_in_bytes?: long non_heap_committed_in_bytes?: long pools?: Record @@ -18254,6 +18390,8 @@ export interface NodesPressureMemory { coordinating_rejections?: long primary_rejections?: long replica_rejections?: long + primary_document_rejections?: long + large_operation_rejections?: long } export interface NodesProcess {