Auto-generated API code (#2625)
This commit is contained in:
@ -4047,8 +4047,8 @@ client.ilm.start({ ... })
|
|||||||
==== Arguments
|
==== Arguments
|
||||||
|
|
||||||
* *Request (object):*
|
* *Request (object):*
|
||||||
** *`master_timeout` (Optional, string | -1 | 0)*
|
** *`master_timeout` (Optional, string | -1 | 0)*: Explicit operation timeout for connection to master node
|
||||||
** *`timeout` (Optional, string | -1 | 0)*
|
** *`timeout` (Optional, string | -1 | 0)*: Explicit operation timeout
|
||||||
|
|
||||||
[discrete]
|
[discrete]
|
||||||
==== stop
|
==== stop
|
||||||
@ -4069,8 +4069,8 @@ client.ilm.stop({ ... })
|
|||||||
==== Arguments
|
==== Arguments
|
||||||
|
|
||||||
* *Request (object):*
|
* *Request (object):*
|
||||||
** *`master_timeout` (Optional, string | -1 | 0)*
|
** *`master_timeout` (Optional, string | -1 | 0)*: Explicit operation timeout for connection to master node
|
||||||
** *`timeout` (Optional, string | -1 | 0)*
|
** *`timeout` (Optional, string | -1 | 0)*: Explicit operation timeout
|
||||||
|
|
||||||
[discrete]
|
[discrete]
|
||||||
=== indices
|
=== indices
|
||||||
@ -5765,7 +5765,7 @@ client.inference.put({ inference_id })
|
|||||||
* *Request (object):*
|
* *Request (object):*
|
||||||
** *`inference_id` (string)*: The inference Id
|
** *`inference_id` (string)*: The inference Id
|
||||||
** *`task_type` (Optional, Enum("sparse_embedding" | "text_embedding" | "rerank" | "completion"))*: The task type
|
** *`task_type` (Optional, Enum("sparse_embedding" | "text_embedding" | "rerank" | "completion"))*: The task type
|
||||||
** *`inference_config` (Optional, { service, service_settings, task_settings })*
|
** *`inference_config` (Optional, { chunking_settings, service, service_settings, task_settings })*
|
||||||
|
|
||||||
[discrete]
|
[discrete]
|
||||||
==== stream_inference
|
==== stream_inference
|
||||||
|
|||||||
@ -12501,7 +12501,15 @@ export type InferenceDenseByteVector = byte[]
|
|||||||
|
|
||||||
export type InferenceDenseVector = float[]
|
export type InferenceDenseVector = float[]
|
||||||
|
|
||||||
|
export interface InferenceInferenceChunkingSettings extends InferenceInferenceEndpoint {
|
||||||
|
max_chunk_size?: integer
|
||||||
|
overlap?: integer
|
||||||
|
sentence_overlap?: integer
|
||||||
|
strategy?: string
|
||||||
|
}
|
||||||
|
|
||||||
export interface InferenceInferenceEndpoint {
|
export interface InferenceInferenceEndpoint {
|
||||||
|
chunking_settings?: InferenceInferenceChunkingSettings
|
||||||
service: string
|
service: string
|
||||||
service_settings: InferenceServiceSettings
|
service_settings: InferenceServiceSettings
|
||||||
task_settings?: InferenceTaskSettings
|
task_settings?: InferenceTaskSettings
|
||||||
|
|||||||
@ -12725,7 +12725,15 @@ export type InferenceDenseByteVector = byte[]
|
|||||||
|
|
||||||
export type InferenceDenseVector = float[]
|
export type InferenceDenseVector = float[]
|
||||||
|
|
||||||
|
export interface InferenceInferenceChunkingSettings extends InferenceInferenceEndpoint {
|
||||||
|
max_chunk_size?: integer
|
||||||
|
overlap?: integer
|
||||||
|
sentence_overlap?: integer
|
||||||
|
strategy?: string
|
||||||
|
}
|
||||||
|
|
||||||
export interface InferenceInferenceEndpoint {
|
export interface InferenceInferenceEndpoint {
|
||||||
|
chunking_settings?: InferenceInferenceChunkingSettings
|
||||||
service: string
|
service: string
|
||||||
service_settings: InferenceServiceSettings
|
service_settings: InferenceServiceSettings
|
||||||
task_settings?: InferenceTaskSettings
|
task_settings?: InferenceTaskSettings
|
||||||
|
|||||||
Reference in New Issue
Block a user