Auto-generated API code (#2716)

This commit is contained in:
Elastic Machine
2025-04-07 21:43:20 +02:00
committed by GitHub
parent 34bb7f5916
commit 1ee6a34e73
4 changed files with 24 additions and 180 deletions

View File

@ -8038,25 +8038,6 @@ client.inference.get({ ... })
** *`task_type` (Optional, Enum("sparse_embedding" | "text_embedding" | "rerank" | "completion" | "chat_completion"))*: The task type
** *`inference_id` (Optional, string)*: The inference Id
[discrete]
==== post_eis_chat_completion
Perform a chat completion task through the Elastic Inference Service (EIS).
Perform a chat completion inference task with the `elastic` service.
{ref}/post-inference-api.html[Endpoint documentation]
[source,ts]
----
client.inference.postEisChatCompletion({ eis_inference_id })
----
[discrete]
==== Arguments
* *Request (object):*
** *`eis_inference_id` (string)*: The unique identifier of the inference endpoint.
** *`chat_completion_request` (Optional, { messages, model, max_completion_tokens, stop, temperature, tool_choice, tools, top_p })*
[discrete]
==== put
Create an inference endpoint.
@ -8277,27 +8258,6 @@ These settings are specific to the `cohere` service.
** *`task_settings` (Optional, { input_type, return_documents, top_n, truncate })*: Settings to configure the inference task.
These settings are specific to the task type you specified.
[discrete]
==== put_eis
Create an Elastic Inference Service (EIS) inference endpoint.
Create an inference endpoint to perform an inference task through the Elastic Inference Service (EIS).
[source,ts]
----
client.inference.putEis({ task_type, eis_inference_id, service, service_settings })
----
[discrete]
==== Arguments
* *Request (object):*
** *`task_type` (Enum("chat_completion"))*: The type of the inference task that the model will perform.
NOTE: The `chat_completion` task type only supports streaming and only through the _stream API.
** *`eis_inference_id` (string)*: The unique identifier of the inference endpoint.
** *`service` (Enum("elastic"))*: The type of service supported for the specified task type. In this case, `elastic`.
** *`service_settings` ({ model_id, rate_limit })*: Settings used to install the inference model. These settings are specific to the `elastic` service.
[discrete]
==== put_elasticsearch
Create an Elasticsearch inference endpoint.