diff --git a/docs/reference.asciidoc b/docs/reference.asciidoc index cea62abd9..9e5291cbb 100644 --- a/docs/reference.asciidoc +++ b/docs/reference.asciidoc @@ -6491,6 +6491,7 @@ client.ml.startTrainedModelDeployment({ model_id }) ** *`cache_size` (Optional, number | string)*: The inference cache size (in memory outside the JVM heap) per node for the model. The default value is the same size as the `model_size_bytes`. To disable the cache, `0b` can be provided. +** *`deployment_id` (Optional, string)*: A unique identifier for the deployment of the model. ** *`number_of_allocations` (Optional, number)*: The number of model allocations on each node where the model is deployed. All allocations on a node share the same copy of the model in memory but use a separate set of threads to evaluate the model. diff --git a/src/api/types.ts b/src/api/types.ts index 0dc17efc8..707e1cc13 100644 --- a/src/api/types.ts +++ b/src/api/types.ts @@ -14106,6 +14106,7 @@ export interface MlStartDatafeedResponse { export interface MlStartTrainedModelDeploymentRequest extends RequestBase { model_id: Id cache_size?: ByteSize + deployment_id?: string number_of_allocations?: integer priority?: MlTrainingPriority queue_capacity?: integer diff --git a/src/api/typesWithBodyKey.ts b/src/api/typesWithBodyKey.ts index 5a4157ac8..dbbe3ea1a 100644 --- a/src/api/typesWithBodyKey.ts +++ b/src/api/typesWithBodyKey.ts @@ -14375,6 +14375,7 @@ export interface MlStartDatafeedResponse { export interface MlStartTrainedModelDeploymentRequest extends RequestBase { model_id: Id cache_size?: ByteSize + deployment_id?: string number_of_allocations?: integer priority?: MlTrainingPriority queue_capacity?: integer