Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Auto-generated code for 8.11 #2376

Merged
merged 1 commit into from
Nov 21, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
4 changes: 4 additions & 0 deletions elasticsearch/_async/client/ml.py
Original file line number Diff line number Diff line change
Expand Up @@ -3582,6 +3582,7 @@ async def start_trained_model_deployment(
*,
model_id: str,
cache_size: t.Optional[t.Union[int, str]] = None,
deployment_id: t.Optional[str] = None,
error_trace: t.Optional[bool] = None,
filter_path: t.Optional[t.Union[str, t.Sequence[str]]] = None,
human: t.Optional[bool] = None,
Expand All @@ -3605,6 +3606,7 @@ async def start_trained_model_deployment(
:param cache_size: The inference cache size (in memory outside the JVM heap)
per node for the model. The default value is the same size as the `model_size_bytes`.
To disable the cache, `0b` can be provided.
:param deployment_id: A unique identifier for the deployment of the model.
:param number_of_allocations: The number of model allocations on each node where
the model is deployed. All allocations on a node share the same copy of the
model in memory but use a separate set of threads to evaluate the model.
Expand All @@ -3631,6 +3633,8 @@ async def start_trained_model_deployment(
__query: t.Dict[str, t.Any] = {}
if cache_size is not None:
__query["cache_size"] = cache_size
if deployment_id is not None:
__query["deployment_id"] = deployment_id
if error_trace is not None:
__query["error_trace"] = error_trace
if filter_path is not None:
Expand Down
4 changes: 4 additions & 0 deletions elasticsearch/_sync/client/ml.py
Original file line number Diff line number Diff line change
Expand Up @@ -3582,6 +3582,7 @@ def start_trained_model_deployment(
*,
model_id: str,
cache_size: t.Optional[t.Union[int, str]] = None,
deployment_id: t.Optional[str] = None,
error_trace: t.Optional[bool] = None,
filter_path: t.Optional[t.Union[str, t.Sequence[str]]] = None,
human: t.Optional[bool] = None,
Expand All @@ -3605,6 +3606,7 @@ def start_trained_model_deployment(
:param cache_size: The inference cache size (in memory outside the JVM heap)
per node for the model. The default value is the same size as the `model_size_bytes`.
To disable the cache, `0b` can be provided.
:param deployment_id: A unique identifier for the deployment of the model.
:param number_of_allocations: The number of model allocations on each node where
the model is deployed. All allocations on a node share the same copy of the
model in memory but use a separate set of threads to evaluate the model.
Expand All @@ -3631,6 +3633,8 @@ def start_trained_model_deployment(
__query: t.Dict[str, t.Any] = {}
if cache_size is not None:
__query["cache_size"] = cache_size
if deployment_id is not None:
__query["deployment_id"] = deployment_id
if error_trace is not None:
__query["error_trace"] = error_trace
if filter_path is not None:
Expand Down