Deployments¶
Deployments ¶
Interact with deployments in the Driverless AI server.
deploy_to_triton_in_local ¶
deploy_to_triton_in_local(
experiment: Experiment,
deploy_predictions: bool = True,
deploy_shapley: bool = False,
deploy_original_shapley: bool = False,
enable_high_concurrency: bool = False,
) -> TritonDeployment
Deploys the model created from an experiment to the local Triton server in the Driverless AI server.
Parameters:
-
experiment
(Experiment
) –Experiment model.
-
deploy_predictions
(bool
, default:True
) –Whether to deploy model predictions or not.
-
deploy_shapley
(bool
, default:False
) –Whether to deploy model Shapley or not.
-
deploy_original_shapley
(bool
, default:False
) –Whether to deploy model original Shapley or not.
-
enable_high_concurrency
(bool
, default:False
) –Whether to enable handling several requests at once
Returns:
-
TritonDeployment
–Deployed Triton deployment.
Beta API
A beta API that is subject to future changes.
deploy_to_triton_in_remote ¶
deploy_to_triton_in_remote(
experiment: Experiment,
deploy_predictions: bool = True,
deploy_shapley: bool = False,
deploy_original_shapley: bool = False,
enable_high_concurrency: bool = False,
) -> TritonDeployment
Deploys the model created from an experiment to a remote Triton server configured in the Driverless AI server.
Parameters:
-
experiment
(Experiment
) –Experiment model.
-
deploy_predictions
(bool
, default:True
) –Whether to deploy model predictions or not.
-
deploy_shapley
(bool
, default:False
) –Whether to deploy model Shapley or not.
-
deploy_original_shapley
(bool
, default:False
) –Whether to deploy model original Shapley or not.
-
enable_high_concurrency
(bool
, default:False
) –Whether to enable handling several requests at once
Returns:
-
TritonDeployment
–Deployed Triton deployment.
Beta API
A beta API that is subject to future changes.
get_from_triton_in_local ¶
get_from_triton_in_local(key: str) -> TritonDeployment
Retrieves a Triton deployment, deployed in the local Triton server configured in the Driverless AI server.
Parameters:
-
key
(str
) –The unique ID of the Triton deployment.
Returns:
-
TritonDeployment
–The Triton deployment corresponding to the key.
get_from_triton_in_remote ¶
get_from_triton_in_remote(key: str) -> TritonDeployment
Retrieves a Triton deployment, deployed in a remote Triton server configured in the Driverless AI server.
Parameters:
-
key
(str
) –The unique ID of the Triton deployment.
Returns:
-
TritonDeployment
–The Triton deployment corresponding to the key.
gui ¶
gui() -> Hyperlink
Returns the full URL to the Deployments page in the Driverless AI server.
Returns:
-
Hyperlink
–The full URL to the Deployments page.
list_triton_deployments ¶
list_triton_deployments(
start_index: int = 0, count: int = None
) -> Sequence[TritonDeployment]
Retrieves Triton deployments in the Driverless AI server.
Parameters:
Returns:
-
Sequence[TritonDeployment]
–Triton deployments.
Beta API
A beta API that is subject to future changes.
TritonDeployment ¶
A deployment in an NVIDIA Triton inference server in the Driverless AI server.
is_local_deployment
property
¶
is_local_deployment: bool
Whether the Triton deployment is in the built-in (local) Triton server in the Driverless AI server or in a remote Triton server.
Returns:
-
bool
–
triton_model
property
¶
triton_model: TritonModel
Triton model created by the Triton deployment.
Beta API
A beta API that is subject to future changes.
Returns:
triton_server_hostname
property
¶
triton_server_hostname: str
Hostname of the Triton server in which the Triton deployment occurred.
Returns:
-
str
–
delete ¶
delete() -> None
Permanently deletes the Triton deployment from the Driverless AI server.
Beta API
A beta API that is subject to future changes.
TritonModel
dataclass
¶
A Triton model created by a Triton deployment.