Source code for mlflow.deployments

"""
Exposes functionality for deploying MLflow models to custom serving tools.

Note: model deployment to AWS Sagemaker can currently be performed via the
:py:mod:`mlflow.sagemaker` module. Model deployment to Azure can be performed by using the
`azureml library <https://pypi.org/project/azureml-mlflow/>`_.

MLflow does not currently provide built-in support for any other deployment targets, but support
for custom targets can be installed via third-party plugins. See a list of known plugins
`here <https://mlflow.org/docs/latest/plugins.html#deployment-plugins>`_.

This page largely focuses on the user-facing deployment APIs. For instructions on implementing
your own plugin for deployment to a custom serving tool, see
`plugin docs <http://mlflow.org/docs/latest/plugins.html#writing-your-own-mlflow-plugins>`_.
"""

import contextlib
import json

from mlflow.deployments.base import BaseDeploymentClient
from mlflow.deployments.databricks import DatabricksDeploymentClient, DatabricksEndpoint
from mlflow.deployments.interface import get_deploy_client, run_local
from mlflow.deployments.openai import OpenAIDeploymentClient
from mlflow.deployments.utils import get_deployments_target, set_deployments_target
from mlflow.exceptions import MlflowException
from mlflow.protos.databricks_pb2 import INVALID_PARAMETER_VALUE

with contextlib.suppress(Exception):
    # MlflowDeploymentClient depends on optional dependencies and can't be imported
    # if they are not installed.
    from mlflow.deployments.mlflow import MlflowDeploymentClient


[docs]class PredictionsResponse(dict): """ Represents the predictions and metadata returned in response to a scoring request, such as a REST API request sent to the ``/invocations`` endpoint of an MLflow Model Server. """
[docs] def get_predictions(self, predictions_format="dataframe", dtype=None): """Get the predictions returned from the MLflow Model Server in the specified format. Args: predictions_format: The format in which to return the predictions. Either ``"dataframe"`` or ``"ndarray"``. dtype: The NumPy datatype to which to coerce the predictions. Only used when the "ndarray" predictions_format is specified. Raises: Exception: If the predictions cannot be represented in the specified format. Returns: The predictions, represented in the specified format. """ import numpy as np import pandas as pd from pandas.core.dtypes.common import is_list_like if predictions_format == "dataframe": predictions = self["predictions"] if isinstance(predictions, str): return pd.DataFrame(data=[predictions]) if isinstance(predictions, dict) and not any( is_list_like(p) and getattr(p, "ndim", 1) == 1 for p in predictions.values() ): return pd.DataFrame(data=predictions, index=[0]) return pd.DataFrame(data=predictions) elif predictions_format == "ndarray": return np.array(self["predictions"], dtype) else: raise MlflowException( f"Unrecognized predictions format: '{predictions_format}'", INVALID_PARAMETER_VALUE, )
[docs] def to_json(self, path=None): """Get the JSON representation of the MLflow Predictions Response. Args: path: If specified, the JSON representation is written to this file path. Returns: If ``path`` is unspecified, the JSON representation of the MLflow Predictions Response. Else, None. """ if path is not None: with open(path, "w") as f: json.dump(dict(self), f) else: return json.dumps(dict(self))
@classmethod def from_json(cls, json_str): try: parsed_response = json.loads(json_str) except Exception as e: raise MlflowException("Predictions response contents are not valid JSON") from e if not isinstance(parsed_response, dict) or "predictions" not in parsed_response: raise MlflowException( f"Invalid response. Predictions response contents must be a dictionary" f" containing a 'predictions' field. Instead, received: {parsed_response}" ) return PredictionsResponse(parsed_response)
__all__ = [ "get_deploy_client", "run_local", "BaseDeploymentClient", "DatabricksDeploymentClient", "OpenAIDeploymentClient", "DatabricksEndpoint", "MlflowDeploymentClient", "PredictionsResponse", "get_deployments_target", "set_deployments_target", ]