Deploying the model

You can use an OpenShift AI model server to deploy the model as an API.

Prerequisites
  • You have saved the model as described in Preparing the model for deployment.

  • You have installed KServe and enabled the model serving platform.

  • You have enabled a preinstalled or custom model-serving runtime.

  • You have obtained values for the following MinIO storage parameters:

    • Access Key

    • Secret Key

    • Endpoint

    • Region

    • Bucket

      To obtain these values, navigate to your project’s Connections tab. For the MyStorage connection, click the action menu (⋮) and then click Edit.

Procedure
  1. In the OpenShift AI dashboard, navigate to the project details page and click the Deployments tab.

  2. Click Deploy model.

    The Deploy a model wizard opens.

  3. In the Model details section, provide information about the model:

    1. For Model location, select S3 object storage.

    2. Enter the following values from your MinIO storage connection:

      • Access Key

      • Secret Key

      • Endpoint

      • Region

      • Bucket

    3. For Path, enter models/fraud.

    4. For Name, enter MyStorage.

    5. For Model type, select Predictive model.

    6. Click Next.

  4. In the Model deployment section, configure the deployment:

    1. For Model deployment name, enter fraud.

    2. For Description, enter a description of your deployment.

    3. For the hardware profile, keep the default value.

    4. For Model framework (name - version), select onnx-1.

    5. For the Serving runtime field, accept the auto-selected runtime, OpenVINO Model Server.

    6. Click Next.

  5. In the Advanced settings section, accept the defaults by clicking Next.

  6. In the Review section, click Deploy model.

Verification
  • Confirm that the deployed model is shown on the Deployments tab for the project, and on the Deployments page of the dashboard with a Started status.

    Deployed model status