(Optional) Creating connections to your own S3-compatible object storage

Skip this procedure if you completed the steps in Running a script to install local object storage buckets and create connections.

If you have existing S3-compatible storage buckets that you want to use for this workshop, you must create a connection to one storage bucket for saving your data and models. If you want to complete the pipelines section of this workshop, create another connection to a different storage bucket for saving pipeline artifacts.

Prerequisites

To create connections to your existing S3-compatible storage buckets, you need the following credential information for the storage buckets:

  • Endpoint URL

  • Access key

  • Secret key

  • Region

  • Bucket name

If you do not have this information, contact your storage administrator.

Procedure
  1. Create a connection for saving your data and models:

    1. In the OpenShift AI dashboard, navigate to the page for your project.

    2. Click the Connections tab, and then click Create connection.

      Add connection
    3. In the Add connection modal, for the Connection type select S3 compatible object storage - v1.

    4. Complete the Add connection form and name your connection My Storage. This connection is for saving your personal work, including data and models.

      Add my storage form
    5. Click Create.

  2. Create a connection for saving pipeline artifacts:

    If you do not intend to complete the pipelines section of the workshop, you can skip this step.
    1. Click Add connection.

    2. Complete the form and name your connection Pipeline Artifacts.

      Add pipeline artifacts form
    3. Click Create.

Verification

In the Connections tab for the project, check to see that your connections are listed.

List of project connections

If your cluster uses self-signed certificates, your OpenShift AI administrator might need to provide a certificate authority (CA) to securely connect to the S3 object storage, as described in Accessing S3-compatible object storage with self-signed certificates (Self-Managed).

Next steps
  • Decide whether you want to complete the pipelines section of this workshop. With OpenShift AI pipelines, you can automate the execution of your notebooks and Python code. You can run long training jobs or retrain your models on a schedule without having to manually run them in a notebook.

    If you want to complete the pipelines section of this workshop, go to Enabling AI pipelines.

  • Decide whether you want to complete the Distributing training jobs with the Training Operator section of this workshop. In that section, you implement a distributed training job by using Kueue for managing job resources.

    If you want to complete the Distributing training jobs with the Training Operator section of this workshop, go to Setting up Kueue resources.

  • Otherwise, skip to Creating a workbench.