Intel AI inference end-to-end solution with RHOCP is based on the Intel® Data Center GPU Flex Series provisioning, Intel® OpenVINO™, and Red Hat OpenShift AI (RHOAI) on RHOCP. There are two AI inference modes verified with Intel® Xeon® processors and Intel Data Center GPU Flex Series with RHOCP.
The Red Hat certified RHOAI operator is published at Red Hat Ecosystem Catalog. You can use the command line interface (CLI) or web console to install it.
Search -> Routes -> rhods-dashboard
from the web console and access the RHOAI UI link.
Note: When installing the operator, the default kfdef
Custom Resource (CR) is created. This CR enables the dashboard for users to browse and launch Jupyter Notebooks projects on an RHOCP cluster. Please refer to this link for more details about kfdef.
The OpenVINO operator is published at Red Hat Ecosystem Catalog. You can use the CLI or web console to install it.
Follow this link to install the operator via the web console.
To enable the interactive mode, the OpenVINO notebook CR needs to be created and integrated with RHOAI.
create Notebook
option from the web console and follow these steps to create the notebook CR.Create AcceleratoProfile in the redhat-ods-applications
namespace
$ oc apply -f https://raw.githubusercontent.com/intel/intel-technology-enabling-for-openshift/main/e2e/inference/accelerator_profile.yaml
openvino-notebooks
ImageStream and add the above created AcceleratorProfile
key to the annotation field, as shown in the image below:Search -> Networking -> Routes
from the web console and access rhods-dashboard
route in the redhat-ods-applications
namespace, as in the image below. Click on the location link to launch RHOAI dashboard.Intel® Data Center GPU Flex Series 140
is shown in the accelerator dropdown menu in rhods-dashboard
. Users can run OpenVINO notebook image with Intel® Data Center GPU Flex Series 140 card.Follow the link for more details on the available Jupyter Notebooks.