Red Hat Previews OpenShift Platform for AI Models
At the Ray Summit this week, Red Hat previewed a platform for running artificial intelligence (AI) workloads based on the Red Hat OpenShift platform the company built atop Kubernetes.
Taneem Ibrahim, manager of software engineering at Red Hat, said Red Hat OpenShift AI will provide IT teams with a consistent distributed computing framework based on Ray clusters, which are designed to run workloads in parallel, combined with CodeFlare, a framework for training AI models.
That approach will enable IT teams to invoke AI models more easily in real-time or process workloads in batch mode using scheduling capabilities that enable them to optimize usage of IT infrastructure, such as graphical processor units (GPUs), as required, said Ibrahim.
The Red Hat OpenShift AI platform provides access to KubeRay, a Kubernetes operator for deploying and managing Ray clusters, and CodeFlare, a Kubernetes operator that deploys and manages the life cycle of the three components of the framework used to manage jobs and dynamically aggregate and scale resources as needed.
In addition, there is KKerve, a Kubernetes Custom Resource Definition for deploying and serving models, a Text Generation Inference Server (TGIS) for loading inference engines, and Caikit, an AI toolkit/runtime that handles the life cycle of the TGIS processes.
There is also OpenShift Serverless, a framework for driving event-driven applications based on the open source Knative project and OpenShift Service Mesh, an instance of the open source Istio service mesh.
Kubernetes platforms such as Red Hat OpenShift are generally preferred when running greenfield AI applications where technical debt is much less of an issue, noted Ibrahim. However, as more organizations begin to embed AI workloads within applications, there will be a more pressing need to integrate DevOps and machine learning operations (MLOps) workflows. Most MLOps workflows today are managed by data scientists, but when it comes to deploying them, an AI model is treated as just another type of software artifact invoked via an application programming interface (API). The Red Hat OpenShift AI platform will make it simpler for those teams to collaborate both before and after AI models are deployed and continuously updated, noted Ibrahim.
It’s still early days as far as operationalizing AI models is concerned, but it’s now only a matter of time before organizations will need to find some way to meld DevOps and MLOps workflows to enable the level of observability that will be required to succeed, said Ibrahim. To rise to that challenge, organizations will need to address a range of technical and cultural issues that inevitably arise when individuals with diverse backgrounds need to collaborate. While data scientists tend to have a lot of experience building AI models, their knowledge of deploying them in production environments is often much less extensive.
In the meantime, IT teams would be well-advised to become a lot more familiar with the nuances of building and deploying AI models, as they will soon be pervasively embedded within every application imaginable.