Red Hat Powers Up AI Scale and Adaptability with Latest Release of Red Hat OpenShift AI

Red Hat OpenShift AI 2.15 delivers enhanced performance and streamlined workflows to fuel AI innovation at scale.

Red Hat, Inc., the world's leading provider of open source solutions, today announced the latest version of Red Hat OpenShift AI, its artificial intelligence (AI) and machine learning (ML) platform built on Red Hat OpenShift that enables enterprises to create and deliver AI-enabled applications at scale across the hybrid cloud. Red Hat OpenShift AI 2.15 is designed to provide greater flexibility, tuning and tracking capabilities, helping to accelerate enterprises’ AI/ML innovation and operational consistency with greater regularity and a stronger security posture at scale across public clouds, datacenters and edge environments.

According to IDC, enterprises included in the Forbes Global 2000 will allocate over 40% of their core IT spend on AI initiatives.1 Furthermore, IDC predicts that by 2026, enterprises will leverage generative AI (gen AI) and automation technologies to drive $1 trillion in productivity gains.2 Red Hat sees this level of investment as one that requires an AI/ML platform that can both manage model lifecycles and build gen AI applications while still being flexible enough to run alongside traditional workloads and applications across the hybrid cloud

Red Hat OpenShift AI 2.15 is aimed to help enterprises address the emerging needs of AI workloads alongside the requirements of the mission-critical, cloud-native applications that power their businesses today. The advanced features delivered with the latest version of Red Hat OpenShift AI include:

  • Model registry, currently provided as technology preview, is the central place to view and manage registered models. This feature provides a structured and organized way to share, version, deploy and track predictive and gen AI models, metadata and model artifacts. The option to provide multiple model registries is also available. Red Hat has also donated the model registry project to the Kubeflow community as a subproject.
  • Data drift detection monitors changes in input data distributions for deployed ML models. This feature allows data scientists to detect when the live data used for model interference significantly deviates from the data upon which the model was trained. Drift detection helps verify model reliability by continuously monitoring input data, keeping the model aligned with real-world data and helping to maintain the accuracy of its predictions over time.
  • Bias detection tools help data scientists and AI engineers monitor whether their models are fair and unbiased, a crucial part of establishing model trust. These tools not only help to provide insights on whether models are unbiased based on the training data but also monitor these models for fairness during real-world deployments. These tools are incorporated from the TrustyAI open source community, which provides a diverse toolkit for responsible AI development and deployment.
  • Efficient fine-tuning with LoRA uses low-rank adapters (LoRA) to enable more efficient fine-tuning of LLMs, such as Llama 3. This allows organizations to scale AI workloads while reducing costs and resource consumption. By optimizing model training and fine-tuning within cloud native environments, this solution enhances both performance and flexibility, making AI deployment more accessible and scalable.
  • Support for NVIDIA NIM, a set of easy-to-use interface microservices that accelerate the delivery of gen AI applications. The integration with NIM, part of the NVIDIA AI Enterprise software platform, will help speed up gen AI deployments while supporting a wide range of AI models to deliver scalable inference on-premises or in the cloud through application programming interfaces (APIs).
  • Support for AMD GPUs enables access to an AMD ROCm workbench image for using AMD GPUs for model development. The new functionality also enables access to images that can be used for serving and training/tuning use cases with AMD GPUs. This support gives organizations additional options for using GPUs to improve performance for computationally intensive activities.

Enhanced model serving

As a comprehensive AI/ML platform, Red Hat OpenShift AI 2.15 also adds new capabilities around gen AI model serving, including vLLM serving runtime for KServe. This new capability brings the popular open source model serving runtime for large language models (LLMs) to the platform. The flexibility and performance of vLLM is an excellent addition to the platform’s currently supported runtimes, with users also able to add their own custom options as business requirements dictate.

The latest version of Red Hat OpenShift AI also adds support for KServe Modelcars, which adds Open Container Initiative (OCI) repositories as an option for storing and accessing containerized model versions. Additionally, private/public route selection for endpoints in KServe enables organizations to enhance the security posture of a model by directing it specifically to internal endpoints when required.

Expanded AI training and experimentation options

Red Hat OpenShift AI 2.15 adds enhancements to data science pipelines and experiment tracking, empowering data scientists to more easily manage, compare and analyze pipeline runs grouped in a logical structure. The platform also adds hyperparameter tuning with Ray Tune, adding advanced optimization algorithms to improve the accuracy and train predictive and gen AI models more efficiently. The base container images for Ray clusters are now included in the newest version of Red Hat OpenShift AI, and training and tuning jobs can be scheduled across distributed workloads in the cluster to speed up jobs and maximize node utilization.

Availability

Red Hat OpenShift 2.15 will be generally available beginning mid-November 2024. For more information on additional features, improvements, bug fixes and how to upgrade to the latest version can be found here.

Supporting Quotes

Joe Fernandes, vice president and general manager, AI business unit, Red Hat

“As enterprises explore the new world of capabilities offered by AI-enabled applications and workloads, we expect interest and demand for underlying platforms to increase as concrete strategies take shape. It’s imperative for enterprises to see returns on these investments via a reliable, scalable and flexible AI platform that runs wherever their data lives across the hybrid cloud. The latest version of Red Hat OpenShift AI offers significant improvements in scalability, performance and operational efficiency while acting as a cornerstone for the overarching model lifecycle, making it possible for IT organizations to gain the benefits of a powerful AI platform while maintaining the ability to build, deploy and run on whatever environment their unique business needs dictate.”

Justin Boitano, vice president, enterprise AI software, NVIDIA

“Enterprises are seeking streamlined solutions to rapidly deploy AI applications. The integration of NVIDIA NIM with Red Hat OpenShift AI 2.15 enhances full-stack performance and scalability across hybrid cloud environments, helping development and IT teams efficiently and securely manage and accelerate their generative AI deployments.”

1

Source: IDC FutureScape Webcast: Worldwide Artificial Intelligence and Automation 2024 Predictions, (#US51901124), March 2024.

2

Source: IDC, FutureScape: Worldwide Future of Work 2024 Predictions (#US49963723), 2023

Additional Resources

Connect with Red Hat

About Red Hat, Inc.

Red Hat is the world’s leading provider of enterprise open source software solutions, using a community-powered approach to deliver reliable and high-performing Linux, hybrid cloud, container, and Kubernetes technologies. Red Hat helps customers integrate new and existing IT applications, develop cloud-native applications, standardize on our industry-leading operating system, and automate, secure, and manage complex environments. Award-winning support, training, and consulting services make Red Hat a trusted adviser to the Fortune 500. As a strategic partner to cloud providers, system integrators, application vendors, customers, and open source communities, Red Hat can help organizations prepare for the digital future.

Forward-Looking Statements

Except for the historical information and discussions contained herein, statements contained in this press release may constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are based on the company’s current assumptions regarding future business and financial performance. These statements involve a number of risks, uncertainties and other factors that could cause actual results to differ materially. Any forward-looking statement in this press release speaks only as of the date on which it is made. Except as required by law, the company assumes no obligation to update or revise any forward-looking statements.

Red Hat, Red Hat Enterprise Linux, the Red Hat logo, JBoss, Ansible, Ceph, CloudForms, Gluster and OpenShift are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the U.S. and other countries. Linux® is the registered trademark of Linus Torvalds in the U.S. and other countries. The OpenStack Word Mark is either a registered trademark/service mark or trademark/service mark of the OpenStack Foundation, in the United States and other countries, and is used with the OpenStack Foundation's permission. Red Hat is not affiliated with, endorsed or sponsored by the OpenStack Foundation, or the OpenStack community.

Contacts

Stock Quote API & Stock News API supplied by www.cloudquote.io
Quotes delayed at least 20 minutes.
By accessing this page, you agree to the following
Privacy Policy and Terms and Conditions.