Artificial Intelligence | News | Insights | AiThority
[bsfp-cryptocurrency style=”widget-18″ align=”marquee” columns=”6″ coins=”selected” coins-count=”6″ coins-selected=”BTC,ETH,XRP,LTC,EOS,ADA,XLM,NEO,LTC,EOS,XEM,DASH,USDT,BNB,QTUM,XVG,ONT,ZEC,STEEM” currency=”USD” title=”Cryptocurrency Widget” show_title=”0″ icon=”” scheme=”light” bs-show-desktop=”1″ bs-show-tablet=”1″ bs-show-phone=”1″ custom-css-class=”” custom-id=”” css=”.vc_custom_1523079266073{margin-bottom: 0px !important;padding-top: 0px !important;padding-bottom: 0px !important;}”]

Red Hat Powers Up AI Scale and Adaptability with Latest Release of Red Hat OpenShift AI

Red Hat OpenShift AI 2.15 delivers enhanced performance and streamlined workflows to fuel AI innovation at scale.

Related Posts
1 of 41,093

Red Hat, Inc., the world’s leading provider of open source solutions, today announced the latest version of Red Hat OpenShift AI, its artificial intelligence (AI) and machine learning (ML) platform built on Red Hat OpenShift that enables enterprises to create and deliver AI-enabled applications at scale across the hybrid cloud. Red Hat OpenShift AI 2.15 is designed to provide greater flexibility, tuning and tracking capabilities, helping to accelerate enterprises’ AI/ML innovation and operational consistency with greater regularity and a stronger security posture at scale across public clouds, datacenters and edge environments.

Also ReadSovereign Digital Identities and Decentralized AI: The Key to Data Control and the Future of Digitalization

“Enterprises are seeking streamlined solutions to rapidly deploy AI applications. The integration of NVIDIA NIM with Red Hat OpenShift AI 2.15 enhances full-stack performance and scalability across hybrid cloud environments, helping development and IT teams efficiently and securely manage and accelerate their generative AI deployments.”

According to IDC, enterprises included in the Forbes Global 2000 will allocate over 40% of their core IT spend on AI initiatives.1 Furthermore, IDC predicts that by 2026, enterprises will leverage generative AI (gen AI) and automation technologies to drive $1 trillion in productivity gains.2 Red Hat sees this level of investment as one that requires an AI/ML platform that can both manage model lifecycles and build gen AI applications while still being flexible enough to run alongside traditional workloads and applications across the hybrid cloud

Red Hat OpenShift AI 2.15 is aimed to help enterprises address the emerging needs of AI workloads alongside the requirements of the mission-critical, cloud-native applications that power their businesses today. The advanced features delivered with the latest version of Red Hat OpenShift AI include:

  • Model registry, currently provided as technology preview, is the central place to view and manage registered models. This feature provides a structured and organized way to share, version, deploy and track predictive and gen AI models, metadata and model artifacts. The option to provide multiple model registries is also available. Red Hat has also donated the model registry project to the Kubeflow community as a subproject.
  • Data drift detection monitors changes in input data distributions for deployed ML models. This feature allows data scientists to detect when the live data used for model interference significantly deviates from the data upon which the model was trained. Drift detection helps verify model reliability by continuously monitoring input data, keeping the model aligned with real-world data and helping to maintain the accuracy of its predictions over time.
  • Bias detection tools help data scientists and AI engineers monitor whether their models are fair and unbiased, a crucial part of establishing model trust. These tools not only help to provide insights on whether models are unbiased based on the training data but also monitor these models for fairness during real-world deployments. These tools are incorporated from the TrustyAI open source community, which provides a diverse toolkit for responsible AI development and deployment.
  • Efficient fine-tuning with LoRA uses low-rank adapters (LoRA) to enable more efficient fine-tuning of LLMs, such as Llama 3. This allows organizations to scale AI workloads while reducing costs and resource consumption. By optimizing model training and fine-tuning within cloud native environments, this solution enhances both performance and flexibility, making AI deployment more accessible and scalable.
  • Support for NVIDIA NIM, a set of easy-to-use interface microservices that accelerate the delivery of gen AI applications. The integration with NIM, part of the NVIDIA AI Enterprise software platform, will help speed up gen AI deployments while supporting a wide range of AI models to deliver scalable inference on-premises or in the cloud through application programming interfaces (APIs).
  • Support for AMD GPUs enables access to an AMD ROCm workbench image for using AMD GPUs for model development. The new functionality also enables access to images that can be used for serving and training/tuning use cases with AMD GPUs. This support gives organizations additional options for using GPUs to improve performance for computationally intensive activities.

Also ReadAiThority Interview with Tiago Azevedo, CIO at Outsystems

Enhanced model serving

As a comprehensive AI/ML platform, Red Hat OpenShift AI 2.15 also adds new capabilities around gen AI model serving, including vLLM serving runtime for KServe. This new capability brings the popular open source model serving runtime for large language models (LLMs) to the platform. The flexibility and performance of vLLM is an excellent addition to the platform’s currently supported runtimes, with users also able to add their own custom options as business requirements dictate.

The latest version of Red Hat OpenShift AI also adds support for KServe Modelcars, which adds Open Container Initiative (OCI) repositories as an option for storing and accessing containerized model versions. Additionally, private/public route selection for endpoints in KServe enables organizations to enhance the security posture of a model by directing it specifically to internal endpoints when required.

Expanded AI training and experimentation options

Red Hat OpenShift AI 2.15 adds enhancements to data science pipelines and experiment tracking, empowering data scientists to more easily manage, compare and analyze pipeline runs grouped in a logical structure. The platform also adds hyperparameter tuning with Ray Tune, adding advanced optimization algorithms to improve the accuracy and train predictive and gen AI models more efficiently. The base container images for Ray clusters are now included in the newest version of Red Hat OpenShift AI, and training and tuning jobs can be scheduled across distributed workloads in the cluster to speed up jobs and maximize node utilization.

Availability

Red Hat OpenShift 2.15 will be generally available beginning mid-November 2024. For more information on additional features, improvements, bug fixes and how to upgrade to the latest version can be found here.

Supporting Quotes

Joe Fernandes, vice president and general manager, AI business unit, Red Hat
“As enterprises explore the new world of capabilities offered by AI-enabled applications and workloads, we expect interest and demand for underlying platforms to increase as concrete strategies take shape. It’s imperative for enterprises to see returns on these investments via a reliable, scalable and flexible AI platform that runs wherever their data lives across the hybrid cloud. The latest version of Red Hat OpenShift AI offers significant improvements in scalability, performance and operational efficiency while acting as a cornerstone for the overarching model lifecycle, making it possible for IT organizations to gain the benefits of a powerful AI platform while maintaining the ability to build, deploy and run on whatever environment their unique business needs dictate.”

Justin Boitano, vice president, enterprise AI software, NVIDIA
“Enterprises are seeking streamlined solutions to rapidly deploy AI applications. The integration of NVIDIA NIM with Red Hat OpenShift AI 2.15 enhances full-stack performance and scalability across hybrid cloud environments, helping development and IT teams efficiently and securely manage and accelerate their generative AI deployments.”

[To share your insights with us as part of editorial or sponsored content, please write to psen@itechseries.com]

Comments are closed.