
Red Hat Optimizes Red Hat AI to Speed Enterprise AI Deployments Across Models, AI Accelerators and Clouds - Middle East Business News and Information
Red Hat AI Inference Server, validated models and integration of Llama Stack and Model Context Protocol help users deliver higher-performing, more consistent AI applications and agents
Red Hat, the world's leading provider of open source solutions, today continues to deliver customer choice in enterprise AI with the introduction of Red Hat AI Inference Server, Red Hat AI third-party validated models and the integration of Llama Stack and Model Context Protocol (MCP) APIs, along with significant updates across the Red Hat AI portfolio. With these developments, Red Hat intends to further advance the capabilities organizations need to accelerate AI adoption while providing greater customer choice and confidence in generative AI (gen AI) production deployments across the hybrid cloud.
According to Forrester, open source software will be the spark for accelerating enterprise AI efforts.1 As the AI landscape grows more complex and dynamic, Red Hat AI Inference Server and third party validated models provide efficient model inference and a tested collection of AI models optimized for performance on the Red Hat AI platform. Coupled with the integration of new APIs for gen AI agent development, including Llama Stack and MCP, Red Hat is working to tackle deployment complexity, empowering IT leaders, data scientists and developers to accelerate AI initiatives with greater control and efficiency.
Efficient inference across the hybrid cloud with Red Hat AI Inference Server:
The Red Hat AI portfolio now includes the new Red Hat AI Inference Server, providing faster, more consistent and cost-effective inference at scale across hybrid cloud environments. This key addition is integrated into the latest releases of Red Hat OpenShift AI and Red Hat Enterprise Linux AI, and is also available as a standalone offering, enabling organizations to deploy intelligent applications with greater efficiency, flexibility and performance.
Tested and optimized models with Red Hat AI third party validated models
Red Hat AI third party validated models, available on Hugging Face, make it easier for enterprises to find the right models for their specific needs. Red Hat AI offers a collection of validated models, as well as deployment guidance to enhance customer confidence in model performance and outcome reproducibility. Select models are also optimized by Red Hat, leveraging model compression techniques to reduce size and increase inference speed, helping to minimize resource consumption and operating costs. Additionally, the ongoing model validation process helps Red Hat AI customers continue to stay at the forefront of optimized gen AI innovation.
Standardized APIs for AI application and agent development with Llama Stack and MCP
Red Hat AI is integrating Llama Stack, initially developed by Meta, along with Anthropic's MCP, to provide users with standardized APIs for building and deploying AI applications and agents. Currently available in developer preview in Red Hat AI, Llama Stack provides a unified API to access inference with vLLM, retrieval-augmented generation (RAG), model evaluation, guardrails and agents, across any gen AI model. MCP enables models to integrate with external tools by providing a standardized interface for connecting APIs, plugins and data sources in agent workflows.
The latest release of Red Hat OpenShift AI (v2.20) delivers additional enhancements for building, training, deploying and monitoring both gen AI and predictive AI models at scale. These include: Optimized model catalog (technology preview) provides easy access to validated Red Hat and third party models, enables the deployment of these models on Red Hat OpenShift AI clusters through the web console interface and manages the lifecycle of those models leveraging Red Hat OpenShift AI's integrated registry.
Distributed training through the KubeFlow Training Operator enables the scheduling and execution of InstructLab model tuning and other PyTorch-based training and tuning workloads, distributed across multiple Red Hat OpenShift nodes and GPUs and includes distributed RDMA networking–acceleration and optimized GPU utilization to reduce costs.
Feature store (technology preview), based on the upstream Kubeflow Feast project, provides a centralized repository for managing and serving data for both model training and inference, streamlining data workflows to improve model accuracy and reusability.
Red Hat Enterprise Linux AI 1.5 brings new updates to Red Hat's foundation model platform for developing, testing and running large language models (LLMs). Key features in version 1.5 include: Google Cloud Marketplace availability, expanding the customer choice for running Red Hat Enterprise Linux AI in public cloud environments–along with AWS and Azure–to help simplify the deployment and management of AI workloads on Google Cloud.
Enhanced multi-language capabilities for Spanish, German, French and Italian via InstructLab, allowing for model customization using native scripts and unlocking new possibilities for multilingual AI applications. Users can also bring their own teacher models for greater control over model customization and testing for specific use cases and languages, with future support planned for Japanese, Hindi and Korean.
The Red Hat AI InstructLab on IBM Cloud service is also now generally available. This new cloud service further streamlines the model customization process, improving scalability and user experience, empowering enterprises to make use of their unique data with greater ease and control.
Red Hat's vision: Any model, any accelerator, any cloud.
The future of AI must be defined by limitless opportunity, not constrained by infrastructure silos. Red Hat sees a horizon where organizations can deploy any model, on any accelerator, across any cloud, delivering an exceptional, more consistent user experience without exorbitant costs. To unlock the true potential of gen AI investments, enterprises require a universal inference platform–a standard for more seamless, high-performance AI innovation, both today and in the years to come.
Red Hat Summit:
Join the Red Hat Summit keynotes to hear the latest from Red Hat executives, customers and partners: Modernized infrastructure meets enterprise-ready AI — Tuesday, May 20, 8-10 a.m. EDT (YouTube)
Hybrid cloud evolves to deliver enterprise innovation — Wednesday, May 21, 8-9:30 a.m. EDT (YouTube)
Supporting Quotes:
Joe Fernandes, vice president and general manager, AI Business Unit, Red Hat
'Faster, more efficient inference is emerging as the newest decision point for gen AI innovation. Red Hat AI, with enhanced inference capabilities through Red Hat AI Inference Server and a new collection of validated third-party models, helps equip organizations to deploy intelligent applications where they need to, how they need to and with the components that best meet their unique needs.'
Michele Rosen, research manager, IDC
'Organizations are moving beyond initial AI explorations and are focused on practical deployments. The key to their continued success lies in the ability to be adaptable with their AI strategies to fit various environments and needs. The future of AI not only demands powerful models, but models that can be deployed with ability and cost-effectiveness. Enterprises seeking to scale their AI initiatives and deliver business value will find this flexibility absolutely essential.'
About Red Hat:
Red Hat is the open hybrid cloud technology leader, delivering a trusted, consistent and comprehensive foundation for transformative IT innovation and AI applications. Its portfolio of cloud, developer, AI, Linux, automation and application platform technologies enables any application, anywhere—from the datacenter to the edge. As the world's leading provider of enterprise open source software solutions, Red Hat invests in open ecosystems and communities to solve tomorrow's IT challenges. Collaborating with partners and customers, Red Hat helps them build, connect, automate, secure and manage their IT environments, supported by consulting services and award-winning training and certification offerings.
Forward-Looking Statements:
Except for the historical information and discussions contained herein, statements contained in this press release may constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are based on the company's current assumptions regarding future business and financial performance. These statements involve a number of risks, uncertainties and other factors that could cause actual results to differ materially. Any forward-looking statement in this press release speaks only as of the date on which it is made. Except as required by law, the company assumes no obligation to update or revise any forward-looking statements.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Mid East Info
5 hours ago
- Mid East Info
Red Hat Powers Modern Virtualization on Microsoft Azure - Middle East Business News and Information
Red Hat OpenShift Virtualization on Microsoft Azure Red Hat OpenShift helps organizations more easily migrate to a modern infrastructure Red Hat, the world's leading provider of open source solutions, announced the public preview of Red Hat OpenShift Virtualization on Microsoft Azure Red Hat OpenShift. Available as a self-managed operator included in Azure Red Hat OpenShift, Red Hat OpenShift Virtualization offers organizations an accelerated path to modernization by streamlining the migration of virtual machines (VMs) from existing virtualization platforms to a scalable, cloud-native platform. Azure Red Hat OpenShift is a turnkey application platform that is jointly managed and supported by Red Hat and Microsoft, designed to help to reduce the complexities associated with managing the underlying infrastructure and empower IT teams to focus time and resources on innovation and modernization rather than routine maintenance. With Red Hat OpenShift Virtualization on Azure Red Hat OpenShift, organizations benefit from a more consistent hybrid cloud stack that can support VMs and containers alike to help significantly streamline application modernization and accelerate cloud-native strategies. This enables organizations to more effectively modernize existing critical VM infrastructure while continuing to evolve with new innovations that meet future business needs. Red Hat OpenShift Virtualization on Azure Red Hat OpenShift empowers organizations to: Accelerate VM migration: Red Hat OpenShift Virtualization on Azure Red Hat OpenShift helps organizations quickly migrate and scale existing VM workloads with built-in migration tooling and automation capabilities, such as those through Red Hat Ansible Automation Platform and Red Hat Advanced Cluster Management, to simplify the migration process, minimize disruption and enable teams to quickly shift to modern infrastructure. Simplify operations: Red Hat OpenShift Virtualization on Azure Red Hat OpenShift provides a unified view of operations to more seamlessly manage both VMs and containers on the same platform across the hybrid cloud. Automated deployment and management of Red Hat OpenShift clusters further reduces complexity and risk. Modernize infrastructure: Red Hat OpenShift Virtualization on Azure Red Hat OpenShift allows customers to build, modernize and deploy applications at scale. By adopting this Kubernetes-based platform, organizations instantly land on modern infrastructure that brings them two steps closer to their cloud-native application modernization goals. Azure Red Hat OpenShift brings modern application development processes and tools to VMs that help expedite the modernization of VM-based applications. Optimize resources: In addition to increasing DevOps productivity and decreasing the time to deploy applications with Azure Red Hat Openshift, further optimization can be realized with Red Hat OpenShift Virtualization by right-sizing VMs to better match workload needs. Built on the industry's leading hybrid cloud application platform powered by Kubernetes and Microsoft Azure's trusted cloud infrastructure, Azure Red Hat OpenShift delivers a future-ready platform with integrated security tooling, automation and management capabilities to extend innovation across the hybrid cloud. Red Hat OpenShift Virtualization is now available in public preview as a self-managed operator on Azure Red Hat OpenShift. Organizations can also apply their Microsoft Azure Consumption Commitment (MACC) and utilize the Azure Migration and Modernization Program (AMMP) for Azure Red Hat OpenShift. Additionally, customers can use the Azure Hybrid Benefit to reuse existing on-premise licenses for both Red Hat Enterprise Linux and Windows Licenses. Supporting Quotes: Chris Wright, senior vice president and chief technology officer, Red Hat 'As organizations continue to modernize and move away from legacy virtualization solutions, it is critical to choose a secure computing foundation for the future that can adapt to their current and evolving multi-infrastructure environments. Building upon our extensive history of collaboration and joint engineering efforts with Microsoft Azure, Red Hat OpenShift Virtualization running on Azure Red Hat OpenShift delivers more consistent orchestration for VMs and containers alike, setting organizations on a clear path to modern application development and deployment.' Brendan Burns, corporate vice president, Azure Compute, Microsoft: 'As customers modernize and move their apps from traditional virtual-machine-based fabrics that are on-premises to modern Kubernetes platforms, some components still need to run on traditional virtual machines for a while. To address this, Microsoft and Red Hat are collaborating to bring open-source innovation from the KubeVirt project into Azure Red Hat OpenShift. What I'm most excited about is how this enables customers to add virtualization capabilities to Azure Red Hat OpenShift, which allows them to modernize at their own pace, and get the best return on investment as they transition to the cloud.' About Red Hat: Red Hat is the open hybrid cloud technology leader, delivering a trusted, consistent, and comprehensive foundation for transformative IT innovation and AI applications. Its portfolio of cloud, developer, AI, Linux, automation, and application platform technologies enables any application, anywhere—from the datacenter to the edge. As the world's leading provider of enterprise open source software solutions, Red Hat invests in open ecosystems and communities to solve tomorrow's IT challenges. Collaborating with partners and customers, Red Hat helps them build, connect, automate, secure, and manage their IT environments, supported by consulting services and award-winning training and certification offerings. Forward-Looking Statements: Except for the historical information and discussions contained herein, statements contained in this press release may constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are based on the company's current assumptions regarding future business and financial performance. These statements involve a number of risks, uncertainties and other factors that could cause actual results to differ materially. Any forward-looking statement in this press release speaks only as of the date on which it is made. Except as required by law, the company assumes no obligation to update or revise any forward-looking statements.


Egypt Independent
3 days ago
- Egypt Independent
Why this leading AI CEO is warning the tech could cause mass unemployment
New York CNN — The chief executive of one of the world's leading artificial intelligence labs is warning that the technology could cause a dramatic spike in unemployment in the very near future. He says policymakers and corporate leaders aren't ready for it. 'AI is starting to get better than humans at almost all intellectual tasks, and we're going to collectively, as a society, grapple with it,' Anthropic CEO Dario Amodei told CNN's Anderson Cooper in an interview on Thursday. 'AI is going to get better at what everyone does, including what I do, including what other CEOs do.' Amodei believes the AI tools that Anthropic and other companies are racing to build could eliminate half of entry-level, white-collar jobs and spike unemployment to as much as 20 percent in the next one to five years, he told Axios on Wednesday. That could mean the US unemployment rate growing fivefold in just a few years; the last time it neared that rate was briefly at the height of the Covid-19 pandemic. It's not the first dire warning about how rapidly advancing AI could upend the economy in the coming years. Academics and economists have also cautioned that AI could replace some jobs or tasks in the coming years, with varying degrees of seriousness. Earlier this year, a World Economic Forum survey showed that 41 percent of employers plan to downsize their workforce because of AI automation by 2030. But Amodei's prediction is notable because it's coming from one of the industry's top leaders and because of the scale of disruption it foretells. It also comes as Anthropic is now selling AI technology on the promise that it can work nearly the length of a typical human workday. The historical narrative about how tech advancement works is that technology would automate lower-paying, lower-skilled jobs, and the displaced human workers can be trained to take more lucrative positions. However, if Amodei is correct, AI could wipe out more specialized white-collar roles that may have required years of expensive training and education — and those workers may not be so easily retrained for equal or higher-paying jobs. Amodei suggested that lawmakers may even need to consider levying a tax on AI companies. 'If AI creates huge total wealth, a lot of that will, by default, go to the AI companies and less to ordinary people,' he said. 'So, you know, it's definitely not in my economic interest to say that, but I think this is something we should consider and I think it shouldn't be a partisan thing.' 'Faster, broader, harder to adapt to' Researchers and economists have forecast that professionals from paralegals and payroll clerks to financial advisers and coders could see their jobs dramatically change – if not eliminated entirely – in the coming years thanks to AI. Meta CEO Mark Zuckerberg said last month that he expects AI to write half the company's code within the next year; Microsoft CEO Satya Nadella said as much as 30 percent of his company's code is currently being written by AI. Amodei told CNN that Anthropic tracks how many people say they use its AI models to augment human jobs versus to entirely automate human jobs. Currently, he said, it's about 60 percent of people using AI for augmentation and 40 percent for automation, but that the latter is growing. Last week, the company released a new AI model that it says can work independently for almost seven hours in a row, taking on more complex tasks with less human oversight. Amodei says most people don't realize just how quickly AI is advancing, but he advises 'ordinary citizens' to 'learn to use AI.' 'People have adapted to past technological changes,' Amodei said. 'But everyone I've talked to has said this technological change looks different, it looks faster, it looks harder to adapt to, it's broader. The pace of progress keeps catching people off guard.' Estimates about just how quickly AI models are improving vary widely. And some skeptics have predicted that as big AI companies run out of high-quality, publicly available data to train their models on, after having already gobbled up much of the internet, the rate of change in the industry may slow. Some who study the technology also say it's more likely that AI will automate certain tasks, rather than entire jobs, giving human workers more time to do complex tasks that computers aren't good at yet. But regardless of where they fall on the prediction scale, most experts agree that it is time for the world to start planning for the economic impacts of AI. 'People sometimes comfort themselves (by) saying, 'Oh, but the economy always creates new jobs,'' University of Virginia business and economics professor Anton Korinek said in an email. 'That's true historically, but unlike in the past, intelligent machines will be able to do the new jobs as well, and probably learn them faster than us humans.' Amodei said he also believes that AI will have positive impacts, such as curing disease. 'I wouldn't be building this technology if I didn't think that it could make the world better,' he said. For the CEO, making this warning now could serve, in some ways, to boost his reputation as a responsible leader in the space. The top AI labs are competing not only to have the most powerful models, but also be perceived as the most trustworthy stewards of the tech transformation, amid growing questions from lawmakers and the public about the technology's efficacy and implications. 'Amodei's message is not just about warning the public. It's part truth-telling, part reputation management, part market positioning, and part policy influence,' tech futurist and Futuremade CEO Tracey Follows told CNN in an email. 'If he makes the claim that this will cause 20 percent unemployment over the next five years, and no-one stops or impedes the ongoing development of this model … then Anthropic cannot be to blame in the future — they warned people.' Amodei told Cooper that he's 'raising the alarm' because other AI leaders 'haven't as much and I think someone needs to say it and to be clear.' 'I don't think we can stop this bus,' Amodei said. 'From the position that I'm in, I can maybe hope to do a little to steer the technology in a direction where we become aware of the harms, we address the harms, and we're still able to achieve the benefits.'


Mid East Info
25-05-2025
- Mid East Info
Red Hat Unlocks Generative AI for Any Model and Any Accelerator Across the Hybrid Cloud with Red Hat AI Inference Server
Red Hat AI Inference Server, powered by vLLM and enhanced with Neural Magic technologies, delivers faster, higher-performing and more cost-efficient AI inference across the hybrid cloud BOSTON – RED HAT SUMMIT – MAY, 2025 — Red Hat, the world's leading provider of open source solutions, announced Red Hat AI Inference Server, a significant step towards democratizing generative AI (gen AI) across the hybrid cloud. A new offering within Red Hat AI, the enterprise-grade inference server is born from the powerful vLLM community project and enhanced by Red Hat's integration of Neural Magic technologies, offering greater speed, accelerator-efficiency and cost-effectiveness to help deliver Red Hat's vision of running any gen AI model on any AI accelerator in any cloud environment. Whether deployed standalone or as an integrated component of Red Hat Enterprise Linux AI (RHEL AI) and Red Hat OpenShift AI, this breakthrough platform empowers organizations to more confidently deploy and scale gen AI in production. Inference is the critical execution engine of AI, where pre-trained models translate data into real-world impact. It's the pivotal point of user interaction, demanding swift and accurate responses. As gen AI models explode in complexity and production deployments scale, inference can become a significant bottleneck, devouring hardware resources and threatening to cripple responsiveness and inflate operational costs. Robust inference servers are no longer a luxury, but a necessity for unlocking the true potential of AI at scale, navigating underlying complexities with greater ease. Red Hat directly addresses these challenges with Red Hat AI Inference Server — an open inference solution engineered for high performance and equipped with leading model compression and optimization tools. This innovation empowers organizations to fully tap into the transformative power of gen AI by delivering dramatically more responsive user experiences and unparalleled freedom in their choice of AI accelerators, models and IT environments. vLLM: Extending inference innovation: Red Hat AI Inference Server builds on the industry-leading vLLM project, which was started by University of California, Berkeley in mid-2023. The community project delivers high-throughput gen AI inference, support for large input context, multi-GPU model acceleration, support for continuous batching and more. vLLM's broad support for publicly available models – coupled with its day zero integration of leading frontier models including DeepSeek, Gemma, Llama, Llama Nemotron, Mistral, Phi and others, as well as open, enterprise-grade reasoning models like Llama Nemotron – positions it as a de facto standard for future AI inference innovation. Leading frontier model providers are increasingly embracing vLLM, solidifying its critical role in shaping gen AI's future. Introducing Red Hat AI Inference Server: Red Hat AI Inference Server packages the leading innovation of vLLM and forges it into the enterprise-grade capabilities of Red Hat AI Inference Server. Red Hat AI Inference Server is available as a standalone containerized offering or as part of both RHEL AI and Red Hat OpenShift AI. Across any deployment environment, Red Hat AI Inference Server provides users with a hardened, supported distribution of vLLM, along with: Intelligent LLM compression tools for dramatically reducing the size of both foundational and fine-tuned AI models, minimizing compute consumption while preserving and potentially enhancing model accuracy. Optimized model repository, hosted in the Red Hat AI organization on Hugging Face, offers instant access to a validated and optimized collection of leading AI models ready for inference deployment, helping to accelerate efficiency by 2-4x without compromising model accuracy. Red Hat's enterprise support and decades of expertise in bringing community projects to production environments. Third-party support for even greater deployment flexibility, enabling Red Hat AI Inference Server to be deployed on non-Red Hat Linux and Kubernetes platforms pursuant to Red Hat's third-party support policy. Red Hat's vision: Any model, any accelerator, any cloud. The future of AI must be defined by limitless opportunity, not constrained by infrastructure silos. Red Hat sees a horizon where organizations can deploy any model, on any accelerator, across any cloud, delivering an exceptional, more consistent user experience without exorbitant costs. To unlock the true potential of gen AI investments, enterprises require a universal inference platform – a standard for more seamless, high-performance AI innovation, both today and in the years to come. Just as Red Hat pioneered the open enterprise by transforming Linux into the bedrock of modern IT, the company is now poised to architect the future of AI inference. vLLM's potential is that of a linchpin for standardized gen AI inference, and Red Hat is committed to building a thriving ecosystem around not just the vLLM community but also llm-d for distributed inference at scale. The vision is clear: regardless of the AI model, the underlying accelerator or the deployment environment, Red Hat intends to make vLLM the definitive open standard for inference across the new hybrid cloud. Red Hat Summit: Join the Red Hat Summit keynotes to hear the latest from Red Hat executives, customers and partners: Modernized infrastructure meets enterprise-ready AI — Tuesday, May 20, 8-10 a.m. EDT (YouTube) Hybrid cloud evolves to deliver enterprise innovation — Wednesday, May 21, 8-9:30 a.m. EDT (YouTube) Supporting Quotes: Joe Fernandes, vice president and general manager, AI Business Unit, Red Hat 'Inference is where the real promise of gen AI is delivered, where user interactions are met with fast, accurate responses delivered by a given model, but it must be delivered in an effective and cost-efficient way. Red Hat AI Inference Server is intended to meet the demand for high-performing, responsive inference at scale while keeping resource demands low, providing a common inference layer that supports any model, running on any accelerator in any environment.' Ramine Roane, corporate vice president, AI Product Management, AMD 'In collaboration with Red Hat, AMD delivers out-of-the-box solutions to drive efficient generative AI in the enterprise. Red Hat AI Inference Server enabled on AMD Instinct™ GPUs equips organizations with enterprise-grade, community-driven AI inference capabilities backed by fully validated hardware accelerators.' Jeremy Foster, senior vice president and general manager, Cisco 'AI workloads need speed, consistency, and flexibility, which is exactly what the Red Hat AI Inference Server is designed to deliver. This innovation offers Cisco and Red Hat opportunities to continue to collaborate on new ways to make AI deployments more accessible, efficient and scalable—helping organizations prepare for what's next.' Bill Pearson, vice president, Data Center & AI Software Solutions and Ecosystem, Intel 'Intel is excited to collaborate with Red Hat to enable Red Hat AI Inference Server on Intel® Gaudi® accelerators. This integration will provide our customers with an optimized solution to streamline and scale AI inference, delivering advanced performance and efficiency for a wide range of enterprise AI applications.' John Fanelli, vice president, Enterprise Software, NVIDIA 'High-performance inference enables models and AI agents not just to answer, but to reason and adapt in real time. With open, full-stack NVIDIA accelerated computing and Red Hat AI Inference Server, developers can run efficient reasoning at scale across hybrid clouds, and deploy with confidence using Red Hat Inference Server with the new NVIDIA Enterprise AI validated design.' About Red Hat: Red Hat is the world's leading provider of enterprise open source software solutions, using a community-powered approach to deliver reliable and high-performing Linux, hybrid cloud, container, and Kubernetes technologies. Red Hat helps customers integrate new and existing IT applications, develop cloud-native applications, standardize on our industry-leading operating system, and automate, secure, and manage complex environments. Award-winning support, training, and consulting services make Red Hat a trusted adviser to the Fortune 500. As a strategic partner to cloud providers, system integrators, application vendors, customers, and open source communities, Red Hat can help organizations prepare for the digital future. Forward-Looking Statements: Except for the historical information and discussions contained herein, statements contained in this press release may constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are based on the company's current assumptions regarding future business and financial performance. These statements involve a number of risks, uncertainties and other factors that could cause actual results to differ materially. Any forward-looking statement in this press release speaks only as of the date on which it is made. Except as required by law, the company assumes no obligation to update or revise any forward-looking statements.