Red Hat recently announced that it has completed its acquisition of Neural Magic, a pioneer in software and algorithms that accelerate generative AI (gen AI) inference workloads. With Neural Magic, Red Hat adds expertise in inference performance engineering and model optimization, helping further the company’s vision of high-performing AI workloads that directly map to unique customer use cases, wherever needed across the hybrid cloud.
The large language models (LLMs) underpinning today’s gen AI use cases, while innovative, are often too expensive and resource-intensive for most organizations to use effectively. To address these challenges, Red Hat views smaller, optimized and open source-licensed models driven by open innovation across compute architectures and deployment environments as key to the future success of AI strategies.
Neural Magic’s commitment to making optimized and efficient AI models a reality furthers Red Hat’s ability to deliver on this vision for AI. Neural Magic is also a leading contributor to vLLM, an open source project developed by UC Berkeley for open model serving, which will help bring even greater choice and accessibility in how organizations build and deploy AI workloads.
The future of hybrid cloud-ready gen AI
With Neural Magic’s technology and performance engineering expertise, Red Hat aims to break through the challenges of wide-scale enterprise AI, using open source innovation to further democratize access to AI’s transformative power via:
- Open source-licensed models, from the 1B to 100’s of billions parameter scale, that can run anywhere and everywhere needed across the hybrid cloud – in corporate data centers, on multiple clouds and at the edge.
- Fine-tuning capabilities that enable organizations to more easily customize LLMs to their private data and uses cases with a stronger security footprint;
- Inference performance engineering expertise, resulting in greater operational and infrastructure efficiencies; and
- A partner and open source ecosystem and support structures that enable broader customer choice, from LLMs and tooling to certified server hardware and underlying chip architectures.
The concept of choice is as crucial for gen AI today as it was cloud-native or containerized applications several years ago: The right environment (cloud, server, edge, etc.), accelerated compute and inference server are all critical for successful gen AI strategies. Red Hat remains firm in its commitment to customer choice across the hybrid cloud, including AI, with the acquisition of Neural Magic furthering supporting this promise.
Red Hat AI: An open source backbone for AI
The expertise and capabilities of Neural Magic will be incorporated into Red Hat AI, Red Hat’s portfolio of gen AI platforms. Built with the hybrid cloud in mind, Red Hat AI encompasses:
- Red Hat Enterprise Linux AI (RHEL AI), a foundation model platform to more seamlessly develop, test and run the IBM Granite family of open source-licensed LLMs for enterprise applications on Linux server deployments;
- Red Hat OpenShift AI, an AI platform that provides tools to rapidly develop, train, serve and monitor machine learning models across distributed Kubernetes environments on-site, in the public cloud or at the edge; and
- InstructLab, an approachable open source AI community project created by Red Hat and IBM that enables anyone to shape the future of gen AI via the collaborative improvement of open source-licensed Granite LLMs using InstructLab’s fine-tuning technology.