- Community Home
- >
- Servers and Operating Systems
- >
- Servers & Systems: The Right Compute
- >
- Qualcomm unveils AI inference for large language m...
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Discussions
Forums
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark as New
- Mark as Read
- Bookmark
- Receive email notifications
- Printer Friendly Page
- Report Inappropriate Content
Qualcomm unveils AI inference for large language models
HPE is collaborating with Qualcomm on the integration of Qualcomm Cloud AI 100 Ultra accelerators in HPE ProLiant Gen11 Servers to deliver exceptional performance for generative AI/LLM inference solutions.
At SC23 Qualcomm announced the Qualcomm Cloud AI 100 Ultra acceleratorโdesigned for deploying LLMsโto their Cloud AI 100 product lineup. In support of this announcement, HPE is collaborating with Qualcomm on integration testing and delivery of Qualcomm Cloud AI 100 Ultra with select HPE ProLiant Gen11 Servers.
This blog delves into how this offering benefits our customers, its remarkable performance capabilities, and who the ideal customers are for this next-gen AI Inference solution.
Introducing Qualcomm Cloud AI 100 Ultra
The Qualcomm Cloud AI 100 Ultra is an advanced AI accelerator designed to deliver exceptional performance and efficiency for Generative AI and LLMs. It's built to address the skyrocketing demand for AI Inference workloads for both enterprise and cloud-service provider customers.
The Qualcomm Cloud AI 100 ULTRA is optimized for a range of AI workloads, from large language models (LLMs), natural language processing (NLP) and computer vision. It's capable of supporting 100B parameter models on a single slot, 150W PCIe card. Larger models are supported with multi-card software stack from Qualcomm.
Performance for deploying LLMs
The Qualcomm Cloud AI 100 Ultra delivers exceptional performance. The accelerator boasts impressive throughput and low-latency AI processing capabilities, making it ideal for time-sensitive AI applications. Its peak performance capabilities and support for LLMs are particularly noteworthy, with significant inferences per second (ips) comparable with AI Inference accelerators from leading GPU vendors. This computational power makes it well-suited for applications that require real-time decision-making, such as text-to-code, chatbots, as text-to-language translation.
Qualcomm Cloud AI 100 Ultra is also energy efficient; with a single-width PCIe design operating at a mere 150W TDPโitโs able to perform at levels similar to AI accelerators operating at twice the Wattage (and price). Thus, the Qualcomm Cloud AI 100 deliver industry-leading AI inference per watt, significantly reducing the total cost of ownership for data centers and cloud service providers. As a result Qualcomm Cloud AI 100 Ultra delivers performance up to 4X of the Qualcomm Cloud AI 100 Standard and Pro models.
The Qualcomm Cloud AI 100 Ultra accelerator also supports leading industry-standard frameworks (eg. PyTorch, ONNX, TensorFlow) and tools, ensuring compatibility with existing AI software ecosystems. This makes the transition to the Qualcomm Cloud AI 100 Ultra smooth for businesses already invested in AI technologies.
Ideal customer
Customers for this AI Inference solution are deploying LLMs, NLP and CV models and require high-performance and energy efficiency. Example customers and industries include:
Cloud Service Providers: Qualcomm Cloud AI 100 Ultra provides an AI inference solution to a wide range of clients, from e-commerce platforms to content streaming services.
Data Centers: Data centers housing massive amounts of data will appreciate the performance and energy efficiency of this AI accelerator. It enables data centers to manage workloads more efficiently and cost-effectively, reducing the environmental impact.
AI Researchers and Developers: For those pushing the boundaries of AI research and development, the Qualcomm Cloud AI 100 Ultra with HPE Servers will offer industry-leading performance for experimentation and innovation. It can accelerate the development of new AI applications and algorithms.
Product availability
Qualcomm Cloud AI 100 Ultra will be offered with select HPE ProLiant Gen11 Servers that can fit 8 x Qualcomm Ultra accelerator card in a single 2U server form factor. Expect HPE integrated product and pricing to be available in H1-2024.
Read the press release.
Server Experts
Hewlett Packard Enterprise
twitter.com/HPE_HPC
linkedin.com/showcase/hpe-servers-and-systems/
hpe.com/servers
- Back to Blog
- Newer Article
- Older Article
- PerryS on: Explore key updates and enhancements for HPE OneVi...
- Dale Brown on: Going beyond large language models with smart appl...
- alimohammadi on: How to choose the right HPE ProLiant Gen11 AMD ser...
- ComputeExperts on: Did you know that liquid cooling is currently avai...
- Jams_C_Servers on: If youโre not using Compute Ops Management yet, yo...
- AmitSharmaAPJ on: HPE servers and AMD EPYCโข 9004X CPUs accelerate te...
- AmandaC1 on: HPE Superdome Flex family earns highest availabili...
- ComputeExperts on: New release: What you need to know about HPE OneVi...
- JimLoi on: 5 things to consider before moving mission-critica...
- Jim Loiacono on: Confused with RISE with SAP S/4HANA options? Let m...
-
COMPOSABLE
77 -
CORE AND EDGE COMPUTE
146 -
CORE COMPUTE
154 -
HPC & SUPERCOMPUTING
137 -
Mission Critical
87 -
SMB
169