HPE Private Cloud AI is a purpose-built solution designed to provide fast and easy deployment of private AI applications with a focus on inferencing, Retrieval-Augmented Generation (RAG), and fine-tuning. HPE Private Cloud AI is a co-developed HPE and NVIDIA enterprise purpose-built solution including a complete infrastructure and software portfolio.

  • Overview

    At A Glance

    HPE Private Cloud AI delivers a unique cloud experience designed to accelerate data science productivity and time to business value. It delivers instant AI productivity by arriving at a customer’s location ready to deploy in three clicks after hardware installation and software onboarding are complete. Once available, multiple personas have self-service access to a diverse set of NVIDIA technologies and open-source tools and models to increase productivity by 90% through an evergreen, cloud-managed experience. 1


    Notes: 1 Source: HPE internal reports. Comparison between using GPT-4 via OpenAI API vs. self-hosted Llama3, assuming an enterprise account with 5,000 users, 5 chat sessions per day, 8,000 tokens per chat.


    AI teams can innovate faster with built-in compliance and explainability to foster model trust, quickly detect model bias, diagnose and improve model performance, and remain compliant with industry regulations.


    Built with enterprise-grade controls means organizations can fearlessly innovate, with a scalable platform — all controlled from a unified dashboard.


    Future-proof your AI journey with HPE. Launch small, scale seamlessly, and invest confidently with our co-developed NVIDIA + HPE solution. One modular architecture protects customers’ investment by ensuring compatibility with future innovations from NVIDIA, HPE, and the open-source world.


    • - Ready to use in three clicks
    • - Evergreen cloud experience with NVIDIA technologies and a rich ecosystem of open-source tools and models
    • - Automated AI pipelines with clear data lineage and verifiable changes empower efficient, accountable development
    • - Robust security, on-demand scalability, and compliance for data and AI models — all managed from a single dashboard

    One of the challenges businesses face is getting AI pilots to production faster.


    HPE Private Cloud AI delivers instant AI productivity with a unique, private cloud experience that accelerates the productivity of data science teams and time to business value with NVIDIA AI Computing.


    HPE Private Cloud for AI offers enterprise customers the ability to leverage NVIDIA AI Enterprise (NVAIE) portfolio, including NVIDIA Inferencing Microservices (NIM), and HPE portfolio of curated market adopted open-source AI tools and platforms with full private control of their data. The solution will enable enterprises to expedite their Machine Learning and AI initiatives starting from creating their private data lakehouses, to data pipeline, model development and fine-tuning, to operationalizing their GenAI workflows.

    What’s New

    • - HPE Private Cloud AI G2 Small configuration featuring NVIDIA® RTX Pro 6000 Blackwell Server Edition GPUs
    • - HPE Private Cloud AI G2 Air-Gapped deployment option

    As agentic AI simplifies automation and provides consistent profitability from AI workloads, the importance of rapid productivity, reliability, and data protection, has never been greater. To accelerate the development and delivery of agentic AI across industries, Hewlett Packard Enterprise and NVIDIA® are introducing the Private Cloud AI G2 Small configuration featuring the NVIDIA RTX Pro 6000 Blackwell Server Edition GPU to their end-to-end enterprise AI platform with support for Physical AI and Visual Computing workloads


    HPE is also supporting secure deployments with the Private Cloud AI G2 medium air-gapped solution. Ensure the protection of sensitive data with private AI model customization by eliminating exposure to external networks.

Category

Description

Platform

- Server support: HPE ProLiant Compute Gen11 and Gen12 servers

- AMD-based HPE ProLiant Compute DL325 Gen 11 Control Nodes

- Intel-based HPE ProLiant Compute DL380a Gen 11 AI Worker Nodes (G1)

- Intel-based HPE ProLiant Compute DL380a Gen 12 AI Worker Nodes (G2)

- Storage support: HPE GreenLake for File with Object Storage enabled for Small/Medium /Large T-shirt sizes

Manageability

- Cloud-based setup and lifecycle management (single-click upgrades)

Analytics & Monitoring

- Cluster and VM capacity and performance, storage health status information

Support

- One call support experience with HPE Services

- HPE Private Cloud AI Smart templates

  • Availability of pre-configured Smart Templates with HPE ProLiant Compute Gen 11 Servers (G1 T-shirt Sizes)
  • Availability of pre-configured Smart Templates with HPE ProLiant Compute Gen 12 Servers (G2 T-shirt Sizes)

HPE Private Cloud AI G1 Family

Feature

Developer System

Small

Medium

Large

Control Node Qty

1

3

3

3

Worker Node QTY

1

1 or 2

2 or 4

4 or 8

Worker Node Generation

HPE ProLiant Compute Gen11

HPE ProLiant Compute Gen11

HPE ProLiant Compute Gen11

HPE ProLiant Compute Gen11

CPO Type / QTY (per node)

2x Xeon 32 Core CPUs

2x Xeon 32 Core CPUs

2x Xeon 32 Core CPUs

2x Xeon 32 Core CPUs

GPU Type / QTY

2x H100NVL

4 or 8x L40S

8 or 16x L40S

16 or 32x H100NVL

Storage

32TB Internal File/Object

109TB GreenLake for File with Object Storage

217TB GreenLake for File with Object Storage

670TB GreenLake for File with Object Storage

Networking Switches

N/A

Nvidia 4600cM Aruba 6300M (oobm)

Nvidia 4700cM Aruba 6300M (oobm)

Nvidia 4700cM Aruba 6300M (oobm)

NIC Speed (AI Network)

200Gb NICs

100Gb NICs

200Gb NICs

400Gb NICs

Rack / PDU

N/A

1x 42U Rack with PDUs

1x 42U Rack with PDUs

2x 42U Rack with PDUs

Install Services Included

N/A

Yes

Yes

Yes

Sales Motion

Traditional or GreenLake

Traditional or GreenLake

Traditional or GreenLake

Traditional or GreenLake

Air-Gapped Deployment Option

No

No

No

No

HPE Private Cloud AI G2 Family

Feature

G2 Small

G2 Medium

G2 Large

G2 Expansion Rack

Control Node Qty

3

3

3

0

Worker Node QTY

1 or 2

2

2

2

Worker Node Generation

HPE ProLiant Gen12

HPE ProLiant Gen12

HPE ProLiant Gen12

HPE ProLiant Gen12

CPU Type / QTY (per node)

2x Xeon 86 Core CPUs

2x Xeon 86 Core CPUs

2x Xeon 86 Core CPUs

2x Xeon 86 Core CPUs

GPU Type / QTY (per solution)

4 or 8x RTX Pro 6000

8x H200

16x H200

16x H200 or

8x RTX Pro 6000

E xpansion Racks Supported

1

1

3

N/A

Storage

109TB GreenLake for File with Object Storage

109TB GreenLake for File with Object Storage

217TB GreenLake for File with Object Storage

N/A

Networking Switches

Nvidia 4700M Aruba 6300M (oobm)

Nvidia 4700M Aruba 6300M (oobm)

Nvidia 4700M Aruba 6300M (oobm)

N/A

NIC Speed (AI Network)

400Gb NICs

400Gb NICs

400Gb NICs

400Gb NICs

Rack / PDU

1x 42U Rack with PDUs

1x 42U Rack with PDUs

1x 42U Rack with PDUs

1x 42U Rack with PDUs

Install Services Included

Yes

Yes

Yes

Yes

Sales Motion

Traditional or GreenLake

Traditional or GreenLake

Traditional or GreenLake

Traditional or GreenLake

Air-Gapped Deployment Option

No

Yes

No

Yes

Key Features and Benefits

HPE Private Cloud AI is turnkey, deployed in minutes, cloud-managed, and ready to use by AI personas and IT operations teams and provides rapid productivity for AI initiatives while protecting data and IP. The key value proposition aligned to customer problems are:


The core feature set includes:

  • - Instant AI productivity: HPE Private Cloud AI provides a unique, private cloud experience that accelerates data science productivity and time to business value with NVIDIA AI Computing. The solution is pre-integrated and ready to run out of the box in minutes. It is not a reference architecture like other solutions in the market.
  • - Unify access to all your data: Secure and Unified access to all your data: HPE simplifies data management and reduces cost and complexity by integrating, organizing, and governing enterprise data for seamless access, data integrity and compliance. Enterprise-grade confidence and control: HPE Private Cloud AI is managed through a simple control plane on HPE GreenLake. Users can easily provision, orchestrate, manage and monitor the private cloud environment and the hybrid cloud landscape it exists within. Comprehensive, multi-layered controls protect sensitive data and models and maintain high performance, reliability and utilization of AI infrastructure.

Cloud experience that keeps data private: HPE Private Cloud AI delivers a true cloud experience through HPE GreenLake. Deployed on-premises and designed for hybrid, HPE Private Cloud Ai provides flexible and modular choices to expand and grow with AI demand. As business needs change, it’s easy for customers to grow the solution. And monthly subscription pricing allows customers to start small financially and grow as their projects prove ROI.

  • Service and Support

    Support is included as part of the subscription for HPE Private Cloud AI. Included with the support is 24x7 telephone and email support for the arrays and hardware components for the chosen subscription term.


    Refer to the HPE Private Cloud AI Data sheet https://www.hpe.com/psnow/doc/a50010051enw?section=Document%20Types for the service deliverables and the shared responsibility model as part of the subscription.

  • Configuration Information

    Easy Configuration through Smart templates


    There are pre-defined smart templates available that allow for quick and easy ways to quote:

    1. HPE Private Cloud AI

    There are pre-defined smart templates that allow for quick and easy way to quote HPE Private Cloud AI.


    Here is an example of a HPE Private Cloud AI Smart template

    Config Name: PrivateCloudAI-Small-1Svr/4xL40S GPU-109TB File/Object-3Phase/NA-Jpn-PDU-1Rack-3yr

    Description : HPE Private Cloud AI Small Single Node-4GPU Solution for AI Inference. 109TB File/Object Storage, 100GbE Networking, Single Rack and 3Phase PDU


    The Smart templates contain the following attributes to choose,

    • 1. T-Shirt Sizing – Developer System, Small, Medium, or Large Configurations
    • 2. Workload Tier
      • a. AI Inference
      • b. Retrieval Augmented Generation (RAG)
      • c. Model Fine Tuning

G1 T-Shirt Size

Entry

Expanded

G1 Developer System

2x H100NVL GPUs and 32TB of integrated Storage for developing low-mid parameter model AI applications

N/A

G1 Small

4x L40s GPUs and 109TB Storage for AI Inference

8x L40s GPUs and 109TB Storage for AI Inference

G1 Medium

8x L40s GPUs and 217TB Storage for AI Inference and RAG

16x L40s GPUs and 217TB Storage for AI Inference and RAG

G1 Large

16x H100NVL GPUs and 670TB Storage for AI Inference, RAG, and Fine Tuning

32x H100NVL GPUs and 670TB Storage for AI Inference, RAG and Fine Tuning

G2 T-Shirt Size

Base

Expanded

G2 Small

4 or 8x RTX Pro 6000 GPUs and 109TB Storage for AI Inference and Visual Computing

8 or 12 RTX Pro 6000 GPUs or 16x H200 GPUs via 1x Expansion Rack

G2 Medium

8x H200 GPUs and 109TB Storage for AI Inference and RAG

24x H200 GPUs via 1x Expansion Rack

G2 Medium Air-Gapped

8x H200 GPUs and 109TB Storage for AI Inference and RAG

24x H200 GPUs via 1x Expansion Rack

G2 Large

16x H200 GPUs and 217TB Storage for AI Inference, RAG, and Fine Tuning

Up to 64x H200 GPUs via 3x Expansion Racks

3. Network Configuration


Network Configuration (Small/Medium/Large)

Detail

Networking equipment included

Two top-of-rack switches and out of band management switches are included along with all transceivers and signal cabling required for l in-rack solution (Customers are responsible for transceivers to connect to core network switches)

Network Configuration

(developer system)

Detail

Networking equipment included

None- Requires customer furnished networking (100GbE or 200GbE recommended)

4. Rack and power Configuration

Rack Configuration

(Small/Medium/Large)

Detail

Rack included

The solution will include a 42U Rack with integrated PDUs for HPE Private Cloud AI.

Rack Dimensions: 600mm (W), 1200mm (D)

Rack Configuration

(developer system)

Detail

Rack included

None- Requires customer furnished rack and PDUs (~2200W Recommended)

Resources and additional links

Shared Responsibility Model (SRM)

HPE Private Cloud AI subscription includes the necessary hardware, software, and services to deliver the service level specified. The service levels offered are based on a foundational shared responsibility model (SRM) depicted below:

HPE Private Cloud AI Small/Medium/Large Configurations


Customer

HPE

Responsible for the connectivity to GreenLake Cloud Platform (GLCP), the administration, and the management of the data/objects

Responsible for the functionality of the service

Site Readiness including datacenter facilities and internet connectivity

Installation of hardware systems & activation of Service

Maintain connectivity to GreenLake Cloud Platform

Data resilience and remote replication

Data backup

Customer Orientation

Access to software, firmware, and documentation updates

Onsite hardware support

Applying recommended software updates & security patches Data Monitoring


Initiating the order of additional capacity beyond total available capacity

Operational guidance through the platform

Red Hat Linux and Rocky Linux OS lifecycle management


NVIDIA and Aruba Switch OS/Firmware lifecycle management

Operational insights and dashboard through the platform

HPE PDU Firmware lifecycle management


NVIDIA GPU Firmware lifecycle management


HPE Private Cloud AI Medium Air-Gapped Configuration

Customer

HPE

Site Readiness including datacenter facilities

Installation of hardware systems & activation of Service

Data resilience and remote replication

Data backup

Customer Orientation

Access to software, firmware, and documentation updates

Onsite hardware support

Applying recommended software updates & security patches Data Monitoring


Initiating the order of additional capacity beyond total available capacity


NVIDIA and Aruba Switch OS/Firmware lifecycle management


HPE PDU Firmware lifecycle management


NVIDIA GPU Firmware lifecycle management


HPE Private Cloud AI developer system

Customer

HPE

Responsible for the connectivity to GreenLake Cloud Platform (GLCP), the administration, and the management of the data/ objects

Responsible for the functionality of the service

Site Readiness including datacenter facilities and internet connectivity


Maintain connectivity to GreenLake Cloud Platform

Data resilience and remote replication

Data backup

Access to software, firmware, and documentation updates

Onsite hardware support

Applying recommended software updates & security patches Data Monitoring


Initiating the order of additional capacity beyond total available capacity

Operational guidance through the platform

Red Hat Linux and Rocky Linux OS lifecycle management


Installation of hardware systems & activation of Service

Operational insights and dashboard through the platform

NVIDIA GPU Firmware lifecycle management


Pre-requisite for HPE Private Cloud AI

As part of the shared responsibility model, the customer is expected to make appropriate decisions including but not limited to:

  • - Rack Infrastructure
    • Space
    • Rails
  • - Power Infrastructure
    • PDU − Cables
  • Summary of Changes

Date

Version History

Action

Description of Change

06-Oct-2025

Changed

HPE Rebranding FY25

02-Sep-2025

Changed

Overview and Configuration Information sections were updated. - AI G2 Small configuration featuring NVIDIA® RTX Pro 6000 Blackwell Server Edition GPUs information was added.

04-Aug-2025

Changed

Overview and Configuration Information sections were updated. - Added G2 Medium Air-Gapped configurations

21-Jul-2025

Changed

Survey link updated.

02-Jun-2025

Changed

Overview and Configuration Information sections were updated.

07-Apr-2025

Changed

Overview and Configuration Information sections were updated.

Updates for developer system

03-Mar-2025

Changed

Overview section was updated

03-Sep-2024

New

New QuickSpecs

Recommended for you