Tesla GPU Optimised Servers

NVIDIA Tesla based GPU optimised servers for AI, Deep Learning and HPC - configurable with up to 10x GPU's for massively parallel computing.

Deployed by some of the planets largest supercomputing enterprises, NVIDIA Tesla is the worlds leading platform for accelerating data centres. Key to this platform is the hugely parallel PU accelerators that provide massively improved throughput for compute-intensive workloads - without increasing cost and physical footprint.


NVIDIA Tesla Elite Partner 25% Discount.

Drive Bay Qty
Drive Bay Size
CyberServe Xeon SP1-110S G3

Ideal for virtualisation, cloud computing, enterprise server. 2x PCI-E 4.0 x16 slots. Intel® Ethernet Controller X550 2x 10GbE RJ45. Redundant power supplies.

Form Factor:
1U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
10
Drive Interface:
SATA , NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
8x 3200MHz
GPU Support:
Tesla GPU Optimised
Features:
Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
1TB
Configure From: $6,299
Configure
CyberServe Xeon SP1-102N G3

Edge Server – 1U 3rd Gen. Intel Xeon Scalable GPU server system, ideal for AI & Edge applications.

Form Factor:
1U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
2
Drive Interface:
SATA , NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
16x 3200MHz
GPU Slots:
1x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
Full Height/Length Expansion, Redundant Power Supply - Standard, Short Depth, Front I/O Ports
Max RAM Capacity:
2TB
Configure From: $6,542
Configure
CyberServe Xeon SP1-104S G4 GPU

Supports 1x double slot GPU card, 4th/5th Gen Intel Xeon Scalable processor, dual 1Gb/s LAN ports, redundant power supply, 4x 3.5" NVMe/SATA hot-swappable bays

Form Factor:
1U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
4
Drive Interface:
SATA , 12Gb/s SAS, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
8x 4800MHz
GPU Slots:
1x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
1TB
Configure From: $6,929
Configure
CyberServe Xeon SP1-110S NVMe G4 GPU

Supports 1x double slot GPU card, 4th/5th Gen Intel Xeon Scalable processor, dual 1Gb/s LAN ports, redundant power supply, 10x 2.5" NVMe/SATA hot-swappable bays

Form Factor:
1U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
10
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
8x 4800MHz
GPU Slots:
1x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
1TB
Configure From: $7,327
Configure
CyberServe Xeon SP1-208S NVMe G4 GPU

Supports 2x double slot GPU cards, 4th/5th Gen Intel Xeon Scalable processor, dual 1Gb/s LAN ports, redundant power supply, 8x 3.5" NVMe/SATA hot-swappable bays

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
8
Drive Interface:
SATA , NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
8x 4800MHz
GPU Slots:
2x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
1TB
Configure From: $7,495
Configure
CyberServe Xeon SP1-202 G4 GPU

4th/5th Gen Intel Xeon Scalable processor, single 1Gb/s LAN port, redundant power supply, 2x 2.5" NVMe/SATA hot-swappable bays

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
2
Drive Interface:
SATA , 12Gb/s SAS, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
16x 4800MHz
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard, Short Depth
Max RAM Capacity:
2TB
Configure From: $7,712
Configure
CyberServe Xeon SP1-212 G4 GPU

Supports up to 3x double slot Gen5 GPU cards, single 1Gb/s LAN port, redundant power supply, 12x 3.5/2.5" SATA/SAS hot-swappable bays, 4th/5th Gen Intel Xeon Scalable processor

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA , 12Gb/s SAS, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
16x 4800MHz
GPU Slots:
3x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $8,903
Configure
3.5" Drives 
CyberServe EPYC EP1-G242-Z11

Up to 4 x NVIDIA ® PCIe Gen4 GPU cards. NVIDIA-Certified system for scalability, functionality, security, and performance. Dedicated management port. Redundant power.

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
4
Drive Interface:
SATA
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
8x 3200MHz
GPU Slots:
4x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, Extra Expansion Slots, Full Height/Length Expansion
Max RAM Capacity:
1TB
Configure From: $9,112
Configure
Short Depth 2.5" Drives NVMe Drives 
CyberServe EPYC EP1 202-NVMe-G G4

Short Depth Single AMD EPYC 9004 Series Edge Server with 2x GPU Slots, 2x 2.5" Gen4 NVMe/ SATA Hot-Swappable bays

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
2
Drive Interface:
SATA , NVMe, M.2
Server Processor:
AMD EPYC 9004 Series
Memory DIMMS:
12x 4800MHz
GPU Slots:
2x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, Full Height/Length Expansion, Redundant Power Supply - Standard, Short Depth
Max RAM Capacity:
1.5TB
Configure From: $9,156
Configure
3.5" Drives NVMe Drives 
CyberServe Xeon SP2-212NS G3

Supports 3x double slot GPU cards, dual 1Gb/s LAN ports, 5x PCIe Gen4 x16 slots, redundant power supply.

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
32x 3200MHz
GPU Slots:
3x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $10,753
Configure
CyberServe EPYC EP1 212-8NVMe G4

Single AMD EPYC 9004 Series, Supports up to 2x FHFL PCIe Gen5 x16 slots - 12x 3.5" NVMe / SATA Drives.

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
AMD EPYC 9004 Series
Memory DIMMS:
12x 4800MHz
GPU Slots:
2x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
1.5TB
Configure From: $10,962
Configure
2.5" Drives NVMe Drives 
CyberServe Xeon SP2-208-2S-SFF-GPU G3

2U GPU server powered by dual-socket 3rd Gen Intel Xeon Scalable processors that supports up to 16 DIMM, four dual-slot GPU, 2 M.2, four NVMe (by SKU), total eleven PCIe 4.0 slots

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
16x 3200MHz
GPU Slots:
4x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $11,639
Configure
2.5" Drives NVMe Drives 
CyberServe EPYC EP1-G292-Z20 GPU Server

8x PCIe Gen4 expansion slots for GPUs, 2 x 10Gb/s SFP+ LAN ports (Mellanox® ConnectX-4 Lx controller), 2 x M.2 with PCIe Gen3 x4/x2 interface

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
8x 3200MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
1TB
Configure From: $11,858
Configure
Ultra High-Performance 2.5" Drives NVMe Drives 10Gb Lan 
CyberServe SP2-104-2S-GPU G3

GPU server optimised for HPC, Scientific Virtualisation and AI. Powered by 3rd Gen Intel Xeon Scalable processors. 6x PCIe Gen 4.0 x16, 1x M.2

Form Factor:
1U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
4
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
16x 3200MHz
GPU Slots:
4x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $12,342
Configure
Rackmount or Tower 3.5" Drives NVMe Drives 10Gb Lan 
Cyberserve Xeon SP2-408-4S GPU G3

Ideal for scientific virtualisation and HPC. 6x PCI-E 4.0 x16 slots. 2x M.2 NVMe or SATA supported. Redundant power supplies.

Form Factor:
Pedestal
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
16x 3200MHz
GPU Slots:
4x Double Width GPU / 8x Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $12,642
Configure
CyberServe Xeon SP2-408S NVMe G4 GPU

Dual 4th/5th Gen Intel Xeon Scalable Gen4 Processor, GPU Computing Pedestal Supercomputer Server, 4x Tesla, RTX GPU Cards

Form Factor:
Pedestal
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
16x 4800MHz
GPU Slots:
4x Double Width GPU / 8x Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $14,266
Configure
Ultra High-Performance 3.5" Drives 2.5" Drives NVMe Drives 
CyberServe EPYC EP2-G292-Z42 GPU Server

8x PCIe Gen3 expansion slots for GPUs, 2x 10Gb/s BASE-T LAN ports (Intel® X550-AT2 controller), 4x NVMe and 4x SATA/SAS 2.5" hot-swappable HDD/SSD bays

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
16x 3200MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $14,760
Configure
3.5" Drives 
CyberServe Xeon SP2-208-2S-GPU G3

2U dual-socket GPU server powered 3rd Gen Intel Xeon Scalable processors that supports up to 16 DIMM, four dual-slot GPU, 4 M.2, eight NVMe (by SKU), total eleven PCIe 4.0 slots.

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
16x 3200MHz
GPU Slots:
4x Double Width GPU / 8x Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
2TB
Configure From: $14,961
Configure
Ultra High-Performance 2.5" Drives NVMe Drives 
CyberServe SP2-G292-280 G3

GPU Server - 2U 8 x GPU Server | Application: AI , AI Training , AI Inference , Visual Computing & HPC. Dual 10Gb/s BASE-T LAN ports.

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
SATA , 12Gb/s SAS, NVMe
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
24x 3200MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
3.1TB
Configure From: $16,539
Configure
Ultra High-Performance 3.5" Drives 10Gb Lan 
CyberServe Xeon SP2-412G-GPU G3

Up to 8x PCIe Gen4 GPGPU cards, dual 10Gb/s LAN ports, redundant power option.

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
32x 3200MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard, Front I/O Ports
Max RAM Capacity:
4.1TB
Configure From: $18,413
Configure
2.5" Drives NVMe Drives 
CyberServe EPYC EP2-G482-Z51 GPU Server

Up to 8 x PCIe Gen4 GPGPU cards, 2 x 10Gb/s BASE-T LAN ports (Intel® X550-AT2), 8-Channel RDIMM/LRDIMM DDR4 per processor, 32 x DIMMs

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
10
Drive Interface:
SATA
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
32x 3200MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $18,592
Configure
CyberServe EPYC EP2-4124GS-TNR GPU Server

8 PCI-E 4.0 x16 + 3 PCI-E 4.0 x8 slots, Up to 24 Hot-swap 2.5" drive bays, 2 GbE LAN ports (rear)

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
24
Drive Interface:
SATA , NVMe
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
32x 3200MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $18,750
Configure
CyberServe Xeon SP2-208-4G NVMe G4 GPU

Supports up to 8x double slot Gen4 GPU cards, dual 10Gb/s BASE-T LAN ports, redundant power supply, 8x 2.5" NVMe/SATA hot-swappable bays. Built for AI & HPC

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
SATA , NVMe
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
24x 4800MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
3.1TB
Configure From: $21,301
Configure
CyberServe EPYC EP2 424-4NVMe-G GPU Server G4

Dual AMD EPYC 9004 Series 8x GPU Server - 20x 2.5" SATA / SAS + 4x NVMe Dedicated Drives

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
24
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
AMD EPYC 9004 Series
Memory DIMMS:
24x 4800MHz
GPU Slots:
8x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
3.1TB
Configure From: $21,644
Configure
Ultra High-Performance 3.5" Drives NVMe Drives 10Gb Lan 
CyberServe Xeon SP2 412-8G GPU G3

Up to 10x PCIe Gen4 GPGPU cards, dual 10Gb/s BASE-T LAN, redundant power supply.

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA , 12Gb/s SAS, NVMe
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
32x 3200MHz
GPU Slots:
10x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard, Front I/O Ports
Max RAM Capacity:
4.1TB
Configure From: $21,942
Configure
Ultra High-Performance 3.5" Drives NVMe Drives 
CyberServe EPYC EP2-G482-Z50 GPU Server

10 x FHFL Gen3 expansion slots for GPU cards, 2 x 10Gb/s BASE-T LAN ports (Intel® X550-AT2), 8 x 2.5" NVMe, 2 x SATA/SAS 2.5" hot-swappable HDD/SSD bays, 12 x 3.5" SATA/SAS hot-swappable HDD/SSD bays

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
22
Drive Interface:
SATA , 12Gb/s SAS, NVMe
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
32x 3200MHz
GPU Slots:
10x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $22,078
Configure
CyberServe Xeon SP2-412T G3 GPU

Supports 10x double slot GPU cards, redundant power supply, 12 x 3.5/2.5" NVMe/SATA hot-swappable bays. Built for AI & HPC

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Xeon Scalable Processor - Gen 3
Memory DIMMS:
32x 3200MHz
GPU Slots:
10x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard, Front I/O Ports
Max RAM Capacity:
4.1TB
Configure From: $26,493
Configure
CyberServe Xeon SP2-412 NVMe G4 GPU

Supports 10x double slot GPU cards, dual 10Gb/s BASE-T LAN ports, redundant power supply, 12x 3.5/2.5" NVMe/SATA hot-swappable bays. Built for AI & HPC

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
3.5" Drives
Qty Drives:
12
Drive Interface:
SATA , 12Gb/s SAS, NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
32x 4800MHz
GPU Slots:
10x Double / Single Width GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard, Front I/O Ports
Max RAM Capacity:
4.1TB
Configure From: $31,383
Configure
Promotion - Test Drive This Server 
CyberServe EPYC EP2-2124GQ-NART GPU Server

High Density 2U System with NVIDIA® HGX™ A100 4-GPU, Direct connect PCI-E Gen4 Platform with NVIDIA® NVLink™, IPMI 2.0 + KVM with dedicated 10G LAN

Form Factor:
2U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
4
Drive Interface:
SATA , 12Gb/s SAS, NVMe
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
32x 3200MHz
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $154,568
Configure
CyberServe Xeon SP2-308 NVMe G4 GPU

Supports 4x SXM5 GPU Modules, dual 10Gb/s BASE-T LAN ports, redundant power supply, 8 x 2.5" NVMe/SATA hot-swappable bays. Built for AI & HPC

Form Factor:
3U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
SATA , NVMe
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
16x 4800MHz
GPU Slots:
4x SXM GPU
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, Extra Expansion Slots, Redundant Power Supply - Standard, Front I/O Ports
Max RAM Capacity:
2TB
Configure From: $297,564
Configure
Promotion - Test Drive This Server 
CyberServe EPYC EP2-4124GO-NART GPU Server

8x NVIDIA A100 Gen4, 6x NVLink Switch Fabric, 2x M.2 on board and 4 Hybrid SATA/Nvme, 8x PCIe x16 Gen4 Slots

Form Factor:
4U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
6
Drive Interface:
NVMe, M.2
Server Processor:
AMD EPYC 7003 Processor
Memory DIMMS:
32x 3200MHz
GPU Support:
Tesla GPU Optimised
Features:
VMware Compatible, High RAM Capacity, Extra Expansion Slots, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $311,605
Configure
CyberServe Xeon SP2-824 NVMe G4 GPU

Supports 8x HGX H100 GPUs, dual 10Gb/s BASE-T LAN ports, redundant power supply, 16x 2.5" NVMe, 8x SATA hot-swappable bays. Built for AI Training and Inferencing.

Form Factor:
8U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
24
Drive Interface:
SATA , NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
Memory DIMMS:
32x 4800MHz
GPU Slots:
8x SXM GPU
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Extra Expansion Slots, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
4.1TB
Configure From: $584,647
Configure
CyberServe EPYC EP2-824 NVMe G4 GPU Server

Supports 8x HGX H100 GPUs, Dual AMD EPYC 9004 Series 8x GPU Server - 16x 2.5" NVMe + 8x SATA Drives Hot-Swappable bays. Built for AI Training and Inferencing.

Form Factor:
8U
Drive Bays:
Hot-Swap Drives
HDD Size:
2.5" Drives
Qty Drives:
24
Drive Interface:
SATA , NVMe, M.2
Server Processor:
AMD EPYC 9004 Series
Memory DIMMS:
24x 4800MHz
GPU Slots:
8x SXM GPU
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Full Height/Length Expansion, Redundant Power Supply - Standard
Max RAM Capacity:
3.1TB
Configure From: $585,929
Configure
NVIDIA DGX H100

NVIDIA DGX H100 with 8x NVIDIA H100 Tensor Core GPUs, Dual Intel® Xeon® Platinum 8480C Processors, 2TB Memory, 2x 1.92TB NVMe M.2 & 8x 3.84TB NVMe U.2.

Form Factor:
8U
Drive Bays:
Fixed Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
GPU Slots:
8x H100 Tensor Core GPUs
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Redundant Power Supply - Standard
Max RAM Capacity:
0GB
Configure From: $686,930
Configure
NVIDIA DGX H200

NVIDIA DGX H200 with 8x NVIDIA H200 141GB SXM5 GPU Server, Dual Intel® Xeon® Platinum Processors, 2TB DDR5 Memory, 2x 1.92TB NVMe M.2 & 8x 3.84TB NVMe SSDs.

Form Factor:
8U
Drive Bays:
Fixed Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
GPU Slots:
8x H200 Tensor Core GPUs
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Redundant Power Supply - Standard
Max RAM Capacity:
0GB
Configure From: $728,520
Configure
NVIDIA DGX B200

NVIDIA DGX B200 with 8x NVIDIA Blackwell GPUs, Dual Intel® Xeon® Platinum 8570 Processors, 4TB DDR5 Memory, 2x 1.92TB NVMe M.2 & 8x 3.84TB NVMe SSDs.

Form Factor:
8U
Drive Bays:
Fixed Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
GPU Slots:
8x NVIDIA Blackwell GPUs
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Redundant Power Supply - Standard
Max RAM Capacity:
0GB
Configure From: $987,870
Configure
NVIDIA DGX GB200

NVIDIA DGX GB200 with 72x NVIDIA Blackwell GPUs, Dual Intel® Xeon® Platinum Processors, 4TB DDR5 Memory, 2x 1.92TB NVMe M.2 & 8x 3.84TB NVMe SSDs.

Form Factor:
8U
Drive Bays:
Fixed Drives
HDD Size:
2.5" Drives
Qty Drives:
8
Drive Interface:
NVMe, M.2
Server Processor:
Intel Xeon Scalable Processor Gen 4/5
GPU Slots:
8x NVIDIA Blackwell GPUs
GPU Support:
Tesla GPU Optimised
Features:
High RAM Capacity, Redundant Power Supply - Standard
Max RAM Capacity:
0GB
Configure From: $15,879,550
Configure

Huge Educational and Research Discount

NVIDIA L4
NVIDIA RTX6000 ADA
NVIDIA L40S
NVIDIA H100
NVIDIA GH200
Application Virtualised Desktop, Graphical and Edge Applications High-end Design, Real-time Rendering, High-performance Compute Workflows Multi-modal Generative AI, and Graphics and Videos Workflows LLMs Inference, AI and Data Analytics Generative AI, LLMs Inference, and Memory Intensive Applications
Architecture Ada Lovelace Ada Lovelace Ada Lovelace Hopper Grace Hopper
SMs 60 142 142 114 144
CUDA Cores 7,424 18,176 18,176 18,432 18,432
Tensor Cores 240 568 568 640 576
Frequency 795 Mhz 915 MHz 1,110 Mhz 1,590 MHz 1,830 MHz
FP32 TFLOPs 30.3 - 91.6 51 67
FP16 TFLOPs 242 91.1 733 1,513 1,979
FP8 TFLOPs 485 - 1,466 3,026 3,958
Cache 48 MB 96 MB 48 MB 50 MB 60 MB
Max. Memory 24 GB 48 GB 48 GB 80 GB 512 GB
Memory B/W 300 GB/s 960 GB/s 864 GB/s 2,000 Gb/s 546 GB/s

The NVIDIA L4 Tensor Core GPU, built on the NVIDIA Ada Lovelace architecture, offers versatile and power-efficient acceleration across a wide range of applications, including video processing, AI, visual computing, graphics, virtualisation, and more. Available in a compact low-profile design, the L4 provides a cost-effective and energy-efficient solution, ensuring high throughput and minimal latency in servers spanning from edge devices to data centers and the cloud.

NVIDIA Tesla L4
Accelerate Workloads Efficiently and Sustainably

The NVIDIA L4 is an integral part of the NVIDIA data center platform. Engineered to support a wide range of applications such as AI, video processing, virtual workstations, graphics rendering, simulations, data science, and data analytics, this platform enhances the performance of more than 3,000 applications. It is accessible across various environments, spanning from data centers to edge computing to the cloud, offering substantial performance improvements and energy-efficient capabilities.

As AI and video technologies become more widespread, there's a growing need for efficient and affordable computing. NVIDIA L4 Tensor Core GPUs offer a substantial boost in AI video performance, up to 120 times better, resulting in a remarkable 99 percent improvement in energy efficiency and lower overall ownership costs when compared to traditional CPU-based systems. This enables businesses to reduce their server space requirements and significantly decrease their environmental impact, all while expanding their data centers to serve more users. Switching from CPUs to NVIDIA L4 GPUs in a 2-megawatt data center can save enough energy to power over 2,000 homes for a year or offset the carbon emissions equivalent to planting 172,000 trees over a decade.

Enterprise Ready: AI Software Streamlines Development and Deployment

As AI becomes commonplace in enterprises, organizations need comprehensive AI-ready infrastructure to prepare for the future. NVIDIA AI Enterprise is a complete cloud-native package of AI and data analytics software, designed to empower all organizations in excelling at AI. It's certified for deployment across various environments, including enterprise data centers and the cloud, and includes global enterprise support to ensure successful AI projects.

NVIDIA AI Enterprise is optimised to streamline AI development and deployment. It comes with tested open-source containers and frameworks, certified to work on standard data center hardware and popular NVIDIA-Certified Systems equipped with NVIDIA L4 Tensor Core GPUs. Plus, it includes support, providing organizations with the benefits of open source transparency and the reliability of global NVIDIA Enterprise Support, offering expertise for both AI practitioners and IT administrators.

NVIDIA AI Enterprise software is an extra license for NVIDIA L4 Tensor Core GPUs, making high-performance AI available to almost any organization for training, inference, and data science tasks. When combined with NVIDIA L4, it simplifies creating an AI-ready platform, speeds up AI development and deployment, and provides the performance, security, and scalability needed to gain insights quickly and realize business benefits sooner.


The new NVIDIA RTX 6000 Ada Generation delivers the features, capabilities, and performance needed to tackle the tasks of modern professional workflows. Using the latest NVIDIA Ada Lovelace GPU architecture, it comes with advanced features and powerful performance. With upgraded RT Cores, Tensor Cores, and CUDA cores, along with 48GB of graphics memory, it offers exceptional rendering, AI processing, graphics work, and computing power. Workstations powered by the NVIDIA RTX 6000 are equipped to help you thrive in today's tough business world..

NVIDIA RTX 6000 ADA
Powering the Next Era of Innovation

The NVIDIA RTX 6000 Ada Generation is the ultimate workstation graphics card designed for professionals who need top-notch performance and reliability to deliver their highest quality work and revolutionary advancements across industries. It offers unmatched performance and capabilities crucial for demanding tasks like high-end design, real-time rendering, AI, and high-performance computing.

Built on the NVIDIA Ada Lovelace architecture, the RTX 6000 merges 142 third-gen RT Cores, 568 fourth-gen Tensor Cores, and 18,176 CUDA cores with 48GB of ECC graphics memory. This combination powers the next era of AI graphics and petaflop inferencing performance, leading to remarkable acceleration in rendering, AI, graphics, and compute workloads.

NVIDIA RTX professional graphics cards are certified for use with an extensive range of professional applications. They've been tested by top independent software vendors (ISVs) and workstation makers and are supported by a global team of experts. With this reliable visual computing solution, you can focus on your important tasks without worries.


Unlock remarkable multi-tasking performance with the NVIDIA L40S GPU. This GPU merges potent AI computing with top-tier graphics and media speed, designed to fuel upcoming data center tasks. From advanced AI and language model processing to 3D graphics, rendering, and video tasks, the L40S GPU is primed for the next level of performance.

NVIDIA L40S
Fourth-Generation Tensor Cores

Hardware support for structural sparsity and optimised TF32 format offer immediate performance improvements for quicker AI and data science model training. Accelerate AI-enhanced graphics capabilities with DLSS to upscale resolution with improved performance in specific applications.

Third-Generation RT Cores

Improved throughput and concurrent ray-tracing and shading capabilities enhance ray-tracing performance, speeding up renders for product design, architecture, engineering, and construction tasks. Experience realistic designs with hardware-accelerated motion blur and impressive real-time animations.

Transformer Engine

The Transformer Engine significantly speeds up AI tasks and enhances memory usage for both training and inference. Using the Ada Lovelace fourth-generation Tensor Cores, Transformer Engine intelligently scans the layers of transformer architecture neural networks and automatically recasts between FP8 and FP16 precisions to boost AI performance and accelerate training and inference.

Data Center Ready

The L40S GPU is fine-tuned for 24/7 enterprise data center operations, and designed,built, tested, and supported by NVIDIA to guarantee top-notch performance, longevity, and operational stability. The L40S GPU meets the latest data center standards, is ready for Network Equipment-Building System (NEBS) Level 3 ready, and features secure boot with root of trust technology, enhancing data center security.

Video Transcoding Performance

The medium of online video is quite possibly the number one way of delivering information in the modern age. As we move forward into the future, the volume of online videos will only continue to grow exponentially. Simultaneously, the demand for answers to how to efficiently search and gain insights from video continues to grow.

T4 provides ground-breaking performance for AI video applications, featuring dedicated hardware transcoding engines which deliver 2x the decoding performance possible with previous-generation GPUs. T4 is able to decode up to nearly 40 full high definition video streams, making it simple to integrate scalable deep learning into video pipelines to provide inventive, smart video services.


Experience remarkable performance, scalability, and security for all tasks using the NVIDIA H100 Tensor Core GPU. The NVIDIA NVLink Switch System allows for connecting up to 256 H100 GPUs to boost exascale workloads. This GPU features a dedicated Transformer Engine to handle trillion-parameter language models. Thanks to these technological advancements, the H100 can accelerate large language models (LLMs) by an impressive 30X compared to the previous generation, establishing it as the leader in conversational AI.

NVIDIA H100
Ready for Enterprise AI?

NVIDIA H100 GPUs for regular servers include a five-year software subscription that encompasses enterprise support for the NVIDIA AI Enterprise software suite. This simplifies the process of adopting AI while ensuring top performance. It grants organizations access to essential AI frameworks and tools to create H100-accelerated AI applications, such as chatbots, recommendation engines, and vision AI. Take advantage of the NVIDIA AI Enterprise software subscription and its associated support for the NVIDIA H100.

Securely Accelerate Workloads From Enterprise to Exascale

The NVIDIA H100 GPUs feature fourth-generation Tensor Cores and the Transformer Engine with FP8 precision, solidifying NVIDIA's AI leadership by achieving up to 4X faster training and an impressive 30X speed boost for inference with large language models. In the realm of high-performance computing (HPC), the H100 triples the floating-point operations per second (FLOPS) for FP64 and introduces dynamic programming (DPX) instructions, resulting in a remarkable 7X performance increase. Equipped with the second-generation Multi-Instance GPU (MIG), built-in NVIDIA confidential computing, and the NVIDIA NVLink Switch System, the H100 provides secure acceleration for all workloads across data centers, ranging from enterprise to exascale.

Exponential Performance Leap with Pascal Architecture

The NVIDIA Pascal architecture enables the Tesla P100 to deliver superior performance for HPC and hyperscale workloads. With more than 21 teraflops of FP16 performance, Pascal is optimised to drive exciting new possibilities in deep learning applications. Pascal also delivers over 5 and 10 teraflops of double and single precision performance for HPC workloads.

Accelerate Every Workload, Everywhere

The NVIDIA H100 is a crucial component of the NVIDIA data center platform, designed to enhance AI, HPC, and data analytics. This platform accelerates more than 3,000 applications and is accessible across various locations, from data centers to edge computing, providing substantial performance improvements and cost-saving possibilities.


The NVIDIA GH200 Grace Hopper Superchip is a revolutionary high-speed CPU built specifically for massive AI and high-performance computing (HPC) tasks. This superchip boosts performance by up to 10 times for applications running terabytes of data, enabling scientists and researchers to find groundbreaking solutions to the toughest global challenges.

NVIDIA Grace Hopper GH100
The Worlds Most Versatile Computing Platform

The NVIDIA Grace Hopper architecture combines the innovative power of the NVIDIA Hopper GPU and the flexibility of the NVIDIA Grace CPU into one advanced superchip. This integration is facilitated by the NVIDIA NVLink Chip-2-Chip (C2C) interconnect, ensuring high-bandwidth and memory coherence between the two components. This unified architecture maximises performance and efficiency, enabling seamless collaboration between GPU and CPU for a wide range of computing tasks.

NVIDIA NVLink-C2C is a memory-coherent, high-bandwidth, and low-latencyinterconnect for superchips. At the core of the GH200 Grace Hopper Superchip, it provides up to 900 gigabytes per second (GB/s) of bandwidth, which is 7 times faster than PCIe Gen5 lanes commonly used in accelerated systems. NVLink-C2C allows applications to use both GPU and CPU memory efficiently. With up to 480GB of LPDDR5X CPU memory per GH200 Grace Hopper Superchip, the GPU has direct access to 7X more fast memory than HMB3 or almost 8X more fast memory with HBM3e. GH200 can be used in standard servers to run a variety of inference, data analytics,and other compute and memory-intensive workloads. GH200 can also be combined with the NVIDIA NVLink Switch System, with all GPU threads running on up to 256 NVLink-connected GPUs and able to access up to 144 terabytes (TB) of memory at high bandwidth.

Power and Efficiency With the Grace CPU

The NVIDIA Grace CPU offers twice the performance per watt compared to traditional x86-64 platforms and stands as the fastest Arm data center CPU worldwide. It's designed for high single-threaded performance, high- memory bandwidth, outstanding data-movement capabilities. The NVIDIA Grace CPU combines 72 Neoverse V2 Armv9 cores and up to 480GB of server-grade LPDDR5X memory with ECC, it achieves an optimal balance between bandwidth, energy efficiency, capacity, and cost. Compared to an eight-channel DDR5 design, the Grace CPU's LPDDR5X memory system delivers 53 percent more bandwidth while using only one-eighth the power per gigabyte per second.

Performance and Speed With the Hopper H100 GPU

The H100 Tensor Core GPU is NVIDIAs latest data center GPU, offering a significant performance boost for large-scale AI and HPC compared to the previous A100 Tensor Core GPU. Built on the new Hopper GPU architecture, the NVIDIA H100 introduces several innovations:

  • New fourth-generation Tensor Cores perform faster matrix computations thanever before, handling a wider range of AI and HPC tasks.
  • A new Transformer Engine enables H100 to deliver AI training speeds up to 9 times faster and AI inference speeds up to 30 times faster than the previous GPU generation.
  • Secure Multi-Instance GPU (MIG) patitions the GPU into separate, appropriately sized sections to enhance quality of service (QoS) for smaller workloads.
Class-Leading Performance for HPC and AI Workloads

The GH200 Grace Hopper Superchip marks the first genuine mixed accelerated platform tailored for HPC tasks. It boosts any application by leveraging the strengths of both GPUs and CPUs, all while offering the simplest and most efficient mixed programming approach yet. This allows scientists and engineers to concentrate on tackling the world's most pressing issues. For AI inference workloads, GH200 Grace Hopper Superchips combines with NVIDIA networking technologies to offer the most cost-effective scaling solutions, empowering users to handle larger datasets, more intricate models, and new tasks with access to up to 624GB of high-speed memory. For AI training, up to 256 NVLink-connected GPUs can access up to 144TB of memory at high bandwidthfor large language model (LLM) or recommender system training.


Broadberry GPU Servers harness the processing power of NVIDIA Ada Lovelace & Hopper graphics processing units for millions of applications such as image and video processing, computational biology and chemistry, fluid dynamics simulation, CT image reconstruction, seismic analysis, ray tracing, and much more.

As computing evolves, and processing moves from the CPU to co-processing between the CPU and GPU's, NVIDIA invented the CUDA parallel computing architecture to harness the performance benefits.

Speak to Broadberry GPU computing experts to find out more.


Call a Broadberry Storage & Server Specialist Now: +61 2 8551 2715

Have a Broadberry Expert Contact You:



Broadberry Celebrating Over 30 Years.


Engineer performing test.Our Rigorous Testing

Before leaving our workshop, all Broadberry server and storage solutions undergo a rigorous 48 hour testing procedure. This, along with the high-quality industry leading components ensures all of our server and storage solutions meet the strictest quality guidelines demanded from us.


Broadberry professional.Un-Equaled Flexibility

Our main objective is to offer great value, high-quality server and storage solutions, we understand that every company has different requirements and as such are able to offer un-equaled flexibility in designing custom server and storage solutions to meet our clients' needs.

Trusted by the World's Biggest Brands

We have established ourselves as one of the biggest storage providers in the UK, and since 1989 supplied our server and storage solutions to the world's biggest brands. Our customers include:

NASA, BBC, ITV, SONY, SKY, Disney, Google logos.