Server rental store

AI in Leatherhead

AI in Leatherhead: Server Configuration Documentation

Welcome to the documentation for the “AI in Leatherhead” server cluster. This document details the hardware and software configuration of our dedicated AI processing environment, intended for users new to the system. This guide will cover the core components, network setup, and software stack. Please read carefully before attempting any modifications or deployments. This server cluster supports a variety of machine learning tasks, including Natural Language Processing, Computer Vision, and Predictive Analytics.

Overview

The “AI in Leatherhead” cluster is a high-performance computing (HPC) environment designed to accelerate AI and machine learning workflows. It comprises multiple interconnected servers, a dedicated network, and a shared storage system. The primary goal of this setup is to provide a scalable and reliable platform for researchers and developers. The system leverages GPU acceleration for computationally intensive tasks.

Hardware Configuration

The cluster consists of three primary server types: Master Nodes, Compute Nodes, and Storage Nodes. Details of each are provided below.

Master Nodes

The Master Nodes manage the cluster, schedule jobs, and monitor resource utilization. We currently have two Master Nodes for redundancy.

Specification Value
CPU Dual Intel Xeon Gold 6338
RAM 256 GB DDR4 ECC Registered
Storage (OS) 1 TB NVMe SSD
Network Interface Dual 100 Gbps InfiniBand

Compute Nodes

The Compute Nodes perform the actual AI/ML computations. We have eight Compute Nodes currently deployed.

Specification Value
CPU Dual AMD EPYC 7763
RAM 512 GB DDR4 ECC Registered
GPU 4x NVIDIA A100 (80GB)
Storage (Local) 2 TB NVMe SSD (for temporary data)
Network Interface Dual 200 Gbps InfiniBand

Storage Nodes

The Storage Nodes provide a shared file system accessible to all nodes in the cluster. This is critical for data-intensive AI workloads.

Specification Value
CPU Intel Xeon Silver 4310
RAM 128 GB DDR4 ECC Registered
Storage (Raw) 2 PB NVMe-oF Array (Redundant)
Network Interface Dual 100 Gbps Ethernet

Network Configuration

The cluster utilizes a dedicated network to minimize latency and maximize bandwidth. The network is segmented into three main parts: Management, InfiniBand, and Ethernet.

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️