Server rental store

AI in Crawley

# AI in Crawley: Server Configuration Documentation

This document details the server configuration supporting the "AI in Crawley" project, designed to provide a resource for new team members and maintainers. This system utilizes a distributed architecture to handle the computational demands of machine learning models processing real-time data from the Crawley area.

Overview

The "AI in Crawley" project involves analyzing various data streams – CCTV footage, traffic patterns, environmental sensors, and social media feeds – to provide insights into city management, public safety, and resource allocation. The server infrastructure is built to be scalable, reliable, and secure. The core components are a cluster of GPU servers for model training and inference, a data lake for storage, and a message queue for data ingestion. The system is monitored using Prometheus and Grafana, with logging handled by the ELK stack. This documentation will cover the hardware, software, and network configuration of these key components.

Hardware Configuration

The core of the system is comprised of four primary server types: Data Ingestion Servers, GPU Training Servers, Inference Servers, and the Central Database Server. Each has specific hardware requirements outlined below.

Data Ingestion Servers

These servers are responsible for receiving and pre-processing data from various sources. They are relatively lightweight in terms of computational needs but require high network bandwidth.

Component Specification
CPU Intel Xeon Silver 4310 (12 cores)
RAM 64 GB DDR4 ECC
Storage 2 x 1 TB NVMe SSD (RAID 1)
Network Interface 10 Gbps Ethernet
Operating System Ubuntu Server 22.04 LTS

GPU Training Servers

These servers are the workhorses for training our machine learning models. They require powerful GPUs and significant RAM. We currently have three of these servers in operation.

Component Specification
CPU AMD EPYC 7763 (64 cores)
RAM 256 GB DDR4 ECC
GPU 4 x NVIDIA A100 (80GB)
Storage 4 x 4 TB NVMe SSD (RAID 0)
Network Interface 100 Gbps InfiniBand
Operating System CentOS Stream 9

Inference Servers

These servers deploy trained models to perform real-time predictions. They need to be highly responsive and efficient. We deploy these servers using Docker containers.

Component Specification
CPU Intel Xeon Gold 6338 (32 cores)
RAM 128 GB DDR4 ECC
GPU 2 x NVIDIA RTX A4000 (16GB)
Storage 1 x 2 TB NVMe SSD
Network Interface 25 Gbps Ethernet
Operating System Ubuntu Server 22.04 LTS

Software Configuration

The software stack is built around open-source technologies, prioritizing flexibility and cost-effectiveness.

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️