AI in Brighton and Hove

From Server rental store
Revision as of 04:47, 16 April 2025 by Admin (talk | contribs) (Automated server configuration article)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search
  1. AI in Brighton and Hove: Server Configuration Overview

This article details the server configuration supporting the "AI in Brighton and Hove" initiative. It's aimed at newcomers to this wiki and provides a technical overview of the hardware and software used. This system is designed to handle large datasets related to local AI projects and provide a platform for model training and deployment. We will cover hardware specifications, software stack, and networking details. It is vital to understand the system architecture before contributing to or modifying the infrastructure. Please refer to the System Administration Guidelines for general wiki policies.

Hardware Infrastructure

The core infrastructure consists of three primary server nodes, each serving a specific role. These are located in a secure data center facility within Brighton and Hove. Detailed specifications are provided below. Consider reviewing the Data Center Security Protocols before physical access is requested.

Server Node Role CPU RAM Storage
Node 1 (Athena) Primary Data Storage & Pre-processing Dual Intel Xeon Gold 6338 256GB DDR4 ECC 6 x 18TB SAS HDDs (RAID 6) + 2 x 1TB NVMe SSDs (OS/Cache)
Node 2 (Hermes) Model Training & Development AMD EPYC 7763 512GB DDR4 ECC 4 x 8TB SAS HDDs (RAID 10) + 2 x 2TB NVMe SSDs (OS/Cache) + 4 x NVIDIA RTX A6000 GPUs
Node 3 (Zeus) Model Deployment & API Server Intel Xeon Silver 4310 128GB DDR4 ECC 2 x 4TB SAS HDDs (RAID 1) + 1 x 1TB NVMe SSD (OS/Cache)

All servers are connected via a dedicated 10GbE network. Power redundancy is provided by dual power supplies and an uninterruptible power supply (UPS). Refer to the Power Management Documentation for further details. We also utilise a dedicated Backup and Disaster Recovery system.

Software Stack

The software stack is built around a Linux operating system, providing a stable and flexible platform. We utilize Debian 11 (Bullseye) as the base OS. The following components are essential to the system's operation.

Component Version Purpose
Operating System Debian 11 (Bullseye) Base OS
Python 3.9 Primary programming language for AI/ML tasks
TensorFlow 2.8.0 Deep learning framework
PyTorch 1.10.0 Deep learning framework
Docker 20.10.7 Containerization platform
Kubernetes 1.23.4 Container orchestration
PostgreSQL 14 Database for metadata and results

All code is version controlled using Git and hosted on a private GitLab instance. Software updates are managed through a controlled release process detailed in the Software Release Management document. Security patches are applied immediately as per the Security Incident Response Plan.

Networking Configuration

The network is segmented to enhance security and performance. Each server node has a static IP address within the 192.168.1.0/24 subnet. A dedicated VLAN is used for communication between the server nodes and the external API gateway.

Server Node IP Address VLAN Firewall Rules
Node 1 (Athena) 192.168.1.10 10 Allows inbound connections from Node 2 & Node 3 on specific ports (PostgreSQL, SSH)
Node 2 (Hermes) 192.168.1.11 10 Allows inbound connections from Node 1 & Node 3 on specific ports (TensorBoard, SSH)
Node 3 (Zeus) 192.168.1.12 20 Allows inbound connections from external API gateway on port 80/443

Firewall rules are managed using `iptables` and are regularly reviewed and updated. See the Network Security Policy for more information. Internal DNS resolution is provided by a local bind9 server. We also utilize Intrusion Detection Systems for security monitoring. Further details on network topology can be found in the Network Diagram.


Future Considerations

We are evaluating the integration of a distributed storage system, such as Ceph, to improve scalability and resilience. We are also exploring the use of more powerful GPUs to accelerate model training. The Roadmap for AI Infrastructure details these planned upgrades. Consider the System Monitoring Procedures for operational stability.



Main Page Server Administration Data Storage Policies Security Best Practices Performance Tuning Guide Troubleshooting Guide Git Workflow Database Administration API Documentation Monitoring Tools Patch Management Incident Reporting Capacity Planning User Access Control System Logs Contact Information Change Management Process Virtualization Strategy


Intel-Based Server Configurations

Configuration Specifications Benchmark
Core i7-6700K/7700 Server 64 GB DDR4, NVMe SSD 2 x 512 GB CPU Benchmark: 8046
Core i7-8700 Server 64 GB DDR4, NVMe SSD 2x1 TB CPU Benchmark: 13124
Core i9-9900K Server 128 GB DDR4, NVMe SSD 2 x 1 TB CPU Benchmark: 49969
Core i9-13900 Server (64GB) 64 GB RAM, 2x2 TB NVMe SSD
Core i9-13900 Server (128GB) 128 GB RAM, 2x2 TB NVMe SSD
Core i5-13500 Server (64GB) 64 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Server (128GB) 128 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Workstation 64 GB DDR5 RAM, 2 NVMe SSD, NVIDIA RTX 4000

AMD-Based Server Configurations

Configuration Specifications Benchmark
Ryzen 5 3600 Server 64 GB RAM, 2x480 GB NVMe CPU Benchmark: 17849
Ryzen 7 7700 Server 64 GB DDR5 RAM, 2x1 TB NVMe CPU Benchmark: 35224
Ryzen 9 5950X Server 128 GB RAM, 2x4 TB NVMe CPU Benchmark: 46045
Ryzen 9 7950X Server 128 GB DDR5 ECC, 2x2 TB NVMe CPU Benchmark: 63561
EPYC 7502P Server (128GB/1TB) 128 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/2TB) 128 GB RAM, 2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/4TB) 128 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/1TB) 256 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/4TB) 256 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 9454P Server 256 GB RAM, 2x2 TB NVMe

Order Your Dedicated Server

Configure and order your ideal server configuration

Need Assistance?

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️