AI in Norwich

From Server rental store
Revision as of 07:25, 16 April 2025 by Admin (talk | contribs) (Automated server configuration article)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search
  1. AI in Norwich: Server Configuration

This document details the server configuration for the "AI in Norwich" project, a local initiative dedicated to exploring the applications of Artificial Intelligence within the city. This guide is aimed at new contributors to the wiki and provides a comprehensive overview of the hardware and software employed. Please read carefully before making any modifications.

Overview

The “AI in Norwich” project utilizes a cluster of servers located within the Norwich Research Park data centre. The servers are primarily used for machine learning model training, data analysis, and hosting web-based AI applications. The system is designed for scalability and redundancy, allowing for future growth and minimizing downtime. This configuration focuses on the core infrastructure. Additional information regarding Data Security Protocols and Network Topology can be found in separate articles.

Hardware Configuration

The cluster consists of three primary server nodes, designated as Node-A, Node-B, and Node-C. Each node is built with similar specifications to ensure consistency and simplify maintenance. A dedicated storage server handles data persistence.

Server Node CPU RAM Storage Network Interface
Node-A Intel Xeon Gold 6248R (24 cores) 256 GB DDR4 ECC Registered 2 x 4TB NVMe SSD (RAID 1) 10 Gigabit Ethernet
Node-B Intel Xeon Gold 6248R (24 cores) 256 GB DDR4 ECC Registered 2 x 4TB NVMe SSD (RAID 1) 10 Gigabit Ethernet
Node-C Intel Xeon Gold 6248R (24 cores) 256 GB DDR4 ECC Registered 2 x 4TB NVMe SSD (RAID 1) 10 Gigabit Ethernet

The storage server, designated 'Storage-1', provides centralized storage for all nodes.

Component Specification
Host Name Storage-1
CPU Intel Xeon Silver 4210 (10 cores)
RAM 64 GB DDR4 ECC Registered
Storage 8 x 16TB SAS HDD (RAID 6)
Network Interface 10 Gigabit Ethernet

Power is supplied via redundant power supplies and a dedicated UPS system, detailed in the Power Management Documentation.

Software Configuration

All server nodes run Ubuntu Server 22.04 LTS. The software stack is designed to facilitate machine learning and data science workflows. The primary software components include Python 3.10, TensorFlow 2.12, PyTorch 2.0, and JupyterLab. A distributed file system, GlusterFS, is used to provide a unified namespace across the cluster.

Software Component Version Purpose
Operating System Ubuntu Server 22.04 LTS Server Operating System
Python 3.10.6 Primary Programming Language
TensorFlow 2.12.0 Machine Learning Framework
PyTorch 2.0.1 Machine Learning Framework
JupyterLab 3.5.0 Interactive Development Environment
GlusterFS 10.1 Distributed File System

Network Configuration

The servers are connected via a dedicated 10 Gigabit Ethernet network. Internal DNS resolution is managed by a local BIND server. Firewall rules are configured using `ufw` to restrict access to essential services. Refer to the Network Security Policy for detailed information. Each node is assigned a static IP address within the 192.168.1.0/24 subnet.

  • Node-A: 192.168.1.10
  • Node-B: 192.168.1.11
  • Node-C: 192.168.1.12
  • Storage-1: 192.168.1.20

The cluster utilizes a load balancer, configured with HAProxy, to distribute traffic to the active nodes.

Monitoring and Logging

Server health and performance are monitored using Prometheus and Grafana. Logs are collected and centralized using the ELK Stack (Elasticsearch, Logstash, Kibana). Alerts are configured to notify administrators of critical issues. Detailed logging configurations can be found in the Logging Standards Guide.

Future Considerations

Planned upgrades include migrating the storage server to NVMe SSDs for improved performance and exploring the use of containerization technologies like Docker and Kubernetes for application deployment. We are also investigating the addition of a dedicated GPU server for accelerated machine learning workloads.


Main Page Server Maintenance Procedures Troubleshooting Guide Data Backup Strategy User Access Control Security Audit Logs Software Update Schedule Hardware Inventory Contact Information Change Management Process Disaster Recovery Plan Incident Response Protocol Network Diagrams Virtualization Platform API Documentation


Intel-Based Server Configurations

Configuration Specifications Benchmark
Core i7-6700K/7700 Server 64 GB DDR4, NVMe SSD 2 x 512 GB CPU Benchmark: 8046
Core i7-8700 Server 64 GB DDR4, NVMe SSD 2x1 TB CPU Benchmark: 13124
Core i9-9900K Server 128 GB DDR4, NVMe SSD 2 x 1 TB CPU Benchmark: 49969
Core i9-13900 Server (64GB) 64 GB RAM, 2x2 TB NVMe SSD
Core i9-13900 Server (128GB) 128 GB RAM, 2x2 TB NVMe SSD
Core i5-13500 Server (64GB) 64 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Server (128GB) 128 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Workstation 64 GB DDR5 RAM, 2 NVMe SSD, NVIDIA RTX 4000

AMD-Based Server Configurations

Configuration Specifications Benchmark
Ryzen 5 3600 Server 64 GB RAM, 2x480 GB NVMe CPU Benchmark: 17849
Ryzen 7 7700 Server 64 GB DDR5 RAM, 2x1 TB NVMe CPU Benchmark: 35224
Ryzen 9 5950X Server 128 GB RAM, 2x4 TB NVMe CPU Benchmark: 46045
Ryzen 9 7950X Server 128 GB DDR5 ECC, 2x2 TB NVMe CPU Benchmark: 63561
EPYC 7502P Server (128GB/1TB) 128 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/2TB) 128 GB RAM, 2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/4TB) 128 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/1TB) 256 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/4TB) 256 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 9454P Server 256 GB RAM, 2x2 TB NVMe

Order Your Dedicated Server

Configure and order your ideal server configuration

Need Assistance?

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️