AI in Woking

From Server rental store
Jump to navigation Jump to search
  1. AI in Woking: Server Configuration and Deployment

This article details the server configuration used to support the "AI in Woking" project, a local initiative exploring applications of Artificial Intelligence within the borough. This document is intended for new system administrators and developers contributing to the project. It outlines the hardware, software, and network setup necessary for successful operation. Please review this document thoroughly before making any changes to the production environment. Refer to the System Administration Manual for general site policies.

Project Overview

The "AI in Woking" project involves several key components: data collection from local sources (e.g., traffic sensors, environmental monitors), model training using a cluster of GPU servers, and a web-based interface for accessing AI-powered insights. The project relies heavily on Python for scripting and data processing, and TensorFlow and PyTorch for machine learning tasks. Data storage is managed using PostgreSQL, and the web interface is built with PHP. Security is paramount; see Security Policies for detailed guidance.

Hardware Configuration

The server infrastructure is hosted in a dedicated rack at a local data center. The primary server roles are divided across several physical machines. The entire infrastructure is monitored using Nagios.

Primary Server Specs

Server Role Model CPU RAM Storage Network Interface
Data Collection Server Dell PowerEdge R750 Intel Xeon Gold 6338 128 GB DDR4 2 x 1 TB NVMe SSD (RAID 1) 10 GbE
Model Training Cluster (Node 1-4) Supermicro SYS-2029U-TR4 AMD EPYC 7763 256 GB DDR4 4 x 4 TB SATA HDD (RAID 5) + 1 x 512GB NVMe SSD 25 GbE
Web Server HP ProLiant DL380 Gen10 Intel Xeon Silver 4210 64 GB DDR4 2 x 480 GB SATA SSD (RAID 1) 1 GbE
Database Server Dell PowerEdge R650 Intel Xeon Gold 6330 64 GB DDR4 4 x 2 TB SATA HDD (RAID 10) 10 GbE

GPU Specifications (Model Training Cluster)

Each node in the model training cluster is equipped with four NVIDIA A100 GPUs.

GPU Model Memory CUDA Cores Tensor Cores Power Consumption
NVIDIA A100 80 GB HBM2e 6912 432 400W

Software Configuration

The operating system across all servers is Ubuntu Server 22.04 LTS. Specific software packages and versions are detailed below. Configuration management is handled using Ansible.

Software Stack

Server Role Operating System Key Software Versions
Data Collection Server Ubuntu Server 22.04 LTS Python, RabbitMQ, InfluxDB 3.10, 3.9, 2.0
Model Training Cluster Ubuntu Server 22.04 LTS Python, TensorFlow, PyTorch, CUDA, NCCL 3.10, 2.12, 2.0, 12.1, 2.12
Web Server Ubuntu Server 22.04 LTS PHP, Apache, MariaDB Client 8.1, 2.4, 10.6
Database Server Ubuntu Server 22.04 LTS PostgreSQL, pgAdmin 14, 4

Network Configuration

The servers are connected to the data center network via a dedicated VLAN. Firewall rules are configured using iptables to restrict access to only necessary ports. A reverse proxy, Nginx, is used on the web server to handle SSL termination and load balancing. Internal DNS is managed using Bind9.

  • **VLAN ID:** 100
  • **Subnet:** 192.168.100.0/24
  • **Gateway:** 192.168.100.1
  • **DNS Servers:** 192.168.100.2, 8.8.8.8

Security Considerations

All servers are regularly patched and updated. Access to the servers is restricted to authorized personnel only, using SSH keys and strong passwords. Regular security audits are conducted. See the Incident Response Plan for details on handling security breaches. Data encryption is employed both in transit and at rest.

Future Expansion

Planned future expansion includes adding more GPUs to the model training cluster and implementing a distributed caching layer using Redis. We also plan to investigate the use of Kubernetes for container orchestration.

Main Page Help:Contents MediaWiki FAQ Manual:Configuration Manual:Installation


Intel-Based Server Configurations

Configuration Specifications Benchmark
Core i7-6700K/7700 Server 64 GB DDR4, NVMe SSD 2 x 512 GB CPU Benchmark: 8046
Core i7-8700 Server 64 GB DDR4, NVMe SSD 2x1 TB CPU Benchmark: 13124
Core i9-9900K Server 128 GB DDR4, NVMe SSD 2 x 1 TB CPU Benchmark: 49969
Core i9-13900 Server (64GB) 64 GB RAM, 2x2 TB NVMe SSD
Core i9-13900 Server (128GB) 128 GB RAM, 2x2 TB NVMe SSD
Core i5-13500 Server (64GB) 64 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Server (128GB) 128 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Workstation 64 GB DDR5 RAM, 2 NVMe SSD, NVIDIA RTX 4000

AMD-Based Server Configurations

Configuration Specifications Benchmark
Ryzen 5 3600 Server 64 GB RAM, 2x480 GB NVMe CPU Benchmark: 17849
Ryzen 7 7700 Server 64 GB DDR5 RAM, 2x1 TB NVMe CPU Benchmark: 35224
Ryzen 9 5950X Server 128 GB RAM, 2x4 TB NVMe CPU Benchmark: 46045
Ryzen 9 7950X Server 128 GB DDR5 ECC, 2x2 TB NVMe CPU Benchmark: 63561
EPYC 7502P Server (128GB/1TB) 128 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/2TB) 128 GB RAM, 2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/4TB) 128 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/1TB) 256 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/4TB) 256 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 9454P Server 256 GB RAM, 2x2 TB NVMe

Order Your Dedicated Server

Configure and order your ideal server configuration

Need Assistance?

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️