Server rental store

AI in the Yellow River

AI in the Yellow River: Server Configuration Documentation

Welcome to the documentation detailing the server configuration for the "AI in the Yellow River" project. This document is designed for newcomers to our MediaWiki site and provides a detailed overview of the hardware and software components powering this initiative. This project focuses on utilizing artificial intelligence to monitor and predict flooding events within the Yellow River basin. Understanding the server infrastructure is crucial for contributing effectively.

Project Overview

The "AI in the Yellow River" project utilizes a distributed server architecture to process massive datasets from various sources, including hydrological sensors, weather stations, and satellite imagery. The primary goal is real-time analysis and accurate flood prediction. Data ingestion, model training, and prediction services are all handled by the server cluster. This necessitates a robust and scalable infrastructure. See also Data Acquisition Process for information on data sources.

Hardware Specifications

Our server cluster consists of three primary node types: Master Nodes, Worker Nodes, and Storage Nodes. Each node type is specialized to perform certain functions. Details are provided below. Refer to Network Topology for a diagram of the network connections.

Master Nodes

Master Nodes manage the cluster, schedule tasks, and monitor the health of other nodes. They are also responsible for initial data distribution. We currently have two Master Nodes for redundancy.

Specification Value
CPU Dual Intel Xeon Gold 6248R (24 cores/48 threads per CPU)
RAM 256 GB DDR4 ECC Registered
Storage (OS) 1 TB NVMe SSD
Network Interface Dual 100 Gbps Ethernet
Operating System CentOS Linux 7

Worker Nodes

Worker Nodes perform the computationally intensive tasks of model training and prediction. These nodes are equipped with powerful GPUs. We have eight Worker Nodes currently deployed. For detailed information on the AI models used, see AI Model Descriptions.

Specification Value
CPU Dual Intel Xeon Silver 4210 (10 cores/20 threads per CPU)
RAM 128 GB DDR4 ECC Registered
GPU 4 x NVIDIA A100 (80GB VRAM each)
Storage (Data) 4 TB NVMe SSD
Network Interface Dual 100 Gbps Ethernet
Operating System Ubuntu Server 20.04

Storage Nodes

Storage Nodes provide persistent storage for the massive datasets used by the project. They are optimized for high throughput and data integrity. We have four Storage Nodes in operation. See Data Backup Strategy for details on data protection.

Specification Value
CPU Intel Xeon E-2224 (6 cores/12 threads)
RAM 64 GB DDR4 ECC Registered
Storage 16 x 16 TB Enterprise SAS HDD (RAID 6)
Network Interface Dual 25 Gbps Ethernet
Operating System FreeBSD 13

Software Stack

The software stack is carefully chosen to support the AI workloads and ensure efficient data processing. See Software Dependencies for a complete list of required packages.

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️