AI in the Palau Rainforest

From Server rental store
Jump to navigation Jump to search

AI in the Palau Rainforest: Server Configuration

This document details the server configuration for the "AI in the Palau Rainforest" project, outlining the hardware, software, and network setup used to support the real-time data processing and analysis of sensor data collected within the Palau rainforest environment. This guide is intended for new system administrators and engineers joining the project. Understanding this setup is crucial for maintenance, troubleshooting, and future scalability. See also System Administration Guide for general server management procedures.

Project Overview

The "AI in the Palau Rainforest" project utilizes a network of sensor nodes deployed throughout the rainforest to collect data on biodiversity, climate conditions, and ecosystem health. This data is streamed back to a central server cluster for processing using machine learning algorithms. The primary goals of the AI processing are species identification from audio recordings, anomaly detection in temperature and humidity data, and predictive modeling of rainforest health. Refer to the Project Goals page for more details. Data storage and access are governed by the Data Management Policy.

Hardware Configuration

The server cluster consists of three primary server types: Data Intake Servers, Processing Servers, and Database Servers. Each type is described below with detailed specifications.

Data Intake Servers

These servers are responsible for receiving data streams from the sensor nodes. They perform initial data validation and buffering before forwarding data to the processing servers. Two Data Intake Servers are currently in operation for redundancy.

Specification Value
Server Model Dell PowerEdge R750
CPU 2 x Intel Xeon Gold 6338 (32 cores/64 threads per CPU)
RAM 128 GB DDR4 ECC Registered
Storage 2 x 1TB NVMe SSD (RAID 1) for OS and buffering
Network Interface 2 x 10 GbE Ethernet
Operating System Ubuntu Server 22.04 LTS

These servers utilize a custom data intake script written in Python to manage the incoming data streams. See Data Intake Script Documentation for details.

Processing Servers

These servers execute the machine learning algorithms and perform data analysis. We currently have four processing servers, optimized for GPU-accelerated computation.

Specification Value
Server Model Supermicro SYS-220M-CT
CPU 2 x AMD EPYC 7763 (64 cores/128 threads per CPU)
RAM 256 GB DDR4 ECC Registered
GPU 4 x NVIDIA A100 (40GB)
Storage 1 x 2TB NVMe SSD (OS) + 4 x 4TB SATA HDD (Data Storage)
Network Interface 2 x 10 GbE Ethernet
Operating System CentOS Stream 9

Machine learning models are developed and deployed using TensorFlow and PyTorch. See Model Deployment Guide for instructions.

Database Servers

These servers store processed data, metadata, and model outputs. We employ a clustered PostgreSQL database for high availability and scalability.

Specification Value
Server Model HP ProLiant DL380 Gen10
CPU 2 x Intel Xeon Silver 4310 (12 cores/24 threads per CPU)
RAM 128 GB DDR4 ECC Registered
Storage 8 x 4TB SAS HDD (RAID 6)
Network Interface 2 x 10 GbE Ethernet
Operating System Debian 11
Database PostgreSQL 14

Database backups are performed nightly and stored offsite. Refer to the Backup and Recovery Plan for details. Data schemas are detailed in the Database Schema Documentation.


Software Configuration

All servers utilize a centralized logging system based on Elasticsearch, Logstash, and Kibana (the ELK stack). This allows for efficient monitoring and troubleshooting. Servers are provisioned using Ansible for automated configuration management. The version control system used for all code and configuration files is Git, hosted on a private GitLab instance. Security is managed using Firewall Configuration and regular vulnerability scanning.

Network Configuration

The server cluster is located within a dedicated VLAN on the research network. Each server has a static IP address. A load balancer distributes traffic to the Data Intake Servers. The network topology is illustrated in the Network Diagram. All communication between servers is encrypted using TLS. Network monitoring is performed using Nagios.

Future Considerations

Future plans include upgrading the Processing Server GPUs to the latest generation and exploring the use of a distributed file system like Hadoop for larger datasets. We are also investigating the integration of a real-time data visualization dashboard using Grafana. Further details about these plans can be found in the Future Development Roadmap.


Intel-Based Server Configurations

Configuration Specifications Benchmark
Core i7-6700K/7700 Server 64 GB DDR4, NVMe SSD 2 x 512 GB CPU Benchmark: 8046
Core i7-8700 Server 64 GB DDR4, NVMe SSD 2x1 TB CPU Benchmark: 13124
Core i9-9900K Server 128 GB DDR4, NVMe SSD 2 x 1 TB CPU Benchmark: 49969
Core i9-13900 Server (64GB) 64 GB RAM, 2x2 TB NVMe SSD
Core i9-13900 Server (128GB) 128 GB RAM, 2x2 TB NVMe SSD
Core i5-13500 Server (64GB) 64 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Server (128GB) 128 GB RAM, 2x500 GB NVMe SSD
Core i5-13500 Workstation 64 GB DDR5 RAM, 2 NVMe SSD, NVIDIA RTX 4000

AMD-Based Server Configurations

Configuration Specifications Benchmark
Ryzen 5 3600 Server 64 GB RAM, 2x480 GB NVMe CPU Benchmark: 17849
Ryzen 7 7700 Server 64 GB DDR5 RAM, 2x1 TB NVMe CPU Benchmark: 35224
Ryzen 9 5950X Server 128 GB RAM, 2x4 TB NVMe CPU Benchmark: 46045
Ryzen 9 7950X Server 128 GB DDR5 ECC, 2x2 TB NVMe CPU Benchmark: 63561
EPYC 7502P Server (128GB/1TB) 128 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/2TB) 128 GB RAM, 2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (128GB/4TB) 128 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/1TB) 256 GB RAM, 1 TB NVMe CPU Benchmark: 48021
EPYC 7502P Server (256GB/4TB) 256 GB RAM, 2x2 TB NVMe CPU Benchmark: 48021
EPYC 9454P Server 256 GB RAM, 2x2 TB NVMe

Order Your Dedicated Server

Configure and order your ideal server configuration

Need Assistance?

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️