AI in the Aruba Rainforest
- AI in the Aruba Rainforest: Server Configuration
This article details the server configuration powering the "AI in the Aruba Rainforest" project, a research initiative utilizing artificial intelligence to monitor and analyze biodiversity within the Aruba rainforest ecosystem. This documentation is intended for new team members and system administrators responsible for maintaining the infrastructure.
Introduction
The AI in the Aruba Rainforest project relies on a distributed server system to process data collected from a network of sensors deployed throughout the rainforest. These sensors generate a continuous stream of data – audio recordings, thermal imagery, and environmental readings – which are analyzed in real-time by machine learning algorithms. The server infrastructure is designed for high availability, scalability, and data integrity. This document outlines the hardware and software components, the network topology, and the security considerations of the system. See also Data Acquisition Protocols for details on the data sources.
Hardware Overview
The core server infrastructure consists of three primary server nodes, designated as the 'Ingestion Node', the 'Processing Node', and the 'Storage Node'. Each node is housed in a climate-controlled server room at the Aruba National Park headquarters. Redundancy is built into the system through hardware replication and failover mechanisms. Refer to Disaster Recovery Plan for further information.
Ingestion Node
The Ingestion Node is responsible for receiving data from the sensor network.
Component | Specification |
---|---|
CPU | Intel Xeon Gold 6248R (24 cores, 3.0 GHz) |
RAM | 128 GB DDR4 ECC Registered |
Network Interface | Dual 10 Gigabit Ethernet |
Storage | 2 x 1 TB NVMe SSD (RAID 1) – for temporary buffering |
Operating System | Ubuntu Server 22.04 LTS |
Processing Node
The Processing Node performs the machine learning analysis on the incoming data. This node requires significant computational power.
Component | Specification |
---|---|
CPU | Dual Intel Xeon Platinum 8380 (40 cores each, 2.3 GHz) |
RAM | 256 GB DDR4 ECC Registered |
GPU | 4 x NVIDIA A100 (80GB VRAM each) |
Network Interface | Dual 10 Gigabit Ethernet |
Storage | 4 x 2 TB NVMe SSD (RAID 0) – for fast data access |
Operating System | Ubuntu Server 22.04 LTS |
Storage Node
The Storage Node provides long-term data storage and archival.
Component | Specification |
---|---|
CPU | Intel Xeon Silver 4310 (12 cores, 2.1 GHz) |
RAM | 64 GB DDR4 ECC Registered |
Network Interface | Dual 10 Gigabit Ethernet |
Storage | 60 TB RAID 6 array (using 18 x 4TB SAS HDDs) |
Operating System | CentOS Stream 9 |
Software Stack
The software stack is designed for efficient data processing and analysis. The core components are detailed below. See Software Dependencies for a complete list of required packages.
- Operating Systems: As detailed above, Ubuntu Server 22.04 LTS and CentOS Stream 9 are used.
- Programming Languages: Python 3.9 is the primary programming language, used for developing the machine learning algorithms and data processing pipelines. R is used for statistical analysis.
- Machine Learning Frameworks: TensorFlow and PyTorch are used for building and training the AI models.
- Data Storage: PostgreSQL is used as the primary database for storing metadata and analysis results. The 60TB RAID array on the Storage Node stores the raw sensor data.
- Message Queue: RabbitMQ is used for asynchronous communication between the Ingestion Node, Processing Node, and Storage Node. See Message Queue Architecture for details.
- Monitoring: Prometheus and Grafana are used for system monitoring and alerting. Monitoring Dashboard Setup provides instructions.
- Containerization: Docker is used to containerize the machine learning applications, ensuring portability and reproducibility. Docker Deployment Guide provides detailed instructions.
Network Configuration
The server nodes are connected via a dedicated 10 Gigabit Ethernet network. A firewall protects the network from unauthorized access. The network topology is a star configuration, with a central switch connecting all three server nodes. The switch is a Cisco Catalyst 9300 Series. See Network Diagram for a visual representation. All communication between nodes is encrypted using TLS. The network is segmented to isolate the server infrastructure from the public internet.
Security Considerations
Security is paramount, given the sensitivity of the collected data. Several measures are in place to protect the system:
- Firewall: A robust firewall (iptables) is configured to restrict access to the server nodes.
- Intrusion Detection System (IDS): Snort is deployed to detect and prevent malicious activity.
- Regular Security Audits: Penetration testing is conducted quarterly to identify and address vulnerabilities. See Security Audit Reports.
- Data Encryption: All data is encrypted both in transit and at rest.
- Access Control: Strict access control policies are enforced, limiting access to the server nodes to authorized personnel only. Use SSH Key Management for secure access.
- Regular Software Updates: All software is kept up to date with the latest security patches.
Future Enhancements
Planned future enhancements include:
- Implementing a distributed file system (e.g., Ceph) for improved scalability and data redundancy.
- Adding a dedicated GPU server for real-time video processing.
- Integrating with a cloud-based data analytics platform for advanced analysis.
Server Maintenance Schedule Contact Information Frequently Asked Questions Data Privacy Policy Troubleshooting Guide Backup and Restore Procedures Glossary of Terms Related Projects Project Documentation Index Change Management Process
Intel-Based Server Configurations
Configuration | Specifications | Benchmark |
---|---|---|
Core i7-6700K/7700 Server | 64 GB DDR4, NVMe SSD 2 x 512 GB | CPU Benchmark: 8046 |
Core i7-8700 Server | 64 GB DDR4, NVMe SSD 2x1 TB | CPU Benchmark: 13124 |
Core i9-9900K Server | 128 GB DDR4, NVMe SSD 2 x 1 TB | CPU Benchmark: 49969 |
Core i9-13900 Server (64GB) | 64 GB RAM, 2x2 TB NVMe SSD | |
Core i9-13900 Server (128GB) | 128 GB RAM, 2x2 TB NVMe SSD | |
Core i5-13500 Server (64GB) | 64 GB RAM, 2x500 GB NVMe SSD | |
Core i5-13500 Server (128GB) | 128 GB RAM, 2x500 GB NVMe SSD | |
Core i5-13500 Workstation | 64 GB DDR5 RAM, 2 NVMe SSD, NVIDIA RTX 4000 |
AMD-Based Server Configurations
Configuration | Specifications | Benchmark |
---|---|---|
Ryzen 5 3600 Server | 64 GB RAM, 2x480 GB NVMe | CPU Benchmark: 17849 |
Ryzen 7 7700 Server | 64 GB DDR5 RAM, 2x1 TB NVMe | CPU Benchmark: 35224 |
Ryzen 9 5950X Server | 128 GB RAM, 2x4 TB NVMe | CPU Benchmark: 46045 |
Ryzen 9 7950X Server | 128 GB DDR5 ECC, 2x2 TB NVMe | CPU Benchmark: 63561 |
EPYC 7502P Server (128GB/1TB) | 128 GB RAM, 1 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (128GB/2TB) | 128 GB RAM, 2 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (128GB/4TB) | 128 GB RAM, 2x2 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (256GB/1TB) | 256 GB RAM, 1 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (256GB/4TB) | 256 GB RAM, 2x2 TB NVMe | CPU Benchmark: 48021 |
EPYC 9454P Server | 256 GB RAM, 2x2 TB NVMe |
Order Your Dedicated Server
Configure and order your ideal server configuration
Need Assistance?
- Telegram: @powervps Servers at a discounted price
⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️