AI in Inclusion
- AI in Inclusion: Server Configuration Guide
This article details the server configuration for the "AI in Inclusion" project, a MediaWiki extension focused on utilizing Artificial Intelligence to improve accessibility and inclusivity within our wiki environment. This guide is intended for newcomers responsible for server maintenance and configuration.
Overview
The “AI in Inclusion” project uses several AI models to enhance the wiki experience. These include models for automated alt-text generation, translation services, content summarization, and bias detection. This demands a robust and scalable server infrastructure. This document outlines the necessary hardware and software configurations required to support these functionalities. We will focus on the core server components and their interaction with the MediaWiki installation. See Manual:Configuration for general MediaWiki configuration information.
Hardware Specifications
The following table details the minimum and recommended hardware specifications for the "AI in Inclusion" server. These specifications are based on anticipated load and the computational demands of the AI models.
Component | Minimum Specification | Recommended Specification |
---|---|---|
CPU | Intel Xeon E5-2680 v4 (14 cores) | Intel Xeon Platinum 8380 (40 cores) |
RAM | 64 GB DDR4 ECC | 256 GB DDR4 ECC |
Storage (OS & MediaWiki) | 500 GB NVMe SSD | 1 TB NVMe SSD |
Storage (AI Models) | 4 TB HDD (7200 RPM) | 8 TB SSD |
Network Interface | 1 Gbps Ethernet | 10 Gbps Ethernet |
GPU | NVIDIA Tesla T4 (for initial testing) | NVIDIA A100 (for production) |
These specifications assume a moderate level of wiki activity. Higher traffic volumes or more complex AI model usage will necessitate increased resources. Consult Help:System requirements for more information about server requirements.
Software Stack
The "AI in Inclusion" project relies on a specific software stack to function correctly. This section details the required operating system, database, web server, and AI-related libraries.
Operating System
We utilize Ubuntu Server 22.04 LTS as the operating system. This provides a stable and well-supported platform. Detailed instructions for Ubuntu Server installation can be found at Help:Installing MediaWiki.
Database
MariaDB 10.6 is used as the database server. The database stores wiki content, user information, and AI model metadata. Proper database configuration is crucial for performance. See Manual:Database for configuration details. The following table outlines key database configuration parameters:
Parameter | Value |
---|---|
`innodb_buffer_pool_size` | 8 GB (minimum), 32 GB (recommended) |
`max_connections` | 200 (minimum), 500 (recommended) |
`query_cache_size` | 64 MB (consider disabling for heavy write loads) |
Web Server
Apache 2.4 is used as the web server, configured with `mod_php` to process PHP scripts. Proper Apache configuration is essential for security and performance. Refer to Manual:Apache for detailed configuration instructions.
AI Libraries & Frameworks
The following Python libraries and frameworks are essential for running the AI models:
- TensorFlow 2.x
- PyTorch 1.x
- Transformers (Hugging Face)
- spaCy
- NLTK
These are managed through a dedicated Python virtual environment to avoid conflicts with other system packages. See Help:Python for virtual environment setup.
Network Configuration
Proper network configuration is vital for accessibility and performance.
- **Firewall:** A firewall (e.g., `ufw`) should be configured to allow access only on necessary ports (80 for HTTP, 443 for HTTPS, 22 for SSH).
- **DNS:** Ensure proper DNS resolution for the wiki domain.
- **Load Balancing:** For high availability and scalability, consider implementing a load balancer (e.g., HAProxy) to distribute traffic across multiple servers. See Help:Load balancing for more details.
- **Caching:** Utilize a caching layer (e.g., Varnish) to reduce server load and improve response times. See Help:Caching for more information.
The following table shows the necessary port configurations:
Port | Protocol | Description |
---|---|---|
80 | TCP | HTTP (unencrypted web traffic) |
443 | TCP | HTTPS (encrypted web traffic) |
22 | TCP | SSH (secure remote access) |
Security Considerations
Security is paramount. Regularly update all software components, implement strong passwords, and monitor server logs for suspicious activity. Enable HTTPS and configure a web application firewall (WAF) to protect against common web attacks. Refer to Manual:Security for comprehensive security guidelines.
Future Scalability
The architecture is designed for scalability. Adding additional servers and distributing the load across them can accommodate increased traffic and AI model complexity. Containerization technologies (e.g., Docker) can simplify deployment and management. See Help:Docker for more information.
Help:Configuration
Manual:Database
Manual:Apache
Help:System requirements
Help:Installing MediaWiki
Help:Python
Help:Load balancing
Help:Caching
Manual:Security
Help:Docker
Extension:AI in Inclusion
Help:Firewall
Manual:Maintenance
Help:Troubleshooting
Help:Accessibility
Manual:Upgrading
Intel-Based Server Configurations
Configuration | Specifications | Benchmark |
---|---|---|
Core i7-6700K/7700 Server | 64 GB DDR4, NVMe SSD 2 x 512 GB | CPU Benchmark: 8046 |
Core i7-8700 Server | 64 GB DDR4, NVMe SSD 2x1 TB | CPU Benchmark: 13124 |
Core i9-9900K Server | 128 GB DDR4, NVMe SSD 2 x 1 TB | CPU Benchmark: 49969 |
Core i9-13900 Server (64GB) | 64 GB RAM, 2x2 TB NVMe SSD | |
Core i9-13900 Server (128GB) | 128 GB RAM, 2x2 TB NVMe SSD | |
Core i5-13500 Server (64GB) | 64 GB RAM, 2x500 GB NVMe SSD | |
Core i5-13500 Server (128GB) | 128 GB RAM, 2x500 GB NVMe SSD | |
Core i5-13500 Workstation | 64 GB DDR5 RAM, 2 NVMe SSD, NVIDIA RTX 4000 |
AMD-Based Server Configurations
Configuration | Specifications | Benchmark |
---|---|---|
Ryzen 5 3600 Server | 64 GB RAM, 2x480 GB NVMe | CPU Benchmark: 17849 |
Ryzen 7 7700 Server | 64 GB DDR5 RAM, 2x1 TB NVMe | CPU Benchmark: 35224 |
Ryzen 9 5950X Server | 128 GB RAM, 2x4 TB NVMe | CPU Benchmark: 46045 |
Ryzen 9 7950X Server | 128 GB DDR5 ECC, 2x2 TB NVMe | CPU Benchmark: 63561 |
EPYC 7502P Server (128GB/1TB) | 128 GB RAM, 1 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (128GB/2TB) | 128 GB RAM, 2 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (128GB/4TB) | 128 GB RAM, 2x2 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (256GB/1TB) | 256 GB RAM, 1 TB NVMe | CPU Benchmark: 48021 |
EPYC 7502P Server (256GB/4TB) | 256 GB RAM, 2x2 TB NVMe | CPU Benchmark: 48021 |
EPYC 9454P Server | 256 GB RAM, 2x2 TB NVMe |
Order Your Dedicated Server
Configure and order your ideal server configuration
Need Assistance?
- Telegram: @powervps Servers at a discounted price
⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️