Server rental store

Machine Learning

# Machine Learning Server Configuration

This article details the recommended server configuration for deploying machine learning workloads within our MediaWiki environment. It is intended for system administrators and engineers responsible for setting up and maintaining the infrastructure. We will cover hardware specifications, software requirements, and key configuration considerations. This builds upon the existing Server Infrastructure Overview and complements the documentation on Database Configuration.

Hardware Requirements

Machine learning tasks are often computationally intensive, particularly during training. Therefore, robust hardware is crucial. The following table outlines the recommended specifications for different tiers of machine learning servers. These specifications assume a primary focus on deep learning applications using frameworks like TensorFlow and PyTorch.

Tier CPU RAM GPU Storage Network
Development Intel Xeon E5-2680 v4 or AMD EPYC 7302P 64 GB DDR4 NVIDIA GeForce RTX 3060 (12GB VRAM) 1 TB NVMe SSD 1 Gbps Ethernet
Production (Small) Intel Xeon Gold 6248R or AMD EPYC 7402P 128 GB DDR4 ECC NVIDIA Tesla T4 (16GB VRAM) 2 TB NVMe SSD (RAID 1) 10 Gbps Ethernet
Production (Large) Dual Intel Xeon Platinum 8280 or Dual AMD EPYC 7763 256 GB DDR4 ECC 4x NVIDIA A100 (80GB VRAM each) 4 TB NVMe SSD (RAID 10) 25 Gbps Ethernet

These are baseline recommendations; specific requirements will vary based on the complexity of the models and the size of the datasets. Consider scaling storage and GPU resources as needed. Refer to the Storage Solutions Guide for more detailed information on storage options.

Software Stack

The software stack for a machine learning server typically includes an operating system, a containerization platform, a machine learning framework, and supporting libraries. We standardize on the following:

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️