Server rental store

AI in the World Bank

# AI in the World Bank: Server Configuration & Architecture

This article details the server configuration supporting Artificial Intelligence (AI) initiatives within the World Bank. It is geared towards new system administrators and developers onboarding to the platform. We will cover hardware, software, and network considerations crucial for maintaining a robust and scalable AI infrastructure. Understanding these details is paramount for successful deployment and maintenance of AI-driven solutions at the World Bank. This document assumes basic familiarity with Linux server administration and networking concepts. See Server Administration Basics for introductory material.

Overview

The World Bank leverages AI for a variety of applications, including risk assessment, fraud detection, project monitoring, and economic forecasting. These applications demand significant computational resources and specialized software. Our current architecture is a hybrid model, utilizing a combination of on-premise servers and cloud resources (primarily Amazon Web Services). This allows for flexibility and scalability while maintaining control over sensitive data. This document focuses specifically on the on-premise components. For information on the cloud infrastructure, please refer to the Cloud Services Documentation.

Hardware Infrastructure

The core of our on-premise AI infrastructure consists of a cluster of high-performance servers. The following table details the specifications of a typical server node:

Component Specification
CPU 2 x Intel Xeon Gold 6338 (32 cores per CPU, 64 total)
RAM 512 GB DDR4 ECC Registered RAM
Storage 4 x 4TB NVMe SSD (RAID 0) for OS and active data 8 x 16TB SAS HDD (RAID 6) for long-term storage
GPU 4 x NVIDIA A100 (80GB HBM2e)
Network Interface 2 x 100 GbE Mellanox ConnectX-6
Power Supply 2 x 2000W Redundant Power Supplies

These servers are housed in a dedicated, climate-controlled server room with redundant power and cooling. Details on the physical security of the server room can be found in the Physical Security Protocol. The network backbone utilizes a high-speed InfiniBand network to minimize latency between nodes during model training.

Software Stack

The software stack is built around a Linux operating system (CentOS 7) and a variety of AI/ML frameworks. Key components include:

⚠️ *Note: All benchmark scores are approximate and may vary based on configuration. Server availability subject to stock.* ⚠️