Current Cluster Information
As of September 2025, the following table summarizes the DSI cluster resources. Some resources are reserved for specific initiatives or grants.
Total Cluster Resources
| Resource Type | Number of Units |
|---|---|
| Number of Nodes | 29 |
| Total Cores | 2,720 |
| Total Storage | 1.5 PB |
| GPUs | 32 NVIDIA A40 48GB 60 NVIDIA A100 80GB 12 NVIDIA L40S 48GB 16 NVIDIA H100 80GB 20 NVIDIA H200 140GB Total: 140 |
Cluster Networking and Topology
The cluster is organized into a series of nodes and storage servers. Storage servers have a mix of different drives and types.
Users connect to the cluster via three login (frontend) nodes: fe01.ds, fe02.ds, and fe03.ds. These provide redundant access — if one is unavailable, users can connect to either of the others.
Note that not all nodes are available on all SLURM partitions and queues.
The cluster has a 10 Gbps line to the internet.
The network between nodes and storage servers runs at 100 Gbps. This internal network is not routable from the campus network or internet.
Space Availability and Server Room
The DSI cluster is housed in the Hinds data center, a UChicago-managed facility. Available rack space and power capacity are managed by the Hinds data center operations team, and our allocation is contingent on the university’s overall use of that space. The server room is power-constrained, which limits the density of GPU nodes per rack and is the primary limiting factor for cluster growth.