The Advanced Research Computing center’s HPC Cluster (ARC) is UC’s large-memory computer cluster. Designed to support data-intensive computing, ARC is particularly well-suited for running software and applications that require large amounts of computer memory and processing. ARC provides specialized deep GPU partitions for researchers with deep learning applications and other applications that require GPUs. ARC currently has 2 major clusters, namely ARC1 and ARC2. The hardware specifications for each are summarized in the table below
ARCC1 (pilot cluster)
Number of nodes | Processors per node | Cores per nodes | Memory per node | Omnipath | Accelerator card | Speed |
36 | 2 Intel Xeon Gold 6148 CPUs | 40 | 192 GB | 100 GB/s | N/A | 2.4-3.7 GHz |
1 | 2 Intel Xeon Gold 6148 CPUs | 40 | 192 GB | 100 GB/s | 2 NVIDIA Tesla V100-32GB GPU | 2.4-3.7 GHz |
ARCC2
Node Type | Number of nodes | Processors per node | Cores per nodes | Memory per node | InfiniBand | Accelerator card | Speed |
RM | 95 | 2 AMD EPYC 7452 CPUS | 64 | 256GB RAM | 100 GB/s | N/A | 2.35-3.35 GHz |
GPU | 10 | 2 AMD EPYC 7452 CPUS | 64 | 1024GB RAM | 100 GB/s | 2 NVIDIA Tesla A100-40GB GPU | 2.35-3.35 GHz |
Large Mem | 1 | 2 AMD EPYC 7452 CPUS | 64 | 2048GB RAM | 100 GB/s | N/A | 2.35-3.35 GHz |
UC’s newest High-Performance Computing, AI and High-Performance Data Analytics (HPDA) system
UC’s newest high-performance computing and data analytics system, available Fall 2022, is funded in part by investments from the Office of Research, IT@UC, colleges and departments and a significant grant from the National Science Foundation’s Major Research Instrumentation (MRI) program (award #2018617). UC is partnered with Hewlett Packard Enterprises (HPE) to architect a purpose-built compute resource for demanding High-Performance Computing (HPC) and Artificial Intelligence (AI) applications.
ARCC2 provides transformative capability for rapidly evolving, computation-intensive and data-intensive research, supporting both traditional and non-traditional research communities and applications. The converged, scalable HPC, machine learning and data tools create an opportunity for collaboration and converged research, prioritizing researcher productivity and ease of use with an easy-to-use web-based interface.
Innovation
- AMD EPYC 7452 CPUs: 64-core 2.35–3.35 GHz
- AI scaling to 20 Tesla A100-40GB GPUs
- Mellanox HDR-100 InfiniBand supports in-network MPI-Direct, RDMA, GPUDirect, SR-IOV, and data encryption
- Cray ClusterStor E1000 Storage System
- Open OnDemand – Web based interface
Regular Memory
Regular Memory (RM) CPU nodes provide extremely powerful general-purpose computing, machine learning and data analytics, AI inferencing, and pre- and post-processing. 95 RM nodes will have:
- Two AMD EPYC CPUS, each with:
- 32 cores
- 2.35-3.35GHz
- 128MB L3
- 256GB of RAM
- 8 memory channels
- SATA SSD (960GB)
- Mellanox ConnectX-6 HDR InfiniBand 100Gb/s Adapter
Large Memory
Large Memory (LM) node will provide 1TB of shared memory for genome sequence assembly, graph analytics, statistics, and other applications requiring a large amount of memory for which distributed-memory implementations are not available.
ARCC2’s 1 LM nodes will consist of:
- Two AMD EPYC CPUS, each with:
- 32 cores
- 2.35-3.35GHz
- 128MB L3
- 8 memory channels
- 1024GB of RAM
- Mellanox ConnectX-6 HDR InfiniBand 100Gb/s Adapter
GPU
10 GPU nodes provide exceptional performance and scalability for deep learning and accelerated computing. Each GPU node will contain:
- Two NVIDIA Tesla A100 40GB GPUs
- Two AMD EPYC CPUS, each with:
- 32 cores
- 2.35-3.35GHz
- 128MB L3
- 8 memory channels
- 1024GB of RAM
- SATA SSD (960GB)
- Mellanox ConnectX-6 HDR InfiniBand 100Gb/s Adapter
Please contact Jane Combs or arc_info@uc.edu with questions