HPCEnvironment

From NJIT-ARCS HPC Wiki
Jump to: navigation, search

Research computing environment overview

HPC Environment Overview
Category Sub-category General
Access
Node Age [1]
(years: %)
DMS-only
Access [2]
Node Age [1]
(years: %)
Private
Access [3]
Node Age [1]
(years: %)
Notes
CPU
Nodes 240 >10: <7
3-10: <90
0-3: 3
31 >10: 0
3-10: 100
0-3: 0
21 >10: 0
3-10: 34
0-3: 76
About 1000 Kong cores are permanently out of service due to hardware failure
Cores 1,896 380 168
RAM, TB 10.5 3.6 4.7
CPU with GPU
Nodes 2 >10: 0
3-10: 100
0-3: 0
2 >10: 0
3-10: 100
0-3: 0
8 >10: 0
3-10: 0
0-3: 100
GPU Cores 10,752 15,320 64,512
CPU Cores 40 44 10
RAM, TB 0.25 0.26 2.0
Node interconnect 13 of 10Gb/sec, 12 of 56 Gb/sec; rest 1Gb/sec 40 Gb/sec 12 of 56Gb/sec; rest 10Gb/sec
Parallel file system
(PFS)
None None None Traditionally used for temporary files, PFS's are now used for all kinds of storage. Researcher request s for storage are routinely 10 to 20TB, compared to 50 to 100GB two to three years ago

Legend
[1] :

  • Applies also to Cores and RAM
  • Red numbers mean that the nodes are Out of Warranty
  • Green numbers mean that the nodes are In Warranty

[2] : "DMS" refers to the Department of Mathematical Sciences. DMS owns the Stheno cluster.

[3] : "Private Access" refers to cluster hardware purchased by individual researchers. That hardware is dedicated to those researchers.

High-speed node interconnect and parallel file system

Roles of internal Network and PFS

HPC cluster storage

Storage accessible to HPC clusters