Cluster Specifications #

Overview #

Compute nodes
39 (of which 9 are communal)
Logical cores
2768 (of which 680 are communal)
RAM
32-1024 GiB/node
Local scratch
0.05-11 TiB/node
Global scratch
240 TiB
User home storage
1 TiB/user (196 TiB in total)

Summary of Compute Environment #

Feature Login Nodes Transfer Nodes Development Nodes Compute Nodes
SSH access from outside of cluster Within UCSF only, incl. UCSF VPN Within UCSF only, incl. UCSF VPN no no
SSH access from inside of cluster (✓) only for troubleshooting
Outbound access No restrictions HTTP/HTTPS, FTP/FTPS, SSH, SFTP, GIT+SSH, Globus Via proxy: HTTP/HTTPS, GIT+SSH no
Network speed 1 Gbps 10 Gbps 1 Gbps 1,10 Gbps
Core software Minimal Minimal CentOS and EPEL packages, compilers and source-code packages Same as development nodes
modules (software stacks) no no
Global file system
Job submission no
Purpose Submit and query jobs. SSH to development nodes. File management. Fast in- & outbound file transfers. File management. Compile and install software. Prototype and test job scripts. Submit and query jobs. Version control (clone, pull, push). File management. Running short and long-running job scripts.

All nodes on the cluster runs CentOS 7.9 which is updated on a regular basis. The job scheduler is Slurm 21.08 (Simple Linux Utility for Resource Management) which provides partitions for both communal and lab-priority tasks.

Details #

Login Nodes #

The cluster can be accessed via SSH to one of two login nodes:

  1. c4-log1.ucsf.edu
  2. c4-log2.ucsf.edu

Data Transfer Nodes #

For transferring large data files, it is recommended to use one of the dedicate data transfer nodes:

  1. c4-dt1.ucsf.edu
  2. c4-dt2.ucsf.edu

which have 10 Gbps connection - providing a file transfer speed of up to (theoretical) 1.25 GB/s = 4.5 TB/h. Please note, as is true for all ethernet, 80% of line speed is doing pretty good. As for the login nodes, the transfer nodes can be accessed via SSH.

Comment: You can also transfer data via the login nodes, but since those only have 1 Gbps connections, you will see much lower transfer rates.

Development Nodes #

The cluster has development nodes for the purpose of validating scripts, prototyping pipelines, compiling software, and more. Development nodes can be accessed from the login nodes.

Node Logical Cores RAM Local /scratch CPU x86-64 level CPU GPU
c4-dev1 48 384 GiB 3.6 TiB x86-64-v1 AMD Opteron Processor 6176  
             
c4-dev2 48 512 GiB 1.1 TiB x86-64-v1 AMD Opteron Processor 6176  
             
c4-dev3 38 128 GiB 5.4 TiB x86-64-v3 Intel Xeon E5-2640 v4 2.40GHz  
c4-gpudev1 104 1024 GiB 3.4 TiB x86-64-v4 Intel Xeon Gold 5320 2.20GHz Nvidia A40 GPU

Comment: Please use the GPU development node only if you need to build or prototype GPU software. The CPU x86-64 level is the x86-64 microarchitecture levels supported by the nodes CPU.

Compute Nodes #

The majority of the compute nodes have Intel processors, while a few have AMD processes. Each compute node has a local /scratch drive (see below for size), which is either a hard disk drive (HDD), a solid state drive (SSD), or a Non-Volatile Memory Express (NVMe) drive. Each node has a tiny /tmp drive (8-8 GiB).

Node Logical Cores RAM Local /scratch/ CPU x86-64 level CPU Notes Priority
c4-n1 64 512 GiB 1.8 TiB x86-64-v4 2.1 GHz SSD /scratch, 10 Gbps ethernet communal
c4-n2 64 512 GiB 1.8 TiB x86-64-v4 2.1 GHz SSD /scratch, 10 Gbps ethernet communal
c4-n3 64 512 GiB 1.8 TiB x86-64-v4 2.1 GHz SSD /scratch, 10 Gbps ethernet communal
c4-n4 64 512 GiB 1.8 TiB x86-64-v4 2.1 GHz SSD /scratch, 10 Gbps ethernet communal
c4-n5 64 512 GiB 2.6 TiB x86-64-v4 2.1 GHz SSD /scratch, 10 Gbps ethernet communal
c4-n6 112 768 GiB 11.0 TiB x86-64-v4 2.2 GHz NVMe /scratch, 10 Gbps ethernet Witte Lab
c4-n7 112 768 GiB 11.0 TiB x86-64-v4 2.2 GHz NVMe /scratch, 10 Gbps ethernet Witte Lab
c4-n8 112 768 GiB 11.0 TiB x86-64-v4 2.2 GHz NVMe /scratch, 10 Gbps ethernet Witte Lab
c4-n9 112 768 GiB 11.0 TiB x86-64-v4 2.2 GHz NVMe /scratch, 10 Gbps ethernet Witte Lab
c4-n10 48 512 GiB 2.3 TiB x86-64-v2 2.1 GHz SATA /scratch, 10 Gbps ethernet communal
c4-n11 128 1024 GiB 5.2 TiB x86-64-v4 2.8 GHz SSD /scratch, 10 Gbps ethernet Costello Lab
c4-n12 64 512 GiB 3.6 TiB x86-64-v2 2.6 GHz SATA /scratch, 10 Gbps ethernet CBI Core
c4-n13 64 512 GiB 3.6 TiB x86-64-v2 2.6 GHz SATA /scratch, 10 Gbps ethernet CBI Core
c4-n14 24 128 GiB 0.9 TiB x86-64-v2 3.5 GHz SATA /scratch, 10 Gbps ethernet Witte Lab
c4-n15 64 256 GiB 1.8 TiB x86-64-v2 2.6 GHz SATA /scratch, 10 Gbps ethernet Witte Lab
c4-n16 64 512 GiB 3.3 TiB x86-64-v4 3.7 GHz SATA /scratch, 10 Gbps ethernet Blelloch Lab
c4-n17 64 512 GiB 2.6 TiB x86-64-v4 3.7 GHz SAS /scratch, 10 Gbps ethernet Francis Lab
c4-n18 64 512 GiB 3.2 TiB x86-64-v3 2.1 GHz SAS /scratch, 10 Gbps ethernet Krummel Lab
c4-n19 64 768 GiB 4.9 TiB x86-64-v4 3.7 GHz SAS /scratch, 10 Gbps ethernet Ziv Lab
c4-n20 48 384 GiB 7.0 TiB x86-64-v3 2.4 GHz SATA /scratch, 10 Gbps ethernet Krummel Lab
c4-n21 64 512 GiB 11.0 TiB x86-64-v3 2.6 GHz SAS /scratch, 10 Gbps ethernet Witte Lab
c4-n22 128 512 GiB 2.6 TiB x86-64-v4 3.7 GHz SSD /scratch, 10 Gbps ethernet Kim Lab
c4-n23 64 512 GiB 2.7 TiB x86-64-v3 2.3 GHz SSD /scratch, 10 Gbps ethernet Shannon Lab
c4-n24 64 512 GiB 2.6 TiB x86-64-v4 3.7 GHz SSD /scratch, 10 Gbps ethernet Shannon Lab
c4-n25 64 512 GiB 2.6 TiB x86-64-v4 3.7 GHz SSD /scratch, 10 Gbps ethernet Bastian Lab
c4-n26 48 256 GiB 5.4 TiB x86-64-v3 2.8 GHz SATA /scratch, 10 Gbps ethernet Bandyopadhyay Lab
c4-n27 64 512 GiB 3.2 TiB x86-64-v4 2.1 GHz SAS /scratch, 10 Gbps ethernet Kriegstein Lab
c4-n28 48 384 GiB 5.4 TiB x86-64-v3 2.8 GHz SATA /scratch, 10 Gbps ethernet Molinaro Lab
c4-n29 48 384 GiB 5.4 TiB x86-64-v3 2.8 GHz SATA /scratch, 10 Gbps ethernet Molinaro Lab
c4-n30 32 192 GiB 0.05 TiB x86-64-v3 1.2 GHz SAS /scratch, 2x1 Gbps ethernet Oldham Lab
c4-n31 64 512 GiB 3.2 TiB x86-64-v4 2.1 GHz SAS /scratch, 10 Gbps ethernet Diaz Lab
c4-n32 64 512 GiB 5.4 TiB x86-64-v3 2.6 GHz SATA /scratch, 10 Gbps ethernet Costello Lab, Diaz Lab, Fung Lab
c4-n33 64 512 GiB 5.4 TiB x86-64-v3 2.6 GHz SATA /scratch, 10 Gbps ethernet Costello Lab, Diaz Lab, Fung Lab
c4-n34 64 512 GiB 5.4 TiB x86-64-v3 2.6 GHz SATA /scratch, 10 Gbps ethernet Costello Lab, Diaz Lab, Fung Lab
c4-n35 64 512 GiB 2.6 TiB x86-64-v4 2.1 GHz SSD /scratch, 10 Gbps ethernet Krummel Lab
c4-n36 48 384 GiB 5.4 TiB x86-64-v3 1.4 GHz SATA /scratch, 10 Gbps ethernet Costello Lab
c4-n37 104 1024 GiB 3.4 TiB x86-64-v4 3.4 GHz SSD /scratch, 10 Gbps ethernet communal
c4-n38 104 1024 GiB 3.4 TiB x86-64-v4 3.4 GHz SSD /scratch, 10 Gbps ethernet Nvidia A100 GPU communal
c4-n39 104 1024 GiB 3.4 TiB x86-64-v4 3.4 GHz SSD /scratch, 10 Gbps ethernet Nvidia A100 GPU communal

The compute nodes can only be utilized by submitting jobs via the scheduler.

File System #

Scratch Storage #

The C4 cluster provides two types of scratch storage:

There are no per-user quotas in these scratch spaces. Files not added or modified during the last two weeks will be automatically deleted on a nightly basis. Note, files with old timestamps that were “added” to the scratch place during this period will not be deleted, which covers the use case where files with old timestamps are extracted from tar.gz file. (Details: tmpwatch --ctime --dirmtime --all --force is used for the cleanup.)

User and Lab Storage #

Each user may use up to 1 TiB disk space in the home directory. It is not possible to expand user’s home directory. Many labs have purchased their own storage servers which were then mounted on the cluster. If you’d like more information please contact us.

Network #

The compute nodes are connected to the local network with 1 Gbps and 10 Gbps network cards.