Skip to content

System Overview

The SRDC Platform supports data storage, management, and computation designed for compute-intensive research on large and sensitive data. SRDC features dedicated infrastructure, housed in the secure campus data center, as well as personnel to support research and enforce the necessary privacy and security policies and procedures.

Technology & Equipment

  • High performance computing, accessible to researchers under their Faculty Computing Allowance. Researchers may purchase additional HPC nodes to add to the system.
  • Windows and Linux virtual machines, accessible to researchers under their Faculty Computing Allowance. Researchers may purchase additional virtual machine hardware to add to the system.
  • A dedicated, large-scale, and high performance storage system. Researchers may purchase additional storage to add to the system.

SRDC High-Performance Computing (HPC) Cluster Overview

The SRDC HPC cluster offers a CPU compute pool comprising 40 nodes, each featuring Intel Xeon Gold 6230 CPUs, 364 GB of RAM, 20 cores running at 2.1 GHz, and support for 16 floating-point operations per clock cycle. Job submissions to these nodes are under the partition name of “srdc”.

GPUs on SRDC HPC

Alongside the CPU-only nodes, the SRDC cluster also features 48 GeForce GTX 1080 Ti GPUs distributed across 12 nodes. Each node contains 8 CPUs and 4 GPUs with a total of 64 GB of CPU RAM and 11 GB of GPU RAM per GPU. SRDC’s GPUs allow for concurrent operations, memory mapping, and coordinated kernel launches.

SRDC HPC Hardware Configuration

Partition Nodes Node List CPU Model # Cores / Node Memory / Node Infiniband Specialty Scheduler Allocation
srdc 40 n00[00-39].srdc.srdc0 Intel Xeon Gold 6230 20 364 GB 4x EDR - By Node
srdc_GTX1080TI 12 n00[40-51].srdc.srdc0 Intel Xeon ES-2623 8 64 GB 4x FDR 4x GeForce GTX 1080 Ti per Node
(11 GB of GPU RAM per GPU)
(48 GPUs total)
By Node

SRDC Virtual Machines (VM) Overview

SRDC also offers Linux and Windows VM computing environments for workflows that necessitate a graphical user interface (GUI) or Windows-specific software. SRDC VMs are hosted on one of eight servers running VMware ESXi bare-metal hypervisors to flexibly meet our VM users' computing needs. The table below provides additional details.

SRDC VM Hardware Configuration

Servers Hypervisor CPU Model # Cores / Host Memory / Host Storage Interconnect Specialty
8 VMware ESXi Intel Xeon Gold 6254 72 (shared) 766 GB (shared) bIT Performance and Utility tiers (managed by SRDC) 32GB Fibre Channel Graphical user interface
Linux VM
Windows VM

Storage on SRDC

Storage allocations for project data are sized based on project needs and storage availability for Linux and Windows SRDC environments.

In Linux SRDC environments, users can utilize working scratch space, subject to a 12TB quota and a purge policy. Larger storage allocations may be available via MOU. Research groups can expand their storage on the GPFS if subject to space limitations. The cost of expansion is $25.8K for 100TB usable as of December 2023.

Please contact us at brc@berkeley.edu for a consultation.

Platform Support & Research Facilitation

The SRDC is managed by system administrators and SRDC consultants, who monitor the OS and software, review security requirements and compliance, identify computational workflows, and help onboard new users. The Information Security Office works closely with SRDC staff to ensure security through monitoring and intrusion detection.