Table 1 Specifications of the hybrid cloud system at the Center for Genomic Medicine, Kyoto University, Japan.

From: Design and implementation of a hybrid cloud system for large-scale human genomic research

Subsystem

Name

Specification

Size

System A

File System

GPFS ESS JBOD (5U84)

2.1 PB

File System

NAS

3.6 PB

Compute Nodes

Intel Xeon Haswell E7-8890 v3 (Haswell, 18 cores 2.5 GHz × 4)/512 GiB

Three nodes

Compute Nodes

Intel Xeon Ivy Bridge E7-4880 v2 (Ivy Bridge, 15 cores 2.5 GHz × 4)/512 GiB

One node

Compute Nodes

Intel Xeon E5-2698 v4 (Broadwell, 20 cores 2.2 GHz × 2)/8× Tesla P100 GPU/512GiB

One node

Job Schedular

Slurm

Container

Singularity v3

Location

Center for Genomic Medicine, Kyoto University, Japan

System B

System name

Laurel 2

File System

Lustre ExaScaler (SFA14K)

0.71(24) PB

Compute Nodes

Intel Xeon E5-2694 v4 (Broadwell, 18 cores 2.1 GHz × 2)/128 GiB

17 (850) nodes

Network

Between network gateways and SINET (100 Gbps)

Between file systems and compute nodes (100 Gbps)

Job Schedular

PBS

Container

Singularity v3

Location

Academic Center for Computing and Media Studies, Kyoto University, Japan

System C

System name

Oakbridge-CX (OBCX)

File System

Lustre ExaScaler (ES18KE × 2)

0.7 (12.4) PB

Compute Nodes

Intel Xeon Platinum 8280 (CascadeLake, 28 cores 2.7 GHz × 2)/192 GiB

256 (1368)

nodes

Network

Between the private network gateway and SINET (10 Gbps)

Between public network gateways to SINET (40 Gbps × 2)

Between file systems and compute nodes (100 Gbps)

Job Schedular

Fujitsu Technical Computing Suite-

Container

Singularity v3

Location

Information Technology Center, the University of Tokyo, Japan

System D

System name

mdx

File System

Lustre File Sytem (NVMe)

0.15 (1) PB

File System

Lustre File Sytem (HDD)

0.3 (16.3) PB

Compute Nodes

Virtualization environment with 608 cores and 1024 GiB in total. 76 cores (152 vCPUs) and 128 GiB are assigned to each compute node. The physical CPU is Intel Xeon Platinum 8368 (IceLake, 38 cores 2.4 GHz × 2).

8 (368)

nodes

Network

Between compute nodes and SINET (25 Gbps)

Job Schedular

Slurm

Container

Singularity v3

Location

Information Technology Center, the University of Tokyo, Japan

System E/P

File System

FSx for Lustre

On-demand

File System

EBS

On-demand

File System

S3

On-demand

Network

Between System A to AWS (AWS Direct Connect via SINET (10 Gbps))

Compute Nodes

AWS assigns requested instance if compute nodes are physically available. For our GATK joint-genotyping in System E, spot instance of r5.large (2 × vCPU (Intel Xeon Platinum 8000 3.1 GHz)/16 GiB)) was used (max 320 nodes).

Available max resources as spot instance of vCPU and memory in AWS region in Japan are hpc6a.48xlarge (AMD EPYC Milan) with 192 vCPUs and x2iedn.32xlarge (Intel Xeon IceLake) 4096 GiB at 1/Jul/2022, respectively.

Spot

Location

AWS region in Japan (Systems E and P are in different virtual private cloud)

  1. PB petabyte, GiB gibibyte.
  2. The value outside the brackets indicates the value currently being rented, and the value inside the brackets indicates the total system.