.. _specs-gina: ******************* Gina specifications ******************* Gina is the Grid cluster at SURFsara; currently all the servers are located at the Digital Realty datacenter. After different tenders these servers belong to different brands (Dell & Fujitsu). Three service nodes are used to host virtual machines for different purposes (creamces, monitoring, installation, dns servers...). All the worker nodes are installed on physical nodes. This page describes the specifications for the Gina Grid cluster at SURFsara: .. contents:: :depth: 4 If you have any questions concerning the technical specifications below, please contact us at helpdesk@surfsara.nl. .. _gina-specs-summary: Quick summary ============= ============================ ===================================================== Gina cluster Capacity ============================ ===================================================== Operating system Linux CentOS 7.x 64bit Total number of cores 7408 Xeon cores at 2.2 to 2.6 GHz Total memory 58.3TB Total scratch space 2258TB ============================ ===================================================== ``Last update: June 2018`` .. _gina-specs-network: Network ============ GinA and Grid storage share a network backbone, connected with a high throughput network. Currently we have seen over 20GB/sec (~170-200Gbit/sec) of peak network traffic to the Grid storage. All workernodes are connected with a single 10Gbit or 25Gbit ethernet connection. .. _gina-specs-wn: Worker Nodes ============ This is the list of the different worker nodes in order of installation/configuration from newest to oldest: Worker Nodes ``f18-{01-12}``, ``f19-{01-11}``, ``f20-{01-12}``:: * Dell R640 * 2x Intel(R) Xeon(R) Gold 6148 CPU @ 2.40GHz (40 cores) * 384GB RAM * 3TB scratch * Type: Skylake architecture * Number of nodes: 35 * Total cores of this type: 1400 * Scratch per core: ~80GB * RAM per core: 8GB Worker Nodes ``f08-{01-16}``, ``f09-{01-16}``, ``f10-{01-16}``, ``f11-{01-16}``, ``f12-{01-17}``:: * Fujitsu RX2530 * 2x Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz (24 cores) * 192GB RAM * ~2.4TB scratch (SSD only) * Type: Broadwell architecture /w Broadwell-EP cores (2016) * Number of nodes: 81 * Total cores of this type: 1944 * Scratch per core: ~100GB * RAM per core: 8GB Worker Nodes ``f03-{01-20}``, ``f04-{01-20}``, ``f05-{01-20}``, ``f06-{01-20}``, ``f07-{01-20}``:: * Dell R630 * 2x Intel(R) Xeon(R) CPU E5-2680 v3 @ 2.50GHz (24 cores) * 192GB RAM * ~8000GB scratch * Type: Haswell architecture /w Haswell-EP cores (2014) * Number of nodes: 100 * Total cores of this type: 2400 * Scratch per core: ~300GB * RAM per core: 8GB Worker Nodes ``f13-{01-24}``, ``f14-{01-20}``, ``f15-{01-20}``, ``f16-{01-20}``, ``f17-{01-20}``:: * Fujitsu CX250 * 2x Intel(R) Xeon(R) CPU E5-2650 v2 @ 2.60GHz (16 cores) * 128GB RAM * ~11TB scratch * Type: Ivy Bridge architecture /w Ivy Bridge-EP cores (2013) * Number of nodes: 104 * Total cores of this type: 1664 * Scratch per core: ~680GB * RAM per core: 8GB Service nodes ============= Service{01,02,03}:: * Dell R420 * 2xIntel(R) Xeon(R) CPU E5-2420 0 @ 1.90GHz (12 cores) * 96GB RAM * Number of nodes: 3 * RAM per core: 8GB CreamCEs ======== All 3 CreamCEs are virtualized and distributed among the 3 Service Nodes. Every CreamCE has 4 cores and 9GB RAM in total. .. _gina-specs-queues: Queues ====== =============== =========================== ============================ Queue Max. Walltime (hh:mm:ss) VOs(group) allowed =============== =========================== ============================ long 96:00:00 geant4 atlas(production) projects.nl(tropomi) virgo lsgrid(mine) medium 36:00:00 astron atlas alice auger bbmri.nl beapps biomed dans enmr.eu esr geant4 lhcb lofar lsgrid nlesc.nl omegac pvier xenon.biggrid.nl projects.nl chem.biggrid.nl euclid-ec.org drihm.eu virgo skatelescope.eu dune short 04:00:00 astron atlas auger bbmri.nl beapps biomed dans enmr.eu esr geant4 lhcb lofar lsgrid nlesc.nl omegac pvier xenon.biggrid.nl projects.nl chem.biggrid.nl euclid-ec.org drihm.eu dune infra 00:30:00 dteam ops pvier =============== =========================== ============================ ``Last update: February 2019`` The Grid jobs submitted to the queues above are restricted by the walltime limit, not the CPU limit. Information on the queues available on the cluster at Nikhef can be found `here`_ .. Links: .. _`here`: https://wiki.nikhef.nl/grid/Grid@Nikhef