Monthly Archives: November 2016

RStudio on Zeus

We have R 3.3.2 installed on zeus (login nodes and compute nodes). You can also access R on one login node via RStudio web interace at http://zeus.coventry.ac.uk/R

Alex Pedcenko

Current (experimental) limits of queues

The total number of nodes (or CPUs) you can use depend on how long your job has to run (i.e. in which queue/partition it was submitted):

[queues are listed from higher to lower priority, i.e. shorter queues have higher priority in the waiting list!]

for up to 4-hours short jobs:

  • short4” queue has unlimited number  of CPUs for up to 4 hours  and can use nodes from any of the queues! (Hint: do not specify how many nodes you need, just specify how many CPU-cores you need for your job, i.e. for 1000 CPU-cores for 4 hours “sbatch -p short4 -n 1000 -t 4:00:00  submitscript.slurm”). Default time (if you do not specify walltime) is 1 hr.

for up to 12hours jobs:

  • You can use up to 144 nodes x 8 CPUs  of “all12”  queue for up to 12 hours.

For up to 24hrs jobs:

  • You can use up to 80 nodes x 8 CPUs  in “all” queue (640 CPUs) for up to 24 hours.
  • You can also use 10 nodes x 32 CPUs of Broadwell queue (another 320 CPUs, use less nodes per job, these are “fat” nodes! ) for up to 36 hours
  • You can also use 20 nodes x 8 CPUs of “all48” (another 160 CPUs) queue for up to 48 hours (lower priority than “all“)

For up to 36hrs jobs:

  • You can use up to 10 nodes x 32CPUs in Broadwell queue
  • You can use up to 5 nodes x 32CPUs (160 CPUs and 10 K80 GPUs) in NGPU queue
  • You can use up to 18 nodes x 12 CPUs in (+ 36 K20 GPUs)  in GPU queue
  • You can use 1 SMP node (32 CPUs and 512 GB RAM) in SMP queue

(specialized queues SMP,GPU,NGPU have higher priority in the waiting list, i.e. if you need to use GPUs on these nodes, you have higher “weight”)

For up to 48hrs jobs:

  • You can use 20 nodes x 8 CPUs (160 CPUs in total) in queue “all48” for up to 48 hours

For >48 hrs long jobs:

  • You can use up to 20 nodes (160 CPUs) in queue “long” for unlimited job time

 

 

 

 

“RequiredNodeNotAvailable” status of the Job

There is a standing reservation of all nodes in “all” queue and “Broadwell+NGPU+Phi” nodes for this Sunday 13/11/16 from 0:00 to 12:00, which is needed for conducting more performance tests before commissioning of HPC. So If your submitted job spans through this time period you will get this message as a reason for “queueing”.

Alex Pedcenko

Zeus Alive!

Just few final tests left after hardware / software upgrades, which took place during couple of weeks at the end of October 2016. The major “full blast heat generation” test will be performed on Monday 7/11/2016 to see if the server room survives. If you want to test some of your codes/jobs before then please ask Alex Pedcenko aa3025@coventry.ac.uk to enable your slurm accounts.

What have we got now:

the HPC specs we will have after the latest upgrade:

Old stuff:

  • 144 nodes with 2xIntel Xeon L5530 @ 2.40GHz CPUs (Nehalem 8 CPU-cores/node), 48 Gb of RAM per node: queues all and long
  • 18 nodes with 2xIntel(R) Xeon(R) CPU E5-2440 0 @ 2.40GHz (Sandy Bridge 12 CPU-cores/node) and 2x NVidia Tesla K20 GPUs, 48 Gb RAM:  queue GPU
  • 1 x 32 CPU SMP Machine Intel(R) Xeon(R) CPU E5-4620 0 @ 2.20GHz (32 CPU-cores), 512 Gb RAM: queue SMP

Compute nodes of HPC are interconnected with High-Speed QDR Infiniband (40 GBps)

New Broadwell CPU based nodes:

  • 44 nodes with 2 x Intel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz (32 CPU-cores/node) 128Gb RAM:   queue Broadwell
  • 10 nodes with 2 x Intel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz (32 CPU-cores/node) 128Gb RAM + 2x NVidia K80 GPU:  queue NGPU
  • 2 nodes  with 2xIntel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz (32 CPU-cores/node) 128Gb RAM + Xeon Phi coprocessor SE10/7120: queue Phi

2 x 15 Tb file servers (/home and /share)

Outside of zeus HPC

  • One “Big Data Analytics” machine Bullion S2 server, 12 Tb storage, Xeon® processor E7 v3 family, 128Gb RAM

Compute nodes of new HPC part  are interconnected with High-Speed FDR Infiniband fabric (54 Gbps)

 

 

Alex Pedcenko

css.php