cluster:166
Back
HPC Users Meeting
Brief history
2006 swallowtail (Dell PE1955, Infiniband, imw, emw)
2010 greentail (HP gen6 blade servers, hp12)
2013 sharptail (Microway storage, K20s, Infiniband, mw256/mwgpu)
2014 mw256fd (Dell 2006 replacement with Supermicro nodes)
2015 tinymem (Supermicro bare metal, expansion for serial jobs)
2017 mw128 (first new faculty startup funds)
2018 6/25 Today's meeting
Since 2006
Grown from 256 to roughly 1,200 physical CPU cores
Processed 3,165,752 jobs (by 18jun2018)
Compute capacity over 60 teraflops (DPFP; 38 cpu side, 25 gpu side)
Total memory footprint is near 7.5 TB
About 500 accounts have been created (incl 22 collaborator and 100 class accounts)
Tidbits
Bought deep U42 rack with AC cooling onboard and two PDUs
Pushed Angstrom rack (bss24) out of our area, ready to recycle that (Done. 06/20/2018)
Currently we have two U42 racks empty with power
Cooling needs to be provided with any new major purchases (provost, ITS, HPC?)
60 TB raw storage purchased for sharptail (/home2 for users with specific needs)
Everything is out of warranty but
All Infiniband ports are in use
Notes
First make a page comparing CPU vs GPU usage which may influence future purchase
CPU vs GPU
$100k quote, 3to5 vendors, data points mid-2018
One node (or all) should have configured on it: amber, gromacs, laamps, namd, latest version
Nvidia latest version, optimal configs cpu:gpu ratios
Amber 1:1 (may be 1:2 in future releases) - amber certified GPU!
Gromacs 10:1 (could ramp up to claiming all resources per node)
Namd 13:1 (could ramp up to claiming all resources per node)
Lammps 2-4:1
128g with enough CPU slots to take over hp12
: double ten core minimum
Anticipated target (also to manage heat exchange)
Back
cluster/166.txt · Last modified: 2018/06/27 11:51 by hmeij07