cluster:110
Differences
This shows you the differences between two versions of the page.
| Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
| cluster:110 [2013/04/15 17:29] – [Specs: EC GPU] hmeij | cluster:110 [2013/05/24 13:39] (current) – [Specs: MW - GPU] hmeij | ||
|---|---|---|---|
| Line 21: | Line 21: | ||
| ===== Round 3 ===== | ===== Round 3 ===== | ||
| + | |||
| + | |||
| + | |||
| + | ==== Specs: MW - GPU ==== | ||
| + | |||
| + | This is what we ended up buying May 2013. | ||
| + | |||
| + | ^ Topic^Description | ||
| + | | General| 10 CPUs (80 cores), 20 GPUs (45,000 cuda cores), 256 gb ram/node (1,280 gb total), plus head node (128 gb)| | ||
| + | | Head Node|1x42U Rackmount System (36 drive bays), 2xXeon E5-2660 2.0 Ghz 20MB Cache 8 cores (total 16 cores)| | ||
| + | | |16x16GB 240-Pin DDR3 1600 MHz ECC (total 256gb, max 512gb), ? | ||
| + | | |2x1TB 7200RPM (Raid 1) + 16x3TB (Raid 6), Areca Raid Controller| | ||
| + | | |Low profile graphics card, ConnectX-3 VPI adapter card, Single-Port, | ||
| + | | |1400w Power Supply 1+1 redundant| | ||
| + | | Nodes|5x 2U Rackmountable Chassis, 5x 2 Xeon E5-2660 2.0 Ghz 20MB Cache 8 cores (16 cores/ | ||
| + | | |5x 16x16GB 240-Pin DDR3 1600 MHz (256gb/node memory, max 256gb)| | ||
| + | | |5x 1x120GB SSD 7200RPM, 5x 4xNVIDIA Tesla K20 5 GB GPUs (4/node), 1CPU-2GPU ratio| | ||
| + | | |? | ||
| + | | |5xConnectX-3 VPI adapter card, Single-Port, | ||
| + | | |5x1620W 1+1 Redundant Power Supplies| | ||
| + | | Network|1x 1U Mellanox InfiniBand QDR Switch (18 ports)& HCAs (single port) + 3m cable QDR to existing Voltaire switch| | ||
| + | | |1x 1U 24 Port Rackmount Switch, 10/ | ||
| + | |Rack |1x42U rack with power distributions (14U used)| | ||
| + | | Power|2xPDU, | ||
| + | | Software| CentOS, Bright Cluster Management (1 year support), MVAPich, OpenMPI, CUDA| | ||
| + | | | scheduler and gnu compilers installed and configured| | ||
| + | | | Amber12 (customer provide license) , Lammps, NAMD, Cuda 4.2 (for apps) & 5 | | ||
| + | | Warranty|3 Year Parts and Labor (lifetime technical support)| | ||
| + | | GPU Teraflops|23.40 double, 70.40 single| | ||
| + | | Quote|< | ||
| + | |Includes | ||
| + | |||
| + | |||
| + | * 16U - estimated draw 6,900 Watts and 23,713 BTUs cooling - $30K/year | ||
| + | * 5 GPU shelves | ||
| + | * 2 PDUs | ||
| + | * 42 TB raw | ||
| + | * FDR interconnects | ||
| + | * 120GB SSD drives on nodes | ||
| + | * 256 gb ram on nodes, 16gb/core | ||
| + | * Areca hardware raid | ||
| + | * Lifetime technical support | ||
| ==== Specs: EC GPU ==== | ==== Specs: EC GPU ==== | ||
| Line 26: | Line 68: | ||
| ^ Topic^Description | ^ Topic^Description | ||
| - | | General| 12 CPUs (96 cores), 20 GPUs (45,000 cuda cores), 128 gb ram/node, plus head node (128gb)| | + | | General| 12 CPUs (96 cores), 20 GPUs (45,000 cuda cores), 128 gb ram/ |
| | Head Node|1x2U Rackmount System, 2xXeon E5-2660 2.20 Ghz 20MB Cache 8 cores| | | Head Node|1x2U Rackmount System, 2xXeon E5-2660 2.20 Ghz 20MB Cache 8 cores| | ||
| | |8x16GB 240-Pin DDR3 1600 MHz ECC (128gb, max 512gb), 2x10/ | | |8x16GB 240-Pin DDR3 1600 MHz ECC (128gb, max 512gb), 2x10/ | ||
| Line 42: | Line 84: | ||
| | Software| CentOS, Bright Cluster Management (1 year support)| | | Software| CentOS, Bright Cluster Management (1 year support)| | ||
| | | Amber12 (cluster install), Lammps (shared filesystem), | | | Amber12 (cluster install), Lammps (shared filesystem), | ||
| - | | Storage| | + | | Storage|3U 52TB Disk Array (28x2TB) Raid 6, cascade cable| |
| | Warranty|3 Year Parts and Labor (EC technical support? | | Warranty|3 Year Parts and Labor (EC technical support? | ||
| | GPU Teraflops|23.40 double, 70.40 single| | | GPU Teraflops|23.40 double, 70.40 single| | ||
| Line 48: | Line 90: | ||
| + | * 20U - estimated draw 7,400 Watts - $30K/year for cooling and power | ||
| * 5 GPU shelves | * 5 GPU shelves | ||
| * 1 CPU shelf | * 1 CPU shelf | ||
| - | * 4 PDUs! | + | * 4 PDU - this could be a problem! |
| - | * 56TB raw | + | * 56TB raw |
| + | * QDR interconnects | ||
| + | * 1 TB disk on node, makes for a large / | ||
| * LSI hardware raid card | * LSI hardware raid card | ||
cluster/110.1366046941.txt.gz · Last modified: by hmeij
