User Tools

Site Tools


cluster:125

This is an old revision of the document!



Back

Dell Racks power off

Soon (Feb/2014), we'll have to power down the Dell Racks and grab one L6-30 circuit supplying power to those racks and use it to power up the new Microway servers.

That leaves some spare L6-30 circuits (the Dell racks use 4 each), so we could contemplate grabbing two and powering up two more shelves of the Blue Sky Studio hardware. That would double the Hadoop cluster and the bss24 queue when needed (total of 100 job slots), and offer access to 1.2 TB of memory. This hardware is generally powered off when not in use.

The new Microway hardware is identical to the GPU-HPC hardware we bought minus the GPUs. A total of 8 1U servers will offer

  • 256 GB of memory per node (2,048 GB total … that's amazing because if you add the GPU nodes memory footprint, the total for the rack becomes 3,328 GB in 18U of rack space).
  • dual 8-core Intel Xeon chips with hyperthreading turned on, so each node presents 32 cores for a toal of 256 cores (job slots). These will be presented as queue mw256f
  • Each core is capable of doing 8 instructions per clock cycle and each core will have access to an average of 8 GB of memory
  • Each node also has a 300 GB 15K RPM hard disk holding the operating system, swap and provides for a /localscratch of 175 GB, hence the fd of mw256fd queue name. It is to be used just like ehwfd.
  • Each node is Infiniband enabled (meaning all our nodes are except the Blue Sky Studio (queue bss24) and /home and /sanscratch are served IPoIB.


Back

cluster/125.1392318963.txt.gz · Last modified: 2014/02/13 14:16 by hmeij