User Tools

Site Tools


cluster:89

This is an old revision of the document!



Back

HP HPC

Notes for the cluster design conference with HP.

“do later” means we tackle after the HP on site visit.

S & H

  • Shipping Address: 5th floor data center
  • No 13'6“ truck, 12'6” is ok or box truck
  • Delivery on standard raised dock, no ways to lift rack out of truck if not docked
  • Freight Elevator and pallet jack available

DM380G7

HP Link (head node)

  • Dual power (one to UPS, one to utility, do later)
  • hostname greentail, another local “tail”, also in reference to HP burning 18-24% more efficient in power/cooling
  • eth0, provision, 192.168.102.254/255.255.0.0 (greentail-eth0, should go to better switch ProCurve 2910)
  • eth1, data/private, 10.10.102.254/255.255.0.0 (greentail-eth1, should go to ProCurve 2610)
  • eth2, public, 129.133.1.226/255.255.255.0 (greentail.wesleyan.edu)
  • eth3, ipmi, 10.10.103.254/255.255.0.0, (greentail-ipmi, do later)
  • ib0, ipoib, 10.10.104.254/255.255.0.0 (greentail-ib0)
  • ib1, ipoib, 10.10.105.254/255.255.0.0 (greentail-ib1, configure, might not have cables!, split traffic across ports?)
  • Raid 1 mirrored disks (2x250gb)
  • /home mount point for home directory volume ~ 10tb
  • /home/apps mount point for software volume ~ 1tb (contains /home/apps/src)
  • /home/sanscratch mount point for sanscratch volume ~ 5 tb
  • logical volume LOCALSCRATCH: mount at /localscratch ~ 100 gb (should match nodes at 160 gb, leave rest for OS)
  • logical volumes ROOT/VAR/BOOT/TMP: defaults

StorageWorks MSA60

HP Link (storage device)

  • Dual power (one to UPS, one to utility, do later)
  • Three volumes to start with:
    • home (raid 6, design a backup path, do later), 10 tb
    • apps (raid 6, design a backup path, do later), 1tb
    • sanscratch (raid 1, no backup), 5 tb
  • Systems Insight Manager (SIM)
    • install, configure, monitor
    • event actions

SL2x170z G6

HP Link (compute nodes)

  • node names hp000, increment by 1
  • eth0, provision, 192.168.102.25(increment by 1)/255.255.0.0 (hp000-eth0, should go to better switch ProCurve 2910)
  • eth1, data/private, 10.10.102.25(increment by 1)/255.255.0.0 (hp000-eth1, should go to ProCurve 2610)
  • eth2, ipmi, 10.10.103.25(increment by 1)/255.255.0.0, (hp000-ipmi, do later)
  • ib0, ipoib, 10.10.104.25(increment by 1)/255.255.0.0 (hp000-ib0)
  • ib1, ipoib, 10.10.105.25(increment by 1)/255.255.0.0 (hp000-ib1, configure, might not have cables!)
  • /home mount point for home directory volume ~ 10tb
  • /home/apps mount point for software volume ~ 1tb (contains /home/apps/src)
  • /home/sanscratch mount point for sanscratch volume ~ 5 tb
  • logical volume LOCALSCRATCH: mount at /localscratch ~ 100 gb (60 gb left for OS)
  • logical volumes ROOT/VAR/BOOT/TMP: defaults

Misc

  • IPoIB
    • configuration, fine tune
    • monitor
  • Cluster Management Utility (CMU)
    • install, configure, monitor
    • golden image capture, deploy (there will initially only be one image)
  • Sun Grid Engine (SGE)
    • install, configure
    • there will only be one queue (hp12)
  • KVM utility
    • functionality
  • Placement
    • where in data center (do later), based on environmental works


Back

cluster/89.1281749668.txt.gz · Last modified: 2010/08/14 01:34 by hmeij