User Tools

Site Tools


cluster:89

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
cluster:89 [2010/08/17 15:31]
hmeij
cluster:89 [2010/08/17 20:10]
hmeij
Line 18: Line 18:
  
 Basically ... Basically ...
 +
 +  * configure all console port switches with an IP
 +    * depending on switch IP in 192.168.102.x or 10.10.102.x
 +    * voltaire console can be stuffed in either
  
   * x.y.z.255 is broadcast   * x.y.z.255 is broadcast
Line 36: Line 40:
   * hostname [[http://www.ct.gov/dep/cwp/view.asp?A=2723&Q=325780|greentail]], another local "tail", also in reference to HP being 18-24% more efficient in power/cooling   * hostname [[http://www.ct.gov/dep/cwp/view.asp?A=2723&Q=325780|greentail]], another local "tail", also in reference to HP being 18-24% more efficient in power/cooling
   * eth0, provision, 192.168.102.254/255.255.0.0 (greentail-eth0, should go to better switch ProCurve 2910)   * eth0, provision, 192.168.102.254/255.255.0.0 (greentail-eth0, should go to better switch ProCurve 2910)
 +    * do we need a iLo eth? in range 192.168.104.254?
   * eth1, data/private, 10.10.102.254/255.255.0.0 (greentail-eth1, should go to ProCurve 2610)   * eth1, data/private, 10.10.102.254/255.255.0.0 (greentail-eth1, should go to ProCurve 2610)
   * eth2, public, 129.133.1.226/255.255.255.0 (greentail.wesleyan.edu)   * eth2, public, 129.133.1.226/255.255.255.0 (greentail.wesleyan.edu)
-  * eth3, ipmi, 192.168.103.254/255.255.0.0,  (greentail-ipmi, should go to better switch ProCurve 2910, do later)+  * eth3 (over eth2), ipmi, 192.168.103.254/255.255.0.0,  (greentail-ipmi, should go to better switch ProCurve 2910, do later) 
 +    * see discussion iLo/IPMI under CMU
   * ib0, ipoib, 10.10.103.254/255.255.0.0 (greentail-ib0)   * ib0, ipoib, 10.10.103.254/255.255.0.0 (greentail-ib0)
   * ib1, ipoib, 10.10.104.254/255.255.0.0 (greentail-ib1, configure, might not have cables!, split traffic across ports?)   * ib1, ipoib, 10.10.104.254/255.255.0.0 (greentail-ib1, configure, might not have cables!, split traffic across ports?)
Line 48: Line 54:
   * logical volume LOCALSCRATCH: mount at /localscratch ~ 100 gb (should match nodes at 160 gb, leave rest for OS)   * logical volume LOCALSCRATCH: mount at /localscratch ~ 100 gb (should match nodes at 160 gb, leave rest for OS)
   * logical volumes ROOT/VAR/BOOT/TMP: defaults   * logical volumes ROOT/VAR/BOOT/TMP: defaults
 +
 +  * IPoIB configuration
 +  * SIM configuration
 +  * CMU configuration
 +  * SGE configuration
  
 =====  StorageWorks MSA60  ===== =====  StorageWorks MSA60  =====
Line 59: Line 70:
     * sanscratch (raid 1, no backup), 5 tb     * sanscratch (raid 1, no backup), 5 tb
  
 +  * SIM
  
  
Line 67: Line 78:
     * node names hp000, increment by 1     * node names hp000, increment by 1
     * eth0, provision, 192.168.102.25(increment by 1)/255.255.0.0 (hp000-eth0, should go to better switch ProCurve 2910)     * eth0, provision, 192.168.102.25(increment by 1)/255.255.0.0 (hp000-eth0, should go to better switch ProCurve 2910)
 +      * do we need an iLo eth? in range 192.168.104.25(increment by 1)
 +      * CMU wants eth0 on NIC1 and PXEboot
     * eth1, data/private, 10.10.102.25(increment by 1)/255.255.0.0 (hp000-eth1, should go to ProCurve 2610)     * eth1, data/private, 10.10.102.25(increment by 1)/255.255.0.0 (hp000-eth1, should go to ProCurve 2610)
-    * eth2, ipmi, 192.168.103.25(increment by 1)/255.255.0.0, (hp000-ipmi, should go to better switch ProCurve 2910, do later)+    * eth2 (over eth1), ipmi, 192.168.103.25(increment by 1)/255.255.0.0, (hp000-ipmi, should go to better switch ProCurve 2910, do later) 
 +      * see discussion iLo/IPMI under CMU
     * ib0, ipoib, 10.10.103.25(increment by 1)/255.255.0.0 (hp000-ib0)     * ib0, ipoib, 10.10.103.25(increment by 1)/255.255.0.0 (hp000-ib0)
     * ib1, ipoib, 10.10.104.25(increment by 1)/255.255.0.0 (hp000-ib1, configure, might not have cables!)     * ib1, ipoib, 10.10.104.25(increment by 1)/255.255.0.0 (hp000-ib1, configure, might not have cables!)
Line 78: Line 92:
     * logical volumes ROOT/VAR/BOOT/TMP: defaults     * logical volumes ROOT/VAR/BOOT/TMP: defaults
  
 +  * SIM
  
 ===== Misc ===== ===== Misc =====
 +
 +  * IPoIB
 +    * configuration, fine tune
 +    * monitor
  
   * Systems Insight Manager (SIM) [[http://h18013.www1.hp.com/products/servers/management/hpsim/index.html?jumpid=go/hpsim|HP Link]] (Linux Install and Configure Guide, and User Guide)   * Systems Insight Manager (SIM) [[http://h18013.www1.hp.com/products/servers/management/hpsim/index.html?jumpid=go/hpsim|HP Link]] (Linux Install and Configure Guide, and User Guide)
     * Do we need a windows box (virtual) to run the Central Management Server on?     * Do we need a windows box (virtual) to run the Central Management Server on?
 +    * SIM + Cluster Monitor (MSCS)?
     * install, configure     * install, configure
     * requires an oracle install? no, hpsmdb is installed with automatic installation (postgresql)     * requires an oracle install? no, hpsmdb is installed with automatic installation (postgresql)
Line 89: Line 109:
     * configure automatic event handling     * configure automatic event handling
  
-  * IPoIB +  * Cluster Management Utility (CMU)[[http://h20338.www2.hp.com/HPC/cache/412128-0-0-0-121.html|HP Link]] (Getting Started - Hardware Preparation, Setup and Install -- Installation Guide v4.2, Users Guides) 
-    * configurationfine tune +  * iLo/IPMI 
-    * monitor +    * HP iLo probably removes the need for IPMI, consult [[http://en.wikipedia.org/wiki/HP_Integrated_Lights-Out|External Link]]do the blades have a management card? 
- +    * well maybe not, IPMI ([[http://en.wikipedia.org/wiki/Ipmi|External Link]]) can be scripted to power on/off, not sure about iLo (all web based)  
-  Cluster Management Utility (CMU) +    is head node the Management server? possibly, needs access to provision and public networks 
-    * install, configure, monitor +    * we may need a iLo eth? in range ... 192.198.104.x? Consult the Hardware Preparation Guide. 
-    * golden image capture, deploy (there will initially only be one image)+    * CMU wants eth0 on NIC1 and PXEboot 
 +    * install CMU management node 
 +    * install X and CMU GUI client node 
 +    * start CMU, start client, scan for nodes, build golden image 
 +    * 
  
   * Sun Grid Engine (SGE)   * Sun Grid Engine (SGE)
     * install, configure     * install, configure
     * there will only be one queue (hp12)     * there will only be one queue (hp12)
 +
 +===== Other =====
  
   * KVM utility   * KVM utility
cluster/89.txt · Last modified: 2010/11/22 19:05 by hmeij