User Tools

Site Tools


cluster:77

Warning: Undefined array key -1 in /usr/share/dokuwiki/inc/html.php on line 1458

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
cluster:77 [2009/06/15 10:18]
hmeij
cluster:77 [2009/09/09 09:21] (current)
hmeij
Line 5: Line 5:
  
  
-===== Migration/Expansion =====+===== Expansion =====
  
-Mental Notes+**[[http://www.connecticutplus.com/cplus/information/news/education/Blue-Sky-Studios-donating-animation-computers-to-UConn-Wesleyan53795379.shtml|Donation of hardware]] by [[http://www.blueskystudios.com|Blue Sky Studios]]** 
 + 
 + 
 +4 racks with blade servers, 52 servers per rack, will be dropped off tomorrow.  Not all will be turned on as 5% of the servers are kaput.  So the first step is to build a fully working rack.  Memory footprint of the servers is either 12 or 24 GB.  Chipset is AMD Operon Model 250 single core dual CPUs at 1.0 GHZ. 
 + 
 +Since there is no budget we will be installing CentOS with Lava as the scheduler.  More details to come. 
 + 
 +===== Pictures ===== 
 + 
 +Two Angstrom racks on each side of Dell racks.   
 + 
 +The closest two racks each hold 52 servers, each with dual AMD Opteron 248 model chips running at 1.0 Ghz with a memory foot print of 12 GB and holds a single 80 GB disk and a dual port NIC.  Each rack holds two HP Procurve 2848 gigabit switches. Memory wise each rack holds 52*12 = 624 GB.  
 + 
 +The two racks on the other side are similar as the front two, except the memory footprint of each server is 24 GB.  That means each rack holds 52*24 = 1,248 GB.  So all four Angstrom racks hold 208 servers that can access a total of 3,744 GB. Compare that to the Dell cluster, which in July 2009, holds a total of 320 GB. The Dell racks offer 288 job slots and the Angstrom could offer 416 job slots. 
 + 
 +Ha! But not so fast. Cooling capacity wise we could not turn them on today, and electric consumption would be critical. Only one rack holding the 12 GB servers servers is currently slated to be turned into a cluster while one of the 24 GB servers racks is slated to become a virtualization farm for an ITS pilot project. 
 + 
 +The rest are for now spare parts. And 5% of the servers have something broken. 
 + 
 +But they are pretty cute looking racks. Sort of animated. Thank You Blue Sky (**[[http://www.connecticutplus.com/cplus/information/news/education/Blue-Sky-Studios-donating-animation-computers-to-UConn-Wesleyan53795379.shtml|Donation of hardware]] by [[http://www.blueskystudios.com|Blue Sky Studios]]**) 
 + 
 +{{:cluster:angstrom_racks2.jpg|}} 
 + 
 + 
 +{{:cluster:angstrom_back2.jpg|}} 
 + 
 +===== Migration ===== 
 + 
 +Mental Notes pour mois ...
  
   * no need to block ssh access to swallowtail or petal tail   * no need to block ssh access to swallowtail or petal tail
Line 15: Line 43:
   * make sure dhcpd is stopped on swallowtail   * make sure dhcpd is stopped on swallowtail
   * remove compute-01   * remove compute-01
-  * back 'good files'+  * back up 'good files'
  
-  * halt jobs, halt compute nodes +  * release nodes on swallowtail 
-    * scp over lsb.accts files from lsf_archive +  * PXEboot 
-    * also the disk and queue log files+  * add to new cluster 
 +  * custom image 
 + 
 +  * scp over lsb.accts files from lsf_archive 
 +  * also the disk and queue log files 
 +  * had to stop|start lsf_daemons to get LSF going
  
   * reroute the power cords on side of rack   * reroute the power cords on side of rack
   * free ports 47*48 on both ethernet switches   * free ports 47*48 on both ethernet switches
- 
-  * reimage nodes using petaltail 
  
   * reconfigure LSF queues (&licenses, remove & add)   * reconfigure LSF queues (&licenses, remove & add)
cluster/77.1245075484.txt.gz · Last modified: 2009/06/15 10:18 by hmeij