User Tools

Site Tools


cluster:123

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
cluster:123 [2013/10/10 14:08]
hmeij [Replace Dell Racks]
cluster:123 [2013/10/10 15:21]
hmeij
Line 9: Line 9:
 Once upon a time, back in 2013,  two Dell racks full of compute nodes, sat noisily chewing away energy on the 5th floor of Science Tower.  They sucked in nicely cooled air from the floor spewing it out the back of the racks at 105-110 degrees (F).  They were giving the three Liebert cooling towers a run for their BTUs.  So much so that if one failed the Dell racks needed to be powered down to avoid the data center reaching temperatures beyond 95 degrees (F). The Dell racks were in a foul mood since that last event not too long ago. And so, day after day, they consumed lots of BTUs, and with the ample supply of Watts coming from their L6-30 roots, converted it all into heat. Tons of heat, making life lousy for the Liebert family. Oh, and they performed some computational work too, but if even they did not, the energy consumption remained the same. That's a fact. They were 6 years old and determined to make it to 12. So the story goes. Once upon a time, back in 2013,  two Dell racks full of compute nodes, sat noisily chewing away energy on the 5th floor of Science Tower.  They sucked in nicely cooled air from the floor spewing it out the back of the racks at 105-110 degrees (F).  They were giving the three Liebert cooling towers a run for their BTUs.  So much so that if one failed the Dell racks needed to be powered down to avoid the data center reaching temperatures beyond 95 degrees (F). The Dell racks were in a foul mood since that last event not too long ago. And so, day after day, they consumed lots of BTUs, and with the ample supply of Watts coming from their L6-30 roots, converted it all into heat. Tons of heat, making life lousy for the Liebert family. Oh, and they performed some computational work too, but if even they did not, the energy consumption remained the same. That's a fact. They were 6 years old and determined to make it to 12. So the story goes.
  
-The Dell racks contain 30 compute nodes, two UPS units, two disks arrays and two switches. We have measured 19 nodes power consumption (pulling one the dual power units out) with a Kill-A-Watt meter for over 775+ hours. The mean power consumption rate is 418.4 watts. That totals to 109,956 KwH/year in power consumption ((watts/1000 Kw per hour) * 24 hours * 365 days * 30 servers). This is a low water mark, it only takes into account thwe compute nodes but that will be the majorioty of energy consumption. We also measured one rack's consumption at the utility panel and Peter's caculation yields 126,000 KwH/year which can be considered a hig water mark+The Dell racks contain 30 compute nodes, two UPS units, two disks arrays and two switches. We have measured 19 nodes power consumption (pulling one the dual power units out) with a Kill-A-Watt meter for over 775+ hours. The mean power consumption rate is 418.4 watts. That totals to 109,956 KwH/year in power consumption ((watts/1000 Kw per hour) * 24 hours * 365 days * 30 servers). This is a low water mark, it only takes into account thwe compute nodes but that will be the majorioty of energy consumption. We also measured one rack's consumption at the utility panel and Peter's caculation yields 126,000 KwH/year which can be considered a high water mark.
- +
  
-Based on 12.5 cents per Kwh (this is an all inclusive cost including natural gas cost, heat recoup costs, distribution, maintenance etc) the hardware burns away $13,744.50 per year. Best guess is cooling costs are at least that (another possible low water mark) so the total cost for all power consumption is $27,489 per year. If we run the hardware for another 3 years that total cost is $82,467.+Next we need to convert to a dollar value A residential electric bill contains a KwH cost as well as generation, distribution, transmission and other cost factors Typically the total KwH cost is 50% of the total bill.  Utility power in the data center comes from our cogen plant as well as CP&L.  The model for cogen attempts to balance both; 15% of CP&L utility and 85% cogen generated. If the cogen is downlike for maintenance this jumps to 100% CP&L. One also needs to factor in natural gas costs, heat reclamation costs, etc. If we set the cogen KwH cost at 6.25 cents we'll need to double that to 12.5 cents and use that.
  
- +Based on 12.5 cents per Kwh (all costs included) the Dell compuet nodes consumed $13,744.50 per year in power. Best guess is cooling costs are at least that (another possible low water mark) so the total cost for both  power and colling consumption is $27,489 per year.
  
-If we could replace, or approximate, the 30 compute nodes' computational power (0.6 teraflops) and job slots (240 cores) with new hardware that consumes 50% less in power, our ROI is 6 years based on the low water mark numbers.  
- 
- 
- 
-So, I've collected quotes and they are rolling in.  I've taken $82K target as a budget so I can downscale from there. Here is a comparison 
  
 +Next step was to collect quotes for a target budget of $82K, 3 years of Dell energy consumption, an arbitrary length of time. That's so we can downscale from there because the new racks ofcourse still consume energy. Four quotes were obtained and they show a similar pattern. Here is the comparison on key features.
    
- +Old hardware: 
-old racks: 30 nodes, 240 job slots12,552 watts, 670 gigaflops (a measure of computational performance) +30 nodes, 2.66 ghz4 mb L-cache (for cpu), 240 cores (job slots)
- +80 gb local drive340 gb total ram12,555 watts (power no cooling)670 gigaflops (actual measure)
-vendor 1: 14 nodes224 job slots, 5,400 watts, 4,659 gigaflops +
- +
-vendor 2: 10 nodes200 job slots4,597 watts, 4,480 gigaflops +
    
- +New hardware: 
-So we can do it with about 40% of the energy consumption. And I can do it with half the number of nodes listed because with hyperthreading the number of job slots can be doubled (just need more memory) Our conclusion was " all power consumption is $27,489 per year". That means $55K is enough and our ROI is 2 years. +14 nodes, 2.60 ghz, 20 mb L-cache (for cpu), 224 cores (job slots)
 +1TB local drive, 1,792 gb  total ram, 5,400 watts (power no cooling)4,659 gigaflops (theoretical)
    
 +In the representative example total energy consumption would be 10,800 watts.  If we reduced the node count to 7 the total energy consumption would be 5,400 watts (47,304 KwH/year) or 43% of the Dell's power consumption (does include cooling). And that's against the low water mark.
  
-My intention is to convince the faculty of this Do you agree?  Then we can take it to Dave Baird.+The total cost of running the new hardware would be $5,913. That would imply a savings rate of $21,576 per yearThe job slot count would be 112 but with hyperthreading technology that can be doubled. We'd still want the 1,792 memory footprint (8 gb/core) and the gigaflops still far exceeds Dell's performance.
  
-  +Within two years, the new hardware (about $41Kwill have paid for itself based on the low water mark (Dell'cost would equal $55K). We still need to adjust some minor issues:
- +
--Henk +
- +
-  +
-From: Meij, Henk +
-Sent: Wednesday, October 02, 2013 12:10 PM +
-To: Staye, Peter +
-Subject: RE: Meeting with Peter 01Oct13 Summary +
- +
-ohbtw the nodes are on 208V utility power, makes no difference in our calculations though. +
- +
--Henk +
- +
-  +
-From: Meij, Henk +
-Sent: Wednesday, October 02, 2013 11:42 AM +
-To: Staye, Peter +
-Subject: Meeting with Peter 01Oct13 Summary +
- +
-My attempt at summarizing our conversation. Make corrections if I got things wrong. +
- +
-  +
- +
--Henk +
- +
-  +
- +
- +
- will save y'll the details but after many measurements and meeting with Peter Staye (physical plant) we are both in agreement that the dell 30 compute nodes "total cost for all energy consumption is $27,489 per year" (power+cooling)+
-  +
-Next step was to collect quotes for a target budget of $82K, 3 years of dell energy consumption.  That's just so that I can downscale from there because the new racks ofcourse still consume energy.  But the details are striking. (I'm getting 4 quotes but the first 3 in are pretty close). +
-  +
-dell: +
-30 nodes, 2.66 ghz, 4 mb L-cache (for cpu), 240 cores, +
-80 gb local drive, 340 gb total ram, 12,555 watts (power no cooling), 670 gigaflops (actual measure) +
-  +
-microway: +
-14 nodes, 2.60 ghz, 20 mb L-cache (for cpu), 224 cores, +
-1TB local drive, 1,792 gb  total ram, 5,400 watts (power no cooling), 4,659 gigaflops (theoretical) +
-  +
-That means microway hardware burns but 40% of the dell'power (not cooling). +
-  +
-If we take half the microway nodes, turn hyperthreading on (like on queue mw256, doubling the core count), raise the memory to 256gb/node (8gb/core) we would be very close to dell's job slot offering (224 vs 240), have 5x more memory, and 7x more gigaflops. +
-  +
-That will easily fit within $55K (two year of Dell power costsand obtain a ROI of 2 years Actually, within 2 years but a max of 2 years. +
-  +
-Peter Staye agrees.  +
  
 +  * there is enough Infiniband ports available to put all new nodes on such a switch (add cards and cables cost)
 +  * the internal disks need to be of higher speed (10K or better) of a larger size (300 GB or larger) mimicking the Dell disk arrays (add costs)
  
 +But it is all very doable within budget of $45-$55K 
  
 The end. The end.
cluster/123.txt · Last modified: 2013/10/23 18:52 by hmeij