User Tools

Site Tools


cluster:127

This is an old revision of the document!



Back

Virtual HPCC services

Thoughts on how to create virtual compute nodes in the HPCC stack. Specifically, trying to solve the need for tiny, but many, compute nodes for the nano physic applications. Like virtual compute nodes with a single core CPU with 128 MB or less of memory running a lean Scientific Linux operating system. Here is a good introduction: http://www.slideshare.net/gpaterno1/comparing-iaas-vmware-vs-openstack-vs-googles-ganeti-28016375

Building Ganeti

I'll select Ganeti to start with as it appears the simplest to setup. I've no need for services like fail-over or migration. Just the ability to rapidly many tiny nodes from a template. It also appears that Ganeti clusters can be embedded with Openstack later.

Update: On hold, these Xen tools (virt-manager, virt-clone, virsh) are very nice, no need for Ganeti up front .

Building Xen

yum install -y --nogpgcheck xen kernel-xen \
virt-manager libvirt libvirt-python python-virtinst

chkconfig xend on

# disable selinux /etc/selinux/config

# add to xen kernel grub line some stuff
title CentOS (2.6.18-371.6.1.el5xen)
        root (hd0,0)
        kernel /xen.gz-2.6.18-371.6.1.el5 dom0_mem=2048M,max:2048M dom0_max_vcpus=1 dom0_vcpus_pin allow_unsafe 
        module /vmlinuz-2.6.18-371.6.1.el5xen ro root=/dev/VolGroup00/LogVol00 rhgb quiet 
        module /initrd-2.6.18-371.6.1.el5xen.img

# hostname fully qualified (in case we go to Ganeti later)

# edit some settings in /etc/xen/xend-config, consult links below

# turn off some services with chkconfig
# cups, iptables, ip6tables, autofs, crond, sysstat, bluetooth
# firstboot, ipmi, iscsi, iscsid

reboot

Then launch virt-manager to look at your Dom0.

Build a base clone and install what you need (like the Lava scheduler files). Give it a static IP via /etc/rc.local so that when you clone it you can send it ssh commands. Then shut it down.

Now you build a script to clone and post prep new clones build from it.

#/bin/bash

# automate this

virt-clone --original vmdemo --name bvm7 --file /var/lib/xen/images/bvm7.img
# wait 3 mins
sleep 180

virsh start bvm7
# wait 2 mins
sleep 120

# clobber vmdemo rc.local that sets up a static ip that you know rather than dhcp for automation
cd /root
scp rc.local vmdemo:/etc

ssh vmdemo "cat /etc/sysconfig/network | sed s/vmdemo/bvm7/g > /tmp/network"
ssh vmdemo "scp /tmp/network /etc/sysconfig/"

ssh vmdemo "cat /etc/sysconfig/network-scripts/ifcfg-eth0 | sed s/192.168.150.0/192.168.150.7/g > /tmp/ifcfg-eth0"
ssh vmdemo "scp /tmp/ifcfg-eth0 /etc/sysconfig/network-scripts/"

# this is the clone
ssh vmdemo reboot
# wait 2 mins
sleep 120

Testing Xen

3d Lennard-Jones melt: for 10,000 steps with 32,000 atoms
Queue, node, HT Jobs/Node, Loop Time Comment
hp12, n15, no-HT 01 jobs, 481
hp12, n15, no-HT 07 jobs, 482
hp12, n2, yes-HT 01 jobs, 470
hp12, n2, yes-HT 16 jobs, 804 known penalty
bss24, many, no-HT 01 jobs, 844 equivalent to hp12, yes-HT
bss24vm, bvm1, VM 01 jobs, 776 1 vcpu, 100 ram
bss24vm, bvm1, VM 02 jobs, 850 2 vcpus, 100 ram
bss24vm, bvm1, VM 04 jobs, 1735 4 vcpus, 512 ram
bss24vm, bvm2, VM 08 jobs, 3582 8 vcpus, 1024 ram
bss24vm, bvm1, VM 32 jobs, 13661 32 vcpus, 1024 ram
bss24vm, bvm1, VM 03 jobs, 1273 3 vcpus, 256 ram, make it 128
bss24vm, bvm1, VM 05 jobs, 2108 5 vcpus, 256 ram
bss24vm, bvm1, VM 31 jobs, 13844 31 vcpus, 21504 ram
multiple vms
bss24vm, bvm1-4, VM 4x01jobs, 1818 4x1vcpu, 4×128 ram
bss24vm, bvm1-8, VM 8x01jobs, 3745 (1 hour) 8x1vcpu, 8×128 ram
bss24vm, bvm1-6, VM 6x32jobs, 82457 (22hrs) 6x32vcpu, 6×1024 ram
optimal physical to virtual cpu ratio for best performance according to Xen
bss24, many, no HT 02 jobs, 826 equivalent to hp12, yes-HT
bss24vm, bvm2-3, VM 2x02jobs, 1708 2x2vcpus, 2×128 ram,
bss24vm, bvm2-5, VM 4x02jobs, 3497 4x2vcpus, 4×128 ram, optimal physical to virtual cpus ratio
bss24vm, bvm1-8, VM 8x02jobs, XXXX 8x2vcpus, 8×128 ram, optimal physical to virtual cpus ratio

Hmmm...

We have some very old hardware here that does not support virtualization in the BIOS (hardware level). So we use paravirtualization with nodes that contain 2 AMD Opteron Model 250 CPUs and 24 gb of memory. The latter was the reasoning for virtualizing. But it seems not to work. I can start multiple VMs (with either 1 vpcu or up to 32 vcpus) but the results are the same. When you run 2 VMs the performance is twice that of 1 VM, and so on. Odd. I can observe the processes running and being actively busy, but the penalty for running more than one is staggering. Must be an underlying hardware limitation. Too bad.

The 6vms with 32vcpus finished work after 22 hours. That would imply one of these nodes could run say 20 VMs with 20×32=640 job slots. And we have 20 servers (12,800 job slots) that could be virtualized but there is no performance gain. Those servers will finish the 12,800 jobs in the same time not being virtualized. Too bad really.

Following that thought, we could virtualize (KVM at hardware level) one of the new Asus servers; at 256 gb of memory this would yield 8,192 job slots for one node 128 mb per vcpu).


Back

cluster/127.1396878389.txt.gz · Last modified: 2014/04/07 13:46 (external edit)