This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision Next revision Both sides next revision | ||
cluster:89 [2010/08/18 14:52] hmeij |
cluster:89 [2010/08/19 15:20] hmeij |
||
---|---|---|---|
Line 23: | Line 23: | ||
* voltaire console can be stuffed in either | * voltaire console can be stuffed in either | ||
- | * head node will be connected to our private network via | + | * head node will be connected to our private network via a two link aggregated ethernet cables in the 10.10.x.y range so current home directories can be mounted somewhere (these dirs will not be available on the back end nodes. |
- | * a two link aggregated ethernet cables in the 10.10.x.y | + | |
- | * range so current home directories can be mounted | + | |
- | * somewhere (these dirs will not be available on the | + | |
- | * back end nodes. | + | |
* x.y.z.255 is broadcast | * x.y.z.255 is broadcast | ||
Line 55: | Line 51: | ||
* Raid 1 mirrored disks (2x250gb) | * Raid 1 mirrored disks (2x250gb) | ||
- | * /home mount point for home directory volume ~ 10tb | + | * /home mount point for home directory volume ~ 10tb (contains / |
- | * /home/apps mount point for software volume ~ 1tb (contains / | + | * /snapshot mount point for snapshot volume ~ 10tb |
- | * /home/sanscratch mount point for sanscratch volume ~ 5 tb | + | * /sanscratch mount point for sanscratch volume ~ 5 tb |
* logical volume LOCALSCRATCH: | * logical volume LOCALSCRATCH: | ||
* logical volumes ROOT/ | * logical volumes ROOT/ | ||
Line 72: | Line 68: | ||
* Three volumes to start with: | * Three volumes to start with: | ||
- | * home (raid 6, design a backup path, do later), 10 tb | + | * home (raid 6), 10 tb |
- | * apps (raid 6, design a backup path, do later), 1tb | + | * snapshot |
- | * sanscratch (raid 1, no backup), 5 tb | + | * sanscratch (raid 1 or 0, no backup), 5 tb |
* SIM | * SIM | ||
Line 92: | Line 88: | ||
* ib1, ipoib, 10.10.104.25(increment by 1)/ | * ib1, ipoib, 10.10.104.25(increment by 1)/ | ||
- | * /home mount point for home directory volume ~ 10tb | + | * /home mount point for home directory volume ~ 10tb (contains / |
- | * /home/apps mount point for software volume ~ 1tb (contains / | + | * /snapshot mount point for snapshot volume ~ 10tb |
- | * /home/sanscratch mount point for sanscratch volume ~ 5 tb | + | * /sanscratch mount point for sanscratch volume ~ 5 tb |
+ | * (next ones must be 50% empty for cloning to work) | ||
* logical volume LOCALSCRATCH: | * logical volume LOCALSCRATCH: | ||
* logical volumes ROOT/ | * logical volumes ROOT/ | ||
Line 106: | Line 103: | ||
* monitor | * monitor | ||
- | * Systems Insight Manager (SIM) [[http:// | + | * Systems Insight Manager (SIM) |
+ | * [[http:// | ||
* Do we need a windows box (virtual) to run the Central Management Server on? | * Do we need a windows box (virtual) to run the Central Management Server on? | ||
* SIM + Cluster Monitor (MSCS)? | * SIM + Cluster Monitor (MSCS)? | ||
Line 115: | Line 113: | ||
* configure automatic event handling | * configure automatic event handling | ||
- | * Cluster Management Utility (CMU)[[http:// | + | * Cluster Management Utility (CMU up to 4,096 nodes) |
- | * iLo/IPMI | + | * [[http:// |
* HP iLo probably removes the need for IPMI, consult [[http:// | * HP iLo probably removes the need for IPMI, consult [[http:// | ||
- | | + | |
+ | * hmm, we can power up/off via CMU so perhaps IPMI is not needed nor this ability via SIM and web browser | ||
* is head node the Management server? possibly, needs access to provision and public networks | * is head node the Management server? possibly, needs access to provision and public networks | ||
* we may need a iLo eth? in range ... 192.198.104.x? | * we may need a iLo eth? in range ... 192.198.104.x? | ||
Line 125: | Line 124: | ||
* install X and CMU GUI client node | * install X and CMU GUI client node | ||
* start CMU, start client, scan for nodes, build golden image | * start CMU, start client, scan for nodes, build golden image | ||
+ | * install monitoring client when building golden image node via CMU GUI | ||
* clone nodes, deploy management agent on nodes | * clone nodes, deploy management agent on nodes | ||
- | | + | |
+ | * pre_reconf.sh (/ | ||
+ | * not sure we can implement CMU HA | ||
* Sun Grid Engine (SGE) | * Sun Grid Engine (SGE) | ||
Line 140: | Line 142: | ||
* where in data center (do later), based on environmental works | * where in data center (do later), based on environmental works | ||
+ | ===== ToDo ===== | ||
+ | |||
+ | All do later. After HP cluster is up. | ||
+ | |||
+ | * Backups. | ||
+ | * Use trickery with linux and rsync to provide snapshots? [[http:// | ||
+ | * Exclude very large files? | ||
+ | * petaltail:/ | ||
+ | * or better [[http:// | ||
+ | |||
+ | * Lava. Install from source and evaluate. | ||
\\ | \\ | ||
**[[cluster: | **[[cluster: |