User Tools

Site Tools


cluster:151

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
cluster:151 [2016/11/29 16:20]
hmeij07 [Resync Data #2]
cluster:151 [2016/12/01 14:38]
hmeij07 [upgrade]
Line 6: Line 6:
 A document for me to recall and make notes of what I read in the manual pages and what needs testing. A document for me to recall and make notes of what I read in the manual pages and what needs testing.
  
-Basically during the Summer of 2016 I investigated if the HPCC could afford enterprise level storage. I wanted 99.999% uptime, snapshots, high availability and other goodies such as parallel NFS. Netapp came the closest but, eh, still at $42K lots of other options show up. The story is detailed here at [[cluster:149|The Storage Problem]]+Basically during the Summer of 2016 I investigated if the HPCC could afford enterprise level storage. I wanted 99.999% uptime, snapshots, high availability and other goodies such as parallel NFS. Netapp came the closest but, eh, still at $42K lots of other options show up. That story is detailed at [[cluster:149|The Storage Problem]]
  
 This page is best read from the bottom up. This page is best read from the bottom up.
  
-==== cluster idea ====+==== beeGFS cluster idea ====
  
-  * Storage servers: buy 2 now 4k+4k then 3rd in July 4k?+  * Storage servers:  
 +    * buy 2 with each 12x2TB slow disk, Raid 6, 20T usable (clustered, parallel file system) 
 +      * create 2 6TB volumes on each, quota at 2TB via XFS, 3 users/server  
 +      * only $HOME changes to ''/mnt/beegfs/home[1|2]'' (migrates ~4.5TB away from /home or ~50%) 
 +      * create 2 buddymirrors; each with primary on one, secondary on the other server (high availability) 
 +    * on UPS 
 +    * on Infiniband
  
-  * move test users over on 2 nodes, test, only change is $HOME+  * Client servers: 
 +    * all compute/login nodes become beegfs clients
  
-  * Home cluster +  * Meta servers: 
-    * cottontail (mngt+admingiu+    * cottontail2 (root meta, on Infiniband) plus n38-n45 nodes (on Infiniband) 
-    * 2-3 new units storage (+snapshots/meta backup+    * all mirrored (total=9) 
-    * cottontail2 meta + n38-n45 metaall mirrored+    * cottontail2 on UPS  
 + 
 +  * Management and Monitor servers 
 +    * cottontail (on UPS, on Infiniband) 
 + 
 +  * Backups (rsnapshot.org via rsync daemons [[cluster:150|Rsync Daemon/Rsnapshot]]
 +    * sharptail:/home --> cottontail 
 +    * serverA:/mnt/beegfs/home1 --> serverB (8TB max
 +    * serverB:/mnt/beegfs/home2 --> serverA (8TB max) 
 + 
 +  * Costs (includes 3 year NBD warranty) 
 +    * Microway $12,500 
 +    * CDW 
  
 ==== beegfs-admin-gui ==== ==== beegfs-admin-gui ====
  
   * ''cottontail:/usr/local/bin/beegfs-admin-gui''   * ''cottontail:/usr/local/bin/beegfs-admin-gui''
 +
 +==== upgrade ====
 +
 +  * [[http://www.beegfs.com/content/updating-upgrading-and-versioning/|External Link]]
 +  * New feature - High Availability for Metadata Servers (self-healing, transparent failover)
 +
 +A bit complicated. 
 +
 +  * Repo base URL baseurl=http://www.beegfs.com/release/beegfs_6/dists/rhel6 via http shows only 6.1-el6
 +    * [   ] beegfs-mgmtd-6.1-el6.x86_64.rpm          2016-11-16 16:27  660K 
 +  * '' yum --disablerepo "*" --enablerepo beegfs repolist'' shows
 +    * beegfs-mgmtd.x86_64                            2015.03.r22-el6            beegfs
 +  * ''yum install --disablerepo "*" --enablerepo beegfs --downloadonly --downloaddir=/sanscratch/tmp/beegfs/beegfs_6/ *x86_64* -y''
 +   * http://www.beegfs.com/release/beegfs_6/dists/rhel6/x86_64/beegfs-mgmtd-2015.03.r22-el6.x86_64.rpm: [Errno 14] PYCURL ERROR 22 - "The requested URL returned error: 404 Not Found" <-- wrong package version
 +
 +
 +
 +
 +
  
 ==== Resync Data #2 ==== ==== Resync Data #2 ====
Line 128: Line 166:
    21701         low     291.2GiB     114.9GiB  39%       18.5M       16.1M  87%    21701         low     291.2GiB     114.9GiB  39%       18.5M       16.1M  87%
    21702         low     291.2GiB     114.9GiB  39%       18.5M       16.1M  87%    21702         low     291.2GiB     114.9GiB  39%       18.5M       16.1M  87%
 +
 +# first rsync drops roughly 5G in both primaries which then get copied to secondaries.
 +# second rsync does the same so both storage servers loose 20G roughly
 +# now shut a storage server down and the whole filesystem can still be accessed (HA)
  
 </code>  </code> 
Line 465: Line 507:
   * made easy [[http://www.beegfs.com/wiki/ManualInstallWalkThrough|External Link]]   * made easy [[http://www.beegfs.com/wiki/ManualInstallWalkThrough|External Link]]
   * rpms pulled from repository via petaltail in ''greentail:/sanscratch/tmp/beegfs''   * rpms pulled from repository via petaltail in ''greentail:/sanscratch/tmp/beegfs''
 +    * ''yum --disablerepo "*" --enablerepo beegfs list available''
 +    * use ''yumdownloader''
  
 <code> <code>
cluster/151.txt · Last modified: 2016/12/06 20:14 by hmeij07