User Tools

Site Tools


cluster:189

Warning: Undefined array key -1 in /usr/share/dokuwiki/inc/html.php on line 1458

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
cluster:189 [2020/02/27 13:12]
hmeij07
cluster:189 [2024/02/12 11:47] (current)
hmeij07 [Priority Policy]
Line 24: Line 24:
 Several months later a pattern emerged.  The Provost would annually contribute $25K **//if//** the HPCC user base raised $15K annually in contributions.  These funds would "roll over". That would amount to $160K in 4 years, enough for a hardware refresh or new hardware acquisition.  Finances also contributed $10K annually for maintenance such as failed disks, network switches, etc ... but these funds do not "roll over". Use it or loose it. All fund cycles restart July 1st. Several months later a pattern emerged.  The Provost would annually contribute $25K **//if//** the HPCC user base raised $15K annually in contributions.  These funds would "roll over". That would amount to $160K in 4 years, enough for a hardware refresh or new hardware acquisition.  Finances also contributed $10K annually for maintenance such as failed disks, network switches, etc ... but these funds do not "roll over". Use it or loose it. All fund cycles restart July 1st.
  
-In order for the HPCC user base to raise $15K annually, CPU and GPU hourly usage monitoring was deployed (using Openlava ''lsacct'' accounting). A dictionary is maintained listing PIs with their associated members (student majors, lab students, grads, phd candidates, collaborators, etc).  Each PI then quarterly contributes to the user fund based on a scheme supposedly yielding $15K annually.+In order for the HPCC user base to raise $15K annually, CPU and GPU hourly usage monitoring was deployed (using scripts parsing the ''lsb.acct'' file). A dictionary is maintained listing PIs with their associated members (student majors, lab students, grads, phd candidates, collaborators, etc).  Each PI then quarterly contributes to the user fund based on a scheme supposedly yielding $15K annually.
  
 Here is queue usage for 2019 [[cluster:188|2019 Queue Usage]] and below is listed  the 2019 contribution scheme. Here is queue usage for 2019 [[cluster:188|2019 Queue Usage]] and below is listed  the 2019 contribution scheme.
Line 39: Line 39:
 A gpu hour of usage is 3x the cpu hourly rate.\\ A gpu hour of usage is 3x the cpu hourly rate.\\
  
-We currently have about 1,450 physical cpu cores (all Xeon), 72 gpus (20x K20, 4x GTX2018Ti, 48x RTX2080S), 520 gb of gpu memory and 8,560 gb of cpu memory. Provided by about 120 compute nodes and login nodes. Scratch spaces are provided local to compute nodes (2-5 tb) or over the network via NFS (55 tb), consult [[cluster:142|Scratch Spaces]]. Home directories are under quota (10 tb) but these will disappear in the future with the TrueNAS/ZFS appliance (190 tb, 475 tb effective assuming a compression rate of 2.5x, consult [[cluster:186|Home Dir Server]] deploy in 2020). A HPCC guide can be found here [[cluster:126|Brief Guide to HPCC]] and the (endless!) software list is located here [[cluster:73|Software Page]].  We run CentOS 6.10 or 7.6 flavors of OS.+We currently have about 1,450 physical cpu cores (all Xeon), 72 gpus (20x K20, 4x GTX2018Ti, 48x RTX2080S), 520 gb of gpu memory and 8,560 gb of cpu memory. Provided by about 120 compute nodes and login nodes. Scratch spaces are provided local to compute nodes (2-5 tb) or over the network via NFS (55 tb), consult [[cluster:142|Scratch Spaces]]. Home directories are under quota (10 tb) but these will disappear in the future with the TrueNAS/ZFS appliance (190 tb, 475 tb effective assuming a compression rate of 2.5x, consult [[cluster:186|Home Dir Server]] deploy in 2020). A HPCC guide can be found here [[cluster:126|Brief Guide to HPCC]] and the (endless!) software list is located here [[cluster:73|Software Page]].  We run CentOS 6.10 or 7.[6|7] flavors of OS.
  
  
 ==== Priority Policy ==== ==== Priority Policy ====
  
-This policy was put in place about 3 years ago to deal with the issues surrounding new monies infusions from for example; new faculty "startup monies", new grant monies (NSF, NIH, DoD, others), or donations made to the HPCC for a specific purpose (as in GTX gpus for Amber). All users have the same priority. All queues have the same priority (except the "test" queue which has the highest priority). Scheduler policy is FIFO overlayed by Round Robin. There is no "wall time" on any queue.+This policy was put in place about 3 years ago (2017) to deal with the issues surrounding new monies infusions from for example; new faculty "startup monies", new grant monies (NSF, NIH, DoD, others), or donations made to the HPCC for a specific purpose (as in GTX gpus for Amber). All users have the same priority. All queues have the same priority (except the "test" queue which has the highest priority). Scheduler policy is FIFO. There is no "wall time" on any queue.
  
 There are few Principles in this Priority Access Policy There are few Principles in this Priority Access Policy
cluster/189.1582827138.txt.gz ยท Last modified: 2020/02/27 13:12 by hmeij07