User Tools

Site Tools


cluster:189

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
cluster:189 [2020/01/12 20:56]
hmeij07 [Funding Policy]
cluster:189 [2020/01/22 13:46]
hmeij07
Line 8: Line 8:
 ==== History ==== ==== History ====
  
-In 2006, 4 Wesleyan faculty members approached ITS with a proposal to centrally manage a high performance computing center (HPCC) seeding the effort with an NSF grant (about $190K, two racks full of Dell PE1950, a total of 256 physical cpu cores on Infiniband). ITS offered 0.5 FTE for a dedicated "hpcadmin". An Advisory Group was formed by these faculty plus hpcadmin (5 members, not necessarily our current "power users"). Another NSF grant reward was added in 2010 (about $105K). An alumni donation followed in 2016 (about $10K).  In 2018 the first instance of "faculty startup monies" was contributed to the HPCC (about $92K, see "Priority Policy" below). In 2019, a TrueNAS/ZFS appliance was purchased (about $40K, [[cluster:186|Home Dir Server]]) followed in 2020 by a GPU expansion project (about $96K, [[cluster:181|2019 GPU Models]]). The latter two were self-funded expenditures, see "Funding Policy" below. To view the NSF grants visit [[cluster:169|Acknowledgement]]+In 2006, 4 Wesleyan faculty members approached ITS with a proposal to centrally manage a high performance computing center (HPCC) seeding the effort with an NSF grant (about $190K, two racks full of Dell PE1950, a total of 256 physical cpu cores on Infiniband). ITS offered 0.5 FTE for a dedicated "hpcadmin". An Advisory Group was formed by these faculty plus hpcadmin (5 members, not necessarily our current "power users"). Another NSF grant reward was added in 2010 (about $105K). An alumni donation followed in 2016 (about $10K).  In 2018 the first instance of "faculty startup monies" was contributed to the HPCC (about $97.4K, see "Priority Policy" below). In 2019, a TrueNAS/ZFS appliance was purchased (about $40K, [[cluster:186|Home Dir Server]]) followed in 2020 by a GPU expansion project (about $96K, [[cluster:181|2019 GPU Models]]). The latter two were self-funded expenditures, see "Funding Policy" below. To view the NSF grants visit [[cluster:169|Acknowledgement]]
  
 The Advisory Group meets with the user base yearly during the reading week of the Spring semester (early May) before everybody scatters for the summer. At this meeting, the hpcadmin reviews the past year, previews the coming year, and the user base are contributing feedback on progress and problems. The Advisory Group meets with the user base yearly during the reading week of the Spring semester (early May) before everybody scatters for the summer. At this meeting, the hpcadmin reviews the past year, previews the coming year, and the user base are contributing feedback on progress and problems.
Line 62: Line 62:
 ==== General ==== ==== General ====
  
-There are 557 lines in ''/etc/passwd'' at this writing. Assume 25 are system accounts, 25 collaboration accounts (hpc01-hpc25, can VPN, for non Wesleyan faculty/PI, AD accounts), 100 temporary/recyclable class accounts (hpc100-hpc200, can not VPN, AD accounts) which then leaves a lifetime user base of roughly 400 user accounts.  Of those, which come and go, 2 to 2 dozen users may be logged in at any time.+There are 557 lines in ''/etc/passwd'' at this writing. Assume 25 are system accounts, 25 collaboration accounts (hpc01-hpc25, can VPN, for non Wesleyan faculty/PI, AD accounts), 100 temporary/recyclable class accounts (hpc100-hpc200, can not VPN, local accounts) which then leaves a lifetime user base of roughly 400 user accounts.  Of those, which come and go, 2 to 2 dozen users may be logged in at any time.
  
-Rstore is a platform for storing research static data. The hope is to move static data off the HPCC and mount it read-only back onto the HPCC login nodes.  440 tb, fully replicated, is provided for this purpose (Supermicro storage boxes using Rsync as replication engine) for HPCC users and other Wesleyan groups.+Rstore is a platform for storing research static data. The hope is to move static data off the HPCC and mount it read-only back onto the HPCC login nodes.  440 tb, fully replicated, is provided for this purpose (Supermicro storage boxes using Rsync as replication engine). For HPCC users and other Wesleyan groups.
  
-The Data Center has recently been renovated so the HPCC has no more cooling power problems (It used to be in the event of a cooling tower failure, within 3 hours the HPCC would push temps above 85F). No more. We have sufficient rack space (5) and power for expansion. For details on that "live renovation" process visit [[cluster:178|Data Center Renovation]]. It turned out the HPCC was consuming 1/3rd of all electric and cooling capacities. Go HPCC.+The Data Center has recently been renovated so the HPCC has no more cooling problems (It used to be in the event of a cooling tower failure, within 3 hours the HPCC would push temps above 85F). No more. We have sufficient rack space (5) and power for expansion. For details on that "live renovation" process visit [[cluster:178|Data Center Renovation]]. It turned out the HPCC was consuming 1/3rd of all electric and cooling capacities. Go HPCC.
  
  
cluster/189.txt ยท Last modified: 2024/02/12 16:47 by hmeij07