User Tools

Site Tools


cluster:116

Warning: Undefined array key -1 in /usr/share/dokuwiki/inc/html.php on line 1458

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
cluster:116 [2013/07/11 15:35]
hmeij
cluster:116 [2014/02/04 13:57] (current)
hmeij
Line 1: Line 1:
 \\ \\
 **[[cluster:0|Back]]** **[[cluster:0|Back]]**
 +
 +Since deployment of sharptail the information below is out of date. /home is now the same across the entire HPCC and served out by sharptail.
 +
 + --- //[[hmeij@wesleyan.edu|Meij, Henk]] 2014/02/04 13:56//
  
 ===== Sharptail Cluster ===== ===== Sharptail Cluster =====
Line 39: Line 43:
 ==== /sanscratch ==== ==== /sanscratch ====
  
-Sharptail will provide the users (and scheduler) with another 5 TB scratch file system.  During this period it is only provided to the sharptail nodes (n33-n37). In the future it will provide this file system to all nodes except greentail nodes (n1-n32).  +Sharptail will provide the users (and scheduler) with another 5 TB scratch file system.  It is only provided to the sharptail nodes (n33-n37). All other nodes will have /sanscratch provided by greentail.  You can follow the progress of your by looking into /sanscratch/JOBPID directory on either greentail or sharptail.
  
   * Please offload as much IO from /home by staging your jobs in /sanscratch   * Please offload as much IO from /home by staging your jobs in /sanscratch
Line 78: Line 82:
  
 In both cases you do not need to target any specific core, the operating system will handle that part of the scheduling. In both cases you do not need to target any specific core, the operating system will handle that part of the scheduling.
 +
 +==== NOTE ====
 +
 +
 +----
 +
 +Instructions below are obsolete, resources are now available via the scheduler.
 +
 +Please read [[cluster:119|Submitting GPU Jobs]]
 +
 + --- //[[hmeij@wesleyan.edu|Meij, Henk]] 2013/08/21 10:46//
 +
 +----
  
  
 ==== CPU-HPC ==== ==== CPU-HPC ====
  
-With hyperthreading on the 5 nodes, it provides for 160 cores.  We need to reserve 20 cores for the GPUs (one per GPU), and lets reserve another 20 cores for the OS (per node).  That still leaves 120 cores for regular jobs like you are used to on greentail.  These 120 cores (24 per node) will show up later as a new queue on greentail/swallowtail; one that is fit for jobs that need much memory. On average 256 gb per node minus 20 gb for 4 GPUs minus 20 gb for OS leaves 5.6 gb ''per core''+With hyperthreading on the 5 nodes, it provides for 160 cores.  We need to reserve 20 cores for the GPUs (one per GPU, 4 per node), and lets reserve another 20 cores for the OS (per node).  That still leaves 120 cores for regular jobs like you are used to on greentail.  These 120 cores (24 per node) will show up later as a new queue on greentail/swallowtail; one that is fit for jobs that need much memory. On average 256 gb per node minus 20 gb for 4 GPUs minus 20 gb for OS leaves 5.6 gb ''per core''
  
 So since there is no scheduler, you need to setup your environment and execute your program.  Here is an example of a program that normally runs on the imw queue.  If your program involves MPI you need to be a bit up to speed on what the lava wrapper actually does for you. So since there is no scheduler, you need to setup your environment and execute your program.  Here is an example of a program that normally runs on the imw queue.  If your program involves MPI you need to be a bit up to speed on what the lava wrapper actually does for you.
Line 240: Line 257:
  
 [[cluster:111|Amber GPU Testing]] ... may help shed some ideas [[cluster:111|Amber GPU Testing]] ... may help shed some ideas
 +
 +Note: ran out of time to get an example running but it should follow the LAMMPS approach of above pretty closely.  The binary is in /cm/share/apps/amber/amber12/bin/pmemd.cuda.MPI
 +
 +Here is quick Amber example
 +
 +<code>
 +
 +[hmeij@sharptail nucleosome]$ export AMBER_HOME=/cm/shared/apps/amber/amber12
 +
 +# find a GPU ID with gpu-info then expose that GPU to pmemd
 +[hmeij@sharptail nucleosome]$ export CUDA_VISIBLE_DEVICES=1
 +
 +# you only need one cpu core
 +[hmeij@sharptail nucleosome]$ mpirun_rsh -ssh -hostfile ~/sharptail/hostfile -np 1 \
 +/cm/shared/apps/amber/amber12/bin/pmemd.cuda.MPI -O -o mdout.1K10 -inf mdinfo.1K10 -x mdcrd.1K10 -r restrt.1K10 -ref inpcrd &
 +
 +</code>
 +
 +
  
 NAMD was compiled with the built-in multi-node networking capabilities, including ibverbs support. NAMD was compiled with the built-in multi-node networking capabilities, including ibverbs support.
cluster/116.1373571305.txt.gz ยท Last modified: 2013/07/11 15:35 by hmeij