Warning: Undefined array key "DOKU_PREFS" in /usr/share/dokuwiki/inc/common.php on line 2082
cluster:140 [DokuWiki]

User Tools

Site Tools


cluster:140

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Last revision Both sides next revision
cluster:140 [2015/06/17 15:51]
hmeij
cluster:140 [2015/06/19 09:23]
hmeij
Line 27: Line 27:
   * #BSUB -R "rusage[tmslow=1]"   * #BSUB -R "rusage[tmslow=1]"
  
-You need to request a consumable for each job slot, so if using say #BSUB -n 4 the '1' becomes a '4' And your job will go PENDing when consumables are exhausted. When would you do this? For example if you do not wish to run on the hptmnodes and are ok with waiting, or if the fabulous new hardware is clogged full of jobs and you wish to immediately bypass those.+You need to request a consumable for each job slot, so if using say #BSUB -n 4 the '1' becomes a '4' And your job will go PENDing when consumables are exhausted. When would you do this? For example if you do not wish to run on the hptmnodes and are ok with waiting, or if the fabulous new hardware is clogged full of jobs and you wish to immediately bypass those. Or, for now, the old hardware is running redhat 5 and the new centos 6.
  
-Doing nothing, that is not using or requesting consumabels, is a perfect strategy too.+Doing nothing, that is not using or requesting consumables, is a perfect strategy too.
  
 Today queues ''hp12'' is closed while we wait for it to empty out. Then it disappears. Today queues ''hp12'' is closed while we wait for it to empty out. Then it disappears.
Line 40: Line 40:
   * ssh node_name top -u user_name -b -n 1   * ssh node_name top -u user_name -b -n 1
  
-**Use top, remotely, in batch.** Look at the VIRT for memory usage the OS thinks you need, if that exceeds the node capacity go to other queues. For example+**Use top, remotely, in batch.** Look at the VIRT for memory usage the OS thinks you need, if that exceeds the node capacity go to other queues. For example (not picking on anybody!)
  
 <code> <code>
Line 62: Line 62:
  
 # and here is a problem, at 10.8 GB per java process # and here is a problem, at 10.8 GB per java process
-# the new hardware can only 3 of these jobs and  +# the new hardware can only handlew 3 of these jobs (32gb max) and  
-# the old hardware can only run 1 of these  +# the old hardware can only run 1 of these (12 gb max) 
-# so you need to either reserve memory in advance (see below) +# so you need to either reserve memory in advance (see below) or 
-or move to another queue like mw256 (and also reserve memory)+(better) move to another queue like mw256 (and also reserve memory)
  
 [hmeij@greentail ~]$ ssh n58 top -u cjustice -b -n 1                                                           [hmeij@greentail ~]$ ssh n58 top -u cjustice -b -n 1                                                          
cluster/140.txt · Last modified: 2015/12/04 13:53 by hmeij