User Tools

Site Tools


cluster:103

Warning: Undefined array key -1 in /usr/share/dokuwiki/inc/html.php on line 1458

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
cluster:103 [2011/12/21 14:39]
hmeij [Submit]
cluster:103 [2011/12/22 14:34] (current)
hmeij [Submit 2]
Line 132: Line 132:
 ==== Submit 2 ==== ==== Submit 2 ====
  
-On the back end compute nodes, unless specified, the job runs inside your home directory.  That job competes with other activities inside /home.  Compute nodes have two other areas where the jobs could be submitted: /localscratch and /sanscratch.  The former is a local filesystem for each node and should be used if file locking essential. The later is a filesystem from greentails disarray of 5 TB served vi IPoIB (that is NFS traffic over fast interconnects switches, the performance should be much better than gigabit ethernet switches).  It is comprised of disks and spindles that are not impacted by happens on /home.  So we're going to use that.+On the back end compute nodes, unless specified, the job runs inside your home directory.  That job competes with other activities inside /home.  Compute nodes have two other areas where the jobs could be submitted: /localscratch and /sanscratch.  The former is a local filesystem for each node and should be used if file locking is essential. The later is a filesystem from greentails diskarray (5 TBserved vi IPoIB (that is NFS traffic over fast interconnects switches, the performance should be much better than gigabit ethernet switches).  It is comprised of disks and spindles that are not impacted by what happens on /home.  So we're going to use that. 
 + 
 +In the SAS program we add the following lines 
 + 
 +<code> 
 +%let jobpid = %sysget(LSB_JOBID); 
 +libname here "/sanscratch/&jobpid"; 
 +</code> 
 + 
 +And change this line to use local disks for storage 
 + 
 +<code> 
 +data here.one; 
 +</code> 
 + 
 +In the submission script we change the following
  
   * new submission file with edits   * new submission file with edits
-  * -n implies reserve job slots (cpu cores) for job (not necesssary, SAS jobs will always be one)+  * -n implies reserve job slots (cpu cores) for job (not necesssary, SAS jobs will always use only one)
   * -R reserves memory, for example, reserve 200 MB of memory on target compute node   * -R reserves memory, for example, reserve 200 MB of memory on target compute node
-  * scheduler creates unique dirs in scratch by JOBPID for you, so we'll satge the job there+  * scheduler creates unique dirs in scratch by JOBPID for you, so we'll stage the job there 
 +  * but now we must copy relevant files //to// scratch dir and results back //to// home dir 
  
 <code> <code>
Line 163: Line 180:
  
   * you can monitor the progress of your jobs from greentail while it runs   * you can monitor the progress of your jobs from greentail while it runs
 +
 <code> <code>
 [hmeij@greentail sas]$ ll /sanscratch/492667/ [hmeij@greentail sas]$ ll /sanscratch/492667/
Line 172: Line 190:
 </code> </code>
  
 +==== Best Practices ====
  
 +  * You may submit as many SAS jobs as you like, just leave enough resources available for others to also get work done 
 +  * Because SAS submission are serial, non-parallel jobs your -n flag is always 1 
 +  * Reserve resources if you know what you need, especially memory 
 +  * Use /sanscratch for large data jobs with heavy read/write operations 
 +  * Queue ehwfd is preferentially for Gaussian users and stay off the stata and matlab queues please 
 +  * Write smart SAS code, for example, use data set indexes and PROC SQL (this can be your best friend) 
 +  * ... suggestions will be added to this page
  
 \\ \\
 **[[cluster:0|Back]]** **[[cluster:0|Back]]**
cluster/103.1324496356.txt.gz · Last modified: 2011/12/21 14:39 by hmeij