Warning: Undefined array key "DOKU_PREFS" in /usr/share/dokuwiki/inc/common.php on line 2082
cluster:207 [DokuWiki]

User Tools

Site Tools


cluster:207

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
cluster:207 [2021/10/12 12:42]
hmeij07
cluster:207 [2021/10/14 15:43]
hmeij07
Line 7: Line 7:
 Getting a head start on our new login node plus two cpu+gpu compute node project. Hardware has been purchased but there is long delivery time. Meanwhile it makes sense to setup a standalone Slurm scheduler and do some testing and have as a backup. Slurm will be running on ''greentail52'' with a some compute nodes. Getting a head start on our new login node plus two cpu+gpu compute node project. Hardware has been purchased but there is long delivery time. Meanwhile it makes sense to setup a standalone Slurm scheduler and do some testing and have as a backup. Slurm will be running on ''greentail52'' with a some compute nodes.
  
 +This page just intended to keep documentation sources handy. Go to the **Users** page [[cluster:208|Slurm Test Env]]
  
 **SLURM documentation** **SLURM documentation**
Line 178: Line 179:
 <code> <code>
  
-#source /share/apps/CENTOS7/amber/miniconda3/etc/profile.d/conda.sh +# cuda 9.2 ...  
-#export PATH=/share/apps/CENTOS7/amber/miniconda3/bin:$PATH +installer found /usr/local/cuda on ''greentail''
-#export LD_LIBRARY_PATH=/share/apps/CENTOS7/amber/miniconda3/lib:$LD_LIBRARY_PATH +
-#which mpirun python conda +
- +
-# cuda 9.2 ... configure finds /usr/local/cuda which points to n37-cuda-9.2 +
-#export CUDAHOME=/usr/local/n37-cuda-9.2 +
-#export PATH=/usr/local/n37-cuda-9.2/bin:$PATH +
-#export LD_LIBRARY_PATH=/usr/local/n37-cuda-9.2/lib64:$LD_LIBRARY_PATH +
-#which nvcc+
  
 # just in case # just in case
Line 199: Line 192:
 --sysconfdir=/usr/local/slurm-21.08.1/etc \ --sysconfdir=/usr/local/slurm-21.08.1/etc \
  | tee -a install.log  | tee -a install.log
-not --with-nvml=/usr/local/n37-cuda-9.2 \ +skip # --with-nvml=/usr/local/n37-cuda-9.2 \ 
-not -with-hdf5=no  \ +skip # -with-hdf5=no  \
- +
 # known hdf5 library problem when including --with-nvml # known hdf5 library problem when including --with-nvml
  
Line 208: Line 199:
 config.status: creating src/plugins/gpu/nvml/Makefile config.status: creating src/plugins/gpu/nvml/Makefile
  
- +====
 Libraries have been installed in: Libraries have been installed in:
    /usr/local/slurm-21.08.1/lib/slurm    /usr/local/slurm-21.08.1/lib/slurm
Line 223: Line 213:
    - use the '-Wl,-rpath -Wl,LIBDIR' linker flag    - use the '-Wl,-rpath -Wl,LIBDIR' linker flag
    - have your system administrator add LIBDIR to '/etc/ld.so.conf'    - have your system administrator add LIBDIR to '/etc/ld.so.conf'
 +====
 +
 +# for now
 +export PATH=/usr/local/slurm/bin:$PATH
 +export LD_LIBRARY_PATH=/usr/local/slurm/lib:$LD_LIBRARY_PATH
  
 </code> </code>
 +
 +
 +For **general accounting** we may rely on simple text file
  
 <code> <code>
  
-export PATH=/usr/local/slurm/bin:$PATH +From job completions file, JOB #3, convert Start and End times to epoch seconds
-export LD_LIBRARY_PATH=/usr/local/slurm/lib:$LD_LIBRARY_PATH +
- +
-From job completions file, JOB #3+
  
 StartTime=2021-10-06T14:32:37 EndTime=2021-10-06T14:37:40 StartTime=2021-10-06T14:32:37 EndTime=2021-10-06T14:37:40
Line 247: Line 242:
  
 **Full Version Slurm Config Tool** **Full Version Slurm Config Tool**
 +
 +  * lets start with this file and build up/out
  
 <code> <code>
Line 396: Line 393:
 # #
 # COMPUTE NODES # COMPUTE NODES
-NodeName=n[110-111] CPUs=2 RealMemory=192 CoresPerSocket=12 ThreadsPerCore=12 State=UNKNOWN+NodeName=n[110-111] CPUs=2 RealMemory=192 CoresPerSocket=12 ThreadsPerCore=State=UNKNOWN
 # #
 # #
cluster/207.txt · Last modified: 2023/10/27 14:47 by hmeij07