Warning: Undefined array key "DOKU_PREFS" in /usr/share/dokuwiki/inc/common.php on line 2082
cluster:207 [DokuWiki]

User Tools

Site Tools


cluster:207

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
cluster:207 [2021/10/12 12:41]
hmeij07
cluster:207 [2021/10/13 13:06]
hmeij07
Line 7: Line 7:
 Getting a head start on our new login node plus two cpu+gpu compute node project. Hardware has been purchased but there is long delivery time. Meanwhile it makes sense to setup a standalone Slurm scheduler and do some testing and have as a backup. Slurm will be running on ''greentail52'' with a some compute nodes. Getting a head start on our new login node plus two cpu+gpu compute node project. Hardware has been purchased but there is long delivery time. Meanwhile it makes sense to setup a standalone Slurm scheduler and do some testing and have as a backup. Slurm will be running on ''greentail52'' with a some compute nodes.
  
 +This page just intended to keep documentation sources handy.
  
 **SLURM documentation** **SLURM documentation**
Line 178: Line 179:
 <code> <code>
  
-#source /share/apps/CENTOS7/amber/miniconda3/etc/profile.d/conda.sh +# cuda 9.2 ...  
-#export PATH=/share/apps/CENTOS7/amber/miniconda3/bin:$PATH +installer found /usr/local/cuda on ''greentail''
-#export LD_LIBRARY_PATH=/share/apps/CENTOS7/amber/miniconda3/lib:$LD_LIBRARY_PATH +
-#which mpirun python conda +
- +
-# cuda 9.2 ... configure finds /usr/local/cuda which points to n37-cuda-9.2 +
-#export CUDAHOME=/usr/local/n37-cuda-9.2 +
-#export PATH=/usr/local/n37-cuda-9.2/bin:$PATH +
-#export LD_LIBRARY_PATH=/usr/local/n37-cuda-9.2/lib64:$LD_LIBRARY_PATH +
-#which nvcc+
  
 # just in case # just in case
Line 199: Line 192:
 --sysconfdir=/usr/local/slurm-21.08.1/etc \ --sysconfdir=/usr/local/slurm-21.08.1/etc \
  | tee -a install.log  | tee -a install.log
-not --with-nvml=/usr/local/n37-cuda-9.2 \ +skip # --with-nvml=/usr/local/n37-cuda-9.2 \ 
-not -with-hdf5=no  \ +skip # -with-hdf5=no  \
- +
 # known hdf5 library problem when including --with-nvml # known hdf5 library problem when including --with-nvml
  
Line 208: Line 199:
 config.status: creating src/plugins/gpu/nvml/Makefile config.status: creating src/plugins/gpu/nvml/Makefile
  
-<code> +====
 Libraries have been installed in: Libraries have been installed in:
    /usr/local/slurm-21.08.1/lib/slurm    /usr/local/slurm-21.08.1/lib/slurm
Line 223: Line 213:
    - use the '-Wl,-rpath -Wl,LIBDIR' linker flag    - use the '-Wl,-rpath -Wl,LIBDIR' linker flag
    - have your system administrator add LIBDIR to '/etc/ld.so.conf'    - have your system administrator add LIBDIR to '/etc/ld.so.conf'
 +====
  
-</code> +# for now
 export PATH=/usr/local/slurm/bin:$PATH export PATH=/usr/local/slurm/bin:$PATH
 export LD_LIBRARY_PATH=/usr/local/slurm/lib:$LD_LIBRARY_PATH export LD_LIBRARY_PATH=/usr/local/slurm/lib:$LD_LIBRARY_PATH
  
-From job completions file, JOB #3+</code> 
 + 
 + 
 +For **general accounting** we may rely on simple text file 
 + 
 +<code> 
 + 
 +From job completions file, JOB #3, convert Start and End times to epoch seconds
  
 StartTime=2021-10-06T14:32:37 EndTime=2021-10-06T14:37:40 StartTime=2021-10-06T14:32:37 EndTime=2021-10-06T14:37:40
Line 245: Line 242:
  
 **Full Version Slurm Config Tool** **Full Version Slurm Config Tool**
 +
 +  * lets start with this file and build up/out
  
 <code> <code>
Line 394: Line 393:
 # #
 # COMPUTE NODES # COMPUTE NODES
-NodeName=n[110-111] CPUs=2 RealMemory=192 CoresPerSocket=12 ThreadsPerCore=12 State=UNKNOWN+NodeName=n[110-111] CPUs=2 RealMemory=192 CoresPerSocket=12 ThreadsPerCore=State=UNKNOWN
 # #
 # #
cluster/207.txt · Last modified: 2023/10/27 14:47 by hmeij07