This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision Next revision Both sides next revision | ||
cluster:175 [2018/09/22 18:25] hmeij07 [P100 vs GTX & K20] |
cluster:175 [2018/09/25 12:35] hmeij07 |
||
---|---|---|---|
Line 5: | Line 5: | ||
==== P100 vs GTX & K20 ==== | ==== P100 vs GTX & K20 ==== | ||
- | ^ ^ P100 ^ GTX ^ K20 ^ | + | ^ ^ P100 ^ GTX ^ K20 |
- | | cores | 3,584 | 3,584 | 2,496 | | + | | cores | 3,584 | 3,584 | 2,496 |
- | | mem | 12/16 | | + | | mem | 12/16 | |
- | | ghz | 2.6 | 1.6 | 0.7 | | + | | ghz | 2.6 | 1.6 | 0.7 |
- | | | + | | |
- | Comparing these GPUs yields the following | + | Comparing these GPUs yields the following |
Credits: This work was made possible, in part, through HPC time donated by Microway, Inc. We gratefully acknowledge Microway for providing access to their GPU-accelerated compute cluster. | Credits: This work was made possible, in part, through HPC time donated by Microway, Inc. We gratefully acknowledge Microway for providing access to their GPU-accelerated compute cluster. | ||
Line 36: | Line 36: | ||
</ | </ | ||
+ | |||
+ | Look at these gpu temperatures, | ||
==== Lammps ==== | ==== Lammps ==== | ||
Line 67: | Line 69: | ||
</ | </ | ||
+ | |||
+ | === WZ === | ||
+ | |||
+ | a) 1GPU with X CPUs | ||
+ | |||
+ | #cpus | ||
+ | 4 89.6 | ||
+ | 2 61.6 | ||
+ | 1 34.3 | ||
+ | |||
+ | b) 4 GPUs with 4 CPUs | ||
+ | 92 ns/day | ||
+ | |||
+ | ---------- | ||
+ | compare these results to | ||
+ | |||
+ | 1 K20 GPU + 4 CPUs 37 ns/day | ||
+ | 1 K20 GPU + 6 CPUs 47 ns/day | ||
+ | 1 GTX GPU + 4 CPUs 73 ns/day | ||
+ | 1 GTX GPU + 6 CPUs 90 ns/day | ||
+ | |||
+ | ---- | ||
+ | Test dir location | ||
+ | / | ||
+ | |||
+ | |||
+ | -------------- | ||
+ | I am running a material called PMMA (https:// | ||
+ | |||
+ | The reason for the different benchmark than yours is that the PMMA simulations require the calculation of molecular bonds, which is not implemented in GPU. | ||
==== Gromacs ==== | ==== Gromacs ==== | ||
Line 72: | Line 104: | ||
Gromacs has shown vastly improved performance between versions. v5 delivered about 20 ns/day per K20 server and 350 ns/day on GTX server. v2018 delivered 75 ns/day per K20 server and 900 ns/day on GTX server. A roughly 3x improvement. | Gromacs has shown vastly improved performance between versions. v5 delivered about 20 ns/day per K20 server and 350 ns/day on GTX server. v2018 delivered 75 ns/day per K20 server and 900 ns/day on GTX server. A roughly 3x improvement. | ||
- | On the P100 test node, I could not invoke the multidir option of gromacs (have run it on GTX, weird). The utilization of the gpu drops as more and more gpus are deployed. | + | On the P100 test node, I could not invoke the multidir option of gromacs (have run it on GTX, weird). The utilization of the gpu drops as more and more gpus are deployed. |
< | < | ||
Line 81: | Line 113: | ||
localhost, | localhost, | ||
localhost, | localhost, | ||
- | gmx_mpi mdrun -gpu_id 0123 -ntmpi 0 \ | + | gmx_mpi mdrun -gpu_id 0123 -ntmpi |
-s topol.tpr -ntomp 4 -npme 1 -nsteps 20000 -pin on -nb gpu | -s topol.tpr -ntomp 4 -npme 1 -nsteps 20000 -pin on -nb gpu | ||