[Wien] slurm script

pluto pluto at physics.ucdavis.edu
Tue Jun 18 10:57:08 CEST 2024


Dear Prof. Blaha,

Thank you for the quick answer. I will be trying to adapt the script 
with the help of our IT experts.

Just to make sure, I have two questions:

- This script will allow running WIEN2k on slurm cluster without 
passwordless ssh login?

- Should this script generate .machines file which is present in the 
CASE directory? Should this .machines file contain a list_of_all_nodes 
times the tasks_per_node? For N=4 and tasks_per_node=8 this should be 32 
lines, 8 lines per each node (like the localhost lines when running on 
desktop machine).

Best,
Lukasz


On 2024-06-17 20:57, Peter Blaha wrote:
> Below some comments:
> 
>> I have gotten the following error:
>> 
>> [lplucin at iffslurm Au-bulk-test]$ more slurm-69063.out
>> iffcluster0414: Using Ethernet for MPI communication.
>> SBATCH: Command not found.
> 
> It looks as if you have a line with
> SBATCH ...      in your script instead of
> #SBATCH ....
> 
> Probably the definition of number of nodes:
> #SBATCH -N 4        # define number of nodes
> 
> PS: For slurm
> #command      are slurm directives,
> ##command     is a comment
> command       would be a Linux command, but not a SLURM directive.
> 
>> DIRECTORY = /local/th1/topo/lplucin/Au-bulk-test
>> WIENROOT = /Users/lplucin/WIEN2k
>> SCRATCH = /SCRATCH
>> SLURM_NTASKS: Undefined variable.
> 
>> 
>> Then I tried to define:
>> 
>> setenv SLURM_NTASKS_PER_NODE 8
>> setenv SLURM_NTASKS 32
> 
> These variables should be set automatically by the
> #SBATCH --tasks-per-node=8
> #SBATCH -N 4        # define number of nodes
> 
> directives.
> 
> 
> PS: Your cluster support should have some information/documents how
> one submits a slurm job and which variables should be set/used.
> 
> I could imagine that different slurm versions/setups work slightly 
> different.
> 
>> 
>> but this did not help much, something started on one node, but crashed 
>> immediately.
>> 
>> Any advice would be appreciated!
>> 
>> Best,
>> Lukasz
>> 
>> 
>> 
>> 
>> #!/bin/tcsh
>> #SBATCH -J compound
>> #SBATCH --tasks-per-node=8
>> ##SBATCH --mail-user=xxx at yyy.zzz
>> ##SBATCH --mail-type=END
>> 
>> ## The next 3 lines may allow some the definition of specific queues 
>> depending on your cluster
>> #SBATCH --partition topo
>> ##SBATCH --qos=XXX
>> ##SBATCH --account=XXX
>> #
>> #########################
>> #SBATCH -N 4 # define number of nodes
>> set mpijob=1 # define number of cores for one lapw1/2 mpi-job
>> ##
>> ## The lines above need to be adapted depending our your case (size of 
>> the
>> ## problem and number of k-points).
>> ## The example requests 64 cores on 4 nodes (because core_per_nodes is 
>> 16),
>> ## and configures WIEN2k such that 4 k-point-parallel jobs will run on 
>> 16-cores
>> ## (one node) in mpi-mode, i.e. it assumes that you have 4, 8, 12 ... 
>> k-points
>> ## and the case is of a size that it runs reasonably on 16 cores.
>> ## For a bigger case and only 1 k-point, you would set mpijob=64 
>> instead....
>> #########################
>> 
>> set cores_per_node=8
>> setenv OMP_NUM_THREADS 1
>> set lapw2_vector_split=4
>> setenv PSM_SHAREDCONTEXTS_MAX 4
>> setenv SCRATCH /SCRATCH                           # set it to 
>> recommended value
>> 
>> _______________________________________________
>> Wien mailing list
>> Wien at zeus.theochem.tuwien.ac.at
>> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
>> SEARCH the MAILING-LIST at:  
>> http://www.mail-archive.com/wien@zeus.theochem.tuwien.ac.at/index.html


More information about the Wien mailing list