[Wien] running jobs without $SCRATCH
jadhikari@clarku.edu
jadhikari at clarku.edu
Fri Apr 6 17:36:55 CEST 2007
Prof. P Blaha,
Thank you very much for the answer.
We have distributed memory system in our cluster with a master node and
about 48 child nodes with 2 processers each. We use $SCRATCH of the nodes
so that no jobs run on the master node (master node alone is not allowed).
I have no idea about running jobs on the nodes without using $SCRATCH
space. I will be very grateful for any suggestion regarding this.
Waiting for the reply.
Subin
> Put granularity:1
> This will evenly distribute the k-points at once (and not one after the
> other, which is usefull for load ballencing, but you cannot use $SCRATCH)
>
>
>
> jadhikari at clarku.edu schrieb:
>> Dear Wien users,
>>
>> I have a question concerning k point distribution.
>> 24 IBZ points are evenly distributed to 3 nodes and 6 processers with
>> each
>> processor getting 4 IBZ points as shown below. (1 node has 2 processers)
>>
>> But this is not the situation always. Sometimes the no of k points that
>> one processor get is more than that of others. And the system always
>> crashes if this happens.
>>
>> Is there a way to control this inhomogeneity? All the processers are of
>> equal speed. The .machines file is shown at the end.
>>
>> Thank you.
>>
>> Subin
>>
>> _________________________________________________________________________
>> node2(1) 9.176u 0.092s 0:09.28 99.7% 0+0k 0+0io 0pf+0w
>> node5(1) 9.715u 0.118s 0:10.50 93.5% 0+0k 0+0io 0pf+0w
>> node9(1) 9.754u 0.130s 0:11.75 84.0% 0+0k 0+0io 0pf+0w
>> node2(1) 10.918u 0.112s 0:17.80 61.9% 0+0k 0+0io 0pf+0w
>> node5(1) 9.453u 0.114s 0:11.28 84.7% 0+0k 0+0io 0pf+0w
>> node9(1) 9.995u 0.117s 0:13.79 73.2% 0+0k 0+0io 0pf+0w
>> node2(1) 9.286u 0.095s 0:09.40 99.6% 0+0k 0+0io 0pf+0w
>> node5(1) 11.702u 0.115s 0:12.99 90.9% 0+0k 0+0io 0pf+0w
>> node9(1) 9.336u 0.110s 0:16.29 57.9% 0+0k 0+0io 0pf+0w
>> node2(1) 9.403u 0.111s 0:15.62 60.8% 0+0k 0+0io 0pf+0w
>> node5(1) 11.607u 0.116s 0:15.94 73.4% 0+0k 0+0io 0pf+0w
>> node9(1) 9.595u 0.119s 0:13.52 71.7% 0+0k 0+0io 0pf+0w
>> node2(1) 9.207u 0.112s 0:10.64 87.5% 0+0k 0+0io 0pf+0w
>> node5(1) 11.135u 0.124s 0:14.81 75.9% 0+0k 0+0io 0pf+0w
>> node9(1) 9.985u 0.114s 0:16.91 59.6% 0+0k 0+0io 0pf+0w
>> node2(1) 10.602u 0.118s 0:18.33 58.4% 0+0k 0+0io 0pf+0w
>> node5(1) 11.476u 0.106s 0:16.98 68.1% 0+0k 0+0io 0pf+0w
>> node9(1) 9.325u 0.100s 0:13.75 68.5% 0+0k 0+0io 0pf+0w
>> node2(1) 9.447u 0.109s 0:10.03 95.1% 0+0k 0+0io 0pf+0w
>> node5(1) 9.997u 0.115s 0:11.08 91.1% 0+0k 0+0io 0pf+0w
>> node9(1) 10.821u 0.119s 0:19.06 57.3% 0+0k 0+0io 0pf+0w
>> node2(1) 9.400u 0.097s 0:13.84 68.5% 0+0k 0+0io 0pf+0w
>> node5(1) 11.749u 0.130s 0:17.38 68.2% 0+0k 0+0io 0pf+0w
>> node9(1) 9.436u 0.112s 0:12.45 76.6% 0+0k 0+0io 0pf+0w
>> Summary of lapw1para:
>> node2 k=8 user=77.439 wallclock=104.94
>> node5 k=8 user=86.834 wallclock=110.96
>> node9 k=8 user=78.247 wallclock=117.52
>> node2 k=8 user=77.439 wallclock=104.94
>> node5 k=8 user=86.834 wallclock=110.96
>> node9 k=8 user=78.247 wallclock=117.52
>> _________________________________________________________
>> .machine file
>>
>> 1:node2
>> 1:node5
>> 1:node9
>> 1:node2
>> 1:node5
>> 1:node9
>> granularity:4
>> extrafine:1
>>
>> _______________________________________________
>> Wien mailing list
>> Wien at zeus.theochem.tuwien.ac.at
>> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
> _______________________________________________
> Wien mailing list
> Wien at zeus.theochem.tuwien.ac.at
> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
>
>
More information about the Wien
mailing list