[Wien] running jobs without $SCRATCH

jadhikari@clarku.edu jadhikari at clarku.edu
Fri Apr 6 17:36:55 CEST 2007


Prof. P Blaha,

Thank you very much for the answer.

We have distributed memory system in our cluster with a master node and
about 48 child nodes with 2 processers each. We use $SCRATCH of the nodes
so that no jobs run on the master node (master node alone is not allowed).

I have no idea about running jobs on the nodes without using $SCRATCH
space. I will be very grateful for any suggestion regarding this.

Waiting for the reply.
Subin



> Put    granularity:1
> This will evenly distribute the k-points at once (and not one after the
> other, which is usefull for load ballencing, but you cannot use $SCRATCH)
>
>
>
> jadhikari at clarku.edu schrieb:
>> Dear Wien users,
>>
>> I have a question concerning k point distribution.
>> 24 IBZ points are evenly distributed to 3 nodes and 6 processers with
>> each
>> processor getting 4 IBZ points as shown below. (1 node has 2 processers)
>>
>> But this is not the situation always. Sometimes the no of k points that
>> one processor get is more than that of others. And the system always
>> crashes if this happens.
>>
>> Is there a way to control this inhomogeneity? All the processers are of
>> equal speed. The .machines file is shown at the end.
>>
>> Thank you.
>>
>> Subin
>>
>> _________________________________________________________________________
>>      node2(1) 9.176u 0.092s 0:09.28 99.7%       0+0k 0+0io 0pf+0w
>>      node5(1) 9.715u 0.118s 0:10.50 93.5%       0+0k 0+0io 0pf+0w
>>      node9(1) 9.754u 0.130s 0:11.75 84.0%       0+0k 0+0io 0pf+0w
>>      node2(1) 10.918u 0.112s 0:17.80 61.9%      0+0k 0+0io 0pf+0w
>>      node5(1) 9.453u 0.114s 0:11.28 84.7%       0+0k 0+0io 0pf+0w
>>      node9(1) 9.995u 0.117s 0:13.79 73.2%       0+0k 0+0io 0pf+0w
>>      node2(1) 9.286u 0.095s 0:09.40 99.6%       0+0k 0+0io 0pf+0w
>>      node5(1) 11.702u 0.115s 0:12.99 90.9%      0+0k 0+0io 0pf+0w
>>      node9(1) 9.336u 0.110s 0:16.29 57.9%       0+0k 0+0io 0pf+0w
>>      node2(1) 9.403u 0.111s 0:15.62 60.8%       0+0k 0+0io 0pf+0w
>>      node5(1) 11.607u 0.116s 0:15.94 73.4%      0+0k 0+0io 0pf+0w
>>      node9(1) 9.595u 0.119s 0:13.52 71.7%       0+0k 0+0io 0pf+0w
>>      node2(1) 9.207u 0.112s 0:10.64 87.5%       0+0k 0+0io 0pf+0w
>>      node5(1) 11.135u 0.124s 0:14.81 75.9%      0+0k 0+0io 0pf+0w
>>      node9(1) 9.985u 0.114s 0:16.91 59.6%       0+0k 0+0io 0pf+0w
>>      node2(1) 10.602u 0.118s 0:18.33 58.4%      0+0k 0+0io 0pf+0w
>>      node5(1) 11.476u 0.106s 0:16.98 68.1%      0+0k 0+0io 0pf+0w
>>      node9(1) 9.325u 0.100s 0:13.75 68.5%       0+0k 0+0io 0pf+0w
>>      node2(1) 9.447u 0.109s 0:10.03 95.1%       0+0k 0+0io 0pf+0w
>>      node5(1) 9.997u 0.115s 0:11.08 91.1%       0+0k 0+0io 0pf+0w
>>      node9(1) 10.821u 0.119s 0:19.06 57.3%      0+0k 0+0io 0pf+0w
>>      node2(1) 9.400u 0.097s 0:13.84 68.5%       0+0k 0+0io 0pf+0w
>>      node5(1) 11.749u 0.130s 0:17.38 68.2%      0+0k 0+0io 0pf+0w
>>      node9(1) 9.436u 0.112s 0:12.45 76.6%       0+0k 0+0io 0pf+0w
>>    Summary of lapw1para:
>>    node2         k=8     user=77.439     wallclock=104.94
>>    node5         k=8     user=86.834     wallclock=110.96
>>    node9         k=8     user=78.247     wallclock=117.52
>>    node2         k=8     user=77.439     wallclock=104.94
>>    node5         k=8     user=86.834     wallclock=110.96
>>    node9         k=8     user=78.247     wallclock=117.52
>> _________________________________________________________
>> .machine file
>>
>> 1:node2
>> 1:node5
>> 1:node9
>> 1:node2
>> 1:node5
>> 1:node9
>> granularity:4
>> extrafine:1
>>
>> _______________________________________________
>> Wien mailing list
>> Wien at zeus.theochem.tuwien.ac.at
>> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
> _______________________________________________
> Wien mailing list
> Wien at zeus.theochem.tuwien.ac.at
> http://zeus.theochem.tuwien.ac.at/mailman/listinfo/wien
>
>



More information about the Wien mailing list