Ilyas,
this has been fixed in newer versions by the following commits:
revision 9.3
date: 2006/08/17 23:10:58;  author: case;  state: Exp;  lines: +1 -0
initialize gmin to 1 to avoid divide by zero
----------------------------
revision 9.8
date: 2007/10/24 22:46:00;  author: steinbrt;  state: Exp;  lines: +19 -33
TBS: Extended and cleaned up Softcore potentials for TI
Please find the .diff attached (I did not test it throughly though).
Have a great night,
  Volodymyr
On Tue, November 11, 2008 14:38, Ilyas Yildirim wrote:
> The test script is as follows:
>
> --------- runsim -------------
> #!/bin/csh -f
>
> set sander = $AMBERHOME/exe/sander.MPI
>
> /bin/rm -f min
>
> cat > min <<EOF
> Initial Minimization of solvent + ions
>  &cntrl
>   imin   = 1,
>   maxcyc = 10,
>   ncyc   = 5,
>   ntb    = 1,
>   ntr    = 0,
>   cut    = 8.0,
>   ntpr = 1, ntx = 1,
>   icfe=1,klambda=6,clambda=0.5
>  /
> EOF
>
> cat > groups_md <<EOF
> -O -i min -p prmtop1 -c inpcrd -o min.out.p1 -r min.rst.p1 -x
> mdcrd_02.traj.p1
> -O -i min -p prmtop2 -c inpcrd -o min.out.p2 -r min.rst.p2 -x
> mdcrd_02.traj.p2
> EOF
>
> mpiexec -np 4 $sander -ng 2 -groupfile groups_md < /dev/null || goto error
> ------------------------------------
>
> The error message is as follows:
>
> ----
> [yildirim.malatya02 ~/l_0.5]# runsim
>
>   Running multisander version of sander amber9
>      Total processors =            4
>      Number of groups =            2
>
>      Looping over processors:
>         WorldRank is the global PE rank
>         NodeID is the local PE rank in current group
>
>         Group     =            0
>         WorldRank =            0
>         NodeID    =            0
>
>         Group     =            1
>         WorldRank =            2
>         NodeID    =            0
>
>         WorldRank =            1
>         NodeID    =            1
>
>         WorldRank =            3
>         NodeID    =            1
>
> [cli_1]: [cli_3]: aborting job:
> Fatal error in MPI_Bcast: Message truncated, error stack:
> MPI_Bcast(784).........................: MPI_Bcast(buf=0xe63888, count=1,
> MPI_INTEGER, root=0, comm=0x84000000) f
> ailed
> MPIR_Bcast(198)........................:
> MPIDI_CH3U_Post_data_receive_found(163): Message from rank 0 and tag 2
> truncated; 144408 bytes received but buffe
> r size is 4
> aborting job:
> Fatal error in MPI_Bcast: Message truncated, error stack:
> MPI_Bcast(784).........................: MPI_Bcast(buf=0xe63888, count=1,
> MPI_INTEGER, root=0, comm=0x84000000) f
> ailed
> MPIR_Bcast(198)........................:
> MPIDI_CH3U_Post_data_receive_found(163): Message from rank 0 and tag 2
> truncated; 144408 bytes received but buffe
> r size is 4
> rank 3 in job 1  malatya02_48011   caused collective abort of all ranks
>   exit status of rank 3: return code 1
> rank 1 in job 1  malatya02_48011   caused collective abort of all ranks
>   exit status of rank 1: return code 1
> error: label not found.
>
> --------------
>
> mpich2version is
>
> [yildirim.malatya02 ~/l_0.5]# mpich2version
> Version:           1.0.5
> Device:            ch3:sock
> Configure Options: '--prefix=/programs/mpich2'
> 'CC=/opt/intel/cce/10.0.026/bin/icc'
> 'CXX=/opt/intel/cce/10.0.026/bin/icpc'
> CC:  /opt/intel/cce/10.0.026/bin/icc
> CXX: /opt/intel/cce/10.0.026/bin/icpc
> F77: /opt/intel/fce/10.0.026/bin/ifort
> F90: ifort
>
> ---------------
>
> Thanks.
>
> On Tue, 11 Nov 2008, Volodymyr Babin wrote:
>
>> On Tue, November 11, 2008 14:18, Ilyas Yildirim wrote:
>> > While the subject has minimization in it, I have a quick question: Is
>> it
>> > possible to use more than 2 CPUs in Thermodynamic Integration
>> > minimization? I could never use 4 CPUs when I wanted to do
>> minimization
>> > with icfe set to either 1 or 2. Everything works fine with 2 CPUs, but
>> not
>> > with 4 CPUs in the minimization process. I use mpich2 and amber9.
>>
>> Could you provide more details on what happens with > 2 cpus?
>> An easily runnable test-case that shows the problem would also be
>> very helpful.
>>
>> Have a great day,
>>   Volodymyr
>>
>>
>
> --
>   Ilyas Yildirim, Ph.D.
>   ---------------------------------------------------------------
>   = Hutchison Hall B#10          - Department of Chemistry      =
>   =                              - University of Rochester      =
>   = 585-275-6766 (office)        -                              =
>   = http://www.pas.rochester.edu/~yildirim/                     =
>   ---------------------------------------------------------------
>
>
Received on Fri Dec 05 2008 - 14:31:02 PST