Problem of running VASP parrelly on different nodes
Posted: Wed Nov 17, 2004 4:02 pm
Hello everybody,
I use pgf90 and lam 7.0.6 to compile vasp.4.6 parallel version.
The program runs pretty well on each of my AMD opteron nodes which have 2 CPU on it.
Then I tryied to run the pvasp on different nodes. First, I magnage to make logging on from one node to another node not require entering password. Then I ran some simple program , like printing "Hello world" from different CPU. It woked fine. However, when I began running pvasp it shows error message.
Input file:
Only change NPAR=4 in INCAR, everything else is the same.
command type:
$ lamboot hostfile
$ mpirun -np 4 pvasp
error message:
FORTRAN STOP
Error reading item 'IMAGES' from file INCAR.
FORTRAN STOP
Error reading item 'IMAGES' from file INCAR.
MPI_Recv: process in local group is dead (rank 0, comm 3)
Rank (0, MPI_COMM_WORLD): Call stack within LAM:
Rank (0, MPI_COMM_WORLD): - MPI_Recv()
Rank (0, MPI_COMM_WORLD): - MPI_Barrier()
Rank (0, MPI_COMM_WORLD): - MPI_Barrier()
Rank (0, MPI_COMM_WORLD): - main()
MPI_Recv: process in local group is dead (rank 1, SSI:coll:smp:local comm for CID 0)
Rank (1, MPI_COMM_WORLD): Call stack within LAM:
Rank (1, MPI_COMM_WORLD): - MPI_Recv()
Rank (1, MPI_COMM_WORLD): - MPI_Bcast()
Rank (1, MPI_COMM_WORLD): - MPI_Barrier()
Rank (1, MPI_COMM_WORLD): - main()
Aside from being confused abou this MPI problem, I am also confused that it looks like the pvasp on the distant node trying to fine the INCAR but failed. But shouldn't the input file INCAR be transfered from local host to the distant node during the process?
I am very grateful if somebody could give a hint of how could I solve the problem Thanks a lot !
I use pgf90 and lam 7.0.6 to compile vasp.4.6 parallel version.
The program runs pretty well on each of my AMD opteron nodes which have 2 CPU on it.
Then I tryied to run the pvasp on different nodes. First, I magnage to make logging on from one node to another node not require entering password. Then I ran some simple program , like printing "Hello world" from different CPU. It woked fine. However, when I began running pvasp it shows error message.
Input file:
Only change NPAR=4 in INCAR, everything else is the same.
command type:
$ lamboot hostfile
$ mpirun -np 4 pvasp
error message:
FORTRAN STOP
Error reading item 'IMAGES' from file INCAR.
FORTRAN STOP
Error reading item 'IMAGES' from file INCAR.
MPI_Recv: process in local group is dead (rank 0, comm 3)
Rank (0, MPI_COMM_WORLD): Call stack within LAM:
Rank (0, MPI_COMM_WORLD): - MPI_Recv()
Rank (0, MPI_COMM_WORLD): - MPI_Barrier()
Rank (0, MPI_COMM_WORLD): - MPI_Barrier()
Rank (0, MPI_COMM_WORLD): - main()
MPI_Recv: process in local group is dead (rank 1, SSI:coll:smp:local comm for CID 0)
Rank (1, MPI_COMM_WORLD): Call stack within LAM:
Rank (1, MPI_COMM_WORLD): - MPI_Recv()
Rank (1, MPI_COMM_WORLD): - MPI_Bcast()
Rank (1, MPI_COMM_WORLD): - MPI_Barrier()
Rank (1, MPI_COMM_WORLD): - main()
Aside from being confused abou this MPI problem, I am also confused that it looks like the pvasp on the distant node trying to fine the INCAR but failed. But shouldn't the input file INCAR be transfered from local host to the distant node during the process?
I am very grateful if somebody could give a hint of how could I solve the problem Thanks a lot !