Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message *** -------------------------------------------------------------------------- mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault). --------------------------------------------------------------------------
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil
Have you tried running a simple MPI code that you compiled yourself using the same MPI installation?
On Dec 17, 2015, at 16:13, D´Souza, Sunil Wilfred sunilwilfred@cpfs.mpg.de wrote:
Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message ***
mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault).
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil
<CMakeCache.txt>
Dear ALPS users, As per your suggestion, I have tried running a simple MPI program by using the following commands
mpic++ -o monte_carlo_mpi monte_carlo_mpi.cpp
mpirun -np 8 monte_carlo_mpi
I have attached the sample MPI program ( computes PI by the Monte Carlo method) which i have used for checking the MPI. Please suggest me has to how to solve the mpi problem in ALPS simulations.
with best regards
Sunil
________________________________ From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of Matthias Troyer [troyer@phys.ethz.ch] Sent: Thursday, December 17, 2015 4:18 PM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Have you tried running a simple MPI code that you compiled yourself using the same MPI installation?
On Dec 17, 2015, at 16:13, D´Souza, Sunil Wilfred <sunilwilfred@cpfs.mpg.demailto:sunilwilfred@cpfs.mpg.de> wrote:
Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message *** -------------------------------------------------------------------------- mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault). --------------------------------------------------------------------------
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil
<CMakeCache.txt>
In my previous mail i forgot to mention that the sample MPI program complies and works fine without any errors. I am kindly requesting you to please suggest how to solve the mpi problem in ALPS which i have mentioned in my previous mails
with best regards Sunil ________________________________ From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of D´Souza, Sunil Wilfred [sunilwilfred@cpfs.mpg.de] Sent: Monday, December 21, 2015 8:10 AM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Dear ALPS users, As per your suggestion, I have tried running a simple MPI program by using the following commands
mpic++ -o monte_carlo_mpi monte_carlo_mpi.cpp
mpirun -np 8 monte_carlo_mpi
I have attached the sample MPI program ( computes PI by the Monte Carlo method) which i have used for checking the MPI. Please suggest me has to how to solve the mpi problem in ALPS simulations.
with best regards
Sunil
________________________________ From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of Matthias Troyer [troyer@phys.ethz.ch] Sent: Thursday, December 17, 2015 4:18 PM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Have you tried running a simple MPI code that you compiled yourself using the same MPI installation?
On Dec 17, 2015, at 16:13, D´Souza, Sunil Wilfred <sunilwilfred@cpfs.mpg.demailto:sunilwilfred@cpfs.mpg.de> wrote:
Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message *** -------------------------------------------------------------------------- mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault). --------------------------------------------------------------------------
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil
<CMakeCache.txt>
You now need to make sure that when building ALPS you use the same MPI version. Are there by any chance multiple MPI versions installed on your machine?
On Dec 21, 2015, at 08:14, D´Souza, Sunil Wilfred sunilwilfred@cpfs.mpg.de wrote:
In my previous mail i forgot to mention that the sample MPI program complies and works fine without any errors. I am kindly requesting you to please suggest how to solve the mpi problem in ALPS which i have mentioned in my previous mails
with best regards Sunil From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of D´Souza, Sunil Wilfred [sunilwilfred@cpfs.mpg.de] Sent: Monday, December 21, 2015 8:10 AM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Dear ALPS users, As per your suggestion, I have tried running a simple MPI program by using the following commands
mpic++ -o monte_carlo_mpi monte_carlo_mpi.cpp
mpirun -np 8 monte_carlo_mpi
I have attached the sample MPI program ( computes PI by the Monte Carlo method) which i have used for checking the MPI. Please suggest me has to how to solve the mpi problem in ALPS simulations.
with best regards Sunil
From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of Matthias Troyer [troyer@phys.ethz.ch] Sent: Thursday, December 17, 2015 4:18 PM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Have you tried running a simple MPI code that you compiled yourself using the same MPI installation?
On Dec 17, 2015, at 16:13, D´Souza, Sunil Wilfred sunilwilfred@cpfs.mpg.de wrote:
Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message ***
mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault).
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil <CMakeCache.txt>
Dear ALPS users, I have used exactly the same compiler links ( /opt/openmpi/1.4.5/gcc/bin/mpicxx ) ( /opt/openmpi/1.4.5/gcc/bin/mpiexec ) as present in the “CmakeCache.txt” configuration file of my ALPS installation to recheck a simple mpi program (monte_carlo_mpi.cpp) and I find that this simple program complies properly and gives a correct parallel execution. but then ALPS does not work in the parallel mode
As far as the different versions of mpi are concerned , I found that there are two versions of mpi on navigating the files in my desktop, In the following extract,I have mentioned the specific versions as well sunil@LIN21:~> /opt/openmpi/1.4.5/gcc/bin/mpirun --version mpirun (Open MPI) 1.4.5
sunil@LIN21:~> /usr/lib64/mpi/gcc/openmpi/bin/mpirun --version mpirun (Open MPI) 1.7.2
Please advice, how to go about in solving the mpi paralleization problem
with best regards Sunil
________________________________ From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of Matthias Troyer [troyer@phys.ethz.ch] Sent: Monday, December 21, 2015 8:27 AM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
You now need to make sure that when building ALPS you use the same MPI version. Are there by any chance multiple MPI versions installed on your machine?
On Dec 21, 2015, at 08:14, D´Souza, Sunil Wilfred <sunilwilfred@cpfs.mpg.demailto:sunilwilfred@cpfs.mpg.de> wrote:
In my previous mail i forgot to mention that the sample MPI program complies and works fine without any errors. I am kindly requesting you to please suggest how to solve the mpi problem in ALPS which i have mentioned in my previous mails
with best regards Sunil ________________________________ From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.chmailto:comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of D´Souza, Sunil Wilfred [sunilwilfred@cpfs.mpg.demailto:sunilwilfred@cpfs.mpg.de] Sent: Monday, December 21, 2015 8:10 AM To: comp-phys-alps-users@lists.phys.ethz.chmailto:comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Dear ALPS users, As per your suggestion, I have tried running a simple MPI program by using the following commands
mpic++ -o monte_carlo_mpi monte_carlo_mpi.cpp
mpirun -np 8 monte_carlo_mpi
I have attached the sample MPI program ( computes PI by the Monte Carlo method) which i have used for checking the MPI. Please suggest me has to how to solve the mpi problem in ALPS simulations.
with best regards
Sunil
________________________________ From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.chmailto:comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of Matthias Troyer [troyer@phys.ethz.chmailto:troyer@phys.ethz.ch] Sent: Thursday, December 17, 2015 4:18 PM To: comp-phys-alps-users@lists.phys.ethz.chmailto:comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Have you tried running a simple MPI code that you compiled yourself using the same MPI installation?
On Dec 17, 2015, at 16:13, D´Souza, Sunil Wilfred <sunilwilfred@cpfs.mpg.demailto:sunilwilfred@cpfs.mpg.de> wrote:
Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message *** -------------------------------------------------------------------------- mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault). --------------------------------------------------------------------------
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil
<CMakeCache.txt>
Does that program work?
On Dec 21, 2015, at 08:10, D´Souza, Sunil Wilfred sunilwilfred@cpfs.mpg.de wrote:
Dear ALPS users, As per your suggestion, I have tried running a simple MPI program by using the following commands
mpic++ -o monte_carlo_mpi monte_carlo_mpi.cpp
mpirun -np 8 monte_carlo_mpi
I have attached the sample MPI program ( computes PI by the Monte Carlo method) which i have used for checking the MPI. Please suggest me has to how to solve the mpi problem in ALPS simulations.
with best regards Sunil
From: Comp-phys-alps-users [comp-phys-alps-users-bounces@lists.phys.ethz.ch] on behalf of Matthias Troyer [troyer@phys.ethz.ch] Sent: Thursday, December 17, 2015 4:18 PM To: comp-phys-alps-users@lists.phys.ethz.ch Subject: Re: [ALPS-users] Problem in running the ALPS (spinmc) code in the parallel mode aith MPI
Have you tried running a simple MPI code that you compiled yourself using the same MPI installation?
On Dec 17, 2015, at 16:13, D´Souza, Sunil Wilfred sunilwilfred@cpfs.mpg.de wrote:
Dear ALPS users, I am trying to run a spinmc simulation in the parallel mode by using the command mpiexec -np 2 spinmc --mpi parm2a.in.xml --Tmin 5
On executing the command I am getting the following error messages
[LIN21:20344] *** Process received signal *** [LIN21:20344] Signal: Segmentation fault (11) [LIN21:20344] Signal code: Address not mapped (1) [LIN21:20344] Failing at address: (nil) [LIN21:20344] [ 0] /lib64/libpthread.so.0(+0xf9f0) [0x7f85f73969f0] [LIN21:20344] [ 1] /opt/openmpi/1.4.5/gcc/lib64/openmpi/mca_pml_v.so(+0x1b62) [0x7f85f1359b62] [LIN21:20344] [ 2] /opt/openmpi/1.4.5/gcc/lib64/libopen-pal.so.0(mca_base_components_close+0x76) [0x7f85f66286c6] [LIN21:20344] [ 3] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(mca_pml_base_select+0x301) [0x7f85f6b2cd11] [LIN21:20344] [ 4] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(+0x39acc) [0x7f85f6aecacc] [LIN21:20344] [ 5] /opt/openmpi/1.4.5/gcc/lib64/libmpi.so.0(MPI_Init+0x16b) [0x7f85f6b0ca8b] [LIN21:20344] [ 6] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9comm_initERiRPPcb+0x58) [0x7f85f8327088] [LIN21:20344] [ 7] /home/sunil/ALPS2.2/lib/libalps.so.2(_ZN4alps9scheduler5startEiPPcRKNS0_7FactoryE+0x37) [0x7f85f85678d7] [LIN21:20344] [ 8] spinmc(main+0x15) [0x449a85] [LIN21:20344] [ 9] /lib64/libc.so.6(__libc_start_main+0xf5) [0x7f85f454bbe5] [LIN21:20344] [10] spinmc() [0x449dd5] [LIN21:20344] *** End of error message ***
mpiexec noticed that process rank 0 with PID 20344 on node LIN21 exited on signal 11 (Segmentation fault).
I have tried running a band structure code in the parallel mode using the mpirun and i am having no problems in running that code. but in the case of ALPS, I am encountering the above problem. I am attaching the cmake txt file for your consideration to make sure that all the fields are correct. I have crosschecked the MPI library path in cmake txt file and i find it to be alright. I am not able to figure out what exactly is causing this problem. I am kindly requesting you to provide me help in solving this issue.
I thank you in advance for your time and consideration.
with best regards Sunil
<CMakeCache.txt>
<monte_carlo_mpi.cpp>
comp-phys-alps-users@lists.phys.ethz.ch