Dear Matthias, I attach input file (parm5c), modified models.xml and lattices.xml, stdout and stderr in separate files (out, out2)
Regards, Mateusz Łącki
On Dec 30, 2011, at 5:36 PM, Matthias Troyer wrote:
Hi,
I can only look into this if you send you input file.
Matthias
On 30 Dec 2011, at 12:05, Mateusz Łącki wrote:
Dear All, I have set up some computation which failed:
MPI_ABORT was invoked on rank 27 in communicator MPI_COMM_WORLD with errorcode -2.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them.
mpiexec has exited due to process rank 27 with PID 15245 on node clone18 exiting without calling "finalize". This may have caused other processes in the application to be terminated by signals sent by mpiexec (as reported here).
just before that (in the output file, not sure about the time):
Avoided problem q = -0 state1 = 0 state2 = 0 bond_type = 0 id 2047013814 2047013814 Spin configuration: Wormheads at 78 0.999997 and 78 0.259301 Site: 0 Kink : [ 0.119336 : 0 ] Kink : [ 0.124065 : 1 ] Kink : [ 0.174815 : 0 ] Kink : [ 0.17605 : 1 ] Kink : [ 0.335094 : 2 ] Kink : [ 0.368865 : 1 ] (...) Site: 299 Kink : [ 0.00590279 : 2 ] Kink : [ 0.0326616 : 1 ] Kink : [ 0.0697665 : 0 ] Kink : [ 0.0977223 : 1 ] Kink : [ 0.254292 : 2 ] Kink : [ 0.256147 : 1 ] Kink : [ 0.328286 : 2 ] Kink : [ 0.329838 : 1 ] Kink : [ 0.405038 : 0 ] Kink : [ 0.438803 : 1 ] Kink : [ 0.487034 : 2 ] Kink : [ 0.503331 : 1 ] Kink : [ 0.812159 : 2 ] Kink : [ 0.827811 : 1 ]
Is this related? I am not sure whether this output indicates an error.
Regards, Mateusz
On Dec 29, 2011, at 9:59 AM, Matthias Troyer wrote:
Yes, indeed
On Dec 29, 2011, at 9:53 AM, Mateusz Łącki wrote:
Dear Matthias,
Thank you for your answer. If I understand correctly the problem is solved and results take into account this special case now?
Regards, Mateusz
This is a debug message which I added a while ago when solving a problem that we had because of finite resolution of floating point numbers. There is a chance of 1e-16 per site and unit imaginary time interval that two bosons hop away from two neighboring sites at exactly the same time. This case needs special consideration, and the notice was added to indicate that this case had happened.
Matthias
On Dec 29, 2011, at 9:39 AM, Mateusz Łącki wrote:
Dear All, While running some QMC (worm) computation by MPI over several nodes I noticed "Avoided problem" message appearing fro time to time. It does not seem particularly dangerous, but is it possible to find out what was the problem in the first place?
Regards, Mateusz Łącki