On Wed, Sep 8, 2010 at 10:51 AM, Sylvain Le Gall wrote: > On 08-09-2010, Eray Ozkural wrote: > > --===============0522474025== > > Content-Type: multipart/alternative; > boundary=0016369204e79d9601048fb8ae36 > > > > --0016369204e79d9601048fb8ae36 > > Content-Type: text/plain; charset=ISO-8859-1 > > > > I'm recently getting errors that are past MPI_Finalize. Since both > > init/final and communicator allocation is managed by ocamlmpi, is it > > possible this is a bug with the library? Have you ever seen something > like > > this? > > > > Using openmpi on OS X. Here is the log message: > > > > *** An error occurred in MPI_Comm_free > > *** after MPI was finalized > > *** MPI_ERRORS_ARE_FATAL (goodbye) > > > > In the code I'm using both point-to-point and collective communication, > and > > as far as I know the code is correct. Could this be due to memory > > corruption, or should this never happen? > > > > Maybe, you can give a minimal code to reproduce this error? > > Hmm, not really its a complex code but I just ran the debug version in parallel with exactly the same parameters and there is absolutely no problem with that. All communication is synchronous so timing cannot be an issue (since the debug build is naturally slower). AFAICT it's not a memory problem because no bound errors are reported in the debug build (was it on by default?). I think it's a lower-level problem than my code. This could happen if some of that resource allocation is done in different threads, for instance. Can you give me any ideas to trace the source of this problem? Best, -- Eray Ozkural, PhD candidate. Comp. Sci. Dept., Bilkent University, Ankara http://groups.yahoo.com/group/ai-philosophy http://myspace.com/arizanesil http://myspace.com/malfunct