caml-list - the Caml user's mailing list
 help / color / mirror / Atom feed
From: "Török Edwin" <edwin+ml-ocaml@etorok.net>
To: caml-list@inria.fr
Subject: Re: [Caml-list] Memory fragmentation
Date: Fri, 20 Apr 2012 19:41:57 +0300	[thread overview]
Message-ID: <4F9191D5.6060809@etorok.net> (raw)
In-Reply-To: <CAHgca+_OD0twkZ6WPc2bkQtEAS6Qnj1ogTDLE-S4esw6GKAV9A@mail.gmail.com>

On 04/20/2012 06:12 PM, SerP wrote:
> Hi!
> We have developped a daemon on ocaml using the lwt lib(libev). It processes about 800 requests per second, but it increases to 2 Gb in memory for a hour of work. We have studied the problem for a long
> time, using mtrace, mallinfo and other tools,
> and we tried to change GC params. We found out that setting up GC.minor_heap_size=10Mb а major_heap_increment=2Мb will make the daemon grow slower.
> mallinfо shows that the memory (1,5G) is allocated using mmap (blkhd field in mallinfo struct) and arena size - about 20M In this case, first calls of GC.compact compress the daemon to 200 Mb (live
> -110Mb) , and mallinfo show decreasing of total size of memory allocated by mmap. The daemon keeps growing, but it allocates the memory to arena (using brk), and calls of GC.compact leads to decrease
> of heap_bytes to 200M, but arena size (1.5 Gb) does not decrease, just doing less uordblks and more fordblks (fileds I.e., after first calls of GC.compact, the daemon starts using brk much more than
> mmap, and cant memory given back to the system.

Sounds like a fragmentation problem in glibc's heap that causes increased VIRT usage.
IIRC glibc has a default threshold of 128k for deciding to use brk() vs mmap(), and OCaml's default major_heap_increment (124k) is just below that,
which would explain the behaviour you see, and also explain why using a larger major_heap_increment (2M) improves the situation:
 brk() cannot free "holes", while mmap arenas can.

You can use malloc_stats() to check whether OCaml gives back all memory to glibc or not, and you can try to use
an alternative malloc implementation (like jemalloc) that might cope better with fragmentation.

However this doesn't explain why memory keeps growing after a Gc.compact, unless calls to malloc
are mixed between the OCaml runtime, Lwt and libev. If that is the case then maybe it would help
to try and allocate memory in the OCaml's runtime by using mmap() directly.

Is your application completely single-threaded (since you're using lwt), or do you also use multiple threads?
I think that glibc caches mmap-ed areas even if you free them (i.e. they're still mapped into your process
but they might be changed via mprotect() to PROT_NONE), which is especially a problem if you use multiple threads.

Best regards,
--Edwin

  parent reply	other threads:[~2012-04-20 16:42 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <CAHgca+9-K71NynNVzS0Stnb2v5Tc6n5ecCnJHZorhuB8y8EdTw@mail.gmail.com>
2012-04-20 15:12 ` SerP
2012-04-20 15:27   ` [Caml-list] " SerP
2012-04-20 16:41   ` Török Edwin [this message]
2012-04-20 18:32     ` [Caml-list] " SerP
2012-04-20 21:32   ` ygrek
2012-04-21  7:26   ` rixed
2012-04-24 19:24     ` SerP

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4F9191D5.6060809@etorok.net \
    --to=edwin+ml-ocaml@etorok.net \
    --cc=caml-list@inria.fr \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).