7 Replies Latest reply on Oct 27, 2010 12:20 PM by LG,x

    YAMLP(Yet Another Memory Leak Problem)


      Hi guys,


      I'm still having big troubles with 3.7.0beta (3.6.4 had the same kind of problem).

      Usually the server handles 400 clients during work hours. I've seen bigger deployments in others posts, so this is ok.

      We have a severe memory leak where the JVM goes up to the 2096MB Java memory limit.

      It appears the same thing happens with just a few clients (35 at the time I'm testing).

      Clients are essentially Spark/Pidgin.


      The server is a Linux/x64 Debian Squeeze with Sun JRE/JDK6, LDAP auth, Database(MySQL) groups.

      We have 3 dozen of groups being shared between each 350 users (enterprise groups).

      It is connected to another Openfire 3.6.4 server which seems to work correctly (~50 clients, Windows2K3x64/SQLLite).


      I've been able to make a heap dump parsed with Eclipse Memory Analyser (very good soft by the way).

      Here are parts of the results(~40 clients)(tested after a fresh start to have the thinest memory footprint) :

      One instance of "java.util.TaskQueue" loaded by "<system class  loader>" occupies 776 748 384 (90,98%) bytes. The instance is  referenced by org.jivesoftware.openfire.pep.PEPService @ 0x7f3b4699f6b0 ,  loaded by "org.jivesoftware.openfire.starter.JiveClassLoader @  0x7f3b11cf5ab8". The memory is accumulated in one instance of  "java.util.TimerTask[]" loaded by "<system class  loader>".



      org.jivesoftware.openfire.starter.JiveClassLoader  @ 0x7f3b11cf5ab8


      Label Number Of Objects Used Heap Size Retained Heap Size
      org.jivesoftware.openfire.pep.PEPService$11 157101 816776 731 944
      The warn log reports a lot of (~1 per minute):
      2010.10.21 23:39:54 Cache Roster was full, shrinked to 90% in 0ms
      Is it a bug?
      Is there a workaround?(property value?)
      Is there a patch?