Hi All,
We are having a strange issue with java out of memory issues at the same time of day every few days.
We are using a Bitnami Azure deployment package that has openfire 4.7.4 (which we realise needs to be upgraded to 4.7.5 due to admin bypass issue.)
Its also configured using mysql-mariadb 10.6.11 and apache 2.4.54
We have around 30 devices connected all the time sending telemetry messages to 3 listener services.
During normal operation the memory usage reported on the openfire server page varies from a couple hundred megabytes to a couple of gigabytes. and ~1% cpu But at around 5:10-5:15 pm +8 (21:10-21:15 UTC) something happens and the memory usage skyrockets and the cpu usage hits 100% every 1,2 or 3 days this causes java to throw an out of memory error and die. Restarting the openfire service doesnt work, it requires rebooting the entire VM to bring it back to life.
There does appear to be increases in network traffic both in and out but mainly out of the server at this time.
In the openfire service logs it appears that about 5pm there are a whole bunch of these logs:
2023.06.21 21:36:02 ESC[33mWARN ESC[m [socket_c2s-thread-8]: org.jivesoftware.openfire.nio.ConnectionHandler - Closing connection due to exception in session: (
0x0000007A: nio socket, server, /[client ip address]:46992 => /[server ip]:5222)
java.io.IOException: Closing session that seems to be stalled. Preventing OOM
[exception information]
then
2023.06.20 09:23:44 ESC[33mWARN ESC[m [Jetty-QTP-AdminConsole-1917]: org.eclipse.jetty.util.thread.QueuedThreadPool -
java.lang.OutOfMemoryError: Java heap space
and thats where the logs stop until the VM is restarted and everything comes back up
We are going to turn on trace logging for a half an hour from 5pm to see if that captures anything more in the logs.
Is there any known issue that would cause something like this at the same time every day or few days?
Is there any obvious things we can try to debug this issue and work out what is giong on?
Kind Regards,
InteliCare team