Tracking memory spikes causing process issues

Hi,

I'm running on FreeBSD 12.2 running a group of python processes under supervisor. We have now twice run into an issue where processes have been caused to enter either T or I status causing access to them to be no longer possible. In both circumstances, kill commands to the processes were ignored. On both occasions a full server restart resolved the issue, but I am looking to track what got us to that point. Currently, under normal workday running, roughly a third of the memory is in use and none of the swap, according to the top command, with around 10% of the CPU in use. When the issue was active, and programs were hanging up, a much larger proportion of both regular and swap memory was in use, and in the circumstance where this was a cause of the problem I would like to take a look at what might have been causing that at the time. Is there a good way of looking at memory usage over time and what process might have been causing the issue?

Many Thanks in advance for any help.
 
Back
Top