Hi everyone, I've been experiencing some repeated behavior on my omv server and I'm hoping to any help or insights as to how to address this.
My setup includes a couple of mergerfs filesystems mounted via NFS. The physical drives are connected via an HBA card and passed through to omv. OMV ver. is 6.4.0-3 (Shaitan) in a debian container. Everything is up to date etc, and persists after reboots/restarting services individually. Mostly video files.
For the most part everything has been stable, but recently I've been having random periods of huge spiking in server load and IO wait, which also makes the NFS shares unresponsive, as well as my VMs which have these shares mounted, which normally results in me having to force kill the process and/or reboot the machine.
During troubleshooting, I noticed a large number of blkid -o full
and omv-engined commands appearing in ps aux
. iotop
sometimes shows a significant IO percentage for mergerfs, although weirdly not for the most recent occurrence.
I noticed multiple instances of the omv-engined
daemon running simultaneously, which seemed unusual.
for example: ps -ef
root 1046841 614244 0 18:32 ? 00:00:00 omv-engined
root 1046847 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046850 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046855 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046859 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046862 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046865 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046867 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046871 1046841 0 18:32 ? 00:00:00 omv-engined
root 1046875 1046841 0 18:32 ? 00:00:00 omv-engined
and my dashboard: