hello,
I am encountering a problem on my OMV version 6 on which I have activated luks encryption on my 4 disks.
I unlock my 4 disks, the RAID reappears but sometimes the file system goes up and sometimes not. I can't understand why.
When the file system does not remount it is impossible for me to remount this system using the openMediavault interface.
After browsing lot of forums it seems that I am not the only one in this case and I tried to apply the pseudo resolution methods: none work :
- adding the nfsver=3 options on the file system
- check by fsk
- state of raid is ok by mdadm command
- change the order of unlocking disk
- ...
When the problem appears i have this message monit[1283]: 'filesystem_srv_dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7' unable to read filesystem '/srv/dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7' state.
I check this filesystem, no right issues
Below are the logs when the system mounts correctly:
QuoteDisplay MoreOct 12 15:37:54 n1nas kernel: [ 78.072369] md/raid:md0: device dm-3 operational as raid disk 2
Oct 12 15:37:54 n1nas kernel: [ 78.072374] md/raid:md0: device dm-2 operational as raid disk 3
Oct 12 15:37:54 n1nas kernel: [ 78.072375] md/raid:md0: device dm-1 operational as raid disk 1
Oct 12 15:37:54 n1nas kernel: [ 78.072376] md/raid:md0: device dm-0 operational as raid disk 0
Oct 12 15:37:54 n1nas kernel: [ 78.073193] md/raid:md0: raid level 5 active with 4 out of 4 devices, algorithm 2
Oct 12 15:37:54 n1nas kernel: [ 78.080607] md0: detected capacity change from 0 to 105469077504
Oct 12 15:37:54 n1nas systemd[1]: Found device /dev/disk/by-uuid/5dab8dff-b5bc-4efc-895a-bff977f392a7.
Oct 12 15:37:54 n1nas systemd[1]: mdadm-last-resort@md0.timer: Succeeded.
Oct 12 15:37:54 n1nas systemd[1]: Stopped Timer to wait for more drives before activating degraded array md0..
Oct 12 15:37:54 n1nas systemd[1]: Started MD array monitor.
Oct 12 15:37:54 n1nas mdadm[2322]: mdadm: No mail address or alert command - not monitoring.
Oct 12 15:37:54 n1nas systemd[1]: Starting File System Check on /dev/disk/by-uuid/5dab8dff-b5bc-4efc-895a-bff977f392a7...
Oct 12 15:37:54 n1nas systemd[1]: mdmonitor.service: Main process exited, code=exited, status=1/FAILURE
and when its not OK:
QuoteDisplay MoreOct 12 15:35:50 n1nas kernel: [ 127.316608] md/raid:md0: device dm-3 operational as raid disk 0
Oct 12 15:35:50 n1nas kernel: [ 127.316618] md/raid:md0: device dm-2 operational as raid disk 2
Oct 12 15:35:50 n1nas kernel: [ 127.316621] md/raid:md0: device dm-1 operational as raid disk 3
Oct 12 15:35:50 n1nas kernel: [ 127.316623] md/raid:md0: device dm-0 operational as raid disk 1
Oct 12 15:35:50 n1nas kernel: [ 127.318250] md/raid:md0: raid level 5 active with 4 out of 4 devices, algorithm 2
Oct 12 15:35:50 n1nas kernel: [ 127.330529] md0: detected capacity change from 0 to 105469077504
Oct 12 15:35:50 n1nas systemd[1]: mdadm-last-resort@md0.timer: Succeeded.
Oct 12 15:35:50 n1nas systemd[1]: Stopped Timer to wait for more drives before activating degraded array md0..
Oct 12 15:35:50 n1nas systemd[1]: Started MD array monitor.
Oct 12 15:35:50 n1nas mdadm[2409]: mdadm: No mail address or alert command - not monitoring.
Oct 12 15:35:50 n1nas systemd[1]: mdmonitor.service: Main process exited, code=exited, status=1/FAILURE
Oct 12 15:35:50 n1nas systemd[1]: mdmonitor.service: Failed with result 'exit-code'.
Oct 12 15:35:50 n1nas monit[1213]: 'filesystem_srv_dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7' trying to restart
Oct 12 15:35:50 n1nas monit[1213]: 'mountpoint_srv_dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7' status failed (1) -- /srv/dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7 is not a mountpoint
Oct 12 15:35:50 n1nas monit[1213]: 'mountpoint_srv_dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7' status failed (1) -- /srv/dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7 is not a mountpoint
As we can see there seems to be a problem with the mdmonitor service.
I tried to run the command: omv-salt deploy run monit. Sometimes this resolves the situation, sometimes not.
But I can in any case mount the file system by hand by: /usr/bin/mount '/srv/dev-disk-by-uuid-5dab8dff-b5bc-4efc-895a-bff977f392a7'
But in itself it is not a solution and does not give me confidence in this system at all. I store the company's backups there. IF one day I definitely can't remount the file system....
If anyone has any leads I'm interested
Thank a lot
Arnaud