Hi,
First of all thanks for the nice work with OMV.
I have a serious problem though... After beeing using OMV for the last 2 1/2 months my systems stopped booting after the first reboot following the last kernel update, I can't recall but I beleive there was a kernel update midle last week. My system runs 24/7 so I didn't do a reboot at the time. But in order to change the place where my system was beeing kept I had to shut it down last wednesday... After that the system never came up by it self anymore. The Raid-6 arrays fail to start and the system stops asking for troubleshooting.
I managed to find out that the problem has something to do with the partition detection system on this new update because if I stop the "bogus" raid arrays and run a partprobe on each drive I can then run a mdam-raid start and the raid arrays get started properly...
Any idea?
here his a boot log and my RAID configuration:
Thu Mar 14 16:44:04 2013: Setting parameters of disc: /dev/sdg.
Thu Mar 14 16:44:04 2013: /dev/sdd ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sdh.
Thu Mar 14 16:44:04 2013: /dev/sdb ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sdi.
Thu Mar 14 16:44:04 2013: /dev/sdc.
Thu Mar 14 16:44:04 2013: /dev/sda ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sde ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sdf ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: Setting preliminary keymap...done.
Thu Mar 14 16:44:04 2013: Activating swap...done.
Thu Mar 14 16:44:04 2013: Checking root file system...fsck from util-linux-ng 2.17.2
Thu Mar 14 16:44:04 2013: /dev/sdc1: clean, 173581/9371648 files, 2110611/37479680 blocks
Thu Mar 14 16:44:04 2013: done.
Thu Mar 14 16:44:04 2013: Cleaning up ifupdown....
Thu Mar 14 16:44:04 2013: Assembling MD array mdANANAS:0...^[[31mfailed (not enough devices).^[[39;49m
Thu Mar 14 16:44:04 2013: Assembling MD array mdANANAS:0_0...^[[31mfailed (not enough devices).^[[39;49m
Thu Mar 14 16:44:04 2013: Generating udev events for MD arrays...done.
Thu Mar 14 16:44:04 2013: Loading kernel modules...done.
Thu Mar 14 16:44:04 2013: Setting up LVM Volume Groups Reading all physical volumes. This may take a while...
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid XyE2h2-IjCT-z1ae-sC5c-RGmW-T0Ax-GBUpe1.
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid 9OrovE-U9hd-H1wi-pLqI-3J4u-OcGA-Ego06a.
Thu Mar 14 16:44:04 2013: Found volume group "StorageVG" using metadata type lvm2
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid XyE2h2-IjCT-z1ae-sC5c-RGmW-T0Ax-GBUpe1.
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid 9OrovE-U9hd-H1wi-pLqI-3J4u-OcGA-Ego06a.
Thu Mar 14 16:44:04 2013: Refusing activation of partial LV StorageLV. Use --partial to override.
Thu Mar 14 16:44:04 2013: Refusing activation of partial LV BackupsLV. Use --partial to override.
Thu Mar 14 16:44:04 2013: 0 logical volume(s) in volume group "StorageVG" now active
Thu Mar 14 16:44:04 2013: ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: Activating lvm and md swap...done.
Thu Mar 14 16:44:04 2013: Checking file systems...fsck from util-linux-ng 2.17.2
Thu Mar 14 16:44:05 2013: fsck.ext4: Unable to resolve 'UUID=0e79526d-6234-4306-8a9c-efe47cee3a93'
Thu Mar 14 16:44:05 2013: fsck.ext4: Unable to resolve 'UUID=9be969eb-e401-4da7-9a29-434633dd7c47'
Thu Mar 14 16:44:05 2013: fsck died with exit status 8
Thu Mar 14 16:44:05 2013: ^[[31mfailed (code 8).^[[39;49m
Thu Mar 14 16:44:05 2013: File system check failed. A log is being saved in /var/log/fsck/checkfs if that location is writable. Please repair the file system manually. ... ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:05 2013: A maintenance shell will now be started. CONTROL-D will terminate this shell and resume system boot. ... ^[[33m(warning).^[[39;49m
Thu Mar 14 16:44:05 2013: Give root password for maintenance
Thu Mar 14 16:44:05 2013: (or type Control-D to continue):
Alles anzeigen
This is the RAID layout after this failed boot:
Thu Mar 14 16:44:50 2013: root@ANANAS:~# cat /proc/mdd\00stat
Thu Mar 14 16:44:58 2013: Personalities : [raid6] [raid5] [raid4]
Thu Mar 14 16:44:58 2013: md125 : inactive sdf1[8](S) sde1[6](S) sdd1[1](S)
Thu Mar 14 16:44:58 2013: 2929680384 blocks super 1.2
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: md126 : inactive sdf2[6](S) sde2[3](S) sdd2[2](S)
Thu Mar 14 16:44:58 2013: 2929686528 blocks super 1.2
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: md127 : active (auto-read-only) raid5 sdi[0] sdb[1]
Thu Mar 14 16:44:58 2013: 3907028992 blocks level 5, 64k chunk, algorithm 2 [3/2] [UU_]
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: md0 : active (auto-read-only) raid6 sdd3[0] sdh1[4] sdg1[3] sdf3[2] sde3[1]
Thu Mar 14 16:44:58 2013: 2930277888 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU]
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: unused devices: <none>
Alles anzeigen
And this is the RAID layout after running partprobe (the RAID's are in read-only on purpose):
This is my correct RAID layout.
root@ANANAS:/var/log# cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4]
md126 : active (auto-read-only) raid6 sdd3[0] sdh1[4] sdg1[3] sdf3[2] sde3[1]
2930277888 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU]
md127 : active (auto-read-only) raid6 sdf1[8] sdi1[5] sde1[6] sdb1[3] sda1[7] sdd1[1]
3906238464 blocks super 1.2 level 6, 512k chunk, algorithm 2 [6/6] [UUUUUU]
md0 : active (auto-read-only) raid6 sda2[0] sdb2[5] sdi2[4] sde2[3] sdd2[2] sdf2[6]
3906246656 blocks super 1.2 level 6, 512k chunk, algorithm 2 [6/6] [UUUUUU]
unused devices: <none>
Alles anzeigen
Any help would be greatly appreciated.
Cheers