Raid is still working because I used Partitions, instead of full drives.
If I had run mdadm --zero-superblock /dev/sda1 then yes it would destroy my array.
But that was not the case.
Raid is still working because I used Partitions, instead of full drives.
If I had run mdadm --zero-superblock /dev/sda1 then yes it would destroy my array.
But that was not the case.
The only other symptom I have is that no disc displays Serial number information in the omv web interface...
But yes apparently that would be the only way to get my system back ...
Help anyone?
Cheers
Ok the principle is:
Performance is not an issue.
RAID-6 is mandatory for double disc failure.
Maximum amount of usable RAID space is desired.
So if you can propose a diferent way to get 10TB Usable out of 3x 3TB Drives and 3x 2TB Drives and 2x 1TB Drives. I would apreciate.
About my problem, system does not boot because some of the partitions are not recognized and do not get to /proc/partitions unless I manually run partprobe after the boot fails.
Hi
Kernel is:
root@ANANAS:~# uname -a
Linux ANANAS 2.6.32-5-amd64 #1 SMP Mon Feb 25 00:26:11 UTC 2013 x86_64 GNU/Linux
but under boot I only have:
root@ANANAS:/boot# ls
config-2.6.32-5-amd64 grub initrd.img-2.6.32-5-amd64 System.map-2.6.32-5-amd64 vmlinuz-2.6.32-5-amd64
UUID is not helping because the issue is that not all the drive partitions are recognized under /proc/partitions until I manually run partprobe.
Thanks for your help.
Hi,
First of all thanks for the nice work with OMV.
I have a serious problem though... After beeing using OMV for the last 2 1/2 months my systems stopped booting after the first reboot following the last kernel update, I can't recall but I beleive there was a kernel update midle last week. My system runs 24/7 so I didn't do a reboot at the time. But in order to change the place where my system was beeing kept I had to shut it down last wednesday... After that the system never came up by it self anymore. The Raid-6 arrays fail to start and the system stops asking for troubleshooting.
I managed to find out that the problem has something to do with the partition detection system on this new update because if I stop the "bogus" raid arrays and run a partprobe on each drive I can then run a mdam-raid start and the raid arrays get started properly...
Any idea?
here his a boot log and my RAID configuration:
Thu Mar 14 16:44:04 2013: Setting parameters of disc: /dev/sdg.
Thu Mar 14 16:44:04 2013: /dev/sdd ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sdh.
Thu Mar 14 16:44:04 2013: /dev/sdb ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sdi.
Thu Mar 14 16:44:04 2013: /dev/sdc.
Thu Mar 14 16:44:04 2013: /dev/sda ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sde ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: /dev/sdf ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: Setting preliminary keymap...done.
Thu Mar 14 16:44:04 2013: Activating swap...done.
Thu Mar 14 16:44:04 2013: Checking root file system...fsck from util-linux-ng 2.17.2
Thu Mar 14 16:44:04 2013: /dev/sdc1: clean, 173581/9371648 files, 2110611/37479680 blocks
Thu Mar 14 16:44:04 2013: done.
Thu Mar 14 16:44:04 2013: Cleaning up ifupdown....
Thu Mar 14 16:44:04 2013: Assembling MD array mdANANAS:0...^[[31mfailed (not enough devices).^[[39;49m
Thu Mar 14 16:44:04 2013: Assembling MD array mdANANAS:0_0...^[[31mfailed (not enough devices).^[[39;49m
Thu Mar 14 16:44:04 2013: Generating udev events for MD arrays...done.
Thu Mar 14 16:44:04 2013: Loading kernel modules...done.
Thu Mar 14 16:44:04 2013: Setting up LVM Volume Groups Reading all physical volumes. This may take a while...
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid XyE2h2-IjCT-z1ae-sC5c-RGmW-T0Ax-GBUpe1.
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid 9OrovE-U9hd-H1wi-pLqI-3J4u-OcGA-Ego06a.
Thu Mar 14 16:44:04 2013: Found volume group "StorageVG" using metadata type lvm2
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid XyE2h2-IjCT-z1ae-sC5c-RGmW-T0Ax-GBUpe1.
Thu Mar 14 16:44:04 2013: Couldn't find device with uuid 9OrovE-U9hd-H1wi-pLqI-3J4u-OcGA-Ego06a.
Thu Mar 14 16:44:04 2013: Refusing activation of partial LV StorageLV. Use --partial to override.
Thu Mar 14 16:44:04 2013: Refusing activation of partial LV BackupsLV. Use --partial to override.
Thu Mar 14 16:44:04 2013: 0 logical volume(s) in volume group "StorageVG" now active
Thu Mar 14 16:44:04 2013: ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:04 2013: Activating lvm and md swap...done.
Thu Mar 14 16:44:04 2013: Checking file systems...fsck from util-linux-ng 2.17.2
Thu Mar 14 16:44:05 2013: fsck.ext4: Unable to resolve 'UUID=0e79526d-6234-4306-8a9c-efe47cee3a93'
Thu Mar 14 16:44:05 2013: fsck.ext4: Unable to resolve 'UUID=9be969eb-e401-4da7-9a29-434633dd7c47'
Thu Mar 14 16:44:05 2013: fsck died with exit status 8
Thu Mar 14 16:44:05 2013: ^[[31mfailed (code 8).^[[39;49m
Thu Mar 14 16:44:05 2013: File system check failed. A log is being saved in /var/log/fsck/checkfs if that location is writable. Please repair the file system manually. ... ^[[31mfailed!^[[39;49m
Thu Mar 14 16:44:05 2013: A maintenance shell will now be started. CONTROL-D will terminate this shell and resume system boot. ... ^[[33m(warning).^[[39;49m
Thu Mar 14 16:44:05 2013: Give root password for maintenance
Thu Mar 14 16:44:05 2013: (or type Control-D to continue):
Alles anzeigen
This is the RAID layout after this failed boot:
Thu Mar 14 16:44:50 2013: root@ANANAS:~# cat /proc/mdd\00stat
Thu Mar 14 16:44:58 2013: Personalities : [raid6] [raid5] [raid4]
Thu Mar 14 16:44:58 2013: md125 : inactive sdf1[8](S) sde1[6](S) sdd1[1](S)
Thu Mar 14 16:44:58 2013: 2929680384 blocks super 1.2
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: md126 : inactive sdf2[6](S) sde2[3](S) sdd2[2](S)
Thu Mar 14 16:44:58 2013: 2929686528 blocks super 1.2
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: md127 : active (auto-read-only) raid5 sdi[0] sdb[1]
Thu Mar 14 16:44:58 2013: 3907028992 blocks level 5, 64k chunk, algorithm 2 [3/2] [UU_]
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: md0 : active (auto-read-only) raid6 sdd3[0] sdh1[4] sdg1[3] sdf3[2] sde3[1]
Thu Mar 14 16:44:58 2013: 2930277888 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU]
Thu Mar 14 16:44:58 2013:
Thu Mar 14 16:44:58 2013: unused devices: <none>
Alles anzeigen
And this is the RAID layout after running partprobe (the RAID's are in read-only on purpose):
This is my correct RAID layout.
root@ANANAS:/var/log# cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4]
md126 : active (auto-read-only) raid6 sdd3[0] sdh1[4] sdg1[3] sdf3[2] sde3[1]
2930277888 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU]
md127 : active (auto-read-only) raid6 sdf1[8] sdi1[5] sde1[6] sdb1[3] sda1[7] sdd1[1]
3906238464 blocks super 1.2 level 6, 512k chunk, algorithm 2 [6/6] [UUUUUU]
md0 : active (auto-read-only) raid6 sda2[0] sdb2[5] sdi2[4] sde2[3] sdd2[2] sdf2[6]
3906246656 blocks super 1.2 level 6, 512k chunk, algorithm 2 [6/6] [UUUUUU]
unused devices: <none>
Alles anzeigen
Any help would be greatly appreciated.
Cheers