Lost Raid 5 after assembling new hard disk drive

  • Hello everybody


    i got a real problem on raid 5.
    How i got the problem
    At first i had an SMART Error
    shut down the system
    unplug fault drive
    got a new hard disk
    plug the new
    reboot


    On the GUI i have no Raid, drive are there


    ##########
    console outputs


    cat /etc/mdadm/mdadm.conf


    ################


    root@nas-omv:~# cat /etc/mdadm/mdadm.conf
    # mdadm.conf
    #
    # Please refer to mdadm.conf(5) for information about this file.
    #


    # by default, scan all partitions (/proc/partitions) for MD superblocks.
    # alternatively, specify devices to scan, using wildcards if desired.
    # Note, if no DEVICE line is present, then "DEVICE partitions" is assumed.
    # To avoid the auto-assembly of RAID devices a pattern that CAN'T match is
    # used if no RAID devices are configured.
    DEVICE partitions


    # auto-create devices with Debian standard permissions
    CREATE owner=root group=disk mode=0660 auto=yes


    # automatically tag new arrays as belonging to the local system
    HOMEHOST <system>


    # definitions of existing MD arrays
    ARRAY /dev/md/FirstRaid5 metadata=1.2 name=NAS-OMV:FirstRaid5 UUID=1fcf9705:1c13cd44:c9bcf8d4:5c66fb11


    # instruct the monitoring daemon where to send mail alerts


    ################


    second output on consol:
    sudo mdadm --misc --examine /dev/sd[abc]


    ################


    root@nas-omv:~# sudo mdadm --misc --examine /dev/sda
    /dev/sda:
    Magic : a92b4efc
    Version : 1.2
    Feature Map : 0x1
    Array UUID : 1fcf9705:1c13cd44:c9bcf8d4:5c66fb11
    Name : NAS-OMV:FirstRaid5
    Creation Time : Fri Nov 4 21:37:43 2016
    Raid Level : raid5
    Raid Devices : 3


    Avail Dev Size : 5860271024 (2794.39 GiB 3000.46 GB)
    Array Size : 5860270080 (5588.79 GiB 6000.92 GB)
    Used Dev Size : 5860270080 (2794.39 GiB 3000.46 GB)
    Data Offset : 262144 sectors
    Super Offset : 8 sectors
    Unused Space : before=262056 sectors, after=944 sectors
    State : clean
    Device UUID : 751ecf5b:6fd057ae:4fbf95d7:939777d0


    Internal Bitmap : 8 sectors from superblock
    Update Time : Tue Feb 12 19:59:12 2019
    Bad Block Log : 512 entries available at offset 72 sectors
    Checksum : 405dfe1a - correct
    Events : 37509


    Layout : left-symmetric
    Chunk Size : 512K


    Device Role : Active device 0
    Array State : AAA ('A' == active, '.' == missing, 'R' == replacing)
    root@nas-omv:~# sudo mdadm --misc --examine /dev/sdb
    /dev/sdb:
    Magic : a92b4efc
    Version : 1.2
    Feature Map : 0x1
    Array UUID : 1fcf9705:1c13cd44:c9bcf8d4:5c66fb11
    Name : NAS-OMV:FirstRaid5
    Creation Time : Fri Nov 4 21:37:43 2016
    Raid Level : raid5
    Raid Devices : 3


    Avail Dev Size : 5860271024 (2794.39 GiB 3000.46 GB)
    Array Size : 5860270080 (5588.79 GiB 6000.92 GB)
    Used Dev Size : 5860270080 (2794.39 GiB 3000.46 GB)
    Data Offset : 262144 sectors
    Super Offset : 8 sectors
    Unused Space : before=262056 sectors, after=944 sectors
    State : clean
    Device UUID : 7254c079:85f1e316:d64fd1da:6fbb95b9


    Internal Bitmap : 8 sectors from superblock
    Update Time : Tue Feb 12 19:59:12 2019
    Bad Block Log : 512 entries available at offset 72 sectors
    Checksum : b335098d - correct
    Events : 37509


    Layout : left-symmetric
    Chunk Size : 512K


    Device Role : Active device 1
    Array State : AAA ('A' == active, '.' == missing, 'R' == replacing)
    root@nas-omv:~# sudo mdadm --misc --examine /dev/sdc
    mdadm: No md superblock detected on /dev/sdc.



    ################


    mdadm --create /dev/md0 --assume-clean --level=5 --verbose --raid-devices=3 missing /dev/sda /dev/sdb /dev/sdc


    ##############
    root@nas-omv:~# mdadm --create /dev/md0 --assume-clean --level=5 --verbose --raid-devices=3 missing /dev/sda /dev/sdb /dev/sdc
    mdadm: You have listed more devices (4) than are in the array(3)!


    ##############



    blkid


    ##############
    root@nas-omv:~# blkid
    /dev/sda: UUID="1fcf9705-1c13-cd44-c9bc-f8d45c66fb11" UUID_SUB="751ecf5b-6fd0-57ae-4fbf-95d7939777d0" LABEL="NAS-OMV:FirstRaid5" TYPE="linux_raid_member"
    /dev/sdb: UUID="1fcf9705-1c13-cd44-c9bc-f8d45c66fb11" UUID_SUB="7254c079-85f1-e316-d64f-d1da6fbb95b9" LABEL="NAS-OMV:FirstRaid5" TYPE="linux_raid_member"
    /dev/sdd1: UUID="b2f11b56-6610-4618-b4e8-037387c66fc4" TYPE="ext4" PARTUUID="afb2ec5b-01"
    /dev/sdd5: UUID="e13aacbb-179e-4577-a52e-16776674f174" TYPE="swap" PARTUUID="afb2ec5b-05"



    ##############


    fdisk -l | grep "Disk "


    ##############


    root@nas-omv:~# fdisk -l | grep "Disk "
    Disk /dev/sda: 2,7 TiB, 3000592982016 bytes, 5860533168 sectors
    Disk /dev/sdb: 2,7 TiB, 3000592982016 bytes, 5860533168 sectors
    Disk /dev/sdc: 2,7 TiB, 3000592982016 bytes, 5860533168 sectors
    Disk /dev/sdd: 14,9 GiB, 15977152512 bytes, 31205376 sectors
    Disk identifier: 0xafb2ec5b
    root@nas-omv:~#



    ##############


    mdadm --detail --scan --verbose


    ##############


    root@nas-omv:~# mdadm --detail --scan --verbose
    root@nas-omv:~#


    ##############



    So i need some advise to fix this issue.


    Thanks!!


    Christian

  • Hello here my outputs
    I can not install the original hard disk anymore. I had to send these back to WD.




    #################


    System:


    Linux nas-omv 4.19.0-0.bpo.1-amd64 #1 SMP Debian 4.19.12-1~bpo9+1 (2018-12-30) x86_64


    The programs included with the Debian GNU/Linux system are free software;
    the exact distribution terms for each program are described in the
    individual files in /usr/share/doc/*/copyright.


    Debian GNU/Linux comes with ABSOLUTELY NO WARRANTY, to the extent
    permitted by applicable law.
    Last login: Fri Feb 22 19:42:53 2019 from 20.20.2.6


    1. #################


    root@nas-omv:~# cat /proc/mdstat

    Code
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
    md127 : inactive sdb[1](S) sda[3](S)
    5860271024 blocks super 1.2
    unused devices: <none>







    2. #################


    root@nas-omv:~# blki

    Code
    /dev/sdb: UUID="1fcf9705-1c13-cd44-c9bc-f8d45c66fb11" UUID_SUB="7254c079-85f1-e316-d64f-d1da6fbb95b9" LABEL="NAS-OMV:FirstRaid5" TYPE="linux_raid_member"
    /dev/sda: UUID="1fcf9705-1c13-cd44-c9bc-f8d45c66fb11" UUID_SUB="751ecf5b-6fd0-57ae-4fbf-95d7939777d0" LABEL="NAS-OMV:FirstRaid5" TYPE="linux_raid_member"
    /dev/sdd1: UUID="b2f11b56-6610-4618-b4e8-037387c66fc4" TYPE="ext4" PARTUUID="afb2ec5b-01"
    /dev/sdd5: UUID="e13aacbb-179e-4577-a52e-16776674f174" TYPE="swap" PARTUUID="afb2ec5b-05"


    3. #################


    root@nas-omv:~# fdisk -l | grep "Disk "



    Code
    Disk /dev/sdc: 2,7 TiB, 3000592982016 bytes, 5860533168 sectors
    Disk /dev/sdb: 2,7 TiB, 3000592982016 bytes, 5860533168 sectors
    Disk /dev/sda: 2,7 TiB, 3000592982016 bytes, 5860533168 sectors
    Disk /dev/sdd: 14,9 GiB, 15977152512 bytes, 31205376 sectors
    Disk identifier: 0xafb2ec5b

    4. #################


    root@nas-omv:~# cat /etc/mdadm/mdadm.conf



    5. #################


    root@nas-omv:~# mdadm --detail --scan --verbose


    Code
    INACTIVE-ARRAY /dev/md127 num-devices=2 metadata=1.2 name=NAS-OMV:FirstRaid5 UUID=1fcf9705:1c13cd44:c9bcf8d4:5c66fb11
    devices=/dev/sda,/dev/sdb
    root@nas-omv:~#


    6. Used Drives: #################


    3x WD Red 3TB
    1x USB System Disk


    Disk /dev/sdc: ==> should be raid its the new one
    Disk /dev/sdb: raid
    Disk /dev/sda: raid
    Disk /dev/sdd: System disk


    7. That happend #################


    New hard drive installed due to SMART failure from the old one.



    Thanks

    • Offizieller Beitrag

    Your mdstat shows the raid as inactive with 2 drives /dev/sda and dev/sdb try the following to bring the raid back up with the two drives mdadm --assemble --verbose --force /dev/md127 /dev/sd[ab]


    That should bring the raid back up as clean/degraded, then in Storage -> Disks select the new drive and "wipe" then in the GUI again Raid Management click on your current Raid then select Recover from the menu. In dialogue box that appears your new drive should be displayed, select it, hit ok and you should see the Raid recovering.

  • Hello,


    thanks for your help.
    So i tried:
    the console output is.


    mdadm: looking for devices for /dev/md127
    mdadm: /dev/sda is busy - skipping
    mdadm: /dev/sdb is busy - skipping


    On GUI i do not see the raid.
    Take this time? Need a restart?


    Edit:


    Now its looking good.
    I had to do frist
    mdadm --stop /dev/md127




    Thanks

Jetzt mitmachen!

Sie haben noch kein Benutzerkonto auf unserer Seite? Registrieren Sie sich kostenlos und nehmen Sie an unserer Community teil!