RAID1 rebuild and DegradedArray events

  • Hello,
    Short story: I replaced a drive with SMART errors. The new drive is the same as the replaced. A 3TB WD red. And now I receive DegradedArray events.


    Long story:
    I degraded the array before I removed the bad drive. I have done this before, to check the drive in a other PC. The recover worked fine every time (when I put back the SAME drive).


    This time I build in a new drive. Same type, same sice, same vendor. I simply clicked Recover in the RAID Management, selected the new drive and the system started to synchronise the data. But now I receive degraded array events:

    And the detail view in the RAID Management shows me the new drive as a spare drive instead a mirror.


    I'm not sure, what to do know. Can I just degrade the array and try to recover it again?
    I suspect there could be a problem in a config file or something like this.
    Maybe it is importend to mention, that also the first drive (the drive with the data) start to throw SMART errors, too. I planned to recover the array and replace also the first drive before I run into this problems.
    And there is a third drive in the PC. sda conatins only the OS and is not part of the Array.


    Here a are the infos, mentioned in the pinned Thread above.


    Code
    root@pluto:~# cat /proc/mdstat
    Personalities : [raid1]
    md127 : active raid1 sdc[3](S) sdb[2]
          2930265424 blocks super 1.2 [2/1] [_U]
    
    
    unused devices: <none>
    Code
    root@pluto:~# blkid
    /dev/sda1: UUID="0c8aa7c5-8de0-4a86-9b04-ec24d889a206" TYPE="ext4"
    /dev/sda5: UUID="b875d1d9-898c-43bc-8ac3-88db0fa4925a" TYPE="swap"
    /dev/md127: LABEL="data" UUID="48f49465-8540-4dd3-b579-a33e0d7f406a" TYPE="ext3" SEC_TYPE="ext2"
    /dev/sdb: UUID="e282c2db-a604-2f4e-af0b-b2767652b009" UUID_SUB="7f0e24fc-87df-a794-fd30-accc7d2bb34b" LABEL="pluto:DATA" TYPE="linux_raid_member"
    /dev/sdc: UUID="e282c2db-a604-2f4e-af0b-b2767652b009" UUID_SUB="79d1c966-1a26-06ad-469e-3c9ce3126d38" LABEL="pluto:DATA" TYPE="linux_raid_member"
    Code
    root@pluto:~# mdadm --detail --scan --verbose
    ARRAY /dev/md/DATA level=raid1 num-devices=2 metadata=1.2 spares=1 name=pluto:DATA UUID=e282c2db:a6042f4e:af0bb276:7652b009
       devices=/dev/sdb,/dev/sdc
  • anyone?


    I tried to set the spare count in the mdadm.conf to 0. The drive synced after this but is still listed as a spare drive.


    I also tried the following comand. The drives synchronised again but /dev/sdc is still listed as a spare.

    Code
    mdadm --assemble --force --no-degraded /dev/md127 /dev/sdc /dev/sdb

    After this I found this article with exact the same problem as mine. I have also bad sectors on /dev/sdb. And maybe this is why, the sync fails. But how can I rescue my single raid device? I'm ok with loosing some bits or bytes. Most of the data are backups and in worst case I lose some jpgs or music files. But how can I make the spare drive become active?

    • Offizieller Beitrag

    Start it as a degraded array by only specifying one drive.

    omv 7.0.4-2 sandworm | 64 bit | 6.5 proxmox kernel

    plugins :: omvextrasorg 7.0 | kvm 7.0.10 | compose 7.1.2 | k8s 7.0-6 | cputemp 7.0 | mergerfs 7.0.3


    omv-extras.org plugins source code and issue tracker - github


    Please try ctrl-shift-R and read this before posting a question.

    Please put your OMV system details in your signature.
    Please don't PM for support... Too many PMs!

  • @ryecoaaron how do I do this? I tried the remove my /dev/sdb and hoped OMV will switch to the spare drive automaticly. But it didn't.


    I looked through the syslog and found this lines:

    Code
    Aug 17 13:31:59 pluto kernel: [11918.526340] sd 5:0:0:0: [sdb]  Add. Sense: Unrecovered read error - auto reallocate failed
    Aug 17 13:31:59 pluto kernel: [11918.526345] sd 5:0:0:0: [sdb] CDB: Read(10): 28 00 ae a8 3c d0 00 00 08 00
    Aug 17 13:31:59 pluto kernel: [11918.526353] end_request: I/O error, dev sdb, sector 2930261200
    Aug 17 13:31:59 pluto kernel: [11918.528186] ata5: EH complete
    Aug 17 13:31:59 pluto kernel: [11918.528196] md/raid1:md127: sdb: unrecoverable I/O read error for block 2930259072
    Aug 17 13:32:01 pluto kernel: [11919.586530] md: md127: recovery done.

    Is there a way to fix the bad sector on /dev/sdb or mark it as unreadable?

    • Offizieller Beitrag

    mdadm --stop /dev/md127
    mdadm --assemble --force /dev/md127 /dev/sdc

    omv 7.0.4-2 sandworm | 64 bit | 6.5 proxmox kernel

    plugins :: omvextrasorg 7.0 | kvm 7.0.10 | compose 7.1.2 | k8s 7.0-6 | cputemp 7.0 | mergerfs 7.0.3


    omv-extras.org plugins source code and issue tracker - github


    Please try ctrl-shift-R and read this before posting a question.

    Please put your OMV system details in your signature.
    Please don't PM for support... Too many PMs!

  • I stopped the array and assembled it as you said. But I got this message:
    mdadm: /dev/md127 assembled from 0 drives and 1 rebuilding - not enough to start the array.

    • Offizieller Beitrag

    What is the output of cat /proc/mdstat. You might have an array rebuilding now.

    omv 7.0.4-2 sandworm | 64 bit | 6.5 proxmox kernel

    plugins :: omvextrasorg 7.0 | kvm 7.0.10 | compose 7.1.2 | k8s 7.0-6 | cputemp 7.0 | mergerfs 7.0.3


    omv-extras.org plugins source code and issue tracker - github


    Please try ctrl-shift-R and read this before posting a question.

    Please put your OMV system details in your signature.
    Please don't PM for support... Too many PMs!

  • It's still inactive.

    Code
    root@pluto:~# cat /proc/mdstat
    Personalities : [raid1]
    md127 : inactive sdc[3](S)
          2930265560 blocks super 1.2
    
    
    unused devices: <none>
    root@pluto:~#

Jetzt mitmachen!

Sie haben noch kein Benutzerkonto auf unserer Seite? Registrieren Sie sich kostenlos und nehmen Sie an unserer Community teil!