Posts by zerozenit

    I would suggest doing a short smart test on each of the drives and check 5, 187, 188, 197, 198 you're looking for any raw values in those drives, if there are any I would replace the drive, likewise if smart in omv's gui shows anything about bad sectors, replace the drive


    At the moment you're back up and running, as for backup start with what you don't want to lose

    Thank you geaves , this is the situation of /dev/sdb (the others are ok):

    Code
      5 Reallocated_Sector_Ct   0x0033   100   100   005    Pre-fail  Always       -       10
    197 Current_Pending_Sector  0x0022   100   100   000    Old_age   Always       -       24
    198 Offline_Uncorrectable   0x0008   100   100   000    Old_age   Offline      -       157

    Is it better to proceed immediately with the disk replacement? The file system is still unmounted. What would be the correct procedure for the replacement? Thanks

    Yep, hopefully is should be OK :) that's why one should have a backup

    Thank you geaves , backing up 24TB is a lot of work but I will have to do it. In the meantime the procedure is complete. :) What do you recommend to do now?

    Code
    ~# cat /proc/mdstat
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] 
    md0 : active raid10 sdd[5] sda[0] sdc[4] sdb[2]
          23437508608 blocks super 1.2 512K chunks 2 near-copies [4/4] [UUUU]
          bitmap: 0/175 pages [0KB], 65536KB chunk

    I will probably have to replace /dev/sdb asap which has some problems.

    Thank you geaves , this is the output:

    Now I think I understand that we can only wait and cross our fingers. :)

    Thank you

    TBH I've never really understood the error (possibly out of date) I've always assumed that some sort of error occurred but mdadm doesn't actually remove the drive from the array.


    Do one thing at a time, the output shows the array as active (auto-read-only) but with only 3 drives, from the cli run;


    mdadm --readwrite /dev/md0 hopefully that will correct the (auto-read-only) cat /proc/mdstat should confirm that

    Thannks geaves this is the output:

    Code
    ~# mdadm --readwrite /dev/md0
    ~# 
    ~# cat /proc/mdstat
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] 
    md0 : active raid10 sda[0] sdc[4] sdb[2]
          23437508608 blocks super 1.2 512K chunks 2 near-copies [4/3] [U_UU]
          bitmap: 0/175 pages [0KB], 65536KB chunk
    
    unused devices: <none>

    Thank you for your support.

    Why was the /dev/sdd disk not added?

    Thank you @geave, this is the output:

    I forgot the command...

    Hello there.


    Somehow all your disks are marked as "spare". That is why you see for example sdc[4](S) (S) means spare.

    You can read up on that. There is a solution. If you can't figure it out, come back again :)

    Thanks @bermuda , as you can see the Update Time and Events are the same only for 2 of the 4 disks and the Array State is AAAA only for one disk, for the others it is A.AA. So I don't know how to deal with a situation like this. So I rely on those who know more. Thanks

    Hi, I had some problems with RAID 10, I can't see the filesystem anymore. I shut down the server and put new SATA cables, and rebooted. Is there any hope of getting the RAID working again?


    Any help from you is appreciated.


    Thank you very much.



    Code
    ~# cat /proc/mdstat
    Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] 
    md0 : inactive sdc[4](S) sda[0](S) sdd[1](S) sdb[2](S)
          46875017216 blocks super 1.2
           
    unused devices: <none>
    Code
    ~# blkid
    /dev/sdb: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="6c9c5433-6838-c39f-abfa-7807205a3238" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sdc: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="28139ffa-50e0-60b6-54ee-d6af74613e6e" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sdd: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="a6bb8aa8-4e9b-7f90-b105-45a9301acbce" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sda: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="3904f2f1-fe1f-bde3-a965-d9dbe0074f66" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sde1: UUID="2218-DC43" BLOCK_SIZE="512" TYPE="vfat" PARTUUID="09f69470-ba7b-4b6b-9456-c09f4c6ad2ee"
    /dev/sde2: UUID="87bfca96-9bee-4725-ae79-d8d7893d5a49" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="3c45a8f0-3106-4ba8-89bc-b15d22e81144"
    /dev/sde3: PARTUUID="fda4b444-cf82-4ae8-b916-01b8244acee3"
    Code
    ~# mdadm --detail --scan --verbose
    INACTIVE-ARRAY /dev/md0 num-devices=4 metadata=1.2 name=pandora:Raid4x12TBWdRed UUID=8b767a7d:c52c068d:c04f1a3c:fd8d4c5f
       devices=/dev/sda,/dev/sdb,/dev/sdc,/dev/sdd

    :D  :D :D

    Um... um...

    I had connected a backup router to the switch instead of the main router following a fiber optic downtime and then, once the downtime was resolved, I forgot to swap them again. When connecting with wi-fi I didn't notice anything, while the devices connected to the switch could no longer work... It's my memory's fault, everything else is alive and well and works as usual. I apologize for the mix-up, thanks for your attention. :saint:

    Hello everyone,

    My server went offline and checking I saw that the network interface is no longer configured. I tried to reconfigure the network interface with omv-firstaid, but both with the manual IP configuration and with DHCP it is not possible to restore the functionality of the network interface, even if the configuration always completes successfully. The server remains offline, it does not see the other hosts in the LAN and is not visible to the other hosts present in the LAN. I changed various network cables, I changed ports on the switch and all other possible attempts, nothing to do. I thought that by configuring the IP manually I would have been able to at least ping the router, but nothing. What can I do to restore the functionality of the network? Thanks

    Everything is fine, but the email is odd; post the output of cat /etc/mdadm/mdadm.conf sometimes a reboot can help, it could be that mdadm has detected a missing drive but hasn't caught up with the change.

    Here below:

    That would be the correct procedure

    I performed the procedure and now the RAID is clean again and everything works. But I get this mail from mdadm:

    It is not clear if everything is ok and how to stop receiving these messages. Thank you.

    Hi everyone,

    after i had some problems with the cables to one of the 4 drives in Raid 10 (unable to open ATA device), i fixed the cable connections and the drive restarted and is visible again (/dev/sde). From the OMV interface the disk is now visible again in Storage -> Disks, and also in S.M.A.R.T., but in Software RAID it is not listed among the devices and the status is "clean, degraded". Do I have to swipe the disk and then recover or is it better to follow another procedure?

    Thank you.

    Code
    cat /proc/mdstat
    Personalities : [raid10] [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] 
    md0 : active raid10 sdb[1] sda[0] sdd[2]
          23437508608 blocks super 1.2 512K chunks 2 near-copies [4/3] [UUU_]
          bitmap: 79/175 pages [316KB], 65536KB chunk
    
    unused devices: <none>
    Code
    blkid
    /dev/sdb: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="a6bb8aa8-4e9b-7f90-b105-45a9301acbce" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sde: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="2c68c265-01f7-dd1b-ffff-6d28eb140780" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sdd: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="6c9c5433-6838-c39f-abfa-7807205a3238" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sda: UUID="8b767a7d-c52c-068d-c04f-1a3cfd8d4c5f" UUID_SUB="3904f2f1-fe1f-bde3-a965-d9dbe0074f66" LABEL="pandora:Raid4x12TBWdRed" TYPE="linux_raid_member"
    /dev/sdc1: UUID="2218-DC43" BLOCK_SIZE="512" TYPE="vfat" PARTUUID="09f69470-ba7b-4b6b-9456-c09f4c6ad2ee"
    /dev/sdc2: UUID="87bfca96-9bee-4725-ae79-d8d7893d5a49" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="3c45a8f0-3106-4ba8-89bc-b15d22e81144"
    /dev/md0: LABEL="REDRAID4X12" UUID="5fd65f52-b922-45e3-a940-eb7c75460446" BLOCK_SIZE="4096" TYPE="ext4"
    /dev/sdc3: PARTUUID="fda4b444-cf82-4ae8-b916-01b8244acee3"
    Code
    mdadm --detail --scan --verbose
    ARRAY /dev/md0 level=raid10 num-devices=4 metadata=1.2 name=pandora:Raid4x12TBWdRed UUID=8b767a7d:c52c068d:c04f1a3c:fd8d4c5f
       devices=/dev/sda,/dev/sdb,/dev/sdd

    In System -> Update Management -> Updates I performed the listed updates:

    openmediavault 6.4.3-1

    openmediavault-compose 6.7.15

    openmediavault-nut 6.0.5-1

    python3-distutils 3.9.2-1

    python3-docker 4.1.0-1.2

    python3-lib2to3 3.9.2-1

    python3-websocket 0.57.0-1


    During Updating workbench configuration files ... I got **CONNECTION LOST**. In the log this is the only error you can see Err:6 https://openmediavault-plugin-developers.github.io/packages/debian shaitan/main amd64 openmediavault-compose all 6.7.15 404 Not Found [IP: 185.199.109.153 443] (I cut some parts):

    I waited a minute and then applied the changes in Pending configuration changes, and got an error.


    Among the notifications I found these two logs:

    Code
    Failed to read from socket: Connection reset by peer
    
    OMV\Rpc\Exception: Failed to read from socket: Connection reset by peer in /usr/share/php/openmediavault/rpc/rpc.inc:172
    Stack trace:
    #0 /usr/share/php/openmediavault/rpc/proxy/json.inc(97): OMV\Rpc\Rpc::call()
    #1 /var/www/openmediavault/rpc.php(45): OMV\Rpc\Proxy\Json->handle()
    #2 {main}

    Of the second I report only the parts referring to the error:

    systemctl status nut-monitor.service

    journalctl -xe

    I understand that the problem is NUT, how can I fix it? Thank you

    So how did you do it without the yaml files to get all your configs back for each of your Docker applications? Or did you start from scratch?

    I solved it getting back all the application configurations. I try to summarize what I recommend to do in these cases:


    1- get a yaml file with autocompose with the command docker run --rm -v /var/run/docker.sock:/var/run/docker.sock red5d/docker-autocompose name_of_your_container


    2- stop the container with docker container stop name_of_your_container


    3- with a text editor change the name of the container contained in the yaml file, something like "container_name: name_of_your_container" becomes "container_name: name_of_your_container-new" (by changing the name of the new container we can keep the old one until everything is went right)


    4- in OMV Compose/File/Create name the new container "name_of_your_container-new" and paste the contents of the new yaml into window 1 (File), ignoring window 2 (Environment)


    5- at this point you need to save and then select the cointainer and give the Up command (up arrow)


    6- during this procedure there will be some errors, by reading them carefully you should understand where to put your hands to fix the file we have pasted. Select in Files "name_of_your_container-new" and click on Edit. Here we can make the changes. Generally it is about removing some parts that are not needed, I list some examples:

    removed:

    Code
    networks:
    - bridge

    removed:

    Code
    networks:
    - hosts

    removed:

    Code
    command:
    - 'dumb-init /etc/openvpn/start.sh'

    modified:

    Code
    devices:
    - CgroupPermissions: rwm
    PathInContainer: /dev/dri
    PathOnHost: /dev/dri

    in

    Code
    devices:
    - /dev/dri:/dev/dri


    7- once the errors have been resolved, the new container will start perfectly keeping all the settings it had previously, and at this point you can delete the old container with docker container rm name_of_your_container

    Ok, thanks, I succeeded. but I had to delete the following lines, otherwise I was getting an error:


    Code
    command:
    - 'dumb-init /etc/openvpn/start.sh'
    
    networks:
    - bridge

    I'll also try to move my lscr.io/linuxserver/plex in a while. Thank you.