multiple BTRFS disks randomly turn to read only mode

  • Hi,


    I am currently testing OMV for my new Aoostar WTR MAX NAS build.

    Before throwing any ciritical data on it, I wanted to test some features and disk-configurations.

    (OMV 8.0.6-1, only OMV-Extras with MergerFS and Snapraid added)


    I currently have 6x4TB WD RED drives installed, 3 old, 2 newer 1 almost unused drive.

    I formated all drives with BTRFS in single mode and created a mergerfs pool with 5 folders, MergerFS1-5 on each drive.

    So far, everything looked good, disks fill up according to policy, SMART is reporting fine on all drives.

    I already tested the 32GB RAM I have installed, no errors.


    At some point 2, 3 or even 4 disks at the same time lock up and get put into read only mode.

    No error in the GUI, no information why this happens, nothing.

    I am pretty lost what to look for here...

    Almost all searches for this behaviour tells the user: BTRFS is the superior filesystem, your drive is broken, no other file system could detect this previously!

    User: OK thanks, I will buy a new drive tomorrow.


    Sorry, but I don't think that 4 drives of different age fail at the same time, same second even...

    I also reinstalled OMV yesterday, reconfigured the drive pool, same result. 4 out of 6 disks went into read only mode in the night during file copy...
    Is there some kind of BTRFS service running in the background that is not capable of working correctly if the drives are constantly in use?


    The result this morning was a lot of errors and the single usable disk in mergerfs getting filled to the last remaining GB available.

    (the 6th drive is currently filled with Snapraid parity data from a previous test that failed, because the drives locked the content files, Snapraid is currently not active)

    In the log I could find some errors, mainly for sda, ironically the drive that was still functioning, to my eyes it looks like that the "bad tree block" error caused multiple disks to lock up at the same time. But I have no clue why... ;(


    The file copy is done with the 10G fiber connection, copying from a 4x10TB RAID that reads with up to 600MB/s, so all drives are under constand writing load with network not being a bottleneck. The system basically is always waiting for the 4TB drives to finish writing.

  • macom

    Approved the thread.
  • govido

    Added the Label OMV 8.x
  • Just a quick update...

    Changing all 6x 4TB with 4x 10TB gave me the exact same issues.

    After about 1 hour of copying data to a fresh formatted disk it goes into read only mode. Same error...

    First copy to drive 1, locked up, restarted OMV, copied the rest.

    Copy to drive 2, about 1 hour later copy failure because of read only, restart OMV.


    any ideas?


    edit:

    Now disk 3 locked up after copying for 1-2 hours, so there is a very specific pattern here.


  • try to identify chipset and google about posible problem with Debian.


    try promox kernel, perhaps you have luck and other kernel have better drivers .


    PD. For me sound like a chipset/driver problem


    PD2: revise temp on disk on that extensive copy of data if > 65ºC perhaps can explain the problem.

  • I tried many things...
    - Reformatting the drives with BTRFS again to get a "clean" state, didn't help

    - Updating the ASM1166 firmware for the SATA controller, no change

    - Reformat the drives with XFS, gave me no errors in the log, but file verify/read failed after about the same time

    - XFS showed no errors in the log, but I discovered a kernel AHCI error

    - Reinstalled OMV entirely with BTRFS again, same errors again on the clean install

    - Changed the kernel to the older version from 6.17 to 6.12


    So far not a single copy job failed (running 10 at the same time to stress test) after about 2 hours. It looks lke 6.12 is playing nicer with the controller and file system.
    Are there known issues with kernel 6.17 and the ASM1166?

    PD. For me sound like a chipset/driver problem

    I guess this is the case, thank you for pointing me in the right direction :) looks like newer kernel is not always better and LTS is there for a reason :D

  • govido

    Added the Label resolved
  • I have a similar situation. I use HM SMR ZONED disks and rebuilt the 6.17 kernel and btrfs-progs with experimental flags (only for RAID1). I thought the problem was with the disks. I replaced BTRFS with F2FS and the stability issues disappeared. My setup is only for tests.

    • New
    • Official Post

    This issue seems like a bug in the ASM1166 driver. It is very possible it was introduced after 6.12. I'm not optimistic it was fixed in 6.18 but could always try. Otherwise, i would stick with the 6.12 kernel or get a different sata controller.

    omv 8.1.1-1 synchrony | 6.17 proxmox kernel

    plugins :: omvextrasorg 8.0.2 | kvm 8.0.7 | compose 8.1.5 | cterm 8.0 | borgbackup 8.1.7 | cputemp 8.0 | mergerfs 8.0 | scripts 8.0.1 | writecache 8.1.1


    omv-extras.org plugins source code and issue tracker - github - changelogs


    Please try ctrl-shift-R and read this before posting a question.

    Please put your OMV system details in your signature.
    Please don't PM for support... Too many PMs!

  • I have another sata hba, not ASM1166, I tried AMD A58 SATA HBA, Adaptec 71685 in HBA SPAN mode and INTEL (1155 generation) pch SATA host - no changes with btrfs.


    6.12 is working fine, 6.17 and 6.17 fails to ro in 6-8 hours.


    Hardware is good. I tried F2FS and BTRFS and found no problems with F2FS. But I have 14TB drives and 20TB drives too, F2FS does not like this :(

    • New
    • Official Post

    i have two single btrfs drives (16 and 18 TB) in my server and I have run the 6.17 kernel for a long time. No issues. So, I can't think of what else it would be. f2fs isn't a CoW filesystem. Have you run memtest?

    omv 8.1.1-1 synchrony | 6.17 proxmox kernel

    plugins :: omvextrasorg 8.0.2 | kvm 8.0.7 | compose 8.1.5 | cterm 8.0 | borgbackup 8.1.7 | cputemp 8.0 | mergerfs 8.0 | scripts 8.0.1 | writecache 8.1.1


    omv-extras.org plugins source code and issue tracker - github - changelogs


    Please try ctrl-shift-R and read this before posting a question.

    Please put your OMV system details in your signature.
    Please don't PM for support... Too many PMs!

  • I have no problems with ram. 6.17 was released in late september 2025. Are you have it about 4-8 weeks? There is no problems with no "heavy loads" in my case (12TB rsync), but i can not make a copy of existing disk on 6.17 or 6.18. Only on 6.12.

    Also, i use debian kernel and stock omv, and you may be use kernel from proxmox. Also i have no quota, compression and other. Only local rsync and smb share on a "flat" single btrfs drive.


    I think, main reasonof problems - large folios

    • New
    • Official Post

    I have no problems with ram

    Remember I don't have your system sitting in front of me. So, all I can do is guess.


    6.17 was released in late september 2025. Are you have it about 4-8 weeks?

    Is your point that that is not a long time? Numerous weeks on a heavily used system is a long time in my Linux admin world.

    Also, i use debian kernel and stock omv, and you may be use kernel from proxmox

    I doubt that the 6.17 kernel from Debian and 6.17 kernel from Proxmox have many or any btrfs code differences.

    Also i have no quota, compression and other. Only local rsync and smb share on a "flat" single btrfs drive.

    I am not using quotas or compression either. I am using rsync and borgbackup on a flat single btrfs drive. Just doing the same backup to two drives.

    I think, main reasonof problems - large folios

    Submit a bug to Debian then.

    omv 8.1.1-1 synchrony | 6.17 proxmox kernel

    plugins :: omvextrasorg 8.0.2 | kvm 8.0.7 | compose 8.1.5 | cterm 8.0 | borgbackup 8.1.7 | cputemp 8.0 | mergerfs 8.0 | scripts 8.0.1 | writecache 8.1.1


    omv-extras.org plugins source code and issue tracker - github - changelogs


    Please try ctrl-shift-R and read this before posting a question.

    Please put your OMV system details in your signature.
    Please don't PM for support... Too many PMs!

Participate now!

Don’t have an account yet? Register yourself now and be a part of our community!