SnapRAID and Union Filesystem No space left on device (28)

  • Hello,
    I finally got my OMV server set up and running.
    A copy of my current filesystem is shown below. I had set up a Union Filesystem with sda1, sdc1,sdf1 called THEKEEP. The details about the drives in the SnapRaid server are also posted below. Other than OMV the only other install I did on this server was the NextCloudPi docker. It keeps its data in a separate drive sde1.


    I ran a SnapRAID sync when I first created the SnapRAID config while the disks were empty. I did t run another SnapRAID sync. (I was going to move all the data first.)


    I had issued an rsync command via ssh from my desktop to copy files to the server from my USB backup drive:
    rsync -av /media/daddy/'Seagate Backup Plus Drive'/THEKEEP/ daddy@192.168.0.7:/srv/1c64b997-f9ee-4570-8357-96d8063f8050/THEKEEP/



    This morning I noticed many errors in the ssh terminal on my desktop running the rsync: xxxxxx No space left on device (28)


    Looking at the Filesystem it appears the UnionFileSystemDrive is full with only 2TB of data even though it has a capacity of 9TB.




    What did I do wrong? Any troubleshooting tips? I am totally confused here.


    Looking at other posts I ran the following du below. I also tried to boot from the GPARTED live and used the shell to try to look at the drives by when I go to /dev/disk/by-label/ I can only see the labels of the physical drives and not look into them. GPARTED did not show any drives full.


    root@SHADOW:~# du -hx --max-depth=1 /
    4.0K /export
    1.3G /var
    7.6M /etc
    4.0K /home
    4.0K /mnt
    8.0K /media
    4.0K /sharedfolders
    168K /root
    1.7G /usr
    16K /opt
    1.7M /srv
    16K /lost+found
    1.4G /boot
    4.3G /






    Screenshot from 2020-03-06 07-39-35.png
    Screenshot from 2020-03-06 07-45-23.png


    Screenshot from 2020-03-06 07-53-42.png

  • Administrators please look into this...

    You are the administrator of your system


    OP, because your pool is using btrfs, your system might be out of space because snapshots are filling the system? Hard to say from the info provided but this definitely isn't caused by OMV or the unionfilesystem/snapraid plugins.

    omv 5.5.17-2 usul | 64 bit | 5.4 proxmox kernel | omvextrasorg 5.4.2
    omv-extras.org plugins source code and issue tracker - github


    Please read this before posting a question.
    Please don't PM for support... Too many PMs!

  • root@nwvault:~# df -i

    Filesystem Inodes IUsed IFree IUse% Mounted on

    udev 2046509 497 2046012 1% /dev

    tmpfs 2051518 777 2050741 1% /run

    /dev/sda1 14221312 133126 14088186 1% /

    tmpfs 2051518 1 2051517 1% /dev/shm

    tmpfs 2051518 3 2051515 1% /run/lock

    tmpfs 2051518 16 2051502 1% /sys/fs/cgroup

    tmpfs 2051518 11 2051507 1% /tmp

    label-Archives_Four:label-Archives_One:id-ata-TOSHIBA_HDWD130_X7V0PPGAS:label-Ar chive_Five 732594176 345206 732248970 1% /srv/0808f8e1-f56e-4184-b872-19fd729 aec5d

    /dev/sdc1 183148544 14 183148530 1% /srv/dev-disk-by-label-Archives_Two

    /dev/sdf 183148544 14 183148530 1% /srv/dev-disk-by-label-Archives_Thre e

    /dev/sdd1 183148544 31798 183116746 1% /sharedfolders/snapfour

    /dev/sde 183148544 5680 183142864 1% /sharedfolders/snaptwo

    /dev/sdb1 183148544 271065 182877479 1% /sharedfolders/snapone

    /dev/sdg 183148544 36663 183111881 1% /srv/dev-disk-by-label-Archive_Five



    does this mean anything?

  • I meant admin for omv, sorry

    As I said, this isn't an OMV problem.

    Any suggestions for how to free up space?

    Are you using the exact same setup as the OP? We have no details of your system.

    omv 5.5.17-2 usul | 64 bit | 5.4 proxmox kernel | omvextrasorg 5.4.2
    omv-extras.org plugins source code and issue tracker - github


    Please read this before posting a question.
    Please don't PM for support... Too many PMs!

  • Just an update...i had 133 gb of log files (syslog &daemon.log)...deleted older ones, was then able to use first-aid, running "apt-clean" "clear wbgui cache", then able to open admin web gui, then i disable clamd "on access scans" for all drives, which stopped continuous erros in both log files... Now back to normal operation...

Participate now!

Don’t have an account yet? Register yourself now and be a part of our community!