Posts by godfuture

    I also have a lot USB hdd issues. I bought a Seagate Deskdrive which sometimes worked at the USB 2.0 port. But very likely failed on USB 3.0.


    But lately also on USB 2.0 I see issues. Then I shucked the drive and bought a Inoteck USB 3.0 SATA adapter. Same problem here.


    My personal issue is a bit that I dont know how to monitor or debug all these situations. I just see EOF errors while copying files via my docker container "cloud commander". And I see io errors in dmesg.


    Summary is...all these USB drives perfectly work on my windows machine, but rarely to never work on my Intel board running openmediavault/Debian with kernel 6.1 (...and also 5.19).


    Edit: I stopped plugging the external drives into the server to move data. I am using a notebook wired to local network and transfer the data via shared folders. I get better speed, no interruptions and no data corruptions.

    I have changed /etc/cron.d/mdadm to "57 0 * 2,5,8,11 0". Lets see if that works :)

    Here is an example using "depends on"


    And here is a link to some documentation.

    https://docs.docker.com/compose/startup-order/

    This made my day. It was an Odyssey with ups and downs. In the beginning, I tried exactly this: postponing the docker startup waiting for my crypt disks, but without success. It went up no matter my depends or wait-for-paths and so on. Then I started individual systemd service units. This worked, but automatic restarts didnt work, only start on boot. Now with your hint, everything works. Everything is started automatically with correct bind mounts in correct order. Happy! :)


    Buuut, still I am scared why systemd unit got broken. Do you have an idea what might have happened?


    Just for the record: I am using fluentd as a docker container for other container. This was an additional dependency to be managed in docker. Therefore I moved from fluentd in docker to deb. My thought is: every basic functionality should come from system, logging is one of it. Since then, I have to manage less dependency in every container. Hope this helps others to find a good architecture.

    And most users don't understand why they are using raid.... I don't use raid on my personal NAS at home but I do maintain hundreds of systems using raid at work. What does that tell you?

    Good question. I would answer it for the following way:

    Important data always needs backup. RAID is not a backup, but kind of risk management. It helps with two issues:

    • Backups are never up to date. So there will be always a chance of files going lost no matter what.
    • Restore from backup costs time.

    In case a disk dies and no raid is applied, both of the problems above kick in.


    But there is a third problem. Cost. The user decides if the cost of data loss is bigger than the cost for backup. In the area of less important data, raid can be a way to balance costs:

    • For less important data, raid is cheaper than full backup.

    So in short, raid balances cost and time.

    My docker setup always has been a bit more complicated, because I have been using luks encrypted drives for some docker container. Unfortunately these luks devices do not get mounted fast enough during boot and docker container start with empty bind mounts.


    My solution was to create systemd service files and add dependency for mount path. This worked, till I have upgraded to omv6.


    Now my systemd service files exist, but cannot be used anymore:

    Code
    Failed to start docker-xyz.service: Unit docker-xyz.service not found.


    I have used "systemd-analyze verify", "systemctl daemon-reload" and and and....but the error above still happens.


    Does somehow know how to tackle this issue? What might be a start?

    Hi,

    I have multiple software raids active. Normally this works really well with OMV, and the software raid webUI is working as expected. But I noticed that everytime the check is going on (I believe also rebuild) the webUI is not responding and I get a "gateway error".


    Are there others with this issue? Or is it because of my devices used or general setup?


    Thankful for every hint in the right direction.

    I tried a few things. It worked with grub commandline. In my case it was southern island, so I had to exchange cik with si:

    Code
    GRUB_CMDLINE_LINUX_DEFAULT="quiet splash radeon.si_support=0 amdgpu.si_support=1"

    Also installing radeontop helped to see clocks and load.


    Now I have less power draw and less heat. Many thanks!

    Hi,

    for my server I have switched from itx board to my old atx in desktop. More SATA, plenty of RAM, good CPU. I also took my old AMD gpu, because it already knows zero-core. I have tested it with a LiveLinux.


    But now built in my OMV case, the gpu reports weird errors in kernel, and runs with 100% load and 90% clock speed:

    Code
    media-server kernel: [2471815.300341] radeon 0000:04:00.0: GPU lockup (current fence id 0x00000000000001b6 last fence id 0x00000000000001b7 on ring 5)

    I have search the internet for open source drivers. But havent found anything that gives me instructions to feel safe. I really do not want to mess my server. From my search I have learnt that the driver is basically a kernel module.


    What module do I need for AMD HD7870 and should it be available for OMV6? How would I enable such a module?


    Many thanks for any help!

    Hi!

    I am trying to create my new data pool with mdadm. I had one filled 8 TB HDD with data. I bought 4 additional HDDs of the same type. I have created a RAID5 4-disk array, copied the data from first old HDD to 4-disk RAID5. After all completed, I added the old disk to RAID5 array with "grow".


    The reshape was running for days and today I needed to read some data. I found in the web that I could pause the sync with "echo frozen /sys/block/md2/md/sync_action" which worked, but once I had finished, I wrote "reshape" back to "/sys/block/md2/md/sync_action". After that it looked really weird. I have now a degraded raid5 array with 6 disks. I can not write "idle" or something into "/sys/block/md2/md/sync_action".


    What do you think had happened? What would you suggest to make my RAID to continue to my last reshape?


    Very happy for every hint here :)

    seems usrmerge is already installed and also not found in buster-backports.


    I am very scared with upgrades. I was happy to have the nice scripts from Dleidert but still there have been bugs which I only found due to testing ahead. So if I want to upgrade, I have to take fresh VDIs and test the whole thing.


    Unfortunately there is no upgrade script from dleidert for migration from 5 to 6.

    Hi!

    I mirrored my server disk to vdi files and started them in virtualbox to play around. I also do this to test upgrades. Now I wanted to test BTRFS raid5, because I need an additional pool.


    In OMV5 (debian buster) I had latest version 4.20.1 of btrfs-progs. This is very old from 2019. This version is lacking the "raid1c3" and "raid1c4" profiles.


    When I install btrfs-progs from backports, it says:

    Code
    mkfs.btrfs
    -bash: /usr/bin/mkfs.btrfs: Datei oder Verzeichnis nicht gefunden

    What I am doing wrong? Is there a way to upgrade to buster-backports version and get it working in buster?

    Maybe you can map the port of your fluentd container to the port on 127.0.0.1 and use localhost instead of the container hostname?

    So the port is not open to the public, but can be reached from the host.

    It works to open the ports on host itself and to refer localhost in fluentd address. This is also okay or even better, because I do not need to attach fluentd container in all networks.

    Now I just need to find out why collabora and nextcloud are not working anymore. I guess this is also related to changed dns behavior.

    Many thanks!

    OMV6 is close to stable. If your system matches your signature, you are still in OMV4. In your case I would consider updating the system. Perhaps this is the root of the problem.

    No. did not update the signature. I am on OMV5, luckily.

    do you use a custom ntwork like my-net>?.

    Yes. And it seems I can resolve those names inside the container. But still when starting, using those names for log driver does not work any longer.

    use this to all of your stacks:

    I think stacks are really useful. But as I am using a lot of encrypted drives, and I had issues with dependencies of crypt partitions used by docker, I have written my own systemd service files.


    I do not really know what logs to provide, as this is happening during the startup phase and the error message is quite clear. So I try to explain the setup a bit:
    My containers are configured with restartPolicy "never". On start, there are systemd services that correctly work together with crypttab dependencies. So systemd is starting the container in correct order. Those docker services with least dependencies are started first like fluentd or portainer. When fluentd is up, other containers are started like mariadb. When mariadb is up, nextcloud and firefox sync goes up. In the past, the logging driver host flag on every container was able to resolve the fluentd container with its name and returned its local ip in network.


    Well, today resolving only works in container. Maybe not trying to fix it, but to use a different approach. I do not want to start using fixed ips. Too much effort and if something changes, nothing works anymore. Does someone know how to achieve the same with help of docker dns?

    Hi guys, I have a very urgent issue.


    Some of my docker container stopped starting because of "no route to host" error. In the past, I have setup a container for fluentd as logging host and configured its container name as host of the logging driver flags in other containers. Basically it is this feature:
    https://stackoverflow.com/ques…ame-instead-of-ip-address


    This setup worked without any flaws since quite a time. Now restarting my server since long runtime, docker does not resolve the internal fluentd container ip anymore, but instead resolves my public ip.


    Do you know where and how I could start tackling this issue? Did something with docker internal dns change recently? I am not using docker-compose.


    Glad for any help!


    Update: now I additionally have different other issues. When opening office documents in Nextcloud with collabora, I get "unauthorized wopi host". Don't know yet how this all comes together...

    Hi guys,

    I have done it. I have upgraded from omv4 to omv5. I have prepared a lot for this. I have moved from nginx-plugin to swag docker. I created backups with clonezilla, restored in virtualbox. trained the upgrade. Reported my issues during upgrade scripts:

    https://github.com/dleidert/openmediavault-upgrade/issues/14

    https://github.com/dleidert/openmediavault-upgrade/issues/15

    https://github.com/dleidert/openmediavault-upgrade/issues/18

    https://github.com/dleidert/openmediavault-upgrade/issues/22


    So the upgrade went smooth. But even migration to swag was done successfully before upgrade, I get now an internal error in nextcloud. As said it worked before with the same config:

    Code
    Interner Serverfehler
    Der Server konnte die Anfrage nicht fertig stellen.
    Sollte dies erneut auftreten, senden Sie bitte die nachfolgenden technischen Einzelheiten  an Ihren Server-Administrator.
    Weitere Details können im Server-Protokoll gefunden werden.
    Technische Details
    Entfernte Adresse: 172.19.0.2
    Anfragekennung: mQoutHj9MFZrtANQDwZc


    The setup is the following: I had my dyndns domain where I initially hosted nextcloud as main application on main domain. With the migration to swag I wanted to move nextcloud to subdomain as all other applications. To be compatible with old clients, I wanted to redirect main domain to subdomain.maindomain.


    What is the best way to do this? I am doing this:


    in another file:


    Maybe this is not the problem at all...could really need your help. Dont want to go back as everything else is working just as it should.


    Update: issue with nc was solved, but help with nginx and the question above still welcome! :)