Posts by kattivius

    Ok, I’ve read it.

    From the documentation, omv-regen only works within the same major OMV release, as stated in its notes:

    Quote

    From "WHAT IS OMV-REGEN"
    NOTE: omv-regen does not support upgrading between major OMV versions (e.g., from OMV 6 to OMV 7). For this, always use the official omv-release-upgrade procedure. omv-regen can only regenerate configurations within the same major OMV version.

    I assume you answered my concern of re-configuring the all system after a re-installation with the omv-regen option.

    When I mentioned a possible re-installation, I was following jgyprime suggesting, and to do so directly to the latest version (OMV8). I think OMV-regen cant be use to regenerate the configuration onto a following major release.

    does your NAS have the possibility to connect a display (HDMI, DP, VGA, whatever) and keyboard (USB)?

    if yes, connect a monitor and login as root user.

    Hi

    it does.
    Once i tried to connect one of my monitors and a keyboard (after it hanged) but I had no signal at that point.
    I dont have a monitor to keep on the nas at all time.
    But I can try to see again if after hanging a monitor shows the prompt and so send the commands you suggested.

    Re-install a fresh OMV is rather annoying.

    I will keep that option as a very last option. I would have to recreate the entire configuration and shares and compose files .... with time, the configuration gets complex to remember and recreate


    I assume that all my OMV daily backup I do will not be useful in a new major release, right?

    Lets see. at the moment is up for at least 4h. Last hang was maybe 1h after reboot.

    I hope is just the UI that is not accessible, how ever, since also SSH is not accessible, is probably more than that.


    Lets see if tomorrow I have connection!


    What can I send on CLI to see if there are some ghost shares still looking for the missing disks?

    you would be better off using mergerfs to pool the drives

    Thanks for the suggestion. I certainly keep it in consideration as I still have to create a pool for my movies and tv-shows. As mentioned, this is not critical mission data. Just dont want to use my internal disks (2 as I only have 2 bays).
    How did I delete the RAID:
    After reading the forum, I figure the only way to do it is to find all the reference to the drive (all shares and other folders pointing to it, even the docker volumes). Once the raid0 was no longer marked 'referenced', I was able to deleted the RAID0. I did this via the GUI.
    I checked a lot of logs and for a while I had an issue with one NFS share and an SMB shareand rsync. I found them and stopped/delete them.
    Once RAID0 was deleted, I shootdown and disconnected all related USB drives.

    At the moment I have 1 RAID1 with 2 nvme that I use for omvbackup and some compose files for fast access.
    And 1 large BTRFS RAID1 (x2 16TB)

    What is the output of cat /proc/mdstat


    Code
    cat /proc/mdstat
    Personalities : [raid1] [raid0] [raid6] [raid5] [raid4] [raid10] 
    md0 : active raid1 nvme1n1[1] nvme0n1[0]
          488254464 blocks super 1.2 [2/2] [UU]
          bitmap: 4/4 pages [16KB], 65536KB chunk

    What is the output of mdadm --detail --scan

    Code
    mdadm --detail --scan
    ARRAY /dev/md0 metadata=1.2 name=ugreen-nas:0 UUID=my_UUID

    On the BTRFS RAID1 /dev/sda and the RAID1 on EXT4 /dev/md0, there are no errors reported whem looking at the details and reports.

    If you need some logs, let me know.

    And thanks for taking the time to help me.

    Hi,

    a couple of days ago I deleted a RAID0 (was just a bunch of disks for my movies and TVshows. no critical data)
    Its was a complex thing as I had to find back all the referenced shares. Also stopped docker with volumes referring to those disks where a problem.
    Eventually I found them all and I was finally able to delete the RAID0.

    Took all disk offline and disconnected them (3 USB external disks)

    Ever since, my OMV is completely unstable.

    keeps stop responding every few hours. forcing me to unplug the power and plug it back to be able to access my nas.
    As I write to you it happened again. and I was checking the activities and there was almost no CPU utilization.

    I checked few logs but I am no expert and so far nothing sounded the alarm bell for me.
    I would need some help how to figure out what is the issue.

    And, since I saw the OMV8 is released, if this update could solve what ever is the issue now. or best is to first sove the problem before upgrading?

    To know:
    I took deleted the raid0 because one disk failed.
    That caused me issues so I decided to replace the disk. Meaning deleting the raid0, change the disk and recreate the raid0 with good disk (I know I lost all the media, but I backed that up so I can restore that too)
    I never had so many hangs before deleting the raid0.


    Where can I look what is the issue?
    Thank you for helping on this.

    NAS details:
    Ugreen DXP2800


    Version
    7.7.24-7 (Sandworm)

    Processor
    Intel(R) N100

    Kernel
    Linux 6.12.57+deb12-amd64


    thank you

    Hi,

    The hardware:
    I have several NFS shares configured. Those shares are in several disks.

    My NAS is a Ugreen hardware NAS where I installed OMV7 on it.

    Its a 2 disks bay.
    The configuration:

    I have the normal RAID1 on my 2 16TB disks for data, docker and a few more things.

    1 USB-C 5TB attached that I use to backup teh most important data from the raid1

    I have 3 disks on RAID0 just for movies and TV-Shows. (I am aware of the danger of RAID0. but that is ok as anything in those disks can be replaced. Is just a buch of disks I had around and wanted to use).

    The situation:

    eventually, one of those 3 disks on RAID0 is failing and creating serious issues. for stability reasons I had to shutdown the NAS, take the 3 disks off the hook.

    When I restart the NAS, non of my NFS share work anymore. Not just (obviously) those on the missing disks, but also those on my RAID1.

    I dont understand why when some NFS shares do not exist anymore, the NFS service stop working.
    The dashboard shows NFS service in red.


    2nd problem:

    It seems it is not possible to just disable an NFS share. It seems the only way to move forward in that direction is delete the NFS share.


    I also cant delete the RAID0 (when active or inactive) so to create a whole new RAID0 with good disks.


    So, my questions:

    anyway I can disable the NFS shares rather than delete them? If yes, how? (I cant find the way).

    How can I destroy my RAID0? is the raid0 not able to destroy till is referenced? Do I have to delete all the NFS shares and SMB if any before able to delete the RAID0?


    I would appreciate some guidance here.

    Thank you.

    Hi jata1,

    eventually anything I do or see point to a high CPU utilization that in turn hangs my NAS.
    I tried everything suggested anywhere I looked on the web.
    I even got to the point that rebooting the NAS, Immich restarted without I realized. I notice my services did not work and sure enough I found my NAS unresponsive and had to hard boot it.

    I completely removed the Immich docker. Never an issue ever since.

    I will try to a different machine with UBUNTU server installed. I also going to try it on a docker there. If that gives the same bad results, I guess eithet Immich in docker is not a good idea or Immich requires its own machine and a lot of resources for itself.
    I like Immich and I really want to use it. I hope will work out.

    Thanks for your help

    nice!

    I will install that

    it seems to be immich related:

    I issued: journalctl -b -1 -e

    among the long list I found this:

    Code
    Oct 17 20:12:55 ... loadavg (5min) of 11.6
    Oct 17 20:13:25 ... loadavg (5min) of 10.8
    Oct 17 20:13:55 ... loadavg (5min) of 10.1

    a load of 10–11 means the CPU is severely overloaded.

    there was also a few cron jobs running (OMV jobs. ) probably the all thing together started to be heavy for a N100 processor.

    The thing is that as long as Immich runs, as soon as I reboot my CPU goes sky high. I trace it all back to Immich machine learning.

    there was a PID python at 280% CPU only when Immich was running.
    So I checked the PID:

    ps -fp 3165

    python -m gunicorn immich_ml.main:app -k immich_ml.config.CustomUvicornWorker ...

    Immich’s machine learning service (immich_ml) running under Gunicorn.

    Immich i’s doing image/video analysis or thumbnail generation, which is CPU-intensive, especially on a low-power N100 CPU... using 280% CPU usage.


    This was already a new Immich clean docker install.
    But yes, I can try to start that clean again and see.


    Thanks for helping. really appreciated.

    Code
         # Prevent CPU overload 
        deploy:
          resources:
            limits:
              cpus: '1.0'


    this is how I have it in my compose file. I guess I did not know how to report it correctly in the other post.

    Since it hanged again, I stopped the container.
    I really don't want to keep pulling the plug.
    I have a RAID0 with a bunch of left over disks that host my TV shows and moves (not mission critical) that since the few hard unplug give me errors (BTRFS). Now scrubbing to make sure what are the errors and will try to fix it.
    It would be much worst if my main 2 Disks on RAID1 would fail.
    Somehow I dont feel safe with Immich on my NAS anymore.

    I am thinking if is better to run it on a separate hardware and point the libraries to the nas... to keep the media 'safe' and have the hard work with possible failure to a different HW.

    What do you think?

    Hi Jata1,

    I did give the 1 core cpu limitation.


    Code
    deploy:
    resources:
    limits:
    cpus: '1.0'

    I must say not so extensive as you mentioned

    I like to share my experience with Immich here.

    Since my original post, I tried a lot of things:

    I start to give some details of my HW first.


    Code
    Model: Intel Alder Lake-N UHD Graphics
    Type: Integrated graphics (built into your CPU)
    Architecture: Modern Intel 12th/13th generation
    PCI Address: 00:02.0
    Hardware Acceleration: Excellent for video transcoding
    AV1 hardware decoding
    VP9

    Not the most powerful.. but is a decent nas!

    Since it is just a N100 CPU and my GPU is not a lot but is there, I managed to pass the instructions in the compose file to use the GPU

    This is the section I added:



    The compose file is checked and fine. Run the compose file and seen little change (a little, yes).

    But my CPU is constantly at 100% with the Python process skyrocketing to 200+%


    So I figure i limit the CPU core utilization and added:

    Code
    # Prevent CPU overload (optional but strongly recommended)
    deploy:
    resources:
    limits:
    cpus: '1.0'


    Sure enough, I went to bed, and this morning I noticed my nas not responding again.

    All I can say is that I figure out the issue but I cant find a solution.

    This way my nas is not usable and unless I find a solution, Immich cant be used.

    It seems Immich need a lot of GPU and/or vRAM and a solid CPU.

    not really a common nas I guess.

    If you have any suggestion, please let me know.

    I am about to delete immish compose file! And for now I keep it down.

    Troubleshooting:

    Code
    journalctl -b -1 -e
    Oct 17 20:12:55 ... loadavg (5min) of 11.6
    Oct 17 20:13:25 ... loadavg (5min) of 10.8
    Oct 17 20:13:55 ... loadavg (5min) of 10.1
    ...


    On a 4-core N100, a load of 10–11 means the CPU is severely overloaded.

    This explains why the system hangs. Tasks are queuing and the kernel struggles to schedule them

    PID 3165, python, 280% CPU → on a 4-core CPU, that basically means 3 cores are fully saturated by this single Python process.

    That’s huge and will cause loadavg spikes, system lag, and hangs.

    The memory usage is moderate (563 MiB), so RAM isn’t the bottleneck, it’s CPU-bound.


    the process that cause the most issue is:


    Code
    ps -fp 3165
    UID PID PPID C STIME TTY TIME CMD
    root 3165 3128 99 22:17 ? 00:18:17 python -m gunicorn immich_ml.main:app -k immich_ml.config.CustomUvicornWorker -c /usr/src/immich_ml/gunicorn_conf.py -b [::]:3003 -w 1 -t 300 --log-config-json /us

    Immich’s machine learning service (immich_ml) running under Gunicorn. The image/video analysis or thumbnail generation.


    I wonder if I can safely desable Machine Learning and have Immich working correctly anyway... I am conerned about not creating image/video analysis or thumbnail generation.

    Well.. I figure is a lot more complicated than that! or anyway different.
    I just checked the immich page for GPU configuration:
    https://docs.immich.app/features/ml-hardware-acceleration/

    I cant get the compose file to run. Has some errors I cant figure out how to fix!

    This require a lot more research and try.
    Mean while my OMV hanged again due to too long high CPU utilization.
    Immich seems to want too much resources.
    I have to decide what to do with it!
    If anyone here used Immich in docker with HW accelleration would be great to know how they modified the yamal.

    Yes, I think thats it!
    I moved years of photos and its taking hours and hours of scan!
    I read I can offload the CPU by using the GPU.
    My Ugreen has some GPU:

    Code
    ls /dev/dri
    by-path card0 renderD128


    I read this means your Intel iGPU (Alder Lake-N UHD Graphics) is recognized by the Linux kernel and fully exposed to user space, the device nodes /dev/dri/card0 and /dev/dri/renderD128 are exactly what Immich needs for hardware acceleration.

    Now the only thing missing is the compute runtime (so that OpenVINO can actually use the GPU).
    The way suggested that should work for OMV is to install intel-opencl-icd intel-level-zero-gpu level-zero

    Code
    sudo apt update
    sudo apt install -y intel-opencl-icd intel-level-zero-gpu level-zero

    Do you think is safe to do so?
    I did search in the plugin list (within OMV) but I cant find anything for Intel.

    Hi.
    I cant be sure, I dont know how to be sure but the pattern seems an hint:

    Code
    Service:     ugreen-nas
    Event:       Resource limit succeeded
    Description: cpu user usage check succeeded [current cpu user usage = 94.2%]

    when I see a few of those messages one after each others, the system becomes inresponsive.
    The emails start arriving at around 23h and the last email (I guess before hanging) was at 00:18.
    I cant SSH and cant access the GUI.
    The last time happened is during Immich docker scanning one of the shares.
    I can see the CPU at 100 %

    Other times I have the "resources limit reached " when the OMV system backup" routine starts or in combination with my linux machine backup to the omv.
    I do get those "resources limit reached " a lot every day
    I think my HW is fairly solid but this is starting to concern me.

    Is there a way to avoid OMV to hang and maybe rather only having the service stopping than the machine hanging?

    Quote

    The HW: Ugreen NASybc DXP2800
    16GB Ram

    Note that when the machine hangs, the power button is useless. I have to pull the cable.

    So, I think is clear the root cause is the CPU utilization.
    Is it normal that backups and dockers (not at the same time) can cause such a drastic result?
    any suggestion to mitigate those outcomes is much appreciated.

    Thanks

    the error was misleading!

    The line 72 was correct.

    The error was in a different place. Once corrected, the container runs.



    Code
    TZ: ${TZ}  <-- Changed from = to :      
    PUID: 1005 <-- Changed from = to :      
    PGID: 101  <-- Changed from = to : 

    For some reason, the error was pointing me to the wrong line!.

    Hi,

    I am trying to install Immich (as the title suggests).
    I followed the instruction from the guide:
    how-to-install-immich-photo-upload-server-with-docker

    When deploying and Checking the compose (using the check button), I keep having an error on line 72.

    My line 72 is: - ${DB_DATA_LOCATION}:/var/lib/postgresql/data

    And did not change as it suggested not to do.


    My environment file related to it is:
    # The location where your uploaded files are stored

    UPLOAD_LOCATION=/srv/dev-disk-by-uuid-<myUUID>/Gallery


    Basically, from the guide, I only changed the DB path.


    Any help/suggestion is welcome.
    If you need more details from me, please ask.

    thanks

    Thank you all for your responses.

    To clarify, I haven’t made any manual or explicit changes to the system configuration, including the fstab entry I referred to. I do use NFS shares, those were intentionally set up for my specific needs.

    For now, I’ll keep my home share as it is, since the system configured setup works fine for my current use. Once I gain a better understanding of how to modify it safely, I’ll making adjustments.

    Thanks again for your help!