Posts by curto

    Hi Guys,


    I have an OMV 6 box been runnign fine for ages.


    Use two external drives in a USB housing - some external event (presumably power issues) has taken both drives out and they no longer show at the command line (FDISK) nor after a reboot of the system


    I have taken them to a dffierent machine (Debian) and tried them internally (they have an ext4 file system) and they are not recognized so i assume they have completely failed.


    They show in the dashboard as missing and when i drill down through the storage menus they are not showing either.


    They had a filesystem on them (ext4) that was shared through NFS


    What are the correct steps to take to remove them full from the system so i can install replacement drives.


    If i go into Storage/Filesystems - they show as missing and referenced and i have not been able to find a combination to remove them


    Craig

    You have linked my thread and in the meanwhile I have been posting in this other one. Check your fstab_tasks.json to see what it contains.

    I think that's what causes it, I just don't know how to fix the config yet.

    My fstab_tasks.json contains this entry related to deleting the root filesystem:

    Code

    Can you check the content of your [tt]fstab_tasks.json[/tt] and post your blkid? Let's see if it's the same issue.

    Here is the content of mine - not sure why it is deleting filesystems ?



    Thanks for taking the time to post - i will read through the other thread and see where that leads me also and post back


    Here is my actual fstab


    Here is the pending configuration changes and this is what happens when i apply them


    And the system then sits there with the 502 bad gateway.


    Any previously established SSH sessions remain open - but can not do anything - however having a shutdown -r 10 command stacked prior to performing the updates does allow the system to restart and returns control.


    Any ideas ?


    Craig

    Yes all of those steps i listed above i perfomed from the CLI i.e. no involvement of OMV. After this weekend when the lastest set of replications is done i can redo them through the GUI and will monitor Journactl


    The point of the above steps was to show there are no hardware issues with the drives or the physical setup of the system.


    The UI is currently sitting there saying the following modules need updating



    Will report back as to the outcome of journalctl once i process those changes


    Craig

    Thanks for taking the time to respond.


    I have just gone through the following manual actions at the CLI (as i needed it in place for this weekend)


    1) gdisk both USB drives and placed a standard linux fd00 partition on them using all the defaults

    2) ran mkfs.ext4 on both drives and created the filesystems tables and inodes etc

    3) created a subdirectory for each drive to mount to in the export folder (which is where i want them to end up when i get the GUI working)

    4) mounted the /dev/sdx1 partition from both drives in /export/USB-1 and /export/USB-2

    5) manually edited the exports file in /etc and added the correct NFS sharing for the drives

    6) shared the NFS export through exportfs -a


    Went to my ESXi console and successfully mounted both drives as Datastores and am now writing VMware Replication jobs to both drives concurrently - so far each drives has written approx 900GB with no issues.


    I think this proves that all the hardware etc is OK


    I just ran journalctl -f and the only thing of note is some postfix errors in there for a misconfigured forwarder


    I am reluctant to restart the server until Monday once i have these copy jobs finalised.


    Is there something else i can try/provide you with in the meantime ?


    Craig

    OK still no change as i continue to try things - still getting the error when applying config changes


    Code
    Http failure response for http://192.168.0.40/rpc.php: 502 Bad Gateway

    Only recovery method is either through the power down manual switch or to have previously scheduled a shutdown/reboot from the CLI


    This is what i am doing each time before trying something


    shutdown -r 10


    Seems to work as the box comes back each time but still none of the config changes have been applied.


    Craig

    Continuuing on with this - i have gone into the plugins section and removed all the optional plugins


    I have rebooted after that and am now trying the whole process of integrating the drives again - still the same thing happening.


    Of interest though - from the CLI before i make any changes i do a shutdown -r 10 - which delays a reboot for 10 minutes. Even though i lose connection to the server and can not do anything at the command line the reboot works and i get control back.


    Craig

    Doing some more reading and there are a couple of other threads where people have raised this issue



    I am working through the steps in there and will add them to this thread as i go for any one else with this



    Hi Guys,


    I am re-experiencing a problem that i have had before - there are two other threads that seem to touch on it




    I have previously chimed in on both of these.


    This problem has come back and bitten me today - i note that i probably forgot about it/left it after the last round - the server prior to today had been up for 301 days.


    This is remote server and has not had an issue in that time - it just sits there and churns away in the background.




    I went back to my notes and realise i had not solved this the last time and just worked around it.


    I have 2 x USB external drives i wish to add to the system.


    From the CLI i can fdisk to find the disks, then mount them (they have an EXT4 filesystem) with no problems - obviously then OMV knows nothing about them in file system terms. They do show up in the GUI under disks.


    I can mount them at the CLI and write to them on an extended basis (they have approx 1TB of data written to them once per week) as a form of secondary backup storage. They have had this working fine with commandline for the last approx 10 months.


    Today i tried to go back and access them through the GUI and set them up in the OMV way - Filesystem, then Shared Folders to be mounted etc.


    As soon as i do this - i get a series of error messages in the gui - culminating in 502 Bad Gateway.


    The CLI is then unable to do anything with reports of systemctl not running.


    Currently, the content of /var/lib/openmediavault/dirtymodules.json is


    Code
    root@omv-afsydney:/var/lib/openmediavault# ls
    apt  borgbackup  cron.d  dirtymodules.json  dirtymodules.json.bkp  engined.sock  fstab_tasks.json  rrd  rsnapshot.d  salt  ssh  workbench
    root@omv-afsydney:/var/lib/openmediavault# cat dirtymodules.json
    [
        "collectd",
        "fstab",
        "monit",
        "quota",
        "nginx"
    ]root@omv-afsydney:/var/lib/openmediavault#



    What other information would you like or what can you suggest i do


    I currently have the pending changes notification on the screen


    I had to get someone to do a hard power off of the machine to retake control and the screenshots etc above are after that restart.


    Craig

    OK - i was just seeing some weird stuff on my switches and wanted to make sure this was the case. So as long as no VLANs are configured - this will be populated (as it is a required Integer field) but will be ignored and will not be appended to the packet being sent.


    The Brocade/Ruckus switches default to VLAN 1 - but as untagged packets - no way to set them up with tagging - so in this case if OMV was sending out the packets with a tagged VLAN of 1 - it breaks the switching of the packets.


    Craig

    Nope i had not done anything with VLANs - and have just set up another interface (the 2nd port on that card) and it did the same thing when looking at the config file)


    It looks like it is putting some default value in there ?


    Craig

    Guys,


    Been trying to setup a 10Gb link to my switch on OMV using an intel Dual port 520DA-2 (intel 82599 chip) network adapter.


    I could not work out why my switch would not bring the port up - so i went digging into the OMV files and found this in the XML file




    And here is the network screen - note this box has not had VLANs ever implemented on it - yet in the config file above VLAN ID 1 has been set - on many switches this VLAN will not be accessible or will be heavily restricted




    Is this by design that this is in the file or has something gone wrong somewhere ?


    Craig

    So by editing the Grub scripts to put nomodeset on the end of each kernel boot entry and then regenerating the grub menus i was able to get my system back and running with all drive loaded - once the raid array disks were back in the system i went and deleted the dirty modules json file inidicated earlier in this thread and was then able to apply updates etc.


    So i can not categorically say this is the solution but it definitely appears to be somehow related to missing disks that are referenced in FSTAB and not being loaded correctly (for whatever reason)


    Craig

    Maybe i can add a bit more to this to try and point a direction for resolution.


    As the OP noted this appeared to have something to do with failed mounts of drives - i too am in this situation


    1) Here is a link to a previous issue i found when using LSI controllers wth a certain kernel and motherboard revision RE: Ryzen 3200G boot issues - ASUS B450 MB - when i performed the upgrade to to OMV 6 it appears this issue has come back - i would have thought this would have been solved by now as it was indicated this was just a timing difference between releases of different chipsets.


    2) So based on the above i have a large array that is referenced in fstab and is not being mounted because the drivers fail to load correctly.


    This seems to be (for me anyway) the root cause of the probem - but i will do some more troubleshooting and report back


    Craig

    Hey guys - wanting to chime in here also as i thought i was going mental - exactly the same behaviour as above. Once a deploy is done - 502 pops up on the console, the SSH session is also not happy and will not allow a shutdown/reboot.


    The only thing that works is a ctrl alt del and then we get the same merry go around again.


    Mine is an ASUS 550m motherboard with an Intel 10GB (Intel 540T-2) and an LSI 8211 flashed to IT mode.


    I will try to remove the dirty modules file and report back also


    Craig

    Just as a follow up to this - does not seem to be anyway to do this through the GUI - as typically a failed arrary will not even mount so i did it all through the CLI.


    Login as root at the CLI


    type

    cat /proc/mdstat


    This should list your arrays and which ones are active - typically it will be inactive if it has a failed drive.


    Note if the drive has failed which one it is which is listed iin the array description


    Check the reports you have from OMV that will list the serial number of the drive that has failed (or is failing)


    run


    hdparm -I /dev/sd* - this will list all the drives that system sees - scroll through and look for the Serial # in question and which drive it is - if it has totally failed - it will not be listed.


    Power down the box - sort through the drives (physically) until you find the correct one - remove it and replace with new drive - i very highly recommend that this is ablank drive - do not leave partitions on there from other systems/OSs - note down the serial # of this drive


    Bring the box back up and once it boots - run the hdparm command above to identify what letter your new drive has been given


    use the cat/proc/mdstat command again to see if the array is active.


    If not make it active


    Then add the new drive to the array


    As the OMV way of putting drives into a MDADM array is to use the whole drive with no partitioning this should work fine and start syncing straight away - you may have issues if drive sizes are different - not sure why OMV insists on full drives rather than partitioning


    You can monitor the progress every now and then by looking at the cat/proc/mdstat device


    Once it is finished i would recommend a reboot so all drives and shares etc are initialised correctly




    Craig

    Hey guys, getting messages through OMV that one of my drives in a RAID10 (4 drive set) is having unrecoverable errors. I am going to the place tomorrow and will be taking a replacement drive with me - what is the OMV method for doing this ?


    In raw linux i would from the command line - fail the drive out of array (firstly noting the serial number), power down the box - then do the physical changeover and from the commandine, add a partition using gdisk - which i would copy from another drive in the set to get the correct layout, then use mdadm to add the drive back into the raid set.


    Whats the OMV way to do this through the GUI ?


    Craig

    It's not up to date yet (the screen shots are for OMV5) but the permissions concepts, in the permissions doc, are the same for OMV6 and the screen presentation is not hugely different.

    I think you'll find what you want -> here. This link leads to a section of the permissions doc where SMB access using IP's is explained. Note that if an IP address is specified, the user of the IP specified workstation will still be required to have access at the shared folder level. It becomes an "AND" operation. Access will require an allowed IP address AND a user within an authorized group.

    Note that if an address is inserted into the "Allow" field, all workstation IP's that are allowed for that share must be added to the field. (And, as noted in the doc, using DHCP can be problematic.)


    Again, while the screen shots are out of date, the concepts are the same. -> Take a look at the Permissions Doc.

    Thanks for the linked document etc - much appreciated.


    Craig

    Hey Guys - this now appears to be quite different in OMV6 ? I can not find any similar screens as shown in the first post in terms of samba user permissions etc ?


    I am trying to do a restricted share on OMV 6 - that will only be accessible from a fixed set of IP addresses and will require a valid login from the OMV users group - just playing around with the options at the moment to get what i want - but they do not seem to be mapping to read me above ?


    Craig

    OK just to document this so it remains fresh in my mind


    1) Dismounted the NFS Datastore on the SSD from my test ESXI 7 host

    2) Deleted the NFS Share on OMV

    3) Deleted the Shared Folder on OMV

    4) Unmounted the Filesystem (EXT4)

    5) Delete the MDADM drive

    6) at the CLI used the hpdarm Secure Erase to clean out the drives


    7) Created a new MDADM Mirror set

    8) Created a new ext4 filesystem and mounted it

    9) Created a new fileshare with the default permissions changed to Everyone Read:write



    10)Created an NFS share and changed the default permissions as follows - changed to Read/Write from default Read, and added Async to the options



    11) Created a new Datastore on ESXi 7 (through Vsphere) - specified NFS3 as the type, changed the default access to Read/Write and Mounted it


    12) copied files from another datastore to this one using the Datastore browser in ESXi - all worked OK !!


    13 When i use WINSCP to look at the file permissions within the Datastore (by pointing at the ESXi host) i see the following


    14) and then to close the loop i look at the filesystem in OMV using WINSCP


    and i can see that nobody has ownership



    So looks like all good now - thanks


    I will put the issues down to needing to reboot the brocade switch after enabling Jumbo frames


    Craig