Beiträge von bbddpp

    I've been running my grounds-up built Dell C2100 rack server since November and while drives have run cooler than ever and have shown nothing but green lights on the SMART screen, I am on what looks like either the second or maybe even the 3rd drive starting to fail out of the blue, looking for some help.


    I do not run a RAID first of all, stubbornly, I like to get the maximum space out of a file system, and I run desktop drives in here. The server just is for serving media to my home and running some apps to search for media, the usual stuff. It's not getting hammered and lives in a cool basement. Drive temps run 26-29 C.


    Anyway, here's what happened this time.


    On a routine SSH session I realized that one of my mountpoints (a 3TB Toshiba drive) had gone "read only" which is a red flag to me that OMV found something wrong in its file system. SMART screen shows green dot next to the drive. I'm afraid to reboot OMV because the last time I did that, I wasn't able to ever re-mount the drive to get the media off. I tried an rsync command to backup the files off the bad drive to another empty drive and it froze. I'm trying a basic cp command now in the shell a few files at a time and still getting Input/Output errors on every file. I tried a short self-test on the device and got this:


    Short INQUIRY response, skip product id
    A mandatory SMART command failed: exiting. To continue, add one or more '-T permissive' options.


    What am I doing wrong here? Why are drives going read-only out of the blue? Am I missing something obvious the way I have things set up? Is there anything I can do to rescue this drive before I reboot OMV or unmount the drive and probably lose all my data on an entire volume AGAIN?


    My drives are all set up as:


    Advanced Power Management: 128 - Minimum power usage without standby (no spindown)
    Automatic Acoustic Management: Maximum performance, maximum acoustic output
    Spindown time: Disabled
    Write cache: off


    I do not run routine SMART self-tests and maybe I should be doing that moving forward (anyone have any suggestions on what tests they run?). All that said, I don't get how drives are just failing in months inside of this thing, without any warning signs.


    Anything else I can do here or am I basically just going to have to expect to lose a drive every few months out of nowhere because I'm not running RAID or using network-level drives?


    Happy to provide any additional detail from logs, fstab, etc.


    Thanks for anyone who can help with this frustration.

    I think a couple plugin experts are lurking about, so I am hoping for some help for a problem I can't seem to resolve.


    Plex Media Server's external visibility ("fully accessible outside your network") seems to randomly stop working, usually every few hours but sometimes shorter, sometimes longer. I've rebooted both the server and my router and am on the latest versions. The server itself is rock solid and never goes down and works perfectly when I am on my Wifi. It's only the external piece of forwarding out that seems to just randomly quit working all day long.


    Going back into the plex interface and hitting the "apply" button on the server's remote access tab in settings will within seconds immediately restore external visibility with no changes to my settings or router settings. I do not specify a port manually or have 32400 forwarding in my router. It just works for a few hours, then, stops.


    If there are any particular logs you can cite that I can refer to I'd be happy to send snippets or full logs, just don't want to go on a goose chase.


    Any suggestions or help would be greatly appreciated!


    Much thanks as always, and thanks for OMV which remains my favorite piece of software of all time because I love and use it daily.

    flmaxey,


    First and foremost thanks for the generosity of your reply and comments. I totally agree that I should have cloned a working booting OMV the moment I had it, even to another SSD drive I could connect in a pinch. Alas, nothing to show for here and it has to be something that happened on one of the many upgrades over the past few months that turned quotas on and made them matter to the point where something is out of whack.


    Alas, I've tried removing all 8 hard drives, leaving only the boot SSD OMV device in there, and not had any luck. The "update-rc.d quota disable" command still throws the same 2 warnings when I run it even when booting with just the OMV volume, and OMV itself doesn't show any interest in booting into the full system even on a drive by itself.


    I think I need a different way to turn quotas off other than the "update-rc.d quota disable" command, or a way to make that command work and take, so that the quota service absolutely does not run.


    Sidenote: OMV spends 1 minute and 30 seconds searching for all the drives before it proceeds when I unplug them, or when they are plugged in, the same amount of time trying to start the quota service on them. It's weird.


    Anyway, hopeful for a miracle cure here, it's so frustrating to not be able to run my server. I guess the nuclear option will be a fresh install and to start totally over which scares me to death. There has to be a way to turn quotas off via config change in the XML or something here without this command, or a way to force this command without it throwing the warnings and not taking hold.


    If anyone has any leads I'd sure appreciate.

    Tried a few more things. I removed the volume it was freezing on at boot and then it just picked 2 more volumes to do the same thing with, so it is definitely the quota script running against the first volume it finds with the "quota issue" I suppose, I had about 4 or 5 volumes in the "red" since I had them practically full.


    I found this command and tried it:


    "update-rc.d quota disable"


    But it doesn't work, it instead throws 2 warning messages from insserv which report that runlevels of script quota are not matching defaults. Doesn't turn quotas off.


    Something in one of these recent updates (again, haven't rebooted in a few months to track down when) is certainly not playing well with my single drives that are near capacity in this thing.


    I also found this thread:


    OMV hangs on boot


    Very similar issues to me, but the command did throw the screenshot of the warnings that appears on page two, but did not indeed turn off any quotas.


    Any other way I can force this quota thing off and out of my life would be appreciated.


    I did see a post saying to remove every drive except the boot drive, boot into the system, run the command, then remount all drives, which I suppose I can try next?


    This is crazy! What caused this?

    First reboot in close to 90 days, been on OMV3.0 for a while and keeping up with patches. Plex server wasn't running right so figured I'd cold boot the server.


    For the record, I just run this with a bunch of drives inside with media, my server supports RAID but I don't use it. Just a bunch of separate ext4 volumes inside. No pooling.


    During the boot, the system stalls while a certain device seems to be trying to initiatilze. I believe this is one of the hard drives. It waits a minute and 30 seconds before giving up. All drives show up as clean. I believe I know which device though it is just giving a disk ID but not a volume name. I removed that device and rebooted and got same result.


    The command that the shell told me to run from console mode when quota mode didn't start was "status quotaon.service". I ran that and got (skipping some basic stuff since I am manually keying this):


    Loaded: loaded
    Active: failed, (Result: exit-code)
    Process: 1679 ExecStart=/sbin/quotaon -aug (code: exited, status=10)


    It then claims it cannot find aquota.user or aquota.group on several of my volumes. I will say that I do not want to track quotas and max these volumes out and I have a feeling the fact that my disks are near full (they showed red in device manager when you looked at them during a working boot)...this may be the problem.


    Unit quotaon.service entered failed state


    Please let me know if I can provide logs or anything else (just need to figure out HOW to get a boot log off a system that won't boot and onto something I can transfer it online for). I assume the quotas are the key here and telling the config how to ignore them?


    Any advice would be most appreciated!

    I have so many boxes around my house that use the same SQL database. Fire TVs, Raspberry Pi, HTPC, Nvidia Shield...I always felt NFS made sense because it used to be fastest (maybe SMB beats it now I dunno)...But NFS was universal and worked everywhere and SMB used to not work on all boxes.


    Maybe that has changed since the last time I set this all up.

    I'm an idiot. That's what happens when you spend your day totally rebuilding your server. You get dumb after 8 hours.


    I was forgetting the "export" in the path. That fixed Kodi right up for me. Kodi still refuses to browse the server at the root level when I add the source in the GUI, but adding the source manually in sources.xml works perfectly.


    I'm not sure if NFS is still the best solution for streaming in home these days, but my variety of Kodi boxes (openelec, nvidia shield, PCs, etc) all seem to agree on NFS so it's just easier.

    So I've got NFS set up and everything as perfect as can be under OMV 3.0 and my darn Apple OSX Mac Mini just refuses to browse the NFS shares.


    I've tried all the uid and pid stuff on the shares. My last 2 tries were:


    subtree_check,all_squash,insecure,anongid=100,anonuid=0
    subtree_check,all_squash,insecure


    With no luck. Kodi refuses to browse the NFS server at all, and won't even use the manual adds of shares I added to the sources.xml.


    When I try and just browse to nfs://192.168.X.XXX on the Mac, it says "You do not have permission to access this server".


    So, I'm stumped. I know NFS doesn't have users and passwords, but it's obvious there needs to be some sort of synchonized user account or LAN setting or something here somewhere on the Mac and/or in OMV to get this working.


    I read somewhere that I should just use SMB, which works great in Kodi on my Mac (where I do all my Kodi Library work)...But I seem to always recall that NFS had a much better reputation for speed when it comes to streaming?


    Every other machine in my house, from Kodi on android, to Kodi on Windows 10, can see OMV's NFS shares no problem. It's just the OSX machine that doesn't connect to them, and that's the most important one! Has Apple just borked up NFS?


    Anyone else solved the NFS problem on an OSX machine to talk to OMV NFS shares?

    Check this out from mysql error.log



    161203 15:53:50 [Note] - '127.0.0.1' resolves to '127.0.0.1';
    161203 15:53:50 [Note] Server socket created on IP: '127.0.0.1'.
    161203 15:53:50 [ERROR] Fatal error: Can't open and lock privilege tables: Table 'mysql.host' doesn't exist
    161203 15:53:50 mysqld_safe mysqld from pid file /var/run/mysqld/mysqld.pid ended
    161203 15:56:30 mysqld_safe Starting mysqld daemon with databases from /media/3ed4d2b9-03c0-4b70-b850-82df1e1757f7/SQL
    161203 15:56:30 [Warning] Using unique option prefix key_buffer instead of key_buffer_size is deprecated and will be removed in a future release. Please use the full name instead.
    161203 15:56:30 [Note] /usr/sbin/mysqld (mysqld 5.5.53-0+deb8u1) starting as process 14916 ...
    161203 15:56:30 [Warning] Using unique option prefix myisam-recover instead of myisam-recover-options is deprecated and will be removed in a future release. Please use the full name instead.
    161203 15:56:30 [Note] Plugin 'FEDERATED' is disabled.
    /usr/sbin/mysqld: Table 'mysql.plugin' doesn't exist
    161203 15:56:30 [ERROR] Can't open the mysql.plugin table. Please run mysql_upgrade to create it.
    161203 15:56:30 InnoDB: The InnoDB memory heap is disabled
    161203 15:56:30 InnoDB: Mutexes and rw_locks use GCC atomic builtins
    161203 15:56:30 InnoDB: Compressed tables use zlib 1.2.8
    161203 15:56:30 InnoDB: Using Linux native AIO
    161203 15:56:30 InnoDB: Initializing buffer pool, size = 128.0M
    161203 15:56:30 InnoDB: Completed initialization of buffer pool
    161203 15:56:30 InnoDB: highest supported file format is Barracuda.
    InnoDB: Log scan progressed past the checkpoint lsn 48941
    161203 15:56:30 InnoDB: Database was not shut down normally!
    InnoDB: Starting crash recovery.

    It's only being used for a Kodi database, which gets its share of use, but not all day and night type of stuff.


    I left that field blank, and the plugin is turne don so I figured I'd just try and restart the service via ssh.


    Here's the output of "service mysql start"


    Job for mysql.service failed. See 'systemctl status mysql.service' and 'journalctl -xn' for details.


    Output of: "systemctl status mysql.service"


    Code
    * mysql.service - LSB: Start and stop the mysql database server daemon
       Loaded: loaded (/etc/init.d/mysql)
       Active: failed (Result: exit-code) since Sat 2016-12-03 15:54:18 EST; 11s ago
      Process: 13439 ExecStart=/etc/init.d/mysql start (code=exited, status=1/FAILURE)
    Dec 03 15:54:18 OMV mysql[13439]: Starting MySQL database server: mysqld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . failed!
    Dec 03 15:54:18 OMV systemd[1]: mysql.service: control process exited, code=exited status=1
    Dec 03 15:54:18 OMV systemd[1]: Failed to start LSB: Start and stop the mysql database server daemon.
    Dec 03 15:54:18 OMV systemd[1]: Unit mysql.service entered failed state.

    Next thoughts? Total uninstall and reinstall? Pretty sure the first time I turned it on, even before I moved the path, it threw an error, so something else might be up here.

    I always thought I should store my databases on something other than the main OMV volume, especially since I am using an SSD. Should I just leave them where they are by default on the SSD OS drive? In that case would I still need to specify an alternate location or just leave that field blank? Because I think when I left it blank, I got an error as well (a different error, but an error all the same).

    Thanks for the push into 3.0, ryecoaaron.


    That fixed all my problems with the segmentation faults.


    Just troubleshooting a MySQL plugin issue now, but will start a thread for that in the proper place.


    3.0 looks nice, all the stuff that made 2.0 great, made even more efficient. And no segmentation faults is awesome too. That was driving me nuts at 2 AM last night! :)

    Checked the forum for similar threads but found no definitive solution, hoping I can get some help. This is on a fresh install just done today of OMV 3.0 beta.


    MySQL plugin can be turned on, but when Data Directory is set, and saved, errors happen.


    datadir = /media/3ed4d2b9-03c0-4b70-b850-82df1e1757f7/SQL


    I figured I'd pull out all stops and make mysql the user and group owner of the above folder with full permissions and created an internal share. Didn't change anything.


    GUI Interface Error:








    Output of: "systemctl status mysql.service"


    What obvious thing am I missing here?

    Is it possible that it's not the drive's fault even though I see the error during a write operation?


    Could it be CPU or Memory causing this to occur? Or an incorrect BIOS setting or unsupported hardware?


    I'm running a memtest next.

    I am seeing segmentation faults on a fresh install of OMV on a new SSD. Smart data is fine, green light and all tests good on SSD. Prior SSD had gone red so I thought a new one would solve it.


    So, things I have tried:


    - Immediately installed flashmemory plugin (first boot)
    - Backports Kernel
    - Changed SSD
    - Uninstalling and re-installing plugins
    - Clean fresh install of OMV
    - Changing SATA data cable to SSD


    I have now tried 3 different SSD drives and all 3 are throwing segmentation faults. What else could cause this and should I be checking? Memory? Something else? Is it possible I am using some sort of system that is incompatible with OMV and will throw segmentation faults?


    Log snippets:




    I'm stumped.

    Log Part 3:


    Log Part 2, Continued:



    Sorry for the cross-post, i realized I posted my initial info in the wrong forum.


    I will admit this is my first full size server and I may be in a bit over my head. I am using a Dell C2100 which has 12 bays and all kinds of hardware inside (backplane, cabling, etc) to make the 12 drives work. I bought this used so my fear is that I have some bad hardware in here (the drive seems to pass all tests).


    3 times now, while performing a copy operation to this drive, the system has frozen, the drive has errored out and come back offline. What worries me is that the log, while it looks mostly greek to me, may be telling me something about the hardware (backplane, etc) failing that can tell me which component is the culprit, but I just can't seem to make heads or tails of it.


    I have the full log at the exact spot it happens, from the moment things go "bad", and am hoping someone might see something in the messages here that spells out what I should be looking for.


    This admitted server newb would appreciate any insight at all, sincerely! If this points to a hardware failure and there are some components I should be checking, that would be a great help in a head start. This used server has a 30 day warranty so perhaps if there is a faulty component somewhere in the drive array I can get it replaced.


    LOG in 3 PARTS, Part 1: