DegradedArray Event

    This site uses cookies. By continuing to browse this site, you are agreeing to our Cookie Policy.

    • DegradedArray Event

      I've enabled messaging and recieved this but cannot make sense of it. Is this something I need to be concerned with, maybe a disk failing?

      Source Code

      1. This is an automatically generated mail message from mdadm
      2. running on OMV
      3. A DegradedArray event had been detected on md device /dev/md/OMVRaid5.
      4. Faithfully yours, etc.
      5. P.S. The /proc/mdstat file currently contains the following:
      6. Personalities : [raid6] [raid5] [raid4]
      7. md127 : active raid5 sda[0] sdd[3] sdb[1]
      8. 8790795264 blocks super 1.2 level 5, 512k chunk, algorithm 2 [4/3] [UU_U]
      9. unused devices: <none>
      Display All
    • Re: DegradedArray Event

      Replace the drive, look what letter it got assigned (sda, sdb, sdc, etc...) and replace it with sdX with the code below: (when sda is your system drive!)

      Source Code

      1. mdadm --stop /dev/md127
      2. mdadm --zero-superblock /dev/sdX
      3. mdadm --assemble /dev/md127 /dev/sd[bcdefghijklm] --verbose --force
      OMV stoneburner | HP Microserver | 256GB Samsung 830 SSD for system | 4x 2TB in a RAID5
      OMV erasmus| Odroid XU4 | 5TB Data drive | 500GB Backup drive
    • Within the gui, used recover and selected the new drive and it started the process great I thought. I'm now receiving the following messages.... I've only got access to the server via a mobile handset right now so no to access info.... have I done something wrong?

      This is an automatically generated mail message from mdadm
      running on OMV

      A DegradedArray event had been detected on md device /dev/md/OMVRaid5.

      Faithfully yours, etc.

      P.S. The /proc/mdstat file currently contains the following:

      Personalities : [raid6] [raid5] [raid4]
      md127 : active raid5 sdd[4](S) sdb[0] sde[3] sdc[1](F)
      8790795264 blocks super 1.2 level 5, 512k chunk, algorithm 2 [4/2] [U__U]

      unused devices: <none>

      Looking at the raid it shows....

      Version : 1.2
      Creation Time : Tue May 28 15:18:10 2013
      Raid Level : raid5
      Array Size : 8790795264 (8383.56 GiB 9001.77 GB)
      Used Dev Size : 2930265088 (2794.52 GiB 3000.59 GB)
      Raid Devices : 4
      Total Devices : 4
      Persistence : Superblock is persistent

      Update Time : Mon Aug 18 13:20:37 2014
      State : clean, FAILED
      Active Devices : 2
      Working Devices : 3
      Failed Devices : 1
      Spare Devices : 1

      Layout : left-symmetric
      Chunk Size : 512K

      Name : N36L-OMV:OMVRaid5
      UUID : 860d598b:36d96569:15b715f8:06ffde88
      Events : 417048

      Number Major Minor RaidDevice State
      0 8 16 0 active sync /dev/sdb
      1 0 0 1 removed
      2 0 0 2 removed
      3 8 64 3 active sync /dev/sde

      1 8 32 - faulty spare /dev/sdc
      4 8 48 - spare /dev/sdd
    • Hers one of the other message I revived
      This email was generated by the smartd daemon running on:

      host name: OMV
      DNS domain: WORKGROUP
      NIS domain: (none)

      The following warning/error was logged by the smartd daemon:

      Device: /dev/disk/by-id/wwn-0x50014ee60338b828 [SAT], 5 Offline uncorrectable sectors


      For details see host's SYSLOG (default: /var/log/syslog).

      You can also use the smartctl utility for further investigation.
      No additional email messages about this problem will be sent.
    • And another while it was recovering the raid..
      This is an automatically generated mail message from mdadm
      running on OMV

      A Fail event had been detected on md device /dev/md/OMVRaid5.

      It could be related to component device /dev/sdc.

      Faithfully yours, etc.

      P.S. The /proc/mdstat file currently contains the following:

      Personalities : [raid6] [raid5] [raid4]
      md127 : active raid5 sdd[4] sdb[0] sde[3] sdc[1](F)
      8790795264 blocks super 1.2 level 5, 512k chunk, algorithm 2 [4/2] [U__U]
      [==================>..] recovery = 93.4% (2738177408/2930265088) finish=335.8min speed=9532K/sec

      unused devices: <none>

      The post was edited 1 time, last by johnyb ().

    • Weird... whats the output of

      'dpkg -l | grep smartmontools'

      Greetings
      David
      "Well... lately this forum has become support for everything except omv" [...] "And is like someone is banning Google from their browsers"

      Only two things are infinite, the universe and human stupidity, and I'm not sure about the former.


      Upload Logfile via WebGUI/CLI
      #openmediavault on freenode IRC | German & English | GMT+1
      Absolutely no Support via PM!

      I host parts of the omv-extras.org Repository, the OpenMediaVault Live Demo and the pre-built PXE Images. If you want you can take part and help covering the costs by having a look at my profile page.
    • Check the local output of you NAS. Maybe it holds because of a missing drive.

      Greetings
      David
      "Well... lately this forum has become support for everything except omv" [...] "And is like someone is banning Google from their browsers"

      Only two things are infinite, the universe and human stupidity, and I'm not sure about the former.


      Upload Logfile via WebGUI/CLI
      #openmediavault on freenode IRC | German & English | GMT+1
      Absolutely no Support via PM!

      I host parts of the omv-extras.org Repository, the OpenMediaVault Live Demo and the pre-built PXE Images. If you want you can take part and help covering the costs by having a look at my profile page.
    • Yeah unfortunately I'm no where near the server right now, so it'll have to wait until next Monday. It's a headless unit so cannot even talk anyone through looking at it. Strange since all I've done is replace the faulty drive, asked our to rebuild the raid 5 and now this. Surely it would be pretty bad luck to get another faulty drive
      Thanks for your help

      The post was edited 1 time, last by johnyb ().

    • So home now and have physical access to the server, results to comand as follows:

      Source Code

      1. root@OMV:~# smartctl -a /dev/sdc
      2. smartctl 5.40 2010-07-12 r3124 [x86_64-unknown-linux-gnu] (local build)
      3. Copyright (C) 2002-10 by Bruce Allen, [url]http://smartmontools.sourceforge.net[/url]
      4. === START OF INFORMATION SECTION ===
      5. Device Model: WDC WD30EFRX-68AX9N0
      6. Serial Number: WD-WMC1T2837415
      7. Firmware Version: 80.00A80
      8. User Capacity: 3,000,592,982,016 bytes
      9. Device is: Not in smartctl database [for details use: -P showall]
      10. ATA Version is: 9
      11. ATA Standard is: Exact ATA specification draft version not indicated
      12. Local Time is: Mon Aug 25 16:00:56 2014 BST
      13. SMART support is: Available - device has SMART capability.
      14. SMART support is: Enabled
      15. === START OF READ SMART DATA SECTION ===
      16. SMART overall-health self-assessment test result: PASSED
      17. General SMART Values:
      18. Offline data collection status: (0x82) Offline data collection activity
      19. was completed without error.
      20. Auto Offline Data Collection: Enabled.
      21. Self-test execution status: ( 0) The previous self-test routine completed
      22. without error or no self-test has ever
      23. been run.
      24. Total time to complete Offline
      25. data collection: (41040) seconds.
      26. Offline data collection
      27. capabilities: (0x7b) SMART execute Offline immediate.
      28. Auto Offline data collection on/off support.
      29. Suspend Offline collection upon new
      30. command.
      31. Offline surface scan supported.
      32. Self-test supported.
      33. Conveyance Self-test supported.
      34. Selective Self-test supported.
      35. SMART capabilities: (0x0003) Saves SMART data before entering
      36. power-saving mode.
      37. Supports SMART auto save timer.
      38. Error logging capability: (0x01) Error logging supported.
      39. General Purpose Logging supported.
      40. Short self-test routine
      41. recommended polling time: ( 2) minutes.
      42. Extended self-test routine
      43. recommended polling time: ( 255) minutes.
      44. Conveyance self-test routine
      45. recommended polling time: ( 5) minutes.
      46. SCT capabilities: (0x70bd) SCT Status supported.
      47. SCT Error Recovery Control supported.
      48. SCT Feature Control supported.
      49. SCT Data Table supported.
      50. SMART Attributes Data Structure revision number: 16
      51. Vendor Specific SMART Attributes with Thresholds:
      52. ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
      53. 1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail Always - 1871
      54. 3 Spin_Up_Time 0x0027 172 172 021 Pre-fail Always - 6366
      55. 4 Start_Stop_Count 0x0032 100 100 000 Old_age Always - 485
      56. 5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail Always - 0
      57. 7 Seek_Error_Rate 0x002e 200 200 000 Old_age Always - 0
      58. 9 Power_On_Hours 0x0032 086 086 000 Old_age Always - 10881
      59. 10 Spin_Retry_Count 0x0032 100 100 000 Old_age Always - 0
      60. 11 Calibration_Retry_Count 0x0032 100 253 000 Old_age Always - 0
      61. 12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 21
      62. 192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age Always - 10
      63. 193 Load_Cycle_Count 0x0032 200 200 000 Old_age Always - 474
      64. 194 Temperature_Celsius 0x0022 123 113 000 Old_age Always - 27
      65. 196 Reallocated_Event_Count 0x0032 200 200 000 Old_age Always - 0
      66. 197 Current_Pending_Sector 0x0032 200 200 000 Old_age Always - 19
      67. 198 Offline_Uncorrectable 0x0030 200 200 000 Old_age Offline - 5
      68. 199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age Always - 0
      69. 200 Multi_Zone_Error_Rate 0x0008 200 200 000 Old_age Offline - 378
      70. SMART Error Log Version: 1
      71. No Errors Logged
      72. SMART Self-test log structure revision number 1
      73. No self-tests have been logged. [To run self-tests, use: smartctl -t]
      74. SMART Selective self-test log data structure revision number 1
      75. SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
      76. 1 0 0 Not_testing
      77. 2 0 0 Not_testing
      78. 3 0 0 Not_testing
      79. 4 0 0 Not_testing
      80. 5 0 0 Not_testing
      81. Selective self-test flags (0x0):
      82. After scanning selected spans, do NOT read-scan remainder of disk.
      83. If Selective self-test is pending on power-up, resume after 0 minute delay.
      Display All

      The post was edited 1 time, last by WastlJ ().