failed drive?

poma pomidorabelisima at gmail.com
Thu Nov 13 20:23:44 UTC 2014


On 13.11.2014 21:18, dustin kempter wrote:
> Hi all, I'm having some issues. I'm a little confused, so I was checking 
> our servers today and saw something strange. cat /proc/mdstat shows that 
> 1 device md0 is inactive. I'm not really sure why. id did a bit more 
> digging and testing using smartctl and it says that the device /dev/sdg 
> (part of md0) is failing, estimated to fail within 24 hrs. but if i do 
> df -h it doesn't even show md0, and was talking to a friend and we 
> disagreed. I believe that based on what smrtctl says the drive is 
> failing but not failed yet. he doesn't think its a problem with the 
> drive. do you have any thoughts on this? and why would the device (md0) 
> suddenly be inactive but still show 2 working devices (sdg, sdh)?
> 
> *(proc/mdstat)*
> [root at csdatastandby3 bin]# cat /proc/mdstat
> Personalities : [raid1] [raid10]
> md125 : active raid10 sdf1[5] sdc1[2] sde1[4] sda1[0] sdb1[1] sdd1[3]
>        11720655360 blocks super 1.2 512K chunks 2 near-copies [6/6] [UUUUUU]
> 
> md126 : active raid1 sdg[1] sdh[0]
>        463992832 blocks super external:/md0/0 [2/2] [UU]
> 
> md0 : inactive sdh[1](S) sdg[0](S)
>        6306 blocks super external:imsm
> 
> unused devices: <none>
> [root at csdatastandby3 bin]#
> 
> 
> *(smartctl)*
> [root at csdatastandby3 bin]# smartctl -H /dev/sdg
> smartctl 5.43 2012-06-30 r3573 [x86_64-linux-2.6.32-431.17.1.el6.x86_64] 
> (local build)
> Copyright (C) 2002-12 by Bruce Allen, http://smartmontools.sourceforge.net
> 
> === START OF READ SMART DATA SECTION ===
> SMART overall-health self-assessment test result: FAILED!
> Drive failure expected in less than 24 hours. SAVE ALL DATA.
> Failed Attributes:
> ID# ATTRIBUTE_NAME          FLAG     VALUE WORST THRESH TYPE UPDATED  
> WHEN_FAILED RAW_VALUE
>    5 Reallocated_Sector_Ct   0x0033   002   002   036    Pre-fail 
> Always   FAILING_NOW 32288
> 
> [root at csdatastandby3 bin]#
> 
> *(df -h)*
> [root at csdatastandby3 bin]# df -h
> Filesystem      Size  Used Avail Use% Mounted on
> /dev/md126p4    404G  4.3G  379G   2% /
> tmpfs            16G  172K   16G   1% /dev/shm
> /dev/md126p2    936M   74M  815M   9% /boot
> /dev/md126p1    350M  272K  350M   1% /boot/efi
> /dev/md125       11T  4.2T  6.1T  41% /data
> [root at csdatastandby3 bin]#
> 
> 
> *(mdadm -D /dev/md0*
> [root at csdatastandby3 bin]# mdadm -D /dev/md0
> /dev/md0:
>          Version : imsm
>       Raid Level : container
>    Total Devices : 2
> 
> Working Devices : 2
> 
> 
>             UUID : 32c1fbb7:4479296b:53c02d9b:666a08f6
>    Member Arrays : /dev/md/Volume0
> 
>      Number   Major   Minor   RaidDevice
> 
>         0       8       96        -        /dev/sdg
>         1       8      112        -        /dev/sdh
> [root at csdatastandby3 bin]#
> 
> 
> thanks
> 
> 
> -dustink
> 
> 
> 

Maybe a coffee filter.




More information about the users mailing list