Software RAID (mdadm) fails on external array after waking from suspend
I have an internal and external software raid. After suspend, the external array fails and starts to rebuild (see below syslog) whereas the internal array is fine.
Mar 31 22:58:56 localhost kernel: [ 3667.077378] md/raid:md126: Disk failure on sdq1, disabling device.
Mar 31 22:58:56 localhost kernel: [ 3667.077378] md/raid:md126: Operation continuing on 9 devices.
Mar 31 22:58:56 localhost kernel: [ 3667.077386] md/raid:md126: Disk failure on sdr1, disabling device.
Are there any additional logs that would be helpful? If I remember correctly the problem is a result of the external drives geting new labels (/sdu1) after waking from suspend. Those drives are all connected to the Silicon Image eSATA card to arrays with port-multipliers (eSATA cables to 2 5-drive arrays).
Relevant info:
/etc/lsb-release
DISTRIB_ID=Ubuntu
DISTRIB_
DISTRIB_
DISTRIB_
lspci
00:1f.2 SATA controller: Intel Corporation 6 Series/C200 Series Chipset Family SATA AHCI Controller (rev 05)
03:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9123 PCIe SATA 6.0 Gb/s controller (rev 11)
03:00.1 IDE interface: Marvell Technology Group Ltd. 88SE91A4 SATA 6Gb/s Controller (rev 11)
04:00.0 RAID bus controller: Silicon Image, Inc. SiI 3132 Serial ATA Raid II Controller (rev 01)
09:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9172 SATA 6Gb/s Controller (rev 11)
mdadm --detail /dev/md126
/dev/md126:
Version : 1.2
Creation Time : Sat Aug 20 22:47:33 2011
Raid Level : raid6
Array Size : 7814070272 (7452.08 GiB 8001.61 GB)
Used Dev Size : 976758784 (931.51 GiB 1000.20 GB)
Raid Devices : 10
Total Devices : 10
Persistence : Superblock is persistent
Intent Bitmap : Internal
Update Time : Sun Apr 1 11:45:57 2012
State : active
Active Devices : 10
Working Devices : 10
Failed Devices : 0
Spare Devices : 0
Layout : left-symmetric
Chunk Size : 512K
Name : :eir
UUID : f52b1d04:
Events : 28616
Number Major Minor RaidDevice State
0 65 1 0 active sync /dev/sdq1
1 8 209 1 active sync /dev/sdn1
2 8 241 2 active sync /dev/sdp1
3 8 225 3 active sync /dev/sdo1
4 65 17 4 active sync /dev/sdr1
5 8 193 5 active sync /dev/sdm1
6 8 177 6 active sync /dev/sdl1
7 8 161 7 active sync /dev/sdk1
8 8 145 8 active sync /dev/sdj1
9 8 129 9 active sync /dev/sdi1
Question information
- Language:
- English Edit question
- Status:
- Answered
- For:
- Ubuntu mdadm Edit question
- Assignee:
- No assignee Edit question
- Last query:
- Last reply:
Can you help with this problem?
Provide an answer of your own, or ask Justin Grevich for more information if necessary.