Software RAID (mdadm) fails on external array after waking from suspend

Asked by Justin Grevich

I have an internal and external software raid. After suspend, the external array fails and starts to rebuild (see below syslog) whereas the internal array is fine.

Mar 31 22:58:56 localhost kernel: [ 3667.077378] md/raid:md126: Disk failure on sdq1, disabling device.
Mar 31 22:58:56 localhost kernel: [ 3667.077378] md/raid:md126: Operation continuing on 9 devices.
Mar 31 22:58:56 localhost kernel: [ 3667.077386] md/raid:md126: Disk failure on sdr1, disabling device.

Are there any additional logs that would be helpful? If I remember correctly the problem is a result of the external drives geting new labels (/sdu1) after waking from suspend. Those drives are all connected to the Silicon Image eSATA card to arrays with port-multipliers (eSATA cables to 2 5-drive arrays).

Relevant info:

/etc/lsb-release

DISTRIB_ID=Ubuntu
DISTRIB_RELEASE=12.04
DISTRIB_CODENAME=precise
DISTRIB_DESCRIPTION="Ubuntu precise (development branch)"

lspci

00:1f.2 SATA controller: Intel Corporation 6 Series/C200 Series Chipset Family SATA AHCI Controller (rev 05)
03:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9123 PCIe SATA 6.0 Gb/s controller (rev 11)
03:00.1 IDE interface: Marvell Technology Group Ltd. 88SE91A4 SATA 6Gb/s Controller (rev 11)
04:00.0 RAID bus controller: Silicon Image, Inc. SiI 3132 Serial ATA Raid II Controller (rev 01)
09:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9172 SATA 6Gb/s Controller (rev 11)

mdadm --detail /dev/md126

/dev/md126:
        Version : 1.2
  Creation Time : Sat Aug 20 22:47:33 2011
     Raid Level : raid6
     Array Size : 7814070272 (7452.08 GiB 8001.61 GB)
  Used Dev Size : 976758784 (931.51 GiB 1000.20 GB)
   Raid Devices : 10
  Total Devices : 10
    Persistence : Superblock is persistent

  Intent Bitmap : Internal

    Update Time : Sun Apr 1 11:45:57 2012
          State : active
 Active Devices : 10
Working Devices : 10
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 512K

           Name : :eir
           UUID : f52b1d04:cd7c8a63:439c9bae:6d2a5bfe
         Events : 28616

    Number Major Minor RaidDevice State
       0 65 1 0 active sync /dev/sdq1
       1 8 209 1 active sync /dev/sdn1
       2 8 241 2 active sync /dev/sdp1
       3 8 225 3 active sync /dev/sdo1
       4 65 17 4 active sync /dev/sdr1
       5 8 193 5 active sync /dev/sdm1
       6 8 177 6 active sync /dev/sdl1
       7 8 161 7 active sync /dev/sdk1
       8 8 145 8 active sync /dev/sdj1
       9 8 129 9 active sync /dev/sdi1

Question information

Language:
English Edit question
Status:
Answered
For:
Ubuntu mdadm Edit question
Assignee:
No assignee Edit question
Last query:
Last reply:
Revision history for this message
actionparsnip (andrew-woodhead666) said :
#1

Report a bug

Precise is not ready and not stable at all.

Revision history for this message
Justin Grevich (jgrevich) said :
#2

Same error occurred in 11.10 and 11.04 for me. I just upgraded to precise.

Revision history for this message
actionparsnip (andrew-woodhead666) said :
#3

But its still an issue in precise so should be reported.

Revision history for this message
Justin Grevich (jgrevich) said :
#4

@actionparsnip do you think I should do more than relate this question to this bug: https://bugs.launchpad.net/ubuntu/+bug/804427

Can you help with this problem?

Provide an answer of your own, or ask Justin Grevich for more information if necessary.

To post a message you must log in.