No F21 on Intel BIOS RAID? Workaround?
Ian Pilcher
arequipeno at gmail.com
Fri Dec 5 03:51:38 UTC 2014
On 12/04/2014 03:35 PM, Chris Murphy wrote:
> On Thu, Dec 4, 2014 at 12:21 PM, Ian Pilcher <arequipeno at gmail.com> wrote:
>> I just tried to install RC5 on my home system, which uses Intel BIOS
>> RAID, and anaconda crashed during its initial storage scan.
>>
>> https://bugzilla.redhat.com/show_bug.cgi?id=1170755
>>
>> Short of major surgery (pulling the RAID drives, breaking the RAID,
>> etc.), can anyone think of a way to get F21 installed on this system?
>
> When booted from any install media, in a shell, what do you get for:
Here's what I get when booted into F20:
> cat /proc/mdstat
Personalities : [raid1]
md126 : active raid1 sdb[1] sdc[0]
976759808 blocks super external:/md127/0 [2/2] [UU]
md127 : inactive sdc[1](S) sdb[0](S)
5288 blocks super external:imsm
unused devices: <none>
> mdadm -D /dev/md126
/dev/md126:
Container : /dev/md/imsm0, member 0
Raid Level : raid1
Array Size : 976759808 (931.51 GiB 1000.20 GB)
Used Dev Size : 976759940 (931.51 GiB 1000.20 GB)
Raid Devices : 2
Total Devices : 2
State : clean
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
UUID : 3d7bd72f:82a8cbcc:2d217397:12f3ff95
Number Major Minor RaidDevice State
1 8 16 0 active sync /dev/sdb
0 8 32 1 active sync /dev/sdc
> mdadm -E /dev/sd[ab]
I assume you mean "mdadm -E /dev/sd[bc]".
/dev/sdb:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.1.00
Orig Family : d7e8a7e3
Family : d7e8a7e3
Generation : 00114ca2
Attributes : All supported
UUID : 1ebd7712:2a74af1f:34298316:cb855b50
Checksum : d1e422ca correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1
Disk00 Serial : MSK5235H29X18G
State : active
Id : 00000001
Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
[Volume0]:
UUID : 3d7bd72f:82a8cbcc:2d217397:12f3ff95
RAID Level : 1
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 0
Array Size : 1953519616 (931.51 GiB 1000.20 GB)
Per Dev Size : 1953519880 (931.51 GiB 1000.20 GB)
Sector Offset : 0
Num Stripes : 7630936
Chunk Size : 64 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : dirty
Disk01 Serial : MSK5235H2PJ7TG
State : active
Id : 00000002
Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
/dev/sdc:
Magic : Intel Raid ISM Cfg Sig.
Version : 1.1.00
Orig Family : d7e8a7e3
Family : d7e8a7e3
Generation : 00114ca2
Attributes : All supported
UUID : 1ebd7712:2a74af1f:34298316:cb855b50
Checksum : d1e422ca correct
MPB Sectors : 1
Disks : 2
RAID Devices : 1
Disk01 Serial : MSK5235H2PJ7TG
State : active
Id : 00000002
Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
[Volume0]:
UUID : 3d7bd72f:82a8cbcc:2d217397:12f3ff95
RAID Level : 1
Members : 2
Slots : [UU]
Failed disk : none
This Slot : 1
Array Size : 1953519616 (931.51 GiB 1000.20 GB)
Per Dev Size : 1953519880 (931.51 GiB 1000.20 GB)
Sector Offset : 0
Num Stripes : 7630936
Chunk Size : 64 KiB
Reserved : 0
Migrate State : idle
Map State : normal
Dirty State : dirty
Disk00 Serial : MSK5235H29X18G
State : active
Id : 00000001
Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
> From the program.log, I'm seeing sd[ab]5 are bcache partitions, and
> looks like they're outside of the imsm container, is that correct? I
> don't know whether this is a supportedlayout, my expectation when
> using firmware RAID is that the firmware (and later the md driver)
> asserts complete control over the entire block device. Any partitions
> are created within the container. If that's correct, the bcache
> partitions would need to be inside the imsm container as well. But in
> any case the installer shouldn't crash, it should inform. The version
> of mdadm in F21 is the same for three months, so if there's a
> regression here it's not recent.
The RAID device (md126) is composed of sdb and sdc. md126p5 is the
bcache backing device. sda is an (un-RAIDed) SSD; sda2 is the bcache
cache device.
As I just posted to the Bugzilla ...
I did some playing around with running "anaconda --askmethod" from a
live CD, and the results were ... interesting. At first, I got the
expected crash, so I edited /usr/lib/python2.7/site-packages/blivet/
devicelibs/mdraid.py and added some additional logging to both
name_from_md_node and md_node_from_name. Just to ensure that the
updated version was used, I also deleted the mdraid.pyc and mdraid.pyo.
Once I did this, the crash went away.
This suggests a couple of different possibilities to me:
1. There was something messed up about the pre-compiled files.
2. (Far more likely) some sort of race condition with udev. Partitions
on top of MD RAID seem to take a particularly long time for all of
the helper programs to run, so perhaps udev simply isn't finished
creating all the symlinks that anaconda needs (until I slow anaconda
down by adding the logging and/or removing the pre-compiled files).
--
========================================================================
Ian Pilcher arequipeno at gmail.com
-------- "I grew up before Mark Zuckerberg invented friendship" --------
========================================================================
More information about the test
mailing list