Hi! I’ve been poking at this for over a week, trying to be as careful a possible and getting the info together. It’s in a Dell Precision 7820 using Intel ISM, and blinking it’s power light accordingly. I think I’m in a fairly good place after /dev/sdb went clicky. Maybe it’ll be just some mdadm.conf edits and an --assemble. It’s just a simple single partition, and it contains my /home, so I’m a little anxious.<\/p>\n
I’ll start with
\ncat /proc/mdstat
\nPersonalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
\nunused devices: <\/p>\n
and
\n/etc/mdadm.conf,<\/p>\n
# definitions of existing MD arrays\nARRAY metadata=imsm UUID=ad4ce666:0ad11124:6baef5eb:73c408df\nARRAY /dev/md/Volume0 container=ad4ce666:0ad11124:6baef5eb:73c408df member=0 UUID=b56693ad:788460f8:a8b93419:703b25c4\n<\/code><\/pre>\nThen I got the info from each drive with ‘mdadm --examine /dev/sd[a-d]’<\/p>\n
/dev/sdb is marked in the Intel ISM BIOS utility as a spare, and it shows up as<\/p>\n
/dev/sdb:\n Magic : Intel Raid ISM Cfg Sig.\n Version : 1.0.00\n Orig Family : 00000000\n Family : abe1d35a\n Generation : 00000001\n Creation Time : Unknown\n Attributes : All supported\n UUID : 00000000:00000000:00000000:00000000\n Checksum : 57c3a6b4 correct\n MPB Sectors : 1\n Disks : 1\n RAID Devices : 0\n\n Disk00 Serial : WD-WCAW32827194\n State : spare\n Id : 01000000\n Usable Size : 1953522958 (931.51 GiB 1000.20 GB)\n\n Disk Serial : WD-WCAW32827194\n State : spare\n Id : 01000000\n Usable Size : 1953522958 (931.51 GiB 1000.20 GB)\n<\/code><\/pre>\nThe other drives all match UUIDs and other info, and seem totally sane and healthy. For example,<\/p>\n
/dev/sda:\n Magic : Intel Raid ISM Cfg Sig.\n Version : 1.2.01\n Orig Family : 6453e91b\n Family : 6453e91b\n Generation : 0003c05c\n Creation Time : Unknown\n Attributes : All supported\n UUID : bb0620fd:8274f3f7:498f64a4:dacbd25f\n Checksum : 8e30356f correct\n MPB Sectors : 2\n Disks : 4\n RAID Devices : 1\n\n Disk00 Serial : WD-WCC3F3ARCLKS\n State : active\n Id : 00000000\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n\n[Volume0]:\n Subarray : 0\n UUID : 5a9f14bb:a252fd06:f08cf7cf:b920b29e\n RAID Level : 10\n Members : 4\n Slots : [__UU]\n Failed disk : 0\n This Slot : 0 (out-of-sync)\n Sector Size : 512\n Array Size : 3906994176 (1863.00 GiB 2000.38 GB)\n Per Dev Size : 1953499136 (931.50 GiB 1000.19 GB)\n Sector Offset : 0\n Num Stripes : 7630848\n Chunk Size : 64 KiB\n Reserved : 0\n Migrate State : idle\n Map State : failed\n Dirty State : clean\n RWH Policy : off\n Volume ID : 1\n\n Disk01 Serial : 57E07H1MS:0\n State : active\n Id : ffffffff\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n\n Disk02 Serial : WD-WCC6Y0RL73N1\n State : active\n Id : 00000002\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n\n Disk03 Serial : WD-WCC6Y0LPDK7T\n State : active\n Id : 00000003\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n<\/code><\/pre>\nblkid shows all 4 drives as TYPE=“isw_raid_member”
\nNone of the /dev/md* stuff exists.<\/p>\n
While getting this together, I did notice that /dev/sdb is 3G/s and the others are 6G/s, so I’ll find another drive and copy the partition table over like I did with this one.<\/p>\n
Here’s where I lose a sense of the order of operation. Do I somehow bring the 3 drives up in Raid 1, then get it to recognize the spare and let it rebuild itself? I know the /dev/md? probably needs to show up before anything useful is going to happen, I’ve tried using “mdasm --assemble --scan /dev/md126 /dev/sd[abcd]” (returns “/dev/sd? not identified in config file” for each drive). I have plenty of tabs open for stackexchange and serverfault with clues, but nothing that’s been close enough to be comfortable moving forward.<\/p>\n
It’ll be great to recover from this and learn how to deal with it in the future. I’ll be upgrading the drives to 3TB by the end of the year, so having a bit more experience should help with that.<\/p>\n
Safety netsand other assistance appreciated.<\/p>","upvoteCount":2,"answerCount":4,"datePublished":"2025-04-15T19:07:40.530Z","author":{"@type":"Person","name":"barkeaterj","url":"https://community.spiceworks.com/u/barkeaterj"},"suggestedAnswer":[{"@type":"Answer","text":"
Hi! I’ve been poking at this for over a week, trying to be as careful a possible and getting the info together. It’s in a Dell Precision 7820 using Intel ISM, and blinking it’s power light accordingly. I think I’m in a fairly good place after /dev/sdb went clicky. Maybe it’ll be just some mdadm.conf edits and an --assemble. It’s just a simple single partition, and it contains my /home, so I’m a little anxious.<\/p>\n
I’ll start with
\ncat /proc/mdstat
\nPersonalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
\nunused devices: <\/p>\n
and
\n/etc/mdadm.conf,<\/p>\n
# definitions of existing MD arrays\nARRAY metadata=imsm UUID=ad4ce666:0ad11124:6baef5eb:73c408df\nARRAY /dev/md/Volume0 container=ad4ce666:0ad11124:6baef5eb:73c408df member=0 UUID=b56693ad:788460f8:a8b93419:703b25c4\n<\/code><\/pre>\nThen I got the info from each drive with ‘mdadm --examine /dev/sd[a-d]’<\/p>\n
/dev/sdb is marked in the Intel ISM BIOS utility as a spare, and it shows up as<\/p>\n
/dev/sdb:\n Magic : Intel Raid ISM Cfg Sig.\n Version : 1.0.00\n Orig Family : 00000000\n Family : abe1d35a\n Generation : 00000001\n Creation Time : Unknown\n Attributes : All supported\n UUID : 00000000:00000000:00000000:00000000\n Checksum : 57c3a6b4 correct\n MPB Sectors : 1\n Disks : 1\n RAID Devices : 0\n\n Disk00 Serial : WD-WCAW32827194\n State : spare\n Id : 01000000\n Usable Size : 1953522958 (931.51 GiB 1000.20 GB)\n\n Disk Serial : WD-WCAW32827194\n State : spare\n Id : 01000000\n Usable Size : 1953522958 (931.51 GiB 1000.20 GB)\n<\/code><\/pre>\nThe other drives all match UUIDs and other info, and seem totally sane and healthy. For example,<\/p>\n
/dev/sda:\n Magic : Intel Raid ISM Cfg Sig.\n Version : 1.2.01\n Orig Family : 6453e91b\n Family : 6453e91b\n Generation : 0003c05c\n Creation Time : Unknown\n Attributes : All supported\n UUID : bb0620fd:8274f3f7:498f64a4:dacbd25f\n Checksum : 8e30356f correct\n MPB Sectors : 2\n Disks : 4\n RAID Devices : 1\n\n Disk00 Serial : WD-WCC3F3ARCLKS\n State : active\n Id : 00000000\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n\n[Volume0]:\n Subarray : 0\n UUID : 5a9f14bb:a252fd06:f08cf7cf:b920b29e\n RAID Level : 10\n Members : 4\n Slots : [__UU]\n Failed disk : 0\n This Slot : 0 (out-of-sync)\n Sector Size : 512\n Array Size : 3906994176 (1863.00 GiB 2000.38 GB)\n Per Dev Size : 1953499136 (931.50 GiB 1000.19 GB)\n Sector Offset : 0\n Num Stripes : 7630848\n Chunk Size : 64 KiB\n Reserved : 0\n Migrate State : idle\n Map State : failed\n Dirty State : clean\n RWH Policy : off\n Volume ID : 1\n\n Disk01 Serial : 57E07H1MS:0\n State : active\n Id : ffffffff\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n\n Disk02 Serial : WD-WCC6Y0RL73N1\n State : active\n Id : 00000002\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n\n Disk03 Serial : WD-WCC6Y0LPDK7T\n State : active\n Id : 00000003\n Usable Size : 1953514766 (931.51 GiB 1000.20 GB)\n<\/code><\/pre>\nblkid shows all 4 drives as TYPE=“isw_raid_member”
\nNone of the /dev/md* stuff exists.<\/p>\n
While getting this together, I did notice that /dev/sdb is 3G/s and the others are 6G/s, so I’ll find another drive and copy the partition table over like I did with this one.<\/p>\n
Here’s where I lose a sense of the order of operation. Do I somehow bring the 3 drives up in Raid 1, then get it to recognize the spare and let it rebuild itself? I know the /dev/md? probably needs to show up before anything useful is going to happen, I’ve tried using “mdasm --assemble --scan /dev/md126 /dev/sd[abcd]” (returns “/dev/sd? not identified in config file” for each drive). I have plenty of tabs open for stackexchange and serverfault with clues, but nothing that’s been close enough to be comfortable moving forward.<\/p>\n
It’ll be great to recover from this and learn how to deal with it in the future. I’ll be upgrading the drives to 3TB by the end of the year, so having a bit more experience should help with that.<\/p>\n
Safety netsand other assistance appreciated.<\/p>","upvoteCount":2,"datePublished":"2025-04-15T19:07:40.598Z","url":"https://community.spiceworks.com/t/mdadm-intel-ism-and-recovering-raid-10-array/1196771/1","author":{"@type":"Person","name":"barkeaterj","url":"https://community.spiceworks.com/u/barkeaterj"}},{"@type":"Answer","text":"
I did find an article called “Rebuilding software RAID without mdadm.conf” that makes sense with a good set. I went ahead and tried --assemble:<\/p>\n
mdadm --assemble --uuid=bb0620fd:8274f3f7:498f64a4:dacbd25f --verbose /dev/md0 /dev/sda /dev/sdc /dev/sdd\n\nmdadm: looking for devices for /dev/md0\nmdadm: /dev/sda is identified as a member of /dev/md0, slot 0.\nmdadm: /dev/sdc is identified as a member of /dev/md0, slot 2.\nmdadm: /dev/sdd is identified as a member of /dev/md0, slot 3.\nmdadm: added /dev/sdc to /dev/md0 as 2\nmdadm: added /dev/sdd to /dev/md0 as 3\nmdadm: added /dev/sda to /dev/md0 as 0\nmdadm: Container /dev/md0 has been assembled with 3 drives\n<\/code><\/pre>\nI noted that it hasn’t been started, just assembled.
\nand then checked the details for /dev/md0<\/p>\n
mdadm --detail /dev/md0\n/dev/md0:\n Version : imsm\n Raid Level : container\n Total Devices : 3\n\n Working Devices : 3\n\n\n UUID : bb0620fd:8274f3f7:498f64a4:dacbd25f\n Member Arrays :\n\n Number Major Minor RaidDevice\n\n - 8 32 - /dev/sdc\n - 8 0 - /dev/sda\n - 8 48 - /dev/sdd\n<\/code><\/pre>\nI’ll assume that’s normal when the array isn’t “running”. When I get a proper replacement drive and use sgdisk to copy the partition table to it, I’ll try to “mdadm --manage /dev/md0 --add /dev/sdb” and see what happens. I won’t run “mdadm --detail --scan >> /etc/mdadm/mdadm.conf” till the array is actually running.<\/p>","upvoteCount":0,"datePublished":"2025-04-16T17:41:59.877Z","url":"https://community.spiceworks.com/t/mdadm-intel-ism-and-recovering-raid-10-array/1196771/2","author":{"@type":"Person","name":"barkeaterj","url":"https://community.spiceworks.com/u/barkeaterj"}},{"@type":"Answer","text":"
Here are a few steps that could<\/em> help. These were pulled from the sources below.<\/p>\nLinux RAID Wiki<\/a>
\nIntel IMSM RAID Recovery – Blog<\/a>
\nArch Wiki on Software RAID<\/a><\/p>\nAssemble the Container<\/strong><\/em>
\nbash
\nmdadm assemble /dev/md0 /dev/sda /dev/sdc /dev/sdd<\/p>\nBring Up the Member Array (e.g., RAID10 volume)<\/strong><\/em>
\nbash
\nmdadm assemble run /dev/md126 /dev/sda /dev/sdc /dev/sdd<\/p>\nAdd a Replacement Disk<\/strong><\/em>
\nbash
\nsgdisk replicate=/dev/sdb /dev/sda
\nsgdisk randomizeguids /dev/sdb
\nmdadm manage /dev/md126 add /dev/sdb<\/p>\nMonitor Rebuild<\/strong><\/em>
\nbash
\nwatch cat /proc/mdstat<\/p>\nAvoid create unless starting fresh — it overwrites metadata
\nUse assemble to avoid data loss
\nAlways verify before adding or removing disks<\/p>\n
Verify State<\/strong><\/em><\/p>\nbash
\ncat /proc/mdstat
\nmdadm detail /dev/md126
\nmdadm examine /dev/sd[ad]<\/p>\n
/dev/md0: IMSM container
\n/dev/md126 (or similar): actual RAID volume (your data lives here)
\nMetadata type: isw_raid_member<\/p>\n
PostRecovery Steps<\/strong><\/em><\/p>\nUpdate mdadm.conf
\nbash
\nmdadm detail scan /etc/mdadm/mdadm.conf<\/p>\n
Rebuild initramfs (for boot arrays)
\nbash
\nupdateinitramfs u<\/p>","upvoteCount":0,"datePublished":"2025-04-23T16:04:32.928Z","url":"https://community.spiceworks.com/t/mdadm-intel-ism-and-recovering-raid-10-array/1196771/3","author":{"@type":"Person","name":"jack-intel","url":"https://community.spiceworks.com/u/jack-intel"}},{"@type":"Answer","text":"
Thanks for the references, Jack!<\/p>\n
I’m spending more time trying to edit my informative reply as to not get “new users can only post 2 links” (there are none) than researching the problem. I’ve been using ‘Intel virtual raid on cpu intel vroc for linux user guide’ pdf for lots of insight.<\/p>\n
One rescued line summarizes my next steps. “The “update-subarray” stuff looks like what I’ll need to research next to figure out [Volume0]'s Failed disk and Disk01’s metadata.”<\/p>\n
The block I pasted above that follows “The other drives all match UUIDs…” shows the result of
\n“mdadm -E / dev/ md0”<\/p>\n
The info that got lost was<\/p>\n
cat / proc/ mdstat<\/code>
\nPersonalities : [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]<\/code>
\nmd0 : inactive sda[2](S) sdd[1](S) sdc[0](S)<\/code>
\n 15603 blocks super external:imsm<\/code>
\nunused devices: <none><\/code><\/p>\nThree drives, all marked as spares. That’s when I checked the metadata with -E and found the metadata issues with Disk01, first seen in my OP.<\/p>","upvoteCount":0,"datePublished":"2025-04-30T01:27:35.600Z","url":"https://community.spiceworks.com/t/mdadm-intel-ism-and-recovering-raid-10-array/1196771/4","author":{"@type":"Person","name":"barkeaterj","url":"https://community.spiceworks.com/u/barkeaterj"}}]}}