//pnn> info c1 Unit UnitType Status %Cmpl Stripe Size(GB) Cache AVerify IgnECC ------------------------------------------------------------------------------ u0 RAID-5 DEGRADED - 64K 1629.74 OFF OFF OFF Port Status Unit Size Blocks Serial --------------------------------------------------------------- p0 OK u0 233.76 GB 490234752 WD-WCANY1850307 p1 OK u0 233.76 GB 490234752 WD-WCANY1790824 p2 OK u0 233.76 GB 490234752 WD-WCANY1851579 p3 OK u0 233.76 GB 490234752 WD-WCANY1789766 p4 OK u0 233.76 GB 490234752 WD-WCANY1796905 p5 NOT-PRESENT - - - - p6 OK u0 233.76 GB 490234752 WD-WCANY1788952 p7 OK u0 233.76 GB 490234752 WD-WCANY1788819 //pnn> /c1 remove p5 Exporting port /c1/p5 ... Failed. (0x0B:0x002E): Port empty
Got the message above, which showed that p5 failed. Since it was already missing, according to the tw_cli software, the remove command failed. So, I put a different disk in (one that had been in another raid) and got the following:
//pnn> info c1 Unit UnitType Status %Cmpl Stripe Size(GB) Cache AVerify IgnECC ------------------------------------------------------------------------------ u0 RAID-5 DEGRADED - 64K 1629.74 OFF OFF OFF u1 RAID-5 INOPERABLE - 64K 1629.74 OFF OFF OFF Port Status Unit Size Blocks Serial --------------------------------------------------------------- p0 OK u0 233.76 GB 490234752 WD-WCANY1850307 p1 OK u0 233.76 GB 490234752 WD-WCANY1790824 p2 OK u0 233.76 GB 490234752 WD-WCANY1851579 p3 OK u0 233.76 GB 490234752 WD-WCANY1789766 p4 OK u0 233.76 GB 490234752 WD-WCANY1796905 p5 OK u1 233.76 GB 490234752 WD-WCANY1787889 p6 OK u0 233.76 GB 490234752 WD-WCANY1788952 p7 OK u0 233.76 GB 490234752 WD-WCANY1788819
Crazy. It now thought that p5 was part of another raid (u1) on this controller.
//pnn> info Ctl Model Ports Drives Units NotOpt RRate VRate BBU ------------------------------------------------------------------------ c0 9550SX-8LP 8 8 1 0 4 4 - c1 9550SX-8LP 8 7 2 2 4 4 -
Yep, it now shows two bad units on c1. I tried rescanning and even removing the disk and reformatting it, but it didn’t matter. The only solution was to delete that second unit, which is always nerve-wracking, for fear that I’ll delete the raid I wanted to keep. So after quadruple-checking the command, I ran:
//pnn> maint deleteunit c1 u1 Deleting unit c1/u1 ...Done. //pnn> info c1 Unit UnitType Status %Cmpl Stripe Size(GB) Cache AVerify IgnECC ------------------------------------------------------------------------------ u0 RAID-5 DEGRADED - 64K 1629.74 OFF OFF OFF Port Status Unit Size Blocks Serial --------------------------------------------------------------- p0 OK u0 233.76 GB 490234752 WD-WCANY1850307 p1 OK u0 233.76 GB 490234752 WD-WCANY1790824 p2 OK u0 233.76 GB 490234752 WD-WCANY1851579 p3 OK u0 233.76 GB 490234752 WD-WCANY1789766 p4 OK u0 233.76 GB 490234752 WD-WCANY1796905 p5 OK - 233.76 GB 490234752 WD-WCANY1787889 p6 OK u0 233.76 GB 490234752 WD-WCANY1788952 p7 OK u0 233.76 GB 490234752 WD-WCANY1788819 //pnn> /c1/u0 start rebuild disk=5 Sending rebuild start request to /c1/u0 on 1 disk(s) [5] ... Done.
That deleted the extra unit and let me start the rebuild on disk 5 of the degraded unit.