[LU-269] Ldisk-fs error (device md41): ldiskfs_check_descriptors: Block bitma p for group 0 not in group Created: 03/May/11  Updated: 28/Jun/11  Resolved: 09/May/11

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 1.8.6
Fix Version/s: None

Type: Bug Priority: Major
Reporter: Dan Ferber (Inactive) Assignee: Cliff White (Inactive)
Resolution: Fixed Votes: 0
Labels: None
Environment:

RHEL 5.5 and Lustre 1.8.0.1


Severity: 2
Rank (Obsolete): 10544

 Description   

MDS1/2 are running and active/active cluster configuration. mgs resides on MDS1 and mdt resides on MDS2. Is this an OK way to do things or not? In the original architecture of the system Oracle stated this was supported, but the customer just found out from Tyian over at Oracle that this is not a supported configuration for the MDS devices.

As of last Thursday customer could see all raid devices by the OS, but for some reason OST11 just simply would not become available. That issue went away with their "bare metal" reboot of the system on Friday morning. What started however, we have yet to fix:

OST15 /dev/md41 resident on OSS4
From /var/log/messages
Ldisk-fs error (device md41): ldiskfs_check_descriptors: Block bitma p for group 0 not in group

This device is not mounting anymore to the OSS at the Operating System level. The raid device can be constructed, but Lustre will not mount it.

Customer contact for questions is tyler.s.wiegers@lmco.com



 Comments   
Comment by Tyler Wiegers (Inactive) [ 03/May/11 ]

Environment should be RHEL 5.3, not 5.5.

We did a clean system startup, cleared messages, and systematically assembled the bitmaps, mounted them, assembled the raid devices, and mounted them for each OST. There were no errors for this OST until attempting to mount the raid device. All other OST's mounted successfully.

The unique data point for this OST is that it's raid device is missing a disk (md41). The disk was reported as unknown after our CAM/firmware upgrades yesterday so we replaced it, but we did not re-insert it into the raid. Would that situation cause the errors that we currently see?

The log output is the following:

oss3# mount -t lustre /dev/md41 /mnt/lustre_ost15
mount.lustre: mount /dev/md41 at /mnt/lustre_ost15 failed: Invalid argument
This may have multiple causes.
Are the mount options correct?
Check the syslog for more info.

/var/log/messages output (trimmed dates/times, I had to re-type this in from hard copy):

LDISKFS-fs error (device md41): ldiskf_check_descriptors: Block bitmap for group 0 not in group (block 134217728)!
LDISKFS-fs: group descriptors corrupted
LustreError: 26127:0:(obd_mount.c:1278:server_kernel_mount()) premount /dev/md41:0x0 ldiskfs failed: -22, ldiskfs2 failed: -19. Is the ldiskfs module available?
LustreError: 26127:0:(obd_mount.c:1278:server_kernel_mount()) Skipped 2 previous similar messages
LustreError: 26127:0:(obd_mount.c:1590:server_fill_super()) Unable to mount device /dev/md41: -22
LustreError: 26127:0:(obd_mount.c:1993:lustre_fill_super()) Unable to mount (-22)

Thanks!

Comment by Cliff White (Inactive) [ 03/May/11 ]

Well,a missing disk without a spare would definitely mess up the raid, i would think.

Was there data on the missing spindle? Has that been recovered?
First, you need to be sure the md41 device is intact and that the local (ldiskfs) filesystem
on that device is intact. If that requires re-inserting the disk, you should do so.

After the md side is healthy, you should run 'fsck -fn' on md41 and see what that reports.
Be sure you have the proper Lustre-aware version of e2fsprogs. (you should have downloaded it
along with the rest of your Lustre RPMS)

Assuming the md41 device is restored, 'fsck -fy' may fix the bitmap issue, but run '-fn' first
-that is a read-only pass to test for errors.

If there are other errors beyond the bitmap, you should attach the results here, but if you only
have the bitmap issue, proceed with the -fy

Comment by Cliff White (Inactive) [ 03/May/11 ]

Also, your first question was un-related - If the MGT and MDT are separate partitions, it okay to have one node active for MGS and the other active for MDS in a failover pair - the MGS is really, really lightweight, so after client mount the MGS node should be more or less idle.

Comment by Tyler Wiegers (Inactive) [ 03/May/11 ]

We inserted the disk back into the raid, it is currently rebuilding. Trying to mount the OST while the disk is rebuilding gives the same error. We've been able to mount OST's while disks are rebuilding in the past, so the core issue doesn't look like it's resolved.

No data should have been lost, we are running an 8+2 raid 6 device, so we can run with 8/10 disks without any data loss.

Comment by Cliff White (Inactive) [ 03/May/11 ]

Okay, after the rebuild, please run 'fsck -fn'

Comment by Johann Lombardi (Inactive) [ 04/May/11 ]

> We inserted the disk back into the raid, it is currently rebuilding. Trying to mount the OST
> while the disk is rebuilding gives the same error. We've been able to mount OST's while disks
> are rebuilding in the past, so the core issue doesn't look like it's resolved.

Based on the following comment:
http://jira.whamcloud.com/browse/LU-270?focusedCommentId=13656&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#action_13656

Could you please tell us what version of the mptsas driver you use?
I think the bug was fixed in versions >= 4.18.20.02.

> No data should have been lost, we are running an 8+2 raid 6 device, so we can run with 8/10 disks without any data loss.

Right.

Comment by Tyler Wiegers (Inactive) [ 04/May/11 ]

This disk finished rebuilding.

Once rebuilt, we attemped to run an e2fsck on the disks, which failed due to the MMP flag being set. We cleared the flag using tune2fs (ref LU-270) and tried the e2fsck again, which failed with another error. We ran tune2fs with the uninit-bg option, which allowed us to run e2fsck. The e2fsck ran for about 3.5 hours before completing.

Once the e2fsck was complete, we were able to successfully mount this OST, which is extremely good news. There were multiple recovered files in lost+found which we will be attempting to recover.

We are in the process of running e2fsck on all of our OSTs. Once complete we are planning a complete power down of all OSS's, MDS's, and disk arrays in order to do a fresh clean startup. I will update no later than tomorrow with hopefully a problem resolved statement.

Comment by Cliff White (Inactive) [ 05/May/11 ]

Great, thanks for keeping us updated
cliffw


cliffw
Support Guy
WhamCloud, Inc.
www.whamcloud.com

Comment by Peter Jones [ 09/May/11 ]

Rob Baker of LMCO has confirmed that the critical situation is over and production is stable. Residual issues will be tracked under a new ticket in the future.

Generated at Sat Feb 10 01:05:23 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.