<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:42:08 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4373] MDS corruption - e2fsck issues</title>
                <link>https://jira.whamcloud.com/browse/LU-4373</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;We had some corruption appear on our MDS (possibly as a result of testing an SSD caching solution - Virident) so we took it offline to run an e2fsck. The check ran out of memory (we have 256Gigs + 90G swap) but not before it managed to do something screwy with the quotas. We could at least mount the FS with the corruption before the e2fsck but now it just bombed out instantly with errors on the quota. I then disabled the ext4 quota feature (we don&apos;t use quotas) and tried to mount but it goes read-only instantly now (originally it took a while to hit the corrupt area of the disk).&lt;/p&gt;

&lt;p&gt;I will probably try e2fsck one more time adding another disk as swap (with ~500 million 4k inodes - how much do we need?) and then resort to reformatting (we can repopulate the fs in a week or so). I&apos;m attaching the relevant output in case there is anything useful here for you.&lt;/p&gt;

&lt;p&gt;First see the corruption:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Dec  5 02:55:49 bmds1 kernel: LDISKFS-fs error (device vgca0_vcache1): ldiskfs_xattr_block_get: inode 299912086: bad block 299969421
Dec  5 02:55:49 bmds1 kernel: Aborting journal on device sdb.
Dec  5 02:55:49 bmds1 kernel: LDISKFS-fs error (device vgca0_vcache1): ldiskfs_journal_start_sb: Detected aborted journal
Dec  5 02:55:49 bmds1 kernel: LDISKFS-fs (vgca0_vcache1): Remounting filesystem read-only
..
..
Dec  5 05:23:10 bmds1 kernel: VFS: cannot write quota structure on device vgca0_vcache1 (error -30). Quota may get out of sync!
Dec  5 05:23:16 bmds1 kernel: LDISKFS-fs (vgca0_vcache1): Quota write (off=99328, len=1024) cancelled because transaction is not started
Dec  5 05:23:16 bmds1 kernel: VFS: Can&apos;t insert quota data block (97) to free entry list.
..
..
Dec  5 06:39:49 bmds1 kernel: LDISKFS-fs warning (device vgca0_vcache1): dx_probe: Unrecognised inode hash code 32 &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; directory #244352420
Dec  5 06:39:49 bmds1 kernel: LDISKFS-fs warning (device vgca0_vcache1): dx_probe: Corrupt dir inode 244352420, running e2fsck is recommended.
Dec  5 06:39:49 bmds1 kernel: LDISKFS-fs warning (device vgca0_vcache1): dx_probe: Unrecognised inode hash code 32 &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; directory #244352420
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Remount without e2fsck and wait until:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Dec  5 18:53:49 bmds1 kernel: LDISKFS-fs error (device md0): mb_free_blocks: &lt;span class=&quot;code-object&quot;&gt;double&lt;/span&gt;-free of inode 0&apos;s block 857326879(bit 17695 in group 26163)
Dec  5 18:53:49 bmds1 kernel: Aborting journal on device sdb.
Dec  5 18:53:49 bmds1 kernel: LDISKFS-fs error (device md0): ldiskfs_journal_start_sb: Detected aborted journal
Dec  5 18:53:49 bmds1 kernel: LDISKFS-fs (md0): Remounting filesystem read-only
Dec  5 18:53:49 bmds1 kernel: LDISKFS-fs (md0): Remounting filesystem read-only
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;So I guess we better run e2fsck then:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;bmds1 /root # e2fsck -fy /dev/md0
e2fsck 1.42.7.wc1 (12-Apr-2013)
Pass 1: Checking inodes, blocks, and sizes
Deleted inode 857268317 has zero dtime.  Fix? yes

Deleted inode 2613636819 has zero dtime.  Fix? yes

Deleted inode 2961739541 has zero dtime.  Fix? yes


Running additional passes to resolve blocks claimed by more than one inode...
Pass 1B: Rescanning &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; multiply-claimed blocks
Multiply-claimed block(s) in inode 268975905: 269046865
Multiply-claimed block(s) in inode 268975912: 269046838
Multiply-claimed block(s) in inode 268998892: 269046838
Multiply-claimed block(s) in inode 268998896: 269046865
Pass 1C: Scanning directories &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; inodes with multiply-claimed blocks
Pass 1D: Reconciling multiply-claimed blocks
(There are 4 inodes containing multiply-claimed blocks.)

File /ROOT/ARCHIVE/dirvish/filers/nfs11/20131203/tree/user_data/FL/rdev_tars/photoshop (inode #268975905, mod time Wed Apr 24 15:31:27 2013) 
  has 1 multiply-claimed block(s), shared with 1 file(s):
        ... (inode #268998896, mod time Wed Jan  9 12:59:03 2013)
Clone multiply-claimed blocks? yes

File /ROOT/ARCHIVE/dirvish/filers/nfs11/20131203/tree/user_data/FL/rdev_tars/REF/paperwork (inode #268975912, mod time Wed Apr 24 15:31:26 2013) 
  has 1 multiply-claimed block(s), shared with 1 file(s):
        ... (inode #268998892, mod time Wed Jan  9 12:59:03 2013)
Clone multiply-claimed blocks? yes

File ... (inode #268998892, mod time Wed Jan  9 12:59:03 2013) 
  has 1 multiply-claimed block(s), shared with 1 file(s):
        /ROOT/ARCHIVE/dirvish/filers/nfs11/20131203/tree/user_data/FL/rdev_tars/REF/paperwork (inode #268975912, mod time Wed Apr 24 15:31:26 2013)
Multiply-claimed blocks already reassigned or cloned.

File ... (inode #268998896, mod time Wed Jan  9 12:59:03 2013) 
  has 1 multiply-claimed block(s), shared with 1 file(s):
        /ROOT/ARCHIVE/dirvish/filers/nfs11/20131203/tree/user_data/FL/rdev_tars/photoshop (inode #268975905, mod time Wed Apr 24 15:31:27 2013)
Multiply-claimed blocks already reassigned or cloned.

Pass 2: Checking directory structure
Error allocating icount structure: Memory allocation failed

bravo-MDT0000: ***** FILE SYSTEM WAS MODIFIED *****
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 0:actual (352486866944, 75720160) != expected (287050371072, 62379940)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 4078:actual (5853184, 1468) != expected (0, 1193)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 5255:actual (32276480, 7735) != expected (24416256, 5884)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 5305:actual (578752512, 2100806) != expected (3155456000, 3376042)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 4080:actual (56586240, 13815) != expected (55222272, 13750)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 6731:actual (241618944, 56584) != expected (263524352, 59180)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 3449:actual (3298160640, 679894) != expected (2682540032, 596461)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 4052:actual (1331068928, 273724) != expected (1242320896, 218991)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 4088:actual (246837248, 60468) != expected (27856896, 19173)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 5339:actual (1207443456, 254756) != expected (1176674304, 270763)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 6699:actual (140206080, 37540) != expected (10919936, 12748)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 3695:actual (68915200, 15943) != expected (56963072, 13213)
..
..
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 7349:actual (64790528, 19964) != expected (52305920, 17445)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 188:actual (20480, 5) != expected (16384, 4)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 1056:actual (15908864, 3640) != expected (15458304, 3552)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 615:actual (180224, 44) != expected (172032, 42)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 513:actual (364544, 89) != expected (32768, 8)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 999:actual (16240640, 3965) != expected (15622144, 3814)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 5001:actual (21745664, 5229) != expected (21049344, 5063)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 1003:actual (0, 0) != expected (36864, 9)
[QUOTA WARNING] Usage inconsistent &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; ID 462:actual (233472, 132) != expected (225280, 123)
Update quota info &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; quota type 1? yes

[ERROR] quotaio_tree.c:357:free_dqentry:: Quota structure has offset to other block (0) than it should (34).
e2fsck: aborted

bravo-MDT0000: ***** FILE SYSTEM WAS MODIFIED *****
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Try to mount now:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Dec  9 09:56:42 bmds1 kernel: LDISKFS-fs (md0): warning: mounting fs with errors, running e2fsck is recommended
Dec  9 09:56:42 bmds1 kernel: LDISKFS-fs (md0): Ignoring delalloc option - requested data journaling mode
Dec  9 09:56:45 bmds1 kernel: LDISKFS-fs (md0): recovery complete
Dec  9 09:56:45 bmds1 kernel: LDISKFS-fs (md0): Can&apos;t enable usage tracking on a filesystem with the QUOTA feature set
Dec  9 09:56:45 bmds1 kernel: LDISKFS-fs (md0): mount failed
Dec  9 09:56:45 bmds1 kernel: ------------[ cut here ]------------
Dec  9 09:56:45 bmds1 kernel: WARNING: at fs/proc/&lt;span class=&quot;code-keyword&quot;&gt;generic&lt;/span&gt;.c:847 remove_proc_entry+0x24f/0x260() (Tainted: P           ---------------   )
Dec  9 09:56:45 bmds1 kernel: Hardware name: PowerEdge R620
Dec  9 09:56:45 bmds1 kernel: remove_proc_entry: removing non-empty directory &lt;span class=&quot;code-quote&quot;&gt;&apos;ldiskfs/md0&apos;&lt;/span&gt;, leaking at least &lt;span class=&quot;code-quote&quot;&gt;&apos;prealloc_table&apos;&lt;/span&gt;
Dec  9 09:56:45 bmds1 kernel: Modules linked in: ldiskfs(U) jbd2 mptctl mptbase ipmi_devintf dell_rbu nfsd exportfs autofs4 nfs lockd fscache auth_rpcgss nfs_acl sunrpc bonding 8021q garp stp llc uinput ipv6 raid1 power_meter sg 
vgcinit(P)(U) vgcdebug(P)(U) shpchp bnx2x libcrc32c mdio dcdbas microcode sb_edac edac_core iTCO_wdt iTCO_vendor_support ext3 jbd mbcache sr_mod cdrom sd_mod crc_t10dif mpt2sas scsi_transport_sas raid_class ahci wmi megaraid_sas 
dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]
Dec  9 09:56:45 bmds1 kernel: Pid: 8092, comm: mount Tainted: P           ---------------    2.6.32-358.18.1.el6_lustre.x86_64 #1
Dec  9 09:56:45 bmds1 kernel: Call Trace:
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff8106e3e7&amp;gt;] ? warn_slowpath_common+0x87/0xc0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff8106e4d6&amp;gt;] ? warn_slowpath_fmt+0x46/0x50
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff811ef7bd&amp;gt;] ? xlate_proc_name+0x4d/0xd0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff811efb1f&amp;gt;] ? remove_proc_entry+0x24f/0x260
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff8116e625&amp;gt;] ? pcpu_free_area+0x165/0x1e0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff8116e755&amp;gt;] ? free_percpu+0xb5/0x140
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffffa049acec&amp;gt;] ? ldiskfs_fill_super+0x23c/0x2a10 [ldiskfs]
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff8118477e&amp;gt;] ? get_sb_bdev+0x18e/0x1d0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffffa049aab0&amp;gt;] ? ldiskfs_fill_super+0x0/0x2a10 [ldiskfs]
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffffa0495388&amp;gt;] ? ldiskfs_get_sb+0x18/0x20 [ldiskfs]
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff81183beb&amp;gt;] ? vfs_kern_mount+0x7b/0x1b0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff81183d92&amp;gt;] ? do_kern_mount+0x52/0x130
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff811a3f52&amp;gt;] ? do_mount+0x2d2/0x8d0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff81139ff4&amp;gt;] ? strndup_user+0x64/0xc0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff811a45e0&amp;gt;] ? sys_mount+0x90/0xe0
Dec  9 09:56:45 bmds1 kernel: [&amp;lt;ffffffff8100b072&amp;gt;] ? system_call_fastpath+0x16/0x1b
Dec  9 09:56:45 bmds1 kernel: ---[ end trace dfa074843fd85142 ]---
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Okay try without the quota feature set:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Dec  9 15:47:56 bmds1 kernel: LDISKFS-fs (md0): warning: mounting fs with errors, running e2fsck is recommended
Dec  9 15:47:56 bmds1 kernel: LDISKFS-fs (md0): Ignoring delalloc option - requested data journaling mode
Dec  9 15:47:59 bmds1 kernel: LDISKFS-fs (md0): recovery complete
Dec  9 15:47:59 bmds1 kernel: LDISKFS-fs (md0): mounted filesystem with journalled data mode. quota=off. Opts: 
Dec  9 15:48:30 bmds1 kernel: LNet: HW CPU cores: 32, npartitions: 4
Dec  9 15:48:30 bmds1 kernel: alg: No test &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; crc32 (crc32-table)
Dec  9 15:48:30 bmds1 kernel: alg: No test &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; adler32 (adler32-zlib)
Dec  9 15:48:30 bmds1 kernel: alg: No test &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; crc32 (crc32-pclmul)
Dec  9 15:48:34 bmds1 kernel: padlock: VIA PadLock Hash Engine not detected.
Dec  9 15:48:34 bmds1 modprobe: FATAL: Error inserting padlock_sha (/lib/modules/2.6.32-358.18.1.el6_lustre.x86_64/kernel/drivers/crypto/padlock-sha.ko): No such device
Dec  9 15:48:38 bmds1 kernel: Lustre: Lustre: Build Version: 2.4.1-RC2--PRISTINE-2.6.32-358.18.1.el6_lustre.x86_64
Dec  9 15:48:38 bmds1 kernel: LNet: Added LNI 10.21.22.50@tcp [8/256/0/180]
Dec  9 15:48:38 bmds1 kernel: LNet: Accept secure, port 988
Dec  9 15:48:39 bmds1 kernel: LDISKFS-fs (md0): barriers disabled
Dec  9 15:48:39 bmds1 kernel: LDISKFS-fs (md0): warning: mounting fs with errors, running e2fsck is recommended
Dec  9 15:48:39 bmds1 kernel: LDISKFS-fs (md0): Ignoring delalloc option - requested data journaling mode
Dec  9 15:48:42 bmds1 kernel: LDISKFS-fs (md0): mounted filesystem with journalled data mode. quota=off. Opts: 
Dec  9 15:48:43 bmds1 kernel: Lustre: bravo-MDT0000: used disk, loading
Dec  9 15:48:43 bmds1 kernel: LDISKFS-fs error (device md0): ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; group 37corrupted: 525 blocks free in bitmap, 524 - in gd
Dec  9 15:48:43 bmds1 kernel: 
Dec  9 15:48:43 bmds1 kernel: Aborting journal on device sdb.
Dec  9 15:48:43 bmds1 kernel: LDISKFS-fs (md0): Remounting filesystem read-only
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

</description>
                <environment></environment>
        <key id="22414">LU-4373</key>
            <summary>MDS corruption - e2fsck issues</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="niu">Niu Yawei</assignee>
                                    <reporter username="daire">Daire Byrne</reporter>
                        <labels>
                    </labels>
                <created>Tue, 10 Dec 2013 11:46:13 +0000</created>
                <updated>Mon, 28 Jul 2014 12:58:02 +0000</updated>
                            <resolved>Mon, 28 Jul 2014 12:58:02 +0000</resolved>
                                    <version>Lustre 2.4.1</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="73221" author="daire" created="Tue, 10 Dec 2013 19:38:29 +0000"  >&lt;p&gt;And another e2fsck run:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;bmds1 /root # e2fsck -fvy /dev/md0
e2fsck 1.42.7.wc1 (12-Apr-2013)
bravo-MDT0000: recovering journal
Pass 1: Checking inodes, blocks, and sizes
Pass 2: Checking directory structure
Invalid inode number &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; &lt;span class=&quot;code-quote&quot;&gt;&apos;.&apos;&lt;/span&gt; in directory inode 268998892.
Fix? yes

Entry &lt;span class=&quot;code-quote&quot;&gt;&apos;lineup&apos;&lt;/span&gt; in /ROOT/ARCHIVE/dirvish/filers/nfs11/20131203/tree/user_data/FL/rdev_tars/REF/??? (268998892) is a link to directory /ROOT/ARCHIVE/dirvish/filers/nfs11/20131203/tree/user_data/FL/rdev_tars/REF/paperwork/lineup (268975913).
Clear? yes

Invalid inode number &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; &lt;span class=&quot;code-quote&quot;&gt;&apos;.&apos;&lt;/span&gt; in directory inode 268998896.
Fix? yes

Entry &lt;span class=&quot;code-quote&quot;&gt;&apos;..&apos;&lt;/span&gt; in .../??? (268998889) has deleted/unused inode 266936598.  Clear? yes

Entry &lt;span class=&quot;code-quote&quot;&gt;&apos;prop-base&apos;&lt;/span&gt; in .../??? (268998889) has deleted/unused inode 269521254.  Clear? yes

Entry &lt;span class=&quot;code-quote&quot;&gt;&apos;props&apos;&lt;/span&gt; in .../??? (268998889) is a link to directory /ROOT/ARCHIVE/dirvish/servers/jenkins1/20131204/tree/usr/src/debug/kernel-2.6.32-279.5.2.el6/linux-2.6.32-279.5.2.el6.x86_64/include/drm (270036736).
Clear? yes

Entry &lt;span class=&quot;code-quote&quot;&gt;&apos;..&apos;&lt;/span&gt; in .../??? (902387214) has deleted/unused inode 882443083.  Clear? yes

ext2fs_get_mem: Cannot allocate memory
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="78333" author="pjones" created="Tue, 4 Mar 2014 14:31:21 +0000"  >&lt;p&gt;Niu&lt;/p&gt;

&lt;p&gt;Could you please look into this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="78434" author="niu" created="Wed, 5 Mar 2014 03:33:24 +0000"  >&lt;p&gt;Looks your memory (256 + 90)G is insufficient to hold all directory structures, you need to enable the &quot;scratch_files&quot; to make e2fsck stores the dir structures in a file instead of memory. (see man e2fsck.conf), hope that&apos;ll complete the fsck.&lt;/p&gt;</comment>
                            <comment id="78515" author="daire" created="Wed, 5 Mar 2014 19:32:31 +0000"  >&lt;p&gt;I believe I tried scratch_files too at the time but still ran out of memory (with a 900G drive?). I guess the question is roughly how much RAM do we need to fsck 300 million dirs/files? I&apos;d rather not find this out by trial and error (900Gigs... 4TB?). 4k X 300 million (dirs) =~ 1.2 TB?&lt;/p&gt;</comment>
                            <comment id="78556" author="niu" created="Thu, 6 Mar 2014 02:01:40 +0000"  >&lt;p&gt;I don&apos;t know exactly if it requires 1.2TB or more, I suggest you use the largest drive you have (4TB?), and monitor the disk usage while running fsck. (&apos;tt&apos; option prints the memory usage for each phase)&lt;/p&gt;</comment>
                            <comment id="82604" author="pjones" created="Mon, 28 Apr 2014 14:25:21 +0000"  >&lt;p&gt;Hi Daire&lt;/p&gt;

&lt;p&gt;Have you tried this procedure yet?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="82716" author="daire" created="Tue, 29 Apr 2014 02:03:41 +0000"  >&lt;p&gt;Peter,&lt;/p&gt;

&lt;p&gt;No I haven&apos;t had the chance yet but we still have the RAID and data so I still plan to do the work soon.&lt;/p&gt;</comment>
                            <comment id="82719" author="pjones" created="Tue, 29 Apr 2014 02:36:08 +0000"  >&lt;p&gt;ok Daire. Thanks for the update. Please keep us posted.&lt;/p&gt;</comment>
                            <comment id="89941" author="daire" created="Thu, 24 Jul 2014 13:39:43 +0000"  >&lt;p&gt;Hey, I finally got around to setting up some hardware (5TB disk for e2fsck scratch) to re-run the e2fsck on this corrupted RAID. I had to remove the external journal but now it looks like e2fsck isn&apos;t getting as far as it did before:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;[root@localhost ~]# e2fsck -tt -fvy /dev/md0
e2fsck 1.42.7.wc1 (12-Apr-2013)
Pass 1: Checking inodes, blocks, and sizes
Signal (11) SIGSEGV si_code=SEGV_MAPERR fault addr=0x7ff5bac5fffc
e2fsck[0x42dbcd]
/lib64/libc.so.6(+0x329a0)[0x7ff4b991d9a0]
/lib64/libc.so.6(memcpy+0x35)[0x7ff4b9974985]
/lib64/libext2fs.so.2(+0x2c046)[0x7ff4ba84e046]
/lib64/libext2fs.so.2(+0x2cc3e)[0x7ff4ba84ec3e]
/lib64/libext2fs.so.2(ext2fs_tdb_store+0x747)[0x7ff4ba8510d7]
/lib64/libext2fs.so.2(+0x20c79)[0x7ff4ba842c79]
/lib64/libext2fs.so.2(ext2fs_icount_store+0x46)[0x7ff4ba842d26]
e2fsck(e2fsck_pass1+0xb58)[0x415468]
e2fsck(e2fsck_run+0x4f)[0x40efff]
e2fsck(main+0xeac)[0x40d1cc]
/lib64/libc.so.6(__libc_start_main+0xfd)[0x7ff4b9909d1d]
e2fsck[0x409eb9]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="90023" author="niu" created="Fri, 25 Jul 2014 02:05:05 +0000"  >&lt;p&gt;It could be &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2627&quot; title=&quot;/bin/ls gets Input/output error&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2627&quot;&gt;&lt;del&gt;LU-2627&lt;/del&gt;&lt;/a&gt;, could you upgrade your e2fsprogs to the latest version? (1.42.9.wc1-7) Thanks.&lt;/p&gt;</comment>
                            <comment id="90143" author="daire" created="Mon, 28 Jul 2014 09:45:23 +0000"  >&lt;p&gt;I guess it may have been an issue with using e2fsck&apos;s scratch files option. It ran to completion using 4TB of swap space. We are happy that we can at least fsck our MDT in future with enough swap configured. We can close this ticket now.&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Pass 5: Memory used: 18014398508741340k/4240k (18014398507398050k/1343291k), time: 442.38/111.08/ 0.00
Pass 5: I/O read: 270MB, write: 1MB, rate: 0.61MB/s

bravo-MDT0000: ***** FILE SYSTEM WAS MODIFIED *****

   429147617 inodes used (13.32%, out of 3221225472)
         189 non-contiguous files (0.0%)
     4422586 non-contiguous directories (0.0%)
             # of inodes with ind/dind/tind blocks: 712347/4105/0
   833979888 blocks used (25.89%, out of 3221225472)
           0 bad blocks
           1 large file

   280344591 regular files
   146575289 directories
          21 character device files
          32 block device files
         123 fifos
  1959990622 links
     2223722 symbolic links (1652788 fast symbolic links)
        3830 sockets
------------
  2389138230 files
Memory used: 18014398508741340k/4240k (18014398507398050k/1343291k), time: 57393.48/10071.86/3111.58
I/O read: 2102044MB, write: 265MB, rate: 36.63MB/s
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="90152" author="pjones" created="Mon, 28 Jul 2014 12:58:02 +0000"  >&lt;p&gt;ok thanks Daire&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwb3b:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>11972</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>