<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 03:35:07 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-17397] mdtest failed (Lustre became read-only) under high stress</title>
                <link>https://jira.whamcloud.com/browse/LU-17397</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We test metadata performance in a simple Lustre environment, where we deploy two servers (#server01, #server02) and both connect to a SAN storage:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;For #server01: we mount a MGT, a MDT, and four OSTs&#160;&lt;/li&gt;
	&lt;li&gt;For #server02: we mount a MDT, and four OSTs&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Here, MDS and OSS run in the same server, and Lustre includes two MDTs and 8 OSTs.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@client02 lustre&amp;#93;&lt;/span&gt;# lfs df -h&lt;br/&gt;
UUID &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; &#160; bytes &#160; &#160; &#160; &#160;Used &#160; Available Use% Mounted on&lt;br/&gt;
l_lfs-MDT0000_UUID &#160; &#160; &#160; &#160; &#160;1.8T &#160; &#160; &#160; 39.2G &#160; &#160; &#160; &#160;1.6T &#160; 3% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;MDT:0&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-MDT0001_UUID &#160; &#160; &#160; &#160; &#160;1.8T &#160; &#160; &#160; 39.4G &#160; &#160; &#160; &#160;1.6T &#160; 3% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;MDT:1&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0000_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.5T &#160; &#160; &#160; &#160;7.8T &#160;31% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:0&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0001_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.6T &#160; &#160; &#160; &#160;7.7T &#160;32% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:1&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0002_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.6T &#160; &#160; &#160; &#160;7.7T &#160;32% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:2&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0003_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.6T &#160; &#160; &#160; &#160;7.7T &#160;32% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:3&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0004_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.8T &#160; &#160; &#160; &#160;7.5T &#160;34% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:4&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0005_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.5T &#160; &#160; &#160; &#160;7.8T &#160;32% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:5&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0006_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.5T &#160; &#160; &#160; &#160;7.8T &#160;31% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:6&amp;#93;&lt;/span&gt;&#160;&lt;br/&gt;
l_lfs-OST0007_UUID &#160; &#160; &#160; &#160; 11.9T &#160; &#160; &#160; &#160;3.6T &#160; &#160; &#160; &#160;7.7T &#160;32% /lustre&lt;span class=&quot;error&quot;&gt;&amp;#91;OST:7&amp;#93;&lt;/span&gt;&#160;&lt;/p&gt;

&lt;p&gt;filesystem_summary: &#160; &#160; &#160; &#160;95.1T &#160; &#160; &#160; 28.6T &#160; &#160; &#160; 61.8T &#160;32% /lustre&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;We leverage mdtest, mpirun with two clients to test metadate performance under the configuration above, the test command is as follows:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;$&amp;gt; mpirun --allow-run-as-root --oversubscribe -mca btl ^openib --mca btl_tcp_if_include 40.40.22.0/24 -np 64 -host client01:32,client02:32 --map-by node mdtest -L -z 3 -b 2 -I 160000 -i 1 -d /lustre/mdtest_demo | tee 2client_64np_3z_2b_160000I.log&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;After stably running around 15 mins, Lustre becomes read-only (blocks the whole test) and generate the sys log as follows:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:29:36 2024&amp;#93;&lt;/span&gt; Lustre: l_lfs-OST0001: deleting orphan objects from 0x440000400:26730785 to 0x440000400:26744321&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt19_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; Aborting journal on device ultrapatha-8.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs (ultrapatha): Remounting filesystem read-only&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): ldiskfs_journal_check_start:61: Detected aborted journal&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt20_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt20_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LustreError: 61165:0:(osd_handler.c:1790:osd_trans_commit_cb()) transaction @0x0000000082b2d9d3 commit error: 2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt08_003: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt21_000: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt20_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt08_003: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:19 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt05_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:20 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt21_000: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:20 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt18_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:24 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error: 355 callbacks suppressed&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:24 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt20_000: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:24 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt21_000: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:24 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt07_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:24 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt20_000: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:24 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt07_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:25 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt19_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:25 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt19_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:25 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt18_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:25 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt07_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Fri Jan &#160;5 17:43:25 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapatha): dx_probe:1169: inode #61343829: block 151386: comm mdt20_000: directory leaf block found instead of index block&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;We repeat the test many times, and still get the similar result (i.e., the LDISKFS-fs error in MDT0 or MDT1), and the workload scale is as follow:&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@client01 lustre&amp;#93;&lt;/span&gt;# lfs quota -u root /lustre/&lt;br/&gt;
Disk quotas for usr root (uid 0):&lt;br/&gt;
&#160; &#160; &#160;Filesystem &#160;kbytes &#160; quota &#160; limit &#160; grace &#160; files &#160; quota &#160; limit &#160; grace&lt;br/&gt;
&#160; &#160; &#160; &#160;/lustre/ 30805903960 &#160; &#160; &#160; 0 &#160; &#160; &#160; 0 &#160; &#160; &#160; - 96453928 &#160; &#160; &#160; 0 &#160; &#160; &#160; 0 &#160; &#160; &#160; -&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Originally, we find this issue with 2.15.0 and we try to upgrade to 2.15.3, but this issue still exists and block our test.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</description>
                <environment>client/server: CentOS-8.5.2111 + Lustre 2.15.3&lt;br/&gt;
Linux 4.18.0-348.2.1.el8_lustre.x86_64 #1 SMP Fri Jun 17 00:10:32 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux</environment>
        <key id="79852">LU-17397</key>
            <summary>mdtest failed (Lustre became read-only) under high stress</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="yzr95924">Zuoru Yang</reporter>
                        <labels>
                    </labels>
                <created>Sat, 6 Jan 2024 03:30:42 +0000</created>
                <updated>Wed, 17 Jan 2024 09:16:19 +0000</updated>
                            <resolved>Wed, 17 Jan 2024 08:42:33 +0000</resolved>
                                    <version>Lustre 2.15.0</version>
                    <version>Lustre 2.15.3</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="398756" author="adilger" created="Sat, 6 Jan 2024 09:04:08 +0000"  >&lt;p&gt;I haven&apos;t specifically checked if the issue is fixed in 2.15.4, but it was just released last week and may help.  It would only need to be installed on the server nodes.&lt;/p&gt;</comment>
                            <comment id="398757" author="adilger" created="Sat, 6 Jan 2024 09:09:46 +0000"  >&lt;p&gt;Just to confirm the test being run, each rank is creating 160000 files in a separate subdirectory from the other ranks, and there are 2^3 leaf subdirectories (branching factor 2, depth 3)?  That would create about 82M files, but it looks like there are some existing files in the filesystem.&lt;/p&gt;

&lt;p&gt;What does e2fsck show when run on the corrupt MDT?&lt;/p&gt;</comment>
                            <comment id="398962" author="JIRAUSER18721" created="Tue, 9 Jan 2024 13:17:56 +0000"  >&lt;p&gt;@Andreas Dilger Sorry for my late reply, we spend some time to check our RAID to ensure that this is not caused by the storage backend. We consider that it might be a bug in EXT4?&lt;/p&gt;

&lt;p&gt;Yes, there exists some files in the filesystem of previous experiments, and we remove them and try again with the same test command this time. The issue still occurs, and the info is as follows (since it does not create all files due to this issue):&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@client02 ~&amp;#93;&lt;/span&gt;# lfs quota -u root /lustre/&lt;br/&gt;
Disk quotas for usr root (uid 0):&lt;br/&gt;
&#160; &#160; &#160;Filesystem &#160;kbytes &#160; quota &#160; limit &#160; grace &#160; files &#160; quota &#160; limit &#160; grace&lt;br/&gt;
&#160; &#160; &#160; &#160;/lustre/ 185332124 &#160; &#160; &#160; 0 &#160; &#160; &#160; 0 &#160; &#160; &#160; - 45025839 &#160; &#160; &#160; 0 &#160; &#160; &#160; 0 &#160; &#160; &#160; -&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt05_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; Aborting journal on device ultrapathb-8.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs (ultrapathb): Remounting filesystem read-only&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): ldiskfs_journal_check_start:61: Detected aborted journal&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LustreError: 260307:0:(osd_handler.c:1790:osd_trans_commit_cb()) transaction @0x0000000069cf0f59 commit error: 2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LustreError: 260307:0:(osd_handler.c:1790:osd_trans_commit_cb()) Skipped 52 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt21_003: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt18_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt10_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt05_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt05_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt18_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt07_001: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt19_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:22 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt18_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error: 180 callbacks suppressed&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt20_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt05_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt18_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt18_002: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt10_003: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt20_004: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt07_003: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt19_000: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt10_003: directory leaf block found instead of index block&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;Tue Jan &#160;9 20:51:27 2024&amp;#93;&lt;/span&gt; LDISKFS-fs error (device ultrapathb): dx_probe:1169: inode #104316384: block 149479: comm mdt05_002: directory leaf block found instead of index block&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Note that device ultrapathb is the backend of MDT1, and the following is the process record when we do e2fsck for device ultrapathb&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Script started on 2024-01-09 21:09:49+08:00&lt;br/&gt;
&lt;sup&gt;[]0;root@server02:~^G&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server02 ~&amp;#93;&lt;/span&gt;# e2fsck -f /dev/ut^H&lt;/sup&gt;[[Klt^Grapathb^M&lt;br/&gt;
e2fsck 1.46.6-wc1 (10-Jan-2023)^M&lt;br/&gt;
MMP interval is 10 seconds and total wait time is 42 seconds. Please wait...^M&lt;br/&gt;
l_lfs-MDT0001: recovering journal^M&lt;br/&gt;
Pass 1: Checking inodes, blocks, and sizes^M&lt;br/&gt;
Pass 2: Checking directory structure^M&lt;br/&gt;
Pass 3: Checking directory connectivity^M&lt;br/&gt;
Pass 4: Checking reference counts^M&lt;br/&gt;
Pass 5: Checking group summary information^M&lt;br/&gt;
Free blocks count wrong (142139109, counted=142167445).^M&lt;br/&gt;
Fix&amp;lt;y&amp;gt;? yes^M&lt;br/&gt;
Free inodes count wrong (412219565, counted=412247109).^M&lt;br/&gt;
Fix&amp;lt;y&amp;gt;? yes^M&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;QUOTA WARNING&amp;#93;&lt;/span&gt; Usage inconsistent for ID 0:actual (72814075904, 17302001) != expected (72899485696, 17302001)^M&lt;br/&gt;
Update quota info for quota type 0&amp;lt;y&amp;gt;? yes^M&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;QUOTA WARNING&amp;#93;&lt;/span&gt; Usage inconsistent for ID 0:actual (72814075904, 17302001) != expected (72899485696, 17302001)^M&lt;br/&gt;
Update quota info for quota type 1&amp;lt;y&amp;gt;? yes^M&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;QUOTA WARNING&amp;#93;&lt;/span&gt; Usage inconsistent for ID 0:actual (72814075904, 17302001) != expected (72899485696, 17302001)^M&lt;br/&gt;
Update quota info for quota type 2&amp;lt;y&amp;gt;? yes^M&lt;br/&gt;
^M&lt;br/&gt;
l_lfs-MDT0001: ***** FILE SYSTEM WAS MODIFIED *****^M&lt;br/&gt;
l_lfs-MDT0001: 17302011/429549120 files (0.0% non-contiguous), 126261419/268428864 blocks^M&lt;br/&gt;
^[]0;root@server02:~^G&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server02 ~&amp;#93;&lt;/span&gt;# exit^M&lt;br/&gt;
exit^M&lt;/p&gt;

&lt;p&gt;Script done on 2024-01-09 21:16:33+08:00&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Is that possible an issue from ext4 with large_dir?&#160;&lt;/p&gt;</comment>
                            <comment id="399273" author="adilger" created="Thu, 11 Jan 2024 12:16:48 +0000"  >&lt;p&gt;Lustre does not modify the on-disk data structures of ldiskfs directly, although it is accessing the filesystem somewhat differently than a regular ext4 mount does.  I don&apos;t think the issue is with large_dir, but more likely with parallel directory locking and updates.  There would need to be some kind of bug in ext4 or the ldiskfs patches applied.  It is not possible for the clients to corrupt the server filesystem directly.&lt;/p&gt;

&lt;p&gt;That said, it appears from the e2fsck output that the on-disk data structures are not corrupted, so it seems like this is some kind of in-memory corruption?  The free blocks/inode counts quota usage messages are normal for a filesystem that is in use.&lt;/p&gt;

&lt;p&gt;There is a tunable parameter to disable the parallel directory locking and updates with &quot;&lt;tt&gt;lctl set_param osd-ldiskfs.lustre-MDT&amp;#42;.pdo=0&lt;/tt&gt;&quot; on the MDS nodes.  Note, that this is never tested and potentially could have some issues, beyond being much slower, but it would be useful to test if this avoids the issue.&lt;/p&gt;</comment>
                            <comment id="399274" author="adilger" created="Thu, 11 Jan 2024 12:18:16 +0000"  >&lt;p&gt;Also, have you tried updating to a newer kernel?  It is possible that the ext4 in the kernel (and ldiskfs that is generated from this) has a bug that has since been fixed.&lt;/p&gt;</comment>
                            <comment id="399281" author="JIRAUSER18721" created="Thu, 11 Jan 2024 13:15:03 +0000"  >&lt;p&gt;@Andreas Dilger Thanks Andreas! We will follow this direction and try the same test with a newer kernel.&lt;/p&gt;</comment>
                            <comment id="399406" author="JIRAUSER18721" created="Fri, 12 Jan 2024 03:39:42 +0000"  >&lt;p&gt;@Andreas Dilger BTW, the reason why I initially consider this issue is related large_dir is this link &lt;a href=&quot;https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1933074&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1933074&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;which also reports &quot;directory leaf block found instead of index block&quot; when there are millions of files on ext4. Never mind, we will test this issue with a newer kernel (e.g., in AlmaLinux 8.8 + 2.15.3)&lt;/p&gt;</comment>
                            <comment id="399978" author="adilger" created="Wed, 17 Jan 2024 04:54:13 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=yzr95924&quot; class=&quot;user-hover&quot; rel=&quot;yzr95924&quot;&gt;yzr95924&lt;/a&gt;, thank you for your launchpad reference.  Indeed that bug looks like it could be related.  That patch is reported included in upstream kernel 5.14 and stable kernel 5.11, and fixing a bug originally in kernel 5.11 (but also backported to the RHEL kernel):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;commit 877ba3f729fd3d8ef0e29bc2a55e57cfa54b2e43
Author:     Theodore Ts&apos;o &amp;lt;tytso@mit.edu&amp;gt;
AuthorDate: Wed Aug 4 14:23:55 2021 -0400

    ext4: fix potential htree corruption when growing large_dir directories
    
    Commit b5776e7524af (&quot;ext4: fix potential htree index checksum
    corruption) removed a required restart when multiple levels of index
    nodes need to be split.  Fix this to avoid directory htree corruptions
    when using the large_dir feature.
    
    Cc: stable@kernel.org # v5.11
    Cc: Artem Blagodarenko &amp;lt;artem.blagodarenko@gmail.com&amp;gt;
    Fixes: b5776e7524af (&quot;ext4: fix potential htree index checksum corruption)
    Reported-by: Denis &amp;lt;denis@voxelsoft.com&amp;gt;
    Signed-off-by: Theodore Ts&apos;o &amp;lt;tytso@mit.edu&amp;gt;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I can confirm that the patch is applied in &lt;tt&gt;4.18.0-425.13.1.el8_7.x86_64&lt;/tt&gt; in &lt;tt&gt;fs/ext4/namei.c&lt;/tt&gt;:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
                        &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (err)
                                &lt;span class=&quot;code-keyword&quot;&gt;goto&lt;/span&gt; journal_error;
                        err = ext4_handle_dirty_dx_node(handle, dir,
                                                        frame-&amp;gt;bh);
                        &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (restart || err)
                                &lt;span class=&quot;code-keyword&quot;&gt;goto&lt;/span&gt; journal_error;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;but I&apos;m not sure whether it is applied in your kernel &lt;tt&gt;4.18.0-348.2.1.el8_lustre.x86_64&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="399990" author="JIRAUSER18721" created="Wed, 17 Jan 2024 07:56:53 +0000"  >&lt;p&gt;@Andreas Dilger, Hi Andreas, thanks for your insights. We double-checked the linux kernel in our env&#160; (actually, we install the kernel package from the Whamcloud with 2.15.0 repo (later upgrade Lustre server to 2.15.3): &lt;a href=&quot;https://downloads.whamcloud.com/public/lustre/lustre-2.15.0-ib/MOFED-5.6-1.0.3.3/el8.5.2111/server/RPMS/x86_64/),&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://downloads.whamcloud.com/public/lustre/lustre-2.15.0-ib/MOFED-5.6-1.0.3.3/el8.5.2111/server/RPMS/x86_64/),&lt;/a&gt; and we confirm that the kernel in the link does not have the patch.&#160;&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;image-wrap&quot; style=&quot;&quot;&gt;&lt;img src=&quot;https://jira.whamcloud.com/secure/attachment/52976/52976_image-2024-01-17-15-52-59-936.png&quot; style=&quot;border: 0px solid black&quot; /&gt;&lt;/span&gt;&lt;/p&gt;</comment>
                            <comment id="399994" author="adilger" created="Wed, 17 Jan 2024 08:42:33 +0000"  >&lt;p&gt;Time to upgrade your server kernel and rebuild in that case. &lt;/p&gt;</comment>
                            <comment id="399999" author="JIRAUSER18721" created="Wed, 17 Jan 2024 09:16:19 +0000"  >&lt;p&gt;@Andreas Dilger Sure, we have evaluated the same test case in AlmaLinux 8.8 + 2.15.3 with the new kernel (4.18.0-477.10.1.el8_lustre.x86_64), now the issue did not occur. Thanks again!&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="51593" name="image-2024-01-09-21-00-00-430.png" size="18130" author="yzr95924" created="Tue, 9 Jan 2024 13:00:01 +0000"/>
                            <attachment id="52976" name="image-2024-01-17-15-52-59-936.png" size="51606" author="yzr95924" created="Wed, 17 Jan 2024 07:53:01 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10040" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic</customfieldname>
                        <customfieldvalues>
                                        <label>ldiskfs</label>
            <label>metadata</label>
            <label>performance</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10030" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic/Theme</customfieldname>
                        <customfieldvalues>
                                        <label>ldiskfs</label>
            <label>lustre-2.15.0</label>
            <label>lustre-2.15.3</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i046hz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>