<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:37:15 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-10680] MDT becoming unresponsive in 2.10.3</title>
                <link>https://jira.whamcloud.com/browse/LU-10680</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;2.10 LTS continues to causing us some trouble, and this time it&apos;s the MDT that&#160;has been&#160;quite unstable lately. A typical problem looks&#160;to start with&#160;this:&#160;&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;[52968.460453] Lustre: 5897:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[53955.181780] Lustre: 8034:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[53955.182270] Lustre: 8034:0:(mdd_device.c:1597:mdd_changelog_clear()) Skipped 2 previous similar messages
[54269.697741] Lustre: 5911:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[54362.202025] Lustre: 7920:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[54362.202511] Lustre: 7920:0:(mdd_device.c:1597:mdd_changelog_clear()) Skipped 1 previous similar message
[54808.564380] perf: interrupt took too &lt;span class=&quot;code-object&quot;&gt;long&lt;/span&gt; (2579 &amp;gt; 2500), lowering kernel.perf_event_max_sample_rate to 77000
[56283.046677] Lustre: 7958:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[56283.047150] Lustre: 7958:0:(mdd_device.c:1597:mdd_changelog_clear()) Skipped 14 previous similar messages
[56867.725405] Lustre: 5397:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[60578.888437] Lustre: 8014:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[60579.796894] Lustre: 8078:0:(mdd_device.c:1597:mdd_changelog_clear()) oak-MDD0000: Failure to clear the changelog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; user 1: -22
[60579.797374] Lustre: 8078:0:(mdd_device.c:1597:mdd_changelog_clear()) Skipped 2 previous similar messages
[62145.590529] LNet: Service thread pid 157540 was inactive &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; 200.33s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; debugging purposes:
[62145.591250] Pid: 157540, comm: mdt_rdpg01_050
[62145.591486] 
               Call Trace:
[62145.591949]  [&amp;lt;ffffffff816a94e9&amp;gt;] schedule+0x29/0x70
[62145.592200]  [&amp;lt;ffffffff816a6ff9&amp;gt;] schedule_timeout+0x239/0x2c0
[62145.592444]  [&amp;lt;ffffffff810c9ef9&amp;gt;] ? select_task_rq_fair+0x549/0x700
[62145.592708]  [&amp;lt;ffffffff816a989d&amp;gt;] wait_for_completion+0xfd/0x140
[62145.592971]  [&amp;lt;ffffffff810c4810&amp;gt;] ? default_wake_function+0x0/0x20
[62145.593223]  [&amp;lt;ffffffff810b07ea&amp;gt;] kthread_create_on_node+0xaa/0x140
[62145.593472]  [&amp;lt;ffffffffc12f0ac0&amp;gt;] ? mdd_chlg_garbage_collect+0x0/0x640 [mdd]
[62145.593760]  [&amp;lt;ffffffffc091346a&amp;gt;] ? llog_add+0x7a/0x1a0 [obdclass]
[62145.594025]  [&amp;lt;ffffffff810ea8ca&amp;gt;] ? __getnstimeofday64+0x3a/0xd0
[62145.594272]  [&amp;lt;ffffffffc12f7ac4&amp;gt;] mdd_changelog_store+0x194/0x540 [mdd]
[62145.594537]  [&amp;lt;ffffffffc130532d&amp;gt;] mdd_changelog_data_store_by_fid+0xed/0x1a0 [mdd]
[62145.595020]  [&amp;lt;ffffffffc130644b&amp;gt;] mdd_changelog_data_store+0x15b/0x210 [mdd]
[62145.595267]  [&amp;lt;ffffffffc130779b&amp;gt;] mdd_close+0x25b/0xcf0 [mdd]
[62145.595540]  [&amp;lt;ffffffffc11c9073&amp;gt;] mdt_mfd_close+0x353/0x610 [mdt]
[62145.595790]  [&amp;lt;ffffffffc11ce651&amp;gt;] mdt_close_internal+0x121/0x220 [mdt]
[62145.596051]  [&amp;lt;ffffffffc11ce970&amp;gt;] mdt_close+0x220/0x780 [mdt]
[62145.596352]  [&amp;lt;ffffffffc0bebda5&amp;gt;] tgt_request_handle+0x925/0x1370 [ptlrpc]
[62145.596633]  [&amp;lt;ffffffffc0b94b16&amp;gt;] ptlrpc_server_handle_request+0x236/0xa90 [ptlrpc]
[62145.597138]  [&amp;lt;ffffffffc0b91148&amp;gt;] ? ptlrpc_wait_event+0x98/0x340 [ptlrpc]
[62145.597385]  [&amp;lt;ffffffff810c4822&amp;gt;] ? default_wake_function+0x12/0x20
[62145.597651]  [&amp;lt;ffffffff810ba588&amp;gt;] ? __wake_up_common+0x58/0x90
[62145.597927]  [&amp;lt;ffffffffc0b98252&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
[62145.598176]  [&amp;lt;ffffffff81029557&amp;gt;] ? __switch_to+0xd7/0x510
[62145.598420]  [&amp;lt;ffffffff816a8f00&amp;gt;] ? __schedule+0x2f0/0x8b0
[62145.598695]  [&amp;lt;ffffffffc0b977c0&amp;gt;] ? ptlrpc_main+0x0/0x1e40 [ptlrpc]
[62145.598955]  [&amp;lt;ffffffff810b098f&amp;gt;] kthread+0xcf/0xe0
[62145.604690]  [&amp;lt;ffffffff810b08c0&amp;gt;] ? kthread+0x0/0xe0
[62145.604932]  [&amp;lt;ffffffff816b4f58&amp;gt;] ret_from_fork+0x58/0x90
[62145.605175]  [&amp;lt;ffffffff810b08c0&amp;gt;] ? kthread+0x0/0xe0

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;... leading to the end of connectivity, but the server doesn&apos;t really crash. So I took a crash dump when the&#160;MDS&#160;was&#160;unresponsive, it is available at the following link:&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;vmcore-oak-md1-s2-2018-02-16-11:02:21.gz:&lt;/p&gt;

&lt;p&gt;[https://stanford.box.com/s/n20wj34xhkmmltiav427l3btdz9qjwvn&lt;/p&gt;

&lt;p&gt;]&lt;/p&gt;

&lt;p&gt;vmlinux-3.10.0-693.2.2.el7_lustre.pl1.x86_64.gz:&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://stanford.box.com/s/r6u6xjzmsgys2kzcq26562bqo6p45xcp&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://stanford.box.com/s/r6u6xjzmsgys2kzcq26562bqo6p45xcp&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;also attaching foreach bt and memory usage&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;

&lt;p&gt;Stephane&lt;/p&gt;</description>
                <environment>3.10.0-693.2.2.el7_lustre.pl1.x86_64</environment>
        <key id="50857">LU-10680</key>
            <summary>MDT becoming unresponsive in 2.10.3</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="sthiell">Stephane Thiell</reporter>
                        <labels>
                    </labels>
                <created>Sat, 17 Feb 2018 00:22:15 +0000</created>
                <updated>Tue, 23 Jan 2024 20:49:53 +0000</updated>
                            <resolved>Tue, 19 Jun 2018 20:08:14 +0000</resolved>
                                    <version>Lustre 2.10.3</version>
                                    <fixVersion>Lustre 2.12.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>10</watches>
                                                                            <comments>
                            <comment id="221238" author="pjones" created="Sat, 17 Feb 2018 12:57:53 +0000"  >&lt;p&gt;Thanks Stephane&lt;/p&gt;</comment>
                            <comment id="221248" author="bfaccini" created="Sun, 18 Feb 2018 00:45:01 +0000"  >&lt;p&gt;Stephane,&lt;br/&gt;
I wonder if according to the stack trace, in the infos you have provided, a thead is in the process to be started in order to do garbage collection (introduced by my patch for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7340&quot; title=&quot;ChangeLogs catalog full condition should be handled more gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7340&quot;&gt;&lt;del&gt;LU-7340&lt;/del&gt;&lt;/a&gt;) of space consumed by changelogs due to a user/consumer (cl1?) being idle since quite a long time ?&lt;/p&gt;

&lt;p&gt;But for some unknown reason kthreadd is not starting it, or has still not reported it did... When on the other hand, the mdd_changelog_clear() error msgs seem to indicate changelogs garbage collection may have already started!&lt;/p&gt;

&lt;p&gt;Thus, I would like to analyze the crash-dump available, but to do so I also need the lustre-debuginfo RPM from your Lustre version, can you also make it available ?&lt;/p&gt;</comment>
                            <comment id="221257" author="sthiell" created="Sun, 18 Feb 2018 06:03:48 +0000"  >&lt;p&gt;Hey Bruno,&lt;/p&gt;

&lt;p&gt;Thanks for checking. I noticed that function `mdd_chlg_garbage_collect()` in the &quot;inactivate thread&quot; stack trace too. But this is weird because it is very unlikely cl1 was actually idle, it&apos;s a robinhood consumer and if I check the robinhood log just a bit before the issue / crash dump, it&#160;seems to be running:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;&#160;2018/02/16 10:29:18 [7022/2] STATS | ==================== Dumping stats at 2018/02/16 10:29:18 =====================
2018/02/16 10:29:18 [7022/2] STATS | ======== General statistics =========
2018/02/16 10:29:18 [7022/2] STATS | Daemon start time: 2018/01/29 11:23:38
2018/02/16 10:29:18 [7022/2] STATS | Started modules: scan,log_reader,policy_run(modeguard)
2018/02/16 10:29:18 [7022/2] STATS | ======== FS scan statistics =========
2018/02/16 10:29:18 [7022/2] STATS | last scan  = 2018/01/30 00:24:56
2018/02/16 10:29:18 [7022/2] STATS | duration    = 13h 01min 18s (46878 s)
2018/02/16 10:29:18 [7022/2] STATS | status      = complete
2018/02/16 10:29:18 [7022/2] STATS | current scan interval = 22.0d
2018/02/16 10:29:18 [7022/2] STATS | ChangeLog reader #0:
2018/02/16 10:29:18 [7022/2] STATS |    fs_name    =   oak
2018/02/16 10:29:18 [7022/2] STATS |    mdt_name   =   MDT0000
2018/02/16 10:29:18 [7022/2] STATS |    reader_id  =   cl1
2018/02/16 10:29:18 [7022/2] STATS |    records read        = 176082077
2018/02/16 10:29:18 [7022/2] STATS |    interesting records = 119680156
2018/02/16 10:29:18 [7022/2] STATS |    suppressed records  = 56401921
2018/02/16 10:29:18 [7022/2] STATS |    records pending     = 999
2018/02/16 10:29:18 [7022/2] STATS |    last received            = 2018/02/16 10:29:17
2018/02/16 10:29:18 [7022/2] STATS |    last read record time    = 2018/02/16 10:29:17.671089
2018/02/16 10:29:18 [7022/2] STATS |    last read record id      = 3459537671
2018/02/16 10:29:18 [7022/2] STATS |    last pushed record id    = 3459535680
2018/02/16 10:29:18 [7022/2] STATS |    last committed record id = 3459535680
2018/02/16 10:29:18 [7022/2] STATS |    last cleared record id   = 3459535680
2018/02/16 10:29:18 [7022/2] STATS |    read speed               = 76977.50 record/sec (513.18 incl. idle time)
2018/02/16 10:29:18 [7022/2] STATS |    processing speed ratio   = 1.00
2018/02/16 10:29:18 [7022/2] STATS |    ChangeLog stats:
2018/02/16 10:29:18 [7022/2] STATS |    MARK: 18414, CREAT: 16458408, MKDIR: 3318418, HLINK: 35763559, SLINK: 120840, MKNOD: 726
2018/02/16 10:29:18 [7022/2] STATS |    UNLNK: 43112348, RMDIR: 3143745, RENME: 6424858, RNMTO: 0, OPEN: 0, CLOSE: 32387229
2018/02/16 10:29:18 [7022/2] STATS |    LYOUT: 118087, TRUNC: 3063071, SATTR: 29921326, XATTR: 0, HSM: 0, MTIME: 695102, CTIME: 1535946
2018/02/16 10:29:18 [7022/2] STATS |    ATIME: 0, MIGRT: 0


&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;I also just checked that we only have one changelog consumer, cl1:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;[root@oak-md1-s2 ~]# cat /proc/fs/lustre/mdd/oak-MDT0000/changelog_users 
current index: 3475865674
ID    index (idle seconds)
cl1   3475865473 (1)
[root@oak-md1-s2 ~]#
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;You should be able to download the&#160;debuginfo RPMs&#160;at the following links:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;lustre-debuginfo-2.10.3_RC1-1.el7.centos.x86_64.rpm: &lt;a href=&quot;https://stanford.box.com/s/e3l44651do73iz92aaftx3ayf4wge254&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://stanford.box.com/s/e3l44651do73iz92aaftx3ayf4wge254&lt;/a&gt;&lt;/li&gt;
	&lt;li&gt;kernel-debuginfo-3.10.0-693.2.2.el7_lustre.pl1.x86_64.rpm: &lt;a href=&quot;https://stanford.box.com/s/6zknljpgd64xpas3xfv9a3rghuhwwi7f&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://stanford.box.com/s/6zknljpgd64xpas3xfv9a3rghuhwwi7f&lt;/a&gt;&lt;/li&gt;
	&lt;li&gt;kernel-debuginfo-common-x86_64-3.10.0-693.2.2.el7_lustre.pl1.x86_64.rpm: &lt;a href=&quot;https://stanford.box.com/s/636nhh2zxxt13o4chlgxx9qyuvc8p5gg&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://stanford.box.com/s/636nhh2zxxt13o4chlgxx9qyuvc8p5gg&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Thanks!&lt;/p&gt;</comment>
                            <comment id="221260" author="bfaccini" created="Sun, 18 Feb 2018 19:09:53 +0000"  >&lt;p&gt;Well after doing some analysis on the crash-dump I feel puzzled by what I seem to have discovered.&lt;br/&gt;
It looks like the assembly generated code of mdd_changelog_store() routine, including my patch for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7340&quot; title=&quot;ChangeLogs catalog full condition should be handled more gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7340&quot;&gt;&lt;del&gt;LU-7340&lt;/del&gt;&lt;/a&gt;, is wrong :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Dump of assembler code for function mdd_changelog_store:
931     int mdd_changelog_store(const struct lu_env *env, struct mdd_device *mdd,
   0xffffffffc12f7972 &amp;lt;+66&amp;gt;:    mov    0x28(%rsi),%r14

932                             struct llog_changelog_rec *rec, struct thandle *th)
933     {
   0xffffffffc12f7930 &amp;lt;+0&amp;gt;:     nopl   0x0(%rax,%rax,1)
   0xffffffffc12f7935 &amp;lt;+5&amp;gt;:     push   %rbp
   0xffffffffc12f7936 &amp;lt;+6&amp;gt;:     mov    %rsp,%rbp
   0xffffffffc12f7939 &amp;lt;+9&amp;gt;:     push   %r15
   0xffffffffc12f793b &amp;lt;+11&amp;gt;:    push   %r14
   0xffffffffc12f793d &amp;lt;+13&amp;gt;:    push   %r13
   0xffffffffc12f793f &amp;lt;+15&amp;gt;:    mov    %rdx,%r13
   0xffffffffc12f7942 &amp;lt;+18&amp;gt;:    push   %r12
   0xffffffffc12f7944 &amp;lt;+20&amp;gt;:    push   %rbx
   0xffffffffc12f7945 &amp;lt;+21&amp;gt;:    mov    %rsi,%rbx
   0xffffffffc12f794f &amp;lt;+31&amp;gt;:    sub    $0x30,%rsp
   0xffffffffc12f7957 &amp;lt;+39&amp;gt;:    mov    %rcx,-0x58(%rbp)
   0xffffffffc12f795b &amp;lt;+43&amp;gt;:    mov    %gs:0x28,%rax
   0xffffffffc12f7964 &amp;lt;+52&amp;gt;:    mov    %rax,-0x30(%rbp)
   0xffffffffc12f7968 &amp;lt;+56&amp;gt;:    xor    %eax,%eax
   0xffffffffc12f796a &amp;lt;+58&amp;gt;:    mov    %rdi,-0x48(%rbp)

934             struct obd_device       *obd = mdd2obd_dev(mdd);
935             struct llog_ctxt        *ctxt;
936             struct thandle          *llog_th;
937             int                      rc;
938             bool                     run_gc_task;
939     
940             rec-&amp;gt;cr_hdr.lrh_len = llog_data_len(sizeof(*rec) +
   0xffffffffc12f79b2 &amp;lt;+130&amp;gt;:   mov    %eax,0x0(%r13)
........................
964     
965             /* time to recover some space ?? */
966             spin_lock(&amp;amp;mdd-&amp;gt;mdd_cl.mc_lock);
967             if (unlikely(mdd-&amp;gt;mdd_changelog_gc &amp;amp;&amp;amp; (ktime_get_real_seconds() -
   0xffffffffc12f7a8f &amp;lt;+351&amp;gt;:   mov    0xf8(%rbx),%eax
   0xffffffffc12f7a95 &amp;lt;+357&amp;gt;:   test   %eax,%eax
   0xffffffffc12f7a97 &amp;lt;+359&amp;gt;:   jne    0xffffffffc12f7dc0 &amp;lt;mdd_changelog_store+1168&amp;gt;
   0xffffffffc12f7dc0 &amp;lt;+1168&amp;gt;:  callq  0xffffffffc07fec90 &amp;lt;ktime_get_real_seconds&amp;gt;
   0xffffffffc12f7dc5 &amp;lt;+1173&amp;gt;:  mov    0x108(%rbx),%edx
   0xffffffffc12f7dcb &amp;lt;+1179&amp;gt;:  sub    0xf0(%rbx),%rax
   0xffffffffc12f7dd2 &amp;lt;+1186&amp;gt;:  cmp    %rdx,%rax
   0xffffffffc12f7dd5 &amp;lt;+1189&amp;gt;:  jle    0xffffffffc12f7a9d &amp;lt;mdd_changelog_store+365&amp;gt;
   0xffffffffc12f7ddb &amp;lt;+1195&amp;gt;:  cmpq   $0x0,0xe8(%rbx)
   0xffffffffc12f7de3 &amp;lt;+1203&amp;gt;:  jne    0xffffffffc12f7a9d &amp;lt;mdd_changelog_store+365&amp;gt;
   0xffffffffc12f7de9 &amp;lt;+1209&amp;gt;:  mov    0x30(%r15),%rdi
   0xffffffffc12f7ded &amp;lt;+1213&amp;gt;:  callq  0xffffffffc0919a20 &amp;lt;llog_cat_free_space&amp;gt;
   0xffffffffc12f7df2 &amp;lt;+1218&amp;gt;:  cmp    0x10c(%rbx),%eax
   0xffffffffc12f7df8 &amp;lt;+1224&amp;gt;:  ja     0xffffffffc12f7a9d &amp;lt;mdd_changelog_store+365&amp;gt;

968                 mdd-&amp;gt;mdd_cl.mc_gc_time &amp;gt; mdd-&amp;gt;mdd_changelog_min_gc_interval) &amp;amp;&amp;amp;
969                 mdd-&amp;gt;mdd_cl.mc_gc_task == NULL &amp;amp;&amp;amp;
970                 llog_cat_free_space(ctxt-&amp;gt;loc_handle) &amp;lt;=
971                                     mdd-&amp;gt;mdd_changelog_min_free_cat_entries)) {
972                     CWARN(&quot;%s: low on changelog_catalog free entries, starting &quot;
   0xffffffffc12f7dfe &amp;lt;+1230&amp;gt;:  lea    0x40(%r14),%rdx
   0xffffffffc12f7e02 &amp;lt;+1234&amp;gt;:  mov    $0xffffffffc1316ec8,%rsi
   0xffffffffc12f7e09 &amp;lt;+1241&amp;gt;:  mov    $0xffffffffc132ba80,%rdi
   0xffffffffc12f7e10 &amp;lt;+1248&amp;gt;:  xor    %eax,%eax
   0xffffffffc12f7e12 &amp;lt;+1250&amp;gt;:  movl   $0x4,0x33c74(%rip)        # 0xffffffffc132ba90 &amp;lt;msgdata.69570+16&amp;gt;
   0xffffffffc12f7e1c &amp;lt;+1260&amp;gt;:  movq   $0xffffffffc1316b28,0x33c59(%rip)        # 0xffffffffc132ba80 &amp;lt;msgdata.69570&amp;gt;
   0xffffffffc12f7e27 &amp;lt;+1271&amp;gt;:  movq   $0xffffffffc1312b70,0x33c56(%rip)        # 0xffffffffc132ba88 &amp;lt;msgdata.69570+8&amp;gt;
   0xffffffffc12f7e32 &amp;lt;+1282&amp;gt;:  movl   $0x3cd,0x33c58(%rip)        # 0xffffffffc132ba94 &amp;lt;msgdata.69570+20&amp;gt;
   0xffffffffc12f7e3c &amp;lt;+1292&amp;gt;:  movq   $0xffffffffc132ba70,0x33c59(%rip)        # 0xffffffffc132baa0 &amp;lt;msgdata.69570+32&amp;gt;
   0xffffffffc12f7e47 &amp;lt;+1303&amp;gt;:  movl   $0x400,0x33c47(%rip)        # 0xffffffffc132ba98 &amp;lt;msgdata.69570+24&amp;gt;
   0xffffffffc12f7e51 &amp;lt;+1313&amp;gt;:  callq  0xffffffffc0809b70 &amp;lt;libcfs_debug_msg&amp;gt;

973                           &quot;ChangeLog garbage collection thread\n&quot;, obd-&amp;gt;obd_name);
974     
975                     /* indicate further kthread run will occur outside right after
976                      * critical section
977                      */
978                     mdd-&amp;gt;mdd_cl.mc_gc_task = (struct task_struct *)(-1);
   0xffffffffc12f7e56 &amp;lt;+1318&amp;gt;:  movq   $0xffffffffffffffff,0xe8(%rbx)
   0xffffffffc12f7e61 &amp;lt;+1329&amp;gt;:  jmpq   0xffffffffc12f7a9d &amp;lt;mdd_changelog_store+365&amp;gt;
   0xffffffffc12f7e66:  nopw   %cs:0x0(%rax,%rax,1)

979                     run_gc_task = true;
980             }
981             spin_unlock(&amp;amp;mdd-&amp;gt;mdd_cl.mc_lock);
982             if (run_gc_task) {
983                     struct task_struct *gc_task;
984     
985                     gc_task = kthread_run(mdd_chlg_garbage_collect, mdd,
   0xffffffffc12f7aa7 &amp;lt;+375&amp;gt;:   xor    %eax,%eax
   0xffffffffc12f7aa9 &amp;lt;+377&amp;gt;:   mov    $0xffffffffc131502e,%rcx
   0xffffffffc12f7ab0 &amp;lt;+384&amp;gt;:   mov    $0xffffffff,%edx
   0xffffffffc12f7ab5 &amp;lt;+389&amp;gt;:   mov    %rbx,%rsi
   0xffffffffc12f7ab8 &amp;lt;+392&amp;gt;:   mov    $0xffffffffc12f0ac0,%rdi
   0xffffffffc12f7abf &amp;lt;+399&amp;gt;:   callq  0xffffffff810b0740 &amp;lt;kthread_create_on_node&amp;gt;
   0xffffffffc12f7ac4 &amp;lt;+404&amp;gt;:   cmp    $0xfffffffffffff000,%rax
   0xffffffffc12f7aca &amp;lt;+410&amp;gt;:   mov    %rax,%r12
   0xffffffffc12f7acd &amp;lt;+413&amp;gt;:   ja     0xffffffffc12f7d55 &amp;lt;mdd_changelog_store+1061&amp;gt;
   0xffffffffc12f7ad3 &amp;lt;+419&amp;gt;:   mov    %rax,%rdi
   0xffffffffc12f7ad6 &amp;lt;+422&amp;gt;:   callq  0xffffffff810c4750 &amp;lt;wake_up_process&amp;gt;
............................
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;where it appears that the compiler has completely avoided run_gc_task variable setting/testing.&lt;br/&gt;
This may have been caused y the fact run_gc_task has been mistakenly forgotten to be initialized, but then it is unclear why there has been no warning/error to be generated at build/compile time (even if we use -Wall option of gcc, but seems that -Wuninitialized specific check may be, silently?, omitted because no optimization is requested as per some forums threads are reporting?) and also why the generated assembly code do not strictly correspond to the source code.&lt;br/&gt;
The side effect of all of this is that an attempt to create a kthread occurs during each ChangeLog record creation (instead of only when garbage-collection conditions have been triggered). And the consequences seems to be aggravated due to the fact now, kthreads are not started in the context of the requestor but requests to do so are queued for &quot;kthreadd&quot; handling which can lead to much trouble if &quot;kthreadd&quot; suffers some issue causing its operations to stop/slow-down.&lt;/p&gt;

&lt;p&gt;And this seems to be somewhat the case in your crash-dump :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; bt 2
PID: 2      TASK: ffff880169448fd0  CPU: 6   COMMAND: &quot;kthreadd&quot;
 #0 [ffff88016946b500] __schedule at ffffffff816a8f65
 #1 [ffff88016946b568] schedule at ffffffff816a94e9
 #2 [ffff88016946b578] wait_transaction_locked at ffffffffc0177085 [jbd2]
 #3 [ffff88016946b5d0] add_transaction_credits at ffffffffc0177368 [jbd2]
 #4 [ffff88016946b630] start_this_handle at ffffffffc01775e1 [jbd2]
 #5 [ffff88016946b6c8] jbd2__journal_start at ffffffffc0177a93 [jbd2]
 #6 [ffff88016946b710] __ldiskfs_journal_start_sb at ffffffffc0f55069 [ldiskfs]
 #7 [ffff88016946b750] ldiskfs_acquire_dquot at ffffffffc0f8b753 [ldiskfs]
 #8 [ffff88016946b770] dqget at ffffffff81267814
 #9 [ffff88016946b7d0] dquot_get_dqblk at ffffffff812685f4
#10 [ffff88016946b7f0] osd_acct_index_lookup at ffffffffc10a58df [osd_ldiskfs]
#11 [ffff88016946b828] lquota_disk_read at ffffffffc1014214 [lquota]
#12 [ffff88016946b858] qsd_refresh_usage at ffffffffc101bbfa [lquota]
#13 [ffff88016946b890] qsd_op_adjust at ffffffffc102a881 [lquota]
#14 [ffff88016946b8d0] osd_object_delete at ffffffffc106ec50 [osd_ldiskfs]
#15 [ffff88016946b910] lu_object_free at ffffffffc095ce3d [obdclass]
#16 [ffff88016946b968] lu_site_purge_objects at ffffffffc095da9e [obdclass]
#17 [ffff88016946ba10] lu_cache_shrink at ffffffffc095e8e9 [obdclass]
#18 [ffff88016946ba60] shrink_slab at ffffffff81195413
#19 [ffff88016946bb00] zone_reclaim at ffffffff81198091
#20 [ffff88016946bba8] get_page_from_freelist at ffffffff8118c264
#21 [ffff88016946bcb8] __alloc_pages_nodemask at ffffffff8118caf6
#22 [ffff88016946bd68] copy_process at ffffffff8108510d
#23 [ffff88016946bdf8] do_fork at ffffffff81086a51
#24 [ffff88016946be70] kernel_thread at ffffffff81086d06
#25 [ffff88016946be80] kthreadd at ffffffff810b1341
#26 [ffff88016946bf50] ret_from_fork at ffffffff816b4f58
crash&amp;gt; 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;and even if it seems to move forward as its last activity timestamp is much recent (-1120s) than the one of thread 157540 still stuck with the same stack.&lt;/p&gt;

&lt;p&gt;So 1st thing I have to do asap is to push a fix to effectivelly initialize run_gc_task variable/boolean to false, and secondly, I will continue crash-dump analysis in order to understand the reason of &quot;kthreadd&quot; pseudo-hang/slow-operations.&lt;/p&gt;
</comment>
                            <comment id="221261" author="gerrit" created="Sun, 18 Feb 2018 19:23:37 +0000"  >&lt;p&gt;Faccini Bruno (bruno.faccini@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/31347&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/31347&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: fix run_gc_task uninitialized&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: df5102382a42a1d366fe38d9da47a1c20ac40dd8&lt;/p&gt;</comment>
                            <comment id="221276" author="sthiell" created="Mon, 19 Feb 2018 17:03:57 +0000"  >&lt;p&gt;Bruno,&lt;/p&gt;

&lt;p&gt;Wow,&#160;what an amazing troubleshooting!&#160;Strange that the uninitialized variable wasn&apos;t detected by&#160;-Wall nor&#160;QA/Maloo indeed... I just built new 2.10.3 RPMs with your patch added that I will deploy&#160;ASAP.&lt;/p&gt;

&lt;p&gt;Thanks much!&lt;/p&gt;</comment>
                            <comment id="221316" author="bfaccini" created="Tue, 20 Feb 2018 18:46:33 +0000"  >&lt;p&gt;Hmm, I think that I have finally understood what has caused the kthreadd thread to hang and then the whole MDT operations to also become blocked/slow.&lt;/p&gt;

&lt;p&gt;We have triggered the following situation/deadlock, where a lot of MDT service threads (mdt&lt;span class=&quot;error&quot;&gt;&amp;#91;0-9&amp;#93;&lt;/span&gt;&lt;b&gt;, mdt_rdpg&lt;/b&gt;) were stuck in uninterruptible state and waiting for kthreadd to spawn with the following kinds of stacks :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;PID: 13741  TASK: ffff88203a31dee0  CPU: 23  COMMAND: &quot;mdt_rdpg01_031&quot;
 #0 [ffff880fe83efb20] __schedule at ffffffff816a8f65
 #1 [ffff880fe83efb88] schedule at ffffffff816a94e9
 #2 [ffff880fe83efb98] schedule_timeout at ffffffff816a6ff9
 #3 [ffff880fe83efc40] wait_for_completion at ffffffff816a989d
 #4 [ffff880fe83efca0] kthread_create_on_node at ffffffff810b07ea
 #5 [ffff880fe83efd58] ptlrpc_start_thread at ffffffffc0b962d5 [ptlrpc]
 #6 [ffff880fe83efde8] ptlrpc_main at ffffffffc0b9803c [ptlrpc]
 #7 [ffff880fe83efec8] kthread at ffffffff810b098f
 #8 [ffff880fe83eff50] ret_from_fork at ffffffff816b4f58
.......................
PID: 157540  TASK: ffff88103a13bf40  CPU: 11  COMMAND: &quot;mdt_rdpg01_050&quot;
 #0 [ffff881cb4b578d0] __schedule at ffffffff816a8f65
 #1 [ffff881cb4b57938] schedule at ffffffff816a94e9
 #2 [ffff881cb4b57948] schedule_timeout at ffffffff816a6ff9
 #3 [ffff881cb4b579f0] wait_for_completion at ffffffff816a989d
 #4 [ffff881cb4b57a50] kthread_create_on_node at ffffffff810b07ea
 #5 [ffff881cb4b57b08] mdd_changelog_store at ffffffffc12f7ac4 [mdd]
 #6 [ffff881cb4b57b70] mdd_changelog_data_store_by_fid at ffffffffc130532d [mdd]
 #7 [ffff881cb4b57bc8] mdd_changelog_data_store at ffffffffc130644b [mdd]
 #8 [ffff881cb4b57be0] mdd_close at ffffffffc130779b [mdd]
 #9 [ffff881cb4b57c40] mdt_mfd_close at ffffffffc11c9073 [mdt]
#10 [ffff881cb4b57c90] mdt_close_internal at ffffffffc11ce651 [mdt]
#11 [ffff881cb4b57cc0] mdt_close at ffffffffc11ce970 [mdt]
#12 [ffff881cb4b57d00] tgt_request_handle at ffffffffc0bebda5 [ptlrpc]
#13 [ffff881cb4b57d48] ptlrpc_server_handle_request at ffffffffc0b94b16 [ptlrpc]
#14 [ffff881cb4b57de8] ptlrpc_main at ffffffffc0b98252 [ptlrpc]
#15 [ffff881cb4b57ec8] kthread at ffffffff810b098f
#16 [ffff881cb4b57f50] ret_from_fork at ffffffff816b4f58
.....................
PID: 193852  TASK: ffff88202234bf40  CPU: 3   COMMAND: &quot;mdt01_098&quot;
 #0 [ffff881aa61fb5e0] __schedule at ffffffff816a8f65
 #1 [ffff881aa61fb648] schedule at ffffffff816a94e9
 #2 [ffff881aa61fb658] schedule_timeout at ffffffff816a6ff9
 #3 [ffff881aa61fb700] wait_for_completion at ffffffff816a989d
 #4 [ffff881aa61fb760] kthread_create_on_node at ffffffff810b07ea
 #5 [ffff881aa61fb818] mdd_changelog_store at ffffffffc12f7ac4 [mdd]
 #6 [ffff881aa61fb880] mdd_changelog_ns_store at ffffffffc12f8096 [mdd]
 #7 [ffff881aa61fb918] mdd_create at ffffffffc12fb40a [mdd]
 #8 [ffff881aa61fba10] mdt_reint_open at ffffffffc11cc9bc [mdt]
 #9 [ffff881aa61fbb00] mdt_reint_rec at ffffffffc11c18a0 [mdt]
#10 [ffff881aa61fbb28] mdt_reint_internal at ffffffffc11a330b [mdt]
#11 [ffff881aa61fbb60] mdt_intent_reint at ffffffffc11a3832 [mdt]
#12 [ffff881aa61fbba0] mdt_intent_policy at ffffffffc11ae59e [mdt]
#13 [ffff881aa61fbbf8] ldlm_lock_enqueue at ffffffffc0b39277 [ptlrpc]
#14 [ffff881aa61fbc50] ldlm_handle_enqueue0 at ffffffffc0b62903 [ptlrpc]
#15 [ffff881aa61fbce0] tgt_enqueue at ffffffffc0be7ea2 [ptlrpc]
#16 [ffff881aa61fbd00] tgt_request_handle at ffffffffc0bebda5 [ptlrpc]
#17 [ffff881aa61fbd48] ptlrpc_server_handle_request at ffffffffc0b94b16 [ptlrpc]
#18 [ffff881aa61fbde8] ptlrpc_main at ffffffffc0b98252 [ptlrpc]
#19 [ffff881aa61fbec8] kthread at ffffffff810b098f
#20 [ffff881aa61fbf50] ret_from_fork at ffffffff816b4f58
...................
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;either trying to spawn new service threads, or due to previous described problem of having one changelog garbage collection thread being spawned for each new changelog record.&lt;/p&gt;

&lt;p&gt;But as I have already indicated, kthreadd is stuck with the following stack :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;PID: 2      TASK: ffff880169448fd0  CPU: 6   COMMAND: &quot;kthreadd&quot;
 #0 [ffff88016946b500] __schedule at ffffffff816a8f65
 #1 [ffff88016946b568] schedule at ffffffff816a94e9
 #2 [ffff88016946b578] wait_transaction_locked at ffffffffc0177085 [jbd2]
 #3 [ffff88016946b5d0] add_transaction_credits at ffffffffc0177368 [jbd2]
 #4 [ffff88016946b630] start_this_handle at ffffffffc01775e1 [jbd2]
 #5 [ffff88016946b6c8] jbd2__journal_start at ffffffffc0177a93 [jbd2]
 #6 [ffff88016946b710] __ldiskfs_journal_start_sb at ffffffffc0f55069 [ldiskfs]
 #7 [ffff88016946b750] ldiskfs_acquire_dquot at ffffffffc0f8b753 [ldiskfs]
 #8 [ffff88016946b770] dqget at ffffffff81267814
 #9 [ffff88016946b7d0] dquot_get_dqblk at ffffffff812685f4
#10 [ffff88016946b7f0] osd_acct_index_lookup at ffffffffc10a58df [osd_ldiskfs]
#11 [ffff88016946b828] lquota_disk_read at ffffffffc1014214 [lquota]
#12 [ffff88016946b858] qsd_refresh_usage at ffffffffc101bbfa [lquota]
#13 [ffff88016946b890] qsd_op_adjust at ffffffffc102a881 [lquota]
#14 [ffff88016946b8d0] osd_object_delete at ffffffffc106ec50 [osd_ldiskfs]
#15 [ffff88016946b910] lu_object_free at ffffffffc095ce3d [obdclass]
#16 [ffff88016946b968] lu_site_purge_objects at ffffffffc095da9e [obdclass]
#17 [ffff88016946ba10] lu_cache_shrink at ffffffffc095e8e9 [obdclass]
#18 [ffff88016946ba60] shrink_slab at ffffffff81195413
#19 [ffff88016946bb00] zone_reclaim at ffffffff81198091
#20 [ffff88016946bba8] get_page_from_freelist at ffffffff8118c264
#21 [ffff88016946bcb8] __alloc_pages_nodemask at ffffffff8118caf6
#22 [ffff88016946bd68] copy_process at ffffffff8108510d
#23 [ffff88016946bdf8] do_fork at ffffffff81086a51
#24 [ffff88016946be70] kernel_thread at ffffffff81086d06
#25 [ffff88016946be80] kthreadd at ffffffff810b1341
#26 [ffff88016946bf50] ret_from_fork at ffffffff816b4f58
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;as upon trying to launch a new thread, it has unfortunately triggered memory reclaim condition with the consequence of shrinkers to be called, including the one for the lu_cache in order to drain unreferenced objects, and doing so it tries to update quotas in quota slave file, but is being blocked waiting to grant journal credits because the current/running transaction has still not committed. And this is because the transaction is awaiting for a lot of its attached/started handles to complete (jbd2_journal_stop!), but this will never occur as they are almost all from the previous threads waiting for kthreadd. Hence the deadlock.&lt;/p&gt;

&lt;p&gt;I think that to fix and prevent such unexpected scenario to re-occur, I need to find a way to start changelog garbage-collection thread outside of the interval/section-of-code where a journal transaction is being filled.&lt;/p&gt;

&lt;p&gt;I will push a new/2nd patch soon to implement this.&lt;/p&gt;


</comment>
                            <comment id="221357" author="bfaccini" created="Wed, 21 Feb 2018 10:45:49 +0000"  >&lt;p&gt;Thinking more about it, it seems to me that khreadd should be excluded to run the memory reclaim mechanism in its context, in this case it would be better to do nothing (and leave next thread that will try to allocate some memory to trigger the same conditions/threshold and start to reclaim in its context) or to simply/only wake-up kswapd to do the job asynchronously ?&lt;/p&gt;

&lt;p&gt;But anyway, we need to protect against such possible cases, and only request for changelog garbage-collection thread creation, outside of any current journal transaction/handle.&lt;/p&gt;
</comment>
                            <comment id="221391" author="sthiell" created="Wed, 21 Feb 2018 19:39:02 +0000"  >&lt;p&gt;Thanks Bruno! Hope you&apos;ll find a way to fix this race condition.&lt;/p&gt;</comment>
                            <comment id="221457" author="gerrit" created="Thu, 22 Feb 2018 15:24:25 +0000"  >&lt;p&gt;Faccini Bruno (bruno.faccini@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/31376&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/31376&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: create gc thread when no current transaction&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 2985625ebbb404c01f88ada84427d5b4386fbfa4&lt;/p&gt;</comment>
                            <comment id="221751" author="gerrit" created="Tue, 27 Feb 2018 03:43:48 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/31347/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/31347/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: fix run_gc_task uninitialized&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 25cdbbb475b8a42e3a1f594898d2ac1406b466f6&lt;/p&gt;</comment>
                            <comment id="222115" author="bfaccini" created="Fri, 2 Mar 2018 08:08:08 +0000"  >&lt;p&gt;Hello Stephane, &lt;br/&gt;
Just one more comment I have in mind and to explain why you have been able to trigger the 2nd problem/deadlock.&lt;br/&gt;
As part of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10709&quot; title=&quot;OSS deadlock in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10709&quot;&gt;LU-10709&lt;/a&gt; I understand that you are running with vm.zone_reclaim_mode=0 also on MDS, so,  according to mm code in this area, this seem to be definitely the reason why you have triggered memory-reclaim, with shrinkers run, during khreadd create of new kthread, instead to grab new memory/pages from preferred/current node+zone free-list.&lt;/p&gt;

&lt;p&gt;Hopefully this specific case should be handled and fixed by my 2nd fix for this ticket.&lt;/p&gt;</comment>
                            <comment id="222165" author="sthiell" created="Fri, 2 Mar 2018 21:54:35 +0000"  >&lt;p&gt;Hey Bruno,&lt;/p&gt;

&lt;p&gt;Thanks for the update! And yes, this happened with vm.zone_reclaim_mode=0 on MDS.&lt;/p&gt;</comment>
                            <comment id="222612" author="gerrit" created="Tue, 6 Mar 2018 19:27:27 +0000"  >&lt;p&gt;John L. Hammond (john.hammond@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/31552&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/31552&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: disable changelog garbage collection by default&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 62e4519de3dcb588b88311950e8c6b7fefcca6a8&lt;/p&gt;</comment>
                            <comment id="222759" author="gerrit" created="Thu, 8 Mar 2018 01:57:56 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/31552/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/31552/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: disable changelog garbage collection by default&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 57719d0c8833b8a8a8b6516d3bd4f2ad2b226285&lt;/p&gt;</comment>
                            <comment id="222866" author="adilger" created="Thu, 8 Mar 2018 18:53:15 +0000"  >&lt;p&gt;Moving over to 2.12 now that the GC feature is no longer enabled by default, and the failing tests have been disabled.  The remaining patch &quot;&lt;tt&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: create gc thread when no current transaction&lt;/tt&gt;&quot; for this ticket can land when it is ready, but no longer blocks the 2.11 release.&lt;/p&gt;</comment>
                            <comment id="229536" author="gerrit" created="Thu, 14 Jun 2018 03:53:53 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/31376/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/31376/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; mdd: create gc thread when no current transaction&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 31fef6845e8be39ae8dfd242714a84001925b7ad&lt;/p&gt;</comment>
                            <comment id="229629" author="pjones" created="Tue, 19 Jun 2018 20:08:14 +0000"  >&lt;p&gt;Landed for 2.12&lt;/p&gt;</comment>
                            <comment id="400887" author="adilger" created="Tue, 23 Jan 2024 20:49:53 +0000"  >&lt;p&gt;&quot;Xing Huang &amp;lt;hxing@ddn.com&amp;gt;&quot; uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/c/fs/lustre-release/+/53773&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/c/fs/lustre-release/+/53773&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10680&quot; title=&quot;MDT becoming unresponsive in 2.10.3&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10680&quot;&gt;&lt;del&gt;LU-10680&lt;/del&gt;&lt;/a&gt; tests: fix interop for sanity test_160h&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_15&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 32f7dd91985cd43e6ad2ed9854f9bed114a0e640&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="32827">LU-7340</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="50979">LU-10734</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="57162">LU-12865</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="57176">LU-12871</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="29587" name="foreach_bt-oak-md1-s2-2018-02-16-11-02-21.txt" size="676940" author="sthiell" created="Sat, 17 Feb 2018 00:22:11 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzzsy7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>