<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:45:22 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4733] All mdt thread stuck in cfs_waitq_wait</title>
                <link>https://jira.whamcloud.com/browse/LU-4733</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;We are seeing all mdt threads on the MDS stuck in &quot;cfs_waitq_wait&quot;. At the same time, we have a lot of rpc request (15k/s).&lt;br/&gt;
Looking closely at the console and the &apos;bt&apos; from crash, we can see that those threads are coming from qos_statfs_update() where they block in l_wait_event and never wake up.&lt;/p&gt;

&lt;p&gt;What is strange is that cfs_time_beforeq_64(max_age, obd-&amp;gt;obd_osfs_age) should be true.&lt;/p&gt;

&lt;p&gt;This issue was hit 4 times during February.&lt;/p&gt;

&lt;p&gt;Please find attached the dmesg and &apos;foreach bt&apos; outputs.&lt;/p&gt;

&lt;p&gt;Sebastien.&lt;/p&gt;</description>
                <environment></environment>
        <key id="23521">LU-4733</key>
            <summary>All mdt thread stuck in cfs_waitq_wait</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="niu">Niu Yawei</assignee>
                                    <reporter username="sebastien.buisson">Sebastien Buisson</reporter>
                        <labels>
                    </labels>
                <created>Fri, 7 Mar 2014 14:20:29 +0000</created>
                <updated>Thu, 2 Oct 2014 14:43:29 +0000</updated>
                            <resolved>Thu, 2 Oct 2014 14:43:29 +0000</resolved>
                                    <version>Lustre 2.1.6</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>11</watches>
                                                                            <comments>
                            <comment id="78706" author="bfaccini" created="Fri, 7 Mar 2014 15:04:22 +0000"  >&lt;p&gt;Looks like a dup of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4335&quot; title=&quot;MDS hangs due to mdt thread hung/inactive&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4335&quot;&gt;&lt;del&gt;LU-4335&lt;/del&gt;&lt;/a&gt; ...&lt;/p&gt;</comment>
                            <comment id="78708" author="sebastien.buisson" created="Fri, 7 Mar 2014 15:11:00 +0000"  >&lt;p&gt;Hi Bruno,&lt;/p&gt;

&lt;p&gt;I was not so sure as problem described in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4335&quot; title=&quot;MDS hangs due to mdt thread hung/inactive&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4335&quot;&gt;&lt;del&gt;LU-4335&lt;/del&gt;&lt;/a&gt; seems to be related to network problems, whereas there is no trace of such here.&lt;/p&gt;</comment>
                            <comment id="78712" author="pjones" created="Fri, 7 Mar 2014 15:27:55 +0000"  >&lt;p&gt;Niu&lt;/p&gt;

&lt;p&gt;Could you please confirm whether this is a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4335&quot; title=&quot;MDS hangs due to mdt thread hung/inactive&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4335&quot;&gt;&lt;del&gt;LU-4335&lt;/del&gt;&lt;/a&gt; or not?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="78720" author="liang" created="Fri, 7 Mar 2014 16:09:26 +0000"  >&lt;p&gt;This issue should have been fixed on master, please see lov_update_set() in master, set-&amp;gt;set_completes are atomic which is safe.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;void lov_update_set(struct lov_request_set *set,
                    struct lov_request *req, int rc)
{
        req-&amp;gt;rq_complete = 1;
        req-&amp;gt;rq_rc = rc;

        cfs_atomic_inc(&amp;amp;set-&amp;gt;set_completes);
        if (rc == 0)
                cfs_atomic_inc(&amp;amp;set-&amp;gt;set_success);

        wake_up(&amp;amp;set-&amp;gt;set_waitq);
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;and please see 2.1.x,  set-&amp;gt;set_completes is just integer which is not safe (yes, increments may be lost under race), which means lov_set_finished() may never return true, and qos_statfs_done() will never be called if this happened&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;void lov_update_set(struct lov_request_set *set,
                    struct lov_request *req, int rc)
{
        req-&amp;gt;rq_complete = 1;
        req-&amp;gt;rq_rc = rc;

        set-&amp;gt;set_completes++;
        if (rc == 0)
                set-&amp;gt;set_success++;

        cfs_waitq_signal(&amp;amp;set-&amp;gt;set_waitq);
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="78825" author="niu" created="Sun, 9 Mar 2014 02:03:17 +0000"  >&lt;p&gt;Liang, 2.1 has only one ptlrpcd thread to process the statfs requests, so I think the ++ operation should be safe. Probably statfs can&apos;t finish on certain OST?&lt;/p&gt;

&lt;p&gt;Sebastien, were there any anbnormal messages on OSTs? and I see following messages on MDT:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;[655444.910448] Uhhuh. NMI received &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; unknown reason 3d on CPU 0.
[655444.916528] Do you have a strange power saving mode enabled?
[655444.922337] Kernel panic - not syncing: NMI: Not continuing
[655444.928060] Pid: 0, comm: swapper Not tainted 2.6.32-220.23.1.bl6.Bull.28.10.x86_64 #1
[655444.936173] Call Trace:
[655444.938768]  &amp;lt;NMI&amp;gt;  [&amp;lt;ffffffff814851a0&amp;gt;] ? panic+0x78/0x143
[655444.944518]  [&amp;lt;ffffffff81488ff0&amp;gt;] ? do_nmi+0x240/0x2b0
[655444.949801]  [&amp;lt;ffffffff81488830&amp;gt;] ? nmi+0x20/0x30
[655444.954655]  [&amp;lt;ffffffff812a9831&amp;gt;] ? intel_idle+0xb1/0x170
[655444.960200]  &amp;lt;&amp;lt;EOE&amp;gt;&amp;gt;  [&amp;lt;ffffffff813aa267&amp;gt;] ? cpuidle_idle_call+0xa7/0x140
[655444.967165]  [&amp;lt;ffffffff81001e06&amp;gt;] ? cpu_idle+0xb6/0x110
[655444.972536]  [&amp;lt;ffffffff8146ff2a&amp;gt;] ? rest_init+0x7a/0x80
[655444.977909]  [&amp;lt;ffffffff81adff8c&amp;gt;] ? start_kernel+0x481/0x48d
[655444.983721]  [&amp;lt;ffffffff81adf31e&amp;gt;] ? x86_64_start_reservations+0x125/0x129
[655444.990651]  [&amp;lt;ffffffff81adf41c&amp;gt;] ? x86_64_start_kernel+0xfa/0x109
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;This looks like a hardware error, could it cause system hang under load?&lt;/p&gt;</comment>
                            <comment id="78835" author="bfaccini" created="Sun, 9 Mar 2014 15:00:14 +0000"  >&lt;p&gt;Niu, my 2 cents this comes from a forced crash-dump attempt or a HA fencing via NMI ...&lt;/p&gt;</comment>
                            <comment id="78867" author="louveta" created="Mon, 10 Mar 2014 13:55:57 +0000"  >&lt;p&gt;The lustre version running at customer site is based on lustre 2.1.6 + the following list of patches&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3240&quot; title=&quot;The link count is not updated after the mkdir&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3240&quot;&gt;&lt;del&gt;LU-3240&lt;/del&gt;&lt;/a&gt; check for all attributes validity in revalidate&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3071&quot; title=&quot;kernel BUG at fs/jbd2/transaction.c:293 (jbd2_journal_start)&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3071&quot;&gt;&lt;del&gt;LU-3071&lt;/del&gt;&lt;/a&gt; call cfs_cap_raise() before journal start&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3929&quot; title=&quot;2.1.6-&amp;gt;2.4.1 rolling upgrade: lustre-MDT0000: recovery is timed out, evict stale exports&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3929&quot;&gt;&lt;del&gt;LU-3929&lt;/del&gt;&lt;/a&gt; deny LWP connection for system upgrade&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3941&quot; title=&quot;(lov_lock.c:671:lov_lock_enqueue()) ASSERTION( sublock-&amp;gt;cll_conflict == ((void *)0) ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3941&quot;&gt;&lt;del&gt;LU-3941&lt;/del&gt;&lt;/a&gt; only modify extent of sub-lock and preserve flags&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4298&quot; title=&quot;&amp;#39;lfs setstripe&amp;#39; with bad values creates an empty file that has no striping information&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4298&quot;&gt;&lt;del&gt;LU-4298&lt;/del&gt;&lt;/a&gt; do not create file with no striping info&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3848&quot; title=&quot;Compute node crashes due to error in Lustre : ASSERTION( me == md-&amp;gt;md_me ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3848&quot;&gt;&lt;del&gt;LU-3848&lt;/del&gt;&lt;/a&gt; fix memory corruption&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3311&quot; title=&quot;Interop 2.1.5&amp;lt;-&amp;gt;2.4 failure on test suite sanity test_27B: (llite_lib.c:1792:ll_update_inode()) ASSERTION( lu_fid_eq(&amp;amp;lli-&amp;gt;lli_fid, &amp;amp;body-&amp;gt;fid1) ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3311&quot;&gt;&lt;del&gt;LU-3311&lt;/del&gt;&lt;/a&gt; mdt: handle -ENOENT &amp;amp;&amp;amp; !MDS_OPEN_CREAT in reint open&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4363&quot; title=&quot;(llite_lib.c:1683:ll_update_inode()) ASSERTION( lu_fid_eq(&amp;amp;lli-&amp;gt;lli_fid, &amp;amp;body-&amp;gt;fid1) ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4363&quot;&gt;&lt;del&gt;LU-4363&lt;/del&gt;&lt;/a&gt; open: backport MDS_OPEN_BY_FID flag&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2665&quot; title=&quot;LBUG while unmounting client&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2665&quot;&gt;&lt;del&gt;LU-2665&lt;/del&gt;&lt;/a&gt; refine &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2665&quot; title=&quot;LBUG while unmounting client&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2665&quot;&gt;&lt;del&gt;LU-2665&lt;/del&gt;&lt;/a&gt; patch for POSIX compliance&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1306&quot; title=&quot;LBUG at (dlm_lock.c:213:ldlm_lock_add_to_lru_nolock()) ASSERTION(lock-&amp;gt;l_resource-&amp;gt;lr_type != LDLM_FLOCK failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1306&quot;&gt;&lt;del&gt;LU-1306&lt;/del&gt;&lt;/a&gt; Protect l_flags with locking to prevent race on signal reception&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2943&quot; title=&quot;LBUG mdt_reconstruct_open()) ASSERTION( (!(rc &amp;lt; 0) || (lustre_msg_get_transno(req-&amp;gt;rq_repmsg) == 0)) )&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2943&quot;&gt;&lt;del&gt;LU-2943&lt;/del&gt;&lt;/a&gt; LBUG in mdt_reconstruct_open() (patch set #3)&lt;/li&gt;
	&lt;li&gt;ORNL-22 general ptlrpcd threads pool support&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1144&quot; title=&quot;implement a NUMA aware ptlrpcd binding policy&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1144&quot;&gt;&lt;del&gt;LU-1144&lt;/del&gt;&lt;/a&gt; implement a NUMA aware ptlrpcd binding policy&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1110&quot; title=&quot;MDS Oops in osd_xattr_get() during file open by FID&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1110&quot;&gt;&lt;del&gt;LU-1110&lt;/del&gt;&lt;/a&gt; MDS Oops in osd_xattr_get() during file open by FID&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2613&quot; title=&quot;opening and closing file can generate &amp;#39;unreclaimable slab&amp;#39; space&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2613&quot;&gt;&lt;del&gt;LU-2613&lt;/del&gt;&lt;/a&gt; to much unreclaimable slab space&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2624&quot; title=&quot;Stop of ptlrpcd threads is long&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2624&quot;&gt;&lt;del&gt;LU-2624&lt;/del&gt;&lt;/a&gt; ptlrpc fix thread stop&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2683&quot; title=&quot;Client deadlock in cl_lock_mutex_get&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2683&quot;&gt;&lt;del&gt;LU-2683&lt;/del&gt;&lt;/a&gt; client deadlock in cl_lock_mutex_get&lt;/li&gt;
	&lt;li&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2665&quot; title=&quot;LBUG while unmounting client&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2665&quot;&gt;&lt;del&gt;LU-2665&lt;/del&gt;&lt;/a&gt; keep resend flocks&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;so:&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;multi ptlrpd is build in&lt;/li&gt;
	&lt;li&gt;while I haven&apos;t identify which patch went with the lov_update_set safe routine, we do have it.&lt;/li&gt;
	&lt;li&gt;the backtrace pointed by Niu is the result of an hardware NMI that was initiated to take the dump (so this is a managed hardware error, not a consequence nor an unexpected error).&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="78965" author="niu" created="Tue, 11 Mar 2014 03:15:58 +0000"  >&lt;p&gt;Alexandre, any error messages on OSTs?&lt;/p&gt;</comment>
                            <comment id="78978" author="patrick.valentin" created="Tue, 11 Mar 2014 08:59:54 +0000"  >&lt;p&gt;Niu,&lt;br/&gt;
To complete the previous note from Alexandre, multi ptlrpcd and lov_update_set() safe routine are coming from ORNL-22 patch, which is a Bull backport in lustre 2.1.6 of branch b2_2 ORNL-22 patch (commit 71350744).&lt;/p&gt;

&lt;p&gt;I have attached the logs provided by support team, corresponding to the two last crashes which occured on 2 and 3 of march.&lt;/p&gt;

&lt;p&gt;here under is a copy of the analyze they have done:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; struct lov_qos.lq_statfs_in_progress 0xffff881eff05a6b0
  lq_statfs_in_progress = 1

crash&amp;gt; struct lov_statfs_data 0xffff881e3b9a2b80
struct lov_statfs_data {
  lsd_oi = {
    oi_policy = {
    .....
    oi_flags = 5, &amp;lt;--- OBD_STATFS_PTLRPCD | OBD_STATFS_NODELAY

crash&amp;gt; p *ptlrpcds
$1 = {
  pd_size = 6352,
  pd_index = 0,
  pd_nthreads = 32,
  pd_thread_rcv = {
    pc_flags = 258,
    pc_lock = {
      raw_lock = {
        slock = 0
      }
    }
.....
crash&amp;gt; struct ptlrpc_request_set 0xffff881fa6c702c0
struct ptlrpc_request_set {
  set_refcount = {
    counter = 1
  },
  set_new_count = {
    counter = 0
  },
  set_remaining = {
    counter = 0
  }
......
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;It seems rpc have been send but not response is received.&lt;/p&gt;

&lt;p&gt;In cb_statfs_update conditions are not satisfied&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;        lov_update_set(lovreq-&amp;gt;rq_rqset, lovreq, rc);
        if (rc)
                GOTO(out, rc);
out:
        if (lovreq-&amp;gt;rq_rqset-&amp;gt;set_oi-&amp;gt;oi_flags &amp;amp; OBD_STATFS_PTLRPCD &amp;amp;&amp;amp;
            lov_set_finished(lovreq-&amp;gt;rq_rqset, 0)) {
               lov_statfs_interpret(NULL, lovreq-&amp;gt;rq_rqset,
                                    lovreq-&amp;gt;rq_rqset-&amp;gt;set_success !=
                                                  lovreq-&amp;gt;rq_rqset-&amp;gt;set_count);
               qos_statfs_done(lov);
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="79837" author="sebastien.buisson" created="Thu, 20 Mar 2014 12:35:43 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;Please find attached the error logs on the OSS.&lt;br/&gt;
We cannot see errors by the time the problem occurred (first occurrence is on March 2nd at 09:45am and second occurrence on March 3rd at 10:04pm).&lt;/p&gt;

&lt;p&gt;Sebastien.&lt;/p&gt;</comment>
                            <comment id="80061" author="niu" created="Mon, 24 Mar 2014 03:51:25 +0000"  >&lt;p&gt;I see some OSTs reported slow read in the log:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;bcluster210.log-20140309.gz:2014-03-02 13:09:40 Lustre: 24804:0:(filter_io.c:475:filter_preprw_read()) scratch3-OST002c: slow start_page_read 30s
bcluster210.log-20140309.gz:2014-03-02 13:09:40 Lustre: 24804:0:(filter_io.c:475:filter_preprw_read()) Skipped 30 previous similar messages
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Maybe the OST is too busy to reply the statfs in time?&lt;/p&gt;

&lt;p&gt;Will the MDT get out of stuck state after the statfs RPC timedout and the burst of create requests from client level off?&lt;/p&gt;</comment>
                            <comment id="80066" author="sebastien.buisson" created="Mon, 24 Mar 2014 09:35:34 +0000"  >&lt;p&gt;I do not have answers to your questions, but given their time of occurrence the &apos;slow&apos; messages seem to be completely independent of the problem when all mdt threads are stuck in &quot;cfs_waitq_wait&quot;, right?&lt;/p&gt;</comment>
                            <comment id="80200" author="niu" created="Tue, 25 Mar 2014 07:01:14 +0000"  >&lt;blockquote&gt;
&lt;p&gt;I do not have answers to your questions, but given their time of occurrence the &apos;slow&apos; messages seem to be completely independent of the problem when all mdt threads are stuck in &quot;cfs_waitq_wait&quot;, right?&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I don&apos;t know for sure.&lt;/p&gt;

&lt;p&gt;I just want to figure out if it&apos;s a deadlock(or missing wakeup) problem (if the threads recovered when system load getting lower, then it&apos;s unlikely a deadlock/missing wakeup problem).&lt;/p&gt;

&lt;p&gt;BTW: Is there anyway to check the code running on servers? (to make sure the all the patches were applied properly)&lt;/p&gt;</comment>
                            <comment id="80202" author="sebastien.buisson" created="Tue, 25 Mar 2014 07:25:01 +0000"  >&lt;p&gt;&amp;gt; BTW: Is there anyway to check the code running on servers? (to make sure the all the patches were applied properly)&lt;/p&gt;

&lt;p&gt;Sure, but what do you suggest? I mean, how can we proceed to &apos;check the code&apos;?&lt;/p&gt;</comment>
                            <comment id="80205" author="niu" created="Tue, 25 Mar 2014 07:56:50 +0000"  >&lt;p&gt;Could you upload the code to the whamcloud ftp server (just like update logs)? Or show me where I can checkout the code? Thanks.&lt;/p&gt;</comment>
                            <comment id="80360" author="sebastien.buisson" created="Thu, 27 Mar 2014 10:03:39 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;On Whamcloud&apos;s FTP (ftp.whamcloud.com), I have created a directory &quot;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4733&quot; title=&quot;All mdt thread stuck in cfs_waitq_wait&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4733&quot;&gt;&lt;del&gt;LU-4733&lt;/del&gt;&lt;/a&gt;&quot; containing a file lustre_216_bull_sources.tgz.&lt;/p&gt;

&lt;p&gt;Sebastien.&lt;/p&gt;</comment>
                            <comment id="80808" author="niu" created="Wed, 2 Apr 2014 06:35:21 +0000"  >&lt;p&gt;Thank you, Sebastien. The code looks correct to me.&lt;/p&gt;

&lt;p&gt;It&apos;s possible that connection between MDT to some OST is broken, so the statfs can&apos;t be updated, and since there were many ongoing create requests, lq_statfs_in_progress could be changed to 1 immediately after a statfs done, so the waiters will go back to wait on qos_statfs_ready().&lt;/p&gt;

&lt;p&gt;When the problem raise again, could you try to collect lustre log on MDT with D_QOS &amp;amp; D_RPCTRACE enabled? Thank you.&lt;/p&gt;</comment>
                            <comment id="81262" author="patrick.valentin" created="Wed, 9 Apr 2014 12:23:29 +0000"  >&lt;p&gt;Hi,&lt;br/&gt;
two new occurences of this issue were detected at TGCC on april 7 and 9.&lt;/p&gt;

&lt;p&gt;Several things were made :&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;list of clients disconnected was made to attempt to access the filesystem, with a df/ls/touch. All attempts have failed.&lt;/li&gt;
	&lt;li&gt;umount of device on MDS failed, the process umount stuck on cfs_waitq_wait&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;The attached file &quot;lctl-dk.tgz&quot; contains the following traces:&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;mantis518_dk.log  (collected with an echo -1 &amp;gt; /proc/sys/lnet/debug)&lt;/li&gt;
	&lt;li&gt;mantis518_rpc.log (collected with an echo +other +rpctrace &amp;gt; /proc/sys/lnet/debug)&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Patrick&lt;/p&gt;</comment>
                            <comment id="82035" author="niu" created="Mon, 21 Apr 2014 02:50:45 +0000"  >&lt;blockquote&gt;
&lt;p&gt;Several things were made :&lt;br/&gt;
list of clients disconnected was made to attempt to access the filesystem, with a df/ls/touch. All attempts have failed.&lt;br/&gt;
umount of device on MDS failed, the process umount stuck on cfs_waitq_wait&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Could you provide the stack trace for all processes on MDS when umount the device?&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;00000100:00000400:2.0F:1396898402.999140:0:12693:0:(service.c:1035:ptlrpc_at_send_early_reply()) @@@ Couldn&apos;t add any time (5/-150), not sending early reply
  req@ffff880f02f9b800 x1463563674626235/t0(0) o400-&amp;gt;4b5b918b-511a-1159-a7ee-db25715da035@JO.BOO.PA.BLW@o2ib2:0/0 lens 192/0 e 0 to 0 dl 1396898407 ref 2 fl New:H/0/ffffffff rc 0/-1
00000100:00000400:2.0:1396898402.999154:0:12693:0:(service.c:1035:ptlrpc_at_send_early_reply()) @@@ Couldn&apos;t add any time (5/-150), not sending early reply
  req@ffff880f02f9bc00 x1463563600312609/t0(0) o400-&amp;gt;23a5a711-176e-4fea-3690-e6e8f81b7ba3@JO.BOO.IW.WW@o2ib2:0/0 lens 192/0 e 0 to 0 dl 1396898407 ref 2 fl New:H/0/ffffffff rc 0/-1
00000100:00000400:2.0:1396898402.999165:0:12693:0:(service.c:1035:ptlrpc_at_send_early_reply()) @@@ Couldn&apos;t add any time (5/-150), not sending early reply
  req@ffff88091d277000 x1463563647245622/t0(0) o400-&amp;gt;e081ff72-cc00-6425-c96f-b093a974f1a4@JO.BOO.WB.AW@o2ib2:0/0 lens 192/0 e 0 to 0 dl 1396898407 ref 2 fl New:H/0/ffffffff rc 0/-1
00000100:00000400:2.0:1396898402.999176:0:12693:0:(service.c:1035:ptlrpc_at_send_early_reply()) @@@ Couldn&apos;t add any time (5/-150), not sending early reply
  req@ffff88091d277400 x1463563646360851/t0(0) o400-&amp;gt;822cdf71-bac5-e141-3e77-a03c858db938@JO.BOO.PA.BBA@o2ib2:0/0 lens 192/0 e 0 to 0 dl 1396898407 ref 2 fl New:H/0/ffffffff rc 0/-1

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;JO.BOO.PA.BLW@o2ib2 is one of the OST? Looks the connections between MDT and some OSTs are not stable. I tend to think this is same issue of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4335&quot; title=&quot;MDS hangs due to mdt thread hung/inactive&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4335&quot;&gt;&lt;del&gt;LU-4335&lt;/del&gt;&lt;/a&gt; &amp;amp; &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4195&quot; title=&quot;MDT Slow with ptlrpcd using 100% cpu.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4195&quot;&gt;&lt;del&gt;LU-4195&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="82456" author="patrick.valentin" created="Fri, 25 Apr 2014 09:31:02 +0000"  >&lt;p&gt;Hi,&lt;br/&gt;
I have attached the stack trace of all processes (2014/04/07) provided by on site support team.&lt;br/&gt;
They also mentionned that JO.BOO.PA.xxx addresses in lctl-dk_rpc file are client&apos;s (compute nodes) adresses, NOT OSS.&lt;/p&gt;</comment>
                            <comment id="86774" author="niu" created="Tue, 17 Jun 2014 02:54:21 +0000"  >&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;bcluster231.log-20140309.gz:2014-03-03 23:18:53 LustreError: 25883:0:(o2iblnd_cb.c:2992:kiblnd_check_txs_locked()) Timed out tx: active_txs, 0 seconds
bcluster231.log-20140309.gz:2014-03-03 23:18:53 LustreError: 25883:0:(o2iblnd_cb.c:3055:kiblnd_check_conns()) Timed out RDMA with JO.BOO.AL.IL@o2ib2 (151): c: 8, oc: 0, rc: 8
bcluster231.log-20140309.gz:2014-03-03 23:18:53 Mar  3 23:18:53 bcluster231 kernel: LustreError: 25883:0:(o2iblnd_cb.c:2992:kiblnd_check_txs_locked()) Timed out tx: active_txs, 0 seconds
bcluster231.log-20140309.gz:2014-03-03 23:18:53 Mar  3 23:18:53 bcluster231 kernel: LustreError: 25883:0:(o2iblnd_cb.c:3055:kiblnd_check_conns()) Timed out RDMA with JO.BOO.AL.IL@o2ib2 (151): c: 8, oc: 0, rc: 8
bcluster231.log-20140309.gz:2014-03-03 23:22:50 Lustre: scratch3-OST01df: haven&lt;span class=&quot;code-quote&quot;&gt;&apos;t heard from client scratch3-MDT0000-mdtlov_UUID (at JO.BOO.AL.IL@o2ib2) in 902 seconds. I think it&apos;&lt;/span&gt;s dead, and I am evicting it. exp ffff880dc9df1000, cur 1393885370 expire 1393884770 last 1393884468
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The OSS log showed that something wrong with network when the problem happend on MDS, could you check your network as Amir suggested in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4195&quot; title=&quot;MDT Slow with ptlrpcd using 100% cpu.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4195&quot;&gt;&lt;del&gt;LU-4195&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="95518" author="sebastien.buisson" created="Thu, 2 Oct 2014 14:39:26 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;The customer has upgraded to Lustre 2.4, and since then is not able to reproduce.&lt;br/&gt;
So this ticket can be closed.&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
Sebastien.&lt;/p&gt;</comment>
                            <comment id="95519" author="pjones" created="Thu, 2 Oct 2014 14:43:29 +0000"  >&lt;p&gt;ok thanks Sebastien&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="14778" name="backtrace-20140407" size="1730560" author="patrick.valentin" created="Fri, 25 Apr 2014 09:31:02 +0000"/>
                            <attachment id="14240" name="bt-all.txt" size="618357" author="sebastien.buisson" created="Fri, 7 Mar 2014 14:20:29 +0000"/>
                            <attachment id="14239" name="dmesg(1).txt" size="176935" author="sebastien.buisson" created="Fri, 7 Mar 2014 14:20:29 +0000"/>
                            <attachment id="14677" name="lctl-dk.tgz" size="329504" author="patrick.valentin" created="Wed, 9 Apr 2014 12:17:46 +0000"/>
                            <attachment id="14265" name="log_2014-03-02-03.tgz" size="1426037" author="patrick.valentin" created="Tue, 11 Mar 2014 08:28:35 +0000"/>
                            <attachment id="14511" name="log_OSS_2014-03-02_04.log2.gz" size="107508" author="sebastien.buisson" created="Thu, 20 Mar 2014 12:35:43 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwh33:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>13008</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>