<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:51:25 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-5428] LNet: Service thread pid completed after 0.00s (DDN SR34734)</title>
                <link>https://jira.whamcloud.com/browse/LU-5428</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This problem was reported against a newly installed system at NOAA (Boulder).  The system was idle at the time:&lt;/p&gt;

&lt;p&gt;Jul 17 04:53:57 lfs-mds-0-1 kernel: : LNet: Service thread pid 29363 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;br/&gt;
Jul 17 05:26:52 lfs-mds-0-1 kernel: : LNet: Service thread pid 29363 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;br/&gt;
Jul 20 04:10:08 lfs-mds-0-1 kernel: : LDISKFS-fs (dm-9): mounted filesystem with ordered data mode. quota=off. Opts: &lt;br/&gt;
Jul 21 01:20:12 lfs-mds-0-1 kernel: : LNet: Service thread pid 13603 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;br/&gt;
Jul 21 14:35:12 lfs-mds-0-1 kernel: : LNet: Service thread pid 13829 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;br/&gt;
Jul 23 05:55:12 lfs-mds-0-1 kernel: : LNet: Service thread pid 29363 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;br/&gt;
Jul 24 11:19:47 lfs-mds-0-1 kernel: : LNet: Service thread pid 13672 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;/p&gt;

&lt;p&gt;Customer stats that he is observing LNet: Service thread pid completed after 0.00s even when the system is idle (they are on testbed (pre-production)).    &lt;/p&gt;

&lt;p&gt;I also saw these same messages on another  idle system that was newly installed (Harvard (HMU)).   &lt;/p&gt;</description>
                <environment>New Installation - Lustre 2.4.3 servers, 1.8.9 Clients</environment>
        <key id="25800">LU-5428</key>
            <summary>LNet: Service thread pid completed after 0.00s (DDN SR34734)</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="liang">Liang Zhen</assignee>
                                    <reporter username="orentas">Oz Rentas</reporter>
                        <labels>
                    </labels>
                <created>Tue, 29 Jul 2014 20:10:37 +0000</created>
                <updated>Tue, 21 Mar 2017 13:54:52 +0000</updated>
                            <resolved>Thu, 28 Aug 2014 19:52:39 +0000</resolved>
                                                    <fixVersion>Lustre 2.7.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="90474" author="pjones" created="Wed, 30 Jul 2014 19:11:37 +0000"  >&lt;p&gt;Liang&lt;/p&gt;

&lt;p&gt;Could you please advise on this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="90558" author="liang" created="Thu, 31 Jul 2014 16:47:15 +0000"  >&lt;p&gt;It is strange that we saw &quot;Service thread pid completed after 0.00s&quot;, because watchdog should complain only if service thread took too long to finish a request, but here we saw 0.00s. I think it could be a bug in our watchdog code, I will look into it. &lt;br/&gt;
Btw, I guess the system should be still working fine besides these fault warning?&lt;/p&gt;</comment>
                            <comment id="90662" author="orentas" created="Fri, 1 Aug 2014 20:46:04 +0000"  >&lt;p&gt;Yes, very strange. I agree.&lt;/p&gt;

&lt;p&gt;&amp;gt;&amp;gt;Btw, I guess the system should be still working fine besides these fault warning?&lt;br/&gt;
Yes&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
Oz&lt;/p&gt;</comment>
                            <comment id="91320" author="orentas" created="Mon, 11 Aug 2014 18:52:32 +0000"  >&lt;p&gt;Any ideas on this one?&lt;/p&gt;</comment>
                            <comment id="91397" author="liang" created="Tue, 12 Aug 2014 15:14:44 +0000"  >&lt;p&gt;Hi, sorry for late response. I have worked out a patch: &lt;a href=&quot;http://review.whamcloud.com/11415&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/11415&lt;/a&gt;&lt;br/&gt;
Briefly, reason of this issue is because there is a race between lc_watchdog_touch and lcw_cb which will generate false alarm, this should be harmless, but it is still good to fix it.&lt;/p&gt;</comment>
                            <comment id="91925" author="liang" created="Tue, 19 Aug 2014 07:41:17 +0000"  >&lt;p&gt;Patch landed to master&lt;/p&gt;</comment>
                            <comment id="92741" author="orentas" created="Thu, 28 Aug 2014 19:46:48 +0000"  >&lt;p&gt;Thanks much.  Go ahead and close this.&lt;/p&gt;</comment>
                            <comment id="92743" author="pjones" created="Thu, 28 Aug 2014 19:52:39 +0000"  >&lt;p&gt;Thanks Oz&lt;/p&gt;</comment>
                            <comment id="143156" author="rustequal" created="Mon, 22 Feb 2016 11:40:05 +0000"  >&lt;p&gt;Will this patch work with earlier versions of Lustre? For example, version 2.1.5 ?&lt;/p&gt;

&lt;p&gt;In version 2.1.5, we are seeing the same symptoms:&lt;/p&gt;

&lt;p&gt;Feb 21 07:29:59 mmp-2 kernel: Lustre: Service thread pid 5040 was inactive for 0.00s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: Pid: 5040, comm: ll_mgs_00&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel:&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: Call Trace:&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: Lustre: Service thread pid 5040 completed after 0.00s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0620af4&amp;gt;&amp;#93;&lt;/span&gt; ? target_send_reply_msg+0x54/0x190 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa03b360e&amp;gt;&amp;#93;&lt;/span&gt; cfs_waitq_wait+0xe/0x10 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa065e9f9&amp;gt;&amp;#93;&lt;/span&gt; ptlrpc_wait_event+0x2b9/0x2c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8105fa40&amp;gt;&amp;#93;&lt;/span&gt; ? default_wake_function+0x0/0x20&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa066657d&amp;gt;&amp;#93;&lt;/span&gt; ptlrpc_main+0x61d/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0665f60&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_main+0x0/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c0ca&amp;gt;&amp;#93;&lt;/span&gt; child_rip+0xa/0x20&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0665f60&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_main+0x0/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0665f60&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_main+0x0/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c0c0&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0x0/0x20&lt;br/&gt;
Feb 21 07:29:59 mmp-2 kernel:&lt;/p&gt;</comment>
                            <comment id="144137" author="liang" created="Mon, 29 Feb 2016 16:30:55 +0000"  >&lt;p&gt;yes, I think it should work for 2.1.5.&lt;/p&gt;</comment>
                            <comment id="169626" author="gerrit" created="Fri, 14 Oct 2016 11:21:20 +0000"  >&lt;p&gt;Hongchao Zhang (hongchao.zhang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/23162&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/23162&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5428&quot; title=&quot;LNet: Service thread pid completed after 0.00s (DDN SR34734)&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5428&quot;&gt;&lt;del&gt;LU-5428&lt;/del&gt;&lt;/a&gt; libcfs: race in lc_watchdog&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 637b75e8ebd2e17c151f32aecb341dbfa336264b&lt;/p&gt;</comment>
                            <comment id="189086" author="pjones" created="Tue, 21 Mar 2017 13:48:33 +0000"  >&lt;p&gt;Patch will be tracked for landing under &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9235&quot; title=&quot;LNet: Service thread pid 13033 was inactive for 0.00s.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9235&quot;&gt;&lt;del&gt;LU-9235&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwskn:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>15111</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>