<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:23:44 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-9156] parallel-scale test_metabench: test failed to respond and timed out</title>
                <link>https://jira.whamcloud.com/browse/LU-9156</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/e072c616-f8f8-11e6-aac4-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/e072c616-f8f8-11e6-aac4-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;From Client 2 console:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;05:47:50:[15560.679336] kworker/0:1     D 0000000000000000     0  6749      2 0x00000080
05:47:50:[15560.680977] Workqueue: events key_garbage_collector
05:47:50:[15560.682439]  ffff88003604bc40 0000000000000046 ffff88007aafce70 ffff88003604bfd8
05:47:50:[15560.684148]  ffff88003604bfd8 ffff88003604bfd8 ffff88007aafce70 ffff88003604bd80
05:47:50:[15560.685840]  ffff88003604bd88 7fffffffffffffff ffff88007aafce70 0000000000000000
05:47:50:[15560.687529] Call Trace:
05:47:50:[15560.688780]  [&amp;lt;ffffffff8168b979&amp;gt;] schedule+0x29/0x70
05:47:50:[15560.690227]  [&amp;lt;ffffffff816893c9&amp;gt;] schedule_timeout+0x239/0x2d0
05:47:50:[15560.691732]  [&amp;lt;ffffffff811dc341&amp;gt;] ? __slab_free+0x81/0x2f0
05:47:50:[15560.693210]  [&amp;lt;ffffffff810a5cdc&amp;gt;] ? insert_work+0x4c/0xa0
05:47:50:[15560.694666]  [&amp;lt;ffffffff811dc341&amp;gt;] ? __slab_free+0x81/0x2f0
05:47:50:[15560.696107]  [&amp;lt;ffffffff8168bd56&amp;gt;] wait_for_completion+0x116/0x170
05:47:50:[15560.697604]  [&amp;lt;ffffffff810c4fd0&amp;gt;] ? wake_up_state+0x20/0x20
05:47:50:[15560.699063]  [&amp;lt;ffffffff81138a30&amp;gt;] ? __call_rcu+0x2c0/0x2c0
05:47:50:[15560.700518]  [&amp;lt;ffffffff810ad1bb&amp;gt;] wait_rcu_gp+0x5b/0x80
05:47:50:[15560.701979]  [&amp;lt;ffffffff810ad140&amp;gt;] ? ftrace_raw_output_rcu_utilization+0x50/0x50
05:47:50:[15560.703560]  [&amp;lt;ffffffff8113862b&amp;gt;] synchronize_sched+0x3b/0x50
05:47:50:[15560.705020]  [&amp;lt;ffffffff8129b47d&amp;gt;] key_garbage_collector+0x1cd/0x390
05:47:50:[15560.706513]  [&amp;lt;ffffffff810a805b&amp;gt;] process_one_work+0x17b/0x470
05:47:50:[15560.707970]  [&amp;lt;ffffffff810a8e96&amp;gt;] worker_thread+0x126/0x410
05:47:50:[15560.709415]  [&amp;lt;ffffffff810a8d70&amp;gt;] ? rescuer_thread+0x460/0x460
05:47:50:[15560.710882]  [&amp;lt;ffffffff810b064f&amp;gt;] kthread+0xcf/0xe0
05:47:50:[15560.712267]  [&amp;lt;ffffffff810b0580&amp;gt;] ? kthread_create_on_node+0x140/0x140
05:47:50:[15560.713767]  [&amp;lt;ffffffff81696818&amp;gt;] ret_from_fork+0x58/0x90
05:47:50:[15560.715157]  [&amp;lt;ffffffff810b0580&amp;gt;] ? kthread_create_on_node+0x140/0x140
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>trevis-36vm1/2/7/8, Full Group test, &lt;br/&gt;
RHEL7.3/zfs, branch master, v2.9.53.19.gdad9ce4, b3534&lt;br/&gt;
</environment>
        <key id="44093">LU-9156</key>
            <summary>parallel-scale test_metabench: test failed to respond and timed out</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jcasper">James Casper</reporter>
                        <labels>
                    </labels>
                <created>Fri, 24 Feb 2017 18:23:22 +0000</created>
                <updated>Mon, 9 Oct 2017 22:09:59 +0000</updated>
                                            <version>Lustre 2.10.0</version>
                    <version>Lustre 2.11.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="186483" author="adilger" created="Tue, 28 Feb 2017 18:50:05 +0000"  >&lt;p&gt;Jim, could you please elaborate on why this particular client stack is the important one?&lt;/p&gt;</comment>
                            <comment id="187663" author="casperjx" created="Thu, 9 Mar 2017 17:20:12 +0000"  >&lt;p&gt;Have also seen &quot;kworker   D&quot; with mmp and sanity-sec.  Latest timeout with mmp:&lt;/p&gt;

&lt;p&gt;test_7: stops at mds umount:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Stopping /mnt/lustre-mds1failover (opts:) on trevis-44vm3
CMD: trevis-44vm3 umount -d /mnt/lustre-mds1failover
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;A passing test_7 continues after the mds umount:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Stopping /mnt/lustre-mds1failover (opts:) on onyx-37vm3
CMD: onyx-37vm3 umount -d /mnt/lustre-mds1failover
CMD: onyx-37vm3 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
CMD: onyx-37vm4 grep -c /mnt/lustre-ost1failover&apos; &apos; /proc/mounts
Stopping /mnt/lustre-ost1failover (opts:) on onyx-37vm4
CMD: onyx-37vm4 umount -d /mnt/lustre-ost1failover
CMD: onyx-37vm4 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
CMD: onyx-37vm7 grep -c /mnt/lustre-mds1&apos; &apos; /proc/mounts
CMD: onyx-37vm7 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
CMD: onyx-37vm8 grep -c /mnt/lustre-ost1&apos; &apos; /proc/mounts
CMD: onyx-37vm8 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
Resetting fail_loc on all nodes...CMD: onyx-37vm1.onyx.hpdd.intel.com,onyx-37vm5,onyx-37vm6,onyx-37vm7,onyx-37vm8 lctl set_param -n fail_loc=0 	    fail_val=0 2&amp;gt;/dev/null
done.
22:35:01 (1485498901) waiting for onyx-37vm1.onyx.hpdd.intel.com network 5 secs ...
22:35:01 (1485498901) network interface is UP
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="188529" author="standan" created="Wed, 15 Mar 2017 22:49:16 +0000"  >&lt;p&gt;An Instance on EE 3.1.1.0 , b_ieel3_0, build# 185&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/8d28da46-0798-11e7-8406-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/8d28da46-0798-11e7-8406-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="196415" author="yujian" created="Thu, 18 May 2017 23:43:14 +0000"  >&lt;p&gt;While testing patch &lt;a href=&quot;https://review.whamcloud.com/25336&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/25336&lt;/a&gt; on master branch, many tests failed with the similar issue to this one.&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/c145c4a3-b193-4734-b4c0-1aec2522f09a&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/c145c4a3-b193-4734-b4c0-1aec2522f09a&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/6fdca020-ca43-42c8-9fc3-3ffac60c1ba1&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/6fdca020-ca43-42c8-9fc3-3ffac60c1ba1&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I did not find any abnormal messages on consoles but only this stack:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;kworker/0:0     D 0000000000000000     0  7535      2 0x00000080
Workqueue: events key_garbage_collector
 ffff88004e1f3c40 0000000000000046 ffff880058ec2f10 ffff88004e1f3fd8
 ffff88004e1f3fd8 ffff88004e1f3fd8 ffff880058ec2f10 ffff88004e1f3d80
 ffff88004e1f3d88 7fffffffffffffff ffff880058ec2f10 0000000000000000
Call Trace:
 [&amp;lt;ffffffff8168c169&amp;gt;] schedule+0x29/0x70
 [&amp;lt;ffffffff81689bc9&amp;gt;] schedule_timeout+0x239/0x2c0
 [&amp;lt;ffffffff811dc861&amp;gt;] ? __slab_free+0x81/0x2f0
 [&amp;lt;ffffffff810b1c17&amp;gt;] ? wake_bit_function_rh+0x37/0x40
 [&amp;lt;ffffffff8168c546&amp;gt;] wait_for_completion+0x116/0x170
 [&amp;lt;ffffffff810c54c0&amp;gt;] ? wake_up_state+0x20/0x20
 [&amp;lt;ffffffff81138f20&amp;gt;] ? __call_rcu+0x2c0/0x2c0
 [&amp;lt;ffffffff810ad5bb&amp;gt;] wait_rcu_gp+0x5b/0x80
 [&amp;lt;ffffffff810ad540&amp;gt;] ? ftrace_raw_output_rcu_utilization+0x50/0x50
 [&amp;lt;ffffffff81138b1b&amp;gt;] synchronize_sched+0x3b/0x50
 [&amp;lt;ffffffff8129b9bd&amp;gt;] key_garbage_collector+0x1cd/0x390
 [&amp;lt;ffffffff810a845b&amp;gt;] process_one_work+0x17b/0x470
 [&amp;lt;ffffffff810a9296&amp;gt;] worker_thread+0x126/0x410
 [&amp;lt;ffffffff810a9170&amp;gt;] ? rescuer_thread+0x460/0x460
 [&amp;lt;ffffffff810b0a4f&amp;gt;] kthread+0xcf/0xe0
 [&amp;lt;ffffffff810b0980&amp;gt;] ? kthread_create_on_node+0x140/0x140
 [&amp;lt;ffffffff816970d8&amp;gt;] ret_from_fork+0x58/0x90
 [&amp;lt;ffffffff810b0980&amp;gt;] ? kthread_create_on_node+0x140/0x140
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="196763" author="casperjx" created="Tue, 23 May 2017 16:46:09 +0000"  >&lt;p&gt;Still looking for guidance with this ticket.  If the kworker hang is not the root cause, please point us in the right direction.  This is the only indication we can find of something that might be wrong.&lt;/p&gt;</comment>
                            <comment id="196994" author="casperjx" created="Wed, 24 May 2017 23:17:05 +0000"  >&lt;p&gt;2.9.57, b3575:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/0800ff00-8d4c-4627-878e-566e8a697c01&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/0800ff00-8d4c-4627-878e-566e8a697c01&lt;/a&gt;&lt;br/&gt;
(Interop w/2.10 on server)&lt;/p&gt;</comment>
                            <comment id="209172" author="yujian" created="Fri, 22 Sep 2017 02:29:13 +0000"  >&lt;p&gt;+1 on master branch:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/401fc94c-9ef8-11e7-b778-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/401fc94c-9ef8-11e7-b778-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzz4w7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>