<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:03:36 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6829] runtests test_1: MDS hung</title>
                <link>https://jira.whamcloud.com/browse/LU-6829</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah_lw &amp;lt;wei3.liu@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/24729d7e-2640-11e5-8b33-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/24729d7e-2640-11e5-8b33-5254006e85c2&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_1 failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;test failed to respond and timed out
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;MDS dmesg&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[ 1244.736431] Lustre: DEBUG MARKER: /usr/sbin/lctl mark copying 880 files from \/etc \/bin to \/mnt\/lustre\/d1.runtests\/etc \/bin at Wed Jul  8 20:30:42 UTC 2015
[ 1244.857091] Lustre: DEBUG MARKER: copying 880 files from /etc /bin to /mnt/lustre/d1.runtests/etc /bin at Wed Jul 8 20:30:42 UTC 2015
[ 1252.974179] Lustre: 2786:0:(client.c:2018:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1436387443/real 1436387443]  req@ffff880069257300 x1506161297721592/t0(0) o400-&amp;gt;lustre-MDT0002-osp-MDT0000@10.1.5.252@tcp:24/4 lens 224/224 e 0 to 1 dl 1436387450 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
[ 1252.979486] Lustre: lustre-MDT0002-osp-MDT0000: Connection to lustre-MDT0002 (at 10.1.5.252@tcp) was lost; in progress operations using this service will wait for recovery to complete
[ 1257.239137] Lustre: 2785:0:(client.c:2018:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1436387448/real 1436387448]  req@ffff880068c86700 x1506161297721644/t0(0) o400-&amp;gt;lustre-MDT0002-osp-MDT0000@10.1.5.252@tcp:24/4 lens 224/224 e 0 to 1 dl 1436387455 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
[ 1257.243885] Lustre: 2785:0:(client.c:2018:ptlrpc_expire_one_request()) Skipped 2 previous similar messages
[ 1258.983125] Lustre: 2784:0:(client.c:2018:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1436387450/real 1436387450]  req@ffff88006912db00 x1506161297721660/t0(0) o38-&amp;gt;lustre-MDT0003-osp-MDT0000@10.1.5.252@tcp:24/4 lens 520/544 e 0 to 1 dl 1436387456 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
[ 1258.988595] Lustre: 2784:0:(client.c:2018:ptlrpc_expire_one_request()) Skipped 2 previous similar messages
[ 1273.982152] Lustre: 2784:0:(client.c:2018:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1436387460/real 1436387460]  req@ffff880069d87300 x1506161297721744/t0(0) o38-&amp;gt;lustre-MDT0001-osp-MDT0000@10.1.5.252@tcp:24/4 lens 520/544 e 0 to 1 dl 1436387471 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
[ 1273.987550] Lustre: 2784:0:(client.c:2018:ptlrpc_expire_one_request()) Skipped 2 previous similar messages
[ 1285.839237] Lustre: 2815:0:(client.c:2018:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1436387476/real 1436387476]  req@ffff880069d87300 x1506161297721892/t0(0) o104-&amp;gt;lustre-MDT0000@10.1.5.252@tcp:15/16 lens 296/224 e 0 to 1 dl 1436387483 ref 1 fl Rpc:X/0/ffffffff rc 0/-1
[ 1285.844932] Lustre: 2815:0:(client.c:2018:ptlrpc_expire_one_request()) Skipped 2 previous similar messages
[ 1292.205644] Lustre: lustre-MDT0000: haven&apos;t heard from client lustre-MDT0000-lwp-MDT0001_UUID (at 10.1.5.252@tcp) in 49 seconds. I think it&apos;s dead, and I am evicting it. exp ffff88007bc28400, cur 1436387490 expire 1436387460 last 1436387441
[ 1293.982194] Lustre: 2784:0:(client.c:2018:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1436387475/real 1436387475]  req@ffff880069d85b00 x1506161297721880/t0(0) o38-&amp;gt;lustre-MDT0001-osp-MDT0000@10.1.5.252@tcp:24/4 lens 520/544 e 0 to 1 dl 1436387491 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
[ 1318.879183] LNet: Service thread pid 2815 was inactive for 40.04s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[ 1318.882589] Pid: 2815, comm: mdt00_001
[ 1318.883358] 
Call Trace:
[ 1318.884177]  [&amp;lt;ffffffff8160a409&amp;gt;] schedule+0x29/0x70
[ 1318.884703]  [&amp;lt;ffffffff816082b5&amp;gt;] schedule_timeout+0x175/0x2d0
[ 1318.885134]  [&amp;lt;ffffffff8107ee80&amp;gt;] ? process_timeout+0x0/0x10
[ 1318.885707]  [&amp;lt;ffffffffa09a8320&amp;gt;] ? ptlrpc_interrupted_set+0x0/0x120 [ptlrpc]
[ 1318.886241]  [&amp;lt;ffffffffa09b2978&amp;gt;] ptlrpc_set_wait+0x4b8/0x9e0 [ptlrpc]
[ 1318.886787]  [&amp;lt;ffffffff810a9650&amp;gt;] ? default_wake_function+0x0/0x20
[ 1318.887253]  [&amp;lt;ffffffffa09b2f1d&amp;gt;] ptlrpc_queue_wait+0x7d/0x220 [ptlrpc]
[ 1318.887764]  [&amp;lt;ffffffffa0f433c5&amp;gt;] osp_remote_sync+0xd5/0x1b0 [osp]
[ 1318.888210]  [&amp;lt;ffffffffa0f2a25d&amp;gt;] osp_attr_get+0x41d/0x690 [osp]
[ 1318.888728]  [&amp;lt;ffffffffa0f25ab4&amp;gt;] osp_object_init+0x114/0x280 [osp]
[ 1318.889220]  [&amp;lt;ffffffffa076058f&amp;gt;] lu_object_alloc+0xdf/0x310 [obdclass]
[ 1318.889785]  [&amp;lt;ffffffffa076099c&amp;gt;] lu_object_find_try+0x17c/0x2c0 [obdclass]
[ 1318.890295]  [&amp;lt;ffffffffa0760b8c&amp;gt;] lu_object_find_at+0xac/0xe0 [obdclass]
[ 1318.890940]  [&amp;lt;ffffffffa0e6ebb5&amp;gt;] ? lod_index_lookup+0x25/0x30 [lod]
[ 1318.891407]  [&amp;lt;ffffffffa0ecc307&amp;gt;] ? __mdd_lookup.isra.17+0x317/0x440 [mdd]
[ 1318.891993]  [&amp;lt;ffffffffa0760bd6&amp;gt;] lu_object_find+0x16/0x20 [obdclass]
[ 1318.892575]  [&amp;lt;ffffffffa0d8a73b&amp;gt;] mdt_object_find+0x4b/0x170 [mdt]
[ 1318.893146]  [&amp;lt;ffffffffa0db0726&amp;gt;] mdt_reint_open+0x2776/0x2d50 [mdt]
[ 1318.893624]  [&amp;lt;ffffffffa07795f9&amp;gt;] ? upcall_cache_get_entry+0x3e9/0x8e0 [obdclass]
[ 1318.894169]  [&amp;lt;ffffffff812ddd72&amp;gt;] ? strlcpy+0x42/0x60
[ 1318.894571]  [&amp;lt;ffffffffa0da46e0&amp;gt;] mdt_reint_rec+0x80/0x210 [mdt]
[ 1318.894985]  [&amp;lt;ffffffffa0d884e9&amp;gt;] mdt_reint_internal+0x5d9/0xb20 [mdt]
[ 1318.895456]  [&amp;lt;ffffffffa0d88b92&amp;gt;] mdt_intent_reint+0x162/0x410 [mdt]
[ 1318.896056]  [&amp;lt;ffffffffa0d921ea&amp;gt;] mdt_intent_policy+0x57a/0xb30 [mdt]
[ 1318.896619]  [&amp;lt;ffffffffa0970223&amp;gt;] ldlm_lock_enqueue+0x353/0x930 [ptlrpc]
[ 1318.897117]  [&amp;lt;ffffffffa0998da2&amp;gt;] ldlm_handle_enqueue0+0x4f2/0x16e0 [ptlrpc]
[ 1318.897725]  [&amp;lt;ffffffffa09c0360&amp;gt;] ? lustre_swab_ldlm_request+0x0/0x30 [ptlrpc]
[ 1318.898274]  [&amp;lt;ffffffffa0a1eed2&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
[ 1318.898762]  [&amp;lt;ffffffffa0a2329b&amp;gt;] tgt_request_handle+0x88b/0x1100 [ptlrpc]
[ 1318.899278]  [&amp;lt;ffffffffa09cafbb&amp;gt;] ptlrpc_server_handle_request+0x21b/0xa90 [ptlrpc]
[ 1318.899926]  [&amp;lt;ffffffffa09c8078&amp;gt;] ? ptlrpc_wait_event+0x98/0x340 [ptlrpc]
[ 1318.900416]  [&amp;lt;ffffffff810a9662&amp;gt;] ? default_wake_function+0x12/0x20
[ 1318.900956]  [&amp;lt;ffffffff810a0898&amp;gt;] ? __wake_up_common+0x58/0x90
[ 1318.901397]  [&amp;lt;ffffffffa09ce900&amp;gt;] ptlrpc_main+0xc00/0x1f60 [ptlrpc]
[ 1318.901953]  [&amp;lt;ffffffffa09cdd00&amp;gt;] ? ptlrpc_main+0x0/0x1f60 [ptlrpc]
[ 1318.902407]  [&amp;lt;ffffffff8109739f&amp;gt;] kthread+0xcf/0xe0
[ 1318.902851]  [&amp;lt;ffffffff810972d0&amp;gt;] ? kthread+0x0/0xe0
[ 1318.903215]  [&amp;lt;ffffffff81614f7c&amp;gt;] ret_from_fork+0x7c/0xb0
[ 1318.903640]  [&amp;lt;ffffffff810972d0&amp;gt;] ? kthread+0x0/0xe0

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>client and server: lustre-master build # 3094 RHEL7 DNE</environment>
        <key id="31031">LU-6829</key>
            <summary>runtests test_1: MDS hung</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="di.wang">Di Wang</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Thu, 9 Jul 2015 18:19:00 +0000</created>
                <updated>Tue, 29 Sep 2015 18:24:03 +0000</updated>
                            <resolved>Wed, 26 Aug 2015 18:04:05 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                    <fixVersion>Lustre 2.8.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="123141" author="sebastien.buisson" created="Tue, 4 Aug 2015 05:58:13 +0000"  >&lt;p&gt;A new instance of the problem:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/ca6cf5aa-3a02-11e5-8f15-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/ca6cf5aa-3a02-11e5-8f15-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="125221" author="di.wang" created="Wed, 26 Aug 2015 17:23:18 +0000"  >&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;https://testing.hpdd.intel.com/test_sets/ca6cf5aa-3a02-11e5-8f15-5254006e85c2
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;This is actually &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6845&quot; title=&quot;conf-sanity test_30b: kernel panic in dt_txn_hook_commit+0x30 during umount&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6845&quot;&gt;&lt;del&gt;LU-6845&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="125241" author="di.wang" created="Wed, 26 Aug 2015 18:03:51 +0000"  >&lt;p&gt;I checked most failures since the last months. All of them are caused by other problem &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7047&quot; title=&quot;runtests test_1: kernel panic during umount -d -f /mnt/mds1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7047&quot;&gt;&lt;del&gt;LU-7047&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7048&quot; title=&quot;runtests test_1: (osc_page.c:314:osc_page_delete()) ASSERTION( 0 ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7048&quot;&gt;&lt;del&gt;LU-7048&lt;/del&gt;&lt;/a&gt;, or &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6578&quot; title=&quot;inodebit locks for remote entries.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6578&quot;&gt;&lt;del&gt;LU-6578&lt;/del&gt;&lt;/a&gt;. So I will close this for now, since it does not happen for more than one month.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="29896">LU-6578</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxhvb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>