<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:36:49 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-10630] recovery-random-scale test_fail_client_mds: client cannot connect to MDS</title>
                <link>https://jira.whamcloud.com/browse/LU-10630</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;recovery-random-scale test_fail_client_mds - Timeout occurred after 1444 mins, last suite running was recovery-random-scale, restarting cluster to continue tests&lt;br/&gt;
^^^^^^^^^^^^^ DO NOT REMOVE LINE ABOVE ^^^^^^^^^^^^^&lt;/p&gt;

&lt;p&gt;This issue was created by maloo for sarah_lw &amp;lt;wei3.liu@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b98d4034-ff51-11e7-a7cd-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b98d4034-ff51-11e7-a7cd-52540065bddc&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;test_fail_client_mds failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Timeout occurred after 1444 mins, last suite running was recovery-random-scale, restarting cluster to continue tests
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[   18.684976] LNet: Accept all, port 7988
[  113.727780] LustreError: 11-0: lustre-MDT0000-mdc-ffff88007b744000: operation mds_connect to node 10.2.8.168@tcp failed: rc = -11
[  128.332376] random: crng init done
[  263.727780] LustreError: 11-0: lustre-MDT0000-mdc-ffff88007b744000: operation mds_connect to node 10.2.8.168@tcp failed: rc = -11
[  413.727721] LustreError: 11-0: lustre-MDT0000-mdc-ffff88007b744000: operation mds_connect to node 10.2.8.168@tcp failed: rc = -11
[  563.727711] LustreError: 11-0: lustre-MDT0000-mdc-ffff88007b744000: operation mds_connect to node 10.2.8.168@tcp failed: rc = -11
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="50691">LU-10630</key>
            <summary>recovery-random-scale test_fail_client_mds: client cannot connect to MDS</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="sarah">Sarah Liu</reporter>
                        <labels>
                    </labels>
                <created>Wed, 7 Feb 2018 17:08:22 +0000</created>
                <updated>Wed, 24 Nov 2021 19:28:25 +0000</updated>
                            <resolved>Wed, 24 Nov 2021 19:28:25 +0000</resolved>
                                    <version>Lustre 2.11.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>2</watches>
                                                                            <comments>
                            <comment id="220375" author="jamesanunez" created="Thu, 8 Feb 2018 00:21:25 +0000"  >&lt;p&gt;There&apos;s not much to look at in the dmesg logs, but in the MDS1 (vm12) console log, we see the following stack trace&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[  618.123147] LNet: Service thread pid 13443 was inactive for 60.04s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[  618.126713] Pid: 13443, comm: mdt00_003
[  618.127520] 
[  618.127520] Call Trace:
[  618.128418]  [&amp;lt;ffffffff816ab6b9&amp;gt;] schedule+0x29/0x70
[  618.129587]  [&amp;lt;ffffffff816a9004&amp;gt;] schedule_timeout+0x174/0x2c0
[  618.130846]  [&amp;lt;ffffffffc0abdd47&amp;gt;] ? libcfs_debug_msg+0x57/0x80 [libcfs]
[  618.132274]  [&amp;lt;ffffffff8109a6c0&amp;gt;] ? process_timeout+0x0/0x10
[  618.133532]  [&amp;lt;ffffffffc0ab2eb1&amp;gt;] ? cfs_block_sigsinv+0x71/0xa0 [libcfs]
[  618.134850]  [&amp;lt;ffffffffc13838d0&amp;gt;] osp_precreate_reserve+0x2e0/0x810 [osp]
[  618.136193]  [&amp;lt;ffffffff810c6440&amp;gt;] ? default_wake_function+0x0/0x20
[  618.137320]  [&amp;lt;ffffffffc1378c53&amp;gt;] osp_declare_create+0x193/0x590 [osp]
[  618.138609]  [&amp;lt;ffffffffc0bea619&amp;gt;] ? lprocfs_counter_add+0xf9/0x160 [obdclass]
[  618.139946]  [&amp;lt;ffffffffc12ca1dc&amp;gt;] lod_sub_declare_create+0xdc/0x210 [lod]
[  618.141282]  [&amp;lt;ffffffffc12c353e&amp;gt;] lod_qos_declare_object_on+0xbe/0x3a0 [lod]
[  618.142569]  [&amp;lt;ffffffffc12c44ba&amp;gt;] lod_alloc_rr.constprop.18+0x70a/0x1000 [lod]
[  618.143974]  [&amp;lt;ffffffffc12c8a8f&amp;gt;] lod_qos_prep_create+0xc0f/0x1830 [lod]
[  618.145265]  [&amp;lt;ffffffffc12c9c0d&amp;gt;] lod_prepare_create+0x25d/0x360 [lod]
[  618.146636]  [&amp;lt;ffffffffc12bbdce&amp;gt;] lod_declare_striped_create+0x1ee/0x970 [lod]
[  618.148097]  [&amp;lt;ffffffffc12ca1dc&amp;gt;] ? lod_sub_declare_create+0xdc/0x210 [lod]
[  618.149544]  [&amp;lt;ffffffffc12c00e4&amp;gt;] lod_declare_create+0x204/0x590 [lod]
[  618.150866]  [&amp;lt;ffffffffc0bea619&amp;gt;] ? lprocfs_counter_add+0xf9/0x160 [obdclass]
[  618.152333]  [&amp;lt;ffffffffc133139f&amp;gt;] mdd_declare_create_object_internal+0xdf/0x2f0 [mdd]
[  618.153810]  [&amp;lt;ffffffffc1321b53&amp;gt;] mdd_declare_create+0x53/0xe20 [mdd]
[  618.155154]  [&amp;lt;ffffffffc1325e69&amp;gt;] mdd_create+0x879/0x1410 [mdd]
[  618.156268]  [&amp;lt;ffffffffc11db305&amp;gt;] mdt_reint_open+0x1a45/0x2890 [mdt]
[  618.157529]  [&amp;lt;ffffffffc0c1e087&amp;gt;] ? upcall_cache_get_entry+0x3f7/0x8f0 [obdclass]
[  618.158905]  [&amp;lt;ffffffffc11beb53&amp;gt;] ? ucred_set_jobid+0x53/0x70 [mdt]
[  618.160143]  [&amp;lt;ffffffffc11cf410&amp;gt;] mdt_reint_rec+0x80/0x210 [mdt]
[  618.161253]  [&amp;lt;ffffffffc11aef8b&amp;gt;] mdt_reint_internal+0x5fb/0x9c0 [mdt]
[  618.162529]  [&amp;lt;ffffffffc11bb457&amp;gt;] mdt_intent_reint+0x157/0x420 [mdt]
[  618.163720]  [&amp;lt;ffffffffc11b20b2&amp;gt;] mdt_intent_opc+0x442/0xad0 [mdt]
[  618.165008]  [&amp;lt;ffffffffc0e3bb90&amp;gt;] ? lustre_swab_ldlm_intent+0x0/0x20 [ptlrpc]
[  618.166314]  [&amp;lt;ffffffffc11b9c73&amp;gt;] mdt_intent_policy+0x1a3/0x360 [mdt]
[  618.167583]  [&amp;lt;ffffffffc0dea2fa&amp;gt;] ldlm_lock_enqueue+0x38a/0x970 [ptlrpc]
[  618.168942]  [&amp;lt;ffffffffc0e13a33&amp;gt;] ldlm_handle_enqueue0+0x8f3/0x1400 [ptlrpc]
[  618.170447]  [&amp;lt;ffffffffc0e3bc10&amp;gt;] ? lustre_swab_ldlm_request+0x0/0x30 [ptlrpc]
[  618.171912]  [&amp;lt;ffffffffc0e99752&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
[  618.173114]  [&amp;lt;ffffffffc0ea1965&amp;gt;] tgt_request_handle+0x925/0x13b0 [ptlrpc]
[  618.173927]  [&amp;lt;ffffffffc0e45c7e&amp;gt;] ptlrpc_server_handle_request+0x24e/0xab0 [ptlrpc]
[  618.174816]  [&amp;lt;ffffffff810bc0f8&amp;gt;] ? __wake_up_common+0x58/0x90
[  618.175639]  [&amp;lt;ffffffffc0e49422&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
[  618.176374]  [&amp;lt;ffffffff810c0d30&amp;gt;] ? finish_task_switch+0x50/0x160
[  618.177202]  [&amp;lt;ffffffffc0e48990&amp;gt;] ? ptlrpc_main+0x0/0x1e40 [ptlrpc]
[  618.177932]  [&amp;lt;ffffffff810b252f&amp;gt;] kthread+0xcf/0xe0
[  618.178689]  [&amp;lt;ffffffff810b2460&amp;gt;] ? kthread+0x0/0xe0
[  618.179434]  [&amp;lt;ffffffff816b8798&amp;gt;] ret_from_fork+0x58/0x90
[  618.180131]  [&amp;lt;ffffffff810b2460&amp;gt;] ? kthread+0x0/0xe0
[  618.180732] 
[  618.180936] LustreError: dumping log to /tmp/lustre-log.1516576126.13443
[  618.379102] Pid: 12418, comm: mdt00_002
[  618.379607] 
[  618.379607] Call Trace:
[  618.380078]  [&amp;lt;ffffffff816ab6b9&amp;gt;] schedule+0x29/0x70
[  618.380662]  [&amp;lt;ffffffff816a9004&amp;gt;] schedule_timeout+0x174/0x2c0
[  618.381362]  [&amp;lt;ffffffffc0abdd47&amp;gt;] ? libcfs_debug_msg+0x57/0x80 [libcfs]
[  618.382286]  [&amp;lt;ffffffff8109a6c0&amp;gt;] ? process_timeout+0x0/0x10
[  618.382976]  [&amp;lt;ffffffffc0ab2eb1&amp;gt;] ? cfs_block_sigsinv+0x71/0xa0 [libcfs]
[  618.383754]  [&amp;lt;ffffffffc13838d0&amp;gt;] osp_precreate_reserve+0x2e0/0x810 [osp]
[  618.384636]  [&amp;lt;ffffffff810c6440&amp;gt;] ? default_wake_function+0x0/0x20
[  618.385399]  [&amp;lt;ffffffffc1378c53&amp;gt;] osp_declare_create+0x193/0x590 [osp]
[  618.386245]  [&amp;lt;ffffffffc0bea619&amp;gt;] ? lprocfs_counter_add+0xf9/0x160 [obdclass]
[  618.387066]  [&amp;lt;ffffffffc12ca1dc&amp;gt;] lod_sub_declare_create+0xdc/0x210 [lod]
[  618.387854]  [&amp;lt;ffffffffc12c353e&amp;gt;] lod_qos_declare_object_on+0xbe/0x3a0 [lod]
[  618.388780]  [&amp;lt;ffffffffc12c44ba&amp;gt;] lod_alloc_rr.constprop.18+0x70a/0x1000 [lod]
[  618.389774]  [&amp;lt;ffffffffc07009d5&amp;gt;] ? dbuf_find+0x1d5/0x1e0 [zfs]
[  618.390494]  [&amp;lt;ffffffffc0604487&amp;gt;] ? tsd_get+0x37/0x60 [spl]
[  618.391235]  [&amp;lt;ffffffffc12c8a8f&amp;gt;] lod_qos_prep_create+0xc0f/0x1830 [lod]
[  618.392049]  [&amp;lt;ffffffffc12c989a&amp;gt;] ? lod_prepare_inuse+0x1ea/0x300 [lod]
[  618.392812]  [&amp;lt;ffffffffc12c9c0d&amp;gt;] lod_prepare_create+0x25d/0x360 [lod]
[  618.393661]  [&amp;lt;ffffffffc12bbdce&amp;gt;] lod_declare_striped_create+0x1ee/0x970 [lod]
[  618.394541]  [&amp;lt;ffffffffc12ca1dc&amp;gt;] ? lod_sub_declare_create+0xdc/0x210 [lod]
[  618.395386]  [&amp;lt;ffffffffc12c00e4&amp;gt;] lod_declare_create+0x204/0x590 [lod]
[  618.396203]  [&amp;lt;ffffffffc133139f&amp;gt;] mdd_declare_create_object_internal+0xdf/0x2f0 [mdd]
[  618.397112]  [&amp;lt;ffffffffc1321b53&amp;gt;] mdd_declare_create+0x53/0xe20 [mdd]
[  618.397925]  [&amp;lt;ffffffffc1325e69&amp;gt;] mdd_create+0x879/0x1410 [mdd]
[  618.398641]  [&amp;lt;ffffffffc11db305&amp;gt;] mdt_reint_open+0x1a45/0x2890 [mdt]
[  618.399495]  [&amp;lt;ffffffffc0c1e087&amp;gt;] ? upcall_cache_get_entry+0x3f7/0x8f0 [obdclass]
[  618.400369]  [&amp;lt;ffffffffc11beb53&amp;gt;] ? ucred_set_jobid+0x53/0x70 [mdt]
[  618.401176]  [&amp;lt;ffffffffc11cf410&amp;gt;] mdt_reint_rec+0x80/0x210 [mdt]
[  618.401861]  [&amp;lt;ffffffffc11aef8b&amp;gt;] mdt_reint_internal+0x5fb/0x9c0 [mdt]
[  618.402628]  [&amp;lt;ffffffffc11bb457&amp;gt;] mdt_intent_reint+0x157/0x420 [mdt]
[  618.403504]  [&amp;lt;ffffffffc11b20b2&amp;gt;] mdt_intent_opc+0x442/0xad0 [mdt]
[  618.404332]  [&amp;lt;ffffffffc0e3bb90&amp;gt;] ? lustre_swab_ldlm_intent+0x0/0x20 [ptlrpc]
[  618.405167]  [&amp;lt;ffffffffc11b9c73&amp;gt;] mdt_intent_policy+0x1a3/0x360 [mdt]
[  618.405990]  [&amp;lt;ffffffffc0dea2fa&amp;gt;] ldlm_lock_enqueue+0x38a/0x970 [ptlrpc]
[  618.406786]  [&amp;lt;ffffffffc0e13a33&amp;gt;] ldlm_handle_enqueue0+0x8f3/0x1400 [ptlrpc]
[  618.407723]  [&amp;lt;ffffffffc0e3bc10&amp;gt;] ? lustre_swab_ldlm_request+0x0/0x30 [ptlrpc]
[  618.408607]  [&amp;lt;ffffffffc0e99752&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
[  618.409414]  [&amp;lt;ffffffffc0ea1965&amp;gt;] tgt_request_handle+0x925/0x13b0 [ptlrpc]
[  618.410241]  [&amp;lt;ffffffffc0e45c7e&amp;gt;] ptlrpc_server_handle_request+0x24e/0xab0 [ptlrpc]
[  618.411197]  [&amp;lt;ffffffff810bc0f8&amp;gt;] ? __wake_up_common+0x58/0x90
[  618.411936]  [&amp;lt;ffffffffc0e49422&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
[  618.412664]  [&amp;lt;ffffffff810c0d30&amp;gt;] ? finish_task_switch+0x50/0x160
[  618.413477]  [&amp;lt;ffffffffc0e48990&amp;gt;] ? ptlrpc_main+0x0/0x1e40 [ptlrpc]
[  618.414215]  [&amp;lt;ffffffff810b252f&amp;gt;] kthread+0xcf/0xe0
[  618.414856]  [&amp;lt;ffffffff810b2460&amp;gt;] ? kthread+0x0/0xe0
[  618.415446]  [&amp;lt;ffffffff816b8798&amp;gt;] ret_from_fork+0x58/0x90
[  618.416169]  [&amp;lt;ffffffff810b2460&amp;gt;] ? kthread+0x0/0xe0
[  618.416780] 
[  618.416974] LustreError: dumping log to /tmp/lustre-log.1516576127.12418
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzzsev:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>