<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:49:57 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-12136] mdt threads blocked in mdd_create</title>
                <link>https://jira.whamcloud.com/browse/LU-12136</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We had an issue yesterday on Oak storage with Lustre 2.10.6. MDT0 didn&apos;t crash but filesystem got stuck. Several stack traces showed up on &lt;tt&gt;oak-md1-s2&lt;/tt&gt; (serving MDT0000). Note: Oak uses DNE1 and another MDT0001 is mounted on &lt;tt&gt;oak-md1-s1&lt;/tt&gt;, but I didn&apos;t find any stack trace on this one. A restart of MDT0000 fixed the issue (after a workaround to mitigate&#160;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8992&quot; title=&quot;osd_object_release() LBUG&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8992&quot;&gt;&lt;del&gt;LU-8992&lt;/del&gt;&lt;/a&gt;).&lt;/p&gt;

&lt;p&gt;My short-term plan is to upgrade Oak to 2.10.7 in a rolling fashion but I thought it would be of interest to have a ticket to track this issue. I&apos;m also attaching kernel logs from this server as oak-md1-s2-kernel.log where all stack traces can be seen.&lt;/p&gt;

&lt;p&gt;First call trace was:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Mar 29 09:38:38 oak-md1-s2 kernel: INFO: task mdt00_003:3491 blocked for more than 120 seconds.
Mar 29 09:38:38 oak-md1-s2 kernel: &quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.
Mar 29 09:38:38 oak-md1-s2 kernel: mdt00_003       D ffffffff00000000     0  3491      2 0x00000080
Mar 29 09:38:38 oak-md1-s2 kernel: ffff88201e3f74b8 0000000000000046 ffff88201e3c3f40 ffff88201e3f7fd8
Mar 29 09:38:38 oak-md1-s2 kernel: ffff88201e3f7fd8 ffff88201e3f7fd8 ffff88201e3c3f40 ffff88201e3c3f40
Mar 29 09:38:38 oak-md1-s2 kernel: ffff88101fc13248 ffff88101fc13240 fffffffe00000001 ffffffff00000000
Mar 29 09:38:38 oak-md1-s2 kernel: Call Trace:
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff816a94e9&amp;gt;] schedule+0x29/0x70
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff816aadd5&amp;gt;] rwsem_down_write_failed+0x225/0x3a0
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff81332047&amp;gt;] call_rwsem_down_write_failed+0x17/0x30
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff816a87cd&amp;gt;] down_write+0x2d/0x3d
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc121b34f&amp;gt;] lod_alloc_qos.constprop.17+0x1af/0x1590 [lod]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0fa49a1&amp;gt;] ? qsd_op_begin0+0x181/0x940 [lquota]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0ed322f&amp;gt;] ? ldiskfs_xattr_ibody_get+0xef/0x1a0 [ldiskfs]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc12204d1&amp;gt;] lod_qos_prep_create+0x1291/0x17f0 [lod]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc1220bf9&amp;gt;] ? lod_prepare_inuse+0x1c9/0x2e0 [lod]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc1220f6d&amp;gt;] lod_prepare_create+0x25d/0x360 [lod]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc121578e&amp;gt;] lod_declare_striped_create+0x1ee/0x970 [lod]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc1217c04&amp;gt;] lod_declare_create+0x1e4/0x540 [lod]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc12828cf&amp;gt;] mdd_declare_create_object_internal+0xdf/0x2f0 [mdd]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc1274023&amp;gt;] mdd_declare_create+0x53/0xe20 [mdd]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc1277ec9&amp;gt;] mdd_create+0x879/0x1400 [mdd]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc114ab93&amp;gt;] mdt_reint_open+0x2173/0x3190 [mdt]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0931dde&amp;gt;] ? upcall_cache_get_entry+0x20e/0x8f0 [obdclass]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc112fad3&amp;gt;] ? ucred_set_jobid+0x53/0x70 [mdt]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc113fa40&amp;gt;] mdt_reint_rec+0x80/0x210 [mdt]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc112131b&amp;gt;] mdt_reint_internal+0x5fb/0x9c0 [mdt]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc1121842&amp;gt;] mdt_intent_reint+0x162/0x430 [mdt]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc112c5ae&amp;gt;] mdt_intent_policy+0x43e/0xc70 [mdt]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0afc12f&amp;gt;] ? ldlm_resource_get+0x9f/0xa30 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0af5277&amp;gt;] ldlm_lock_enqueue+0x387/0x970 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b1e9e3&amp;gt;] ldlm_handle_enqueue0+0x9c3/0x1680 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b46bc0&amp;gt;] ? lustre_swab_ldlm_lock_desc+0x30/0x30 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0ba3e92&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0ba7d95&amp;gt;] tgt_request_handle+0x925/0x1370 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b50bf6&amp;gt;] ptlrpc_server_handle_request+0x236/0xa90 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b4d228&amp;gt;] ? ptlrpc_wait_event+0x98/0x340 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff810c4822&amp;gt;] ? default_wake_function+0x12/0x20
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff810ba588&amp;gt;] ? __wake_up_common+0x58/0x90
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b54332&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b538a0&amp;gt;] ? ptlrpc_register_service+0xe30/0xe30 [ptlrpc]
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff810b098f&amp;gt;] kthread+0xcf/0xe0
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff810b08c0&amp;gt;] ? insert_kthread_work+0x40/0x40
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff816b4f58&amp;gt;] ret_from_fork+0x58/0x90
Mar 29 09:38:38 oak-md1-s2 kernel: [&amp;lt;ffffffff810b08c0&amp;gt;] ? insert_kthread_work+0x40/0x40
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>3.10.0-693.2.2.el7_lustre.pl1.x86_64</environment>
        <key id="55309">LU-12136</key>
            <summary>mdt threads blocked in mdd_create</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="sthiell">Stephane Thiell</reporter>
                        <labels>
                    </labels>
                <created>Sat, 30 Mar 2019 17:02:58 +0000</created>
                <updated>Mon, 10 Oct 2022 23:29:44 +0000</updated>
                                            <version>Lustre 2.10.6</version>
                                                        <due></due>
                            <votes>1</votes>
                                    <watches>13</watches>
                                                                            <comments>
                            <comment id="245014" author="bzzz" created="Mon, 1 Apr 2019 10:51:31 +0000"  >&lt;p&gt;please, attach logs from OST.&lt;/p&gt;</comment>
                            <comment id="245186" author="jgmitter" created="Wed, 3 Apr 2019 14:00:32 +0000"  >&lt;p&gt;Hi Stephane,&lt;/p&gt;

&lt;p&gt;Any luck in getting the OST logs as requested?&lt;/p&gt;

&lt;p&gt;Thanks.&lt;/p&gt;

&lt;p&gt;Joe&lt;/p&gt;</comment>
                            <comment id="245209" author="sthiell" created="Wed, 3 Apr 2019 18:01:02 +0000"  >&lt;p&gt;Hi Alex and Joe &#8211; I just attached the OSS kernel logs, thanks!&lt;br/&gt;
Note: We upgraded Oak servers to 2.10.7 last night.&lt;/p&gt;</comment>
                            <comment id="255923" author="adilger" created="Fri, 4 Oct 2019 18:25:47 +0000"  >&lt;p&gt;Alex, did you ever get a chance to look into this code?  It seems like we should not be holding this global mutex over RPC calls. If the OST is unavailable, or doesn&apos;t have any objects, then the call should trigger the precreated thread and return immediately, rather than blocking the MDS thread. &lt;/p&gt;</comment>
                            <comment id="256002" author="bzzz" created="Mon, 7 Oct 2019 16:57:04 +0000"  >&lt;p&gt;hmm, the cause for reboot was LBUG, AFAICS:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Mar 29 10:28:48 oak-md1-s2 kernel: LustreError: 118086:0:(osd_handler.c:1982:osd_object_release()) LBUG
Mar 29 10:28:48 oak-md1-s2 kernel: Pid: 118086, comm: tgt_recover_0 3.10.0-693.2.2.el7_lustre.pl1.x86_64 #1 SMP Sat Sep 30 00:44:25 PDT 2017
Mar 29 10:28:48 oak-md1-s2 kernel: Call Trace:
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffff8103a212&amp;gt;] save_stack_trace_tsk+0x22/0x40
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc08247cc&amp;gt;] libcfs_call_trace+0x8c/0xc0 [libcfs]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc082487c&amp;gt;] lbug_with_loc+0x4c/0xa0 [libcfs]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc0fe3660&amp;gt;] osd_get_ldiskfs_dirent_param+0x0/0x130 [osd_ldiskfs]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc096ac4d&amp;gt;] lu_object_put+0x17d/0x3d0 [obdclass]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc0c0e738&amp;gt;] distribute_txn_replay_handle+0x418/0xc00 [ptlrpc]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b54987&amp;gt;] replay_request_or_update.isra.21+0x247/0x8c0 [ptlrpc]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffffc0b55616&amp;gt;] target_recovery_thread+0x616/0x1360 [ptlrpc]
Mar 29 10:28:48 oak-md1-s2 kernel: [&amp;lt;ffffffff810b098f&amp;gt;] kthread+0xcf/0xe0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="256007" author="bzzz" created="Mon, 7 Oct 2019 18:10:15 +0000"  >&lt;p&gt;Andreas, normally we don&apos;t wait on RPC in that path - when specific OST is out of precreated objects, but once we hit it then the calling thread wats for obd_timeout..&lt;/p&gt;</comment>
                            <comment id="256971" author="dauchy" created="Wed, 23 Oct 2019 17:10:18 +0000"  >&lt;p&gt;Any update on getting a fix for this issue?&#160; We have hit it at NOAA at least twice, with no sign of any LBUG.&#160; This is with lustre-2.10.7_ddn8.&#160; Here is the call trace in case it helps:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Oct 21 23:08:11 lfs-mds-1-1 kernel: INFO: task mdt00_002:17259 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.
Oct 21 23:08:11 lfs-mds-1-1 kernel: &lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.
Oct 21 23:08:11 lfs-mds-1-1 kernel: mdt00_002       D ffff96927956d140     0 17259      2 0x00000000
Oct 21 23:08:11 lfs-mds-1-1 kernel: Call Trace:
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa4b68e19&amp;gt;] schedule+0x29/0x70
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa4b6a705&amp;gt;] rwsem_down_write_failed+0x225/0x3a0
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa47874a7&amp;gt;] call_rwsem_down_write_failed+0x17/0x30
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa4b6814d&amp;gt;] down_write+0x2d/0x3d
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1624cef&amp;gt;] lod_alloc_qos.constprop.17+0x1af/0x1590 [lod]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc12d7a47&amp;gt;] ? qsd_op_begin0+0xc7/0x940 [lquota]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa467966d&amp;gt;] ? __getblk+0x2d/0x300
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1628d51&amp;gt;] lod_qos_prep_create+0x12b1/0x1810 [lod]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1629479&amp;gt;] ? lod_prepare_inuse+0x1c9/0x2e0 [lod]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc16297ed&amp;gt;] lod_prepare_create+0x25d/0x360 [lod]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc161e11e&amp;gt;] lod_declare_striped_create+0x1ee/0x970 [lod]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1620594&amp;gt;] lod_declare_create+0x1e4/0x550 [lod]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc168dc72&amp;gt;] mdd_declare_create_object_internal+0xe2/0x2f0 [mdd]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc167d1a3&amp;gt;] mdd_declare_create+0x53/0xe40 [mdd]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1681739&amp;gt;] mdd_create+0x9c9/0x1420 [mdd]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1552685&amp;gt;] mdt_reint_open+0x2175/0x3190 [mdt]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0ff2037&amp;gt;] ? lustre_msg_add_version+0x27/0xa0 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1547403&amp;gt;] mdt_reint_rec+0x83/0x210 [mdt]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc152818b&amp;gt;] mdt_reint_internal+0x5fb/0x9c0 [mdt]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc15286b2&amp;gt;] mdt_intent_reint+0x162/0x480 [mdt]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc152b50b&amp;gt;] mdt_intent_opc+0x1eb/0xaf0 [mdt]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0ff4060&amp;gt;] ? lustre_swab_ldlm_policy_data+0x30/0x30 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1533e70&amp;gt;] mdt_intent_policy+0x130/0x320 [mdt]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0fa2226&amp;gt;] ldlm_lock_enqueue+0x396/0x960 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0fcbcd7&amp;gt;] ldlm_handle_enqueue0+0xa97/0x1640 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0ff40e0&amp;gt;] ? lustre_swab_ldlm_lock_desc+0x30/0x30 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1054882&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc10587ba&amp;gt;] tgt_request_handle+0x92a/0x1370 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0ffe12b&amp;gt;] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc0ffa758&amp;gt;] ? ptlrpc_wait_event+0x98/0x340 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa44d6af2&amp;gt;] ? default_wake_function+0x12/0x20
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa44cbb8b&amp;gt;] ? __wake_up_common+0x5b/0x90
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1001872&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffc1000de0&amp;gt;] ? ptlrpc_register_service+0xe30/0xe30 [ptlrpc]
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa44c1d21&amp;gt;] kthread+0xd1/0xe0
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa44c1c50&amp;gt;] ? insert_kthread_work+0x40/0x40
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa4b75c37&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
Oct 21 23:08:11 lfs-mds-1-1 kernel:  [&amp;lt;ffffffffa44c1c50&amp;gt;] ? insert_kthread_work+0x40/0x40
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="256976" author="sthiell" created="Wed, 23 Oct 2019 17:41:16 +0000"  >&lt;p&gt;Hi Nathan,&lt;/p&gt;

&lt;p&gt;Just FYI, we&apos;ve not seen this problem on our 2.10 servers since we use 2.10.8 + 3 patches (&lt;a href=&quot;https://github.com/stanford-rc/lustre/commits/b2_10_oak&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/stanford-rc/lustre/commits/b2_10_oak&lt;/a&gt;). I&apos;m wondering if you have &quot;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12018&quot; title=&quot;deadlock on OSS: quota reintegration vs memory release&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12018&quot;&gt;&lt;del&gt;LU-12018&lt;/del&gt;&lt;/a&gt; quota: do not start a thread under memory pressure&quot;. It&apos;s a critical patch that is missing from 2.10.8 IMHO.&lt;/p&gt;</comment>
                            <comment id="268974" author="gerrit" created="Thu, 30 Apr 2020 13:28:10 +0000"  >&lt;p&gt;Alex Zhuravlev (bzzz@whamcloud.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/38428&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/38428&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12136&quot; title=&quot;mdt threads blocked in mdd_create&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12136&quot;&gt;LU-12136&lt;/a&gt; osp: blocked-in-precreation stats&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: a3c054c4b39c49a750666698c52385d4f7133a98&lt;/p&gt;</comment>
                            <comment id="349134" author="bzzz" created="Mon, 10 Oct 2022 15:13:57 +0000"  >&lt;p&gt;the number of improvements have been made in this area (&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8367&quot; title=&quot;delete orphan phase isn&amp;#39;t stated for multistriped file&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8367&quot;&gt;&lt;del&gt;LU-8367&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15082&quot; title=&quot;osp can stop refreshing statfs data&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15082&quot;&gt;&lt;del&gt;LU-15082&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15724&quot; title=&quot;MDT failover hang&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15724&quot;&gt;&lt;del&gt;LU-15724&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-13073&quot; title=&quot;Multiple MDS deadlocks (in lod_qos_prep_create) after OSS crash&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-13073&quot;&gt;&lt;del&gt;LU-13073&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-13513&quot; title=&quot;osp_precreate_reserve() should not take d-&amp;gt;opd_pre_status == -EIO as fatal&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-13513&quot;&gt;&lt;del&gt;LU-13513&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12397&quot; title=&quot;osp: race around opd_new_connection&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12397&quot;&gt;&lt;del&gt;LU-12397&lt;/del&gt;&lt;/a&gt;, etc). probably makes sense to check is this still a problem with the newer versions?&lt;/p&gt;</comment>
                            <comment id="349191" author="sthiell" created="Mon, 10 Oct 2022 23:29:44 +0000"  >&lt;p&gt;Thanks Alex! Agreed that this is an old ticket and is probably not relevant anymore.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="50900">LU-10697</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="32379" name="LU-12136-OSS-logs.tar.gz" size="49948" author="sthiell" created="Wed, 3 Apr 2019 17:58:05 +0000"/>
                            <attachment id="32352" name="oak-md1-s2-kernel.log" size="460427" author="sthiell" created="Sat, 30 Mar 2019 17:03:15 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i00e8v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>