<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:53:13 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-12510] mds server hangs cv_wait_common</title>
                <link>https://jira.whamcloud.com/browse/LU-12510</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We are hitting an issue after the upgrade to lustre-2.12.2 last weekend where the MDS servers will start to hang, and the lustre_health check will report NOT HEALTHY because of long waiting requests. We think this is related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10250&quot; title=&quot;replay-single test_74:  hang and timed out&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10250&quot;&gt;LU-10250&lt;/a&gt;, but wanted to create a seperate ticket to track this to push the priority up. It is taking an MDS down about once a day causing production outages.&lt;/p&gt;

&lt;p&gt;2019-07-05T02:16:36.434782-04:00 f2-mds2.ncrc.gov kernel: Lustre: f2-OST0035-osc-MDT0001: Connection to f2-OST0035 (at 10.10.33.50@o2ib2) was l&lt;/p&gt;

&lt;p&gt;ost; in progress operations using this service will wait for recovery to complete&lt;/p&gt;

&lt;p&gt;2019-07-05T02:17:26.605024-04:00 f2-mds2.ncrc.gov kernel: Lustre: f2-OST0035-osc-MDT0001: Connection restored to 10.10.33.50@o2ib2 (at 10.10.33.50@o2ib2)&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.360456-04:00 f2-mds2.ncrc.gov kernel: INFO: task txg_quiesce:40218 blocked for more than 120 seconds.&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.360500-04:00 f2-mds2.ncrc.gov kernel: &quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.371558-04:00 f2-mds2.ncrc.gov kernel: txg_quiesce &#160; &#160; D ffff99aec932a080 &#160; &#160; 0 40218&#160; &#160; &#160; 2 0x00000000&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.371580-04:00 f2-mds2.ncrc.gov kernel: Call Trace:&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.371599-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff83967c49&amp;gt;&amp;#93;&lt;/span&gt; schedule+0x29/0x70&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.384261-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0995325&amp;gt;&amp;#93;&lt;/span&gt; cv_wait_common+0x125/0x150 &lt;span class=&quot;error&quot;&gt;&amp;#91;spl&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.384280-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff832c2d40&amp;gt;&amp;#93;&lt;/span&gt; ? wake_up_atomic_t+0x30/0x30&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.397218-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0995365&amp;gt;&amp;#93;&lt;/span&gt; __cv_wait+0x15/0x20 &lt;span class=&quot;error&quot;&gt;&amp;#91;spl&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.397238-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0bc863b&amp;gt;&amp;#93;&lt;/span&gt; txg_quiesce_thread+0x2cb/0x3c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;zfs&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.411080-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0bc8370&amp;gt;&amp;#93;&lt;/span&gt; ? txg_init+0x2b0/0x2b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;zfs&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.411101-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc099cb93&amp;gt;&amp;#93;&lt;/span&gt; thread_generic_wrapper+0x73/0x80 &lt;span class=&quot;error&quot;&gt;&amp;#91;spl&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.425331-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc099cb20&amp;gt;&amp;#93;&lt;/span&gt; ? __thread_exit+0x20/0x20 &lt;span class=&quot;error&quot;&gt;&amp;#91;spl&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.425350-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff832c1c71&amp;gt;&amp;#93;&lt;/span&gt; kthread+0xd1/0xe0&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.430937-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff832c1ba0&amp;gt;&amp;#93;&lt;/span&gt; ? insert_kthread_work+0x40/0x40&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.437770-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff83974c1d&amp;gt;&amp;#93;&lt;/span&gt; ret_from_fork_nospec_begin+0x7/0x21&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.444951-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff832c1ba0&amp;gt;&amp;#93;&lt;/span&gt; ? insert_kthread_work+0x40/0x40&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.459337-04:00 f2-mds2.ncrc.gov kernel: INFO: task mdt04_000:42947 blocked for more than 120 seconds.&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.459357-04:00 f2-mds2.ncrc.gov kernel: &quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.478994-04:00 f2-mds2.ncrc.gov kernel: mdt04_000 &#160; &#160; &#160; D ffff99aecc7ae180 &#160; &#160; 0 42947&#160; &#160; &#160; 2 0x00000000&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.479015-04:00 f2-mds2.ncrc.gov kernel: Call Trace:&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.479037-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff83967c49&amp;gt;&amp;#93;&lt;/span&gt; schedule+0x29/0x70&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.491665-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0995325&amp;gt;&amp;#93;&lt;/span&gt; cv_wait_common+0x125/0x150 &lt;span class=&quot;error&quot;&gt;&amp;#91;spl&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.491685-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff832c2d40&amp;gt;&amp;#93;&lt;/span&gt; ? wake_up_atomic_t+0x30/0x30&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.504560-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0995365&amp;gt;&amp;#93;&lt;/span&gt; __cv_wait+0x15/0x20 &lt;span class=&quot;error&quot;&gt;&amp;#91;spl&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.504580-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0b675fb&amp;gt;&amp;#93;&lt;/span&gt; dmu_tx_wait+0x20b/0x3b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;zfs&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.517986-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc0b67831&amp;gt;&amp;#93;&lt;/span&gt; dmu_tx_assign+0x91/0x490 &lt;span class=&quot;error&quot;&gt;&amp;#91;zfs&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.518007-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc176d049&amp;gt;&amp;#93;&lt;/span&gt; osd_trans_start+0x199/0x440 &lt;span class=&quot;error&quot;&gt;&amp;#91;osd_zfs&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.532397-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc18ce837&amp;gt;&amp;#93;&lt;/span&gt; mdt_empty_transno+0xf7/0x850 &lt;span class=&quot;error&quot;&gt;&amp;#91;mdt&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.532416-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc18d1eee&amp;gt;&amp;#93;&lt;/span&gt; mdt_mfd_open+0x8de/0xe70 &lt;span class=&quot;error&quot;&gt;&amp;#91;mdt&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.546398-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc18a4ba2&amp;gt;&amp;#93;&lt;/span&gt; ? mdt_pack_acl2body+0x1c2/0x9f0 &lt;span class=&quot;error&quot;&gt;&amp;#91;mdt&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;2019-07-05T02:28:56.546419-04:00 f2-mds2.ncrc.gov kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffc18d2acb&amp;gt;&amp;#93;&lt;/span&gt; mdt_finish_open+0x64b/0x760 &lt;span class=&quot;error&quot;&gt;&amp;#91;mdt&amp;#93;&lt;/span&gt;&lt;/p&gt;</description>
                <environment>RHEL7.6, lustre-2.12.2, ZFS-0.8.1-1</environment>
        <key id="56275">LU-12510</key>
            <summary>mds server hangs cv_wait_common</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="dustb100">Dustin Leverman</reporter>
                        <labels>
                            <label>LLNL</label>
                            <label>ORNL</label>
                    </labels>
                <created>Fri, 5 Jul 2019 13:52:52 +0000</created>
                <updated>Sat, 8 May 2021 15:11:03 +0000</updated>
                            <resolved>Wed, 24 Jul 2019 20:10:16 +0000</resolved>
                                    <version>Lustre 2.12.2</version>
                                    <fixVersion>Lustre 2.13.0</fixVersion>
                    <fixVersion>Lustre 2.12.3</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>17</watches>
                                                                            <comments>
                            <comment id="250723" author="dustb100" created="Fri, 5 Jul 2019 14:11:56 +0000"  >&lt;p&gt;We are working on fixing an issue with kdump, and when this crashes again tonight, we plan on capturing a kdump.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Thanks,&lt;/p&gt;

&lt;p&gt;Dustin&lt;/p&gt;</comment>
                            <comment id="250725" author="simmonsja" created="Fri, 5 Jul 2019 14:27:47 +0000"  >&lt;p&gt;Opened&#160;&lt;a href=&quot;https://github.com/zfsonlinux/zfs/issues/8994&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/zfsonlinux/zfs/issues/8994&lt;/a&gt;&#160;for the ZFS guys as well.&lt;/p&gt;</comment>
                            <comment id="250767" author="pjones" created="Fri, 5 Jul 2019 17:59:48 +0000"  >&lt;p&gt;Alex&lt;/p&gt;

&lt;p&gt;Could you please investigate?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="250771" author="bzzz" created="Fri, 5 Jul 2019 18:14:30 +0000"  >&lt;p&gt;I&apos;m dowloading the archive.. will take some time. there must be another thread stuck somewhere else, but holding TXG open (and preventing new TXG). it would be great to have all backtraces and/or the crashdump.&lt;/p&gt;</comment>
                            <comment id="250772" author="dustb100" created="Fri, 5 Jul 2019 18:21:09 +0000"  >&lt;p&gt;Thanks Alex!&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;I plan on gathering the crashdump next time this happens along with some perf top tracing of the txg_quiesce process.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;-Dustin&lt;/p&gt;</comment>
                            <comment id="250773" author="adilger" created="Fri, 5 Jul 2019 18:34:39 +0000"  >&lt;p&gt;Dustin, grabbing &quot;&lt;tt&gt;echo t &amp;gt; /proc/sysrq-trigger&lt;/tt&gt;&quot; would be useful when the threads are stuck, before you trigger the crashdump, since that will be easier/possible to attach to the ticket and may provide some information about why they are stuck.&lt;/p&gt;</comment>
                            <comment id="250778" author="dustb100" created="Fri, 5 Jul 2019 20:03:21 +0000"  >&lt;p&gt;I&apos;ll make sure to grab that as well. Thanks!&lt;/p&gt;</comment>
                            <comment id="250783" author="pjones" created="Fri, 5 Jul 2019 21:49:04 +0000"  >&lt;p&gt;BTW do you have any patches applied to 2.12.2 or are you just running the vanilla release?&lt;/p&gt;</comment>
                            <comment id="250803" author="simmonsja" created="Sun, 7 Jul 2019 13:42:40 +0000"  >&lt;p&gt;We are running&#160; vanilla 2.12 at git commit&#160;&lt;/p&gt;

&lt;p&gt;v2_12_2-30-g989217d.&lt;/p&gt;

&lt;p&gt;with ZFS 0.8.1 release&lt;/p&gt;</comment>
                            <comment id="250824" author="dustb100" created="Mon, 8 Jul 2019 12:17:39 +0000"  >&lt;p&gt;We had another crash last night, and I gathered some debug data: kernel traces from sysrq, kernel logs, perf record data, and a couple of other things. There is still an outstanding issue with kdump, so I wasn&apos;t able to get the crashdump.&lt;/p&gt;

&lt;p&gt;Please let us know if there are additional things you would like for us to gather during the next crash.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Thanks,&lt;/p&gt;

&lt;p&gt;Dustin&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;nobr&quot;&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/attachment/33082/33082_f2-mds4_lustre_unhealthy_20190707.tar.gz&quot; title=&quot;f2-mds4_lustre_unhealthy_20190707.tar.gz attached to LU-12510&quot;&gt;f2-mds4_lustre_unhealthy_20190707.tar.gz&lt;sup&gt;&lt;img class=&quot;rendericon&quot; src=&quot;https://jira.whamcloud.com/images/icons/link_attachment_7.gif&quot; height=&quot;7&quot; width=&quot;7&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/sup&gt;&lt;/a&gt;&lt;/span&gt;&lt;/p&gt;</comment>
                            <comment id="250907" author="curtispb" created="Tue, 9 Jul 2019 17:11:27 +0000"  >&lt;p&gt;We encountered this issue again on this same mds. We were able to resolve the issue with the crashdump in between and I have gathered some other recommended debug info.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Thanks,&lt;/p&gt;

&lt;p&gt;Philip&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;nobr&quot;&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/attachment/33092/33092_f2-mds4_20190709.tgz&quot; title=&quot;f2-mds4_20190709.tgz attached to LU-12510&quot;&gt;f2-mds4_20190709.tgz&lt;sup&gt;&lt;img class=&quot;rendericon&quot; src=&quot;https://jira.whamcloud.com/images/icons/link_attachment_7.gif&quot; height=&quot;7&quot; width=&quot;7&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/sup&gt;&lt;/a&gt;&lt;/span&gt;&lt;/p&gt;</comment>
                            <comment id="250910" author="bzzz" created="Tue, 9 Jul 2019 17:21:48 +0000"  >&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
[ 3857.825288]  [&amp;lt;ffffffffc12937a3&amp;gt;] dnode_hold_impl+0x873/0xcb0 [zfs]
[ 3857.832059]  [&amp;lt;ffffffffc127d256&amp;gt;] dmu_object_alloc_impl+0x216/0x400 [zfs]
[ 3857.839339]  [&amp;lt;ffffffffc127d4f9&amp;gt;] dmu_object_alloc_dnsize+0x39/0x40 [zfs]
[ 3857.846591]  [&amp;lt;ffffffffc18b9662&amp;gt;] __osd_object_create+0x82/0x170 [osd_zfs]
[ 3857.853934]  [&amp;lt;ffffffffc18b99cd&amp;gt;] osd_mkreg+0x7d/0x210 [osd_zfs]
[ 3857.860402]  [&amp;lt;ffffffffc18b5cd6&amp;gt;] osd_create+0x346/0xb20 [osd_zfs]
[ 3857.867040]  [&amp;lt;ffffffffc1acd7e5&amp;gt;] lod_sub_create+0x1f5/0x480 [lod]
[ 3857.873688]  [&amp;lt;ffffffffc1abe049&amp;gt;] lod_create+0x69/0x340 [lod]
[ 3857.879895]  [&amp;lt;ffffffffc1b37ef8&amp;gt;] mdd_create_object_internal+0xb8/0x280 [mdd]
[ 3857.887486]  [&amp;lt;ffffffffc1b217e5&amp;gt;] mdd_create_object+0x75/0x820 [mdd]
[ 3857.894292]  [&amp;lt;ffffffffc1b2bd90&amp;gt;] mdd_create+0xe00/0x14b0 [mdd]
[ 3857.900663]  [&amp;lt;ffffffffc19faf60&amp;gt;] mdt_reint_open+0x19d0/0x27d0 [mdt]
[ 3857.907475]  [&amp;lt;ffffffffc19edfa3&amp;gt;] mdt_reint_rec+0x83/0x210 [mdt]
[ 3857.913933]  [&amp;lt;ffffffffc19cc1b3&amp;gt;] mdt_reint_internal+0x6e3/0xaf0 [mdt]
[ 3857.920902]  [&amp;lt;ffffffffc19d8972&amp;gt;] mdt_intent_open+0x82/0x3a0 [mdt]
[ 3857.927524]  [&amp;lt;ffffffffc19d6a18&amp;gt;] mdt_intent_policy+0x2e8/0xd00 [mdt]
[ 3857.934406]  [&amp;lt;ffffffffc167dd26&amp;gt;] ldlm_lock_enqueue+0x366/0xa60 [ptlrpc]
[ 3857.941582]  [&amp;lt;ffffffffc16a6587&amp;gt;] ldlm_handle_enqueue0+0xa47/0x15a0 [ptlrpc]
[ 3857.949097]  [&amp;lt;ffffffffc172e882&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;seem to be related to &lt;a href=&quot;https://github.com/zfsonlinux/zfs/issues/8994&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/zfsonlinux/zfs/issues/8994&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="250916" author="dustb100" created="Tue, 9 Jul 2019 17:32:30 +0000"  >&lt;p&gt;Alex,&#160;&lt;/p&gt;

&lt;p&gt;&#160;&#160;&#160;&#160;&#160;&#160;&#160;&#160; James created a ticket with the ZFS developers as well to get their eyes on it in addition to yours.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;-Dustin&lt;/p&gt;</comment>
                            <comment id="250918" author="simmonsja" created="Tue, 9 Jul 2019 17:47:19 +0000"  >&lt;p&gt;Yes its the same issue. I have a earlier comment&#160; that states I opened that ticket on zfsforlinux issue tracker.&lt;/p&gt;</comment>
                            <comment id="250957" author="simmonsja" created="Wed, 10 Jul 2019 16:46:27 +0000"  >&lt;p&gt;So a potential work around is setting&#160;dnodesize=legacy. Any opinions on this? Any know issues if this is done?&lt;/p&gt;</comment>
                            <comment id="250961" author="dustb100" created="Wed, 10 Jul 2019 18:12:24 +0000"  >&lt;p&gt;To add to what James is asking in the previous message, we are considering testing the recommendation from the ZFS folks (Brian) to set&#160; dnodesize=legacy to reduce lock contention. We wanted to run this by Whamcloud first since you are hold our support contract.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;-Dustin&lt;/p&gt;</comment>
                            <comment id="250975" author="curtispb" created="Wed, 10 Jul 2019 21:08:58 +0000"  >&lt;p&gt;As we have just hit this issue again, it would be a good time to test this change if it is safe to make and see if it keeps it from happening.&lt;/p&gt;

&lt;p&gt;-Philip&lt;/p&gt;</comment>
                            <comment id="250982" author="behlendorf" created="Wed, 10 Jul 2019 22:25:15 +0000"  >&lt;p&gt;&amp;gt; So a potential work around is setting dnodesize=legacy. Any opinions on this? Any know issues if this is done?&lt;/p&gt;

&lt;p&gt;From the ZFS side the consequences of changing this would be that we don&apos;t store Lustre&apos;s xattrs along with the dnode on disk for newly allocated objects.&#160; These objects (for as long as they exist) will incur an additional I/O to access the xattrs (even Lustre&apos;s internal ones).&#160; For an OST that&apos;s going to be less of a performance limitation than on a MDT.&#160; The additional tiny blocks to store the xattrs can contribute to fragmenting free space too.&#160; There may be additional considerations at the Lustre level.&lt;/p&gt;

&lt;p&gt;I suggested this as possible stop gap which doesn&apos;t require any code change until we have a proper fix.&#160; I&apos;m working on this now.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="251093" author="curtispb" created="Thu, 11 Jul 2019 15:54:18 +0000"  >&lt;p&gt;I updated the zfs issue, but I will update here as well. It looks like setting the dnodesize=legacy on the MDS hosts that were hitting this issue may not be working. We just hit it again on the f2-mds4 host.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;One thought I had, is that dnodesize something that can be set on the fly or is it something that I would need to reimport to take effect? I did not do that last night when I set it.&lt;/p&gt;</comment>
                            <comment id="251157" author="behlendorf" created="Thu, 11 Jul 2019 23:05:51 +0000"  >&lt;p&gt;Alex, I believe I have spotted the problem.&#160; Can you take a look at this to make sure I correctly understand the design and the analysis makes sense.&lt;/p&gt;

&lt;p&gt;As I understand it the Lustre osd-zfs code always acquires the zrl lock for a dnode handle when it&apos;s created, via osd_create(), or when an existing object needs to be initialized, via osd_object_init().&#160; It then doesn&apos;t release the zrl lock until osd_object_delete() is called to release the resources.&#160; Which means the zrl lock in the dnode handle can, and likely will, be held for a long time.&lt;/p&gt;

&lt;p&gt;Normally, this doesn&apos;t cause a problem since DB_DNODE_ENTER (which takes the&#160; lock) is called after you have a pointer to the dnode and Lustre won&apos;t try and hold it again.&#160; Doing so would result in exactly this deadlock if two process concurrently tried to create or initialize the same dnode.&#160; This is the situation described by the comment in the&#160;DNODE_MUST_BE_ALLOCATED case of dnode_hold_impl().&#160; However, that logic depends on the unwritten assumption that the dnode handle&apos;s zrl lock will only even be held long enough to access the dnh_dnode field.&#160;&#160; As is the case in the rest of the ZFS code.&lt;/p&gt;

&lt;p&gt;I&apos;d suggest that the best way to resolve this issue is to update the osd-zfs so that it never directly manipulates the zrl locks.&#160; The dnode_hold and dnode_rele functions are the intended interfaces to ensure that a pointer to a dnode remains valid.&#160; It looks like perhaps DB_DNODE_ENTER and DB_DNODE_EXIT were used originally since the dnode_ functions were not exported.&#160; In fact, dnode_rele() still doesn&apos;t appear to be exported but it definitely can be along with any other symbols we need.&lt;/p&gt;

&lt;p&gt;Putting that change together will take some time, if a quicker fix is needed we could perhaps consider using the existing obj-&amp;gt;oo_guard to prevent this.&#160; Though, I&apos;d expect that concurrent allocations of new objects would still be vulnerable.&lt;/p&gt;

&lt;p&gt;Additionally, I did find one code path where osd_dnode_rele() is not called for an error condition in osd_mkreg().&#160; In which case the zrl lock won&apos;t be released.&#160; That would also explain the symptoms we&apos;re seeing and should be fixed, but I didn&apos;t find the matching &quot;can&apos;t change blocksize:&quot; CERROR in the console log, so I&apos;m pretty sure we didn&apos;t hit this.&lt;/p&gt;</comment>
                            <comment id="251218" author="bzzz" created="Fri, 12 Jul 2019 07:34:41 +0000"  >&lt;p&gt;thanks, Brian. working on the patch...&lt;/p&gt;</comment>
                            <comment id="251283" author="behlendorf" created="Fri, 12 Jul 2019 18:07:13 +0000"  >&lt;p&gt;Alex I&apos;ve opened PR &lt;a href=&quot;https://github.com/zfsonlinux/zfs/pull/9027&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/zfsonlinux/zfs/pull/9027&lt;/a&gt; which exports the dnode symbols you&apos;ll need, plus a few additional ones I thought you might find useful.&#160; Let me know if there are additional symbols I should add to that PR.&#160; A previous commit has already exported dmu_object_alloc_hold() and&#160; zap_create_hold() which return the held dnode as part of the create.&lt;/p&gt;</comment>
                            <comment id="251309" author="simmonsja" created="Fri, 12 Jul 2019 21:11:51 +0000"  >&lt;p&gt;We are trying to reproduce this in our test bed and currently we can&apos;t seem to reproduce this easily. Can you recommend a setup that would expose this issue easily. I think this is mostly due to the small scale we are running in the test bed.&lt;/p&gt;</comment>
                            <comment id="251449" author="bzzz" created="Tue, 16 Jul 2019 03:41:22 +0000"  >&lt;p&gt;please, try &lt;a href=&quot;https://review.whamcloud.com/#/c/35524/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/#/c/35524/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="251478" author="behlendorf" created="Tue, 16 Jul 2019 15:39:47 +0000"  >&lt;p&gt;James, I&apos;d expect that a large parallel create and unlink workload on an aged mdt would have the best chance of hitting this.&lt;/p&gt;</comment>
                            <comment id="251717" author="simmonsja" created="Fri, 19 Jul 2019 16:25:34 +0000"  >&lt;p&gt;So far patch 35524 is holding up well with our production file system. The current patch is a work around until a proper solution is done so don&apos;t close this ticket once the patch lands.&lt;/p&gt;</comment>
                            <comment id="251831" author="pfarrell" created="Mon, 22 Jul 2019 19:53:49 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=bzzz&quot; class=&quot;user-hover&quot; rel=&quot;bzzz&quot;&gt;bzzz&lt;/a&gt;,&lt;/p&gt;

&lt;p&gt;I&apos;m wondering if you think the failure here:&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/612acaa2-ac97-11e9-a0be-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/612acaa2-ac97-11e9-a0be-52540065bddc&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Which is due to the client being evicted from OST0, which happens because OST0 is stuck like this:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[ 9758.094925] Lustre: ll_ost00_016: service thread pid 2789 was inactive for 40.093 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one.
[ 9767.054545] Lustre: ll_ost00_073: service thread pid 3361 was inactive for 40.069 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[ 9767.054551] Pid: 3233, comm: ll_ost00_055 3.10.0-957.21.3.el7_lustre.x86_64 #1 SMP Tue Jul 16 17:30:03 UTC 2019
[ 9767.054551] Call Trace:
[ 9767.054645]  [&amp;lt;ffffffffc02ee2d5&amp;gt;] cv_wait_common+0x125/0x150 [spl]
[ 9767.054650]  [&amp;lt;ffffffffc02ee315&amp;gt;] __cv_wait+0x15/0x20 [spl]
[ 9767.054685]  [&amp;lt;ffffffffc04542bf&amp;gt;] txg_wait_synced+0xef/0x140 [zfs]
[ 9767.054736]  [&amp;lt;ffffffffc112c9ab&amp;gt;] osd_trans_stop+0x53b/0x5e0 [osd_zfs]
[ 9767.054771]  [&amp;lt;ffffffffc1270cd5&amp;gt;] ofd_trans_stop+0x25/0x60 [ofd]
[ 9767.054776]  [&amp;lt;ffffffffc12752c5&amp;gt;] ofd_destroy+0x2c5/0x960 [ofd]
[ 9767.054780]  [&amp;lt;ffffffffc126d534&amp;gt;] ofd_destroy_by_fid+0x1f4/0x4a0 [ofd]
[ 9767.054784]  [&amp;lt;ffffffffc1262057&amp;gt;] ofd_destroy_hdl+0x267/0x970 [ofd]
[ 9767.055141]  [&amp;lt;ffffffffc0f5e51a&amp;gt;] tgt_request_handle+0x91a/0x15c0 [ptlrpc]
[ 9767.055179]  [&amp;lt;ffffffffc0f03646&amp;gt;] ptlrpc_server_handle_request+0x256/0xb00 [ptlrpc]
[ 9767.055212]  [&amp;lt;ffffffffc0f0717c&amp;gt;] ptlrpc_main+0xbac/0x1560 [ptlrpc]
[ 9767.055252]  [&amp;lt;ffffffffa9cc1da1&amp;gt;] kthread+0xd1/0xe0
[ 9767.055280]  [&amp;lt;ffffffffaa375c37&amp;gt;] ret_from_fork_nospec_end+0x0/0x39
[ 9767.055288]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;We&apos;ve seen similar failures in a few tests recently, where the force_sync command apparently takes forever.&#160; I&apos;m guessing they have the same detailed signature as this.&lt;/p&gt;</comment>
                            <comment id="251855" author="bzzz" created="Tue, 23 Jul 2019 03:43:21 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=pfarrell&quot; class=&quot;user-hover&quot; rel=&quot;pfarrell&quot;&gt;pfarrell&lt;/a&gt; hard to say, but I&apos;d not very likely, few reported processes weren&apos;t running, but got stuck with IO:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
[16250.115937] Pid: 2829, comm: ll_ost00_029 3.10.0-957.21.3.el7_lustre.x86_64 #1 SMP Tue Jul 16 17:30:03 UTC 2019
[16250.117636] Call Trace:
[16250.118155]  [&amp;lt;ffffffffc02ee262&amp;gt;] cv_wait_common+0xb2/0x150 [spl]
[16250.119447]  [&amp;lt;ffffffffc02ee338&amp;gt;] __cv_wait_io+0x18/0x20 [spl]
[16250.120491]  [&amp;lt;ffffffffc04aca3b&amp;gt;] zio_wait+0x11b/0x1c0 [zfs]
[16250.121621]  [&amp;lt;ffffffffc03ebfce&amp;gt;] dbuf_read+0x67e/0x9e0 [zfs]
[16250.122689]  [&amp;lt;ffffffffc03f8060&amp;gt;] dmu_buf_hold_by_dnode+0x50/0x80 [zfs]
[16250.123910]  [&amp;lt;ffffffffc046b6c2&amp;gt;] zap_get_leaf_byblk.isra.11+0x92/0x260 [zfs]
[16250.125193]  [&amp;lt;ffffffffc046b978&amp;gt;] zap_deref_leaf+0xe8/0x100 [zfs]
[16250.126336]  [&amp;lt;ffffffffc046c9b2&amp;gt;] fzap_lookup+0x62/0x180 [zfs]
[16250.127455]  [&amp;lt;ffffffffc04712b1&amp;gt;] zap_lookup_impl+0xd1/0x200 [zfs]
[16250.128629]  [&amp;lt;ffffffffc0471c2e&amp;gt;] zap_lookup_norm+0x8e/0xd0 [zfs]
[16250.129773]  [&amp;lt;ffffffffc0471ca3&amp;gt;] zap_lookup+0x33/0x40 [zfs]
[16250.130823]  [&amp;lt;ffffffffc113d4c1&amp;gt;] osd_fid_lookup+0x3e1/0x4c0 [osd_zfs]
[16250.132040]  [&amp;lt;ffffffffc1135669&amp;gt;] osd_object_init+0x389/0xca0 [osd_zfs]
[16250.133221]  [&amp;lt;ffffffffc0c17b25&amp;gt;] lu_object_alloc+0xe5/0x320 [obdclass]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="251911" author="gerrit" created="Wed, 24 Jul 2019 04:20:52 +0000"  >&lt;p&gt;Oleg Drokin (green@whamcloud.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/35524/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/35524/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12510&quot; title=&quot;mds server hangs cv_wait_common&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12510&quot;&gt;&lt;del&gt;LU-12510&lt;/del&gt;&lt;/a&gt; osd: osd-zfs to release zrlock quickly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 88b329ac2ab568a25251f3f7c3a7e0c7367cb36f&lt;/p&gt;</comment>
                            <comment id="251926" author="gerrit" created="Wed, 24 Jul 2019 13:08:45 +0000"  >&lt;p&gt;James Simmons (jsimmons@infradead.org) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/35600&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/35600&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12510&quot; title=&quot;mds server hangs cv_wait_common&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12510&quot;&gt;&lt;del&gt;LU-12510&lt;/del&gt;&lt;/a&gt; osd: osd-zfs to release zrlock quickly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_12&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: d844b66bf8d58708237547598d016b356a07a01d&lt;/p&gt;</comment>
                            <comment id="251975" author="pjones" created="Wed, 24 Jul 2019 20:10:16 +0000"  >&lt;p&gt;Landed for 2.13&lt;/p&gt;</comment>
                            <comment id="252015" author="simmonsja" created="Thu, 25 Jul 2019 13:39:00 +0000"  >&lt;p&gt;The above patch is only a work around. A more complete solution is needed with ZFS 0.8.2. We should either open another ticket or reopen this ticket.&lt;/p&gt;</comment>
                            <comment id="252022" author="pjones" created="Thu, 25 Jul 2019 14:23:25 +0000"  >&lt;p&gt;James&lt;/p&gt;

&lt;p&gt;Do you mean that further Lustre changes will be needed to adjust to the ZFS changes in 0.8.2?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="252023" author="bzzz" created="Thu, 25 Jul 2019 15:19:37 +0000"  >&lt;p&gt;I&apos;ll be working on a patch re-implementing this using new DMU API, but this can wait a bit, IMO.&lt;br/&gt;
the landed patch is fully functional, it just keep using not-best-approach to access DMU functionality.&lt;br/&gt;
that was made this way as part of performance improvements due to missing bits in DMU API.&lt;/p&gt;</comment>
                            <comment id="252024" author="pjones" created="Thu, 25 Jul 2019 15:21:45 +0000"  >&lt;p&gt;In that case let&apos;s create another JIRA ticket to track that work and leave this as resolved and get the fix onto the LTS branch too (I think this is already in flight)&lt;/p&gt;</comment>
                            <comment id="252225" author="gerrit" created="Tue, 30 Jul 2019 04:03:30 +0000"  >&lt;p&gt;Oleg Drokin (green@whamcloud.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/35600/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/35600/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12510&quot; title=&quot;mds server hangs cv_wait_common&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12510&quot;&gt;&lt;del&gt;LU-12510&lt;/del&gt;&lt;/a&gt; osd: osd-zfs to release zrlock quickly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_12&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: ec6b9a60c87767236804865da51a5c1c425db01c&lt;/p&gt;</comment>
                            <comment id="252642" author="gerrit" created="Tue, 6 Aug 2019 23:27:24 +0000"  >&lt;p&gt;Andreas Dilger (adilger@whamcloud.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/35707&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/35707&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12510&quot; title=&quot;mds server hangs cv_wait_common&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12510&quot;&gt;&lt;del&gt;LU-12510&lt;/del&gt;&lt;/a&gt; osd: osd-zfs to release zrlock quickly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 678e24e01ea4427b84d6c4b2a4d679edf077d3cc&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="56402">LU-12544</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="49351">LU-10250</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="33078" name="LU12510.tar.gz" size="73641648" author="dustb100" created="Fri, 5 Jul 2019 14:11:08 +0000"/>
                            <attachment id="33092" name="f2-mds4_20190709.tgz" size="72441830" author="curtispb" created="Tue, 9 Jul 2019 17:11:22 +0000"/>
                            <attachment id="33082" name="f2-mds4_lustre_unhealthy_20190707.tar.gz" size="2449256" author="dustb100" created="Mon, 8 Jul 2019 12:17:30 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10040" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic</customfieldname>
                        <customfieldvalues>
                                        <label>server</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i00j8f:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10021"><![CDATA[2]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>