<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:33:17 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-10239] Lustre crash (client): The first extent to be fit in a RPC contains 17 chunks, which is over the limit 16.</title>
                <link>https://jira.whamcloud.com/browse/LU-10239</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Our clients have max_rpcs_in_flight set to 16. Some of our clients hit this bug at times, which crashes these nodes:&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
2017-10-27T20:22:40-05:00 node0748 kernel: [14998.118253] LustreError: 1708:0:(osc_cache.c:1931:try_to_add_extent_for_io()) extent ffff882e6c434000@{[60858 -&amp;gt; 64953/64953], [1|0|+|lockdone|wSu|ffff882dc5cce4d0], [0|4096|+|-| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)|4096| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)]} The first extent to be fit in a RPC contains 17 chunks, which is over the limit 16.
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.154188] LustreError: 1708:0:(osc_cache.c:1226:osc_extent_tree_dump0()) Dump object ffff882dc5cce4d0 extents at try_to_add_extent_for_io:1931, mppr: 4096.
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.181833] LustreError: 1708:0:(osc_cache.c:1239:osc_extent_tree_dump0()) extent ffff882e6c434000@{[60858 -&amp;gt; 64953/64953], [1|0|+|lockdone|wSu|ffff882dc5cce4d0], [0|4096|+|-| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)|4096| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)]}urgent 1.
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.216119] LustreError: 1708:0:(osc_cache.c:1931:try_to_add_extent_for_io()) ASSERTION( data-&amp;gt;erd_page_count != 0 || chunk_count &amp;lt;= data-&amp;gt;erd_max_chunks ) failed: 
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.244827] LustreError: 1708:0:(osc_cache.c:1931:try_to_add_extent_for_io()) LBUG
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.259546] Pid: 1708, comm: ptlrpcd_00_86
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.270365] 
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.270365] Call Trace:
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.287055] [&amp;lt;ffffffffa07537f3&amp;gt;] libcfs_debug_dumpstack+0x53/0x80 [libcfs]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.301366] [&amp;lt;ffffffffa0753861&amp;gt;] lbug_with_loc+0x41/0xb0 [libcfs]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.314695] [&amp;lt;ffffffffa0c69498&amp;gt;] try_to_add_extent_for_io.isra.24+0xf58/0x12e0 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.328249] [&amp;lt;ffffffffa0c6b9dd&amp;gt;] osc_io_unplug0+0x3fd/0x1950 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.339730] [&amp;lt;ffffffff810d2372&amp;gt;] ? load_balance+0x192/0x990
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.350013] [&amp;lt;ffffffff810ce46c&amp;gt;] ? dequeue_entity+0x11c/0x5d0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.360444] [&amp;lt;ffffffffa0c6db30&amp;gt;] osc_io_unplug+0x10/0x20 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.370874] [&amp;lt;ffffffffa0c49441&amp;gt;] brw_queue_work+0x31/0xd0 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.381510] [&amp;lt;ffffffffa0a5e3d7&amp;gt;] work_interpreter+0x37/0xf0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.392538] [&amp;lt;ffffffffa0a5b0b5&amp;gt;] ptlrpc_check_set.part.23+0x425/0x1dd0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.405132] [&amp;lt;ffffffffa0a5cabb&amp;gt;] ptlrpc_check_set+0x5b/0xe0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.416411] [&amp;lt;ffffffffa0a88a3b&amp;gt;] ptlrpcd_check+0x4db/0x5d0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.427367] [&amp;lt;ffffffffa0a88d57&amp;gt;] ptlrpcd+0x227/0x560 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.437727] [&amp;lt;ffffffff810c4fd0&amp;gt;] ? default_wake_function+0x0/0x20
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.448479] [&amp;lt;ffffffffa0a88b30&amp;gt;] ? ptlrpcd+0x0/0x560 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.458669] [&amp;lt;ffffffff810b064f&amp;gt;] kthread+0xcf/0xe0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.467800] [&amp;lt;ffffffff810b0580&amp;gt;] ? kthread+0x0/0xe0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.477111] [&amp;lt;ffffffff81696818&amp;gt;] ret_from_fork+0x58/0x90
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.486513] [&amp;lt;ffffffff810b0580&amp;gt;] ? kthread+0x0/0xe0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.495492] 
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.500760] Kernel panic - not syncing: LBUG
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.509021] CPU: 203 PID: 1708 Comm: ptlrpcd_00_86 Tainted: P OE ------------ 3.10.0-514.6.1.el7.x86_64 #1
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.524591] Hardware name: Penguin Computing Relion 1904GT/S7200AP, BIOS S72C610.86B.01.02.0001.112820162103 11/28/2016
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.540144] ffffffffa0770ccc 00000000e8cd63f9 ffff882ed87a7958 ffffffff816862ac
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.551919] ffff882ed87a79d8 ffffffff8167f6b3 ffffffff00000008 ffff882ed87a79e8
&#160;
2017-10-27T20:22:40-05:00 node0748 kernel: [14998.118253] LustreError: 1708:0:(osc_cache.c:1931:try_to_add_extent_for_io()) extent ffff882e6c434000@{[60858 -&amp;gt; 64953/64953], [1|0|+|lockdone|wSu|ffff882dc5cce4d0], [0|4096|+|-| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)|4096| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)]} The first extent to be fit in a RPC contains 17 chunks, which is over the limit 16.
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.154188] LustreError: 1708:0:(osc_cache.c:1226:osc_extent_tree_dump0()) Dump object ffff882dc5cce4d0 extents at try_to_add_extent_for_io:1931, mppr: 4096.
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.181833] LustreError: 1708:0:(osc_cache.c:1239:osc_extent_tree_dump0()) extent ffff882e6c434000@{[60858 -&amp;gt; 64953/64953], [1|0|+|lockdone|wSu|ffff882dc5cce4d0], [0|4096|+|-| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)|4096| (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)]}urgent 1.
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.216119] LustreError: 1708:0:(osc_cache.c:1931:try_to_add_extent_for_io()) ASSERTION( data-&amp;gt;erd_page_count != 0 || chunk_count &amp;lt;= data-&amp;gt;erd_max_chunks ) failed: 
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.244827] LustreError: 1708:0:(osc_cache.c:1931:try_to_add_extent_for_io()) LBUG
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.259546] Pid: 1708, comm: ptlrpcd_00_86
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.270365] 
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.270365] Call Trace:
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.287055] [&amp;lt;ffffffffa07537f3&amp;gt;] libcfs_debug_dumpstack+0x53/0x80 [libcfs]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.301366] [&amp;lt;ffffffffa0753861&amp;gt;] lbug_with_loc+0x41/0xb0 [libcfs]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.314695] [&amp;lt;ffffffffa0c69498&amp;gt;] try_to_add_extent_for_io.isra.24+0xf58/0x12e0 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.328249] [&amp;lt;ffffffffa0c6b9dd&amp;gt;] osc_io_unplug0+0x3fd/0x1950 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.339730] [&amp;lt;ffffffff810d2372&amp;gt;] ? load_balance+0x192/0x990
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.350013] [&amp;lt;ffffffff810ce46c&amp;gt;] ? dequeue_entity+0x11c/0x5d0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.360444] [&amp;lt;ffffffffa0c6db30&amp;gt;] osc_io_unplug+0x10/0x20 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.370874] [&amp;lt;ffffffffa0c49441&amp;gt;] brw_queue_work+0x31/0xd0 [osc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.381510] [&amp;lt;ffffffffa0a5e3d7&amp;gt;] work_interpreter+0x37/0xf0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.392538] [&amp;lt;ffffffffa0a5b0b5&amp;gt;] ptlrpc_check_set.part.23+0x425/0x1dd0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.405132] [&amp;lt;ffffffffa0a5cabb&amp;gt;] ptlrpc_check_set+0x5b/0xe0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.416411] [&amp;lt;ffffffffa0a88a3b&amp;gt;] ptlrpcd_check+0x4db/0x5d0 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.427367] [&amp;lt;ffffffffa0a88d57&amp;gt;] ptlrpcd+0x227/0x560 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.437727] [&amp;lt;ffffffff810c4fd0&amp;gt;] ? default_wake_function+0x0/0x20
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.448479] [&amp;lt;ffffffffa0a88b30&amp;gt;] ? ptlrpcd+0x0/0x560 [ptlrpc]
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.458669] [&amp;lt;ffffffff810b064f&amp;gt;] kthread+0xcf/0xe0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.467800] [&amp;lt;ffffffff810b0580&amp;gt;] ? kthread+0x0/0xe0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.477111] [&amp;lt;ffffffff81696818&amp;gt;] ret_from_fork+0x58/0x90
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.486513] [&amp;lt;ffffffff810b0580&amp;gt;] ? kthread+0x0/0xe0
 2017-10-27T20:22:40-05:00 node0748 kernel: [14998.495492] 
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.500760] Kernel panic - not syncing: LBUG
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.509021] CPU: 203 PID: 1708 Comm: ptlrpcd_00_86 Tainted: P OE ------------ 3.10.0-514.6.1.el7.x86_64 #1
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.524591] Hardware name: Penguin Computing Relion 1904GT/S7200AP, BIOS S72C610.86B.01.02.0001.112820162103 11/28/2016
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.540144] ffffffffa0770ccc 00000000e8cd63f9 ffff882ed87a7958 ffffffff816862ac
 2017-10-27T20:22:41-05:00 node0748 kernel: [14998.551919] ffff882ed87a79d8 ffffffff8167f6b3 ffffffff00000008 ffff882ed87a79e8
&#160;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Seems to be related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8680&quot; title=&quot;replay-single test_20b: BUG: soft lockup - osc_makes_rpc()&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8680&quot;&gt;&lt;del&gt;LU-8680&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</description>
                <environment></environment>
        <key id="49310">LU-10239</key>
            <summary>Lustre crash (client): The first extent to be fit in a RPC contains 17 chunks, which is over the limit 16.</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bobijam">Zhenyu Xu</assignee>
                                    <reporter username="ma256">Murshid Azman</reporter>
                        <labels>
                    </labels>
                <created>Tue, 14 Nov 2017 08:15:57 +0000</created>
                <updated>Wed, 26 Feb 2020 15:40:24 +0000</updated>
                            <resolved>Fri, 4 Jan 2019 05:03:27 +0000</resolved>
                                    <version>Lustre 2.9.0</version>
                                    <fixVersion>Lustre 2.13.0</fixVersion>
                                        <due></due>
                            <votes>1</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="214094" author="phils@dugeo.com" created="Sun, 19 Nov 2017 14:23:33 +0000"  >&lt;p&gt;We just ran a job that blew up 1,600 nodes with this issue, so we&apos;re highly motivated to solve or work around it.&lt;/p&gt;

&lt;p&gt;My theory &#8211; which is admittedly&#160;not&#160;very deep &#8211; goes something like this:&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;osc_io_submit() builds up a list of pending pages. It honours max_pages_per_rpc, but I don&apos;t see it making any effort to limit the number of chunks? (our max_pages_per_rpc on most nodes = 4096)&lt;/li&gt;
	&lt;li&gt;osc_queue_sync_pages() walks the list of pending items, building up an extent.&lt;/li&gt;
	&lt;li&gt;that extent gets added to the osc_object-&amp;gt;oo_reading_exts&lt;/li&gt;
	&lt;li&gt;osc_send_read_rpc() iterates over the oo_reading_exts extents&lt;/li&gt;
	&lt;li&gt;each extent gets passed into try_to_add_extent_for_io()&lt;/li&gt;
	&lt;li&gt;try_to_add_extent_for_io() sees that it has 17 chunks, greater than the 16 limit, and blows up&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Am I wrong? What prevents osc_io_submit() from building up a 4096-page I/O that contains more than 16 chunks?&lt;/p&gt;

&lt;p&gt;This is a&#160;2.9.52 client, but I checked master and it didn&apos;t look like things were materially different in this area.&lt;/p&gt;

&lt;p&gt;I could see that cutting max_pages_per_rpc to 16 might work around it, but&#160;that would be&#160;an extremely harsh workaround, resulting in a lot of extremely tiny I/Os...&lt;/p&gt;

&lt;p&gt;Does anyone have&#160;a clue to spare, perhaps, to point us in the right direction?&lt;/p&gt;</comment>
                            <comment id="216867" author="jamesanunez" created="Wed, 20 Dec 2017 18:19:29 +0000"  >&lt;p&gt;Bobijam - &lt;/p&gt;

&lt;p&gt;Would you please look into this issue?&lt;/p&gt;

&lt;p&gt;Thank you &lt;/p&gt;</comment>
                            <comment id="216937" author="gerrit" created="Thu, 21 Dec 2017 12:29:39 +0000"  >&lt;p&gt;Bobi Jam (bobijam@hotmail.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/30627&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/30627&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10239&quot; title=&quot;Lustre crash (client): The first extent to be fit in a RPC contains 17 chunks, which is over the limit 16.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10239&quot;&gt;&lt;del&gt;LU-10239&lt;/del&gt;&lt;/a&gt; osc: limit chunk size of write submit&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: cae710b2f9d2e257add0a572ebc999fd407a07aa&lt;/p&gt;</comment>
                            <comment id="239321" author="gerrit" created="Fri, 4 Jan 2019 04:44:54 +0000"  >&lt;p&gt;Oleg Drokin (green@whamcloud.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/30627/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/30627/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10239&quot; title=&quot;Lustre crash (client): The first extent to be fit in a RPC contains 17 chunks, which is over the limit 16.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10239&quot;&gt;&lt;del&gt;LU-10239&lt;/del&gt;&lt;/a&gt; osc: limit chunk number of write submit&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 93ef6e7863b4e3aadfb3171f7f4958c5040446ec&lt;/p&gt;</comment>
                            <comment id="239366" author="pjones" created="Fri, 4 Jan 2019 05:03:27 +0000"  >&lt;p&gt;Landed for 2.13&lt;/p&gt;</comment>
                            <comment id="262929" author="tomaka" created="Sat, 8 Feb 2020 12:23:36 +0000"  >&lt;p&gt;Hey, Any chances to backport this fix to 2.12.4?&lt;/p&gt;</comment>
                            <comment id="264088" author="tomaka" created="Wed, 26 Feb 2020 15:40:24 +0000"  >&lt;p&gt;Hmm?&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="36888">LU-8135</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="53458">LU-11449</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzznrj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>