<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:05:28 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7039] llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&gt;lrh_index == tail-&gt;lrt_index ) failed:</title>
                <link>https://jira.whamcloud.com/browse/LU-7039</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Running tip of master with SWL and DNE. &lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;2015-08-25 12:19:21 Lustre: lustre-MDT0001: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-450
2015-08-25 12:19:21 LustreError: 6378:0:(llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:
2015-08-25 12:19:21 LustreError: 6384:0:(llog_osd.c:788:llog_osd_next_block()) lustre-MDT000b-osp-MDT0001: invalid llog tail at log id 0x3:2147484674/0 offset 3407872
2015-08-25 12:19:21 LustreError: 6384:0:(lod_dev.c:392:lod_sub_recovery_thread()) lustre-MDT000b-osp-MDT0001 getting update log failed: rc = -22
2015-08-25 12:19:21 LustreError: 6378:0:(llog_osd.c:778:llog_osd_next_block()) LBUG
2015-08-25 12:19:21 Pid: 6378, comm: lod0001_rec0005
2015-08-25 12:19:21 Aug 25 12:19:21
2015-08-25 12:19:21 iws12 kernel: LuCall Trace:
2015-08-25 12:19:21 streError: 6378: [&amp;lt;ffffffffa04a2875&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
2015-08-25 12:19:21 0:(llog_osd.c:77 [&amp;lt;ffffffffa04a2e77&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
2015-08-25 12:19:21 8:llog_osd_next_ [&amp;lt;ffffffffa08dad15&amp;gt;] llog_osd_next_block+0xb75/0xbf0 [obdclass]
2015-08-25 12:19:21 block()) ASSERTI [&amp;lt;ffffffffa08ccb4e&amp;gt;] llog_process_thread+0xInitializing cgroup subsys cpuset
2015-08-25 12:19:21 Initializing cgroup subsys cpu
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Attempting to recreate and get a dump&lt;/p&gt;</description>
                <environment>Hyperion SWL test</environment>
        <key id="31663">LU-7039</key>
            <summary>llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&gt;lrh_index == tail-&gt;lrt_index ) failed:</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="di.wang">Di Wang</assignee>
                                    <reporter username="cliffw">Cliff White</reporter>
                        <labels>
                    </labels>
                <created>Tue, 25 Aug 2015 19:25:21 +0000</created>
                <updated>Mon, 8 Jun 2020 17:00:33 +0000</updated>
                            <resolved>Wed, 27 Jan 2016 00:55:05 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                    <fixVersion>Lustre 2.8.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>13</watches>
                                                                            <comments>
                            <comment id="125105" author="cliffw" created="Tue, 25 Aug 2015 19:38:29 +0000"  >&lt;p&gt;Repeat, got the full stack this time:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;2015-08-25 12:33:20 Lustre: lustre-MDT0001: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-450
2015-08-25 12:33:20 LustreError: 6140:0:(llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:
2015-08-25 12:33:20 LustreError: 6134:0:(llog_osd.c:788:llog_osd_next_block()) lustre-MDT0005-osp-MDT0001: invalid llog tail at log id 0x3:1073742851/0 offset 3407872
2015-08-25 12:33:20 LustreError: 6134:0:(lod_dev.c:392:lod_sub_recovery_thread()) lustre-MDT0005-osp-MDT0001 getting update log failed: rc = -22
2015-08-25 12:33:20 LustreError: 6140:0:(llog_osd.c:778:llog_osd_next_block()) LBUG
2015-08-25 12:33:20 Pid: 6140, comm: lod0001_rec000b
2015-08-25 12:33:20
2015-08-25 12:33:20 Call Trace:
2015-08-25 12:33:20  [&amp;lt;ffffffffa04a2875&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
2015-08-25 12:33:20  [&amp;lt;ffffffffa04a2e77&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05f4d15&amp;gt;] llog_osd_next_block+0xb75/0xbf0 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05e6b4e&amp;gt;] llog_process_thread+0x2de/0xfc0 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05e78ed&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa12bc990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05eadb8&amp;gt;] llog_cat_process_cb+0x458/0x600 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05e71ba&amp;gt;] llog_process_thread+0x94a/0xfc0 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05e78ed&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05ea960&amp;gt;] ? llog_cat_process_cb+0x0/0x600 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05e961d&amp;gt;] llog_cat_process_or_fork+0x1ad/0x300 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa12bc990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
2015-08-25 12:33:20  [&amp;lt;ffffffffa12bc990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
2015-08-25 12:33:20  [&amp;lt;ffffffffa04aebf1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
2015-08-25 12:33:20  [&amp;lt;ffffffffa05e9789&amp;gt;] llog_cat_process+0x19/0x20 [obdclass]
2015-08-25 12:33:20  [&amp;lt;ffffffffa12c1e4a&amp;gt;] lod_sub_recovery_thread+0x69a/0xbc0 [lod]
2015-08-25 12:33:20  [&amp;lt;ffffffffa12c17b0&amp;gt;] ? lod_sub_recovery_thread+0x0/0xbc0 [lod]
2015-08-25 12:33:20  [&amp;lt;ffffffff8109e78e&amp;gt;] kthread+0x9e/0xc0
2015-08-25 12:33:20  [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
2015-08-25 12:33:20  [&amp;lt;ffffffff8109e6f0&amp;gt;] ? kthread+0x0/0xc0
2015-08-25 12:33:20  [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Lustre log attached&lt;/p&gt;</comment>
                            <comment id="125106" author="cliffw" created="Tue, 25 Aug 2015 19:39:05 +0000"  >&lt;p&gt;Lustre log&lt;/p&gt;</comment>
                            <comment id="125127" author="jamesanunez" created="Tue, 25 Aug 2015 21:29:58 +0000"  >&lt;p&gt;I saw this on some master failures from last night. I attributed the failures to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3263&quot; title=&quot;llog_osd_next_block(): ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3263&quot;&gt;&lt;del&gt;LU-3263&lt;/del&gt;&lt;/a&gt;. The assertions and stack traces look alike.&lt;/p&gt;</comment>
                            <comment id="125152" author="gerrit" created="Wed, 26 Aug 2015 07:29:31 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16084&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16084&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; llog: debug patches&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: de2d20f45a8e41a6e1510b596097e6c8ab270396&lt;/p&gt;</comment>
                            <comment id="125153" author="di.wang" created="Wed, 26 Aug 2015 07:30:46 +0000"  >&lt;p&gt;Cliff: could you please try this debug patch and get the debug log(-1 level) for me? thanks.&lt;/p&gt;</comment>
                            <comment id="126196" author="jgmitter" created="Thu, 3 Sep 2015 16:51:59 +0000"  >&lt;p&gt;Per Cliff:  we have tried the debug patch for a couple days and didn&apos;t get a debug log.  Di is aware of this and will provide a new build.&lt;/p&gt;</comment>
                            <comment id="126461" author="di.wang" created="Fri, 4 Sep 2015 21:51:56 +0000"  >&lt;p&gt;cliff: could you please try the test with this build &lt;a href=&quot;http://review.whamcloud.com/#/c/13942/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/13942/&lt;/a&gt; , I added a few debug information and also some fixes, which are not landed yet. &lt;/p&gt;</comment>
                            <comment id="129107" author="heckes" created="Fri, 2 Oct 2015 07:22:40 +0000"  >&lt;p&gt;Same error happened during soak testing of build &lt;a href=&quot;http://review.whamcloud.com/#/c/16685/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/16685/&lt;/a&gt;&lt;br/&gt;
(see &lt;a href=&quot;https://wiki.hpdd.intel.com/pages/viewpage.action?spaceKey=Releases&amp;amp;title=Soak+Testing+on+Lola#SoakTestingonLola-20150930&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/pages/viewpage.action?spaceKey=Releases&amp;amp;title=Soak+Testing+on+Lola#SoakTestingonLola-20150930&lt;/a&gt;)&lt;br/&gt;
Here&apos;s the stack trace:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Oct  1 20:06:27 lola-10 kernel: LustreError: 9078:0:(llog_osd.c:856:llog_osd_next_block()) ASSERTION( last_rec
-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed: 
Oct  1 20:06:27 lola-10 kernel: LustreError: 9078:0:(llog_osd.c:856:llog_osd_next_block()) LBUG
Oct  1 20:06:27 lola-10 kernel: Pid: 9078, comm: lod0004_rec0000
Oct  1 20:06:27 lola-10 kernel: 
Oct  1 20:06:27 lola-10 kernel: Call Trace:
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa05c1875&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa05c1e77&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06d5b90&amp;gt;] llog_osd_next_block+0xa40/0xc80 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06c6c0e&amp;gt;] llog_process_thread+0x2de/0xfc0 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06c79ad&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa11ac990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06cae78&amp;gt;] llog_cat_process_cb+0x458/0x600 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06c727a&amp;gt;] llog_process_thread+0x94a/0xfc0 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06c79ad&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06caa20&amp;gt;] ? llog_cat_process_cb+0x0/0x600 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06c96dd&amp;gt;] llog_cat_process_or_fork+0x1ad/0x300 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa11d8209&amp;gt;] ? lod_sub_prep_llog+0x4f9/0x7a0 [lod]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa11ac990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa06c9849&amp;gt;] llog_cat_process+0x19/0x20 [obdclass]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa11b1e2a&amp;gt;] lod_sub_recovery_thread+0x69a/0xbc0 [lod]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffffa11b1790&amp;gt;] ? lod_sub_recovery_thread+0x0/0xbc0 [lod]
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffff8109e78e&amp;gt;] kthread+0x9e/0xc0
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffff8109e6f0&amp;gt;] ? kthread+0x0/0xc0
Oct  1 20:06:27 lola-10 kernel: [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
Oct  1 20:06:27 lola-10 kernel: 
Oct  1 20:06:27 lola-10 kernel: LustreError: dumping log to /tmp/lustre-log.1443755187.9078
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Also attached the debug log. I hope it contains some usable informations.&lt;/p&gt;</comment>
                            <comment id="129108" author="heckes" created="Fri, 2 Oct 2015 07:26:23 +0000"  >&lt;p&gt;debug log of MDS node lola-10&lt;/p&gt;</comment>
                            <comment id="129116" author="heckes" created="Fri, 2 Oct 2015 12:25:28 +0000"  >&lt;p&gt;The node continuously panics with same error at the moment. On thing to add for the test set-up. AS we hit LDEV-193 &lt;br/&gt;
MDTs are mounted with option &apos;&lt;tt&gt;abort_recovery&lt;/tt&gt;&apos;.&lt;/p&gt;</comment>
                            <comment id="129195" author="gerrit" created="Fri, 2 Oct 2015 21:47:59 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16714&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16714&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; llog: remove LASSERT check for record index&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: f07a1980bd1e5cb08a00108ef286adc6bad4b4ea&lt;/p&gt;</comment>
                            <comment id="129196" author="di.wang" created="Fri, 2 Oct 2015 21:49:26 +0000"  >&lt;p&gt;Frank, Cliff: I add some debug information in this patch. Could you please run test with this patch and collect -1 level debug log for me? Thank you.&lt;/p&gt;</comment>
                            <comment id="129302" author="heckes" created="Mon, 5 Oct 2015 12:44:33 +0000"  >&lt;p&gt;yep, we&apos;ll create a build containing your patch and the fix for LDEV-24. Otherwise this bug has no chance to occur.&lt;/p&gt;</comment>
                            <comment id="129334" author="heckes" created="Mon, 5 Oct 2015 17:24:13 +0000"  >&lt;p&gt;lola is updated with your patch, but MDSes crash on first mount.&lt;/p&gt;</comment>
                            <comment id="129429" author="heckes" created="Tue, 6 Oct 2015 08:07:15 +0000"  >&lt;p&gt;A single debug message beside a LBUG was written:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Oct  5 09:51:39 lola-10 kernel: LustreError: 14865:0:(llog_osd.c:866:llog_osd_next_block()) soaked-MDT0000-osp
-MDT0004: invalid llog head/tail at log id 0x2:165027/0 offset 163840 last_rec 1515870810/1515870810/151587081
0/1515870810tail 0/0
Oct  5 09:51:39 lola-10 kernel: LustreError: 14865:0:(llog_osd.c:868:llog_osd_next_block()) LBUG
Oct  5 09:51:39 lola-10 kernel: Pid: 14865, comm: lod0004_rec0000
Oct  5 09:51:39 lola-10 kernel: 
Oct  5 09:51:39 lola-10 kernel: Call Trace:
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa05c1875&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa05c1e77&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06d5c23&amp;gt;] llog_osd_next_block+0xad3/0xd70 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06c6c0e&amp;gt;] llog_process_thread+0x2de/0xfc0 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06c79ad&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa11ae990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06cae78&amp;gt;] llog_cat_process_cb+0x458/0x600 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06c727a&amp;gt;] llog_process_thread+0x94a/0xfc0 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06c79ad&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06caa20&amp;gt;] ? llog_cat_process_cb+0x0/0x600 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06c96dd&amp;gt;] llog_cat_process_or_fork+0x1ad/0x300 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa11da549&amp;gt;] ? lod_sub_prep_llog+0x4f9/0x7a0 [lod]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa11ae990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa06c9849&amp;gt;] llog_cat_process+0x19/0x20 [obdclass]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa11b3e2a&amp;gt;] lod_sub_recovery_thread+0x69a/0xbc0 [lod]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffffa11b3790&amp;gt;] ? lod_sub_recovery_thread+0x0/0xbc0 [lod]
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffff8109e78e&amp;gt;] kthread+0x9e/0xc0
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffff8109e6f0&amp;gt;] ? kthread+0x0/0xc0
Oct  5 09:51:39 lola-10 kernel: [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
Oct  5 09:51:39 lola-10 kernel:
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="129644" author="di.wang" created="Tue, 6 Oct 2015 21:02:42 +0000"  >&lt;p&gt;According to the debug log &lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000004:00000010:25.0:1444148492.324048:0:4548:0:(osp_trans.c:215:osp_update_request_destroy()) kfreed &apos;our&apos;: 72 at ffff880816a30740.
00000004:00000040:25.0:1444148492.324050:0:4548:0:(osp_md_object.c:1208:osp_md_read()) soaked-MDT0000-osp-MDT0004: total read [0x2000284a3:0x2:0x0] pos 163840 len 32768
00000004:00000001:25.0:1444148492.324059:0:4548:0:(osp_md_object.c:1209:osp_md_read()) Process leaving via out (rc=32768 : 32768 : 0x8000)
00000040:00020000:25.0:1444148492.324062:0:4548:0:(llog_osd.c:866:llog_osd_next_block()) soaked-MDT0000-osp-MDT0004: invalid llog head/tail at log id 0x2:165027/0 offset 163840 last_rec 1515870810/1515870810/1515870810/1515870810tail 0/0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;tail-&amp;gt;lrt_len = 0 &amp;amp; tail-&amp;gt;lrt_index = 0.  Hmm, it seems update log is corrupted.&lt;/p&gt;

&lt;p&gt;Frank: could you please mount MDT0 as ldiskfs,&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;mount -t ldiskfs mdt0-dev  /mnt/mds1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;and get me this file  /mnt/mds1/update_log_dir/&lt;span class=&quot;error&quot;&gt;&amp;#91;0x2000284a3:0x2:0x0&amp;#93;&lt;/span&gt; and post it here. Thanks.&lt;/p&gt;</comment>
                            <comment id="129654" author="di.wang" created="Wed, 7 Oct 2015 01:02:17 +0000"  >&lt;p&gt;Sigh, according the debug log and real llog update file, it seems llog file is corrupted, and the end of llog file is filled with junks. Unfortunately there are no enough information to know why it is corrupted. &lt;/p&gt;

&lt;p&gt;To fix the filesystem, it needs to mount all of MDTs as ldiskfs, and delete these files&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@testnode tests]# mount -t ldiskfs -o loop /tmp/lustre-mdtn /mnt/mdsn
[root@testnode tests]# rm -rf /mnt/mdsn/update_log
[root@testnode tests]# rm -rf /mnt/mdsn/update_log_dir
[root@testnode tests]# umount /mnt/mdsn
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Then restart lustre. &lt;/p&gt;</comment>
                            <comment id="129658" author="gerrit" created="Wed, 7 Oct 2015 04:09:14 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16740&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16740&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; llog: skip to next chunk for corrupt record&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: c612e7c78344b2ff88c0d5803b558c9991ab766f&lt;/p&gt;</comment>
                            <comment id="129659" author="di.wang" created="Wed, 7 Oct 2015 04:09:59 +0000"  >&lt;p&gt;This is not a real fix. Just fix log_reader to make it skip the corrupt records.&lt;/p&gt;</comment>
                            <comment id="129841" author="tappro" created="Thu, 8 Oct 2015 16:18:37 +0000"  >&lt;p&gt;the same issue with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7100&quot; title=&quot;conf-sanity test_84 LBUGS with &#8220;(llog_osd.c:811:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index )&#8221;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7100&quot;&gt;&lt;del&gt;LU-7100&lt;/del&gt;&lt;/a&gt;, but different llog&lt;/p&gt;</comment>
                            <comment id="129866" author="tappro" created="Thu, 8 Oct 2015 18:51:54 +0000"  >&lt;p&gt;Di, I notice that changes made by you some time ago:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;	lgi-&amp;gt;lgi_off = max_t(__u64, lgi-&amp;gt;lgi_attr.la_size, lgi-&amp;gt;lgi_off);
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Not sure why that was made, could you explain that a bit? IIRC, that was done so because la_size may be not yet updated after first write, e.g. header, right?&lt;br/&gt;
If so then the same can happen with pad record, consider that pad was written, the lgi_off with offset is lost after that - it is rewriten by 0 for header update, but la_size may be not yet updated, so the record itself will overwrite the pad record.&lt;/p&gt;

&lt;p&gt;The solution could be saving the lgi_off after pad writting in local variable and use it.&lt;/p&gt;</comment>
                            <comment id="129870" author="tappro" created="Thu, 8 Oct 2015 18:53:10 +0000"  >&lt;p&gt;I&apos;ve just cook the patch if that make sense: &lt;a href=&quot;http://review.whamcloud.com/16771&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16771&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="129879" author="di.wang" created="Thu, 8 Oct 2015 21:06:46 +0000"  >&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;If so then the same can happen with pad record, consider that pad was written, the lgi_off with offset is lost after that - it is rewriten by 0 for header update, but la_size may be not yet updated, so the record itself will overwrite the pad record.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Hmm, we suppose to update the size in osp cache, see osp_md_write(). Note: this is only for remote llog object, for local llog object, adding pad record will update la_size in time, since you also see this problem in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7100&quot; title=&quot;conf-sanity test_84 LBUGS with &#8220;(llog_osd.c:811:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index )&#8221;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7100&quot;&gt;&lt;del&gt;LU-7100&lt;/del&gt;&lt;/a&gt; for local log, so I doubt this is the reason.&lt;/p&gt;</comment>
                            <comment id="129921" author="tappro" created="Fri, 9 Oct 2015 05:31:10 +0000"  >&lt;p&gt;If so, then we don&apos;t need max_t(__u64, lgi-&amp;gt;lgi_attr.la_size, lgi-&amp;gt;lgi_off); anymore, right? it can be just la_size always.&lt;/p&gt;</comment>
                            <comment id="129927" author="tappro" created="Fri, 9 Oct 2015 07:16:21 +0000"  >&lt;p&gt;As I can see that obj-&amp;gt;ooa may not exists, and it seems in most cases that is so, because it it is created just upon create of file what is not case with llog. Nevertheless, I agree that my suggestion doesn&apos;t explain corruption in local llog, I&apos;ve added some debug in patch, let&apos;s wait for new issues.&lt;/p&gt;</comment>
                            <comment id="129949" author="heckes" created="Fri, 9 Oct 2015 14:20:36 +0000"  >&lt;p&gt;Both patches installed for this bug + LDEV-24 patch that landed after creation of the build (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20151007&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20151007&lt;/a&gt;) and the cleanup procedure was executed.&lt;br/&gt;
The error didn&apos;t happened since now.&lt;br/&gt;
The following error message was printed during failover of MDT-0004 from &lt;tt&gt;lola-10&lt;/tt&gt; --&amp;gt; &lt;tt&gt;lola-11&lt;/tt&gt; and on other MDSes, too:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Oct  8 16:23:23 lola-11 kernel: LDISKFS-fs (dm-6): recovery complete
Oct  8 16:23:23 lola-11 kernel: LDISKFS-fs (dm-6): mounted filesystem with ordered data mode. quota=on. Opts: 
Oct  8 16:23:23 lola-11 kernel: Lustre: soaked-MDT0006: Connection restored to 192.168.1.111@o2ib10 (at 0@lo)
Oct  8 16:23:23 lola-11 kernel: Lustre: Skipped 3 previous similar messages
Oct  8 16:23:23 lola-11 kernel: LustreError: 11-0: soaked-MDT0003-osp-MDT0005: operation mds_connect to node 1
92.168.1.109@o2ib10 failed: rc = -114
Oct  8 16:23:23 lola-11 kernel: LustreError: Skipped 2 previous similar messages
Oct  8 16:23:23 lola-11 kernel: Lustre: soaked-MDT0005: Imperative Recovery enabled, recovery window shrunk fr
om 300-900 down to 150-450
Oct  8 16:23:23 lola-11 sshd[5890]: Received disconnect from 10.4.0.116: 11: disconnected by user
Oct  8 16:23:23 lola-11 sshd[5890]: pam_unix(sshd:session): session closed for user root
Oct  8 16:23:23 lola-11 kernel: LustreError: 5981:0:(llog.c:534:llog_process_thread()) soaked-MDT0006-osp-MDT0005: Invalid record: index 8 but expected 7
Oct  8 16:23:23 lola-11 kernel: LustreError: 5981:0:(lod_dev.c:392:lod_sub_recovery_thread()) soaked-MDT0006-osp-MDT0005 getting update log failed: rc = -34
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Not sure whether it is related to this bug.&lt;/p&gt;</comment>
                            <comment id="130121" author="gerrit" created="Mon, 12 Oct 2015 16:57:22 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16797&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16797&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; tgt: Delete txn_callback correctly in tgt_init()&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 57daeb1473bbf2cf3c3cfde83890818beee55f28&lt;/p&gt;</comment>
                            <comment id="130553" author="gerrit" created="Thu, 15 Oct 2015 18:48:07 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16839&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16839&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; osp: set update req error once import invalid&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 946fbe320ab6f093885247ec8234f2e4a28b50dc&lt;/p&gt;</comment>
                            <comment id="130602" author="heckes" created="Fri, 16 Oct 2015 08:46:25 +0000"  >&lt;p&gt;soak ran with build for CentOS-6.6 of build associated with change &lt;a href=&quot;http://review.whamcloud.com/#/c/16838/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/16838/&lt;/a&gt;&lt;br/&gt;
After reset of single MDS and remount the error occurred again:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Oct 15 16:08:03 lola-10 kernel: LDISKFS-fs warning (device dm-5): ldiskfs_multi_mount_protect: MMP interval 42 higher than expected, please wait.
Oct 15 16:08:03 lola-10 kernel: 
Oct 15 16:09:00 lola-10 kernel: LDISKFS-fs (dm-5): recovery complete
Oct 15 16:09:00 lola-10 kernel: LDISKFS-fs (dm-5): mounted filesystem with ordered data mode. quota=on. Opts: 
Oct 15 16:09:01 lola-10 kernel: LustreError: 137-5: soaked-MDT0004_UUID: not available for connect from 0@lo (no target). If you are running an HA pair check that the target is mounted on the other server.
Oct 15 16:09:01 lola-10 kernel: Lustre: soaked-MDT0005: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-450
Oct 15 16:09:01 lola-10 kernel: Lustre: 4390:0:(llog.c:520:llog_process_thread()) invalid length 0 in llog record for index 0/43
Oct 15 16:09:01 lola-10 kernel: LustreError: 4390:0:(lod_dev.c:402:lod_sub_recovery_thread()) soaked-MDT0002-osp-MDT0005 getting update log failed: rc = -22
Oct 15 16:09:01 lola-10 kernel: LustreError: 4242:0:(mdt_handler.c:5596:mdt_iocontrol()) soaked-MDT0005: Aborting recovery for device
Oct 15 16:09:01 lola-10 kernel: LustreError: 4242:0:(ldlm_lib.c:2473:target_stop_recovery_thread()) soaked-MDT0005: Aborting recovery
Oct 15 16:09:01 lola-10 kernel: Lustre: 4395:0:(ldlm_lib.c:1946:target_recovery_overseer()) recovery is aborted, evict exports in recovery
Oct 15 16:09:01 lola-10 kernel: Lustre: soaked-MDT0005: disconnecting 16 stale clients
Oct 15 16:09:07 lola-10 kernel: Lustre: soaked-MDT0005: Denying connection for new client e20f50f6-1966-798c-abf7-2d3bde20760a(at 192.168.1.126@o2ib100), waiting for 16 known clients (0 recovered, 0 in progress, and 16 evicted) to recover in 21116921:47
Oct 15 16:09:08 lola-10 kernel: LustreError: 4393:0:(llog_osd.c:856:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed: 
Oct 15 16:09:08 lola-10 kernel: LustreError: 4393:0:(llog_osd.c:856:llog_osd_next_block()) LBUG
Oct 15 16:09:08 lola-10 kernel: Pid: 4393, comm: lod0005_rec0006
Oct 15 16:09:08 lola-10 kernel: 
Oct 15 16:09:08 lola-10 kernel: Call Trace:
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa05c1875&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa05c1e77&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06d5bab&amp;gt;] llog_osd_next_block+0xa4b/0xc90 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06c6c2e&amp;gt;] llog_process_thread+0x2de/0xfc0 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06c79cd&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa11a7990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06cae98&amp;gt;] llog_cat_process_cb+0x458/0x600 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06c729a&amp;gt;] llog_process_thread+0x94a/0xfc0 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa07168b4&amp;gt;] ? dt_read+0x14/0x50 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06c79cd&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06caa40&amp;gt;] ? llog_cat_process_cb+0x0/0x600 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06c96fd&amp;gt;] llog_cat_process_or_fork+0x1ad/0x300 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa11d3589&amp;gt;] ? lod_sub_prep_llog+0x4f9/0x7a0 [lod]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa11a7990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa06c9869&amp;gt;] llog_cat_process+0x19/0x20 [obdclass]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa11a8c8e&amp;gt;] lod_sub_recovery_thread+0x26e/0xb90 [lod]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffffa11a8a20&amp;gt;] ? lod_sub_recovery_thread+0x0/0xb90 [lod]
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffff8109e78e&amp;gt;] kthread+0x9e/0xc0
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffff8109e6f0&amp;gt;] ? kthread+0x0/0xc0
Oct 15 16:09:08 lola-10 kernel: [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
Oct 15 16:09:08 lola-10 kernel: 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="130714" author="di.wang" created="Sat, 17 Oct 2015 08:41:31 +0000"  >&lt;p&gt;I just figured out a way to reproduce the problem, and also prove my thought that the failure of llog records cause the corruption. I will update the patch.&lt;/p&gt;</comment>
                            <comment id="130725" author="di.wang" created="Sun, 18 Oct 2015 06:37:45 +0000"  >&lt;p&gt;I update the fix along with other DNE patches to &lt;a href=&quot;http://review.whamcloud.com/16838&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16838&lt;/a&gt; . After the test approved this actually works, I will update 16389.&lt;/p&gt;</comment>
                            <comment id="130743" author="tappro" created="Mon, 19 Oct 2015 06:29:28 +0000"  >&lt;p&gt;Di, could you describe what caused llog corruption? Was it wrong usage of llog by external caller or problem inside llog itself?&lt;/p&gt;</comment>
                            <comment id="130745" author="di.wang" created="Mon, 19 Oct 2015 06:49:48 +0000"  >&lt;p&gt;Please check the test case replay single 117  in  patch 16838. Basically it because of the llog header is not protected by lock when used by OSP for update log.  Basically, it happens like this&lt;/p&gt;

&lt;p&gt;1. For example OSP adds 3 updates into the OSP sending list. All of them will write to the same remote llog file. So they will share the same bitmap and log index etc. Though the bitmap in 2nd will include 1st&apos;s bit, and the 3rd includes both 1st and 2nd bits.&lt;/p&gt;

&lt;p&gt;2. And the 1 updates failed for some reason, which might happen especially when we restart 2 MDTs at the same time in soak-test.  &lt;/p&gt;

&lt;p&gt;3. After recovery, it will keep writing 2nd and 3 rd updates, which might include 1st bitmap, but there are no 1st record because of the failure of step 2.&lt;/p&gt;

&lt;p&gt;So right now, I will let all of updates in the sending list fail once this happen, then refresh the header and also llog size for the new RPC.&lt;/p&gt;

&lt;p&gt;That said this probably not the reason for local log corruption.  And another thing I do not understand is that why this just start to happen recently (since last 2 weeks)? I checked the recent commit for llog, and did not find any suspicious patch? Do you have any clue?&lt;/p&gt;</comment>
                            <comment id="131505" author="tappro" created="Mon, 26 Oct 2015 09:51:09 +0000"  >&lt;p&gt;I don&apos;t see how the dt_attr_get() synchronized with dt_write() over OSP. As I can see there is no cache for attributes if llog wasn&apos;t created but opened. That means the dt_attr_get() will cause RPC to another server, but dt_write() might not be even applied remotely, so returned size might be smaller and llog will be corrupted as result. Is that true or I miss something?&lt;/p&gt;</comment>
                            <comment id="131565" author="di.wang" created="Mon, 26 Oct 2015 16:55:46 +0000"  >&lt;p&gt;the llog update (write + size update) is serialized by rpc_version, see osp_send_update_thread(). &lt;/p&gt;</comment>
                            <comment id="131604" author="tappro" created="Mon, 26 Oct 2015 19:53:09 +0000"  >&lt;p&gt;It is not about updates but about attr_get() which is not serialized with updates, so can return smaller size&lt;/p&gt;</comment>
                            <comment id="131611" author="di.wang" created="Mon, 26 Oct 2015 21:40:23 +0000"  >&lt;p&gt;They are serialize by lgh_lock in llog_cat_add_rec(). &lt;/p&gt;</comment>
                            <comment id="131631" author="tappro" created="Tue, 27 Oct 2015 05:50:46 +0000"  >&lt;p&gt;I don&apos;t see how lgh_lock helps here. Consider the following scenario:&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;llog_osd_write() issues new record, write RPC is scheduled for another server&lt;/li&gt;
	&lt;li&gt;new llog_osd_write() is called immediately and issues attr_get() RPC for the same llog to the remote server&lt;/li&gt;
	&lt;li&gt;since these two RPCs are not serialized and were issued almost together technically the attr_get() RPC might return size value before write will be applied on remote server&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;This is not the case if llog write RPC is synchronous but it is not so, isn&apos;t it?&lt;/p&gt;</comment>
                            <comment id="131642" author="di.wang" created="Tue, 27 Oct 2015 08:18:10 +0000"  >&lt;p&gt;Hmm, Mike, I think you are right. We need update the size for the OSP object after write.&lt;/p&gt;</comment>
                            <comment id="131651" author="tappro" created="Tue, 27 Oct 2015 11:53:54 +0000"  >&lt;p&gt;Do you mean to make opo_ooa exists all time? That would solve this problem I think.&lt;/p&gt;</comment>
                            <comment id="131683" author="di.wang" created="Tue, 27 Oct 2015 16:39:15 +0000"  >&lt;p&gt;Yes, I believe so. I added this fix to 16838 (along with a few other changes) and soak-test will try to see if this can resolve the corrupt issue. &lt;/p&gt;</comment>
                            <comment id="131779" author="gerrit" created="Tue, 27 Oct 2015 23:52:55 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16969&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16969&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; llog: update llog header and size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 9191d9a7f79c739d71dc7652333b9f07456218ad&lt;/p&gt;</comment>
                            <comment id="131807" author="gerrit" created="Wed, 28 Oct 2015 13:48:22 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/16797/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16797/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; tgt: Delete txn_callback correctly in tgt_init()&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 44e9ec0b46fc46cc72bebbdc35e4a59a0397a81c&lt;/p&gt;</comment>
                            <comment id="131985" author="heckes" created="Thu, 29 Oct 2015 12:54:53 +0000"  >&lt;p&gt;Di: For build &apos;20151027&apos; (&lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20151027&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20151027&lt;/a&gt;)&lt;br/&gt;
which doesn&apos;t include change 16797, as far as I can see, the problem is still present:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Oct 28 07:10:59 lola-8 kernel: LustreError: 6954:0:(llog_osd.c:874:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == 
tail-&amp;gt;lrt_index ) failed: 
Oct 28 07:10:59 lola-8 kernel: LustreError: 6954:0:(llog_osd.c:874:llog_osd_next_block()) LBUG
Oct 28 07:10:59 lola-8 kernel: Pid: 6954, comm: lod0003_rec0007
Oct 28 07:10:59 lola-8 kernel: 
Oct 28 07:10:59 lola-8 kernel: Call Trace:
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa07fc875&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa07fce77&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa0917a7b&amp;gt;] llog_osd_next_block+0xa4b/0xc90 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffff81174450&amp;gt;] ? cache_alloc_refill+0x1c0/0x240
Oct 28 07:10:59 lola-8 kernel: LustreError: 6948:0:(llog.c:534:llog_process_thread()) soaked-MDT0000-osp-MDT0003: Invalid re
cord: index 9421 but expected 9420
Oct 28 07:10:59 lola-8 kernel: LustreError: 6948:0:(lod_dev.c:402:lod_sub_recovery_thread()) soaked-MDT0000-osp-MDT0003 gett
ing update log failed: rc = -34
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa0906d3e&amp;gt;] llog_process_thread+0x2de/0xfc0 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa0904d5c&amp;gt;] ? llog_init_handle+0x11c/0x950 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa0907add&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa1374990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa090b108&amp;gt;] llog_cat_process_cb+0x458/0x600 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa09073aa&amp;gt;] llog_process_thread+0x94a/0xfc0 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa0907add&amp;gt;] llog_process_or_fork+0xbd/0x5d0 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa090acb0&amp;gt;] ? llog_cat_process_cb+0x0/0x600 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa090996d&amp;gt;] llog_cat_process_or_fork+0x1ad/0x300 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa13a0589&amp;gt;] ? lod_sub_prep_llog+0x4f9/0x7a0 [lod]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa1374990&amp;gt;] ? lod_process_recovery_updates+0x0/0x420 [lod]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa0909ad9&amp;gt;] llog_cat_process+0x19/0x20 [obdclass]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa1375c8e&amp;gt;] lod_sub_recovery_thread+0x26e/0xb90 [lod]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffffa1375a20&amp;gt;] ? lod_sub_recovery_thread+0x0/0xb90 [lod]
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffff8109e78e&amp;gt;] kthread+0x9e/0xc0
Oct 28 07:10:59 lola-8 kernel: [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="132104" author="di.wang" created="Thu, 29 Oct 2015 21:59:33 +0000"  >&lt;p&gt;Frank: I updated &lt;a href=&quot;http://review.whamcloud.com/16838&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16838&lt;/a&gt;, and added more debug information there. Could you please retry with the patch. Thanks.&lt;/p&gt;</comment>
                            <comment id="132172" author="gerrit" created="Fri, 30 Oct 2015 16:39:20 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/16740/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16740/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; llog: skip to next chunk for corrupt record&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 04f4023cf59b6e5a1634ba492cd813dcb1af0c7c&lt;/p&gt;</comment>
                            <comment id="133557" author="gerrit" created="Sun, 15 Nov 2015 06:52:20 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/17199&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17199&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; recovery: abort update recovery once fails&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 70905df1d7ea16d50c927b6af9957bced89a0f3b&lt;/p&gt;</comment>
                            <comment id="134091" author="di.wang" created="Fri, 20 Nov 2015 18:04:51 +0000"  >&lt;p&gt;Just update, it seems corruption disappears in the build of 20151120, though we need run more test to confirm this. Currently the soak-test is blocked by &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7456&quot; title=&quot;kernel panic on ldiskfs_journal_commit_callback+0x4a/0x80 &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7456&quot;&gt;&lt;del&gt;LU-7456&lt;/del&gt;&lt;/a&gt;, and we will continue soak-test to check this problem once 7456 is fixed.&lt;/p&gt;</comment>
                            <comment id="134443" author="di.wang" created="Tue, 24 Nov 2015 19:36:58 +0000"  >&lt;p&gt;Just reminder, &lt;a href=&quot;http://review.whamcloud.com/16969&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16969&lt;/a&gt; and &lt;a href=&quot;http://review.whamcloud.com/17199&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17199&lt;/a&gt; are key fixes for this problem.&lt;/p&gt;</comment>
                            <comment id="135208" author="heckes" created="Fri, 4 Dec 2015 13:11:47 +0000"  >&lt;p&gt;The error below happens during soak testing of change 16838 patch set #31 (no Wiki entry for build exits, yet) on cluster &lt;em&gt;lola&lt;/em&gt;. DNE is enabled and MDSes are configured in active-active HA failover configuration. &lt;/p&gt;

&lt;ul&gt;
	&lt;li&gt;Primary resources of MDT &lt;tt&gt;lola-11&lt;/tt&gt; were failed back at Dec, 3 20:18.&lt;br/&gt;
  The allocation of slabs increased continuously till ~ 31 GB till crash &lt;/li&gt;
	&lt;li&gt;MDS node &lt;tt&gt;lola-11&lt;/tt&gt; crashed with oom-killer at Dec, 4 00:21 (local time).  (see also &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7432&quot; title=&quot;oom-killer started on MDSes&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7432&quot;&gt;&lt;del&gt;LU-7432&lt;/del&gt;&lt;/a&gt;)&lt;/li&gt;
	&lt;li&gt;&lt;tt&gt;ptlrpc_cache&lt;/tt&gt; seems to be the biggest consumer&lt;br/&gt;
Attached &lt;tt&gt;lola-11&lt;/tt&gt;&apos;s  messages, console log, vmcore-dmesg file, collectl (version V4.0.2-1) files (for time interval specified above). Also&lt;br/&gt;
attached files containing extracted counters for memory, slab totals and per slab allocation.&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;The crash dump has been saved to &lt;tt&gt;lola-1:/scratch/crashdumps/lu-7039/127.0.0.1-2015-12-04-00\:22\:36&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="135210" author="heckes" created="Fri, 4 Dec 2015 13:22:48 +0000"  >&lt;p&gt;It turned out that the  &lt;tt&gt;collectl&lt;/tt&gt; raw files are to big to be uploaded to Jira. I saved them to &lt;tt&gt;lola-1:/scratch/crashdumps/lu-7039&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="135246" author="di.wang" created="Fri, 4 Dec 2015 17:27:36 +0000"  >&lt;p&gt;This OOM is not caused by llog corruption, so it is a new problem (&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7517&quot; title=&quot;oom killer active after failback of MDS resources&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7517&quot;&gt;&lt;del&gt;LU-7517&lt;/del&gt;&lt;/a&gt;).&lt;/p&gt;</comment>
                            <comment id="139362" author="sarah" created="Tue, 19 Jan 2016 23:53:37 +0000"  >&lt;p&gt;another instance on master DNE mode&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b57ce146-bbfd-11e5-8506-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b57ce146-bbfd-11e5-8506-5254006e85c2&lt;/a&gt;&lt;br/&gt;
client and server: lustre-master build#3305&lt;/p&gt;</comment>
                            <comment id="140103" author="gerrit" created="Tue, 26 Jan 2016 20:20:01 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/16969/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16969/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7039&quot; title=&quot;llog_osd.c:778:llog_osd_next_block()) ASSERTION( last_rec-&amp;gt;lrh_index == tail-&amp;gt;lrt_index ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7039&quot;&gt;&lt;del&gt;LU-7039&lt;/del&gt;&lt;/a&gt; llog: update llog header and size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: e1745ed18d8e28f3cf3d72df3b7ef50d83f36601&lt;/p&gt;</comment>
                            <comment id="140140" author="jgmitter" created="Wed, 27 Jan 2016 00:55:05 +0000"  >&lt;p&gt;Landed for 2.8.0&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="33009">LU-7391</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="31909">LU-7100</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="31033">LU-6831</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="34308">LU-7715</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="31450">LU-6994</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="33262">LU-7455</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="18754" name="LU-7039.llog.txt.gz" size="3701121" author="cliffw" created="Tue, 25 Aug 2015 19:39:05 +0000"/>
                            <attachment id="19797" name="console.log.bz2" size="194520" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                            <attachment id="19166" name="lola-10-lustre-log.1444148492.4548-dm-minus-one.log.bz2" size="1045642" author="heckes" created="Tue, 6 Oct 2015 16:38:49 +0000"/>
                            <attachment id="19044" name="lustre-log.1443755187.9078" size="729590" author="heckes" created="Fri, 2 Oct 2015 07:26:23 +0000"/>
                            <attachment id="19798" name="memory-counter-lola-11.dat.bz2" size="25254" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                            <attachment id="19799" name="messages-lola-11.log.bz2" size="309631" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                            <attachment id="19800" name="slab-details-lola-11.dat.bz2" size="894137" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                            <attachment id="19801" name="slab-details-one-file-per-slab.tar.bz2" size="631494" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                            <attachment id="19802" name="slab-total-lola-11.dat.bz2" size="28753" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                            <attachment id="19803" name="vmcore-dmesg.txt.bz2" size="28332" author="heckes" created="Fri, 4 Dec 2015 13:13:40 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10490" key="com.atlassian.jira.plugin.system.customfieldtypes:datepicker">
                        <customfieldname>End date</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>Fri, 4 Dec 2015 19:25:21 +0000</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                            <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxlbb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                        <customfield id="customfield_10493" key="com.atlassian.jira.plugin.system.customfieldtypes:datepicker">
                        <customfieldname>Start date</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>Tue, 25 Aug 2015 19:25:21 +0000</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                    </customfields>
    </item>
</channel>
</rss>