<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:16:31 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-8320] :(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:</title>
                <link>https://jira.whamcloud.com/browse/LU-8320</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;MDS crash with LBUG.&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;0&amp;gt;LustreError: 39313:0:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed: ^M
&amp;lt;0&amp;gt;LustreError: 39313:0:(llog_osd.c:338:llog_osd_write_rec()) LBUG^M
&amp;lt;4&amp;gt;Pid: 39313, comm: mdt02_049^M
&amp;lt;4&amp;gt;^M
&amp;lt;4&amp;gt;Call Trace:^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa048b895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa048be97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa05bed55&amp;gt;] llog_osd_write_rec+0xfb5/0x1370 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0d46ecb&amp;gt;] ? dynlock_unlock+0x16b/0x1d0 [osd_ldiskfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0d2e5d2&amp;gt;] ? iam_path_release+0x42/0x70 [osd_ldiskfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0590438&amp;gt;] llog_write_rec+0xc8/0x290 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa059910d&amp;gt;] llog_cat_add_rec+0xad/0x480 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0590231&amp;gt;] llog_add+0x91/0x1d0 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fd04f7&amp;gt;] osp_sync_add_rec+0x247/0xad0 [osp]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fd0e2b&amp;gt;] osp_sync_add+0x7b/0x80 [osp]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fc27d6&amp;gt;] osp_object_destroy+0x106/0x150 [osp]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f068e7&amp;gt;] lod_object_destroy+0x1a7/0x350 [lod]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f74880&amp;gt;] mdd_finish_unlink+0x210/0x3d0 [mdd]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f65d35&amp;gt;] ? mdd_attr_check_set_internal+0x275/0x2c0 [mdd]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f75306&amp;gt;] mdd_unlink+0x8c6/0xca0 [mdd]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e37788&amp;gt;] mdo_unlink+0x18/0x50 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e3b005&amp;gt;] mdt_reint_unlink+0x835/0x1030 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e37571&amp;gt;] mdt_reint_rec+0x41/0xe0 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1ced3&amp;gt;] mdt_reint_internal+0x4c3/0x780 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1d1d4&amp;gt;] mdt_reint+0x44/0xe0 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1fada&amp;gt;] mdt_handle_common+0x52a/0x1470 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e5c5f5&amp;gt;] mds_regular_handle+0x15/0x20 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa07750c5&amp;gt;] ptlrpc_server_handle_request+0x385/0xc00 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa048c5ae&amp;gt;] ? cfs_timer_arm+0xe/0x10 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa049d8d5&amp;gt;] ? lc_watchdog_touch+0x65/0x170 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa076da69&amp;gt;] ? ptlrpc_wait_event+0xa9/0x2d0 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81057779&amp;gt;] ? __wake_up_common+0x59/0x90^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa077789d&amp;gt;] ptlrpc_main+0xafd/0x1780 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0776da0&amp;gt;] ? ptlrpc_main+0x0/0x1780 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20^M
&amp;lt;4&amp;gt;^M
&amp;lt;0&amp;gt;Kernel panic - not syncing: LBUG^M
&amp;lt;4&amp;gt;Pid: 39313, comm: mdt02_049 Tainted: G           ---------------  T 2.6.32-504.30.3.el6.20151008.x86_64.lustre253 #1^M
&amp;lt;4&amp;gt;Call Trace:^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81564fb9&amp;gt;] ? panic+0xa7/0x190^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa048beeb&amp;gt;] ? lbug_with_loc+0x9b/0xb0 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa05bed55&amp;gt;] ? llog_osd_write_rec+0xfb5/0x1370 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0d46ecb&amp;gt;] ? dynlock_unlock+0x16b/0x1d0 [osd_ldiskfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0d2e5d2&amp;gt;] ? iam_path_release+0x42/0x70 [osd_ldiskfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0590438&amp;gt;] ? llog_write_rec+0xc8/0x290 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa059910d&amp;gt;] ? llog_cat_add_rec+0xad/0x480 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0590231&amp;gt;] ? llog_add+0x91/0x1d0 [obdclass]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fd04f7&amp;gt;] ? osp_sync_add_rec+0x247/0xad0 [osp]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fd0e2b&amp;gt;] ? osp_sync_add+0x7b/0x80 [osp]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fc27d6&amp;gt;] ? osp_object_destroy+0x106/0x150 [osp]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f068e7&amp;gt;] ? lod_object_destroy+0x1a7/0x350 [lod]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f74880&amp;gt;] ? mdd_finish_unlink+0x210/0x3d0 [mdd]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f65d35&amp;gt;] ? mdd_attr_check_set_internal+0x275/0x2c0 [mdd]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f75306&amp;gt;] ? mdd_unlink+0x8c6/0xca0 [mdd]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e37788&amp;gt;] ? mdo_unlink+0x18/0x50 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e3b005&amp;gt;] ? mdt_reint_unlink+0x835/0x1030 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e37571&amp;gt;] ? mdt_reint_rec+0x41/0xe0 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1ced3&amp;gt;] ? mdt_reint_internal+0x4c3/0x780 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1d1d4&amp;gt;] ? mdt_reint+0x44/0xe0 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1fada&amp;gt;] ? mdt_handle_common+0x52a/0x1470 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e5c5f5&amp;gt;] ? mds_regular_handle+0x15/0x20 [mdt]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa07750c5&amp;gt;] ? ptlrpc_server_handle_request+0x385/0xc00 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa048c5ae&amp;gt;] ? cfs_timer_arm+0xe/0x10 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa049d8d5&amp;gt;] ? lc_watchdog_touch+0x65/0x170 [libcfs]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa076da69&amp;gt;] ? ptlrpc_wait_event+0xa9/0x2d0 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81057779&amp;gt;] ? __wake_up_common+0x59/0x90^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa077789d&amp;gt;] ? ptlrpc_main+0xafd/0x1780 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c28a&amp;gt;] ? child_rip+0xa/0x20^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0776da0&amp;gt;] ? ptlrpc_main+0x0/0x1780 [ptlrpc]^M
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20^M
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="37784">LU-8320</key>
            <summary>:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="tappro">Mikhail Pershin</assignee>
                                    <reporter username="mhanafi">Mahmoud Hanafi</reporter>
                        <labels>
                    </labels>
                <created>Thu, 23 Jun 2016 18:13:10 +0000</created>
                <updated>Thu, 14 Jun 2018 21:41:19 +0000</updated>
                            <resolved>Mon, 1 Aug 2016 05:45:12 +0000</resolved>
                                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="156742" author="mhanafi" created="Thu, 23 Jun 2016 18:41:34 +0000"  >&lt;p&gt;After the reboot and remount we hit the same LBUG!&lt;/p&gt;
</comment>
                            <comment id="156746" author="mhanafi" created="Thu, 23 Jun 2016 18:50:55 +0000"  >&lt;p&gt;This is lustre 2.5.3&lt;/p&gt;</comment>
                            <comment id="156749" author="pjones" created="Thu, 23 Jun 2016 19:01:23 +0000"  >&lt;p&gt;Looking into details supplied&lt;/p&gt;</comment>
                            <comment id="156752" author="green" created="Thu, 23 Jun 2016 19:08:07 +0000"  >&lt;p&gt;any other messages before this?&lt;/p&gt;</comment>
                            <comment id="156755" author="mhanafi" created="Thu, 23 Jun 2016 19:18:15 +0000"  >&lt;p&gt;no. &lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;[-- MARK -- Thu Jun 23 09:00:00 2016]^M
format at ldlm_pool.c:628:ldlm_pool_recalc doesn&apos;t end in newline^M
LustreError: 39313:0:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed: ^M
LustreError: 39313:0:(llog_osd.c:338:llog_osd_write_rec()) LBUG^M
Pid: 39313, comm: mdt02_049^M
^M
Call Trace:^M
 [&amp;lt;ffffffffa048b895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]^M
 [&amp;lt;ffffffffa048be97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]^M
^M
Entering kdb (current=0xffff8834a41b8ab0, pid 39313) on processor 16 Oops: (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;)^M
due to oops @ 0x0^M
kdba_dumpregs: pt_regs not available, use bt* or pid to select a different task^M
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="156767" author="jaylan" created="Thu, 23 Jun 2016 20:18:02 +0000"  >&lt;p&gt;Our git repo for 2.5.3 is at &lt;br/&gt;
&lt;a href=&quot;https://github.com/NASAEarthExchange/lustre-nas-fe/commits/nas-2.5.3&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/NASAEarthExchange/lustre-nas-fe/commits/nas-2.5.3&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="156768" author="mhanafi" created="Thu, 23 Jun 2016 20:18:12 +0000"  >&lt;p&gt;Do you require any additional info?&lt;/p&gt;</comment>
                            <comment id="156774" author="mhanafi" created="Thu, 23 Jun 2016 20:39:20 +0000"  >&lt;p&gt;Any ideas on how we can at least get the mdt mounted and bring up the filesystem?&lt;/p&gt;</comment>
                            <comment id="156775" author="green" created="Thu, 23 Jun 2016 20:39:52 +0000"  >&lt;p&gt;can you see local variables easily?&lt;br/&gt;
We wonder what is loghandle-&amp;gt;lgh_id value?&lt;br/&gt;
The current suspicion is you have a corrupted llog file somehow and in order to analyze and repair it we need to know what&apos;s the id so that we don&apos;t need to nuke all of them.&lt;/p&gt;</comment>
                            <comment id="156782" author="mhanafi" created="Thu, 23 Jun 2016 21:18:51 +0000"  >&lt;p&gt;What was the LU that the tool was part of?&lt;/p&gt;</comment>
                            <comment id="156784" author="green" created="Thu, 23 Jun 2016 21:25:38 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6696&quot; title=&quot;ASSERTION( rc == 0 || rc == LLOG_PROC_BREAK ) failed: 0 changes, 0 in progress, 0 in flight: -5&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6696&quot;&gt;&lt;del&gt;LU-6696&lt;/del&gt;&lt;/a&gt; &lt;/p&gt;</comment>
                            <comment id="156787" author="mhanafi" created="Thu, 23 Jun 2016 21:40:38 +0000"  >&lt;p&gt;I am a bit confused how exactly to use the tool. You need me to run it on file in /O/?&lt;/p&gt;</comment>
                            <comment id="156788" author="green" created="Thu, 23 Jun 2016 21:45:58 +0000"  >&lt;p&gt;Yes, pretty much.&lt;br/&gt;
If you can get the llog id from the variable - on that one.&lt;br/&gt;
Otherwise - I guess on all of them (do you have many?) in some reverse data order, I imagine until it fixes something (it should print a corresponding message)&lt;/p&gt;</comment>
                            <comment id="156790" author="mhanafi" created="Thu, 23 Jun 2016 21:53:05 +0000"  >&lt;p&gt;Most of the files are returning this&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;nbp2-mds /mnt/lustre/nbp2-mdt/O/1/d2 # llog_catfix 12226
Header size : 8192
Time : Thu Jun 23 09:04:29 2016
&lt;span class=&quot;code-object&quot;&gt;Number&lt;/span&gt; of records: 1544
-----------------------
Llog is not catalog, llh_size: 0, need 64
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="156793" author="green" created="Thu, 23 Jun 2016 22:06:41 +0000"  >&lt;p&gt;hm, yes, the tool only fixes catalog files.&lt;br/&gt;
Anyway it can tell them apart.&lt;br/&gt;
Do you have too many to go through all of them?&lt;/p&gt;</comment>
                            <comment id="156794" author="mhanafi" created="Thu, 23 Jun 2016 22:09:03 +0000"  >&lt;p&gt;yes there are 11k files under /O&lt;/p&gt;

&lt;p&gt;can we just delete all?  will it not regenerate them. &lt;/p&gt;</comment>
                            <comment id="156797" author="green" created="Thu, 23 Jun 2016 22:15:29 +0000"  >&lt;p&gt;You can delete them but hat would lead to llog records not replayed - in other words - space leakage on your OSTs.&lt;/p&gt;

&lt;p&gt;The alternative is to find the llog id of the log that&apos;s bad - either from the crash or since it always crashes - you can just insert a printk there rebuild just for the MDS and run it again to get it?&lt;/p&gt;</comment>
                            <comment id="156799" author="ndauchy" created="Thu, 23 Jun 2016 22:17:30 +0000"  >&lt;p&gt;Can we assume it is one of the files modified in the last day or so that is the problem, and hit them all with something like this?&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;find /mnt/lustre/nbp2-mdt/O/1 -type f -mtime -2 | grep -v LAST_ID | xargs -t -I {} llog_catfix {} 2&amp;gt;&amp;amp;1 | tee /tmp/llog_catfix.out
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="156803" author="green" created="Thu, 23 Jun 2016 22:32:06 +0000"  >&lt;p&gt;I think that&apos;s reasonable to assume, but there&apos;s no guarantee. The records might have been there since last reboot too.&lt;/p&gt;

&lt;p&gt;In the worst case you can always remove the CATALOGS file I imagine and it&apos;ll forget all about it (with the leaked space problem of course - so save that file too if you do that).&lt;br/&gt;
Best of all is to find the llog id of the problematic llog and threat just that.&lt;/p&gt;</comment>
                            <comment id="156810" author="green" created="Thu, 23 Jun 2016 22:51:05 +0000"  >&lt;p&gt;so that lgh_id=&lt;/p&gt;
{...}
&lt;p&gt; - what&apos;s inside?&lt;/p&gt;

&lt;p&gt;Actually, this must be a wrong one? lgh_hdr is not NULL here.&lt;/p&gt;

&lt;p&gt;Also, surprisingly the log record is a configuration one somehow:&lt;br/&gt;
        OBD_CFG_REC             = LLOG_OP_MAGIC | 0x20000,&lt;/p&gt;</comment>
                            <comment id="156812" author="mhanafi" created="Thu, 23 Jun 2016 22:57:43 +0000"  >&lt;p&gt;I used systemtap to trace the function and this is when we hit the lbug&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt; 0 llog_process_th(9252):-&amp;gt;llog_osd_write_rec env={.le_ctx={...}, .le_ses=0xffff881c7aca7ba0} loghandle={.lgh_lock={...}, .lgh_hdr_lock={...}, .lgh_id={...}, .lgh_hdr=0x0, .lgh_obj=0xffff883e7ccb9d40, .lgh_last_idx=0, .lgh_cur_idx=0, .lgh_cur_offset=0, .lgh_ctxt=0xffff883e7d6e8640, .u={...}, .lgh_name=&lt;span class=&quot;code-quote&quot;&gt;&quot;&amp;lt;unknown&amp;gt;&quot;&lt;/span&gt;, .private_data=0xffff883e8410a340, .lgh_logops=0xffffffffa1013a80, .lgh_refcount={...}} rec={.lrh_len=64, .lrh_index=0, .lrh_type=274989056, .lrh_id=0} reccookie={.lgc_lgl={...}, .lgc_subsys=0, .lgc_index=0, .lgc_padding=0} cookiecount=1 buf=

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="156814" author="green" created="Thu, 23 Jun 2016 23:14:44 +0000"  >&lt;p&gt;Can you get the .lgh_id= content please? That&apos;s what should tell us the filename&lt;/p&gt;</comment>
                            <comment id="156817" author="mhanafi" created="Thu, 23 Jun 2016 23:31:42 +0000"  >&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;0 llog_process_th(9553):-&amp;gt;llog_osd_write_rec env={.le_ctx={.lc_tags=268435457, .lc_state=2, .lc_thread=0x0, .lc_value=0xffff881d24328600, .lc_remember={.next=0xffff883e8f777b78, .prev=0xffff883e8f777b78}, .lc_version=37, .lc_cookie=0}, .le_ses=0xffff883e8f777ba0} loghandle={.lgh_lock={.count=-4294967295, .wait_lock={.raw_lock={.slock=0}}, .wait_list={.next=0xffff881c6e34fbd0, .prev=0xffff881c6e34fbd0}}, .lgh_hdr_lock={.raw_lock={.slock=0}}, .lgh_id={.lgl_oi={&amp;lt;union&amp;gt;={.oi={.oi_id=12744, .oi_seq=1}, .oi_fid={.f_seq=12744, .f_oid=1, .f_ver=0}}}, .lgl_og


&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="156819" author="green" created="Thu, 23 Jun 2016 23:46:34 +0000"  >&lt;p&gt;So the file we are loking at is named either 12744 or 31c8.&lt;/p&gt;

&lt;p&gt;Try the fixing program on it and if it does not help, just move it away (please retain it still for further analysis.&lt;/p&gt;</comment>
                            <comment id="156820" author="mhanafi" created="Thu, 23 Jun 2016 23:51:39 +0000"  >&lt;p&gt;found this file &lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;nbp2-mds /mnt/lustre/nbp2-mdt/O # find . -name &lt;span class=&quot;code-quote&quot;&gt;&quot;12744&quot;&lt;/span&gt;
./1/d8/12744
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt; llog_catfix ./1/d8/12744
Header size : 8192
Time : Tue Feb 23 14:44:48 2016
&lt;span class=&quot;code-object&quot;&gt;Number&lt;/span&gt; of records: 8
-----------------------
Llog is not catalog, llh_size: 0, need 64
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;removed the file and mounted without lbug, so far. osp-sync threads are running&lt;/p&gt;</comment>
                            <comment id="156823" author="mhanafi" created="Fri, 24 Jun 2016 00:08:53 +0000"  >&lt;p&gt;Filesystem is back up. Would you like to get a copy of the bad file?&lt;/p&gt;</comment>
                            <comment id="156825" author="ndauchy" created="Fri, 24 Jun 2016 00:24:05 +0000"  >&lt;p&gt;Since we have seen a form of this issue a couple times now, can we get the work that was started in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6696&quot; title=&quot;ASSERTION( rc == 0 || rc == LLOG_PROC_BREAK ) failed: 0 changes, 0 in progress, 0 in flight: -5&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6696&quot;&gt;&lt;del&gt;LU-6696&lt;/del&gt;&lt;/a&gt; and &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7011&quot; title=&quot;Kernel part of llog subsystem can do self-repairing in some cases&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7011&quot;&gt;LU-7011&lt;/a&gt; updated and landed?&lt;/p&gt;</comment>
                            <comment id="156826" author="green" created="Fri, 24 Jun 2016 00:25:59 +0000"  >&lt;p&gt;Yes, please attach the file here&lt;/p&gt;</comment>
                            <comment id="156847" author="ndauchy" created="Fri, 24 Jun 2016 13:32:37 +0000"  >&lt;p&gt;FYI, it looks like we hit the same bug again:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;lt;0&amp;gt;LustreError: 37590:0:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed: 
&amp;lt;0&amp;gt;LustreError: 37590:0:(llog_osd.c:338:llog_osd_write_rec()) LBUG
&amp;lt;4&amp;gt;Pid: 37590, comm: mdt03_120
&amp;lt;4&amp;gt;
&amp;lt;4&amp;gt;Call Trace:
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0487895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0487e97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa05bad55&amp;gt;] llog_osd_write_rec+0xfb5/0x1370 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0d49ecb&amp;gt;] ? dynlock_unlock+0x16b/0x1d0 [osd_ldiskfs]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0d315d2&amp;gt;] ? iam_path_release+0x42/0x70 [osd_ldiskfs]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa058c438&amp;gt;] llog_write_rec+0xc8/0x290 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa059510d&amp;gt;] llog_cat_add_rec+0xad/0x480 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa058c231&amp;gt;] llog_add+0x91/0x1d0 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fd34f7&amp;gt;] osp_sync_add_rec+0x247/0xad0 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fd3e2b&amp;gt;] osp_sync_add+0x7b/0x80 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0fc57d6&amp;gt;] osp_object_destroy+0x106/0x150 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f098e7&amp;gt;] lod_object_destroy+0x1a7/0x350 [lod]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f77880&amp;gt;] mdd_finish_unlink+0x210/0x3d0 [mdd]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f68d35&amp;gt;] ? mdd_attr_check_set_internal+0x275/0x2c0 [mdd]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0f78306&amp;gt;] mdd_unlink+0x8c6/0xca0 [mdd]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e3a788&amp;gt;] mdo_unlink+0x18/0x50 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e3e005&amp;gt;] mdt_reint_unlink+0x835/0x1030 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e3a571&amp;gt;] mdt_reint_rec+0x41/0xe0 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e1fed3&amp;gt;] mdt_reint_internal+0x4c3/0x780 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e201d4&amp;gt;] mdt_reint+0x44/0xe0 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e22ada&amp;gt;] mdt_handle_common+0x52a/0x1470 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0e5f5f5&amp;gt;] mds_regular_handle+0x15/0x20 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa07710c5&amp;gt;] ptlrpc_server_handle_request+0x385/0xc00 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa04998d5&amp;gt;] ? lc_watchdog_touch+0x65/0x170 [libcfs]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0769a69&amp;gt;] ? ptlrpc_wait_event+0xa9/0x2d0 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa077389d&amp;gt;] ptlrpc_main+0xafd/0x1780 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0772da0&amp;gt;] ? ptlrpc_main+0x0/0x1780 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
&amp;lt;4&amp;gt;
&amp;lt;0&amp;gt;Kernel panic - not syncing: LBUG
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;We will try to use yesterday&apos;s process to identify and fix a bad llog.&lt;/p&gt;

&lt;p&gt;Please stand by for more info and let us know if there are additional debug steps we should take.&lt;/p&gt;</comment>
                            <comment id="156853" author="mhanafi" created="Fri, 24 Jun 2016 14:18:47 +0000"  >&lt;p&gt;This is the file with bad record causing the LBUG.&lt;/p&gt;</comment>
                            <comment id="156854" author="mhanafi" created="Fri, 24 Jun 2016 14:20:22 +0000"  >&lt;p&gt;Today&apos;s crash was also from a file dated in Feb. Since we can&apos;t tell what files are bad I delete all files from Feb.&lt;/p&gt;</comment>
                            <comment id="156895" author="mhanafi" created="Fri, 24 Jun 2016 19:04:14 +0000"  >&lt;p&gt;I think we must have more of these corrupted files. Is there a quick method for finding them? &lt;/p&gt;</comment>
                            <comment id="156912" author="pjones" created="Fri, 24 Jun 2016 20:54:08 +0000"  >&lt;p&gt;Mike&lt;/p&gt;

&lt;p&gt;Please can you advise as to what is not handled in the attached llog file and how this could be better handled?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="156922" author="mhanafi" created="Fri, 24 Jun 2016 22:08:00 +0000"  >&lt;p&gt;Would it make sense to deal with this assertion rather than just an LBUG. &lt;/p&gt;</comment>
                            <comment id="156923" author="mhanafi" created="Fri, 24 Jun 2016 22:28:12 +0000"  >&lt;p&gt;If llog_catfix can&apos;t read the file and errors out&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Llog is not catalog, llh_size: 0, need 64
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;means that the file is corrupted  and can potentially trigger this lbug.&lt;/p&gt;

</comment>
                            <comment id="156957" author="tappro" created="Sun, 26 Jun 2016 16:36:34 +0000"  >&lt;p&gt;First of all the llog_init_handle() should handle llog_read_header() errors correctly, not allowing the further use of such llog_handle. This will prevent similar lbugs(). It is the origin of that issue. &lt;/p&gt;</comment>
                            <comment id="156958" author="mhanafi" created="Sun, 26 Jun 2016 19:20:53 +0000"  >&lt;p&gt;Can we expect a patch? Any idea what is causing the corruption?&lt;/p&gt;</comment>
                            <comment id="156960" author="tappro" created="Sun, 26 Jun 2016 20:14:15 +0000"  >&lt;p&gt;Mahmoud, the message about &apos;Llog is not catalog&apos; is not an error or corruption, it is just saying that llog file is not catalog as expected by llog_catfix tool. I am checking 12744.save now. It looks like plain llog  with unlink records in it, I am still looking through it.&lt;/p&gt;

&lt;p&gt;yes, I am working on patch at least to handle llog_read_header() errors gracefully.&lt;/p&gt;</comment>
                            <comment id="157176" author="mhanafi" created="Tue, 28 Jun 2016 19:29:59 +0000"  >&lt;p&gt;Any updates?&lt;/p&gt;</comment>
                            <comment id="157304" author="tappro" created="Wed, 29 Jun 2016 18:27:24 +0000"  >&lt;p&gt;Hello, am I right that you are using 2.5.3 Lustre with additional patches? Are there patches related to llog subsystem? Could you make a list of them?&lt;/p&gt;

&lt;p&gt;I have found one suspicious thing in corrupted llog, checking the code now to understand how that may happen.&lt;/p&gt;</comment>
                            <comment id="157305" author="pjones" created="Wed, 29 Jun 2016 18:34:24 +0000"  >&lt;p&gt;Mike is setting up a github account. I will provide the details to Jay so he can grant the appropriate access to see NASA&apos;s tree&lt;/p&gt;</comment>
                            <comment id="157310" author="mhanafi" created="Wed, 29 Jun 2016 19:19:30 +0000"  >&lt;p&gt;FYI, we have a mix of 2.5.3, 2.7.1, and 2.7.2. We are moving all of them to 2.7.2 but that will take some time so we would need patch for 2.5.3 and 2.7.1. Jay will give you access to the repo.&lt;/p&gt;
</comment>
                            <comment id="157465" author="tappro" created="Thu, 30 Jun 2016 21:17:04 +0000"  >&lt;p&gt;Mahmoud, in your very first comment about systemtap info, the llog_process_th is 9252 but in the next one it is 9553. Can that be this is another thread and related llog file is different from that one causing assertion? Do you still have any data related to that or similar lbug?&lt;/p&gt;</comment>
                            <comment id="157472" author="mhanafi" created="Thu, 30 Jun 2016 23:04:58 +0000"  >&lt;p&gt;The 9252 was for a previous crash where I didn&apos;t print out the full data structure. The second time (9553) I printing the correct structures.   &lt;br/&gt;
We did have a crash the next day where I used the systemtap script to find the file and delete it. But i didn&apos;t save a copy.&lt;/p&gt;</comment>
                            <comment id="157510" author="tappro" created="Fri, 1 Jul 2016 13:47:02 +0000"  >&lt;p&gt;OK, got it. Another question then - there should be CATALOG file with just list of catalog llog ids in it, could you look in it and provide those catalog llogs for inspection? No need to remove them, I want just check how big they are and what is inside.&lt;/p&gt;</comment>
                            <comment id="157544" author="gerrit" created="Fri, 1 Jul 2016 17:08:14 +0000"  >&lt;p&gt;Mike Pershin (mike.pershin@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/21128&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/21128&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8320&quot; title=&quot;:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8320&quot;&gt;&lt;del&gt;LU-8320&lt;/del&gt;&lt;/a&gt; llog: prevent llog ID re-use.&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_5&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: a051f8c40337d9a577bee88dab1b50477d84061d&lt;/p&gt;</comment>
                            <comment id="157547" author="tappro" created="Fri, 1 Jul 2016 17:15:58 +0000"  >&lt;p&gt;I think I&apos;ve found the reason of this issue, it looks like the newly generated llog ID may cycle through zero and match some very old llog files. In that case it is considered as new llog (not initialized yet, but since it exists on disk several check are passed while they shouldn&apos;t and we have that assertion. That scenario fits well with what I see in llog file provided and assertion with llh == NULL, as well as with your comment that problem files are very old.&lt;/p&gt;</comment>
                            <comment id="157552" author="mhanafi" created="Fri, 1 Jul 2016 17:22:28 +0000"  >&lt;p&gt;Great!. do you still need to the CATALOGS file? Is the an issue with 2.7? if so we will need a patch for that as well.&lt;/p&gt;

&lt;p&gt;Why would we have old llog files still around?&lt;/p&gt;</comment>
                            <comment id="157554" author="tappro" created="Fri, 1 Jul 2016 17:26:24 +0000"  >&lt;p&gt;Also I&apos;d recommend to apply patch from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt; in addition to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7079&quot; title=&quot;OSP shouldn&amp;#39;t discard requests due to imp_peer_committed_transno&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7079&quot;&gt;&lt;del&gt;LU-7079&lt;/del&gt;&lt;/a&gt; (already in your tree) which might help to avoid having very old llogs.&lt;br/&gt;
Consider also the patch from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4528&quot; title=&quot;osd_trans_exec_op()) ASSERTION( oti-&amp;gt;oti_declare_ops_rb[rb] &amp;gt; 0 ) failed: rb = 0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4528&quot;&gt;&lt;del&gt;LU-4528&lt;/del&gt;&lt;/a&gt; (&lt;a href=&quot;http://review.whamcloud.com/#/c/11751/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/11751/&lt;/a&gt;) which wasn&apos;t merged to 2.5 - it helps to avoid several types of corruptions we had in past. It is already in 2.7 so maybe this is not so critical if you are moving to 2.7.&lt;/p&gt;</comment>
                            <comment id="157555" author="tappro" created="Fri, 1 Jul 2016 17:29:21 +0000"  >&lt;p&gt;Mahmoud, there was a bug in OSP which caused some llogs to be not processed after some moment, so they stays forever. This was fixed by &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7079&quot; title=&quot;OSP shouldn&amp;#39;t discard requests due to imp_peer_committed_transno&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7079&quot;&gt;&lt;del&gt;LU-7079&lt;/del&gt;&lt;/a&gt; which was merged in your tree at May 25, may it be so that failed node had no such patch applied? Also patch from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt; solved similar problem which may cause stuck llog files.&lt;/p&gt;

&lt;p&gt;I don&apos;t need those CATALOGS file right now, but if you will have time, I&apos;d like to look at llog files in it to check how many plain llogs are in use.&lt;/p&gt;</comment>
                            <comment id="157557" author="gerrit" created="Fri, 1 Jul 2016 17:32:43 +0000"  >&lt;p&gt;Mike Pershin (mike.pershin@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/21130&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/21130&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8320&quot; title=&quot;:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8320&quot;&gt;&lt;del&gt;LU-8320&lt;/del&gt;&lt;/a&gt; llog: prevent llog ID re-use.&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_7&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: d5e1cfbd9bd5bfdcd6d5c6b029e4cf578c9e75b8&lt;/p&gt;</comment>
                            <comment id="157563" author="jaylan" created="Fri, 1 Jul 2016 17:55:28 +0000"  >&lt;p&gt;The 2.5.3 lustre server still running 2.5.3-6nasS, and &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7079&quot; title=&quot;OSP shouldn&amp;#39;t discard requests due to imp_peer_committed_transno&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7079&quot;&gt;&lt;del&gt;LU-7079&lt;/del&gt;&lt;/a&gt; patch was included in 2.5.3-6.1nasS.&lt;/p&gt;</comment>
                            <comment id="157566" author="jaylan" created="Fri, 1 Jul 2016 18:14:59 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt; patch has not landed to b2_7_fe yet. I cherry-picked it to our nas-2.7.1 and nas-2.7.2 anyway (locally, not pushed to github yet.) OTOH, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt; patch caused conflicts in nas-2.5.3. Since there is a workaround (ie, find the offending file and remove it), I think we do not need &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt; on nas-2.5.3.&lt;/p&gt;

&lt;p&gt;So, we have &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4528&quot; title=&quot;osd_trans_exec_op()) ASSERTION( oti-&amp;gt;oti_declare_ops_rb[rb] &amp;gt; 0 ) failed: rb = 0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4528&quot;&gt;&lt;del&gt;LU-4528&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7079&quot; title=&quot;OSP shouldn&amp;#39;t discard requests due to imp_peer_committed_transno&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7079&quot;&gt;&lt;del&gt;LU-7079&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6696&quot; title=&quot;ASSERTION( rc == 0 || rc == LLOG_PROC_BREAK ) failed: 0 changes, 0 in progress, 0 in flight: -5&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6696&quot;&gt;&lt;del&gt;LU-6696&lt;/del&gt;&lt;/a&gt;, and &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt; on nas-2.7.1 and nas-2.7.2. I tried &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8320&quot; title=&quot;:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8320&quot;&gt;&lt;del&gt;LU-8320&lt;/del&gt;&lt;/a&gt; and it was applied cleanly on nas-2.7.x, but I will wait until it gets reviews.&lt;/p&gt;</comment>
                            <comment id="157612" author="gerrit" created="Mon, 4 Jul 2016 17:51:37 +0000"  >&lt;p&gt;Mike Pershin (mike.pershin@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/21144&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/21144&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8320&quot; title=&quot;:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8320&quot;&gt;&lt;del&gt;LU-8320&lt;/del&gt;&lt;/a&gt; llog: prevent llog ID re-use.&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 722f308635f118d00a5c4a44fa72d18986ccdac9&lt;/p&gt;</comment>
                            <comment id="160393" author="gerrit" created="Mon, 1 Aug 2016 02:10:04 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/21144/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/21144/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8320&quot; title=&quot;:(llog_osd.c:338:llog_osd_write_rec()) ASSERTION( llh ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8320&quot;&gt;&lt;del&gt;LU-8320&lt;/del&gt;&lt;/a&gt; llog: prevent llog ID re-use.&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: a93ede18ababa3fe1ae8f4a5f92e868589a58cb6&lt;/p&gt;</comment>
                            <comment id="160402" author="pjones" created="Mon, 1 Aug 2016 05:45:12 +0000"  >&lt;p&gt;Landed for 2.9&lt;/p&gt;</comment>
                            <comment id="163463" author="ndauchy" created="Mon, 29 Aug 2016 19:10:50 +0000"  >&lt;p&gt;Please re-open until the backport patch lands to 2.7 FE.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="22006" name="12744.save" size="4153280" author="mhanafi" created="Fri, 24 Jun 2016 14:18:47 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzyflb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10020"><![CDATA[1]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>