<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:22:01 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2059] mgc to backup configuration on osd-based llogs</title>
                <link>https://jira.whamcloud.com/browse/LU-2059</link>
                <project id="10000" key="LU">Lustre</project>
                    <description></description>
                <environment></environment>
        <key id="16192">LU-2059</key>
            <summary>mgc to backup configuration on osd-based llogs</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="jfc">John Fuchs-Chesney</assignee>
                                    <reporter username="bzzz">Alex Zhuravlev</reporter>
                        <labels>
                            <label>yuc2</label>
                    </labels>
                <created>Mon, 1 Oct 2012 02:22:49 +0000</created>
                <updated>Mon, 29 May 2017 06:16:48 +0000</updated>
                            <resolved>Mon, 29 May 2017 06:16:48 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                    <version>Lustre 2.4.1</version>
                    <version>Lustre 2.5.0</version>
                    <version>Lustre 2.6.0</version>
                                    <fixVersion>Lustre 2.4.0</fixVersion>
                    <fixVersion>Lustre 2.5.0</fixVersion>
                    <fixVersion>Lustre 2.6.0</fixVersion>
                    <fixVersion>Lustre 2.5.4</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>12</watches>
                                                                            <comments>
                            <comment id="45791" author="bzzz" created="Mon, 1 Oct 2012 03:02:50 +0000"  >&lt;p&gt;&lt;tt&gt;mgc_process_cfg_log()&lt;/tt&gt; do not make a local copy of configuration if zfs is in use. this is because of &lt;tt&gt;lsi_srv_mnt == NULL&lt;/tt&gt; in:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;        /* Copy the setup log locally &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; we can. Don&lt;span class=&quot;code-quote&quot;&gt;&apos;t mess around &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; we&apos;&lt;/span&gt;re
           running an MGS though (logs are already local). */
	&lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (lctxt &amp;amp;&amp;amp; lsi &amp;amp;&amp;amp; IS_SERVER(lsi) &amp;amp;&amp;amp;
            (lsi-&amp;gt;lsi_srv_mnt == cli-&amp;gt;cl_mgc_vfsmnt) &amp;amp;&amp;amp;
	    !IS_MGS(lsi) &amp;amp;&amp;amp; lsi-&amp;gt;lsi_srv_mnt) {
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;this is done as &lt;tt&gt;mgc_copy_llog()&lt;/tt&gt; still rely on lvfs, see &lt;tt&gt;mgc_llog_init()&lt;/tt&gt;:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;	rc = llog_setup(NULL, obd, olg, LLOG_CONFIG_ORIG_CTXT, tgt,
			&amp;amp;llog_lvfs_ops);
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;this should be changed to &lt;tt&gt;llog_osd_ops&lt;/tt&gt;&lt;/p&gt;
</comment>
                            <comment id="46282" author="adilger" created="Tue, 9 Oct 2012 17:26:11 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/4234&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4234&lt;/a&gt; skips insanity.sh test_2 and test_4 for ZFS configs, and removes skip for conf-sanity.sh test_5d, test_19b, test_21b, test_27a for ldiskfs.&lt;/p&gt;</comment>
                            <comment id="47627" author="adilger" created="Fri, 9 Nov 2012 03:48:52 +0000"  >&lt;p&gt;Moving over to OSD API for the MGC backup will also allow us to delete a few thousand lines of obsolete code in lustre/lvfs/fsfilt*.c and lustre/lvfs/lvfs*.c.&lt;/p&gt;</comment>
                            <comment id="50831" author="tappro" created="Fri, 18 Jan 2013 14:29:21 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/5049&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5049&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="52113" author="simmonsja" created="Mon, 11 Feb 2013 08:01:07 +0000"  >&lt;p&gt;Patch needs to be rebased.&lt;/p&gt;</comment>
                            <comment id="52121" author="tappro" created="Mon, 11 Feb 2013 13:48:40 +0000"  >&lt;p&gt;Done&lt;/p&gt;</comment>
                            <comment id="53063" author="utopiabound" created="Tue, 26 Feb 2013 22:24:11 +0000"  >&lt;p&gt;Tests are listed for zfs as ALWAYS_EXCEPTED under bug &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2778&quot; title=&quot;Test failure on test suite conf-sanity, subtest test_57b&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2778&quot;&gt;&lt;del&gt;LU-2778&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="53144" author="tappro" created="Wed, 27 Feb 2013 21:01:40 +0000"  >&lt;p&gt;Maloo failed at replay-ost-single test_7. As I can see that happens because config logs are bigger than 40 blocks as test expects but 44. Need to find out is that just because logs become really so big and test must be fixed or this is patch issue&lt;/p&gt;</comment>
                            <comment id="53262" author="simmonsja" created="Mon, 4 Mar 2013 08:27:54 +0000"  >&lt;p&gt;Mikhail does &lt;a href=&quot;http://review.whamcloud.com/#change,5569&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,5569&lt;/a&gt; replace &lt;a href=&quot;http://review.whamcloud.com/#change,5049&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,5049&lt;/a&gt; ?&lt;/p&gt;</comment>
                            <comment id="53275" author="tappro" created="Mon, 4 Mar 2013 12:57:18 +0000"  >&lt;p&gt;James, it is the same as 5049 but with debug in replay-ost-single to find out how big config logs can be. See my previous comment.&lt;/p&gt;</comment>
                            <comment id="54681" author="simmonsja" created="Fri, 22 Mar 2013 18:17:35 +0000"  >&lt;p&gt;Any updates on this?&lt;/p&gt;</comment>
                            <comment id="55132" author="tappro" created="Sat, 30 Mar 2013 17:17:21 +0000"  >&lt;p&gt;James, I&apos;ve found object leaks with the latest patch. I am figuring out what is wrong.&lt;/p&gt;</comment>
                            <comment id="55193" author="simmonsja" created="Mon, 1 Apr 2013 16:14:05 +0000"  >&lt;p&gt;Excellent, looking forward to your patch.&lt;/p&gt;</comment>
                            <comment id="55218" author="pjones" created="Mon, 1 Apr 2013 19:26:56 +0000"  >&lt;p&gt;Is this a must-have fix for 2.4.0 or can it be tidied up in 2.4.1?&lt;/p&gt;</comment>
                            <comment id="55285" author="pjones" created="Tue, 2 Apr 2013 14:01:44 +0000"  >&lt;p&gt;dropping priority after chatting with Mike.&lt;/p&gt;</comment>
                            <comment id="55312" author="morrone" created="Tue, 2 Apr 2013 17:25:32 +0000"  >&lt;p&gt;Could you elaborate?  This doesn&apos;t seem to me to be the kind of thing we should be doing on maintenance branches.&lt;/p&gt;</comment>
                            <comment id="55317" author="pjones" created="Tue, 2 Apr 2013 17:44:39 +0000"  >&lt;p&gt;Chris&lt;/p&gt;

&lt;p&gt;Mike basically said that this ticket was created to track some technical debt type cleanup and that it would be ok to live with the present situation for 2.4. If, as you say, when the changes are too large when ready then they would end up in 2.5 rather than 2.4.1&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="56970" author="simmonsja" created="Wed, 24 Apr 2013 19:23:49 +0000"  >&lt;p&gt;With the latest test I&apos;m getting this error. I do have a few other patches running with this one so tomorrow I will try a stock 2.3.64 with only this patch to see if the problem still exist.&lt;/p&gt;

&lt;p&gt;Apr 24 17:05:13 barry-oss1 kernel: [  481.702376] INFO: task mount.lustre:8891 blocked for more than 120 seconds.&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.720497] &quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.736916] mount.lustre  D 0000000000000001     0  8891   8890 0x00000080&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.751549]  ffff8803e085b798 0000000000000086 0000000000000000 ffff8803c8cee000&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.766831]  ffff8803e085b748 ffffc9001ecf3030 0000000000000246 0000000000000246&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.782091]  ffff8803e06a7af8 ffff8803e085bfd8 000000000000fb88 ffff8803e06a7af8&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.797351] Call Trace:&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.809126]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa04586fe&amp;gt;&amp;#93;&lt;/span&gt; cfs_waitq_wait+0xe/0x10 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.822999]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b5e43&amp;gt;&amp;#93;&lt;/span&gt; lu_object_find_at+0xb3/0x360 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.837409]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81063310&amp;gt;&amp;#93;&lt;/span&gt; ? default_wake_function+0x0/0x20&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.851209]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b851a&amp;gt;&amp;#93;&lt;/span&gt; dt_locate_at+0x3a/0x140 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.865294]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0598333&amp;gt;&amp;#93;&lt;/span&gt; local_oid_storage_init+0x513/0x1200 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.887428]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0d390bc&amp;gt;&amp;#93;&lt;/span&gt; ? osd_index_ea_lookup+0x2fc/0x770 &lt;span class=&quot;error&quot;&gt;&amp;#91;osd_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.908544]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa058ff7a&amp;gt;&amp;#93;&lt;/span&gt; llog_osd_setup+0xfa/0x360 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.922742]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa056a1d2&amp;gt;&amp;#93;&lt;/span&gt; llog_setup+0x352/0x920 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.938678]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0db2fe9&amp;gt;&amp;#93;&lt;/span&gt; mgc_set_info_async+0x1249/0x1960 &lt;span class=&quot;error&quot;&gt;&amp;#91;mgc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.953709]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa04682a1&amp;gt;&amp;#93;&lt;/span&gt; ? libcfs_debug_msg+0x41/0x50 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.967880]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05d7ab0&amp;gt;&amp;#93;&lt;/span&gt; server_mgc_set_fs+0x120/0x520 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  481.988772]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05de4e5&amp;gt;&amp;#93;&lt;/span&gt; server_start_targets+0x85/0x19c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.003376]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0458b2e&amp;gt;&amp;#93;&lt;/span&gt; ? cfs_free+0xe/0x10 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.016577]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05af3c5&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_start_mgc+0x4b5/0x1ee0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.031047]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa057caf8&amp;gt;&amp;#93;&lt;/span&gt; ? class_new_export+0x718/0x940 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.048057]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05e09cc&amp;gt;&amp;#93;&lt;/span&gt; server_fill_super+0xbac/0x1660 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.062188]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b0fc8&amp;gt;&amp;#93;&lt;/span&gt; lustre_fill_super+0x1d8/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.076384]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b0df0&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_fill_super+0x0/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.090313]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811842bf&amp;gt;&amp;#93;&lt;/span&gt; get_sb_nodev+0x5f/0xa0&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.102667]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05a8965&amp;gt;&amp;#93;&lt;/span&gt; lustre_get_sb+0x25/0x30 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.123386]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811838fb&amp;gt;&amp;#93;&lt;/span&gt; vfs_kern_mount+0x7b/0x1b0&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.138730]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81183aa2&amp;gt;&amp;#93;&lt;/span&gt; do_kern_mount+0x52/0x130&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.152137]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811a3d52&amp;gt;&amp;#93;&lt;/span&gt; do_mount+0x2d2/0x8d0&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.166646]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811a43e0&amp;gt;&amp;#93;&lt;/span&gt; sys_mount+0x90/0xe0&lt;br/&gt;
Apr 24 17:05:13 barry-oss1 kernel: [  482.178209]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100b072&amp;gt;&amp;#93;&lt;/span&gt; system_call_fastpath+0x16/0x1b&lt;br/&gt;
Apr 24 17:05:15 barry-oss1 kernel: [  483.945364] LustreError: 137-5: UUID &apos;lustre-OST0014_UUID&apos; is not available for connect (no target)&lt;br/&gt;
Apr 24 17:05:15 barry-oss1 kernel: [  483.962848] LustreError: Skipped 261 previous similar messages&lt;/p&gt;
</comment>
                            <comment id="57023" author="simmonsja" created="Thu, 25 Apr 2013 11:45:09 +0000"  >&lt;p&gt;With run plain 2.3.64 with this patch alone still produces this error.&lt;/p&gt;</comment>
                            <comment id="57036" author="tappro" created="Thu, 25 Apr 2013 13:54:20 +0000"  >&lt;p&gt;James, how to reproduce this error? Some specific tests or sequence of actions? Is this first mount?&lt;/p&gt;</comment>
                            <comment id="57038" author="simmonsja" created="Thu, 25 Apr 2013 13:58:36 +0000"  >&lt;p&gt;This is the first mount.&lt;/p&gt;</comment>
                            <comment id="57040" author="tappro" created="Thu, 25 Apr 2013 14:10:03 +0000"  >&lt;p&gt;James, is that with zfs or doesn&apos;t matter?&lt;/p&gt;</comment>
                            <comment id="57042" author="simmonsja" created="Thu, 25 Apr 2013 14:43:25 +0000"  >&lt;p&gt;ldiskfs. No zfs setup here.&lt;/p&gt;</comment>
                            <comment id="57048" author="tappro" created="Thu, 25 Apr 2013 15:32:39 +0000"  >&lt;p&gt;OK, I can reproduce that after patch re-base over the latest master. &lt;/p&gt;</comment>
                            <comment id="57050" author="simmonsja" created="Thu, 25 Apr 2013 16:00:13 +0000"  >&lt;p&gt;Time for git bisect &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;</comment>
                            <comment id="57190" author="tappro" created="Sat, 27 Apr 2013 18:55:39 +0000"  >&lt;p&gt;OK, I found the conflict and solve it, please try again, patch is re-pushed&lt;/p&gt;</comment>
                            <comment id="57226" author="simmonsja" created="Mon, 29 Apr 2013 11:31:04 +0000"  >&lt;p&gt;Yes the patch working great now. One comment to point out to testers of this patch. You need to umount the file system and reformat the MGT. After you remount the file system should be good to go.&lt;/p&gt;</comment>
                            <comment id="57250" author="morrone" created="Mon, 29 Apr 2013 18:30:09 +0000"  >&lt;p&gt;That last comment about reformatting is disturbing.  Hopefully that was only if you were running previous version of &lt;a href=&quot;http://review.whamcloud.com/5049&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;change 5049&lt;/a&gt;?&lt;/p&gt;</comment>
                            <comment id="57429" author="simmonsja" created="Wed, 1 May 2013 13:11:15 +0000"  >&lt;p&gt;The format for the changelog is altered with this patch. Because of this you need to delete the MGT change logs. Yes reformatting my MGT was a bit heavy handed. You still need to unmount your file system and then use tunefs.lustre --writeconf ... to erase your old data. Then remount and the file system should be good to go.&lt;/p&gt;</comment>
                            <comment id="57448" author="morrone" created="Wed, 1 May 2013 17:15:22 +0000"  >&lt;p&gt;We aren&apos;t going to land something that requires all existing users of Lustre to writeconf their filesystems, right?&lt;/p&gt;</comment>
                            <comment id="58875" author="simmonsja" created="Mon, 20 May 2013 14:39:45 +0000"  >&lt;p&gt;Looks like this is a 2.5 thing. Peter can you change the Affected version to 2.5&lt;/p&gt;</comment>
                            <comment id="58877" author="pjones" created="Mon, 20 May 2013 14:44:52 +0000"  >&lt;p&gt;ok done&lt;/p&gt;</comment>
                            <comment id="58907" author="tappro" created="Mon, 20 May 2013 17:23:31 +0000"  >&lt;p&gt;Christopher, the format of changelog was changed not in this patch, it was just rebased over the latest master.&lt;/p&gt;</comment>
                            <comment id="61848" author="simmonsja" created="Fri, 5 Jul 2013 15:39:39 +0000"  >&lt;p&gt;Testing with the latest patch I&apos;m getting the following error when attempting to bring up a file system.&lt;/p&gt;

&lt;p&gt;Jul  5 11:36:30 barry-oss1 kernel: [  530.526152] LDISKFS-fs (dm-9): recovery complete&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.531930] LDISKFS-fs (dm-9): mounted filesystem with ordered data mode. quota=on. Opts:&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.714659] LustreError: 3465:0:(fld_handler.c:123:fld_server_lookup()) ASSERTION( info != ((void *)0) )&lt;br/&gt;
failed: &lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.725093] LustreError: 3465:0:(fld_handler.c:123:fld_server_lookup()) LBUG&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.732252] Pid: 3465, comm: mount.lustre&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.736331] &lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.736332] Call Trace:&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.740414]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0472895&amp;gt;&amp;#93;&lt;/span&gt; libcfs_debug_dumpstack+0x55/0x80 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.747488]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0472e97&amp;gt;&amp;#93;&lt;/span&gt; lbug_with_loc+0x47/0xb0 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.753779]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa095dddd&amp;gt;&amp;#93;&lt;/span&gt; fld_server_lookup+0x2bd/0x3d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;fld&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.760312]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8119c87e&amp;gt;&amp;#93;&lt;/span&gt; ? generic_detach_inode+0x18e/0x1f0&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.766784]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0d77d91&amp;gt;&amp;#93;&lt;/span&gt; osd_fld_lookup+0x71/0x1d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;osd_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.773667]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8119c6e2&amp;gt;&amp;#93;&lt;/span&gt; ? iput+0x62/0x70&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.778559]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0d77f8a&amp;gt;&amp;#93;&lt;/span&gt; osd_remote_fid+0x9a/0x280 &lt;span class=&quot;error&quot;&gt;&amp;#91;osd_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.785456]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0d84ca1&amp;gt;&amp;#93;&lt;/span&gt; osd_index_ea_lookup+0x521/0x850 &lt;span class=&quot;error&quot;&gt;&amp;#91;osd_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.792949]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05ed34f&amp;gt;&amp;#93;&lt;/span&gt; dt_lookup_dir+0x6f/0x130 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.799515]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05cbd25&amp;gt;&amp;#93;&lt;/span&gt; llog_osd_open+0x475/0xbb0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.806156]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa059832a&amp;gt;&amp;#93;&lt;/span&gt; llog_open+0xba/0x2c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.812362]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa059bf81&amp;gt;&amp;#93;&lt;/span&gt; llog_backup+0x61/0x500 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.818741]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81281d00&amp;gt;&amp;#93;&lt;/span&gt; ? sprintf+0x40/0x50&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.829312]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0e046e2&amp;gt;&amp;#93;&lt;/span&gt; mgc_process_log+0x1192/0x18e0 &lt;span class=&quot;error&quot;&gt;&amp;#91;mgc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.841291]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0dfe370&amp;gt;&amp;#93;&lt;/span&gt; ? mgc_blocking_ast+0x0/0x7e0 &lt;span class=&quot;error&quot;&gt;&amp;#91;mgc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.853282]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa074ccf0&amp;gt;&amp;#93;&lt;/span&gt; ? ldlm_completion_ast+0x0/0x960 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.865730]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0e062c4&amp;gt;&amp;#93;&lt;/span&gt; mgc_process_config+0x594/0xed0 &lt;span class=&quot;error&quot;&gt;&amp;#91;mgc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.877804]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05e14a6&amp;gt;&amp;#93;&lt;/span&gt; lustre_process_log+0x256/0xaa0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.890296]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b0922&amp;gt;&amp;#93;&lt;/span&gt; ? class_name2dev+0x42/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.902398]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81167d73&amp;gt;&amp;#93;&lt;/span&gt; ? kmem_cache_alloc_trace+0x1a3/0x1b0&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.914407]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b09ce&amp;gt;&amp;#93;&lt;/span&gt; ? class_name2obd+0xe/0x30 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.926353]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0615121&amp;gt;&amp;#93;&lt;/span&gt; server_start_targets+0x1821/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.939063]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05e4ac3&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_start_mgc+0x493/0x1e90 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.951484]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05b4638&amp;gt;&amp;#93;&lt;/span&gt; ? class_new_export+0x708/0x930 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.963790]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05dca10&amp;gt;&amp;#93;&lt;/span&gt; ? class_config_llog_handler+0x0/0x1820 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.976755]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0618ccc&amp;gt;&amp;#93;&lt;/span&gt; server_fill_super+0xbac/0x1a04 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  530.989081]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05e6698&amp;gt;&amp;#93;&lt;/span&gt; lustre_fill_super+0x1d8/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.001275]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05e64c0&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_fill_super+0x0/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.013381]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8118430f&amp;gt;&amp;#93;&lt;/span&gt; get_sb_nodev+0x5f/0xa0&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.023912]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa05de335&amp;gt;&amp;#93;&lt;/span&gt; lustre_get_sb+0x25/0x30 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.035448]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8118394b&amp;gt;&amp;#93;&lt;/span&gt; vfs_kern_mount+0x7b/0x1b0&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.046172]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81183af2&amp;gt;&amp;#93;&lt;/span&gt; do_kern_mount+0x52/0x130&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.056713]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811a3d22&amp;gt;&amp;#93;&lt;/span&gt; do_mount+0x2d2/0x8d0&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.066857]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811a43b0&amp;gt;&amp;#93;&lt;/span&gt; sys_mount+0x90/0xe0&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.076861]  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100b072&amp;gt;&amp;#93;&lt;/span&gt; system_call_fastpath+0x16/0x1b&lt;br/&gt;
Jul  5 11:36:30 barry-oss1 kernel: [  531.087802]&lt;/p&gt;</comment>
                            <comment id="62965" author="utopiabound" created="Thu, 25 Jul 2013 13:35:42 +0000"  >&lt;p&gt;Re-enable all tests EXCEPTed or skipped:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/7113&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7113&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="65945" author="yujian" created="Fri, 6 Sep 2013 14:51:04 +0000"  >&lt;p&gt;Lustre build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_4/44/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_4/44/&lt;/a&gt; (2.4.1 RC1)&lt;br/&gt;
Distro/Arch: RHEL6.4/x86_64&lt;br/&gt;
FSTYPE=zfs&lt;br/&gt;
MDSCOUNT=4&lt;/p&gt;

&lt;p&gt;insanity test 1 failed as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Starting mds2: -o user_xattr,acl  lustre-mdt2/mdt2 /mnt/mds2
CMD: wtm-29vm7 mkdir -p /mnt/mds2; mount -t lustre -o user_xattr,acl  		                   lustre-mdt2/mdt2 /mnt/mds2
wtm-29vm7: mount.lustre: mount lustre-mdt2/mdt2 at /mnt/mds2 failed: Input/output error
wtm-29vm7: Is the MGS running?
Start of lustre-mdt2/mdt2 on mds2 failed 5
 insanity test_1: @@@@@@ FAIL: test_1 failed with 2 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Console log on wtm-29vm7:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;12:50:34:Lustre: DEBUG MARKER: mkdir -p /mnt/mds2; mount -t lustre -o user_xattr,acl  		                   lustre-mdt2/mdt2 /mnt/mds2
12:50:34:LustreError: 16886:0:(client.c:1052:ptlrpc_import_delay_req()) @@@ send limit expired   req@ffff88007073e000 x1445162105843372/t0(0) o253-&amp;gt;MGC10.10.17.33@tcp@10.10.17.33@tcp:26/25 lens 4768/4768 e 0 to 0 dl 0 ref 2 fl Rpc:W/0/ffffffff rc 0/-1
12:50:34:LustreError: 16886:0:(obd_mount_server.c:1124:server_register_target()) lustre-MDT0001: error registering with the MGS: rc = -5 (not fatal)
12:50:34:LustreError: 16886:0:(client.c:1052:ptlrpc_import_delay_req()) @@@ send limit expired   req@ffff8800740bb400 x1445162105843380/t0(0) o101-&amp;gt;MGC10.10.17.33@tcp@10.10.17.33@tcp:26/25 lens 328/344 e 0 to 0 dl 0 ref 2 fl Rpc:W/0/ffffffff rc 0/-1
12:50:34:LustreError: 16886:0:(client.c:1052:ptlrpc_import_delay_req()) @@@ send limit expired   req@ffff8800740bb400 x1445162105843384/t0(0) o101-&amp;gt;MGC10.10.17.33@tcp@10.10.17.33@tcp:26/25 lens 328/344 e 0 to 0 dl 0 ref 2 fl Rpc:W/0/ffffffff rc 0/-1
12:50:34:LustreError: 15c-8: MGC10.10.17.33@tcp: The configuration from log &apos;lustre-MDT0001&apos; failed (-5). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
12:50:34:LustreError: 16886:0:(obd_mount_server.c:1258:server_start_targets()) failed to start server lustre-MDT0001: -5
12:50:34:LustreError: 16886:0:(obd_mount_server.c:1700:server_fill_super()) Unable to start targets: -5
12:50:34:LustreError: 16886:0:(obd_mount_server.c:849:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0001: Can&apos;t end config log lustre-client.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/059ee4a8-164c-11e3-aa2a-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/059ee4a8-164c-11e3-aa2a-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="69303" author="jlevi" created="Fri, 18 Oct 2013 16:54:37 +0000"  >&lt;p&gt;Re-enabling the tests is being tracked in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4121&quot; title=&quot;re-enable tests disabled due to LU-2059&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4121&quot;&gt;&lt;del&gt;LU-4121&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="69304" author="tappro" created="Fri, 18 Oct 2013 16:54:59 +0000"  >&lt;p&gt;patch was merged, are there more work to do in context of this bug?&lt;/p&gt;</comment>
                            <comment id="71776" author="utopiabound" created="Mon, 18 Nov 2013 13:52:17 +0000"  >&lt;p&gt;Skipping insanity/10 because it fails in the same way as the other always_excepted bugs fail&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/8316&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8316&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="71828" author="adilger" created="Mon, 18 Nov 2013 21:21:36 +0000"  >&lt;p&gt;Reopening this bug, since it doesn&apos;t seem that the landed patch actually allows the OST to start when the MGS is down.  I guess the config log might be replicated, but the MGC is somehow blocked waiting for the MGS?&lt;/p&gt;</comment>
                            <comment id="79588" author="jlevi" created="Tue, 18 Mar 2014 17:58:39 +0000"  >&lt;p&gt;Mike,&lt;br/&gt;
Can you please comment on the sanity test 10 failures?&lt;br/&gt;
Thank you!&lt;/p&gt;</comment>
                            <comment id="81452" author="tappro" created="Fri, 11 Apr 2014 20:03:39 +0000"  >&lt;p&gt;yes, I have to investigate this first&lt;/p&gt;</comment>
                            <comment id="82954" author="adilger" created="Wed, 30 Apr 2014 23:37:36 +0000"  >&lt;p&gt;I think the first step is to run a local test to verify manually that the OST is actually creating a local copy of the config llog on the first mount, and then check that the OST can start from the local llog if the MGS is not running. &lt;/p&gt;</comment>
                            <comment id="83155" author="tappro" created="Mon, 5 May 2014 07:00:12 +0000"  >&lt;p&gt;Well, I did that for sure when did the initial patch for this ticket which makes local config llogs. Nevertheless I agree this is first step to do now, just to make sure it still works. Meanwhile I tend to think you are right that problem might be in locking or something else what cause MGC to wait for MGS&lt;/p&gt;</comment>
                            <comment id="83156" author="bzzz" created="Mon, 5 May 2014 07:18:25 +0000"  >&lt;p&gt;hmm, there were tests in conf-sanity to verify OST can start with inactive MGS.&lt;/p&gt;</comment>
                            <comment id="83414" author="adilger" created="Wed, 7 May 2014 17:33:21 +0000"  >&lt;p&gt;I noticed that several conf-sanity.sh tests are being skipped for ZFS because of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2059&quot; title=&quot;mgc to backup configuration on osd-based llogs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2059&quot;&gt;&lt;del&gt;LU-2059&lt;/del&gt;&lt;/a&gt; &quot;no local config for ZFS&quot;, but I think that is no longer true?&lt;/p&gt;

&lt;p&gt;conf-sanity.sh: 5d, 19b, 21b, 27a&lt;br/&gt;
insanity.sh: 1, 2, 4, 10&lt;/p&gt;

&lt;p&gt;There is also conf-sanity.sh test_50h that doesn&apos;t give a reason, and sanity-quota.sh test_7d.&lt;/p&gt;

&lt;p&gt;It would be good to submit a patch or two to enable these tests along with a &lt;tt&gt;Test-parameters:&lt;/tt&gt; line that runs the relevant test ten times to see what the success rate is.&lt;/p&gt;</comment>
                            <comment id="83973" author="tappro" created="Tue, 13 May 2014 11:20:14 +0000"  >&lt;p&gt;the local config logs are created as expected, meanwhile there was problem with &apos;params&apos; log, it can be empty and that case shouldn&apos;t be critical. I&apos;ve enabled all excepted tests and they passed with both ldiskfs and zfs except insanity test_1. It failed only with zfs as following:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;May 13 14:56:02 nodez kernel: LustreError: 13691:0:(mgc_request.c:516:do_requeue()) failed processing log: -5
May 13 14:56:02 nodez kernel: LustreError: 13691:0:(mgc_request.c:516:do_requeue()) Skipped 2 previous similar messages
May 13 14:56:15 nodez kernel: INFO: task mount.lustre:15680 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.
May 13 14:56:15 nodez kernel:      Tainted: P           ---------------    2.6.32 #11
May 13 14:56:15 nodez kernel: &lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.
May 13 14:56:15 nodez kernel: mount.lustre  D ffff880001e0f7e0  3592 15680  15679 0x00000080
May 13 14:56:15 nodez kernel: ffff8800241167f0 0000000000000082 0000000000000000 ffffffff814ffbf6
May 13 14:56:15 nodez kernel: ffff88002d7a3b38 ffffffff81a2d020 ffff880024116ac0 000000000000f7e0
May 13 14:56:15 nodez kernel: ffff88001eaabfd8 ffff880024116ac0 ffff8800241167f0 ffff88001eaab6c8
May 13 14:56:15 nodez kernel: Call Trace:
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff814ffbf6&amp;gt;] ? __mutex_lock_slowpath+0x1f6/0x2a0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff814ff325&amp;gt;] ? schedule_timeout+0x1c5/0x280
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff81048804&amp;gt;] ? task_rq_lock+0x54/0xa0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff810512cd&amp;gt;] ? set_cpus_allowed_ptr+0xad/0x1b0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff814ff325&amp;gt;] ? schedule_timeout+0x1c5/0x280
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff814fe319&amp;gt;] ? wait_for_common+0x139/0x1a0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff81050f80&amp;gt;] ? default_wake_function+0x0/0x20
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0eed770&amp;gt;] ? client_lwp_config_process+0x0/0x1540 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0e72d9c&amp;gt;] ? llog_process_or_fork+0x2bc/0x4c0 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0e72faf&amp;gt;] ? llog_process+0xf/0x20 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ea9f79&amp;gt;] ? class_config_parse_llog+0x1e9/0x340 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0e78b02&amp;gt;] ? __llog_ctxt_put+0xf2/0x150 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa027948c&amp;gt;] ? mgc_process_cfg_log+0x56c/0xe90 [mgc]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa027bdfa&amp;gt;] ? mgc_process_log+0x59a/0x1390 [mgc]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0dc7f6f&amp;gt;] ? libcfs_debug_msg+0x3f/0x50 [libcfs]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa027ccc5&amp;gt;] ? do_config_log_add+0xd5/0x5b0 [mgc]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0276dd0&amp;gt;] ? mgc_blocking_ast+0x0/0x800 [mgc]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0300240&amp;gt;] ? ldlm_completion_ast+0x0/0x910 [ptlrpc]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa027db8f&amp;gt;] ? mgc_process_config+0x69f/0x1290 [mgc]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ec1dc9&amp;gt;] ? lustre_process_log+0x219/0x1000 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ec2106&amp;gt;] ? lustre_process_log+0x556/0x1000 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff811283e2&amp;gt;] ? kmem_cache_alloc_trace+0x1c2/0x1d0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ee89c6&amp;gt;] ? lustre_start_lwp+0x146/0x700 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ee8a1f&amp;gt;] ? lustre_start_lwp+0x19f/0x700 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0eead66&amp;gt;] ? server_start_targets+0xeb6/0x1d30 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ebe4c6&amp;gt;] ? lustre_start_mgc+0x456/0x29b0 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0eb6af0&amp;gt;] ? class_config_llog_handler+0x0/0x1e50 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ef1a3f&amp;gt;] ? server_fill_super+0xbff/0x15c0 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ec0a20&amp;gt;] ? lustre_fill_super+0x0/0x520 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ec0bf8&amp;gt;] ? lustre_fill_super+0x1d8/0x520 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ec0a20&amp;gt;] ? lustre_fill_super+0x0/0x520 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0ec0a20&amp;gt;] ? lustre_fill_super+0x0/0x520 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff8113c9a9&amp;gt;] ? get_sb_nodev+0x59/0xb0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffffa0eb8b5c&amp;gt;] ? lustre_get_sb+0x1c/0x30 [obdclass]
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff8113bf16&amp;gt;] ? vfs_kern_mount+0x76/0x1b0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff8113c0c3&amp;gt;] ? do_kern_mount+0x53/0x130
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff81157d32&amp;gt;] ? do_mount+0x302/0x8f0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff811583ba&amp;gt;] ? sys_mount+0x9a/0xf0
May 13 14:56:15 nodez kernel: [&amp;lt;ffffffff81009f42&amp;gt;] ? system_call_fastpath+0x16/0x1b
May 13 14:57:14 nodez kernel: LustreError: 137-5: lustre-MDT0000_UUID: not available &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; connect from 0@lo (no target). If you are running an HA pair check that the target is mounted on the other server.
May 13 14:57:14 nodez kernel: LustreError: Skipped 129 previous similar messages
May 13 14:57:24 nodez kernel: LustreError: 11-0: lustre-MDT0000-osp-MDT0001: Communicating with 0@lo, operation mds_connect failed with -19.
May 13 14:57:24 nodez kernel: LustreError: Skipped 51 previous similar messages
May 13 14:58:15 nodez kernel: INFO: task ll_cfg_requeue:13691 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.
May 13 14:58:15 nodez kernel:      Tainted: P           ---------------    2.6.32 #11
May 13 14:58:15 nodez kernel: &lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.
May 13 14:58:15 nodez kernel: ll_cfg_requeu D ffff880001e0f7e0  4760 13691      2 0x00000080
May 13 14:58:15 nodez kernel: ffff88001f96c040 0000000000000046 0000000000000000 ffff880001e0f920
May 13 14:58:15 nodez kernel: 0000000000000018 ffffffff81a2d020 ffff88001f96c310 000000000000f7e0
May 13 14:58:15 nodez kernel: ffff880028f99fd8 ffff88001f96c310 ffff88001f96c040 ffff880001e0f920
May 13 14:58:15 nodez kernel: Call Trace:
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff812bc54a&amp;gt;] ? string+0x3a/0xf0
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff812bdbf7&amp;gt;] ? vsnprintf+0x357/0x600
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff814ffb17&amp;gt;] ? __mutex_lock_slowpath+0x117/0x2a0
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff8150031f&amp;gt;] ? mutex_lock+0x1f/0x40
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffffa027b8c3&amp;gt;] ? mgc_process_log+0x63/0x1390 [mgc]
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff81501377&amp;gt;] ? _spin_lock_irqsave+0x17/0x20
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff81068353&amp;gt;] ? lock_timer_base+0x33/0x70
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff8106899e&amp;gt;] ? try_to_del_timer_sync+0x6e/0xd0
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffffa0dc7f6f&amp;gt;] ? libcfs_debug_msg+0x3f/0x50 [libcfs]
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffffa027ea8d&amp;gt;] ? mgc_requeue_thread+0x30d/0x7a0 [mgc]
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff81050f80&amp;gt;] ? default_wake_function+0x0/0x20
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffffa027e780&amp;gt;] ? mgc_requeue_thread+0x0/0x7a0 [mgc]
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff81077e9e&amp;gt;] ? kthread+0x8e/0xa0
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff8100b0ea&amp;gt;] ? child_rip+0xa/0x20
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff81077e10&amp;gt;] ? kthread+0x0/0xa0
May 13 14:58:15 nodez kernel: [&amp;lt;ffffffff8100b0e0&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;That looks like deadlock, not sure why it happens with ZFS only though&lt;/p&gt;</comment>
                            <comment id="83986" author="tappro" created="Tue, 13 May 2014 14:47:43 +0000"  >&lt;p&gt;Patch in gerrit with last fixes related to &apos;params&apos; log handling and enabled tests:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/10311&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10311&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The test_1 in insanity is still excluded, it failed due to deadlock of some kind and need more investigation. Maybe it worths to file another bug for that or reopen &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3562&quot; title=&quot;Test failure on test suite insanity, subtest test_1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3562&quot;&gt;&lt;del&gt;LU-3562&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="84802" author="jlevi" created="Fri, 23 May 2014 18:44:21 +0000"  >&lt;p&gt;Patch landed to Master. Please reopen ticket if more work is needed.&lt;/p&gt;</comment>
                            <comment id="87211" author="utopiabound" created="Fri, 20 Jun 2014 20:53:21 +0000"  >&lt;p&gt;Port to b2_4:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/10771&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10771&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="95379" author="adilger" created="Wed, 1 Oct 2014 06:44:54 +0000"  >&lt;p&gt;sanity-quota test_7d is still being skipped for ZFS filesystems, reportedly because of this bug.&lt;/p&gt;</comment>
                            <comment id="95398" author="adilger" created="Wed, 1 Oct 2014 07:28:09 +0000"  >&lt;p&gt;insanity.sh test_1 is also being skipped for ZFS filesystems due to this bug.&lt;/p&gt;</comment>
                            <comment id="97497" author="yujian" created="Sat, 25 Oct 2014 05:40:11 +0000"  >&lt;p&gt;Here are the back-ported patches for Lustre b2_5 branch:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/12427&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/12427&lt;/a&gt; (from &lt;a href=&quot;http://review.whamcloud.com/10311&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10311&lt;/a&gt;)&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/12428&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/12428&lt;/a&gt; (from &lt;a href=&quot;http://review.whamcloud.com/10589&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10589&lt;/a&gt;)&lt;/p&gt;</comment>
                            <comment id="98843" author="yujian" created="Tue, 11 Nov 2014 01:42:18 +0000"  >&lt;p&gt;The above patches were merged into Lustre b2_5 branch for 2.5.4 release.&lt;/p&gt;</comment>
                            <comment id="100238" author="adilger" created="Fri, 28 Nov 2014 18:39:44 +0000"  >&lt;p&gt;The conf-sanity test_57b is still being skipped and patch &lt;a href=&quot;http://review.whamcloud.com/10357&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10357&lt;/a&gt; which enables it is faint testing. This bug is definitely not fixed. &lt;/p&gt;</comment>
                            <comment id="101688" author="tappro" created="Tue, 16 Dec 2014 08:28:58 +0000"  >&lt;p&gt;Andreas, the patch you mentioned passed test_57b successfully. Did I miss something?&lt;/p&gt;</comment>
                            <comment id="101689" author="tappro" created="Tue, 16 Dec 2014 08:36:44 +0000"  >&lt;p&gt;as for insanity test_1, it is not &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2059&quot; title=&quot;mgc to backup configuration on osd-based llogs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2059&quot;&gt;&lt;del&gt;LU-2059&lt;/del&gt;&lt;/a&gt; issue and has different nature, more related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3562&quot; title=&quot;Test failure on test suite insanity, subtest test_1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3562&quot;&gt;&lt;del&gt;LU-3562&lt;/del&gt;&lt;/a&gt;. It worths to open separate ticket for this or reopen &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3562&quot; title=&quot;Test failure on test suite insanity, subtest test_1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3562&quot;&gt;&lt;del&gt;LU-3562&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="17885">LU-2959</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="15949">LU-1931</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="19698">LU-3562</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="17885">LU-2959</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="21508">LU-4121</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="27794">LU-5984</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="19026">LU-3347</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="22929">LU-4565</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="29208">LU-6395</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="17497">LU-2778</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="22169">LU-4274</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="27281">LU-5796</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="11345">LU-2158</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv4s7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4308</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>