<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:08:49 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7428] conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!</title>
                <link>https://jira.whamcloud.com/browse/LU-7428</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Andreas Dilger &amp;lt;andreas.dilger@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/5d42a610-8187-11e5-a41e-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/5d42a610-8187-11e5-a41e-5254006e85c2&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_84 failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: shadow-10vm4 e2label /dev/lvm-Role_MDS/P1 				2&amp;gt;/dev/null | grep -E &apos;:[a-zA-Z]{3}[0-9]{4}&apos;
CMD: shadow-10vm4 e2label /dev/lvm-Role_MDS/P1 				2&amp;gt;/dev/null | grep -E &apos;:[a-zA-Z]{3}[0-9]{4}&apos;
CMD: shadow-10vm4 e2label /dev/lvm-Role_MDS/P1 				2&amp;gt;/dev/null | grep -E &apos;:[a-zA-Z]{3}[0-9]{4}&apos;
Update not seen after 90s: wanted &apos;&apos; got &apos;lustre:MDT0000&apos;
 conf-sanity test_84: @@@@@@ FAIL: /dev/lvm-Role_MDS/P1 failed to initialize! 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4843:error()
  = /usr/lib64/lustre/tests/test-framework.sh:1270:mount_facet()
  = /usr/lib64/lustre/tests/test-framework.sh:1188:mount_facets()
  = /usr/lib64/lustre/tests/test-framework.sh:2513:facet_failover()
  = /usr/lib64/lustre/tests/conf-sanity.sh:5594:test_84()
  = /usr/lib64/lustre/tests/test-framework.sh:5090:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:5127:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:4992:run_test()
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Please provide additional information about the failure here.&lt;/p&gt;

&lt;p&gt;Info required for matching: conf-sanity 84&lt;br/&gt;
Info required for matching: replay-dual 0a&lt;/p&gt;</description>
                <environment></environment>
        <key id="33136">LU-7428</key>
            <summary>conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="simmonsja">James A Simmons</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>p4hc</label>
                    </labels>
                <created>Sun, 15 Nov 2015 05:28:17 +0000</created>
                <updated>Mon, 29 May 2017 05:58:11 +0000</updated>
                            <resolved>Mon, 26 Sep 2016 15:44:03 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>18</watches>
                                                                            <comments>
                            <comment id="133555" author="adilger" created="Sun, 15 Nov 2015 05:38:39 +0000"  >&lt;p&gt;This is failing about twice per day on master.&lt;/p&gt;</comment>
                            <comment id="134634" author="gerrit" created="Fri, 27 Nov 2015 07:39:37 +0000"  >&lt;p&gt;Andreas Dilger (andreas.dilger@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/17371&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17371&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; tests: write superblock in conf-sanity test_84&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: d3e516e59697f7c48e5cb97054ee04cf97dc7132&lt;/p&gt;</comment>
                            <comment id="134663" author="yong.fan" created="Sat, 28 Nov 2015 04:42:52 +0000"  >&lt;p&gt;It is another failure instance &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7169&quot; title=&quot;conf-sanity 84 restart mds1 failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7169&quot;&gt;&lt;del&gt;LU-7169&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="134667" author="adilger" created="Sat, 28 Nov 2015 06:07:28 +0000"  >&lt;p&gt;Please don&apos;t close this bug, as I&apos;ve got a patch tracked here that may fix the problem.&lt;/p&gt;</comment>
                            <comment id="134769" author="gerrit" created="Mon, 30 Nov 2015 18:48:28 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/17371/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17371/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; tests: write superblock in conf-sanity test_84&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 5fda01f3002e7e742a206ce149652c6b78356828&lt;/p&gt;</comment>
                            <comment id="135186" author="adilger" created="Fri, 4 Dec 2015 03:18:11 +0000"  >&lt;p&gt;It looks like conf-sanity test_84 is still failing in some cases, even with this patch applied:&lt;/p&gt;

&lt;p&gt;It still fails consistently with the new e2fsprogs patches base on master build d059b3c01 &quot;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6693&quot; title=&quot;object_update_param_get inconsistent return value&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6693&quot;&gt;&lt;del&gt;LU-6693&lt;/del&gt;&lt;/a&gt; out: not return NULL in object_update_param_get&quot;, though I can&apos;t really see why except that the superblock label never gets updated:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/17151&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17151&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/17150&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17150&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/17152&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17152&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/17153&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17153&lt;/a&gt;&lt;/p&gt;
</comment>
                            <comment id="135275" author="adilger" created="Fri, 4 Dec 2015 19:41:50 +0000"  >&lt;p&gt;Bob, this test is failing too often (see many different linked bugs), and doesn&apos;t provide much value for testing in comparison.  Could you please add it to the ALWAYS_EXCEPT list until it can be fixed.  My last attempt didn&apos;t seem to resolve the problem.&lt;/p&gt;</comment>
                            <comment id="135279" author="gerrit" created="Fri, 4 Dec 2015 19:53:27 +0000"  >&lt;p&gt;Bob Glossman (bob.glossman@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/17482&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17482&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: disable conf-sanity, test_84&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: a26001d568432d1fa646c6fc850e0ab66e41f97f&lt;/p&gt;</comment>
                            <comment id="135352" author="gerrit" created="Sun, 6 Dec 2015 21:15:23 +0000"  >&lt;p&gt;Andreas Dilger (andreas.dilger@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/17482/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17482/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: disable conf-sanity, test_84&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 74d95a078c6725884f67a2737ea5b7e55fab1087&lt;/p&gt;</comment>
                            <comment id="135412" author="adilger" created="Mon, 7 Dec 2015 18:37:34 +0000"  >&lt;p&gt;It looks like all of the e2fsprogs test failures are on CentOS 7.  I don&apos;t know if that means the failures are &lt;em&gt;only&lt;/em&gt; on CentOS 7, or if it is because the new e2fsprogs patches are not tested on any other distro.&lt;/p&gt;</comment>
                            <comment id="135827" author="hongchao.zhang" created="Thu, 10 Dec 2015 15:35:55 +0000"  >&lt;p&gt;Status update:&lt;br/&gt;
I have tried to reproduce the issue on CentOS7 with the latest e2fsprogs (branch remotes/origin/master-lustre), and I can&apos;t reproduce it.&lt;/p&gt;</comment>
                            <comment id="135956" author="standan" created="Thu, 10 Dec 2015 22:07:27 +0000"  >&lt;p&gt;master, build# 3264, 2.7.64 tag&lt;br/&gt;
recovery-mds-scale test_failover_ost failed with same issue.&lt;br/&gt;
Hard Failover: EL7 Server/Client&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/cf43bf1c-9e9a-11e5-b163-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/cf43bf1c-9e9a-11e5-b163-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hard Failover: EL7 Server/SLES11 SP3 Client&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a39034e8-a077-11e5-8d69-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a39034e8-a077-11e5-8d69-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="136065" author="standan" created="Fri, 11 Dec 2015 18:10:20 +0000"  >&lt;p&gt;master, build# 3264, 2.7.64 tag&lt;br/&gt;
Regression:EL7.1 Server/EL7.1 Client&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/7704baac-9f37-11e5-ba94-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/7704baac-9f37-11e5-ba94-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="136082" author="standan" created="Fri, 11 Dec 2015 18:56:26 +0000"  >&lt;p&gt;master, build# 3264, 2.7.64 tag&lt;br/&gt;
Regression:EL7.1 Server/SLES11 SP3 Client&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/18ce6c0a-9f2b-11e5-bf9b-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/18ce6c0a-9f2b-11e5-bf9b-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="136424" author="pjones" created="Tue, 15 Dec 2015 20:00:00 +0000"  >&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;Have the tests ever run successfully on autotest? Perhaps there is something missing from the TEI environment for running these tests that you can identify so the TEI team can correct?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="136697" author="hongchao.zhang" created="Thu, 17 Dec 2015 14:15:28 +0000"  >&lt;p&gt;Yes, there are successful test on autotest,&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;https://testing.hpdd.intel.com/sub_tests/query?commit=Update+results&amp;amp;page=3&amp;amp;sub_test%5Bquery_bugs%5D=&amp;amp;sub_test%5Bstatus%5D=&amp;amp;sub_test%5Bsub_test_script_id%5D=5e30cae0-7f56-11e4-b7e8-5254006e85c2&amp;amp;test_node%5Barchitecture_type_id%5D=&amp;amp;test_node%5Bdistribution_type_id%5D=0dcf0e82-e30f-11e4-9cb2-5254006e85c2&amp;amp;test_node%5Bfile_system_type_id%5D=&amp;amp;test_node%5Blustre_branch_id%5D=&amp;amp;test_node%5Bos_type_id%5D=&amp;amp;test_node_network%5Bnetwork_type_id%5D=&amp;amp;test_session%5Bend_date%5D=&amp;amp;test_session%5Bquery_recent_period%5D=&amp;amp;test_session%5Bstart_date%5D=&amp;amp;test_session%5Btest_group%5D=&amp;amp;test_session%5Btest_host%5D=&amp;amp;test_session%5Buser_id%5D=&amp;amp;test_set%5Btest_set_script_id%5D=7f66aa20-3db2-11e0-80c0-52540025f9af&amp;amp;utf8=&#10003;&amp;amp;warn%5Bnotice%5D=true
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I&apos;m not sure whether the problem is related to TEI or not.&lt;/p&gt;</comment>
                            <comment id="136945" author="standan" created="Sat, 19 Dec 2015 00:04:58 +0000"  >&lt;p&gt;Another instance for EL7.1 Server/EL7.1 Client - DNE&lt;br/&gt;
Master , Build# 3270&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/6cae7cf8-a26d-11e5-bdef-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/6cae7cf8-a26d-11e5-bdef-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="137309" author="adilger" created="Wed, 23 Dec 2015 19:43:07 +0000"  >&lt;p&gt;Hongchao, it looks like the main problem is that the e2label command run as part of mkfs.lustre to update the filesystem label from &lt;tt&gt;fsname:MDT0000&lt;/tt&gt; to &lt;tt&gt;fsname-MDT0000&lt;/tt&gt; is failing to be written to disk.  It isn&apos;t clear if that is a problem with the way the dev_rdonly code works, that prevents the superblock update from being persistent, or if there is something in the VM that is dropping the writes from the guest.&lt;/p&gt;

&lt;p&gt;What would make sense is to add some debugging to conf-sanity.sh test_84 (since this one seems to hit this failure most often) that runs &lt;tt&gt;e2label $MDSDEV&lt;/tt&gt; to print the filesystem label after the initial mount + sync + sleep but before &lt;tt&gt;replay_barrier&lt;/tt&gt;, and then print it again in &lt;tt&gt;mount_facet&lt;/tt&gt; if the mount has failed so that we can see what the current label is.&lt;/p&gt;</comment>
                            <comment id="137473" author="parinay" created="Mon, 28 Dec 2015 10:30:59 +0000"  >&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;We do see similar failure during relay-dual tests on
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Configuration : 4 node cluster . 1 MDS/ 1 OSS / 2 Clients.
Release
3.10.0_229.20.1.el7_lustremaster_master__81.x86_64_g70bb27b
Server 2.7.64
Client 2.7.642.7.64  
kernel - 3.10.0_229.20.1.el7
git hash - 70bb27b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
&lt;/ul&gt;


&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;stdout.log
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== replay-dual test 0a: expired recovery with lost client ============================================ 05:44:27 (1450935867)
Check file is LU482_FAILED=/tmp/replay-dual.lu482.wNwVYw
Filesystem                 1K-blocks  Used Available Use% Mounted on
192.168.113.21@tcp:/lustre   1345184 35144   1209424   3% /mnt/lustre
total: 50 creates in 0.14 seconds: 355.12 creates/second
fail_loc=0x80000514
Failing mds1 on fre1321
Stopping /mnt/mds1 (opts:) on fre1321
pdsh@fre1323: fre1321: ssh exited with exit code 1
reboot facets: mds1
Failover mds1 to fre1321
05:44:45 (1450935885) waiting for fre1321 network 900 secs ...
05:44:45 (1450935885) network interface is UP
mount facets: mds1
Starting mds1: -o rw,user_xattr  /dev/vdb /mnt/mds1
Waiting 90 secs for update
Waiting 80 secs for update
Waiting 70 secs for update
Waiting 60 secs for update
Waiting 50 secs for update
Waiting 40 secs for update
Waiting 30 secs for update
Waiting 20 secs for update
Waiting 10 secs for update
Update not seen after 90s: wanted &apos;&apos; got &apos;lustre:MDT0000&apos;
 replay-dual test_0a: @@@@@@ FAIL: /dev/vdb failed to initialize! 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4822:error_noexit()
  = /usr/lib64/lustre/tests/test-framework.sh:4853:error()
  = /usr/lib64/lustre/tests/test-framework.sh:1270:mount_facet()
  = /usr/lib64/lustre/tests/test-framework.sh:1188:mount_facets()
  = /usr/lib64/lustre/tests/test-framework.sh:2523:facet_failover()
  = /usr/lib64/lustre/tests/replay-dual.sh:66:test_0a()
  = /usr/lib64/lustre/tests/test-framework.sh:5100:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:5137:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:4954:run_test()
  = /usr/lib64/lustre/tests/replay-dual.sh:76:main()
Dumping lctl log to /tmp/test_logs/1450935862/replay-dual.test_0a.*.1450935987.log
fre1324: Warning: Permanently added &apos;fre1323,192.168.113.23&apos; (ECDSA) to the list of known hosts.

fre1322: Warning: Permanently added &apos;fre1323,192.168.113.23&apos; (ECDSA) to the list of known hosts.

fre1321: Warning: Permanently added &apos;fre1323,192.168.113.23&apos; (ECDSA) to the list of known hosts.

FAIL 0a (121s)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;Its reproducible. please note the kernel version is 3.10.x&lt;/li&gt;
	&lt;li&gt;Attaching the logs, let me know, if more information is required.&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Can somebody confirm, if its the same issue ?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;</comment>
                            <comment id="139391" author="standan" created="Wed, 20 Jan 2016 01:50:00 +0000"  >&lt;p&gt;Another instance found for hardfailover: EL7 Server/Client&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/285d11ca-bc00-11e5-a592-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/285d11ca-bc00-11e5-a592-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="140194" author="gerrit" created="Wed, 27 Jan 2016 13:52:48 +0000"  >&lt;p&gt;Hongchao Zhang (hongchao.zhang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/18178&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18178&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: debug patch&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 4474b31c11f7b9034a25371abbaafe61371e9b85&lt;/p&gt;</comment>
                            <comment id="141048" author="standan" created="Wed, 3 Feb 2016 19:39:03 +0000"  >&lt;p&gt;Encountered same issue for tag 2.7.66 for FULL- EL7.1 Server/EL6.7 Client , master , build# 3314 for replay-vbr.&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/85c15246-ca91-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/85c15246-ca91-11e5-9609-5254006e85c2&lt;/a&gt;&lt;br/&gt;
Encountered another instance for tag 2.7.66 for FUL - EL7.1 Server/EL7.1 Client , master , build# 3314.&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b9fde76c-ca88-11e5-84d3-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b9fde76c-ca88-11e5-84d3-5254006e85c2&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Update not seen after 90s: wanted &apos;&apos; got &apos;lustre:MDT0000&apos;
 replay-vbr test_1b: @@@@@@ FAIL: /dev/lvm-Role_MDS/P1 failed to initialize! 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="141087" author="standan" created="Wed, 3 Feb 2016 23:42:17 +0000"  >&lt;p&gt;Another failure for master : Tag 2.7.66 FULL - EL7.1 Server/SLES11 SP3 Client, build# 3314 for replay-single.&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/9fad2e16-ca7b-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/9fad2e16-ca7b-11e5-9609-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance for FULL - EL7.1 Server/EL7.1 Client - DNE, master, build# 3314&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a973fa52-cac5-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a973fa52-cac5-11e5-9609-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/9ece523c-cac5-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/9ece523c-cac5-11e5-9609-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="141702" author="standan" created="Tue, 9 Feb 2016 23:55:20 +0000"  >&lt;p&gt;Another instance found for hardfailover : EL7 Server/Client, tag 2.7.66, master build 3314&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/8d13249a-ca8f-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/8d13249a-ca8f-11e5-9609-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for hardfailover : EL7 Server/SLES11 SP3 Client, tag 2.7.66, master build 3316&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/2fbf67e4-cd4c-11e5-b1fa-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/2fbf67e4-cd4c-11e5-b1fa-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for Full tag 2.7.66 - EL7.1 Server/EL6.7 Client, build# 3314&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/85c15246-ca91-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/85c15246-ca91-11e5-9609-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for Full tag 2.7.66 - EL7.1 Server/EL7.1 Client, build# 3314&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b9fde76c-ca88-11e5-84d3-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b9fde76c-ca88-11e5-84d3-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for Full tag 2.7.66 -EL7.1 Server/SLES11 SP3 Client, build# 3314&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a63f3418-ca7b-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a63f3418-ca7b-11e5-9609-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/9fad2e16-ca7b-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/9fad2e16-ca7b-11e5-9609-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for Full tag 2.7.66 -EL7.1 Server/EL7.1 Client - DNE, build# 3314&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a973fa52-cac5-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a973fa52-cac5-11e5-9609-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/9ece523c-cac5-11e5-9609-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/9ece523c-cac5-11e5-9609-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="143568" author="standan" created="Wed, 24 Feb 2016 16:55:39 +0000"  >&lt;p&gt;Another instance found on b2_8 for failover testing , build# 6.&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/eb9f29ec-d8da-11e5-83e2-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/eb9f29ec-d8da-11e5-83e2-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/2f0aa9f6-d5a5-11e5-9cc2-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/2f0aa9f6-d5a5-11e5-9cc2-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="145216" author="gerrit" created="Fri, 11 Mar 2016 04:40:01 +0000"  >&lt;p&gt;Hongchao Zhang (hongchao.zhang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/18871&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18871&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: commit the label change to disk&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 9c6eb1c7f48e205168bb9f3a011af8c957b19616&lt;/p&gt;</comment>
                            <comment id="147937" author="gerrit" created="Wed, 6 Apr 2016 01:39:33 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/18871/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18871/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: commit the label change to disk&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 4635d6235f8c8f4bb212aa59710c4f68db6acd7a&lt;/p&gt;</comment>
                            <comment id="148379" author="jgmitter" created="Mon, 11 Apr 2016 13:47:21 +0000"  >&lt;p&gt;Additional fix has landed to master for 2.9.0&lt;/p&gt;</comment>
                            <comment id="151575" author="adilger" created="Mon, 9 May 2016 22:55:22 +0000"  >&lt;p&gt;The patch landed to try the new workaround, but test 84 is still in the ALWAYS_EXCEPT list in conf-sanity.sh, so until that is removed there is no way to know whether this problem was actually fixed.&lt;/p&gt;</comment>
                            <comment id="152218" author="gerrit" created="Fri, 13 May 2016 14:50:42 +0000"  >&lt;p&gt;Hongchao Zhang (hongchao.zhang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/20194&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20194&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: remove test 84 from ALWAYS_EXCEPT&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 589c22a31431c94fcd981c95847d295f841fb1e8&lt;/p&gt;</comment>
                            <comment id="152961" author="adilger" created="Fri, 20 May 2016 07:45:27 +0000"  >&lt;p&gt;It might be worthwhile to test the patch from &lt;a href=&quot;https://github.com/Xyratex/lustre-stable/commit/6197a27f174e683d3c66137db8976bddc7ef179b&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/Xyratex/lustre-stable/commit/6197a27f174e683d3c66137db8976bddc7ef179b&lt;/a&gt; to see if that is fixing the problem?  I think that patch could be simplified to just call &lt;tt&gt;sb-&amp;gt;s_op-&amp;gt;s_freeze()&lt;/tt&gt; before marking the device read-only.&lt;/p&gt;</comment>
                            <comment id="154202" author="gerrit" created="Wed, 1 Jun 2016 02:43:36 +0000"  >&lt;p&gt;Gu Zheng (gzheng@ddn.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/20535&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20535&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; osd: freeze fs before set device readonly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: cb2a762f99cef9ff86dc76445248f91b40f0199b&lt;/p&gt;</comment>
                            <comment id="154408" author="gerrit" created="Thu, 2 Jun 2016 11:01:11 +0000"  >&lt;p&gt;Hongchao Zhang (hongchao.zhang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/20586&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20586&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; osd: set rdonly correctly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 521b8290fbf0b47d4ad03272a206d038f648db2d&lt;/p&gt;</comment>
                            <comment id="154409" author="hongchao.zhang" created="Thu, 2 Jun 2016 11:02:42 +0000"  >&lt;p&gt;the patch ported from MRP-2135 (&lt;a href=&quot;https://github.com/Xyratex/lustre-stable/commit/6197a27f174e683d3c66137db8976bddc7ef179b&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/Xyratex/lustre-stable/commit/6197a27f174e683d3c66137db8976bddc7ef179b&lt;/a&gt;) &lt;br/&gt;
is tracked at &lt;a href=&quot;http://review.whamcloud.com/20586&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20586&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="155004" author="simmonsja" created="Tue, 7 Jun 2016 23:05:21 +0000"  >&lt;p&gt;Does this patch mean we don&apos;t need &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-684&quot; title=&quot;replace dev_rdonly kernel patch with dm-flakey&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-684&quot;&gt;&lt;del&gt;LU-684&lt;/del&gt;&lt;/a&gt; anymore?&lt;/p&gt;</comment>
                            <comment id="155012" author="adilger" created="Tue, 7 Jun 2016 23:28:56 +0000"  >&lt;p&gt;No, this won&apos;t replace &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-684&quot; title=&quot;replace dev_rdonly kernel patch with dm-flakey&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-684&quot;&gt;&lt;del&gt;LU-684&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;This patch is to (hopefully) fix a problem where the device is sync&apos;d and set read-only, but loses some recent writes, for an unknown reason.  This shows up with a variety of different symptoms, and may be a result of bad interactions with LVM and VM virtual block devices, or it may be caused by the dev readonly patches.&lt;/p&gt;</comment>
                            <comment id="155595" author="gerrit" created="Tue, 14 Jun 2016 03:49:18 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/20586/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20586/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; osd: set device read-only correctly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: a079ade7913b923b795ea5c01df4e69bf1a87691&lt;/p&gt;</comment>
                            <comment id="158647" author="pjones" created="Wed, 13 Jul 2016 17:50:30 +0000"  >&lt;p&gt;Landed for 2.9&lt;/p&gt;</comment>
                            <comment id="165481" author="jhammond" created="Fri, 9 Sep 2016 15:13:02 +0000"  >&lt;p&gt;Did the landing of &lt;a href=&quot;http://review.whamcloud.com/20586/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20586/&lt;/a&gt; resolve this issue? I see that we still have 84 in &lt;tt&gt;ALWAYS_EXCEPT&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="165810" author="hongchao.zhang" created="Tue, 13 Sep 2016 09:55:35 +0000"  >&lt;p&gt;the patch &lt;a href=&quot;http://review.whamcloud.com/#/c/20194/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/20194/&lt;/a&gt; to remove test 84 has been refreshed, and have passed the tests in Maloo now.&lt;/p&gt;</comment>
                            <comment id="166110" author="adilger" created="Thu, 15 Sep 2016 07:19:25 +0000"  >&lt;p&gt;Reopen until patch enabling test_84 actually lands.&lt;/p&gt;</comment>
                            <comment id="167245" author="gerrit" created="Mon, 26 Sep 2016 15:19:30 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/20194/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/20194/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7428&quot; title=&quot;conf-sanity test_84, replay-dual 0a: /dev/lvm-Role_MDS/P1 failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7428&quot;&gt;&lt;del&gt;LU-7428&lt;/del&gt;&lt;/a&gt; test: remove test 84 from ALWAYS_EXCEPT&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: e40a3fd8a531ed60528ca82e02ce41918b1be6ba&lt;/p&gt;</comment>
                            <comment id="167265" author="pjones" created="Mon, 26 Sep 2016 15:44:03 +0000"  >&lt;p&gt;Test re-enabled for 2.9&lt;/p&gt;</comment>
                            <comment id="168952" author="yujian" created="Mon, 10 Oct 2016 16:40:21 +0000"  >&lt;p&gt;Hi Hongchao,&lt;/p&gt;

&lt;p&gt;With patch &lt;a href=&quot;http://review.whamcloud.com/7200&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7200&lt;/a&gt; on master branch, conf-sanity test 84 failed as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: onyx-31vm7 e2label /dev/mapper/mds1_flakey 				2&amp;gt;/dev/null | grep -E &apos;:[a-zA-Z]{3}[0-9]{4}&apos;
Update not seen after 90s: wanted &apos;&apos; got &apos;lustre:MDT0000&apos;
 conf-sanity test_84: @@@@@@ FAIL: /dev/mapper/mds1_flakey failed to initialize! 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/e88a61c2-89bf-11e6-a8b7-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/e88a61c2-89bf-11e6-a8b7-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Could you please advise? Thank you.&lt;/p&gt;</comment>
                            <comment id="170265" author="yujian" created="Wed, 19 Oct 2016 04:07:00 +0000"  >&lt;p&gt;Hi Hongchao,&lt;br/&gt;
Since the above failure is specific to dm-flakey patch, I filed &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8729&quot; title=&quot;conf-sanity test_84: FAIL: /dev/mapper/mds1_flakey failed to initialize!&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8729&quot;&gt;&lt;del&gt;LU-8729&lt;/del&gt;&lt;/a&gt; to track it.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="33340">LU-7481</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="32929">LU-7364</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="27929">LU-6029</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="32154">LU-7169</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="31903">LU-7097</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="31448">LU-6992</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="32358">LU-7222</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="32950">LU-7368</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="30892">LU-6789</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="32919">LU-7361</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="33362">LU-7492</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="31909">LU-7100</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="33407">LU-7509</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="20008" name="reaply-dual-all-log.zip" size="2212483" author="parinay" created="Mon, 28 Dec 2015 10:32:14 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxt2f:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>