<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:20:14 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-1850] replay-vbr test_1b: FAIL: open succeeded unexpectedly</title>
                <link>https://jira.whamcloud.com/browse/LU-1850</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;&lt;a href=&quot;https://maloo.whamcloud.com/test_sessions/a693a99c-f682-11e1-8eb0-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sessions/a693a99c-f682-11e1-8eb0-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://maloo.whamcloud.com/test_logs/04f30584-f686-11e1-8eb0-52540035b04c/show_text&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_logs/04f30584-f686-11e1-8eb0-52540035b04c/show_text&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;HOST:client-29vm6&amp;#93;&lt;/span&gt; &lt;span class=&quot;error&quot;&gt;&amp;#91;old_quota_type:3&amp;#93;&lt;/span&gt; &lt;span class=&quot;error&quot;&gt;&amp;#91;new_quota_type:ug3&amp;#93;&lt;/span&gt;&lt;br/&gt;
open /mnt/lustre failed: Stale NFS file handle (116)&lt;br/&gt;
quotacheck failed: Stale NFS file handle&lt;br/&gt;
 metadata-updates : @@@@@@ FAIL: quotacheck has failed &lt;br/&gt;
 &#160;Trace dump:&lt;br/&gt;
 &#160;= /usr/lib64/lustre/tests/test-framework.sh:3638:error_noexit()&lt;br/&gt;
 &#160;= /usr/lib64/lustre/tests/test-framework.sh:3660:error()&lt;br/&gt;
 &#160;= /usr/lib64/lustre/tests/test-framework.sh:989:quota_save_version()&lt;br/&gt;
 &#160;= /u&lt;/p&gt;</description>
                <environment></environment>
        <key id="15726">LU-1850</key>
            <summary>replay-vbr test_1b: FAIL: open succeeded unexpectedly</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="sarah">Sarah Liu</assignee>
                                    <reporter username="pjones">Peter Jones</reporter>
                        <labels>
                    </labels>
                <created>Thu, 6 Sep 2012 19:38:22 +0000</created>
                <updated>Fri, 12 Aug 2016 05:21:18 +0000</updated>
                            <resolved>Fri, 28 Sep 2012 11:00:56 +0000</resolved>
                                    <version>Lustre 2.3.0</version>
                                    <fixVersion>Lustre 2.3.0</fixVersion>
                    <fixVersion>Lustre 2.4.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="44325" author="pjones" created="Thu, 6 Sep 2012 19:38:56 +0000"  >&lt;p&gt;Niu&lt;/p&gt;

&lt;p&gt;Could you please look at this failure and see if you can work out why this has happened?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="44474" author="niu" created="Mon, 10 Sep 2012 02:27:37 +0000"  >&lt;p&gt;The error message is indicating that the /mnt/lustre isn&apos;t accessiable (got -ESTALE when trying to open it), it&apos;s not related to quota.&lt;/p&gt;

&lt;p&gt;I think it could probably caused by the insane testing environment after replay-vbr failed.&lt;/p&gt;

&lt;p&gt;We can see from the maloo report:&lt;br/&gt;
1. replay-vbr failed;&lt;br/&gt;
2. /mnt/lustre become inaccessiable, tests followed replay-vbr failed for -ESTALE when accessing /mnt/lustre; (from insanity to performance-sanity)&lt;br/&gt;
3. testing environment recovered after node reinitalization; (after performance-sanity)&lt;/p&gt;

&lt;p&gt;I think it should be a test script issue (we&apos;d make sure the testing environment is clean after each failed test?) but not a lustre bug.&lt;/p&gt;</comment>
                            <comment id="44514" author="pjones" created="Mon, 10 Sep 2012 12:50:11 +0000"  >&lt;p&gt;Thanks Niu. Sarah will look at making the necessary test script improvements&lt;/p&gt;</comment>
                            <comment id="44563" author="sarah" created="Tue, 11 Sep 2012 01:04:38 +0000"  >&lt;p&gt;Try unmount and remount client, make sure the system is ready for the following tests.&lt;br/&gt;
Patch for b2_3 is here &lt;a href=&quot;http://review.whamcloud.com/#change,3929&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,3929&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="44652" author="yujian" created="Tue, 11 Sep 2012 23:40:38 +0000"  >&lt;p&gt;The initial issue is that replay-vbr test_1b failed as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: client-29vm6 checkstat -v -a /mnt/lustre/d0.replay-vbr/d1/f.replay-vbr.1b
Can&apos;t lstat /mnt/lustre/d0.replay-vbr/d1/f.replay-vbr.1b: Stale NFS file handle
 replay-vbr test_1b: @@@@@@ FAIL: open succeeded unexpectedly 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Dmesg on Client client-29vm6:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[33466.651998] Lustre: DEBUG MARKER: checkstat -v -a /mnt/lustre/d0.replay-vbr/d1/f.replay-vbr.1b
[33466.659487] LustreError: 3642:0:(file.c:2328:ll_inode_revalidate_fini()) failure -116 inode 524300
[33466.724390] Lustre: DEBUG MARKER: /usr/sbin/lctl mark  replay-vbr test_1b: @@@@@@ FAIL: open succeeded unexpectedly 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Dmesg on MDS client-29vm3:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 8705:0:(llog_lvfs.c:610:llog_lvfs_create()) error looking up logfile 0x60:0x3f5bbfdd: rc -116
LustreError: 8705:0:(llog_cat.c:168:llog_cat_id2handle()) error opening log id 0x60:3f5bbfdd: rc -116
LustreError: 8705:0:(llog_obd.c:314:cat_cancel_cb()) Cannot find handle for log 0x60
LustreError: 8676:0:(llog_obd.c:387:llog_obd_origin_setup()) llog_process() with cat_cancel_cb failed: -116
LustreError: 8676:0:(llog_obd.c:214:llog_setup_named()) obd lustre-OST0003-osc-MDT0000 ctxt 2 lop_setup=ffffffffa03cbdf0 failed -116
LustreError: 8676:0:(osc_request.c:3281:__osc_llog_init()) failed LLOG_MDS_OST_ORIG_CTXT
LustreError: 8676:0:(osc_request.c:3298:__osc_llog_init()) osc &apos;lustre-OST0003-osc-MDT0000&apos; tgt &apos;mdd_obd-lustre-MDT0000&apos; catid ffff880065dd78d0 rc=-116
LustreError: 8676:0:(osc_request.c:3300:__osc_llog_init()) logid 0x58:0x3f5bbfd2
LustreError: 8676:0:(osc_request.c:3328:osc_llog_init()) rc: -116
LustreError: 8676:0:(lov_log.c:243:lov_llog_init()) error osc_llog_init idx 3 osc &apos;lustre-OST0003-osc-MDT0000&apos; tgt &apos;mdd_obd-lustre-MDT0000&apos; (rc=-116)
LustreError: 8676:0:(llog_lvfs.c:610:llog_lvfs_create()) error looking up logfile 0x59:0x3f5bbfd3: rc -116
LustreError: 8676:0:(llog_obd.c:214:llog_setup_named()) obd lustre-OST0004-osc-MDT0000 ctxt 2 lop_setup=ffffffffa03cbdf0 failed -116
LustreError: 8676:0:(osc_request.c:3281:__osc_llog_init()) failed LLOG_MDS_OST_ORIG_CTXT
LustreError: 8676:0:(osc_request.c:3298:__osc_llog_init()) osc &apos;lustre-OST0004-osc-MDT0000&apos; tgt &apos;mdd_obd-lustre-MDT0000&apos; catid ffff880065dd78d0 rc=-116
LustreError: 8676:0:(osc_request.c:3300:__osc_llog_init()) logid 0x59:0x3f5bbfd3
LustreError: 8676:0:(osc_request.c:3328:osc_llog_init()) rc: -116
LustreError: 8676:0:(lov_log.c:243:lov_llog_init()) error osc_llog_init idx 4 osc &apos;lustre-OST0004-osc-MDT0000&apos; tgt &apos;mdd_obd-lustre-MDT0000&apos; (rc=-116)
LustreError: 8706:0:(lov_log.c:155:lov_llog_origin_connect()) error osc_llog_connect tgt 3 (-107)
LustreError: 8706:0:(mds_lov.c:873:__mds_lov_synchronize()) lustre-OST0003_UUID failed at llog_origin_connect: -107
LustreError: 8706:0:(mds_lov.c:903:__mds_lov_synchronize()) lustre-OST0003_UUID sync failed -107, deactivating
LustreError: 8707:0:(lov_log.c:155:lov_llog_origin_connect()) error osc_llog_connect tgt 4 (-107)
LustreError: 8707:0:(mds_lov.c:873:__mds_lov_synchronize()) lustre-OST0004_UUID failed at llog_origin_connect: -107
LustreError: 8707:0:(mds_lov.c:903:__mds_lov_synchronize()) lustre-OST0004_UUID sync failed -107, deactivating
Lustre: DEBUG MARKER: PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/u
Lustre: DEBUG MARKER: e2label /dev/lvm-MDS/P1 2&amp;gt;/dev/null
Lustre: DEBUG MARKER: /usr/sbin/lctl mark  replay-vbr test_1b: @@@@@@ FAIL: open succeeded unexpectedly 
Lustre: DEBUG MARKER: replay-vbr test_1b: @@@@@@ FAIL: open succeeded unexpectedly
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;It was really a Lustre issue. We need fix this Lustre issue first.&lt;/p&gt;</comment>
                            <comment id="44654" author="niu" created="Tue, 11 Sep 2012 23:54:52 +0000"  >&lt;p&gt;don&apos;t we have ticket for the replay-vbr test_1b failure? If not I think we&apos;d open another ticket for it or change the description/summary of this ticket and fix it here.&lt;/p&gt;</comment>
                            <comment id="44656" author="yujian" created="Wed, 12 Sep 2012 00:54:41 +0000"  >&lt;blockquote&gt;&lt;p&gt;don&apos;t we have ticket for the replay-vbr test_1b failure? If not I think we&apos;d open another ticket for it or change the description/summary of this ticket and fix it here.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I just searched for this failure on Jira and Maloo, and found it did not occur in the past. So I changed the summary of this ticket.&lt;/p&gt;

&lt;p&gt;For the test script, after the above Lustre failure occurs, I&apos;m not sure whether trying to mount /mnt/lustre again will succeed or not. Li Wei, could you please review the above patch created by Sarah to see whether that will reduce the affection on the subsequent sub-tests/tests after the Lustre failure occurs? Thanks.&lt;/p&gt;</comment>
                            <comment id="44657" author="liwei" created="Wed, 12 Sep 2012 00:58:08 +0000"  >&lt;p&gt;OK, will do the review in a couple of hours.&lt;/p&gt;</comment>
                            <comment id="44661" author="niu" created="Wed, 12 Sep 2012 01:56:20 +0000"  >&lt;p&gt;I highly suspect the replay-vbr test_1b failure is another instance of LVM/VM flush problem, like &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1012&quot; title=&quot;replay-vbr: test_1b failure.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1012&quot;&gt;&lt;del&gt;LU-1012&lt;/del&gt;&lt;/a&gt;. From the log, looks like the llog object creation isn&apos;t committed before replay-barrier set the fs as read_only.&lt;/p&gt;</comment>
                            <comment id="44937" author="yujian" created="Sat, 15 Sep 2012 02:42:12 +0000"  >&lt;p&gt;Here are the historical Maloo reports for replay-vbr test 1b failures:&lt;br/&gt;
&lt;a href=&quot;http://tinyurl.com/8tybkum&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://tinyurl.com/8tybkum&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As we can see, the failure in this ticket did not occur regularly. And per the analysis from Niu, this is very likely another instance of LVM/VM flush issue. So, we can lower the priority of this ticket.&lt;/p&gt;

&lt;p&gt;We need land the test script patch created by Sarah to reduce the affection on the subsequent sub-tests/tests after the failure occurs.&lt;/p&gt;</comment>
                            <comment id="45300" author="sarah" created="Thu, 20 Sep 2012 17:05:35 +0000"  >&lt;p&gt;patch for master &lt;a href=&quot;http://review.whamcloud.com/#change,4063&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,4063&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="45723" author="jlevi" created="Fri, 28 Sep 2012 11:00:47 +0000"  >&lt;p&gt;Reopening ticket to add Fix Version 2.3&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv5i7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4426</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>