<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:11:03 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7686] Interop 2.7.1&lt;-&gt;master - sanity-scrub test_4a: (4) Expected &apos;inconsistent&apos; on mds1, but got &apos;recreated,inconsistent&apos;</title>
                <link>https://jira.whamcloud.com/browse/LU-7686</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Saurabh Tandan &amp;lt;saurabh.tandan@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/39652058-bad8-11e5-87b4-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/39652058-bad8-11e5-87b4-5254006e85c2&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_4a failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;(4) Expected &apos;inconsistent&apos; on mds1, but got &apos;recreated,inconsistent&apos;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;test log:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity-scrub test 4a: Auto trigger OI scrub if bad OI mapping was found (1) == 10:46:38 (1452681998)
preparing... Wed Jan 13 10:46:38 UTC 2016
creating 0 files on mds1
prepared Wed Jan 13 10:46:39 UTC 2016.
CMD: shadow-2vm5.shadow.whamcloud.com,shadow-2vm6 running=\$(grep -c /mnt/lustre&apos; &apos; /proc/mounts);
if [ \$running -ne 0 ] ; then
echo Stopping client \$(hostname) /mnt/lustre opts:;
lsof /mnt/lustre || need_kill=no;
if [ x != x -a x\$need_kill != xno ]; then
    pids=\$(lsof -t /mnt/lustre | sort -u);
    if [ -n \&quot;\$pids\&quot; ]; then
             kill -9 \$pids;
    fi
fi;
while umount  /mnt/lustre 2&amp;gt;&amp;amp;1 | grep -q busy; do
    echo /mnt/lustre is still busy, wait one second &amp;amp;&amp;amp; sleep 1;
done;
fi
stop mds1
CMD: shadow-2vm12 grep -c /mnt/mds1&apos; &apos; /proc/mounts
CMD: shadow-2vm12 umount -d /mnt/mds1
CMD: shadow-2vm12 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
CMD: shadow-2vm12 test -b /dev/lvm-Role_MDS/P1
file-level backup/restore on mds1:/dev/lvm-Role_MDS/P1
CMD: shadow-2vm12 mkdir -p /mnt/brpt
CMD: shadow-2vm12 rm -f /tmp/backup_restore.ea /tmp/backup_restore.tgz
CMD: shadow-2vm12 mount -t ldiskfs /dev/lvm-Role_MDS/P1 /mnt/brpt
backup EA
CMD: shadow-2vm12 cd /mnt/brpt &amp;amp;&amp;amp; getfattr -R -d -m &apos;.*&apos; -P . &amp;gt; /tmp/backup_restore.ea &amp;amp;&amp;amp; cd -
/usr/lib64/lustre/tests
backup data
CMD: shadow-2vm12 umount -d /mnt/brpt
reformat new device
CMD: shadow-2vm12 grep -c /mnt/mds1&apos; &apos; /proc/mounts
CMD: shadow-2vm12 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
CMD: shadow-2vm12 mkfs.lustre --mgs --fsname=lustre --mdt --index=0 --param=sys.timeout=20 --param=lov.stripesize=1048576 --param=lov.stripecount=0 --param=mdt.identity_upcall=/usr/sbin/l_getidentity --backfstype=ldiskfs --device-size=200000 --backfstype ldiskfs --reformat /dev/lvm-Role_MDS/P1
CMD: shadow-2vm12 mount -t ldiskfs /dev/lvm-Role_MDS/P1 /mnt/brpt
restore data
restore EA
CMD: shadow-2vm12 cd /mnt/brpt &amp;amp;&amp;amp; setfattr --restore=/tmp/backup_restore.ea &amp;amp;&amp;amp; cd - 
/usr/lib64/lustre/tests
remove recovery logs
CMD: shadow-2vm12 rm -fv /mnt/brpt/OBJECTS/* /mnt/brpt/CATALOGS
removed `/mnt/brpt/CATALOGS&apos;
CMD: shadow-2vm12 umount -d /mnt/brpt
CMD: shadow-2vm12 rm -f /tmp/backup_restore.ea /tmp/backup_restore.tgz
CMD: shadow-2vm12 e2label /dev/lvm-Role_MDS/P1 lustre-MDT0000
starting MDTs with OI scrub disabled
CMD: shadow-2vm12 mkdir -p /mnt/mds1
CMD: shadow-2vm12 test -b /dev/lvm-Role_MDS/P1
CMD: shadow-2vm12 mkdir -p /mnt/mds1; mount -t lustre -o user_xattr,noscrub  		                   /dev/lvm-Role_MDS/P1 /mnt/mds1
CMD: shadow-2vm12 PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/qt-3.3/bin:/usr/lib64/openmpi/bin:/usr/bin:/bin:/usr/sbin:/sbin::/sbin:/bin:/usr/sbin: NAME=autotest_config sh rpc.sh set_default_debug \&quot;vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck\&quot; \&quot;all -lnet -lnd -pinger\&quot; 4 
CMD: shadow-2vm12 e2label /dev/lvm-Role_MDS/P1 2&amp;gt;/dev/null
CMD: shadow-2vm12 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0000.oi_scrub
 sanity-scrub test_4a: @@@@@@ FAIL: (4) Expected &apos;inconsistent&apos; on mds1, but got &apos;recreated,inconsistent&apos;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>Server: master, build# 3303, RHEL 6.7&lt;br/&gt;
Client: 2.7.1, b2_7_fe/34</environment>
        <key id="34184">LU-7686</key>
            <summary>Interop 2.7.1&lt;-&gt;master - sanity-scrub test_4a: (4) Expected &apos;inconsistent&apos; on mds1, but got &apos;recreated,inconsistent&apos;</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="2">Won&apos;t Fix</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Tue, 19 Jan 2016 22:19:33 +0000</created>
                <updated>Thu, 2 Jun 2016 23:43:03 +0000</updated>
                            <resolved>Thu, 11 Feb 2016 04:07:53 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="141864" author="standan" created="Wed, 10 Feb 2016 22:08:15 +0000"  >&lt;p&gt;Another instance found for interop tag 2.7.66 - EL6.7 Server/2.7.1 Client, build# 3316&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/535a0f2e-cc98-11e5-b80c-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/535a0f2e-cc98-11e5-b80c-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for interop tag 2.7.66 - EL6.7 Server/2.5.5 Client, build# 3316&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/ad6dd9b2-cc9f-11e5-963e-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/ad6dd9b2-cc9f-11e5-963e-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for interop tag 2.7.66 - EL7 Server/2.5.5 Client, build# 3316&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/781e3562-cc46-11e5-901d-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/781e3562-cc46-11e5-901d-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="141945" author="yong.fan" created="Thu, 11 Feb 2016 04:07:53 +0000"  >&lt;p&gt;It is NOT necessary to test sanity-scrub/sanity-lfsck under interoperability mode. See &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7144&quot; title=&quot;Interop 2.7.0&amp;lt;-&amp;gt;master- sanity-scrub test_14: (6) Some entry under /lost+found should be repaired&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7144&quot;&gt;&lt;del&gt;LU-7144&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxykf:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>