<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:08:40 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7410] After downgrade from 2.8 to 2.5.5, hit unsupported incompat filesystem feature(s) 400</title>
                <link>https://jira.whamcloud.com/browse/LU-7410</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;1. upgrade system from 2.5.5 RHEL6.6 to master RHEL6.7 PASS&lt;br/&gt;
2. downgrade system from master RHEL6.7 to 2.5.5 6.6 FAIL &lt;/p&gt;

&lt;p&gt;mount MDS failed&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: DEBUG MARKER: == upgrade-downgrade End == 15:01:41 (1447110101)
LDISKFS-fs (sdb1): mounted filesystem with ordered data mode. quota=on. Opts: 
Lustre: MGC10.2.4.47@tcp: Connection restored to MGS (at 0@lo)
Lustre: lustre-MDT0000: used disk, loading
LustreError: 12684:0:(mdt_recovery.c:263:mdt_server_data_init()) lustre-MDT0000: unsupported incompat filesystem feature(s) 400
LustreError: 12684:0:(obd_config.c:572:class_setup()) setup lustre-MDT0000 failed (-22)
LustreError: 12684:0:(obd_config.c:1629:class_config_llog_handler()) MGC10.2.4.47@tcp: cfg command failed: rc = -22
Lustre:    cmd=cf003 0:lustre-MDT0000  1:lustre-MDT0000_UUID  2:0  3:lustre-MDT0000-mdtlov  4:f  
LustreError: 15b-f: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos;failed from the MGS (-22).  Make sure this client and the MGS are running compatible versions of Lustre.
LustreError: 15c-8: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos; failed (-22). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
LustreError: 12589:0:(obd_mount_server.c:1254:server_start_targets()) failed to start server lustre-MDT0000: -22
LustreError: 12589:0:(obd_mount_server.c:1737:server_fill_super()) Unable to start targets: -22
LustreError: 12589:0:(obd_mount_server.c:847:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client.
LustreError: 12589:0:(obd_mount_server.c:1422:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2)
LustreError: 12589:0:(obd_config.c:619:class_cleanup()) Device 5 not setup
Lustre: 12589:0:(client.c:1943:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1447110105/real 1447110105]  req@ffff8808352bac00 x1517404919169064/t0(0) o251-&amp;gt;MGC10.2.4.47@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1447110111 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-MDT0000 complete
LustreError: 12589:0:(obd_mount.c:1330:lustre_fill_super()) Unable to mount  (-22)
Lustre: DEBUG MARKER: Using TIMEOUT=100
Lustre: DEBUG MARKER: upgrade-downgrade : @@@@@@ FAIL: NAME=ncli not mounted
LDISKFS-fs (sdb1): mounted filesystem with ordered data mode. quota=on. Opts: 
Lustre: MGC10.2.4.47@tcp: Connection restored to MGS (at 0@lo)
Lustre: lustre-MDT0000: used disk, loading
LustreError: 13112:0:(mdt_recovery.c:263:mdt_server_data_init()) lustre-MDT0000: unsupported incompat filesystem feature(s) 400
LustreError: 13112:0:(obd_config.c:572:class_setup()) setup lustre-MDT0000 failed (-22)
LustreError: 13112:0:(obd_config.c:1629:class_config_llog_handler()) MGC10.2.4.47@tcp: cfg command failed: rc = -22
Lustre:    cmd=cf003 0:lustre-MDT0000  1:lustre-MDT0000_UUID  2:0  3:lustre-MDT0000-mdtlov  4:f  
LustreError: 15b-f: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos;failed from the MGS (-22).  Make sure this client and the MGS are running compatible versions of Lustre.
LustreError: 15c-8: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos; failed (-22). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
LustreError: 13025:0:(obd_mount_server.c:1254:server_start_targets()) failed to start server lustre-MDT0000: -22
LustreError: 13025:0:(obd_mount_server.c:1737:server_fill_super()) Unable to start targets: -22
LustreError: 13025:0:(obd_mount_server.c:847:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client.
LustreError: 13025:0:(obd_mount_server.c:1422:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2)
LustreError: 13025:0:(obd_config.c:619:class_cleanup()) Device 5 not setup
Lustre: 13025:0:(client.c:1943:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1447110256/real 1447110256]  req@ffff88081d67dc00 x1517404919169104/t0(0) o251-&amp;gt;MGC10.2.4.47@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1447110262 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-MDT0000 complete
LustreError: 13025:0:(obd_mount.c:1330:lustre_fill_super()) Unable to mount  (-22)
Lustre: DEBUG MARKER: Using TIMEOUT=100
Lustre: DEBUG MARKER: upgrade-downgrade : @@@@@@ FAIL: NAME=ncli not mounted
[root@onyx-25 ~]# 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>before upgrade: lustre-master #3226 RHEL6.7&lt;br/&gt;
after upgrade: lustre-b2_5_fe #62  RHEL6.6</environment>
        <key id="33072">LU-7410</key>
            <summary>After downgrade from 2.8 to 2.5.5, hit unsupported incompat filesystem feature(s) 400</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="2">Won&apos;t Fix</resolution>
                                        <assignee username="pichong">Gregoire Pichon</assignee>
                                    <reporter username="sarah">Sarah Liu</reporter>
                        <labels>
                    </labels>
                <created>Mon, 9 Nov 2015 23:35:47 +0000</created>
                <updated>Thu, 20 Jul 2017 20:26:16 +0000</updated>
                            <resolved>Wed, 17 Aug 2016 19:20:45 +0000</resolved>
                                                                        <due></due>
                            <votes>0</votes>
                                    <watches>10</watches>
                                                                            <comments>
                            <comment id="133097" author="adilger" created="Tue, 10 Nov 2015 03:07:20 +0000"  >&lt;p&gt;This is caused by &lt;tt&gt;OBD_INCOMPAT_MULTI_RPCS&lt;/tt&gt; being set on the MDS.  It should be cleared if the MDS is unmounted cleanly.&lt;/p&gt;</comment>
                            <comment id="133170" author="sarah" created="Tue, 10 Nov 2015 19:23:40 +0000"  >&lt;p&gt;before downgrade the system, the script did cleanupall to umount&lt;/p&gt;</comment>
                            <comment id="133206" author="sarah" created="Wed, 11 Nov 2015 05:41:08 +0000"  >&lt;p&gt;Hit the same issue when downgrading from master RHEL7 to 2.5.5 RHEL6.6. &lt;/p&gt;

&lt;p&gt;before downgrade, the MDS is unmounted&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[14588.476494] Lustre: DEBUG MARKER: == upgrade-downgrade Start clean downgrade == 20:19:16 (1447215556)
[14588.857361] Lustre: DEBUG MARKER: == upgrade-downgrade Shutdown the entire Lustre filesystem == 20:19:16 (1447215556)
[14592.877840] LustreError: 3346:0:(client.c:1138:ptlrpc_import_delay_req()) @@@ IMP_CLOSED   req@ffff88040e18c800 x1517502900474884/t0(0) o13-&amp;gt;lustre-OST0000-osc-MDT0000@10.2.4.56@tcp:7/4 lens 224/368 e 0 to 0 dl 0 ref 1 fl Rpc:/0/ffffffff rc 0/-1
[14594.489735] Lustre: lustre-MDT0000: Not available for connect from 10.2.4.56@tcp (stopping)
[14602.944947] Lustre: 28852:0:(client.c:2039:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1447215564/real 1447215564]  req@ffff880427623f00 x1517502900474904/t0(0) o251-&amp;gt;MGC10.2.4.47@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1447215570 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
[14603.506132] Lustre: server umount lustre-MDT0000 complete
[14627.485965] Lustre: DEBUG MARKER: == upgrade-downgrade downgrade the Lustre servers all at once == 20:19:46 (1447215586)
[  OK  ] Started Show Plymouth Power Off Screen.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; 
</comment>
                            <comment id="133483" author="pjones" created="Fri, 13 Nov 2015 19:02:07 +0000"  >&lt;p&gt;Gregoire&lt;/p&gt;

&lt;p&gt;Could you please advise on this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="133579" author="pichong" created="Mon, 16 Nov 2015 08:23:23 +0000"  >&lt;p&gt;To have the &lt;tt&gt;OBD_INCOMPAT_MULTI_RPCS&lt;/tt&gt; incompatiblity flag cleaned, the MDT target must have no client connected when it is unmounted.&lt;br/&gt;
Connected clients include both Lustre clients and other MDT targets.&lt;/p&gt;

&lt;p&gt;Therefore, if the file system has only one MDT target, unmounting the Lustre clients first will allow clearing the incompatibility flag at MDT target unmount.&lt;/p&gt;

&lt;p&gt;If the file system has several MDT targets, then it is required to unmount all MDT targets, and then, for each MDT target (one by one), mount with &lt;tt&gt;abort_recovery&lt;/tt&gt; option and unmount the target. This is mentioned in the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5319&quot; title=&quot;Support multiple slots per client in last_rcvd file&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5319&quot;&gt;&lt;del&gt;LU-5319&lt;/del&gt;&lt;/a&gt; &lt;a href=&quot;https://jira.hpdd.intel.com/secure/attachment/18436/MDTReplyReconstructionImprovement.testplan.pdf&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;test plan&lt;/a&gt;, at section &quot;Upgrade / Downgrade&quot;.&lt;/p&gt;

&lt;p&gt;After clearing the incompatiblity flag, the server can be downgraded to a lower Lustre version.&lt;/p&gt;</comment>
                            <comment id="133876" author="sarah" created="Wed, 18 Nov 2015 22:44:39 +0000"  >&lt;p&gt;Hi Gregoire,&lt;/p&gt;

&lt;p&gt;I hit this issue when doing a clean downgrade from master to 2.5.5, which umount all the servers and clients and downgrade them all at once, then tried to mount the system again and failed.&lt;/p&gt;

&lt;p&gt;I will try with 2.7.0 and see how it goes.&lt;/p&gt;</comment>
                            <comment id="133914" author="pichong" created="Thu, 19 Nov 2015 08:25:49 +0000"  >&lt;p&gt;Sarah,&lt;/p&gt;

&lt;p&gt;The issue will be the same with a downgrade to 2.7.0, if you don&apos;t perform the additional operation that clears the incompatibility flag.&lt;/p&gt;

&lt;p&gt;The important point is that to have the &lt;tt&gt;OBD_INCOMPAT_MULTI_RPCS&lt;/tt&gt; incompatibility flag cleared on the MDT servers, it must unmount all the clients, then unmount the servers, &lt;b&gt;and then&lt;/b&gt; additionally operate the &quot;mount with &lt;tt&gt;abort_recovery&lt;/tt&gt; option and unmount of each MDT one by one&quot;. After that the nodes can be downgraded to a lower Lustre version.&lt;/p&gt;</comment>
                            <comment id="133967" author="sarah" created="Thu, 19 Nov 2015 18:17:28 +0000"  >&lt;p&gt;Ah I see  thank you for the clarification!&lt;/p&gt;</comment>
                            <comment id="134468" author="pjones" created="Tue, 24 Nov 2015 23:10:01 +0000"  >&lt;p&gt;If I understand correctly, this is not a bug&lt;/p&gt;</comment>
                            <comment id="142496" author="sarah" created="Wed, 17 Feb 2016 20:54:54 +0000"  >&lt;p&gt;Hello Gregoire,&lt;/p&gt;

&lt;p&gt;I hit the same issue recently, on master/tag-2.7.66 and b2_8/tag-2.7.90.  I did remount the MDS with option &quot;abort_recovery&quot; and umount it again before downgrading, here is what I saw. The same test passed on tag-2.7.64, do you have any idea why this happens? &lt;/p&gt;

&lt;p&gt;on MDS&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-25 ~]# mount -t lustre -o abort_recovery /dev/sdb1 /mnt/mds1
LDISKFS-fs (sdb1): mounted filesystem with ordered data mode. quota=on. Opts: 
Lustre: MGS: Connection restored to MGC10.2.4.47@tcp_0 (at 0@lo)
Lustre: Skipped 4 previous similar messages
LustreError: 45919:0:(mdt_handler.c:5735:mdt_iocontrol()) lustre-MDT0000: Aborting recovery for device
[root@onyx-25 ~]# mount
/dev/sda1 on / type ext3 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw)
nfsd on /proc/fs/nfsd type nfsd (rw)
/dev/sdb1 on /mnt/mds1 type lustre (rw,abort_recovery)
[root@onyx-25 ~]# Lustre: 23885:0:(client.c:2063:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1455732424/real 1455732424]  req@ffff8808074dfcc0 x1526383585120268/t0(0) o8-&amp;gt;lustre-OST0000-osc-MDT0000@10.2.4.56@tcp:28/4 lens 520/544 e 0 to 1 dl 1455732429 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: lustre-MDT0000: Connection restored to MGC10.2.4.47@tcp_0 (at 0@lo)


[root@onyx-25 ~]# umount /mnt/mds1
Lustre: Failing over lustre-MDT0000

Lustre: 46030:0:(client.c:2063:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1455732461/real 1455732461]  req@ffff88080d158cc0 x1526383585120452/t0(0) o251-&amp;gt;MGC10.2.4.47@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1455732467 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-MDT0000 complete
[root@onyx-25 ~]# 
[root@onyx-25 ~]# mount
/dev/sda1 on / type ext3 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw)
nfsd on /proc/fs/nfsd type nfsd (rw)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;dmesg of MDS&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LDISKFS-fs (sdb1): mounted filesystem with ordered data mode. quota=on. Opts: 
Lustre: MGC10.2.4.47@tcp: Connection restored to MGS (at 0@lo)
Lustre: lustre-MDT0000: used disk, loading
LustreError: 10899:0:(mdt_recovery.c:263:mdt_server_data_init()) lustre-MDT0000: unsupported incompat filesystem feature(s) 400
LustreError: 10899:0:(obd_config.c:572:class_setup()) setup lustre-MDT0000 failed (-22)
LustreError: 10899:0:(obd_config.c:1629:class_config_llog_handler()) MGC10.2.4.47@tcp: cfg command failed: rc = -22
Lustre:    cmd=cf003 0:lustre-MDT0000  1:lustre-MDT0000_UUID  2:0  3:lustre-MDT0000-mdtlov  4:f  
LustreError: 15b-f: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos;failed from the MGS (-22).  Make sure this client and the MGS are running compatible versions of Lustre.
LustreError: 15c-8: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos; failed (-22). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
LustreError: 10804:0:(obd_mount_server.c:1254:server_start_targets()) failed to start server lustre-MDT0000: -22
LustreError: 10804:0:(obd_mount_server.c:1737:server_fill_super()) Unable to start targets: -22
LustreError: 10804:0:(obd_mount_server.c:847:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client.
LustreError: 10804:0:(obd_mount_server.c:1422:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2)
LustreError: 10804:0:(obd_config.c:619:class_cleanup()) Device 5 not setup
Lustre: 10804:0:(client.c:1943:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1455737214/real 1455737214]  req@ffff8808181aec00 x1526451090227240/t0(0) o251-&amp;gt;MGC10.2.4.47@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1455737220 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-MDT0000 complete
LustreError: 10804:0:(obd_mount.c:1330:lustre_fill_super()) Unable to mount  (-22)
Lustre: DEBUG MARKER: Using TIMEOUT=100
Lustre: DEBUG MARKER: upgrade-downgrade : @@@@@@ FAIL: NAME=ncli not mounted
Slow work thread pool: Starting up
Slow work thread pool: Ready
FS-Cache: Loaded
NFS: Registering the id_resolver key type
FS-Cache: Netfs &apos;nfs&apos; registered for caching
LDISKFS-fs (sdb1): mounted filesystem with ordered data mode. quota=on. Opts: 
Lustre: MGC10.2.4.47@tcp: Connection restored to MGS (at 0@lo)
Lustre: lustre-MDT0000: used disk, loading
LustreError: 34262:0:(mdt_recovery.c:263:mdt_server_data_init()) lustre-MDT0000: unsupported incompat filesystem feature(s) 400
LustreError: 34262:0:(obd_config.c:572:class_setup()) setup lustre-MDT0000 failed (-22)
LustreError: 34262:0:(obd_config.c:1629:class_config_llog_handler()) MGC10.2.4.47@tcp: cfg command failed: rc = -22
Lustre:    cmd=cf003 0:lustre-MDT0000  1:lustre-MDT0000_UUID  2:0  3:lustre-MDT0000-mdtlov  4:f  
LustreError: 15b-f: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos;failed from the MGS (-22).  Make sure this client and the MGS are running compatible versions of Lustre.
LustreError: 15c-8: MGC10.2.4.47@tcp: The configuration from log &apos;lustre-MDT0000&apos; failed (-22). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
LustreError: 34110:0:(obd_mount_server.c:1254:server_start_targets()) failed to start server lustre-MDT0000: -22
LustreError: 34110:0:(obd_mount_server.c:1737:server_fill_super()) Unable to start targets: -22
LustreError: 34110:0:(obd_mount_server.c:847:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client.
LustreError: 34110:0:(obd_mount_server.c:1422:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2)
LustreError: 34110:0:(obd_config.c:619:class_cleanup()) Device 5 not setup
Lustre: 34110:0:(client.c:1943:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1455737367/real 1455737367]  req@ffff880412fb3800 x1526451090227280/t0(0) o251-&amp;gt;MGC10.2.4.47@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1455737373 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-MDT0000 complete
LustreError: 34110:0:(obd_mount.c:1330:lustre_fill_super()) Unable to mount  (-22)
Lustre: DEBUG MARKER: Using TIMEOUT=100
Lustre: DEBUG MARKER: upgrade-downgrade : @@@@@@ FAIL: NAME=ncli not mounted
[root@onyx-25 ~]# 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="142934" author="pichong" created="Fri, 19 Feb 2016 08:45:52 +0000"  >&lt;p&gt;Could you provide the complete test case that was executed ?&lt;br/&gt;
How is designed the filesystem (nodes hosting the MGT, MDTs, OSTs, client nodes...) ?&lt;br/&gt;
It would be helpful to also provide the full MDS lustre log, including the log before downgrade and log after downgrade.&lt;/p&gt;

&lt;p&gt;thanks.&lt;/p&gt;</comment>
                            <comment id="144242" author="sarah" created="Tue, 1 Mar 2016 05:57:25 +0000"  >&lt;p&gt;the complete case is:&lt;br/&gt;
1. format and setup the system with 1 MDS(1MDT), 1 OSS(1 OST) and 2 clients with lustre 2.5.5 RHEL6.6; create some data &lt;br/&gt;
2. shundown the whole system, umount all nodes&lt;br/&gt;
3. upgrade the whole system to b2_8/build #8; only clear the boot disk, keep data disk untouched&lt;br/&gt;
4. remount the whole system, check the data, works fine;&lt;br/&gt;
5. shudown the whole system again, umount all nodes&lt;br/&gt;
6. do additional step, remounting the MDS with abort_recovery option&lt;br/&gt;
7. umount the MDS again&lt;br/&gt;
8. downgrade all servers and clients to 2.5.5 again without touching the data disk&lt;br/&gt;
9. mount MDS failed as above. &lt;/p&gt;

&lt;p&gt;Please find the attached for more logs. &apos;before means before downgrade; after means after downgrade&apos;&lt;/p&gt;</comment>
                            <comment id="144243" author="sarah" created="Tue, 1 Mar 2016 05:58:05 +0000"  >&lt;p&gt;MDS logs before and after downgrade&lt;/p&gt;

&lt;p&gt;update:  I tried today with b2_8/build #11, manually ran those steps without using script, it doesn&apos;t hit the problem.&lt;/p&gt;</comment>
                            <comment id="144379" author="pichong" created="Wed, 2 Mar 2016 08:01:20 +0000"  >&lt;p&gt;Could you add to the script some calls to the command &quot;lr_reader &amp;lt;mdt-target-device&amp;gt;&quot; at different places (between steps 5-6, 6-7, 7-8 and 8-9 for example) ?&lt;br/&gt;
This could help identify why the incompatibility flag is not cleared.&lt;/p&gt;

&lt;p&gt;The output of the command looks like this:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader  /dev/sdc
last_rcvd:
  uuid: fs3-MDT0000_UUID
  feature_compat: 0x8
  feature_incompat: 0x61c
  feature_rocompat: 0x1
  last_transaction: 30064771072
  target_index: 0
  mount_count: 44
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;The flag &lt;tt&gt;OBD_INCOMPAT_MULTI_RPCS = 0x400&lt;/tt&gt; can be checked within &lt;tt&gt;feature_incompat&lt;/tt&gt; value.&lt;/p&gt;</comment>
                            <comment id="162147" author="yujian" created="Wed, 17 Aug 2016 01:27:50 +0000"  >&lt;p&gt;Hi Gregoire,&lt;/p&gt;

&lt;p&gt;I performed a basic clean upgrading/downing testing from EE 2.4.2.2 (tag 2.5.42.15) to EE 3.0.1.0 (tag 2.7.16.5) with the following steps:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;format and mount EE 2.4.2.2 filesystem with 1 MGS/MDS (1 MDT), 1 OSS (1 OST) and 1 Client
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x21c
Feature rocompat=0x1
Last transaction 4294967296
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;unmount the whole filesystem
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x21c
Feature rocompat=0x1
Last transaction 4294967296
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;upgrade the whole filesystem to EE 3.0.1.0&lt;/li&gt;
	&lt;li&gt;mount the whole filesystem&lt;/li&gt;
	&lt;li&gt;unmount the whole filesystem
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x61c
  feature_rocompat: 0x1
  last_transaction: 8589934592
  target_index: 0
  mount_count: 2
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;remount MDS and OSS with &quot;-o abort_recovery&quot; option
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x61c
  feature_rocompat: 0x1
  last_transaction: 12884901888
  target_index: 0
  mount_count: 3
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;unmount MDS and OSS
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x61c
  feature_rocompat: 0x1
  last_transaction: 12884901888
  target_index: 0
  mount_count: 3
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;downgrade the whole filesystem to EE 2.4.2.2
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x61c
Feature rocompat=0x1
Last transaction 12884901888
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
	&lt;li&gt;mount MDS still failed:
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 24312:0:(mdt_recovery.c:263:mdt_server_data_init()) lustre-MDT0000: unsupported incompat filesystem feature(s) 400
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lr_reader /dev/sdc
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x61c
Feature rocompat=0x1
Last transaction 12884901888
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
&lt;/ol&gt;
</comment>
                            <comment id="162154" author="niu" created="Wed, 17 Aug 2016 06:21:16 +0000"  >&lt;p&gt;Hi, Sarah&lt;/p&gt;

&lt;p&gt;How did you umount the MDT in 7th step?&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;6. &lt;span class=&quot;code-keyword&quot;&gt;do&lt;/span&gt; additional step, remounting the MDS with abort_recovery option
7. umount the MDS again
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;If you didn&apos;t use &quot;umount -f&quot;, could you try with &quot;umount -f&quot; to see if the problem can be reproduced?&lt;/p&gt;</comment>
                            <comment id="162203" author="sarah" created="Wed, 17 Aug 2016 16:05:52 +0000"  >&lt;p&gt;Hi Niu,&lt;/p&gt;

&lt;p&gt;No I didn&apos;t use &quot;-f&quot;, I will try today and get back to you.  thank you!&lt;/p&gt;</comment>
                            <comment id="162248" author="sarah" created="Wed, 17 Aug 2016 17:34:00 +0000"  >&lt;p&gt;Thank you YuJian for the information.&lt;/p&gt;

&lt;p&gt;Niu, I tried with &quot;-f&quot; option(step 7) and it seems work, upgrade from EE2.4.2.2 RHEL6.8 to EE3.0.1 RHEL7.2 and downgrade again:&lt;br/&gt;
MDS&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# lr_reader /dev/sdb1
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x21c
Feature rocompat=0x1
Last transaction 17179869184
target index 0
MDS, index 0
[root@onyx-27 ~]# mount
/dev/sda1 on / type ext3 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw)
nfsd on /proc/fs/nfsd type nfsd (rw)
onyx-4.onyx.hpdd.intel.com:/export/scratch on /scratch type nfs (rw,vers=4,addr=10.2.0.2,clientaddr=10.2.4.65)
/dev/sdb1 on /mnt/mds1 type lustre (rw,acl,user_xattr)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I also record the lc_reader value for all steps for reference:&lt;br/&gt;
1. first time mount system under EE2.4.2.2 RHEL6.8&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# lr_reader /dev/sdb1
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x21c
Feature rocompat=0x1
Last transaction 4294967296
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;2. umount MDS&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# lr_reader /dev/sdb1
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x21c
Feature rocompat=0x1
Last transaction 4294967307
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;3. after upgrade to EE3.0.1 RHEL7 and remount&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# lr_reader /dev/sdb1
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x61c
  feature_rocompat: 0x1
  last_transaction: 8589934592
  target_index: 0
  mount_count: 2
[root@onyx-27 ~]# 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;4. umount again&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# lr_reader /dev/sdb1
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x21c
  feature_rocompat: 0x1
  last_transaction: 8589934594
  target_index: 0
  mount_count: 2
[root@onyx-27 ~]# 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;5. remount with abort_recovery&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# mount -t lustre -o abort_recovery /dev/sdb1 /mnt/mds1
[ 1098.170025] LDISKFS-fs (sdb1): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro,no_mbcache
[ 1098.827347] LustreError: 23424:0:(mdt_handler.c:5840:mdt_iocontrol()) lustre-MDT0000: Aborting recovery for device
[root@onyx-27 ~]# mountg[ 1103.554471] Lustre: 23226:0:(client.c:2029:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1471453302/real 1471453302]  req@ffff8807fd682d00 x1542930245353916/t0(0) o8-&amp;gt;lustre-OST0000-osc-MDT0000@10.2.4.74@tcp:28/4 lens 520/544 e 0 to 1 dl 1471453307 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
[ 1103.595823] Lustre: 23226:0:(client.c:2029:ptlrpc_expire_one_request()) Skipped 1 previous similar message
[root@onyx-27 ~]# lr_reader /dev/sdb1
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x21c
  feature_rocompat: 0x1
  last_transaction: 12884901888
  target_index: 0
  mount_count: 3
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;6. umount with &quot;-f&quot;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# umount -f /mnt/mds1
[ 1321.101538] Lustre: server umount lustre-MDT0000 complete
[root@onyx-27 ~]# lr_reader /dev/sdb1
last_rcvd:
  uuid: lustre-MDT0000_UUID
  feature_compat: 0xc
  feature_incompat: 0x21c
  feature_rocompat: 0x1
  last_transaction: 12884901888
  target_index: 0
  mount_count: 3
[root@onyx-27 ~]# 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;7. downgrade the system to EE2.4.2.2 and mount again&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@onyx-27 ~]# lr_reader /dev/sdb1
Reading last_rcvd
UUID lustre-MDT0000_UUID
Feature compat=0xc
Feature incompat=0x21c
Feature rocompat=0x1
Last transaction 17179869184
target index 0
MDS, index 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="162258" author="pjones" created="Wed, 17 Aug 2016 19:20:46 +0000"  >&lt;p&gt;I think that we can safely close this out from a community release point of view. Upgrade/downgrade from 2.5.x to 2.9 is outside the official scope of the release and there is a viable workaround for those who want to try this anyway.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="25521">LU-5319</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="47379">LU-9788</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="20607" name="debug-after" size="43551" author="sarah" created="Tue, 1 Mar 2016 05:58:05 +0000"/>
                            <attachment id="20609" name="dmesg-after" size="94905" author="sarah" created="Tue, 1 Mar 2016 05:58:05 +0000"/>
                            <attachment id="20610" name="dmesg-before" size="95608" author="sarah" created="Tue, 1 Mar 2016 05:58:05 +0000"/>
                            <attachment id="20608" name="trace-after" size="493739" author="sarah" created="Tue, 1 Mar 2016 05:58:05 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxson:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>