<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:41:42 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4325] Config failover between 2 Lustre serves, simulating one server crashed, the other server crashed unexpected when it take the task of the crashed one</title>
                <link>https://jira.whamcloud.com/browse/LU-4325</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;1.mount 1 MDT and 4 OSTs on the Lustre Server1.&lt;br/&gt;
2.mount 4 OSTs on the Lustre Server2.&lt;br/&gt;
3.Config Lustre failover between the 2 Lustre Servers.&lt;br/&gt;
4.mount the Lustre File System on the Lustre Client.&lt;br/&gt;
5.Write and Read datas on the Lustre Client.&lt;br/&gt;
6.Simulating the Lustre Server1 crashed.&lt;br/&gt;
7.The Lustre Server2 crashed unexpectedly when it take the task of the Lustre Server1 ,the call trace info as follow:&lt;br/&gt;
LustreError: 137-5: lustre-OST0000_UUID: not available for connect from 192.168.22.202@tcp (no target) &lt;br/&gt;
LustreError: Skipped 3 previous similar messages &lt;br/&gt;
LDISKFS-fs (sde): recovery complete &lt;br/&gt;
LDISKFS-fs (sde): mounted filesystem with ordered data mode. quota=on. Opts: &lt;br/&gt;
LustreError: 10026:0:(genops.c:320:class_newdev()) Device MGC192.168.22.50@tcp already exists at 2, won&apos;t add &lt;br/&gt;
LustreError: 10026:0:(obd_config.c:374:class_attach()) Cannot create device MGC192.168.22.50@tcp of type mgc : -17 &lt;br/&gt;
LustreError: 10026:0:(obd_mount.c:196:lustre_start_simple()) MGC192.168.22.50@tcp attach error -17 &lt;br/&gt;
LustreError: 10026:0:(obd_mount_server.c:844:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client. &lt;br/&gt;
LustreError: 10026:0:(obd_mount_server.c:1426:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2) &lt;br/&gt;
LustreError: 10026:0:(obd_mount_server.c:1456:server_put_super()) no obd lustre-MDT0000 &lt;br/&gt;
LustreError: 10026:0:(obd_mount_server.c:135:server_deregister_mount()) lustre-MDT0000 not registered &lt;br/&gt;
LustreError: 10026:0:(genops.c:1570:obd_exports_barrier()) ASSERTION( list_empty(&amp;amp;obd-&amp;gt;obd_exports) ) failed: &lt;br/&gt;
LustreError: 10026:0:(genops.c:1570:obd_exports_barrier()) LBUG &lt;br/&gt;
Pid: 10026, comm: mount.lustre &lt;/p&gt;

&lt;p&gt;Call Trace: &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa070f8a5&amp;gt;&amp;#93;&lt;/span&gt; libcfs_debug_dumpstack+0x55/0x80 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa070feb7&amp;gt;&amp;#93;&lt;/span&gt; lbug_with_loc+0x47/0xb0 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0818d91&amp;gt;&amp;#93;&lt;/span&gt; obd_exports_barrier+0x181/0x190 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0f23886&amp;gt;&amp;#93;&lt;/span&gt; mgs_device_fini+0xf6/0x5c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;mgs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0843837&amp;gt;&amp;#93;&lt;/span&gt; class_cleanup+0x817/0xe00 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa081ce2c&amp;gt;&amp;#93;&lt;/span&gt; ? class_name2dev+0x7c/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0847e9b&amp;gt;&amp;#93;&lt;/span&gt; class_process_config+0x1b6b/0x2f60 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0710b90&amp;gt;&amp;#93;&lt;/span&gt; ? cfs_alloc+0x30/0x60 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0849723&amp;gt;&amp;#93;&lt;/span&gt; class_manual_cleanup+0x493/0xe80 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8147a1fe&amp;gt;&amp;#93;&lt;/span&gt; ? _read_unlock+0xe/0x10 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa081ce2c&amp;gt;&amp;#93;&lt;/span&gt; ? class_name2dev+0x7c/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0884b9d&amp;gt;&amp;#93;&lt;/span&gt; server_put_super+0x42d/0x2580 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0887440&amp;gt;&amp;#93;&lt;/span&gt; server_fill_super+0x750/0x1580 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0854c98&amp;gt;&amp;#93;&lt;/span&gt; lustre_fill_super+0x1d8/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0854ac0&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_fill_super+0x0/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8114d21f&amp;gt;&amp;#93;&lt;/span&gt; get_sb_nodev+0x5f/0xa0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa084c3f5&amp;gt;&amp;#93;&lt;/span&gt; lustre_get_sb+0x25/0x30 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8114c74b&amp;gt;&amp;#93;&lt;/span&gt; vfs_kern_mount+0x7b/0x1b0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8114c8f2&amp;gt;&amp;#93;&lt;/span&gt; do_kern_mount+0x52/0x130 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81168912&amp;gt;&amp;#93;&lt;/span&gt; do_mount+0x2d2/0x8c0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81168f90&amp;gt;&amp;#93;&lt;/span&gt; sys_mount+0x90/0xe0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81002f5b&amp;gt;&amp;#93;&lt;/span&gt; system_call_fastpath+0x16/0x1b &lt;/p&gt;


&lt;p&gt;Message fromKernel panic - not syncing: LBUG &lt;br/&gt;
Pid: 10026, comm: mount.lustre Tainted: PF --------------- 2.6.32-358.6.2.l2.08 #2 &lt;br/&gt;
Call Trace: &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81476fa7&amp;gt;&amp;#93;&lt;/span&gt; ? panic+0xa1/0x163 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa070ff0b&amp;gt;&amp;#93;&lt;/span&gt; ? lbug_with_loc+0x9b/0xb0 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0818d91&amp;gt;&amp;#93;&lt;/span&gt; ? obd_exports_barrier+0x181/0x190 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0f23886&amp;gt;&amp;#93;&lt;/span&gt; ? mgs_device_fini+0xf6/0x5c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;mgs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0843837&amp;gt;&amp;#93;&lt;/span&gt; ? class_cleanup+0x817/0xe00 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa081ce2c&amp;gt;&amp;#93;&lt;/span&gt; ? class_name2dev+0x7c/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0847e9b&amp;gt;&amp;#93;&lt;/span&gt; ? class_process_config+0x1b6b/0x2f60 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 syslogd@50:B3:4 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0710b90&amp;gt;&amp;#93;&lt;/span&gt; ? cfs_alloc+0x30/0x60 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0849723&amp;gt;&amp;#93;&lt;/span&gt; ? class_manual_cleanup+0x493/0xe80 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
2:00:01:01 at Se &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8147a1fe&amp;gt;&amp;#93;&lt;/span&gt; ? _read_unlock+0xe/0x10 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa081ce2c&amp;gt;&amp;#93;&lt;/span&gt; ? class_name2dev+0x7c/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0884b9d&amp;gt;&amp;#93;&lt;/span&gt; ? server_put_super+0x42d/0x2580 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0887440&amp;gt;&amp;#93;&lt;/span&gt; ? server_fill_super+0x750/0x1580 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
p 22 12:53:12 .. &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0854c98&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_fill_super+0x1d8/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0854ac0&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_fill_super+0x0/0x530 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8114d21f&amp;gt;&amp;#93;&lt;/span&gt; ? get_sb_nodev+0x5f/0xa0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa084c3f5&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_get_sb+0x25/0x30 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt; &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8114c74b&amp;gt;&amp;#93;&lt;/span&gt; ? vfs_kern_mount+0x7b/0x1b0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8114c8f2&amp;gt;&amp;#93;&lt;/span&gt; ? do_kern_mount+0x52/0x130 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81168912&amp;gt;&amp;#93;&lt;/span&gt; ? do_mount+0x2d2/0x8c0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81168f90&amp;gt;&amp;#93;&lt;/span&gt; ? sys_mount+0x90/0xe0 &lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81002f5b&amp;gt;&amp;#93;&lt;/span&gt; ? system_call_fastpath+0x16/0x1b &lt;br/&gt;
*******&lt;b&gt;show para for nt_memcpy16&lt;/b&gt;******* &lt;br/&gt;
src: ffff880285fc4f00, dst: ffffc90112030e70, len: 56 &lt;br/&gt;
*******&lt;b&gt;show para for panic done&lt;/b&gt;******* &lt;br/&gt;
ODSP:MSG:BUGON: This stack is bug. &lt;br/&gt;
ODSP:MSG:BUGON: Local was taken over by peer. Suspend CPU. &lt;br/&gt;
ODSP:MSG:BUGON: Local was taken over by peer. Suspend CPU. &lt;/p&gt;</description>
                <environment>2 Lustre server +1 client server</environment>
        <key id="22256">LU-4325</key>
            <summary>Config failover between 2 Lustre serves, simulating one server crashed, the other server crashed unexpected when it take the task of the crashed one</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="yueyuling">yueyuling</reporter>
                        <labels>
                    </labels>
                <created>Thu, 28 Nov 2013 08:25:21 +0000</created>
                <updated>Sat, 9 Oct 2021 05:50:34 +0000</updated>
                            <resolved>Sat, 9 Oct 2021 05:50:34 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="73147" author="green" created="Mon, 9 Dec 2013 23:22:48 +0000"  >&lt;p&gt;How is this ticked different from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4190&quot; title=&quot;LustreError: 18166:0:(genops.c:1570:obd_exports_barrier()) ASSERTION( list_empty(&amp;amp;obd-&amp;gt;obd_exports) ) failed: &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4190&quot;&gt;&lt;del&gt;LU-4190&lt;/del&gt;&lt;/a&gt; that you filed on Oct 30th and that has exactly the same stacktrace?&lt;/p&gt;</comment>
                            <comment id="73164" author="yueyuling" created="Tue, 10 Dec 2013 01:33:07 +0000"  >&lt;p&gt;Thank you for your attention to the two problems.&lt;br/&gt;
You are right, the two problems have the same stacktrace. Because I pay more attention to the phenomenon of the problem, not the stacktrace. I think the phenomenon of the two problems are different , so the two problems are different.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                    <customfield id="customfield_10030" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic/Theme</customfieldname>
                        <customfieldvalues>
                                        <label>test</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwa6v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>11824</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>