<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:02:11 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6665] Interop 2.7.0&lt;-&gt;master conf-sanity test_80: (import.c:293:ptlrpc_invalidate_import()) ASSERTION( imp-&gt;imp_invalid ) failed</title>
                <link>https://jira.whamcloud.com/browse/LU-6665</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah_lw &amp;lt;wei3.liu@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/6d359bc8-0035-11e5-a922-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/6d359bc8-0035-11e5-a922-5254006e85c2&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_80 failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;test failed to respond and timed out
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;OST console show:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;03:25:51:Lustre: DEBUG MARKER: /usr/sbin/lctl set_param fail_val=10 fail_loc=0x906
03:25:51:LustreError: 11-0: MGC10.1.4.201@tcp: operation obd_ping to node 10.1.4.201@tcp failed: rc = -107
03:25:51:LustreError: Skipped 7 previous similar messages
03:26:22:LustreError: 166-1: MGC10.1.4.201@tcp: Connection to MGS (at 10.1.4.201@tcp) was lost; in progress operations using this service will fail
03:26:22:Lustre: 14127:0:(client.c:1939:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1432178732/real 1432178732]  req@ffff880077ed96c0 x1501746053519508/t0(0) o250-&amp;gt;MGC10.1.4.201@tcp@10.1.4.201@tcp:26/25 lens 400/544 e 0 to 1 dl 1432178738 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
03:26:22:Lustre: 14127:0:(client.c:1939:ptlrpc_expire_one_request()) Skipped 8 previous similar messages
03:26:22:Lustre: Evicted from MGS (at 10.1.4.201@tcp) after server handle changed from 0x7a35d8e1992e29fd to 0x7a35d8e1992e2aac
03:26:22:LustreError: 4602:0:(fail.c:132:__cfs_fail_timeout_set()) cfs_fail_timeout id 906 sleeping for 15000ms
03:26:22:Lustre: DEBUG MARKER: mkdir -p /mnt/ost2
03:26:22:Lustre: DEBUG MARKER: test -b /dev/lvm-Role_OSS/P2
03:26:22:Lustre: DEBUG MARKER: mkdir -p /mnt/ost2; mount -t lustre   		                   /dev/lvm-Role_OSS/P2 /mnt/ost2
03:26:22:LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. quota=on. Opts: 
03:26:22:LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. quota=on. Opts: 
03:26:22:LustreError: 4746:0:(fail.c:132:__cfs_fail_timeout_set()) cfs_fail_timeout id 906 sleeping for 15000ms
03:26:22:LustreError: 4602:0:(fail.c:136:__cfs_fail_timeout_set()) cfs_fail_timeout id 906 awake
03:26:22:Lustre: MGC10.1.4.201@tcp: Connection restored to MGS (at 10.1.4.201@tcp)
03:26:22:LustreError: 4746:0:(fail.c:136:__cfs_fail_timeout_set()) cfs_fail_timeout id 906 awake
03:26:22:LustreError: 4746:0:(import.c:293:ptlrpc_invalidate_import()) ASSERTION( imp-&amp;gt;imp_invalid ) failed: 
03:26:22:LustreError: 4746:0:(import.c:293:ptlrpc_invalidate_import()) LBUG
03:26:22:Pid: 4746, comm: mount.lustre
03:26:22:
03:26:22:Call Trace:
03:26:22: [&amp;lt;ffffffffa0820895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
03:26:22: [&amp;lt;ffffffffa0820e97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
03:26:22: [&amp;lt;ffffffffa0c3f06d&amp;gt;] ptlrpc_invalidate_import+0x85d/0x930 [ptlrpc]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa0c440f6&amp;gt;] ? ptlrpc_set_import_discon+0xf6/0x5b0 [ptlrpc]
03:26:22: [&amp;lt;ffffffffa0c445e3&amp;gt;] ptlrpc_reconnect_import+0x33/0x1b0 [ptlrpc]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa12ea2ea&amp;gt;] mgc_set_info_async+0x5ea/0x1940 [mgc]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa0a006d1&amp;gt;] obd_set_info_async.clone.2+0xf1/0x360 [obdclass]
03:26:22: [&amp;lt;ffffffffa0a06c18&amp;gt;] lustre_start_mgc+0x14c8/0x1e00 [obdclass]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa0a356f2&amp;gt;] server_fill_super+0x5c2/0x1690 [obdclass]
03:26:22: [&amp;lt;ffffffffa082b818&amp;gt;] ? libcfs_log_return+0x28/0x40 [libcfs]
03:26:22: [&amp;lt;ffffffffa0a07ab0&amp;gt;] lustre_fill_super+0x560/0xa80 [obdclass]
03:26:22: [&amp;lt;ffffffffa0a07550&amp;gt;] ? lustre_fill_super+0x0/0xa80 [obdclass]
03:26:22: [&amp;lt;ffffffff811917af&amp;gt;] get_sb_nodev+0x5f/0xa0
03:26:22: [&amp;lt;ffffffffa09feb05&amp;gt;] lustre_get_sb+0x25/0x30 [obdclass]
03:26:22: [&amp;lt;ffffffff81190deb&amp;gt;] vfs_kern_mount+0x7b/0x1b0
03:26:22: [&amp;lt;ffffffff81190f92&amp;gt;] do_kern_mount+0x52/0x130
03:26:22: [&amp;lt;ffffffff811b2b9b&amp;gt;] do_mount+0x2fb/0x930
03:26:22: [&amp;lt;ffffffff811b3260&amp;gt;] sys_mount+0x90/0xe0
03:26:22: [&amp;lt;ffffffff8100b072&amp;gt;] system_call_fastpath+0x16/0x1b
03:26:22:
03:26:22:Kernel panic - not syncing: LBUG
03:26:22:Pid: 4746, comm: mount.lustre Not tainted 2.6.32-504.8.1.el6_lustre.x86_64 #1
03:26:22:Call Trace:
03:26:22: [&amp;lt;ffffffff81529b76&amp;gt;] ? panic+0xa7/0x16f
03:26:22: [&amp;lt;ffffffffa0820eeb&amp;gt;] ? lbug_with_loc+0x9b/0xb0 [libcfs]
03:26:22: [&amp;lt;ffffffffa0c3f06d&amp;gt;] ? ptlrpc_invalidate_import+0x85d/0x930 [ptlrpc]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa0c440f6&amp;gt;] ? ptlrpc_set_import_discon+0xf6/0x5b0 [ptlrpc]
03:26:22: [&amp;lt;ffffffffa0c445e3&amp;gt;] ? ptlrpc_reconnect_import+0x33/0x1b0 [ptlrpc]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa12ea2ea&amp;gt;] ? mgc_set_info_async+0x5ea/0x1940 [mgc]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa0a006d1&amp;gt;] ? obd_set_info_async.clone.2+0xf1/0x360 [obdclass]
03:26:22: [&amp;lt;ffffffffa0a06c18&amp;gt;] ? lustre_start_mgc+0x14c8/0x1e00 [obdclass]
03:26:22: [&amp;lt;ffffffffa08311c1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
03:26:22: [&amp;lt;ffffffffa0a356f2&amp;gt;] ? server_fill_super+0x5c2/0x1690 [obdclass]
03:26:22: [&amp;lt;ffffffffa082b818&amp;gt;] ? libcfs_log_return+0x28/0x40 [libcfs]
03:26:22: [&amp;lt;ffffffffa0a07ab0&amp;gt;] ? lustre_fill_super+0x560/0xa80 [obdclass]
03:26:22: [&amp;lt;ffffffffa0a07550&amp;gt;] ? lustre_fill_super+0x0/0xa80 [obdclass]
03:26:22: [&amp;lt;ffffffff811917af&amp;gt;] ? get_sb_nodev+0x5f/0xa0
03:26:22: [&amp;lt;ffffffffa09feb05&amp;gt;] ? lustre_get_sb+0x25/0x30 [obdclass]
03:26:22: [&amp;lt;ffffffff81190deb&amp;gt;] ? vfs_kern_mount+0x7b/0x1b0
03:26:22: [&amp;lt;ffffffff81190f92&amp;gt;] ? do_kern_mount+0x52/0x130
03:26:22: [&amp;lt;ffffffff811b2b9b&amp;gt;] ? do_mount+0x2fb/0x930
03:26:22: [&amp;lt;ffffffff811b3260&amp;gt;] ? sys_mount+0x90/0xe0
03:26:22: [&amp;lt;ffffffff8100b072&amp;gt;] ? system_call_fastpath+0x16/0x1b
03:26:22:Initializing cgroup subsys cpuset
03:26:22:Initializing cgroup subsys cpu
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>server: 2.7.0&lt;br/&gt;
client: lustre-master #3029</environment>
        <key id="30433">LU-6665</key>
            <summary>Interop 2.7.0&lt;-&gt;master conf-sanity test_80: (import.c:293:ptlrpc_invalidate_import()) ASSERTION( imp-&gt;imp_invalid ) failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="tappro">Mikhail Pershin</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Fri, 29 May 2015 22:37:05 +0000</created>
                <updated>Sun, 16 Jan 2022 08:19:01 +0000</updated>
                            <resolved>Sun, 16 Jan 2022 08:19:01 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="117071" author="adilger" created="Mon, 1 Jun 2015 17:22:41 +0000"  >&lt;p&gt;Sarah, is this a repeatable failure or only intermittent?&lt;/p&gt;</comment>
                            <comment id="120674" author="sarah" created="Wed, 8 Jul 2015 05:54:27 +0000"  >&lt;p&gt;Hi Andreas, &lt;/p&gt;

&lt;p&gt;this is a repeatable issue:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b07b41e2-1211-11e5-a1d3-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b07b41e2-1211-11e5-a1d3-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/e36e29c2-250b-11e5-8009-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/e36e29c2-250b-11e5-8009-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="123736" author="paf" created="Mon, 10 Aug 2015 15:58:34 +0000"  >&lt;p&gt;This sure looks like &lt;a href=&quot;https://jira.hpdd.intel.com/browse/LU-4913&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://jira.hpdd.intel.com/browse/LU-4913&lt;/a&gt;.  It&apos;s being reproduced by the test added for that issue.&lt;/p&gt;

&lt;p&gt;It seems the race there is not completely closed.  (And I suspect this isn&apos;t related to interop.) Cray has seen this in our testing of 2.5 with the patch from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4913&quot; title=&quot;mgc import reconnect race&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4913&quot;&gt;&lt;del&gt;LU-4913&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="141857" author="standan" created="Wed, 10 Feb 2016 21:59:36 +0000"  >&lt;p&gt;Another instance found for interop tag 2.7.66 - 2.7.1 Server/EL7 Client, build# 3316&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/7f66f230-ccde-11e5-8b0e-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/7f66f230-ccde-11e5-8b0e-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another instance found for interop tag 2.7.66 - 2.7.1 Server/EL6.7 Client, build# 3316&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/ddac30e0-ccdd-11e5-b80c-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/ddac30e0-ccdd-11e5-b80c-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="143597" author="standan" created="Wed, 24 Feb 2016 17:21:59 +0000"  >&lt;p&gt;Another instance found for interop - 2.7.1 Server/EL6.7 Client, tag 2.7.90. &lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/f371534e-d573-11e5-bc47-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/f371534e-d573-11e5-bc47-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="322869" author="tappro" created="Sun, 16 Jan 2022 08:19:01 +0000"  >&lt;p&gt;outdated&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxejj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>