<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:44:10 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4595] lod_device_free()) ASSERTION( atomic_read(&amp;lu-&gt;ld_ref) == 0 ) failed</title>
                <link>https://jira.whamcloud.com/browse/LU-4595</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Running racer against today&apos;s master (2.5.55-4-gb6a1b94) on a single node with MDSCOUNT=4 and OSTCOUNT=2 I see these LBUGs during umount. &lt;/p&gt;

&lt;p&gt;This loop reproduced the LBUG after 3 iterations:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;cd ~/lustre-release
export MDSCOUNT=4
export MOUNT_2=y
for ((i = 0; i &amp;lt; 10; i++)); do
  echo -e &quot;\n\n\n########### $i $(date) ############\n\n\n&quot;
  llmount.sh
  sh lustre/tests/racer.sh
  umount /mnt/lustre /mnt/lustre2
  umount /mnt/mds{1..4} /mnt/ost{1..2}
  llmountcleanup.sh
done
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: DEBUG MARKER: == racer test complete, duration 314 sec == 10:26:08 (1391703968)
Lustre: Unmounted lustre-client
Lustre: Unmounted lustre-client
Lustre: Failing over lustre-MDT0000
Lustre: server umount lustre-MDT0000 complete
LustreError: 11-0: lustre-MDT0000-lwp-MDT0001: Communicating with 0@lo, operation mds_disconnect failed with -107.
Lustre: Failing over lustre-MDT0001
Lustre: server umount lustre-MDT0001 complete
Lustre: Failing over lustre-MDT0002
LustreError: 3307:0:(lod_dev.c:711:lod_device_free()) ASSERTION( atomic_read(&amp;amp;lu-&amp;gt;ld_ref) == 0 ) failed: 
LustreError: 27074:0:(mdt_handler.c:4256:mdt_fini()) ASSERTION( atomic_read(&amp;amp;d-&amp;gt;ld_ref) == 0 ) failed: 
LustreError: 27074:0:(mdt_handler.c:4256:mdt_fini()) LBUG
Pid: 27074, comm: umount

Call Trace:
 [&amp;lt;ffffffffa0968895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
 [&amp;lt;ffffffffa0968e97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
 [&amp;lt;ffffffffa0b55cdf&amp;gt;] mdt_device_fini+0xd5f/0xda0 [mdt]
 [&amp;lt;ffffffffa0e2dee6&amp;gt;] ? class_disconnect_exports+0x116/0x2f0 [obdclass]
 [&amp;lt;ffffffffa0e532b3&amp;gt;] class_cleanup+0x573/0xd30 [obdclass]
 [&amp;lt;ffffffffa0e2b836&amp;gt;] ? class_name2dev+0x56/0xe0 [obdclass]
 [&amp;lt;ffffffffa0e54fda&amp;gt;] class_process_config+0x156a/0x1ad0 [obdclass]
 [&amp;lt;ffffffffa0e4d2b3&amp;gt;] ? lustre_cfg_new+0x2d3/0x6e0 [obdclass]
 [&amp;lt;ffffffffa0e556b9&amp;gt;] class_manual_cleanup+0x179/0x6f0 [obdclass]
 [&amp;lt;ffffffffa0e2b836&amp;gt;] ? class_name2dev+0x56/0xe0 [obdclass]
 [&amp;lt;ffffffffa0e8ea19&amp;gt;] server_put_super+0x8e9/0xe40 [obdclass]
 [&amp;lt;ffffffff81184c3b&amp;gt;] generic_shutdown_super+0x5b/0xe0
 [&amp;lt;ffffffff81184d26&amp;gt;] kill_anon_super+0x16/0x60
 [&amp;lt;ffffffffa0e57576&amp;gt;] lustre_kill_super+0x36/0x60 [obdclass]
 [&amp;lt;ffffffff811854c7&amp;gt;] deactivate_super+0x57/0x80
 [&amp;lt;ffffffff811a375f&amp;gt;] mntput_no_expire+0xbf/0x110
 [&amp;lt;ffffffff811a41cb&amp;gt;] sys_umount+0x7b/0x3a0
 [&amp;lt;ffffffff8100b072&amp;gt;] system_call_fastpath+0x16/0x1b

Kernel panic - not syncing: LBUG
Pid: 27074, comm: umount Not tainted 2.6.32-358.18.1.el6.lustre.x86_64 #1
Call Trace:
 [&amp;lt;ffffffff8150f018&amp;gt;] ? panic+0xa7/0x16f
 [&amp;lt;ffffffffa0968eeb&amp;gt;] ? lbug_with_loc+0x9b/0xb0 [libcfs]
 [&amp;lt;ffffffffa0b55cdf&amp;gt;] ? mdt_device_fini+0xd5f/0xda0 [mdt]
 [&amp;lt;ffffffffa0e2dee6&amp;gt;] ? class_disconnect_exports+0x116/0x2f0 [obdclass]
 [&amp;lt;ffffffffa0e532b3&amp;gt;] ? class_cleanup+0x573/0xd30 [obdclass]
 [&amp;lt;ffffffffa0e2b836&amp;gt;] ? class_name2dev+0x56/0xe0 [obdclass]
 [&amp;lt;ffffffffa0e54fda&amp;gt;] ? class_process_config+0x156a/0x1ad0 [obdclass]
 [&amp;lt;ffffffffa0e4d2b3&amp;gt;] ? lustre_cfg_new+0x2d3/0x6e0 [obdclass]
 [&amp;lt;ffffffffa0e556b9&amp;gt;] ? class_manual_cleanup+0x179/0x6f0 [obdclass]
 [&amp;lt;ffffffffa0e2b836&amp;gt;] ? class_name2dev+0x56/0xe0 [obdclass]
 [&amp;lt;ffffffffa0e8ea19&amp;gt;] ? server_put_super+0x8e9/0xe40 [obdclass]
 [&amp;lt;ffffffff81184c3b&amp;gt;] ? generic_shutdown_super+0x5b/0xe0
 [&amp;lt;ffffffff81184d26&amp;gt;] ? kill_anon_super+0x16/0x60
 [&amp;lt;ffffffffa0e57576&amp;gt;] ? lustre_kill_super+0x36/0x60 [obdclass]
 [&amp;lt;ffffffff811854c7&amp;gt;] ? deactivate_super+0x57/0x80
 [&amp;lt;ffffffff811a375f&amp;gt;] ? mntput_no_expire+0xbf/0x110
 [&amp;lt;ffffffff811a41cb&amp;gt;] ? sys_umount+0x7b/0x3a0
 [&amp;lt;ffffffff8100b072&amp;gt;] ? system_call_fastpath+0x16/0x1b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; bt
PID: 27074  TASK: ffff880196477500  CPU: 1   COMMAND: &quot;umount&quot;
 #0 [ffff8801beb939b0] machine_kexec at ffffffff81035d6b
 #1 [ffff8801beb93a10] crash_kexec at ffffffff810c0e22
 #2 [ffff8801beb93ae0] panic at ffffffff8150f01f
 #3 [ffff8801beb93b60] lbug_with_loc at ffffffffa0968eeb [libcfs]
 #4 [ffff8801beb93b80] mdt_device_fini at ffffffffa0b55cdf [mdt]
 #5 [ffff8801beb93bf0] class_cleanup at ffffffffa0e532b3 [obdclass]
 #6 [ffff8801beb93c70] class_process_config at ffffffffa0e54fda [obdclass]
 #7 [ffff8801beb93d00] class_manual_cleanup at ffffffffa0e556b9 [obdclass]
 #8 [ffff8801beb93dc0] server_put_super at ffffffffa0e8ea19 [obdclass]
 #9 [ffff8801beb93e30] generic_shutdown_super at ffffffff81184c3b
#10 [ffff8801beb93e50] kill_anon_super at ffffffff81184d26
#11 [ffff8801beb93e70] lustre_kill_super at ffffffffa0e57576 [obdclass]
#12 [ffff8801beb93e90] deactivate_super at ffffffff811854c7
#13 [ffff8801beb93eb0] mntput_no_expire at ffffffff811a375f
#14 [ffff8801beb93ee0] sys_umount at ffffffff811a41cb
#15 [ffff8801beb93f80] system_call_fastpath at ffffffff8100b072
    RIP: 00007ff7634689a7  RSP: 00007fff84d7d120  RFLAGS: 00010202
    RAX: 00000000000000a6  RBX: ffffffff8100b072  RCX: 00007ff763d55009
    RDX: 0000000000000000  RSI: 0000000000000000  RDI: 00007ff765c4cb90
    RBP: 00007ff765c4cb70   R8: 0000000000000000   R9: 0000000000000000
    R10: 0000000000000000  R11: 0000000000000246  R12: 0000000000000000
    R13: 0000000000000000  R14: 0000000000000000  R15: 00007ff765c4cbf0
    ORIG_RAX: 00000000000000a6  CS: 0033  SS: 002b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="23034">LU-4595</key>
            <summary>lod_device_free()) ASSERTION( atomic_read(&amp;lu-&gt;ld_ref) == 0 ) failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="ys">Yang Sheng</assignee>
                                    <reporter username="jhammond">John Hammond</reporter>
                        <labels>
                            <label>lod</label>
                            <label>mdt</label>
                    </labels>
                <created>Thu, 6 Feb 2014 16:37:07 +0000</created>
                <updated>Mon, 12 Sep 2016 17:03:08 +0000</updated>
                            <resolved>Mon, 12 Sep 2016 17:03:08 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="91278" author="ys" created="Mon, 11 Aug 2014 14:40:46 +0000"  >&lt;p&gt;I hit this issue when run 2.6 conf-sanity test-24a on rhel7 kernel.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[11703.385741] LustreError: Skipped 1 previous similar message
[11707.178453] LustreError: 12463:0:(mdt_handler.c:4379:mdt_fini()) ASSERTION( atomic_read(&amp;amp;d-&amp;gt;ld_ref) == 0 ) failed: 
[11707.178821] LustreError: 12463:0:(mdt_handler.c:4379:mdt_fini()) LBUG
[11707.180103] LustreError: 10181:0:(mdd_device.c:1158:mdd_device_free()) ASSERTION( atomic_read(&amp;amp;lu-&amp;gt;ld_ref) == 0 ) failed: 
[11707.180465] LustreError: 10181:0:(mdd_device.c:1158:mdd_device_free()) LBUG
[11707.181712] Kernel panic - not syncing: LBUG
[11707.182035] CPU: 0 PID: 12463 Comm: umount Tainted: GF       W  O--------------   3.10.0-123.el7.x86_64 #1
[11707.182035] Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
[11707.182035]  ffffffffa038ac8d 00000000b55e544c ffff88001dc99b08 ffffffff815e19ba
[11707.182035]  ffff88001dc99b88 ffffffff815db549 ffffffff00000008 ffff88001dc99b98
[11707.182035]  ffff88001dc99b38 00000000b55e544c ffffffffa0d39703 ffff88001f7b6660
[11707.182035] Call Trace:
[11707.182035]  [&amp;lt;ffffffff815e19ba&amp;gt;] dump_stack+0x19/0x1b
[11707.182035]  [&amp;lt;ffffffff815db549&amp;gt;] panic+0xd8/0x1e7
[11707.182035]  [&amp;lt;ffffffffa0365e6b&amp;gt;] lbug_with_loc+0xab/0xc0 [libcfs]
[11707.182035]  [&amp;lt;ffffffffa0ce2221&amp;gt;] mdt_device_fini+0xe61/0xe70 [mdt]
[11707.182035]  [&amp;lt;ffffffffa04d397f&amp;gt;] class_cleanup+0x8ef/0xcc0 [obdclass]
[11707.182035]  [&amp;lt;ffffffffa04d97f8&amp;gt;] class_process_config+0x1898/0x29e0 [obdclass]
[11707.182035]  [&amp;lt;ffffffffa0376047&amp;gt;] ? libcfs_debug_msg+0x57/0x80 [libcfs]
[11707.182035]  [&amp;lt;ffffffffa0370914&amp;gt;] ? libcfs_log_return+0x24/0x30 [libcfs]
[11707.182035]  [&amp;lt;ffffffffa04daa2f&amp;gt;] class_manual_cleanup+0xef/0x6b0 [obdclass]
[11707.182035]  [&amp;lt;ffffffffa0515e6b&amp;gt;] server_put_super+0x86b/0xe30 [obdclass]
[11707.182035]  [&amp;lt;ffffffff811b1fd6&amp;gt;] generic_shutdown_super+0x56/0xe0
[11707.182035]  [&amp;lt;ffffffff811b2242&amp;gt;] kill_anon_super+0x12/0x20
[11707.188420]  [&amp;lt;ffffffffa04ddda2&amp;gt;] lustre_kill_super+0x32/0x50 [obdclass]
[11707.188420]  [&amp;lt;ffffffff811b265d&amp;gt;] deactivate_locked_super+0x3d/0x60
[11707.188420]  [&amp;lt;ffffffff811b26c6&amp;gt;] deactivate_super+0x46/0x60
[11707.188420]  [&amp;lt;ffffffff811cf455&amp;gt;] mntput_no_expire+0xc5/0x120
[11707.188420]  [&amp;lt;ffffffff811d058f&amp;gt;] SyS_umount+0x9f/0x3c0
[11707.188420]  [&amp;lt;ffffffff815f2119&amp;gt;] system_call_fastpath+0x16/0x1b
[11707.188420] Shutting down cpus with NMI
[11707.188420] drm_kms_helper: panic occurred, switching back to text console

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="135503" author="jamesanunez" created="Tue, 8 Dec 2015 15:40:18 +0000"  >&lt;p&gt;Look like we hit this in master:&lt;br/&gt;
2015-12-07 18:10:15 - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/19d4b21c-9d41-11e5-a4d7-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/19d4b21c-9d41-11e5-a4d7-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="165712" author="ys" created="Mon, 12 Sep 2016 17:03:08 +0000"  >&lt;p&gt;I don&apos;t encounter it a long time. Close it first.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="26894">LU-5713</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwefj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>12555</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>