<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:24:53 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2400] BUG: unable to handle kernel paging request in lprocfs_srch+0x48/0x80</title>
                <link>https://jira.whamcloud.com/browse/LU-2400</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;I was running the &lt;tt&gt;replay-single&lt;/tt&gt; test in a loop and eventually hit this:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;
BUG: unable to handle kernel paging request at ffffffffffffffde
IP: [&amp;lt;ffffffffa12d4f28&amp;gt;] lprocfs_srch+0x48/0x80 [obdclass]
PGD 1a87067 PUD 1a88067 PMD 0 
Oops: 0000 [#1] SMP 
last sysfs file: /sys/module/ptlrpc/initstate
CPU 0 
Modules linked in: lustre(U) ofd(U) osp(U) lod(U) ost(U) mdt(U) mdd(U) mgs(U) osd_zfs(U) lquota(U) obdecho(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) libcfs(U) jbd sha512_generic sha256_generic ebtable_nat ebtables ipt_MASQUERADE iptable_nat nf_nat nf_conntrack_ipv4 nf_defrag_ipv4 ipt_REJECT xt_CHECKSUM iptable_mangle iptable_filter ip_tables bridge stp llc fuse autofs4 nfs lockd fscache nfs_acl auth_rpcgss sunrpc ip6t_REJECT nf_conntrack_ipv6 nf_defrag_ipv6 xt_state nf_conntrack ip6table_filter ip6_tables ipv6 zfs(P)(U) zcommon(P)(U) znvpair(P)(U) zavl(P)(U) zunicode(P)(U) spl(U) zlib_deflate vhost_net macvtap macvlan tun kvm virtio_balloon snd_hda_intel snd_hda_codec snd_hwdep snd_seq snd_seq_device snd_pcm snd_timer snd soundcore snd_page_alloc virtio_net i2c_piix4 i2c_core ext4 mbcache jbd2 virtio_blk pata_acpi ata_generic ata_piix virtio_pci virtio_ring virtio dm_mirror dm_region_hash dm_log dm_mod [last unloaded: libcfs]

Pid: 2886, comm: mount.lustre Tainted: P           ---------------    2.6.32-279.9.1.1chaos.ch5.1.x86_64 #1 Bochs Bochs
RIP: 0010:[&amp;lt;ffffffffa12d4f28&amp;gt;]  [&amp;lt;ffffffffa12d4f28&amp;gt;] lprocfs_srch+0x48/0x80 [obdclass]
RSP: 0018:ffff88001cc599a8  EFLAGS: 00010286
RAX: ffffffffa135ccc0 RBX: 0000000000000000 RCX: 0000000000000000
RDX: 0000000000000000 RSI: ffff88001cc59ab8 RDI: ffffffffa135ccc0
RBP: ffff88001cc599c8 R08: 00000000fffffffe R09: 0000000000000000
R10: 000000000000000f R11: 000000000000000f R12: ffff88001cc59ab8
R13: ffffffffffffff8e R14: 0000000000000000 R15: ffff88001cc59ab8
FS:  00007fa9dee12700(0000) GS:ffff880002200000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: ffffffffffffffde CR3: 000000001bf30000 CR4: 00000000000006f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process mount.lustre (pid: 2886, threadinfo ffff88001cc58000, task ffff880002053540)
Stack:
 ffffffffffffff8e ffff88001cc59ab8 ffffffffffffff8e 0000000000000000
&amp;lt;d&amp;gt; ffff88001cc59a18 ffffffffa12d7084 ffff88001cc599e8 ffff8800195f6000
&amp;lt;d&amp;gt; ffff88001cc59a88 ffffffff8127d0a6 ffff88001c874000 0000000000000006
Call Trace:
 [&amp;lt;ffffffffa12d7084&amp;gt;] lprocfs_register+0x34/0x100 [obdclass]
 [&amp;lt;ffffffff8127d0a6&amp;gt;] ? vsnprintf+0x2b6/0x5f0
 [&amp;lt;ffffffffa109e3ce&amp;gt;] lprocfs_register_mountpoint+0x12e/0xb80 [lustre]
 [&amp;lt;ffffffffa108f226&amp;gt;] client_common_fill_super+0x1a6/0x4e70 [lustre]
 [&amp;lt;ffffffff8127d420&amp;gt;] ? sprintf+0x40/0x50
 [&amp;lt;ffffffffa1094c64&amp;gt;] ll_fill_super+0xd74/0x1500 [lustre]
 [&amp;lt;ffffffffa130fb6c&amp;gt;] lustre_fill_super+0x12c/0x1ae0 [obdclass]
 [&amp;lt;ffffffff8117cb16&amp;gt;] ? set_anon_super+0x56/0x100
 [&amp;lt;ffffffff8116a03c&amp;gt;] ? pcpu_alloc+0x3ac/0xa50
 [&amp;lt;ffffffff8127a24a&amp;gt;] ? strlcpy+0x4a/0x60
 [&amp;lt;ffffffff8117d9e3&amp;gt;] ? sget+0x3e3/0x480
 [&amp;lt;ffffffff8117cac0&amp;gt;] ? set_anon_super+0x0/0x100
 [&amp;lt;ffffffffa130fa40&amp;gt;] ? lustre_fill_super+0x0/0x1ae0 [obdclass]
 [&amp;lt;ffffffff8117df2f&amp;gt;] get_sb_nodev+0x5f/0xa0
 [&amp;lt;ffffffffa12fb185&amp;gt;] lustre_get_sb+0x25/0x30 [obdclass]
 [&amp;lt;ffffffff8117db8b&amp;gt;] vfs_kern_mount+0x7b/0x1b0
 [&amp;lt;ffffffff8117dd32&amp;gt;] do_kern_mount+0x52/0x130
 [&amp;lt;ffffffff8119c442&amp;gt;] do_mount+0x2d2/0x8d0
 [&amp;lt;ffffffff8119cad0&amp;gt;] sys_mount+0x90/0xe0
 [&amp;lt;ffffffff8100b0f2&amp;gt;] system_call_fastpath+0x16/0x1b
Code: a1 49 89 f4 e8 ca b0 22 e0 4d 85 ed 75 1d 48 c7 c7 c0 cc 35 a1 e8 69 2b dc df 48 83 c4 08 48 89 d8 5b 41 5c 41 5d c9 c3 0f 1f 00 &amp;lt;49&amp;gt; 8b 5d 50 48 85 db 75 10 eb d8 0f 1f 44 00 00 48 8b 5b 40 48 
RIP  [&amp;lt;ffffffffa12d4f28&amp;gt;] lprocfs_srch+0x48/0x80 [obdclass]
 RSP &amp;lt;ffff88001cc599a8&amp;gt;
CR2: ffffffffffffffde
---[ end trace 0ce3c640bfa74514 ]---
Kernel panic - not syncing: Fatal exception
Pid: 2886, comm: mount.lustre Tainted: P      D    ---------------    2.6.32-279.9.1.1chaos.ch5.1.x86_64 #1
Call Trace:
 [&amp;lt;ffffffff814fdceb&amp;gt;] ? panic+0xa0/0x168
 [&amp;lt;ffffffff81501e84&amp;gt;] ? oops_end+0xe4/0x100
 [&amp;lt;ffffffff81043bfb&amp;gt;] ? no_context+0xfb/0x260
 [&amp;lt;ffffffff81043e85&amp;gt;] ? __bad_area_nosemaphore+0x125/0x1e0
 [&amp;lt;ffffffff81043f53&amp;gt;] ? bad_area_nosemaphore+0x13/0x20
 [&amp;lt;ffffffff810446b1&amp;gt;] ? __do_page_fault+0x321/0x480
 [&amp;lt;ffffffff8127c23d&amp;gt;] ? pointer+0x8d/0x830
 [&amp;lt;ffffffffa06ff799&amp;gt;] ? _ldlm_lock_debug+0x189/0x5d0 [ptlrpc]
 [&amp;lt;ffffffff81503e3e&amp;gt;] ? do_page_fault+0x3e/0xa0
 [&amp;lt;ffffffff815011f5&amp;gt;] ? page_fault+0x25/0x30
 [&amp;lt;ffffffffa12d4f28&amp;gt;] ? lprocfs_srch+0x48/0x80 [obdclass]
 [&amp;lt;ffffffffa12d4f06&amp;gt;] ? lprocfs_srch+0x26/0x80 [obdclass]
 [&amp;lt;ffffffffa12d7084&amp;gt;] ? lprocfs_register+0x34/0x100 [obdclass]
 [&amp;lt;ffffffff8127d0a6&amp;gt;] ? vsnprintf+0x2b6/0x5f0
 [&amp;lt;ffffffffa109e3ce&amp;gt;] ? lprocfs_register_mountpoint+0x12e/0xb80 [lustre]
 [&amp;lt;ffffffffa108f226&amp;gt;] ? client_common_fill_super+0x1a6/0x4e70 [lustre]
 [&amp;lt;ffffffff8127d420&amp;gt;] ? sprintf+0x40/0x50
 [&amp;lt;ffffffffa1094c64&amp;gt;] ? ll_fill_super+0xd74/0x1500 [lustre]
 [&amp;lt;ffffffffa130fb6c&amp;gt;] ? lustre_fill_super+0x12c/0x1ae0 [obdclass]
 [&amp;lt;ffffffff8117cb16&amp;gt;] ? set_anon_super+0x56/0x100
 [&amp;lt;ffffffff8116a03c&amp;gt;] ? pcpu_alloc+0x3ac/0xa50
 [&amp;lt;ffffffff8127a24a&amp;gt;] ? strlcpy+0x4a/0x60
 [&amp;lt;ffffffff8117d9e3&amp;gt;] ? sget+0x3e3/0x480
 [&amp;lt;ffffffff8117cac0&amp;gt;] ? set_anon_super+0x0/0x100
 [&amp;lt;ffffffffa130fa40&amp;gt;] ? lustre_fill_super+0x0/0x1ae0 [obdclass]
 [&amp;lt;ffffffff8117df2f&amp;gt;] ? get_sb_nodev+0x5f/0xa0
 [&amp;lt;ffffffffa12fb185&amp;gt;] ? lustre_get_sb+0x25/0x30 [obdclass]
 [&amp;lt;ffffffff8117db8b&amp;gt;] ? vfs_kern_mount+0x7b/0x1b0
 [&amp;lt;ffffffff8117dd32&amp;gt;] ? do_kern_mount+0x52/0x130
 [&amp;lt;ffffffff8119c442&amp;gt;] ? do_mount+0x2d2/0x8d0
 [&amp;lt;ffffffff8119cad0&amp;gt;] ? sys_mount+0x90/0xe0
 [&amp;lt;ffffffff8100b0f2&amp;gt;] ? system_call_fastpath+0x16/0x1b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The reproducer:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;$ while true; sudo FSTYPE=zfs ONLY=&quot;59 60&quot; sh ./lustre/tests/replay-single.sh; done
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;It took about 6 hours to hit in my single node VM.&lt;/p&gt;

&lt;p&gt;There are also some other messages on the console just prior to the crash which I don&apos;t totally understand but I&apos;ll post here just for completeness:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;SLAB: cache with size 256 has lost its name
... snip ...
SLAB: cache with size 1152 has lost its name
LustreError: 958:0:(lprocfs_status.c:492:lprocfs_register())  Lproc: Attempting to register llite more than once 
SLAB: cache with size 256 has lost its name
... snip ...
SLAB: cache with size 1152 has lost its name
kmem_cache_create: duplicate cache lustre_inode_cache
SLAB: cache with size 256 has lost its name
... snip ...
SLAB: cache with size 1152 has lost its name
LustreError: 2522:0:(lprocfs_status.c:492:lprocfs_register())  Lproc: Attempting to register llite more than once 
SLAB: cache with size 256 has lost its name
... snip ...
SLAB: cache with size 1152 has lost its name
LustreError: 11-0: lustre-MDT0000-osp-MDT0000: Communicating with 0@lo, operation mds_connect failed with -11
LustreError: 11-0: lustre-MDT0000-osp-OST0000: Communicating with 0@lo, operation mds_connect failed with -11
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="16793">LU-2400</key>
            <summary>BUG: unable to handle kernel paging request in lprocfs_srch+0x48/0x80</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="2">Won&apos;t Fix</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="prakash">Prakash Surya</reporter>
                        <labels>
                            <label>llnl</label>
                    </labels>
                <created>Wed, 28 Nov 2012 11:52:34 +0000</created>
                <updated>Mon, 23 Jul 2018 09:04:41 +0000</updated>
                            <resolved>Mon, 23 Jul 2018 09:04:41 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="48542" author="pjones" created="Thu, 29 Nov 2012 14:30:02 +0000"  >&lt;p&gt;Alex, what do you think about this one?&lt;/p&gt;</comment>
                            <comment id="48619" author="adilger" created="Fri, 30 Nov 2012 14:57:21 +0000"  >&lt;p&gt;Looks like there was a leak in some of the slabs, and then remounting caused the client to be unhappy.  I don&apos;t think this really has anything to do with Orion code changes.&lt;/p&gt;</comment>
                            <comment id="50636" author="bzzz" created="Thu, 17 Jan 2013 02:12:40 +0000"  >&lt;p&gt;Oleg fixed a leak ( see &lt;a href=&quot;http://review.whamcloud.com/#change,4150&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,4150&lt;/a&gt; ), but it was on Oct 6 .. &lt;br/&gt;
any new instances?&lt;/p&gt;</comment>
                            <comment id="50668" author="prakash" created="Thu, 17 Jan 2013 10:29:38 +0000"  >&lt;p&gt;No, I only hit this the one time, but I haven&apos;t been trying to hit it either. I could let the reproducer run over night in a VM if you&apos;d like, to see if it triggers. Would that be useful?&lt;/p&gt;</comment>
                            <comment id="50670" author="bzzz" created="Thu, 17 Jan 2013 10:42:20 +0000"  >&lt;p&gt;yes, please do if you have a chance.&lt;/p&gt;</comment>
                            <comment id="230730" author="bzzz" created="Mon, 23 Jul 2018 09:04:41 +0000"  >&lt;p&gt;no more reports&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10490" key="com.atlassian.jira.plugin.system.customfieldtypes:datepicker">
                        <customfieldname>End date</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>Thu, 26 Jun 2014 11:52:34 +0000</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                            <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvcz3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5695</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                        <customfield id="customfield_10493" key="com.atlassian.jira.plugin.system.customfieldtypes:datepicker">
                        <customfieldname>Start date</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>Wed, 28 Nov 2012 11:52:34 +0000</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                    </customfields>
    </item>
</channel>
</rss>