<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:06:08 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-347] Failure on test suite conf-sanity.sh test_48</title>
                <link>https://jira.whamcloud.com/browse/LU-347</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/e5b0eb78-8228-11e0-b4df-52540025f9af&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/e5b0eb78-8228-11e0-b4df-52540025f9af&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;09:18:42:Lustre: DEBUG MARKER: == conf-sanity test 48: too many acls on file ======================================================== 02:18:41 (1305796721)&lt;br/&gt;
09:18:45:LDISKFS-fs (sdb): warning: maximal mount count reached, running e2fsck is recommended&lt;br/&gt;
09:18:50:Lustre: lustre-MDT0000: new disk, initializing&lt;br/&gt;
09:18:51:LustreError: 11651:0:(ldlm_lib.c:2118:target_send_reply_msg()) @@@ processing error (&lt;del&gt;11)  req@ffff81040cfb4800 x1369227103567881/t0(0) o-1&lt;/del&gt;&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 368/0 e 0 to 0 dl 1305796751 ref 1 fl Interpret:/ffffffff/ffffffff rc -11/-1&lt;br/&gt;
09:18:51:LustreError: 11651:0:(ldlm_lib.c:2118:target_send_reply_msg()) Skipped 25 previous similar messages&lt;br/&gt;
09:18:58:Lustre: 11831:0:(quota_master.c:793:close_quota_files()) quota&lt;span class=&quot;error&quot;&gt;&amp;#91;0&amp;#93;&lt;/span&gt; is off already&lt;br/&gt;
09:18:58:Lustre: 11831:0:(quota_master.c:793:close_quota_files()) Skipped 35 previous similar messages&lt;br/&gt;
09:18:59:Unable to handle kernel paging request at 0000000000100100 RIP: &lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff889ae134&amp;gt;&amp;#93;&lt;/span&gt; :ptlrpc:ptlrpc_unregister_service+0x1e4/0xbd0&lt;br/&gt;
09:18:59:PGD 0 &lt;br/&gt;
09:18:59:Oops: 0000 &lt;span class=&quot;error&quot;&gt;&amp;#91;1&amp;#93;&lt;/span&gt; SMP &lt;br/&gt;
09:18:59:last sysfs file: /block/sdb/queue/max_sectors_kb&lt;br/&gt;
09:18:59:CPU 3 &lt;br/&gt;
09:18:59:Modules linked in: nfs(U) fscache(U) nfs_acl(U) cmm(U) osd_ldiskfs(U) mdt(U) mdd(U) mds(U) fsfilt_ldiskfs(U) exportfs(U) mgs(U) mgc(U) lustre(U) lov(U) osc(U) lquota(U) mdc(U) fid(U) fld(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) libcfs(U) ldiskfs(U) jbd2(U) crc16(U) autofs4(U) hidp(U) rfcomm(U) l2cap(U) bluetooth(U) lockd(U) sunrpc(U) cpufreq_ondemand(U) powernow_k8(U) freq_table(U) be2iscsi(U) iscsi_tcp(U) bnx2i(U) cnic(U) uio(U) cxgb3i(U) iw_cxgb3(U) cxgb3(U) libiscsi_tcp(U) ib_iser(U) libiscsi2(U) scsi_transport_iscsi2(U) scsi_transport_iscsi(U) ib_srp(U) rds(U) ib_sdp(U) ib_ipoib(U) ipoib_helper(U) ipv6(U) xfrm_nalgo(U) crypto_api(U) rdma_ucm(U) rdma_cm(U) ib_ucm(U) ib_uverbs(U) ib_umad(U) ib_cm(U) iw_cm(U) ib_addr(U) ib_sa(U) loop(U) dm_mirror(U) dm_multipath(U) scsi_dh(U) video(U) backlight(U) sbs(U) power_meter(U) hwmon(U) i2c_ec(U) dell_wmi(U) wmi(U) button(U) battery(U) asus_acpi(U) acpi_memhotplug(U) ac(U) parport_pc(U) lp(U) parport(U) mlx4_ib(U) ib_mad(U) ib_core(U) mlx4_en(U) shpchp(U) igb(U) 8021q(U) sg(U) mlx4_core(U) dca(U) i2c_piix4(U) i2c_core(U) amd64_edac_mod(U) edac_mc(U) pcspkr(U) serio_raw(U) dm_raid45(U) dm_message(U) dm_region_hash(U) dm_log(U) dm_mod(U) dm_mem_cache(U) ahci(U) libata(U) sd_mod(U) scsi_mod(U) ext3(U) jbd(U) uhci_hcd(U) ohci_hcd(U) ehci_hcd(U)&lt;br/&gt;
09:18:59:Pid: 5695, comm: obd_zombid Tainted: G      2.6.18-194.17.1.el5_lustre.gce09fce #1&lt;br/&gt;
09:18:59:RIP: 0010:&lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff889ae134&amp;gt;&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff889ae134&amp;gt;&amp;#93;&lt;/span&gt; :ptlrpc:ptlrpc_unregister_service+0x1e4/0xbd0&lt;br/&gt;
09:18:59:RSP: 0018:ffff81021eb6dcd0  EFLAGS: 00010246&lt;br/&gt;
09:18:59:RAX: 00000000fffffffe RBX: 0000000000100100 RCX: ffff8100765189d0&lt;br/&gt;
09:18:59:RDX: ffff81031e11d250 RSI: 0000000000000001 RDI: 0000000001406495&lt;br/&gt;
09:18:59:RBP: ffff810078fe0c00 R08: 0000000000000001 R09: 0000000000000001&lt;br/&gt;
09:18:59:R10: 00000000ffffffff R11: 00000000000000d0 R12: ffff81010a4266a8&lt;br/&gt;
09:18:59:R13: 0000000000000000 R14: 0000000000000000 R15: ffff810078fe0ca8&lt;br/&gt;
09:18:59:FS:  00002b36a618f6e0(0000) GS:ffff8102239255c0(0000) knlGS:0000000000000000&lt;br/&gt;
09:18:59:CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b&lt;br/&gt;
09:18:59:CR2: 0000000000100100 CR3: 0000000000201000 CR4: 00000000000006e0&lt;br/&gt;
09:18:59:Process obd_zombid (pid: 5695, threadinfo ffff81021eb6c000, task ffff81010537c080)&lt;br/&gt;
09:18:59:Stack:  0000000000000000 0000000000000000 ffffffff88a04a34 ffffffff88ccc670&lt;br/&gt;
09:18:59: ffffffff88a6bbd0 0000000000000000 ffff810078fe0dd8 ffff810078fe0ce0&lt;br/&gt;
09:18:59: ffff810078fe0db8 ffffffff88a6bbd0 0000000000000000 0000000000000001&lt;br/&gt;
09:18:59:Call Trace:&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8002e244&amp;gt;&amp;#93;&lt;/span&gt; __wake_up+0x38/0x4f&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff88ca70bb&amp;gt;&amp;#93;&lt;/span&gt; :mgs:mgs_cleanup+0xeb/0x220&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff888b82cf&amp;gt;&amp;#93;&lt;/span&gt; :obdclass:class_decref+0x43f/0x5b0&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff88ca72d6&amp;gt;&amp;#93;&lt;/span&gt; :mgs:mgs_destroy_export+0xe6/0xf0&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8889e302&amp;gt;&amp;#93;&lt;/span&gt; :obdclass:obd_zombie_impexp_cull+0x402/0x4f0&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff888a5847&amp;gt;&amp;#93;&lt;/span&gt; :obdclass:obd_zombie_impexp_thread+0x1f7/0x2a0&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8008cf99&amp;gt;&amp;#93;&lt;/span&gt; default_wake_function+0x0/0xe&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8005dfb1&amp;gt;&amp;#93;&lt;/span&gt; child_rip+0xa/0x11&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff888a5650&amp;gt;&amp;#93;&lt;/span&gt; :obdclass:obd_zombie_impexp_thread+0x0/0x2a0&lt;br/&gt;
09:18:59: &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8005dfa7&amp;gt;&amp;#93;&lt;/span&gt; child_rip+0x0/0x11&lt;br/&gt;
09:18:59:&lt;br/&gt;
09:18:59:&lt;br/&gt;
09:18:59:Code: 48 8b 03 0f 18 08 48 39 5c 24 38 75 bf 4c 8d bd a8 00 00 00 &lt;br/&gt;
09:18:59:RIP  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff889ae134&amp;gt;&amp;#93;&lt;/span&gt; :ptlrpc:ptlrpc_unregister_service+0x1e4/0xbd0&lt;br/&gt;
09:18:59: RSP &amp;lt;ffff81021eb6dcd0&amp;gt;&lt;br/&gt;
09:18:59:CR2: 0000000000100100&lt;br/&gt;
09:18:59: &amp;lt;0&amp;gt;Kernel panic - not syncing: Fatal exception&lt;/p&gt;</description>
                <environment></environment>
        <key id="10954">LU-347</key>
            <summary>Failure on test suite conf-sanity.sh test_48</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="bobijam">Zhenyu Xu</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Thu, 19 May 2011 11:39:39 +0000</created>
                <updated>Fri, 27 May 2011 07:01:37 +0000</updated>
                            <resolved>Fri, 27 May 2011 06:47:29 +0000</resolved>
                                    <version>Lustre 2.1.0</version>
                                    <fixVersion>Lustre 2.1.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>0</watches>
                                                                            <comments>
                            <comment id="15089" author="pjones" created="Thu, 26 May 2011 06:20:03 +0000"  >&lt;p&gt;Bobijam&lt;/p&gt;

&lt;p&gt;Could you please look into this one? It is showing up regularly during testing&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="15166" author="bobijam" created="Fri, 27 May 2011 06:43:00 +0000"  >&lt;p&gt;from &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/9b40afc6-871c-11e0-b4df-52540025f9af&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/9b40afc6-871c-11e0-b4df-52540025f9af&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;17:25:56: [] :ptlrpc:ptlrpc_unregister_service+0x580/0xbd0&lt;br/&gt;
17:25:56: [] __wake_up+0x38/0x4f&lt;br/&gt;
17:25:56: [] :mgs:mgs_cleanup+0xeb/0x220&lt;br/&gt;
17:25:56: [] :obdclass:class_decref+0x43f/0x5b0&lt;br/&gt;
17:25:56: [] :obdclass:class_incref+0x7a/0x90&lt;br/&gt;
17:25:56: [] :obdclass:obd_devlist_next+0x65/0x80&lt;/p&gt;

&lt;p&gt;and my local test ran similar panic &lt;/p&gt;

&lt;p&gt;crash&amp;gt; bt&lt;br/&gt;
PID: 17830  TASK: ffff81000d61f790  CPU: 1   COMMAND: &quot;lctl&quot;&lt;br/&gt;
 #0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1a20&amp;#93;&lt;/span&gt; crash_kexec at ffffffff800af38f&lt;br/&gt;
 #1 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1ae0&amp;#93;&lt;/span&gt; __die at ffffffff80065117&lt;br/&gt;
 #2 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1b20&amp;#93;&lt;/span&gt; do_page_fault at ffffffff8006748d&lt;br/&gt;
 #3 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1c10&amp;#93;&lt;/span&gt; error_exit at ffffffff8005dde9&lt;br/&gt;
    &lt;span class=&quot;error&quot;&gt;&amp;#91;exception RIP: ptlrpc_unregister_service+115&amp;#93;&lt;/span&gt;&lt;br/&gt;
    RIP: ffffffff88646ff3  RSP: ffff810015fa1cc8  RFLAGS: 00010296&lt;br/&gt;
    RAX: 0000000000000001  RBX: ffff81000a55e700  RCX: 0000000000000001&lt;br/&gt;
    RDX: 0000000000000d32  RSI: ffffffff8868595a  RDI: ffff810016ed8080&lt;br/&gt;
    RBP: 0000000000000000   R8: 0000000000000000   R9: 0000000000000a34&lt;br/&gt;
    R10: 00000000ffffffff  R11: 0000000000000000  R12: ffff81000a55e1b8&lt;br/&gt;
    R13: 0000000000000000  R14: 0000000000000000  R15: 00002b0b5e5b4000&lt;br/&gt;
    ORIG_RAX: ffffffffffffffff  CS: 0010  SS: 0018&lt;br/&gt;
 #4 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1cc0&amp;#93;&lt;/span&gt; ptlrpc_unregister_service at ffffffff88646ff3&lt;br/&gt;
 #5 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1d80&amp;#93;&lt;/span&gt; ost_cleanup at ffffffff88a88caf&lt;br/&gt;
 #6 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1dd0&amp;#93;&lt;/span&gt; class_decref at ffffffff8855425f&lt;br/&gt;
 #7 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1e70&amp;#93;&lt;/span&gt; obd_devlist_next at ffffffff8853b1a5&lt;br/&gt;
 #8 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff810015fa1e90&amp;#93;&lt;/span&gt; obd_device_list_seq_next at ffffffff885230c7&lt;/p&gt;

&lt;p&gt;Looks like obd device is accessed during its cleanup. Lai Siyao&apos;s patch in &lt;a href=&quot;http://review.whamcloud.com/#patch,sidebyside,326,13,lustre/obdclass/genops.c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#patch,sidebyside,326,13,lustre/obdclass/genops.c&lt;/a&gt; will handle it (it&apos;s patch for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-106&quot; title=&quot;unable to handle kernel paging request in lprocfs_stats_collect()&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-106&quot;&gt;&lt;del&gt;LU-106&lt;/del&gt;&lt;/a&gt;)&lt;/p&gt;</comment>
                            <comment id="15167" author="bobijam" created="Fri, 27 May 2011 06:47:30 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-106&quot; title=&quot;unable to handle kernel paging request in lprocfs_stats_collect()&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-106&quot;&gt;&lt;del&gt;LU-106&lt;/del&gt;&lt;/a&gt;&apos;s patch can fix this error, resolve it as dup of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-106&quot; title=&quot;unable to handle kernel paging request in lprocfs_stats_collect()&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-106&quot;&gt;&lt;del&gt;LU-106&lt;/del&gt;&lt;/a&gt;, and feel free to reopen it if the issue reoccurs after &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-106&quot; title=&quot;unable to handle kernel paging request in lprocfs_stats_collect()&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-106&quot;&gt;&lt;del&gt;LU-106&lt;/del&gt;&lt;/a&gt;&apos;s patch got landed.&lt;/p&gt;

&lt;p&gt;root cause:&lt;br/&gt;
mgs was in clean up phase, while conf-sanity.sh was checking /proc/fs/lustre/devices and find the mgs device, obd_devlist_next() class_incref() it, afterward class_decref() it which called obd_cleanup() again, causing invalid memory access violation.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv4hj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4260</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>