<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:18:48 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-8580] general protection fault: osd_xattr_get+0x32c/0x5b0 [osd_ldiskfs]</title>
                <link>https://jira.whamcloud.com/browse/LU-8580</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Error happened during soak testing of build &apos;20160902&apos; (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160902&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160902&lt;/a&gt;)&lt;br/&gt;
Configuration reads as:&lt;br/&gt;
4 MDS with 1 MDT / MDS, backend FS &lt;em&gt;ldiskfs&lt;/em&gt;, nodes configured pairwise in active-active HA configuration&lt;br/&gt;
6 OSS with 4 OSTs / OSS, backend FS &lt;em&gt;zfs&lt;/em&gt;, nodes configured in pairwise in active-active HA configuration&lt;/p&gt;

&lt;p&gt;MDS crashed two times with the following message:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;lt;4&amp;gt;general protection fault: 0000 [#1] SMP 
&amp;lt;4&amp;gt;last sysfs file: /sys/devices/system/cpu/online
&amp;lt;4&amp;gt;CPU 26 
&amp;lt;4&amp;gt;Modules linked in: osp(U) mdd(U) lod(U) mdt(U) lfsck(U) mgc(U) osd_ldiskfs(U) ldiskfs(U) jbd2 lquota(U) lustre(U) lov(U) mdc(U) fid(U) lmv(U) fld(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) sha512_generic crc32c_intel libcfs(U) 8021q garp stp llc nfsd exportfs nfs lockd fscache auth_rpcgss nfs_acl sunrpc cpufreq_ondemand acpi_cpufreq freq_table mperf ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm dm_round_robin dm_multipath microcode iTCO_wdt iTCO_vendor_support zfs(P)(U) zcommon(P)(U) znvpair(P)(U) spl(U) zlib_deflate zavl(P)(U) zunicode(P)(U) sb_edac edac_core joydev lpc_ich mfd_core i2c_i801 ioatdma sg igb dca i2c_algo_bit i2c_core ext3 jbd mbcache sd_mod crc_t10dif ahci wmi isci libsas mpt2sas scsi_transport_sas raid_class mlx4_ib ib_sa ib_mad ib_core ib_addr ipv6 mlx4_en ptp pps_core mlx4_core dm_mirror dm_region_hash dm_log dm_mod scsi_dh_rdac [last unloaded: scsi_wait_scan]
&amp;lt;4&amp;gt;
&amp;lt;4&amp;gt;Pid: 6399, comm: mdt02_008 Tainted: P           -- ------------    2.6.32-573.26.1.el6_lustre.x86_64 #1 Intel Corporation S2600GZ ........../S2600GZ
&amp;lt;4&amp;gt;RIP: 0010:[&amp;lt;ffffffffa1083c9c&amp;gt;]  [&amp;lt;ffffffffa1083c9c&amp;gt;] osd_xattr_get+0x32c/0x5b0 [osd_ldiskfs]
&amp;lt;4&amp;gt;RSP: 0018:ffff8803fb9bf960  EFLAGS: 00010206
&amp;lt;4&amp;gt;RAX: 00000000ffffffff RBX: ffff8803fc661cc0 RCX: dead000000100100
&amp;lt;4&amp;gt;RDX: 0000000000000003 RSI: ffff88080efacdf8 RDI: ffffffffa12f59e4
&amp;lt;4&amp;gt;RBP: ffff8803fb9bf9b0 R08: 000000000000000b R09: ffff8803fc661d78
&amp;lt;4&amp;gt;R10: ffff88082a2e509c R11: 0000000000000000 R12: ffff88081392f010
&amp;lt;4&amp;gt;R13: ffffffffa12f59e4 R14: ffff8803de29bb70 R15: ffff880813931000
&amp;lt;4&amp;gt;FS:  0000000000000000(0000) GS:ffff88044e540000(0000) knlGS:0000000000000000
&amp;lt;4&amp;gt;CS:  0010 DS: 0018 ES: 0018 CR0: 000000008005003b
&amp;lt;4&amp;gt;CR2: 0000003d4feacd90 CR3: 0000000001a8d000 CR4: 00000000000407e0
&amp;lt;4&amp;gt;DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
&amp;lt;4&amp;gt;DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
&amp;lt;4&amp;gt;Process mdt02_008 (pid: 6399, threadinfo ffff8803fb9bc000, task ffff8803fb946040)
&amp;lt;4&amp;gt;Stack:
&amp;lt;4&amp;gt; ffff8803fc661d78 ffff88080efacdc0 000000000000000b ffff88080efacdf8
&amp;lt;4&amp;gt;&amp;lt;d&amp;gt; ffff8803fb9bf9c0 ffff88081392f000 ffff8803fc661cc0 ffff88082cca66c0
&amp;lt;4&amp;gt;&amp;lt;d&amp;gt; ffffffffa12f59e4 ffff88081392f010 ffff8803fb9bf9f0 ffffffffa12ccdc3
&amp;lt;4&amp;gt;Call Trace:
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa12ccdc3&amp;gt;] lod_get_ea+0xc3/0x530 [lod]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa12de6bc&amp;gt;] lod_ah_init+0x6cc/0x980 [lod]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1359e49&amp;gt;] mdd_object_make_hint+0x139/0x180 [mdd]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1084f08&amp;gt;] ? osd_object_read_unlock+0x88/0xd0 [osd_ldiskfs]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1356251&amp;gt;] mdd_create+0x6f1/0x1770 [mdd]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa090ca41&amp;gt;] ? lu_object_find_at+0xb1/0xe0 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1212b94&amp;gt;] ? mdt_version_save+0x84/0x1a0 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa121cc4c&amp;gt;] mdt_reint_create+0xbdc/0xfe0 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa120e30c&amp;gt;] ? mdt_root_squash+0x2c/0x3f0 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0b058db&amp;gt;] ? lustre_pack_reply_v2+0x1eb/0x280 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81299b7a&amp;gt;] ? strlcpy+0x4a/0x60
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa120f97a&amp;gt;] ? old_init_ucred_common+0xda/0x2b0 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1211ead&amp;gt;] mdt_reint_rec+0x5d/0x200 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa11fd5db&amp;gt;] mdt_reint_internal+0x62b/0xa50 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa11fdeab&amp;gt;] mdt_reint+0x6b/0x120 [mdt]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0b69ccc&amp;gt;] tgt_request_handle+0x8ec/0x1440 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0b16501&amp;gt;] ptlrpc_main+0xd31/0x1800 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8106ee50&amp;gt;] ? pick_next_task_fair+0xd0/0x130
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81539896&amp;gt;] ? schedule+0x176/0x3a0
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0b157d0&amp;gt;] ? ptlrpc_main+0x0/0x1800 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff810a138e&amp;gt;] kthread+0x9e/0xc0
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
&amp;lt;4&amp;gt; [&amp;lt;ffffffff810a12f0&amp;gt;] ? kthread+0x0/0xc0
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
&amp;lt;4&amp;gt;Code: e0 48 8b 8b b8 00 00 00 4c 8d 8b b8 00 00 00 49 89 c0 4c 39 c9 75 14 e9 8a 01 00 00 0f 1f 00 48 8b 09 49 39 c9 0f 84 7b 01 00 00 &amp;lt;4c&amp;gt; 3b 41 18 75 ee 48 8d 41 38 4c 89 c2 4c 89 ef 48 89 4d b8 4c 
&amp;lt;1&amp;gt;RIP  [&amp;lt;ffffffffa1083c9c&amp;gt;] osd_xattr_get+0x32c/0x5b0 [osd_ldiskfs]
&amp;lt;4&amp;gt; RSP &amp;lt;ffff8803fb9bf960&amp;gt;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Sequence of events:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-09-02-12:58:40   First crash; happened during &apos;normal&apos; operations while no fault had been injected&lt;/li&gt;
	&lt;li&gt;2016-09-02-16:11:03   Second crash; happened after MDS restart&lt;/li&gt;
	&lt;li&gt;Both incidents occurred on the same node (&lt;tt&gt;lola-11&lt;/tt&gt;)&lt;/li&gt;
	&lt;li&gt;No errors on other nodes can be correlate to these events&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Attached files: messages, console, vmcore files of node &lt;tt&gt;lola-11&lt;/tt&gt; of both crashes. (Note console have time stamps printed in 5 min intervals)&lt;br/&gt;
Crash dump files are available.&lt;/p&gt;</description>
                <environment>lola&lt;br/&gt;
&lt;a href=&quot;https://build.hpdd.intel.com/job/lustre-master/3431/&quot;&gt;https://build.hpdd.intel.com/job/lustre-master/3431/&lt;/a&gt;   (tag 2.8.57 for el6.7)</environment>
        <key id="39608">LU-8580</key>
            <summary>general protection fault: osd_xattr_get+0x32c/0x5b0 [osd_ldiskfs]</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="heckes">Frank Heckes</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Mon, 5 Sep 2016 13:32:10 +0000</created>
                <updated>Fri, 7 Oct 2016 12:15:51 +0000</updated>
                            <resolved>Fri, 7 Oct 2016 12:15:51 +0000</resolved>
                                    <version>Lustre 2.9.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="164912" author="heckes" created="Mon, 5 Sep 2016 13:55:19 +0000"  >&lt;p&gt;Crash dump files have been saved to the subdirectories &lt;tt&gt;127.0.0.1-2016-09-02-12:58:40, 127.0.0.1-2016-09-02-16:11:03&lt;/tt&gt; of &lt;tt&gt;lhn.hpdd.intel.com:/scratch/crashdumps/lu-8580/lola-11&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="164941" author="heckes" created="Tue, 6 Sep 2016 10:19:37 +0000"  >&lt;p&gt;The error happens quite often. 5 more incidents on all MDS nodes.&lt;/p&gt;</comment>
                            <comment id="165282" author="heckes" created="Thu, 8 Sep 2016 11:19:43 +0000"  >&lt;p&gt;one more crash during last nights session.&lt;/p&gt;</comment>
                            <comment id="165460" author="bzzz" created="Fri, 9 Sep 2016 12:08:01 +0000"  >&lt;p&gt;Lai, can this be similar to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8581&quot; title=&quot;Kernel Panic - osd_oxc_lookup+0x38/0x70 [osd_ldiskfs]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8581&quot;&gt;&lt;del&gt;LU-8581&lt;/del&gt;&lt;/a&gt; ? we started to use RCU in OSD quite recently and the code is very similar.&lt;/p&gt;</comment>
                            <comment id="165609" author="laisiyao" created="Mon, 12 Sep 2016 03:29:35 +0000"  >&lt;p&gt;yes, the backtrace of &quot;osd_xattr_get+0x32c&quot; is exactly in osd_oxc_get() list traversal.&lt;/p&gt;</comment>
                            <comment id="165612" author="pjones" created="Mon, 12 Sep 2016 04:06:41 +0000"  >&lt;p&gt;ok then let&apos;s track the fix under &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8581&quot; title=&quot;Kernel Panic - osd_oxc_lookup+0x38/0x70 [osd_ldiskfs]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8581&quot;&gt;&lt;del&gt;LU-8581&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="165613" author="pjones" created="Mon, 12 Sep 2016 04:43:47 +0000"  >&lt;p&gt;Ooops. My misunderstanding - it&apos;s a similar issue but not a duplicate&lt;/p&gt;</comment>
                            <comment id="165786" author="laisiyao" created="Tue, 13 Sep 2016 02:56:45 +0000"  >&lt;p&gt;IMO this is a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8581&quot; title=&quot;Kernel Panic - osd_oxc_lookup+0x38/0x70 [osd_ldiskfs]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8581&quot;&gt;&lt;del&gt;LU-8581&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="165824" author="bzzz" created="Tue, 13 Sep 2016 13:50:25 +0000"  >&lt;p&gt;well, we can&apos;t fix this issue with the patch for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8581&quot; title=&quot;Kernel Panic - osd_oxc_lookup+0x38/0x70 [osd_ldiskfs]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8581&quot;&gt;&lt;del&gt;LU-8581&lt;/del&gt;&lt;/a&gt;. so technically it&apos;s not a duplicate. but it&apos;s very similar, of course.&lt;/p&gt;</comment>
                            <comment id="166366" author="heckes" created="Mon, 19 Sep 2016 09:56:18 +0000"  >&lt;p&gt;Error occurs with frequency 1 - 2 hours for build &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160916&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160916&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="168652" author="bzzz" created="Fri, 7 Oct 2016 12:15:51 +0000"  >&lt;p&gt;a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8581&quot; title=&quot;Kernel Panic - osd_oxc_lookup+0x38/0x70 [osd_ldiskfs]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8581&quot;&gt;&lt;del&gt;LU-8581&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="39609">LU-8581</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="22948" name="console-lola-11.log.bz2" size="44237" author="heckes" created="Mon, 5 Sep 2016 13:58:36 +0000"/>
                            <attachment id="22949" name="lola-11-vmcore-dmesg.txt-2016-09-02-125840.bz2" size="21410" author="heckes" created="Mon, 5 Sep 2016 13:58:36 +0000"/>
                            <attachment id="22950" name="lola-11-vmcore-dmesg.txt-2016-09-02-161103.bz2" size="21653" author="heckes" created="Mon, 5 Sep 2016 13:58:36 +0000"/>
                            <attachment id="22947" name="messages-lola-11.log.bz2" size="124087" author="heckes" created="Mon, 5 Sep 2016 13:58:36 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzyndr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>