<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:13:11 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7935] MDS crash with NULL pointer dereference at 0000000000000010</title>
                <link>https://jira.whamcloud.com/browse/LU-7935</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Error happens during soak testing of build &apos;20160324&apos; (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160324&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160324&lt;/a&gt;). DNE is enabled. MDTs had been formatted with &lt;em&gt;ldiskfs&lt;/em&gt;, OSTs using &lt;em&gt;zfs&lt;/em&gt;. MDSes are configured in HA active-active failover configuration with &lt;b&gt;1&lt;/b&gt; MDT per MDS.&lt;br/&gt;
Nodes &lt;tt&gt;lola-8&lt;/tt&gt; and &lt;tt&gt;lola-9&lt;/tt&gt; form a HA cluster.&lt;/p&gt;

&lt;p&gt;Sequence of events:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-03-28 11:38:34  triggering fault mds_failover (lola-8)&lt;br/&gt;
                                      rebooting node lola-8&lt;/li&gt;
	&lt;li&gt;2016-03-28 11:44:55  lola-8 up again&lt;/li&gt;
	&lt;li&gt;2016-03-28 11:46:04  MDT0000 mounted on lola-9&lt;/li&gt;
	&lt;li&gt;2016-03-28 11:47:06,056:fsmgmt.fsmgmt:INFO     Node lola-9: &apos;soaked-MDT0000&apos; recovery completed&lt;/li&gt;
	&lt;li&gt;2016-03-28 11:47:06,056:fsmgmt.fsmgmt:INFO     Failing back soaked-MDT0000 ...   (aka umount MDT0000 on lola-9)&lt;/li&gt;
	&lt;li&gt;lola-9 crashed with message:
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;lt;4&amp;gt;NULL pointer dereference at 0000000000000010
&amp;lt;1&amp;gt;IP: [&amp;lt;ffffffffa084bbb7&amp;gt;] lu_context_key_get+0x17/0x60 [obdclass]
&amp;lt;4&amp;gt;PGD 0 
&amp;lt;4&amp;gt;Oops: 0000 [#1] SMP 
&amp;lt;4&amp;gt;last sysfs file: /sys/devices/system/cpu/online
&amp;lt;4&amp;gt;CPU 10 
&amp;lt;4&amp;gt;Modules linked in: mgs(U) osp(U) mdd(U) lod(U) mdt(U) lfsck(U) mgc(U) osd_ldiskfs(U) ldiskfs(U) jbd2 lquota(U) lustre(U) lov(U) mdc(U) fid(U) lmv(U) fld(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) sha512_generic crc32c_intel libcfs(U) 8021q garp stp llc nfsd exportfs nfs lockd fscache auth_rpcgss nfs_acl sunrpc cpufreq_ondemand acpi_cpufreq freq_table mperf ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm scsi_dh_rdac dm_round_robin dm_multipath microcode iTCO_wdt iTCO_vendor_support zfs(P)(U) zcommon(P)(U) znvpair(P)(U) spl(U) zlib_deflate zavl(P)(U) zunicode(P)(U) sb_edac edac_core lpc_ich mfd_core i2c_i801 ioatdma sg igb dca i2c_algo_bit i2c_core ptp pps_core ext3 jbd mbcache sd_mod crc_t10dif ahci isci libsas wmi mpt2sas scsi_transport_sas raid_class mlx4_ib ib_sa ib_mad ib_core ib_addr ipv6 mlx4_core dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]
&amp;lt;4&amp;gt;
&amp;lt;4&amp;gt;Pid: 11450, comm: osp_up2-0 Tainted: P           ---------------    2.6.32-504.30.3.el6_lustre.g2aa02ca.x86_64 #1 Intel Corporation S2600GZ ........../S2600GZ
&amp;lt;4&amp;gt;RIP: 0010:[&amp;lt;ffffffffa084bbb7&amp;gt;]  [&amp;lt;ffffffffa084bbb7&amp;gt;] lu_context_key_get+0x17/0x60 [obdclass]
&amp;lt;4&amp;gt;RSP: 0018:ffff8807ae5997f0  EFLAGS: 00010246
&amp;lt;4&amp;gt;RAX: 0000000000000008 RBX: 0000000000000000 RCX: ffff8807ae5998b0
&amp;lt;4&amp;gt;RDX: 0000000280023695 RSI: ffffffffa0cdf9a0 RDI: 0000000000000000
&amp;lt;4&amp;gt;RBP: ffff8807ae5997f0 R08: ffff8807ae599918 R09: ffff8807e5962290
&amp;lt;4&amp;gt;R10: 0000000000000007 R11: 2000000000000000 R12: ffff8803f4ed6cc0
&amp;lt;4&amp;gt;R13: 0000000280023695 R14: ffff8807ae5998b0 R15: ffff8807b15a1b50
&amp;lt;4&amp;gt;FS:  0000000000000000(0000) GS:ffff88044e440000(0000) knlGS:0000000000000000
&amp;lt;4&amp;gt;CS:  0010 DS: 0018 ES: 0018 CR0: 000000008005003b
&amp;lt;4&amp;gt;CR2: 0000000000000010 CR3: 0000000001a85000 CR4: 00000000000407e0
&amp;lt;4&amp;gt;DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
&amp;lt;4&amp;gt;DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
&amp;lt;4&amp;gt;Process osp_up2-0 (pid: 11450, threadinfo ffff8807ae598000, task ffff8807af36cab0)
&amp;lt;4&amp;gt;Stack:
&amp;lt;4&amp;gt; ffff8807ae599840 ffffffffa0cd883f ffff880000033c28 0006125000000001
&amp;lt;4&amp;gt;&amp;lt;d&amp;gt; 0000000000000246 ffff8803d9081498 0000000280023695 ffff8803f4ed6cc0
&amp;lt;4&amp;gt;&amp;lt;d&amp;gt; ffff8807ae5998b0 0000000000000000 ffff8807ae599890 ffffffffa0cd8c83
&amp;lt;4&amp;gt;Call Trace:
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0cd883f&amp;gt;] fld_local_lookup+0x4f/0x290 [fld]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0cd8c83&amp;gt;] fld_server_lookup+0x53/0x330 [fld]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa123738f&amp;gt;] lod_fld_lookup+0x34f/0x520 [lod]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff811753da&amp;gt;] ? kmem_cache_alloc+0x18a/0x190
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa124d243&amp;gt;] lod_object_init+0x103/0x3c0 [lod]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa084f1f8&amp;gt;] lu_object_alloc+0xd8/0x320 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa08505e1&amp;gt;] lu_object_find_try+0x151/0x260 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa08507a1&amp;gt;] lu_object_find_at+0xb1/0xe0 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa084f093&amp;gt;] ? lu_object_free+0x113/0x1a0 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa085080f&amp;gt;] lu_object_find_slice+0x1f/0x80 [obdclass]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1342a4e&amp;gt;] osp_trans_stop_cb+0x1be/0x2d0 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa13442be&amp;gt;] osp_update_interpret+0x21e/0x4a0 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8108742c&amp;gt;] ? lock_timer_base+0x3c/0x70
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0a600e5&amp;gt;] ptlrpc_check_set+0x615/0x1da0 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8152b22a&amp;gt;] ? schedule_timeout+0x19a/0x2e0
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0a61bca&amp;gt;] ptlrpc_set_wait+0x35a/0x960 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81064c00&amp;gt;] ? default_wake_function+0x0/0x20
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0a6de85&amp;gt;] ? lustre_msg_set_jobid+0xf5/0x130 [ptlrpc]
&amp;lt;3&amp;gt;LustreError: 11-0: soaked-MDT0000-osp-MDT0001: operation out_update to node 0@lo failed: rc = -107
&amp;lt;3&amp;gt;LustreError: Skipped 752 previous similar messages
&amp;lt;4&amp;gt;Lustre: soaked-MDT0000-osp-MDT0001: Connection to soaked-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa0a62251&amp;gt;] ptlrpc_queue_wait+0x81/0x220 [ptlrpc]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa13449c6&amp;gt;] osp_send_update_req+0x256/0x850 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa134130c&amp;gt;] ? osp_get_next_request+0xfc/0x1a0 [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa134563f&amp;gt;] osp_send_update_thread+0x20f/0x7ac [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff81064c00&amp;gt;] ? default_wake_function+0x0/0x20
&amp;lt;4&amp;gt; [&amp;lt;ffffffffa1345430&amp;gt;] ? osp_send_update_thread+0x0/0x7ac [osp]
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8109e78e&amp;gt;] kthread+0x9e/0xc0
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8109e6f0&amp;gt;] ? kthread+0x0/0xc0
&amp;lt;4&amp;gt; [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
&amp;lt;4&amp;gt;Code: c4 38 5b 41 5c 41 5d c9 c3 66 66 2e 0f 1f 84 00 00 00 00 00 55 48 89 e5 0f 1f 44 00 00 48 63 46 20 48 3b 34 c5 e0 26 8d a0 75 0a &amp;lt;48&amp;gt; 8b 57 10 48 8b 04 c2 c9 c3 48 c7 c7 00 b3 8a a0 48 c7 c2 c8 
&amp;lt;1&amp;gt;RIP  [&amp;lt;ffffffffa084bbb7&amp;gt;] lu_context_key_get+0x17/0x60 [obdclass]
&amp;lt;4&amp;gt; RSP &amp;lt;ffff8807ae5997f0&amp;gt;
&amp;lt;4&amp;gt;CR2: 0000000000000010
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Attached files: &lt;br/&gt;
console, messages and vmcore-dmesg.txt of node &lt;tt&gt;lola-9&lt;/tt&gt;&lt;/p&gt;</description>
                <environment>lola&lt;br/&gt;
build: 2.8 GA + patches</environment>
        <key id="35660">LU-7935</key>
            <summary>MDS crash with NULL pointer dereference at 0000000000000010</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="laisiyao">Lai Siyao</assignee>
                                    <reporter username="heckes">Frank Heckes</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Tue, 29 Mar 2016 13:04:45 +0000</created>
                <updated>Tue, 22 Nov 2016 03:25:03 +0000</updated>
                            <resolved>Tue, 22 Nov 2016 03:25:03 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="147166" author="heckes" created="Tue, 29 Mar 2016 13:25:00 +0000"  >&lt;p&gt;Crash dump has been saved to directory &lt;tt&gt;lhn.lola.hpdd.intel.com: /scratch/crashdumps/lu-7935/lola-9/127.0.0.1-2016-03-28-11:47:23&lt;/tt&gt;&lt;/p&gt;</comment>
                            <comment id="147223" author="jgmitter" created="Tue, 29 Mar 2016 17:36:58 +0000"  >&lt;p&gt;Hi Lai,&lt;br/&gt;
Can you have a look into this issue?&lt;br/&gt;
Thanks.&lt;br/&gt;
Joe&lt;/p&gt;</comment>
                            <comment id="147319" author="heckes" created="Wed, 30 Mar 2016 07:43:05 +0000"  >&lt;p&gt;Error is reproducible and happened again at 2016-03-29-14:53:37 for the same HA pair and the same node. &lt;br/&gt;
I saved the crash files to &lt;tt&gt;lhn.lola.hpdd/intel.com:/scratch/crashdumps/lu-7935/lola-9/127.0.0.1-2016-03-29-14:53:37&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="147323" author="di.wang" created="Wed, 30 Mar 2016 09:07:56 +0000"  >&lt;p&gt;Hmm, we probably need create env here, which might be related with 7782&lt;/p&gt;</comment>
                            <comment id="147417" author="laisiyao" created="Thu, 31 Mar 2016 07:27:23 +0000"  >&lt;p&gt;I agree with Di, &lt;a href=&quot;http://review.whamcloud.com/#/c/18493/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/18493/&lt;/a&gt; is the fix for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7782&quot; title=&quot;sanity-scrub test_2: NULL pointer dereference at 0x10 in lu_context_key_get() on mds2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7782&quot;&gt;&lt;del&gt;LU-7782&lt;/del&gt;&lt;/a&gt;, which has been landed to latest master.&lt;/p&gt;</comment>
                            <comment id="174590" author="laisiyao" created="Tue, 22 Nov 2016 03:25:03 +0000"  >&lt;p&gt;Duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7782&quot; title=&quot;sanity-scrub test_2: NULL pointer dereference at 0x10 in lu_context_key_get() on mds2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7782&quot;&gt;&lt;del&gt;LU-7782&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="34728">LU-7782</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="20922" name="console-lola-9.log.bz2" size="48229" author="heckes" created="Tue, 29 Mar 2016 13:19:06 +0000"/>
                            <attachment id="20923" name="messages-lola-9.log.bz2" size="92798" author="heckes" created="Tue, 29 Mar 2016 13:19:06 +0000"/>
                            <attachment id="20924" name="vmcore-dmesg.txt.bz2" size="37597" author="heckes" created="Tue, 29 Mar 2016 13:19:06 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy5yv:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>