<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:54:34 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-12663] SOAK: OSS hit general protection fault: 0000 [#1] SMP</title>
                <link>https://jira.whamcloud.com/browse/LU-12663</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;2 OSS hit the &quot;general protection fault: 0000 &lt;a href=&quot;#1&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;1&lt;/a&gt; SMP&quot; in failover test when running for about 2 days&lt;/p&gt;

&lt;p&gt;soak-4 and soak-5 all shows&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[  320.525059] LustreError: Skipped 5 previous similar messages
[  320.554484] Lustre: Skipped 6 previous similar messages
[  321.739504] Lustre: soaked-OST0000: Connection restored to cccf9223-12de-f413-f873-1abd2bca972e (at 172.16.1.29@o2ib1)
[  321.751469] Lustre: Skipped 3 previous similar messages
[  322.819814] LustreError: 137-5: soaked-OST0004_UUID: not available for connect from 192.168.1.111@o2ib (no target). If you are running an HA pair check that the target is
 mounted on the other server.
[  322.839645] LustreError: Skipped 4 previous similar messages
[  325.343054] Lustre: soaked-OST0004: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-900
[  326.721349] Lustre: soaked-OST0008: Connection restored to b6d310e5-89b5-4823-cd5f-bf69ac139c6a (at 172.16.1.26@o2ib1)
[  326.721352] Lustre: soaked-OST0000: Connection restored to b6d310e5-89b5-4823-cd5f-bf69ac139c6a (at 172.16.1.26@o2ib1)
[  326.721358] Lustre: Skipped 3 previous similar messages
[  330.480118] Lustre: soaked-OST0004: Will be in recovery for at least 2:30, or until 28 clients reconnect
[  330.543092] Lustre: soaked-OST000c: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-900
[  335.247236] Lustre: soaked-OST000c: Connection restored to  (at 172.16.1.22@o2ib1)
[  335.255753] Lustre: Skipped 20 previous similar messages
[  351.605041] Lustre: soaked-OST0000: Connection restored to d33f0a29-feff-6666-e008-19168b75e455 (at 172.16.1.38@o2ib1)
[  351.617080] Lustre: Skipped 65 previous similar messages
[  354.213237] Lustre: soaked-OST0004: Recovery over after 0:23, of 28 clients 28 recovered and 0 were evicted.
[  354.248446] Lustre: soaked-OST0004: deleting orphan objects from 0x0:201673384 to 0x0:201676484
[  354.337426] Lustre: soaked-OST0004: deleting orphan objects from 0x400000402:131183875 to 0x400000402:131188552
[  354.344896] Lustre: soaked-OST0004: deleting orphan objects from 0x400000401:194940686 to 0x400000401:194951539
[  355.092341] Lustre: soaked-OST0008: Recovery over after 0:35, of 28 clients 28 recovered and 0 were evicted.
[  355.112233] Lustre: soaked-OST0008: deleting orphan objects from 0x0:201739882 to 0x0:201746221
[  355.439495] Lustre: soaked-OST0008: deleting orphan objects from 0x500000401:195199465 to 0x500000401:195206220
[  355.439890] Lustre: soaked-OST0008: deleting orphan objects from 0x500000402:131227465 to 0x500000402:131237293
[  430.413849] Lustre: soaked-OST0004: deleting orphan objects from 0x400000400:140327007 to 0x400000400:140331081
[  430.413851] Lustre: soaked-OST0008: deleting orphan objects from 0x500000400:140581975 to 0x500000400:140583563
[  430.493859] Lustre: soaked-OST0008: Connection restored to f9470b7c-9158-fc3c-884e-b494778ee289 (at 172.16.1.31@o2ib1)
[  430.505816] Lustre: Skipped 3 previous similar messages
[  503.499925] LustreError: 34183:0:(ldlm_lockd.c:256:expired_lock_main()) ### lock callback timer expired after 150s: evicting client at 172.16.1.35@o2ib1  ns: filter-soake
d-OST0004_UUID lock: ffff9b2f0e700b40/0x39ef56938122067c lrc: 3/0,0 mode: PW/PW res: [0x400000400:0x85d13f5:0x0].0x0 rrc: 7 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;18446
744073709551615) flags: 0x60000400000020 nid: 172.16.1.35@o2ib1 remote: 0x4d7feaf65d51590b expref: 8 pid: 52997 timeout: 503 lvb_type: 0
[  506.075756] Lustre: soaked-OST0004: Connection restored to 2f6488d4-9684-6185-e787-5c39dc9ffacd (at 172.16.1.35@o2ib1)
[  506.087771] Lustre: Skipped 5 previous similar messages
[  506.322934] Lustre: soaked-OST0000: recovery is timed out, evict stale exports
[  506.331066] Lustre: soaked-OST0000: disconnecting 3 stale clients
[  508.810352] Lustre: soaked-OST0000: Recovery over after 3:09, of 28 clients 25 recovered and 3 were evicted.
[  508.836797] Lustre: soaked-OST0000: deleting orphan objects from 0x0:201827014 to 0x0:201838552
[  509.128301] Lustre: soaked-OST0000: deleting orphan objects from 0x300000402:131294344 to 0x300000402:131298527
[  509.182439] Lustre: soaked-OST0000: deleting orphan objects from 0x300000401:195138776 to 0x300000401:195149783
[  517.586983] Lustre: soaked-OST000c: recovery is timed out, evict stale exports
[  517.595101] Lustre: soaked-OST000c: disconnecting 1 stale clients
[  519.522063] Lustre: soaked-OST000c: Recovery over after 3:09, of 28 clients 27 recovered and 1 was evicted.
[  519.522711] Lustre: soaked-OST000c: deleting orphan objects from 0x600000401:140374471 to 0x600000401:140375679
[  519.544629] general protection fault: 0000 [#1] SMP 
[  519.550190] Modules linked in: osp(OE) ofd(OE) lfsck(OE) ost(OE) mgc(OE) osd_zfs(OE) lquota(OE) fid(OE) fld(OE) ko2iblnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE)[  519.565747] Lustre: soaked-OST000c: deleting orphan objects from 0x0:201738135 to 0x0:201745341
 libcfs(OE) rpcsec_gss_krb5 nfsv4 dns_resolver nfs lockd grace fscache rdma_ucm(OE) ib_ucm(OE) rdma_cm(OE) iw_cm(OE) ib_ipoib(OE) ib_cm(OE) ib_umad(OE) mlx5_ib(OE) mlx5_core(OE) mlxfw(OE) mlx4_en(OE) sb_edac intel_powerclamp coretemp intel_rapl iosf_mbi kvm irqbypass crc32_pclmul ghash_clmulni_intel aesni_intel iTCO_wdt lrw iTCO_vendor_support gf128mul glue_helper ablk_helper cryptd pcspkr dm_round_robin zfs(POE) zunicode(POE) zavl(POE) icp(POE) zcommon(POE) znvpair(POE) spl(OE) joydev ses enclosure ipmi_ssif lpc_ich mei_me sg ioatdma i2c_i801 mei ipmi_si ipmi_devintf ipmi_msghandler wmi dm_multipath dm_mod auth_rpcgss sunrpc ip_tables ext4 mbcache jbd2 sd_mod crc_t10dif crct10dif_generic mlx4_ib(OE) ib_uverbs(OE) ib_core(OE) mgag200 drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops ttm ahci isci igb drm mlx4_core(OE) crct10dif_pclmul libsas libahci crct10dif_common ptp mpt2sas crc32c_intel devlink pps_core libata raid_class dca scsi_transport_sas mlx_compat(OE) drm_panel_orientation_quirks i2c_algo_bit
[  519.677136] CPU: 3 PID: 60300 Comm: tgt_recover_12 Kdump: loaded Tainted: P           OE  ------------   3.10.0-957.21.3.el7_lustre.x86_64 #1
[  519.691306] Hardware name: Intel Corporation S2600GZ ........../S2600GZ, BIOS SE5C600.86B.01.08.0003.022620131521 02/26/2013
[  519.703834] task: ffff9b2f27cb0000 ti: ffff9b2f20b18000 task.ti: ffff9b2f20b18000
[  519.712184] RIP: 0010:[&amp;lt;ffffffffc0fdeefc&amp;gt;]  [&amp;lt;ffffffffc0fdeefc&amp;gt;] keys_fill+0x5c/0x180 [obdclass]
[  519.722046] RSP: 0018:ffff9b2f20b1bad0  EFLAGS: 00010246
[  519.727974] RAX: 5a5a5a5a5a5a5a5a RBX: 0000000000000000 RCX: ffff9b2f20b1bfd8
[  519.735936] RDX: ffff9b2f20b1baf8 RSI: 0000000000000002 RDI: ffffffffc1044080
[  519.743898] RBP: ffff9b2f20b1baf0 R08: 0000000000000000 R09: ffff9b287fc07b00
[  519.751868] R10: ffffffffc14cf797 R11: ffff9b2f12860c00 R12: ffffffffc1044140
[  519.759836] R13: ffff9b2f076a3120 R14: 0000000000000013 R15: ffff9b2f15fe62c8
[  519.767799] FS:  0000000000000000(0000) GS:ffff9b2b2e0c0000(0000) knlGS:0000000000000000
[  519.776834] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[  519.783244] CR2: 00007f81db1e2140 CR3: 000000023c410000 CR4: 00000000000607e0
[  519.791207] Call Trace:
[  519.793962]  [&amp;lt;ffffffffc0fe3961&amp;gt;] lu_context_refill+0x41/0x50 [obdclass]
[  519.801466]  [&amp;lt;ffffffffc0fe39f4&amp;gt;] lu_env_refill+0x24/0x30 [obdclass]
[  519.808579]  [&amp;lt;ffffffffc14cf831&amp;gt;] ofd_lvbo_init+0x2a1/0x7f0 [ofd]
[  519.815426]  [&amp;lt;ffffffffc12aa0fd&amp;gt;] ldlm_server_completion_ast+0x5fd/0x980 [ptlrpc]
[  519.823809]  [&amp;lt;ffffffffc12a9b00&amp;gt;] ? ldlm_server_blocking_ast+0xa40/0xa40 [ptlrpc]
[  519.832181]  [&amp;lt;ffffffffc127c748&amp;gt;] ldlm_work_cp_ast_lock+0xa8/0x1d0 [ptlrpc]
[  519.839981]  [&amp;lt;ffffffffc12c3bf2&amp;gt;] ptlrpc_set_wait+0x72/0x790 [ptlrpc]
[  519.847171]  [&amp;lt;ffffffffa401d75d&amp;gt;] ? kmem_cache_alloc_node_trace+0x11d/0x210
[  519.854957]  [&amp;lt;ffffffffc0fc1a79&amp;gt;] ? lprocfs_counter_add+0xf9/0x160 [obdclass]
[  519.862942]  [&amp;lt;ffffffffc127c6a0&amp;gt;] ? ldlm_work_gl_ast_lock+0x3a0/0x3a0 [ptlrpc]
[  519.871032]  [&amp;lt;ffffffffc12ba472&amp;gt;] ? ptlrpc_prep_set+0xd2/0x280 [ptlrpc]
[  519.878452]  [&amp;lt;ffffffffc1281f25&amp;gt;] ldlm_run_ast_work+0xd5/0x3a0 [ptlrpc]
[  519.885847]  [&amp;lt;ffffffffc12833e1&amp;gt;] __ldlm_reprocess_all+0x101/0x340 [ptlrpc]
[  519.893651]  [&amp;lt;ffffffffc1283986&amp;gt;] ldlm_reprocess_res+0x26/0x30 [ptlrpc]
[  519.901043]  [&amp;lt;ffffffffc0cf4fb0&amp;gt;] cfs_hash_for_each_relax+0x250/0x450 [libcfs]
[  519.909127]  [&amp;lt;ffffffffc1283960&amp;gt;] ? ldlm_lock_mode_downgrade+0x320/0x320 [ptlrpc]
[  519.917499]  [&amp;lt;ffffffffc1283960&amp;gt;] ? ldlm_lock_mode_downgrade+0x320/0x320 [ptlrpc]
[  519.925861]  [&amp;lt;ffffffffc0cf8345&amp;gt;] cfs_hash_for_each_nolock+0x75/0x1c0 [libcfs]
[  519.933945]  [&amp;lt;ffffffffc12839cc&amp;gt;] ldlm_reprocess_recovery_done+0x3c/0x110 [ptlrpc]
[  519.942416]  [&amp;lt;ffffffffc1296211&amp;gt;] target_recovery_thread+0xcd1/0x1160 [ptlrpc]
[  519.950516]  [&amp;lt;ffffffffc1295540&amp;gt;] ? replay_request_or_update.isra.23+0x8c0/0x8c0 [ptlrpc]
[  519.959660]  [&amp;lt;ffffffffa3ec1da1&amp;gt;] kthread+0xd1/0xe0
[  519.965102]  [&amp;lt;ffffffffa3ec1cd0&amp;gt;] ? insert_kthread_work+0x40/0x40
[  519.971919]  [&amp;lt;ffffffffa4575c37&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[  519.979204]  [&amp;lt;ffffffffa3ec1cd0&amp;gt;] ? insert_kthread_work+0x40/0x40
[  519.986002] Code: ab 51 06 00 0f 1f 00 31 db eb 15 0f 1f 40 00 48 83 c3 08 48 81 fb 40 01 00 00 0f 84 9f 00 00 00 49 8b 45 10 4c 8b a3 e0 bf 10 c1 &amp;lt;48&amp;gt; 83 3c 18 00 75 dd 4d 85 e4 74 d8 41 8b 04 24 41 8b 55 00 85 
[  520.007678] RIP  [&amp;lt;ffffffffc0fdeefc&amp;gt;] keys_fill+0x5c/0x180 [obdclass]
[  520.014893]  RSP &amp;lt;ffff9b2f20b1bad0&amp;gt;
[    0.000000] Initializing cgroup subsys cpuset
[    0.000000] Initializing cgroup subsys cpu
[    0.000000] Initializing cgroup subsys cpuacct
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>lustre-b2_12-ib build #31 version=2.12.2_105_gec6b9a6</environment>
        <key id="56659">LU-12663</key>
            <summary>SOAK: OSS hit general protection fault: 0000 [#1] SMP</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="sarah">Sarah Liu</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Tue, 13 Aug 2019 18:27:29 +0000</created>
                <updated>Wed, 14 Aug 2019 15:27:02 +0000</updated>
                                            <version>Lustre 2.12.3</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="253026" author="jamesanunez" created="Wed, 14 Aug 2019 15:27:02 +0000"  >&lt;p&gt;Here is what was loaded on soak and issues seen prior to this crash:&lt;br/&gt;
2019-8-3:  Soak started with  lustre-b2_12-ib #31&lt;br/&gt;
2019-8-7:  OSS hit &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9845&quot; title=&quot;ost-pools test_22 hangs with &#8216;WARNING: Pool &amp;#39;lustre-mdt1&amp;#39; has encountered an uncorrectable I/O failure and has been suspended.&#8217;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9845&quot;&gt;&lt;del&gt;LU-9845&lt;/del&gt;&lt;/a&gt; which caused the whole testing hung, stop and restart soak&lt;br/&gt;
2019-8-12: OSS hit &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-12663&quot; title=&quot;SOAK: OSS hit general protection fault: 0000 [#1] SMP&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-12663&quot;&gt;LU-12663&lt;/a&gt;, stop running soak&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i00l7j:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>