<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 03:23:11 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-16009] BUG: unable to handle kernel paging request</title>
                <link>https://jira.whamcloud.com/browse/LU-16009</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Repeated kernel crash (BUG: unable to handle kernel paging reques) with zfs and without zfs. I have multiple crash dump I can upload.&lt;/p&gt;

&lt;p&gt;Here are some of the stack traces&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
127.0.0.1-2022-07-12-12\:07\:42/vmcore
crash&amp;gt; bt 21199
PID: 21199 &#160;TASK: ffff90997c843000 &#160;CPU: 28 &#160;COMMAND: &lt;span class=&quot;code-quote&quot;&gt;&quot;fio&quot;&lt;/span&gt;
&#160;#0 [ffff9fe0e357fb88] machine_kexec at fffffffface641ce
&#160;#1 [ffff9fe0e357fbe0] __crash_kexec at ffffffffacf9df1d
&#160;#2 [ffff9fe0e357fca8] crash_kexec at ffffffffacf9ee0d
&#160;#3 [ffff9fe0e357fcc0] oops_end at fffffffface2613d
&#160;#4 [ffff9fe0e357fce0] no_context at fffffffface7562f
&#160;#5 [ffff9fe0e357fd38] __bad_area_nosemaphore at fffffffface7598c
&#160;#6 [ffff9fe0e357fd80] do_page_fault at fffffffface76267
&#160;#7 [ffff9fe0e357fdb0] page_fault at ffffffffad80111e
&#160; &#160; [exception RIP: __kmalloc_node+418]
&#160; &#160; RIP: ffffffffad0fd382 &#160;RSP: ffff9fe0e357fe68 &#160;RFLAGS: 00010246
&#160; &#160; RAX: ffff909c032f3e78 &#160;RBX: ffff9fe0e357fed8 &#160;RCX: 0000000000000000
&#160; &#160; RDX: 0000000000623f36 &#160;RSI: 00000000006000c0 &#160;RDI: 000000000002f040
&#160; &#160; RBP: 00000000006000c0 &#160; R8: ffff909bbd32f040 &#160; R9: ffff908ff2efae40
&#160; &#160; R10: ffff905c80004400 &#160;R11: ffffffffae45a410 &#160;R12: 0000000000000008
&#160; &#160; R13: 00000000ffffffff &#160;R14: ffff905c80004400 &#160;R15: ffffffffad75537b
&#160; &#160; ORIG_RAX: ffffffffffffffff &#160;CS: 0010 &#160;SS: 0018
&#160;#8 [ffff9fe0e357feb8] alloc_cpumask_var_node at ffffffffad75537b
&#160;#9 [ffff9fe0e357fec8] sched_setaffinity at ffffffffacf1eb4d
#10 [ffff9fe0e357ff08] __x64_sys_sched_setaffinity at ffffffffacf1edef
#11 [ffff9fe0e357ff38] do_syscall_64 at fffffffface042bb
#12 [ffff9fe0e357ff50] entry_SYSCALL_64_after_hwframe at ffffffffad8000ad
&#160; &#160; RIP: 00007ff5a7c8e81d &#160;RSP: 00007ff546cd46f8 &#160;RFLAGS: 00000203
&#160; &#160; RAX: ffffffffffffffda &#160;RBX: 000055a5798b4170 &#160;RCX: 00007ff5a7c8e81d
&#160; &#160; RDX: 00007ff546cd4700 &#160;RSI: 0000000000000080 &#160;RDI: 00000000000052cf
&#160; &#160; RBP: 00007ff546cd4700 &#160; R8: 00007ff546cd7700 &#160; R9: 0000000000000100
&#160; &#160; R10: 00007ff546cd7700 &#160;R11: 0000000000000203 &#160;R12: 00007ffcf7d4db5e
&#160; &#160; R13: 00007ffcf7d4db5f &#160;R14: 0000000000000000 &#160;R15: 00007ff546cd4880
&#160; &#160; ORIG_RAX: 00000000000000cb &#160;CS: 0033 &#160;SS: 002b



127.0.0.1-2022-07-12-23\:43\:02/vmcore
crash&amp;gt; bt 2
PID: 2 &#160; &#160; &#160;TASK: ffff9012c36b4800 &#160;CPU: 18 &#160;COMMAND: &lt;span class=&quot;code-quote&quot;&gt;&quot;kthreadd&quot;&lt;/span&gt;
&#160;#0 [ffff9e5d0c4bb708] machine_kexec at ffffffff9e8641ce
&#160;#1 [ffff9e5d0c4bb760] __crash_kexec at ffffffff9e99df1d
&#160;#2 [ffff9e5d0c4bb828] crash_kexec at ffffffff9e99ee0d
&#160;#3 [ffff9e5d0c4bb840] oops_end at ffffffff9e82613d
&#160;#4 [ffff9e5d0c4bb860] no_context at ffffffff9e87562f
&#160;#5 [ffff9e5d0c4bb8b8] __bad_area_nosemaphore at ffffffff9e87598c
&#160;#6 [ffff9e5d0c4bb900] do_page_fault at ffffffff9e876267
&#160;#7 [ffff9e5d0c4bb930] page_fault at ffffffff9f20111e
&#160; &#160; [exception RIP: deactivate_slab+158]
&#160; &#160; RIP: ffffffff9eafad0e &#160;RSP: ffff9e5d0c4bb9e0 &#160;RFLAGS: 00010086
&#160; &#160; RAX: 0000000000000010 &#160;RBX: ffff904244a66908 &#160;RCX: 00000000000000cc
&#160; &#160; RDX: ffff904244a66248 &#160;RSI: ffff905244a66e00 &#160;RDI: ffff904244a66000
&#160; &#160; RBP: ffff9e5d0c4bbab0 &#160; R8: 0000000000000001 &#160; R9: 00000000000000cc
&#160; &#160; R10: 0000000000000000 &#160;R11: 0000000000000000 &#160;R12: ffffd01702129980
&#160; &#160; R13: ffff9021fffaf040 &#160;R14: ffff905244a66e00 &#160;R15: ffff9012c0004400
&#160; &#160; ORIG_RAX: ffffffffffffffff &#160;CS: 0010 &#160;SS: 0018
&#160;#8 [ffff9e5d0c4bbab8] ___slab_alloc at ffffffff9eafbd31
&#160;#9 [ffff9e5d0c4bbaf0] memcg_alloc_page_obj_cgroups at ffffffff9eb18dea
#10 [ffff9e5d0c4bbb30] perf_output_end at ffffffff9ea6bf9a
#11 [ffff9e5d0c4bbb38] perf_event_task_output at ffffffff9ea5b92c
#12 [ffff9e5d0c4bbcf8] copy_process at ffffffff9e8e96b9
#13 [ffff9e5d0c4bbd70] copy_process at ffffffff9e8e96b9
#14 [ffff9e5d0c4bbdd0] __switch_to_asm at ffffffff9f2001e5
#15 [ffff9e5d0c4bbe30] _do_fork at ffffffff9e8eb09f
#16 [ffff9e5d0c4bbea8] kernel_thread at ffffffff9e8eb415
#17 [ffff9e5d0c4bbeb0] kthreadd at ffffffff9e910c04
#18 [ffff9e5d0c4bbf50] ret_from_fork at ffffffff9f200242



127.0.0.1-2022-07-12-23:31:06/vmcore
crash&amp;gt; bt 2
PID: 2 &#160; &#160; &#160;TASK: ffff8ce78317c800 &#160;CPU: 18 &#160;COMMAND: &lt;span class=&quot;code-quote&quot;&gt;&quot;kthreadd&quot;&lt;/span&gt;
&#160;#0 [ffff9a144c4bb708] machine_kexec at ffffffff8fa641ce
&#160;#1 [ffff9a144c4bb760] __crash_kexec at ffffffff8fb9df1d
&#160;#2 [ffff9a144c4bb828] crash_kexec at ffffffff8fb9ee0d
&#160;#3 [ffff9a144c4bb840] oops_end at ffffffff8fa2613d
&#160;#4 [ffff9a144c4bb860] no_context at ffffffff8fa7562f
&#160;#5 [ffff9a144c4bb8b8] __bad_area_nosemaphore at ffffffff8fa7598c
&#160;#6 [ffff9a144c4bb900] do_page_fault at ffffffff8fa76267
&#160;#7 [ffff9a144c4bb930] page_fault at ffffffff9040111e
&#160; &#160; [exception RIP: deactivate_slab+158]
&#160; &#160; RIP: ffffffff8fcfad0e &#160;RSP: ffff9a144c4bb9e0 &#160;RFLAGS: 00010082
&#160; &#160; RAX: 0000000000000010 &#160;RBX: ffff8ce78436eff8 &#160;RCX: 00000000000000ae
&#160; &#160; RDX: ffff8ce78436e248 &#160;RSI: ffff8cf78436e038 &#160;RDI: ffff8ce78436e000
&#160; &#160; RBP: ffff9a144c4bbab0 &#160; R8: 0000000000000001 &#160; R9: 00000000000000ae
&#160; &#160; R10: 0000000000000000 &#160;R11: 0000000000000000 &#160;R12: ffffee500210db80
&#160; &#160; R13: ffff8cc73ffaf040 &#160;R14: ffff8cf78436e038 &#160;R15: ffff8cb800004400
&#160; &#160; ORIG_RAX: ffffffffffffffff &#160;CS: 0010 &#160;SS: 0018
&#160;#8 [ffff9a144c4bbab8] ___slab_alloc at ffffffff8fcfbd31
&#160;#9 [ffff9a144c4bbaf0] memcg_alloc_page_obj_cgroups at ffffffff8fd18dea
#10 [ffff9a144c4bbb30] perf_output_end at ffffffff8fc6bf9a
#11 [ffff9a144c4bbb38] perf_event_task_output at ffffffff8fc5b92c
#12 [ffff9a144c4bbcf8] copy_process at ffffffff8fae96b9
#13 [ffff9a144c4bbd70] copy_process at ffffffff8fae96b9
#14 [ffff9a144c4bbdd0] __switch_to_asm at ffffffff904001e5
#15 [ffff9a144c4bbe30] _do_fork at ffffffff8faeb09f
#16 [ffff9a144c4bbea8] kernel_thread at ffffffff8faeb415
#17 [ffff9a144c4bbeb0] kthreadd at ffffffff8fb10c04
#18 [ffff9a144c4bbf50] ret_from_fork at ffffffff90400242
 

127.0.0.1-2022-07-13-10_12_31/vmcore
crash&amp;gt; bt 106937
PID: 106937 &#160;TASK: ffff9b71c54c9800 &#160;CPU: 27 &#160;COMMAND: &lt;span class=&quot;code-quote&quot;&gt;&quot;ll_ost_io02_016&quot;&lt;/span&gt;
&#160;#0 [ffffad02a1ba7488] machine_kexec at ffffffff93c641ce
&#160;#1 [ffffad02a1ba74e0] __crash_kexec at ffffffff93d9df1d
&#160;#2 [ffffad02a1ba75a8] crash_kexec at ffffffff93d9ee0d
&#160;#3 [ffffad02a1ba75c0] oops_end at ffffffff93c2613d
&#160;#4 [ffffad02a1ba75e0] no_context at ffffffff93c7562f
&#160;#5 [ffffad02a1ba7638] __bad_area_nosemaphore at ffffffff93c7598c
&#160;#6 [ffffad02a1ba7680] do_page_fault at ffffffff93c76267
&#160;#7 [ffffad02a1ba76b0] page_fault at ffffffff9460111e
&#160; &#160; [exception RIP: __kmalloc_node+418]
&#160; &#160; RIP: ffffffff93efd382 &#160;RSP: ffffad02a1ba7768 &#160;RFLAGS: 00010246
&#160; &#160; RAX: 0000001000000000 &#160;RBX: 000000000060c2c0 &#160;RCX: 0000000000000000
&#160; &#160; RDX: 000000000001eca3 &#160;RSI: 000000000060c2c0 &#160;RDI: 000000000002f040
&#160; &#160; RBP: 000000000060c2c0 &#160; R8: ffff9b917fdef040 &#160; R9: ffffad02a1ba786c
&#160; &#160; R10: ffff9b6240004400 &#160;R11: 0000000000000000 &#160;R12: 0000000000000008
&#160; &#160; R13: 00000000ffffffff &#160;R14: ffff9b6240004400 &#160;R15: ffffffffc01f2563
&#160; &#160; ORIG_RAX: ffffffffffffffff &#160;CS: 0010 &#160;SS: 0018
&#160;#8 [ffffad02a1ba77b8] spl_kmem_zalloc at ffffffffc01f2563 [spl]
&#160;#9 [ffffad02a1ba77e8] dmu_buf_hold_array_by_dnode at ffffffffc03e9bbd [zfs]
#10 [ffffad02a1ba7850] dmu_write_by_dnode at ffffffffc03ea2ca [zfs]
#11 [ffffad02a1ba78a0] osd_write at ffffffffc255ebe8 [osd_zfs]
#12 [ffffad02a1ba78e0] dt_record_write at ffffffffc1b684b2 [obdclass]
#13 [ffffad02a1ba78f0] tgt_server_data_write at ffffffffc193e460 [ptlrpc]
#14 [ffffad02a1ba7920] tgt_txn_stop_cb at ffffffffc19469c0 [ptlrpc]
#15 [ffffad02a1ba7988] dt_txn_hook_stop at ffffffffc1b6b563 [obdclass]
#16 [ffffad02a1ba79b0] osd_trans_stop at ffffffffc25509f6 [osd_zfs]
#17 [ffffad02a1ba7a00] ofd_commitrw_write at ffffffffc203a4d3 [ofd]
#18 [ffffad02a1ba7aa0] ofd_commitrw at ffffffffc203f831 [ofd]
#19 [ffffad02a1ba7b60] obd_commitrw at ffffffffc194b47c [ptlrpc]
#20 [ffffad02a1ba7bd0] tgt_brw_write at ffffffffc1953a80 [ptlrpc]
#21 [ffffad02a1ba7d50] tgt_request_handle at ffffffffc1955053 [ptlrpc]
#22 [ffffad02a1ba7dd0] ptlrpc_server_handle_request at ffffffffc1901983 [ptlrpc]
#23 [ffffad02a1ba7e38] ptlrpc_main at ffffffffc1903486 [ptlrpc]
#24 [ffffad02a1ba7f10] kthread at ffffffff93d0f726
#25 [ffffad02a1ba7f50] ret_from_fork at ffffffff94600242&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>4.18.0-348.2.1.el8_lustre.x86_64 redhat8.5</environment>
        <key id="71132">LU-16009</key>
            <summary>BUG: unable to handle kernel paging request</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="yujian">Jian Yu</assignee>
                                    <reporter username="mhanafi">Mahmoud Hanafi</reporter>
                        <labels>
                    </labels>
                <created>Wed, 13 Jul 2022 17:34:46 +0000</created>
                <updated>Tue, 23 Jan 2024 08:05:30 +0000</updated>
                                            <version>Lustre 2.15.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="340283" author="mhanafi" created="Wed, 13 Jul 2022 17:55:13 +0000"  >&lt;p&gt;Where should I upload the crash dumps.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="340284" author="JIRAUSER17312" created="Wed, 13 Jul 2022 18:03:13 +0000"  >&lt;p&gt;Hey,&lt;/p&gt;

&lt;p&gt;Can you detail your process for building lustre with ZFS? In &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16006&quot; title=&quot;can&amp;#39;t install kmod-lustre-osd-zfs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16006&quot;&gt;LU-16006&lt;/a&gt; there seems to be some suggestion that it was not successful?&lt;/p&gt;

&lt;p&gt;Also, can you upload the vmcore to our FTP site: ftp.whamcloud.com (anonymous / your email) in the /uploads directory?&lt;/p&gt;

&lt;p&gt;Thank you!&lt;/p&gt;</comment>
                            <comment id="340309" author="mhanafi" created="Wed, 13 Jul 2022 20:27:20 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16006&quot; title=&quot;can&amp;#39;t install kmod-lustre-osd-zfs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16006&quot;&gt;LU-16006&lt;/a&gt; is very odd. Even though rpm install complains about the missing ksym, if you force install the rpm. The modules load without errors. But this crash was happening&#160;before switching to zfs.&#160;&lt;/p&gt;</comment>
                            <comment id="340318" author="mhanafi" created="Wed, 13 Jul 2022 20:57:47 +0000"  >&lt;p&gt;Uploaded the following crash dumps to the ftp site.&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16009&quot; title=&quot;BUG: unable to handle kernel paging request&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16009&quot;&gt;LU-16009&lt;/a&gt;_127.0.0.1-2022-07-10-00_00_41.tgz&lt;br/&gt;
&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16009&quot; title=&quot;BUG: unable to handle kernel paging request&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16009&quot;&gt;LU-16009&lt;/a&gt;_127.0.0.1-2022-07-12-12_07_42.tgz&lt;br/&gt;
&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16009&quot; title=&quot;BUG: unable to handle kernel paging request&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16009&quot;&gt;LU-16009&lt;/a&gt;_127.0.0.1-2022-07-12-23_00_45.tgz&lt;br/&gt;
&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16009&quot; title=&quot;BUG: unable to handle kernel paging request&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16009&quot;&gt;LU-16009&lt;/a&gt;_127.0.0.1-2022-07-12-23_31_06.tgz&lt;br/&gt;
&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16009&quot; title=&quot;BUG: unable to handle kernel paging request&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16009&quot;&gt;LU-16009&lt;/a&gt;_127.0.0.1-2022-07-12-23_43_02.tgz&lt;br/&gt;
&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-16009&quot; title=&quot;BUG: unable to handle kernel paging request&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-16009&quot;&gt;LU-16009&lt;/a&gt;_127.0.0.1-2022-07-13-10_12_31.tgz&lt;/p&gt;</comment>
                            <comment id="340399" author="pjones" created="Thu, 14 Jul 2022 14:18:57 +0000"  >&lt;p&gt;Jian&lt;/p&gt;

&lt;p&gt;Could you please assist here?&lt;/p&gt;

&lt;p&gt;thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="340468" author="yujian" created="Thu, 14 Jul 2022 22:07:33 +0000"  >&lt;p&gt;Hi Mahmoud,&lt;br/&gt;
Could you please list the commands/operations you performed to trigger the kernel crash?&lt;br/&gt;
While testing Lustre 2.15.0 release, the fio test passed on RHEL 8.5 with kernel 4.18.0-348.2.1.el8_lustre.x86_64:&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/sub_tests/5604d12c-25d6-4a56-a60d-7a399f9abda0&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/sub_tests/5604d12c-25d6-4a56-a60d-7a399f9abda0&lt;/a&gt; (ldiskfs)&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/sub_tests/5bdcac4a-e8f3-45f5-a481-dad2bf747d78&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/sub_tests/5bdcac4a-e8f3-45f5-a481-dad2bf747d78&lt;/a&gt; (ldiskfs)&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/sub_tests/67c9fa6b-8abe-4cae-aad7-7f09a3017a7a&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/sub_tests/67c9fa6b-8abe-4cae-aad7-7f09a3017a7a&lt;/a&gt; (zfs-2.0.7)&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/sub_tests/ba9745bb-1809-4803-a338-d1d0afc67569&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/sub_tests/ba9745bb-1809-4803-a338-d1d0afc67569&lt;/a&gt; (zfs-2.0.7) &lt;/p&gt;</comment>
                            <comment id="340617" author="mhanafi" created="Fri, 15 Jul 2022 20:26:55 +0000"  >&lt;p&gt;Our setup is a Single OSS attached to external Netapp via ISER. It also has 5 on board NVME devices. The NVME devices are raid using MD or zpool. &#160;&#160;The server and client both run on the same server.&lt;/p&gt;

&lt;p&gt;The attached script will trigger a crash with-in 2 or 3 iterations. I am trying to setup a second server with rhel7 and lustre2.12.9 to test also.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;nobr&quot;&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/attachment/44430/44430_fio_loop.sh&quot; title=&quot;fio_loop.sh attached to LU-16009&quot;&gt;fio_loop.sh&lt;sup&gt;&lt;img class=&quot;rendericon&quot; src=&quot;https://jira.whamcloud.com/images/icons/link_attachment_7.gif&quot; height=&quot;7&quot; width=&quot;7&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/sup&gt;&lt;/a&gt;&lt;/span&gt;&lt;/p&gt;</comment>
                            <comment id="340949" author="yujian" created="Wed, 20 Jul 2022 08:03:09 +0000"  >&lt;p&gt;Hi &lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=mhanafi&quot; class=&quot;user-hover&quot; rel=&quot;mhanafi&quot;&gt;mhanafi&lt;/a&gt;,&lt;br/&gt;
I set up a single-node test environment and ran fio_loop.sh for over 12 iterations. The node did not crash.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;# lsb_release -d
Description:    CentOS Linux release 8.5.2111

# uname -r
4.18.0-348.2.1.el8_lustre.x86_64

# rpm -q lustre
lustre-2.15.0-1.el8.x86_64

# lfs df -h
UUID                       bytes        Used   Available Use% Mounted on
lustre-MDT0000_UUID         4.2G        1.9M        3.8G   1% /mnt/lustre[MDT:0]
lustre-OST0000_UUID        55.3G        1.2M       52.4G   1% /mnt/lustre[OST:0]

filesystem_summary:        55.3G        1.2M       52.4G   1% /mnt/lustre

# FS=mnt/lustre bash -x fio_loop.sh |&amp;amp; tee fio_loop.log

# ls /mnt/lustre/final/
20220720_06_46  20220720_06_47  20220720_06_48  20220720_06_49  20220720_06_50  20220720_06_52  20220720_06_53  20220720_06_54  20220720_06_56  20220720_06_57  20220720_06_58  20220720_06_59  20220720_07_00  20220720_07_01
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; 

&lt;p&gt;From the fio crash backtrace in this ticket&apos;s description, we can see:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[exception RIP: __kmalloc_node+418]
RIP: ffffffffad0fd382 ......
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;The crash was triggered while allocating memory.&lt;/p&gt;</comment>
                            <comment id="343180" author="mhanafi" created="Wed, 10 Aug 2022 03:28:21 +0000"  >&lt;p&gt;This may be * Kernel bug on mm/slub.c:314 (BZ#2102251)&#160;&lt;/p&gt;

&lt;p&gt;I can&apos;t view the details of the bug.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="400734" author="yujian" created="Tue, 23 Jan 2024 08:05:30 +0000"  >&lt;p&gt;Hi &lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=mhanafi&quot; class=&quot;user-hover&quot; rel=&quot;mhanafi&quot;&gt;mhanafi&lt;/a&gt;,&lt;br/&gt;
Did you hit the crash with newer kernel? Can we close this ticket?&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="67415">LU-15308</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="44430" name="fio_loop.sh" size="1188" author="mhanafi" created="Fri, 15 Jul 2022 20:20:50 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i02uhz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10021"><![CDATA[2]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>