<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 03:08:54 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-14341] hard LOCKUP lustre servers with kernel-3.10.0-1160.11.1</title>
                <link>https://jira.whamcloud.com/browse/LU-14341</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;hard LOCKUP and panic.&#160; Most frequently observed on OSTs after mdtest completes or after OST mount , a few seconds after&#160;&quot;deleting orphan objects from&quot; console log messages.&#160;&lt;/p&gt;

&lt;p&gt;This appeared to be due to kernel timer behavior changes introduced between kernel-3.10.0-1160.6.1 and kernel-3.10.0-1160.11.1.&#160;&lt;/p&gt;

&lt;p&gt;Fix in progress.&#160; See&#160;&lt;a href=&quot;https://bugzilla.redhat.com/show_bug.cgi?id=1914011&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://bugzilla.redhat.com/show_bug.cgi?id=1914011&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;For brevity, only the bottoms of the stacks, are listed below.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Kernel panic - not syncing: Hard LOCKUP
CPU: 14 PID: 0 Comm: swapper/14 Kdump: loaded Tainted: P        W  OE  ------------   3.10.0-1160.11.1.1chaos.ch6.x86_64 #1
...
Call Trace:
 &amp;lt;NMI&amp;gt;  [&amp;lt;ffffffffa47ae072&amp;gt;] dump_stack+0x19/0x1b
 [&amp;lt;ffffffffa47a71e7&amp;gt;] panic+0xe8/0x21f
...
 [&amp;lt;ffffffffa40b1edc&amp;gt;] ? run_timer_softirq+0xbc/0x370
 &amp;lt;EOE&amp;gt;  &amp;lt;IRQ&amp;gt;  [&amp;lt;ffffffffa40a82fd&amp;gt;] __do_softirq+0xfd/0x2c0
 [&amp;lt;ffffffffa47c56ec&amp;gt;] call_softirq+0x1c/0x30
 [&amp;lt;ffffffffa4030995&amp;gt;] do_softirq+0x65/0xa0
 [&amp;lt;ffffffffa40a86d5&amp;gt;] irq_exit+0x105/0x110
 [&amp;lt;ffffffffa47c6c88&amp;gt;] smp_apic_timer_interrupt+0x48/0x60
 [&amp;lt;ffffffffa47c31ba&amp;gt;] apic_timer_interrupt+0x16a/0x170
 &amp;lt;EOI&amp;gt;  [&amp;lt;ffffffffa40b3113&amp;gt;] ? get_next_timer_interrupt+0x103/0x270
 [&amp;lt;ffffffffa45eace7&amp;gt;] ? cpuidle_enter_state+0x57/0xd0
 [&amp;lt;ffffffffa45eae3e&amp;gt;] cpuidle_idle_call+0xde/0x270
 [&amp;lt;ffffffffa403919e&amp;gt;] arch_cpu_idle+0xe/0xc0
 [&amp;lt;ffffffffa410856a&amp;gt;] cpu_startup_entry+0x14a/0x1e0
 [&amp;lt;ffffffffa405cbb7&amp;gt;] start_secondary+0x207/0x280
 [&amp;lt;ffffffffa40000d5&amp;gt;] start_cpu+0x5/0x14
 
Another one that we see is quite similar to what is happening on cpu 21 in the original BZ.

Call Trace:
 &amp;lt;NMI&amp;gt;  [&amp;lt;ffffffff85fae072&amp;gt;] dump_stack+0x19/0x1b
 [&amp;lt;ffffffff85fa71e7&amp;gt;] panic+0xe8/0x21f
...
 [&amp;lt;ffffffff8591f4e8&amp;gt;] ? native_queued_spin_lock_slowpath+0x158/0x200
 &amp;lt;EOE&amp;gt;  [&amp;lt;ffffffff85fa7dd2&amp;gt;] queued_spin_lock_slowpath+0xb/0xf
 [&amp;lt;ffffffff85fb7197&amp;gt;] _raw_spin_lock_irqsave+0x47/0x50
 [&amp;lt;ffffffff858b1b8b&amp;gt;] lock_timer_base.isra.38+0x2b/0x50
 [&amp;lt;ffffffff858b244f&amp;gt;] try_to_del_timer_sync+0x2f/0x90
 [&amp;lt;ffffffff858b2502&amp;gt;] del_timer_sync+0x52/0x60
 [&amp;lt;ffffffff85fb1920&amp;gt;] schedule_timeout+0x180/0x320
 [&amp;lt;ffffffff858b1870&amp;gt;] ? requeue_timers+0x1f0/0x1f0 &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>kernel-3.10.0-1160.11.1&lt;br/&gt;
lustre-2.12.6_2.llnl-1.ch6.x86_64</environment>
        <key id="62397">LU-14341</key>
            <summary>hard LOCKUP lustre servers with kernel-3.10.0-1160.11.1</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="pjones">Peter Jones</assignee>
                                    <reporter username="ofaaland">Olaf Faaland</reporter>
                        <labels>
                            <label>llnl</label>
                    </labels>
                <created>Wed, 20 Jan 2021 00:45:23 +0000</created>
                <updated>Sun, 31 Dec 2023 18:47:31 +0000</updated>
                            <resolved>Wed, 5 May 2021 21:12:32 +0000</resolved>
                                                                        <due></due>
                            <votes>1</votes>
                                    <watches>17</watches>
                                                                            <comments>
                            <comment id="289878" author="ofaaland" created="Wed, 20 Jan 2021 00:47:02 +0000"  >&lt;p&gt;For my reference, my local ticket is&#160;TOSS-4985&lt;/p&gt;</comment>
                            <comment id="289879" author="ofaaland" created="Wed, 20 Jan 2021 00:50:18 +0000"  >&lt;p&gt;Redhat has a fix which we tested successfully on our system.&#160; The patch is moving through their system.&#160; See their bugzilla for details.&lt;/p&gt;</comment>
                            <comment id="289881" author="sthiell" created="Wed, 20 Jan 2021 01:02:05 +0000"  >&lt;p&gt;We are seeing the same hard lockups on new servers (Dell R6525) with Lustre 2.12.6 on CentOS 7.9 using kernel&#160;3.10.0-1160.11.1.el7 + Lustre patches for rhel7.9. It&apos;s very easy to trigger, just one client doing some moderate I/Os. I was just about to report it too! Olaf was faster &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;

&lt;p&gt;&#160;Example of crash dump:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; ps | grep &quot;&amp;gt;&quot;
...
&amp;gt;&#160; 4588&#160; &#160; &#160; 2&#160; 51&#160; ffff9c8d1dbed280&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost01_000]
&amp;gt;&#160; 4589&#160; &#160; &#160; 2&#160; 28&#160; ffff9c8d1dbee300&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost01_001]
&amp;gt;&#160; 4590&#160; &#160; &#160; 2&#160; 60&#160; ffff9c8d2dfc8000&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost01_002]
&amp;gt;&#160; 4651&#160; &#160; &#160; 2&#160; 50&#160; ffff9c8d15ea5280&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost01_005]
&amp;gt;&#160; 4655&#160; &#160; &#160; 2&#160; 18&#160; ffff9c73e47f8000&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost01_008]
&amp;gt;&#160; 4656&#160; &#160; &#160; 2&#160; 17&#160; ffff9c73e47fd280&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost01_009]
&amp;gt;&#160; 4663&#160; &#160; &#160; 2&#160; 16&#160; ffff9c8d349ab180&#160; RU &#160; 0.0 &#160; &#160; &#160; 0&#160; &#160; &#160; 0&#160; [ll_ost_io01_006]&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;I don&apos;t have access to this RedHat bz, but I can also see &quot;lock_timer_base&quot; in one of the backtrace (for 4588 ll_ost01_000):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; bt 4588
PID: 4588   TASK: ffff9c8d1dbed280  CPU: 51  COMMAND: &quot;ll_ost01_000&quot;
 #0 [ffff9cb1ff0c8b58] crash_nmi_callback at ffffffffad85855b
 #1 [ffff9cb1ff0c8b68] nmi_panic_self_stop at ffffffffad858c86
 #2 [ffff9cb1ff0c8b80] nmi_panic at ffffffffad89aefb
 #3 [ffff9cb1ff0c8b90] watchdog_overflow_callback at ffffffffad94ef81
 #4 [ffff9cb1ff0c8ba8] __perf_event_overflow at ffffffffad9a8b07
 #5 [ffff9cb1ff0c8be0] perf_event_overflow at ffffffffad9b2304
 #6 [ffff9cb1ff0c8bf0] x86_pmu_handle_irq at ffffffffad805595
 #7 [ffff9cb1ff0c8e20] amd_pmu_handle_irq at ffffffffad8066f5
 #8 [ffff9cb1ff0c8e38] perf_event_nmi_handler at ffffffffadf8b031
 #9 [ffff9cb1ff0c8e58] nmi_handle at ffffffffadf8c93c
#10 [ffff9cb1ff0c8eb0] do_nmi at ffffffffadf8cb5d
#11 [ffff9cb1ff0c8ef0] end_repeat_nmi at ffffffffadf8bd9c
    [exception RIP: native_queued_spin_lock_slowpath+464]
    RIP: ffffffffad917920  RSP: ffff9c8d2dfbf6e8  RFLAGS: 00000002
    RAX: 0000000000810101  RBX: 0000000000000282  RCX: 0000000000000001
    RDX: 0000000000000101  RSI: 0000000000000001  RDI: ffff9cb1ff013940
    RBP: ffff9c8d2dfbf6e8   R8: 0000000000000101   R9: 00000000000000c0
    R10: 0000000000000000  R11: 0000000000000000  R12: ffff9cb1ff013940
    R13: ffffffffc158a138  R14: ffff9c8d2dfbf758  R15: 0000000000000000
    ORIG_RAX: ffffffffffffffff  CS: 0010  SS: 0018
--- &amp;lt;NMI exception stack&amp;gt; ---
#12 [ffff9c8d2dfbf6e8] native_queued_spin_lock_slowpath at ffffffffad917920
#13 [ffff9c8d2dfbf6f0] queued_spin_lock_slowpath at ffffffffadf7bfe5
#14 [ffff9c8d2dfbf700] _raw_spin_lock_irqsave at ffffffffadf8a777
#15 [ffff9c8d2dfbf718] lock_timer_base at ffffffffad8adbeb
#16 [ffff9c8d2dfbf748] mod_timer at ffffffffad8aed14
#17 [ffff9c8d2dfbf798] __ldlm_add_waiting_lock at ffffffffc146fac7 [ptlrpc]
#18 [ffff9c8d2dfbf7c0] ldlm_add_waiting_lock at ffffffffc146ffb3 [ptlrpc]
#19 [ffff9c8d2dfbf7e8] ldlm_server_blocking_ast at ffffffffc147533c [ptlrpc]
#20 [ffff9c8d2dfbf828] tgt_blocking_ast at ffffffffc15018a9 [ptlrpc]
#21 [ffff9c8d2dfbf8e8] ldlm_work_bl_ast_lock at ffffffffc144a26c [ptlrpc]
#22 [ffff9c8d2dfbf970] ptlrpc_set_wait at ffffffffc1490f42 [ptlrpc]
#23 [ffff9c8d2dfbfa18] ldlm_run_ast_work at ffffffffc144e145 [ptlrpc]
#24 [ffff9c8d2dfbfa48] ldlm_handle_conflict_lock at ffffffffc144ef50 [ptlrpc]
#25 [ffff9c8d2dfbfa80] ldlm_lock_enqueue at ffffffffc144f5e0 [ptlrpc]
#26 [ffff9c8d2dfbfaf8] ldlm_cli_enqueue_local at ffffffffc146a227 [ptlrpc]
#27 [ffff9c8d2dfbfb98] ofd_destroy_by_fid at ffffffffc1bafb21 [ofd]
#28 [ffff9c8d2dfbfc70] ofd_destroy_hdl at ffffffffc1ba6797 [ofd]
#29 [ffff9c8d2dfbfcd0] tgt_request_handle at ffffffffc1506f1a [ptlrpc]
#30 [ffff9c8d2dfbfd58] ptlrpc_server_handle_request at ffffffffc14ab88b [ptlrpc]
#31 [ffff9c8d2dfbfdf8] ptlrpc_main at ffffffffc14af1f4 [ptlrpc]
#32 [ffff9c8d2dfbfec8] kthread at ffffffffad8c5e71
#33 [ffff9c8d2dfbff50] ret_from_fork_nospec_begin at ffffffffadf94ddd
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="289882" author="ofaaland" created="Wed, 20 Jan 2021 01:02:25 +0000"  >&lt;p&gt;Related to&#160;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-14226&quot; title=&quot;kernel update [RHEL7.9 3.10.0-1160.11.1.el7]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-14226&quot;&gt;&lt;del&gt;LU-14226&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="289883" author="pjones" created="Wed, 20 Jan 2021 01:02:55 +0000"  >&lt;p&gt;Olaf&lt;/p&gt;

&lt;p&gt;Do I read this correctly that you are mostly just opening this ticket for the benefit of others and not expecting action on our part at this time?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="289966" author="ofaaland" created="Wed, 20 Jan 2021 19:05:39 +0000"  >&lt;p&gt;Peter,&lt;/p&gt;

&lt;p&gt;Sorry I didn&apos;t see your comment earlier.&#160; Yes, this ticket is for the benefit of others.&lt;/p&gt;</comment>
                            <comment id="291403" author="sthiell" created="Mon, 8 Feb 2021 04:11:27 +0000"  >&lt;p&gt;FYI I tried Lustre (server) with kernel-3.10.0-1160.15.2.el7 and immediately got the same hard lockups, so it looks like this issue hasn&apos;t been fixed by Redhat yet...&lt;/p&gt;

&lt;p&gt;Does anyone know if this incompatibility with Lustre since 3.10.0-1160.11.1 was introduced by the following patch?&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Loop in __run_timers() because base-&amp;gt;timer_jiffies is very far behind causes a lockup condition. (BZ#1849716)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="291425" author="jfilizetti" created="Mon, 8 Feb 2021 12:02:41 +0000"  >&lt;p&gt;My investigation found it was introduced in 3.10.0-1160.8.1.el7 by&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;kernel&amp;#93;&lt;/span&gt; timer: Fix lockup in __run_timers() caused by large jiffies/timer_jiffies delta (Waiman Long) &lt;span class=&quot;error&quot;&gt;&amp;#91;1849716&amp;#93;&lt;/span&gt;&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="292791" author="adilger" created="Tue, 23 Feb 2021 17:32:06 +0000"  >&lt;p&gt;What is the commit hash for that patch?  I can&apos;t find it in the upstream kernel. &lt;/p&gt;</comment>
                            <comment id="292793" author="jfilizetti" created="Tue, 23 Feb 2021 17:59:40 +0000"  >&lt;p&gt;In the digging I did I found no indication of an upstream patch for that.  It looks like it was Redhat only to address some customer issue.&lt;/p&gt;
</comment>
                            <comment id="292798" author="adilger" created="Tue, 23 Feb 2021 19:29:24 +0000"  >&lt;p&gt;Is there a copy of that patch somewhere?  I was going to revert it and see if that fixes the problem. &lt;/p&gt;</comment>
                            <comment id="293361" author="jfilizetti" created="Sat, 27 Feb 2021 16:48:48 +0000"  >&lt;p&gt;This is what I believe the patch was from diffing between the kernels.  I have a hotfix kernel from RHEL which is supposed to fix the problem but have not tested it yet.&lt;br/&gt;
 &lt;span class=&quot;nobr&quot;&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/attachment/37626/37626_timer.patch&quot; title=&quot;timer.patch attached to LU-14341&quot;&gt;timer.patch&lt;sup&gt;&lt;img class=&quot;rendericon&quot; src=&quot;https://jira.whamcloud.com/images/icons/link_attachment_7.gif&quot; height=&quot;7&quot; width=&quot;7&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/sup&gt;&lt;/a&gt;&lt;/span&gt; &lt;/p&gt;</comment>
                            <comment id="293600" author="jfilizetti" created="Tue, 2 Mar 2021 14:31:54 +0000"  >&lt;p&gt;FWIW I have not seen any issues since upgrading to the hotfix kernel.&lt;/p&gt;</comment>
                            <comment id="293640" author="simmonsja" created="Tue, 2 Mar 2021 17:01:37 +0000"  >&lt;p&gt;Which kernel did this start showing up?&lt;/p&gt;</comment>
                            <comment id="293659" author="jfilizetti" created="Tue, 2 Mar 2021 17:52:06 +0000"  >&lt;p&gt;3.10.0-1160.8.1.el7&lt;/p&gt;</comment>
                            <comment id="293701" author="gerrit" created="Wed, 3 Mar 2021 03:12:03 +0000"  >&lt;p&gt;The patch for master branch needs to be incorporated into &lt;a href=&quot;https://review.whamcloud.com/41822&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/41822&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="293927" author="knweiss" created="Thu, 4 Mar 2021 09:41:35 +0000"  >&lt;p&gt;I&apos;ve noticed that &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-14395&quot; title=&quot;kernel update [RHEL7.9 3.10.0-1160.15.2.el7]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-14395&quot;&gt;&lt;del&gt;LU-14395&lt;/del&gt;&lt;/a&gt; (kernel: kernel update RHEL7.9 &lt;span class=&quot;error&quot;&gt;&amp;#91;3.10.0-1160.15.2.el7&amp;#93;&lt;/span&gt;) was merged in the b2_12 branch. It contains a patch that reverts the timer patch that was introduced upstream in kernel 3.10.0-1160.8.1.el7. How is this issue going to be fixed for Lustre servers with patchless kernels?&lt;/p&gt;</comment>
                            <comment id="293930" author="adilger" created="Thu, 4 Mar 2021 11:09:19 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=knweiss&quot; class=&quot;user-hover&quot; rel=&quot;knweiss&quot;&gt;knweiss&lt;/a&gt; those users will have to patch their kernel anyway, or stick with 3.10.0-1160-6.1.el7 or earlier until RHEL fixes the bug. There isn&apos;t anything that can be done in Lustre to avoid this, since it is in a core part of the kernel.  Since the affected code was working fine for many years without the 1160.8.1 change, there is no reason to expect that reverting it will cause any problems.&lt;/p&gt;</comment>
                            <comment id="300540" author="knweiss" created="Wed, 5 May 2021 12:43:53 +0000"  >&lt;p&gt;Olaf, since a couple of weeks passed may I ask: Did anything happen on Red Hat&apos;s side since you&apos;ve opened Red Hat bug #1914011? Unfortunately, I&apos;m not authorized to access it myself. Is Red Hat aware that their timer change causes this regression?&lt;/p&gt;

&lt;p&gt;Andreas, wouldn&apos;t it be a good idea to at least mention &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-14341&quot; title=&quot;hard LOCKUP lustre servers with kernel-3.10.0-1160.11.1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-14341&quot;&gt;&lt;del&gt;LU-14341&lt;/del&gt;&lt;/a&gt; (i.e. this regression regarding patchless kernels) in the lustre/ChangeLog for the coming Lustre 2.12.7? AFAICS the timer revert is still necessary.&lt;/p&gt;</comment>
                            <comment id="300602" author="adilger" created="Wed, 5 May 2021 20:41:18 +0000"  >&lt;p&gt;Karsten, feel free to submit a patch to that effect.&lt;/p&gt;

&lt;p&gt;Unlike some issues, I don&apos;t think there is anything we can do in the Lustre code to avoid the use of kernel timers completely (e.g. make our own &quot;&lt;tt&gt;cfs_&amp;#42;&lt;/tt&gt;&quot; version of some function and call that instead) because the timers are used all over the kernel and the lockups don&apos;t appear to be in code that Lustre is calling directly.&lt;/p&gt;</comment>
                            <comment id="300605" author="ofaaland" created="Wed, 5 May 2021 20:52:05 +0000"  >&lt;p&gt;Karsten, it looks to me like this is fixed in kernel-3.10.0-1160.20.1.el7&lt;/p&gt;</comment>
                            <comment id="300608" author="yujian" created="Wed, 5 May 2021 21:07:36 +0000"  >&lt;p&gt;Yes, the issue was fixed in kernel 3.10.0-1160.20.1.el7:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;- [kernel] timer: Fix potential bug in requeue_timers() (Waiman Long) [1914011] 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;And in &lt;a href=&quot;https://review.whamcloud.com/42089&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/42089&lt;/a&gt; for Lustre 2.12.7, the revert-fix-lockup-in-run_timers-large-jiffies-delta.patch was removed.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="300610" author="adilger" created="Wed, 5 May 2021 21:12:23 +0000"  >&lt;p&gt;Remove the revert patch was landed as part of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-14527&quot; title=&quot;kernel update [RHEL7.9 3.10.0-1160.21.1.el7]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-14527&quot;&gt;&lt;del&gt;LU-14527&lt;/del&gt;&lt;/a&gt;. &lt;/p&gt;</comment>
                            <comment id="300653" author="knweiss" created="Thu, 6 May 2021 06:28:32 +0000"  >&lt;p&gt;Awesome, thank you all!&lt;/p&gt;

&lt;p&gt;(Sorry for the confusion but for whatever reason my b2_12 checkout was not up-to-date when I checked the status yesterday and thus I missed the decisive commit. I&apos;m looking forward to test Lustre 2.12.7 with (patchless) kernel 3.10.0-1160.21.1.el7 next!)&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="62014">LU-14226</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="62669">LU-14395</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="63367">LU-14527</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="37626" name="timer.patch" size="3458" author="jfilizetti" created="Sat, 27 Feb 2021 16:48:46 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i01jnr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>