<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:58:42 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-13137] User process segfaults since 2.13 client upgrade</title>
                <link>https://jira.whamcloud.com/browse/LU-13137</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We&apos;re not 100% sure this comes from Lustre but this is the main hint that we have so far. We started to upgrade our clients to Lustre 2.13 on Sherlock in mid-December 2019 and it this had taken about 3 weeks to upgrade the whole cluster in a rolling fashion. Now all clients are 2.13. Note that no other changes have been done since then (eg. no system, kernel nor OFED upgrades). Users have been reporting random segmentation faults from all clients. In the few cases that we checked, the binary executed was stored in Lustre, although we&apos;re not sure at this point that it&apos;s always the case.&lt;/p&gt;

&lt;p&gt;Example of segfaults are:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;2020-01-14T10:47:50-08:00 sh-116-04 kernel: python[44357]: segfault at a4996 ip 00000000000a4996 sp 00007fc7c945dee8 error 14

2020-01-14T10:46:22-08:00 sh-108-38 kernel: angsd[335614]: segfault at 0 ip           (null) sp 00007ffeec25ab08 error 14

2020-01-14T10:33:27-08:00 sh-107-31 kernel: minimap2[100552]: segfault at 2206 ip 0000000000002206 sp 00007f942ce0cbf8 error 14

2020-01-14T10:33:23-08:00 sh-107-10 kernel: samtools[305687]: segfault at f936 ip 000000000000f936 sp 00007ffebc4dfd18 error 14 in samtools[5631f2a6d000+7000]

2020-01-14T10:32:49-08:00 sh-ln06 kernel: conda[15953]: segfault at 3206 ip 0000000000003206 sp 00007fe3c0495a18 error 14 in python3.7[5590a194e000+5b000]

2020-01-14T10:25:19-08:00 sh-27-30 kernel: gatk[123897]: segfault at 0 ip           (null) sp 00007ffea874efe8 error 14 in python3.8[556487517000+5e000]

2020-01-14T10:25:08-08:00 sh-27-29 kernel: bwa[39718]: segfault at 0 ip           (null) sp 00007f1eebffec08 error 14 in bwa[400000+59000]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Most are error 14. From &lt;a href=&quot;https://utcc.utoronto.ca/~cks/space/blog/linux/KernelSegfaultErrorCodes:&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://utcc.utoronto.ca/~cks/space/blog/linux/KernelSegfaultErrorCodes:&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;error 14: attempt to execute code from an unmapped area.
This is the sign of trying to call through a mangled function pointer (or a NULL one), or perhaps returning from a call when the stack is in an unexpected or corrupted state so that the return address isn&apos;t valid. One source of mangled function pointers is use-after-free issues where the (freed) object contains embedded function pointers.

(Error 14 with a faulting address of 0 often means a function call through a NULL pointer, which in turn often means &apos;making an indirect call to a function without checking that it&apos;s defined&apos;. There are various larger scale causes of this in code.)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;From a core dump of such segfaults, I couldn&apos;t get any more insight neither:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;$ gdb -c core.118324 /scratch/users/xunger08/run_debug_2018/dqmc_stack
GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-114.el7
Copyright (C) 2013 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later &amp;lt;http://gnu.org/licenses/gpl.html&amp;gt;
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.  Type &quot;show copying&quot;
and &quot;show warranty&quot; for details.
This GDB was configured as &quot;x86_64-redhat-linux-gnu&quot;.
For bug reporting instructions, please see:
&amp;lt;http://www.gnu.org/software/gdb/bugs/&amp;gt;...
Reading symbols from /scratch/users/xunger08/run_debug_2018/dqmc_stack...done.
[New LWP 118364]
[New LWP 118324]
Core was generated by `/scratch/users/xunger08/run_debug_2018/./dqmc_stack -t 35000 /scratch/users/xun&apos;.
Program terminated with signal 11, Segmentation fault.
#0  0x0000000000000000 in ?? ()
(gdb) thread apply all bt

Thread 2 (LWP 118324):
#0  0x00007f393383cf19 in ?? ()
#1  0xcccccccccccccccc in ?? ()
#2  0x00007f3931ceae00 in ?? ()
#3  0x00007f3900000001 in ?? ()
#4  0xcccccccc00000001 in ?? ()
#5  0x00007ffd624a05e0 in ?? ()
#6  0x00007f3900000000 in ?? ()
#7  0x00007f3931ce9d00 in ?? ()
#8  0x00007ffd00000001 in ?? ()
#9  0x00007ffd00000001 in ?? ()
#10 0x00007f3931cea000 in ?? ()
#11 0x0009b4232aa60776 in ?? ()
#12 0x00007ffd00000000 in ?? ()
#13 0x00007f3931cea000 in ?? ()
#14 0x00007ffd00000001 in ?? ()
#15 0x000000005c1b37a8 in ?? ()
#16 0x00007ffd624a09f0 in ?? ()
#17 0x00007ffd00000000 in ?? ()
#18 0x0000000000000001 in ?? ()
#19 0x0000000000000000 in ?? ()

Thread 1 (LWP 118364):
#0  0x0000000000000000 in ?? ()
#1  0x0000000000000000 in ?? ()
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;We&apos;re strongly wondering if this could be related to this recent 2.12 to 2.13 client upgrade. Checking our logs through Splunk shows that this issue started at the same time we began to upgrade our Lustre clients to 2.13, that is why I&apos;m opening this ticket. Could this be a come-back of an exec-in-lustre issue? Any suggestions are welcomed, otherwise, we&apos;ll try to start downgrading to 2.12 LTS and see if that fixes the problem.&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;</description>
                <environment>CentOS 7.6, Lustre client 2.13.0 from WC</environment>
        <key id="57803">LU-13137</key>
            <summary>User process segfaults since 2.13 client upgrade</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="qian_wc">Qian Yingjin</assignee>
                                    <reporter username="sthiell">Stephane Thiell</reporter>
                        <labels>
                    </labels>
                <created>Wed, 15 Jan 2020 00:06:43 +0000</created>
                <updated>Thu, 14 May 2020 13:45:53 +0000</updated>
                            <resolved>Thu, 14 May 2020 13:45:53 +0000</resolved>
                                    <version>Lustre 2.13.0</version>
                                    <fixVersion>Lustre 2.14.0</fixVersion>
                                        <due></due>
                            <votes>1</votes>
                                    <watches>14</watches>
                                                                            <comments>
                            <comment id="261231" author="adilger" created="Wed, 15 Jan 2020 06:20:55 +0000"  >&lt;p&gt;Stephane, looking at the commit messages and affected files for patches that were landed in 2.13, I see the following that are related to mmap:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;https://review.whamcloud.com/35500  LU-12400 llite: Use the new vm_fault_t type
https://review.whamcloud.com/34247  LU-11403 llite: ll_fault fixes  [in 2.12.3]
https://review.whamcloud.com/34242  LU-8299 llite: ll_fault should fail for insane file offsets
https://review.whamcloud.com/32966  LU-10092 pcc: Non-blocking PCC caching
https://review.whamcloud.com/32963  LU-10092 llite: Add persistent cache on client
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The first one looks harmless (mostly a compile fix for new kernels).  The second one looks like it is the most likely to affect the mmap/page fault code, but it was also backported to 2.12.3, so I think you would already have been running it on your clients?  If you were previously running 2.12.2 on your clients then this would be a likely candidate, but otherwise it seems unlikely.  The PCC patches did modify &lt;tt&gt;llite_mmap.c&lt;/tt&gt; but &lt;em&gt;shouldn&apos;t&lt;/em&gt; have affected the functionality unless PCC was configured on the client.&lt;/p&gt;</comment>
                            <comment id="261281" author="sthiell" created="Wed, 15 Jan 2020 17:59:36 +0000"  >&lt;p&gt;Hi Andreas,&lt;/p&gt;

&lt;p&gt;We&apos;re still investigating this and since yesterday:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;one user confirmed that after moving his binaries from Lustre (Fir) to the $HOME NFS filesystem, his code doesn&apos;t crash anymore&lt;/li&gt;
	&lt;li&gt;we found a binary that has crashed which was stored on Oak (2.10 server)&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Users are reporting that crashes are quite random and happen after hours of run time. Some of them seem to be able to reproduce it.&lt;/p&gt;

&lt;p&gt;Also we have checked and we were using 2.12.3 on the clients before upgrading to 2.13, and we haven&apos;t noticed that amount of segfaults with 2.12.3.&lt;/p&gt;</comment>
                            <comment id="261325" author="pjones" created="Thu, 16 Jan 2020 15:22:14 +0000"  >&lt;p&gt;Bobijam&lt;/p&gt;

&lt;p&gt;Can you please advise?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="261335" author="srcc" created="Thu, 16 Jan 2020 15:53:48 +0000"  >&lt;p&gt;Hi all, &lt;/p&gt;

&lt;p&gt;I just wanted to add that we had several users who confirmed that they moved their binaries from Lustre (Fir or Oak) to NFS filesystems and that they haven&apos;t encountered any segfault since then.&lt;/p&gt;

&lt;p&gt;Thanks!&lt;br/&gt;
--&#160;&lt;/p&gt;

&lt;p&gt;Kilian&lt;/p&gt;</comment>
                            <comment id="261440" author="sthiell" created="Fri, 17 Jan 2020 17:59:08 +0000"  >&lt;p&gt;Good news. Kilian and I were able to reproduce the problem this morning!&lt;/p&gt;

&lt;p&gt;We have figured out that this problem is related to client-side LRU cached locks because most of the binaries on Lustre crashed after 65 minutes, which is the default value of &lt;tt&gt;lru_max_age&lt;/tt&gt;.&lt;/p&gt;

&lt;p&gt;A segmentation fault can easily be reproduced with a Lustre 2.13 client by executing a binary in Lustre and then clearing the locks (MDT only seems to be enough, even without DoM):&lt;/p&gt;

&lt;p&gt;1. execute any binary in Lustre&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;./binary
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;2. at the same time, clear the locks:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lctl set_param ldlm.namespaces.*MDT*.lru_size=clear
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;This is also reproducible by reducing &lt;tt&gt;lru_max_age&lt;/tt&gt; to low value, something like 30s, and the binary will segfault exactly after 30 seconds.&lt;/p&gt;

&lt;p&gt;As a partial workaround, we have increased &lt;tt&gt;lru_max_age&lt;/tt&gt; on all Sherlock clients to &lt;tt&gt;1209600s&lt;/tt&gt; (14 days), which matches our max allowed job time, but I don&apos;t think this will be enough to avoid all segfaults.&lt;/p&gt;</comment>
                            <comment id="261483" author="adilger" created="Sat, 18 Jan 2020 11:57:56 +0000"  >&lt;p&gt;Bobijam, probably the easiest way to debug this would be to make a patch with a sanity test for master that triggers the failure as described by Stephane, run only that test with &lt;tt&gt;Test-Paramweters: fortestonly testlist=sanity envdefinitions=ONLY=N&lt;/tt&gt;, then use &lt;tt&gt;git bisect&lt;/tt&gt; to cherry-pick the test to different versions of Lustre between 2.12.50 and 2.13.50.&lt;/p&gt;

&lt;p&gt;One idea that has been proposed in the past is to write a script similar to &lt;tt&gt;git bisect&lt;/tt&gt; that submits several versions of the patch (maybe 33 by default, with different &lt;tt&gt;Change-Id:&lt;/tt&gt; values) to Gerrit so that they can be run in parallel.  This would likely be able to able to isolate the problem within a few hours, since &quot;&lt;tt&gt;git log --pretty=oneline  2.12.50..2.13.50&lt;/tt&gt;&quot; reports 1049 patches.  The first set of 33 patches should result in the first patch at 2.12.50 passing, and the last patch at 2.13.50 should fail, and somewhere inbetween, leaving 32 or 33 patches between the passing and failing patches.  A second set of patches could be generated to isolate the failure to a single patch if it isn&apos;t clear what the source of the problem is.&lt;/p&gt;

&lt;p&gt;Having a script to do this would be useful for isolating all kinds of problems in the future, so should be checked in to Git as well.  It might be necessary to run the subtest multiple times to ensure we get a strong verification on whether the subtest is passing or failing, in case it is an intermittent failure, so it would be useful to allow &lt;tt&gt;run_one_logged()&lt;/tt&gt; to call the same subtest multiple times in a row (e.g. with &lt;tt&gt;ONLY_REPEAT=N&lt;/tt&gt; in &lt;tt&gt;envdefinitions&lt;/tt&gt;).&lt;/p&gt;</comment>
                            <comment id="261506" author="gerrit" created="Sun, 19 Jan 2020 14:15:12 +0000"  >&lt;p&gt;Bobi Jam (bobijam@hotmail.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/37278&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/37278&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-13137&quot; title=&quot;User process segfaults since 2.13 client upgrade&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-13137&quot;&gt;&lt;del&gt;LU-13137&lt;/del&gt;&lt;/a&gt; debug: exec binary while clear locks&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 8164da42211f5185e6ecb1e8a7f7dea5c8a80bff&lt;/p&gt;</comment>
                            <comment id="261507" author="gerrit" created="Sun, 19 Jan 2020 14:17:35 +0000"  >&lt;p&gt;Bobi Jam (bobijam@hotmail.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/37279&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/37279&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-13137&quot; title=&quot;User process segfaults since 2.13 client upgrade&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-13137&quot;&gt;&lt;del&gt;LU-13137&lt;/del&gt;&lt;/a&gt; debug: exec binary while clear lock&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 0f995b39cf3f243e4634402e53fd7498d08fc6ad&lt;/p&gt;</comment>
                            <comment id="261556" author="bobijam" created="Tue, 21 Jan 2020 08:48:41 +0000"  >&lt;p&gt;$ git bisect good f172b116885753d0f316549a2fb9d451e9b4bd2e&lt;br/&gt;
58d744e3eaab358ef346e51ff4aa17e9f08efbb3 is the first bad commit&lt;br/&gt;
commit 58d744e3eaab358ef346e51ff4aa17e9f08efbb3&lt;br/&gt;
Author: Qian Yingjin &amp;lt;qian@ddn.com&amp;gt;&lt;br/&gt;
Date:   Thu Jul 5 14:43:46 2018 +0800&lt;/p&gt;

&lt;p&gt;    &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10092&quot; title=&quot;PCC: Lustre Persistent Client Cache&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10092&quot;&gt;&lt;del&gt;LU-10092&lt;/del&gt;&lt;/a&gt; pcc: Non-blocking PCC caching&lt;/p&gt;</comment>
                            <comment id="261731" author="sthiell" created="Thu, 23 Jan 2020 16:35:26 +0000"  >&lt;p&gt;Hello! Is a patch against master being actively worked on? Even with our &lt;tt&gt;lru_max_age&lt;/tt&gt;&#160;workaround, we are getting up to 200 segfaults per hour on Sherlock. We told our users to move all executables out of Lustre (we have an NFS filer for that), but some have large software setups on Lustre. I&apos;m asking to know if we should downgrade all our clients to 2.12 or if we&apos;ll have a chance to apply a patch on top of 2.13 (within the next days) that would fix this issue. Thanks much!&lt;/p&gt;</comment>
                            <comment id="261741" author="adilger" created="Thu, 23 Jan 2020 19:22:02 +0000"  >&lt;p&gt;Stephane, I&apos;m assuming you are not using PCC, so if we supply a temporary patch that breaks PCC functionality to get the executables working again, that would be OK?  I have an idea what the culprit is, so I could &quot;fix&quot; this, at the likely cost of breaking PCC. &lt;/p&gt;</comment>
                            <comment id="261742" author="sthiell" created="Thu, 23 Jan 2020 19:25:05 +0000"  >&lt;p&gt;Hi Andreas,&lt;/p&gt;

&lt;p&gt;No we&apos;re not using PCC (yet &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/wink.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;), so a patch that would break PCC would be perfectly fine at this point.&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;</comment>
                            <comment id="261743" author="adilger" created="Thu, 23 Jan 2020 19:26:17 +0000"  >&lt;p&gt;In particular, I think it is the following hunk that is causing the problem:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
{@@ -487,7 +499,7 @@ &lt;span class=&quot;code-object&quot;&gt;int&lt;/span&gt; ll_teardown_mmaps(struct address_space *mapping, __u64 first, __u64 last)
         &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (mapping_mapped(mapping)) {
                 rc = 0;
                unmap_mapping_range(mapping, first + PAGE_SIZE - 1,
-                                    last - first + 1, 0);
+                                   last - first + 1, 1);
         }
 
         RETURN(rc);
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;this is causing &lt;b&gt;all&lt;/b&gt; mmapped pages to be cleared, which is probably not the right thing.  I will submit a patch to test that theory shortly, but you may be able to test/deploy a patch to revert that change faster than me. &lt;/p&gt;</comment>
                            <comment id="261757" author="adilger" created="Thu, 23 Jan 2020 21:18:11 +0000"  >&lt;p&gt;I&apos;ve updated &lt;a href=&quot;https://review.whamcloud.com/37278&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/37278&lt;/a&gt; (the test case) with a potential fix.  We&apos;ll see in an hour or so whether it solves the problem, since the test case could reproduce the problem easily.&lt;/p&gt;</comment>
                            <comment id="261759" author="sthiell" created="Thu, 23 Jan 2020 21:27:01 +0000"  >&lt;p&gt;Andreas, I built a 2.13 + the revert of that change and just tested on a client but I&apos;m still able to reproduce the segfault when executing a binary and clearing the LRU cache.&lt;/p&gt;</comment>
                            <comment id="261760" author="adilger" created="Thu, 23 Jan 2020 21:58:14 +0000"  >&lt;p&gt;It looks like there was a second case of &lt;tt&gt;unmap_mapping_range()&lt;/tt&gt; being changed by the PCC patch.  With v12 of the patch that reverts both of those changes. the patch is now passing the regression test which previously failed.  Please skip v11 of the patch with only the &lt;tt&gt;ll_teardown_mmaps()&lt;/tt&gt; change, you need the change in &lt;tt&gt;vvp_conf_set()&lt;/tt&gt; also.&lt;/p&gt;</comment>
                            <comment id="261764" author="sthiell" created="Thu, 23 Jan 2020 22:28:03 +0000"  >&lt;p&gt;Andreas, I tested with the two changes as in v12 on top of 2.13 and it seems to fix the problem: no binary segfault anymore when clearing the LRU manually or reducing &lt;tt&gt;lru_max_age&lt;/tt&gt;. We might perform additional tests and wait for the test results from Gerrit before deploying it. Thanks &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="261776" author="adilger" created="Fri, 24 Jan 2020 01:12:46 +0000"  >&lt;p&gt;You can see from the initial test results that the patch is passing the fault-specific regression tests (it runs each newly-added test 10x for &amp;lt;ldiskfs,ZFS&amp;gt;x&amp;lt;single,DNE&amp;gt; in the &quot;special&quot; sessions, and each of those tests runs exec+lock cancel 10x), in addition to the 3x custom tests, so this test has already passed a few hundred times when it was otherwise failing 100% of the time:&lt;br/&gt;
&lt;a href=&quot;https://testing-archive.whamcloud.com/gerrit-janitor/5943/results.html&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing-archive.whamcloud.com/gerrit-janitor/5943/results.html&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sessions/db6415c7-4af7-4700-a86e-19f39f36e695&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sessions/db6415c7-4af7-4700-a86e-19f39f36e695&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sessions/84c2c40a-8ec9-451f-8d87-2b8de45fc112&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sessions/84c2c40a-8ec9-451f-8d87-2b8de45fc112&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sessions/42fd313e-1c96-4d67-91db-f9bdaa2050e7&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sessions/42fd313e-1c96-4d67-91db-f9bdaa2050e7&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Also, since this is just reverting the behaviour to that of 2.12.x and earlier the risk is very low.&lt;/p&gt;

&lt;p&gt;Note that there is currently a totally unrelated regression in ZFS testing (&lt;a href=&quot;https://review.whamcloud.com/37320&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/37320&lt;/a&gt;) that is causing all ZFS test sessions to fail, so do not be alarmed when you see the patch fail all of the ZFS test sessions.  &lt;/p&gt;</comment>
                            <comment id="267497" author="qian_wc" created="Tue, 14 Apr 2020 01:41:31 +0000"  >&lt;p&gt;This is a mmap problem for PCC when multiple client access on a shared mmapped file. Fixing it seems different if not patch the kernel of a client. As Lixi suggested, it seems reasonable to release the semantics of mmap when using PCC on a shared file among different clients.&lt;/p&gt;</comment>
                            <comment id="270143" author="gerrit" created="Thu, 14 May 2020 05:38:26 +0000"  >&lt;p&gt;Oleg Drokin (green@whamcloud.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/37278/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/37278/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-13137&quot; title=&quot;User process segfaults since 2.13 client upgrade&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-13137&quot;&gt;&lt;del&gt;LU-13137&lt;/del&gt;&lt;/a&gt; llite: do not flush COW pages from mapping&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 13a0066afb8d89b12653ff72f7311fd5e03ef6b4&lt;/p&gt;</comment>
                            <comment id="270192" author="pjones" created="Thu, 14 May 2020 13:45:53 +0000"  >&lt;p&gt;Landed for 2.14&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="48631">LU-10092</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="34128" name="Screen Shot 2020-01-14 at 3.50.24 PM.png" size="116995" author="sthiell" created="Wed, 15 Jan 2020 00:02:11 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i00s1b:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10021"><![CDATA[2]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>