<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:26:52 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-9514] sanity test 51f times out due to out of memory</title>
                <link>https://jira.whamcloud.com/browse/LU-9514</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;sanity test_51f times out with the usual message&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&apos;test failed to respond and timed out&apos; 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The only information in the test_log is&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity test 51f: check many open files limit ====================================================== 08:37:34 (1494923854)
MDT0 numfree=711956, max=100000
changed ulimit from 1024 to 100020
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;There are two cases where sanity test 51f times out with oom. The stack trace is found in the client console&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;08:38:17:[ 1565.371695] Lustre: DEBUG MARKER: == sanity test 51f: check many open files limit ====================================================== 08:37:34 (1494923854)
08:38:17:[ 1602.443283] automount invoked oom-killer: gfp_mask=0x200da, order=0, oom_score_adj=0
08:38:17:[ 1602.446098] automount cpuset=/ mems_allowed=0
08:38:17:[ 1602.448513] CPU: 0 PID: 1010 Comm: automount Tainted: G           OE  ------------   3.10.0-514.16.1.el7.x86_64 #1
08:38:17:[ 1602.451566] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2007
08:38:17:
08:38:17:[ 1602.454058]  ffff8800794eaf10 000000003457cee2 ffff88007b9e79f8 ffffffff81686ac3
08:38:17:[ 1602.459021]  ffff88007b9e7a88 ffffffff81681a6e ffff88007b9e7af0 ffff88007b9e7a48
08:38:17:[ 1602.461680]  ffffffff81692b6c ffffffff81a33cc0 0000000000000000 0000000000000000
08:38:17:[ 1602.464276] Call Trace:
08:38:17:[ 1602.466395]  [&amp;lt;ffffffff81686ac3&amp;gt;] dump_stack+0x19/0x1b
08:38:17:[ 1602.468694]  [&amp;lt;ffffffff81681a6e&amp;gt;] dump_header+0x8e/0x225
08:38:17:[ 1602.470982]  [&amp;lt;ffffffff81692b6c&amp;gt;] ? notifier_call_chain+0x4c/0x70
08:38:17:[ 1602.473299]  [&amp;lt;ffffffff810b6d98&amp;gt;] ? __blocking_notifier_call_chain+0x58/0x70
08:38:17:[ 1602.475696]  [&amp;lt;ffffffff81184e1e&amp;gt;] check_panic_on_oom+0x2e/0x60
08:38:17:[ 1602.477970]  [&amp;lt;ffffffff8118523b&amp;gt;] out_of_memory+0x23b/0x4f0
08:38:17:[ 1602.480183]  [&amp;lt;ffffffff81682577&amp;gt;] __alloc_pages_slowpath+0x5d7/0x725
08:38:17:[ 1602.482651]  [&amp;lt;ffffffff8118b5c5&amp;gt;] __alloc_pages_nodemask+0x405/0x420
08:38:17:[ 1602.484867]  [&amp;lt;ffffffff811d272a&amp;gt;] alloc_pages_vma+0x9a/0x150
08:38:17:[ 1602.487039]  [&amp;lt;ffffffff811c351b&amp;gt;] read_swap_cache_async+0xeb/0x160
08:38:17:[ 1602.489192]  [&amp;lt;ffffffff811c3638&amp;gt;] swapin_readahead+0xa8/0x110
08:38:17:[ 1602.491293]  [&amp;lt;ffffffff811b189c&amp;gt;] handle_mm_fault+0xb1c/0xfe0
08:38:17:[ 1602.493385]  [&amp;lt;ffffffff81692644&amp;gt;] __do_page_fault+0x154/0x450
08:38:17:[ 1602.495536]  [&amp;lt;ffffffff81692975&amp;gt;] do_page_fault+0x35/0x90
08:38:17:[ 1602.497619]  [&amp;lt;ffffffff8168eb88&amp;gt;] page_fault+0x28/0x30
08:38:17:[ 1602.499598] Mem-Info:
08:38:17:[ 1602.501313] active_anon:350 inactive_anon:385 isolated_anon:0
08:38:17:[ 1602.501313]  active_file:0 inactive_file:525 isolated_file:0
08:38:17:[ 1602.501313]  unevictable:0 dirty:0 writeback:142 unstable:0
08:38:17:[ 1602.501313]  slab_reclaimable:4058 slab_unreclaimable:36202
08:38:17:[ 1602.501313]  mapped:0 shmem:52 pagetables:1613 bounce:0
08:38:17:[ 1602.501313]  free:12916 free_pcp:7 free_cma:0
08:38:17:[ 1602.512345] Node 0 DMA free:7020kB min:416kB low:520kB high:624kB active_anon:400kB inactive_anon:640kB active_file:0kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15984kB managed:15892kB mlocked:0kB dirty:0kB writeback:36kB mapped:0kB shmem:228kB slab_reclaimable:108kB slab_unreclaimable:712kB kernel_stack:48kB pagetables:20kB unstable:0kB bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:15736 all_unreclaimable? yes
08:38:17:[ 1602.521429] lowmem_reserve[]: 0 1663 1663 1663
08:38:17:[ 1602.523300] Node 0 DMA32 free:44644kB min:44636kB low:55792kB high:66952kB active_anon:1000kB inactive_anon:900kB active_file:0kB inactive_file:2100kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:2080756kB managed:1705000kB mlocked:0kB dirty:0kB writeback:532kB mapped:0kB shmem:0kB slab_reclaimable:16124kB slab_unreclaimable:144096kB kernel_stack:2656kB pagetables:6432kB unstable:0kB bounce:0kB free_pcp:28kB local_pcp:0kB free_cma:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? no
08:38:17:[ 1602.534026] lowmem_reserve[]: 0 0 0 0
08:38:17:[ 1602.535888] Node 0 DMA: 39*4kB (EM) 32*8kB (UEM) 25*16kB (UM) 12*32kB (UEM) 3*64kB (EM) 10*128kB (UM) 3*256kB (EM) 3*512kB (UEM) 2*1024kB (UE) 0*2048kB 0*4096kB = 7020kB
08:38:17:[ 1602.540391] Node 0 DMA32: 995*4kB (UEM) 623*8kB (UEM) 319*16kB (UEM) 861*32kB (UEM) 46*64kB (M) 0*128kB 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 44564kB
08:38:17:[ 1602.544824] Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0 hugepages_size=1048576kB
08:38:17:[ 1602.547006] Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0 hugepages_size=2048kB
08:38:17:[ 1602.549124] 975 total pagecache pages
08:38:17:[ 1602.550875] 375 pages in swap cache
08:38:17:[ 1602.552582] Swap cache stats: add 26733, delete 26358, find 248/330
08:38:17:[ 1602.554487] Free swap  = 3566956kB
08:38:17:[ 1602.556164] Total swap = 3671036kB
08:38:17:[ 1602.557826] 524185 pages RAM
08:38:17:[ 1602.559435] 0 pages HighMem/MovableOnly
08:38:17:[ 1602.561132] 93962 pages reserved
08:38:17:[ 1602.562747] [ pid ]   uid  tgid total_vm      rss nr_ptes swapents oom_score_adj name
08:38:17:[ 1602.564748] [  343]     0   343     9303        1      20       92             0 systemd-journal
08:38:17:[ 1602.566845] [  365]     0   365    29723        0      26       85             0 lvmetad
08:38:17:[ 1602.568879] [  382]     0   382    10895        1      21      134         -1000 systemd-udevd
08:38:17:[ 1602.570936] [  460]     0   460    13854        0      28      108         -1000 auditd
08:38:17:[ 1602.572986] [  492]     0   492     4793       18      14       37             0 irqbalance
08:38:17:[ 1602.575027] [  493]   998   493   134941        0      53     2131             0 polkitd
08:38:17:[ 1602.577041] [  496]    81   496     9197        0      20      129          -900 dbus-daemon
08:38:17:[ 1602.579085] [  509]   997   509    28962        0      28       98             0 chronyd
08:38:17:[ 1602.581105] [  559]     0   559    67227        0      43      657             0 gssproxy
08:38:17:[ 1602.583118] [  603]     0   603    64343        0      74      266             0 sssd
08:38:17:[ 1602.585096] [  604]     0   604    90858        0     118      570             0 sssd_be
08:38:17:[ 1602.587095] [  605]     0   605   127883        0      68     1074             0 NetworkManager
08:38:17:[ 1602.589144] [  612]     0   612    30625        1      30      207             0 rpc.gssd
08:38:17:[ 1602.591150] [  619]     0   619    65752        5      84      227             0 sssd_nss
08:38:17:[ 1602.593151] [  620]     0   620    60653        0      73      238             0 sssd_pam
08:38:17:[ 1602.595417] [  621]     0   621    59047        5      71      221             0 sssd_ssh
08:38:17:[ 1602.598098] [  622]     0   622    67553        0      83      330             0 sssd_pac
08:38:17:[ 1602.600145] [  635]     0   635     6048        1      16       77             0 systemd-logind
08:38:17:[ 1602.602201] [  641]     0   641    28220        1      58     3120             0 dhclient
08:38:17:[ 1602.604259] [  859]     0   859   138287       62      87     2611             0 tuned
08:38:17:[ 1602.606279] [  860]     0   860    77477        0      46      285             0 rsyslogd
08:38:17:[ 1602.608319] [  861]     0   861    26370        1      54      248         -1000 sshd
08:38:17:[ 1602.610259] [  868]     0   868     6776        1      17       62             0 xinetd
08:38:17:[ 1602.612235] [  875]   995   875    56858        0      23      263             0 munged
08:38:17:[ 1602.614120] [  989]     0   989    22780        0      43      262             0 master
08:38:17:[ 1602.615980] [  991]    89   991    25347        0      45      253             0 pickup
08:38:17:[ 1602.617764] [  992]    89   992    25364        0      47      254             0 qmgr
08:38:17:[ 1602.619536] [ 1007]     0  1007    31555        0      18      155             0 crond
08:38:17:[ 1602.621255] [ 1009]     0  1009   149996        0      63      498             0 automount
08:38:17:[ 1602.623021] [ 1012]     0  1012     6461        0      17       51             0 atd
08:38:17:[ 1602.624633] [ 1043]     0  1043    27509        1      10       32             0 agetty
08:38:17:[ 1602.626296] [ 1045]     0  1045    27509        1      12       32             0 agetty
08:38:17:[ 1602.627878] [ 2247]     0  2247    40560        0      79      347             0 sshd
08:38:17:[ 1602.629439] [ 2249]     0  2249    28283        0      14       58             0 run_test.sh
08:38:17:[ 1602.630986] [ 2461]     0  2461    29424       28      15      671             0 bash
08:38:17:[ 1602.632479] [22429]     0 22429    29424       23      13      675             0 bash
08:38:17:[ 1602.633907] [22430]     0 22430    26975        0       9       27             0 tee
08:38:17:[ 1602.635313] [22600]     0 22600    29793       39      14     1016             0 bash
08:38:17:[ 1602.636706] [17680]     0 17680    29793       59      14      999             0 bash
08:38:17:[ 1602.638107] [17681]     0 17681    26975        1       9       26             0 tee
08:38:17:[ 1602.639460] [17855]     0 17855     2076        9       9       19             0 createmany
08:38:17:[ 1602.640891] Kernel panic - not syncing: Out of memory: system-wide panic_on_oom is enabled
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Logs for these two cases are at:&lt;br/&gt;
2017-05-16 - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/4f3d47bc-3a42-11e7-b81e-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/4f3d47bc-3a42-11e7-b81e-5254006e85c2&lt;/a&gt;&lt;br/&gt;
2017-05-16 - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/6d1be1a2-3a43-11e7-b861-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/6d1be1a2-3a43-11e7-b861-5254006e85c2&lt;/a&gt;&lt;/p&gt;</description>
                <environment>autotest</environment>
        <key id="46127">LU-9514</key>
            <summary>sanity test 51f times out due to out of memory</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="yong.fan">nasf</assignee>
                                    <reporter username="jamesanunez">James Nunez</reporter>
                        <labels>
                    </labels>
                <created>Tue, 16 May 2017 16:45:24 +0000</created>
                <updated>Wed, 24 Apr 2019 17:00:14 +0000</updated>
                            <resolved>Wed, 19 Jul 2017 03:29:54 +0000</resolved>
                                    <version>Lustre 2.10.0</version>
                                    <fixVersion>Lustre 2.10.1</fixVersion>
                    <fixVersion>Lustre 2.11.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>10</watches>
                                                                            <comments>
                            <comment id="196156" author="bogl" created="Wed, 17 May 2017 13:47:26 +0000"  >&lt;p&gt;more on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/2d824ac6-3abc-11e7-91db-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/2d824ac6-3abc-11e7-91db-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b573cb72-3b56-11e7-b3f6-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b573cb72-3b56-11e7-b3f6-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/f864ac5e-3c18-11e7-b3f6-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/f864ac5e-3c18-11e7-b3f6-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="196347" author="gerrit" created="Thu, 18 May 2017 14:25:05 +0000"  >&lt;p&gt;James Nunez (james.a.nunez@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/27189&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27189&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9514&quot; title=&quot;sanity test 51f times out due to out of memory&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9514&quot;&gt;&lt;del&gt;LU-9514&lt;/del&gt;&lt;/a&gt; test: sanity 51f add to ALWAYS_EXCEPT&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 9f6e57118a8af37a0184303740d9d3c43d4fced9&lt;/p&gt;</comment>
                            <comment id="196365" author="jay" created="Thu, 18 May 2017 16:55:24 +0000"  >&lt;p&gt;it seems this bug outboke after the patch &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9305&quot; title=&quot;Running File System Aging create write checksum errors &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9305&quot;&gt;&lt;del&gt;LU-9305&lt;/del&gt;&lt;/a&gt; is landed.&lt;/p&gt;</comment>
                            <comment id="196366" author="pjones" created="Thu, 18 May 2017 17:07:42 +0000"  >&lt;p&gt;Jinshan&lt;/p&gt;

&lt;p&gt;Do you think that we should revert &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9305&quot; title=&quot;Running File System Aging create write checksum errors &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9305&quot;&gt;&lt;del&gt;LU-9305&lt;/del&gt;&lt;/a&gt; until you have this more fully understood?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="196373" author="jay" created="Thu, 18 May 2017 17:36:32 +0000"  >&lt;p&gt;Another possible culprit would be &quot;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7473&quot; title=&quot;Enlarge ACL entry limitation&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7473&quot;&gt;&lt;del&gt;LU-7473&lt;/del&gt;&lt;/a&gt; acl: increase ACL entries limitation&quot; where the sanity:51f was not run at all: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/3dfbffd0-e181-11e6-8cc2-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/3dfbffd0-e181-11e6-8cc2-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="196376" author="jay" created="Thu, 18 May 2017 18:02:10 +0000"  >&lt;p&gt;Peter - no, it&apos;s less likely to be &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9305&quot; title=&quot;Running File System Aging create write checksum errors &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9305&quot;&gt;&lt;del&gt;LU-9305&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="196389" author="yujian" created="Thu, 18 May 2017 20:17:51 +0000"  >&lt;p&gt;By reverting the patch &lt;a href=&quot;https://review.whamcloud.com/19790/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/19790/&lt;/a&gt; for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7473&quot; title=&quot;Enlarge ACL entry limitation&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7473&quot;&gt;&lt;del&gt;LU-7473&lt;/del&gt;&lt;/a&gt; from the latest master branch, sanity test 51f passed with FSTYPE=zfs:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/1f66dfc9-5d3a-4cf0-a264-6bfabef2d825&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/1f66dfc9-5d3a-4cf0-a264-6bfabef2d825&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="196407" author="green" created="Thu, 18 May 2017 22:56:07 +0000"  >&lt;p&gt;I believe the actual problem is test 51f itself - it puts the system under so much strain that any slightest deviation in memory allocation tips it over with our minimalmemory test VMs.&lt;br/&gt;
e.g. if we enable debug i nthe kernel - that kills it too: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/e4ce6544-2a8a-11e7-9073-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/e4ce6544-2a8a-11e7-9073-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="196421" author="adilger" created="Fri, 19 May 2017 02:05:13 +0000"  >&lt;p&gt;I think the root of the problem is that we are probably allocating too much memory for each open file.  For a 2GB node, allocating 100k files is about 21KB per file.  Not outrageous, but I suspect we could do better with some investigation.&lt;/p&gt;

&lt;p&gt;I ran a test on my VM (2GB RAM, 72k files) and didn&apos;t have any problems with the patch, but i did get about 88MB into swap during the run.  If our test nodes don&apos;t have swap space they would fail when running out of memory.&lt;/p&gt;

&lt;p&gt;In the meantime, it might make sense to limit this to &quot;safer&quot; parameters like 50000 files max when running in a VM?&lt;/p&gt;</comment>
                            <comment id="196432" author="gerrit" created="Fri, 19 May 2017 04:21:27 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/27189/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27189/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9514&quot; title=&quot;sanity test 51f times out due to out of memory&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9514&quot;&gt;&lt;del&gt;LU-9514&lt;/del&gt;&lt;/a&gt; test: sanity 51f add to ALWAYS_EXCEPT&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: cfed34264ac3ab96507a55973e6ffe1d648ca527&lt;/p&gt;</comment>
                            <comment id="196435" author="yong.fan" created="Fri, 19 May 2017 06:38:38 +0000"  >&lt;p&gt;Part of the reason is known now.&lt;br/&gt;
Because the client does not know how many ACL entries will be for the file to be opened, it will create big enough buffer to hold the open RPC reply. On the other hand, open RPC is replayable, so before the application close the file, the open RPC is held on the client, then its large reply buffer is held also. With large ACL patch introduced, the open RPC reply buffer is larger than original case. The  &quot;createmany -o -k&quot; operation keeps a lot of files opened on the client. As to a lot of RAM are occupied by the unused reply buffer. That is why the client out of RAM in sanity-51f.&lt;/p&gt;

&lt;p&gt;I will make patch to handle that.&lt;/p&gt;</comment>
                            <comment id="196436" author="gerrit" created="Fri, 19 May 2017 06:38:52 +0000"  >&lt;p&gt;Fan Yong (fan.yong@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/27208&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27208&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9514&quot; title=&quot;sanity test 51f times out due to out of memory&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9514&quot;&gt;&lt;del&gt;LU-9514&lt;/del&gt;&lt;/a&gt; ptlrpc: free reply buffer for replay RPC&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 0dfb0b541e104c3a9b6cc5db7910052d91a4984b&lt;/p&gt;</comment>
                            <comment id="196536" author="pjones" created="Sat, 20 May 2017 14:08:33 +0000"  >&lt;p&gt;If I understand correctly, the failing test has been added to always_except for now so this is not a &quot;must have&quot; for 2.10, but can land if ready on time. Otherwise it can be fixed in 2.10.1&lt;/p&gt;</comment>
                            <comment id="198882" author="yong.fan" created="Mon, 12 Jun 2017 10:57:23 +0000"  >&lt;p&gt;The sanity test_51f has been disabled, so it will not affect Maloo tests. But the issue still needs to be fixed, otherwise it will affect the client opened files count. (current implementation will occupy too much RAM on client for opened files, then less files can be opened concurrently by the client)&lt;/p&gt;</comment>
                            <comment id="202563" author="yong.fan" created="Wed, 19 Jul 2017 02:38:55 +0000"  >&lt;p&gt;According to the discussion with Oleg, there are some others can be improved. It is unnecessary for the client to prepare very large buffer for unknown server side RPC reply. In fact, for most of cases, the real reply is quite small. So we can make the client to prepare a relative small buffer the RPC reply. If the MDT finds that it is not big enough, it will tell the client via the RPC reply. Then the client can prepare larger buffer and retry the RPC. In theory, some others may changed the target file and cause the second RPC failed again, but such case will be very rare, so retrying RPC should not bring too much performance trouble.&lt;/p&gt;</comment>
                            <comment id="202568" author="gerrit" created="Wed, 19 Jul 2017 03:28:01 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/27208/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27208/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9514&quot; title=&quot;sanity test 51f times out due to out of memory&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9514&quot;&gt;&lt;del&gt;LU-9514&lt;/del&gt;&lt;/a&gt; ptlrpc: free reply buffer earlier for open RPC&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: c8e3992acf3039b2824725d41f90d9a3be3be921&lt;/p&gt;</comment>
                            <comment id="202594" author="pjones" created="Wed, 19 Jul 2017 03:29:54 +0000"  >&lt;p&gt;Landed for 2.11&lt;/p&gt;</comment>
                            <comment id="203600" author="gerrit" created="Wed, 26 Jul 2017 16:54:30 +0000"  >&lt;p&gt;Minh Diep (minh.diep@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/28217&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/28217&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9514&quot; title=&quot;sanity test 51f times out due to out of memory&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9514&quot;&gt;&lt;del&gt;LU-9514&lt;/del&gt;&lt;/a&gt; ptlrpc: free reply buffer earlier for open RPC&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: be8e18e899c029be0ff85c044c7e7da80c12523d&lt;/p&gt;</comment>
                            <comment id="204381" author="gerrit" created="Thu, 3 Aug 2017 21:40:35 +0000"  >&lt;p&gt;John L. Hammond (john.hammond@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/28217/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/28217/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9514&quot; title=&quot;sanity test 51f times out due to out of memory&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9514&quot;&gt;&lt;del&gt;LU-9514&lt;/del&gt;&lt;/a&gt; ptlrpc: free reply buffer earlier for open RPC&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: a9d4e4b6a17bec691ce32460fdf9057ec080a738&lt;/p&gt;</comment>
                            <comment id="246307" author="bzzz" created="Wed, 24 Apr 2019 17:00:14 +0000"  >&lt;p&gt;still hitting this locally sometimes, on master.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="33314">LU-7473</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="48917">LU-10155</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="27743">LU-5964</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzzcwv:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>