<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:37:59 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3910] Interop 2.4.0&lt;-&gt;2.5 failure on test suite parallel-scale-nfsv4 test_iorssf: MDS OOM</title>
                <link>https://jira.whamcloud.com/browse/LU-3910</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;http://maloo.whamcloud.com/test_sets/f9f7ec36-15c3-11e3-a83f-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/f9f7ec36-15c3-11e3-a83f-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_iorssf failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;test failed to respond and timed out&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;MDS console:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;12:37:48:Lustre: DEBUG MARKER: == parallel-scale-nfsv4 test iorssf: iorssf == 12:37:39 (1378323459)
12:37:48:Lustre: DEBUG MARKER: lfs setstripe /mnt/lustre/d0.ior.ssf -c -1
12:38:00:Lustre: MGS: Client a0cabda3-b9a7-2ed3-58fd-f9e8cebfe558 (at 10.10.4.199@tcp) reconnecting
12:38:00:Lustre: lustre-MDT0000: Client lustre-MDT0000-lwp-OST0000_UUID (at 10.10.4.199@tcp) reconnecting
12:43:36:ptlrpcd_0: page allocation failure. order:1, mode:0x40
12:43:36:Pid: 2733, comm: ptlrpcd_0 Not tainted 2.6.32-358.18.1.el6_lustre.x86_64 #1
12:43:37:Call Trace:
12:43:37: [&amp;lt;ffffffff8112c257&amp;gt;] ? __alloc_pages_nodemask+0x757/0x8d0
12:43:37: [&amp;lt;ffffffffa0767b0f&amp;gt;] ? ptlrpc_set_add_new_req+0xcf/0x150 [ptlrpc]
12:43:37: [&amp;lt;ffffffff81166d92&amp;gt;] ? kmem_getpages+0x62/0x170
12:43:37: [&amp;lt;ffffffff811679aa&amp;gt;] ? fallback_alloc+0x1ba/0x270
12:43:37: [&amp;lt;ffffffff811673ff&amp;gt;] ? cache_grow+0x2cf/0x320
12:43:37: [&amp;lt;ffffffff81167729&amp;gt;] ? ____cache_alloc_node+0x99/0x160
12:43:37: [&amp;lt;ffffffffa053cea7&amp;gt;] ? LNetMDAttach+0x157/0x5a0 [lnet]
12:43:37: [&amp;lt;ffffffff811684f9&amp;gt;] ? __kmalloc+0x189/0x220
12:43:38: [&amp;lt;ffffffffa053cea7&amp;gt;] ? LNetMDAttach+0x157/0x5a0 [lnet]
12:43:38: [&amp;lt;ffffffffa0775935&amp;gt;] ? ptlrpc_register_bulk+0x265/0x9d0 [ptlrpc]
12:43:38: [&amp;lt;ffffffffa07777f2&amp;gt;] ? ptl_send_rpc+0x232/0xc40 [ptlrpc]
12:43:38: [&amp;lt;ffffffff81281b74&amp;gt;] ? snprintf+0x34/0x40
12:43:38: [&amp;lt;ffffffffa0489951&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
12:43:38: [&amp;lt;ffffffffa076c3f4&amp;gt;] ? ptlrpc_send_new_req+0x454/0x790 [ptlrpc]
12:43:38: [&amp;lt;ffffffffa0770148&amp;gt;] ? ptlrpc_check_set+0x888/0x1b30 [ptlrpc]
12:43:38: [&amp;lt;ffffffffa079bb3b&amp;gt;] ? ptlrpcd_check+0x53b/0x560 [ptlrpc]
12:43:38: [&amp;lt;ffffffff8109715c&amp;gt;] ? remove_wait_queue+0x3c/0x50
12:43:38: [&amp;lt;ffffffffa079bfc0&amp;gt;] ? ptlrpcd+0x190/0x380 [ptlrpc]
12:43:38: [&amp;lt;ffffffff81063410&amp;gt;] ? default_wake_function+0x0/0x20
12:43:38: [&amp;lt;ffffffffa079be30&amp;gt;] ? ptlrpcd+0x0/0x380 [ptlrpc]
12:43:39: [&amp;lt;ffffffff81096a36&amp;gt;] ? kthread+0x96/0xa0
12:43:40: [&amp;lt;ffffffff8100c0ca&amp;gt;] ? child_rip+0xa/0x20
12:43:40: [&amp;lt;ffffffff810969a0&amp;gt;] ? kthread+0x0/0xa0
12:43:40: [&amp;lt;ffffffff8100c0c0&amp;gt;] ? child_rip+0x0/0x20
12:43:41:Mem-Info:
12:43:41:Node 0 DMA per-cpu:
12:43:41:CPU    0: hi:    0, btch:   1 usd:   0
12:43:41:Node 0 DMA32 per-cpu:
12:43:41:CPU    0: hi:  186, btch:  31 usd: 202
12:43:42:active_anon:2089 inactive_anon:2411 isolated_anon:0
12:43:42: active_file:60821 inactive_file:275748 isolated_file:32
12:43:42: unevictable:0 dirty:34820 writeback:16128 unstable:0
12:43:42: free:17978 slab_reclaimable:5665 slab_unreclaimable:87698
12:43:42: mapped:2558 shmem:41 pagetables:793 bounce:0
12:43:43:Node 0 DMA free:8276kB min:332kB low:412kB high:496kB active_anon:0kB inactive_anon:0kB active_file:304kB inactive_file:5516kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15324kB mlocked:0kB dirty:0kB writeback:0kB mapped:0kB shmem:0kB slab_reclaimable:16kB slab_unreclaimable:1620kB kernel_stack:0kB pagetables:0kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? no
12:43:43:lowmem_reserve[]: 0 2003 2003 2003
12:43:43:Node 0 DMA32 free:63636kB min:44720kB low:55900kB high:67080kB active_anon:8356kB inactive_anon:9644kB active_file:242980kB inactive_file:1097476kB unevictable:0kB isolated(anon):0kB isolated(file):128kB present:2052064kB mlocked:0kB dirty:139280kB writeback:64512kB mapped:10232kB shmem:164kB slab_reclaimable:22644kB slab_unreclaimable:349172kB kernel_stack:2024kB pagetables:3172kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:32 all_unreclaimable? no
12:43:43:lowmem_reserve[]: 0 0 0 0
12:43:43:Node 0 DMA: 47*4kB 5*8kB 21*16kB 9*32kB 10*64kB 5*128kB 6*256kB 1*512kB 2*1024kB 1*2048kB 0*4096kB = 8276kB
12:43:43:Node 0 DMA32: 13749*4kB 30*8kB 7*16kB 7*32kB 4*64kB 15*128kB 1*256kB 1*512kB 1*1024kB 0*2048kB 1*4096kB = 63636kB
12:43:43:299076 total pagecache pages
12:43:44:0 pages in swap cache
12:43:44:Swap cache stats: add 0, delete 0, find 0/0
12:43:45:Free swap  = 4128760kB
12:43:45:Total swap = 4128760kB
12:43:45:524284 pages RAM
12:43:45:43669 pages reserved
12:43:46:330512 pages shared
12:43:46:160098 pages non-shared
12:43:46:LNetError: 2733:0:(lib-lnet.h:457:lnet_md_alloc()) LNET: out of memory at /var/lib/jenkins/workspace/lustre-master/arch/x86_64/build_type/server/distro/el6/ib_stack/inkernel/BUILD/BUILD/lustre-2.4.92/lnet/include/lnet/lib-lnet.h:457 (tried to alloc &apos;(md)&apos; = 4208)
12:43:46:LNetError: 2733:0:(lib-lnet.h:457:lnet_md_alloc()) LNET: 55454455 total bytes allocated by lnet
12:43:46:LustreError: 2733:0:(niobuf.c:376:ptlrpc_register_bulk()) lustre-OST0001-osc-ffff88006a434c00: LNetMDAttach failed x1445262125420992/0: rc = -12
12:43:46:Lustre: 2733:0:(client.c:1896:ptlrpc_expire_one_request()) @@@ Request sent has failed due to network error: [sent 0/real 0]  req@ffff880024150400 x1445262125420992/t0(0) o4-&amp;gt;lustre-OST0001-osc-ffff88006a434c00@10.10.4.199@tcp:6/4 lens 488/448 e 0 to 1 dl 0 ref 2 fl Rpc:X/0/ffffffff rc -12/-1
12:43:47:Lustre: lustre-OST0001-osc-ffff88006a434c00: Connection to lustre-OST0001 (at 10.10.4.199@tcp) was lost; in progress operations using this service will wait for recovery to complete
12:43:47:LustreError: 11-0: lustre-OST0001-osc-ffff88006a434c00: Communicating with 10.10.4.199@tcp, operation ost_connect failed with -16.
12:43:47:LustreError: Skipped 1 previous similar message
12:43:47:Lustre: lustre-OST0001-osc-ffff88006a434c00: Connection restored to lustre-OST0001 (at 10.10.4.199@tcp)
12:44:19:nfsd: page allocation failure. order:1, mode:0x40
12:44:19:Pid: 10656, comm: nfsd Not tainted 2.6.32-358.18.1.el6_lustre.x86_64 #1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>client: 2.4.0&lt;br/&gt;
server: lustre-master build # 1652</environment>
        <key id="20861">LU-3910</key>
            <summary>Interop 2.4.0&lt;-&gt;2.5 failure on test suite parallel-scale-nfsv4 test_iorssf: MDS OOM</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Sun, 8 Sep 2013 23:24:40 +0000</created>
                <updated>Thu, 13 Feb 2014 22:15:25 +0000</updated>
                            <resolved>Thu, 13 Feb 2014 22:15:25 +0000</resolved>
                                    <version>Lustre 2.5.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="66088" author="adilger" created="Mon, 9 Sep 2013 16:40:47 +0000"  >&lt;p&gt;There are two problems here:&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;NFS server is running on the MDS, which only has 1.83GB of memory&lt;/li&gt;
	&lt;li&gt;client on NFS server is holding lots of data pages pinned like &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2139&quot; title=&quot;Tracking unstable pages&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2139&quot;&gt;&lt;del&gt;LU-2139&lt;/del&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;I&apos;m going to leave this bug to focus on changing this NFS test to run on MDS nodes with more memory, so that the testing does not fail. &lt;/p&gt;</comment>
                            <comment id="66090" author="adilger" created="Mon, 9 Sep 2013 16:48:12 +0000"  >&lt;p&gt;Please note that interop testing NFS with the MDS as NFS server is not useful, since the Lustre 2.4.0 client will not be used for anything. It would be better to run the NFS server on the Lustre 2.4.0 client, which is more likely how it will be used in real life. This will also help avoid memory problems on the MDS. &lt;/p&gt;

&lt;p&gt;How is a change like this done to the testing system?&lt;/p&gt;</comment>
                            <comment id="77029" author="adilger" created="Thu, 13 Feb 2014 22:08:25 +0000"  >&lt;p&gt;Shows mode:0x40 == __GFP_IO, but missing __GFP_WAIT from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4357&quot; title=&quot;page allocation failure. mode:0x40 caused by missing __GFP_WAIT flag&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4357&quot;&gt;&lt;del&gt;LU-4357&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="22373">LU-4357</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="15971">LU-2139</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzw1dz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>10306</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>