<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:07:06 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7230] memory leak in sanityn.sh 90 &amp; 91</title>
                <link>https://jira.whamcloud.com/browse/LU-7230</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@testnode tests]# MDSCOUNT=4 sh llmount.sh 
Stopping clients: testnode /mnt/lustre (opts:)
Stopping clients: testnode /mnt/lustre2 (opts:)
Loading modules from /work/lustre-release_new/lustre/tests/..
detected 8 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=vfstrace rpctrace dlmtrace neterror ha config 		      ioctl super lfsck
subsystem_debug=all -lnet -lnd -pinger
quota/lquota options: &apos;hash_lqs_cur_bits=3&apos;
Formatting mgs, mds, osts
Format mds1: /tmp/lustre-mdt1
Format mds2: /tmp/lustre-mdt2
Format mds3: /tmp/lustre-mdt3
Format mds4: /tmp/lustre-mdt4
Format ost1: /tmp/lustre-ost1
Format ost2: /tmp/lustre-ost2
Checking servers environments
Checking clients testnode environments
Loading modules from /work/lustre-release_new/lustre/tests/..
detected 8 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=vfstrace rpctrace dlmtrace neterror ha config 		      ioctl super lfsck
subsystem_debug=all -lnet -lnd -pinger
Setup mgs, mdt, osts
Starting mds1:   -o loop /tmp/lustre-mdt1 /mnt/mds1
Started lustre-MDT0000
Starting mds2:   -o loop /tmp/lustre-mdt2 /mnt/mds2
Started lustre-MDT0001
Starting mds3:   -o loop /tmp/lustre-mdt3 /mnt/mds3
Started lustre-MDT0002
Starting mds4:   -o loop /tmp/lustre-mdt4 /mnt/mds4
Started lustre-MDT0003
Starting ost1:   -o loop /tmp/lustre-ost1 /mnt/ost1
Started lustre-OST0000
Starting ost2:   -o loop /tmp/lustre-ost2 /mnt/ost2
Started lustre-OST0001
Starting client: testnode:  -o user_xattr,flock testnode@tcp:/lustre /mnt/lustre
Using TIMEOUT=20
seting jobstats to procname_uid
Setting lustre.sys.jobid_var from disable to procname_uid
Waiting 90 secs for update
Updated after 3s: wanted &apos;procname_uid&apos; got &apos;procname_uid&apos;
disable quota as required
[root@testnode tests]# MDSCOUNT=4 ONLY=&quot;90 91&quot; sh sanityn.sh
Logging to shared log directory: /tmp/test_logs/1443440481
Starting client testnode:  -o user_xattr,flock testnode@tcp:/lustre /mnt/lustre2
Started clients testnode: 
testnode@tcp:/lustre on /mnt/lustre2 type lustre (rw,user_xattr,flock)
testnode: Checking config lustre mounted on /mnt/lustre
Checking servers environments
Checking clients testnode environments
Using TIMEOUT=20
disable quota as required
osd-ldiskfs.track_declares_assert=1
1+0 records in
1+0 records out
1048576 bytes (1.0 MB) copied, 0.073335 s, 14.3 MB/s
running as uid/gid/euid/egid 500/500/500/500, groups:
 [touch] [/mnt/lustre/d0_runas_test/f6126]
excepting tests: 14b 18c 19 28 29 35
skipping tests SLOW=no: 33a
192.168.1.61@tcp:/lustre /mnt/lustre2 lustre rw,flock,user_xattr 0 0


== sanityn test 90: open/create and unlink striped directory == 04:41:22 (1443440482)
start pid 7559 to open/create under striped directory
start pid 7560 to unlink striped directory
sanityn.sh: line 3395:  7559 Terminated              ( cd $DIR1; while true; do
    $LFS mkdir -c$MDSCOUNT $tdir &amp;gt; /dev/null 2&amp;gt;&amp;amp;1; touch $tdir/f{0..3} &amp;gt; /dev/null 2&amp;gt;&amp;amp;1;
done )
sanityn.sh: line 3395:  7560 Terminated              ( cd $DIR2; while true; do
    rm -rf $tdir &amp;gt; /dev/null 2&amp;gt;&amp;amp;1;
done )
Resetting fail_loc on all nodes...done.
PASS 90 (180s)

== sanityn test 91: chmod and unlink striped directory == 04:44:22 (1443440662)
start pid 83746 to chmod striped directory
start pid 83747 to unlink striped directory
sanityn.sh: line 3432: 83746 Terminated              ( cd $DIR1; while true; do
    $LFS mkdir -c$MDSCOUNT $tdir &amp;gt; /dev/null 2&amp;gt;&amp;amp;1; chmod go+w $tdir &amp;gt; /dev/null 2&amp;gt;&amp;amp;1;
done )
sanityn.sh: line 3432: 83747 Terminated              ( cd $DIR2; while true; do
    rm -rf $tdir &amp;gt; /dev/null 2&amp;gt;&amp;amp;1;
done )
Resetting fail_loc on all nodes...done.
PASS 91 (180s)
cleanup: ======================================================
== sanityn test complete, duration 362 sec == 04:47:23 (1443440843)
Stopping clients: testnode /mnt/lustre2 (opts:)
[root@testnode tests]# MDSCOUNT=4 sh llmountcleanup.sh 
Stopping clients: testnode /mnt/lustre (opts:-f)
Stopping client testnode /mnt/lustre opts:-f
Stopping clients: testnode /mnt/lustre2 (opts:-f)
Stopping /mnt/mds1 (opts:-f) on testnode
Stopping /mnt/mds2 (opts:-f) on testnode
Stopping /mnt/mds3 (opts:-f) on testnode
Stopping /mnt/mds4 (opts:-f) on testnode
Stopping /mnt/ost1 (opts:-f) on testnode
Stopping /mnt/ost2 (opts:-f) on testnode
waited 0 for 31 ST ost OSS OSS_uuid 0
LustreError: 19915:0:(class_obd.c:638:cleanup_obdclass()) obd_memory max: 947765874, leaked: 4111712

Memory leaks detected
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="32388">LU-7230</key>
            <summary>memory leak in sanityn.sh 90 &amp; 91</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="di.wang">Di Wang</assignee>
                                    <reporter username="di.wang">Di Wang</reporter>
                        <labels>
                    </labels>
                <created>Wed, 30 Sep 2015 01:34:06 +0000</created>
                <updated>Tue, 3 Nov 2015 22:01:34 +0000</updated>
                            <resolved>Tue, 3 Nov 2015 22:01:34 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                    <fixVersion>Lustre 2.8.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="128848" author="gerrit" created="Wed, 30 Sep 2015 01:56:23 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/16677&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16677&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7230&quot; title=&quot;memory leak in sanityn.sh 90 &amp;amp; 91&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7230&quot;&gt;&lt;del&gt;LU-7230&lt;/del&gt;&lt;/a&gt; llite: clear dir stripe md in ll_iget&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 6a7dd3bcb58d65f3e8fe1dd4c8973f5a88f21b5b&lt;/p&gt;</comment>
                            <comment id="129082" author="di.wang" created="Fri, 2 Oct 2015 01:22:06 +0000"  >&lt;p&gt;This patch fix some quite serious top-&amp;gt;sub transaction issue. (the original implementation forget to transfer th_local/th_tag to sub_trans for local transaction). Let&apos;s make it critical.&lt;/p&gt;</comment>
                            <comment id="132541" author="gerrit" created="Tue, 3 Nov 2015 20:02:02 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/16677/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16677/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7230&quot; title=&quot;memory leak in sanityn.sh 90 &amp;amp; 91&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7230&quot;&gt;&lt;del&gt;LU-7230&lt;/del&gt;&lt;/a&gt; llite: clear dir stripe md in ll_iget&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: d05051c34077c74e3c08f3723bf6084554c0daf8&lt;/p&gt;</comment>
                            <comment id="132556" author="pjones" created="Tue, 3 Nov 2015 22:01:34 +0000"  >&lt;p&gt;Landed for 2.8&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="32389">LU-7231</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxp2v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>