<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 03:20:33 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-15700] conf-sanity test 115 does not cleanup on mkfs failure</title>
                <link>https://jira.whamcloud.com/browse/LU-15700</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;conf-sanity test 115 does not cleanup when it runs out of space leading to all subsequent tests failing due to no space left in /tmp&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== conf-sanity test 115: Access large xattr with inodes number over 2TB ========================================================== 05:31:00 (1648373460)

oleg219-server: 
oleg219-server: mkfs.lustre FATAL: Unable to build fs /dev/loop0 (256)
oleg219-server: 
oleg219-server: mkfs.lustre FATAL: mkfs failed 256
pdsh@oleg219-client: oleg219-server: ssh exited with exit code 1

   Permanent disk data:
Target:     lustre:MDT0000
Index:      0
Lustre FS:  lustre
Mount type: ldiskfs
Flags:      0x65
              (MDT MGS first_time update )
Persistent mount opts: user_xattr,errors=remount-ro
Parameters: sys.timeout=20 mdt.identity_upcall=/home/green/git/lustre-release/lustre/utils/l_getidentity

device size = 3145728MB
formatting backing filesystem ldiskfs on /dev/loop0
	target name   lustre:MDT0000
	kilobytes     3221225472
	options        -N 2247484000  -J size=4096 -I 1024 -q -O ea_inode,^resize_inode,meta_bg,dirdata,uninit_bg,^extents,dir_nlink,quota,project,huge_file,large_dir,flex_bg -E lazy_itable_init,lazy_journal_init=&quot;0&quot; -F
mkfs_cmd = mke2fs -j -b 4096 -L lustre:MDT0000  -N 2247484000  -J size=4096 -I 1024 -q -O ea_inode,^resize_inode,meta_bg,dirdata,uninit_bg,^extents,dir_nlink,quota,project,huge_file,large_dir,flex_bg -E lazy_itable_init,lazy_journal_init=&quot;0&quot; -F /dev/loop0 3221225472k

 SKIP: conf-sanity test_115 format large MDT failed

== conf-sanity test 116: big size MDT support ============ 05:45:12 (1648374312)
/usr/sbin/mkfs.xfs
Stopping clients: oleg219-client.virtnet /mnt/lustre (opts:)
Stopping clients: oleg219-client.virtnet /mnt/lustre2 (opts:)
Loading modules from /home/green/git/lustre-release/lustre
detected 4 online CPUs by sysfs
Force libcfs to create 2 CPU partitions
ptlrpc/ptlrpc options: &apos;lbug_on_grant_miscount=1&apos;
quota/lquota options: &apos;hash_lqs_cur_bits=3&apos;
loading modules on: &apos;oleg219-server&apos;
oleg219-server: /home/green/git/lustre-release/lustre/tests/test-framework.sh: line 469: cannot create temp file for here-document: No space left on device
...
 conf-sanity test_116: @@@@@@ FAIL: mkfs temporary xfs image &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="69281">LU-15700</key>
            <summary>conf-sanity test 115 does not cleanup on mkfs failure</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="green">Oleg Drokin</reporter>
                        <labels>
                    </labels>
                <created>Sun, 27 Mar 2022 15:34:17 +0000</created>
                <updated>Wed, 4 May 2022 23:00:54 +0000</updated>
                            <resolved>Wed, 4 May 2022 23:00:54 +0000</resolved>
                                    <version>Lustre 2.15.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="330423" author="JIRAUSER17312" created="Mon, 28 Mar 2022 17:38:49 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=artem_blagodarenko&quot; class=&quot;user-hover&quot; rel=&quot;artem_blagodarenko&quot;&gt;artem_blagodarenko&lt;/a&gt; &amp;#8211; heads up.&lt;/p&gt;</comment>
                            <comment id="330424" author="JIRAUSER17312" created="Mon, 28 Mar 2022 17:43:11 +0000"  >&lt;p&gt;&quot;Test was being skipped in Janitor testing, probably should not block 2.15 release&quot;&lt;/p&gt;</comment>
                            <comment id="332300" author="artem_blagodarenko" created="Tue, 19 Apr 2022 10:31:39 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=cfaber&quot; class=&quot;user-hover&quot; rel=&quot;cfaber&quot;&gt;cfaber&lt;/a&gt;, it looks strange. There is&#160;cleanup_115() that executed on skip and do &quot;rm -f $TMP/$tdir/lustre-mdt&quot; so large file should be deleted.&lt;/p&gt;

&lt;p&gt;There can I see more logs for such a test session?&#160;&lt;/p&gt;

&lt;p&gt;Thanks.&lt;/p&gt;</comment>
                            <comment id="332328" author="adilger" created="Tue, 19 Apr 2022 15:36:20 +0000"  >&lt;p&gt;Artem, you could probably reproduce this easily in Oleg&apos;s Gerrit Janitor test system by pushing a patch that modified test_115 and test_116. &lt;/p&gt;</comment>
                            <comment id="332341" author="gerrit" created="Tue, 19 Apr 2022 17:02:13 +0000"  >&lt;p&gt;&quot;Artem Blagodarenko &amp;lt;artem.blagodarenko@hpe.com&amp;gt;&quot; uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/47093&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/47093&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15700&quot; title=&quot;conf-sanity test 115 does not cleanup on mkfs failure&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15700&quot;&gt;&lt;del&gt;LU-15700&lt;/del&gt;&lt;/a&gt; tests: check how conf_sanity_115 cleanup works&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 037e49dbf030101b5eb0312f3b5b8656948bdcb6&lt;/p&gt;</comment>
                            <comment id="332342" author="artem_blagodarenko" created="Tue, 19 Apr 2022 17:10:23 +0000"  >&lt;p&gt;&#160;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=adilger&quot; class=&quot;user-hover&quot; rel=&quot;adilger&quot;&gt;adilger&lt;/a&gt; , thank you very much for the idea. I have created &#160;&lt;a href=&quot;https://review.whamcloud.com/47093&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/47093&lt;/a&gt; that force &quot;skip&quot; in conf_sanity_115 and modify conf_sanity_116. Let&apos;s see resutls.&lt;/p&gt;</comment>
                            <comment id="333122" author="artem_blagodarenko" created="Wed, 27 Apr 2022 12:46:05 +0000"  >&lt;p&gt;The issue can be addressed by &lt;a href=&quot;https://review.whamcloud.com/47153&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/47153&lt;/a&gt; from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15789&quot; title=&quot;conf-sanity  test_115()  cleanup_115() defect&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15789&quot;&gt;&lt;del&gt;LU-15789&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="333824" author="adilger" created="Wed, 4 May 2022 23:00:54 +0000"  >&lt;p&gt;Use &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15789&quot; title=&quot;conf-sanity  test_115()  cleanup_115() defect&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15789&quot;&gt;&lt;del&gt;LU-15789&lt;/del&gt;&lt;/a&gt; for the fix.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="54751">LU-11915</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="70035">LU-15789</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i02ltz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>