<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:33:17 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-10238] adding new OSTs causes quota reporting error</title>
                <link>https://jira.whamcloud.com/browse/LU-10238</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We have a Lustre 2.10.0 filesystem which was built with two OSSes containing 5 OSTs each. Last week I added a third OSS (same exact hardware, slightly newer OS software except the kernel &amp;amp; lustre). When I created the OSTs with mkfs.lustre, the filesystem seemed to grow correctly. We currently only set and enforce group quotas.&lt;/p&gt;

&lt;p&gt;Later that day, we noticed the output of `lfs quota -g $GROUP /center1` was showing bad values and an error message. Here&apos;s an example.&lt;/p&gt;

&lt;p&gt;chinook02:PENGUIN$ sudo lfs quota -g penguin /center1&lt;br/&gt;
Disk quotas for grp penguin (gid 12738):&lt;br/&gt;
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace&lt;br/&gt;
       /center1   &lt;span class=&quot;error&quot;&gt;&amp;#91;214&amp;#93;&lt;/span&gt;  1073741824 1181116006       -      13       0       0       -&lt;br/&gt;
Some errors happened when getting quota info. Some devices may be not working or deactivated. The data in &quot;[]&quot; is inaccurate.&lt;/p&gt;

&lt;p&gt;We found a workaround. As soon as the group has data written to the new OSTs, `lfs quota` seems to work fine.&lt;/p&gt;

&lt;p&gt;chinook02:PENGUIN$ lfs setstripe -i -1 -c -1 loforbes&lt;br/&gt;
chinook02:PENGUIN$ dd of=loforbes/testfile if=/dev/urandom bs=1M count=15&lt;br/&gt;
15+0 records in&lt;br/&gt;
15+0 records out&lt;br/&gt;
15728640 bytes (16 MB) copied, 1.80694 s, 8.7 MB/s&lt;br/&gt;
chinook02:PENGUIN$ sudo lfs quota -g penguin /center1&lt;br/&gt;
Disk quotas for grp penguin (gid 12738):&lt;br/&gt;
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace&lt;br/&gt;
       /center1 671997883  1073741824 1181116006       -      13       0       0       -&lt;br/&gt;
chinook02:PENGUIN$ lfs getstripe loforbes/testfile &lt;br/&gt;
loforbes/testfile&lt;br/&gt;
lmm_stripe_count:  15&lt;br/&gt;
lmm_stripe_size:   1048576&lt;br/&gt;
lmm_pattern:       1&lt;br/&gt;
lmm_layout_gen:    0&lt;br/&gt;
lmm_stripe_offset: 12&lt;br/&gt;
	obdidx		 objid		 objid		 group&lt;br/&gt;
	    12	         31981	       0x7ced	             0&lt;br/&gt;
	     7	      62233208	    0x3b59a78	             0&lt;br/&gt;
	    14	         32068	       0x7d44	             0&lt;br/&gt;
	     8	      72183233	    0x44d6dc1	             0&lt;br/&gt;
	    10	         31854	       0x7c6e	             0&lt;br/&gt;
	    11	         31849	       0x7c69	             0&lt;br/&gt;
	     2	      68917015	    0x41b9717	             0&lt;br/&gt;
	     5	      71171215	    0x43dfc8f	             0&lt;br/&gt;
	     1	      69395583	    0x422e47f	             0&lt;br/&gt;
	    13	         32088	       0x7d58	             0&lt;br/&gt;
	     9	      68211489	    0x410d321	             0&lt;br/&gt;
	     6	      70389457	    0x4320ed1	             0&lt;br/&gt;
	     4	      70225352	    0x42f8dc8	             0&lt;br/&gt;
	     3	      66783438	    0x3fb08ce	             0&lt;br/&gt;
	     0	      65674625	    0x3ea1d81	             0&lt;/p&gt;

&lt;p&gt;We figured out it&apos;s not really necessary to have data on the 10 original OSTs, just the 5 new ones for this to work. I&apos;ve implemented this workaround for all projects using our lustre filesystem.&lt;/p&gt;

&lt;p&gt;Before implementing the workaround, we tried &quot;deleting&quot; a group&apos;s quota and recreating. That didn&apos;t seem to impact the issue. We also tried unmounting and remounting the filesystem on a client. Again, no change. Removing all files owned by a group that have data on the new OSTs results in `lfs quota` showing the error again.&lt;/p&gt;

&lt;p&gt;We are considering a Lustre 2.10.1 update sometime soon.&lt;/p&gt;

&lt;p&gt;Regards,&lt;br/&gt;
-liam&lt;/p&gt;</description>
                <environment>CentOS 7 servers&lt;br/&gt;
&amp;nbsp;&amp;nbsp;kernel-3.10.0-514.21.1.el7_lustre.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;lustre-2.10.0-1.el7.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;lustre-dkms-2.10.0-1.el7.noarch&lt;br/&gt;
&amp;nbsp;&amp;nbsp;lustre-osd-zfs-mount-2.10.0-1.el7.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;lustre-resource-agents-2.10.0-1.el7.x86_64&lt;br/&gt;
CentOS 6 clients&lt;br/&gt;
&amp;nbsp;&amp;nbsp;lustre-client-2.10.0-1.el6.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;lustre-client-dkms-2.10.0-1.el6.noarch&lt;br/&gt;
ZFS for OSTs &amp;amp; MDT&lt;br/&gt;
&amp;nbsp;&amp;nbsp;libzfs2-0.7.3-1.el7_3.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;libzfs2-devel-0.7.3-1.el7_3.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;zfs-0.7.3-1.el7_3.x86_64&lt;br/&gt;
&amp;nbsp;&amp;nbsp;zfs-dkms-0.7.3-1.el7_3.noarch&lt;br/&gt;
&amp;nbsp;&amp;nbsp;zfs-release-1-4.el7_3.centos.noarch&lt;br/&gt;
DKMS kernel modules</environment>
        <key id="49306">LU-10238</key>
            <summary>adding new OSTs causes quota reporting error</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="hongchao.zhang">Hongchao Zhang</assignee>
                                    <reporter username="loforbes">Liam Forbes</reporter>
                        <labels>
                    </labels>
                <created>Mon, 13 Nov 2017 23:41:28 +0000</created>
                <updated>Fri, 2 Feb 2018 04:17:12 +0000</updated>
                                            <version>Lustre 2.10.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="216868" author="jamesanunez" created="Wed, 20 Dec 2017 18:22:27 +0000"  >&lt;p&gt;Hongchao - &lt;/p&gt;

&lt;p&gt;Would you please look into this issue?&lt;/p&gt;

&lt;p&gt;Thank you&lt;/p&gt;</comment>
                            <comment id="217241" author="hongchao.zhang" created="Fri, 29 Dec 2017 08:53:58 +0000"  >&lt;p&gt;Hi Liam,&lt;/p&gt;

&lt;p&gt;I can&apos;t reproduce the issue in my local VMs, could you please attach the logs (syslog and debug log) when the issue occurred?&lt;br/&gt;
Thanks!&lt;/p&gt;

&lt;p&gt;btw, please add quota to the debug log by &quot;lctl set_param debug=+quota&quot;.&lt;/p&gt;</comment>
                            <comment id="218862" author="loforbes" created="Mon, 22 Jan 2018 22:57:51 +0000"  >&lt;p&gt;Hongchao,&lt;/p&gt;

&lt;p&gt;I&apos;m attaching the syslog file from the two days when we added the new OSS (oss09) to the filesystem. Unfortunately, I can&apos;t say exactly what time that occurred. Also unfortunately, I don&apos;t seem to have the syslogs from that OSS on that day either.&lt;/p&gt;

&lt;p&gt;Here are the system logs that occur when we get the error message in the `lfs quota` output.&lt;/p&gt;

&lt;p&gt;From a client:&lt;br/&gt;
Jan 22 13:51:13 chinook02 kernel: LustreError: 30907:0:(osc_quota.c:291:osc_quotactl()) ptlrpc_queue_wait failed, rc: -2&lt;br/&gt;
Jan 22 13:51:13 chinook02 kernel: LustreError: 30907:0:(osc_quota.c:291:osc_quotactl()) Skipped 4 previous similar messages&lt;/p&gt;

&lt;p&gt;No messages occur on the MDS or OSS. Could this be an LNET issue?&lt;/p&gt;

&lt;p&gt;Regards,&lt;br/&gt;
-liam &lt;span class=&quot;nobr&quot;&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/attachment/29270/29270_mdsLogs.tar.gz&quot; title=&quot;mdsLogs.tar.gz attached to LU-10238&quot;&gt;mdsLogs.tar.gz&lt;sup&gt;&lt;img class=&quot;rendericon&quot; src=&quot;https://jira.whamcloud.com/images/icons/link_attachment_7.gif&quot; height=&quot;7&quot; width=&quot;7&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/sup&gt;&lt;/a&gt;&lt;/span&gt; &lt;/p&gt;
</comment>
                            <comment id="219813" author="hongchao.zhang" created="Fri, 2 Feb 2018 04:17:12 +0000"  >&lt;p&gt;Hi Liam,&lt;/p&gt;

&lt;p&gt;The issue is related to the OSS, could you please get the quota usage of some non-existing group (say, 20000) on your site&lt;br/&gt;
to check whether this issue can be triggered or not, and if so, please collect the logs (debug log by running &apos;lctl dk &amp;gt;log_fiile&apos;)&lt;br/&gt;
on the OSS? Thanks!&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="29270" name="mdsLogs.tar.gz" size="1424421" author="loforbes" created="Mon, 22 Jan 2018 22:57:47 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzznqn:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>