<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:17:55 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-8480] Server syslog: ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 69347328 != fo_tot_granted 102901760</title>
                <link>https://jira.whamcloud.com/browse/LU-8480</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Error message seen in system log on at least one production server:&lt;br/&gt;
(running lustre-2.7.2-1.1nasS_mofed32v1)&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;May 20 18:38:58 nbp1-oss1 kernel: LustreError: 38115:0:(ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 98847872 != fo_tot_granted 100945024
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Reproduced on test system, (running lustre-2.7.2-2nasS_mofed32v1.el67.20160517v2), by mounting just 8 clients, running IOR write from 1, IOR read from 1, mdtest on the other 6.  Then upon unmount of all 8 clients simultaneously, get the messages on each (of the two) OSS nodes:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Aug  5 08:06:36 service320 kernel: LustreError: 76476:0:(ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 69347328 != fo_tot_granted 102901760
Aug  5 08:06:36 service323 kernel: LustreError: 76171:0:(ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 69347328 != fo_tot_granted 102901760
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;What does this message mean, and is it cause for major concern?&lt;/p&gt;

&lt;p&gt;We are testing in preparation for upgrade to all remaining production file systems on 8/15, so need to know whether to proceed ASAP.&lt;/p&gt;</description>
                <environment>Mostly CentOS-6.8, with 6.7 kernel 2.6.32_573.26.1.el6&lt;br/&gt;
Lustre 2.7.2-2nasS_mofed32v1.el67.20160517v2</environment>
        <key id="38640">LU-8480</key>
            <summary>Server syslog: ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 69347328 != fo_tot_granted 102901760</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="yujian">Jian Yu</assignee>
                                    <reporter username="ndauchy">Nathan Dauchy</reporter>
                        <labels>
                    </labels>
                <created>Fri, 5 Aug 2016 20:57:46 +0000</created>
                <updated>Thu, 15 Jun 2017 23:44:53 +0000</updated>
                            <resolved>Mon, 29 Aug 2016 13:02:03 +0000</resolved>
                                    <version>Lustre 2.7.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="161154" author="adilger" created="Mon, 8 Aug 2016 17:48:49 +0000"  >&lt;p&gt;It looks like the delta between the grant amounts is about 32MB, which is the amount of grant held by a single client after it has done some writes to the OST.  It looks like there is a race window between traversing all of the client exports to accumulate their grant, and the running total that is kept for the whole OST when a client is being unmounted, and is possible to be triggered when multiple clients are being unmounted.  This doesn&apos;t seem harmful in itself, and for systems with more than 100 clients this check is skipped because it slows down the unmount too much (it is O(n^2) to accumulate the per-export grants as each client unmounts).&lt;/p&gt;

&lt;p&gt;It looks like there may be a very simple fix, for which I can push a patch if you could give it a test.&lt;/p&gt;</comment>
                            <comment id="161155" author="gerrit" created="Mon, 8 Aug 2016 17:51:37 +0000"  >&lt;p&gt;Andreas Dilger (andreas.dilger@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/21813&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/21813&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8480&quot; title=&quot;Server syslog: ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 69347328 != fo_tot_granted 102901760&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8480&quot;&gt;&lt;del&gt;LU-8480&lt;/del&gt;&lt;/a&gt; ofd: hold obd_dev_lock across grant comparison&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 14d7415e9048a287aabde9a5d03dd2d9b6c48ae9&lt;/p&gt;</comment>
                            <comment id="161174" author="ndauchy" created="Mon, 8 Aug 2016 18:53:56 +0000"  >&lt;p&gt;Thanks for the quick review.  Since this shouldn&apos;t be harmful, and is skipped for more than 100 clients anyway, we may not rush to get a rebuild into testing prior to the upgrade planned for next week.  Will test as soon as time permits though.&lt;/p&gt;

&lt;p&gt;NOTE:  we will need a backport to 2.7.2.&lt;/p&gt;</comment>
                            <comment id="161176" author="pjones" created="Mon, 8 Aug 2016 19:01:02 +0000"  >&lt;p&gt;Jian&lt;/p&gt;

&lt;p&gt;Could you please port this fix to 2.7 FE once it has landed to master&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="161235" author="yujian" created="Tue, 9 Aug 2016 05:28:46 +0000"  >&lt;p&gt;Here is the back-ported patch for Lustre 2.7 FE: &lt;a href=&quot;http://review.whamcloud.com/22018&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/22018&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="162811" author="ndauchy" created="Tue, 23 Aug 2016 13:55:52 +0000"  >&lt;p&gt;For some reason, I am unable to view the backported patch.  Is the link correct, or is there perhaps a permissions issue with my gerrit account?&lt;/p&gt;</comment>
                            <comment id="162813" author="pjones" created="Tue, 23 Aug 2016 14:16:14 +0000"  >&lt;p&gt;Nathan&lt;/p&gt;

&lt;p&gt;It must be the latter. Please send me an email about this issue and I&apos;ll get it sorted out.&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="163366" author="gerrit" created="Mon, 29 Aug 2016 03:52:12 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/21813/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/21813/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8480&quot; title=&quot;Server syslog: ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 69347328 != fo_tot_granted 102901760&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8480&quot;&gt;&lt;del&gt;LU-8480&lt;/del&gt;&lt;/a&gt; ofd: hold obd_dev_lock across grant comparison&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: cf54bda257dc3287458a9157bb4647f05d8f8469&lt;/p&gt;</comment>
                            <comment id="163391" author="pjones" created="Mon, 29 Aug 2016 13:02:03 +0000"  >&lt;p&gt;Landed for 2.9&lt;/p&gt;</comment>
                            <comment id="163460" author="ndauchy" created="Mon, 29 Aug 2016 19:10:01 +0000"  >&lt;p&gt;Please re-open until the backport patch lands to 2.7 FE.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                                        </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzyjn3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>