<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:10:23 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-785] Gettting found wrong generation error for the same inode</title>
                <link>https://jira.whamcloud.com/browse/LU-785</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We are getting the following error on our mds server repeatedly. This error continues to appear even after a MDS reboot.&lt;/p&gt;

&lt;p&gt;Oct 21 12:09:18 nbp30-mds kernel: LustreError: 8021:0:(handler.c:275:mds_fid2dentry()) Skipped 599 previous similar messages&lt;br/&gt;
Oct 21 12:19:18 nbp30-mds kernel: LustreError: 8020:0:(handler.c:275:mds_fid2dentry()) found wrong generation: inode 209323401, link: 2, count: 1, generation 3543922068/3529758994&lt;br/&gt;
Oct 21 12:19:18 nbp30-mds kernel: LustreError: 8020:0:(handler.c:275:mds_fid2dentry()) Skipped 599 previous similar messages&lt;br/&gt;
Oct 21 12:29:19 nbp30-mds kernel: LustreError: 9206:0:(handler.c:275:mds_fid2dentry()) found wrong generation: inode 209323401, link: 2, count: 1, generation 3543922068/3529770923&lt;br/&gt;
Oct 21 12:29:19 nbp30-mds kernel: LustreError: 9206:0:(handler.c:275:mds_fid2dentry()) Skipped 598 previous similar messages&lt;br/&gt;
Oct 21 12:39:20 nbp30-mds kernel: LustreError: 9293:0:(handler.c:275:mds_fid2dentry()) found wrong generation: inode 209323401, link: 2, count: 1, generation 3543922068/3529758994&lt;br/&gt;
Oct 21 12:39:20 nbp30-mds kernel: LustreError: 9293:0:(handler.c:275:mds_fid2dentry()) Skipped 599 previous similar messages&lt;br/&gt;
Oct 21 12:49:20 nbp30-mds kernel: LustreError: 9238:0:(handler.c:275:mds_fid2dentry()) found wrong generation: inode 209323401, link: 2, count: 1, generation 3543922068/3529758994&lt;br/&gt;
Oct 21 12:49:20 nbp30-mds kernel: LustreError: 9238:0:(handler.c:275:mds_fid2dentry()) Skipped 599 previous similar messages&lt;br/&gt;
Oct 21 12:59:21 nbp30-mds kernel: LustreError: 9204:0:(handler.c:275:mds_fid2dentry()) found wrong generation: inode 209323401, link: 2, count: 1, generation 3543922068/3529758994&lt;br/&gt;
Oct 21 12:59:21 nbp30-mds kernel: LustreError: 9204:0:(handler.c:275:mds_fid2dentry()) Skipped 598 previous similar messages&lt;/p&gt;</description>
                <environment>THIS IS FOR NASA AMES.&lt;br/&gt;
MDS running &lt;br/&gt;
SUSE Linux Enterprise Server 10 (x86_64)&lt;br/&gt;
VERSION = 10&lt;br/&gt;
PATCHLEVEL = 2&lt;br/&gt;
</environment>
        <key id="12218">LU-785</key>
            <summary>Gettting found wrong generation error for the same inode</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="laisiyao">Lai Siyao</assignee>
                                    <reporter username="mhanafi">Mahmoud Hanafi</reporter>
                        <labels>
                    </labels>
                <created>Fri, 21 Oct 2011 16:06:32 +0000</created>
                <updated>Tue, 29 Oct 2013 18:32:52 +0000</updated>
                            <resolved>Tue, 29 Oct 2013 18:32:52 +0000</resolved>
                                    <version>Lustre 1.8.x (1.8.0 - 1.8.5)</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="21657" author="pjones" created="Fri, 21 Oct 2011 16:10:49 +0000"  >&lt;p&gt;Mahmoud&lt;/p&gt;

&lt;p&gt;Is the filesystem down or are you just concerned by the error messages?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="21661" author="mhanafi" created="Fri, 21 Oct 2011 16:39:34 +0000"  >&lt;p&gt;The filesystem is not down. Just want to know if this should be cause &lt;br/&gt;
for concern or a need for possible fsck?&lt;/p&gt;</comment>
                            <comment id="21662" author="pjones" created="Fri, 21 Oct 2011 16:41:48 +0000"  >&lt;p&gt;Understood. Are you running the Oracle 1.8.6 or the Whamcloud 1.8.6-wc1 release and do you have any patches applied?&lt;/p&gt;</comment>
                            <comment id="21672" author="adilger" created="Fri, 21 Oct 2011 17:31:54 +0000"  >&lt;p&gt;Do you know of any unusual application running at this time?&lt;/p&gt;

&lt;p&gt;Concievably this could be a &quot;unusual but harmless&quot; error if clients are deleting and recreating files one is being assigned the same inode number, but a new generation number (which is the point of the generation number).  The clients could be revalidating the attributes of an old version of the inode, but that inode was deleted and a new one is in its place.&lt;/p&gt;

&lt;p&gt;What is unusual here (and the cause of my speculation) is that the &quot;current&quot; inode/generation number (209323401/3543922068) is staying constant, but the generation number that is being requested is changing (either 3529758994 or 3529770923, which are earlier values).&lt;/p&gt;

&lt;p&gt;Also, are you exporting the filesystem via NFS to other clients?  They may keep the inode+generation in the NFS file handle for a long time, and this could generate error messages like this also.&lt;/p&gt;

&lt;p&gt;It may well be that this error message is unnecessary and can be changed to be an internal debugging message.&lt;/p&gt;

&lt;p&gt;What is a bit interesting is that the errors appear to be happening exactly once per second (~600 every 10 minutes), so this may be some kind of application that is polling the filesystem every second?  You could determine what the filename in question by running on the MDS &quot;debugfs -c -R &apos;ncheck 209323401&apos; /dev/&lt;/p&gt;
{MDSDEV}
&lt;p&gt;&quot;, and perhaps discover which user and/or application is causing this message, and whether they are experiencing any problems.&lt;/p&gt;</comment>
                            <comment id="21673" author="jaylan" created="Fri, 21 Oct 2011 17:49:31 +0000"  >&lt;p&gt;Mahmoud, nbp30-mds runs sles10sp2, lustre-1.8.2-3.4nas_ofed151.&lt;br/&gt;
The lustre server was built by Jason with git source based on LLNL 1.8.2&apos;s.&lt;/p&gt;

&lt;p&gt;The git source can be found at &lt;a href=&quot;https://github.com/jlan/lustre-nas&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/jlan/lustre-nas&lt;/a&gt;&lt;br/&gt;
branch b1_8-server-nas, tag 1.8.2-3.4nas.&lt;/p&gt;</comment>
                            <comment id="21845" author="pjones" created="Tue, 25 Oct 2011 12:53:09 +0000"  >&lt;p&gt;Thanks Jay. When do you\Mahmoud expect to be able to answer Andreas&apos;s other questions?&lt;/p&gt;</comment>
                            <comment id="22056" author="mhanafi" created="Thu, 27 Oct 2011 13:32:26 +0000"  >&lt;p&gt;This inode, 209323401, points to a empty directory &quot;/ROOT/yham/output_ran7/Y1994/19941011&quot; This user is not running any jobs and is not logged in. So it would appear that this directory is static. &lt;/p&gt;

&lt;p&gt;drwxr-xr-x 2 yham xxxxx 4096 Oct 11 08:34 /nobackupp30/yham/output_ran7/Y1994/19941011&lt;/p&gt;</comment>
                            <comment id="22161" author="adilger" created="Fri, 28 Oct 2011 19:25:00 +0000"  >&lt;p&gt;It looks like some client(s) cache an old version of this information about this directory for some reason.&lt;/p&gt;

&lt;p&gt;If the message was repeating fairly regularly, it would be possible to collect RPCTRACE logs on the MDS (via &quot;lctl set_param debug=+rpctrace&quot;) and then dump them quickly after the message appeared (via &quot;lctl dk /tmp/debug.log&quot;) to see which client this was coming from.  Then that client could collect RPCTRACE and VFSTRACE logs to see what operation it is doing to trigger this (to debug it) or just flush its cache (via &quot;lctl set_param ldlm.namespaces.&lt;b&gt;MDT&lt;/b&gt;.lru_size=clear&quot;) to get rid of the problem.  The problem should also clear up if the clients are restarted.&lt;/p&gt;</comment>
                            <comment id="22299" author="mhanafi" created="Wed, 2 Nov 2011 16:19:23 +0000"  >&lt;p&gt;Found the client that the request were coming from. Got debug logs then clear the cache. the issue remained. Gathered debug after the flush. See attached file debug.log.tgz&lt;/p&gt;</comment>
                            <comment id="22702" author="pjones" created="Tue, 8 Nov 2011 15:43:51 +0000"  >&lt;p&gt;Lai&lt;/p&gt;

&lt;p&gt;Could you please review the latest information from NASA?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="24611" author="laisiyao" created="Tue, 13 Dec 2011 11:12:20 +0000"  >&lt;p&gt;log of process 21071&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000080:00200000:6:1320263638.353865:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263640.357859:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263642.362592:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263644.369859:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263646.373762:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263648.377838:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263650.381701:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263652.385898:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
00000080:00200000:6:1320263654.389734:0:21071:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529767979(ffff810189be75a0), inode mode 41c0 mask 1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;logs of process 898:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000080:00200000:7:1320263650.433778:0:898:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529770923(ffff810049e77ce0), inode mode 41c0 mask 1
00000080:00200000:7:1320263650.433790:0:898:0:(namei.c:557:ll_lookup_it()) VFS Op:name=COMPLETED.f15_Mach2,dir=209323401/3529770923(ffff810049e77ce0),intent=getattr
00000100:00100000:7:1320263650.433833:0:898:0:(client.c:2066:ptlrpc_queue_wait()) Sending RPC pname:cluuid:pid:xid:nid:opc runOverflow.csh:2bb186c5-12bb-59bc-5cbf-2eb3fb75f6d0:898:x1380095751772954:10.151.25.176@o2ib:101
00000100:00100000:7:1320263650.433957:0:898:0:(client.c:2171:ptlrpc_queue_wait()) Completed RPC pname:cluuid:pid:xid:nid:opc runOverflow.csh:2bb186c5-12bb-59bc-5cbf-2eb3fb75f6d0:898:x1380095751772954:10.151.25.176@o2ib:101
00000100:00100000:7:1320263650.433965:0:898:0:(events.c:114:reply_in_callback()) @@@ unlink  req@ffff8101aa92c000 x1380095751772954/t0 o101-&amp;gt;nbp30-MDT0000_UUID@10.151.25.176@o2ib:12/10 lens 560/3920 e 0 to 1 dl 1320263992 ref 1 fl Rpc:R/0/0 rc 0/301
00000100:00100000:7:1320263650.433983:0:898:0:(client.c:1891:ptlrpc_free_committed()) nbp30-MDT0000-mdc-ffff8101e335ac00: skip recheck: last_committed 210979479427
00000002:00100000:7:1320263650.434001:0:898:0:(mdc_locks.c:505:mdc_finish_enqueue()) @@@ op: 8 disposition: 3, status: -2  req@ffff8101aa92c000 x1380095751772954/t0 o101-&amp;gt;nbp30-MDT0000_UUID@10.151.25.176@o2ib:12/10 lens 560/3920 e 0 to 1 dl 1320263992 ref 1 fl Interpret:R/0/0 rc 0/301
00000080:00200000:7:1320263652.437757:0:898:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529770923(ffff810049e77ce0), inode mode 41c0 mask 1
00000080:00200000:7:1320263652.437766:0:898:0:(namei.c:557:ll_lookup_it()) VFS Op:name=COMPLETED.f15_Mach2,dir=209323401/3529770923(ffff810049e77ce0),intent=getattr
00000100:00100000:7:1320263652.437792:0:898:0:(client.c:2066:ptlrpc_queue_wait()) Sending RPC pname:cluuid:pid:xid:nid:opc runOverflow.csh:2bb186c5-12bb-59bc-5cbf-2eb3fb75f6d0:898:x1380095751774603:10.151.25.176@o2ib:101
00000100:00100000:7:1320263652.437900:0:898:0:(client.c:2171:ptlrpc_queue_wait()) Completed RPC pname:cluuid:pid:xid:nid:opc runOverflow.csh:2bb186c5-12bb-59bc-5cbf-2eb3fb75f6d0:898:x1380095751774603:10.151.25.176@o2ib:101
00000100:00100000:7:1320263652.437904:0:898:0:(events.c:114:reply_in_callback()) @@@ unlink  req@ffff8101c62cc400 x1380095751774603/t0 o101-&amp;gt;nbp30-MDT0000_UUID@10.151.25.176@o2ib:12/10 lens 560/3920 e 0 to 1 dl 1320263994 ref 1 fl Rpc:R/0/0 rc 0/301
00000100:00100000:7:1320263652.437911:0:898:0:(client.c:1891:ptlrpc_free_committed()) nbp30-MDT0000-mdc-ffff8101e335ac00: skip recheck: last_committed 210979479427
00000002:00100000:7:1320263652.437921:0:898:0:(mdc_locks.c:505:mdc_finish_enqueue()) @@@ op: 8 disposition: 3, status: -2  req@ffff8101c62cc400 x1380095751774603/t0 o101-&amp;gt;nbp30-MDT0000_UUID@10.151.25.176@o2ib:12/10 lens 560/3920 e 0 to 1 dl 1320263994 ref 1 fl Interpret:R/0/0 rc 0/301
00000080:00200000:7:1320263654.441728:0:898:0:(file.c:3500:ll_inode_permission()) VFS Op:inode=209323401/3529770923(ffff810049e77ce0), inode mode 41c0 mask 1
00000080:00200000:7:1320263654.441741:0:898:0:(namei.c:557:ll_lookup_it()) VFS Op:name=COMPLETED.f15_Mach2,dir=209323401/3529770923(ffff810049e77ce0),intent=getattr
00000100:00100000:7:1320263654.441787:0:898:0:(client.c:2066:ptlrpc_queue_wait()) Sending RPC pname:cluuid:pid:xid:nid:opc runOverflow.csh:2bb186c5-12bb-59bc-5cbf-2eb3fb75f6d0:898:x1380095751776847:10.151.25.176@o2ib:101
00000100:00100000:7:1320263654.441965:0:898:0:(client.c:2171:ptlrpc_queue_wait()) Completed RPC pname:cluuid:pid:xid:nid:opc runOverflow.csh:2bb186c5-12bb-59bc-5cbf-2eb3fb75f6d0:898:x1380095751776847:10.151.25.176@o2ib:101
00000100:00100000:7:1320263654.441973:0:898:0:(events.c:114:reply_in_callback()) @@@ unlink  req@ffff8101c7f90000 x1380095751776847/t0 o101-&amp;gt;nbp30-MDT0000_UUID@10.151.25.176@o2ib:12/10 lens 560/3920 e 0 to 1 dl 1320263996 ref 1 fl Rpc:R/0/0 rc 0/301
00000100:00100000:7:1320263654.441990:0:898:0:(client.c:1891:ptlrpc_free_committed()) nbp30-MDT0000-mdc-ffff8101e335ac00: skip recheck: last_committed 210979479427
00000002:00100000:7:1320263654.442013:0:898:0:(mdc_locks.c:505:mdc_finish_enqueue()) @@@ op: 8 disposition: 3, status: -2  req@ffff8101c7f90000 x1380095751776847/t0 o101-&amp;gt;nbp30-MDT0000_UUID@10.151.25.176@o2ib:12/10 lens 560/3920 e 0 to 1 dl 1320263996 ref 1 fl Interpret:R/0/0 rc 0/301
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;These logs show some patterns: &lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;process 21071 calls ll_inode_permission() on inode 209323401 (a directory) every 2 seconds, and it always uses i_generation 3529767979, but it doesn&apos;t do any fs operations other than that.&lt;/li&gt;
	&lt;li&gt;process 898 calls lookup on a file under inode 209323401, and it always uses i_generation 3529770923. And this lookup will fail because MDS thinks the generation is wrong.&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Mahmoud, if you could give some hint of what these two processes are doing, that will help. &lt;/p&gt;</comment>
                            <comment id="25640" author="pjones" created="Wed, 4 Jan 2012 12:57:07 +0000"  >&lt;p&gt;Mahmoud&lt;/p&gt;

&lt;p&gt;Are you able to provide any further information about the processes Lai has highlighted?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="70147" author="mhanafi" created="Tue, 29 Oct 2013 18:07:12 +0000"  >&lt;p&gt;This case can be closed&lt;/p&gt;</comment>
                            <comment id="70159" author="pjones" created="Tue, 29 Oct 2013 18:32:52 +0000"  >&lt;p&gt;Thanks Mahmoud&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="10571" name="debug.log.tgz" size="3602879" author="mhanafi" created="Wed, 2 Nov 2011 16:17:11 +0000"/>
                            <attachment id="10570" name="debug.log.tgz" size="3602879" author="mhanafi" created="Wed, 2 Nov 2011 16:16:06 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10040" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic</customfieldname>
                        <customfieldvalues>
                                        <label>metadata</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvkfb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>7038</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>