<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:45:39 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4765] Failure on test suite sanity test_39l: atime is not updated</title>
                <link>https://jira.whamcloud.com/browse/LU-4765</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;http://maloo.whamcloud.com/test_sets/5b796592-aa08-11e3-b4b1-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/5b796592-aa08-11e3-b4b1-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_39l failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;atime is not updated from future: 1426094124, 1394558125&amp;lt;atime&amp;lt;1394558125&lt;/p&gt;&lt;/blockquote&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity test 39l: directory atime update =========================================================== 10:15:24 (1394558124)
CMD: shadow-8vm12 lctl get_param -n mdd.*MDT0000*.atime_diff
 sanity test_39l: @@@@@@ FAIL: atime is not updated from future: 1426094124, 1394558125&amp;lt;atime&amp;lt;1394558125 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>server: lustre-master build # 1937&lt;br/&gt;
client: SLES11 SP3</environment>
        <key id="23626">LU-4765</key>
            <summary>Failure on test suite sanity test_39l: atime is not updated</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="bogl">Bob Glossman</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Thu, 13 Mar 2014 21:26:54 +0000</created>
                <updated>Thu, 25 Feb 2021 00:38:59 +0000</updated>
                            <resolved>Thu, 25 Feb 2021 00:38:59 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                    <version>Lustre 2.9.0</version>
                    <version>Lustre 2.10.0</version>
                    <version>Lustre 2.11.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="79290" author="bogl" created="Thu, 13 Mar 2014 21:45:16 +0000"  >&lt;p&gt;seems to be something wrong with the client/server combo being run here.  the server Centos rpms are from a recent build, lustre-master #1937, but the sles11sp3 client rpms appear to be much older.  Is this a version interop run of some sort?  If not why aren&apos;t the client and server rpms from the same build?&lt;/p&gt;</comment>
                            <comment id="79529" author="adilger" created="Mon, 17 Mar 2014 18:08:02 +0000"  >&lt;p&gt;This may relate to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1783&quot; title=&quot;sanity test_39l failed: atime is not updated&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1783&quot;&gt;LU-1783&lt;/a&gt; which was also a problem on newer kernels with atime updates from the future.  Yang Sheng had made a patch for upstream, but it wasn&apos;t accepted for a variety of minor reasons.  It should probably be refreshed and tried again.&lt;/p&gt;</comment>
                            <comment id="82681" author="bogl" created="Mon, 28 Apr 2014 19:39:35 +0000"  >&lt;p&gt;another:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/63de2e74-cf07-11e3-a250-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/63de2e74-cf07-11e3-a250-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="82699" author="bogl" created="Mon, 28 Apr 2014 22:10:50 +0000"  >&lt;p&gt;seeing a lot of these in autotest test runs with sles11sp3 clients.  have been able to reproduce this bug manually, but only if I explicitly mount a client with the -o relatime option.  as far as I can tell this option isn&apos;t typically used, mount.lustre sets MS_STRICTATIME by default.  Is there something in autotest that is forcing the relatime option in client mounts?  I haven&apos;t found anything in logs that would indicate that is happening.&lt;/p&gt;</comment>
                            <comment id="82753" author="bogl" created="Tue, 29 Apr 2014 16:30:38 +0000"  >&lt;p&gt;Following a suspicion that mount options on SP3 were out of wack I took a close look at some #include files currently on our sles11sp3 builder nodes.  I discovered that in fact the #include files there were seriously out of date.  In particular in mount.h on the builder node MS_RDONLY is #defined but MS_STRICTATIME isn&apos;t.  I&apos;m pretty sure this causes the mount.lustre built in our sles11sp3 artifacts to not have the special override code making MS_STRICTATIME the default in all mounts.   This would explain why the 39l is failing in autotest runs, but not in my manual local runs where mount.lustre is built with more up to date tools and #includes.&lt;/p&gt;

&lt;p&gt;I will file a TEI ticket requesting the environment on SP3 builders be updated. Fairly certain that will make this problem go away.&lt;/p&gt;</comment>
                            <comment id="82818" author="bogl" created="Tue, 29 Apr 2014 23:39:54 +0000"  >&lt;p&gt;test only patch to verify TEI-2052 changes&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/10156&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10156&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="82922" author="bogl" created="Wed, 30 Apr 2014 21:39:19 +0000"  >&lt;p&gt;&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/a9add412-d0ac-11e3-b9d4-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/a9add412-d0ac-11e3-b9d4-52540035b04c&lt;/a&gt; shows test 39l now passing.  The only remaining failures in sanity,sh on sles11sp3 clients are known bugs &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4713&quot; title=&quot;Failure on test suite sanity test_237: check_fhandle_syscalls failed &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4713&quot;&gt;&lt;del&gt;LU-4713&lt;/del&gt;&lt;/a&gt; and &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4341&quot; title=&quot;Failure on test suite sanity test_170: expected 31 bad lines, but got 34&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4341&quot;&gt;&lt;del&gt;LU-4341&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="85444" author="mmansk" created="Mon, 2 Jun 2014 14:41:49 +0000"  >&lt;p&gt;We&apos;re seeing this failure as well for 2.6.  Not sure if/when we&apos;ll be updating our sles11sp3.  Our current klibc is at 1.5.18.  &lt;/p&gt;

&lt;p&gt;Any way to workaround this besides comment out the test?&lt;/p&gt;
</comment>
                            <comment id="85447" author="bogl" created="Mon, 2 Jun 2014 15:03:38 +0000"  >&lt;p&gt;currently not reproducible in our 2.6 builds anymore.&lt;br/&gt;
klibc?  do you mean glibc?  as far as I know the relevant #include files are part of glibc-devel.  I believe current version of glibc-devel in sles11sp3 is 2.11.3, so if you have 1.&amp;lt;anything&amp;gt; you are seriously out of date.&lt;/p&gt;

&lt;p&gt;Are you rolling your own builds?  Again, as far as I know you should not be having problems if you just install and use our prebuilt rpms.&lt;/p&gt;</comment>
                            <comment id="94017" author="jay" created="Mon, 15 Sep 2014 18:24:09 +0000"  >&lt;p&gt;I took a look at atime problem in Lustre when I was investigating POSIX compliance test failure at &lt;a href=&quot;http://review.whamcloud.com/11844&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/11844&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The conclusion is that POSIX atime is not strictly honored when the reading data is already in cache. To fix this problem, we need to send a RPC to the MDT to update atime for every read syscall, this will introduce huge performance loss.&lt;/p&gt;

&lt;p&gt;This issue can be easily reproduce as follows:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@mercury lustre]# cat passwd &amp;gt; /dev/null
[root@mercury lustre]# stat passwd 
  File: `passwd&apos;
  Size: 1533      	Blocks: 8          IO Block: 4194304 regular file
Device: 2c54f966h/743766374d	Inode: 144115205255725455  Links: 1
Access: (0644/-rw-r--r--)  Uid: (    0/    root)   Gid: (    0/    root)
Access: 2014-09-15 07:26:39.000000000 -0700
Modify: 2014-09-15 06:34:54.000000000 -0700
Change: 2014-09-15 06:34:54.000000000 -0700
[root@mercury lustre]# cat passwd &amp;gt; /dev/null
[root@mercury lustre]# stat passwd 
  File: `passwd&apos;
  Size: 1533      	Blocks: 8          IO Block: 4194304 regular file
Device: 2c54f966h/743766374d	Inode: 144115205255725455  Links: 1
Access: (0644/-rw-r--r--)  Uid: (    0/    root)   Gid: (    0/    root)
Access: 2014-09-15 07:26:48.000000000 -0700
Modify: 2014-09-15 06:34:54.000000000 -0700
Change: 2014-09-15 06:34:54.000000000 -0700
[root@mercury lustre]# lctl set_param ldlm.namespaces.*.lru_size=0
ldlm.namespaces.MGC172.30.24.132@tcp.lru_size=0
ldlm.namespaces.lustre-MDT0000-lwp-MDT0000.lru_size=0
ldlm.namespaces.lustre-MDT0000-lwp-OST0000.lru_size=0
ldlm.namespaces.lustre-MDT0000-lwp-OST0001.lru_size=0
ldlm.namespaces.lustre-MDT0000-mdc-ffff88014ce49400.lru_size=0
ldlm.namespaces.lustre-OST0000-osc-MDT0000.lru_size=0
ldlm.namespaces.lustre-OST0000-osc-ffff88014ce49400.lru_size=0
ldlm.namespaces.lustre-OST0001-osc-MDT0000.lru_size=0
ldlm.namespaces.lustre-OST0001-osc-ffff88014ce49400.lru_size=0
[root@mercury lustre]# stat passwd 
  File: `passwd&apos;
  Size: 1533      	Blocks: 8          IO Block: 4194304 regular file
Device: 2c54f966h/743766374d	Inode: 144115205255725455  Links: 1
Access: (0644/-rw-r--r--)  Uid: (    0/    root)   Gid: (    0/    root)
Access: 2014-09-15 07:26:39.000000000 -0700
Modify: 2014-09-15 06:34:54.000000000 -0700
Change: 2014-09-15 06:34:54.000000000 -0700
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The atime of the 2nd read &quot;cat passwd &amp;gt; /dev/null&quot; was not honored. It only updated atime in memory but never updated it on the OST. Therefore, newly refreshed read kept the old atime of the first read.&lt;/p&gt;</comment>
                            <comment id="94021" author="adilger" created="Mon, 15 Sep 2014 18:57:16 +0000"  >&lt;p&gt;What your test was missing is that the MDS does not write the atime update to disk unless the atime is at least 60s old.  The atime is updated on the MDT at close time when the atime exceeds the &lt;tt&gt;mdd.*.atime_diff&lt;/tt&gt; tunable, so that this doesn&apos;t cause numerous updates at close time when the same file is opened by thousands of clients.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;mds# lctl get_param mdd.*.atime_diff
mdd.myth-MDT0000.atime_diff=60
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;In Jinshan&apos;s example, the atime happened to be only 9s old (07:26:39 vs. 07:26:48) so it was not updated on the MDT on disk.  Testing this on an old file I have on my filesystem (which you can see has already had the atime updated due to previous accesses):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;$ stat /myth/tmp/f9
  File: `/myth/tmp/f9&apos;
:
Access: 2014-06-23 12:59:48.000000000 -0600
Modify: 2013-05-28 22:04:47.000000000 -0600
Change: 2013-11-27 15:26:07.000000000 -0700
$ cat /myth/tmp/f9 &amp;gt; /dev/null
$ stat /myth/tmp/f9
:
Access: 2014-09-15 12:38:12.000000000 -0600
$ sudo lctl set_param ldlm.namespaces.*.lru_size=clear
ldlm.namespaces.MGC192.168.20.1@tcp.lru_size=clear
ldlm.namespaces.myth-MDT0000-mdc-ffff8800377fc000.lru_size=clear
ldlm.namespaces.myth-OST0000-osc-ffff8800377fc000.lru_size=clear
ldlm.namespaces.myth-OST0001-osc-ffff8800377fc000.lru_size=clear
ldlm.namespaces.myth-OST0002-osc-ffff8800377fc000.lru_size=clear
ldlm.namespaces.myth-OST0003-osc-ffff8800377fc000.lru_size=clear
ldlm.namespaces.myth-OST0004-osc-ffff8800377fc000.lru_size=clear
$ stat /myth/tmp/f9
:
Access: 2014-09-15 12:38:12.000000000 -0600
Modify: 2013-05-28 22:04:47.000000000 -0600
Change: 2013-11-27 15:26:07.000000000 -0700
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I also verified with debugfs that this inode&apos;s atime was updated on disk on the MDT.&lt;/p&gt;

&lt;p&gt;The original problem reported in this bug is quite different.  In newer kernels, setting the atime into the past (e.g. after accessing a file with &quot;tar&quot; for backup and trying to reset the atime to the original value) does not affect filesystems mounted using the &quot;relatime&quot; option.  This bug has not been fixed in upstream kernels yet, and may never be.&lt;/p&gt;

&lt;p&gt;PS: setting &quot;lru_size=0&quot; doesn&apos;t necessarily clear the DLM locks from the client, but rather disables the static DLM LRU maximum and enables dynamic LRU sizing. Use &quot;lctl set_param ldlm.namespaces.*.lru_size=clear&quot; to clear the DLM lock cache on the client.&lt;/p&gt;</comment>
                            <comment id="102315" author="bogl" created="Fri, 26 Dec 2014 00:43:58 +0000"  >&lt;p&gt;another one seen on sles11sp3:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/14d6386e-8c7e-11e4-b81b-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/14d6386e-8c7e-11e4-b81b-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="155573" author="sarah" created="Mon, 13 Jun 2016 23:52:55 +0000"  >&lt;p&gt;failure seen in interop testing between 2.9 server and 2.8 client:&lt;br/&gt;
server: master/#3385 RHEL7.2&lt;br/&gt;
client: 2.8.0 RHEL6.7&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a9ca904c-2d62-11e6-80b9-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a9ca904c-2d62-11e6-80b9-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="169714" author="standan" created="Fri, 14 Oct 2016 19:34:04 +0000"  >&lt;p&gt;This issue has occurred around 44 times in past 30 days.&lt;/p&gt;</comment>
                            <comment id="169802" author="adilger" created="Fri, 14 Oct 2016 23:26:40 +0000"  >&lt;p&gt;In general, I think 30-day results are less useful than, say, 7-day results, because a problem may have happened many times three weeks ago but was fixed and hasn&apos;t happened at all in the past week.  If you are reporting on a problem that is still happening regularly (i.e. it also happened 8 times in the past week) then please include that information in the summary.&lt;/p&gt;

&lt;p&gt;For example, the sanityn test_77&lt;span class=&quot;error&quot;&gt;&amp;#91;a-e&amp;#93;&lt;/span&gt; failures appear even in the 7-day results, but not at all in the 3-day results since the problem was resolved earlier in the week.  &lt;/p&gt;</comment>
                            <comment id="178956" author="bogl" created="Fri, 23 Dec 2016 15:13:12 +0000"  >&lt;p&gt;another on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/3f9ace90-c8ca-11e6-8911-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/3f9ace90-c8ca-11e6-8911-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="178967" author="bogl" created="Fri, 23 Dec 2016 16:08:04 +0000"  >&lt;p&gt;wondering if the recent rash of these failures have something to do with the mod landed this year; &quot;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8041&quot; title=&quot;Close should update atime increasing only&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8041&quot;&gt;&lt;del&gt;LU-8041&lt;/del&gt;&lt;/a&gt; mdd: increasing only atime update on close&quot;&lt;/p&gt;

&lt;p&gt;I note this includes changes to test 39l&lt;/p&gt;</comment>
                            <comment id="188232" author="bogl" created="Tue, 14 Mar 2017 13:05:17 +0000"  >&lt;p&gt;another on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/3c35c808-08ae-11e7-9053-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/3c35c808-08ae-11e7-9053-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="190262" author="bogl" created="Fri, 31 Mar 2017 13:52:06 +0000"  >&lt;p&gt;another on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/6bb96e0e-15b0-11e7-8920-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/6bb96e0e-15b0-11e7-8920-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="292958" author="adilger" created="Thu, 25 Feb 2021 00:38:59 +0000"  >&lt;p&gt;Close as duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1739&quot; title=&quot;parallel-scale-nfsv4: compilebench: IOError: [Errno 5] Input/output error&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1739&quot;&gt;&lt;del&gt;LU-1739&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="15578">LU-1783</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwhlz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>13108</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>