<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:40:21 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4176] Test failure on test suite sanity-hsm, subtest test_31a</title>
                <link>https://jira.whamcloud.com/browse/LU-4176</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Nathaniel Clark &amp;lt;nathaniel.l.clark@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;http://maloo.whamcloud.com/test_sets/6bbd2b34-407a-11e3-af95-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/6bbd2b34-407a-11e3-af95-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_31a failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;request on &lt;span class=&quot;error&quot;&gt;&amp;#91;0x200000402:0xf2:0x0&amp;#93;&lt;/span&gt; is not SUCCEED&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Info required for matching: sanity-hsm 31a&lt;/p&gt;

&lt;p&gt;This test was EXCEPT&apos;d prior to &lt;a href=&quot;http://review.whamcloud.com/7374&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7374&lt;/a&gt; (&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3815&quot; title=&quot;Fix sanity-hsm test 34: &amp;quot;Remove file during restore&amp;quot;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3815&quot;&gt;&lt;del&gt;LU-3815&lt;/del&gt;&lt;/a&gt;) re-enabled most of the EXCEPT&apos;d tests.&lt;/p&gt;</description>
                <environment></environment>
        <key id="21702">LU-4176</key>
            <summary>Test failure on test suite sanity-hsm, subtest test_31a</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>HSM</label>
                    </labels>
                <created>Tue, 29 Oct 2013 13:28:55 +0000</created>
                <updated>Sun, 3 Apr 2016 03:33:31 +0000</updated>
                            <resolved>Fri, 23 Jan 2015 09:55:54 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                                    <fixVersion>Lustre 2.6.0</fixVersion>
                    <fixVersion>Lustre 2.5.1</fixVersion>
                    <fixVersion>Lustre 2.7.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="70568" author="bfaccini" created="Sat, 2 Nov 2013 15:37:45 +0000"  >&lt;p&gt;After I found similar problem with sanity-hsm/test_31c during test session &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/bf8b7f8a-4396-11e3-942a-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/bf8b7f8a-4396-11e3-942a-52540035b04c&lt;/a&gt;, I wanted to create a new ticket but since I thought that both sub-tests 31a/31b may have experienced the same problem, after some search I 1st found &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3969&quot; title=&quot;Test failure on test suite sanity-hsm, subtest test_31b&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3969&quot;&gt;&lt;del&gt;LU-3969&lt;/del&gt;&lt;/a&gt; for test_31b similar problem, which has been dup&apos;ed to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3761&quot; title=&quot;sanity-hsm 55: request on 0x2000013a3:0x2:0x0 is not FAILED&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3761&quot;&gt;&lt;del&gt;LU-3761&lt;/del&gt;&lt;/a&gt; with no real conclusion about the exact scenario, and also found this one for same error with test_31a !!&lt;/p&gt;

&lt;p&gt;The issue seems to come from the same sub-function restore_and_check_size() all, and only, these 3 sub-tests are using to check the restore of a big file :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;restore_and_check_size() {
        local f=$1
        local fid=$2
        local s=$(stat -c &quot;%s&quot; $f)
        local n=$s
        local st=$(get_hsm_flags $f)
        local err=0
        local cpt=0
        $LFS hsm_restore $f
        while [[ &quot;$st&quot; != &quot;0x00000009&quot; &amp;amp;&amp;amp; $cpt -le 10 ]]
        do
                n=$(stat -c &quot;%s&quot; $f)
                # we echo in both cases to show stat is not
                # hang
                if [[ $n != $s ]]; then
                        echo &quot;size seen is $n != $s&quot;
                        err=1
                else
                        echo &quot;size seen is right: $n == $s&quot;
                fi
                st=$(get_hsm_flags $f)
                sleep 10
                cpt=$((cpt + 1))
        done
        if [[ $cpt -lt 10 ]]; then
                echo &quot; restore is too long&quot;
        else
                echo &quot; &quot;done
        fi
        wait_request_state $fid RESTORE SUCCEED
        return $err
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;and if we look at the tests error logs they all look similar and here is 31a log for the error that caused this ticket creation :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity-hsm test 31a: Import a large file and check size during restore == 22:35:58 (1383024958)
CMD: wtm-10vm5 pkill -CONT -x lhsmtool_posix
Purging archive on wtm-10vm5
CMD: wtm-10vm5 rm -rf /home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1/*
Starting copytool agt1 on wtm-10vm5
CMD: wtm-10vm5 mkdir -p /home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1
CMD: wtm-10vm5 lhsmtool_posix  --daemon --hsm-root /home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1 --bandwidth 1 /mnt/lustre &amp;lt; /dev/null &amp;gt; /logdir/test_logs/2013-10-28/lustre-reviews-el6-x86_64--review--1_2_1__19069__-69843391629520-112320/sanity-hsm.test_31a.copytool_log.wtm-10vm5.log 2&amp;gt;&amp;amp;1
CMD: wtm-10vm5 mkdir -p /home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1/d0.sanity-hsm/d31
CMD: wtm-10vm5 dd if=/dev/urandom of=/home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1/d0.sanity-hsm/d31/f.sanity-hsm.31a count=32 bs=1000000
wtm-10vm5: 32+0 records in
wtm-10vm5: 32+0 records out
wtm-10vm5: 32000000 bytes (32 MB) copied, 4.95968 s, 6.5 MB/s
CMD: wtm-10vm5 lhsmtool_posix --archive 2 --hsm-root /home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1		--import d0.sanity-hsm/d31/f.sanity-hsm.31a /mnt/lustre/d0.sanity-hsm/d31/f.sanity-hsm.31a /mnt/lustre
wtm-10vm5: lhsmtool_posix[23455]: action=1 src=d0.sanity-hsm/d31/f.sanity-hsm.31a dst=/mnt/lustre/d0.sanity-hsm/d31/f.sanity-hsm.31a mount_point=/mnt/lustre
wtm-10vm5: lhsmtool_posix[23455]: importing &apos;/mnt/lustre/d0.sanity-hsm/d31/f.sanity-hsm.31a&apos; from &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1/d0.sanity-hsm/d31/f.sanity-hsm.31a&apos;
wtm-10vm5: lhsmtool_posix[23455]: imported &apos;/mnt/lustre/d0.sanity-hsm/d31/f.sanity-hsm.31a&apos; from &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1/00f2/0000/0402/0000/0002/0000/0x200000402:0xf2:0x0&apos;==&apos;/home/cgearing/.autotest/shared_dir/2013-10-28/112321-69843391629520/arc1/d0.sanity-hsm/d31/f.sanity-hsm.31a&apos;
wtm-10vm5: lhsmtool_posix[23455]: process finished, errs: 0 major, 0 minor, rc=0 (Success)
CMD: wtm-10vm5 pkill -CONT -x lhsmtool_posix
Wakeup copytool agt1 on wtm-10vm5
size seen is right: 32000000 == 32000000
size seen is right: 32000000 == 32000000
size seen is right: 32000000 == 32000000
size seen is right: 32000000 == 32000000
size seen is right: 32000000 == 32000000
 restore is too long
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 100 secs for update
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 90 secs for update
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=

&#8230;&#8230;&#8230;&#8230;. &amp;lt;&amp;lt;&amp;lt;&amp;lt; spending the whole 100s printing the same

CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 10 secs for update
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-10vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;[0x200000402:0xf2:0x0]&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Update not seen after 100s: wanted &apos;SUCCEED&apos; got &apos;&apos;
 sanity-hsm test_31a: @@@@@@ FAIL: request on [0x200000402:0xf2:0x0] is not SUCCEED 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4289:error_noexit()
  = /usr/lib64/lustre/tests/test-framework.sh:4320:error()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:470:wait_request_state()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1891:restore_and_check_size()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1907:test_31a()
  = /usr/lib64/lustre/tests/test-framework.sh:4575:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:4609:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:4463:run_test()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1914:main()
Dumping lctl log to /logdir/test_logs/2013-10-28/lustre-reviews-el6-x86_64--review--1_2_1__19069__-69843391629520-112320/sanity-hsm.test_31a.*.1383025137.log
CMD: wtm-10vm3,wtm-10vm4,wtm-10vm5,wtm-10vm6.rosso.whamcloud.com /usr/sbin/lctl dk &amp;gt; /logdir/test_logs/2013-10-28/lustre-reviews-el6-x86_64--review--1_2_1__19069__-69843391629520-112320/sanity-hsm.test_31a.debug_log.\$(hostname -s).1383025137.log;
         dmesg &amp;gt; /logdir/test_logs/2013-10-28/lustre-reviews-el6-x86_64--review--1_2_1__19069__-69843391629520-112320/sanity-hsm.test_31a.dmesg.\$(hostname -s).1383025137.log
CMD: wtm-10vm5 pkill -INT -x lhsmtool_posix
Copytool is stopped on wtm-10vm5
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;here is the 31b log for the error in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3969&quot; title=&quot;Test failure on test suite sanity-hsm, subtest test_31b&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3969&quot;&gt;&lt;del&gt;LU-3969&lt;/del&gt;&lt;/a&gt; :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity-hsm test 31b: Restore a large unaligned file and check size during restore == 13:50:02 (1379451002)
CMD: wtm-14vm5 pkill -CONT -x lhsmtool_posix
Purging archive on wtm-14vm5
CMD: wtm-14vm5 rm -rf /home/cgearing/.autotest/shared_dir/2013-09-17/024917-69984331480660/arc1/*
Starting copytool agt1 on wtm-14vm5
CMD: wtm-14vm5 mkdir -p /home/cgearing/.autotest/shared_dir/2013-09-17/024917-69984331480660/arc1
CMD: wtm-14vm5 lhsmtool_posix  --daemon --hsm-root /home/cgearing/.autotest/shared_dir/2013-09-17/024917-69984331480660/arc1 --bandwidth 1 /mnt/lustre &amp;lt; /dev/null &amp;gt; /logdir/test_logs/2013-09-17/lustre-reviews-el6-x86_64--review-zfs--1_3_1__18234__-69984331480660-024916/sanity-hsm.test_31b.copytool_log.wtm-14vm5.log 2&amp;gt;&amp;amp;1
39+0 records in
39+0 records out
39000000 bytes (39 MB) copied, 15.8537 s, 2.5 MB/s
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 0s: from &apos;&apos; to &apos;STARTED&apos;
Waiting 100 secs for update
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 90 secs for update
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=

&#8230;&#8230;&#8230;&#8230;&#8230; &amp;lt;&amp;lt;&amp;lt;&amp;lt; spending the whole 100s printing the same

CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 10 secs for update
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-14vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x21:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Update not seen after 100s: wanted &apos;SUCCEED&apos; got &apos;&apos;
 sanity-hsm test_31b: @@@@@@ FAIL: request on 0x200008101:0x21:0x0 is not SUCCEED 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4266:error_noexit()
  = /usr/lib64/lustre/tests/test-framework.sh:4293:error()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:444:wait_request_state()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1696:restore_and_check_size()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1734:test_31b()
  = /usr/lib64/lustre/tests/test-framework.sh:4547:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:4580:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:4435:run_test()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1741:main()
Dumping lctl log to /logdir/test_logs/2013-09-17/lustre-reviews-el6-x86_64--review-zfs--1_3_1__18234__-69984331480660-024916/sanity-hsm.test_31b.*.1379451241.log
CMD: wtm-14vm3,wtm-14vm4,wtm-14vm5,wtm-14vm6.rosso.whamcloud.com /usr/sbin/lctl dk &amp;gt; /logdir/test_logs/2013-09-17/lustre-reviews-el6-x86_64--review-zfs--1_3_1__18234__-69984331480660-024916/sanity-hsm.test_31b.debug_log.\$(hostname -s).1379451241.log;
         dmesg &amp;gt; /logdir/test_logs/2013-09-17/lustre-reviews-el6-x86_64--review-zfs--1_3_1__18234__-69984331480660-024916/sanity-hsm.test_31b.dmesg.\$(hostname -s).1379451241.log
CMD: wtm-14vm5 pkill -INT -x lhsmtool_posix
Copytool is stopped on wtm-14vm5
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;and here is the 31c log for the error I just reported :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity-hsm test 31c: Restore a large aligned file and check size during restore == 18:03:34 (1383354214)
CMD: wtm-19vm5 pkill -CONT -x lhsmtool_posix
Purging archive on wtm-19vm5
CMD: wtm-19vm5 rm -rf /home/cgearing/.autotest/shared_dir/2013-11-01/053728-70102366359140/arc1/*
Starting copytool agt1 on wtm-19vm5
CMD: wtm-19vm5 mkdir -p /home/cgearing/.autotest/shared_dir/2013-11-01/053728-70102366359140/arc1
CMD: wtm-19vm5 lhsmtool_posix  --daemon --hsm-root /home/cgearing/.autotest/shared_dir/2013-11-01/053728-70102366359140/arc1 --bandwidth 1 /mnt/lustre &amp;lt; /dev/null &amp;gt; /logdir/test_logs/2013-11-01/lustre-reviews-el6-x86_64--review--1_2_1__19165__-70102366359140-053726/sanity-hsm.test_31c.copytool_log.wtm-19vm5.log 2&amp;gt;&amp;amp;1
/usr/lib64/lustre/tests/sanity-hsm.sh: line 402: [: /mnt/lustre: integer expression expected
33+0 records in
33+0 records out
34603008 bytes (35 MB) copied, 5.01832 s, 6.9 MB/s
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 0s: from &apos;&apos; to &apos;STARTED&apos;
Waiting 100 secs for update
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 90 secs for update
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 80 secs for update
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Updated after 29s: wanted &apos;SUCCEED&apos; got &apos;SUCCEED&apos;
size seen is right: 34603008 == 34603008
size seen is right: 34603008 == 34603008
size seen is right: 34603008 == 34603008
size seen is right: 34603008 == 34603008
size seen is right: 34603008 == 34603008
 restore is too long
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 100 secs for update
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 90 secs for update
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=

&#8230;&#8230;&#8230;&#8230;&#8230; &amp;lt;&amp;lt;&amp;lt;&amp;lt; spending the whole 100s printing the same

CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 10 secs for update
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-19vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x28:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Update not seen after 100s: wanted &apos;SUCCEED&apos; got &apos;&apos;
 sanity-hsm test_31c: @@@@@@ FAIL: request on 0x200000401:0x28:0x0 is not SUCCEED 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4289:error_noexit()
  = /usr/lib64/lustre/tests/test-framework.sh:4320:error()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:470:wait_request_state()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1891:restore_and_check_size()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1950:test_31c()
  = /usr/lib64/lustre/tests/test-framework.sh:4575:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:4609:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:4463:run_test()
  = /usr/lib64/lustre/tests/sanity-hsm.sh:1957:main()
Dumping lctl log to /logdir/test_logs/2013-11-01/lustre-reviews-el6-x86_64--review--1_2_1__19165__-70102366359140-053726/sanity-hsm.test_31c.*.1383354428.log
CMD: wtm-19vm3,wtm-19vm4,wtm-19vm5,wtm-19vm6.rosso.whamcloud.com /usr/sbin/lctl dk &amp;gt; /logdir/test_logs/2013-11-01/lustre-reviews-el6-x86_64--review--1_2_1__19165__-70102366359140-053726/sanity-hsm.test_31c.debug_log.\$(hostname -s).1383354428.log;
         dmesg &amp;gt; /logdir/test_logs/2013-11-01/lustre-reviews-el6-x86_64--review--1_2_1__19165__-70102366359140-053726/sanity-hsm.test_31c.dmesg.\$(hostname -s).1383354428.log
CMD: wtm-19vm5 pkill -INT -x lhsmtool_posix
Copytool is stopped on wtm-19vm5
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;This clearly shows that :&lt;/p&gt;

&lt;p&gt;       _ restore_and_check_size() must be fixed to exchange &quot;too long&quot; and &quot;done&quot; outputs to reflect the real restore status.&lt;/p&gt;

&lt;p&gt;       _ the following+common &quot;wait_request_state $fid RESTORE SUCCEED&quot; may only be required for the &quot;too long&quot; case, since it seems to be unnecessary and buggy if the RESTORE/SUCCESS record has already been deleted or never written ??&#8230; This is what seems to happen in all 3 cases where it is obvious that the file hsm_flags as it HSM_RELEASED bit cleared, meaning the restore was completed.&lt;/p&gt;

&lt;p&gt;       _ thus a possible way to fix could be to only &quot;wait_request_state $fid RESTORE SUCCEED&quot; for &quot;too long&quot; restore case. An other would be to ensure llog RESTORE/SUCCESS record is written/updated (which should not occur only upon request timeout or upon some llog layer error) and will be kept long enough.&lt;/p&gt;</comment>
                            <comment id="70601" author="jcl" created="Mon, 4 Nov 2013 11:59:59 +0000"  >&lt;p&gt;As the function tests the hsm flags, and trigs error if not as expected, we can remove the  wait_request_state $fid RESTORE SUCCEED, so this will manage the &quot;deleted record&quot; case&lt;/p&gt;</comment>
                            <comment id="70604" author="bfaccini" created="Mon, 4 Nov 2013 14:11:53 +0000"  >&lt;p&gt;Submitted patch &lt;a href=&quot;http://review.whamcloud.com/8154&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8154&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="71593" author="jlevi" created="Thu, 14 Nov 2013 23:24:44 +0000"  >&lt;p&gt;Patch landed to Master. IF more work is needed in this ticket, let m know and I&apos;ll reopen this ticket.&lt;/p&gt;</comment>
                            <comment id="72472" author="yujian" created="Thu, 28 Nov 2013 11:42:04 +0000"  >&lt;p&gt;The same failure occurred on Lustre b2_5 branch:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/4bc47d48-57d4-11e3-919d-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/4bc47d48-57d4-11e3-919d-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="74326" author="jamesanunez" created="Fri, 3 Jan 2014 23:54:33 +0000"  >&lt;p&gt;Patch for b2_5 at &lt;a href=&quot;http://review.whamcloud.com/#/c/8720/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/8720/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="74329" author="yujian" created="Sat, 4 Jan 2014 08:52:33 +0000"  >&lt;p&gt;The same failure occurred on Lustre b2_5 branch again:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/5d3e042e-74ce-11e3-8ad9-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/5d3e042e-74ce-11e3-8ad9-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="74449" author="bogl" created="Tue, 7 Jan 2014 00:20:54 +0000"  >&lt;p&gt;seen again in b2_5:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/2181559c-772d-11e3-b181-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/2181559c-772d-11e3-b181-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="78981" author="bfaccini" created="Tue, 11 Mar 2014 11:22:01 +0000"  >&lt;p&gt;Even if patches/fixes have been landed for both b2_5 and master, test_31a (only!! and surprisingly not test_31b and test_31c also when they did trigger this same problem ...) is still disabled respectively after patches #9134 and #8122 (for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4178&quot; title=&quot;Test failure on test suite sanity-hsm, subtest test_200&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4178&quot;&gt;&lt;del&gt;LU-4178&lt;/del&gt;&lt;/a&gt;) have been landed to these 2 branches.&lt;/p&gt;

&lt;p&gt;Patch for master to re-enable test_31a is at &lt;a href=&quot;http://review.whamcloud.com/9577&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/9577&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="95491" author="adilger" created="Thu, 2 Oct 2014 05:42:53 +0000"  >&lt;p&gt;Reopen bug until patch to enable test is landed. &lt;/p&gt;</comment>
                            <comment id="98747" author="bfaccini" created="Mon, 10 Nov 2014 08:41:57 +0000"  >&lt;p&gt;It just has, so do we need to wait some more and verify its behavior in auto-tests before to close ?&lt;/p&gt;</comment>
                            <comment id="104319" author="bfaccini" created="Thu, 22 Jan 2015 11:50:50 +0000"  >&lt;p&gt;Andreas, do you agree than we can close this ticket, now that test_31&lt;span class=&quot;error&quot;&gt;&amp;#91;a,b,c&amp;#93;&lt;/span&gt; have been all re-enabled within master since November and none of these 3 sub-tests have reported the failure addressed in this ticket until now?&lt;/p&gt;</comment>
                            <comment id="105930" author="gerrit" created="Thu, 5 Feb 2015 21:07:54 +0000"  >&lt;p&gt;James Nunez (james.a.nunez@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/13662&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13662&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4176&quot; title=&quot;Test failure on test suite sanity-hsm, subtest test_31a&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4176&quot;&gt;&lt;del&gt;LU-4176&lt;/del&gt;&lt;/a&gt; tests: re-enable sanity-hsm/test_31a&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_5&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 5f3b35ef6b1673c7ae5e23cd3c5a00f9cc07b3e5&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzw76f:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>11307</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>