<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:39:31 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4086] Test failure on test suite sanity-hsm, subtest test_33</title>
                <link>https://jira.whamcloud.com/browse/LU-4086</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for John Hammond &amp;lt;john.hammond@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;http://maloo.whamcloud.com/test_sets/78b59cb0-3123-11e3-8ec6-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/78b59cb0-3123-11e3-8ec6-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_33 failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;request on 0x200000401:0x28:0x0 is not CANCELED&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Info required for matching: sanity-hsm 33&lt;/p&gt;</description>
                <environment></environment>
        <key id="21352">LU-4086</key>
            <summary>Test failure on test suite sanity-hsm, subtest test_33</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>HSM</label>
                    </labels>
                <created>Thu, 10 Oct 2013 14:40:09 +0000</created>
                <updated>Wed, 8 Oct 2014 18:27:57 +0000</updated>
                            <resolved>Mon, 21 Apr 2014 22:57:04 +0000</resolved>
                                    <version>Lustre 2.5.0</version>
                                    <fixVersion>Lustre 2.6.0</fixVersion>
                    <fixVersion>Lustre 2.5.4</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="70367" author="yujian" created="Thu, 31 Oct 2013 14:13:32 +0000"  >&lt;p&gt;Another instance: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/14278fd4-40fe-11e3-a1e8-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/14278fd4-40fe-11e3-a1e8-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="71199" author="bfaccini" created="Sat, 9 Nov 2013 13:27:32 +0000"  >&lt;p&gt;+1: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/c4f891ea-4924-11e3-a79b-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/c4f891ea-4924-11e3-a79b-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="71200" author="bfaccini" created="Sat, 9 Nov 2013 14:24:53 +0000"  >&lt;p&gt;And according to the copytool log :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lhsmtool_posix[18608]: &apos;[0x200000401:0x29:0x0]&apos; action CANCEL reclen 72, cookie=0x527df027
lhsmtool_posix[18608]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[18608]: cancel not implemented for file system &apos;/mnt/lustre&apos;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;which may be a good start to explain why the RESTORE request never reached the CANCELED state !!&#8230;&lt;/p&gt;

&lt;p&gt;Also interesting, the &quot;md5sum&quot; command intended to start the implicit RESTORE has not been killed as expected too :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Updated after 34s: wanted &apos;SUCCEED&apos; got &apos;SUCCEED&apos;
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 0s: from &apos;&apos; to &apos;STARTED&apos;
Waiting 100 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 90 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 80 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 70 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
/usr/lib64/lustre/tests/test-framework.sh: line 1904: 21375 Terminated              md5sum $f &amp;gt; /dev/null
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 32s: from &apos;STARTED&apos; to &apos;SUCCEED&apos;
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 60 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 49s: from &apos;SUCCEED&apos; to &apos;&apos;
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 50 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 40 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 30 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 20 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Waiting 10 secs for update
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: client-26vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200000401:0x29:0x0&apos;.*action=&apos;RESTORE&apos;/ {print \$13}&apos; | cut -f2 -d=
Update not seen after 100s: wanted &apos;CANCELED&apos; got &apos;&apos;
 sanity-hsm test_33: @@@@@@ FAIL: request on 0x200000401:0x29:0x0 is not CANCELED 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4289:error_noexit()
&#8230;&#8230;..

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;But finally, since ps command output is trashed, is it really expected to be ??&lt;/p&gt;

&lt;p&gt;For the record/history and to highlight my previous comments here is current sanity-hsm/test_33() sub-test code :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;test_33() {
        # test needs a running copytool
        copytool_setup

        mkdir -p $DIR/$tdir

        local f=$DIR/$tdir/$tfile
        local fid=$(make_large_for_progress $f)
        $LFS hsm_archive --archive $HSM_ARCHIVE_NUMBER $f
        wait_request_state $fid ARCHIVE SUCCEED
        $LFS hsm_release $f

        md5sum $f &amp;gt;/dev/null &amp;amp;
        local pid=$!
        wait_request_state $fid RESTORE STARTED

        kill -15 $pid
        sleep 1

        # Check restore trigger process was killed
        local killed=$(ps -o pid,comm hp $pid &amp;gt;/dev/null)

        $LFS hsm_cancel $f

        wait_request_state $fid RESTORE CANCELED
        wait_request_state $fid CANCEL SUCCEED

        [ -z $killed ] ||
                error &quot;Cannot kill process waiting for restore ($killed)&quot;

        copytool_cleanup
}
run_test 33 &quot;Kill a restore waiting process&quot;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="71262" author="jcl" created="Mon, 11 Nov 2013 20:06:10 +0000"  >&lt;p&gt;The logic of unsupported CANCEl is a little bit tricky:&lt;br/&gt;
1) start a request with a CT&lt;br/&gt;
2) CANCEL received, CDT marks the request CANCELED and send CANCEL to CT, CT returns not supported (and print message)&lt;br/&gt;
3) CT doing the request send a progress to CDT, CDT informs the CT the request has been canceled, and mark CANCEL request as done&lt;/p&gt;

</comment>
                            <comment id="72328" author="bfaccini" created="Tue, 26 Nov 2013 17:06:33 +0000"  >&lt;p&gt;J-C,&lt;br/&gt;
Thanks for these highlights on CANCEL process!&lt;br/&gt;
BTW, it helped me to better understand the copytool logs of the failing auto-tests sessions for this ticket, and seems that what happen here is that the CANCEL requests reaches copytool when RESTORE already completed.&lt;/p&gt;

&lt;p&gt;From &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/78b59cb0-3123-11e3-8ec6-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/78b59cb0-3123-11e3-8ec6-52540035b04c&lt;/a&gt; :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lhsmtool_posix[23890]: waiting for message from kernel
lhsmtool_posix[23893]: &apos;[0x200000401:0x28:0x0]&apos; action RESTORE reclen 72, cookie=0x52556a97 &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE request received
lhsmtool_posix[23893]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[23893]: reading stripe rules from &apos;/home/autotest/.autotest/shared_dir/2013-10-08/203349-69989071297660/arc1/0028/0000/0401/0000/0002/0000/0x200000401:0x28:0x0.lov&apos; for &apos;/home/autotest/.autotest/shared_dir/2013-10-08/203349-69989071297660/arc1/0028/0000/0401/0000/0002/0000/0x200000401:0x28:0x0&apos;
lhsmtool_posix[23893]: restoring data from &apos;/home/autotest/.autotest/shared_dir/2013-10-08/203349-69989071297660/arc1/0028/0000/0401/0000/0002/0000/0x200000401:0x28:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf3:0x0]&apos;
lhsmtool_posix[23893]: going to copy data from &apos;/home/autotest/.autotest/shared_dir/2013-10-08/203349-69989071297660/arc1/0028/0000/0401/0000/0002/0000/0x200000401:0x28:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf3:0x0]&apos;
lhsmtool_posix[23893]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[23893]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[23893]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[23893]: %83 
lhsmtool_posix[23893]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[23893]: data restore from &apos;/home/autotest/.autotest/shared_dir/2013-10-08/203349-69989071297660/arc1/0028/0000/0401/0000/0002/0000/0x200000401:0x28:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf3:0x0]&apos; done
lhsmtool_posix[23893]: Action completed, notifying coordinator cookie=0x52556a97, FID=[0x200000401:0x28:0x0], hp_flags=0 err=0
lhsmtool_posix[23890]: copytool fs=lustre archive#=2 item_count=1
lhsmtool_posix[23890]: waiting for message from kernel
lhsmtool_posix[23893]: llapi_hsm_action_end() on &apos;/mnt/lustre/.lustre/fid/0x200000401:0x28:0x0&apos; ok (rc=0) &amp;lt;&amp;lt;&amp;lt;&amp;lt; RETORE done
lhsmtool_posix[23898]: &apos;[0x200000401:0x28:0x0]&apos; action CANCEL reclen 72, cookie=0x52556a97 &amp;lt;&amp;lt;&amp;lt;&amp;lt; CANCEL request received
lhsmtool_posix[23898]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[23898]: cancel not implemented for file system &apos;/mnt/lustre&apos;
exiting: Interrupt
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;From &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/14278fd4-40fe-11e3-a1e8-52540035b04c:&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/14278fd4-40fe-11e3-a1e8-52540035b04c:&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lhsmtool_posix[2462]: waiting for message from kernel
lhsmtool_posix[2466]: &apos;[0x2000088d1:0x29:0x0]&apos; action RESTORE reclen 72, cookie=0x527042b9 &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE request received
lhsmtool_posix[2466]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[2466]: reading stripe rules from &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/222039-69843390548440/arc1/0029/0000/88d1/0000/0002/0000/0x2000088d1:0x29:0x0.lov&apos; for &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/222039-69843390548440/arc1/0029/0000/88d1/0000/0002/0000/0x2000088d1:0x29:0x0&apos;
lhsmtool_posix[2466]: restoring data from &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/222039-69843390548440/arc1/0029/0000/88d1/0000/0002/0000/0x2000088d1:0x29:0x0&apos; to &apos;{VOLATILE}=[0x2000088d2:0xf6:0x0]&apos;
lhsmtool_posix[2466]: going to copy data from &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/222039-69843390548440/arc1/0029/0000/88d1/0000/0002/0000/0x2000088d1:0x29:0x0&apos; to &apos;{VOLATILE}=[0x2000088d2:0xf6:0x0]&apos;
lhsmtool_posix[2466]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[2466]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[2466]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[2466]: %86 
lhsmtool_posix[2466]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[2466]: data restore from &apos;/home/cgearing/.autotest/shared_dir/2013-10-28/222039-69843390548440/arc1/0029/0000/88d1/0000/0002/0000/0x2000088d1:0x29:0x0&apos; to &apos;{VOLATILE}=[0x2000088d2:0xf6:0x0]&apos; done
lhsmtool_posix[2466]: Action completed, notifying coordinator cookie=0x527042b9, FID=[0x2000088d1:0x29:0x0], hp_flags=0 err=0
lhsmtool_posix[2466]: llapi_hsm_action_end() on &apos;/mnt/lustre/.lustre/fid/0x2000088d1:0x29:0x0&apos; ok (rc=0) &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE done
lhsmtool_posix[2462]: copytool fs=lustre archive#=2 item_count=1
lhsmtool_posix[2462]: waiting for message from kernel
lhsmtool_posix[2469]: &apos;[0x2000088d1:0x29:0x0]&apos; action CANCEL reclen 72, cookie=0x527042b9 &amp;lt;&amp;lt;&amp;lt;&amp;lt; CANCEL request received
lhsmtool_posix[2469]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[2469]: cancel not implemented for file system &apos;/mnt/lustre&apos;
exiting: Interrupt&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;From &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/c4f891ea-4924-11e3-a79b-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/c4f891ea-4924-11e3-a79b-52540035b04c&lt;/a&gt; :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lhsmtool_posix[18601]: copytool fs=lustre archive#=2 item_count=1
lhsmtool_posix[18601]: waiting for message from kernel
lhsmtool_posix[18606]: &apos;[0x200000401:0x29:0x0]&apos; action RESTORE reclen 72, cookie=0x527df027 &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE request received
lhsmtool_posix[18606]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[18606]: reading stripe rules from &apos;/home/autotest/.autotest/shared_dir/2013-11-08/154450-70330946069980/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0.lov&apos; for &apos;/home/autotest/.autotest/shared_dir/2013-11-08/154450-70330946069980/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos;
lhsmtool_posix[18606]: restoring data from &apos;/home/autotest/.autotest/shared_dir/2013-11-08/154450-70330946069980/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos;
lhsmtool_posix[18606]: going to copy data from &apos;/home/autotest/.autotest/shared_dir/2013-11-08/154450-70330946069980/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos;
lhsmtool_posix[18606]: bandwith control: excess=2.026880E+05 sleep for 193298us
lhsmtool_posix[18606]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[18606]: bandwith control: excess=2.026880E+05 sleep for 193298us
lhsmtool_posix[18606]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[18606]: bandwith control: excess=2.026880E+05 sleep for 193298us
lhsmtool_posix[18606]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[18606]: %83 
lhsmtool_posix[18606]: bandwith control: excess=2.026880E+05 sleep for 193298us
lhsmtool_posix[18606]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[18606]: data restore from &apos;/home/autotest/.autotest/shared_dir/2013-11-08/154450-70330946069980/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos; done
lhsmtool_posix[18606]: Action completed, notifying coordinator cookie=0x527df027, FID=[0x200000401:0x29:0x0], hp_flags=0 err=0
lhsmtool_posix[18601]: copytool fs=lustre archive#=2 item_count=1
lhsmtool_posix[18601]: waiting for message from kernel
lhsmtool_posix[18606]: llapi_hsm_action_end() on &apos;/mnt/lustre/.lustre/fid/0x200000401:0x29:0x0&apos; ok (rc=0)  &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE done
lhsmtool_posix[18608]: &apos;[0x200000401:0x29:0x0]&apos; action CANCEL reclen 72, cookie=0x527df027 &amp;lt;&amp;lt;&amp;lt;&amp;lt; CANCEL request received
lhsmtool_posix[18608]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[18608]: cancel not implemented for file system &apos;/mnt/lustre&apos;
exiting: Interrupt
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;When copytool log for a successful run of test_33 looks more like :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lhsmtool_posix[21009]: waiting for message from kernel
lhsmtool_posix[21011]: &apos;[0x200000401:0x29:0x0]&apos; action RESTORE reclen 72, cookie=0x52904448 &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE request received
lhsmtool_posix[21011]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[21011]: reading stripe rules from &apos;/home/cgearing/.autotest/shared_dir/2013-11-22/125116-70304641718120/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0.lov&apos; for &apos;/home/cgearing/.autotest/shared_dir/2013-11-22/125116-70304641718120/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos;
lhsmtool_posix[21011]: restoring data from &apos;/home/cgearing/.autotest/shared_dir/2013-11-22/125116-70304641718120/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos;
lhsmtool_posix[21011]: going to copy data from &apos;/home/cgearing/.autotest/shared_dir/2013-11-22/125116-70304641718120/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos;
lhsmtool_posix[21009]: copytool fs=lustre archive#=2 item_count=1
lhsmtool_posix[21009]: waiting for message from kernel
lhsmtool_posix[21012]: &apos;[0x200000401:0x29:0x0]&apos; action CANCEL reclen 72, cookie=0x52904448 &amp;lt;&amp;lt;&amp;lt;&amp;lt; CANCEL request received
lhsmtool_posix[21012]: processing file &apos;d0.sanity-hsm/d33/f.sanity-hsm.33&apos;
lhsmtool_posix[21012]: cancel not implemented for file system &apos;/mnt/lustre&apos;
lhsmtool_posix[21011]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[21011]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[21011]: bandwith control: excess=1.251264E+06 sleep for 1193298us
lhsmtool_posix[21011]: %86 
lhsmtool_posix[21011]: progress ioctl for copy &apos;/home/cgearing/.autotest/shared_dir/2013-11-22/125116-70304641718120/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos;-&amp;gt;&apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos; failed: Operation canceled (125)
lhsmtool_posix[21011]: cannot copy data from &apos;/home/cgearing/.autotest/shared_dir/2013-11-22/125116-70304641718120/arc1/0029/0000/0401/0000/0002/0000/0x200000401:0x29:0x0&apos; to &apos;{VOLATILE}=[0x200000402:0xf4:0x0]&apos;: Operation canceled (125)
lhsmtool_posix[21011]: Action completed, notifying coordinator cookie=0x52904448, FID=[0x200000401:0x29:0x0], hp_flags=0 err=125 &amp;lt;&amp;lt;&amp;lt;&amp;lt; RESTORE request end with ECANCELED errno
lhsmtool_posix[21011]: llapi_hsm_action_end() on &apos;/mnt/lustre/.lustre/fid/0x200000401:0x29:0x0&apos; ok (rc=0)
exiting: Interrupt
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="72515" author="bfaccini" created="Fri, 29 Nov 2013 10:10:02 +0000"  >&lt;p&gt;I am puzzled with the &quot;waiting for message from kernel&quot; and &quot;copytool fs=lustre archive#=2 item_count=1&quot; copytool main-thread msgs sequence here, there should be some buffering (with CT_TRACE()/llapi_&lt;span class=&quot;error&quot;&gt;&amp;#91;error,printf&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;_callback&amp;#93;&lt;/span&gt;() usage) causing some out-of-order log recording of msgs from copytool main and dynamically/async created threads to serve requests.&lt;/p&gt;

&lt;p&gt;But even with that in mind, how can we explain that copytool main thread/loop could be blocked during more than 30s allowing the RESTORE to finish and before to receive/handle the CANCEL request ?? Or may be this delay comes from the Client requesting the CANCEL and or the MDT/CDT side ??&lt;/p&gt;

&lt;p&gt;BTW, the failed auto-tests sessions Lustre debug-logs do not help very much since they do not trace HSM activity, and particularly the main copytool thread. As far as I can see CANCEL request leave Client on-time, but then I am a bit lost with the sequence of llog record creation/update and communucations between CDT and CT &#8230;&lt;/p&gt;

&lt;p&gt;Anyway will try to find a way to retrieve where the CANCEL request handling delay comes from.&lt;/p&gt;</comment>
                            <comment id="73357" author="bfaccini" created="Thu, 12 Dec 2013 14:03:17 +0000"  >&lt;p&gt;Now I see 2 possible ways to fix this issue :&lt;/p&gt;

&lt;p&gt;      _ in case there are some &quot;orphan&quot; active_requests from previous sub-test, add +1 to max_requests before running the hsm_cancel command. This may handle the case where there is only one &quot;real&quot; slot left that will cause the Restore and Cancel operations to be serialized.&lt;/p&gt;

&lt;p&gt;      _ Use &quot;wait_all_done()&quot; instead of &quot;wait_request_state $fid RESTORE CANCELED&quot; + &quot;&quot;wait_request_state $fid CANCEL SUCCEED&quot;, and then check/handle case of RESTORE is SUCCEED/CANCELED and CANCEL is SUCCEED/FAILED with get_request_state(). This will allow to detect if Cancel was too late thus not fail. Will require a cdt_purge at beginning of test to ensure wait_all_done success.&lt;/p&gt;
</comment>
                            <comment id="79115" author="bfaccini" created="Wed, 12 Mar 2014 13:18:11 +0000"  >&lt;p&gt;Seem that all failures of sanity-hsm/test_33 subtest, since mid-december, are not single errors but part of bunch of subtests errors caused by some other issue.&lt;/p&gt;

&lt;p&gt;But anyway, in order to strengthen test_33 subtest, I pushed a patch, to implement the 2nd way I described before, at &lt;a href=&quot;http://review.whamcloud.com/9600&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/9600&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="82106" author="bfaccini" created="Mon, 21 Apr 2014 22:56:47 +0000"  >&lt;p&gt;Patch has landed. No more Maloo/auto-tests failures. Marking ticket as solved. To be re-opened if problem re-occurs.&lt;/p&gt;</comment>
                            <comment id="86581" author="jamesanunez" created="Fri, 13 Jun 2014 16:56:56 +0000"  >&lt;p&gt;Patch for b2_5 at &lt;a href=&quot;http://review.whamcloud.com/#/c/10712/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/10712/&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzw5c7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>10982</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>