<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:39:35 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4093] sanity-hsm test_24d: wanted &apos;SUCCEED&apos; got &apos;WAITING&apos;</title>
                <link>https://jira.whamcloud.com/browse/LU-4093</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Nathaniel Clark &amp;lt;nathaniel.l.clark@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;http://maloo.whamcloud.com/test_sets/04c602fa-3258-11e3-9de6-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/04c602fa-3258-11e3-9de6-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_24d failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Cannot send HSM request (use of /mnt/lustre2/d0.sanity-hsm/d24/f.sanity-hsm.24d): Read-only file system&lt;br/&gt;
:&lt;br/&gt;
:&lt;br/&gt;
CMD: wtm-27vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.actions | awk &apos;/&apos;0x200008101:0x1f:0x0&apos;.*action=&apos;ARCHIVE&apos;/ { print \$13 }&apos; | cut -f2 -d=&lt;br/&gt;
Update not seen after 100s: wanted &apos;SUCCEED&apos; got &apos;WAITING&apos;&lt;br/&gt;
 sanity-hsm test_24d: @@@@@@ FAIL: request on 0x200008101:0x1f:0x0 is not SUCCEED &lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Info required for matching: sanity-hsm 24d&lt;/p&gt;</description>
                <environment></environment>
        <key id="21374">LU-4093</key>
            <summary>sanity-hsm test_24d: wanted &apos;SUCCEED&apos; got &apos;WAITING&apos;</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>HSM</label>
                            <label>zfs</label>
                    </labels>
                <created>Fri, 11 Oct 2013 16:18:49 +0000</created>
                <updated>Mon, 15 Sep 2014 17:09:06 +0000</updated>
                            <resolved>Tue, 10 Dec 2013 09:16:56 +0000</resolved>
                                    <version>Lustre 2.5.0</version>
                                    <fixVersion>Lustre 2.6.0</fixVersion>
                    <fixVersion>Lustre 2.5.1</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="69395" author="bfaccini" created="Mon, 21 Oct 2013 13:45:48 +0000"  >&lt;p&gt;In fact it is not only sanity-hsm/test_24d which fails due to this &quot;FAIL: request on 0x200008101:0x1f:0x0 is not SUCCEED&quot; error, but also sub-tests test_26/test_27b/test_28 !!&lt;/p&gt;

&lt;p&gt;This error always happen during a file Archive, when similar actions in intermediate sub-tests were successful.&lt;/p&gt;

&lt;p&gt;sanity-hsm/test_33 also failed during the same session for issue (&quot;request on 0x200000401:0x28:0x0 is not CANCELED&quot;) addressed by &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4086&quot; title=&quot;Test failure on test suite sanity-hsm, subtest test_33&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4086&quot;&gt;&lt;del&gt;LU-4086&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;I tried to setup a test platform (2 Clients, single MGS+MDS, single OSS, ZFS-only, running this particular build), but I am currently unable to reproduce.&lt;/p&gt;</comment>
                            <comment id="69563" author="bfaccini" created="Tue, 22 Oct 2013 18:08:39 +0000"  >&lt;p&gt;I have been able to reproduce problem on the local platform I described before, this by running &quot;sanity-hsm.sh 24c 24d 26 27b 28&quot; test/sub-tests in a loop. And it appears the failure (ie requests/actions staying in &quot;WAITING&quot; state from their start/send and during 100s) is caused by &quot;hsm/max_requests&quot; (default=3) orphan requests which still appear with action &quot;STARTED&quot; state (and also in &quot;/proc/fs/lustre/mdt/&amp;lt;MDT&amp;gt;/hsm/requests&quot;) preventing new requests to start.&lt;/p&gt;

&lt;p&gt;Doing more debugging I found that the orphan requests come from test_24c and are the latest hsm_archive requests in the sub-test that do not wait end of request and call copytool_cleanup in a raw which may cause this last request to never end nor be cleared from log.&lt;/p&gt;

&lt;p&gt;Thus every following hsm_archive will fail in following sub-tests like 24d/26/27b/28 with the same msg/error symptom than for this ticket, this until test_40 clears all orphans using/calling &quot;cdt_purge&quot; (ie &quot;echo purge &amp;gt; /proc/fs/lustre/mdt/&amp;lt;MDT&amp;gt;/hsm_control&quot;) and/or request_timeout expires ...&lt;/p&gt;

&lt;p&gt;Will try to add a systematic cdt_shutdown/cdt_enable (ie, &quot;echo &lt;span class=&quot;error&quot;&gt;&amp;#91;shutdown,enable&amp;#93;&lt;/span&gt; &amp;gt; /proc/fs/lustre/mdt/&amp;lt;MDT&amp;gt;/hsm_control&quot;) in copytool_&lt;span class=&quot;error&quot;&gt;&amp;#91;setup,cleanup&amp;#93;&lt;/span&gt;() and see if it fixes as I expect.&lt;/p&gt;</comment>
                            <comment id="69633" author="bfaccini" created="Wed, 23 Oct 2013 14:26:41 +0000"  >&lt;p&gt;Found more infos that confirm or better explain earlier findings :&lt;/p&gt;

&lt;p&gt;       _ in fact cdt_purge only clear actions but not requests ... Thus need to cdt_&lt;span class=&quot;error&quot;&gt;&amp;#91;shutdown+enable&amp;#93;&lt;/span&gt; (or better cdt_restart!) to clear orphan requests immediately.&lt;/p&gt;

&lt;p&gt;       _ orphan requests can also clear after timeout, this is confirmed by the timings found in the test suite run that caused this ticket to be created, and where I found that some point of time (around 1 hour/3600s later than test_24c ran and test_24d 1st triggered the WAITING hsm_archive requests state) during test_30 hsm_archive requests restarted to be handled ... :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;.......
 [/usr/bin/lfs] [hsm_archive] [/mnt/lustre/d0.sanity-hsm/d24/f.sanity-hsm.24c]
2+0 records in
2+0 records out
2097152 bytes (2.1 MB) copied, 2.39885 s, 874 kB/s
0x200008101:0x1e:0x0
CMD: wtm-21vm3 /usr/sbin/lctl set_param -n mdt.lustre-MDT0000.hsm.other_request_mask
running as uid/gid/euid/egid 500/500/500/500, groups:
 [/usr/bin/lfs] [hsm_archive] [/mnt/lustre/d0.sanity-hsm/d24/f.sanity-hsm.24c]
Cannot send HSM request (use of /mnt/lustre/d0.sanity-hsm/d24/f.sanity-hsm.24c): Operation not permitted
CMD: wtm-21vm3 /usr/sbin/lctl set_param -n mdt.lustre-MDT0000.hsm.other_request_mask=archive
running as uid/gid/euid/egid 500/500/500/500, groups:
 [/usr/bin/lfs] [hsm_archive] [/mnt/lustre/d0.sanity-hsm/d24/f.sanity-hsm.24c]  &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt; last archive request in test_24c

CMD: wtm-21vm5 pkill -INT -x lhsmtool_posix  &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt; immediate copytool_cleanup
Copytool is stopped on wtm-21vm5
CMD: wtm-21vm3 /usr/sbin/lctl set_param -n mdt.lustre-MDT0000.hsm.user_request_mask=RESTORE
CMD: wtm-21vm3 /usr/sbin/lctl set_param -n mdt.lustre-MDT0000.hsm.group_request_mask=RESTORE
CMD: wtm-21vm3 /usr/sbin/lctl set_param -n mdt.lustre-MDT0000.hsm.other_request_mask=RESTORE
Resetting fail_loc on all nodes...CMD: wtm-21vm3,wtm-21vm4,wtm-21vm5,wtm-21vm6.rosso.whamcloud.com lctl set_param -n fail_loc=0 2&amp;gt;/dev/null || true
done.
CMD: wtm-21vm3,wtm-21vm4,wtm-21vm5 rc=\$([ -f /proc/sys/lnet/catastrophe ] &amp;amp;&amp;amp;
		echo \$(&amp;lt; /proc/sys/lnet/catastrophe) || echo 0);
		if [ \$rc -ne 0 ]; then echo \$(hostname): \$rc; fi
		exit \$rc
PASS 24c (11s)

== sanity-hsm test 24d: check that read-only mounts are respected == 23:38:25 (1381473505)
CMD: wtm-21vm5 pkill -CONT -x lhsmtool_posix
Wakeup copytool agt1 on wtm-21vm5
2+0 records in
2+0 records out
2097152 bytes (2.1 MB) copied, 1.79625 s, 1.2 MB/s
Cannot send HSM request (use of /mnt/lustre2/d0.sanity-hsm/d24/f.sanity-hsm.24d): Read-only file system
CMD: wtm-21vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x1f:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 0s: from &apos;&apos; to &apos;WAITING&apos;  &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt; 1st valid archive request in test_24d is WAITING direct!!!
Waiting 100 secs for update
CMD: wtm-21vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x1f:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=

.......

Update not seen after 100s: wanted &apos;SUCCEED&apos; got &apos;WAITING&apos;
 sanity-hsm test_24d: @@@@@@ FAIL: request on 0x200008101:0x1f:0x0 is not SUCCEED  &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt; test_24d fails at 1st archive

.......

Same for test_26/27b/28 (only following tests with archives...) at their 1st archive request !!!

.......

== sanity-hsm test 30b: Restore at exec (release case) == 00:38:49 (1381477129)
CMD: wtm-21vm5 pkill -CONT -x lhsmtool_posix
Purging archive on wtm-21vm5
CMD: wtm-21vm5 rm -rf /home/cgearing/.autotest/shared_dir/2013-10-10/082939-69981897718180/arc1/*
Starting copytool agt1 on wtm-21vm5
CMD: wtm-21vm5 mkdir -p /home/cgearing/.autotest/shared_dir/2013-10-10/082939-69981897718180/arc1
CMD: wtm-21vm5 lhsmtool_posix  --daemon --hsm-root /home/cgearing/.autotest/shared_dir/2013-10-10/082939-69981897718180/arc1 --bandwidth 1 /mnt/lustre &amp;lt; /dev/null &amp;gt; /logdir/test_logs/2013-10-10/lustre-reviews-el6-x86_64--review-zfs--1_3_1__18725__-69981897718180-082939/sanity-hsm.test_30b.copytool_log.wtm-21vm5.log 2&amp;gt;&amp;amp;1
CMD: wtm-21vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x24:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 0s: from &apos;&apos; to &apos;WAITING&apos;
Waiting 100 secs for update

.......

CMD: wtm-21vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x24:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
CMD: wtm-21vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x24:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Changed after 96s: from &apos;WAITING&apos; to &apos;STARTED&apos;  &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt; archives only restarted during test_30b about 1h/3600s (default timeout!)
CMD: wtm-21vm3 /usr/sbin/lctl get_param -n mdt.lustre-MDT0000.hsm.agent_actions | awk &apos;/&apos;0x200008101:0x24:0x0&apos;.*action=&apos;ARCHIVE&apos;/ {print \$13}&apos; | cut -f2 -d=
Updated after 97s: wanted &apos;SUCCEED&apos; got &apos;SUCCEED&apos;

........

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;       _ what is the best way to fix this ? As I already thought and commented &quot;add a systematic cdt_shutdown/cdt_enable (or cdt_restart) in copytool_cleanup()&quot; or only ensure to allow copytool to get enough time to handle requests before cleanup (like in test_24c) ?&lt;/p&gt;

&lt;p&gt;       _ also I need to check, if it is it allowed to kill/end copytool when there are requests being processed and thus if it is expected for these requests to stay orphan until they reach their timeout or CDT restarts ? BTW, running tests in this configuration causes CDT to complain with these (normal/expected?) msgs :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 22856:0:(mdt_coordinator.c:338:mdt_coordinator_cb()) lustrez-MDT0000: Cannot cleanup timeouted request: [0x200000401:0x2590:0x0] for cookie 0x52678bbd action=ARCHIVE
LustreError: 22856:0:(mdt_coordinator.c:338:mdt_coordinator_cb()) Skipped 1 previous similar message
LustreError: 22856:0:(mdt_coordinator.c:1448:mdt_hsm_update_request_state()) lustrez-MDT0000: Cannot find running request for cookie 0x52678e26 on fid=[0x200000401:0x25a2:0x0]
LustreError: 22856:0:(mdt_coordinator.c:1448:mdt_hsm_update_request_state()) Skipped 1 previous similar message
LustreError: 22856:0:(mdt_coordinator.c:338:mdt_coordinator_cb()) lustrez-MDT0000: Cannot cleanup timeouted request: [0x200000401:0x25a2:0x0] for cookie 0x52678e26 action=ARCHIVE
LustreError: 22856:0:(mdt_coordinator.c:338:mdt_coordinator_cb()) Skipped 1 previous similar message
LustreError: 22856:0:(mdt_coordinator.c:1448:mdt_hsm_update_request_state()) lustrez-MDT0000: Cannot find running request for cookie 0x5267908a on fid=[0x200000401:0x25b4:0x0]
LustreError: 22856:0:(mdt_coordinator.c:1448:mdt_hsm_update_request_state()) Skipped 1 previous similar message
LustreError: 22856:0:(mdt_coordinator.c:338:mdt_coordinator_cb()) lustrez-MDT0000: Cannot cleanup timeouted request: [0x200000401:0x25b4:0x0] for cookie 0x5267908a action=ARCHIVE
LustreError: 22856:0:(mdt_coordinator.c:338:mdt_coordinator_cb()) Skipped 1 previous similar message
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="70611" author="bfaccini" created="Mon, 4 Nov 2013 15:07:00 +0000"  >&lt;p&gt;Fix to prevent zombie requests during CT/copytool restart is at &lt;a href=&quot;http://review.whamcloud.com/8157&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8157&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="71829" author="adilger" created="Mon, 18 Nov 2013 21:31:26 +0000"  >&lt;p&gt;Patch 8157 has landed on 2013-11-13, but I still see tests failing with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4093&quot; title=&quot;sanity-hsm test_24d: wanted &amp;#39;SUCCEED&amp;#39; got &amp;#39;WAITING&amp;#39;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4093&quot;&gt;&lt;del&gt;LU-4093&lt;/del&gt;&lt;/a&gt;/&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4126&quot; title=&quot;sanity-hsm test_15 failure:  &amp;#39;requests did not complete&amp;#39; &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4126&quot;&gt;&lt;del&gt;LU-4126&lt;/del&gt;&lt;/a&gt; in the past few days.  Is that just because the test queue is so long that the results we are seeing today are for patches that were based on code not including the fix?&lt;/p&gt;

&lt;p&gt;If that can be verified by checking the parent commit of recent test failures does NOT include change 8157, then I guess this bug can be closed.  It looks from the results that &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4093&quot; title=&quot;sanity-hsm test_24d: wanted &amp;#39;SUCCEED&amp;#39; got &amp;#39;WAITING&amp;#39;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4093&quot;&gt;&lt;del&gt;LU-4093&lt;/del&gt;&lt;/a&gt; shows up as lustre-hsm passing only 73/91 tests.  There are still cases with 90/91 tests passing, so that needs to be a separate bug.&lt;/p&gt;</comment>
                            <comment id="71871" author="bfaccini" created="Tue, 19 Nov 2013 11:19:50 +0000"  >&lt;p&gt;Humm thanks Andreas to chase this, seems that original patch #8157 (patch-set #3) has a typo (MDT_HSMCTRL vs mdt_hsmctrl) and an inverted test (&lt;span class=&quot;error&quot;&gt;&amp;#91;echo $oldstate | grep stop || continue&amp;#93;&lt;/span&gt; vs &lt;span class=&quot;error&quot;&gt;&amp;#91;echo $oldstate | grep stop &amp;amp;&amp;amp; continue&amp;#93;&lt;/span&gt;) that should prevent it to work as expected and rather with reverted logic (only stopped CDT will be stopped) &#8230; &lt;br/&gt;
The fact itself passed auto-tests could be due to the condition it is intended to fix did not show-up during its own testing !!&lt;br/&gt;
Just pushed &lt;a href=&quot;http://review.whamcloud.com/8329&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8329&lt;/a&gt; to fix this.&lt;/p&gt;</comment>
                            <comment id="72477" author="bfaccini" created="Thu, 28 Nov 2013 13:45:57 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/8329&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8329&lt;/a&gt; has just landed, so need to wait a week or 2 to verify its effects and close.&lt;/p&gt;
</comment>
                            <comment id="73182" author="bfaccini" created="Tue, 10 Dec 2013 09:16:56 +0000"  >&lt;p&gt;No more related failures for master builds reported since Nov 18th. So as expected the 2 changes for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4093&quot; title=&quot;sanity-hsm test_24d: wanted &amp;#39;SUCCEED&amp;#39; got &amp;#39;WAITING&amp;#39;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4093&quot;&gt;&lt;del&gt;LU-4093&lt;/del&gt;&lt;/a&gt; made it.&lt;/p&gt;</comment>
                            <comment id="82293" author="jamesanunez" created="Wed, 23 Apr 2014 16:07:28 +0000"  >&lt;p&gt;It looks like I just hit this issue with b2_5 in review-zfs at &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/e11b4944-c822-11e3-888b-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/e11b4944-c822-11e3-888b-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="93933" author="bfaccini" created="Sat, 13 Sep 2014 14:53:35 +0000"  >&lt;p&gt;Hello Jian, humm looks different I think, at least because in all cases you noticed the copytool_log is not present for test_24d !&lt;/p&gt;

&lt;p&gt;And if I look precisely into the logs/msgs, &quot;Wakeup copytool agt1 on ...&quot; should not be printed after a successfull copytool_cleanup, from previous sub-test, and copytool_setup, within new sub-test, in sequence because the implied &quot;pkill&quot; command should have fail if no copytool thread from previous sub-test can be found after its stop/kill in copytool_cleanup.&lt;br/&gt;
So a possible scenario is that some copytool thread from previous test_24c sub-test must be somewhat stuck and still not terminated at the time test_24d tries to restart copytool ... But finally dies, and in fact we may finally run test_24d with no copytool started !&lt;/p&gt;

&lt;p&gt;So this should be addressed in a new ticket, and the fix should be to ensure full death of ALL copytool threads in copytool_cleanup().&lt;br/&gt;
If you agree, I will create ticket, assign to myself, add you in watchers list, and push a fix soon.&lt;/p&gt;


</comment>
                            <comment id="93943" author="yujian" created="Sun, 14 Sep 2014 06:30:22 +0000"  >&lt;blockquote&gt;&lt;p&gt;If you agree, I will create ticket, assign to myself, add you in watchers list, and push a fix soon.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Thank you, Bruno. Please do.&lt;/p&gt;</comment>
                            <comment id="93981" author="bfaccini" created="Mon, 15 Sep 2014 15:28:03 +0000"  >&lt;p&gt;Just created &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5622&quot; title=&quot;copytool_cleanup function should check/wait for copytool death&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5622&quot;&gt;&lt;del&gt;LU-5622&lt;/del&gt;&lt;/a&gt; to address this new problem.&lt;/p&gt;

&lt;p&gt;Just want to add that the fact it seems to occur more frequently between test_24c and test_24d sub-tests switch may come from specific/enhanced cleanup actions in test_24c ...&lt;/p&gt;

&lt;p&gt;Also, I think we need to change the link of the failures you listed from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4093&quot; title=&quot;sanity-hsm test_24d: wanted &amp;#39;SUCCEED&amp;#39; got &amp;#39;WAITING&amp;#39;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4093&quot;&gt;&lt;del&gt;LU-4093&lt;/del&gt;&lt;/a&gt; to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5622&quot; title=&quot;copytool_cleanup function should check/wait for copytool death&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5622&quot;&gt;&lt;del&gt;LU-5622&lt;/del&gt;&lt;/a&gt;, what do you think ?&lt;/p&gt;
</comment>
                            <comment id="94002" author="yujian" created="Mon, 15 Sep 2014 17:09:06 +0000"  >&lt;blockquote&gt;&lt;p&gt;Also, I think we need to change the link of the failures you listed from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4093&quot; title=&quot;sanity-hsm test_24d: wanted &amp;#39;SUCCEED&amp;#39; got &amp;#39;WAITING&amp;#39;&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4093&quot;&gt;&lt;del&gt;LU-4093&lt;/del&gt;&lt;/a&gt; to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5622&quot; title=&quot;copytool_cleanup function should check/wait for copytool death&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5622&quot;&gt;&lt;del&gt;LU-5622&lt;/del&gt;&lt;/a&gt;, what do you think ?&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Agreed. Just changed. Thank you, Bruno.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="21948">LU-4235</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="21520">LU-4126</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzw5f3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>10999</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>