<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:28:30 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2822] softlockups, evictions during recovery-scale</title>
                <link>https://jira.whamcloud.com/browse/LU-2822</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;running the mds-recovery-scale test. Not really sure what is going on. The test completes one failover, then activity ceases. Appear to be having soft lockups on the clients, pdflush is blocked, followed/prceeded by evictions from OSTs. &lt;br/&gt;
Example, this is during an MDS failover event:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Feb 15 09:59:11 ehyperion607 Lustre: 893:0:(&lt;span class=&quot;code-keyword&quot;&gt;import&lt;/span&gt;.c:517:import_select_connection()) lustre-MDT0000-mdc-ffff8101bf048400: tried all connections, increasing latency to 8s
Feb 15 09:59:11 ehyperion607 Lustre: 893:0:(&lt;span class=&quot;code-keyword&quot;&gt;import&lt;/span&gt;.c:517:import_select_connection()) Skipped 16 previous similar messages
Feb 15 09:59:11 ehyperion607 Lustre: 892:0:(client.c:1529:ptlrpc_expire_one_request()) @@@ Request x1426985036308729 sent from lustre-OST0006-osc-ffff8101bf048400 to NID 192.168.120.86@o2ib 12s ago has timed out (12s prior to deadline).
Feb 15 09:59:11 ehyperion607 req@ffff8101a84e8400 x1426985036308729/t0 o8-&amp;gt;lustre-OST0006_UUID@192.168.120.86@o2ib:28/4 lens 368/584 e 0 to 1 dl 1360951115 ref 2 fl Rpc:N/0/0 rc 0/0
Feb 15 09:59:11 ehyperion607 Lustre: 892:0:(client.c:1529:ptlrpc_expire_one_request()) Skipped 30 previous similar messages
Feb 15 09:59:11 ehyperion607 BUG: soft lockup - CPU#0 stuck &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; 60s! [events/0:50]
Feb 15 09:59:11 ehyperion607 CPU 0:
Feb 15 09:59:11 ehyperion607 Modules linked in: mgc(U) lustre(U) lov(U) mdc(U) lquota(U) osc(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) libcfs(U) acpi_cpufreq freq_table mperf ib_ipoib rdma_ucm rdma_cm iw_cm ib_addr ib_ucm ib_uverbs ib_umad mlx4_ib mlx4_core ipoib_helper ib_cm ib_sa ib_mad ib_core ipv6 xfrm_nalgo crypto_api dm_mirror dm_log dm_multipath scsi_dh dm_mod video backlight sbs power_meter hwmon i2c_ec dell_wmi wmi button battery asus_acpi acpi_memhotplug ac parport_pc lp parport joydev shpchp ahci i7core_edac edac_mc libata i2c_i801 pcspkr i2c_core scsi_mod uhci_hcd tpm_tis tpm tpm_bios nfs nfs_acl lockd sunrpc e1000e
Feb 15 09:59:11 ehyperion607 Pid: 50, comm: events/0 Tainted: G     ---- 2.6.18-348.1.1.el5 #1
Feb 15 09:59:11 ehyperion607 RIP: 0010:[&amp;lt;ffffffff80077925&amp;gt;]  [&amp;lt;ffffffff80077925&amp;gt;] __smp_call_function_many+0x96/0xbc
Feb 15 09:59:11 ehyperion607 RSP: 0018:ffff8101bf953d40  EFLAGS: 00000297
Feb 15 09:59:11 ehyperion607 RAX: 0000000000000006 RBX: 0000000000000007 RCX: 0000000000000000
Feb 15 09:59:11 ehyperion607 RDX: 00000000000000ff RSI: 00000000000000ff RDI: 00000000000000c0
Feb 15 09:59:11 ehyperion607 RBP: ffff8101bf953e10 R08: 0000000000000008 R09: 0000000000000038
Feb 15 09:59:11 ehyperion607 R10: ffff8101bf953ce0 R11: ffff8101c571c500 R12: 00000002ffffffff
Feb 15 09:59:11 ehyperion607 R13: ffffffff8008dedf R14: 00000001ffffffff R15: ffffffff80157cde
Feb 15 09:59:11 ehyperion607 FS:  0000000000000000(0000) GS:ffffffff80435000(0000) knlGS:0000000000000000
Feb 15 09:59:11 ehyperion607 CS:  0010 DS: 0018 ES: 0018 CR0: 000000008005003b
Feb 15 09:59:11 ehyperion607 CR2: 00002aaaabb7c088 CR3: 00000001b73bb000 CR4: 00000000000006a0
Feb 15 09:59:11 ehyperion607
Feb 15 09:59:11 ehyperion607 Call Trace:
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80073de0&amp;gt;] mcheck_check_cpu+0x0/0x30
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80073de0&amp;gt;] mcheck_check_cpu+0x0/0x30
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80077a27&amp;gt;] smp_call_function_many+0x38/0x4c
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80073de0&amp;gt;] mcheck_check_cpu+0x0/0x30
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80077b18&amp;gt;] smp_call_function+0x4e/0x5e
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80073de0&amp;gt;] mcheck_check_cpu+0x0/0x30
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80073020&amp;gt;] mcheck_timer+0x0/0x6c
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80096be9&amp;gt;] on_each_cpu+0x10/0x22
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8007303c&amp;gt;] mcheck_timer+0x1c/0x6c
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8004d8b9&amp;gt;] run_workqueue+0x9e/0xfb
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8004a102&amp;gt;] worker_thread+0x0/0x122
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8004a1f2&amp;gt;] worker_thread+0xf0/0x122
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8008f38d&amp;gt;] default_wake_function+0x0/0xe
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80032c2c&amp;gt;] kthread+0xfe/0x132
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8005dfc1&amp;gt;] child_rip+0xa/0x11
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff80032b2e&amp;gt;] kthread+0x0/0x132
Feb 15 09:59:11 ehyperion607 [&amp;lt;ffffffff8005dfb7&amp;gt;] child_rip+0x0/0x11
Feb 15 09:59:11 ehyperion607
Feb 15 09:59:11 ehyperion607 Lustre: 893:0:(&lt;span class=&quot;code-keyword&quot;&gt;import&lt;/span&gt;.c:517:import_select_connection()) lustre-OST0006-osc-ffff8101bf048400: tried all connections, increasing latency to 9s
Feb 15 09:59:11 ehyperion607 Lustre: 893:0:(&lt;span class=&quot;code-keyword&quot;&gt;import&lt;/span&gt;.c:517:import_select_connection()) Skipped 25 previous similar messages
Feb 15 09:59:54 ehyperion607 INFO: task pdflush:543 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.
Feb 15 09:59:54 ehyperion607 &lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.
Feb 15 09:59:54 ehyperion607 pdflush       D ffff8101c546a7a0     0   543    171           544   542 (L-TLB)
Feb 15 09:59:54 ehyperion607 ffff81033f57fb30 0000000000000046 ffff81033f57fa80 0000000000000000
Feb 15 09:59:54 ehyperion607 ffffffff88718220 000000000000000a ffff8101bf566820 ffff8101c56de7a0
Feb 15 09:59:54 ehyperion607 00009ea40c4605ce 00000000001a7575 ffff8101bf566a08 0000000705c2b6a0
Feb 15 09:59:54 ehyperion607 Call Trace:
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8006ed48&amp;gt;] do_gettimeofday+0x40/0x90
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff80029173&amp;gt;] sync_page+0x0/0x43
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800637de&amp;gt;] io_schedule+0x3f/0x67
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800291b1&amp;gt;] sync_page+0x3e/0x43
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff80063922&amp;gt;] __wait_on_bit_lock+0x36/0x66
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8003ff85&amp;gt;] __lock_page+0x5e/0x64
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800a3c31&amp;gt;] wake_bit_function+0x0/0x23
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8001d54e&amp;gt;] mpage_writepages+0x14f/0x37d
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff88846c30&amp;gt;] :lustre:ll_writepage_26+0x0/0x10
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8005aee2&amp;gt;] do_writepages+0x20/0x2f
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8002ff0f&amp;gt;] __writeback_single_inode+0x1a2/0x31c
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8002159a&amp;gt;] sync_sb_inodes+0x1b7/0x271
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800a39eb&amp;gt;] keventd_create_kthread+0x0/0xc4
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8005130b&amp;gt;] writeback_inodes+0x82/0xd8
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800cd802&amp;gt;] wb_kupdate+0xf0/0x16a
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800568b2&amp;gt;] pdflush+0x0/0x1fb
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff80056a03&amp;gt;] pdflush+0x151/0x1fb
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800cd712&amp;gt;] wb_kupdate+0x0/0x16a
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff80032c2c&amp;gt;] kthread+0xfe/0x132
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8005dfc1&amp;gt;] child_rip+0xa/0x11
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff800a39eb&amp;gt;] keventd_create_kthread+0x0/0xc4
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff80032b2e&amp;gt;] kthread+0x0/0x132
Feb 15 09:59:54 ehyperion607 [&amp;lt;ffffffff8005dfb7&amp;gt;] child_rip+0x0/0x11
Feb 15 09:59:54 ehyperion607
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>CHAOS4/RHEL5 Old Hyperion system</environment>
        <key id="17590">LU-2822</key>
            <summary>softlockups, evictions during recovery-scale</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="niu">Niu Yawei</assignee>
                                    <reporter username="cliffw">Cliff White</reporter>
                        <labels>
                    </labels>
                <created>Fri, 15 Feb 2013 13:36:23 +0000</created>
                <updated>Tue, 29 Sep 2015 08:26:51 +0000</updated>
                            <resolved>Tue, 29 Sep 2015 08:26:51 +0000</resolved>
                                    <version>Lustre 1.8.9</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="52494" author="pjones" created="Fri, 15 Feb 2013 17:43:59 +0000"  >&lt;p&gt;Niu&lt;/p&gt;

&lt;p&gt;Could you please look into this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="52495" author="green" created="Fri, 15 Feb 2013 17:49:01 +0000"  >&lt;p&gt;So, pdflush blocking is expected - the logs show disconnction from OST, so no way to flush those dirty pages until reconnect happens, and I assuem t&apos;s not happening? Any further messages to clue us why?&lt;/p&gt;

&lt;p&gt;Teh soft lockup stack is strange, I don&apos;t believe we actually use events API in lustre at all so I am not sure why would it lock up there.&lt;/p&gt;</comment>
                            <comment id="52523" author="yujian" created="Sat, 16 Feb 2013 00:24:21 +0000"  >&lt;p&gt;Lustre Tag: v1_8_9_WC1_RC1&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b1_8/256&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b1_8/256&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL5.9/x86_64(server), RHEL6.3/x86_64(client)&lt;/p&gt;

&lt;p&gt;The large-scale test 3a (failing over MDS while running mdsrate on clients) also hung over IB network (&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2823&quot; title=&quot;large-scale test_3a hung: LBUG: ASSERTION(ergo(!obd-&amp;gt;obd_recovering, diff &amp;gt;= 0)) failed: lustre-OST0001: 1013476 - 1024134 = -10658&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2823&quot;&gt;&lt;del&gt;LU-2823&lt;/del&gt;&lt;/a&gt;). The logs also showed that connection to OST was lost.&lt;/p&gt;

&lt;p&gt;The same test run by autotest passed over TCP network: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/ad2ba008-7683-11e2-bc2f-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/ad2ba008-7683-11e2-bc2f-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The recovery-mds-scale test run by autotest also passed over TCP network with hard failing over MDS 95 times in 24 hours: &lt;a href=&quot;https://maloo.whamcloud.com/sub_tests/9caf800a-77dc-11e2-abae-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/sub_tests/9caf800a-77dc-11e2-abae-52540035b04c&lt;/a&gt; &lt;/p&gt;</comment>
                            <comment id="52551" author="niu" created="Sat, 16 Feb 2013 23:26:13 +0000"  >&lt;p&gt;Cliff, is there any log from OSS and MDS? Thanks.&lt;/p&gt;</comment>
                            <comment id="52595" author="yujian" created="Mon, 18 Feb 2013 02:46:06 +0000"  >&lt;p&gt;Lustre Tag: v1_8_9_WC1_RC1&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b1_8/256&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b1_8/256&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL5.9/x86_64&lt;br/&gt;
Network: IB (in-kernel OFED)&lt;br/&gt;
ENABLE_QUOTA=yes&lt;br/&gt;
FAILURE_MODE=HARD&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;MGS/MDS Nodes: fat-amd-1-ib(active), fat-amd-2-ib(passive)
                                  \  /
                            1 combined MGS/MDT

OSS Nodes: fat-amd-3-ib(active), fat-amd-4-ib(active)
                              \  /
                              6 OSTs

Client Nodes: fat-intel-2-ib,client-[1,3,6,17]-ib
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I started a manual test run of recovery-mds-scale over IB network with the above configuration on Toro cluster (the number of client is far less than that on Hyperion cluster). Finally, the MDS failed over 24 times in 6 hours without issue, and the OST failed over 1 time with hitting &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-463&quot; title=&quot;orphan recovery happens too late, causing writes to fail with ENOENT after recovery&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-463&quot;&gt;&lt;del&gt;LU-463&lt;/del&gt;&lt;/a&gt;:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/d7e82752-79db-11e2-8fd2-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/d7e82752-79db-11e2-8fd2-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="52623" author="pjones" created="Mon, 18 Feb 2013 09:06:25 +0000"  >&lt;p&gt;Could it be worth testing on the OpenSFS test cluster? That should give us more clients to play with (though still not quite as many as Hyperion)&lt;/p&gt;</comment>
                            <comment id="52631" author="cliffw" created="Mon, 18 Feb 2013 11:07:13 +0000"  >&lt;p&gt;MDS log:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Feb 15 09:56:52 hyperion-mds3 Lustre: MDT lustre-MDT0000 has stopped.
Feb 15 09:56:52 hyperion-mds3 LustreError: 137-5: UUID &lt;span class=&quot;code-quote&quot;&gt;&apos;lustre-MDT0000_UUID&apos;&lt;/span&gt; is not available &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; connect (stopping)
Feb 15 09:56:52 hyperion-mds3 LustreError: Skipped 2 previous similar messages
Feb 15 09:56:52 hyperion-mds3 LustreError: 13540:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (-
19)  req@ffff81023f2b2c50 x1426985035882333/t0 o38-&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 368/0 e 0 to 0 dl 1360951112 ref 1 fl Interpret
:/0/0 rc -19/0
Feb 15 09:56:52 hyperion-mds3 LustreError: 13540:0:(ldlm_lib.c:1921:target_send_reply_msg()) Skipped 1 previous simi
lar message
Feb 15 09:56:52 hyperion-mds3 LustreError: 13549:0:(handler.c:1515:mds_handle()) operation 400 on unconnected MDS fr
om 12345-192.168.116.64@o2ib
Feb 15 09:56:53 hyperion-mds3 LustreError: 137-5: UUID &lt;span class=&quot;code-quote&quot;&gt;&apos;lustre-MDT0000_UUID&apos;&lt;/span&gt; is not available &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; connect (no target
)
Feb 15 09:56:53 hyperion-mds3 LustreError: Skipped 2 previous similar messages
Feb 15 09:56:53 hyperion-mds3 Lustre: server umount lustre-MDT0000 complete
Feb 15 09:56:53 hyperion-mds3 mrshd[14785]: root@ehyperion582 as root: cmd=&apos;PATH=/admin/scripts:/admin/bin:/bin:/usr
/bin:/sbin:/usr/sbin;(PATH=$PATH:/usr/lib64/lustre/utils:/usr/lib64/lustre/tests:/sbin:/usr/sbin; cd /usr/lib64/lust
re/tests; sh -c &lt;span class=&quot;code-quote&quot;&gt;&quot;lsmod | grep lnet &amp;gt; /dev/&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt; &amp;amp;&amp;amp; lctl dl | grep &lt;span class=&quot;code-quote&quot;&gt;&apos; ST &apos;&lt;/span&gt;&quot;&lt;/span&gt;);echo XXRETCODE:$?&apos;
Feb 15 09:57:03 hyperion-mds3 mrshd[14795]: root@ehyperion582 as root: cmd=&apos;PATH=/admin/scripts:/admin/bin:/bin:/usr
/bin:/sbin:/usr/sbin;(PATH=$PATH:/usr/lib64/lustre/utils:/usr/lib64/lustre/tests:/sbin:/usr/sbin; cd /usr/lib64/lust
re/tests; sh -c &lt;span class=&quot;code-quote&quot;&gt;&quot;hostname&quot;&lt;/span&gt;);echo XXRETCODE:$?&apos;
Feb 15 09:57:03 hyperion-mds3 mrshd[14801]: root@ehyperion582 as root: cmd=&apos;PATH=/admin/scripts:/admin/bin:/bin:/usr
/bin:/sbin:/usr/sbin;(PATH=$PATH:/usr/lib64/lustre/utils:/usr/lib64/lustre/tests:/sbin:/usr/sbin; cd /usr/lib64/lust
re/tests; sh -c &lt;span class=&quot;code-quote&quot;&gt;&quot;mkdir -p /p/mds; mount -t lustre -o user_xattr  /dev/md1 /p/mds&quot;&lt;/span&gt;);echo XXRETCODE:$?&apos;
Feb 15 09:57:04 hyperion-mds3 LDISKFS-fs (md1): mounted filesystem with ordered data mode
Feb 15 09:57:04 hyperion-mds3 JBD: barrier-based sync failed on md1-8 - disabling barriers
Feb 15 09:57:05 hyperion-mds3 LDISKFS-fs (md1): warning: maximal mount count reached, running e2fsck is recommended
Feb 15 09:57:06 hyperion-mds3 LDISKFS-fs (md1): mounted filesystem with ordered data mode
Feb 15 09:57:06 hyperion-mds3 Lustre: Enabling user_xattr
Feb 15 09:57:06 hyperion-mds3 Lustre: Skipped 1 previous similar message
Feb 15 09:57:06 hyperion-mds3 Lustre: Enabling user_xattr
Feb 15 09:57:06 hyperion-mds3 Lustre: 14918:0:(mds_fs.c:677:mds_init_server_data()) RECOVERY: service lustre-MDT0000
, 70 recoverable clients, 0 delayed clients, last_transno 38654758668
Feb 15 09:57:06 hyperion-mds3 JBD: barrier-based sync failed on md1-8 - disabling barriers
Feb 15 09:57:06 hyperion-mds3 Lustre: lustre-MDT0000: Now serving lustre-MDT0000 on /dev/md1 with recovery enabled
Feb 15 09:57:06 hyperion-mds3 Lustre: lustre-MDT0000: Will be in recovery &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; at least 5:00, or until 70 clients rec
onnect
Feb 15 09:57:06 hyperion-mds3 Lustre: 14918:0:(lproc_mds.c:271:lprocfs_wr_group_upcall()) lustre-MDT0000: group upca
ll set to /usr/sbin/l_getgroups
Feb 15 09:57:06 hyperion-mds3 Lustre: 14918:0:(mds_lov.c:1155:mds_notify()) MDS lustre-MDT0000: add target lustre-OS
T0000_UUID
Feb 15 09:57:06 hyperion-mds3 Lustre: 14918:0:(mds_lov.c:1155:mds_notify()) Skipped 4 previous similar messages
Feb 15 09:57:06 hyperion-mds3 Lustre: 3083:0:(mds_lov.c:1191:mds_notify()) MDS lustre-MDT0000: in recovery, not rese
tting orphans on lustre-OST0000_UUID
Feb 15 09:57:06 hyperion-mds3 Lustre: 3083:0:(mds_lov.c:1191:mds_notify()) Skipped 1 previous similar message
Feb 15 09:57:06 hyperion-mds3 Lustre: 14918:0:(mds_lov.c:1155:mds_notify()) MDS lustre-MDT0000: add target lustre-OS
T0004_UUID
Feb 15 09:57:06 hyperion-mds3 Lustre: 14918:0:(mds_lov.c:1155:mds_notify()) Skipped 3 previous similar messages
Feb 15 09:57:06 hyperion-mds3 Lustre: 3083:0:(mds_lov.c:1191:mds_notify()) MDS lustre-MDT0000: in recovery, not rese
tting orphans on lustre-OST0005_UUID
Feb 15 09:57:06 hyperion-mds3 Lustre: 3083:0:(mds_lov.c:1191:mds_notify()) Skipped 4 previous similar messages
Feb 15 09:57:06 hyperion-mds3 mrshd[14929]: root@ehyperion582 as root: cmd=&apos;PATH=/admin/scripts:/admin/bin:/bin:/usr
/bin:/sbin:/usr/sbin;(PATH=$PATH:/usr/lib64/lustre/utils:/usr/lib64/lustre/tests:/sbin:/usr/sbin; cd /usr/lib64/lust
re/tests; sh -c &quot;PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/ioz
one/bin:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/tests:/opt/mdtest-1.7.4/bin:/opt/mib-1.9.10/bin:/opt/ior-2.1
0.2/bin:/opt/simul-1.15/bin:/opt/mvapich-gnu-gen2-1.2/bin:/admin/bin:/admin/sbin:/admin/scripts:/usr/lib64/qt-3.3/bi
n:/bin:/usr/bin:/usr/bin/X11:/sbin:/usr/sbin:/usr/kerberos/bin:/usr/kerberos/sbin:/usr/lib64/lustre/utils:/usr/lib64
/lustre/../lustre-iokit/sgpdd-survey::/usr/lib64/lustre/tests/mpi NAME=ncli sh rpc.sh set_default_debug \&quot;0x33f1504\
&lt;span class=&quot;code-quote&quot;&gt;&quot; \&quot;&lt;/span&gt; 0xffb7e3ff\&lt;span class=&quot;code-quote&quot;&gt;&quot; 16 &quot;&lt;/span&gt;);echo XXRETCODE:$?&apos;
Feb 15 09:57:06 hyperion-mds3 mrshd[15014]: root@ehyperion582 as root: cmd=&apos;PATH=/admin/scripts:/admin/bin:/bin:/usr
/bin:/sbin:/usr/sbin;(PATH=$PATH:/usr/lib64/lustre/utils:/usr/lib64/lustre/tests:/sbin:/usr/sbin; cd /usr/lib64/lust
re/tests; sh -c &lt;span class=&quot;code-quote&quot;&gt;&quot;e2label /dev/md1&quot;&lt;/span&gt;);echo XXRETCODE:$?&apos;
Feb 15 09:57:07 hyperion-mds3 Lustre: 14854:0:(ldlm_lib.c:1817:target_queue_last_replay_reply()) lustre-MDT0000: 69 
recoverable clients remain
Feb 15 09:57:07 hyperion-mds3 Lustre: 14854:0:(ldlm_lib.c:1817:target_queue_last_replay_reply()) Skipped 2 previous 
similar messages
Feb 15 09:57:07 hyperion-mds3 mrshd[15022]: root@ehyperion582 as root: cmd=&apos;PATH=/admin/scripts:/admin/bin:/bin:/usr
/bin:/sbin:/usr/sbin;(PATH=$PATH:/usr/lib64/lustre/utils:/usr/lib64/lustre/tests:/sbin:/usr/sbin; cd /usr/lib64/lust
re/tests; sh -c &lt;span class=&quot;code-quote&quot;&gt;&quot;/usr/sbin/lctl mark ==== Checking the clients loads AFTER failover -- failure NOT OK&quot;&lt;/span&gt;);echo XXRETCO
DE:$?&apos;
Feb 15 09:57:07 hyperion-mds3 Lustre: DEBUG MARKER: ==== Checking the clients loads AFTER failover -- failure NOT OK
Feb 15 09:57:07 hyperion-mds3 LustreError: 14871:0:(handler.c:1515:mds_handle()) operation 400 on unconnected MDS fr
om 12345-192.168.116.10@o2ib
Feb 15 09:57:07 hyperion-mds3 LustreError: 14871:0:(handler.c:1515:mds_handle()) Skipped 14 previous similar message
s
Feb 15 09:57:07 hyperion-mds3 LustreError: 14871:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (-
107)  req@ffff810f919ca000 x1426985034582787/t0 o400-&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 192/0 e 0 to 0 dl 1360951043 ref 1 fl Interpr
et:H/0/0 rc -107/0
Feb 15 09:57:07 hyperion-mds3 LustreError: 14871:0:(ldlm_lib.c:1921:target_send_reply_msg()) Skipped 21 previous sim
ilar messages
Feb 15 09:57:11 hyperion-mds3 Lustre: 14866:0:(ldlm_lib.c:1817:target_queue_last_replay_reply()) lustre-MDT0000: 25 
recoverable clients remain
Feb 15 09:57:11 hyperion-mds3 Lustre: 14866:0:(ldlm_lib.c:1817:target_queue_last_replay_reply()) Skipped 48 previous
 similar messages
Feb 15 09:57:16 hyperion-mds3 LustreError: 14865:0:(handler.c:1515:mds_handle()) operation 400 on unconnected MDS fr
om 12345-192.168.116.78@o2ib
Feb 15 09:57:16 hyperion-mds3 LustreError: 14865:0:(handler.c:1515:mds_handle()) Skipped 9 previous similar messages
Feb 15 09:57:16 hyperion-mds3 LustreError: 14865:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (-
107)  req@ffff810837941400 x1426985038316154/t0 o400-&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 192/0 e 0 to 0 dl 1360951052 ref 1 fl Interpr
et:H/0/0 rc -107/0
Feb 15 09:57:16 hyperion-mds3 LustreError: 14865:0:(ldlm_lib.c:1921:target_send_reply_msg()) Skipped 9 previous simi
lar messages
Feb 15 09:57:19 hyperion-mds3 Lustre: 14850:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:19 hyperion-mds3 Lustre: 14865:0:(ldlm_lib.c:1817:target_queue_last_replay_reply()) lustre-MDT0000: 5 r
ecoverable clients remain

Feb 15 09:57:19 hyperion-mds3 Lustre: 14865:0:(ldlm_lib.c:1817:target_queue_last_replay_reply()) Skipped 9 previous 
similar messages
Feb 15 09:57:21 hyperion-mds3 Lustre: 14872:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: e1802d77-7b
80-d67a-2994-6269f294be6d reconnecting
Feb 15 09:57:27 hyperion-mds3 Lustre: 14877:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: 2d202d86-7c
6c-ef27-da63-3094a685e606 reconnecting
Feb 15 09:57:35 hyperion-mds3 Lustre: 14860:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: 2d202d86-7c
6c-ef27-da63-3094a685e606 reconnecting
Feb 15 09:57:35 hyperion-mds3 Lustre: 14860:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 2 previous similar 
messages
Feb 15 09:57:44 hyperion-mds3 Lustre: 14852:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: 2d202d86-7c
6c-ef27-da63-3094a685e606 reconnecting
Feb 15 09:57:44 hyperion-mds3 Lustre: 14852:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 4 previous similar 
messages
Feb 15 10:00:01 hyperion-mds3 mrshd[15031]: root@ehyperion0 as root: cmd=&lt;span class=&quot;code-quote&quot;&gt;&apos;rdistd -S&apos;&lt;/span&gt;
Feb 15 10:01:33 hyperion-mds3 Lustre: MGS: haven&apos;t heard from client 2d23bcad-fae7-5d3d-38a1-4db19f074eb8 (at 192.16
8.116.30@o2ib) in 228 seconds. I think it&apos;s dead, and I am evicting it.
Feb 15 10:01:33 hyperion-mds3 Lustre: MGS: haven&apos;t heard from client 1fdb7e20-6f21-c410-56a3-6c818466b316 (at 192.16
8.116.32@o2ib) in 226 seconds. I think it&apos;s dead, and I am evicting it.
Feb 15 10:03:40 hyperion-mds3 Lustre: 14857:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: 31a45b78-d0
37-7471-33be-0b77d9441939 reconnecting
Feb 15 10:03:40 hyperion-mds3 Lustre: 14857:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 3 previous similar 
messages
Feb 15 10:03:40 hyperion-mds3 Lustre: 14857:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
Feb 15 10:03:40 hyperion-mds3 LustreError: 14857:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (-
16)  req@ffff810c46390000 x1426985036859689/t0 o38-&amp;gt;31a45b78-d037-7471-33be-0b77d9441939@NET_0x50000c0a8741b_UUID:0/
0 lens 368/264 e 0 to 0 dl 1360951520 ref 1 fl Interpret:/0/0 rc -16/0
Feb 15 10:03:40 hyperion-mds3 LustreError: 14857:0:(ldlm_lib.c:1921:target_send_reply_msg()) Skipped 7 previous simi
lar messages
Feb 15 10:03:48 hyperion-mds3 Lustre: 14854:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
Feb 15 10:03:56 hyperion-mds3 Lustre: 14870:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
Feb 15 10:04:04 hyperion-mds3 Lustre: 14865:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
Feb 15 10:04:12 hyperion-mds3 Lustre: 14872:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-MDT0000: 31a45b78-d0
37-7471-33be-0b77d9441939 reconnecting
Feb 15 10:04:12 hyperion-mds3 Lustre: 14872:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 3 previous similar 
messages
Feb 15 10:04:12 hyperion-mds3 Lustre: 14872:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
Feb 15 10:04:12 hyperion-mds3 LustreError: 14872:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (-
16)  req@ffff8105ddcde400 x1426985036859702/t0 o38-&amp;gt;31a45b78-d037-7471-33be-0b77d9441939@NET_0x50000c0a8741b_UUID:0/
0 lens 368/264 e 0 to 0 dl 1360951552 ref 1 fl Interpret:/0/0 rc -16/0
Feb 15 10:04:12 hyperion-mds3 LustreError: 14872:0:(ldlm_lib.c:1921:target_send_reply_msg()) Skipped 3 previous simi
lar messages
Feb 15 10:04:20 hyperion-mds3 Lustre: 14856:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
Feb 15 10:04:36 hyperion-mds3 Lustre: 14866:0:(ldlm_lib.c:876:target_handle_connect()) lustre-MDT0000: refuse reconn
ection from 31a45b78-d037-7471-33be-0b77d9441939@192.168.116.27@o2ib to 0xffff8103bf910e00; still busy with 1 active
 RPCs
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Let me know if you need more&lt;/p&gt;</comment>
                            <comment id="52632" author="cliffw" created="Mon, 18 Feb 2013 11:09:23 +0000"  >&lt;p&gt;Typical OST:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Feb 15 09:57:07 ehyperion1158 Lustre: DEBUG MARKER: ==== Checking the clients loads AFTER failover -- failure NOT OK
Feb 15 09:57:40 ehyperion1158 Lustre: 30886:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0000: e1802d77-7b
80-d67a-2994-6269f294be6d reconnecting
Feb 15 09:57:41 ehyperion1158 Lustre: 30840:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0007: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:41 ehyperion1158 Lustre: 30840:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 1 previous similar 
message
Feb 15 09:57:43 ehyperion1158 Lustre: 30852:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0000: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:45 ehyperion1158 Lustre: 30935:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0000: 2d202d86-7c
6c-ef27-da63-3094a685e606 reconnecting
Feb 15 09:57:45 ehyperion1158 Lustre: 30935:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 1 previous similar 
message
Feb 15 09:57:48 ehyperion1158 Lustre: 30932:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0007: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:48 ehyperion1158 Lustre: 30932:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 2 previous similar 
messages
Feb 15 09:57:52 ehyperion1158 Lustre: 30887:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0000: 2d202d86-7c
6c-ef27-da63-3094a685e606 reconnecting
Feb 15 09:57:52 ehyperion1158 Lustre: 30887:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 2 previous similar 
messages
Feb 15 10:00:01 ehyperion1158 mrshd[7679]: root@ehyperion0 as root: cmd=&lt;span class=&quot;code-quote&quot;&gt;&apos;rdistd -S&apos;&lt;/span&gt;
Feb 15 10:01:42 ehyperion1158 Lustre: lustre-OST0000: haven&apos;t heard from client e1802d77-7b80-d67a-2994-6269f294be6d
 (at 192.168.116.32@o2ib) in 227 seconds. I think it&apos;s dead, and I am evicting it.
Feb 15 10:01:42 ehyperion1158 Lustre: Skipped 1 previous similar message
Feb 15 10:01:42 ehyperion1158 Lustre: lustre-OST0000: haven&apos;t heard from client d18624c2-39f3-9e9c-8a23-42b31c59c704
 (at 192.168.116.28@o2ib) in 222 seconds. I think it&apos;s dead, and I am evicting it.
Feb 15 10:07:07 ehyperion1158 Lustre: lustre-OST0000: received MDS connection from 192.168.120.126@o2ib
Feb 15 10:07:07 ehyperion1158 Lustre: 30878:0:(filter.c:3129:filter_destroy_precreated()) lustre-OST0000: deleting o
rphan objects from 683938 to 683981, orphan objids won&apos;t be reused any more.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="52633" author="cliffw" created="Mon, 18 Feb 2013 11:10:14 +0000"  >&lt;p&gt;One OST did report an issue during this time.&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Feb 15 09:57:07 ehyperion1157 Lustre: DEBUG MARKER: ==== Checking the clients loads AFTER failover -- failure NOT OK
Feb 15 09:57:41 ehyperion1157 Lustre: 31000:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0003: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:41 ehyperion1157 Lustre: 31000:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 2 previous similar 
messages
Feb 15 09:57:48 ehyperion1157 Lustre: 30938:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0003: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:48 ehyperion1157 Lustre: 30938:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 1 previous similar 
message
Feb 15 09:57:56 ehyperion1157 Lustre: 31047:0:(ldlm_lib.c:576:target_handle_reconnect()) lustre-OST0003: d18624c2-39
f3-9e9c-8a23-42b31c59c704 reconnecting
Feb 15 09:57:56 ehyperion1157 Lustre: 31047:0:(ldlm_lib.c:576:target_handle_reconnect()) Skipped 1 previous similar 
message
Feb 15 09:58:07 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 4, status -103, 
desc ffff8104da95e000
Feb 15 09:58:07 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 2, status -103, 
desc ffff8104da95e000
Feb 15 09:58:07 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 4, status -103, 
desc ffff8102d9e48000
Feb 15 09:58:07 ehyperion1157 LustreError: 32307:0:(ost_handler.c:1078:ost_brw_write()) @@@ network error on bulk GE
T 0(1048576)  req@ffff81031378f850 x1426985036338408/t0 o4-&amp;gt;2d202d86-7c6c-ef27-da63-3094a685e606@NET_0x50000c0a8741e
_UUID:0/0 lens 448/416 e 1 to 0 dl 1360951137 ref 1 fl Interpret:/0/0 rc 0/0
Feb 15 09:58:07 ehyperion1157 Lustre: 32307:0:(ost_handler.c:1229:ost_brw_write()) lustre-OST0003: ignoring bulk IO 
comm error with 2d202d86-7c6c-ef27-da63-3094a685e606@NET_0x50000c0a8741e_UUID id 12345-192.168.116.30@o2ib - client 
will retry
Feb 15 09:58:08 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 2, status -103, 
desc ffff8102d9e48000
Feb 15 09:58:08 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 4, status -103, 
desc ffff8102e61e6000
Feb 15 09:58:08 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 2, status -103, 
desc ffff8102e61e6000
Feb 15 09:58:08 ehyperion1157 LustreError: 32276:0:(ost_handler.c:1078:ost_brw_write()) @@@ network error on bulk GE
T 0(1048576)  req@ffff810625080000 x1426985036338409/t0 o4-&amp;gt;2d202d86-7c6c-ef27-da63-3094a685e606@NET_0x50000c0a8741e
_UUID:0/0 lens 448/416 e 1 to 0 dl 1360951137 ref 1 fl Interpret:/0/0 rc 0/0
Feb 15 09:58:08 ehyperion1157 LustreError: 30804:0:(events.c:387:server_bulk_callback()) event type 4, status -103, 
desc ffff810544376000
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="52653" author="johann" created="Mon, 18 Feb 2013 17:22:28 +0000"  >&lt;p&gt;What about Client&apos;s sysrq-t (or even better crash dump)?&lt;/p&gt;</comment>
                            <comment id="52692" author="cliffw" created="Tue, 19 Feb 2013 12:19:29 +0000"  >&lt;p&gt;I will attempt to reproduce today and get sysrq. &lt;/p&gt;</comment>
                            <comment id="128721" author="adilger" created="Tue, 29 Sep 2015 08:26:51 +0000"  >&lt;p&gt;Closing old 1.8 bug.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvj7z:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>6828</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>