<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:50:42 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-5347] replay-ost-single test_3: tee: standard output: Input/output error</title>
                <link>https://jira.whamcloud.com/browse/LU-5347</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;replay-ost-single test 3 failed as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== replay-ost-single test 3: Fail OST during write, with verification == 20:09:46 (1400123386)
Failing ost1 on shadow-49vm8
CMD: shadow-49vm8 grep -c /mnt/ost1&apos; &apos; /proc/mounts
Stopping /mnt/ost1 (opts:) on shadow-49vm8
CMD: shadow-49vm8 umount -d /mnt/ost1
CMD: shadow-49vm8 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
reboot facets: ost1
Failover ost1 to shadow-49vm8
20:10:04 (1400123404) waiting for shadow-49vm8 network 900 secs ...
20:10:04 (1400123404) network interface is UP
CMD: shadow-49vm8 hostname
mount facets: ost1
CMD: shadow-49vm8 test -b /dev/lvm-Role_OSS/P1
Starting ost1:   /dev/lvm-Role_OSS/P1 /mnt/ost1
CMD: shadow-49vm8 mkdir -p /mnt/ost1; mount -t lustre   		                   /dev/lvm-Role_OSS/P1 /mnt/ost1
CMD: shadow-49vm8 PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/bin:/usr/bin:/bin:/sbin:/usr/sbin::/sbin:/bin:/usr/sbin: NAME=autotest_config sh rpc.sh set_default_debug \&quot;vfstrace rpctrace dlmtrace neterror ha config ioctl super\&quot; \&quot;all -lnet -lnd -pinger\&quot; 4 
CMD: shadow-49vm8 e2label /dev/lvm-Role_OSS/P1 2&amp;gt;/dev/null
Started lustre-OST0000
CMD: shadow-49vm1,shadow-49vm2.shadow.whamcloud.com PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/bin:/usr/bin:/bin:/sbin:/usr/sbin::/sbin:/bin:/usr/sbin: NAME=autotest_config sh rpc.sh wait_import_state_mount FULL osc.lustre-OST0000-osc-*.ost_server_uuid 
shadow-49vm1: CMD: shadow-49vm1.shadow.whamcloud.com lctl get_param -n at_max
shadow-49vm2: CMD: shadow-49vm2.shadow.whamcloud.com lctl get_param -n at_max
shadow-49vm1: osc.lustre-OST0000-osc-*.ost_server_uuid in FULL state after 95 sec
tee: standard output: Input/output error
shadow-49vm2: osc.lustre-OST0000-osc-*.ost_server_uuid in FULL state after 105 sec
1280+0 records in
1280+0 records out
5242880 bytes (5.2 MB) copied, 125.153 s, 41.9 kB/s
tee: write error
 replay-ost-single test_3: @@@@@@ FAIL: test_3 failed with 1 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Debug log on client:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000080:00200000:0.0:1400123511.721003:0:28098:0:(file.c:1229:ll_file_io_generic()) iot: 1, result: -5
00000100:00100000:1.0:1400123511.726097:0:20485:0:(client.c:1851:ptlrpc_check_set()) Completed RPC pname:cluuid:pid:xid:nid:opc ptlrpcd_0:c890eb56-964e-2a0e-1f85-b5461794e3dc:0:1468135597178740:10.1.6.62@tcp:4
00000008:00080000:1.0:1400123511.726121:0:20486:0:(osc_request.c:1769:brw_interpret()) lustre-OST0000-osc-ffff88006b95d000: resend cross eviction for object: 0x0:15338, rc = -5.
00000100:00100000:1.0:1400123511.726335:0:20486:0:(client.c:1851:ptlrpc_check_set()) Completed RPC pname:cluuid:pid:xid:nid:opc ptlrpcd_1:c890eb56-964e-2a0e-1f85-b5461794e3dc:0:1468135597149296:10.1.6.62@tcp:4
00000008:00080000:1.0:1400123511.726341:0:20486:0:(osc_request.c:1769:brw_interpret()) lustre-OST0000-osc-ffff88006b95d000: resend cross eviction for object: 0x0:15338, rc = -5.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;http://testing.hpdd.intel.com/test_sets/1786067c-dc5a-11e3-8b61-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://testing.hpdd.intel.com/test_sets/1786067c-dc5a-11e3-8b61-52540035b04c&lt;/a&gt;&lt;/p&gt;</description>
                <environment>Lustre Build: &lt;a href=&quot;https://build.hpdd.intel.com/job/lustre-master/2048/&quot;&gt;https://build.hpdd.intel.com/job/lustre-master/2048/&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL6.5/x86_64</environment>
        <key id="25579">LU-5347</key>
            <summary>replay-ost-single test_3: tee: standard output: Input/output error</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="yujian">Jian Yu</assignee>
                                    <reporter username="yujian">Jian Yu</reporter>
                        <labels>
                    </labels>
                <created>Tue, 15 Jul 2014 00:12:17 +0000</created>
                <updated>Mon, 27 Feb 2017 23:36:03 +0000</updated>
                            <resolved>Mon, 27 Feb 2017 23:36:03 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                    <version>Lustre 2.5.2</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>10</watches>
                                                                            <comments>
                            <comment id="88996" author="yujian" created="Tue, 15 Jul 2014 00:27:37 +0000"  >&lt;p&gt;More instances on Lustre master and b2_5 branches:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/53a5c16c-07ff-11e4-8e5a-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/53a5c16c-07ff-11e4-8e5a-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/1576f466-edaf-11e3-a130-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/1576f466-edaf-11e3-a130-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/128130c2-f86e-11e3-8846-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/128130c2-f86e-11e3-8846-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The failures were vetted as &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2285&quot; title=&quot;Test failure on replay-ost-single test_3: write page inode failed -2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2285&quot;&gt;&lt;del&gt;LU-2285&lt;/del&gt;&lt;/a&gt;. However, it looks like they are not the same issue.&lt;/p&gt;</comment>
                            <comment id="89070" author="adilger" created="Tue, 15 Jul 2014 17:44:35 +0000"  >&lt;p&gt;I see in &lt;a href=&quot;https://testing.hpdd.intel.com/test_logs/a1c7f374-07ff-11e4-8e5a-5254006e85c2/show_text&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_logs/a1c7f374-07ff-11e4-8e5a-5254006e85c2/show_text&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;19:34:11:LustreError: 2177:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x2cb:0x0: rc = -2
19:34:11:LustreError: 2177:0:(ldlm_resource.c:1150:ldlm_resource_get()) Skipped 583 previous similar messages
19:34:11:LustreError: 2177:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x2cb:0x0: rc = -2
19:34:11:LustreError: 2177:0:(ldlm_resource.c:1150:ldlm_resource_get()) Skipped 303 previous similar messages
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Di, I think this bug was already fixed on master?  It had to do with incorrect decoding of osd_id or similar.  I suspect that patch needs to be backported to b2_5.&lt;/p&gt;</comment>
                            <comment id="91686" author="di.wang" created="Thu, 14 Aug 2014 23:55:39 +0000"  >&lt;p&gt;Andreas: Not sure if you mean this patch &lt;a href=&quot;http://review.whamcloud.com/#/c/5820/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/5820/&lt;/a&gt;, but it is already landed to 2.4.  So you may mean sth else? Besides this problem also exists on b2_6, probably needs further investigation. &lt;/p&gt;</comment>
                            <comment id="91687" author="yujian" created="Fri, 15 Aug 2014 00:00:46 +0000"  >&lt;p&gt;Lustre Build: &lt;a href=&quot;https://build.hpdd.intel.com/job/lustre-b2_6/2/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://build.hpdd.intel.com/job/lustre-b2_6/2/&lt;/a&gt; (2.6.0 RC2)&lt;br/&gt;
Distro/Arch: SLES11SP3/x86_64 (client), RHEL6.5/x86_64 (server)&lt;/p&gt;

&lt;p&gt;The same failure occurred:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/48921180-22b8-11e4-94dd-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/48921180-22b8-11e4-94dd-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/11666c4e-2063-11e4-9d86-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/11666c4e-2063-11e4-9d86-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/58d0aa7e-1422-11e4-88ed-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/58d0aa7e-1422-11e4-88ed-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="91797" author="yujian" created="Sat, 16 Aug 2014 01:14:13 +0000"  >&lt;p&gt;For &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/48921180-22b8-11e4-94dd-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/48921180-22b8-11e4-94dd-5254006e85c2&lt;/a&gt;, after OST was failed over, console log showed that:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00:38:44:Lustre: DEBUG MARKER: e2label /dev/lvm-Role_OSS/P1 2&amp;gt;/dev/null
00:38:44:LustreError: 17775:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x3aaf:0x0: rc = -2
00:40:19:Lustre: lustre-OST0000: Client ad3fb862-29c2-d5fd-4885-211ab6e3d1bd (at 10.2.4.94@tcp) reconnecting, waiting for 3 clients in recovery for 1:22
00:40:19:Lustre: Skipped 2 previous similar messages
00:40:20:Lustre: ost: This server is not able to keep up with request traffic (cpu-bound).
00:40:21:Lustre: Skipped 2 previous similar messages
00:40:21:Lustre: 13466:0:(service.c:1511:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=21, delay=0(jiff)
00:40:21:Lustre: 13466:0:(service.c:1511:ptlrpc_at_check_timed()) Skipped 2 previous similar messages
00:40:21:Lustre: 13466:0:(service.c:1308:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)?  req@ffff8800410a4800 x1476204951412392/t0(0) o400-&amp;gt;lustre-MDT0000-mdtlov_UUID@10.2.4.99@tcp:0/0 lens 224/0 e 283854 to 0 dl 1407829125 ref 2 fl Complete:H/c0/ffffffff rc 0/-1
00:40:22:Lustre: 13466:0:(service.c:1308:ptlrpc_at_send_early_reply()) Skipped 2 previous similar messages
00:40:24:Lustre: lustre-OST0000: Client lustre-MDT0000-mdtlov_UUID (at 10.2.4.99@tcp) reconnecting, waiting for 3 clients in recovery for 0:47
00:40:24:Lustre: Skipped 9 previous similar messages
00:40:24:LustreError: 17775:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x3ab0:0x0: rc = -2
00:40:24:LustreError: 17775:0:(ldlm_resource.c:1150:ldlm_resource_get()) Skipped 46 previous similar messages
00:40:25:LustreError: 17775:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x3ab0:0x0: rc = -2
00:40:25:LustreError: 17775:0:(ldlm_resource.c:1150:ldlm_resource_get()) Skipped 239 previous similar messages
00:40:25:LustreError: 17775:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x3aaf:0x0: rc = -2
00:40:25:Lustre: lustre-OST0001: Client lustre-MDT0000-mdtlov_UUID (at 10.2.4.99@tcp) reconnecting
00:40:26:Lustre: Skipped 19 previous similar messages
00:40:26:Lustre: 13028:0:(client.c:1926:ptlrpc_expire_one_request()) @@@ Request sent has timed out for sent delay: [sent 1407829163/real 0]  req@ffff880019d12800 x1476204961866504/t0(0) o400-&amp;gt;lustre-MDT0000-lwp-OST0002@10.2.4.99@tcp:12/10 lens 224/224 e 0 to 1 dl 1407829172 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
00:40:26:Lustre: 13028:0:(client.c:1926:ptlrpc_expire_one_request()) Skipped 65 previous similar messages
00:40:27:Lustre: lustre-MDT0000-lwp-OST0002: Connection to lustre-MDT0000 (at 10.2.4.99@tcp) was lost; in progress operations using this service will wait for recovery to complete
00:40:27:Lustre: Skipped 48 previous similar messages
00:40:27:LustreError: 166-1: MGC10.2.4.99@tcp: Connection to MGS (at 10.2.4.99@tcp) was lost; in progress operations using this service will fail
00:40:27:LustreError: Skipped 6 previous similar messages
00:40:28:Lustre: lustre-OST0001: deleting orphan objects from 0x0:14116 to 0x0:14209
00:40:28:Lustre: lustre-OST0003: deleting orphan objects from 0x0:14148 to 0x0:14273
00:40:29:Lustre: lustre-OST0004: deleting orphan objects from 0x0:14148 to 0x0:14273
00:40:29:Lustre: lustre-OST0002: deleting orphan objects from 0x0:14212 to 0x0:14305
00:40:29:Lustre: lustre-OST0005: deleting orphan objects from 0x0:14180 to 0x0:14305
00:40:30:Lustre: lustre-OST0006: deleting orphan objects from 0x0:14179 to 0x0:14305
00:40:30:Lustre: lustre-OST0000: Client lustre-MDT0000-mdtlov_UUID (at 10.2.4.99@tcp) reconnecting, waiting for 3 clients in recovery for 0:04
00:40:30:Lustre: Skipped 60 previous similar messages
00:40:31:Lustre: lustre-OST0000: recovery is timed out, evict stale exports
00:40:31:Lustre: lustre-OST0000: disconnecting 1 stale clients
00:40:31:LustreError: 17775:0:(ofd_grant.c:183:ofd_grant_sanity_check()) ofd_obd_disconnect: tot_granted 2097152 != fo_tot_granted 2523136
00:40:32:Lustre: 17775:0:(ldlm_lib.c:2071:target_recovery_thread()) too long recovery - read logs
00:40:32:LustreError: dumping log to /tmp/lustre-log.1407829195.17775
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Unfortunately, the debug log /tmp/lustre-log.1407829195.17775 was not gathered. Still investigating.&lt;/p&gt;</comment>
                            <comment id="92116" author="yujian" created="Thu, 21 Aug 2014 07:25:13 +0000"  >&lt;p&gt;The failure has not occurred on Lustre b2_5 branch since build #76.&lt;/p&gt;</comment>
                            <comment id="186377" author="adilger" created="Mon, 27 Feb 2017 23:36:03 +0000"  >&lt;p&gt;Close old bug.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwrfr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>14913</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>