<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:57:17 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6108] Interop 2.6.0&lt;-&gt;2.7 sanity-scrub test_12: OST oops</title>
                <link>https://jira.whamcloud.com/browse/LU-6108</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/767fbc34-9624-11e4-acfc-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/767fbc34-9624-11e4-acfc-5254006e85c2&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_12 failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;test failed to respond and timed out
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;


&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;18:56:07:Lustre: DEBUG MARKER: == sanity-scrub test 12: OI scrub can rebuild invalid /O entries == 18:55:49 (1420512949)
18:56:07:Lustre: DEBUG MARKER: /usr/sbin/lctl set_param fail_loc=0x195
18:56:07:Lustre: *** cfs_fail_loc=195, val=0***
18:56:07:Lustre: DEBUG MARKER: grep -c /mnt/ost1&apos; &apos; /proc/mounts
18:56:07:Lustre: DEBUG MARKER: umount -d /mnt/ost1
18:56:07:Lustre: Failing over lustre-OST0000
18:56:07:Lustre: Skipped 2 previous similar messages
18:56:07:Lustre: server umount lustre-OST0000 complete
18:56:07:Lustre: Skipped 3 previous similar messages
18:56:07:Lustre: DEBUG MARKER: lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
18:56:07:LustreError: 137-5: lustre-OST0000_UUID: not available for connect from 10.2.4.99@tcp (no target). If you are running an HA pair check that the target is mounted on the other server.
18:56:07:LustreError: Skipped 3 previous similar messages
18:56:07:Lustre: DEBUG MARKER: /usr/sbin/lctl set_param fail_loc=0x233
18:56:07:Lustre: DEBUG MARKER: mkdir -p /mnt/ost1
18:56:07:Lustre: DEBUG MARKER: test -b /dev/lvm-Role_OSS/P1
18:56:07:Lustre: DEBUG MARKER: mkdir -p /mnt/ost1; mount -t lustre -o user_xattr,noscrub  		                   /dev/lvm-Role_OSS/P1 /mnt/ost1
18:56:07:LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. quota=on. Opts: 
18:56:07:Lustre: DEBUG MARKER: PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/u
18:56:07:Lustre: lustre-OST0000: Will be in recovery for at least 1:00, or until 2 clients reconnect
18:56:07:Lustre: Skipped 1 previous similar message
18:56:08:Lustre: DEBUG MARKER: e2label /dev/lvm-Role_OSS/P1 2&amp;gt;/dev/null
18:56:08:Lustre: lustre-OST0000: Denying connection for new client 2bf68531-a8c6-f88c-5c44-2fce40112099 (at 10.2.4.94@tcp), waiting for all 2 known clients (1 recovered, 0 in progress, and 0 evicted) to recover in 1:05
18:56:08:Lustre: lustre-OST0000: Recovery over after 0:05, of 2 clients 2 recovered and 0 were evicted.
18:56:08:Lustre: Skipped 2 previous similar messages
18:56:08:Lustre: *** cfs_fail_loc=233, val=0***
18:56:08:LustreError: 9129:0:(ldlm_resource.c:1150:ldlm_resource_get()) lustre-OST0000: lvbo_init failed for resource 0x26b:0x0: rc = -78
18:56:08:LustreError: 9129:0:(ldlm_resource.c:1150:ldlm_resource_get()) Skipped 2 previous similar messages
18:56:08:BUG: unable to handle kernel paging request at 000000005a5a5a5a
18:56:08:IP: [&amp;lt;ffffffffa07ea251&amp;gt;] ldlm_lock_create+0x201/0xd70 [ptlrpc]
18:56:08:PGD 0 
18:56:08:Oops: 0000 [#1] SMP 
18:56:08:last sysfs file: /sys/devices/system/cpu/online
18:56:08:CPU 0 
18:56:08:Modules linked in: osp(U) ofd(U) lfsck(U) ost(U) mgc(U) osd_ldiskfs(U) lquota(U) lustre(U) lov(U) mdc(U) fid(U) lmv(U) fld(U) ksocklnd(U) ptlrpc(U) obdclass(U) lnet(U) sha512_generic sha256_generic libcfs(U) ldiskfs(U) jbd2 nfs fscache nfsd lockd nfs_acl auth_rpcgss sunrpc exportfs autofs4 ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode virtio_balloon 8139too 8139cp mii i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk pata_acpi ata_generic ata_piix virtio_pci virtio_ring virtio dm_mirror dm_region_hash dm_log dm_mod [last unloaded: obdecho]
18:56:08:
18:56:08:Pid: 11802, comm: ll_ost00_057 Tainted: G        W  ---------------    2.6.32-431.20.3.el6_lustre.x86_64 #1 Red Hat KVM
18:56:08:RIP: 0010:[&amp;lt;ffffffffa07ea251&amp;gt;]  [&amp;lt;ffffffffa07ea251&amp;gt;] ldlm_lock_create+0x201/0xd70 [ptlrpc]
18:56:08:RSP: 0018:ffff880040213c60  EFLAGS: 00010246
18:56:08:RAX: ffff88004f297ad8 RBX: ffff88004f297900 RCX: ffff88007c022840
18:56:08:RDX: 0000000000000001 RSI: 0000000000000000 RDI: ffff88004f297a10
18:56:09:RBP: ffff880040213cb0 R08: 0000000000000005 R09: 0000000000000000
18:56:09:R10: ffff88004f297900 R11: 0000000000000200 R12: ffff88004f297900
18:56:09:R13: ffffffffa0916ea0 R14: ffff88006d5a0000 R15: 000000005a5a5a5a
18:56:09:FS:  0000000000000000(0000) GS:ffff880002200000(0000) knlGS:0000000000000000
18:56:09:CS:  0010 DS: 0018 ES: 0018 CR0: 000000008005003b
18:56:09:CR2: 000000005a5a5a5a CR3: 0000000079c56000 CR4: 00000000000006f0
18:56:09:DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
18:56:09:DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
18:56:09:Process ll_ost00_057 (pid: 11802, threadinfo ffff880040212000, task ffff88004020eaa0)
18:56:09:Stack:
18:56:09: ffff880000000010 0000000000000000 0000000b00000004 ffffffffa05d5e31
18:56:09:&amp;lt;d&amp;gt; ffff88003deddcc8 ffff880049cea000 ffff88003deddda8 ffffffffa0916ea0
18:56:09:&amp;lt;d&amp;gt; ffff88006d5a0000 0000000000000001 ffff880040213d20 ffffffffa0811c74
18:56:09:Call Trace:
18:56:09: [&amp;lt;ffffffffa05d5e31&amp;gt;] ? lprocfs_counter_add+0x151/0x1c0 [obdclass]
18:56:09: [&amp;lt;ffffffffa0811c74&amp;gt;] ldlm_handle_enqueue0+0x174/0x11d0 [ptlrpc]
18:56:09: [&amp;lt;ffffffffa08947c2&amp;gt;] tgt_enqueue+0x62/0x1d0 [ptlrpc]
18:56:09: [&amp;lt;ffffffffa0894b6c&amp;gt;] tgt_request_handle+0x23c/0xac0 [ptlrpc]
18:56:09: [&amp;lt;ffffffffa084426a&amp;gt;] ptlrpc_main+0xd1a/0x1980 [ptlrpc]
18:56:09: [&amp;lt;ffffffffa0843550&amp;gt;] ? ptlrpc_main+0x0/0x1980 [ptlrpc]
18:56:09: [&amp;lt;ffffffff8109abf6&amp;gt;] kthread+0x96/0xa0
18:56:09: [&amp;lt;ffffffff8100c20a&amp;gt;] child_rip+0xa/0x20
18:56:09: [&amp;lt;ffffffff8109ab60&amp;gt;] ? kthread+0x0/0xa0
18:56:09: [&amp;lt;ffffffff8100c200&amp;gt;] ? child_rip+0x0/0x20
18:56:09:Code: 00 00 49 8d 84 24 d8 01 00 00 49 c7 84 24 90 00 00 00 00 00 00 00 ba 01 00 00 00 49 89 84 24 d8 01 00 00 49 89 84 24 e0 01 00 00 &amp;lt;49&amp;gt; 8b 07 48 8b 00 48 8b b8 88 01 00 00 e8 7d ba de ff 4d 89 24 
18:56:09:RIP  [&amp;lt;ffffffffa07ea251&amp;gt;] ldlm_lock_create+0x201/0xd70 [ptlrpc]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Info required for matching: sanity-scrub 12&lt;/p&gt;</description>
                <environment>server: 2.6.0&lt;br/&gt;
client: lustre-master build #2808</environment>
        <key id="28125">LU-6108</key>
            <summary>Interop 2.6.0&lt;-&gt;2.7 sanity-scrub test_12: OST oops</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Fri, 9 Jan 2015 23:40:01 +0000</created>
                <updated>Sat, 10 Jan 2015 02:00:31 +0000</updated>
                            <resolved>Sat, 10 Jan 2015 02:00:31 +0000</resolved>
                                    <version>Lustre 2.7.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>2</watches>
                                                                            <comments>
                            <comment id="103126" author="yong.fan" created="Sat, 10 Jan 2015 02:00:31 +0000"  >&lt;p&gt;It is another failure instance of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5904&quot; title=&quot;Interop 2.6.0&amp;lt;-&amp;gt;master sanity-scrub test_13: BUG: ldlm_lock_create() on OST&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5904&quot;&gt;&lt;del&gt;LU-5904&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzx3n3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>16995</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>