<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:20:20 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-1863] Test failure with MDS spontaneous rebooting (test suite sanity, subtest test_32n)</title>
                <link>https://jira.whamcloud.com/browse/LU-1863</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Minh Diep &amp;lt;mdiep@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/f3243a50-f9b2-11e1-b8d8-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/f3243a50-f9b2-11e1-b8d8-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_32n failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;test failed to respond and timed out&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;The console on the OSS shows it rebooted for no reason and conman fail to capture the issue&lt;/p&gt;

&lt;p&gt;Lustre: DEBUG MARKER: /usr/sbin/lctl mark == sanity test 32n: open d32n\/symlink-\&amp;gt;tmp\/symlink-\&amp;gt;lustre-root ====================================== 16:04:09 &amp;#40;1346972649&amp;#41;^M&lt;br/&gt;
Lustre: DEBUG MARKER: == sanity test 32n: open d32n/symlink-&amp;gt;tmp/symlink-&amp;gt;lustre-root ====================================== 16:04:09 (1346972649)^M&lt;br/&gt;
Lustre: 3770:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346972670/real 1346972670&amp;#93;&lt;/span&gt;  req@ffff880071728c00 x1412399426961989/t0(0) o400-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 224/224 e 0 to 1 dl 1346972677 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
LustreError: 166-1: MGC10.10.4.222@tcp: Connection to MGS (at 10.10.4.222@tcp) was lost; in progress operations using this service will fail^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346972702/real 1346972702&amp;#93;&lt;/span&gt;  req@ffff880078a46400 x1412399426961991/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346972713 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 1 previous similar message^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346972727/real 1346972729&amp;#93;&lt;/span&gt;  req@ffff880074d4bc00 x1412399426961992/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346972743 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346972777/real 1346972777&amp;#93;&lt;/span&gt;  req@ffff8800791f2400 x1412399426961994/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346972803 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 1 previous similar message^M&lt;br/&gt;
Lustre: lustre-OST0000: haven&apos;t heard from client lustre-MDT0000-mdtlov_UUID (at 10.10.4.222@tcp) in 235 seconds. I think it&apos;s dead, and I am evicting it. exp ffff88007178e400, cur 1346972885 expire 1346972735 last 1346972650^M&lt;br/&gt;
Lustre: lustre-OST0002: haven&apos;t heard from client lustre-MDT0000-mdtlov_UUID (at 10.10.4.222@tcp) in 235 seconds. I think it&apos;s dead, and I am evicting it. exp ffff880061d38400, cur 1346972885 expire 1346972735 last 1346972650^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346972877/real 1346972880&amp;#93;&lt;/span&gt;  req@ffff8800791f2c00 x1412399426961996/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346972913 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 1 previous similar message^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346973002/real 1346973005&amp;#93;&lt;/span&gt;  req@ffff880074050800 x1412399426961999/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346973053 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 2 previous similar messages^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346973302/real 1346973305&amp;#93;&lt;/span&gt;  req@ffff880072541c00 x1412399426962003/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346973357 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 3 previous similar messages^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346973827/real 1346973830&amp;#93;&lt;/span&gt;  req@ffff88007171f800 x1412399426962010/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346973882 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 6 previous similar messages^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346974502/real 1346974505&amp;#93;&lt;/span&gt;  req@ffff880078a46800 x1412399426962019/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346974557 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 8 previous similar messages^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346975177/real 1346975180&amp;#93;&lt;/span&gt;  req@ffff88005ecfac00 x1412399426962028/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346975232 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 8 previous similar messages^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: &lt;span class=&quot;error&quot;&gt;&amp;#91;sent 1346975852/real 1346975855&amp;#93;&lt;/span&gt;  req@ffff880078a46800 x1412399426962037/t0(0) o250-&amp;gt;MGC10.10.4.222@tcp@10.10.4.222@tcp:26/25 lens 400/544 e 0 to 1 dl 1346975907 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1^M&lt;br/&gt;
Lustre: 3768:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 8 previous similar messages^M&lt;br/&gt;
^M&lt;br/&gt;
&amp;lt;ConMan&amp;gt; Console &lt;span class=&quot;error&quot;&gt;&amp;#91;client-19vm4&amp;#93;&lt;/span&gt; disconnected from &amp;lt;client-19:6003&amp;gt; at 09-06 17:05.^M&lt;br/&gt;
^M&lt;br/&gt;
&amp;lt;ConMan&amp;gt; Console &lt;span class=&quot;error&quot;&gt;&amp;#91;client-19vm4&amp;#93;&lt;/span&gt; connected to &amp;lt;client-19:6003&amp;gt; at 09-06 17:05.^M&lt;br/&gt;
^MPress any key to continue.^M&lt;br/&gt;
^MPress any key to continue.^M&lt;br/&gt;
^MPress any key to continue.^M&lt;br/&gt;
^MPress any key to continue.^M&lt;br/&gt;
^MPress any key to continue.^M&lt;br/&gt;
&lt;sup&gt;[[H&lt;/sup&gt;[[J^M&lt;/p&gt;


&lt;p&gt;Info required for matching: sanity 32n&lt;/p&gt;</description>
                <environment></environment>
        <key id="15838">LU-1863</key>
            <summary>Test failure with MDS spontaneous rebooting (test suite sanity, subtest test_32n)</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="bobijam">Zhenyu Xu</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>USE_OFD</label>
                            <label>zfs</label>
                    </labels>
                <created>Sat, 8 Sep 2012 11:24:47 +0000</created>
                <updated>Wed, 31 Jul 2013 15:20:06 +0000</updated>
                            <resolved>Wed, 31 Jul 2013 15:20:06 +0000</resolved>
                                    <version>Lustre 2.3.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="44446" author="pjones" created="Sun, 9 Sep 2012 10:49:52 +0000"  >&lt;p&gt;Not just limited to OFD runs&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/957f03fe-fa6f-11e1-887d-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/957f03fe-fa6f-11e1-887d-52540035b04c&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/a41657fe-fa70-11e1-a03c-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/a41657fe-fa70-11e1-a03c-52540035b04c&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/75760d88-fa50-11e1-887d-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/75760d88-fa50-11e1-887d-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="44447" author="pjones" created="Sun, 9 Sep 2012 10:52:38 +0000"  >&lt;p&gt;Bobijam&lt;/p&gt;

&lt;p&gt;Could you please look into this failure?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="44466" author="bobijam" created="Sun, 9 Sep 2012 23:37:28 +0000"  >&lt;p&gt;Something happened on MDS make it out of service, there is no log from MDS though, I&apos;ll trying to reproduce it on Toro nodes.&lt;/p&gt;</comment>
                            <comment id="44508" author="yujian" created="Mon, 10 Sep 2012 12:02:48 +0000"  >&lt;p&gt;Hi Bobi,&lt;/p&gt;

&lt;p&gt;FYI, I also found some failures on the latest b2_3 sessions related to MDS out of service issue:&lt;br/&gt;
conf-sanity test 45: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/aaac51d6-fa70-11e1-887d-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/aaac51d6-fa70-11e1-887d-52540035b04c&lt;/a&gt;&lt;br/&gt;
parallel-scale-nfsv3: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/9fe57158-fa73-11e1-887d-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/9fe57158-fa73-11e1-887d-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;More failures are reported in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1926&quot; title=&quot;Reboots during test runs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1926&quot;&gt;&lt;del&gt;TT-851&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="44510" author="adilger" created="Mon, 10 Sep 2012 12:30:35 +0000"  >&lt;p&gt;I think the conf-sanity.sh test_45 failure reported above should be filed in a separate bug.  I think it is just a failure in that test, and does not contain any report of the MDS spontaneous rebooting.&lt;/p&gt;</comment>
                            <comment id="44565" author="yujian" created="Tue, 11 Sep 2012 01:59:46 +0000"  >&lt;blockquote&gt;&lt;p&gt;I think the conf-sanity.sh test_45 failure reported above should be filed in a separate bug. I think it is just a failure in that test, and does not contain any report of the MDS spontaneous rebooting.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;My fault. I saw that &quot;Connection to lustre-MDT0000 (at 10.10.4.182@tcp) was lost;&quot; on the client node and thought that MDS was out of service, but did not realize that the MDS was unmounted before. &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1882&quot; title=&quot;conf-sanity test_45: umount2: Device or resource busy&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1882&quot;&gt;&lt;del&gt;LU-1882&lt;/del&gt;&lt;/a&gt; was created.&lt;/p&gt;</comment>
                            <comment id="44675" author="yujian" created="Wed, 12 Sep 2012 07:34:35 +0000"  >&lt;p&gt;Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_3/17&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_3/17&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Still failed: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/d8e2ddcc-fc82-11e1-a4a6-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/d8e2ddcc-fc82-11e1-a4a6-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="44681" author="yujian" created="Wed, 12 Sep 2012 09:28:55 +0000"  >&lt;p&gt;Lustre Server Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_3/17&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_3/17&lt;/a&gt;&lt;br/&gt;
Lustre Client Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_1/121&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_1/121&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;sanity test 32n also hung: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/47eebb26-fbf9-11e1-a4a6-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/47eebb26-fbf9-11e1-a4a6-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;However, this time I found the following message on the console log of MDS (client-20):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00:15:57:Lustre: DEBUG MARKER: == sanity test 32n: open d32n/symlink-&amp;gt;tmp/symlink-&amp;gt;lustre-root ======== 00:15:55 (1347347755)
00:15:57:------------[ cut here ]------------
00:15:57:WARNING: at kernel/sched_fair.c:132 load_balance_next_fair+0x6a/0x80() (Not tainted)
00:15:57:Hardware name: X8DTT
00:17:18:Modules linked in: nfs fscache cmm(U) osd_ldiskfs(U) mdt(U) mdd(U) mds(U) (T)&#65533;�[0m�[1;30;40m�[01;00H
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="44791" author="yujian" created="Thu, 13 Sep 2012 09:20:13 +0000"  >&lt;p&gt;Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_3/17&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_3/17&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Only running sanity test 32 passed: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/6b92dd2a-fd9b-11e1-afe5-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/6b92dd2a-fd9b-11e1-afe5-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Running sanity from start hung at test 32g: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/a52af1ca-fda1-11e1-afe5-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/a52af1ca-fda1-11e1-afe5-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Console log on MDS (fat-intel-2):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: DEBUG MARKER: == sanity test 32g: stat d32g/symlink-&amp;gt;tmp/symlink-&amp;gt;lustre-subdir/2 == 05:39:13 (1347539953)
BUG: unable to handle kernel NULL pointer dereference at 0000000000000040
IP: [&amp;lt;ffffffff8150057e&amp;gt;] _spin_lock+0xe/0x30
PGD 0 
Oops: 0002 [#1] SMP 
last sysfs file: /sys/devices/system/cpu/cpu23/cache/index2/shared_cpu_map
CPU 18 
Modules linked in: nfs fscache cmm(U) osd_ldiskfs(U) mdt(U) mdd(U) mds(U) fsfilt_ldiskfs(U) mgs(U) mgc(U) ldiskfs(U) jbd2 lustre(U) lquota(U) lov(U) osc(U) mdc(U) fid(U) fld(U) ksocklnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) sha512_generic sha256_generic libcfs(U) nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc cpufreq_ondemand acpi_cpufreq freq_table mperf ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa mlx4_ib ib_mad ib_core mlx4_en mlx4_core e1000e microcode serio_raw i2c_i801 i2c_core sg iTCO_wdt iTCO_vendor_support ioatdma dca i7core_edac edac_core shpchp ext3 jbd mbcache sd_mod crc_t10dif ahci dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]

Pid: 0, comm: swapper Not tainted 2.6.32-279.5.1.el6_lustre.g634f764.x86_64 #1 Supermicro X8DTT-H/X8DTT-H
RIP: 0010:[&amp;lt;ffffffff8150057e&amp;gt;]  [&amp;lt;ffffffff8150057e&amp;gt;] _spin_lock+0xe/0x30
RSP: 0018:ffff88034acc3dd0  EFLAGS: 00010002
RAX: 0000000000010000 RBX: 000000000000f14c RCX: ffff880336b68800
RDX: 0040000000000080 RSI: ffff8806304dc080 RDI: 0000000000000040
RBP: ffff88034acc3dd0 R08: 0000000000000000 R09: ffff880337eefe68
R10: 000002936cfd7c69 R11: 0000000000000000 R12: 0000000000000000
R13: 0000000000000286 R14: ffff88033fcd0340 R15: 000000000000e140
FS:  0000000000000000(0000) GS:ffff88034acc0000(0000) knlGS:0000000000000000
CS:  0010 DS: 0018 ES: 0018 CR0: 000000008005003b
CR2: 0000000000000040 CR3: 00000003303fc000 CR4: 00000000000006e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="44870" author="yujian" created="Fri, 14 Sep 2012 10:47:01 +0000"  >&lt;p&gt;Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_3/19&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_3/19&lt;/a&gt; (with fix for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1881&quot; title=&quot;sanity test 116 soft lockup&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1881&quot;&gt;&lt;del&gt;LU-1881&lt;/del&gt;&lt;/a&gt;)&lt;/p&gt;

&lt;p&gt;sanity tests passed: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/19ee1aae-fe7a-11e1-a707-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/19ee1aae-fe7a-11e1-a707-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="44947" author="pjones" created="Sat, 15 Sep 2012 09:37:12 +0000"  >&lt;p&gt;Lowering priority as this now only occurs with USE_OFED=yes&lt;/p&gt;</comment>
                            <comment id="44948" author="yujian" created="Sat, 15 Sep 2012 09:40:56 +0000"  >&lt;p&gt;Hi Minh,&lt;/p&gt;

&lt;p&gt;While I running sanity tests on Lustre build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_3/19&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_3/19&lt;/a&gt; with the following variables:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;USE_OFD=yes
OSTFSTYPE=zfs
LOAD_MODULES_REMOTE=true
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I hit the following module loading issue:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;fat-intel-4: FATAL: Error inserting osd_zfs (/lib/modules/2.6.32-279.5.1.el6_lustre.x86_64/updates/kernel/fs/lustre/osd_zfs.ko): Unknown symbol in module, or unknown parameter (see dmesg)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Dmesg on fat-intel-4:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: Lustre: Build Version: 2.2.96--PRISTINE-2.6.32-279.5.1.el6_lustre.x86_64
LNet: Added LNI 10.10.4.131@tcp [8/256/0/180]
LNet: Accept secure, port 988
Lustre: Echo OBD driver; http://www.lustre.org/
osd_zfs: Unknown symbol zap_cursor_serialize
osd_zfs: Unknown symbol zap_remove
osd_zfs: Unknown symbol dmu_tx_hold_write
osd_zfs: Unknown symbol zfs_attr_table
osd_zfs: Unknown symbol nvpair_value_nvlist
osd_zfs: Unknown symbol sa_spill_rele
osd_zfs: Unknown symbol zap_cursor_advance
osd_zfs: Unknown symbol kmem_alloc_debug
osd_zfs: Unknown symbol dmu_read
osd_zfs: Unknown symbol nvlist_add_byte_array
osd_zfs: Unknown symbol zap_cursor_init_serialized
......
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Did you hit the above issue before?&lt;/p&gt;</comment>
                            <comment id="44960" author="liwei" created="Sat, 15 Sep 2012 22:16:42 +0000"  >&lt;p&gt;Yu Jian,&lt;/p&gt;

&lt;p&gt;Those symbols should be provided by spl/zfs.  Unlike LDiskFS, spl/zfs source trees are not included in lustre.  Jenkins pulls and builds them when building lustre, and generates spl/zfs RPMs.  I&apos;d check if spl-modules and zfs-modules are installed in this case.  (Historically, loadjenkinsbuild did not automatically install spl/zfs RPMs that came with requested lustre builds.)&lt;/p&gt;</comment>
                            <comment id="45211" author="yujian" created="Wed, 19 Sep 2012 06:32:18 +0000"  >&lt;p&gt;Thanks Li Wei.&lt;/p&gt;

&lt;p&gt;After installing spl-modules and zfs-modules packages, the sanity tests went forward and finished on Lustre b2_3 build #19 with USE_OFD=yes:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/ff130926-0241-11e2-ab94-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/ff130926-0241-11e2-ab94-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The issue in this ticket did not occur.&lt;/p&gt;</comment>
                            <comment id="63375" author="utopiabound" created="Wed, 31 Jul 2013 15:20:06 +0000"  >&lt;p&gt;This is a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1926&quot; title=&quot;Reboots during test runs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1926&quot;&gt;&lt;del&gt;LU-1926&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="15634">LU-1926</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv4jr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4270</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>