<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:58:20 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6222] LustreError (statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;entry-&gt;se_list) )</title>
                <link>https://jira.whamcloud.com/browse/LU-6222</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We seem to be having an issue similar to the one described in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5883&quot; title=&quot;DNE II testing: LustreError: 6618:0:(statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;amp;entry-&amp;gt;se_list) ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5883&quot;&gt;&lt;del&gt;LU-5883&lt;/del&gt;&lt;/a&gt;. We&apos;re using 2.6.93 on the clients and 2.6.92 on the servers. The problem can be reproduced reliably on our end, so I&apos;d be happy to provide additional logs/diagnostic information (Is there a standard procedure for this? I couldn&apos;t seem to find anything via google.) I&apos;ve included the kernel log messages that were dumped to the console as well as the stack trace below.&lt;/p&gt;

&lt;p&gt;kernel:LustreError: 13007:0:(statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;amp;entry-&amp;gt;se_list) ) failed:&lt;br/&gt;
kernel:LustreError: 13007:0:(statahead.c:262:sa_kill()) LBUG&lt;br/&gt;
PID: 13007 TASK: ffff880b0aea2040 CPU: 25 COMMAND: &quot;rsync&quot;&lt;br/&gt;
#0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef16f0&amp;#93;&lt;/span&gt; machine_kexec at ffffffff8103b5bb&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/arch/x86/kernel/machine_kexec_64.c: 336&lt;br/&gt;
#1 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1750&amp;#93;&lt;/span&gt; crash_kexec at ffffffff810c9852&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/kernel/kexec.c: 1106&lt;br/&gt;
#2 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1820&amp;#93;&lt;/span&gt; panic at ffffffff8152927e&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/kernel/panic.c: 111&lt;br/&gt;
#3 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef18a0&amp;#93;&lt;/span&gt; lbug_with_loc at ffffffffa03c5eeb &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
/usr/src/debug/lustre-2.6.93/libcfs/libcfs/linux/linux-debug.c: 175&lt;br/&gt;
#4 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef18c0&amp;#93;&lt;/span&gt; revalidate_statahead_dentry at ffffffffa0a0e31d &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
/usr/src/debug/lustre-2.6.93/lustre/llite/statahead.c: 263&lt;br/&gt;
#5 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1a00&amp;#93;&lt;/span&gt; ll_statahead at ffffffffa0a0e642 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
/usr/src/debug/lustre-2.6.93/libcfs/include/libcfs/libcfs_debug.h: 219&lt;br/&gt;
#6 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1ae0&amp;#93;&lt;/span&gt; ll_lookup_it at ffffffffa09f8907 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
/usr/src/debug/lustre-2.6.93/lustre/llite/namei.c: 541&lt;br/&gt;
#7 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1ba0&amp;#93;&lt;/span&gt; ll_lookup_nd at ffffffffa09f9029 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
/usr/src/debug/lustre-2.6.93/lustre/llite/namei.c: 771&lt;br/&gt;
#8 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1bf0&amp;#93;&lt;/span&gt; do_lookup at ffffffff8119dc65&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/namei.c: 1063&lt;br/&gt;
#9 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1c50&amp;#93;&lt;/span&gt; __link_path_walk at ffffffff8119e8f4&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/namei.c: 1239&lt;br/&gt;
#10 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1d30&amp;#93;&lt;/span&gt; path_walk at ffffffff8119f40a&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/namei.c: 558&lt;br/&gt;
#11 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1d70&amp;#93;&lt;/span&gt; filename_lookup at ffffffff8119f61b&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/namei.c: 1375&lt;br/&gt;
#12 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1db0&amp;#93;&lt;/span&gt; user_path_at at ffffffff811a0747&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/namei.c: 1597&lt;br/&gt;
#13 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1e80&amp;#93;&lt;/span&gt; vfs_fstatat at ffffffff81193bc0&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/stat.c: 84&lt;br/&gt;
#14 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1ee0&amp;#93;&lt;/span&gt; vfs_lstat at ffffffff81193c7e&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/stat.c: 107&lt;br/&gt;
#15 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1ef0&amp;#93;&lt;/span&gt; sys_newlstat at ffffffff81193ca4&lt;br/&gt;
/usr/src/debug/kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/fs/stat.c: 257&lt;br/&gt;
#16 &lt;span class=&quot;error&quot;&gt;&amp;#91;ffff881c20ef1f80&amp;#93;&lt;/span&gt; system_call_fastpath at ffffffff8100b072&lt;br/&gt;
/usr/src/debug////////kernel-2.6.32-504.3.3.el6/linux-2.6.32-504.3.3.el6.x86_64/arch/x86/kernel/entry_64.S: 489&lt;br/&gt;
RIP: 00000031f78daf75 RSP: 00007fff51e03da8 RFLAGS: 00010206&lt;br/&gt;
RAX: 0000000000000006 RBX: ffffffff8100b072 RCX: 00007fff51e040b0&lt;br/&gt;
RDX: 00007fff51dffb10 RSI: 00007fff51dffb10 RDI: 00007fff51e00ba0&lt;br/&gt;
RBP: 0000000000000000 R8: 00007fff51e00be3 R9: 0000000000000000&lt;br/&gt;
R10: 0000000000000000 R11: 0000000000000246 R12: 00007fff51e00ba0&lt;br/&gt;
R13: 00007fff51dffb10 R14: 00007fff51dffb10 R15: 00007fff51e00ba0&lt;br/&gt;
ORIG_RAX: 0000000000000006 CS: 0033 SS: 002b&lt;/p&gt;</description>
                <environment>2.6.93 on the clients and 2.6.92 on the servers, centos with rpms from lustre-master jenkins tree.</environment>
        <key id="28591">LU-6222</key>
            <summary>LustreError (statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;entry-&gt;se_list) )</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="laisiyao">Lai Siyao</assignee>
                                    <reporter username="azenk">Andrew Zenk</reporter>
                        <labels>
                    </labels>
                <created>Fri, 6 Feb 2015 17:36:23 +0000</created>
                <updated>Mon, 8 Jun 2015 07:50:48 +0000</updated>
                            <resolved>Fri, 13 Feb 2015 05:20:27 +0000</resolved>
                                                    <fixVersion>Lustre 2.7.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="106055" author="jlevi" created="Fri, 6 Feb 2015 18:10:12 +0000"  >&lt;p&gt;Lai,&lt;br/&gt;
Could you please have a look at this one and comment?&lt;br/&gt;
Thank you!&lt;/p&gt;</comment>
                            <comment id="106056" author="green" created="Fri, 6 Feb 2015 18:11:19 +0000"  >&lt;p&gt;Can you please detail your reproduction steps if this is something we can easily replicate?&lt;br/&gt;
Do you use DNE? What&apos;s your exact configuration of Lustre?&lt;/p&gt;</comment>
                            <comment id="106076" author="azenk" created="Fri, 6 Feb 2015 18:58:19 +0000"  >&lt;p&gt;We do not use DNE.  In this case a user is running an rsync as follows &quot;rsync --size-only --progress -av --prune-empty-dirs --include=/ --exclude=.man --exclude=&lt;b&gt;_eot.txt --exclude=.MAN --exclude=_EOT.TXT --include=052903541090_01&lt;/b&gt;* --exclude=* /lustre_mountpoint/staging/orig/_uploads/DG_A11281 /lustre_mountpoint/somepath/northslope/&quot;  This command is run several times  in a serial fashion from a single client with slight variations of the include and srcdir for each run.  At a some, seemingly random point during the sequence of rsync jobs, the kernel on the client node panics. &lt;/p&gt;

&lt;p&gt;Configuration:&lt;br/&gt;
We&apos;re using a single MDS with nine OSSs.  Each OSS contains 3 direct attached targets.  Our system isn&apos;t designed for failover.  The MDS and OSS nodes are booted from a common image built using centos 6.6 with this build: &lt;a href=&quot;https://build.hpdd.intel.com/job/lustre-master/arch=x86_64,build_type=server,distro=el6.6,ib_stack=inkernel/2832/artifact/artifacts/RPMS/x86_64/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://build.hpdd.intel.com/job/lustre-master/arch=x86_64,build_type=server,distro=el6.6,ib_stack=inkernel/2832/artifact/artifacts/RPMS/x86_64/&lt;/a&gt;.  &lt;/p&gt;

&lt;p&gt;I&apos;m happy to supply exact specs on raid configurations and disk counts if you feel that it&apos;s important, but we&apos;ll skip that for now.  The MDS is using a single target on a ssd raid10.  The OSTs are sata of various types.   All servers are connected to our QDR IB fabric as well as a gigabit VLAN.  The latter is used for connecting 3 clients that aren&apos;t experiencing any issues.  &lt;/p&gt;

&lt;p&gt;There are approximately 20 clients, which are also using centos 6.6.  The lustre clients are installed via the pre-built rpms from lustre-master, just like the servers.  Though the clients are of slightly mixed build versions.  The two clients that we&apos;ve reproduced the issue on were both running build #2835.&lt;/p&gt;

&lt;p&gt;The entire filesystem has a stripe count of 1.&lt;/p&gt;

&lt;p&gt;Let me know if you need any additional information.  Thanks!&lt;/p&gt;

</comment>
                            <comment id="106226" author="laisiyao" created="Mon, 9 Feb 2015 06:02:48 +0000"  >&lt;p&gt;could you list all process backtraces in the dump?&lt;/p&gt;</comment>
                            <comment id="106362" author="azenk" created="Mon, 9 Feb 2015 21:47:04 +0000"  >&lt;p&gt;Attached output from foreach bt -l &lt;/p&gt;</comment>
                            <comment id="106433" author="gerrit" created="Tue, 10 Feb 2015 14:01:20 +0000"  >&lt;p&gt;Lai Siyao (lai.siyao@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/13708&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13708&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6222&quot; title=&quot;LustreError (statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;amp;entry-&amp;gt;se_list) )&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6222&quot;&gt;&lt;del&gt;LU-6222&lt;/del&gt;&lt;/a&gt; statahead: add to list before make ready&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: c77d1c16d45850e3cd4492d27746b30a89ba2beb&lt;/p&gt;</comment>
                            <comment id="106434" author="laisiyao" created="Tue, 10 Feb 2015 14:03:29 +0000"  >&lt;p&gt;Andrew, I just uploaded a fix for this issue, will you apply it and test again?&lt;/p&gt;</comment>
                            <comment id="106448" author="azenk" created="Tue, 10 Feb 2015 15:49:48 +0000"  >&lt;p&gt;Thanks!  We&apos;re testing it now.&lt;/p&gt;</comment>
                            <comment id="106628" author="azenk" created="Wed, 11 Feb 2015 15:26:19 +0000"  >&lt;p&gt;That seems to have fixed it.  The rsync script that consistently caused the issue after a minute or two has been running flawlessly for many hours.  Thanks again.&lt;/p&gt;</comment>
                            <comment id="106896" author="gerrit" created="Fri, 13 Feb 2015 00:54:38 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/13708/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13708/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6222&quot; title=&quot;LustreError (statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;amp;entry-&amp;gt;se_list) )&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6222&quot;&gt;&lt;del&gt;LU-6222&lt;/del&gt;&lt;/a&gt; statahead: add to list before make ready&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: c59baf3ff10beca9841bc8aae211af120ab913dc&lt;/p&gt;</comment>
                            <comment id="106903" author="pjones" created="Fri, 13 Feb 2015 05:20:27 +0000"  >&lt;p&gt;Landed for 2.7&lt;/p&gt;</comment>
                            <comment id="117716" author="gerrit" created="Mon, 8 Jun 2015 07:50:48 +0000"  >&lt;p&gt;Lai Siyao (lai.siyao@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/15178&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/15178&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6222&quot; title=&quot;LustreError (statahead.c:262:sa_kill()) ASSERTION( !list_empty(&amp;amp;entry-&amp;gt;se_list) )&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6222&quot;&gt;&lt;del&gt;LU-6222&lt;/del&gt;&lt;/a&gt; statahead: add to list before make ready&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_5&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 1091da447db103549ca8e20d5c5fb97679f7080c&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="16922" name="all_bt.out" size="819640" author="azenk" created="Mon, 9 Feb 2015 21:47:04 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzx5uf:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>17399</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>