<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:14:37 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-1217] ASSERTION(cfs_atomic_read(&amp;lock-&gt;cll_ref) &gt; 1 || olock-&gt;ols_hold == 0) failed</title>
                <link>https://jira.whamcloud.com/browse/LU-1217</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;system crashed with following assertion &lt;/p&gt;

&lt;p&gt;LustreError: 35225:0:(osc_page.c:293:osc_page_putref_lock()) ASSERTION(cfs_atomic_read(&amp;amp;lock-&amp;gt;cll_ref) &amp;gt; 1 || olock-&amp;gt;ols_hold == 0) failed&lt;br/&gt;
LustreError: 35225:0:(osc_page.c:293:osc_page_putref_lock()) LBUG&lt;br/&gt;
Pid: 35225, comm: ptlrpcd_6&lt;/p&gt;

&lt;p&gt;Call Trace:&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0369855&amp;gt;&amp;#93;&lt;/span&gt; libcfs_debug_dumpstack+0x55/0x80 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0369e95&amp;gt;&amp;#93;&lt;/span&gt; lbug_with_loc+0x75/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0375fe6&amp;gt;&amp;#93;&lt;/span&gt; libcfs_assertion_failed+0x66/0x70 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa064ca06&amp;gt;&amp;#93;&lt;/span&gt; osc_page_putref_lock+0xb6/0xc0 &lt;span class=&quot;error&quot;&gt;&amp;#91;osc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa064ca28&amp;gt;&amp;#93;&lt;/span&gt; osc_page_completion_read+0x18/0x20 &lt;span class=&quot;error&quot;&gt;&amp;#91;osc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa04555af&amp;gt;&amp;#93;&lt;/span&gt; cl_page_completion+0x21f/0x550 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa045610b&amp;gt;&amp;#93;&lt;/span&gt; ? cl_page_put+0x1ab/0x440 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa053b604&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_msg_get_opc+0x94/0x100 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa045f1e8&amp;gt;&amp;#93;&lt;/span&gt; ? cl_req_page_done+0x38/0x180 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa064d145&amp;gt;&amp;#93;&lt;/span&gt; osc_completion+0x165/0x380 &lt;span class=&quot;error&quot;&gt;&amp;#91;osc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa055d734&amp;gt;&amp;#93;&lt;/span&gt; ? at_measured+0x114/0x360 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa054099e&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_master_callback+0x3e/0xc0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0634f48&amp;gt;&amp;#93;&lt;/span&gt; osc_ap_completion+0x128/0x300 &lt;span class=&quot;error&quot;&gt;&amp;#91;osc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0641cb9&amp;gt;&amp;#93;&lt;/span&gt; brw_interpret+0x4a9/0x710 &lt;span class=&quot;error&quot;&gt;&amp;#91;osc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa052c066&amp;gt;&amp;#93;&lt;/span&gt; ? after_reply+0x7b6/0xe30 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff810017ac&amp;gt;&amp;#93;&lt;/span&gt; ? __switch_to+0x1ac/0x320&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81041760&amp;gt;&amp;#93;&lt;/span&gt; ? fair___dequeue_entity+0x30/0x50&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0530c0a&amp;gt;&amp;#93;&lt;/span&gt; ptlrpc_check_set+0x48a/0x1ba0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff810657ec&amp;gt;&amp;#93;&lt;/span&gt; ? lock_timer_base+0x3c/0x70&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0565147&amp;gt;&amp;#93;&lt;/span&gt; ptlrpcd_check+0x547/0x560 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa056543c&amp;gt;&amp;#93;&lt;/span&gt; ptlrpcd+0x2dc/0x460 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8104c780&amp;gt;&amp;#93;&lt;/span&gt; ? default_wake_function+0x0/0x20&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0565160&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpcd+0x0/0x460 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff810041aa&amp;gt;&amp;#93;&lt;/span&gt; child_rip+0xa/0x20&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0565160&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpcd+0x0/0x460 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0565160&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpcd+0x0/0x460 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff810041a0&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0x0/0x20&lt;/p&gt;

&lt;p&gt;Unless I did something really wrong, &lt;br/&gt;
cfs_atomic_read(&amp;amp;lock-&amp;gt;cll_ref) = cfs_atomic_read(0xffff881dea5c9e78) = 258&lt;/p&gt;

&lt;p&gt;At time of crash, the system was higly loaded with a lot of interactive commands and did suffer of disconnection from some OST &amp;amp; MDT.&lt;/p&gt;</description>
                <environment></environment>
        <key id="13569">LU-1217</key>
            <summary>ASSERTION(cfs_atomic_read(&amp;lock-&gt;cll_ref) &gt; 1 || olock-&gt;ols_hold == 0) failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="jay">Jinshan Xiong</assignee>
                                    <reporter username="louveta">Alexandre Louvet</reporter>
                        <labels>
                    </labels>
                <created>Wed, 14 Mar 2012 15:55:49 +0000</created>
                <updated>Mon, 4 Jun 2012 04:43:29 +0000</updated>
                            <resolved>Mon, 4 Jun 2012 04:43:29 +0000</resolved>
                                    <version>Lustre 2.1.0</version>
                                    <fixVersion>Lustre 2.3.0</fixVersion>
                    <fixVersion>Lustre 2.1.2</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="31217" author="pjones" created="Wed, 14 Mar 2012 16:10:45 +0000"  >&lt;p&gt;Bobi&lt;/p&gt;

&lt;p&gt;Could you please look into this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="31246" author="jay" created="Wed, 14 Mar 2012 18:21:15 +0000"  >&lt;p&gt;it&apos;s impossible for this atomic count to be 258 if it hit this assertion. I don&apos;t know why it&apos;s `cfs_atomic_read(&amp;amp;lock-&amp;gt;cll_ref) &amp;gt; 1&apos; instead of `&amp;gt; 0&apos;.&lt;/p&gt;</comment>
                            <comment id="31975" author="pjones" created="Fri, 23 Mar 2012 09:25:36 +0000"  >&lt;p&gt;Alexandre&lt;/p&gt;

&lt;p&gt;You mentioned on yesterday&apos;s call that it would be possible to supply a crash dump for this ticket. That would definitely help expedite us understanding this scenario.&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="31976" author="bobijam" created="Fri, 23 Mar 2012 09:31:13 +0000"  >&lt;p&gt;Would you mind uploading crash image/kernel image/symbol files to our ftp site?&lt;/p&gt;</comment>
                            <comment id="31989" author="nedbass" created="Fri, 23 Mar 2012 14:38:34 +0000"  >&lt;p&gt;We hit this today as well on one of our classified 2.1 clients.  We have a crash dump, but can&apos;t upload it.  I&apos;m happy to do any analysis of the dump at your direction.&lt;/p&gt;</comment>
                            <comment id="32098" author="bobijam" created="Mon, 26 Mar 2012 01:24:37 +0000"  >&lt;p&gt;first, we need to know the troublesome cl_lock&apos;s content.&lt;/p&gt;</comment>
                            <comment id="32136" author="nedbass" created="Mon, 26 Mar 2012 14:44:40 +0000"  >&lt;blockquote&gt;&lt;p&gt;first, we need to know the troublesome cl_lock&apos;s content.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I&apos;ve only been able to find one address on the stack that looks like a cl_lock.  But, with lock-&amp;gt;cll_ref = 99 it shouldn&apos;t have hit the assertion.  Maybe something touched that memory after the atomic read?&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;struct cl_lock {
  cll_ref = { 
    counter = 99
  },  
  cll_layers = { 
    next = 0xffff88...
    prev = 0xffff88...
  },  
  cll_linkage = { 
    next = 0xffff88...
    prev = 0xffff88...
  },  
  cll_descr = { 
    cld_obj = 0xffff88...
    cld_start = 0,
    cld_end = 18446744073709551615,
    cld_gid = 0,
    cld_mode = CLM_READ,
    cld_enq_flags = 0 
  },  
  cll_state = CLS_HELD,
  cll_wq = { 
    lock = { 
       raw_lock = { 
          slock = 169675293
       }   
    },  
    task_list = { 
      next = 0xffff88...
      prev = 0xffff88...
    }   
  },  
  cll_guard = { 
    count = { 
      counter = 1 
    },  
    wait_lock = { 
       raw_lock = { 
          slock = 1179666
       }   
    },  
    wait_list = { 
      next = 0xffff88...
      prev = 0xffff88...
    },  
    owner = 0x0 
  },  
  cll_guarder = 0x0,
  cll_depth = 0,
  cll_intransit_owner = 0x0,
  cll_error = 0,
  cll_holds = 1,
  cll_users = 1,
  cll_flags = 0,
  cll_inclosure = { 
      next = 0xffff88...
      prev = 0xffff88...
  },  
  cll_conflict = 0x0,
  cll_reference = {&amp;lt;No data fields&amp;gt;},
  cll_holders = {&amp;lt;No data fields&amp;gt;},
  cll_obj_ref = 0x0 
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="32139" author="jay" created="Mon, 26 Mar 2012 15:28:02 +0000"  >&lt;p&gt;let&apos;s try to print the lock state precisely with patch: &lt;a href=&quot;http://review.whamcloud.com/#change,2383&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2383&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="32143" author="nedbass" created="Mon, 26 Mar 2012 16:02:48 +0000"  >&lt;p&gt;OK, but it will be several weeks before we can update this system, and even then the bug may be rare to reproduce.&lt;/p&gt;</comment>
                            <comment id="33648" author="jay" created="Fri, 6 Apr 2012 01:55:13 +0000"  >&lt;p&gt;from the attached crash dump, the lock looks okay. I don&apos;t know why.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; p *(struct osc_lock *)0xffff881fc381a140
$28 = {
  ols_cl = {
    cls_lock = 0xffff881dea5c9e78, 
    cls_obj = 0xffff881f30d9e4e0, 
    cls_ops = 0xffffffffa06537e0, 
    cls_linkage = {
      next = 0xffff881dea5c9e80, 
      prev = 0xffff881e8c1e8938
    }
  }, 
  ols_lock = 0xffff881e974df240, 
  ols_lvb = {
    lvb_size = 1443889168, 
    lvb_mtime = 1329614555, 
    lvb_atime = 1329614548, 
    lvb_ctime = 1329614555, 
    lvb_blocks = 2820112
  }, 
  ols_flags = 1052673, 
  ols_handle = {
    cookie = 8399946539139540594
  }, 
  ols_einfo = {
    ei_type = 11, 
    ei_mode = 4, 
    ei_cb_bl = 0xffffffffa064ee10, 
    ei_cb_cp = 0xffffffffa064fcb0, 
    ei_cb_gl = 0xffffffffa064ea50, 
    ei_cb_wg = 0xffffffffa064e7a0, 
    ei_cbdata = 0xffff881fc381a140, 
    ei_async = 0
  }, 
  ols_state = OLS_GRANTED, 
  ols_pageref = {
    counter = 256
  }, 
  ols_hold = 1, 
  ols_has_ref = 1, 
  ols_locklessable = 0, 
  ols_ast_wait = 0, 
  ols_flush = 0, 
  ols_glimpse = 0, 
  ols_owner = 0xffff881aac5471d0
}
crash&amp;gt; p *(struct cl_lock *)0xffff881dea5c9e78
$29 = {
  cll_ref = {
    counter = 258
  }, 
  cll_layers = {
    next = 0xffff881e8c1e8938, 
    prev = 0xffff881fc381a158
  }, 
  cll_linkage = {
    next = 0xffff8819d4c0df38, 
    prev = 0xffff8819d4c0df38
  }, 
  cll_descr = {
    cld_obj = 0xffff8819d4c0df58, 
    cld_start = 0, 
    cld_end = 18446744073709551615, 
    cld_gid = 0, 
    cld_mode = CLM_READ, 
    cld_enq_flags = 0
  }, 
  cll_state = CLS_HELD, 
  cll_wq = {
    lock = {
      raw_lock = {
        slock = 76350605
      }
    }, 
    task_list = {
      next = 0xffff881dea5c9ed8, 
      prev = 0xffff881dea5c9ed8
    }
  }, 
  cll_guard = {
    count = {
      counter = 1
    }, 
    wait_lock = {
      raw_lock = {
        slock = 0
      }
    }, 
    wait_list = {
      next = 0xffff881dea5c9ef0, 
      prev = 0xffff881dea5c9ef0
    }, 
    owner = 0x0
  }, 
  cll_guarder = 0x0, 
  cll_depth = 0, 
  cll_intransit_owner = 0x0, 
  cll_error = 0, 
  cll_holds = 1, 
  cll_users = 1, 
  cll_flags = 0, 
  cll_inclosure = {
    next = 0xffff881dea5c9f38, 
    prev = 0xffff881dea5c9f38
  }, 
  cll_conflict = 0x0, 
  cll_reference = {&amp;lt;No data fields&amp;gt;}, 
  cll_holders = {&amp;lt;No data fields&amp;gt;}, 
  cll_obj_ref = 0x0
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="33652" author="green" created="Fri, 6 Apr 2012 02:13:01 +0000"  >&lt;p&gt;Ned, do you also apply multi-threaded ptlrpcd patch on top of your 2.1 tree?&lt;/p&gt;</comment>
                            <comment id="33814" author="morrone" created="Fri, 6 Apr 2012 14:05:22 +0000"  >&lt;p&gt;No, we do not have multi-threaded pltrpcd.&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://github.com/chaos/lustre/tree/2.1.0-llnl&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/chaos/lustre/tree/2.1.0-llnl&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="33838" author="jay" created="Fri, 6 Apr 2012 20:57:13 +0000"  >&lt;p&gt;I have no idea what happened. Can you please apply the patch I mentioned so that we can see what was exactly the lock state?&lt;/p&gt;</comment>
                            <comment id="33839" author="nedbass" created="Fri, 6 Apr 2012 21:04:44 +0000"  >&lt;p&gt;Hi Jinshan, we have applied the patch to our tree and it will by rolled out in the coming weeks.&lt;/p&gt;</comment>
                            <comment id="35425" author="morrone" created="Tue, 24 Apr 2012 18:12:53 +0000"  >&lt;p&gt;A sysadmin got a hit on a client on our secure network with your debugging patch applied.  He could not cut-and-paste.  What I know so far is just that your patch reported that the olock pointer is non-null and hold is 1.  If there is something specific that you want from CL_LOCK_DEBUG, let me know and I&apos;ll either have the admin transcribe it or a developer can do it when we return from LUG.&lt;/p&gt;</comment>
                            <comment id="35428" author="jay" created="Tue, 24 Apr 2012 19:38:24 +0000"  >&lt;p&gt;I will need the detailed info of the lock to address the problem.&lt;/p&gt;</comment>
                            <comment id="35441" author="morrone" created="Wed, 25 Apr 2012 10:13:36 +0000"  >&lt;p&gt;Ah, yes, I see all values that cl_lock_print includes.  We&apos;ll need Ned to copy those into this ticket when he gets back from LUG.&lt;/p&gt;</comment>
                            <comment id="35458" author="jay" created="Wed, 25 Apr 2012 14:19:58 +0000"  >&lt;p&gt;Yes, I knew this from your previous comment either. Let&apos;s do it after you guys come back from LUG.&lt;/p&gt;</comment>
                            <comment id="35529" author="nedbass" created="Thu, 26 Apr 2012 17:22:43 +0000"  >&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lock@ffff8803948bc978 [2 3 0 1 1 00000000] R(1):[0, 18446744073709551615]@[0x100cf000:0x3e138b:0x0] {
    lovsub@ffff8804340bd9a0 [0 ffff8804340db9a0: [ 0 ffff8801679d2d60 R(1):[256, 18446744073709551615]@[0x3f5c87bae:0x6a:0x0]]
    osc@ffff88041a6671f8: ffff880429e7eb40 00101001 0xa29f718631a3c48d 3 ffff8803b773f2b8 size: 367001600 mtime: 1335282921 atime: 1335282807 ctime: 1335282921 blocks: 716808
} lock@ffff8803948bc978
olock: ffff88041a6671f8, hold: 1.

PID: 14280 COMMAND: &quot;ptlrpcd-brw&quot;
machine_kexec
crash_kexec
panic
lbug_with_loc
libcfs_assertion_failed
osc_page_putref_lock
osc_page_completion_read
cl_page_completion
osc_completion
osc_ap_completion
brw_interpret
ptlrpc_check_set
ptlrpcd_check
ptlrpcd
kernel_thread
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="35532" author="jay" created="Thu, 26 Apr 2012 18:47:17 +0000"  >&lt;p&gt;Ned, thank you for the information.&lt;/p&gt;

&lt;p&gt;I realize this problem is a race between lock use and assertion check as following:&lt;br/&gt;
1. this page held the last refcount of the lock;&lt;br/&gt;
2. in osc_page_putref_lock(), cfs_atomic_read(&amp;amp;lock-&amp;gt;cll_ref) == 1 was true, but it was delayed for a while before checking ols_hold due to interrupt...&lt;br/&gt;
3. another proc calls cl_use() to bring this lock into CLS_HELD state, so ols_hold is set;&lt;br/&gt;
4. osc_page_putref_lock() continued to check ols_hold, definitely it was set and of course lock&apos;s refcount should be greater 1 for sure at this moment.&lt;/p&gt;

&lt;p&gt;Fix this problem is easy - just to remove the assertion.&lt;/p&gt;</comment>
                            <comment id="35533" author="jay" created="Thu, 26 Apr 2012 18:56:24 +0000"  >&lt;p&gt;Please apply patch: &lt;a href=&quot;http://review.whamcloud.com/2604&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/2604&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="35614" author="morrone" created="Fri, 27 Apr 2012 17:45:59 +0000"  >&lt;p&gt;Will do.&lt;/p&gt;</comment>
                            <comment id="39903" author="pjones" created="Mon, 4 Jun 2012 04:43:29 +0000"  >&lt;p&gt;Landed for 2.1.2 and 2.3&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv6mv:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4609</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>