<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:27:37 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2720] osc_page_delete()) ASSERTION(0) failed</title>
                <link>https://jira.whamcloud.com/browse/LU-2720</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;it was already posted to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1723&quot; title=&quot;osc_page_delete()) ASSERTION(0) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1723&quot;&gt;&lt;del&gt;LU-1723&lt;/del&gt;&lt;/a&gt;, but seems to be a different issue from the original one in that ticket, so moving it here.&lt;/p&gt;

&lt;p&gt;2012-11-06T14:57:10.872333-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_cache.c:2367:osc_teardown_async_page()) extent ffff88060eedfe58@&lt;/p&gt;
{[23 -&amp;gt; 23/255], [2|0|-|cache|wi|ffff88020e18f8c8], [4096|1|+|-|ffff8801f9ed9c18|256| (null)]}
&lt;p&gt; trunc at 23.&lt;br/&gt;
2012-11-06T14:57:10.872389-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) page@ffff88080e8c8bc0&lt;span class=&quot;error&quot;&gt;&amp;#91;2 ffff88020cefad08:23 ^ (null)_ffff88080e8c8b00 4 0 1 (null) (null) 0x0&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:10.902445-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) page@ffff88080e8c8b00&lt;span class=&quot;error&quot;&gt;&amp;#91;1 ffff88020d8c1f58:23 ^ffff88080e8c8bc0_ (null) 4 0 1 (null) (null) 0x0&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:10.902494-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) vvp-page@ffff88080e8cf5a0(0:0:0) vm@ffffea001c6a99b8 e00000000000063 7:0 0 23 lru&lt;br/&gt;
2012-11-06T14:57:10.902508-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) lov-page@ffff8808050a7888&lt;br/&gt;
2012-11-06T14:57:10.958005-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) osc-page@ffff88080d78ed20: 1&amp;lt; 0x845fed 258 0 + - &amp;gt; 2&amp;lt; 94208 0 4096 0x0 0x520 | (null) ffff880804350700 ffff88020e18f8c8 &amp;gt; 3&amp;lt; + ffff8801f8f577b0 0 0 0 &amp;gt; 4&amp;lt; 0 0 8 33824768 - | - - + - &amp;gt; 5&amp;lt; - - + - | 0 - | 1 - -&amp;gt;&lt;br/&gt;
2012-11-06T14:57:10.958049-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) end page@ffff88080e8c8bc0&lt;br/&gt;
2012-11-06T14:57:10.983504-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:411:osc_page_delete()) Trying to teardown failed: -16&lt;br/&gt;
2012-11-06T14:57:10.983536-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:412:osc_page_delete()) ASSERTION( 0 ) failed:&lt;br/&gt;
2012-11-06T14:57:10.983549-06:00 c0-0c1s6n0 LustreError: 5270:0:(osc_page.c:412:osc_page_delete()) LBUG&lt;br/&gt;
2012-11-06T14:57:10.983570-06:00 c0-0c1s6n0 Pid: 5270, comm: fsx-linux&lt;br/&gt;
2012-11-06T14:57:10.983582-06:00 c0-0c1s6n0 Call Trace:&lt;br/&gt;
2012-11-06T14:57:10.983605-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff810063b1&amp;gt;&amp;#93;&lt;/span&gt; try_stack_unwind+0x161/0x1a0&lt;br/&gt;
2012-11-06T14:57:11.009114-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81004bf9&amp;gt;&amp;#93;&lt;/span&gt; dump_trace+0x89/0x440&lt;br/&gt;
2012-11-06T14:57:11.009138-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa014e887&amp;gt;&amp;#93;&lt;/span&gt; libcfs_debug_dumpstack+0x57/0x80 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.009161-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa014ede7&amp;gt;&amp;#93;&lt;/span&gt; lbug_with_loc+0x47/0xc0 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.009173-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0671d41&amp;gt;&amp;#93;&lt;/span&gt; osc_page_delete+0x2d1/0x2e0 &lt;span class=&quot;error&quot;&gt;&amp;#91;osc&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.034700-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa02b0095&amp;gt;&amp;#93;&lt;/span&gt; cl_page_delete0+0xd5/0x4e0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.034737-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa02b04e2&amp;gt;&amp;#93;&lt;/span&gt; cl_page_delete+0x42/0x120 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.034762-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07f2e2d&amp;gt;&amp;#93;&lt;/span&gt; ll_invalidatepage+0x8d/0x170 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.034774-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07ea290&amp;gt;&amp;#93;&lt;/span&gt; ll_page_mkwrite+0x7c0/0x840 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
2012-11-06T14:57:11.034820-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81107cb7&amp;gt;&amp;#93;&lt;/span&gt; __do_fault+0xe7/0x570&lt;br/&gt;
2012-11-06T14:57:11.034833-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811081e4&amp;gt;&amp;#93;&lt;/span&gt; handle_pte_fault+0xa4/0xcd0&lt;br/&gt;
2012-11-06T14:57:11.060561-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81108fbe&amp;gt;&amp;#93;&lt;/span&gt; handle_mm_fault+0x1ae/0x240&lt;br/&gt;
2012-11-06T14:57:11.060588-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81025471&amp;gt;&amp;#93;&lt;/span&gt; do_page_fault+0x191/0x410&lt;br/&gt;
2012-11-06T14:57:11.060600-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81301b5f&amp;gt;&amp;#93;&lt;/span&gt; page_fault+0x1f/0x30&lt;br/&gt;
2012-11-06T14:57:11.060647-06:00 c0-0c1s6n0 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;00000000200422b3&amp;gt;&amp;#93;&lt;/span&gt; 0x200422b3&lt;br/&gt;
2012-11-06T14:57:11.060660-06:00 c0-0c1s6n0 Kernel panic - not syncing: LBUG&lt;/p&gt;</description>
                <environment></environment>
        <key id="17379">LU-2720</key>
            <summary>osc_page_delete()) ASSERTION(0) failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="keith">Keith Mannthey</assignee>
                                    <reporter username="vitaly_fertman">Vitaly Fertman</reporter>
                        <labels>
                            <label>LB</label>
                            <label>patch</label>
                    </labels>
                <created>Thu, 31 Jan 2013 07:35:38 +0000</created>
                <updated>Fri, 15 Apr 2016 17:19:11 +0000</updated>
                            <resolved>Sat, 23 Feb 2013 14:28:52 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                    <fixVersion>Lustre 2.4.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="51523" author="vitaly_fertman" created="Thu, 31 Jan 2013 07:41:58 +0000"  >&lt;p&gt;1. The ENODATA handling code in ll_page_mkwrite0 writes the page and invalidates it, but the page could have been re-added to the cache in between these 2 steps.&lt;br/&gt;
2. lock could be busy, so we get ENODATA on trylock, whereas a racy ll_fault waits for lock mutex and re-adds the page to the cache;&lt;/p&gt;


&lt;p&gt;At the same time, it is not clear why we wanted to just PEEK the lock, not waiting here for a new one. we want a writable page, thus we need a lock. if old one is cancelled, we will have to request a new one anyway. i.e.&lt;br/&gt;
1. if to be cancelled and to request a new lock, the new one will be granted only after canceling the old one on ost. thus, page will not get under old lock and will not be written, won&apos;t it ? &lt;/p&gt;

&lt;p&gt;2. whereas mkwrite will finish must faster with no new lock request, later we still want to make the page writable and anyway we have to request a new lock - so not faster in general, isn&apos;t it?&lt;/p&gt;

&lt;p&gt;so the PEEK &quot;optimisation&quot; is not clear and troublesome, so the patch I send drops it.&lt;/p&gt;</comment>
                            <comment id="51528" author="vitaly_fertman" created="Thu, 31 Jan 2013 07:53:42 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/5222&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5222&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="51813" author="adilger" created="Tue, 5 Feb 2013 15:53:07 +0000"  >&lt;p&gt;Vitaly, can you please include some information about how this problem was initially hit (e.g. test load, frequency of being hit, etc).&lt;/p&gt;</comment>
                            <comment id="51981" author="wang" created="Thu, 7 Feb 2013 12:33:36 +0000"  >&lt;p&gt;We hit this often when running fsx-linux from LTP. Usually it happens during a stress run in an hour or two. We haven&apos;t seen this bug after applying this patch together with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2722&quot; title=&quot;LustreError: 30217:0:(rw26.c:485:ll_direct_IO_26()) ASSERTION( obj-&amp;gt;cob_transient_pages == 0 ) failed:&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2722&quot;&gt;&lt;del&gt;LU-2722&lt;/del&gt;&lt;/a&gt;/2723 patches fixing the DIO issues.&lt;/p&gt;</comment>
                            <comment id="52041" author="vitaly_fertman" created="Fri, 8 Feb 2013 12:31:58 +0000"  >&lt;p&gt;after a talk to Jay, we decided not to change cl_lock_peek, because it may return REPEAT  only for lock canceling or glimpse ast. at the same time, lock canceling may be long and we do not sleep here, so this looping will consume CPU resources. as after this patch, cl_lock_peek is used for SOM only, it may result the only ioepoch holder does not provide attibute update in done_writing and mds re-asks for them. if there will be a need in minimizing the amount of these RPCs, a sleeping version of cl_lock_peek is to be implemented.&lt;/p&gt;</comment>
                            <comment id="52050" author="keith" created="Fri, 8 Feb 2013 13:46:30 +0000"  >&lt;p&gt;Are you going to drop &lt;a href=&quot;http://review.whamcloud.com/5222&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5222&lt;/a&gt; ?  &lt;/p&gt;</comment>
                            <comment id="52929" author="pjones" created="Sat, 23 Feb 2013 14:28:52 +0000"  >&lt;p&gt;Landed for 2.4&lt;/p&gt;</comment>
                            <comment id="53470" author="spitzcor" created="Wed, 6 Mar 2013 14:24:05 +0000"  >&lt;p&gt;Can someone please comment about change #5222 landing considering the comments and questions from 08/Feb/13?  Cray is seeing significant CPU spinning in 2.4 RC testing, but Wally would have to confirm if change #5222 is the cause.&lt;/p&gt;</comment>
                            <comment id="53478" author="green" created="Wed, 6 Mar 2013 16:45:12 +0000"  >&lt;p&gt;Vitaly comments relates  to patchset #2, he since added patchset #3 that fixes the potential issue discussed.&lt;br/&gt;
As for the cpu spinning you are seeing, we need more details to make any educated guess there.&lt;/p&gt;</comment>
                            <comment id="53480" author="spitzcor" created="Wed, 6 Mar 2013 16:55:13 +0000"  >&lt;p&gt;Ah, thanks Oleg.  Sure, we&apos;ll keep looking at our spin issue.&lt;/p&gt;</comment>
                            <comment id="53482" author="pjones" created="Wed, 6 Mar 2013 16:57:35 +0000"  >&lt;p&gt;Corey&lt;/p&gt;

&lt;p&gt;It would be best to open a new ticket with details of the CPU spinning issue that you are seeing&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="57949" author="spitzcor" created="Wed, 8 May 2013 21:30:41 +0000"  >&lt;p&gt;Thanks, it is &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3217&quot; title=&quot;Client hung in cl_io_loop/cl_io_lock path&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3217&quot;&gt;&lt;del&gt;LU-3217&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="17062">LU-2557</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                            <subtask id="17519">LU-2788</subtask>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzviav:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>6614</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>