<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:33:38 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3406] Submit raid5-mmp-unplug-dev patch upstream</title>
                <link>https://jira.whamcloud.com/browse/LU-3406</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;In order to submit the raid5-mmp-unplug-dev patch upstream, this needs to be updated for the latest kernel.  Unfortunately, the affected code seems to have changed since the patch was written, so I&apos;m not sure whether a simple &quot;best guess update&quot; of the patch will be correct.&lt;/p&gt;

&lt;p&gt;I&apos;ll attach my &quot;best guess&quot; patch, but it needs to be verified by someone who actually understands the MD RAID code better.&lt;/p&gt;</description>
                <environment></environment>
        <key id="19168">LU-3406</key>
            <summary>Submit raid5-mmp-unplug-dev patch upstream</summary>
                <type id="7" iconUrl="https://jira.whamcloud.com/images/icons/issuetypes/task_agile.png">Technical task</type>
                            <parent id="10111">LU-20</parent>
                                    <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="2">Won&apos;t Fix</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="adilger">Andreas Dilger</reporter>
                        <labels>
                    </labels>
                <created>Mon, 27 May 2013 19:57:47 +0000</created>
                <updated>Mon, 24 Apr 2017 21:37:40 +0000</updated>
                            <resolved>Mon, 24 Apr 2017 20:05:01 +0000</resolved>
                                    <version>Lustre 2.5.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="59394" author="adilger" created="Mon, 27 May 2013 19:59:55 +0000"  >&lt;p&gt;Prototype patch against current Linux Git HEAD.&lt;/p&gt;</comment>
                            <comment id="59431" author="simmonsja" created="Tue, 28 May 2013 14:45:09 +0000"  >&lt;p&gt;Since this patch is more a bug fix it might even make it into 3.10. Time to send it to dm-devel@redhat.com&lt;/p&gt;</comment>
                            <comment id="59508" author="adilger" created="Wed, 29 May 2013 08:55:34 +0000"  >&lt;p&gt;James, a problem I&apos;ve just noticed since I looked at this patch in detail for upstream submission is that overloading READ_SYNC/REQ_SYNC to mean &quot;bypass MD RAID cache&quot; is that this can negatively impact other filesystems such as Btrfs, OCFS, and others that now use READ_SYNC to distinguish reads that are being waited upon by processes from readahead requests.&lt;/p&gt;

&lt;p&gt;That is fine for Lustre server kernels, but isn&apos;t necessarily fine for general usage.  One option is to make a separate REQ_NOCACHE flag or similar, but that needs a much bigger patch that is only useful for ext4 MMP.&lt;/p&gt;

&lt;p&gt;Hopefully someone who understands the RAID and block layer details better can come up with a solution. &lt;/p&gt;</comment>
                            <comment id="59815" author="bfaccini" created="Fri, 31 May 2013 22:14:47 +0000"  >&lt;p&gt;Why not using _META flag for this purpose ?&lt;/p&gt;</comment>
                            <comment id="59840" author="adilger" created="Sat, 1 Jun 2013 14:24:13 +0000"  >&lt;p&gt;I don&apos;t think _META is less used than _SYNC. It is often used for metadata IO requests to increase the priority in the scheduler.  I think it is probably best to try with a new REQ flag and see what the upstream MD maintainer thinks.  This is a generic bug with ext4 MMP, and not Lustre specific, so there should be some kind of solution possible. &lt;/p&gt;</comment>
                            <comment id="60020" author="bfaccini" created="Wed, 5 Jun 2013 09:47:48 +0000"  >&lt;p&gt;Ok, so let&apos;s go for a new flag.&lt;br/&gt;
But having a look to the MD/Raid5 source code I am now concerned about the real need for md_wakeup_thread() call at the end of make_request() if flag is set, seems to me that it should have been already called within release_stripe&lt;span class=&quot;error&quot;&gt;&amp;#91;_plug&amp;#93;&lt;/span&gt;() call and underlying routines. Thus, even if ineffective it could be considered as useless and costly.&lt;/p&gt;
</comment>
                            <comment id="60039" author="bfaccini" created="Wed, 5 Jun 2013 16:32:36 +0000"  >&lt;p&gt;BTW, within current supported Kernels and current patch version, my earlier comment applies to md_raid5_unplug_device() instead of md_wakeup_thread(). And it was already in original patch submitted for BZ #17895 by Jinshan, may be he can help me to confirm if it is necessary or not.&lt;/p&gt;</comment>
                            <comment id="60348" author="bfaccini" created="Tue, 11 Jun 2013 14:31:58 +0000"  >&lt;p&gt;I spent some time digging in latest/3.9.4 kernel sources and I can confirm there are still no way to bypass the MD/Raid5 stripe-cache upon a read request.&lt;br/&gt;
I am 1st testing a patch (which introduces a new flag) against current Lustre-Server supported Kernel version and to be exposed under HA/mmp tests.&lt;/p&gt;</comment>
                            <comment id="63476" author="bfaccini" created="Thu, 1 Aug 2013 14:50:23 +0000"  >&lt;p&gt;Oops, forgot to indicate patch is at &lt;a href=&quot;http://review.whamcloud.com/6652&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/6652&lt;/a&gt;.&lt;br/&gt;
Also auto-tests never started for patch-set #1 for unexplained reasons, I submitted patch-set #2 with less restrictive test-parameters just in case.&lt;br/&gt;
On the other hand I am working on a local test platform + use-cases to ensure MMP work fine over SW-Raid.&lt;/p&gt;
</comment>
                            <comment id="88382" author="simmonsja" created="Mon, 7 Jul 2014 23:06:03 +0000"  >&lt;p&gt;Patch &lt;a href=&quot;http://review.whamcloud.com/6652&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/6652&lt;/a&gt; has been updated to latest master. If it proves stable we should look to pushing it upstream to get feedback to see what the final result is.&lt;/p&gt;</comment>
                            <comment id="88416" author="bfaccini" created="Tue, 8 Jul 2014 11:27:55 +0000"  >&lt;p&gt;James,&lt;br/&gt;
I need to apologize to have not updated this ticket, and associated change #6652 too, since months now, even if I have been assigned to higher priority tasks since... In fact what I have really forgotten is to already give a detailled update on where I was on this, so will try to do it now !&lt;/p&gt;

&lt;p&gt;After I had pushed patch-set #1 of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6652&quot; title=&quot;replay-dual test 18 statmany wrong file&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6652&quot;&gt;&lt;del&gt;LU-6652&lt;/del&gt;&lt;/a&gt;, I ran local tests on an ad-hoc HA platform to verify patch&apos;s functionality and correct behavior, but then I discovered that there was no debug trace generated out from raid456 module upon MMP block reads !!!! And this when there was during MMP block writes, which seems just impossible when reading both ext4/ldiskfs and md/raid5 source code! But also confirmed by iostat/blktrace monitoring.&lt;br/&gt;
This is the reason why next patch-sets 2-5 (don&apos;t remember why I removed the &quot;fortestonly&quot; param ...) are only adding more debug stuff/traces to help understand what&apos;s going-on ...&lt;/p&gt;

&lt;p&gt;BTW, at this time and before to give-up due to higher priorities..., I tried to verify the current/original patch behavior, and it exhibited tha same unexpected results.&lt;/p&gt;

&lt;p&gt;So here I was and still I am on this, so if you pursue in re-basing my patch What I strongly suggest is to again verify patch functionality/behavior at the lowest level.&lt;/p&gt;

</comment>
                            <comment id="98227" author="simmonsja" created="Mon, 3 Nov 2014 21:21:59 +0000"  >&lt;p&gt;I submitted a similar patch to dm-revel@redhat.com. You can see the message at &lt;a href=&quot;https://www.redhat.com/archives/dm-devel/2014-November/msg00004.html&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://www.redhat.com/archives/dm-devel/2014-November/msg00004.html&lt;/a&gt;. I like to see what the feedback is so we can develop a approach acceptable upstream and then back port it to supported distros.&lt;/p&gt;</comment>
                            <comment id="193287" author="simmonsja" created="Mon, 24 Apr 2017 21:19:53 +0000"  >&lt;p&gt;Since this will not be fixed I suggest we delete the current patches we carry. Another option is to cache the patches in contrib until some one wants to work on a version to get accepted upstream. Especially since upstream show potential corruption with our current patch.&lt;/p&gt;</comment>
                            <comment id="193288" author="adilger" created="Mon, 24 Apr 2017 21:37:40 +0000"  >&lt;p&gt;I think that makes sense.  Alternately, the patch could just be removed from the patch series files and left in &lt;tt&gt;kernel_patches&lt;/tt&gt; in case anyone wants to use it.  That would be easier to find than in the &lt;tt&gt;contrib&lt;/tt&gt; directory.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="10111">LU-20</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="12941" name="0001-md-submit-MMP-reads-REQ_SYNC-to-bypass-RAID5-cache.patch" size="1620" author="adilger" created="Mon, 27 May 2013 19:59:55 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                    <customfield id="customfield_10020" key="com.atlassian.jira.plugin.system.customfieldtypes:float">
                        <customfieldname>Bugzilla ID</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>17895.0</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvrxr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>8423</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                                                                                </customfields>
    </item>
</channel>
</rss>