<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:11:46 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7771] ZFS MDT running with high fragmentation</title>
                <link>https://jira.whamcloud.com/browse/LU-7771</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;After six months of production the LFS MDT (only one MDT, no DNE) is running a zpool fragmentation level of 86%.&lt;/p&gt;

&lt;p&gt;The customer has concerns about the high fragmentation.&lt;/p&gt;

&lt;p&gt;Q) Is there a long term risk, other than reduced performance, of running at a high fragmentation level?&lt;/p&gt;

&lt;p&gt;Also, the only way I know to clear the fragmentation is to quiesce the LFS, go offline, zfs send the MDT pool to a remote pool. Destroy and recreate the MDT pool and zfs send the remote pool back to local. &lt;/p&gt;

&lt;p&gt;Q) Is there another way of accomplishing this? &lt;/p&gt;

&lt;p&gt;Running the mdt:zfs_send-&amp;gt;remote_pool-destroy/recreate-remotepool:zfs_send_&amp;gt;new_mdt_pool process is akin to doing a block level dd backup of an ldiskfs MDT. I am all for the conservative path (with user data) but I want to inquire if there is a better path developed to reach the same result.&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;</description>
                <environment>CentOS 6.6, Lustre 2.7.64, ZFS 0.6.5.3</environment>
        <key id="34595">LU-7771</key>
            <summary>ZFS MDT running with high fragmentation</summary>
                <type id="9" iconUrl="https://jira.whamcloud.com/images/icons/issuetypes/undefined.png">Question/Request</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="6">Not a Bug</resolution>
                                        <assignee username="adilger">Andreas Dilger</assignee>
                                    <reporter username="aeonjeffj">Jeff Johnson</reporter>
                        <labels>
                            <label>zfs</label>
                    </labels>
                <created>Wed, 10 Feb 2016 00:47:35 +0000</created>
                <updated>Thu, 11 Feb 2016 12:21:56 +0000</updated>
                            <resolved>Thu, 11 Feb 2016 12:21:55 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="141813" author="aeonjeffj" created="Wed, 10 Feb 2016 19:27:58 +0000"  >&lt;p&gt;I&apos;ve come to learn that this is expected with ZFS MDT operation. The zfs send_recv process is the only way to remedy it and that other than performance degradation, less so with SSDs, it&apos;s not a data integrity risk.  &lt;/p&gt;

&lt;p&gt;Once large_dnode lands and gets tested to production safe levels this will subside.&lt;/p&gt;

&lt;p&gt;You can close this question.&lt;/p&gt;</comment>
                            <comment id="141821" author="adilger" created="Wed, 10 Feb 2016 19:48:27 +0000"  >&lt;p&gt;The ZFS &quot;fragmentation percentage&quot; is a bit misleading since this doesn&apos;t indicate the fragmentation of the file data but rather the fragmentation of free space, see &lt;a href=&quot;https://utcc.utoronto.ca/~cks/space/blog/solaris/ZFSZpoolFragmentationMeaning&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://utcc.utoronto.ca/~cks/space/blog/solaris/ZFSZpoolFragmentationMeaning&lt;/a&gt; for details.  A fragmentation score of 86% means the average free blocksize is about 32KB, though I&apos;m not sure if that is mean or median.&lt;/p&gt;

&lt;p&gt;IIRC, for the MDT the largest blocks it will ever allocate are 32KB (or possibly 64KB, don&apos;t recall offhand) for ZAP leaf blocks and interior tree blocks and dnode blocks are also 32KB so having a higher fragmentation score is expected for the MDT and not in itself a sign of problems.  With SSD storage for the MDT this is even less of a concern, though there is some overhead at the SSD level due to larger erase block size and write amplification.&lt;/p&gt;

&lt;p&gt;Is there a noticeable performance degradation that is observed, or is the concern only about the reported fragmentation percentage?  Doing the send/recv is only going to resolve this issue for a short time and will require an outage (though it could be minimized with some preparation), so I don&apos;t think it is necessarily a valuable exercise.  That said, I &lt;em&gt;would&lt;/em&gt; recommend regularly making and keeping the MDT backup with &lt;tt&gt;zfs send&lt;/tt&gt; on a periodic basis for disaster recovery purposes.&lt;/p&gt;</comment>
                            <comment id="141824" author="aeonjeffj" created="Wed, 10 Feb 2016 19:55:41 +0000"  >&lt;p&gt;I think this is a reaction to seeing the zpool report a large fragmentation percentage and the age old wisdom that fragmentation is bad. No performance issues reported and on SSDs the performance hit would be must less noticeable than with rotating disks.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                            <customfield id="customfield_10191" key="com.atlassian.jira.plugin.system.customfieldtypes:multiselect">
                        <customfieldname>IEEL Options</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10253"><![CDATA[Robin Hood]]></customfieldvalue>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy0vz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                                                        <customfield id="customfield_10190" key="com.atlassian.jira.plugin.system.customfieldtypes:textfield">
                        <customfieldname>Site Affected:</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>MSU (Michigan State Univ)</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                            </customfields>
    </item>
</channel>
</rss>