<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:43:38 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-11411] Lustre/ZFS snapshots mount error from llog - enhancement of snapshot-mount logic</title>
                <link>https://jira.whamcloud.com/browse/LU-11411</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;(creating an LU based on email traffic, see&#160;&lt;a href=&quot;http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/2018-September/015898.html&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/2018-September/015898.html&lt;/a&gt;)&lt;/p&gt;

&lt;p&gt;We have two filesystems, fsA &amp;amp; fsB (eadc below). Both of which get snapshots taken daily, rotated over a week. It&apos;s a beautiful feature we&apos;ve been using in production ever since it was introduced with 2.10.&lt;/p&gt;

&lt;p&gt;-) We&apos;ve got Lustre/ZFS 2.10.4 on CentOS 7.5.&lt;br/&gt;
 -) Both fsA &amp;amp; fsB have changelogs active.&lt;br/&gt;
 -) fsA has combined mgt/mdt on a single ZFS filesystem.&lt;br/&gt;
 -) fsB has a single mdt on a single ZFS filesystem.&lt;br/&gt;
 -) for fsA, I have no issues mounting any of the snapshots via lctl.&lt;br/&gt;
 -) for fsB, I can mount the most three recent snapshots, then encounter errors:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt; [root at hpfs-fsl-mds0 ~]# lctl snapshot_mount -F eadc -n eadc_AutoSS-Mon
 mounted the snapshot eadc_AutoSS-Mon with fsname 3d40bbc
 [root at hpfs-fsl-mds0 ~]# lctl snapshot_umount -F eadc -n
 eadc_AutoSS-Mon
 [root at hpfs-fsl-mds0 ~]# lctl snapshot_mount -F eadc -n eadc_AutoSS-Sun
 mounted the snapshot eadc_AutoSS-Sun with fsname 584c07a
 [root at hpfs-fsl-mds0 ~]# lctl snapshot_umount -F eadc -n
 eadc_AutoSS-Sun
 [root at hpfs-fsl-mds0 ~]# lctl snapshot_mount -F eadc -n eadc_AutoSS-Sat
 mounted the snapshot eadc_AutoSS-Sat with fsname 4e646fe
 [root at hpfs-fsl-mds0 ~]# lctl snapshot_umount -F eadc -n
 eadc_AutoSS-Sat
 [root at hpfs-fsl-mds0 ~]# lctl snapshot_mount -F eadc -n eadc_AutoSS-Fri
 mount.lustre: mount metadata/meta-eadc at eadc_AutoSS-Fri at
 /mnt/eadc_AutoSS-Fri_MDT0000 failed: Read-only file system Can&apos;t mount
 the snapshot eadc_AutoSS-Fri: Read-only file system


&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;The relevant bits from dmesg are&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt; [1353434.417762] Lustre: 3d40bbc-MDT0000: set dev_rdonly on this
 device [1353434.417765] Lustre: Skipped 3 previous similar messages
 [1353434.649480] Lustre: 3d40bbc-MDT0000: Imperative Recovery enabled,
 recovery window shrunk from 300-900 down to 150-900 [1353434.649484]
 Lustre: Skipped 3 previous similar messages [1353434.866228] Lustre:
 3d40bbc-MDD0000: changelog on [1353434.866233] Lustre: Skipped 1
 previous similar message [1353435.427744] Lustre: 3d40bbc-MDT0000:
 Connection restored to ... at tcp&amp;lt;[!/images/icons/mail_small.gif|width=13,height=12,align=absmiddle!|https://jira.whamcloud.com/secure/mailto:]... at tcp&amp;gt; (at ... at tcp&amp;lt;[!/images/icons/mail_small.gif|width=13,height=12,align=absmiddle!|https://jira.whamcloud.com/secure/mailto:]... at tcp&amp;gt;) [1353435.427747] Lustre:
 Skipped 23 previous similar messages [1353445.255899] Lustre: Failing
 over 3d40bbc-MDT0000 [1353445.255903] Lustre: Skipped 3 previous
 similar messages [1353445.256150] LustreError: 11-0:
 3d40bbc-OST0000-osc-MDT0000: operation ost_disconnect to node ... at tcp&amp;lt;[!/images/icons/mail_small.gif|width=13,height=12,align=absmiddle!|https://jira.whamcloud.com/secure/mailto:]... at tcp&amp;gt;
 failed: rc = -107 [1353445.257896] LustreError: Skipped 23 previous
 similar messages [1353445.353874] Lustre: server umount
 3d40bbc-MDT0000 complete [1353445.353877] Lustre: Skipped 3 previous
 similar messages [1353475.302224] Lustre: 4e646fe-MDD0000: changelog
 on [1353475.302228] Lustre: Skipped 1 previous similar message [1353498.964016] LustreError: 25582:0:(osd_handler.c:341:osd_trans_create()) 36ca26b-MDT0000-osd: someone try to start transaction under readonly mode, should be disabled.
 [1353498.967260] LustreError: 25582:0:(osd_handler.c:341:osd_trans_create()) Skipped 1 previous similar message
 [1353498.968829] CPU: 6 PID: 25582 Comm: mount.lustre Kdump: loaded Tainted: P OE ------------ 3.10.0-862.6.3.el7.x86_64 #1
 [1353498.968830] Hardware name: Supermicro SYS-6027TR-D71FRF/X9DRT,
 BIOS 3.2a 08/04/2015 [1353498.968832] Call Trace:
 [1353498.968841] [&amp;lt;ffffffffb5b0e80e&amp;gt;] dump_stack+0x19/0x1b
 [1353498.968851] [&amp;lt;ffffffffc0cbe5db&amp;gt;] osd_trans_create+0x38b/0x3d0
 [osd_zfs] [1353498.968876] [&amp;lt;ffffffffc1116044&amp;gt;]
 llog_destroy+0x1f4/0x3f0 [obdclass] [1353498.968887]
 [&amp;lt;ffffffffc111f0f6&amp;gt;] llog_cat_reverse_process_cb+0x246/0x3f0
 [obdclass] [1353498.968897] [&amp;lt;ffffffffc111a32c&amp;gt;]
 llog_reverse_process+0x38c/0xaa0 [obdclass] [1353498.968910]
 [&amp;lt;ffffffffc111eeb0&amp;gt;] ? llog_cat_process_cb+0x4e0/0x4e0 [obdclass]
 [1353498.968922] [&amp;lt;ffffffffc111af69&amp;gt;]
 llog_cat_reverse_process+0x179/0x270 [obdclass] [1353498.968932]
 [&amp;lt;ffffffffc1115585&amp;gt;] ? llog_init_handle+0xd5/0x9a0 [obdclass]
 [1353498.968943] [&amp;lt;ffffffffc1116e78&amp;gt;] ? llog_open_create+0x78/0x320
 [obdclass] [1353498.968949] [&amp;lt;ffffffffc12e55f0&amp;gt;] ?
 mdd_root_get+0xf0/0xf0 [mdd] [1353498.968954] [&amp;lt;ffffffffc12ec7af&amp;gt;]
 mdd_prepare+0x13ff/0x1c70 [mdd] [1353498.968966] [&amp;lt;ffffffffc166b037&amp;gt;]
 mdt_prepare+0x57/0x3b0 [mdt] [1353498.968983] [&amp;lt;ffffffffc1183afd&amp;gt;]
 server_start_targets+0x234d/0x2bd0 [obdclass] [1353498.968999]
 [&amp;lt;ffffffffc1153500&amp;gt;] ? class_config_dump_handler+0x7e0/0x7e0
 [obdclass] [1353498.969012] [&amp;lt;ffffffffc118541d&amp;gt;]
 server_fill_super+0x109d/0x185a [obdclass] [1353498.969025]
 [&amp;lt;ffffffffc115cef8&amp;gt;] lustre_fill_super+0x328/0x950 [obdclass]
 [1353498.969038] [&amp;lt;ffffffffc115cbd0&amp;gt;] ?
 lustre_common_put_super+0x270/0x270 [obdclass] [1353498.969041]
 [&amp;lt;ffffffffb561f3bf&amp;gt;] mount_nodev+0x4f/0xb0 [1353498.969053]
 [&amp;lt;ffffffffc1154f18&amp;gt;] lustre_mount+0x38/0x60 [obdclass]
 [1353498.969055] [&amp;lt;ffffffffb561ff3e&amp;gt;] mount_fs+0x3e/0x1b0 [1353498.969060] [&amp;lt;ffffffffb563d4b7&amp;gt;] vfs_kern_mount+0x67/0x110 [1353498.969062] [&amp;lt;ffffffffb563fadf&amp;gt;] do_mount+0x1ef/0xce0 [1353498.969066] [&amp;lt;ffffffffb55f7c2c&amp;gt;] ? kmem_cache_alloc_trace+0x3c/0x200 [1353498.969069] [&amp;lt;ffffffffb5640913&amp;gt;] SyS_mount+0x83/0xd0 [1353498.969074] [&amp;lt;ffffffffb5b20795&amp;gt;] system_call_fastpath+0x1c/0x21 [1353498.969079] LustreError: 25582:0:(llog_cat.c:1027:llog_cat_reverse_process_cb()) 36ca26b-MDD0000: fail to destroy empty log: rc = -30
 [1353498.970785] CPU: 6 PID: 25582 Comm: mount.lustre Kdump: loaded Tainted: P OE ------------ 3.10.0-862.6.3.el7.x86_64 #1
 [1353498.970786] Hardware name: Supermicro SYS-6027TR-D71FRF/X9DRT,
 BIOS 3.2a 08/04/2015 [1353498.970787] Call Trace:
 [1353498.970790] [&amp;lt;ffffffffb5b0e80e&amp;gt;] dump_stack+0x19/0x1b
 [1353498.970795] [&amp;lt;ffffffffc0cbe5db&amp;gt;] osd_trans_create+0x38b/0x3d0
 [osd_zfs] [1353498.970807] [&amp;lt;ffffffffc1117921&amp;gt;]
 llog_cancel_rec+0xc1/0x880 [obdclass] [1353498.970817]
 [&amp;lt;ffffffffc111e13b&amp;gt;] llog_cat_cleanup+0xdb/0x380 [obdclass]
 [1353498.970827] [&amp;lt;ffffffffc111f14d&amp;gt;]
 llog_cat_reverse_process_cb+0x29d/0x3f0 [obdclass] [1353498.970838]
 [&amp;lt;ffffffffc111a32c&amp;gt;] llog_reverse_process+0x38c/0xaa0 [obdclass]
 [1353498.970848] [&amp;lt;ffffffffc111eeb0&amp;gt;] ?
 llog_cat_process_cb+0x4e0/0x4e0 [obdclass] [1353498.970858]
 [&amp;lt;ffffffffc111af69&amp;gt;] llog_cat_reverse_process+0x179/0x270 [obdclass]
 [1353498.970868] [&amp;lt;ffffffffc1115585&amp;gt;] ? llog_init_handle+0xd5/0x9a0
 [obdclass] [1353498.970878] [&amp;lt;ffffffffc1116e78&amp;gt;] ?
 llog_open_create+0x78/0x320 [obdclass] [1353498.970883]
 [&amp;lt;ffffffffc12e55f0&amp;gt;] ? mdd_root_get+0xf0/0xf0 [mdd] [1353498.970887]
 [&amp;lt;ffffffffc12ec7af&amp;gt;] mdd_prepare+0x13ff/0x1c70 [mdd] [1353498.970894]
 [&amp;lt;ffffffffc166b037&amp;gt;] mdt_prepare+0x57/0x3b0 [mdt] [1353498.970908]
 [&amp;lt;ffffffffc1183afd&amp;gt;] server_start_targets+0x234d/0x2bd0 [obdclass]
 [1353498.970924] [&amp;lt;ffffffffc1153500&amp;gt;] ?
 class_config_dump_handler+0x7e0/0x7e0 [obdclass] [1353498.970938]
 [&amp;lt;ffffffffc118541d&amp;gt;] server_fill_super+0x109d/0x185a [obdclass]
 [1353498.970950] [&amp;lt;ffffffffc115cef8&amp;gt;] lustre_fill_super+0x328/0x950
 [obdclass] [1353498.970962] [&amp;lt;ffffffffc115cbd0&amp;gt;] ?
 lustre_common_put_super+0x270/0x270 [obdclass] [1353498.970964]
 [&amp;lt;ffffffffb561f3bf&amp;gt;] mount_nodev+0x4f/0xb0 [1353498.970976]
 [&amp;lt;ffffffffc1154f18&amp;gt;] lustre_mount+0x38/0x60 [obdclass]
 [1353498.970978] [&amp;lt;ffffffffb561ff3e&amp;gt;] mount_fs+0x3e/0x1b0
 [1353498.970980] [&amp;lt;ffffffffb563d4b7&amp;gt;] vfs_kern_mount+0x67/0x110
 [1353498.970982] [&amp;lt;ffffffffb563fadf&amp;gt;] do_mount+0x1ef/0xce0
 [1353498.970984] [&amp;lt;ffffffffb55f7c2c&amp;gt;] ?
 kmem_cache_alloc_trace+0x3c/0x200 [1353498.970986]
 [&amp;lt;ffffffffb5640913&amp;gt;] SyS_mount+0x83/0xd0 [1353498.970989]
 [&amp;lt;ffffffffb5b20795&amp;gt;] system_call_fastpath+0x1c/0x21 [1353498.970996]
 LustreError: 25582:0:(mdd_device.c:354:mdd_changelog_llog_init())
 36ca26b-MDD0000: changelog init failed: rc = -30 [1353498.972790]
 LustreError: 25582:0:(mdd_device.c:427:mdd_changelog_init())
 36ca26b-MDD0000: changelog setup during init failed: rc = -30
 [1353498.974525] LustreError:
 25582:0:(mdd_device.c:1061:mdd_prepare()) 36ca26b-MDD0000: failed to
 initialize changelog: rc = -30 [1353498.976229] LustreError:
 25582:0:(obd_mount_server.c:1879:server_fill_super()) Unable to start
 targets: -30 [1353499.072002] LustreError:
 25582:0:(obd_mount.c:1582:lustre_fill_super()) Unable to mount (-30)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;I&apos;m hoping those traces mean something to someone - any ideas?&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;</description>
                <environment>CentOS 7.5, triple homed Ethernet/FDR/EDR servers</environment>
        <key id="53376">LU-11411</key>
            <summary>Lustre/ZFS snapshots mount error from llog - enhancement of snapshot-mount logic</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="bkirk">Benjamin Kirk</reporter>
                        <labels>
                    </labels>
                <created>Thu, 20 Sep 2018 18:51:07 +0000</created>
                <updated>Tue, 25 Sep 2018 21:37:45 +0000</updated>
                            <resolved>Tue, 25 Sep 2018 21:37:45 +0000</resolved>
                                    <version>Lustre 2.10.4</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>2</watches>
                                                                            <comments>
                            <comment id="233828" author="utopiabound" created="Thu, 20 Sep 2018 20:01:50 +0000"  >&lt;p&gt;I believe this is a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-11193&quot; title=&quot;lsnapshot mount fails with DNE&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-11193&quot;&gt;&lt;del&gt;LU-11193&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The ZFS system has to have DNE enabled for the mount of the snapshot to fail.&#160;&lt;/p&gt;

&lt;p&gt;There is a patch that appears to fix this issue:&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://review.whamcloud.com/33157&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/33157&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="233830" author="bkirk" created="Thu, 20 Sep 2018 20:17:47 +0000"  >&lt;p&gt;We only have a single MDT on each filesystem; we just happen to have two separate filesystems hosted on the same servers.&#160; So it&apos;s not clear to me we have DNE in the equation.&lt;/p&gt;</comment>
                            <comment id="233860" author="utopiabound" created="Fri, 21 Sep 2018 12:50:10 +0000"  >&lt;p&gt;Okay.&#160; It&apos;s the same code path.&#160; I guess one of your snapshots had llog data that needed to be cleaned up.&#160; DNE always has llog data to cleanup.&lt;/p&gt;</comment>
                            <comment id="233862" author="bkirk" created="Fri, 21 Sep 2018 13:11:50 +0000"  >&lt;p&gt;Ahh good. I&#8217;m glad it is something you can repeat in your test environment!&lt;/p&gt;</comment>
                            <comment id="233985" author="bkirk" created="Tue, 25 Sep 2018 18:36:39 +0000"  >&lt;p&gt;I can confirm that the referenced patch (&lt;a href=&quot;https://review.whamcloud.com/33157&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/33157)&lt;/a&gt;&#160;on top of 2.10.5 allows me to mount all 11 snapshots from fsA and all 7 from fsB.&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="233988" author="adilger" created="Tue, 25 Sep 2018 21:37:45 +0000"  >&lt;p&gt;Close as a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-11193&quot; title=&quot;lsnapshot mount fails with DNE&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-11193&quot;&gt;&lt;del&gt;LU-11193&lt;/del&gt;&lt;/a&gt;, whiuch already has a patch.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="52858">LU-11193</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                    <customfield id="customfield_10030" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic/Theme</customfieldname>
                        <customfieldvalues>
                                        <label>zfs</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i002tj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>