<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:04:27 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6923] writing process hung at txg_wait_open</title>
                <link>https://jira.whamcloud.com/browse/LU-6923</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;I saw an iozone writing process hung at waiting transaction to open. The backtrace is as follows:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[10459.403319] iozone          D ffff88082f353440     0  6714      1 0x00000004
[10459.403321]  ffff880085e83b28 0000000000000082 ffff88055b805180 ffff880085e83fd8
[10459.403323]  0000000000013440 0000000000013440 ffff880803713d20 ffff8807f3701368
[10459.403326]  ffff8807f3701220 ffff8807f3701370 ffff8807f3701248 0000000000000000
[10459.403328] Call Trace:
[10459.403330]  [&amp;lt;ffffffff81766c59&amp;gt;] schedule+0x29/0x70
[10459.403334]  [&amp;lt;ffffffffc09b4b45&amp;gt;] cv_wait_common+0xe5/0x120 [spl]
[10459.403337]  [&amp;lt;ffffffff810b4cd0&amp;gt;] ? prepare_to_wait_event+0x100/0x100
[10459.403341]  [&amp;lt;ffffffffc09b4b95&amp;gt;] __cv_wait+0x15/0x20 [spl]
[10459.403360]  [&amp;lt;ffffffffc0ab2f43&amp;gt;] txg_wait_open+0x83/0xd0 [zfs]
[10459.403379]  [&amp;lt;ffffffffc0a6e760&amp;gt;] dmu_tx_wait+0x380/0x390 [zfs]
[10459.403387]  [&amp;lt;ffffffff817690b2&amp;gt;] ? mutex_lock+0x12/0x2f
[10459.403406]  [&amp;lt;ffffffffc0a6e80a&amp;gt;] dmu_tx_assign+0x9a/0x510 [zfs]
[10459.403423]  [&amp;lt;ffffffffc0a5f69c&amp;gt;] dmu_free_long_range+0x18c/0x240 [zfs]
[10459.403449]  [&amp;lt;ffffffffc0ad266d&amp;gt;] zfs_rmnode+0x5d/0x340 [zfs]
[10459.403473]  [&amp;lt;ffffffffc0af2bb8&amp;gt;] zfs_zinactive+0x168/0x180 [zfs]
[10459.403494]  [&amp;lt;ffffffffc0aecc30&amp;gt;] zfs_inactive+0x60/0x200 [zfs]
[10459.403518]  [&amp;lt;ffffffffc0b034c3&amp;gt;] zpl_evict_inode+0x43/0x60 [zfs]
[10459.403521]  [&amp;lt;ffffffff811ee864&amp;gt;] evict+0xb4/0x180
[10459.403523]  [&amp;lt;ffffffff811ef045&amp;gt;] iput+0xf5/0x180
[10459.403525]  [&amp;lt;ffffffff811e3a83&amp;gt;] do_unlinkat+0x193/0x2c0
[10459.403528]  [&amp;lt;ffffffff811d8cbf&amp;gt;] ? SYSC_newstat+0x2f/0x40
[10459.403530]  [&amp;lt;ffffffff811e4b06&amp;gt;] SyS_unlink+0x16/0x20
[10459.403532]  [&amp;lt;ffffffff8176aced&amp;gt;] system_call_fastpath+0x1a/0x1f
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I will attach the backtrace of all running processes soon&lt;/p&gt;</description>
                <environment>ZFS</environment>
        <key id="31259">LU-6923</key>
            <summary>writing process hung at txg_wait_open</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jay">Jinshan Xiong</reporter>
                        <labels>
                            <label>zfs</label>
                    </labels>
                <created>Tue, 28 Jul 2015 19:44:13 +0000</created>
                <updated>Sat, 15 Dec 2018 18:07:06 +0000</updated>
                                                                                <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="122467" author="bzzz" created="Tue, 28 Jul 2015 19:47:25 +0000"  >&lt;p&gt;probably this will be fixed with &lt;a href=&quot;http://review.whamcloud.com/#/c/15481/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/15481/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="122482" author="jay" created="Tue, 28 Jul 2015 21:24:33 +0000"  >&lt;p&gt;Hi Alex,&lt;/p&gt;

&lt;p&gt;No, I was benchmarking ZFS master and there is no Lustre code involved.&lt;/p&gt;</comment>
                            <comment id="122483" author="jay" created="Tue, 28 Jul 2015 21:27:40 +0000"  >&lt;p&gt;txg_sync has been running for ages with the following backtrace:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[16932.253158] txg_sync        R  running task        0 26338      2 0x00000000
[16932.253166]  ffff8807b1c77c38 ffff8807f659a3e8 ffffc900086ad7f0 000000000ccd7dca
[16932.253177]  ffffffffc09adb20 ffffffffc09adb20 0000000000000000 000000000000c210
[16932.253185]  0000000000001000 0000000000000000 0000000000000000 ffff8807b1c77c68
[16932.253187] Call Trace:
[16932.253190]  [&amp;lt;ffffffff811b6cf9&amp;gt;] ? __kmalloc_node+0x1c9/0x2a0
[16932.253193]  [&amp;lt;ffffffffc09adb20&amp;gt;] ? spl_kmem_zalloc+0xc0/0x170 [spl]
[16932.253197]  [&amp;lt;ffffffffc09adb20&amp;gt;] ? spl_kmem_zalloc+0xc0/0x170 [spl]
[16932.253201]  [&amp;lt;ffffffffc09b4bd8&amp;gt;] ? __cv_wait_io+0x18/0x20 [spl]
[16932.253226]  [&amp;lt;ffffffffc0afbe83&amp;gt;] ? zio_wait+0x123/0x210 [zfs]
[16932.253240]  [&amp;lt;ffffffffc0a5d9da&amp;gt;] ? ddt_get_dedup_stats+0x3a/0x60 [zfs]
[16932.253243]  [&amp;lt;ffffffff8107c2ca&amp;gt;] ? mod_timer+0x12a/0x1e0
[16932.253260]  [&amp;lt;ffffffffc0a875b1&amp;gt;] ? dsl_pool_sync+0xb1/0x470 [zfs]
[16932.253267]  [&amp;lt;ffffffffc09ad76a&amp;gt;] ? spl_kmem_free+0x2a/0x40 [spl]
[16932.253292]  [&amp;lt;ffffffffc0aae346&amp;gt;] ? spa_update_dspace+0x26/0x40 [zfs]
[16932.253315]  [&amp;lt;ffffffffc0aa1d32&amp;gt;] ? spa_sync+0x3a2/0xb20 [zfs]
[16932.253321]  [&amp;lt;ffffffff810b4ce2&amp;gt;] ? autoremove_wake_function+0x12/0x40
[16932.253327]  [&amp;lt;ffffffff8101bdc9&amp;gt;] ? read_tsc+0x9/0x20
[16932.253350]  [&amp;lt;ffffffffc0ab3d8b&amp;gt;] ? txg_sync_thread+0x36b/0x630 [zfs]
[16932.253357]  [&amp;lt;ffffffff8101c3b9&amp;gt;] ? sched_clock+0x9/0x10
[16932.253380]  [&amp;lt;ffffffffc0ab3a20&amp;gt;] ? txg_quiesce_thread+0x380/0x380 [zfs]
[16932.253388]  [&amp;lt;ffffffffc09aff61&amp;gt;] ? thread_generic_wrapper+0x71/0x80 [spl]
[16932.253395]  [&amp;lt;ffffffffc09afef0&amp;gt;] ? __thread_exit+0x20/0x20 [spl]
[16932.253402]  [&amp;lt;ffffffff810912c2&amp;gt;] ? kthread+0xd2/0xf0
[16932.253408]  [&amp;lt;ffffffff810911f0&amp;gt;] ? kthread_create_on_node+0x1c0/0x1c0
[16932.253413]  [&amp;lt;ffffffff8176ac3c&amp;gt;] ? ret_from_fork+0x7c/0xb0
[16932.253415]  [&amp;lt;ffffffff810911f0&amp;gt;] ? kthread_create_on_node+0x1c0/0x1c0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; </comment>
                            <comment id="122485" author="jay" created="Tue, 28 Jul 2015 21:40:42 +0000"  >&lt;p&gt;filed a ticket at: &lt;a href=&quot;https://github.com/zfsonlinux/zfs/issues/3645&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/zfsonlinux/zfs/issues/3645&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="122488" author="dbrady" created="Tue, 28 Jul 2015 21:53:41 +0000"  >&lt;p&gt;Could you supply the config (zdb -C &amp;lt;poolname&amp;gt;) and any recent pool events (zpool events -v)?  Thanks.&lt;/p&gt;</comment>
                            <comment id="122527" author="bzzz" created="Wed, 29 Jul 2015 05:02:38 +0000"  >&lt;p&gt;sorry, my wording wasn&apos;t good enough. see the patch, it switches Lustre to use 0.6.4.2, which:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;Fix hard lockup due to infinite loop in zfs_zget() (zfsonlinux/zfs#3349)&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="122528" author="jay" created="Wed, 29 Jul 2015 05:08:54 +0000"  >&lt;p&gt;Hi Alex,&lt;/p&gt;

&lt;p&gt;If you&apos;re referring to commit 6102d037 for issue #3349, it has already been included in my code base for testing.&lt;/p&gt;

&lt;p&gt;Don,&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;jxiong@titan:~/work/zfs/zfs$ sudo zdb -C zfstest

MOS Configuration:
        version: 5000
        name: &apos;zfstest&apos;
        state: 0
        txg: 101
        pool_guid: 4724987114355670347
        errata: 0
        hostname: &apos;titan&apos;
        vdev_children: 1
        vdev_tree:
            type: &apos;root&apos;
            id: 0
            guid: 4724987114355670347
            children[0]:
                type: &apos;disk&apos;
                id: 0
                guid: 4902936745656769906
                path: &apos;/dev/nvme0n1p1&apos;
                whole_disk: 0
                metaslab_array: 34
                metaslab_shift: 29
                ashift: 9
                asize: 99993780224
                is_log: 0
                create_txg: 4
        features_for_read:
            com.delphix:hole_birth
            com.delphix:embedded_data
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="140489" author="yong.fan" created="Fri, 29 Jan 2016 05:36:18 +0000"  >&lt;p&gt;We hit similar trouble on lola as described in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7221&quot; title=&quot;replay-ost-single test_3: ASSERTION( __v &amp;gt; 0 &amp;amp;&amp;amp; __v &amp;lt; ((int)0x5a5a5a5a5a5a5a5a) ) failed: value: 0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7221&quot;&gt;&lt;del&gt;LU-7221&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="34340">LU-7721</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxj8f:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>