<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:05:42 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-302] ll_ost_io_* threads hung</title>
                <link>https://jira.whamcloud.com/browse/LU-302</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;While running runtests test, the ll_ost_io_* threads hung as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: DEBUG MARKER: copying files from /etc /bin to /mnt/lustre/runtest.5368/etc /bin at Tue May 10 02:14:07 PDT 2011
Lustre: Service thread pid 6575 was inactive for 40.00s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
Pid: 6575, comm: ll_ost_io_03

Call Trace:
 [&amp;lt;ffffffff8006466c&amp;gt;] __down_read+0x7a/0x92
 [&amp;lt;ffffffff88bb2b0f&amp;gt;] ldiskfs_ext_walk_space+0xdf/0x2d0 [ldiskfs]
 [&amp;lt;ffffffff88c0bf10&amp;gt;] ldiskfs_ext_new_extent_cb+0x0/0x650 [fsfilt_ldiskfs]
 [&amp;lt;ffffffff8006456b&amp;gt;] __down_write_nested+0x12/0x92
 [&amp;lt;ffffffff88c0846d&amp;gt;] fsfilt_map_nblocks+0xfd/0x150 [fsfilt_ldiskfs]
 [&amp;lt;ffffffff88c69a7d&amp;gt;] filter_direct_io+0x46d/0xd50 [obdfilter]
 [&amp;lt;ffffffff88c08be7&amp;gt;] fsfilt_ldiskfs_setattr+0x1a7/0x250 [fsfilt_ldiskfs]
 [&amp;lt;ffffffff88c6c840&amp;gt;] filter_commitrw_write+0x1800/0x2be0 [obdfilter]
 [&amp;lt;ffffffff8005c33c&amp;gt;] cache_alloc_refill+0x106/0x186
 [&amp;lt;ffffffff88c24eed&amp;gt;] ost_checksum_bulk+0x37d/0x5a0 [ost]
 [&amp;lt;ffffffff88c2bd09&amp;gt;] ost_brw_write+0x1c99/0x2480 [ost]
 [&amp;lt;ffffffff8001aa2d&amp;gt;] vsnprintf+0x5df/0x627
 [&amp;lt;ffffffff88945f25&amp;gt;] lustre_msg_get_opc+0x35/0xf0 [ptlrpc]
 [&amp;lt;ffffffff889460d8&amp;gt;] lustre_msg_check_version_v2+0x8/0x20 [ptlrpc]
 [&amp;lt;ffffffff88c2f09e&amp;gt;] ost_handle+0x2bae/0x55b0 [ost]
 [&amp;lt;ffffffff8890019a&amp;gt;] lock_res_and_lock+0xba/0xd0 [ptlrpc]
 [&amp;lt;ffffffff887b4a87&amp;gt;] libcfs_next_nidstring+0x37/0x50 [libcfs]
 [&amp;lt;ffffffff889556f9&amp;gt;] ptlrpc_server_handle_request+0x989/0xe00 [ptlrpc]
 [&amp;lt;ffffffff88955e55&amp;gt;] ptlrpc_wait_event+0x2e5/0x310 [ptlrpc]
 [&amp;lt;ffffffff8008c86f&amp;gt;] __wake_up_common+0x3e/0x68
 [&amp;lt;ffffffff88956de6&amp;gt;] ptlrpc_main+0xf66/0x1120 [ptlrpc]
 [&amp;lt;ffffffff8005dfb1&amp;gt;] child_rip+0xa/0x11
 [&amp;lt;ffffffff88955e80&amp;gt;] ptlrpc_main+0x0/0x1120 [ptlrpc]
 [&amp;lt;ffffffff8005dfa7&amp;gt;] child_rip+0x0/0x11

LustreError: dumping log to /tmp/lustre-log.1305018888.6575
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Please refer to the following Maloo report for more logs:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/ecc7177c-7aec-11e0-b5bf-52540025f9af&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/ecc7177c-7aec-11e0-b5bf-52540025f9af&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The issue is blocking the testing on b1_8 branch.&lt;/p&gt;</description>
                <environment>Lustre Branch: b1_8&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://newbuild.whamcloud.com/job/lustre-reviews/363/arch=x86_64,build_type=server,distro=el5,ib_stack=inkernel/&quot;&gt;http://newbuild.whamcloud.com/job/lustre-reviews/363/arch=x86_64,build_type=server,distro=el5,ib_stack=inkernel/&lt;/a&gt;&lt;br/&gt;
Kernel Version: 2.6.18-238.9.1.el5_lustre.20110509050254</environment>
        <key id="10777">LU-302</key>
            <summary>ll_ost_io_* threads hung</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="ys">Yang Sheng</assignee>
                                    <reporter username="yujian">Jian Yu</reporter>
                        <labels>
                    </labels>
                <created>Tue, 10 May 2011 04:11:50 +0000</created>
                <updated>Wed, 11 May 2011 08:33:02 +0000</updated>
                            <resolved>Wed, 11 May 2011 08:33:02 +0000</resolved>
                                    <version>Lustre 1.8.6</version>
                                    <fixVersion>Lustre 1.8.6</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="14097" author="pjones" created="Tue, 10 May 2011 07:05:12 +0000"  >&lt;p&gt;Johann&lt;/p&gt;

&lt;p&gt;Yu Jian suspects that this issue may be due to a recent landing. Could you please have a quick look and then assign as necessary? This is blocking all 1.8.x testing atm&lt;/p&gt;

&lt;p&gt;Regards&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="14102" author="adilger" created="Tue, 10 May 2011 08:29:00 +0000"  >&lt;p&gt;This is probably the ext_walk_space locking change that hit on some of the newer kernels also, probably because the RHEL 5 kernel backported some change.&lt;/p&gt;

&lt;p&gt;I think we need a better configure check to determine whether ext_walk_space needs to be locked by the caller or internally. I proposed a way to do this using &quot;grep -A&quot; in the original RHEL6 bug in bugzilla that Kalpak was working on. &lt;/p&gt;</comment>
                            <comment id="14134" author="pjones" created="Tue, 10 May 2011 13:40:24 +0000"  >&lt;p&gt;YangSheng&lt;/p&gt;

&lt;p&gt;It seems that this is something that you have been working on&lt;/p&gt;

&lt;p&gt;Regards&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="14148" author="ys" created="Wed, 11 May 2011 06:07:11 +0000"  >&lt;p&gt;Looks REHL5.6 backport upstream commit fab3a549e204172236779f502eccb4f9bf0dc87d(ext4: Fix potential fiemap deadlock (mmap_sem vs. i_data_sem)). So we may need landed fix in &lt;a href=&quot;https://bugzilla.lustre.org/show_bug.cgi?id=23780&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://bugzilla.lustre.org/show_bug.cgi?id=23780&lt;/a&gt;. Original patch against SLES11 SP1. But as further solution. How about change down_read() to down_write() in ext4_ext_walk_space:&lt;br/&gt;
               /* find extent for this block */&lt;br/&gt;
               down_read(&amp;amp;EXT4_I(inode)-&amp;gt;i_data_sem);&lt;br/&gt;
               path = ext4_ext_find_extent(inode, block, path);&lt;br/&gt;
               up_read(&amp;amp;EXT4_I(inode)-&amp;gt;i_data_sem);&lt;br/&gt;
and trying to push this change to upstream. This change with WALK_SPACE_HAS_I_DATA_SEM config check. We can deal with this situation. &lt;/p&gt;</comment>
                            <comment id="14169" author="johann" created="Wed, 11 May 2011 08:15:42 +0000"  >&lt;p&gt;This should be addressed by &lt;a href=&quot;http://review.whamcloud.com/#change,491&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,491&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="14170" author="ys" created="Wed, 11 May 2011 08:33:02 +0000"  >&lt;p&gt;So just port lu-216 to b1_8 is enough to resolved this issue.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzw09j:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>10122</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>