<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:23:16 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2208] deadlock in add_lsmref</title>
                <link>https://jira.whamcloud.com/browse/LU-2208</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Hit this in sanity 118k&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[19970.051046] Lustre: DEBUG MARKER: == sanity test 118k: bio alloc -ENOMEM and IO TERM handling =========== 22:02:17 (1350525737)
[20160.612156] INFO: task flush-lustre-5:18802 blocked for more than 120 seconds.
[20160.612995] &quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.
[20160.613824] flush-lustre- D 0000000000000000  3504 18802      2 0x00000000
[20160.614395]  ffff880047a39958 0000000000000046 0000000000000000 ffff880047a39900
[20160.615163]  ffff880047a399a0 0000000100000000 0000000000000020 ffff88005f620c20
[20160.615529]  ffff880020730738 ffff880047a39fd8 000000000000fba8 ffff880020730738
[20160.615893] Call Trace:
[20160.616057]  [&amp;lt;ffffffff814fabbd&amp;gt;] rwsem_down_failed_common+0x8d/0x1d0
[20160.616272]  [&amp;lt;ffffffff814fad56&amp;gt;] rwsem_down_read_failed+0x26/0x30
[20160.616483]  [&amp;lt;ffffffff8127c104&amp;gt;] call_rwsem_down_read_failed+0x14/0x30
[20160.616718]  [&amp;lt;ffffffff814f9ec7&amp;gt;] ? down_read+0x37/0x40
[20160.616954]  [&amp;lt;ffffffffa1527b24&amp;gt;] lov_lsm_addref+0x34/0x150 [lov]
[20160.617186]  [&amp;lt;ffffffffa1528043&amp;gt;] lov_io_init+0x73/0x160 [lov]
[20160.617417]  [&amp;lt;ffffffffa10eb4e8&amp;gt;] cl_io_init0+0x98/0x160 [obdclass]
[20160.617647]  [&amp;lt;ffffffffa10ee2a4&amp;gt;] cl_io_init+0x64/0x100 [obdclass]
[20160.617869]  [&amp;lt;ffffffffa095e9ce&amp;gt;] cl_sync_file_range+0x11e/0x560 [lustre]
[20160.618793]  [&amp;lt;ffffffffa0984342&amp;gt;] ll_writepages+0x72/0x1b0 [lustre]
[20160.619027]  [&amp;lt;ffffffff81127f44&amp;gt;] do_writepages+0x24/0x40
[20160.619228]  [&amp;lt;ffffffff811a5594&amp;gt;] writeback_single_inode+0xe4/0x2d0
[20160.619439]  [&amp;lt;ffffffff811a5a13&amp;gt;] writeback_sb_inodes+0xd3/0x190
[20160.619646]  [&amp;lt;ffffffff811a5b4b&amp;gt;] writeback_inodes_wb+0x7b/0x1a0
[20160.619854]  [&amp;lt;ffffffff811a5efb&amp;gt;] wb_writeback+0x28b/0x3d0
[20160.620060]  [&amp;lt;ffffffff8107c6e2&amp;gt;] ? del_timer_sync+0x22/0x30
[20160.620266]  [&amp;lt;ffffffff811a61e5&amp;gt;] wb_do_writeback+0x1a5/0x250
[20160.620474]  [&amp;lt;ffffffff811a62f3&amp;gt;] bdi_writeback_task+0x63/0x1b0
[20160.620714]  [&amp;lt;ffffffff8108fc27&amp;gt;] ? bit_waitqueue+0x17/0xd0
[20160.620928]  [&amp;lt;ffffffff81136dc0&amp;gt;] ? bdi_start_fn+0x0/0x100
[20160.621172]  [&amp;lt;ffffffff81136e46&amp;gt;] bdi_start_fn+0x86/0x100
[20160.621396]  [&amp;lt;ffffffff81136dc0&amp;gt;] ? bdi_start_fn+0x0/0x100
[20160.621598]  [&amp;lt;ffffffff8108fa16&amp;gt;] kthread+0x96/0xa0
[20160.621790]  [&amp;lt;ffffffff8100c14a&amp;gt;] child_rip+0xa/0x20
[20160.621985]  [&amp;lt;ffffffff8108f980&amp;gt;] ? kthread+0x0/0xa0
[20160.622191]  [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20
[20160.622411] INFO: task dd:26845 blocked for more than 120 seconds.
[20160.622631] &quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.
[20160.622987] dd            D 0000000000000007  2608 26845      1 0x00000000
[20160.623220]  ffff88005bea3470 0000000000000086 0000000000000000 0000000000000286
[20160.623585]  0000000000000092 ffff88003b6a8548 000000015bea3458 ffffffff81f61808
[20160.623949]  ffff880046f00b78 ffff88005bea3fd8 000000000000fba8 ffff880046f00b78
[20160.624318] Call Trace:
[20160.624476]  [&amp;lt;ffffffff81044f4e&amp;gt;] ? kernel_map_pages+0xfe/0x110
[20160.624700]  [&amp;lt;ffffffff814fabbd&amp;gt;] rwsem_down_failed_common+0x8d/0x1d0
[20160.624978]  [&amp;lt;ffffffffa08d1afe&amp;gt;] ? cfs_mem_cache_free+0xe/0x10 [libcfs]
[20160.625319]  [&amp;lt;ffffffff814fad56&amp;gt;] rwsem_down_read_failed+0x26/0x30
[20160.625557]  [&amp;lt;ffffffff8127c104&amp;gt;] call_rwsem_down_read_failed+0x14/0x30
[20160.625774]  [&amp;lt;ffffffff814f9ec7&amp;gt;] ? down_read+0x37/0x40
[20160.625992]  [&amp;lt;ffffffffa1527b24&amp;gt;] lov_lsm_addref+0x34/0x150 [lov]
[20160.626238]  [&amp;lt;ffffffffa1528043&amp;gt;] lov_io_init+0x73/0x160 [lov]
[20160.626467]  [&amp;lt;ffffffffa10eb4e8&amp;gt;] cl_io_init0+0x98/0x160 [obdclass]
[20160.626690]  [&amp;lt;ffffffffa10ee2a4&amp;gt;] cl_io_init+0x64/0x100 [obdclass]
[20160.626922]  [&amp;lt;ffffffffa149b6d3&amp;gt;] osc_lru_shrink+0x4a3/0x8c0 [osc]
[20160.627143]  [&amp;lt;ffffffff8116133a&amp;gt;] ? cache_alloc_debugcheck_after+0x14a/0x210
[20160.627378]  [&amp;lt;ffffffffa149bfa8&amp;gt;] osc_page_init+0x4b8/0xb40 [osc]
[20160.627610]  [&amp;lt;ffffffffa10de7b5&amp;gt;] ? cl_page_slice_add+0x55/0x140 [obdclass]
[20160.627856]  [&amp;lt;ffffffffa10e2e7b&amp;gt;] cl_page_find0+0x2db/0x900 [obdclass]
[20160.628081]  [&amp;lt;ffffffff8116133a&amp;gt;] ? cache_alloc_debugcheck_after+0x14a/0x210
[20160.628327]  [&amp;lt;ffffffffa10e34b8&amp;gt;] cl_page_find_sub+0x18/0x20 [obdclass]
[20160.628575]  [&amp;lt;ffffffffa152a503&amp;gt;] lov_page_init_raid0+0x1a3/0x780 [lov]
[20160.628806]  [&amp;lt;ffffffffa1527f58&amp;gt;] lov_page_init+0x68/0xe0 [lov]
[20160.629038]  [&amp;lt;ffffffffa10e2e7b&amp;gt;] cl_page_find0+0x2db/0x900 [obdclass]
[20160.629268]  [&amp;lt;ffffffffa08e83e2&amp;gt;] ? cfs_hash_lookup+0x82/0xa0 [libcfs]
[20160.629492]  [&amp;lt;ffffffff811704f5&amp;gt;] ? mem_cgroup_charge_common+0xa5/0xd0
[20160.629728]  [&amp;lt;ffffffffa10e34d1&amp;gt;] cl_page_find+0x11/0x20 [obdclass]
[20160.629962]  [&amp;lt;ffffffffa0985ab4&amp;gt;] ll_cl_init+0x154/0x5b0 [lustre]
[20160.630190]  [&amp;lt;ffffffff814faeee&amp;gt;] ? _spin_unlock_irq+0xe/0x20
[20160.630430]  [&amp;lt;ffffffffa0986163&amp;gt;] ll_prepare_write+0x53/0x1a0 [lustre]
[20160.630672]  [&amp;lt;ffffffffa099dfbe&amp;gt;] ll_write_begin+0x7e/0x1a0 [lustre]
[20160.630898]  [&amp;lt;ffffffff81112d23&amp;gt;] generic_file_buffered_write+0x123/0x300
[20160.631130]  [&amp;lt;ffffffff8106fea7&amp;gt;] ? current_fs_time+0x27/0x30
[20160.631341]  [&amp;lt;ffffffff811147e0&amp;gt;] __generic_file_aio_write+0x250/0x480
[20160.631563]  [&amp;lt;ffffffff81114a7f&amp;gt;] generic_file_aio_write+0x6f/0xe0
[20160.631790]  [&amp;lt;ffffffffa09b171c&amp;gt;] vvp_io_write_start+0x9c/0x240 [lustre]
[20160.632034]  [&amp;lt;ffffffffa10eb26a&amp;gt;] cl_io_start+0x6a/0x140 [obdclass]
[20160.632285]  [&amp;lt;ffffffffa10efa54&amp;gt;] cl_io_loop+0xb4/0x1b0 [obdclass]
[20160.632529]  [&amp;lt;ffffffffa095e05b&amp;gt;] ll_file_io_generic+0x42b/0x550 [lustre]
[20160.632763]  [&amp;lt;ffffffffa095ef4c&amp;gt;] ll_file_aio_write+0x13c/0x2c0 [lustre]
[20160.632994]  [&amp;lt;ffffffffa095f239&amp;gt;] ll_file_write+0x169/0x2a0 [lustre]
[20160.633222]  [&amp;lt;ffffffff8117b2e8&amp;gt;] vfs_write+0xb8/0x1a0
[20160.633423]  [&amp;lt;ffffffff8117bbb1&amp;gt;] sys_write+0x51/0x90
[20160.633624]  [&amp;lt;ffffffff8100b0f2&amp;gt;] system_call_fastpath+0x16/0x1b
[20160.633837] INFO: task dd:26851 blocked for more than 120 seconds.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;There&apos;s a bunch of other dd processes hung in there.&lt;br/&gt;
Jinshan, I am leaving centos6-1 node in this state for you to look when you have time. Feel free to crash it with echo c &amp;gt;/proc/sysrq-trigger if you feel you would benefit from using crash to further investigate it that way.&lt;/p&gt;</description>
                <environment></environment>
        <key id="16396">LU-2208</key>
            <summary>deadlock in add_lsmref</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="jay">Jinshan Xiong</assignee>
                                    <reporter username="green">Oleg Drokin</reporter>
                        <labels>
                    </labels>
                <created>Thu, 18 Oct 2012 03:51:33 +0000</created>
                <updated>Tue, 13 Nov 2012 15:09:27 +0000</updated>
                            <resolved>Tue, 13 Nov 2012 15:09:27 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="46711" author="green" created="Thu, 18 Oct 2012 03:52:55 +0000"  >&lt;p&gt;Might be related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1612&quot; title=&quot;Test failure on test suite parallel-scale-nfsv3, subtest test_iorssf&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1612&quot;&gt;&lt;del&gt;LU-1612&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="46713" author="green" created="Thu, 18 Oct 2012 03:54:16 +0000"  >&lt;p&gt;Oh, and test output to date (hanging at this stage for many hours already):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity test 118k: bio alloc -ENOMEM and IO TERM handling =========== 22:02:17 (1350525737)
fail_loc=0x20e
10+0 records in
10+0 records out
10485760 bytes (10 MB) copied, 0.0580369 s, 181 MB/s
sanity.sh: line 6428: kill: (26840) - No such process
sanity.sh: line 6429: 26843 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
10+0 records in
10+0 records out
10485760 bytes (10 MB) copied, 0.0606453 s, 173 MB/s
sanity.sh: line 6428: kill: (26846) - No such process
sanity.sh: line 6429: 26849 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
10+0 records in
10+0 records out
10485760 bytes (10 MB) copied, 0.0595693 s, 176 MB/s
sanity.sh: line 6428: kill: (26852) - No such process
sanity.sh: line 6429: 26855 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
sanity.sh: line 6429: 26858 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
sanity.sh: line 6429: 26861 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
sanity.sh: line 6429: 26864 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
sanity.sh: line 6429: 26867 Terminated              ( dd if=/dev/zero of=$DIR/$tdir/$tfile-$i bs=1M count=10 || error &quot;dd to $DIR/$tdir/$tfile-$i failed&quot; )
fail_loc=0
10+0 records in
10+0 records out
10485760 bytes (10 MB) copied, 2.63154 s, 4.0 MB/s
10+0 records in
10+0 records out
10485760 bytes (10 MB) copied, 1.63022 s, 6.4 MB/s
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="46994" author="green" created="Sun, 28 Oct 2012 18:49:25 +0000"  >&lt;p&gt;Hit this again, here&apos;s the debug log&lt;/p&gt;</comment>
                            <comment id="47260" author="jay" created="Thu, 1 Nov 2012 12:06:09 +0000"  >&lt;p&gt;patch is at &lt;a href=&quot;http://review.whamcloud.com/4416&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4416&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="47741" author="jay" created="Tue, 13 Nov 2012 15:09:27 +0000"  >&lt;p&gt;This problem was imported by layout lock client implementation. Set this as duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1876&quot; title=&quot;Layout Lock Server Patch Landings to Master&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1876&quot;&gt;&lt;del&gt;LU-1876&lt;/del&gt;&lt;/a&gt; as it will be fixed over there.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="15169">LU-1612</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="11993" name="lu2208.txt.gz" size="447251" author="green" created="Sun, 28 Oct 2012 18:49:25 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvalr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5253</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>