<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:25:46 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2505] lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]</title>
                <link>https://jira.whamcloud.com/browse/LU-2505</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;lfsck test hung as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;lt;~snip~&amp;gt;
creating /mnt/lustre/d0.lfsck/testfile.30
creating files in /mnt/lustre/d0.lfsck/d9
creating files in /mnt/lustre/d0.lfsck/d10
creating files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
linking files in /mnt/lustre/d0.lfsck/d11
CMD: client-26vm8 lctl get_param -n obdfilter.*.uuid
CMD: client-26vm8 lctl get_param -n obdfilter.lustre-OST0000.mntdev
Stopping clients: client-26vm1,client-26vm2.lab.whamcloud.com /mnt/lustre (opts:-f)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Console log on client-26vm2 showed that:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: 4267:0:(llite_lib.c:2290:ll_dirty_page_discard_warn()) dirty page discard: 10.10.4.154@tcp:/lustre/fid: [0x200002b13:0x1224:0x0]/ may get corrupted (rc -108)
Lustre: 4267:0:(llite_lib.c:2290:ll_dirty_page_discard_warn()) dirty page discard: 10.10.4.154@tcp:/lustre/fid: [0x200002b13:0x1237:0x0]/ may get corrupted (rc -108)
Lustre: 22194:0:(cl_lock.c:2026:cl_lock_page_out()) Writing 1 pages error: -108
BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]
Modules linked in: lustre(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) lquota(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) libcfs(U) ext2 nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode virtio_balloon 8139too 8139cp mii i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: libcfs]
CPU 0 
Modules linked in: lustre(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) lquota(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) libcfs(U) ext2 nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode virtio_balloon 8139too 8139cp mii i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: libcfs]

Pid: 22194, comm: umount Not tainted 2.6.32-279.14.1.el6.x86_64 #1 Red Hat KVM
RIP: 0010:[&amp;lt;ffffffff8150098e&amp;gt;]  [&amp;lt;ffffffff8150098e&amp;gt;] _spin_lock+0x1e/0x30
RSP: 0018:ffff88007b7ef508  EFLAGS: 00000206
RAX: 0000000000000001 RBX: ffff88007b7ef508 RCX: 0000000000000000
RDX: 0000000000000000 RSI: ffff88007ae2c400 RDI: ffff880076b3a734
RBP: ffffffff8100bc0e R08: 0000000000000001 R09: 0000000000000001
R10: 0000000000000000 R11: 0000000000000001 R12: ffff8800760154e0
R13: ffff8800760154e0 R14: ffff880024017cf0 R15: ffff88007b7ef488
FS:  00007f7464baa740(0000) GS:ffff880002200000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007f74642c81b0 CR3: 0000000078444000 CR4: 00000000000006f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process umount (pid: 22194, threadinfo ffff88007b7ee000, task ffff88007d059500)
Stack:
 ffff88007b7ef5a8 ffffffffa08ee29c ffff8800754f3d98 ffff880073c08400
&amp;lt;d&amp;gt; ffff88007ae2c400 ffff880078d76c38 ffff88007ae2c510 ffff88007ae2c4f0
&amp;lt;d&amp;gt; ffff880076b3a648 0000000100000001 ffff88007ae2c4f8 ffff88007d0449e8
Call Trace:
 [&amp;lt;ffffffffa08ee29c&amp;gt;] ? osc_io_submit+0x31c/0x540 [osc]
 [&amp;lt;ffffffffa081c148&amp;gt;] ? cl_io_submit_rw+0x78/0x180 [obdclass]
 [&amp;lt;ffffffffa0958ad5&amp;gt;] ? lov_io_submit+0x355/0xc30 [lov]
 [&amp;lt;ffffffffa0813c65&amp;gt;] ? cl_page_put+0x1e5/0x580 [obdclass]
 [&amp;lt;ffffffffa081680a&amp;gt;] ? pgoff_at_lock+0x1a/0x70 [obdclass]
 [&amp;lt;ffffffffa081c148&amp;gt;] ? cl_io_submit_rw+0x78/0x180 [obdclass]
 [&amp;lt;ffffffffa081f29f&amp;gt;] ? cl_io_submit_sync+0x9f/0x210 [obdclass]
 [&amp;lt;ffffffffa08166c7&amp;gt;] ? cl_lock_page_out+0x207/0x330 [obdclass]
 [&amp;lt;ffffffffa0816860&amp;gt;] ? pageout_cb+0x0/0x100 [obdclass]
 [&amp;lt;ffffffffa08ea99f&amp;gt;] ? osc_lock_flush+0x4f/0x90 [osc]
 [&amp;lt;ffffffffa08eaaaf&amp;gt;] ? osc_lock_cancel+0xcf/0x1b0 [osc]
 [&amp;lt;ffffffffa08152a5&amp;gt;] ? cl_lock_cancel0+0x75/0x160 [obdclass]
 [&amp;lt;ffffffffa0815f8b&amp;gt;] ? cl_lock_cancel+0x13b/0x140 [obdclass]
 [&amp;lt;ffffffffa08ebb8a&amp;gt;] ? osc_ldlm_blocking_ast+0x13a/0x380 [osc]
 [&amp;lt;ffffffffa0420cc0&amp;gt;] ? ldlm_cancel_callback+0x60/0x100 [ptlrpc]
 [&amp;lt;ffffffffa043b14b&amp;gt;] ? ldlm_cli_cancel_local+0x7b/0x380 [ptlrpc]
 [&amp;lt;ffffffffa043efd8&amp;gt;] ? ldlm_cli_cancel+0x58/0x3a0 [ptlrpc]
 [&amp;lt;ffffffffa0428db8&amp;gt;] ? cleanup_resource+0x168/0x300 [ptlrpc]
 [&amp;lt;ffffffffa06baefe&amp;gt;] ? cfs_hash_spin_lock+0xe/0x10 [libcfs]
 [&amp;lt;ffffffffa0428f7a&amp;gt;] ? ldlm_resource_clean+0x2a/0x50 [ptlrpc]
 [&amp;lt;ffffffffa06bb18f&amp;gt;] ? cfs_hash_for_each_relax+0x17f/0x380 [libcfs]
 [&amp;lt;ffffffffa0428f50&amp;gt;] ? ldlm_resource_clean+0x0/0x50 [ptlrpc]
 [&amp;lt;ffffffffa0428f50&amp;gt;] ? ldlm_resource_clean+0x0/0x50 [ptlrpc]
 [&amp;lt;ffffffffa06bcbaf&amp;gt;] ? cfs_hash_for_each_nolock+0x7f/0x1c0 [libcfs]
 [&amp;lt;ffffffffa0425a49&amp;gt;] ? ldlm_namespace_cleanup+0x29/0xb0 [ptlrpc]
 [&amp;lt;ffffffffa08e2122&amp;gt;] ? osc_import_event+0x3d2/0x1470 [osc]
 [&amp;lt;ffffffffa0487c59&amp;gt;] ? ptlrpc_invalidate_import+0x2a9/0x8d0 [ptlrpc]
 [&amp;lt;ffffffffa08e283c&amp;gt;] ? osc_import_event+0xaec/0x1470 [osc]
 [&amp;lt;ffffffffa045f53d&amp;gt;] ? ptlrpc_set_import_active+0x33d/0x480 [ptlrpc]
 [&amp;lt;ffffffffa08dd070&amp;gt;] ? osc_iocontrol+0xc80/0x11b0 [osc]
 [&amp;lt;ffffffff8127cf26&amp;gt;] ? vsnprintf+0x2b6/0x5f0
 [&amp;lt;ffffffffa092f7ff&amp;gt;] ? lov_iocontrol+0x27f/0x1e90 [lov]
 [&amp;lt;ffffffffa064eadf&amp;gt;] ? lmv_iocontrol+0x25f/0x1cf0 [lmv]
 [&amp;lt;ffffffff81188961&amp;gt;] ? path_put+0x31/0x40
 [&amp;lt;ffffffff81188325&amp;gt;] ? putname+0x35/0x50
 [&amp;lt;ffffffffa09dd640&amp;gt;] ? ll_umount_begin+0x2c0/0xc30 [lustre]
 [&amp;lt;ffffffff8119b52a&amp;gt;] ? sys_umount+0x11a/0x3a0
 [&amp;lt;ffffffff810d6d42&amp;gt;] ? audit_syscall_entry+0x272/0x2a0
 [&amp;lt;ffffffff81503ade&amp;gt;] ? do_page_fault+0x3e/0xa0
 [&amp;lt;ffffffff8100b0f2&amp;gt;] ? system_call_fastpath+0x16/0x1b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The full console log is attached.&lt;/p&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/679bdff0-4828-11e2-a406-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/679bdff0-4828-11e2-a406-52540035b04c&lt;/a&gt;&lt;/p&gt;</description>
                <environment>Lustre Tag: v2_1_4_RC1&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_1/159/&quot;&gt;http://build.whamcloud.com/job/lustre-b2_1/159/&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL6.3/x86_64 (kernel version: 2.6.32_279.14.1.el6)&lt;br/&gt;
</environment>
        <key id="16958">LU-2505</key>
            <summary>lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="jay">Jinshan Xiong</assignee>
                                    <reporter username="yujian">Jian Yu</reporter>
                        <labels>
                    </labels>
                <created>Tue, 18 Dec 2012 02:00:06 +0000</created>
                <updated>Sat, 22 Dec 2012 13:41:38 +0000</updated>
                            <resolved>Sat, 22 Dec 2012 13:41:38 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                    <version>Lustre 2.1.4</version>
                                    <fixVersion>Lustre 2.4.0</fixVersion>
                    <fixVersion>Lustre 2.1.4</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="49360" author="yujian" created="Tue, 18 Dec 2012 02:39:37 +0000"  >&lt;p&gt;RHEL6.3/x86_64 (2.3.0 Server + 2.1.4 RC1 Client):&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/29296da0-486e-11e2-8cdc-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/29296da0-486e-11e2-8cdc-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49407" author="pjones" created="Tue, 18 Dec 2012 16:34:59 +0000"  >&lt;p&gt;Bob is looking into this&lt;/p&gt;</comment>
                            <comment id="49418" author="yujian" created="Tue, 18 Dec 2012 23:41:24 +0000"  >&lt;p&gt;I reproduced this issue manually on Toro client-12vm&lt;span class=&quot;error&quot;&gt;&amp;#91;1,3,4&amp;#93;&lt;/span&gt; against Lustre v2_1_4_RC1 and got the kernel dump file of client-12vm1.&lt;/p&gt;

&lt;p&gt;Client:  client-12vm1&lt;br/&gt;
MGS/MDS: client-12vm3&lt;br/&gt;
OSS:     client-12vm4&lt;/p&gt;

&lt;p&gt;The dump file is /scratch/logs/2.1.4/&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;/client-12vm1.dump on brent node.&lt;/p&gt;

&lt;p&gt;I also put the kernel-debuginfo-* packages into that directory.&lt;/p&gt;

&lt;p&gt;The console log client-12vm1 is attached.&lt;/p&gt;</comment>
                            <comment id="49420" author="jay" created="Wed, 19 Dec 2012 00:23:03 +0000"  >&lt;p&gt;This should be the same problem of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2451&quot; title=&quot;recovery-small test_24b: BUG: soft lockup - CPU#0 stuck for 67s! [ll_imp_inval:4791]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2451&quot;&gt;&lt;del&gt;LU-2451&lt;/del&gt;&lt;/a&gt; because it also spin on client obd list lock. The patch is at: &lt;a href=&quot;http://review.whamcloud.com/4806&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4806&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49421" author="jay" created="Wed, 19 Dec 2012 00:49:29 +0000"  >&lt;p&gt;yujian: what&apos;s the build # you&apos;re running?&lt;/p&gt;</comment>
                            <comment id="49423" author="yujian" created="Wed, 19 Dec 2012 00:54:11 +0000"  >&lt;blockquote&gt;&lt;p&gt;yujian: what&apos;s the build # you&apos;re running?&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Lustre v2_1_4_RC1: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_1/159/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_1/159/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I&apos;m re-provisioning the test nodes to verify &lt;a href=&quot;http://review.whamcloud.com/4806&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4806&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="49425" author="jay" created="Wed, 19 Dec 2012 01:41:52 +0000"  >&lt;p&gt;I took a look at the kernel dump - thank you for the dump file, yujian.&lt;/p&gt;

&lt;p&gt;Apparently the problem is in ll_dirty_page_discard_warn() where it allocates a page with GFP_KERNEL. And this function can be called inside client obd list lock as follows:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;PID: 10150  TASK: ffff88007d11f540  CPU: 0   COMMAND: &lt;span class=&quot;code-quote&quot;&gt;&quot;ptlrpcd-brw&quot;&lt;/span&gt;
 #0 [ffff88007d075840] schedule at ffffffff814fe052
 #1 [ffff88007d075908] __cond_resched at ffffffff8106244a
 #2 [ffff88007d075928] _cond_resched at ffffffff814fea60
 #3 [ffff88007d075938] __alloc_pages_nodemask at ffffffff811272e5
 #4 [ffff88007d075a58] alloc_pages_current at ffffffff8115c7ea
 #5 [ffff88007d075a88] __get_free_pages at ffffffff8112572e
 #6 [ffff88007d075a98] ll_dirty_page_discard_warn at ffffffffa0a3d252 [lustre]
 #7 [ffff88007d075b08] vvp_page_completion_common at ffffffffa0a7adfc [lustre]
 #8 [ffff88007d075b38] vvp_page_completion_write_common at ffffffffa0a7ae6b [lustre]
 #9 [ffff88007d075b68] vvp_page_completion_write at ffffffffa0a7b83e [lustre]
#10 [ffff88007d075ba8] cl_page_completion at ffffffffa05eed8f [obdclass]
#11 [ffff88007d075c48] osc_completion at ffffffffa0880812 [osc]
#12 [ffff88007d075c98] osc_ap_completion at ffffffffa086a544 [osc]
#13 [ffff88007d075cf8] brw_interpret at ffffffffa0876d69 [osc]
#14 [ffff88007d075d98] ptlrpc_check_set at ffffffffa0736d5b [ptlrpc]
#15 [ffff88007d075e38] ptlrpcd_check at ffffffffa0767fe0 [ptlrpc]
#16 [ffff88007d075e68] ptlrpcd at ffffffffa076827e [ptlrpc]
#17 [ffff88007d075f48] kernel_thread at ffffffff8100c14a
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The fix is pretty simple, use GFP_ATOMIC instead.&lt;/p&gt;</comment>
                            <comment id="49428" author="jay" created="Wed, 19 Dec 2012 01:48:46 +0000"  >&lt;p&gt;patch is at: &lt;a href=&quot;http://review.whamcloud.com/4866&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4866&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49432" author="pjones" created="Wed, 19 Dec 2012 02:44:47 +0000"  >&lt;p&gt;Jinshan&lt;/p&gt;

&lt;p&gt;Does this issue also affect master?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="49436" author="adilger" created="Wed, 19 Dec 2012 04:37:21 +0000"  >&lt;p&gt;This is caused by &lt;a href=&quot;http://review.whamcloud.com/1908&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/1908&lt;/a&gt; from &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-921&quot; title=&quot;generate warnings in case of discarding dirty pages&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-921&quot;&gt;&lt;del&gt;LU-921&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="49457" author="jay" created="Wed, 19 Dec 2012 12:18:44 +0000"  >&lt;p&gt;patch for master is at: &lt;a href=&quot;http://review.whamcloud.com/4870&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4870&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49471" author="pjones" created="Wed, 19 Dec 2012 19:49:33 +0000"  >&lt;p&gt;Jinshan has a fix for this&lt;/p&gt;</comment>
                            <comment id="49609" author="pjones" created="Sat, 22 Dec 2012 13:41:38 +0000"  >&lt;p&gt;Landed for 2.1.4 and 2.4&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="16881">LU-2451</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="12661">LU-921</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="12108" name="client-12vm1-console.log" size="104160" author="yujian" created="Tue, 18 Dec 2012 23:41:24 +0000"/>
                            <attachment id="12105" name="client-26vm2-console.log" size="750811" author="yujian" created="Tue, 18 Dec 2012 02:00:06 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvdyv:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5875</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>