<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:33:21 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3373] ldiskfs patches for FC19</title>
                <link>https://jira.whamcloud.com/browse/LU-3373</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This ticket intended to track ldiskfs patches work on FC18 3.7 kernel.&lt;/p&gt;</description>
                <environment></environment>
        <key id="19084">LU-3373</key>
            <summary>ldiskfs patches for FC19</summary>
                <type id="4" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11310&amp;avatarType=issuetype">Improvement</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="ys">Yang Sheng</assignee>
                                    <reporter username="ys">Yang Sheng</reporter>
                        <labels>
                    </labels>
                <created>Tue, 21 May 2013 14:12:08 +0000</created>
                <updated>Fri, 15 Nov 2019 20:29:00 +0000</updated>
                            <resolved>Tue, 1 Jul 2014 01:13:48 +0000</resolved>
                                                    <fixVersion>Lustre 2.7.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="59052" author="dmiter" created="Wed, 22 May 2013 14:05:30 +0000"  >&lt;p&gt;FC18 have 3.9.2 kernel already. Why 3.7 kernel is going to be supported?&lt;/p&gt;</comment>
                            <comment id="59054" author="simmonsja" created="Wed, 22 May 2013 14:11:41 +0000"  >&lt;p&gt;Can you post a link to download the kernel source. I will gladly update the kernel patch in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1812&quot; title=&quot;3.6/FC18 Server Patches&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1812&quot;&gt;&lt;del&gt;LU-1812&lt;/del&gt;&lt;/a&gt;. This is of interest since one of my goals for 2.5+ is to create a patchless ldiskfs.&lt;/p&gt;</comment>
                            <comment id="59058" author="dmiter" created="Wed, 22 May 2013 14:30:45 +0000"  >&lt;p&gt;&lt;a href=&quot;http://mirrors.med.harvard.edu/fedora/updates/18/SRPMS/kernel-3.9.2-200.fc18.src.rpm&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://mirrors.med.harvard.edu/fedora/updates/18/SRPMS/kernel-3.9.2-200.fc18.src.rpm&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="59059" author="ys" created="Wed, 22 May 2013 14:42:16 +0000"  >&lt;p&gt;I think work start from 3.7 can keep max compatible for further support.&lt;/p&gt;</comment>
                            <comment id="63792" author="ys" created="Wed, 7 Aug 2013 17:55:45 +0000"  >&lt;p&gt;First patchset: &lt;a href=&quot;http://review.whamcloud.com/#/c/7263/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7263/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I&apos;ll continue update it after 3.9.5 kernel work done.&lt;/p&gt;</comment>
                            <comment id="63799" author="bogl" created="Wed, 7 Aug 2013 18:21:40 +0000"  >&lt;p&gt;I think this a little behind the curve.  current kernel in fc18 is 3.9.11.  In fc19 it&apos;s 3.10.4&lt;/p&gt;

&lt;p&gt;As far as I know everything needed for 3.9 has already landed.  Not so true of 3.10.&lt;/p&gt;

&lt;p&gt;I don&apos;t see any mods to ldiskfs autoconf here to automatically choose the right (new) patch series.&lt;/p&gt;</comment>
                            <comment id="66009" author="ys" created="Sat, 7 Sep 2013 18:15:21 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/#/c/7263/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7263/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="66343" author="ys" created="Wed, 11 Sep 2013 14:27:52 +0000"  >&lt;p&gt;sanity test fail list:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;sanity: FAIL: test_17m e2fsck should not report error upon  short/long symlink MDT: rc=4
FAIL 17m (43s)
sanity: FAIL: test_24A Expected 5000 files, got 5005 (5005 unique)
FAIL 24A (5s)
sanity: FAIL: test_27k  &amp;gt; 4194304
FAIL 27k (1s)
sanity: FAIL: test_33a create
FAIL 33a (1s)
sanity: FAIL: test_36c 
FAIL 36c (1s)
sanity: FAIL: test_36d 
FAIL 36d (1s)
sanity: FAIL: test_102c setstripe failed
FAIL 102c (2s)
sanity: FAIL: test_103 permissions failed
FAIL 103 (17s)
sanity: FAIL: test_104b lfs check servers test failed
FAIL 104b (1s)
sanity: FAIL: test_129 exceeded dir size limit 4096 x 1 4096 : 12288 bytes
FAIL 129 (9s)
sanity: FAIL: test_133c The counter for destroy on ost was not incremented
FAIL 133c (16s)
sanity: FAIL: test_180c failed to load module obdecho
FAIL 180c (2s)
sanity: FAIL: test_183 test_183 failed with 1
FAIL 183 (3s)
sanity: FAIL: test_184c concurrent write on /mnt/lustre/d0.sanity/d184/184c/file1 failed
FAIL 184c (448s)
sanity: FAIL: test_218 multiop failed while creating a file
FAIL 218 (3s)
sanity: FAIL: test_219 test_219 failed with 1
FAIL 219 (2s)
FAIL 228a (8s)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="69655" author="simmonsja" created="Wed, 23 Oct 2013 17:35:59 +0000"  >&lt;p&gt;With the patches for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3319&quot; title=&quot;Adapt to 3.10 upstream kernel proc_dir_entry change&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3319&quot;&gt;&lt;del&gt;LU-3319&lt;/del&gt;&lt;/a&gt;, &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3974&quot; title=&quot;Support for linux 3.11 kernel&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3974&quot;&gt;&lt;del&gt;LU-3974&lt;/del&gt;&lt;/a&gt;, and this ticket we can almost get ldiskfs/osd-ldiskfs to build. As you can see from the below errors most of the errors are proc related which will be addressed for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3319&quot; title=&quot;Adapt to 3.10 upstream kernel proc_dir_entry change&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3319&quot;&gt;&lt;del&gt;LU-3319&lt;/del&gt;&lt;/a&gt;. The other bug deals with the reader -&amp;gt; iterate migrate discussed on &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3974&quot; title=&quot;Support for linux 3.11 kernel&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3974&quot;&gt;&lt;del&gt;LU-3974&lt;/del&gt;&lt;/a&gt;. We are almost there &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;

&lt;p&gt;lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c:66:&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_internal.h:625: error: &apos;struct lprocfs_static_vars&apos; declared inside parameter list&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_internal.h:625: error: its scope is only this definition or declaration, which is probably not what you want&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c: In function &apos;osd_ldiskfs_it_fill&apos;:&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c:4676: error: &apos;const struct file_operations&apos; has no member named &apos;readdir&apos;&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c: In function &apos;osd_mod_init&apos;:&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c:5806: error: storage size of &apos;lvars&apos; isn&apos;t known&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c:5817: error: passing argument 4 of &apos;class_register_type&apos; from incompatible pointer type&lt;br/&gt;
lustre-2.5.50/lustre/include/obd_class.h:89: note: expected &apos;struct lu_device_type *&apos; but argument is of type &apos;char *&apos;&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c:5817: error: too many arguments to function &apos;class_register_type&apos;&lt;br/&gt;
lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.c:5806: error: unused variable &apos;lvars&apos;&lt;br/&gt;
make&lt;span class=&quot;error&quot;&gt;&amp;#91;6&amp;#93;&lt;/span&gt;: *** &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre-2.5.50/lustre/osd-ldiskfs/osd_handler.o&amp;#93;&lt;/span&gt; Error 1&lt;br/&gt;
make&lt;span class=&quot;error&quot;&gt;&amp;#91;5&amp;#93;&lt;/span&gt;: *** &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre-2.5.50/lustre/osd-ldiskfs&amp;#93;&lt;/span&gt; Error 2&lt;/p&gt;</comment>
                            <comment id="70098" author="ys" created="Tue, 29 Oct 2013 08:32:05 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/7948&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7948&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/7794&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7794&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/7727&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7727&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="70373" author="ys" created="Thu, 31 Oct 2013 14:54:48 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/8110&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8110&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/8116&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8116&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="70573" author="ys" created="Sun, 3 Nov 2013 14:39:16 +0000"  >&lt;p&gt;Latest sanity status: &lt;/p&gt;

&lt;p&gt;sanity: FAIL: test_17m e2fsck should not report error upon  short/long symlink MDT: rc=4&lt;br/&gt;
FAIL 17m (44s)&lt;br/&gt;
sanity: FAIL: test_103 permissions failed&lt;br/&gt;
FAIL 103 (6s)&lt;br/&gt;
sanity: FAIL: test_133c The counter for destroy on ost was not incremented&lt;br/&gt;
FAIL 133c (14s)&lt;br/&gt;
sanity: FAIL: test_133e Bad write_bytes sum, expected 1376256, got 0&lt;br/&gt;
FAIL 133e (3s)&lt;/p&gt;</comment>
                            <comment id="71236" author="simmonsja" created="Mon, 11 Nov 2013 16:19:58 +0000"  >&lt;p&gt;I also had a patch for osd-ldiskfs iterate change but mine was unstable. Now I see it was due to me not updating the position field in struct dir_context. Your patch doesn&apos;t currently build for me on normal rhel6.4 but I working on fixing that. I should have something ready in the next few hours. Thanks Yang for finding that issue.&lt;/p&gt;</comment>
                            <comment id="71264" author="bogl" created="Mon, 11 Nov 2013 20:45:22 +0000"  >&lt;p&gt;I think there is still one more call to ldiskfs_journal_start_sb() in osd_handler.c that hasn&apos;t been switched to osd_journal_start_sb() by the existing patches.  Maybe could be taken care of by a refresh of #7794?&lt;/p&gt;

&lt;p&gt;Besides that all that&apos;s missing is the readdir -&amp;gt; iterate conversion that James has already mentioned.  And of course finding all the bugs.  Much of the previous hole has been filled in by recent server patch #8232 from James.&lt;/p&gt;

&lt;p&gt;We&apos;re getting very close to having a full ldiskfs server build.&lt;/p&gt;</comment>
                            <comment id="71298" author="ys" created="Tue, 12 Nov 2013 04:37:59 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/8231&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8231&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="71322" author="simmonsja" created="Tue, 12 Nov 2013 15:08:21 +0000"  >&lt;p&gt;It lives!!!! Behold ldiskfs on a 3.11.1 kernel. I had to fix up some of the ldiskfs patches but it works.&lt;/p&gt;

&lt;p&gt;jsimmons@spoon46:~$ uname -r;df&lt;br/&gt;
3.11.1&lt;br/&gt;
Filesystem           1K-blocks      Used Available Use% Mounted on&lt;br/&gt;
unionfs               88763392  62004224  22251520  74% /&lt;br/&gt;
none                       100         0       100   0% /dev/shm&lt;br/&gt;
none                   2938844     18352   2920492   1% /tmp&lt;br/&gt;
/dev/loop0              133560      1576    118624   2% /tmp/lustre/mds1&lt;br/&gt;
/dev/loop1              171080      9380    147700   6% /tmp/lustre/ost1&lt;br/&gt;
/dev/loop2              171080      9380    147700   6% /tmp/lustre/ost2&lt;br/&gt;
10.37.248.48@o2ib1:/lustre&lt;br/&gt;
                        342160     18760    295400   6% /lustre/barry&lt;/p&gt;</comment>
                            <comment id="71326" author="bogl" created="Tue, 12 Nov 2013 15:42:07 +0000"  >&lt;p&gt;with the refresh of #7794 and the addition of #8231 I can now complete a full server build with ldiskfs on the 3.11.7 kernel that is current in fc19 too.&lt;/p&gt;

&lt;p&gt;major milestone!!&lt;/p&gt;

&lt;p&gt;I&apos;m almost afraid to try some functional tests.&lt;br/&gt;
James, can you describe the changes you needed in ldiskfs patches?  They worked for me as is.&lt;/p&gt;</comment>
                            <comment id="71340" author="simmonsja" created="Tue, 12 Nov 2013 17:31:09 +0000"  >&lt;p&gt;Yes with the refresh now all I have to do is set the ldiskfs series manually. Also you need patch &lt;a href=&quot;http://review.whamcloud.com/#/c/8237&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/8237&lt;/a&gt; which I just pushed. I also updated &lt;a href=&quot;http://review.whamcloud.com/#/c/8116&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/8116&lt;/a&gt; locally to test if ext4_map_blocks() is available.&lt;/p&gt;</comment>
                            <comment id="71342" author="bogl" created="Tue, 12 Nov 2013 17:36:19 +0000"  >&lt;p&gt;I was using Peng&apos;s original upstream patch.  I see #8237 now.&lt;/p&gt;

&lt;p&gt;By the way I do have a prototype autoconf patch that selects the ldiskfs series. I haven&apos;t pushed it because I&apos;m not sure it&apos;s production worthy.&lt;/p&gt;</comment>
                            <comment id="71363" author="simmonsja" created="Tue, 12 Nov 2013 19:55:42 +0000"  >&lt;p&gt;Send it my way. I will try it out. Tomorrow I will setup a real file system using a 3.11 kernel to see how it holds up. If it holds up I plan to mount  it on our cray test bed and run a bunch of jobs to stress test it.&lt;/p&gt;</comment>
                            <comment id="71463" author="bogl" created="Wed, 13 Nov 2013 19:49:56 +0000"  >&lt;p&gt;it lives, take II&lt;br/&gt;
I&apos;m seeing first signs of life now with ldiskfs in fc19:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@fedora19 tests]# llmount.sh
Stopping clients: fedora19 /mnt/lustre (opts:)
Stopping clients: fedora19 /mnt/lustre2 (opts:)
Loading modules from /usr/lib64/lustre/tests/..
detected 1 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=vfstrace rpctrace dlmtrace neterror ha config ioctl super
subsystem_debug=all -lnet -lnd -pinger
gss/krb5 is not supported
quota/lquota options: &apos;hash_lqs_cur_bits=3&apos;
Formatting mgs, mds, osts
Format mds1: /tmp/lustre-mdt1
Format ost1: /tmp/lustre-ost1
Format ost2: /tmp/lustre-ost2
Checking servers environments
Checking clients fedora19 environments
Loading modules from /usr/lib64/lustre/tests/..
detected 1 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=vfstrace rpctrace dlmtrace neterror ha config ioctl super
subsystem_debug=all -lnet -lnd -pinger
gss/krb5 is not supported
Setup mgs, mdt, osts
Starting mds1:   -o loop /tmp/lustre-mdt1 /mnt/mds1
Started lustre-MDT0000
Starting ost1:   -o loop /tmp/lustre-ost1 /mnt/ost1
Started lustre-OST0000
Starting ost2:   -o loop /tmp/lustre-ost2 /mnt/ost2
Started lustre-OST0001
Starting client: fedora19: -o user_xattr,flock fedora19@tcp:/lustre /mnt/lustre
Using TIMEOUT=20
seting jobstats to procname_uid
Setting lustre.sys.jobid_var from disable to procname_uid
Changed after 0s: from &apos;&apos; to &apos;disable&apos;
Waiting 90 secs for update
Updated after 9s: wanted &apos;procname_uid&apos; got &apos;procname_uid&apos;
disable quota as required

[root@fedora19 tests]# lfs df
UUID                   1K-blocks        Used   Available Use% Mounted on
lustre-MDT0000_UUID       133560        1576      118624   1% /mnt/lustre[MDT:0]
lustre-OST0000_UUID       171080        9380      147700   6% /mnt/lustre[OST:0]
lustre-OST0001_UUID       171080        9380      147700   6% /mnt/lustre[OST:1]

filesystem summary:       342160       18760      295400   6% /mnt/lustre
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
</comment>
                            <comment id="71473" author="bogl" created="Wed, 13 Nov 2013 21:38:16 +0000"  >&lt;p&gt;One odd quirk.  lustre-ldiskfs.spec is setup with a Requires: modutils. However there isn&apos;t any rpm in fedora that provides modutils.  I could only install lustre with a --force option on the command line.&lt;/p&gt;

&lt;p&gt;Looks like in earlier distro modutils was one of the Provides: in the module-init-tools rpm.  The fedora rpm that contains similar files as the old module-init-tools rpm is named kmod, but it has no Provides: modutils in it.&lt;/p&gt;</comment>
                            <comment id="71878" author="simmonsja" created="Tue, 19 Nov 2013 12:03:06 +0000"  >&lt;p&gt;Ran several jobs (various IOR, mdtest configurations and a science app) against the 3.11 base lustre file system using ldiskfs. It was successful so I would say we are in pretty good shape.&lt;/p&gt;</comment>
                            <comment id="72246" author="ys" created="Mon, 25 Nov 2013 18:21:18 +0000"  >&lt;p&gt;O_LOV_DELAY_CREATE conflict with __O_TMPFILE. Need change it. lod and osp procfs patch need fix some bug. I have commented in gerrit.&lt;/p&gt;</comment>
                            <comment id="72709" author="bogl" created="Tue, 3 Dec 2013 17:02:10 +0000"  >&lt;p&gt;I think &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4209&quot; title=&quot;O_LOV_DELAY_CREATE conflict with __O_TMPFILE&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4209&quot;&gt;&lt;del&gt;LU-4209&lt;/del&gt;&lt;/a&gt; already exists to address the problem with O_LOV_DELAY_CREATE.&lt;/p&gt;</comment>
                            <comment id="72987" author="simmonsja" created="Fri, 6 Dec 2013 16:11:58 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/#/c/8116&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/8116&lt;/a&gt; looks ready for inspection and possible merger.&lt;/p&gt;</comment>
                            <comment id="73475" author="simmonsja" created="Fri, 13 Dec 2013 15:11:56 +0000"  >&lt;p&gt;I see patch 8116 has been abandon. I believe the point of the patch was to limit the port of patches for fc19 support. Do this mean the ldiskfs patches will be updated or will the rhel6.5 patch be enough?&lt;/p&gt;</comment>
                            <comment id="74168" author="simmonsja" created="Mon, 30 Dec 2013 19:02:38 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/#/c/8231&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/8231&lt;/a&gt; is ready for inspection and possible merger.&lt;/p&gt;</comment>
                            <comment id="87860" author="jlevi" created="Tue, 1 Jul 2014 01:13:48 +0000"  >&lt;p&gt;Patches landed to Master. Final patch &lt;a href=&quot;http://review.whamcloud.com/#/c/8116&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/8116&lt;/a&gt; is being tracked under &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5276&quot; title=&quot;ldiskfs patches for FC19&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5276&quot;&gt;&lt;del&gt;LU-5276&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="18540">LU-3228</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="21854">LU-4209</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="20128">LU-3675</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="29458">LU-6446</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="21039">LU-3974</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="25383">LU-5276</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvrh3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>8342</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                                                                                </customfields>
    </item>
</channel>
</rss>