<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:29:52 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-9854] Lustre 2.10.0 mmap() issues</title>
                <link>https://jira.whamcloud.com/browse/LU-9854</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;At my company we use mmap() exclusively for accessing our data on Lustre.  For starters we are seeing some very weird (or maybe expected) poor random read/write performance for these types of access patterns.  I decided to give Lustre 2.10.0 a try with ZFS 0.7.0 as the backend instead of ldisk and after compiling and building the RPMs, the filesystem mounted up just fine.  I then started doing some iozone runs to test the stability of the filesystem and although izone does complete uts benchmark I am seeing a lot of stack traces coming out of various kernel threads.  Note I am only seeing this when using mmap().  We also ran our application as well just to verify.  I am also going to try with an ldiskfs format as well to see if this changes anything. &lt;/p&gt;

&lt;p&gt;My ZFS settings are modest, with 50% of memory allocated to the ARC:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;options zfs zfs_arc_max=3921674240 zfs_prefetch_disable=1
recordsize=1M
compression=on
dedupe=off
xattr=sa
dnodesize=auto
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Below is the output from the stack trace:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Aug  8 09:38:04 dev-gc01-oss001 kernel: BUG: Bad page state: 87 messages suppressed
Aug  8 09:38:04 dev-gc01-oss001 kernel: BUG: Bad page state in process socknal_sd00_01  pfn:1cbac1
Aug  8 09:38:04 dev-gc01-oss001 kernel: page:ffffea00072eb040 count:0 mapcount:-1 mapping:          (&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;) index:0x0
Aug  8 09:38:04 dev-gc01-oss001 kernel: page flags: 0x2fffff00008000(tail)
Aug  8 09:38:04 dev-gc01-oss001 kernel: page dumped because: nonzero mapcount
Aug  8 09:38:04 dev-gc01-oss001 kernel: Modules linked in: 8021q garp mrp stp llc osp(OE) ofd(OE) lfsck(OE) ost(OE) mgc(OE) osd_zfs(OE) lquota(OE) fid(OE) fld(OE) ksocklnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE) libcfs(OE) iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd ppdev sg i2c_piix4 parport_pc i2c_core parport pcspkr nfsd nfs_acl lockd grace binfmt_misc auth_rpcgss sunrpc ip_tables xfs libcrc32c zfs(POE) zunicode(POE) zavl(POE) icp(POE) zcommon(POE) znvpair(POE) spl(OE) zlib_deflate sd_mod crc_t10dif crct10dif_generic virtio_net virtio_scsi crct10dif_pclmul crct10dif_common crc32c_intel serio_raw virtio_pci virtio_ring virtio
Aug  8 09:38:04 dev-gc01-oss001 kernel: CPU: 0 PID: 2558 Comm: socknal_sd00_01 Tainted: P    B      OE  ------------   3.10.0-514.26.2.el7.x86_64 #1
Aug  8 09:38:04 dev-gc01-oss001 kernel: Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/01/2011
Aug  8 09:38:04 dev-gc01-oss001 kernel: ffffea00072eb040 00000000005e265e ffff8800b879f5a8 ffffffff81687133
Aug  8 09:38:04 dev-gc01-oss001 kernel: ffff8800b879f5d0 ffffffff81682368 ffffea00072eb040 0000000000000000
Aug  8 09:38:04 dev-gc01-oss001 kernel: 000fffff00000000 ffff8800b879f618 ffffffff8118946d fff00000fe000000
Aug  8 09:38:04 dev-gc01-oss001 kernel: Call Trace:
Aug  8 09:38:04 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81687133&amp;gt;] dump_stack+0x19/0x1b
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81682368&amp;gt;] bad_page.part.75+0xdf/0xfc
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8118946d&amp;gt;] free_pages_prepare+0x16d/0x190
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff811897b9&amp;gt;] __free_pages_ok+0x19/0xd0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8118988b&amp;gt;] free_compound_page+0x1b/0x20
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81683526&amp;gt;] __put_compound_page+0x1f/0x22
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81683698&amp;gt;] put_compound_page+0x16f/0x17d
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8118edfc&amp;gt;] put_page+0x4c/0x60
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8155ec1f&amp;gt;] skb_release_data+0x8f/0x140
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8155ecf4&amp;gt;] skb_release_all+0x24/0x30
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8155f1ec&amp;gt;] consume_skb+0x2c/0x80
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8156f06d&amp;gt;] __dev_kfree_skb_any+0x3d/0x50
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffffa0019afb&amp;gt;] free_old_xmit_skbs.isra.32+0x6b/0xc0 [virtio_net]
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffffa0019baf&amp;gt;] start_xmit+0x5f/0x4f0 [virtio_net]
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8156f9a1&amp;gt;] dev_hard_start_xmit+0x171/0x3b0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81597574&amp;gt;] sch_direct_xmit+0x104/0x200
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8157252c&amp;gt;] __dev_queue_xmit+0x23c/0x570
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81572870&amp;gt;] dev_queue_xmit+0x10/0x20
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815b6876&amp;gt;] ip_finish_output+0x466/0x750
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815b7873&amp;gt;] ip_output+0x73/0xe0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815b5531&amp;gt;] ip_local_out_sk+0x31/0x40
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815b58a3&amp;gt;] ip_queue_xmit+0x143/0x3a0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815cf04f&amp;gt;] tcp_transmit_skb+0x4af/0x990
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815cf68a&amp;gt;] tcp_write_xmit+0x15a/0xce0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815d048e&amp;gt;] __tcp_push_pending_frames+0x2e/0xc0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815bed2c&amp;gt;] tcp_push+0xec/0x120
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815c25b8&amp;gt;] tcp_sendmsg+0xc8/0xc40
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff815ed854&amp;gt;] inet_sendmsg+0x64/0xb0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81555ff0&amp;gt;] sock_sendmsg+0xb0/0xf0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff8168ea3b&amp;gt;] ? _raw_spin_unlock_bh+0x1b/0x40
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81556067&amp;gt;] kernel_sendmsg+0x37/0x50
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffffa09f40d9&amp;gt;] ksocknal_lib_send_iov+0xd9/0x140 [ksocklnd]
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffffa09ed32f&amp;gt;] ksocknal_process_transmit+0x2af/0xb90 [ksocklnd]
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffffa09f1b84&amp;gt;] ksocknal_scheduler+0x204/0x670 [ksocklnd]
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff810b1b20&amp;gt;] ? wake_up_atomic_t+0x30/0x30
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffffa09f1980&amp;gt;] ? ksocknal_recv+0x2a0/0x2a0 [ksocklnd]
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff810b0a4f&amp;gt;] kthread+0xcf/0xe0
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff810b0980&amp;gt;] ? kthread_create_on_node+0x140/0x140
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff81697758&amp;gt;] ret_from_fork+0x58/0x90
Aug  8 09:38:05 dev-gc01-oss001 kernel: [&amp;lt;ffffffff810b0980&amp;gt;] ? kthread_create_on_node+0x140/0x140
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;-Chris&lt;/p&gt;</description>
                <environment>CentOS 7.3 - 3.10.0-514.26.2.el7.x86_64&lt;br/&gt;
Lustre 2.10.0&lt;br/&gt;
ZFS 0.7.0</environment>
        <key id="47745">LU-9854</key>
            <summary>Lustre 2.10.0 mmap() issues</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="4">Incomplete</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="chjohnst">Christopher Johnston</reporter>
                        <labels>
                            <label>zfs</label>
                    </labels>
                <created>Thu, 10 Aug 2017 14:40:12 +0000</created>
                <updated>Fri, 17 Nov 2017 21:13:54 +0000</updated>
                            <resolved>Thu, 26 Oct 2017 14:28:17 +0000</resolved>
                                    <version>Lustre 2.10.0</version>
                                    <fixVersion>Lustre 2.11.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="205045" author="pjones" created="Thu, 10 Aug 2017 16:33:17 +0000"  >&lt;p&gt;Alex&lt;/p&gt;

&lt;p&gt;Is this related to the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9305&quot; title=&quot;Running File System Aging create write checksum errors &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9305&quot;&gt;&lt;del&gt;LU-9305&lt;/del&gt;&lt;/a&gt; change?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="205118" author="bzzz" created="Fri, 11 Aug 2017 01:50:34 +0000"  >&lt;p&gt;Hi, did you get these traces on the client side?&lt;/p&gt;</comment>
                            <comment id="208447" author="pjones" created="Thu, 14 Sep 2017 23:03:43 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/secure/ViewProfile.jspa?name=chjohnst&quot; class=&quot;user-hover&quot; rel=&quot;chjohnst&quot;&gt;chjohnst&lt;/a&gt; are you able to supply some more detail here?&lt;/p&gt;</comment>
                            <comment id="212070" author="pjones" created="Thu, 26 Oct 2017 14:28:17 +0000"  >&lt;p&gt;ok this does not seem to be a priority for you so we&apos;ll just close the ticket. If someone else encounters this and wants to engage we can reopen the ticket&lt;/p&gt;</comment>
                            <comment id="214060" author="chjohnst" created="Fri, 17 Nov 2017 21:13:54 +0000"  >&lt;p&gt;Sorry for the delay in response, other priorities at work took precedence.  This is still likely an issue and I have not tried to use 2.10.1 which is recently released but there was still an issue with my test cluster when we tried it recently.  I will look to stand up a latest and greatest version over the weekend and retest as using ZFS is becoming more important for me to take advantage of the compression and snapshots. &lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="40183">LU-9305</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                    <customfield id="customfield_10030" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic/Theme</customfieldname>
                        <customfieldvalues>
                                        <label>zfs</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzzi87:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>