<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:24:43 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2379] 40GigE LNet performance</title>
                <link>https://jira.whamcloud.com/browse/LU-2379</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Warpmech got LNet performance issue over 40GigE:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;between two 40GigE nodes, selftest can only get half of bandwidth
	&lt;ul&gt;
		&lt;li&gt;it&apos;s not a big surprise to me, because socklnd connection is bound with exact one thread, and receiving side doesn&apos;t have Zero-copy, which means only one core is receiving from a 40GigE link if there is only one peer, so I suspect performance of 1:1 40GigE is CPU bound.&lt;/li&gt;
		&lt;li&gt;one possible solution is reusing CONTROL link to transfer bulk data as well with some special policy&lt;/li&gt;
		&lt;li&gt;I know that enable kiov vmap can help on receiving performance of some NICs, but not sure it will work for this case.&lt;/li&gt;
	&lt;/ul&gt;
	&lt;/li&gt;
	&lt;li&gt;while running BRW tests between 10GigE clients and 40GigE server, one direction can work well (saturate the bandwidth), another direction can only get half.
	&lt;ul&gt;
		&lt;li&gt;8 clients read can saturate the link pretty well, but I don&apos;t know if 4 clients can do the same.&lt;/li&gt;
		&lt;li&gt;8 clients write can&apos;t saturate the link, it can get half of bandwidth. I actually found that one 10GigE client can only get 250-300MB/sec on write against 40GigE server, which matches aggregation write performance of 8 clients (2GB/sec)&lt;/li&gt;
		&lt;li&gt;network latency is quite low (I don&apos;t remember the number)&lt;/li&gt;
	&lt;/ul&gt;
	&lt;/li&gt;
&lt;/ul&gt;
</description>
                <environment></environment>
        <key id="16756">LU-2379</key>
            <summary>40GigE LNet performance</summary>
                <type id="4" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11310&amp;avatarType=issuetype">Improvement</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="liang">Liang Zhen</reporter>
                        <labels>
                    </labels>
                <created>Fri, 23 Nov 2012 00:45:34 +0000</created>
                <updated>Thu, 11 Jun 2020 09:52:13 +0000</updated>
                            <resolved>Thu, 11 Jun 2020 09:52:13 +0000</resolved>
                                                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="48762" author="isaac" created="Tue, 4 Dec 2012 17:17:19 +0000"  >&lt;p&gt;1. point to point 40GE tests:&lt;br/&gt;
I heard that at least a fully utilized 5GHz CPU would be needed to saturate one direction of a 10GE link, if nothing is off loaded from host CPU. That depends on message size. Have they tuned jumbo frames and TCP send/recv buffers?&lt;/p&gt;

&lt;p&gt;2. BRW tests between 10GigE clients and 40GigE server:&lt;/p&gt;

&lt;p&gt;I assumed that it was 8 clients vs one server. If true, then when 8 clients are reading, the overhead of memory copying on incoming data was spread over the 8 clients, while in the write case, the server had to handle all the memory copy overhead. If that&apos;s the cause, then it&apos;d be interesting to see the results with more servers, e.g. 8 clients write to 2 servers or more.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvcpj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5648</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                                                                                </customfields>
    </item>
</channel>
</rss>