<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:11:58 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7794] tgt_clients_data_init()) soaked-MDT0001: duplicate export for client generation 1</title>
                <link>https://jira.whamcloud.com/browse/LU-7794</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We saw this during DNE failover soak-test.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: 4374:0:(client.c:2063:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1455822992/real 1455822992]  req@ffff8807bba3b980 x1526529651962748/t0(0) o250-&amp;gt;MGC192.168.1.108@o2ib10@0@lo:26/25 lens 520/544 e 0 to 1 dl 1455823038 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: 4374:0:(client.c:2063:ptlrpc_expire_one_request()) Skipped 37 previous similar messages
LustreError: 8115:0:(tgt_lastrcvd.c:1464:tgt_clients_data_init()) soaked-MDT0001: duplicate export for client generation 3
LustreError: 8115:0:(obd_config.c:578:class_setup()) setup soaked-MDT0001 failed (-114)
LustreError: 8115:0:(obd_config.c:1666:class_config_llog_handler()) MGC192.168.1.108@o2ib10: cfg command failed: rc = -114
Lustre:    cmd=cf003 0:soaked-MDT0001  1:soaked-MDT0001_UUID  2:1  3:soaked-MDT0001-mdtlov  4:f  
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;And it cause the failover MDT can not be mounted on the new MDS. Similar as &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7430&quot; title=&quot;General protection fault: 0000 upon mounting MDT&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7430&quot;&gt;&lt;del&gt;LU-7430&lt;/del&gt;&lt;/a&gt;, but no panic this time.&lt;/p&gt;</description>
                <environment></environment>
        <key id="34782">LU-7794</key>
            <summary>tgt_clients_data_init()) soaked-MDT0001: duplicate export for client generation 1</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="pichong">Gregoire Pichon</assignee>
                                    <reporter username="di.wang">Di Wang</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Thu, 18 Feb 2016 19:54:10 +0000</created>
                <updated>Thu, 3 Nov 2022 03:10:07 +0000</updated>
                            <resolved>Thu, 3 Nov 2022 03:10:07 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="142861" author="di.wang" created="Thu, 18 Feb 2016 20:01:04 +0000"  >&lt;p&gt;debuglog snippet &lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000020:01000000:8.0:1455822976.430423:0:8057:0:(obd_config.c:1562:class_config_llog_handler()) For 2.x interoperability, rename obd type from lov to lod (soaked-MDT0001)
00000020:00000080:8.0:1455822976.430426:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf001
00000020:00000080:8.0:1455822976.430429:0:8057:0:(obd_config.c:362:class_attach()) attach type lod name: soaked-MDT0001-mdtlov uuid: soaked-MDT0001-mdtlov_UUID
00000020:00000080:8.0:1455822976.430526:0:8057:0:(genops.c:371:class_newdev()) Adding new device soaked-MDT0001-mdtlov (ffff8807b479f078)
00000020:00000080:8.0:1455822976.430530:0:8057:0:(obd_config.c:432:class_attach()) OBD: dev 59 attached type lod with refcount 1
00000020:00000080:8.0:1455822976.430534:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf003
00080000:01000000:8.0:1455822976.430613:0:8057:0:(osd_handler.c:6597:osd_obd_connect()) connect #1
00000020:00000080:8.0:1455822976.430617:0:8057:0:(genops.c:1167:class_connect()) connect: client soaked-MDT0001-osd_UUID, cookie 0xe84502d68e3d0d9a
00000020:00000080:8.0:1455822976.430707:0:8057:0:(obd_config.c:552:class_setup()) finished setup of obd soaked-MDT0001-mdtlov (uuid soaked-MDT0001-mdtlov_UUID)
00000020:01000000:8.0:1455822976.430711:0:8057:0:(obd_config.c:1487:class_config_llog_handler()) Marker, inst_flg=0x2 mark_flg=0x2
00000020:00000080:8.0:1455822976.430714:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf010
00000020:00000080:8.0:1455822976.430715:0:8057:0:(obd_config.c:1218:class_process_config()) marker 8 (0x2) soaked-MDT0001-m lov setup
00000020:01000000:8.0:1455822976.430717:0:8057:0:(obd_config.c:1487:class_config_llog_handler()) Marker, inst_flg=0x0 mark_flg=0x1
00000020:00000080:8.0:1455822976.430720:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf010
00000020:00000080:8.0:1455822976.430721:0:8057:0:(obd_config.c:1218:class_process_config()) marker 9 (0x1) soaked-MDT0001 add mdt
00000020:00000080:8.0:1455822976.430724:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf001
00000020:00000080:8.0:1455822976.430725:0:8057:0:(obd_config.c:362:class_attach()) attach type mdt name: soaked-MDT0001 uuid: soaked-MDT0001_UUID
00000020:00000080:8.0:1455822976.430816:0:8057:0:(genops.c:371:class_newdev()) Adding new device soaked-MDT0001 (ffff8807a97fd038)
00000020:00000080:8.0:1455822976.430819:0:8057:0:(obd_config.c:432:class_attach()) OBD: dev 60 attached type mdt with refcount 1
00000020:00000080:8.0:1455822976.430822:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf007
00000020:00000080:8.0:1455822976.430823:0:8057:0:(obd_config.c:1176:class_process_config()) mountopt: profile soaked-MDT0001 osc soaked-MDT0001-mdtlov mdc (null)
00000020:01000000:8.0:1455822976.430826:0:8057:0:(obd_config.c:873:class_add_profile()) Add profile soaked-MDT0001
00000020:00000080:8.0:1455822976.430831:0:8057:0:(obd_config.c:1148:class_process_config()) processing cmd: cf003
00000020:01000004:8.0:1455822976.430883:0:8057:0:(obd_mount_server.c:175:server_get_mount()) get mount ffff8807c8232400 from soaked-MDT0001, refs=2
00000020:00000080:8.0:1455822976.430908:0:8057:0:(obd_config.c:362:class_attach()) attach type mdd name: soaked-MDD0001 uuid: soaked-MDD0001_UUID
00000020:00000080:8.0:1455822976.430998:0:8057:0:(genops.c:371:class_newdev()) Adding new device soaked-MDD0001 (ffff8807ea567078)
00000020:00000080:8.0:1455822976.431001:0:8057:0:(obd_config.c:432:class_attach()) OBD: dev 61 attached type mdd with refcount 1
00000004:01000000:8.0:1455822976.431060:0:8057:0:(lod_dev.c:1687:lod_obd_connect()) connect #0
00000020:00000080:8.0:1455822976.431064:0:8057:0:(genops.c:1167:class_connect()) connect: client soaked-MDT0001-mdtlov_UUID, cookie 0xe84502d68e3d0daf
00000020:00000080:8.0:1455822976.431096:0:8057:0:(obd_config.c:552:class_setup()) finished setup of obd soaked-MDD0001 (uuid soaked-MDD0001_UUID)
00000004:01000000:8.0:1455822976.431103:0:8057:0:(mdd_device.c:1178:mdd_obd_connect()) connect #000000020:00000080:8.0:1455822976.431106:0:8057:0:(genops.c:1167:class_connect()) connect: client soaked-MDD0001_UUID, cookie 0xe84502d68e3d0db6
00080000:01000000:8.0:1455822976.431113:0:8057:0:(osd_handler.c:6597:osd_obd_connect()) connect #2
00000020:00000080:8.0:1455822976.431115:0:8057:0:(genops.c:1167:class_connect()) connect: client soaked-MDT0001-osd_UUID, cookie 0xe84502d68e3d0dbd
00000001:00080000:8.0:1455822976.461440:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0000-mdtlov_UUID idx: 0 lr: 17182107883 srv lr: 17180780327 lx: 1526530761809700 gen 0
00000001:00080000:8.0:1455822976.461464:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0003-mdtlov_UUID idx: 1 lr: 17181767205 srv lr: 17180780327 lx: 1526529647176604 gen 0
00000001:00080000:8.0:1455822976.461475:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0002-mdtlov_UUID idx: 2 lr: 17182149935 srv lr: 17180780327 lx: 1526529649871472 gen 0
00000001:00080000:8.0:1455822976.461483:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0005-mdtlov_UUID idx: 3 lr: 17182105205 srv lr: 17180780327 lx: 1526525227761264 gen 0
00000001:00080000:8.0:1455822976.461493:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0004-mdtlov_UUID idx: 4 lr: 17182152956 srv lr: 17180780327 lx: 1526525229666900 gen 0
00000001:00080000:8.0:1455822976.461503:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0007-mdtlov_UUID idx: 5 lr: 17182138513 srv lr: 17180780327 lx: 1526536609097660 gen 0
00000001:00080000:8.0:1455822976.461512:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: soaked-MDT0006-mdtlov_UUID idx: 6 lr: 17181782621 srv lr: 17180780327 lx: 1526536607769984 gen 0
00000001:00080000:8.0:1455822976.461520:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: c2781877-e222-f9a8-07f4-a1250eba2af6 idx: 7 lr: 0 srv lr: 17180780327 lx: 0 gen 1
00000001:00080000:8.0:1455822976.461530:0:8057:0:(tgt_lastrcvd.c:1427:tgt_clients_data_init()) RCVRNG CLIENT uuid: 150208e6-71a8-375e-d139-d478eec5b761 idx: 8 lr: 0 srv lr: 17180780327 lx: 0 gen 1
00000001:00020000:8.0:1455822976.461539:0:8057:0:(tgt_lastrcvd.c:1464:tgt_clients_data_init()) soaked-MDT0001: duplicate export for client generation 1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="143013" author="pjones" created="Fri, 19 Feb 2016 19:09:18 +0000"  >&lt;p&gt;Gregoire&lt;/p&gt;

&lt;p&gt;Could you please advise?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="143060" author="pichong" created="Fri, 19 Feb 2016 22:27:32 +0000"  >&lt;p&gt;Honestly, I have no idea what could lead to that duplicate export situation.&lt;/p&gt;

&lt;p&gt;Could you detail what tests were running on the filesystem ?&lt;/p&gt;

&lt;p&gt;Do you think you could rebuild the operations (client connections/disconnections, MDT mount/unmount, other MDTs connections/disconnections, ...) that occurred on the filesystem from the node logs ?&lt;/p&gt;</comment>
                            <comment id="143096" author="di.wang" created="Sat, 20 Feb 2016 02:52:08 +0000"  >&lt;p&gt;Hmm, I am not sure we can reproduce the problem easily. I am not sure there are node logs. If you need debug log, maybe I can upload it somewhere.&lt;/p&gt;</comment>
                            <comment id="143111" author="pichong" created="Sat, 20 Feb 2016 17:22:31 +0000"  >&lt;p&gt;Yes, could upload debug logs as attachment to this JIRA ticket ?&lt;/p&gt;

&lt;p&gt;Are the D_INFO messages written to the logs ?&lt;br/&gt;
It would be helpful to look for the D_INFO messages logged in tgt_client_new()&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;CDEBUG(D_INFO, &lt;span class=&quot;code-quote&quot;&gt;&quot;%s: &lt;span class=&quot;code-keyword&quot;&gt;new&lt;/span&gt; client at index %d (%llu) with UUID &lt;span class=&quot;code-quote&quot;&gt;&apos;%s&apos;&lt;/span&gt; &quot;&lt;/span&gt;
        &lt;span class=&quot;code-quote&quot;&gt;&quot;generation %d\n&quot;&lt;/span&gt;,
        tgt-&amp;gt;lut_obd-&amp;gt;obd_name, ted-&amp;gt;ted_lr_idx, ted-&amp;gt;ted_lr_off,
        ted-&amp;gt;ted_lcd-&amp;gt;lcd_uuid, ted-&amp;gt;ted_lcd-&amp;gt;lcd_generation);
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;And see in which context they were called.&lt;/p&gt;

&lt;p&gt;And again, could you detail what tests were running on the filesystem, please ?&lt;/p&gt;</comment>
                            <comment id="143138" author="di.wang" created="Mon, 22 Feb 2016 03:57:15 +0000"  >&lt;p&gt;I upload the debug log to ftp.hpdd.intel.com:/uploads/lu-7794/&lt;/p&gt;

&lt;p&gt;The whole process does like this, 4 MDS, each MDS has 2 MDTs.  MDS0 (MDT0/1), MDS1(MDT2/3), MDS2(MDT4/5), MDS3(MDT6/7). MDS0 and MDS1 are configured as active/active failover, and MDS2 and MDS3 are configured as active/active failover.&lt;/p&gt;

&lt;p&gt;And the test is randomly chosen one of MDS to reboot.  then its MDTs will be failover to its pair MDS. In this case, it is MDS0 is restarted, then MDT0/MDT1 should be mounted on MDS1, but failed because of this.  No, I do not think this can be easily produced.&lt;/p&gt;

&lt;p&gt;Maybe Frank or Cliff will know more.&lt;/p&gt;</comment>
                            <comment id="143163" author="heckes" created="Mon, 22 Feb 2016 13:52:02 +0000"  >&lt;p&gt;Gregoire: I hope the info&apos;s below will answer you&apos;re question. Please let me know if something is missing or unclear.&lt;/p&gt;
&lt;h3&gt;&lt;a name=&quot;smallenvironmentinfoaddition&quot;&gt;&lt;/a&gt;small environment info addition&lt;/h3&gt;
&lt;ul&gt;
	&lt;li&gt;&lt;tt&gt;lola-8&lt;/tt&gt; --&amp;gt; MDS0, &lt;tt&gt;lola-9&lt;/tt&gt; --&amp;gt; MDS1, &lt;tt&gt;lola-10&lt;/tt&gt; --&amp;gt; MDS2,  &lt;tt&gt;lola-11&lt;/tt&gt; --&amp;gt; MDS3&lt;/li&gt;
	&lt;li&gt;MDTs formatted with &lt;em&gt;ldiskfs&lt;/em&gt; , OSTs using &lt;em&gt;zfs&lt;/em&gt;&lt;/li&gt;
	&lt;li&gt;Failover procedure
	&lt;ul&gt;
		&lt;li&gt;Triggered by automated framework&lt;/li&gt;
		&lt;li&gt;(Random) selected node is powercycled at (randomly) chosen time&lt;/li&gt;
		&lt;li&gt;Wait till node is up again&lt;/li&gt;
		&lt;li&gt;Mount resources on failover partner:&lt;br/&gt;
    Mount MDT resources in sequence &lt;br/&gt;
    Wait for mount command to complete&lt;br/&gt;
    If error occurrs retry to mount Lustre MDT(s) again.&lt;/li&gt;
		&lt;li&gt;Framework is configured for test session NOT to  wait for RECOVERY process to complete&lt;/li&gt;
		&lt;li&gt;umount MDTs on secondary node&lt;/li&gt;
		&lt;li&gt;mount MDTs on primary node again&lt;/li&gt;
	&lt;/ul&gt;
	&lt;/li&gt;
	&lt;li&gt;NOTE:&lt;br/&gt;
I checked the soak framework. I&apos;m very sure that the implementation &lt;b&gt;won&apos;t&apos;&lt;/b&gt;  execute multiple &lt;br/&gt;
&lt;tt&gt;mount&lt;/tt&gt; commands at the same time or start a new &lt;tt&gt;mount&lt;/tt&gt; command while the one started before&lt;br/&gt;
haven&apos;t finished, yet. &lt;br/&gt;
The framework &lt;b&gt;didn&apos;t&lt;/b&gt; check whether there&apos;s already a mount command running executed from outside the framework.&lt;/li&gt;
&lt;/ul&gt;


&lt;h3&gt;&lt;a name=&quot;Concerning%27sequenceofevents%27&quot;&gt;&lt;/a&gt;Concerning &apos;sequence of events&apos;&lt;/h3&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;2016-02-18 11:07:14,437:fsmgmt.fsmgmt:INFO     triggering fault mds_failover
2016-02-18 11:07:14,438:fsmgmt.fsmgmt:INFO     reseting MDS node lola-8     (--&amp;gt; Node was powercycled!)
2016-02-18 11:07:14,439:fsmgmt.fsmgmt:INFO     executing cmd pm -h powerman -c lola-8&amp;gt; /dev/null
2016-02-18 11:07:28,291:fsmgmt.fsmgmt:INFO     trying to connect to lola-8 ...
2016-02-18 11:07:38,307:fsmgmt.fsmgmt:INFO     trying to connect to lola-8 ...
2016-02-18 11:07:46,410:fsmgmt.fsmgmt:INFO     trying to connect to lola-8 ...
...
...
2016-02-18 11:13:36,132:fsmgmt.fsmgmt:INFO     trying to connect to lola-8 ...
2016-02-18 11:13:37,060:fsmgmt.fsmgmt:INFO     lola-8 is up!!!
2016-02-18 11:13:48,072:fsmgmt.fsmgmt:INFO     Failing over soaked-MDT0001 ...
2016-02-18 11:13:48,073:fsmgmt.fsmgmt:INFO     Mounting soaked-MDT0001 on lola-9 ...
2016-02-18 11:16:16,760:fsmgmt.fsmgmt:ERROR    ... mount of soaked-MDT0001 on lola-9 failed with 114, retrying ...
2016-02-18 11:16:16,760:fsmgmt.fsmgmt:INFO     mount.lustre: increased /sys/block/dm-10/queue/max_sectors_kb from 1024 to 16383
mount.lustre: increased /sys/block/dm-8/queue/max_sectors_kb from 1024 to 16383
mount.lustre: increased /sys/block/sdg/queue/max_sectors_kb from 1024 to 16383
mount.lustre: increased /sys/block/sdb/queue/max_sectors_kb from 1024 to 16383
mount.lustre: mount /dev/mapper/360080e50002ffd820000024f52013094p1 at /mnt/soaked-mdt1 failed: Operation already in progress
The target service is already running. (/dev/mapper/360080e50002ffd820000024f52013094p1)
2016-02-18 11:18:12,738:fsmgmt.fsmgmt:ERROR    ... mount of soaked-MDT0001 on lola-9 failed with 114, retrying ...
2016-02-18 11:18:12,739:fsmgmt.fsmgmt:INFO     mount.lustre: mount /dev/mapper/360080e50002ffd820000024f52013094p1 at /mnt/soaked-mdt1 failed: Operation already in progress
The target service is already running. (/dev/mapper/360080e50002ffd820000024f52013094p1)
2016-02-18 11:26:58,512:fsmgmt.fsmgmt:INFO     ... soaked-MDT0001 mounted successfully on lola-9
2016-02-18 11:26:58,513:fsmgmt.fsmgmt:INFO     ... soaked-MDT0001 failed over
2016-02-18 11:26:58,513:fsmgmt.fsmgmt:INFO     Failing over soaked-MDT0000 ...
2016-02-18 11:26:58,513:fsmgmt.fsmgmt:INFO     Mounting soaked-MDT0000 on lola-9 ...
2016-02-18 11:27:51,049:fsmgmt.fsmgmt:INFO     ... soaked-MDT0000 mounted successfully on lola-9
2016-02-18 11:27:51,049:fsmgmt.fsmgmt:INFO     ... soaked-MDT0000 failed over
2016-02-18 11:28:11,430:fsmgmt.fsmgmt:DEBUG    Recovery Result Record: {&apos;lola-9&apos;: {&apos;soaked-MDT0001&apos;: &apos;RECOVERING&apos;, &apos;soaked-MDT0000&apos;: &apos;RECOVERING&apos;, &apos;soaked-MDT0003&apos;: &apos;COMPLETE&apos;, &apos;soaked-MDT0002&apos;: &apos;COMPLETE&apos;}}
2016-02-18 11:28:11,431:fsmgmt.fsmgmt:INFO     soaked-MDT0001 in status &apos;RECOVERING&apos;.
2016-02-18 11:28:11,431:fsmgmt.fsmgmt:INFO     soaked-MDT0000 in status &apos;RECOVERING&apos;.
2016-02-18 11:28:11,431:fsmgmt.fsmgmt:INFO     Don&apos;t wait for recovery to complete. Failback MDT&apos;s immediately
2016-02-18 11:28:11,431:fsmgmt.fsmgmt:INFO     Failing back soaked-MDT0001 ...
2016-02-18 11:28:11,431:fsmgmt.fsmgmt:INFO     Unmounting soaked-MDT0001 on lola-9 ...
2016-02-18 11:28:12,078:fsmgmt.fsmgmt:INFO     ... soaked-MDT0001 unmounted successfully on lola-9
2016-02-18 11:28:12,079:fsmgmt.fsmgmt:INFO     Mounting soaked-MDT0001 on lola-8 ...
2016-02-18 11:29:03,122:fsmgmt.fsmgmt:INFO     ... soaked-MDT0001 mounted successfully on lola-8
2016-02-18 11:29:03,122:fsmgmt.fsmgmt:INFO     ... soaked-MDT0001 failed back
2016-02-18 11:29:03,123:fsmgmt.fsmgmt:INFO     Failing back soaked-MDT0000 ...
2016-02-18 11:29:03,123:fsmgmt.fsmgmt:INFO     Unmounting soaked-MDT0000 on lola-9 ...
2016-02-18 11:29:09,942:fsmgmt.fsmgmt:INFO     ... soaked-MDT0000 unmounted successfully on lola-9
2016-02-18 11:29:09,942:fsmgmt.fsmgmt:INFO     Mounting soaked-MDT0000 on lola-8 ...
2016-02-18 11:29:24,023:fsmgmt.fsmgmt:INFO     ... soaked-MDT0000 mounted successfully on lola-8
2016-02-18 11:29:24,023:fsmgmt.fsmgmt:INFO     ... soaked-MDT0000 failed back
2016-02-18 11:29:24,024:fsmgmt.fsmgmt:INFO     mds_failover just completed
2016-02-18 11:29:24,024:fsmgmt.fsmgmt:INFO     next fault in 1898s
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;The error message at &lt;tt&gt;2016-02-18 11:16:16,760  ...  failed with 114 ...&lt;/tt&gt; is strange as (stated above) the soak framework&lt;br/&gt;
don&apos;t execute multiple mounts of the same device file. &lt;br/&gt;
I&apos;m not sure whether some manual mount (outside the soak framework) of MDT1 took place. I&apos;m sure I didn&apos;t execute anything on the node at this time.&lt;/p&gt;

&lt;h3&gt;&lt;a name=&quot;applicationsexecuted&quot;&gt;&lt;/a&gt;applications executed&lt;/h3&gt;
&lt;ul&gt;
	&lt;li&gt;mdtest  (single shared file, file per process)&lt;/li&gt;
	&lt;li&gt;IOR (single shared file, file per process)&lt;/li&gt;
	&lt;li&gt;simul&lt;/li&gt;
	&lt;li&gt;blogbench&lt;/li&gt;
	&lt;li&gt;kcompile&lt;/li&gt;
	&lt;li&gt;pct (producer, consumer inhouse application)&lt;br/&gt;
All applications are initiated with random size, file count, block size. If needed I could provide the (slurm) list of active jobs at the time the error occured.&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="143233" author="bzzz" created="Mon, 22 Feb 2016 19:03:26 +0000"  >&lt;p&gt;it&apos;s interesting that in this case and in LDEV-180 the duplicated entry had generation=1. meaning that for a reason the counter got reset. &lt;/p&gt;</comment>
                            <comment id="144101" author="pichong" created="Mon, 29 Feb 2016 09:23:12 +0000"  >&lt;p&gt;Di Wang,&lt;/p&gt;

&lt;p&gt;Unfortunately, I cannot get the debug log from ftp.hpdd.intel.com:/uploads/lu-7794/, since the anonymous login only provides write access.&lt;/p&gt;

&lt;p&gt;Would it be another way to provide these logs ?&lt;/p&gt;</comment>
                            <comment id="144104" author="pichong" created="Mon, 29 Feb 2016 10:33:42 +0000"  >&lt;p&gt;Why the error reported in the description of the ticket (&lt;tt&gt;soaked-MDT0001: duplicate export for client generation 3&lt;/tt&gt;) is not the same than the error reported in the first comment from Di Wang (&lt;tt&gt;soaked-MDT0001: duplicate export for client generation 1&lt;/tt&gt;) ?&lt;/p&gt;

&lt;p&gt;Does it mean there has been several occurences of the issue ?&lt;br/&gt;
Could it be possible to have the logs of the first occurence ?&lt;/p&gt;


&lt;p&gt;By the way, how many Lustre clients are connected to the filesystem ?&lt;br/&gt;
Are the two clients with uuid &lt;tt&gt;c2781877-e222-f9a8-07f4-a1250eba2af6&lt;/tt&gt; and &lt;tt&gt;150208e6-71a8-375e-d139-d478eec5b761&lt;/tt&gt; different from others ?&lt;/p&gt;
</comment>
                            <comment id="144189" author="cliffw" created="Mon, 29 Feb 2016 20:35:34 +0000"  >&lt;p&gt;There are 8 clients connected to the filesystem. I will check on client differences, but afaik they are uniform.&lt;/p&gt;</comment>
                            <comment id="144191" author="pjones" created="Mon, 29 Feb 2016 20:40:38 +0000"  >&lt;p&gt;Gr&#233;goire&lt;/p&gt;

&lt;p&gt;To turn the question around - is there someone you would like the logs uploaded to so that you can access them?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="144192" author="di.wang" created="Mon, 29 Feb 2016 20:50:02 +0000"  >&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Does it mean there has been several occurences of the issue ?
Could it be possible to have the logs of the first occurrence ?
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Yes, there are several occurrences according to the log. You can tell me one place, and I can upload the debug log there. Unfortunately the debug log is too big (more 100M) to be uploaded here.&lt;/p&gt;</comment>
                            <comment id="145123" author="pichong" created="Thu, 10 Mar 2016 12:31:18 +0000"  >&lt;p&gt;I have been able to access to the log yesterday. Unfortunately, I have not found any usefull information that could help identify the cause of the problem.&lt;/p&gt;

&lt;p&gt;Anyway, I wonder if the &lt;tt&gt;tgt_client_new()&lt;/tt&gt; routine could be called after &lt;tt&gt;tgt_init()&lt;/tt&gt; and before &lt;tt&gt;tgt_clients_data_init()&lt;/tt&gt; leading to assign the client lcd_generation using a wrong &lt;tt&gt;lut_client_generation&lt;/tt&gt; value. The &lt;tt&gt;lut_client_generation&lt;/tt&gt; value is initialized to 0 in &lt;tt&gt;tgt_init()&lt;/tt&gt; but updated with the highest client generation read from last_rcvd file in case of recovery.&lt;/p&gt;

&lt;p&gt;In case of recovery, if a client connects to the MDT in parallel of the target initialization, are its &lt;tt&gt;obd_export&lt;/tt&gt; and &lt;tt&gt;tg_export_data&lt;/tt&gt; structures immediately initialized ?&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="70713">LU-15935</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="33139">LU-7430</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy1s7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>