<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:25:34 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2480] divide error in ldiskfs_mb_normalize_request</title>
                <link>https://jira.whamcloud.com/browse/LU-2480</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We get OSS crash on any attempt to write data to our Lustre FS. The file system is created from scratch with version 2.1.3 package. We have tried all kernel version from the Env. field.&lt;/p&gt;

&lt;p&gt;Initially I thought this was a kernel bug fixed in RHEL kernels-2.6.32-279.10.1.el6 &lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;kernel&amp;#93;&lt;/span&gt; sched: fix divide by zero at 
{thread_group,task}
&lt;p&gt;_times (Stanislaw Gruszka) &lt;span class=&quot;error&quot;&gt;&amp;#91;856703 843771&amp;#93;&lt;/span&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;



&lt;p&gt;On write attempt we get OSS crashes with the following console message :&lt;/p&gt;

&lt;p&gt;divide error: 0000 &lt;a href=&quot;#1&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;1&lt;/a&gt; SMP &lt;br/&gt;
last sysfs file: /sys/devices/system/cpu/cpu7/cache/index2/shared_cpu_map&lt;br/&gt;
CPU 7 &lt;br/&gt;
Modules linked in: obdfilter(U) fsfilt_ldiskfs(U) ost(U) mgc(U) ldiskfs(U) lustre(U) lov(U) osc(U) lquota(U) mdc(U) fid(U) fld(U) ksocklnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) libcfs(U) nfs fscache xt_multiport nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ipt_REJECT nf_conntrack_ipv4 nf_defrag_ipv4 xt_state nf_conntrack iptable_filter ip_tables ip6t_REJECT ip6table_filter ip6_tables ipv6 power_meter dcdbas microcode serio_raw ixgbe dca mdio k10temp amd64_edac_mod edac_core edac_mce_amd i2c_piix4 i2c_core sg ses enclosure bnx2 ext4 mbcache jbd2 sr_mod cdrom sd_mod crc_t10dif ahci qla2xxx scsi_transport_fc scsi_tgt megaraid_sas dm_mirror dm_region_hash dm_log dm_mod &lt;span class=&quot;error&quot;&gt;&amp;#91;last unloaded: mperf&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;Pid: 29280, comm: ll_ost_io_127 Not tainted 2.6.32-279.14.1.el6.x86_64 #1 Dell Inc. PowerEdge R715/0C5MMK&lt;br/&gt;
RIP: 0010:&lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0bb9c24&amp;gt;&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0bb9c24&amp;gt;&amp;#93;&lt;/span&gt; ldiskfs_mb_normalize_request+0xf4/0x3d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
RSP: 0018:ffff8804141a73e0  EFLAGS: 00010246&lt;br/&gt;
RAX: 0000000000020000 RBX: ffff88041c783898 RCX: 0000000000000003&lt;br/&gt;
RDX: 0000000000000000 RSI: 0000000000020100 RDI: 0000000000000000&lt;br/&gt;
RBP: ffff8804141a7430 R08: 0000000000000000 R09: 0000000000020000&lt;br/&gt;
R10: ffff88041c00e540 R11: 0000000000000000 R12: 0000000000000100&lt;br/&gt;
R13: ffff8804141a7500 R14: ffff88041c09dc00 R15: ffff8803908729c8&lt;br/&gt;
FS:  00007f5a66dde700(0000) GS:ffff880323c20000(0000) knlGS:0000000000000000&lt;br/&gt;
CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b&lt;br/&gt;
CR2: 00000000006dbf98 CR3: 000000041c04f000 CR4: 00000000000406e0&lt;br/&gt;
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000&lt;br/&gt;
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400&lt;br/&gt;
Process ll_ost_io_127 (pid: 29280, threadinfo ffff8804141a6000, task ffff8804141a2aa0)&lt;br/&gt;
Stack:&lt;br/&gt;
 ffff8804141a7400 ffffffffa0bb3c7f ffff8804141a7430 ffffffffa0bba212&lt;br/&gt;
&amp;lt;d&amp;gt; ffff8804141a7430 ffff8804141a7500 0000000000000100 ffff880418bfd000&lt;br/&gt;
&amp;lt;d&amp;gt; ffff88041c09dc00 ffff88041c783898 ffff8804141a74d0 ffffffffa0bc13aa&lt;br/&gt;
Call Trace:&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0bb3c7f&amp;gt;&amp;#93;&lt;/span&gt; ? ldiskfs_dirty_inode+0x4f/0x60 &lt;span class=&quot;error&quot;&gt;&amp;#91;ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0bba212&amp;gt;&amp;#93;&lt;/span&gt; ? ldiskfs_mb_initialize_context+0x82/0x1f0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0bc13aa&amp;gt;&amp;#93;&lt;/span&gt; ldiskfs_mb_new_blocks+0x42a/0x660 &lt;span class=&quot;error&quot;&gt;&amp;#91;ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811adb49&amp;gt;&amp;#93;&lt;/span&gt; ? __find_get_block+0xa9/0x200&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811adccc&amp;gt;&amp;#93;&lt;/span&gt; ? __getblk+0x2c/0x2e0&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811639bc&amp;gt;&amp;#93;&lt;/span&gt; ? __kmalloc+0x20c/0x220&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c68fca&amp;gt;&amp;#93;&lt;/span&gt; ldiskfs_ext_new_extent_cb+0x59a/0x6d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;fsfilt_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0ba869f&amp;gt;&amp;#93;&lt;/span&gt; ldiskfs_ext_walk_space+0x14f/0x340 &lt;span class=&quot;error&quot;&gt;&amp;#91;ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c68a30&amp;gt;&amp;#93;&lt;/span&gt; ? ldiskfs_ext_new_extent_cb+0x0/0x6d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;fsfilt_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c68758&amp;gt;&amp;#93;&lt;/span&gt; fsfilt_map_nblocks+0xd8/0x100 &lt;span class=&quot;error&quot;&gt;&amp;#91;fsfilt_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa00fced5&amp;gt;&amp;#93;&lt;/span&gt; ? start_this_handle+0xe5/0x500 &lt;span class=&quot;error&quot;&gt;&amp;#91;jbd2&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c68893&amp;gt;&amp;#93;&lt;/span&gt; fsfilt_ldiskfs_map_ext_inode_pages+0x113/0x220 &lt;span class=&quot;error&quot;&gt;&amp;#91;fsfilt_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c68a25&amp;gt;&amp;#93;&lt;/span&gt; fsfilt_ldiskfs_map_inode_pages+0x85/0x90 &lt;span class=&quot;error&quot;&gt;&amp;#91;fsfilt_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0cae99b&amp;gt;&amp;#93;&lt;/span&gt; filter_do_bio+0xdcb/0x18f0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdfilter&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c67580&amp;gt;&amp;#93;&lt;/span&gt; ? fsfilt_ldiskfs_brw_start+0x280/0x5a0 &lt;span class=&quot;error&quot;&gt;&amp;#91;fsfilt_ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0cb115e&amp;gt;&amp;#93;&lt;/span&gt; filter_commitrw_write+0x145e/0x2e78 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdfilter&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa04d8c1b&amp;gt;&amp;#93;&lt;/span&gt; ? lnet_send+0x29b/0xa60 &lt;span class=&quot;error&quot;&gt;&amp;#91;lnet&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8107ebe2&amp;gt;&amp;#93;&lt;/span&gt; ? del_timer_sync+0x22/0x30&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff814ff1ca&amp;gt;&amp;#93;&lt;/span&gt; ? schedule_timeout+0x19a/0x2e0&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0ca4252&amp;gt;&amp;#93;&lt;/span&gt; filter_commitrw+0x272/0x290 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdfilter&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c35bdd&amp;gt;&amp;#93;&lt;/span&gt; obd_commitrw+0x11d/0x3c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ost&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c3dd94&amp;gt;&amp;#93;&lt;/span&gt; ost_brw_write+0xcc4/0x1600 &lt;span class=&quot;error&quot;&gt;&amp;#91;ost&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06a2000&amp;gt;&amp;#93;&lt;/span&gt; ? target_bulk_timeout+0x0/0xc0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0c42e37&amp;gt;&amp;#93;&lt;/span&gt; ost_handle+0x2b77/0x4270 &lt;span class=&quot;error&quot;&gt;&amp;#91;ost&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06e077c&amp;gt;&amp;#93;&lt;/span&gt; ? lustre_msg_get_transno+0x8c/0x100 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06e7bfb&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_update_export_timer+0x4b/0x470 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06ef7eb&amp;gt;&amp;#93;&lt;/span&gt; ptlrpc_main+0xc4b/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06eeba0&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_main+0x0/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c14a&amp;gt;&amp;#93;&lt;/span&gt; child_rip+0xa/0x20&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06eeba0&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_main+0x0/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06eeba0&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_main+0x0/0x1a40 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c140&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0x0/0x20&lt;br/&gt;
Code: 8b 04 ca 4c 39 c0 0f 86 2b 02 00 00 83 c7 01 48 63 cf 48 39 d1 72 e8 48 8d 04 cd f8 ff ff ff 4d 63 04 02 31 d2 44 89 c8 44 89 c7 &amp;lt;48&amp;gt; f7 f7 31 d2 89 c1 8d 46 ff 41 0f af c8 48 f7 f7 89 ca 48 83 &lt;br/&gt;
RIP  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0bb9c24&amp;gt;&amp;#93;&lt;/span&gt; ldiskfs_mb_normalize_request+0xf4/0x3d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ldiskfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
 RSP &amp;lt;ffff8804141a73e0&amp;gt;&lt;/p&gt;

&lt;p&gt;This error stops us from deploying new lustre setup. Any help is greatly appreciated.&lt;/p&gt;</description>
                <environment>CPU: AMD Opteron(tm) Processor 6204&lt;br/&gt;
CentOS 6.3, kernels:&lt;br/&gt;
kernel-2.6.32-279.14.1.el6.x86_64&lt;br/&gt;
kernel-2.6.32-279.2.1.el6_lustre.gc46c389.x86_64&lt;br/&gt;
kernel-2.6.32-220.el6_lustre.g4554b65.x86_64&lt;br/&gt;
kernel-2.6.32-279.5.1.el6_lustre.gb16fe80.x86_64</environment>
        <key id="16906">LU-2480</key>
            <summary>divide error in ldiskfs_mb_normalize_request</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="sazs">Alexander Oltu</reporter>
                        <labels>
                    </labels>
                <created>Wed, 12 Dec 2012 05:50:45 +0000</created>
                <updated>Thu, 9 Jan 2020 06:17:34 +0000</updated>
                            <resolved>Thu, 9 Jan 2020 06:17:34 +0000</resolved>
                                    <version>Lustre 2.3.0</version>
                    <version>Lustre 2.1.1</version>
                    <version>Lustre 2.1.3</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="49111" author="bfaccini" created="Wed, 12 Dec 2012 06:50:24 +0000"  >&lt;p&gt;Hello,&lt;br/&gt;
Did you get any other messages just prior this Oops ???&lt;br/&gt;
And also have you tried to &lt;span class=&quot;error&quot;&gt;&amp;#91;e2&amp;#93;&lt;/span&gt;fsck (with &quot;-n&quot; option for &quot;no changes&quot; !!) the OSTs on this OSS ??&lt;/p&gt;</comment>
                            <comment id="49113" author="sazs" created="Wed, 12 Dec 2012 07:22:49 +0000"  >&lt;p&gt;No messages prior to crash.&lt;br/&gt;
This happens with any OST which gets data to be written. &lt;/p&gt;

&lt;p&gt;Fsck is OK:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@eik-oss1 ~&amp;#93;&lt;/span&gt;# e2fsck LABEL=workcmn1-OST0000&lt;br/&gt;
e2fsck 1.42.3.wc3 (15-Aug-2012)&lt;br/&gt;
workcmn1-OST0000: recovering journal&lt;br/&gt;
Setting free blocks count to 1952053240 (was 1952104696)&lt;br/&gt;
workcmn1-OST0000: clean, 121/15261952 files, 1467400/1953520640 blocks&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@eik-oss1 ~&amp;#93;&lt;/span&gt;# e2fsck -n LABEL=workcmn1-OST0000&lt;br/&gt;
e2fsck 1.42.3.wc3 (15-Aug-2012)&lt;br/&gt;
workcmn1-OST0000: clean, 121/15261952 files, 1467400/1953520640 blocks&lt;/p&gt;

&lt;p&gt; e2fsck -f LABEL=workcmn1-OST0000&lt;br/&gt;
e2fsck 1.42.3.wc3 (15-Aug-2012)&lt;br/&gt;
Pass 1: Checking inodes, blocks, and sizes&lt;br/&gt;
Pass 2: Checking directory structure&lt;br/&gt;
Pass 3: Checking directory connectivity&lt;br/&gt;
Pass 4: Checking reference counts&lt;br/&gt;
Pass 5: Checking group summary information&lt;br/&gt;
workcmn1-OST0000: 121/15261952 files (1.7% non-contiguous), 1467400/1953520640 blocks&lt;/p&gt;</comment>
                            <comment id="49115" author="bfaccini" created="Wed, 12 Dec 2012 08:11:13 +0000"  >&lt;p&gt;Ok, thank&apos;s, and BTW I forgot to ask if any crash-dump has been taken at the time of the crashes ??&lt;/p&gt;</comment>
                            <comment id="49117" author="sazs" created="Wed, 12 Dec 2012 08:24:03 +0000"  >&lt;p&gt;Yes, I have crash dumps they are around 100 MB. I can make a new crash dump with whamcloud kernel so that you have all debug symbols and I can provide vmcore file. Plz let me know where I can upload the file.&lt;/p&gt;

&lt;p&gt;BTW. I have tried mounting OST as an ext4 and writing files locally on the OSS is happening fine. But as soon as I mount it as ldiskfs it crashes.&lt;/p&gt;</comment>
                            <comment id="49120" author="sazs" created="Wed, 12 Dec 2012 08:53:31 +0000"  >&lt;p&gt;Bruno, I have dd&apos;ed with zeros begging of OST, reformatted with additional mkfs.options, disabled max_sectors optimizations and used mkfsoptions like stride and stripe_width with smaller values. The OST stopped crashing.&lt;/p&gt;

&lt;p&gt;I am going to add optimizations and reformat with proper stride, stripe_width and see.&lt;/p&gt;</comment>
                            <comment id="49122" author="bfaccini" created="Wed, 12 Dec 2012 09:09:28 +0000"  >&lt;p&gt;That was my next question, after your previous comment about ext4 mounts beeing ok but not ldiskfs mounts !!..., how did you format your OSTs ???&lt;/p&gt;
</comment>
                            <comment id="49127" author="sazs" created="Wed, 12 Dec 2012 09:41:09 +0000"  >&lt;p&gt;I didn&apos;t use any mkfsoptions initially, because I couldn&apos;t find real block size for our DDN S2A9550. And surprisingly the default is :&lt;br/&gt;
RAID stripe width:        16384 !!!&lt;/p&gt;

&lt;p&gt;So now I am using our cache block size on DDNs of 1024k and RAID 8+2 so using the following options:&lt;br/&gt;
mkfs.lustre --mkfsoptions=&quot;-m0 -E stripe_width=2048,stride=256&quot; --fsname=workcmn1 --ost --mgsnode=192.168.110.11@tcp0 --mgsnode=192.168.111.11@tcp1 /dev/sdb&lt;/p&gt;

&lt;p&gt;It is not crashing now. So I will reformat all our OSTs and will make a try.&lt;br/&gt;
I am just afraid that the stride is too big (2048 Segments of 1024 Kbytes per controller) and we are going to have 6 OSTs per controller . Maybe I can use LUN block size of 4096 (-E stripe_width=8,stride=1). Bruno, can you advise?&lt;br/&gt;
thanks,&lt;br/&gt;
Alex.&lt;/p&gt;</comment>
                            <comment id="49188" author="bfaccini" created="Thu, 13 Dec 2012 07:46:55 +0000"  >&lt;p&gt;Have a look to the &quot;Lustre Operations Manual&quot;, chapter &quot;Configuring Storage on a Lustre File System&quot; and adapt with your DDN S2A9550 OST/LUN design. But just to answer your last comment/question, the &quot;stride&quot; should be the size/4K-blocks to be written at a time on each of the disks which makes your RAID6/8+2 OSTs, and the &quot;stripe-width&quot; should be the optimal IO-size to fit your RAID design, so 2048=256x8 appear ok for me.&lt;/p&gt;</comment>
                            <comment id="49189" author="sazs" created="Thu, 13 Dec 2012 08:00:45 +0000"  >&lt;p&gt;Bruno, thank you for suggestions! I ended up using stripe_width=2048,stride=256.&lt;/p&gt;

&lt;p&gt;The only 2 open questions are:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;Why when I mkfs.lustre --ost without &quot;stripe_width&quot; and &quot;stride&quot; get &quot;stripe_width&quot; of 16384&lt;/li&gt;
	&lt;li&gt;Why ldiskfs crashes with &quot;divide zero&quot; when using such a big &quot;stripe_width&quot;&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="260827" author="adilger" created="Thu, 9 Jan 2020 06:17:34 +0000"  >&lt;p&gt;Close old ticket.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvdo7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5821</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10023"><![CDATA[4]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>