[LU-4192] NFS server crash while fsx runs on clients Created: 30/Oct/13  Updated: 07/Oct/16  Resolved: 07/Oct/16

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.5.0
Fix Version/s: None

Type: Bug Priority: Minor
Reporter: James Nunez (Inactive) Assignee: WC Triage
Resolution: Duplicate Votes: 0
Labels: None
Environment:

Lustre 2.5.0-RC1 build #2, RHEL 6, OpenSFS cluster with one MDS/MGS, one OST with two OSSs, one NFS server, four clients.


Issue Links:
Duplicate
is duplicated by LU-5263 ll_read_ahead_pages() ASSERTION( page... Resolved
Severity: 3
Rank (Obsolete): 11347

 Description   

Running 2.5.0-RC1, with one NFS server exporting the Lustre file system with 4 clients running "./fsx", "./fsx", "./fsx -c 3 -l 8000000 -r 75 -w 4096 /misc/export/nfs_x_4" and "./fsx -c 5 -l 500000 -o 2048". The export options are rw,async,insecure. After a few hours of running fsx, the NFS server crashed.

From the console of the NFS server (c06):

Message from syslogd@c06 at Oct 30 14:52:40 ...
 kernel:LustreError: 6653:0:(rw.c:695:ll_read_ahead_pages()) ASSERTION( page_idx > ria->ria_stoff ) failed: Invalid page_idx 1025rs 1025 re 1534 ro 1279 rl 256 rp 1

Message from syslogd@c06 at Oct 30 14:52:40 ...
 kernel:LustreError: 6653:0:(rw.c:695:ll_read_ahead_pages()) LBUG

From vmcore-dmesg.txt on the NFS server (c06):

<0>LustreError: 6653:0:(rw.c:695:ll_read_ahead_pages()) ASSERTION( page_idx > ri
a->ria_stoff ) failed: Invalid page_idx 1025rs 1025 re 1534 ro 1279 rl 256 rp 1
<0>LustreError: 6653:0:(rw.c:695:ll_read_ahead_pages()) LBUG
<4>Pid: 6653, comm: nfsd
<4>
<4>Call Trace:
<4> [<ffffffffa044f895>] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
<4> [<ffffffffa044fe97>] lbug_with_loc+0x47/0xb0 [libcfs]
<4> [<ffffffffa0a607d6>] ll_readahead+0x10c6/0x10f0 [lustre]
<4> [<ffffffff8127d847>] ? radix_tree_insert+0x1c7/0x220
<4> [<ffffffffa0a8cd75>] vvp_io_read_page+0x305/0x340 [lustre]
<4> [<ffffffffa05d5c7d>] cl_io_read_page+0x8d/0x170 [obdclass]
<4> [<ffffffffa05c9a77>] ? cl_page_assume+0xf7/0x220 [obdclass]
<4> [<ffffffffa0a5f176>] ll_readpage+0x96/0x1a0 [lustre]
<4> [<ffffffff811b0cf8>] __generic_file_splice_read+0x3a8/0x560
<4> [<ffffffff81168712>] ? kmem_cache_alloc+0x182/0x190
<4> [<ffffffffa0464e47>] ? cfs_hash_bd_lookup_intent+0x37/0x130 [libcfs]
<4> [<ffffffffa0464e47>] ? cfs_hash_bd_lookup_intent+0x37/0x130 [libcfs]
<4> [<ffffffffa05cd6db>] ? cl_lock_fits_into+0x6b/0x190 [obdclass]
<4> [<ffffffffa09b38d9>] ? lov_lock_fits_into+0x409/0x560 [lov]
<4> [<ffffffff811aeb90>] ? spd_release_page+0x0/0x20
<4> [<ffffffff811b0efa>] generic_file_splice_read+0x4a/0x90
<4> [<ffffffffa0a8e905>] vvp_io_read_start+0x3c5/0x470 [lustre]
<4> [<ffffffffa05d37ea>] cl_io_start+0x6a/0x140 [obdclass]
<4> [<ffffffffa05d7ef4>] cl_io_loop+0xb4/0x1b0 [obdclass]
<4> [<ffffffffa0a30f9f>] ll_file_io_generic+0x33f/0x610 [lustre]
<4> [<ffffffffa0a319c0>] ll_file_splice_read+0xb0/0x1d0 [lustre]
<4> [<ffffffff811af15b>] do_splice_to+0x6b/0xa0
<4> [<ffffffff811af45f>] splice_direct_to_actor+0xaf/0x1c0
<4> [<ffffffffa03aa3b0>] ? nfsd_direct_splice_actor+0x0/0x20 [nfsd]
<4> [<ffffffffa03aae70>] nfsd_vfs_read+0x1a0/0x1c0 [nfsd]
<4> [<ffffffffa03ac4a0>] nfsd_read_file+0x90/0xb0 [nfsd]
<4> [<ffffffffa03c00ef>] nfsd4_encode_read+0x13f/0x240 [nfsd]
<4> [<ffffffffa03c5dd4>] ? nfs4_preprocess_stateid_op+0x284/0x310 [nfsd]
<4> [<ffffffffa03b9d35>] nfsd4_encode_operation+0x75/0x180 [nfsd]
<4> [<ffffffffa03b7d35>] nfsd4_proc_compound+0x195/0x490 [nfsd]
<4> [<ffffffffa03a543e>] nfsd_dispatch+0xfe/0x240 [nfsd]
<4> [<ffffffffa02d0614>] svc_process_common+0x344/0x640 [sunrpc]
<4> [<ffffffff81063410>] ? default_wake_function+0x0/0x20
<4> [<ffffffffa02d0c50>] svc_process+0x110/0x160 [sunrpc]
<4> [<ffffffffa03a5b62>] nfsd+0xc2/0x160 [nfsd]
<4> [<ffffffffa03a5aa0>] ? nfsd+0x0/0x160 [nfsd]
<4> [<ffffffff81096a36>] kthread+0x96/0xa0
<4> [<ffffffff8100c0ca>] child_rip+0xa/0x20
<4> [<ffffffff810969a0>] ? kthread+0x0/0xa0
<4> [<ffffffff8100c0c0>] ? child_rip+0x0/0x20
<4>
<0>Kernel panic - not syncing: LBUG
<4>Pid: 6653, comm: nfsd Not tainted 2.6.32-358.18.1.el6.x86_64 #1
<4>Call Trace:
<4> [<ffffffff8150da18>] ? panic+0xa7/0x16f
<4> [<ffffffffa044feeb>] ? lbug_with_loc+0x9b/0xb0 [libcfs]
<4> [<ffffffffa0a607d6>] ? ll_readahead+0x10c6/0x10f0 [lustre]
<4> [<ffffffff8127d847>] ? radix_tree_insert+0x1c7/0x220
<4> [<ffffffffa0a8cd75>] ? vvp_io_read_page+0x305/0x340 [lustre]
<4> [<ffffffffa05d5c7d>] ? cl_io_read_page+0x8d/0x170 [obdclass]
<4> [<ffffffffa05c9a77>] ? cl_page_assume+0xf7/0x220 [obdclass]
<4> [<ffffffffa0a5f176>] ? ll_readpage+0x96/0x1a0 [lustre]
<4> [<ffffffff811b0cf8>] ? __generic_file_splice_read+0x3a8/0x560
<4> [<ffffffff81168712>] ? kmem_cache_alloc+0x182/0x190
<4> [<ffffffffa0464e47>] ? cfs_hash_bd_lookup_intent+0x37/0x130 [libcfs]
<4> [<ffffffffa0464e47>] ? cfs_hash_bd_lookup_intent+0x37/0x130 [libcfs]
<4> [<ffffffffa05cd6db>] ? cl_lock_fits_into+0x6b/0x190 [obdclass]
<4> [<ffffffffa09b38d9>] ? lov_lock_fits_into+0x409/0x560 [lov]
<4> [<ffffffff811aeb90>] ? spd_release_page+0x0/0x20
<4> [<ffffffff811b0efa>] ? generic_file_splice_read+0x4a/0x90
<4> [<ffffffffa0a8e905>] ? vvp_io_read_start+0x3c5/0x470 [lustre]
<4> [<ffffffffa05d37ea>] ? cl_io_start+0x6a/0x140 [obdclass]
<4> [<ffffffffa05d7ef4>] ? cl_io_loop+0xb4/0x1b0 [obdclass]
<4> [<ffffffffa0a30f9f>] ? ll_file_io_generic+0x33f/0x610 [lustre]
<4> [<ffffffffa0a319c0>] ? ll_file_splice_read+0xb0/0x1d0 [lustre]
<4> [<ffffffff811af15b>] ? do_splice_to+0x6b/0xa0
<4> [<ffffffff811af45f>] ? splice_direct_to_actor+0xaf/0x1c0
<4> [<ffffffffa03aa3b0>] ? nfsd_direct_splice_actor+0x0/0x20 [nfsd]
<4> [<ffffffffa03aae70>] ? nfsd_vfs_read+0x1a0/0x1c0 [nfsd]
<4> [<ffffffffa03ac4a0>] ? nfsd_read_file+0x90/0xb0 [nfsd]
<4> [<ffffffffa03c00ef>] ? nfsd4_encode_read+0x13f/0x240 [nfsd]
<4> [<ffffffffa03c5dd4>] ? nfs4_preprocess_stateid_op+0x284/0x310 [nfsd]
<4> [<ffffffffa03b9d35>] ? nfsd4_encode_operation+0x75/0x180 [nfsd]
<4> [<ffffffffa03b7d35>] ? nfsd4_proc_compound+0x195/0x490 [nfsd]
<4> [<ffffffffa03a543e>] ? nfsd_dispatch+0xfe/0x240 [nfsd]
<4> [<ffffffffa02d0614>] ? svc_process_common+0x344/0x640 [sunrpc]
<4> [<ffffffff81063410>] ? default_wake_function+0x0/0x20
<4> [<ffffffffa02d0c50>] ? svc_process+0x110/0x160 [sunrpc]
<4> [<ffffffffa03a5b62>] ? nfsd+0xc2/0x160 [nfsd]
<4> [<ffffffffa03a5aa0>] ? nfsd+0x0/0x160 [nfsd]
<4> [<ffffffff81096a36>] ? kthread+0x96/0xa0
<4> [<ffffffff8100c0ca>] ? child_rip+0xa/0x20
<4> [<ffffffff810969a0>] ? kthread+0x0/0xa0
<4> [<ffffffff8100c0c0>] ? child_rip+0x0/0x20

From dmesg on the MDS/MGS (c04):

Lustre: MGS: haven't heard from client 69d18973-27a2-0441-1720-b772711670c2 (at 192.168.2.106@o2ib) in 236 seconds. I think it's dead, and I am evicting it. exp ffff880664582400, cur 1383170178 expire 1383170028 last 1383169942
Lustre: scratch-MDT0000: haven't heard from client 63f330f5-0403-f077-ec5e-2a25cace70a6 (at 192.168.2.106@o2ib) in 239 seconds. I think it's dead, and I am evicting it. exp ffff88082e2c9c00, cur 1383170203 expire 1383170053 last 138316996

dmesg on one of the clients(c03) running fsx:

nf_conntrack version 0.5.0 (16384 buckets, 65536 max)
nfs: server c06-ib.lab.opensfs.org not responding, still trying
nfs: server c06-ib.lab.opensfs.org OK

dmesg on another client(c11) running fsx:

nf_conntrack version 0.5.0 (16384 buckets, 65536 max)
ip_tables: (C) 2000-2006 Netfilter Core Team
nf_conntrack version 0.5.0 (16384 buckets, 65536 max)
INFO: task fsx:7077 blocked for more than 120 seconds.
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
fsx           D 0000000000000002     0  7077   6831 0x00000080
 ffff880822b21c28 0000000000000082 ffff880822b21ba8 ffffffffa02cbcd0
 ffff880823ec7c00 ffff880822b21bd8 ffff880823ec7cb0 ffff880823161408
 ffff88082482a638 ffff880822b21fd8 000000000000fb88 ffff88082482a638
Call Trace:
 [<ffffffffa02cbcd0>] ? rpc_execute+0x50/0xa0 [sunrpc]
 [<ffffffff810a2431>] ? ktime_get_ts+0xb1/0xf0
 [<ffffffff81119e10>] ? sync_page+0x0/0x50
 [<ffffffff8150e8c3>] io_schedule+0x73/0xc0
 [<ffffffff81119e4d>] sync_page+0x3d/0x50
 [<ffffffff8150f27f>] __wait_on_bit+0x5f/0x90
 [<ffffffff8111a083>] wait_on_page_bit+0x73/0x80
 [<ffffffff81096de0>] ? wake_bit_function+0x0/0x50
 [<ffffffff8112f0e5>] ? pagevec_lookup_tag+0x25/0x40
 [<ffffffff8111a4ab>] wait_on_page_writeback_range+0xfb/0x190
 [<ffffffff8111a56f>] filemap_fdatawait+0x2f/0x40
 [<ffffffff8111ab94>] filemap_write_and_wait+0x44/0x60
 [<ffffffffa032ddfd>] nfs_getattr+0x10d/0x120 [nfs]
 [<ffffffff81186d21>] vfs_getattr+0x51/0x80
 [<ffffffff81186fdf>] vfs_fstat+0x3f/0x60
 [<ffffffff81187024>] sys_newfstat+0x24/0x40
 [<ffffffff810dc937>] ? audit_syscall_entry+0x1d7/0x200
 [<ffffffff810dc685>] ? __audit_syscall_exit+0x265/0x290
 [<ffffffff8100b072>] system_call_fastpath+0x16/0x1b
nfs: server c06-ib.lab.opensfs.org not responding, still trying
INFO: task fsx:7077 blocked for more than 120 seconds.
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
fsx           D 0000000000000002     0  7077   6831 0x00000080
 ffff880822b21c28 0000000000000082 ffff880822b21ba8 ffffffffa02cbcd0
 ffff880823ec7c00 ffff880822b21bd8 ffff880823ec7cb0 ffff880823161408
 ffff88082482a638 ffff880822b21fd8 000000000000fb88 ffff88082482a638
Call Trace:
 [<ffffffffa02cbcd0>] ? rpc_execute+0x50/0xa0 [sunrpc]
 [<ffffffff810a2431>] ? ktime_get_ts+0xb1/0xf0
 [<ffffffff81119e10>] ? sync_page+0x0/0x50
 [<ffffffff8150e8c3>] io_schedule+0x73/0xc0
 [<ffffffff81119e4d>] sync_page+0x3d/0x50
 [<ffffffff8150f27f>] __wait_on_bit+0x5f/0x90
 [<ffffffff8111a083>] wait_on_page_bit+0x73/0x80
 [<ffffffff81096de0>] ? wake_bit_function+0x0/0x50
 [<ffffffff8112f0e5>] ? pagevec_lookup_tag+0x25/0x40
 [<ffffffff8111a4ab>] wait_on_page_writeback_range+0xfb/0x190
 [<ffffffff8111a56f>] filemap_fdatawait+0x2f/0x40
 [<ffffffff8111ab94>] filemap_write_and_wait+0x44/0x60
 [<ffffffffa032ddfd>] nfs_getattr+0x10d/0x120 [nfs]
 [<ffffffff81186d21>] vfs_getattr+0x51/0x80
 [<ffffffff81186fdf>] vfs_fstat+0x3f/0x60
 [<ffffffff81187024>] sys_newfstat+0x24/0x40
 [<ffffffff810dc937>] ? audit_syscall_entry+0x1d7/0x200
 [<ffffffff810dc685>] ? __audit_syscall_exit+0x265/0x290
 [<ffffffff8100b072>] system_call_fastpath+0x16/0x1b
INFO: task fsx:7077 blocked for more than 120 seconds.
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
fsx           D 0000000000000002     0  7077   6831 0x00000080
 ffff880822b21c28 0000000000000082 ffff880822b21ba8 ffffffffa02cbcd0
 ffff880823ec7c00 ffff880822b21bd8 ffff880823ec7cb0 ffff880823161408
 ffff88082482a638 ffff880822b21fd8 000000000000fb88 ffff88082482a638
Call Trace:
 [<ffffffffa02cbcd0>] ? rpc_execute+0x50/0xa0 [sunrpc]
 [<ffffffff810a2431>] ? ktime_get_ts+0xb1/0xf0
 [<ffffffff81119e10>] ? sync_page+0x0/0x50
 [<ffffffff8150e8c3>] io_schedule+0x73/0xc0
 [<ffffffff81119e4d>] sync_page+0x3d/0x50
 [<ffffffff8150f27f>] __wait_on_bit+0x5f/0x90
 [<ffffffff8111a083>] wait_on_page_bit+0x73/0x80
 [<ffffffff81096de0>] ? wake_bit_function+0x0/0x50
 [<ffffffff8112f0e5>] ? pagevec_lookup_tag+0x25/0x40
 [<ffffffff8111a4ab>] wait_on_page_writeback_range+0xfb/0x190
 [<ffffffff8111a56f>] filemap_fdatawait+0x2f/0x40
 [<ffffffff8111ab94>] filemap_write_and_wait+0x44/0x60
 [<ffffffffa032ddfd>] nfs_getattr+0x10d/0x120 [nfs]
 [<ffffffff81186d21>] vfs_getattr+0x51/0x80
 [<ffffffff81186fdf>] vfs_fstat+0x3f/0x60
 [<ffffffff81187024>] sys_newfstat+0x24/0x40
 [<ffffffff810dc937>] ? audit_syscall_entry+0x1d7/0x200
 [<ffffffff810dc685>] ? __audit_syscall_exit+0x265/0x290
 [<ffffffff8100b072>] system_call_fastpath+0x16/0x1b
nfs: server c06-ib.lab.opensfs.org OK

On the above two clients running fsx, there is some information from fsx on the console.

From c11:

21166179: 1383169965.726634 TRUNCATE DOWN       from 0x3f79b to 0x1faa0
21166180: 1383169965.727509 WRITE    0xf5 thru 0xe785 (0xe691 bytes)
21166181: 1383169965.728551 READ     0xcc40 thru 0x1048d (0x384e bytes)
21166182: 1383169965.728579 TRUNCATE UP from 0x1faa0 to 0x3d30f
21166183: 1383169965.729270 WRITE    0x1efa1 thru 0x1ff22 (0xf82 bytes)
21166184: 1383169965.729569 MAPWRITE 0x24b11 thru 0x3137f (0xc86f bytes)
21166185: 1383169965.732139 WRITE    0x3d715 thru 0x3ffff (0x28eb bytes) HOLE
21166186: 1383169965.732569 READ     0x36ae3 thru 0x3bebc (0x53da bytes)
21166187: 1383169965.732607 WRITE    0x2bde3 thru 0x2edce (0x2fec bytes)
21166188: 1383169965.733019 TRUNCATE DOWN       from 0x40000 to 0x22c5a
21166189: 1383169965.733905 READ     0x5aa2 thru 0x111fd (0xb75c bytes)
21166190: 1383169965.733984 MAPREAD  0x8954 thru 0xf80c (0x6eb9 bytes)
21166191: 1383169965.734046 TRUNCATE UP from 0x22c5a to 0x2d1ff
21166192: 1383169965.734796 TRUNCATE DOWN       from 0x2d1ff to 0x14b1c
21166193: 1383169965.735632 READ     0x13752 thru 0x14b1b (0x13ca bytes)
21166194: 1383169965.735643 READ     0xec90 thru 0x14b1b (0x5e8c bytes)
21166195: 1383169965.735682 MAPWRITE 0x16fcc thru 0x22e5b (0xbe90 bytes)
21166196: 1383169965.738445 TRUNCATE DOWN       from 0x22e5c to 0x946a
21166197: 1383169965.739279 MAPWRITE 0x1f64b thru 0x23e6a (0x4820 bytes)
21166198: 1383169965.742263 MAPREAD  0x14df4 thru 0x1855c (0x3769 bytes)
21166199: 1383169965.742303 MAPWRITE 0x11972 thru 0x21497 (0xfb26 bytes)
Correct content saved for comparison
(maybe hexdump "/misc/export/nfs_x_2" vs "/misc/export/nfs_x_2.fsxgood")

on c03 a very long listing ending in:

…
2170216: 1383169960.525813 MAPWRITE 0x32e000 thru 0x33c2c1 (0xe2c2 bytes)
2170217: 1383169960.526777 TRUNCATE DOWN	from 0x743f40 to 0x1fe27c
2170218: 1383169960.531574 MAPREAD  0x1c02b4 thru 0x1cc215 (0xbf62 bytes)
2170219: 1383169960.531687 CLOSE/OPEN
2170220: 1383169960.532293 TRUNCATE UP	from 0x1fe27c to 0x68dab9
2170221: 1383169960.533532 MAPWRITE 0x343000 thru 0x34c4d4 (0x94d5 bytes)
2170222: 1383169960.800438 CLOSE/OPEN
2170223: 1383169960.800480 MAPREAD  0x61c245 thru 0x620e19 (0x4bd5 bytes)
2170224: 1383169960.800536 CLOSE/OPEN
2170225: 1383169960.800563 WRITE    0x7c000 thru 0x8b4de (0xf4df bytes)
2170226: 1383169960.801359 CLOSE/OPEN
2170227: 1383169960.801386 TRUNCATE DOWN	from 0x68dab9 to 0x206e69
Correct content saved for comparison
(maybe hexdump "/misc/export/nfs_x_4" vs "/misc/export/nfs_x_4.fsxgood")


 Comments   
Comment by Phil Schwan (Inactive) [ 28/Dec/13 ]

We're seeing this – often – in a non-NFS-server situation, just with standard pread64() calls.

LustreError: 18743:0:(rw.c:695:ll_read_ahead_pages()) ASSERTION( page_idx > ria->ria_stoff ) failed: invalid page_idx 176690rs 176690 re 176895 ro 176779 rl 89 rp 1
LustreError: 18743:0:(rw.c:695:ll_read_ahead_pages()) LBUG

Call Trace:
[<ffffffffa00c4905>] libcfs_debug_dumpstack
[<ffffffffa00c4f07>] lbug_with_loc
[<ffffffffa06ba4c6>] ll_readahead+0x10c6/0x10f0 [lustre]
[<ffffffff8126efe7>] ? radix_tree_insert
[<ffffffffa06e6d45>] vvp_io_read_page+0x305/0x340 [lustre]
[<ffffffffa023ba8d>] cl_io_read_page+0x8d/0x170 [obdclass]
[<ffffffffa022f887>] ? cl_read_page_assume+0xf7/0x220 [obdclass]
[<ffffffffa06b8e66>] ll_readpage+0x96/0x1a0 [lustre]
[<ffffffff811117ec>] generic_file_aio_read+0x1fc/0x700
[<ffffffffa06e8767>] vvp_io_read_start+0x257/0x470 [lustre]
[<ffffffffa02395fa>] cl_io_start+0x6a/0x140 [obdclass]
[<ffffffffa023dd04>] cl_io_loop+0xb4/0x1b0 [obdclass]
[<ffffffffa068a3cf>] ll_file_io_generic+0x3ff/0x610 [lustre]
[<ffffffffa068a7df>] ll_file_aio_read+0x13f/0x2c0 [lustre]
[<ffffffffa068aeac>] ll_file_read+0x16c/0x2a0 [lustre]
[<ffffffff81176cb5>] vfs_read
[<ffffffff81176fe2>] sys_pread64

One particular application seems to reproduce it fairly reliably; in 8 hours running on 35 nodes, 4 nodes crashed in this way.

I don't think this is very minor

Lustre: Lustre: Build Version: 2.5.0-RC1--PRISTINE-2.6.32-279.14.1.el6.x86_64
Linux hnod0032 2.6.32-279.19.1.el6.x86_64 #1 SMP Wed Dec 19 07:05:20 UTC 2012 x86_64 x86_64 x86_64 GNU/Linux

Comment by Phil Schwan (Inactive) [ 29/Dec/13 ]

We've lost 10 more nodes to this.

I'm going to try setting max_read_ahead_mb to 0, in the meantime, just to see if it keeps the cluster alive.

Comment by Phil Schwan (Inactive) [ 30/Dec/13 ]

Well, the good news is that it does indeed stop crashing if I disable readahead. I guess that's predictable.

The bad news is our apps were getting a lot of juice from readahead: a part that used to take 4 minutes and represent about 3% of the runtime now takes 80 minutes and represents 40% of runtime. So that's not really tenable...

Comment by Henri Doreau (Inactive) [ 28/Apr/14 ]

Was this further investigated? We see it regularly at CEA on a 2.1 instance which is re-exported to NFS.

Comment by Dan Tascione [ 08/May/14 ]

We are seeing the same problem daily on our Lustre 2.4.2 setup. This is also in a non-NFS situation, and we are just doing standard pread calls.

The interesting part is that it appears to only be happening for us when we access one particular file. Every time a client panics, it is in relation to reading from this file. However, it doesn't seem to happen every time we access the file, just occasionally. Only a single node reads or writes to this file at a time, and the read/write load (for that file) is typically very small.

I haven't tried turning off read ahead yet – that has had a very negative impact on performance for us in the past.

Comment by Zhenyu Xu [ 01/Jul/14 ]

Would you please try this patch http://review.whamcloud.com/10915 , it can also apply to 2.5 lustre code.

Comment by Dan Tascione [ 15/Aug/14 ]

We applied that patch here (on 2.4.2) and have been running for a few weeks – so far, so good!

Comment by Andreas Dilger [ 07/Oct/16 ]

Fixed as a duplicate of LU-5263.

Generated at Sat Feb 10 01:40:30 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.