[LU-2468] MDS out of memory, blocked in ldlm_pools_shrink() Created: 11/Dec/12 Updated: 28/Oct/15 Resolved: 28/Oct/15 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.4.0 |
| Fix Version/s: | Lustre 2.4.0 |
| Type: | Bug | Priority: | Major |
| Reporter: | Maloo | Assignee: | Sarah Liu |
| Resolution: | Fixed | Votes: | 0 |
| Labels: | None | ||
| Attachments: |
|
||||||||||||
| Issue Links: |
|
||||||||||||
| Severity: | 3 | ||||||||||||
| Rank (Obsolete): | 5814 | ||||||||||||
| Description |
|
This issue was created by maloo for Andreas Dilger <adilger@whamcloud.com> This issue relates to the following test suite run: https://maloo.whamcloud.com/test_sets/35a3873c-4166-11e2-af91-52540035b04c. The sub-test test_24a failed with the following error on the MDS console:
It is likely that this is only a symptom of something else consuming memory on the MDS, and not the root cause. Info required for matching: conf-sanity 24a |
| Comments |
| Comment by Oleg Drokin [ 11/Dec/12 ] |
|
It looks like we have some low-memory deadlock between ldlm pool thread and something else. And in addition somethign started to use more RAM of course. her's my report from tonight's testing, OOM suddenly started in the middle of sanity test60, which is a kernel-module test and should not be leaking anything I would imagine: PID: 11332 TASK: ffff880026c964c0 CPU: 1 COMMAND: "mdt00_004" #0 [ffff8800104837d0] schedule at ffffffff814f7c98 #1 [ffff880010483828] _spin_lock at ffffffff814fafde #2 [ffff880010483908] mutex_lock at ffffffff814f97e1 #3 [ffff880010483928] ldlm_pools_shrink at ffffffffa0ef515c [ptlrpc] #4 [ffff880010483978] ldlm_pools_cli_shrink at ffffffffa0ef5423 [ptlrpc] #5 [ffff880010483988] shrink_slab at ffffffff8112b5aa #6 [ffff8800104839e8] do_try_to_free_pages at ffffffff8112d5cf #7 [ffff880010483a78] try_to_free_pages at ffffffff8112d9dd #8 [ffff880010483b28] __alloc_pages_nodemask at ffffffff8112570c #9 [ffff880010483c38] kmem_getpages at ffffffff81160a62 #10 [ffff880010483c68] fallback_alloc at ffffffff8116349c #11 [ffff880010483ce8] ____cache_alloc_node at ffffffff81163188 #12 [ffff880010483d58] __kmalloc at ffffffff81163838 #13 [ffff880010483da8] cfs_alloc at ffffffffa08f7c10 [libcfs] #14 [ffff880010483dd8] ptlrpc_at_check_timed at ffffffffa0f1e361 [ptlrpc] #15 [ffff880010483e98] ptlrpc_main at ffffffffa0f1fd30 [ptlrpc] #16 [ffff880010483f48] kernel_thread at ffffffff8100c14a PID: 5875 TASK: ffff880098692400 CPU: 2 COMMAND: "lc_watchdogd" #0 [ffff88001d9ab600] schedule at ffffffff814f7c98 #1 [ffff88001d9ab678] cfs_hash_bd_lookup_intent at ffffffffa090ce97 [libcfs] #2 [ffff88001d9ab6d8] _spin_lock at ffffffff814fafde #3 [ffff88001d9ab6e8] cfs_hash_spin_lock at ffffffffa090d89e [libcfs] #4 [ffff88001d9ab6f8] lu_site_stats_get at ffffffffa0dd5b70 [obdclass] #5 [ffff88001d9ab758] lu_cache_shrink at ffffffffa0dd782d [obdclass] #6 [ffff88001d9ab7d8] shrink_slab at ffffffff8112b5aa #7 [ffff88001d9ab838] do_try_to_free_pages at ffffffff8112d5cf #8 [ffff88001d9ab8c8] try_to_free_pages at ffffffff8112d9dd #9 [ffff88001d9ab978] __alloc_pages_nodemask at ffffffff8112570c #10 [ffff88001d9aba88] kmem_getpages at ffffffff81160a62 #11 [ffff88001d9abab8] fallback_alloc at ffffffff8116349c #12 [ffff88001d9abb38] ____cache_alloc_node at ffffffff81163188 #13 [ffff88001d9abba8] kmem_cache_alloc at ffffffff81163c6b #14 [ffff88001d9abbf8] get_empty_filp at ffffffff8117c829 #15 [ffff88001d9abc18] do_filp_open at ffffffff8118bd35 #16 [ffff88001d9abd78] filp_open at ffffffff8118c88d #17 [ffff88001d9abd88] cfs_filp_open at ffffffffa08fc434 [libcfs] #18 [ffff88001d9abdb8] cfs_tracefile_dump_all_pages at ffffffffa09080f4 [libcfs] #19 [ffff88001d9abe48] libcfs_debug_dumplog_internal at ffffffffa0902445 [libcfs] #20 [ffff88001d9abe78] lc_watchdog_dumplog at ffffffffa09089a1 [libcfs] #21 [ffff88001d9abe88] lcw_dispatch_main at ffffffffa090a0db [libcfs] #22 [ffff88001d9abf48] kernel_thread at ffffffff8100c14a I am not really sure how come spinlock comes after that path or why tasks with schedule as the last call sit on CPU (this is bt -a output in crash) |
| Comment by Sarah Liu [ 13/Dec/12 ] |
|
another failure https://maloo.whamcloud.com/test_sets/92d0fe6c-44a9-11e2-8c8b-52540035b04c |
| Comment by Sarah Liu [ 13/Dec/12 ] |
|
another MDS OOM error seen in performance-sanity test: 10:57:59:Lustre: DEBUG MARKER: == performance-sanity test 3: small file create/open/delete ======== 10:57:53 (1355252273) 10:57:59:Lustre: DEBUG MARKER: /usr/sbin/lctl mark ===== mdsrate-create-small.sh ====== 10:57:59:Lustre: DEBUG MARKER: ===== mdsrate-create-small.sh ====== 10:57:59:Lustre: DEBUG MARKER: running=$(grep -c /mnt/mds1' ' /proc/mounts); 10:57:59:mpts=$(mount | grep -c /mnt/mds1' '); 10:57:59:if [ $running -ne $mpts ]; then 10:57:59: echo $(hostname) env are INSANE!; 10:57:59: exit 1; 10:57:59:fi 10:57:59:Lustre: DEBUG MARKER: running=$(grep -c /mnt/mds1' ' /proc/mounts); 10:57:59:mpts=$(mount | grep -c /mnt/mds1' '); 10:57:59:if [ $running -ne $mpts ]; then 10:57:59: echo $(hostname) env are INSANE!; 10:57:59: exit 1; 10:57:59:fi 10:57:59:Lustre: DEBUG MARKER: e2label /dev/lvm-MDS/P1 2>/dev/null 10:58:00:Lustre: DEBUG MARKER: cat /proc/mounts 10:58:11:Lustre: DEBUG MARKER: lctl get_param -n timeout 10:58:11:Lustre: DEBUG MARKER: /usr/sbin/lctl mark Using TIMEOUT=20 10:58:11:Lustre: DEBUG MARKER: Using TIMEOUT=20 10:58:11:Lustre: DEBUG MARKER: lctl dl | grep ' IN osc ' 2>/dev/null | wc -l 10:58:11:Lustre: DEBUG MARKER: /usr/sbin/lctl conf_param lustre.sys.jobid_var=procname_uid 10:58:11:Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n version 10:58:12:Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0000.quota_slave.enabled 10:58:12:Lustre: DEBUG MARKER: /usr/sbin/lctl conf_param lustre.quota.mdt=ug3 10:58:12:Lustre: DEBUG MARKER: /usr/sbin/lctl conf_param lustre.quota.ost=ug3 10:58:12:Lustre: DEBUG MARKER: PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/us 10:58:12:__ratelimit: 1 callbacks suppressed 10:58:12:cannot allocate a tage (259) 10:58:12:LNet: 12909:0:(debug.c:324:libcfs_debug_str2mask()) You are trying to use a numerical value for the mask - this will be deprecated in a future release. 10:58:12:LNet: 12909:0:(debug.c:324:libcfs_debug_str2mask()) Skipped 1 previous similar message 10:58:12:cannot allocate a tage (259) 10:58:12:Lustre: ctl-lustre-MDT0000: super-sequence allocation rc = 0 [0x0000000200000400-0x0000000240000400):0:mdt 10:58:12:cannot allocate a tage (259) 10:58:12:Lustre: DEBUG MARKER: /usr/sbin/lctl mark ===== mdsrate-create-small.sh ### 1 NODE CREATE ### 10:58:12:cannot allocate a tage (259) 10:58:12:Lustre: DEBUG MARKER: ===== mdsrate-create-small.sh 10:58:12:cannot allocate a tage (259) 10:58:12:cannot allocate a tage (259) 10:58:12:cannot allocate a tage (259) 10:58:12:cannot allocate a tage (259) 10:58:12:cannot allocate a tage (259) 10:58:12:cannot allocate a tage (259) 11:02:36:Lustre: DEBUG MARKER: /usr/sbin/lctl mark ===== mdsrate-create-small.sh ### 1 NODE UNLINK ### 11:02:37:Lustre: DEBUG MARKER: ===== mdsrate-create-small.sh 11:08:21:Lustre: DEBUG MARKER: /usr/sbin/lctl mark ===== mdsrate-create-small.sh ### 2 NODES CREATE with 3 threads per client ### 11:08:21:Lustre: DEBUG MARKER: ===== mdsrate-create-small.sh 11:09:42:__ratelimit: 12 callbacks suppressed 11:09:42:sssd_nss invoked oom-killer: gfp_mask=0x201da, order=0, oom_adj=0, oom_score_adj=0 11:09:42:sssd_nss cpuset=/ mems_allowed=0 11:09:42:Pid: 1831, comm: sssd_nss Not tainted 2.6.32-279.14.1.el6_lustre.x86_64 #1 11:09:42:Call Trace: 11:09:42: [<ffffffff810c4cd1>] ? cpuset_print_task_mems_allowed+0x91/0xb0 11:09:42: [<ffffffff81117440>] ? dump_header+0x90/0x1b0 11:09:42: [<ffffffff81500d07>] ? _spin_unlock_irqrestore+0x17/0x20 11:09:42: [<ffffffff811178c2>] ? oom_kill_process+0x82/0x2a0 11:09:42: [<ffffffff811177be>] ? select_bad_process+0x9e/0x120 11:09:42: [<ffffffff81117d00>] ? out_of_memory+0x220/0x3c0 11:09:42: [<ffffffff81127a1e>] ? __alloc_pages_nodemask+0x89e/0x940 11:09:42: [<ffffffff8115c7ea>] ? alloc_pages_current+0xaa/0x110 11:09:42: [<ffffffff81114847>] ? __page_cache_alloc+0x87/0x90 11:09:42: [<ffffffff8112a46b>] ? __do_page_cache_readahead+0xdb/0x210 11:09:42: [<ffffffff8112a5c1>] ? ra_submit+0x21/0x30 11:09:42: [<ffffffff81115b73>] ? filemap_fault+0x4c3/0x500 11:09:42: [<ffffffff8113ef74>] ? __do_fault+0x54/0x510 11:09:42: [<ffffffff8113f527>] ? handle_pte_fault+0xf7/0xb50 11:09:42: [<ffffffff811401ba>] ? handle_mm_fault+0x23a/0x310 11:09:42: [<ffffffff810444c9>] ? __do_page_fault+0x139/0x480 11:09:42: [<ffffffff81039688>] ? pvclock_clocksource_read+0x58/0xd0 11:09:42: [<ffffffff8103877c>] ? kvm_clock_read+0x1c/0x20 11:09:42: [<ffffffff81038789>] ? kvm_clock_get_cycles+0x9/0x10 11:09:42: [<ffffffff8109cb30>] ? getnstimeofday+0x60/0xf0 11:09:42: [<ffffffff81503e0e>] ? do_page_fault+0x3e/0xa0 11:09:42: [<ffffffff815011c5>] ? page_fault+0x25/0x30 11:09:43:Mem-Info: 11:09:43:Node 0 DMA per-cpu: 11:09:43:CPU 0: hi: 0, btch: 1 usd: 0 11:09:43:CPU 1: hi: 0, btch: 1 usd: 0 11:09:43:CPU 2: hi: 0, btch: 1 usd: 0 11:09:43:CPU 3: hi: 0, btch: 1 usd: 0 11:09:43:CPU 4: hi: 0, btch: 1 usd: 0 11:09:43:CPU 5: hi: 0, btch: 1 usd: 0 11:09:43:CPU 6: hi: 0, btch: 1 usd: 0 11:09:43:CPU 7: hi: 0, btch: 1 usd: 0 11:09:43:CPU 8: hi: 0, btch: 1 usd: 0 11:09:43:CPU 9: hi: 0, btch: 1 usd: 0 11:09:43:CPU 10: hi: 0, btch: 1 usd: 0 11:09:43:CPU 11: hi: 0, btch: 1 usd: 0 11:09:43:CPU 12: hi: 0, btch: 1 usd: 0 11:09:43:CPU 13: hi: 0, btch: 1 usd: 0 11:09:43:CPU 14: hi: 0, btch: 1 usd: 0 11:09:43:CPU 15: hi: 0, btch: 1 usd: 0 11:09:43:CPU 16: hi: 0, btch: 1 usd: 0 11:09:43:CPU 17: hi: 0, btch: 1 usd: 0 11:09:43:CPU 18: hi: 0, btch: 1 usd: 0 11:09:43:CPU 19: hi: 0, btch: 1 usd: 0 11:09:43:CPU 20: hi: 0, btch: 1 usd: 0 11:09:43:CPU 21: hi: 0, btch: 1 usd: 0 11:09:43:CPU 22: hi: 0, btch: 1 usd: 0 11:09:43:CPU 23: hi: 0, btch: 1 usd: 0 11:09:43:Node 0 DMA32 per-cpu: 11:09:43:CPU 0: hi: 186, btch: 31 usd: 0 11:09:43:CPU 1: hi: 186, btch: 31 usd: 0 11:09:43:CPU 2: hi: 186, btch: 31 usd: 0 11:09:43:CPU 3: hi: 186, btch: 31 usd: 0 11:09:43:CPU 4: hi: 186, btch: 31 usd: 0 11:09:43:CPU 5: hi: 186, btch: 31 usd: 0 11:09:43:CPU 6: hi: 186, btch: 31 usd: 0 11:09:43:CPU 7: hi: 186, btch: 31 usd: 0 11:09:43:CPU 8: hi: 186, btch: 31 usd: 0 11:09:43:CPU 9: hi: 186, btch: 31 usd: 0 11:09:43:CPU 10: hi: 186, btch: 31 usd: 0 11:09:43:CPU 11: hi: 186, btch: 31 usd: 0 11:09:43:CPU 12: hi: 186, btch: 31 usd: 0 11:09:43:CPU 13: hi: 186, btch: 31 usd: 0 11:09:43:CPU 14: hi: 186, btch: 31 usd: 0 11:09:43:CPU 15: hi: 186, btch: 31 usd: 0 11:09:43:CPU 16: hi: 186, btch: 31 usd: 0 11:09:43:CPU 17: hi: 186, btch: 31 usd: 0 11:09:43:CPU 18: hi: 186, btch: 31 usd: 0 11:09:43:CPU 19: hi: 186, btch: 31 usd: 0 11:09:43:CPU 20: hi: 186, btch: 31 usd: 0 11:09:43:CPU 21: hi: 186, btch: 31 usd: 0 11:09:43:CPU 22: hi: 186, btch: 31 usd: 0 11:09:43:CPU 23: hi: 186, btch: 31 usd: 0 11:09:43:active_anon:53 inactive_anon:100 isolated_anon:0 11:09:43: active_file:8727 inactive_file:8692 isolated_file:0 11:09:43: unevictable:0 dirty:0 writeback:0 unstable:0 11:09:43: free:13216 slab_reclaimable:33653 slab_unreclaimable:33437 11:09:43: mapped:36 shmem:0 pagetables:750 bounce:0 11:09:43:Node 0 DMA free:8352kB min:332kB low:412kB high:496kB active_anon:8kB inactive_anon:8kB active_file:640kB inactive_file:196kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15324kB mlocked:0kB dirty:0kB writeback:0kB mapped:4kB shmem:0kB slab_reclaimable:4852kB slab_unreclaimable:916kB kernel_stack:24kB pagetables:0kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:1284 all_unreclaimable? yes 11:09:43:lowmem_reserve[]: 0 2003 2003 2003 11:09:43:Node 0 DMA32 free:44512kB min:44720kB low:55900kB high:67080kB active_anon:204kB inactive_anon:392kB active_file:34268kB inactive_file:34572kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:2052060kB mlocked:0kB dirty:0kB writeback:0kB mapped:140kB shmem:0kB slab_reclaimable:129760kB slab_unreclaimable:132832kB kernel_stack:5624kB pagetables:3000kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:148106 all_unreclaimable? yes 11:09:43:lowmem_reserve[]: 0 0 0 0 11:09:43:Node 0 DMA: 147*4kB 124*8kB 68*16kB 28*32kB 9*64kB 3*128kB 1*256kB 3*512kB 0*1024kB 1*2048kB 0*4096kB = 8364kB 11:09:43:Node 0 DMA32: 3090*4kB 2086*8kB 673*16kB 90*32kB 2*64kB 1*128kB 1*256kB 1*512kB 1*1024kB 1*2048kB 0*4096kB = 46792kB 11:09:43:17779 total pagecache pages 11:09:43:38 pages in swap cache 11:09:43:Swap cache stats: add 4557, delete 4519, find 87/119 11:09:43:Free swap = 4111556kB 11:09:43:Total swap = 4128760kB 11:09:43:524283 pages RAM 11:09:43:44357 pages reserved 11:09:43:17679 pages shared 11:09:43:437072 pages non-shared 11:09:43:[ pid ] uid tgid total_vm rss cpu oom_adj oom_score_adj name 11:09:43:[ 707] 0 707 2761 0 23 -17 -1000 udevd 11:09:43:[ 1740] 0 1740 6909 1 4 -17 -1000 auditd 11:09:43:[ 1756] 0 1756 62271 1 0 0 0 rsyslogd 11:09:43:[ 1785] 0 1785 2286 1 4 0 0 irqbalance 11:09:43:[ 1799] 32 1799 4743 1 4 0 0 rpcbind 11:09:43:[ 1811] 0 1811 45434 38 17 0 0 sssd 11:09:43:[ 1824] 0 1824 50962 4 13 0 0 sssd_be 11:09:43:[ 1829] 29 1829 6354 1 0 0 0 rpc.statd 11:09:43:[ 1831] 0 1831 42905 23 3 0 0 sssd_nss 11:09:43:[ 1832] 0 1832 42854 2 5 0 0 sssd_pam 11:09:43:[ 1980] 81 1980 5868 1 4 0 0 dbus-daemon 11:09:43:[ 2013] 0 2013 1019 0 5 0 0 acpid 11:09:43:[ 2022] 68 2022 6806 1 4 0 0 hald 11:09:43:[ 2023] 0 2023 4525 1 4 0 0 hald-runner 11:09:43:[ 2051] 0 2051 5054 1 0 0 0 hald-addon-inpu 11:09:43:[ 2061] 68 2061 4450 1 5 0 0 hald-addon-acpi 11:09:43:[ 2082] 0 2082 167751 1 0 0 0 automount 11:09:43:[ 2123] 0 2123 26826 0 9 0 0 rpc.rquotad 11:09:43:[ 2127] 0 2127 5413 0 1 0 0 rpc.mountd 11:09:43:[ 2176] 0 2176 6290 1 1 0 0 rpc.idmapd 11:09:43:[ 2201] 0 2201 16017 0 1 -17 -1000 sshd 11:09:43:[ 2209] 0 2209 5522 1 1 0 0 xinetd 11:09:43:[ 2217] 38 2217 7004 1 4 0 0 ntpd 11:09:43:[ 2233] 0 2233 22181 35 2 0 0 sendmail 11:09:43:[ 2241] 51 2241 19527 0 1 0 0 sendmail 11:09:43:[ 2263] 0 2263 27017 1 4 0 0 abrt-dump-oops 11:09:43:[ 2271] 0 2271 29301 8 4 0 0 crond 11:09:43:[ 2282] 0 2282 5363 0 0 0 0 atd 11:09:43:[ 2295] 0 2295 1018 1 0 0 0 agetty 11:09:43:[ 2297] 0 2297 1015 1 7 0 0 mingetty 11:09:43:[ 2299] 0 2299 1015 1 16 0 0 mingetty 11:09:43:[ 2301] 0 2301 1015 1 7 0 0 mingetty 11:09:43:[ 2303] 0 2303 1015 1 16 0 0 mingetty 11:09:43:[ 2305] 0 2305 2760 0 8 -17 -1000 udevd 11:09:43:[ 2306] 0 2306 1015 1 7 0 0 mingetty 11:09:43:[ 2307] 0 2307 2760 0 12 -17 -1000 udevd 11:09:43:[ 2309] 0 2309 1015 1 21 0 0 mingetty 11:09:43:[12989] 0 12989 4324 0 4 0 0 anacron 11:09:43:Out of memory: Kill process 1756 (rsyslogd) score 1 or sacrifice child 11:09:43:Killed process 1756, UID 0, (rsyslogd) total-vm:249084kB, anon-rss:0kB, file-rss:4kB 11:09:43:sssd_be invoked oom-killer: gfp_mask=0x200da, order=0, oom_adj=0, oom_score_adj=0 11:09:43:sssd_be cpuset=/ mems_allowed=0 11:09:43:Pid: 1824, comm: sssd_be Not tainted 2.6.32-279.14.1.el6_lustre.x86_64 #1 |
| Comment by Keith Mannthey (Inactive) [ 13/Dec/12 ] |
|
Failure Rate: 31.00% of last 100 executions [all branches] |
| Comment by Peter Jones [ 27/Dec/12 ] |
|
Bobijam could you please look into this one? |
| Comment by Zhenyu Xu [ 28/Dec/12 ] |
|
found a glitch along the ldlm_shrink path, ldlm_pool_shrink() Shrinker path should not allocate memory again. While It's not the deadlock root cause though. |
| Comment by Niu Yawei (Inactive) [ 08/Jan/13 ] |
|
In cfs_hash_bd_lookup_intent(), shouldn't we use cfs_hlist_for_each_safe() instead of cfs_hlist_for_each()? Seems the item is possible to be deleted from the list, that'll corrupt the list. |
| Comment by Niu Yawei (Inactive) [ 08/Jan/13 ] |
|
oops, the function (cfs_hash_bd_lookup_intent()) returns immediately after deletion, so that's fine. |
| Comment by nasf (Inactive) [ 19/Feb/13 ] |
|
Another failure instance: https://maloo.whamcloud.com/test_sets/b1455ce2-7a57-11e2-b5c8-52540035b04c |
| Comment by Andreas Dilger [ 19/Feb/13 ] |
|
I pushed http://review.whamcloud.com/5470 for quieting the spurious "cannot allocate a tage" message, but this does not solve the root of the problem, which is that the MDS is consuming more and more memory. |
| Comment by Jodi Levi (Inactive) [ 13/Mar/13 ] |
|
Per discussions with BobiJam, lowering priority since http://review.whamcloud.com/4940 landed and believed to help. Will raise priority if further testing proves otherwise. |
| Comment by Sarah Liu [ 15/Mar/13 ] |
|
Found OOM in the interop test between 2.3 server and the latest tag-2.3.62 https://maloo.whamcloud.com/test_sets/422ac24e-8b95-11e2-abec-52540035b04c MDS console: 00:03:57:Lustre: DEBUG MARKER: == conf-sanity test 24a: Multiple MDTs on a single node == 00:03:40 (1363071820) 00:03:57:Lustre: DEBUG MARKER: grep -c /mnt/fs2mds' ' /proc/mounts 00:03:57:Lustre: DEBUG MARKER: lsmod | grep lnet > /dev/null && lctl dl | grep ' ST ' 00:03:57:Lustre: DEBUG MARKER: mkfs.lustre --mgsnode=client-31vm3@tcp --fsname=lustre --mdt --index=0 --param=sys.timeout=20 --param=lov.stripesize=1048576 --param=lov.stripecount=0 --param=mdt.identity_upcall=/usr/sbin/l_getidentity --backfstype=ldiskfs --device-size=2097152 --mkfsopt 00:03:57:LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. quota=off. Opts: 00:03:57:Lustre: DEBUG MARKER: mkdir -p /mnt/mds1 00:03:57:Lustre: DEBUG MARKER: test -b /dev/lvm-MDS/P1 00:03:57:Lustre: DEBUG MARKER: mkdir -p /mnt/mds1; mount -t lustre -o user_xattr,acl /dev/lvm-MDS/P1 /mnt/mds1 00:03:57:LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. quota=off. Opts: 00:03:57:LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. quota=off. Opts: 00:03:57:Lustre: DEBUG MARKER: PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/usr/lib64/lustre/tests//usr/lib64/lustre/tests:/usr/lib64/lustre/tests:/usr/lib64/lustre/tests/../utils:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lust 00:03:57:Lustre: DEBUG MARKER: e2label /dev/lvm-MDS/P1 2>/dev/null 00:04:08:Lustre: 3472:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1363071835/real 1363071835] req@ffff880066563800 x1429282250294701/t0(0) o8->lustre-OST0000-osc-MDT0000@10.10.4.191@tcp:28/4 lens 400/544 e 0 to 1 dl 1363071840 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1 00:04:08:Lustre: 3472:0:(client.c:1917:ptlrpc_expire_one_request()) Skipped 12 previous similar messages 00:04:10:Lustre: DEBUG MARKER: mkdir -p /mnt/fs2mds 00:04:11:Lustre: DEBUG MARKER: test -b /dev/lvm-MDS/S1 00:04:12:Lustre: DEBUG MARKER: mkdir -p /mnt/fs2mds; mount -t lustre -o user_xattr,acl /dev/lvm-MDS/S1 /mnt/fs2mds 00:04:12:LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. quota=off. Opts: 00:04:12:LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. quota=off. Opts: 00:04:12:Lustre: Setting parameter test1234-MDT0000-mdtlov.lov.stripesize in log test1234-MDT0000 00:04:12:Lustre: Skipped 4 previous similar messages 00:04:12:Lustre: test1234-MDT0000: new disk, initializing 00:04:12:__ratelimit: 23 callbacks suppressed 00:04:12:llog_process_th invoked oom-killer: gfp_mask=0xd2, order=0, oom_adj=-17, oom_score_adj=0 00:04:12:llog_process_th cpuset=/ mems_allowed=0 00:04:12:Pid: 900, comm: llog_process_th Not tainted 2.6.32-279.5.1.el6_lustre.gb16fe80.x86_64 #1 00:04:12:Call Trace: 00:04:12: [<ffffffff810c4aa1>] ? cpuset_print_task_mems_allowed+0x91/0xb0 00:04:12: [<ffffffff81117210>] ? dump_header+0x90/0x1b0 00:04:12: [<ffffffff81117692>] ? oom_kill_process+0x82/0x2a0 00:04:12: [<ffffffff8111758e>] ? select_bad_process+0x9e/0x120 00:04:12: [<ffffffff81117ad0>] ? out_of_memory+0x220/0x3c0 00:04:12: [<ffffffff811277ee>] ? __alloc_pages_nodemask+0x89e/0x940 00:04:12: [<ffffffff8114d789>] ? __vmalloc_area_node+0xb9/0x190 00:04:12: [<ffffffffa04df9e0>] ? cfs_cpt_vmalloc+0x20/0x30 [libcfs] 00:04:12: [<ffffffff8114d65d>] ? __vmalloc_node+0xad/0x120 00:04:12: [<ffffffffa04df9e0>] ? cfs_cpt_vmalloc+0x20/0x30 [libcfs] 00:04:12: [<ffffffff8114d9a9>] ? vmalloc_node+0x29/0x30 00:04:12: [<ffffffffa04df9e0>] ? cfs_cpt_vmalloc+0x20/0x30 [libcfs] 00:04:12: [<ffffffffa07e726e>] ? ptlrpc_alloc_rqbd+0x13e/0x690 [ptlrpc] 00:04:12: [<ffffffffa07e7825>] ? ptlrpc_grow_req_bufs+0x65/0x1b0 [ptlrpc] 00:04:12: [<ffffffffa07e8e07>] ? ptlrpc_register_service+0x1497/0x1770 [ptlrpc] 00:04:12: [<ffffffffa0e38c62>] ? mdt_init0+0x1202/0x23f0 [mdt] 00:04:12: [<ffffffffa0e39f43>] ? mdt_device_alloc+0xf3/0x220 [mdt] 00:04:12: [<ffffffffa064b0d7>] ? obd_setup+0x1d7/0x2f0 [obdclass] 00:04:12: [<ffffffffa064b3f8>] ? class_setup+0x208/0x890 [obdclass] 00:04:12: [<ffffffffa065308c>] ? class_process_config+0xc0c/0x1c30 [obdclass] 00:04:12: [<ffffffffa04ea088>] ? libcfs_log_return+0x28/0x40 [libcfs] 00:04:12: [<ffffffffa064cef1>] ? lustre_cfg_new+0x391/0x7e0 [obdclass] 00:04:12: [<ffffffffa065515b>] ? class_config_llog_handler+0x9bb/0x1610 [obdclass] 00:04:12: [<ffffffffa061e1f8>] ? llog_process_thread+0x888/0xd00 [obdclass] 00:04:12: [<ffffffffa061d970>] ? llog_process_thread+0x0/0xd00 [obdclass] 00:04:12: [<ffffffff8100c14a>] ? child_rip+0xa/0x20 00:04:12: [<ffffffffa061d970>] ? llog_process_thread+0x0/0xd00 [obdclass] 00:04:12: [<ffffffffa061d970>] ? llog_process_thread+0x0/0xd00 [obdclass] 00:04:12: [<ffffffff8100c140>] ? child_rip+0x0/0x20 00:04:12:Mem-Info: 00:04:12:Node 0 DMA per-cpu: 00:04:12:CPU 0: hi: 0, btch: 1 usd: 0 00:04:12:CPU 1: hi: 0, btch: 1 usd: 0 00:04:12:CPU 2: hi: 0, btch: 1 usd: 0 00:04:12:CPU 3: hi: 0, btch: 1 usd: 0 00:04:12:CPU 4: hi: 0, btch: 1 usd: 0 00:04:12:CPU 5: hi: 0, btch: 1 usd: 0 00:04:12:CPU 6: hi: 0, btch: 1 usd: 0 00:04:12:CPU 7: hi: 0, btch: 1 usd: 0 00:04:12:CPU 8: hi: 0, btch: 1 usd: 0 00:04:12:CPU 9: hi: 0, btch: 1 usd: 0 00:04:12:CPU 10: hi: 0, btch: 1 usd: 0 00:04:12:CPU 11: hi: 0, btch: 1 usd: 0 00:04:12:CPU 12: hi: 0, btch: 1 usd: 0 00:04:12:CPU 13: hi: 0, btch: 1 usd: 0 00:04:12:CPU 14: hi: 0, btch: 1 usd: 0 00:04:12:CPU 15: hi: 0, btch: 1 usd: 0 00:04:12:CPU 16: hi: 0, btch: 1 usd: 0 00:04:12:CPU 17: hi: 0, btch: 1 usd: 0 00:04:12:CPU 18: hi: 0, btch: 1 usd: 0 00:04:12:CPU 19: hi: 0, btch: 1 usd: 0 00:04:12:CPU 20: hi: 0, btch: 1 usd: 0 00:04:12:CPU 21: hi: 0, btch: 1 usd: 0 00:04:12:CPU 22: hi: 0, btch: 1 usd: 0 00:04:13:CPU 23: hi: 0, btch: 1 usd: 0 00:04:13:Node 0 DMA32 per-cpu: 00:04:13:CPU 0: hi: 186, btch: 31 usd: 64 00:04:13:CPU 1: hi: 186, btch: 31 usd: 0 00:04:13:CPU 2: hi: 186, btch: 31 usd: 0 00:04:13:CPU 3: hi: 186, btch: 31 usd: 0 00:04:13:CPU 4: hi: 186, btch: 31 usd: 0 00:04:13:CPU 5: hi: 186, btch: 31 usd: 0 00:04:13:CPU 6: hi: 186, btch: 31 usd: 0 00:04:13:CPU 7: hi: 186, btch: 31 usd: 0 00:04:13:CPU 8: hi: 186, btch: 31 usd: 0 00:04:13:CPU 9: hi: 186, btch: 31 usd: 0 00:04:13:CPU 10: hi: 186, btch: 31 usd: 0 00:04:13:CPU 11: hi: 186, btch: 31 usd: 0 00:04:13:CPU 12: hi: 186, btch: 31 usd: 0 00:04:13:CPU 13: hi: 186, btch: 31 usd: 0 00:04:13:CPU 14: hi: 186, btch: 31 usd: 0 00:04:13:CPU 15: hi: 186, btch: 31 usd: 0 00:04:13:CPU 16: hi: 186, btch: 31 usd: 0 00:04:13:CPU 17: hi: 186, btch: 31 usd: 0 00:04:13:CPU 18: hi: 186, btch: 31 usd: 0 00:04:13:CPU 19: hi: 186, btch: 31 usd: 0 00:04:13:CPU 20: hi: 186, btch: 31 usd: 0 00:04:13:CPU 21: hi: 186, btch: 31 usd: 0 00:04:13:CPU 22: hi: 186, btch: 31 usd: 0 00:04:13:CPU 23: hi: 186, btch: 31 usd: 58 00:04:13:active_anon:3696 inactive_anon:1307 isolated_anon:0 00:04:13: active_file:105 inactive_file:47 isolated_file:0 00:04:13: unevictable:0 dirty:72 writeback:927 unstable:0 00:04:13: free:13329 slab_reclaimable:2530 slab_unreclaimable:28916 00:04:13: mapped:0 shmem:41 pagetables:818 bounce:0 00:04:13:Node 0 DMA free:8348kB min:332kB low:412kB high:496kB active_anon:0kB inactive_anon:8kB active_file:188kB inactive_file:196kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:15324kB mlocked:0kB dirty:4kB writeback:208kB mapped:0kB shmem:0kB slab_reclaimable:176kB slab_unreclaimable:452kB kernel_stack:0kB pagetables:0kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:4552 all_unreclaimable? yes 00:04:13:lowmem_reserve[]: 0 2003 2003 2003 00:04:13:Node 0 DMA32 free:44968kB min:44720kB low:55900kB high:67080kB active_anon:14784kB inactive_anon:5220kB active_file:232kB inactive_file:0kB unevictable:0kB isolated(anon):0kB isolated(file):0kB present:2052060kB mlocked:0kB dirty:284kB writeback:3500kB mapped:0kB shmem:164kB slab_reclaimable:9944kB slab_unreclaimable:115212kB kernel_stack:5760kB pagetables:3272kB unstable:0kB bounce:0kB writeback_tmp:0kB pages_scanned:0 all_unreclaimable? no 00:04:13:lowmem_reserve[]: 0 0 0 0 00:04:13:Node 0 DMA: 49*4kB 43*8kB 38*16kB 23*32kB 11*64kB 5*128kB 2*256kB 3*512kB 1*1024kB 1*2048kB 0*4096kB = 8348kB 00:04:13:Node 0 DMA32: 1692*4kB 1057*8kB 583*16kB 269*32kB 91*64kB 16*128kB 5*256kB 1*512kB 1*1024kB 1*2048kB 0*4096kB = 45896kB 00:04:13:1404 total pagecache pages 00:04:13:1080 pages in swap cache 00:04:13:Swap cache stats: add 1506, delete 426, find 0/0 00:04:13:Free swap = 4122736kB 00:04:13:Total swap = 4128760kB 00:04:13:524283 pages RAM 00:04:13:44356 pages reserved 00:04:13:301 pages shared 00:04:13:455288 pages non-shared 00:04:13:[ pid ] uid tgid total_vm rss cpu oom_adj oom_score_adj name 00:04:13:[ 723] 0 723 2694 35 22 -17 -1000 udevd 00:04:13:[ 1779] 0 1779 6909 33 7 -17 -1000 auditd 00:04:13:[ 1795] 0 1795 62270 121 0 0 0 rsyslogd 00:04:13:[ 1824] 0 1824 2286 42 10 0 0 irqbalance 00:04:13:[ 1838] 32 1838 4743 17 1 0 0 rpcbind 00:04:13:[ 1856] 29 1856 5836 3 0 0 0 rpc.statd 00:04:13:[ 2005] 81 2005 5868 34 0 0 0 dbus-daemon 00:04:13:[ 2038] 0 2038 1019 0 13 0 0 acpid 00:04:13:[ 2047] 68 2047 6805 118 0 0 0 hald 00:04:13:[ 2048] 0 2048 4525 1 5 0 0 hald-runner 00:04:13:[ 2076] 0 2076 5054 1 0 0 0 hald-addon-inpu 00:04:13:[ 2083] 68 2083 4450 1 0 0 0 hald-addon-acpi 00:04:13:[ 2097] 0 2097 45434 37 0 0 0 sssd 00:04:13:[ 2100] 0 2100 50963 186 0 0 0 sssd_be 00:04:13:[ 2109] 0 2109 43037 114 1 0 0 sssd_nss 00:04:13:[ 2110] 0 2110 42892 1 2 0 0 sssd_pam 00:04:13:[ 2121] 0 2121 167752 323 0 0 0 automount 00:04:13:[ 2162] 0 2162 26826 1 5 0 0 rpc.rquotad 00:04:13:[ 2166] 0 2166 5413 144 11 0 0 rpc.mountd 00:04:13:[ 2215] 0 2215 6290 62 1 0 0 rpc.idmapd 00:04:13:[ 2258] 498 2258 56709 255 2 0 0 munged 00:04:13:[ 2273] 0 2273 16017 166 1 -17 -1000 sshd 00:04:13:[ 2281] 0 2281 5522 60 0 0 0 xinetd 00:04:13:[ 2289] 38 2289 7004 113 0 0 0 ntpd 00:04:13:[ 2305] 0 2305 22183 415 1 0 0 sendmail 00:04:13:[ 2313] 51 2313 19528 348 1 0 0 sendmail 00:04:13:[ 2335] 0 2335 27017 60 3 0 0 abrt-dump-oops 00:04:13:[ 2343] 0 2343 29303 152 6 0 0 crond 00:04:13:[ 2354] 0 2354 5363 45 0 0 0 atd 00:04:13:[ 2380] 0 2380 1018 23 0 0 0 agetty 00:04:13:[ 2381] 0 2381 1015 22 0 0 0 mingetty 00:04:13:[ 2383] 0 2383 1015 22 0 0 0 mingetty 00:04:13:[ 2385] 0 2385 1015 21 6 0 0 mingetty 00:04:14:[ 2387] 0 2387 1015 22 9 0 0 mingetty 00:04:15:[ 2389] 0 2389 1015 21 6 0 0 mingetty 00:04:15:[ 2390] 0 2390 2706 46 2 -17 -1000 udevd 00:04:15:[ 2391] 0 2391 2706 39 5 -17 -1000 udevd 00:04:17:[ 2393] 0 2393 1015 21 6 0 0 mingetty 00:04:17:[ 806] 0 806 15354 173 0 0 0 in.mrshd 00:04:17:[ 807] 0 807 26514 66 15 0 0 bash 00:04:17:[ 828] 0 828 26514 66 16 0 0 bash 00:04:17:[ 829] 0 829 26515 49 20 0 0 sh 00:04:17:[ 831] 0 831 27935 39 22 0 0 mount 00:04:17:[ 832] 0 832 2092 42 0 0 0 mount.lustre 00:04:18:Out of memory: Kill process 1795 (rsyslogd) score 1 or sacrifice child 00:04:18:Killed process 1795, UID 0, (rsyslogd) total-vm:249080kB, anon-rss:480kB, file-rss:4kB 00:04:18:llog_process_th invoked oom-killer: gfp_mask=0xd2, order=0, oom_adj=-17, oom_score_adj=0 00:04:18:llog_process_th cpuset=/ mems_allowed=0 00:04:18:Pid: 900, comm: llog_process_th Not tainted 2.6.32-279.5.1.el6_lustre.gb16fe80.x86_64 #1 00:04:19:Call Trace: 00:04:19: [<ffffffff810c4aa1>] ? cpuset_print_task_mems_allowed+0x91/0xb0 00:04:19: [<ffffffff81117210>] ? dump_header+0x90/0x1b0 00:04:19: [<ffffffff810e368e>] ? __delayacct_freepages_end+0x2e/0x30 00:04:19: [<ffffffff8121489c>] ? security_real_capable_noaudit+0x3c/0x70 00:04:19: [<ffffffff81117692>] ? oom_kill_process+0x82/0x2a0 00:04:20: [<ffffffff8111758e>] ? select_bad_process+0x9e/0x120 00:04:20: [<ffffffff81117ad0>] ? out_of_memory+0x220/0x3c0 00:04:20: [<ffffffff811277ee>] ? __alloc_pages_nodemask+0x89e/0x940 00:04:20: [<ffffffff8114d789>] ? __vmalloc_area_node+0xb9/0x190 00:04:20: [<ffffffffa04df9e0>] ? cfs_cpt_vmalloc+0x20/0x30 [libcfs] 00:04:20: [<ffffffff8114d65d>] ? __vmalloc_node+0xad/0x120 00:04:20: [<ffffffffa04df9e0>] ? cfs_cpt_vmalloc+0x20/0x30 [libcfs] 00:04:20: [<ffffffff8114d9a9>] ? vmalloc_node+0x29/0x30 00:04:20: [<ffffffffa04df9e0>] ? cfs_cpt_vmalloc+0x20/0x30 [libcfs] 00:04:20: [<ffffffffa07e726e>] ? ptlrpc_alloc_rqbd+0x13e/0x690 [ptlrpc] 00:04:20: [<ffffffffa07e7825>] ? ptlrpc_grow_req_bufs+0x65/0x1b0 [ptlrpc] 00:04:20: [<ffffffffa07e8e07>] ? ptlrpc_register_service+0x1497/0x1770 [ptlrpc] 00:04:20: [<ffffffffa0e38c62>] ? mdt_init0+0x1202/0x23f0 [mdt] 00:04:20: [<ffffffffa0e39f43>] ? mdt_device_alloc+0xf3/0x220 [mdt] 00:04:20: [<ffffffffa064b0d7>] ? obd_setup+0x1d7/0x2f0 [obdclass] 00:04:20: [<ffffffffa064b3f8>] ? class_setup+0x208/0x890 [obdclass] 00:04:20: [<ffffffffa065308c>] ? class_process_config+0xc0c/0x1c30 [obdclass] 00:04:20: [<ffffffffa04ea088>] ? libcfs_log_return+0x28/0x40 [libcfs] 00:04:20: [<ffffffffa064cef1>] ? lustre_cfg_new+0x391/0x7e0 [obdclass] 00:04:20: [<ffffffffa065515b>] ? class_config_llog_handler+0x9bb/0x1610 [obdclass] 00:04:20: [<ffffffffa061e1f8>] ? llog_process_thread+0x888/0xd00 [obdclass] 00:04:20: [<ffffffffa061d970>] ? llog_process_thread+0x0/0xd00 [obdclass] 00:04:20: [<ffffffff8100c14a>] ? child_rip+0xa/0x20 00:04:20: [<ffffffffa061d970>] ? llog_process_thread+0x0/0xd00 [obdclass] 00:04:20: [<ffffffffa061d970>] ? llog_process_thread+0x0/0xd00 [obdclass] 00:04:20: [<ffffffff8100c140>] ? child_rip+0x0/0x20 another failure: |
| Comment by Sarah Liu [ 18/Mar/13 ] |
|
I think this should be raise priority since it affected many interop tests |
| Comment by Andreas Dilger [ 18/Mar/13 ] |
|
Sarah to submit a patch to skip conf-sanity test_24a: [ $(lustre_version_code mds) -eq $(version_code 2.3.0) ] &&
skip "skipping test for 2.3.0 MDS due to LU-2468" && return 0
|
| Comment by Sarah Liu [ 19/Mar/13 ] |
|
Andreas, got it, thanks! |
| Comment by Liang Zhen (Inactive) [ 22/Mar/13 ] |
|
I guess we didn't submit anything to 2.3, which means recent patches can't help for 2.3 MDS? |
| Comment by Patrick Valentin (Inactive) [ 18/Apr/13 ] |
|
A Bull customer with lustre 2.1.3 had the same hang on a lustre client (loggin node). |
| Comment by Sebastien Buisson (Inactive) [ 05/Sep/13 ] |
|
Hi, Do you have any news on this one? Each time, we can see a lot OOM messages in the syslog of the dump files. TIA, |
| Comment by Andreas Dilger [ 28/Oct/15 ] |
|
This was fixed for Lustre 2.4.0 by http://review.whamcloud.com/4940 landing. |