[LU-9706] conf-sanity test_53a: MDS soft lockup - CPU#0 stuck for 22s in osd_trans_commit_cb() Created: 22/Jun/17  Updated: 01/Apr/19  Resolved: 27/Feb/19

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.10.0, Lustre 2.11.0, Lustre 2.12.0
Fix Version/s: Lustre 2.13.0, Lustre 2.10.7, Lustre 2.12.1

Type: Bug Priority: Minor
Reporter: Maloo Assignee: Alex Zhuravlev
Resolution: Fixed Votes: 0
Labels: zfs

Issue Links:
Duplicate
duplicates LU-6814 conf-sanity test 23b hangs on unmount Resolved
duplicates LU-11528 sanity-lfsck test_11a: soft lockup - ... Resolved
Severity: 3
Rank (Obsolete): 9223372036854775807

 Description   

This issue was created by maloo for sarah_lw <wei3.liu@intel.com>

This issue relates to the following test suite run: https://testing.hpdd.intel.com/test_sets/c25bf48e-53d1-11e7-a749-5254006e85c2.

The sub-test test_53a failed with the following error:

test failed to respond and timed out

env: tag-2.9.59 #3603 ZFS

MDS console

05:24:43:[40506.686671] Lustre: MGS: Connection restored to 1a7ddb58-f96d-6d7b-dd19-c36909e007a9 (at 0@lo)
05:24:43:[40506.826682] Lustre: lustre-MDT0000: Imperative Recovery not enabled, recovery window 60-180
05:24:43:[40506.829961] LustreError: 16342:0:(osd_oi.c:503:osd_oid()) lustre-MDT0000-osd: unsupported quota oid: 0x16
05:24:43:[40507.031480] Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n health_check
05:24:43:[40507.336736] Lustre: DEBUG MARKER: PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/usr/lib64/lustre/tests//usr/lib64/lustre/tests:/usr/lib64/lustre/tests:/usr/lib64/lustre/tests/../utils:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lust
05:24:43:[40507.892722] Lustre: DEBUG MARKER: /usr/sbin/lctl mark trevis-46vm7.trevis.hpdd.intel.com: executing set_default_debug -1 all 4
05:24:43:[40507.893599] Lustre: DEBUG MARKER: /usr/sbin/lctl mark trevis-46vm7.trevis.hpdd.intel.com: executing set_default_debug -1 all 4
05:24:43:[40508.120605] Lustre: DEBUG MARKER: trevis-46vm7.trevis.hpdd.intel.com: executing set_default_debug -1 all 4
05:24:43:[40508.128355] Lustre: DEBUG MARKER: trevis-46vm7.trevis.hpdd.intel.com: executing set_default_debug -1 all 4
05:24:43:[40508.289854] Lustre: DEBUG MARKER: lctl set_param -n mdt.lustre*.enable_remote_dir=1
05:24:43:[40508.591867] Lustre: DEBUG MARKER: zfs get -H -o value 				lustre:svname lustre-mdt1/mdt1 2>/dev/null | 				grep -E ':[a-zA-Z]{3}[0-9]{4}'
05:24:43:[40508.901939] Lustre: DEBUG MARKER: zfs get -H -o value 				lustre:svname lustre-mdt1/mdt1 2>/dev/null | 				grep -E ':[a-zA-Z]{3}[0-9]{4}'
05:24:43:[40509.212232] Lustre: DEBUG MARKER: zfs get -H -o value lustre:svname 		                           lustre-mdt1/mdt1 2>/dev/null
05:24:43:[40509.520874] Lustre: DEBUG MARKER: lctl set_param -n mdt.lustre*.enable_remote_dir=1
05:24:43:[40511.168162] Lustre: MGS: Connection restored to 00faafbe-57f8-e637-4480-1f8b3bfa5137 (at 10.9.6.35@tcp)
05:24:43:[40511.800039] Lustre: 15064:0:(client.c:2114:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1497677067/real 1497677067]  req@ffff8800363c0300 x1570428225061024/t0(0) o8->lustre-OST0000-osc-MDT0000@10.9.6.35@tcp:28/4 lens 520/544 e 0 to 1 dl 1497677072 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
05:24:43:[40512.608633] Lustre: DEBUG MARKER: /usr/sbin/lctl mark trevis-46vm8.trevis.hpdd.intel.com: executing set_default_debug -1 all 4
05:25:17:[40512.800854] Lustre: DEBUG MARKER: trevis-46vm8.trevis.hpdd.intel.com: executing set_default_debug -1 all 4
05:25:17:[40513.793184] Lustre: MGS: Connection restored to 7310674c-545b-174e-1751-630f6aaa5a8a (at 10.9.6.28@tcp)
05:25:17:[40513.796051] Lustre: Skipped 2 previous similar messages
05:25:17:[40516.831820] LustreError: 11-0: lustre-OST0000-osc-MDT0000: operation ost_statfs to node 10.9.6.35@tcp failed: rc = -107
05:25:17:[40516.835465] Lustre: lustre-OST0000-osc-MDT0000: Connection to lustre-OST0000 (at 10.9.6.35@tcp) was lost; in progress operations using this service will wait for recovery to complete
05:25:17:[40518.877430] Lustre: DEBUG MARKER: grep -c /mnt/lustre-mds1' ' /proc/mounts
05:25:17:[40519.195771] Lustre: DEBUG MARKER: umount -d -f /mnt/lustre-mds1
05:25:17:[40544.062002] NMI watchdog: BUG: soft lockup - CPU#1 stuck for 22s! [tx_commit_cb:16414]
05:25:17:
05:25:17:[40544.062002] Modules linked in: osp(OE) mdd(OE) lod(OE) mdt(OE) lfsck(OE) mgs(OE) osd_zfs(OE) lquota(OE) lustre(OE) obdecho(OE) mgc(OE) lov(OE) osc(OE) mdc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) zfs(POE) zunicode(POE) zavl(POE) zcommon(POE) znvpair(POE) spl(OE) zlib_deflate dm_mod rpcsec_gss_krb5 nfsv4 dns_resolver nfs fscache rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod crc_t10dif crct10dif_generic ib_srp scsi_transport_srp scsi_tgt ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel ppdev lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon i2c_piix4 parport_pc parport nfsd nfs_acl lockd grace auth_rpcgss sunrpc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi virtio_blk cirrus crct10dif_pclmul crct10dif_common drm_kms_helper syscopyarea sysfillrect crc32c_intel 8139too serio_raw sysimgblt virtio_pci fb_sys_fops virtio_ring ttm virtio ata_piix 8139cp mii drm libata i2c_core floppy [last unloaded: libcfs]
05:25:17:[40544.062002] CPU: 1 PID: 16414 Comm: tx_commit_cb Tainted: P           OE  ------------   3.10.0-514.21.1.el7_lustre.x86_64 #1
05:25:17:[40544.062002] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2007
05:25:17:[40544.062002] task: ffff88005d50bec0 ti: ffff88004361c000 task.ti: ffff88004361c000
05:25:17:[40544.062002] RIP: 0010:[<ffffffffa08c4fe4>]  [<ffffffffa08c4fe4>] dt_txn_hook_commit+0x34/0x60 [obdclass]
05:25:17:[40544.062002] RSP: 0018:ffff88004361fd78  EFLAGS: 00000297
05:25:17:[40544.062002] RAX: 0000000000000000 RBX: ffff8800694b94a0 RCX: ffff880069efc000
05:25:17:[40544.062002] RDX: ffff880069efc048 RSI: ffffffffa1287600 RDI: ffff88005e026e00
05:25:17:[40544.062002] RBP: ffff88004361fd88 R08: 20737365636f7250 R09: 0a64657265746e65
05:25:17:[40544.062002] R10: ffff880014dc2535 R11: 0a64657265746e65 R12: 0000000000000000
05:25:17:[40544.062002] R13: ffffffff810ce9ac R14: ffff88004361fd20 R15: 00000000a57bc182
05:25:17:[40544.062002] FS:  0000000000000000(0000) GS:ffff88007fd00000(0000) knlGS:0000000000000000
05:25:17:[40544.062002] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
05:25:17:[40544.062002] CR2: 00007f23d5a9e000 CR3: 000000007b0d7000 CR4: 00000000000406e0
05:25:17:[40544.062002] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
05:25:17:[40544.062002] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
05:25:17:[40544.062002] Stack:
05:25:17:[40544.062002]  ffff8800694b94a0 ffff88005cbd0600 ffff88004361fdd0 ffffffffa0f676a8
05:25:17:[40544.062002]  ffff880069efc000 ffff88007c0c0000 ffff8800694b94a0 ffff8800694b90f0
05:25:17:[40544.062002]  0000000000000000 ffff8800694b90e0 ffff880079f78800 ffff88004361fe00
05:25:17:[40544.062002] Call Trace:
05:25:17:[40544.062002]  [<ffffffffa0f676a8>] osd_trans_commit_cb+0xe8/0x490 [osd_zfs]
05:25:17:[40544.062002]  [<ffffffffa17d05e4>] dmu_tx_do_callbacks+0x44/0x70 [zfs]
05:25:17:[40544.062002]  [<ffffffffa1814c74>] txg_do_callbacks+0x14/0x30 [zfs]
05:25:17:[40544.062002]  [<ffffffffa06d36de>] taskq_thread+0x22e/0x440 [spl]
05:25:17:[40544.062002]  [<ffffffff810c54e0>] ? wake_up_state+0x20/0x20
05:25:17:[40544.062002]  [<ffffffffa06d34b0>] ? taskq_thread_spawn+0x60/0x60 [spl]
05:25:17:[40544.062002]  [<ffffffff810b0a4f>] kthread+0xcf/0xe0
05:25:17:[40544.062002]  [<ffffffff810b0980>] ? kthread_create_on_node+0x140/0x140
05:25:17:[40544.062002]  [<ffffffff81697798>] ret_from_fork+0x58/0x90
05:25:17:[40544.062002]  [<ffffffff810b0980>] ? kthread_create_on_node+0x140/0x140
05:25:17:[40544.062002] Code: 41 54 49 89 fc 53 f6 47 48 02 75 42 48 8b 0f 48 8b 41 48 48 8d 51 48 48 39 d0 48 8d 58 d8 74 2e 0f 1f 80 00 00 00 00 48 8b 43 10 <48> 85 c0 74 0d 48 8b 73 18 4c 89 e7 ff d0 49 8b 0c 24 48 8b 43 
05:25:17:[40544.062002] Kernel panic - not syncing: softlockup: hung tasks
05:25:17:[40544.062002] CPU: 1 PID: 16414 Comm: tx_commit_cb Tainted: P           OEL ------------   3.10.0-514.21.1.el7_lustre.x86_64 #1
05:25:17:[40544.062002] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2007
05:25:17:[40544.062002]  ffffffff818da7c9 00000000a57bc182 ffff88007fd03e18 ffffffff8168717f
05:25:17:[40544.062002]  ffff88007fd03e98 ffffffff816805aa 0000000000000008 ffff88007fd03ea8
05:25:17:[40544.062002]  ffff88007fd03e48 00000000a57bc182 ffff88007fd03e67 0000000000000000
05:25:17:[40544.062002] Call Trace:
05:25:17:[40544.062002]  <IRQ>  [<ffffffff8168717f>] dump_stack+0x19/0x1b
05:25:17:[40544.062002]  [<ffffffff816805aa>] panic+0xe3/0x1f2
05:25:17:[40544.062002]  [<ffffffff8112f444>] watchdog_timer_fn+0x234/0x240
05:25:17:[40544.062002]  [<ffffffff8112f210>] ? watchdog+0x50/0x50
05:25:17:[40544.062002]  [<ffffffff810b4d72>] __hrtimer_run_queues+0xd2/0x260
05:25:17:[40544.062002]  [<ffffffff810b5310>] hrtimer_interrupt+0xb0/0x1e0
05:25:17:[40544.062002]  [<ffffffff816992dc>] ? call_softirq+0x1c/0x30
05:25:17:[40544.062002]  [<ffffffff81050ff7>] local_apic_timer_interrupt+0x37/0x60
05:25:17:[40544.062002]  [<ffffffff81699f4f>] smp_apic_timer_interrupt+0x3f/0x60
05:25:17:[40544.062002]  [<ffffffff8169849d>] apic_timer_interrupt+0x6d/0x80
05:25:17:[40544.062002]  <EOI>  [<ffffffffa08c4fe4>] ? dt_txn_hook_commit+0x34/0x60 [obdclass]
05:25:17:[40544.062002]  [<ffffffffa08c4ff2>] ? dt_txn_hook_commit+0x42/0x60 [obdclass]
05:25:17:[40544.062002]  [<ffffffffa0f676a8>] osd_trans_commit_cb+0xe8/0x490 [osd_zfs]
05:25:17:[    0.000000] Initializing cgroup subsys cpuset
05:25:17:[    0.000000] Initializing cgroup subsys cpu
05:25:17:[    0.000000] Initializing cgroup subsys cpuacct
05:25:17:[    0.000000] Linux version 3.10.0-514.21.1.el7_lustre.x86_64 (jenkins@trevis-310-el7


 Comments   
Comment by James Nunez (Inactive) [ 07/Nov/17 ]

I've seen the same hang on MDS for conf-sanity test 53b. Logs are at https://testing.hpdd.intel.com/test_sets/1d538b38-c347-11e7-88ab-52540065bddc. This is the only time 53b hung on MDS umount in the past two months.

Comment by Jian Yu [ 16/Nov/17 ]

conf-sanity test 39 in ZFS test session also hit the same failure: https://testing.hpdd.intel.com/test_sets/63f771f2-cae6-11e7-9840-52540065bddc

Console log on MDS:

Lustre: DEBUG MARKER: umount -d -f /mnt/lustre-mds1
NMI watchdog: BUG: soft lockup - CPU#1 stuck for 23s! [tx_commit_cb:27046]
Modules linked in: osd_zfs(OE) zfs(POE) obdecho(OE) ptlrpc_gss(OE) ofd(OE) ost(OE) lustre(OE) lmv(OE) mdc(OE) osc(OE) lov(OE) osp(OE) mdd(OE) lod(OE) mdt(OE) lfsck(OE) mgs(OE) mgc(OE) lquota(OE) fid(OE) fld(OE) ksocklnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE) zunicode(POE) zavl(POE) icp(POE) zcommon(POE) znvpair(POE) spl(OE) libcfs(OE) dm_mod rpcsec_gss_krb5 nfsv4 dns_resolver nfs fscache rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod crc_t10dif crct10dif_generic ib_srp scsi_transport_srp scsi_tgt ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_core iosf_mbi crc32_pclmul ghash_clmulni_intel ppdev aesni_intel lrw gf128mul glue_helper ablk_helper cryptd joydev parport_pc pcspkr virtio_balloon parport i2c_piix4 i2c_core nfsd nfs_acl lockd auth_rpcgss grace sunrpc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi virtio_blk ata_piix libata crct10dif_pclmul crct10dif_common 8139too crc32c_intel floppy virtio_pci virtio_ring serio_raw virtio 8139cp mii [last unloaded: zfs] 
CPU: 1 PID: 27046 Comm: tx_commit_cb Tainted: P           OE  ------------   3.10.0-693.5.2.el7_lustre.x86_64 #1
Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
task: ffff880067184f10 ti: ffff88007ac34000 task.ti: ffff88007ac34000
RIP: 0010:[<ffffffffc0beb664>]  [<ffffffffc0beb664>] dt_txn_hook_commit+0x34/0x60 [obdclass]
RSP: 0018:ffff88007ac37d70  EFLAGS: 00000216
RAX: 0000000000000000 RBX: ffff880067184f78 RCX: ffff8800503c8000
RDX: ffff8800503c8048 RSI: 0000000000000000 RDI: ffff880060fc1000
RBP: ffff88007ac37d80 R08: 0000000000000001 R09: 000000010076d1ab
R10: ffff88004ecb6088 R11: 7fffffffffffffff R12: ffff88004f57a3e0
R13: ffff880050461d80 R14: ffffffff810cdc8c R15: ffff88007ac37d18
FS:  0000000000000000(0000) GS:ffff88007fd00000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 000055aba7b561d8 CR3: 00000000019f2000 CR4: 00000000000406e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Stack:
 ffff88004f57a3e0 ffff880060fc1000 ffff88007ac37dc8 ffffffffc07be6e8
 ffff8800503c8000 ffff88007c0a4000 ffff88004f57a3e0 ffff88004f57a490
 0000000000000000 ffff88004f57a480 0000000000000000 ffff88007ac37df8
Call Trace:
 [<ffffffffc07be6e8>] osd_trans_commit_cb+0xe8/0x480 [osd_zfs]
 [<ffffffffc198fce4>] dmu_tx_do_callbacks+0x44/0x70 [zfs]
 [<ffffffffc19db364>] txg_do_callbacks+0x14/0x30 [zfs]
 [<ffffffffc0659ed6>] taskq_thread+0x246/0x470 [spl]
 [<ffffffff810c4820>] ? wake_up_state+0x20/0x20
 [<ffffffffc0659c90>] ? taskq_thread_spawn+0x60/0x60 [spl]
 [<ffffffff810b099f>] kthread+0xcf/0xe0
 [<ffffffff810bf114>] ? finish_task_switch+0x54/0x160
 [<ffffffff810b08d0>] ? insert_kthread_work+0x40/0x40
 [<ffffffff816b4fd8>] ret_from_fork+0x58/0x90
 [<ffffffff810b08d0>] ? insert_kthread_work+0x40/0x40
Code: 41 54 49 89 fc 53 f6 47 14 02 75 42 48 8b 0f 48 8b 41 48 48 8d 51 48 48 39 d0 48 8d 58 d8 74 2e 0f 1f 80 00 00 00 00 48 8b 43 10 <48> 85 c0 74 0d 48 8b 73 18 4c 89 e7 ff d0 49 8b 0c 24 48 8b 43
Kernel panic - not syncing: softlockup: hung tasks
CPU: 1 PID: 27046 Comm: tx_commit_cb Tainted: P           OEL ------------   3.10.0-693.5.2.el7_lustre.x86_64 #1
Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
 0000000000000000 000000006a1707aa ffff88007fd03e20 ffffffff816a3e51
 ffff88007fd03ea0 ffffffff8169dd14 0000000000000008 ffff88007fd03eb0
 ffff88007fd03e50 000000006a1707aa ffffffff8102d8af 0000000000000000
Call Trace:
 <IRQ>  [<ffffffff816a3e51>] dump_stack+0x19/0x1b
 [<ffffffff8169dd14>] panic+0xe8/0x20d
 [<ffffffff8102d8af>] ? show_regs+0x5f/0x220
 [<ffffffff8112f661>] watchdog_timer_fn+0x221/0x230
 [<ffffffff8112f440>] ? watchdog+0x40/0x40
 [<ffffffff810b4af4>] __hrtimer_run_queues+0xd4/0x260
 [<ffffffff810b508f>] hrtimer_interrupt+0xaf/0x1d0
 [<ffffffff81053895>] local_apic_timer_interrupt+0x35/0x60
 [<ffffffff816b777d>] smp_apic_timer_interrupt+0x3d/0x50
 [<ffffffff816b5cdd>] apic_timer_interrupt+0x6d/0x80
 <EOI>  [<ffffffffc0beb664>] ? dt_txn_hook_commit+0x34/0x60 [obdclass]
 [<ffffffffc07be6e8>] osd_trans_commit_cb+0xe8/0x480 [osd_zfs]
 [<ffffffffc198fce4>] dmu_tx_do_callbacks+0x44/0x70 [zfs]
 [<ffffffffc19db364>] txg_do_callbacks+0x14/0x30 [zfs]
 [<ffffffffc0659ed6>] taskq_thread+0x246/0x470 [spl]
 [<ffffffff810c4820>] ? wake_up_state+0x20/0x20
 [<ffffffffc0659c90>] ? taskq_thread_spawn+0x60/0x60 [spl]
 [<ffffffff810b099f>] kthread+0xcf/0xe0
 [<ffffffff810bf114>] ? finish_task_switch+0x54/0x160
 [<ffffffff810b08d0>] ? insert_kthread_work+0x40/0x40
 [<ffffffff816b4fd8>] ret_from_fork+0x58/0x90
 [<ffffffff810b08d0>] ? insert_kthread_work+0x40/0x40
Comment by nasf (Inactive) [ 07/Dec/17 ]

+1 on master:
https://testing.hpdd.intel.com/test_sets/c66d5546-db2f-11e7-9840-52540065bddc

Comment by Andreas Dilger [ 13/Apr/18 ]

+1 on master when testing sanity-lfsck:
https://testing.hpdd.intel.com/test_sets/79d0a996-3f51-11e8-960d-52540065bddc

Comment by Sarah Liu [ 25/Apr/18 ]

+1 on master tag-2.11.51 DNE zfs sanity-lfsck test_11b

https://testing.hpdd.intel.com/test_sets/22afde2a-477b-11e8-95c0-52540065bddc

Comment by Nathaniel Clark [ 04/Jun/18 ]

Is this a duplicate of LU-6845? Or maybe even/also LU-6814?

Comment by James Nunez (Inactive) [ 17/Sep/18 ]

Same crash on sanity-quota test 35 at https://testing.whamcloud.com/test_sets/9d3c27ca-b995-11e8-8c12-52540065bddc for Lustre 2.11.55.

Comment by Jian Yu [ 28/Sep/18 ]

+1 on master branch when testing conf-sanity test 90a:
https://testing.whamcloud.com/test_sets/7d3b2c1e-c313-11e8-a2b9-52540065bddc

Comment by Alex Zhuravlev [ 07/Feb/19 ]

I also saw this few times in the reports. looking at the source - it's not clear why dt_txn_callback_add() and dt_txn_callback_del() don't use any locks.

Comment by Alex Zhuravlev [ 07/Feb/19 ]

at initialisation tgt_init() is called before mdt_fs_setup(). thus dt_txn_hook_commit() can race against dt_txn_callback_add(). similarly dt_txn_callback_del() from mdt_fs_cleanup() can race as there is no barrier for in-flight transaction.
the trivial solution would be to use a semaphore (we can't use spinlocks there), but ZFS may call commit callbacks on few cores and that can be a performance issue.

Comment by Alex Zhuravlev [ 07/Feb/19 ]

$ findsrc dtc_txn_commit.*=
./lustre/target/tgt_main.c: lut->lut_txn_cb.dtc_txn_commit = NULL;
./lustre/mdt/mdt_recovery.c: mdt->mdt_txn_cb.dtc_txn_commit = NULL;

I guess we can remove dt_txn_hook_commit()

Comment by Andreas Dilger [ 07/Feb/19 ]

+1 on b2_10: https://testing.whamcloud.com/test_sets/af0788a8-2693-11e9-b901-52540065bddc

Comment by Gerrit Updater [ 08/Feb/19 ]

Alex Zhuravlev (bzzz@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/34212
Subject: LU-9706 dt: remove dt_txn_hook_commit()
Project: fs/lustre-release
Branch: master
Current Patch Set: 1
Commit: b46892faa8c9b1f174d767f00ba4268715b4030e

Comment by Minh Diep [ 16/Feb/19 ]

+1 https://testing.whamcloud.com/test_sessions/f220c6df-ec28-4946-ad8e-c861d57e5a64

Comment by Alex Zhuravlev [ 18/Feb/19 ]

https://testing.whamcloud.com/test_sessions/7794d909-1bc3-4be5-a75f-d01c4b79a1b5/retest

Comment by Gerrit Updater [ 18/Feb/19 ]

Alex Zhuravlev (bzzz@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/34274
Subject: LU-9706 dt: remove dt_txn_hook_commit()
Project: fs/lustre-release
Branch: b2_10
Current Patch Set: 1
Commit: 2b82e3c36bde3ec0d2c1b314b48b392bdfc92518

Comment by Patrick Farrell (Inactive) [ 20/Feb/19 ]

Also saw this with conf-sanity 5e:
https://testing.whamcloud.com/test_sessions/4040ad08-d8aa-4066-994d-28e0bf82036f

Comment by Patrick Farrell (Inactive) [ 25/Feb/19 ]

Master:
https://testing.whamcloud.com/test_sessions/4040ad08-d8aa-4066-994d-28e0bf82036f

Comment by Gerrit Updater [ 27/Feb/19 ]

Oleg Drokin (green@whamcloud.com) merged in patch https://review.whamcloud.com/34212/
Subject: LU-9706 dt: remove dt_txn_hook_commit()
Project: fs/lustre-release
Branch: master
Current Patch Set:
Commit: e763467ebe00913e8d03f855dc4b918b95099931

Comment by Peter Jones [ 27/Feb/19 ]

Landed for 2.13

Comment by Gerrit Updater [ 02/Mar/19 ]

Oleg Drokin (green@whamcloud.com) merged in patch https://review.whamcloud.com/34274/
Subject: LU-9706 dt: remove dt_txn_hook_commit()
Project: fs/lustre-release
Branch: b2_10
Current Patch Set:
Commit: 4b239e9138aa69eb996ee47b31f075c085efa1e5

Comment by Alex Zhuravlev [ 02/Mar/19 ]

Peter, ldiskfs was vulnerable as well.

Comment by Gerrit Updater [ 23/Mar/19 ]

Minh Diep (mdiep@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/34496
Subject: LU-9706 dt: remove dt_txn_hook_commit()
Project: fs/lustre-release
Branch: b2_12
Current Patch Set: 1
Commit: 3a518aa8871ce41c03c44d7daeb6b8eb6c077d97

Comment by Gerrit Updater [ 01/Apr/19 ]

Oleg Drokin (green@whamcloud.com) merged in patch https://review.whamcloud.com/34496/
Subject: LU-9706 dt: remove dt_txn_hook_commit()
Project: fs/lustre-release
Branch: b2_12
Current Patch Set:
Commit: 6c2ecb0c72510cad32210ec327122c2b140eb6cc

Generated at Sat Feb 10 02:28:31 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.