Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-8388

LustreError: 59548:0:(osc_page.c:308:osc_page_delete()) ASSERTION( 0 ) failed: LBUG

    XMLWordPrintable

Details

    • Bug
    • Resolution: Unresolved
    • Minor
    • None
    • Lustre 2.9.0
    • None
    • 2.8.55-85-g9d84696, Solo setup.
    • 3
    • 9223372036854775807

    Description

      • stdout.log
      # sh llmount.sh
      Stopping clients: localhost.localdomain /mnt/lustre (opts:)
      Stopping clients: localhost.localdomain /mnt/lustre2 (opts:)
      Loading modules from /root/lustre-release/lustre/tests/..
      detected 4 online CPUs by sysfs
      Force libcfs to create 2 CPU partitions
      ../libcfs/libcfs/libcfs options: 'cpu_npartitions=2'
      debug=vfstrace rpctrace dlmtrace neterror ha config 		      ioctl super lfsck
      subsystem_debug=all -lnet -lnd -pinger
      ../lnet/lnet/lnet options: 'networks=tcp(eth3) accept=all'
      gss/krb5 is not supported
      quota/lquota options: 'hash_lqs_cur_bits=3'
      Formatting mgs, mds, osts
      Format mds1: /tmp/lustre-mdt1
      Format ost1: /tmp/lustre-ost1
      Format ost2: /tmp/lustre-ost2
      Checking servers environments
      Checking clients localhost.localdomain environments
      Loading modules from /root/lustre-release/lustre/tests/..
      detected 4 online CPUs by sysfs
      Force libcfs to create 2 CPU partitions
      debug=vfstrace rpctrace dlmtrace neterror ha config 		      ioctl super lfsck
      subsystem_debug=all -lnet -lnd -pinger
      gss/krb5 is not supported
      Setup mgs, mdt, osts
      Starting mds1:   -o loop /tmp/lustre-mdt1 /mnt/lustre-mds1
      Started lustre-MDT0000
      Starting ost1:   -o loop /tmp/lustre-ost1 /mnt/lustre-ost1
      Started lustre-OST0000
      Starting ost2:   -o loop /tmp/lustre-ost2 /mnt/lustre-ost2
      Started lustre-OST0001
      Starting client: localhost.localdomain:  -o user_xattr,flock localhost.localdomain@tcp:/lustre /mnt/lustre
      Using TIMEOUT=20
      seting jobstats to procname_uid
      Setting lustre.sys.jobid_var from disable to procname_uid
      Waiting 90 secs for update
      Updated after 8s: wanted 'procname_uid' got 'procname_uid'
      disable quota as required
      [root@localhost tests]#
      
      [root@localhost tests]# df -h
      Filesystem            Size  Used Avail Use% Mounted on
      /dev/sda3              44G   31G   13G  72% /
      tmpfs                 1.9G     0  1.9G   0% /dev/shm
      /dev/sda1             2.0G  350M  1.6G  19% /boot
      /dev/loop0            139M   18M  112M  14% /mnt/lustre-mds1
      /dev/loop1            359M   30M  310M   9% /mnt/lustre-ost1
      /dev/loop2            359M   30M  310M   9% /mnt/lustre-ost2
      localhost.localdomain@tcp:/lustre
                            717M   59M  620M   9% /mnt/lustre
      [root@localhost tests]#
      
      
      • lctl blockdev*
      [root@localhost lustre-release]# insmod lustre/llite/llite_lloop.ko
      [root@localhost lustre-release]# MD=/mnt/; dd if=/dev/zero of=$MD/testfile bs=4096 count=5000; lctl blockdev_attach $MD/testfile /dev/loop3; lctl blockdev_info /dev/loop3
      5000+0 records in
      5000+0 records out
      20480000 bytes (20 MB) copied, 0.244223 s, 83.9 MB/s
      attach error(Inappropriate ioctl for device)
      error: Invalid argument
      [root@localhost lustre-release]#
      
      [root@localhost lustre-release]# MD=/mnt/lustre; dd if=/dev/zero of=$MD/testfile bs=4096 count=5000; lctl blockdev_attach $MD/testfile loop3; lctl blockdev_info loop3
      5000+0 records in
      5000+0 records out
      20480000 bytes (20 MB) copied, 1.13594 s, 18.0 MB/s
      
      Message from syslogd@localhost at Jul 12 16:28:50 ...
       kernel:LustreError: 59548:0:(osc_page.c:308:osc_page_delete()) ASSERTION( 0 ) failed: 
      
      Message from syslogd@localhost at Jul 12 16:28:50 ...
       kernel:LustreError: 59548:0:(osc_page.c:308:osc_page_delete()) LBUG
      
      • dmesg
      1805 <4>Lustre: DEBUG MARKER: Using TIMEOUT=20
      1806 <4>Lustre: 59548:0:(lloop.c:720:lloop_ioctl()) Enter llop_ioctl
      1807 <3>LustreError: 59548:0:(osc_cache.c:2488:osc_teardown_async_page()) extent ffff8800871bdd08@{[4864 -> 4999/5119], [2|0|-|cache|wi|ffff880087763ea8], [581632|136|+|-|ffff8800877ffd00|256|(null)]} trunc      at 4864.
      1808 <3>LustreError: 59548:0:(osc_cache.c:2488:osc_teardown_async_page()) ### extent: ffff8800871bdd08 ns: lustre-OST0001-osc-ffff880139dae800 lock: ffff8800877ffd00/0x8c9f98a4b6bd6cd1 lrc: 2/0,0 mode: PW/PW      res: [0x2:0x0:0x0].0x0 rrc: 1 type: EXT [0->18446744073709551615] (req 0->4095) flags: 0x20000000000 nid: local remote: 0x8c9f98a4b6bd6cd8 expref: -99 pid: 59546 timeout: 0 lvb_type: 1
      1809 <3>LustreError: 59548:0:(osc_page.c:307:osc_page_delete()) page@ffff88008328de00[2 ffff88008706eb38 4 1 (null)]
      1810 <3>
      1811 <3>LustreError: 59548:0:(osc_page.c:307:osc_page_delete()) vvp-page@ffff88008328de50(0:0) vm@ffffea0001c901a0 2000000000083d 3:0 ffff88008328de00 4864 lru
      1812 <3>
      1813 <3>LustreError: 59548:0:(osc_page.c:307:osc_page_delete()) lov-page@ffff88008328de90, raid0
      1814 <3>
      1815 <3>LustreError: 59548:0:(osc_page.c:307:osc_page_delete()) osc-page@ffff88008328def8 4864: 1< 0x845fed 258 0 + - > 2< 19922944 0 4096 0x0 0x520 | (null) ffff88008829b4f0 ffff880087763ea8 > 3< 0 0 0 > 4<      0 0 8 21954560 - | - - + - > 5< - - + - | 0 - | 136 - ->
      1816 <3>
      1817 <3>LustreError: 59548:0:(osc_page.c:307:osc_page_delete()) end page@ffff88008328de00
      1818 <3>
      1819 <3>LustreError: 59548:0:(osc_page.c:307:osc_page_delete()) Trying to teardown failed: -16
      1820 <0>LustreError: 59548:0:(osc_page.c:308:osc_page_delete()) ASSERTION( 0 ) failed:
      1821 <0>LustreError: 59548:0:(osc_page.c:308:osc_page_delete()) LBUG
      1822 <4>Pid: 59548, comm: lctl
      1823 <4>
      1824 <4>Call Trace:
      1825 <4> [<ffffffffa0cb1875>] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
      1826 <4> [<ffffffffa0cb1e77>] lbug_with_loc+0x47/0xb0 [libcfs]
      1827 <4> [<ffffffffa15cb4ae>] osc_page_delete+0x46e/0x4e0 [osc]
      1828 <4> [<ffffffffa113662d>] cl_page_delete0+0x7d/0x210 [obdclass]
      1829 <4> [<ffffffffa11367fd>] cl_page_delete+0x3d/0x110 [obdclass]
      1830 <4> [<ffffffffa0afc34d>] ll_invalidatepage+0x8d/0x160 [lustre]
      1831 <4> [<ffffffff811372e5>] do_invalidatepage+0x25/0x30
      1832 <4> [<ffffffff81137602>] truncate_inode_page+0xa2/0xc0
      1833 <4> [<ffffffff811379af>] truncate_inode_pages_range+0x16f/0x500
      1834 <4> [<ffffffff8128596a>] ? kobject_get+0x1a/0x30
      1835 <4> [<ffffffff81137dd5>] truncate_inode_pages+0x15/0x20
      1836 <4> [<ffffffffa0167900>] lloop_ioctl+0x5a0/0x780 [llite_lloop]
      1837 <4> [<ffffffffa0acb267>] ll_file_ioctl+0x667/0x3eb0 [lustre]
      1838 <4> [<ffffffff811865b1>] ? nameidata_to_filp+0x31/0x70
      1839 <4> [<ffffffff8119c3c8>] ? do_filp_open+0x798/0xd20
      1840 <4> [<ffffffff8119e972>] vfs_ioctl+0x22/0xa0
      1841 <4> [<ffffffff8119ee3a>] do_vfs_ioctl+0x3aa/0x580
      1842 <4> [<ffffffff81196dd6>] ? final_putname+0x26/0x50
      1843 <4> [<ffffffff8119f091>] sys_ioctl+0x81/0xa0
      1844 <4> [<ffffffff810e202e>] ? __audit_syscall_exit+0x25e/0x290
      1845 <4> [<ffffffff8100b072>] system_call_fastpath+0x16/0x1b
      

      @andreas
      This is seen on my solo VM setup, I want to debug and give the fix for this, if its fine.

      Thanks.

      Attachments

        Activity

          People

            wc-triage WC Triage
            parinay parinay v kondekar (Inactive)
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

            Dates

              Created:
              Updated: