mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-25 05:17:28 +07:00
66d204a16c
Very sporadically I had test case btrfs/069 from fstests hanging (for years, it is not a recent regression), with the following traces in dmesg/syslog: [162301.160628] BTRFS info (device sdc): dev_replace from /dev/sdd (devid 2) to /dev/sdg started [162301.181196] BTRFS info (device sdc): scrub: finished on devid 4 with status: 0 [162301.287162] BTRFS info (device sdc): dev_replace from /dev/sdd (devid 2) to /dev/sdg finished [162513.513792] INFO: task btrfs-transacti:1356167 blocked for more than 120 seconds. [162513.514318] Not tainted 5.9.0-rc6-btrfs-next-69 #1 [162513.514522] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [162513.514747] task:btrfs-transacti state:D stack: 0 pid:1356167 ppid: 2 flags:0x00004000 [162513.514751] Call Trace: [162513.514761] __schedule+0x5ce/0xd00 [162513.514765] ? _raw_spin_unlock_irqrestore+0x3c/0x60 [162513.514771] schedule+0x46/0xf0 [162513.514844] wait_current_trans+0xde/0x140 [btrfs] [162513.514850] ? finish_wait+0x90/0x90 [162513.514864] start_transaction+0x37c/0x5f0 [btrfs] [162513.514879] transaction_kthread+0xa4/0x170 [btrfs] [162513.514891] ? btrfs_cleanup_transaction+0x660/0x660 [btrfs] [162513.514894] kthread+0x153/0x170 [162513.514897] ? kthread_stop+0x2c0/0x2c0 [162513.514902] ret_from_fork+0x22/0x30 [162513.514916] INFO: task fsstress:1356184 blocked for more than 120 seconds. [162513.515192] Not tainted 5.9.0-rc6-btrfs-next-69 #1 [162513.515431] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [162513.515680] task:fsstress state:D stack: 0 pid:1356184 ppid:1356177 flags:0x00004000 [162513.515682] Call Trace: [162513.515688] __schedule+0x5ce/0xd00 [162513.515691] ? _raw_spin_unlock_irqrestore+0x3c/0x60 [162513.515697] schedule+0x46/0xf0 [162513.515712] wait_current_trans+0xde/0x140 [btrfs] [162513.515716] ? finish_wait+0x90/0x90 [162513.515729] start_transaction+0x37c/0x5f0 [btrfs] [162513.515743] btrfs_attach_transaction_barrier+0x1f/0x50 [btrfs] [162513.515753] btrfs_sync_fs+0x61/0x1c0 [btrfs] [162513.515758] ? __ia32_sys_fdatasync+0x20/0x20 [162513.515761] iterate_supers+0x87/0xf0 [162513.515765] ksys_sync+0x60/0xb0 [162513.515768] __do_sys_sync+0xa/0x10 [162513.515771] do_syscall_64+0x33/0x80 [162513.515774] entry_SYSCALL_64_after_hwframe+0x44/0xa9 [162513.515781] RIP: 0033:0x7f5238f50bd7 [162513.515782] Code: Bad RIP value. [162513.515784] RSP: 002b:00007fff67b978e8 EFLAGS: 00000206 ORIG_RAX: 00000000000000a2 [162513.515786] RAX: ffffffffffffffda RBX: 000055b1fad2c560 RCX: 00007f5238f50bd7 [162513.515788] RDX: 00000000ffffffff RSI: 000000000daf0e74 RDI: 000000000000003a [162513.515789] RBP: 0000000000000032 R08: 000000000000000a R09: 00007f5239019be0 [162513.515791] R10: fffffffffffff24f R11: 0000000000000206 R12: 000000000000003a [162513.515792] R13: 00007fff67b97950 R14: 00007fff67b97906 R15: 000055b1fad1a340 [162513.515804] INFO: task fsstress:1356185 blocked for more than 120 seconds. [162513.516064] Not tainted 5.9.0-rc6-btrfs-next-69 #1 [162513.516329] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [162513.516617] task:fsstress state:D stack: 0 pid:1356185 ppid:1356177 flags:0x00000000 [162513.516620] Call Trace: [162513.516625] __schedule+0x5ce/0xd00 [162513.516628] ? _raw_spin_unlock_irqrestore+0x3c/0x60 [162513.516634] schedule+0x46/0xf0 [162513.516647] wait_current_trans+0xde/0x140 [btrfs] [162513.516650] ? finish_wait+0x90/0x90 [162513.516662] start_transaction+0x4d7/0x5f0 [btrfs] [162513.516679] btrfs_setxattr_trans+0x3c/0x100 [btrfs] [162513.516686] __vfs_setxattr+0x66/0x80 [162513.516691] __vfs_setxattr_noperm+0x70/0x200 [162513.516697] vfs_setxattr+0x6b/0x120 [162513.516703] setxattr+0x125/0x240 [162513.516709] ? lock_acquire+0xb1/0x480 [162513.516712] ? mnt_want_write+0x20/0x50 [162513.516721] ? rcu_read_lock_any_held+0x8e/0xb0 [162513.516723] ? preempt_count_add+0x49/0xa0 [162513.516725] ? __sb_start_write+0x19b/0x290 [162513.516727] ? preempt_count_add+0x49/0xa0 [162513.516732] path_setxattr+0xba/0xd0 [162513.516739] __x64_sys_setxattr+0x27/0x30 [162513.516741] do_syscall_64+0x33/0x80 [162513.516743] entry_SYSCALL_64_after_hwframe+0x44/0xa9 [162513.516745] RIP: 0033:0x7f5238f56d5a [162513.516746] Code: Bad RIP value. [162513.516748] RSP: 002b:00007fff67b97868 EFLAGS: 00000202 ORIG_RAX: 00000000000000bc [162513.516750] RAX: ffffffffffffffda RBX: 0000000000000001 RCX: 00007f5238f56d5a [162513.516751] RDX: 000055b1fbb0d5a0 RSI: 00007fff67b978a0 RDI: 000055b1fbb0d470 [162513.516753] RBP: 000055b1fbb0d5a0 R08: 0000000000000001 R09: 00007fff67b97700 [162513.516754] R10: 0000000000000004 R11: 0000000000000202 R12: 0000000000000004 [162513.516756] R13: 0000000000000024 R14: 0000000000000001 R15: 00007fff67b978a0 [162513.516767] INFO: task fsstress:1356196 blocked for more than 120 seconds. [162513.517064] Not tainted 5.9.0-rc6-btrfs-next-69 #1 [162513.517365] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [162513.517763] task:fsstress state:D stack: 0 pid:1356196 ppid:1356177 flags:0x00004000 [162513.517780] Call Trace: [162513.517786] __schedule+0x5ce/0xd00 [162513.517789] ? _raw_spin_unlock_irqrestore+0x3c/0x60 [162513.517796] schedule+0x46/0xf0 [162513.517810] wait_current_trans+0xde/0x140 [btrfs] [162513.517814] ? finish_wait+0x90/0x90 [162513.517829] start_transaction+0x37c/0x5f0 [btrfs] [162513.517845] btrfs_attach_transaction_barrier+0x1f/0x50 [btrfs] [162513.517857] btrfs_sync_fs+0x61/0x1c0 [btrfs] [162513.517862] ? __ia32_sys_fdatasync+0x20/0x20 [162513.517865] iterate_supers+0x87/0xf0 [162513.517869] ksys_sync+0x60/0xb0 [162513.517872] __do_sys_sync+0xa/0x10 [162513.517875] do_syscall_64+0x33/0x80 [162513.517878] entry_SYSCALL_64_after_hwframe+0x44/0xa9 [162513.517881] RIP: 0033:0x7f5238f50bd7 [162513.517883] Code: Bad RIP value. [162513.517885] RSP: 002b:00007fff67b978e8 EFLAGS: 00000206 ORIG_RAX: 00000000000000a2 [162513.517887] RAX: ffffffffffffffda RBX: 000055b1fad2c560 RCX: 00007f5238f50bd7 [162513.517889] RDX: 0000000000000000 RSI: 000000007660add2 RDI: 0000000000000053 [162513.517891] RBP: 0000000000000032 R08: 0000000000000067 R09: 00007f5239019be0 [162513.517893] R10: fffffffffffff24f R11: 0000000000000206 R12: 0000000000000053 [162513.517895] R13: 00007fff67b97950 R14: 00007fff67b97906 R15: 000055b1fad1a340 [162513.517908] INFO: task fsstress:1356197 blocked for more than 120 seconds. [162513.518298] Not tainted 5.9.0-rc6-btrfs-next-69 #1 [162513.518672] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [162513.519157] task:fsstress state:D stack: 0 pid:1356197 ppid:1356177 flags:0x00000000 [162513.519160] Call Trace: [162513.519165] __schedule+0x5ce/0xd00 [162513.519168] ? _raw_spin_unlock_irqrestore+0x3c/0x60 [162513.519174] schedule+0x46/0xf0 [162513.519190] wait_current_trans+0xde/0x140 [btrfs] [162513.519193] ? finish_wait+0x90/0x90 [162513.519206] start_transaction+0x4d7/0x5f0 [btrfs] [162513.519222] btrfs_create+0x57/0x200 [btrfs] [162513.519230] lookup_open+0x522/0x650 [162513.519246] path_openat+0x2b8/0xa50 [162513.519270] do_filp_open+0x91/0x100 [162513.519275] ? find_held_lock+0x32/0x90 [162513.519280] ? lock_acquired+0x33b/0x470 [162513.519285] ? do_raw_spin_unlock+0x4b/0xc0 [162513.519287] ? _raw_spin_unlock+0x29/0x40 [162513.519295] do_sys_openat2+0x20d/0x2d0 [162513.519300] do_sys_open+0x44/0x80 [162513.519304] do_syscall_64+0x33/0x80 [162513.519307] entry_SYSCALL_64_after_hwframe+0x44/0xa9 [162513.519309] RIP: 0033:0x7f5238f4a903 [162513.519310] Code: Bad RIP value. [162513.519312] RSP: 002b:00007fff67b97758 EFLAGS: 00000246 ORIG_RAX: 0000000000000055 [162513.519314] RAX: ffffffffffffffda RBX: 00000000ffffffff RCX: 00007f5238f4a903 [162513.519316] RDX: 0000000000000000 RSI: 00000000000001b6 RDI: 000055b1fbb0d470 [162513.519317] RBP: 00007fff67b978c0 R08: 0000000000000001 R09: 0000000000000002 [162513.519319] R10: 00007fff67b974f7 R11: 0000000000000246 R12: 0000000000000013 [162513.519320] R13: 00000000000001b6 R14: 00007fff67b97906 R15: 000055b1fad1c620 [162513.519332] INFO: task btrfs:1356211 blocked for more than 120 seconds. [162513.519727] Not tainted 5.9.0-rc6-btrfs-next-69 #1 [162513.520115] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [162513.520508] task:btrfs state:D stack: 0 pid:1356211 ppid:1356178 flags:0x00004002 [162513.520511] Call Trace: [162513.520516] __schedule+0x5ce/0xd00 [162513.520519] ? _raw_spin_unlock_irqrestore+0x3c/0x60 [162513.520525] schedule+0x46/0xf0 [162513.520544] btrfs_scrub_pause+0x11f/0x180 [btrfs] [162513.520548] ? finish_wait+0x90/0x90 [162513.520562] btrfs_commit_transaction+0x45a/0xc30 [btrfs] [162513.520574] ? start_transaction+0xe0/0x5f0 [btrfs] [162513.520596] btrfs_dev_replace_finishing+0x6d8/0x711 [btrfs] [162513.520619] btrfs_dev_replace_by_ioctl.cold+0x1cc/0x1fd [btrfs] [162513.520639] btrfs_ioctl+0x2a25/0x36f0 [btrfs] [162513.520643] ? do_sigaction+0xf3/0x240 [162513.520645] ? find_held_lock+0x32/0x90 [162513.520648] ? do_sigaction+0xf3/0x240 [162513.520651] ? lock_acquired+0x33b/0x470 [162513.520655] ? _raw_spin_unlock_irq+0x24/0x50 [162513.520657] ? lockdep_hardirqs_on+0x7d/0x100 [162513.520660] ? _raw_spin_unlock_irq+0x35/0x50 [162513.520662] ? do_sigaction+0xf3/0x240 [162513.520671] ? __x64_sys_ioctl+0x83/0xb0 [162513.520672] __x64_sys_ioctl+0x83/0xb0 [162513.520677] do_syscall_64+0x33/0x80 [162513.520679] entry_SYSCALL_64_after_hwframe+0x44/0xa9 [162513.520681] RIP: 0033:0x7fc3cd307d87 [162513.520682] Code: Bad RIP value. [162513.520684] RSP: 002b:00007ffe30a56bb8 EFLAGS: 00000202 ORIG_RAX: 0000000000000010 [162513.520686] RAX: ffffffffffffffda RBX: 0000000000000004 RCX: 00007fc3cd307d87 [162513.520687] RDX: 00007ffe30a57a30 RSI: 00000000ca289435 RDI: 0000000000000003 [162513.520689] RBP: 0000000000000000 R08: 0000000000000000 R09: 0000000000000000 [162513.520690] R10: 0000000000000008 R11: 0000000000000202 R12: 0000000000000003 [162513.520692] R13: 0000557323a212e0 R14: 00007ffe30a5a520 R15: 0000000000000001 [162513.520703] Showing all locks held in the system: [162513.520712] 1 lock held by khungtaskd/54: [162513.520713] #0: ffffffffb40a91a0 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x15/0x197 [162513.520728] 1 lock held by in:imklog/596: [162513.520729] #0: ffff8f3f0d781400 (&f->f_pos_lock){+.+.}-{3:3}, at: __fdget_pos+0x4d/0x60 [162513.520782] 1 lock held by btrfs-transacti/1356167: [162513.520784] #0: ffff8f3d810cc848 (&fs_info->transaction_kthread_mutex){+.+.}-{3:3}, at: transaction_kthread+0x4a/0x170 [btrfs] [162513.520798] 1 lock held by btrfs/1356190: [162513.520800] #0: ffff8f3d57644470 (sb_writers#15){.+.+}-{0:0}, at: mnt_want_write_file+0x22/0x60 [162513.520805] 1 lock held by fsstress/1356184: [162513.520806] #0: ffff8f3d576440e8 (&type->s_umount_key#62){++++}-{3:3}, at: iterate_supers+0x6f/0xf0 [162513.520811] 3 locks held by fsstress/1356185: [162513.520812] #0: ffff8f3d57644470 (sb_writers#15){.+.+}-{0:0}, at: mnt_want_write+0x20/0x50 [162513.520815] #1: ffff8f3d80a650b8 (&type->i_mutex_dir_key#10){++++}-{3:3}, at: vfs_setxattr+0x50/0x120 [162513.520820] #2: ffff8f3d57644690 (sb_internal#2){.+.+}-{0:0}, at: start_transaction+0x40e/0x5f0 [btrfs] [162513.520833] 1 lock held by fsstress/1356196: [162513.520834] #0: ffff8f3d576440e8 (&type->s_umount_key#62){++++}-{3:3}, at: iterate_supers+0x6f/0xf0 [162513.520838] 3 locks held by fsstress/1356197: [162513.520839] #0: ffff8f3d57644470 (sb_writers#15){.+.+}-{0:0}, at: mnt_want_write+0x20/0x50 [162513.520843] #1: ffff8f3d506465e8 (&type->i_mutex_dir_key#10){++++}-{3:3}, at: path_openat+0x2a7/0xa50 [162513.520846] #2: ffff8f3d57644690 (sb_internal#2){.+.+}-{0:0}, at: start_transaction+0x40e/0x5f0 [btrfs] [162513.520858] 2 locks held by btrfs/1356211: [162513.520859] #0: ffff8f3d810cde30 (&fs_info->dev_replace.lock_finishing_cancel_unmount){+.+.}-{3:3}, at: btrfs_dev_replace_finishing+0x52/0x711 [btrfs] [162513.520877] #1: ffff8f3d57644690 (sb_internal#2){.+.+}-{0:0}, at: start_transaction+0x40e/0x5f0 [btrfs] This was weird because the stack traces show that a transaction commit, triggered by a device replace operation, is blocking trying to pause any running scrubs but there are no stack traces of blocked tasks doing a scrub. After poking around with drgn, I noticed there was a scrub task that was constantly running and blocking for shorts periods of time: >>> t = find_task(prog, 1356190) >>> prog.stack_trace(t) #0 __schedule+0x5ce/0xcfc #1 schedule+0x46/0xe4 #2 schedule_timeout+0x1df/0x475 #3 btrfs_reada_wait+0xda/0x132 #4 scrub_stripe+0x2a8/0x112f #5 scrub_chunk+0xcd/0x134 #6 scrub_enumerate_chunks+0x29e/0x5ee #7 btrfs_scrub_dev+0x2d5/0x91b #8 btrfs_ioctl+0x7f5/0x36e7 #9 __x64_sys_ioctl+0x83/0xb0 #10 do_syscall_64+0x33/0x77 #11 entry_SYSCALL_64+0x7c/0x156 Which corresponds to: int btrfs_reada_wait(void *handle) { struct reada_control *rc = handle; struct btrfs_fs_info *fs_info = rc->fs_info; while (atomic_read(&rc->elems)) { if (!atomic_read(&fs_info->reada_works_cnt)) reada_start_machine(fs_info); wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0, (HZ + 9) / 10); } (...) So the counter "rc->elems" was set to 1 and never decreased to 0, causing the scrub task to loop forever in that function. Then I used the following script for drgn to check the readahead requests: $ cat dump_reada.py import sys import drgn from drgn import NULL, Object, cast, container_of, execscript, \ reinterpret, sizeof from drgn.helpers.linux import * mnt_path = b"/home/fdmanana/btrfs-tests/scratch_1" mnt = None for mnt in for_each_mount(prog, dst = mnt_path): pass if mnt is None: sys.stderr.write(f'Error: mount point {mnt_path} not found\n') sys.exit(1) fs_info = cast('struct btrfs_fs_info *', mnt.mnt.mnt_sb.s_fs_info) def dump_re(re): nzones = re.nzones.value_() print(f're at {hex(re.value_())}') print(f'\t logical {re.logical.value_()}') print(f'\t refcnt {re.refcnt.value_()}') print(f'\t nzones {nzones}') for i in range(nzones): dev = re.zones[i].device name = dev.name.str.string_() print(f'\t\t dev id {dev.devid.value_()} name {name}') print() for _, e in radix_tree_for_each(fs_info.reada_tree): re = cast('struct reada_extent *', e) dump_re(re) $ drgn dump_reada.py re at 0xffff8f3da9d25ad8 logical 38928384 refcnt 1 nzones 1 dev id 0 name b'/dev/sdd' $ So there was one readahead extent with a single zone corresponding to the source device of that last device replace operation logged in dmesg/syslog. Also the ID of that zone's device was 0 which is a special value set in the source device of a device replace operation when the operation finishes (constant BTRFS_DEV_REPLACE_DEVID set at btrfs_dev_replace_finishing()), confirming again that device /dev/sdd was the source of a device replace operation. Normally there should be as many zones in the readahead extent as there are devices, and I wasn't expecting the extent to be in a block group with a 'single' profile, so I went and confirmed with the following drgn script that there weren't any single profile block groups: $ cat dump_block_groups.py import sys import drgn from drgn import NULL, Object, cast, container_of, execscript, \ reinterpret, sizeof from drgn.helpers.linux import * mnt_path = b"/home/fdmanana/btrfs-tests/scratch_1" mnt = None for mnt in for_each_mount(prog, dst = mnt_path): pass if mnt is None: sys.stderr.write(f'Error: mount point {mnt_path} not found\n') sys.exit(1) fs_info = cast('struct btrfs_fs_info *', mnt.mnt.mnt_sb.s_fs_info) BTRFS_BLOCK_GROUP_DATA = (1 << 0) BTRFS_BLOCK_GROUP_SYSTEM = (1 << 1) BTRFS_BLOCK_GROUP_METADATA = (1 << 2) BTRFS_BLOCK_GROUP_RAID0 = (1 << 3) BTRFS_BLOCK_GROUP_RAID1 = (1 << 4) BTRFS_BLOCK_GROUP_DUP = (1 << 5) BTRFS_BLOCK_GROUP_RAID10 = (1 << 6) BTRFS_BLOCK_GROUP_RAID5 = (1 << 7) BTRFS_BLOCK_GROUP_RAID6 = (1 << 8) BTRFS_BLOCK_GROUP_RAID1C3 = (1 << 9) BTRFS_BLOCK_GROUP_RAID1C4 = (1 << 10) def bg_flags_string(bg): flags = bg.flags.value_() ret = '' if flags & BTRFS_BLOCK_GROUP_DATA: ret = 'data' if flags & BTRFS_BLOCK_GROUP_METADATA: if len(ret) > 0: ret += '|' ret += 'meta' if flags & BTRFS_BLOCK_GROUP_SYSTEM: if len(ret) > 0: ret += '|' ret += 'system' if flags & BTRFS_BLOCK_GROUP_RAID0: ret += ' raid0' elif flags & BTRFS_BLOCK_GROUP_RAID1: ret += ' raid1' elif flags & BTRFS_BLOCK_GROUP_DUP: ret += ' dup' elif flags & BTRFS_BLOCK_GROUP_RAID10: ret += ' raid10' elif flags & BTRFS_BLOCK_GROUP_RAID5: ret += ' raid5' elif flags & BTRFS_BLOCK_GROUP_RAID6: ret += ' raid6' elif flags & BTRFS_BLOCK_GROUP_RAID1C3: ret += ' raid1c3' elif flags & BTRFS_BLOCK_GROUP_RAID1C4: ret += ' raid1c4' else: ret += ' single' return ret def dump_bg(bg): print() print(f'block group at {hex(bg.value_())}') print(f'\t start {bg.start.value_()} length {bg.length.value_()}') print(f'\t flags {bg.flags.value_()} - {bg_flags_string(bg)}') bg_root = fs_info.block_group_cache_tree.address_of_() for bg in rbtree_inorder_for_each_entry('struct btrfs_block_group', bg_root, 'cache_node'): dump_bg(bg) $ drgn dump_block_groups.py block group at 0xffff8f3d673b0400 start 22020096 length 16777216 flags 258 - system raid6 block group at 0xffff8f3d53ddb400 start 38797312 length 536870912 flags 260 - meta raid6 block group at 0xffff8f3d5f4d9c00 start 575668224 length 2147483648 flags 257 - data raid6 block group at 0xffff8f3d08189000 start 2723151872 length 67108864 flags 258 - system raid6 block group at 0xffff8f3db70ff000 start 2790260736 length 1073741824 flags 260 - meta raid6 block group at 0xffff8f3d5f4dd800 start 3864002560 length 67108864 flags 258 - system raid6 block group at 0xffff8f3d67037000 start 3931111424 length 2147483648 flags 257 - data raid6 $ So there were only 2 reasons left for having a readahead extent with a single zone: reada_find_zone(), called when creating a readahead extent, returned NULL either because we failed to find the corresponding block group or because a memory allocation failed. With some additional and custom tracing I figured out that on every further ocurrence of the problem the block group had just been deleted when we were looping to create the zones for the readahead extent (at reada_find_extent()), so we ended up with only one zone in the readahead extent, corresponding to a device that ends up getting replaced. So after figuring that out it became obvious why the hang happens: 1) Task A starts a scrub on any device of the filesystem, except for device /dev/sdd; 2) Task B starts a device replace with /dev/sdd as the source device; 3) Task A calls btrfs_reada_add() from scrub_stripe() and it is currently starting to scrub a stripe from block group X. This call to btrfs_reada_add() is the one for the extent tree. When btrfs_reada_add() calls reada_add_block(), it passes the logical address of the extent tree's root node as its 'logical' argument - a value of 38928384; 4) Task A then enters reada_find_extent(), called from reada_add_block(). It finds there isn't any existing readahead extent for the logical address 38928384, so it proceeds to the path of creating a new one. It calls btrfs_map_block() to find out which stripes exist for the block group X. On the first iteration of the for loop that iterates over the stripes, it finds the stripe for device /dev/sdd, so it creates one zone for that device and adds it to the readahead extent. Before getting into the second iteration of the loop, the cleanup kthread deletes block group X because it was empty. So in the iterations for the remaining stripes it does not add more zones to the readahead extent, because the calls to reada_find_zone() returned NULL because they couldn't find block group X anymore. As a result the new readahead extent has a single zone, corresponding to the device /dev/sdd; 4) Before task A returns to btrfs_reada_add() and queues the readahead job for the readahead work queue, task B finishes the device replace and at btrfs_dev_replace_finishing() swaps the device /dev/sdd with the new device /dev/sdg; 5) Task A returns to reada_add_block(), which increments the counter "->elems" of the reada_control structure allocated at btrfs_reada_add(). Then it returns back to btrfs_reada_add() and calls reada_start_machine(). This queues a job in the readahead work queue to run the function reada_start_machine_worker(), which calls __reada_start_machine(). At __reada_start_machine() we take the device list mutex and for each device found in the current device list, we call reada_start_machine_dev() to start the readahead work. However at this point the device /dev/sdd was already freed and is not in the device list anymore. This means the corresponding readahead for the extent at 38928384 is never started, and therefore the "->elems" counter of the reada_control structure allocated at btrfs_reada_add() never goes down to 0, causing the call to btrfs_reada_wait(), done by the scrub task, to wait forever. Note that the readahead request can be made either after the device replace started or before it started, however in pratice it is very unlikely that a device replace is able to start after a readahead request is made and is able to complete before the readahead request completes - maybe only on a very small and nearly empty filesystem. This hang however is not the only problem we can have with readahead and device removals. When the readahead extent has other zones other than the one corresponding to the device that is being removed (either by a device replace or a device remove operation), we risk having a use-after-free on the device when dropping the last reference of the readahead extent. For example if we create a readahead extent with two zones, one for the device /dev/sdd and one for the device /dev/sde: 1) Before the readahead worker starts, the device /dev/sdd is removed, and the corresponding btrfs_device structure is freed. However the readahead extent still has the zone pointing to the device structure; 2) When the readahead worker starts, it only finds device /dev/sde in the current device list of the filesystem; 3) It starts the readahead work, at reada_start_machine_dev(), using the device /dev/sde; 4) Then when it finishes reading the extent from device /dev/sde, it calls __readahead_hook() which ends up dropping the last reference on the readahead extent through the last call to reada_extent_put(); 5) At reada_extent_put() it iterates over each zone of the readahead extent and attempts to delete an element from the device's 'reada_extents' radix tree, resulting in a use-after-free, as the device pointer of the zone for /dev/sdd is now stale. We can also access the device after dropping the last reference of a zone, through reada_zone_release(), also called by reada_extent_put(). And a device remove suffers the same problem, however since it shrinks the device size down to zero before removing the device, it is very unlikely to still have readahead requests not completed by the time we free the device, the only possibility is if the device has a very little space allocated. While the hang problem is exclusive to scrub, since it is currently the only user of btrfs_reada_add() and btrfs_reada_wait(), the use-after-free problem affects any path that triggers readhead, which includes btree_readahead_hook() and __readahead_hook() (a readahead worker can trigger readahed for the children of a node) for example - any path that ends up calling reada_add_block() can trigger the use-after-free after a device is removed. So fix this by waiting for any readahead requests for a device to complete before removing a device, ensuring that while waiting for existing ones no new ones can be made. This problem has been around for a very long time - the readahead code was added in 2011, device remove exists since 2008 and device replace was introduced in 2013, hard to pick a specific commit for a git Fixes tag. CC: stable@vger.kernel.org # 4.4+ Reviewed-by: Josef Bacik <josef@toxicpanda.com> Signed-off-by: Filipe Manana <fdmanana@suse.com> Reviewed-by: David Sterba <dsterba@suse.com> Signed-off-by: David Sterba <dsterba@suse.com>
1070 lines
26 KiB
C
1070 lines
26 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2011 STRATO. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/writeback.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/workqueue.h>
|
|
#include "ctree.h"
|
|
#include "volumes.h"
|
|
#include "disk-io.h"
|
|
#include "transaction.h"
|
|
#include "dev-replace.h"
|
|
#include "block-group.h"
|
|
|
|
#undef DEBUG
|
|
|
|
/*
|
|
* This is the implementation for the generic read ahead framework.
|
|
*
|
|
* To trigger a readahead, btrfs_reada_add must be called. It will start
|
|
* a read ahead for the given range [start, end) on tree root. The returned
|
|
* handle can either be used to wait on the readahead to finish
|
|
* (btrfs_reada_wait), or to send it to the background (btrfs_reada_detach).
|
|
*
|
|
* The read ahead works as follows:
|
|
* On btrfs_reada_add, the root of the tree is inserted into a radix_tree.
|
|
* reada_start_machine will then search for extents to prefetch and trigger
|
|
* some reads. When a read finishes for a node, all contained node/leaf
|
|
* pointers that lie in the given range will also be enqueued. The reads will
|
|
* be triggered in sequential order, thus giving a big win over a naive
|
|
* enumeration. It will also make use of multi-device layouts. Each disk
|
|
* will have its on read pointer and all disks will by utilized in parallel.
|
|
* Also will no two disks read both sides of a mirror simultaneously, as this
|
|
* would waste seeking capacity. Instead both disks will read different parts
|
|
* of the filesystem.
|
|
* Any number of readaheads can be started in parallel. The read order will be
|
|
* determined globally, i.e. 2 parallel readaheads will normally finish faster
|
|
* than the 2 started one after another.
|
|
*/
|
|
|
|
#define MAX_IN_FLIGHT 6
|
|
|
|
struct reada_extctl {
|
|
struct list_head list;
|
|
struct reada_control *rc;
|
|
u64 generation;
|
|
};
|
|
|
|
struct reada_extent {
|
|
u64 logical;
|
|
struct btrfs_key top;
|
|
struct list_head extctl;
|
|
int refcnt;
|
|
spinlock_t lock;
|
|
struct reada_zone *zones[BTRFS_MAX_MIRRORS];
|
|
int nzones;
|
|
int scheduled;
|
|
};
|
|
|
|
struct reada_zone {
|
|
u64 start;
|
|
u64 end;
|
|
u64 elems;
|
|
struct list_head list;
|
|
spinlock_t lock;
|
|
int locked;
|
|
struct btrfs_device *device;
|
|
struct btrfs_device *devs[BTRFS_MAX_MIRRORS]; /* full list, incl
|
|
* self */
|
|
int ndevs;
|
|
struct kref refcnt;
|
|
};
|
|
|
|
struct reada_machine_work {
|
|
struct btrfs_work work;
|
|
struct btrfs_fs_info *fs_info;
|
|
};
|
|
|
|
static void reada_extent_put(struct btrfs_fs_info *, struct reada_extent *);
|
|
static void reada_control_release(struct kref *kref);
|
|
static void reada_zone_release(struct kref *kref);
|
|
static void reada_start_machine(struct btrfs_fs_info *fs_info);
|
|
static void __reada_start_machine(struct btrfs_fs_info *fs_info);
|
|
|
|
static int reada_add_block(struct reada_control *rc, u64 logical,
|
|
struct btrfs_key *top, u64 generation);
|
|
|
|
/* recurses */
|
|
/* in case of err, eb might be NULL */
|
|
static void __readahead_hook(struct btrfs_fs_info *fs_info,
|
|
struct reada_extent *re, struct extent_buffer *eb,
|
|
int err)
|
|
{
|
|
int nritems;
|
|
int i;
|
|
u64 bytenr;
|
|
u64 generation;
|
|
struct list_head list;
|
|
|
|
spin_lock(&re->lock);
|
|
/*
|
|
* just take the full list from the extent. afterwards we
|
|
* don't need the lock anymore
|
|
*/
|
|
list_replace_init(&re->extctl, &list);
|
|
re->scheduled = 0;
|
|
spin_unlock(&re->lock);
|
|
|
|
/*
|
|
* this is the error case, the extent buffer has not been
|
|
* read correctly. We won't access anything from it and
|
|
* just cleanup our data structures. Effectively this will
|
|
* cut the branch below this node from read ahead.
|
|
*/
|
|
if (err)
|
|
goto cleanup;
|
|
|
|
/*
|
|
* FIXME: currently we just set nritems to 0 if this is a leaf,
|
|
* effectively ignoring the content. In a next step we could
|
|
* trigger more readahead depending from the content, e.g.
|
|
* fetch the checksums for the extents in the leaf.
|
|
*/
|
|
if (!btrfs_header_level(eb))
|
|
goto cleanup;
|
|
|
|
nritems = btrfs_header_nritems(eb);
|
|
generation = btrfs_header_generation(eb);
|
|
for (i = 0; i < nritems; i++) {
|
|
struct reada_extctl *rec;
|
|
u64 n_gen;
|
|
struct btrfs_key key;
|
|
struct btrfs_key next_key;
|
|
|
|
btrfs_node_key_to_cpu(eb, &key, i);
|
|
if (i + 1 < nritems)
|
|
btrfs_node_key_to_cpu(eb, &next_key, i + 1);
|
|
else
|
|
next_key = re->top;
|
|
bytenr = btrfs_node_blockptr(eb, i);
|
|
n_gen = btrfs_node_ptr_generation(eb, i);
|
|
|
|
list_for_each_entry(rec, &list, list) {
|
|
struct reada_control *rc = rec->rc;
|
|
|
|
/*
|
|
* if the generation doesn't match, just ignore this
|
|
* extctl. This will probably cut off a branch from
|
|
* prefetch. Alternatively one could start a new (sub-)
|
|
* prefetch for this branch, starting again from root.
|
|
* FIXME: move the generation check out of this loop
|
|
*/
|
|
#ifdef DEBUG
|
|
if (rec->generation != generation) {
|
|
btrfs_debug(fs_info,
|
|
"generation mismatch for (%llu,%d,%llu) %llu != %llu",
|
|
key.objectid, key.type, key.offset,
|
|
rec->generation, generation);
|
|
}
|
|
#endif
|
|
if (rec->generation == generation &&
|
|
btrfs_comp_cpu_keys(&key, &rc->key_end) < 0 &&
|
|
btrfs_comp_cpu_keys(&next_key, &rc->key_start) > 0)
|
|
reada_add_block(rc, bytenr, &next_key, n_gen);
|
|
}
|
|
}
|
|
|
|
cleanup:
|
|
/*
|
|
* free extctl records
|
|
*/
|
|
while (!list_empty(&list)) {
|
|
struct reada_control *rc;
|
|
struct reada_extctl *rec;
|
|
|
|
rec = list_first_entry(&list, struct reada_extctl, list);
|
|
list_del(&rec->list);
|
|
rc = rec->rc;
|
|
kfree(rec);
|
|
|
|
kref_get(&rc->refcnt);
|
|
if (atomic_dec_and_test(&rc->elems)) {
|
|
kref_put(&rc->refcnt, reada_control_release);
|
|
wake_up(&rc->wait);
|
|
}
|
|
kref_put(&rc->refcnt, reada_control_release);
|
|
|
|
reada_extent_put(fs_info, re); /* one ref for each entry */
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
int btree_readahead_hook(struct extent_buffer *eb, int err)
|
|
{
|
|
struct btrfs_fs_info *fs_info = eb->fs_info;
|
|
int ret = 0;
|
|
struct reada_extent *re;
|
|
|
|
/* find extent */
|
|
spin_lock(&fs_info->reada_lock);
|
|
re = radix_tree_lookup(&fs_info->reada_tree,
|
|
eb->start >> PAGE_SHIFT);
|
|
if (re)
|
|
re->refcnt++;
|
|
spin_unlock(&fs_info->reada_lock);
|
|
if (!re) {
|
|
ret = -1;
|
|
goto start_machine;
|
|
}
|
|
|
|
__readahead_hook(fs_info, re, eb, err);
|
|
reada_extent_put(fs_info, re); /* our ref */
|
|
|
|
start_machine:
|
|
reada_start_machine(fs_info);
|
|
return ret;
|
|
}
|
|
|
|
static struct reada_zone *reada_find_zone(struct btrfs_device *dev, u64 logical,
|
|
struct btrfs_bio *bbio)
|
|
{
|
|
struct btrfs_fs_info *fs_info = dev->fs_info;
|
|
int ret;
|
|
struct reada_zone *zone;
|
|
struct btrfs_block_group *cache = NULL;
|
|
u64 start;
|
|
u64 end;
|
|
int i;
|
|
|
|
zone = NULL;
|
|
spin_lock(&fs_info->reada_lock);
|
|
ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone,
|
|
logical >> PAGE_SHIFT, 1);
|
|
if (ret == 1 && logical >= zone->start && logical <= zone->end) {
|
|
kref_get(&zone->refcnt);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
return zone;
|
|
}
|
|
|
|
spin_unlock(&fs_info->reada_lock);
|
|
|
|
cache = btrfs_lookup_block_group(fs_info, logical);
|
|
if (!cache)
|
|
return NULL;
|
|
|
|
start = cache->start;
|
|
end = start + cache->length - 1;
|
|
btrfs_put_block_group(cache);
|
|
|
|
zone = kzalloc(sizeof(*zone), GFP_KERNEL);
|
|
if (!zone)
|
|
return NULL;
|
|
|
|
ret = radix_tree_preload(GFP_KERNEL);
|
|
if (ret) {
|
|
kfree(zone);
|
|
return NULL;
|
|
}
|
|
|
|
zone->start = start;
|
|
zone->end = end;
|
|
INIT_LIST_HEAD(&zone->list);
|
|
spin_lock_init(&zone->lock);
|
|
zone->locked = 0;
|
|
kref_init(&zone->refcnt);
|
|
zone->elems = 0;
|
|
zone->device = dev; /* our device always sits at index 0 */
|
|
for (i = 0; i < bbio->num_stripes; ++i) {
|
|
/* bounds have already been checked */
|
|
zone->devs[i] = bbio->stripes[i].dev;
|
|
}
|
|
zone->ndevs = bbio->num_stripes;
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
ret = radix_tree_insert(&dev->reada_zones,
|
|
(unsigned long)(zone->end >> PAGE_SHIFT),
|
|
zone);
|
|
|
|
if (ret == -EEXIST) {
|
|
kfree(zone);
|
|
ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone,
|
|
logical >> PAGE_SHIFT, 1);
|
|
if (ret == 1 && logical >= zone->start && logical <= zone->end)
|
|
kref_get(&zone->refcnt);
|
|
else
|
|
zone = NULL;
|
|
}
|
|
spin_unlock(&fs_info->reada_lock);
|
|
radix_tree_preload_end();
|
|
|
|
return zone;
|
|
}
|
|
|
|
static struct reada_extent *reada_find_extent(struct btrfs_fs_info *fs_info,
|
|
u64 logical,
|
|
struct btrfs_key *top)
|
|
{
|
|
int ret;
|
|
struct reada_extent *re = NULL;
|
|
struct reada_extent *re_exist = NULL;
|
|
struct btrfs_bio *bbio = NULL;
|
|
struct btrfs_device *dev;
|
|
struct btrfs_device *prev_dev;
|
|
u64 length;
|
|
int real_stripes;
|
|
int nzones = 0;
|
|
unsigned long index = logical >> PAGE_SHIFT;
|
|
int dev_replace_is_ongoing;
|
|
int have_zone = 0;
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
re = radix_tree_lookup(&fs_info->reada_tree, index);
|
|
if (re)
|
|
re->refcnt++;
|
|
spin_unlock(&fs_info->reada_lock);
|
|
|
|
if (re)
|
|
return re;
|
|
|
|
re = kzalloc(sizeof(*re), GFP_KERNEL);
|
|
if (!re)
|
|
return NULL;
|
|
|
|
re->logical = logical;
|
|
re->top = *top;
|
|
INIT_LIST_HEAD(&re->extctl);
|
|
spin_lock_init(&re->lock);
|
|
re->refcnt = 1;
|
|
|
|
/*
|
|
* map block
|
|
*/
|
|
length = fs_info->nodesize;
|
|
ret = btrfs_map_block(fs_info, BTRFS_MAP_GET_READ_MIRRORS, logical,
|
|
&length, &bbio, 0);
|
|
if (ret || !bbio || length < fs_info->nodesize)
|
|
goto error;
|
|
|
|
if (bbio->num_stripes > BTRFS_MAX_MIRRORS) {
|
|
btrfs_err(fs_info,
|
|
"readahead: more than %d copies not supported",
|
|
BTRFS_MAX_MIRRORS);
|
|
goto error;
|
|
}
|
|
|
|
real_stripes = bbio->num_stripes - bbio->num_tgtdevs;
|
|
for (nzones = 0; nzones < real_stripes; ++nzones) {
|
|
struct reada_zone *zone;
|
|
|
|
dev = bbio->stripes[nzones].dev;
|
|
|
|
/* cannot read ahead on missing device. */
|
|
if (!dev->bdev)
|
|
continue;
|
|
|
|
zone = reada_find_zone(dev, logical, bbio);
|
|
if (!zone)
|
|
continue;
|
|
|
|
re->zones[re->nzones++] = zone;
|
|
spin_lock(&zone->lock);
|
|
if (!zone->elems)
|
|
kref_get(&zone->refcnt);
|
|
++zone->elems;
|
|
spin_unlock(&zone->lock);
|
|
spin_lock(&fs_info->reada_lock);
|
|
kref_put(&zone->refcnt, reada_zone_release);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
}
|
|
if (re->nzones == 0) {
|
|
/* not a single zone found, error and out */
|
|
goto error;
|
|
}
|
|
|
|
/* Insert extent in reada tree + all per-device trees, all or nothing */
|
|
down_read(&fs_info->dev_replace.rwsem);
|
|
ret = radix_tree_preload(GFP_KERNEL);
|
|
if (ret) {
|
|
up_read(&fs_info->dev_replace.rwsem);
|
|
goto error;
|
|
}
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
ret = radix_tree_insert(&fs_info->reada_tree, index, re);
|
|
if (ret == -EEXIST) {
|
|
re_exist = radix_tree_lookup(&fs_info->reada_tree, index);
|
|
re_exist->refcnt++;
|
|
spin_unlock(&fs_info->reada_lock);
|
|
radix_tree_preload_end();
|
|
up_read(&fs_info->dev_replace.rwsem);
|
|
goto error;
|
|
}
|
|
if (ret) {
|
|
spin_unlock(&fs_info->reada_lock);
|
|
radix_tree_preload_end();
|
|
up_read(&fs_info->dev_replace.rwsem);
|
|
goto error;
|
|
}
|
|
radix_tree_preload_end();
|
|
prev_dev = NULL;
|
|
dev_replace_is_ongoing = btrfs_dev_replace_is_ongoing(
|
|
&fs_info->dev_replace);
|
|
for (nzones = 0; nzones < re->nzones; ++nzones) {
|
|
dev = re->zones[nzones]->device;
|
|
|
|
if (dev == prev_dev) {
|
|
/*
|
|
* in case of DUP, just add the first zone. As both
|
|
* are on the same device, there's nothing to gain
|
|
* from adding both.
|
|
* Also, it wouldn't work, as the tree is per device
|
|
* and adding would fail with EEXIST
|
|
*/
|
|
continue;
|
|
}
|
|
if (!dev->bdev)
|
|
continue;
|
|
|
|
if (test_bit(BTRFS_DEV_STATE_NO_READA, &dev->dev_state))
|
|
continue;
|
|
|
|
if (dev_replace_is_ongoing &&
|
|
dev == fs_info->dev_replace.tgtdev) {
|
|
/*
|
|
* as this device is selected for reading only as
|
|
* a last resort, skip it for read ahead.
|
|
*/
|
|
continue;
|
|
}
|
|
prev_dev = dev;
|
|
ret = radix_tree_insert(&dev->reada_extents, index, re);
|
|
if (ret) {
|
|
while (--nzones >= 0) {
|
|
dev = re->zones[nzones]->device;
|
|
BUG_ON(dev == NULL);
|
|
/* ignore whether the entry was inserted */
|
|
radix_tree_delete(&dev->reada_extents, index);
|
|
}
|
|
radix_tree_delete(&fs_info->reada_tree, index);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
up_read(&fs_info->dev_replace.rwsem);
|
|
goto error;
|
|
}
|
|
have_zone = 1;
|
|
}
|
|
if (!have_zone)
|
|
radix_tree_delete(&fs_info->reada_tree, index);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
up_read(&fs_info->dev_replace.rwsem);
|
|
|
|
if (!have_zone)
|
|
goto error;
|
|
|
|
btrfs_put_bbio(bbio);
|
|
return re;
|
|
|
|
error:
|
|
for (nzones = 0; nzones < re->nzones; ++nzones) {
|
|
struct reada_zone *zone;
|
|
|
|
zone = re->zones[nzones];
|
|
kref_get(&zone->refcnt);
|
|
spin_lock(&zone->lock);
|
|
--zone->elems;
|
|
if (zone->elems == 0) {
|
|
/*
|
|
* no fs_info->reada_lock needed, as this can't be
|
|
* the last ref
|
|
*/
|
|
kref_put(&zone->refcnt, reada_zone_release);
|
|
}
|
|
spin_unlock(&zone->lock);
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
kref_put(&zone->refcnt, reada_zone_release);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
}
|
|
btrfs_put_bbio(bbio);
|
|
kfree(re);
|
|
return re_exist;
|
|
}
|
|
|
|
static void reada_extent_put(struct btrfs_fs_info *fs_info,
|
|
struct reada_extent *re)
|
|
{
|
|
int i;
|
|
unsigned long index = re->logical >> PAGE_SHIFT;
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
if (--re->refcnt) {
|
|
spin_unlock(&fs_info->reada_lock);
|
|
return;
|
|
}
|
|
|
|
radix_tree_delete(&fs_info->reada_tree, index);
|
|
for (i = 0; i < re->nzones; ++i) {
|
|
struct reada_zone *zone = re->zones[i];
|
|
|
|
radix_tree_delete(&zone->device->reada_extents, index);
|
|
}
|
|
|
|
spin_unlock(&fs_info->reada_lock);
|
|
|
|
for (i = 0; i < re->nzones; ++i) {
|
|
struct reada_zone *zone = re->zones[i];
|
|
|
|
kref_get(&zone->refcnt);
|
|
spin_lock(&zone->lock);
|
|
--zone->elems;
|
|
if (zone->elems == 0) {
|
|
/* no fs_info->reada_lock needed, as this can't be
|
|
* the last ref */
|
|
kref_put(&zone->refcnt, reada_zone_release);
|
|
}
|
|
spin_unlock(&zone->lock);
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
kref_put(&zone->refcnt, reada_zone_release);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
}
|
|
|
|
kfree(re);
|
|
}
|
|
|
|
static void reada_zone_release(struct kref *kref)
|
|
{
|
|
struct reada_zone *zone = container_of(kref, struct reada_zone, refcnt);
|
|
|
|
radix_tree_delete(&zone->device->reada_zones,
|
|
zone->end >> PAGE_SHIFT);
|
|
|
|
kfree(zone);
|
|
}
|
|
|
|
static void reada_control_release(struct kref *kref)
|
|
{
|
|
struct reada_control *rc = container_of(kref, struct reada_control,
|
|
refcnt);
|
|
|
|
kfree(rc);
|
|
}
|
|
|
|
static int reada_add_block(struct reada_control *rc, u64 logical,
|
|
struct btrfs_key *top, u64 generation)
|
|
{
|
|
struct btrfs_fs_info *fs_info = rc->fs_info;
|
|
struct reada_extent *re;
|
|
struct reada_extctl *rec;
|
|
|
|
/* takes one ref */
|
|
re = reada_find_extent(fs_info, logical, top);
|
|
if (!re)
|
|
return -1;
|
|
|
|
rec = kzalloc(sizeof(*rec), GFP_KERNEL);
|
|
if (!rec) {
|
|
reada_extent_put(fs_info, re);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
rec->rc = rc;
|
|
rec->generation = generation;
|
|
atomic_inc(&rc->elems);
|
|
|
|
spin_lock(&re->lock);
|
|
list_add_tail(&rec->list, &re->extctl);
|
|
spin_unlock(&re->lock);
|
|
|
|
/* leave the ref on the extent */
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* called with fs_info->reada_lock held
|
|
*/
|
|
static void reada_peer_zones_set_lock(struct reada_zone *zone, int lock)
|
|
{
|
|
int i;
|
|
unsigned long index = zone->end >> PAGE_SHIFT;
|
|
|
|
for (i = 0; i < zone->ndevs; ++i) {
|
|
struct reada_zone *peer;
|
|
peer = radix_tree_lookup(&zone->devs[i]->reada_zones, index);
|
|
if (peer && peer->device != zone->device)
|
|
peer->locked = lock;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* called with fs_info->reada_lock held
|
|
*/
|
|
static int reada_pick_zone(struct btrfs_device *dev)
|
|
{
|
|
struct reada_zone *top_zone = NULL;
|
|
struct reada_zone *top_locked_zone = NULL;
|
|
u64 top_elems = 0;
|
|
u64 top_locked_elems = 0;
|
|
unsigned long index = 0;
|
|
int ret;
|
|
|
|
if (dev->reada_curr_zone) {
|
|
reada_peer_zones_set_lock(dev->reada_curr_zone, 0);
|
|
kref_put(&dev->reada_curr_zone->refcnt, reada_zone_release);
|
|
dev->reada_curr_zone = NULL;
|
|
}
|
|
/* pick the zone with the most elements */
|
|
while (1) {
|
|
struct reada_zone *zone;
|
|
|
|
ret = radix_tree_gang_lookup(&dev->reada_zones,
|
|
(void **)&zone, index, 1);
|
|
if (ret == 0)
|
|
break;
|
|
index = (zone->end >> PAGE_SHIFT) + 1;
|
|
if (zone->locked) {
|
|
if (zone->elems > top_locked_elems) {
|
|
top_locked_elems = zone->elems;
|
|
top_locked_zone = zone;
|
|
}
|
|
} else {
|
|
if (zone->elems > top_elems) {
|
|
top_elems = zone->elems;
|
|
top_zone = zone;
|
|
}
|
|
}
|
|
}
|
|
if (top_zone)
|
|
dev->reada_curr_zone = top_zone;
|
|
else if (top_locked_zone)
|
|
dev->reada_curr_zone = top_locked_zone;
|
|
else
|
|
return 0;
|
|
|
|
dev->reada_next = dev->reada_curr_zone->start;
|
|
kref_get(&dev->reada_curr_zone->refcnt);
|
|
reada_peer_zones_set_lock(dev->reada_curr_zone, 1);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int reada_tree_block_flagged(struct btrfs_fs_info *fs_info, u64 bytenr,
|
|
int mirror_num, struct extent_buffer **eb)
|
|
{
|
|
struct extent_buffer *buf = NULL;
|
|
int ret;
|
|
|
|
buf = btrfs_find_create_tree_block(fs_info, bytenr);
|
|
if (IS_ERR(buf))
|
|
return 0;
|
|
|
|
set_bit(EXTENT_BUFFER_READAHEAD, &buf->bflags);
|
|
|
|
ret = read_extent_buffer_pages(buf, WAIT_PAGE_LOCK, mirror_num);
|
|
if (ret) {
|
|
free_extent_buffer_stale(buf);
|
|
return ret;
|
|
}
|
|
|
|
if (test_bit(EXTENT_BUFFER_CORRUPT, &buf->bflags)) {
|
|
free_extent_buffer_stale(buf);
|
|
return -EIO;
|
|
} else if (extent_buffer_uptodate(buf)) {
|
|
*eb = buf;
|
|
} else {
|
|
free_extent_buffer(buf);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int reada_start_machine_dev(struct btrfs_device *dev)
|
|
{
|
|
struct btrfs_fs_info *fs_info = dev->fs_info;
|
|
struct reada_extent *re = NULL;
|
|
int mirror_num = 0;
|
|
struct extent_buffer *eb = NULL;
|
|
u64 logical;
|
|
int ret;
|
|
int i;
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
if (dev->reada_curr_zone == NULL) {
|
|
ret = reada_pick_zone(dev);
|
|
if (!ret) {
|
|
spin_unlock(&fs_info->reada_lock);
|
|
return 0;
|
|
}
|
|
}
|
|
/*
|
|
* FIXME currently we issue the reads one extent at a time. If we have
|
|
* a contiguous block of extents, we could also coagulate them or use
|
|
* plugging to speed things up
|
|
*/
|
|
ret = radix_tree_gang_lookup(&dev->reada_extents, (void **)&re,
|
|
dev->reada_next >> PAGE_SHIFT, 1);
|
|
if (ret == 0 || re->logical > dev->reada_curr_zone->end) {
|
|
ret = reada_pick_zone(dev);
|
|
if (!ret) {
|
|
spin_unlock(&fs_info->reada_lock);
|
|
return 0;
|
|
}
|
|
re = NULL;
|
|
ret = radix_tree_gang_lookup(&dev->reada_extents, (void **)&re,
|
|
dev->reada_next >> PAGE_SHIFT, 1);
|
|
}
|
|
if (ret == 0) {
|
|
spin_unlock(&fs_info->reada_lock);
|
|
return 0;
|
|
}
|
|
dev->reada_next = re->logical + fs_info->nodesize;
|
|
re->refcnt++;
|
|
|
|
spin_unlock(&fs_info->reada_lock);
|
|
|
|
spin_lock(&re->lock);
|
|
if (re->scheduled || list_empty(&re->extctl)) {
|
|
spin_unlock(&re->lock);
|
|
reada_extent_put(fs_info, re);
|
|
return 0;
|
|
}
|
|
re->scheduled = 1;
|
|
spin_unlock(&re->lock);
|
|
|
|
/*
|
|
* find mirror num
|
|
*/
|
|
for (i = 0; i < re->nzones; ++i) {
|
|
if (re->zones[i]->device == dev) {
|
|
mirror_num = i + 1;
|
|
break;
|
|
}
|
|
}
|
|
logical = re->logical;
|
|
|
|
atomic_inc(&dev->reada_in_flight);
|
|
ret = reada_tree_block_flagged(fs_info, logical, mirror_num, &eb);
|
|
if (ret)
|
|
__readahead_hook(fs_info, re, NULL, ret);
|
|
else if (eb)
|
|
__readahead_hook(fs_info, re, eb, ret);
|
|
|
|
if (eb)
|
|
free_extent_buffer(eb);
|
|
|
|
atomic_dec(&dev->reada_in_flight);
|
|
reada_extent_put(fs_info, re);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
static void reada_start_machine_worker(struct btrfs_work *work)
|
|
{
|
|
struct reada_machine_work *rmw;
|
|
int old_ioprio;
|
|
|
|
rmw = container_of(work, struct reada_machine_work, work);
|
|
|
|
old_ioprio = IOPRIO_PRIO_VALUE(task_nice_ioclass(current),
|
|
task_nice_ioprio(current));
|
|
set_task_ioprio(current, BTRFS_IOPRIO_READA);
|
|
__reada_start_machine(rmw->fs_info);
|
|
set_task_ioprio(current, old_ioprio);
|
|
|
|
atomic_dec(&rmw->fs_info->reada_works_cnt);
|
|
|
|
kfree(rmw);
|
|
}
|
|
|
|
/* Try to start up to 10k READA requests for a group of devices */
|
|
static int reada_start_for_fsdevs(struct btrfs_fs_devices *fs_devices)
|
|
{
|
|
u64 enqueued;
|
|
u64 total = 0;
|
|
struct btrfs_device *device;
|
|
|
|
do {
|
|
enqueued = 0;
|
|
list_for_each_entry(device, &fs_devices->devices, dev_list) {
|
|
if (atomic_read(&device->reada_in_flight) <
|
|
MAX_IN_FLIGHT)
|
|
enqueued += reada_start_machine_dev(device);
|
|
}
|
|
total += enqueued;
|
|
} while (enqueued && total < 10000);
|
|
|
|
return total;
|
|
}
|
|
|
|
static void __reada_start_machine(struct btrfs_fs_info *fs_info)
|
|
{
|
|
struct btrfs_fs_devices *fs_devices = fs_info->fs_devices, *seed_devs;
|
|
int i;
|
|
u64 enqueued = 0;
|
|
|
|
mutex_lock(&fs_devices->device_list_mutex);
|
|
|
|
enqueued += reada_start_for_fsdevs(fs_devices);
|
|
list_for_each_entry(seed_devs, &fs_devices->seed_list, seed_list)
|
|
enqueued += reada_start_for_fsdevs(seed_devs);
|
|
|
|
mutex_unlock(&fs_devices->device_list_mutex);
|
|
if (enqueued == 0)
|
|
return;
|
|
|
|
/*
|
|
* If everything is already in the cache, this is effectively single
|
|
* threaded. To a) not hold the caller for too long and b) to utilize
|
|
* more cores, we broke the loop above after 10000 iterations and now
|
|
* enqueue to workers to finish it. This will distribute the load to
|
|
* the cores.
|
|
*/
|
|
for (i = 0; i < 2; ++i) {
|
|
reada_start_machine(fs_info);
|
|
if (atomic_read(&fs_info->reada_works_cnt) >
|
|
BTRFS_MAX_MIRRORS * 2)
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void reada_start_machine(struct btrfs_fs_info *fs_info)
|
|
{
|
|
struct reada_machine_work *rmw;
|
|
|
|
rmw = kzalloc(sizeof(*rmw), GFP_KERNEL);
|
|
if (!rmw) {
|
|
/* FIXME we cannot handle this properly right now */
|
|
BUG();
|
|
}
|
|
btrfs_init_work(&rmw->work, reada_start_machine_worker, NULL, NULL);
|
|
rmw->fs_info = fs_info;
|
|
|
|
btrfs_queue_work(fs_info->readahead_workers, &rmw->work);
|
|
atomic_inc(&fs_info->reada_works_cnt);
|
|
}
|
|
|
|
#ifdef DEBUG
|
|
static void dump_devs(struct btrfs_fs_info *fs_info, int all)
|
|
{
|
|
struct btrfs_device *device;
|
|
struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
|
|
unsigned long index;
|
|
int ret;
|
|
int i;
|
|
int j;
|
|
int cnt;
|
|
|
|
spin_lock(&fs_info->reada_lock);
|
|
list_for_each_entry(device, &fs_devices->devices, dev_list) {
|
|
btrfs_debug(fs_info, "dev %lld has %d in flight", device->devid,
|
|
atomic_read(&device->reada_in_flight));
|
|
index = 0;
|
|
while (1) {
|
|
struct reada_zone *zone;
|
|
ret = radix_tree_gang_lookup(&device->reada_zones,
|
|
(void **)&zone, index, 1);
|
|
if (ret == 0)
|
|
break;
|
|
pr_debug(" zone %llu-%llu elems %llu locked %d devs",
|
|
zone->start, zone->end, zone->elems,
|
|
zone->locked);
|
|
for (j = 0; j < zone->ndevs; ++j) {
|
|
pr_cont(" %lld",
|
|
zone->devs[j]->devid);
|
|
}
|
|
if (device->reada_curr_zone == zone)
|
|
pr_cont(" curr off %llu",
|
|
device->reada_next - zone->start);
|
|
pr_cont("\n");
|
|
index = (zone->end >> PAGE_SHIFT) + 1;
|
|
}
|
|
cnt = 0;
|
|
index = 0;
|
|
while (all) {
|
|
struct reada_extent *re = NULL;
|
|
|
|
ret = radix_tree_gang_lookup(&device->reada_extents,
|
|
(void **)&re, index, 1);
|
|
if (ret == 0)
|
|
break;
|
|
pr_debug(" re: logical %llu size %u empty %d scheduled %d",
|
|
re->logical, fs_info->nodesize,
|
|
list_empty(&re->extctl), re->scheduled);
|
|
|
|
for (i = 0; i < re->nzones; ++i) {
|
|
pr_cont(" zone %llu-%llu devs",
|
|
re->zones[i]->start,
|
|
re->zones[i]->end);
|
|
for (j = 0; j < re->zones[i]->ndevs; ++j) {
|
|
pr_cont(" %lld",
|
|
re->zones[i]->devs[j]->devid);
|
|
}
|
|
}
|
|
pr_cont("\n");
|
|
index = (re->logical >> PAGE_SHIFT) + 1;
|
|
if (++cnt > 15)
|
|
break;
|
|
}
|
|
}
|
|
|
|
index = 0;
|
|
cnt = 0;
|
|
while (all) {
|
|
struct reada_extent *re = NULL;
|
|
|
|
ret = radix_tree_gang_lookup(&fs_info->reada_tree, (void **)&re,
|
|
index, 1);
|
|
if (ret == 0)
|
|
break;
|
|
if (!re->scheduled) {
|
|
index = (re->logical >> PAGE_SHIFT) + 1;
|
|
continue;
|
|
}
|
|
pr_debug("re: logical %llu size %u list empty %d scheduled %d",
|
|
re->logical, fs_info->nodesize,
|
|
list_empty(&re->extctl), re->scheduled);
|
|
for (i = 0; i < re->nzones; ++i) {
|
|
pr_cont(" zone %llu-%llu devs",
|
|
re->zones[i]->start,
|
|
re->zones[i]->end);
|
|
for (j = 0; j < re->zones[i]->ndevs; ++j) {
|
|
pr_cont(" %lld",
|
|
re->zones[i]->devs[j]->devid);
|
|
}
|
|
}
|
|
pr_cont("\n");
|
|
index = (re->logical >> PAGE_SHIFT) + 1;
|
|
}
|
|
spin_unlock(&fs_info->reada_lock);
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* interface
|
|
*/
|
|
struct reada_control *btrfs_reada_add(struct btrfs_root *root,
|
|
struct btrfs_key *key_start, struct btrfs_key *key_end)
|
|
{
|
|
struct reada_control *rc;
|
|
u64 start;
|
|
u64 generation;
|
|
int ret;
|
|
struct extent_buffer *node;
|
|
static struct btrfs_key max_key = {
|
|
.objectid = (u64)-1,
|
|
.type = (u8)-1,
|
|
.offset = (u64)-1
|
|
};
|
|
|
|
rc = kzalloc(sizeof(*rc), GFP_KERNEL);
|
|
if (!rc)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
rc->fs_info = root->fs_info;
|
|
rc->key_start = *key_start;
|
|
rc->key_end = *key_end;
|
|
atomic_set(&rc->elems, 0);
|
|
init_waitqueue_head(&rc->wait);
|
|
kref_init(&rc->refcnt);
|
|
kref_get(&rc->refcnt); /* one ref for having elements */
|
|
|
|
node = btrfs_root_node(root);
|
|
start = node->start;
|
|
generation = btrfs_header_generation(node);
|
|
free_extent_buffer(node);
|
|
|
|
ret = reada_add_block(rc, start, &max_key, generation);
|
|
if (ret) {
|
|
kfree(rc);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
reada_start_machine(root->fs_info);
|
|
|
|
return rc;
|
|
}
|
|
|
|
#ifdef DEBUG
|
|
int btrfs_reada_wait(void *handle)
|
|
{
|
|
struct reada_control *rc = handle;
|
|
struct btrfs_fs_info *fs_info = rc->fs_info;
|
|
|
|
while (atomic_read(&rc->elems)) {
|
|
if (!atomic_read(&fs_info->reada_works_cnt))
|
|
reada_start_machine(fs_info);
|
|
wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0,
|
|
5 * HZ);
|
|
dump_devs(fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0);
|
|
}
|
|
|
|
dump_devs(fs_info, atomic_read(&rc->elems) < 10 ? 1 : 0);
|
|
|
|
kref_put(&rc->refcnt, reada_control_release);
|
|
|
|
return 0;
|
|
}
|
|
#else
|
|
int btrfs_reada_wait(void *handle)
|
|
{
|
|
struct reada_control *rc = handle;
|
|
struct btrfs_fs_info *fs_info = rc->fs_info;
|
|
|
|
while (atomic_read(&rc->elems)) {
|
|
if (!atomic_read(&fs_info->reada_works_cnt))
|
|
reada_start_machine(fs_info);
|
|
wait_event_timeout(rc->wait, atomic_read(&rc->elems) == 0,
|
|
(HZ + 9) / 10);
|
|
}
|
|
|
|
kref_put(&rc->refcnt, reada_control_release);
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
void btrfs_reada_detach(void *handle)
|
|
{
|
|
struct reada_control *rc = handle;
|
|
|
|
kref_put(&rc->refcnt, reada_control_release);
|
|
}
|
|
|
|
/*
|
|
* Before removing a device (device replace or device remove ioctls), call this
|
|
* function to wait for all existing readahead requests on the device and to
|
|
* make sure no one queues more readahead requests for the device.
|
|
*
|
|
* Must be called without holding neither the device list mutex nor the device
|
|
* replace semaphore, otherwise it will deadlock.
|
|
*/
|
|
void btrfs_reada_remove_dev(struct btrfs_device *dev)
|
|
{
|
|
struct btrfs_fs_info *fs_info = dev->fs_info;
|
|
|
|
/* Serialize with readahead extent creation at reada_find_extent(). */
|
|
spin_lock(&fs_info->reada_lock);
|
|
set_bit(BTRFS_DEV_STATE_NO_READA, &dev->dev_state);
|
|
spin_unlock(&fs_info->reada_lock);
|
|
|
|
/*
|
|
* There might be readahead requests added to the radix trees which
|
|
* were not yet added to the readahead work queue. We need to start
|
|
* them and wait for their completion, otherwise we can end up with
|
|
* use-after-free problems when dropping the last reference on the
|
|
* readahead extents and their zones, as they need to access the
|
|
* device structure.
|
|
*/
|
|
reada_start_machine(fs_info);
|
|
btrfs_flush_workqueue(fs_info->readahead_workers);
|
|
}
|
|
|
|
/*
|
|
* If when removing a device (device replace or device remove ioctls) an error
|
|
* happens after calling btrfs_reada_remove_dev(), call this to undo what that
|
|
* function did. This is safe to call even if btrfs_reada_remove_dev() was not
|
|
* called before.
|
|
*/
|
|
void btrfs_reada_undo_remove_dev(struct btrfs_device *dev)
|
|
{
|
|
spin_lock(&dev->fs_info->reada_lock);
|
|
clear_bit(BTRFS_DEV_STATE_NO_READA, &dev->dev_state);
|
|
spin_unlock(&dev->fs_info->reada_lock);
|
|
}
|