btrfs: improve batch deletion of delayed dir index items

Currently we group delayed dir index items for deletion in a single batch
(single btree operation) as long as they all exist in the same leaf and as
long as their keys are sequential in the key space. For example if we have
a leaf that has dir index items with offsets:

    2, 3, 4, 6, 7, 10

And we have delayed dir index items for deleting all these indexes, and
no delayed items for any other index keys in between, then we end up
deleting in 3 batches:

1) First batch for indexes 2, 3 and 4;
2) Second batch for indexes 6 and 7;
3) Third batch for index 10.

This is a waste because we can delete all the index keys in a single
batch. What matters is that each consecutive delayed index key matches
each consecutive dir index key in a leaf.

So update the logic at btrfs_batch_delete_items() to check only for a
key match between delayed dir index items and dir index items in a leaf.
Also avoid the useless first iteration on comparing the key of the
first slot to delete with the key of the first delayed item, as it's
silly since they always match, as the delayed item's key was used for
the btree search that gave us the path we have.

This is more efficient and reduces runtime of running delayed items, as
well as lock contention on the subvolume's tree.

For example, the following test script:

  $ cat test.sh
  #!/bin/bash

  DEV=/dev/sdj
  MNT=/mnt/sdj

  mkfs.btrfs -f $DEV
  mount $DEV $MNT

  NUM_FILES=1000

  mkdir $MNT/testdir
  for ((i = 1; i <= $NUM_FILES; i++)); do
      echo -n > $MNT/testdir/file_$i
  done

  # Now delete every other file, to create gaps in the dir index keys.
  for ((i = 1; i <= $NUM_FILES; i += 2)); do
      rm -f $MNT/testdir/file_$i
  done

  # Sync to force any delayed items to be flushed to the tree.
  sync

  start=$(date +%s%N)
  rm -fr $MNT/testdir
  end=$(date +%s%N)
  dur=$(( (end - start) / 1000000 ))

  echo -e "\nrm -fr took $dur milliseconds"

  umount $MNT

Running that test script while having the following bpftrace script
running in another shell:

  $ cat bpf-measure.sh
  #!/usr/bin/bpftrace

  /* Add 'noinline' to btrfs_delete_delayed_items()'s definition. */
  k:btrfs_delete_delayed_items
  {
      @start_delete_delayed_items[tid] = nsecs;
  }

  k:btrfs_del_items
  /@start_delete_delayed_items[tid]/
  {
      @delete_batches = count();
  }

  kr:btrfs_delete_delayed_items
  /@start_delete_delayed_items[tid]/
  {
      $dur = (nsecs - @start_delete_delayed_items[tid]) / 1000;
      @btrfs_delete_delayed_items_total_time = sum($dur);
      delete(@start_delete_delayed_items[tid]);
  }

Before this change:

@btrfs_delete_delayed_items_total_time: 9563
@delete_batches: 1001

After this change:

@btrfs_delete_delayed_items_total_time: 7328
@delete_batches: 509

Reviewed-by: Nikolay Borisov <nborisov@suse.com>
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Reviewed-by: David Sterba <dsterba@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
This commit is contained in:
Filipe Manana 2022-05-31 16:06:38 +01:00 committed by David Sterba
parent 36baa2c751
commit 4bd02d9012
1 changed files with 27 additions and 37 deletions

View File

@ -798,68 +798,58 @@ static int btrfs_batch_delete_items(struct btrfs_trans_handle *trans,
{
struct btrfs_delayed_item *curr, *next;
struct extent_buffer *leaf = path->nodes[0];
struct btrfs_key key;
struct list_head head;
int nitems, i, last_item;
int ret = 0;
LIST_HEAD(batch_list);
int nitems, slot, last_slot;
int ret;
ASSERT(leaf != NULL);
i = path->slots[0];
last_item = btrfs_header_nritems(leaf) - 1;
slot = path->slots[0];
last_slot = btrfs_header_nritems(leaf) - 1;
/*
* Our caller always gives us a path pointing to an existing item, so
* this can not happen.
*/
ASSERT(i <= last_item);
if (WARN_ON(i > last_item))
ASSERT(slot <= last_slot);
if (WARN_ON(slot > last_slot))
return -ENOENT;
next = item;
INIT_LIST_HEAD(&head);
btrfs_item_key_to_cpu(leaf, &key, i);
nitems = 0;
/*
* count the number of the dir index items that we can delete in batch
*/
while (btrfs_comp_cpu_keys(&next->key, &key) == 0) {
list_add_tail(&next->tree_list, &head);
nitems++;
nitems = 1;
curr = item;
list_add_tail(&curr->tree_list, &batch_list);
/*
* Keep checking if the next delayed item matches the next item in the
* leaf - if so, we can add it to the batch of items to delete from the
* leaf.
*/
while (slot < last_slot) {
struct btrfs_key key;
curr = next;
next = __btrfs_next_delayed_item(curr);
if (!next)
break;
if (!btrfs_is_continuous_delayed_item(curr, next))
slot++;
btrfs_item_key_to_cpu(leaf, &key, slot);
if (btrfs_comp_cpu_keys(&next->key, &key) != 0)
break;
i++;
if (i > last_item)
break;
btrfs_item_key_to_cpu(leaf, &key, i);
nitems++;
curr = next;
list_add_tail(&curr->tree_list, &batch_list);
}
/*
* Our caller always gives us a path pointing to an existing item, so
* this can not happen.
*/
ASSERT(nitems >= 1);
if (nitems < 1)
return -ENOENT;
ret = btrfs_del_items(trans, root, path, path->slots[0], nitems);
if (ret)
goto out;
return ret;
list_for_each_entry_safe(curr, next, &head, tree_list) {
list_for_each_entry_safe(curr, next, &batch_list, tree_list) {
btrfs_delayed_item_release_metadata(root, curr);
list_del(&curr->tree_list);
btrfs_release_delayed_item(curr);
}
out:
return ret;
return 0;
}
static int btrfs_delete_delayed_items(struct btrfs_trans_handle *trans,