radix-tree: rename radix_tree_is_indirect_ptr()

As with indirect_to_ptr(), ptr_to_indirect() and
RADIX_TREE_INDIRECT_PTR, change radix_tree_is_indirect_ptr() to
radix_tree_is_internal_node().

Signed-off-by: Matthew Wilcox <willy@linux.intel.com>
Cc: Konstantin Khlebnikov <koct9i@gmail.com>
Cc: Kirill Shutemov <kirill.shutemov@linux.intel.com>
Cc: Jan Kara <jack@suse.com>
Cc: Neil Brown <neilb@suse.de>
Cc: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Matthew Wilcox 2016-05-20 17:03:30 -07:00 committed by Linus Torvalds
parent 4dd6c0987c
commit b194d16c27
3 changed files with 31 additions and 31 deletions

View File

@ -57,7 +57,7 @@
#define RADIX_DAX_ENTRY(sector, pmd) ((void *)((unsigned long)sector << \
RADIX_DAX_SHIFT | (pmd ? RADIX_DAX_PMD : RADIX_DAX_PTE)))
static inline int radix_tree_is_indirect_ptr(void *ptr)
static inline int radix_tree_is_internal_node(void *ptr)
{
return (int)((unsigned long)ptr & RADIX_TREE_INTERNAL_NODE);
}
@ -224,7 +224,7 @@ static inline void *radix_tree_deref_slot_protected(void **pslot,
*/
static inline int radix_tree_deref_retry(void *arg)
{
return unlikely(radix_tree_is_indirect_ptr(arg));
return unlikely(radix_tree_is_internal_node(arg));
}
/**
@ -259,7 +259,7 @@ static inline int radix_tree_exception(void *arg)
*/
static inline void radix_tree_replace_slot(void **pslot, void *item)
{
BUG_ON(radix_tree_is_indirect_ptr(item));
BUG_ON(radix_tree_is_internal_node(item));
rcu_assign_pointer(*pslot, item);
}
@ -468,7 +468,7 @@ radix_tree_next_slot(void **slot, struct radix_tree_iter *iter, unsigned flags)
if (unlikely(!iter->tags))
return NULL;
while (IS_ENABLED(CONFIG_RADIX_TREE_MULTIORDER) &&
radix_tree_is_indirect_ptr(slot[1])) {
radix_tree_is_internal_node(slot[1])) {
if (entry_to_node(slot[1]) == canon) {
iter->tags >>= 1;
iter->index = __radix_tree_iter_add(iter, 1);
@ -498,7 +498,7 @@ radix_tree_next_slot(void **slot, struct radix_tree_iter *iter, unsigned flags)
iter->index = __radix_tree_iter_add(iter, 1);
if (IS_ENABLED(CONFIG_RADIX_TREE_MULTIORDER) &&
radix_tree_is_indirect_ptr(*slot)) {
radix_tree_is_internal_node(*slot)) {
if (entry_to_node(*slot) == canon)
continue;
iter->next_index = iter->index;

View File

@ -100,7 +100,7 @@ static unsigned radix_tree_descend(struct radix_tree_node *parent,
void **entry = rcu_dereference_raw(parent->slots[offset]);
#ifdef CONFIG_RADIX_TREE_MULTIORDER
if (radix_tree_is_indirect_ptr(entry)) {
if (radix_tree_is_internal_node(entry)) {
unsigned long siboff = get_slot_offset(parent, entry);
if (siboff < RADIX_TREE_MAP_SIZE) {
offset = siboff;
@ -232,7 +232,7 @@ static void dump_node(struct radix_tree_node *node, unsigned long index)
entry, i,
*(void **)entry_to_node(entry),
first, last);
} else if (!radix_tree_is_indirect_ptr(entry)) {
} else if (!radix_tree_is_internal_node(entry)) {
pr_debug("radix entry %p offset %ld indices %ld-%ld\n",
entry, i, first, last);
} else {
@ -247,7 +247,7 @@ static void radix_tree_dump(struct radix_tree_root *root)
pr_debug("radix root: %p rnode %p tags %x\n",
root, root->rnode,
root->gfp_mask >> __GFP_BITS_SHIFT);
if (!radix_tree_is_indirect_ptr(root->rnode))
if (!radix_tree_is_internal_node(root->rnode))
return;
dump_node(entry_to_node(root->rnode), 0);
}
@ -302,7 +302,7 @@ radix_tree_node_alloc(struct radix_tree_root *root)
ret = kmem_cache_alloc(radix_tree_node_cachep,
gfp_mask | __GFP_ACCOUNT);
out:
BUG_ON(radix_tree_is_indirect_ptr(ret));
BUG_ON(radix_tree_is_internal_node(ret));
return ret;
}
@ -421,7 +421,7 @@ static unsigned radix_tree_load_root(struct radix_tree_root *root,
*nodep = node;
if (likely(radix_tree_is_indirect_ptr(node))) {
if (likely(radix_tree_is_internal_node(node))) {
node = entry_to_node(node);
*maxindex = node_maxindex(node);
return node->shift + RADIX_TREE_MAP_SHIFT;
@ -467,7 +467,7 @@ static int radix_tree_extend(struct radix_tree_root *root,
node->offset = 0;
node->count = 1;
node->parent = NULL;
if (radix_tree_is_indirect_ptr(slot))
if (radix_tree_is_internal_node(slot))
entry_to_node(slot)->parent = node;
node->slots[0] = slot;
slot = node_to_entry(node);
@ -535,7 +535,7 @@ int __radix_tree_create(struct radix_tree_root *root, unsigned long index,
} else
rcu_assign_pointer(root->rnode,
node_to_entry(slot));
} else if (!radix_tree_is_indirect_ptr(slot))
} else if (!radix_tree_is_internal_node(slot))
break;
/* Go a level down */
@ -585,7 +585,7 @@ int __radix_tree_insert(struct radix_tree_root *root, unsigned long index,
void **slot;
int error;
BUG_ON(radix_tree_is_indirect_ptr(item));
BUG_ON(radix_tree_is_internal_node(item));
error = __radix_tree_create(root, index, order, &node, &slot);
if (error)
@ -637,7 +637,7 @@ void *__radix_tree_lookup(struct radix_tree_root *root, unsigned long index,
if (index > maxindex)
return NULL;
while (radix_tree_is_indirect_ptr(node)) {
while (radix_tree_is_internal_node(node)) {
unsigned offset;
if (node == RADIX_TREE_RETRY)
@ -720,7 +720,7 @@ void *radix_tree_tag_set(struct radix_tree_root *root,
shift = radix_tree_load_root(root, &node, &maxindex);
BUG_ON(index > maxindex);
while (radix_tree_is_indirect_ptr(node)) {
while (radix_tree_is_internal_node(node)) {
unsigned offset;
shift -= RADIX_TREE_MAP_SHIFT;
@ -770,7 +770,7 @@ void *radix_tree_tag_clear(struct radix_tree_root *root,
parent = NULL;
while (radix_tree_is_indirect_ptr(node)) {
while (radix_tree_is_internal_node(node)) {
shift -= RADIX_TREE_MAP_SHIFT;
offset = (index >> shift) & RADIX_TREE_MAP_MASK;
@ -835,7 +835,7 @@ int radix_tree_tag_get(struct radix_tree_root *root,
if (node == NULL)
return 0;
while (radix_tree_is_indirect_ptr(node)) {
while (radix_tree_is_internal_node(node)) {
int offset;
shift -= RADIX_TREE_MAP_SHIFT;
@ -900,7 +900,7 @@ void **radix_tree_next_chunk(struct radix_tree_root *root,
if (index > maxindex)
return NULL;
if (radix_tree_is_indirect_ptr(rnode)) {
if (radix_tree_is_internal_node(rnode)) {
rnode = entry_to_node(rnode);
} else if (rnode) {
/* Single-slot tree */
@ -957,7 +957,7 @@ void **radix_tree_next_chunk(struct radix_tree_root *root,
if ((slot == NULL) || (slot == RADIX_TREE_RETRY))
goto restart;
if (!radix_tree_is_indirect_ptr(slot))
if (!radix_tree_is_internal_node(slot))
break;
node = entry_to_node(slot);
@ -1039,7 +1039,7 @@ unsigned long radix_tree_range_tag_if_tagged(struct radix_tree_root *root,
*first_indexp = last_index + 1;
return 0;
}
if (!radix_tree_is_indirect_ptr(slot)) {
if (!radix_tree_is_internal_node(slot)) {
*first_indexp = last_index + 1;
root_tag_set(root, settag);
return 1;
@ -1059,7 +1059,7 @@ unsigned long radix_tree_range_tag_if_tagged(struct radix_tree_root *root,
if (!tag_get(node, iftag, offset))
goto next;
/* Sibling slots never have tags set on them */
if (radix_tree_is_indirect_ptr(slot)) {
if (radix_tree_is_internal_node(slot)) {
node = entry_to_node(slot);
shift -= RADIX_TREE_MAP_SHIFT;
continue;
@ -1152,7 +1152,7 @@ radix_tree_gang_lookup(struct radix_tree_root *root, void **results,
results[ret] = rcu_dereference_raw(*slot);
if (!results[ret])
continue;
if (radix_tree_is_indirect_ptr(results[ret])) {
if (radix_tree_is_internal_node(results[ret])) {
slot = radix_tree_iter_retry(&iter);
continue;
}
@ -1235,7 +1235,7 @@ radix_tree_gang_lookup_tag(struct radix_tree_root *root, void **results,
results[ret] = rcu_dereference_raw(*slot);
if (!results[ret])
continue;
if (radix_tree_is_indirect_ptr(results[ret])) {
if (radix_tree_is_internal_node(results[ret])) {
slot = radix_tree_iter_retry(&iter);
continue;
}
@ -1311,7 +1311,7 @@ static unsigned long __locate(struct radix_tree_node *slot, void *item,
rcu_dereference_raw(slot->slots[i]);
if (node == RADIX_TREE_RETRY)
goto out;
if (!radix_tree_is_indirect_ptr(node)) {
if (!radix_tree_is_internal_node(node)) {
if (node == item) {
info->found_index = index;
info->stop = true;
@ -1357,7 +1357,7 @@ unsigned long radix_tree_locate_item(struct radix_tree_root *root, void *item)
do {
rcu_read_lock();
node = rcu_dereference_raw(root->rnode);
if (!radix_tree_is_indirect_ptr(node)) {
if (!radix_tree_is_internal_node(node)) {
rcu_read_unlock();
if (node == item)
info.found_index = 0;
@ -1398,7 +1398,7 @@ static inline bool radix_tree_shrink(struct radix_tree_root *root)
struct radix_tree_node *to_free = root->rnode;
struct radix_tree_node *slot;
if (!radix_tree_is_indirect_ptr(to_free))
if (!radix_tree_is_internal_node(to_free))
break;
to_free = entry_to_node(to_free);
@ -1412,10 +1412,10 @@ static inline bool radix_tree_shrink(struct radix_tree_root *root)
slot = to_free->slots[0];
if (!slot)
break;
if (!radix_tree_is_indirect_ptr(slot) && to_free->shift)
if (!radix_tree_is_internal_node(slot) && to_free->shift)
break;
if (radix_tree_is_indirect_ptr(slot))
if (radix_tree_is_internal_node(slot))
entry_to_node(slot)->parent = NULL;
/*
@ -1445,7 +1445,7 @@ static inline bool radix_tree_shrink(struct radix_tree_root *root)
* also results in a stale slot). So tag the slot as indirect
* to force callers to retry.
*/
if (!radix_tree_is_indirect_ptr(slot))
if (!radix_tree_is_internal_node(slot))
to_free->slots[0] = RADIX_TREE_RETRY;
radix_tree_node_free(to_free);

View File

@ -193,7 +193,7 @@ static int verify_node(struct radix_tree_node *slot, unsigned int tag,
void verify_tag_consistency(struct radix_tree_root *root, unsigned int tag)
{
struct radix_tree_node *node = root->rnode;
if (!radix_tree_is_indirect_ptr(node))
if (!radix_tree_is_internal_node(node))
return;
verify_node(node, tag, !!root_tag_get(root, tag));
}
@ -222,7 +222,7 @@ void tree_verify_min_height(struct radix_tree_root *root, int maxindex)
{
unsigned shift;
struct radix_tree_node *node = root->rnode;
if (!radix_tree_is_indirect_ptr(node)) {
if (!radix_tree_is_internal_node(node)) {
assert(maxindex == 0);
return;
}