btrfs-progs: fix all variable shadowing

There are quite some variable shadowing in btrfs-progs, most of them are
just reusing some common names like tmp.
And those are quite safe and the shadowed one are even different type.

But there are some exceptions:

- @end in traverse_tree_blocks()
  There is already an @end with the same type, but a different meaning
  (the end of the current extent buffer passed in).
  Just rename it to @child_end.

- @start in generate_new_data_csums_range()
  Just rename it to @csum_start.

- @size of fixup_chunk_tree_block()
  This one is particularly bad, we declare a local @size and initialize
  it to -1, then before we really utilize the variable @size, we
  immediately reset it to 0, then pass it to logical_to_physical().
  Then there is a location to check if @size is -1, which will always be
  true.

  According to the code in logical_to_physical(), @size would be clamped
  down by its original value, thus our local @size will always be 0.

  This patch would rename the local @size to @found_size, and only set
  it to -1.
  The call site is only to pass something as logical_to_physical()
  requires a non-NULL pointer.
  We don't really need to bother the returned value.

- duplicated @ref declaration in run_delayed_tree_ref()
- duplicated @super_flags in change_meta_csums()
  Just delete the duplicated one.

Signed-off-by: Qu Wenruo <wqu@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
This commit is contained in:
Qu Wenruo 2023-09-11 20:10:33 +09:30 committed by David Sterba
parent c788977878
commit 930c6362d1
9 changed files with 36 additions and 39 deletions

View File

@ -7597,9 +7597,9 @@ static int find_possible_backrefs(struct btrfs_path *path,
cache = lookup_cache_extent(extent_cache, bytenr, 1); cache = lookup_cache_extent(extent_cache, bytenr, 1);
if (cache) { if (cache) {
struct extent_record *tmp; struct extent_record *extent;
tmp = container_of(cache, struct extent_record, cache); extent = container_of(cache, struct extent_record, cache);
/* /*
* If we found an extent record for the bytenr for this * If we found an extent record for the bytenr for this
@ -7609,7 +7609,7 @@ static int find_possible_backrefs(struct btrfs_path *path,
* extent tree since they likely belong to this record * extent tree since they likely belong to this record
* and we need to fix it if it doesn't match bytenrs. * and we need to fix it if it doesn't match bytenrs.
*/ */
if (tmp->found_rec) if (extent->found_rec)
continue; continue;
} }

View File

@ -5027,7 +5027,7 @@ static int walk_down_tree(struct btrfs_root *root, struct btrfs_path *path,
next = btrfs_find_tree_block(gfs_info, bytenr, gfs_info->nodesize); next = btrfs_find_tree_block(gfs_info, bytenr, gfs_info->nodesize);
if (!next || !btrfs_buffer_uptodate(next, ptr_gen, 0)) { if (!next || !btrfs_buffer_uptodate(next, ptr_gen, 0)) {
struct btrfs_tree_parent_check check = { struct btrfs_tree_parent_check tree_check = {
.owner_root = btrfs_header_owner(cur), .owner_root = btrfs_header_owner(cur),
.transid = ptr_gen, .transid = ptr_gen,
.level = *level - 1, .level = *level - 1,
@ -5035,7 +5035,7 @@ static int walk_down_tree(struct btrfs_root *root, struct btrfs_path *path,
free_extent_buffer(next); free_extent_buffer(next);
reada_walk_down(root, cur, path->slots[*level]); reada_walk_down(root, cur, path->slots[*level]);
next = read_tree_block(gfs_info, bytenr, &check); next = read_tree_block(gfs_info, bytenr, &tree_check);
if (!extent_buffer_uptodate(next)) { if (!extent_buffer_uptodate(next)) {
struct btrfs_key node_key; struct btrfs_key node_key;

View File

@ -408,7 +408,7 @@ static int account_one_extent(struct ulist *roots, u64 bytenr, u64 num_bytes)
int ret; int ret;
u64 id, nr_roots, nr_refs; u64 id, nr_roots, nr_refs;
struct qgroup_count *count; struct qgroup_count *count;
struct ulist *counts = ulist_alloc(0); struct ulist *local_counts = ulist_alloc(0);
struct ulist *tmp = ulist_alloc(0); struct ulist *tmp = ulist_alloc(0);
struct ulist_iterator uiter; struct ulist_iterator uiter;
struct ulist_iterator tmp_uiter; struct ulist_iterator tmp_uiter;
@ -416,8 +416,8 @@ static int account_one_extent(struct ulist *roots, u64 bytenr, u64 num_bytes)
struct ulist_node *tmp_unode; struct ulist_node *tmp_unode;
struct btrfs_qgroup_list *glist; struct btrfs_qgroup_list *glist;
if (!counts || !tmp) { if (!local_counts || !tmp) {
ulist_free(counts); ulist_free(local_counts);
ulist_free(tmp); ulist_free(tmp);
return ENOMEM; return ENOMEM;
} }
@ -435,7 +435,7 @@ static int account_one_extent(struct ulist *roots, u64 bytenr, u64 num_bytes)
continue; continue;
BUG_ON(!is_fstree(unode->val)); BUG_ON(!is_fstree(unode->val));
ret = ulist_add(counts, count->qgroupid, ptr_to_u64(count), 0); ret = ulist_add(local_counts, count->qgroupid, ptr_to_u64(count), 0);
if (ret < 0) if (ret < 0)
goto out; goto out;
@ -462,7 +462,7 @@ static int account_one_extent(struct ulist *roots, u64 bytenr, u64 num_bytes)
BUG_ON(!count); BUG_ON(!count);
ret = ulist_add(counts, id, ptr_to_u64(parent), ret = ulist_add(local_counts, id, ptr_to_u64(parent),
0); 0);
if (ret < 0) if (ret < 0)
goto out; goto out;
@ -480,7 +480,7 @@ static int account_one_extent(struct ulist *roots, u64 bytenr, u64 num_bytes)
*/ */
nr_roots = roots->nnodes; nr_roots = roots->nnodes;
ULIST_ITER_INIT(&uiter); ULIST_ITER_INIT(&uiter);
while ((unode = ulist_next(counts, &uiter))) { while ((unode = ulist_next(local_counts, &uiter))) {
count = u64_to_ptr(unode->aux); count = u64_to_ptr(unode->aux);
nr_refs = group_get_cur_refcnt(count); nr_refs = group_get_cur_refcnt(count);
@ -506,7 +506,7 @@ static int account_one_extent(struct ulist *roots, u64 bytenr, u64 num_bytes)
inc_qgroup_seq(roots->nnodes); inc_qgroup_seq(roots->nnodes);
ret = 0; ret = 0;
out: out:
ulist_free(counts); ulist_free(local_counts);
ulist_free(tmp); ulist_free(tmp);
return ret; return ret;
} }
@ -922,7 +922,7 @@ static int add_qgroup_relation(u64 memberid, u64 parentid)
} }
static void read_qgroup_status(struct btrfs_fs_info *info, struct extent_buffer *eb, static void read_qgroup_status(struct btrfs_fs_info *info, struct extent_buffer *eb,
int slot, struct counts_tree *counts) int slot, struct counts_tree *ct)
{ {
struct btrfs_qgroup_status_item *status_item; struct btrfs_qgroup_status_item *status_item;
u64 flags; u64 flags;
@ -930,16 +930,15 @@ static void read_qgroup_status(struct btrfs_fs_info *info, struct extent_buffer
status_item = btrfs_item_ptr(eb, slot, struct btrfs_qgroup_status_item); status_item = btrfs_item_ptr(eb, slot, struct btrfs_qgroup_status_item);
flags = btrfs_qgroup_status_flags(eb, status_item); flags = btrfs_qgroup_status_flags(eb, status_item);
if (counts->simple == 1) if (ct->simple == 1)
counts->enable_gen = btrfs_qgroup_status_enable_gen(eb, status_item); ct->enable_gen = btrfs_qgroup_status_enable_gen(eb, status_item);
/* /*
* Since qgroup_inconsist/rescan_running is just one bit, * Since qgroup_inconsist/rescan_running is just one bit,
* assign value directly won't work. * assign value directly won't work.
*/ */
counts->qgroup_inconsist = !!(flags & ct->qgroup_inconsist = !!(flags & BTRFS_QGROUP_STATUS_FLAG_INCONSISTENT);
BTRFS_QGROUP_STATUS_FLAG_INCONSISTENT); ct->rescan_running = !!(flags & BTRFS_QGROUP_STATUS_FLAG_RESCAN);
counts->rescan_running = !!(flags & BTRFS_QGROUP_STATUS_FLAG_RESCAN); ct->scan_progress = btrfs_qgroup_status_rescan(eb, status_item);
counts->scan_progress = btrfs_qgroup_status_rescan(eb, status_item);
} }
static int load_quota_info(struct btrfs_fs_info *info) static int load_quota_info(struct btrfs_fs_info *info)

View File

@ -169,10 +169,10 @@ static int traverse_tree_blocks(struct extent_io_tree *tree,
if (ret) if (ret)
return ret; return ret;
} else { } else {
u64 end; u64 child_end;
bytenr = btrfs_node_blockptr(eb, i); bytenr = btrfs_node_blockptr(eb, i);
end = bytenr + fs_info->nodesize - 1; child_end = bytenr + fs_info->nodesize - 1;
/* If we aren't the tree root don't read the block */ /* If we aren't the tree root don't read the block */
if (level == 1 && !tree_root) { if (level == 1 && !tree_root) {
@ -180,7 +180,8 @@ static int traverse_tree_blocks(struct extent_io_tree *tree,
btrfs_pin_extent(fs_info, bytenr, btrfs_pin_extent(fs_info, bytenr,
fs_info->nodesize); fs_info->nodesize);
else else
set_extent_dirty(tree, bytenr, end, GFP_NOFS); set_extent_dirty(tree, bytenr,
child_end, GFP_NOFS);
continue; continue;
} }

View File

@ -537,11 +537,11 @@ static int print_filesystem_usage_overall(int fd, struct chunk_info *chunkinfo,
* As mixed mode is not supported in zoned mode, this * As mixed mode is not supported in zoned mode, this
* will account for all profile types * will account for all profile types
*/ */
u64 tmp; u64 unusable;
tmp = device_get_zone_unusable(fd, flags); unusable = device_get_zone_unusable(fd, flags);
if (tmp != DEVICE_ZONE_UNUSABLE_UNKNOWN) if (unusable != DEVICE_ZONE_UNUSABLE_UNKNOWN)
zone_unusable += tmp; zone_unusable += unusable;
} }
if (flags & BTRFS_BLOCK_GROUP_DATA) { if (flags & BTRFS_BLOCK_GROUP_DATA) {
r_data_used += sargs->spaces[i].used_bytes * ratio; r_data_used += sargs->spaces[i].used_bytes * ratio;

View File

@ -1411,7 +1411,7 @@ static int btrfs_list_subvols(int fd, struct rb_root *root_lookup)
n = rb_first(root_lookup); n = rb_first(root_lookup);
while (n) { while (n) {
struct root_info *entry; struct root_info *entry;
int ret;
entry = to_root_info(n); entry = to_root_info(n);
ret = lookup_ino_path(fd, entry); ret = lookup_ino_path(fd, entry);
if (ret && ret != -ENOENT) if (ret && ret != -ENOENT)

View File

@ -227,15 +227,15 @@ static int fixup_chunk_tree_block(struct mdrestore_struct *mdres,
for (i = 0; i < btrfs_header_nritems(eb); i++) { for (i = 0; i < btrfs_header_nritems(eb); i++) {
struct btrfs_chunk *chunk; struct btrfs_chunk *chunk;
struct btrfs_key key; struct btrfs_key key;
u64 type, physical, physical_dup, size = (u64)-1; u64 type, physical, physical_dup;
u64 found_size = (u64)-1;
btrfs_item_key_to_cpu(eb, &key, i); btrfs_item_key_to_cpu(eb, &key, i);
if (key.type != BTRFS_CHUNK_ITEM_KEY) if (key.type != BTRFS_CHUNK_ITEM_KEY)
continue; continue;
size = 0;
physical = logical_to_physical(mdres, key.offset, physical = logical_to_physical(mdres, key.offset,
&size, &physical_dup); &found_size, &physical_dup);
if (!physical_dup) if (!physical_dup)
truncate_item(eb, i, sizeof(*chunk)); truncate_item(eb, i, sizeof(*chunk));
@ -254,9 +254,7 @@ static int fixup_chunk_tree_block(struct mdrestore_struct *mdres,
btrfs_set_chunk_num_stripes(eb, chunk, 1); btrfs_set_chunk_num_stripes(eb, chunk, 1);
btrfs_set_chunk_sub_stripes(eb, chunk, 0); btrfs_set_chunk_sub_stripes(eb, chunk, 0);
btrfs_set_stripe_devid_nr(eb, chunk, 0, mdres->devid); btrfs_set_stripe_devid_nr(eb, chunk, 0, mdres->devid);
if (size != (u64)-1) btrfs_set_stripe_offset_nr(eb, chunk, 0, physical);
btrfs_set_stripe_offset_nr(eb, chunk, 0,
physical);
/* update stripe 2 offset */ /* update stripe 2 offset */
if (physical_dup) if (physical_dup)
btrfs_set_stripe_offset_nr(eb, chunk, 1, btrfs_set_stripe_offset_nr(eb, chunk, 1,

View File

@ -3768,7 +3768,6 @@ static int run_delayed_tree_ref(struct btrfs_trans_handle *trans,
BUG_ON(!extent_op || !extent_op->update_flags); BUG_ON(!extent_op || !extent_op->update_flags);
ret = alloc_reserved_tree_block(trans, node, extent_op); ret = alloc_reserved_tree_block(trans, node, extent_op);
} else if (node->action == BTRFS_DROP_DELAYED_REF) { } else if (node->action == BTRFS_DROP_DELAYED_REF) {
struct btrfs_delayed_tree_ref *ref = btrfs_delayed_node_to_tree_ref(node);
ret = __free_extent(trans, node->bytenr, node->num_bytes, ret = __free_extent(trans, node->bytenr, node->num_bytes,
ref->parent, ref->root, ref->level, 0, 1); ref->parent, ref->root, ref->level, 0, 1);
} else { } else {

View File

@ -248,7 +248,7 @@ static int generate_new_data_csums_range(struct btrfs_fs_info *fs_info, u64 star
} }
while (cur < last_csum) { while (cur < last_csum) {
u64 start; u64 csum_start;
u64 len; u64 len;
u32 item_size; u32 item_size;
@ -276,14 +276,14 @@ static int generate_new_data_csums_range(struct btrfs_fs_info *fs_info, u64 star
assert(key.offset >= cur); assert(key.offset >= cur);
item_size = btrfs_item_size(path.nodes[0], path.slots[0]); item_size = btrfs_item_size(path.nodes[0], path.slots[0]);
start = key.offset; csum_start = key.offset;
len = item_size / fs_info->csum_size * fs_info->sectorsize; len = item_size / fs_info->csum_size * fs_info->sectorsize;
read_extent_buffer(path.nodes[0], csum_buffer, read_extent_buffer(path.nodes[0], csum_buffer,
btrfs_item_ptr_offset(path.nodes[0], path.slots[0]), btrfs_item_ptr_offset(path.nodes[0], path.slots[0]),
item_size); item_size);
btrfs_release_path(&path); btrfs_release_path(&path);
ret = generate_new_csum_range(trans, start, len, new_csum_type, ret = generate_new_csum_range(trans, csum_start, len, new_csum_type,
csum_buffer); csum_buffer);
if (ret < 0) if (ret < 0)
goto out; goto out;
@ -303,7 +303,7 @@ static int generate_new_data_csums_range(struct btrfs_fs_info *fs_info, u64 star
goto out; goto out;
} }
} }
cur = start + len; cur = csum_start + len;
} }
ret = btrfs_commit_transaction(trans, csum_root); ret = btrfs_commit_transaction(trans, csum_root);
if (inject_error(0x4de02239)) if (inject_error(0x4de02239))
@ -628,7 +628,7 @@ out:
struct btrfs_root *tree_root = fs_info->tree_root; struct btrfs_root *tree_root = fs_info->tree_root;
struct btrfs_trans_handle *trans; struct btrfs_trans_handle *trans;
u64 super_flags = btrfs_super_flags(fs_info->super_copy); super_flags = btrfs_super_flags(fs_info->super_copy);
btrfs_set_super_csum_type(fs_info->super_copy, new_csum_type); btrfs_set_super_csum_type(fs_info->super_copy, new_csum_type);
super_flags &= ~(BTRFS_SUPER_FLAG_CHANGING_DATA_CSUM | super_flags &= ~(BTRFS_SUPER_FLAG_CHANGING_DATA_CSUM |