mirror of
https://github.com/torvalds/linux.git
synced 2025-11-30 23:16:01 +07:00
Merge tag 'for-6.18-rc5-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux
Pull btrfs fixes from David Sterba: - fix new inode name tracking in tree-log - fix conventional zone and stripe calculations in zoned mode - fix bio reference counts on error paths in relocation and scrub * tag 'for-6.18-rc5-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux: btrfs: release root after error in data_reloc_print_warning_inode() btrfs: scrub: put bio after errors in scrub_raid56_parity_stripe() btrfs: do not update last_log_commit when logging inode due to a new name btrfs: zoned: fix stripe width calculation btrfs: zoned: fix conventional zone capacity calculation
This commit is contained in:
@@ -177,8 +177,10 @@ static int data_reloc_print_warning_inode(u64 inum, u64 offset, u64 num_bytes,
|
||||
return ret;
|
||||
}
|
||||
ret = paths_from_inode(inum, ipath);
|
||||
if (ret < 0)
|
||||
if (ret < 0) {
|
||||
btrfs_put_root(local_root);
|
||||
goto err;
|
||||
}
|
||||
|
||||
/*
|
||||
* We deliberately ignore the bit ipath might have been too small to
|
||||
|
||||
@@ -2203,6 +2203,7 @@ static int scrub_raid56_parity_stripe(struct scrub_ctx *sctx,
|
||||
ret = btrfs_map_block(fs_info, BTRFS_MAP_WRITE, full_stripe_start,
|
||||
&length, &bioc, NULL, NULL);
|
||||
if (ret < 0) {
|
||||
bio_put(bio);
|
||||
btrfs_put_bioc(bioc);
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
goto out;
|
||||
@@ -2212,6 +2213,7 @@ static int scrub_raid56_parity_stripe(struct scrub_ctx *sctx,
|
||||
btrfs_put_bioc(bioc);
|
||||
if (!rbio) {
|
||||
ret = -ENOMEM;
|
||||
bio_put(bio);
|
||||
btrfs_bio_counter_dec(fs_info);
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -7122,7 +7122,7 @@ log_extents:
|
||||
* a power failure unless the log was synced as part of an fsync
|
||||
* against any other unrelated inode.
|
||||
*/
|
||||
if (inode_only != LOG_INODE_EXISTS)
|
||||
if (!ctx->logging_new_name && inode_only != LOG_INODE_EXISTS)
|
||||
inode->last_log_commit = inode->last_sub_trans;
|
||||
spin_unlock(&inode->lock);
|
||||
|
||||
|
||||
@@ -1317,6 +1317,7 @@ static int btrfs_load_zone_info(struct btrfs_fs_info *fs_info, int zone_idx,
|
||||
if (!btrfs_dev_is_sequential(device, info->physical)) {
|
||||
up_read(&dev_replace->rwsem);
|
||||
info->alloc_offset = WP_CONVENTIONAL;
|
||||
info->capacity = device->zone_info->zone_size;
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -1522,6 +1523,8 @@ static int btrfs_load_block_group_raid0(struct btrfs_block_group *bg,
|
||||
u64 last_alloc)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = bg->fs_info;
|
||||
u64 stripe_nr = 0, stripe_offset = 0;
|
||||
u32 stripe_index = 0;
|
||||
|
||||
if ((map->type & BTRFS_BLOCK_GROUP_DATA) && !fs_info->stripe_root) {
|
||||
btrfs_err(fs_info, "zoned: data %s needs raid-stripe-tree",
|
||||
@@ -1529,28 +1532,26 @@ static int btrfs_load_block_group_raid0(struct btrfs_block_group *bg,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (last_alloc) {
|
||||
u32 factor = map->num_stripes;
|
||||
|
||||
stripe_nr = last_alloc >> BTRFS_STRIPE_LEN_SHIFT;
|
||||
stripe_offset = last_alloc & BTRFS_STRIPE_LEN_MASK;
|
||||
stripe_nr = div_u64_rem(stripe_nr, factor, &stripe_index);
|
||||
}
|
||||
|
||||
for (int i = 0; i < map->num_stripes; i++) {
|
||||
if (zone_info[i].alloc_offset == WP_MISSING_DEV)
|
||||
continue;
|
||||
|
||||
if (zone_info[i].alloc_offset == WP_CONVENTIONAL) {
|
||||
u64 stripe_nr, full_stripe_nr;
|
||||
u64 stripe_offset;
|
||||
int stripe_index;
|
||||
|
||||
stripe_nr = div64_u64(last_alloc, map->stripe_size);
|
||||
stripe_offset = stripe_nr * map->stripe_size;
|
||||
full_stripe_nr = div_u64(stripe_nr, map->num_stripes);
|
||||
div_u64_rem(stripe_nr, map->num_stripes, &stripe_index);
|
||||
|
||||
zone_info[i].alloc_offset =
|
||||
full_stripe_nr * map->stripe_size;
|
||||
zone_info[i].alloc_offset = btrfs_stripe_nr_to_offset(stripe_nr);
|
||||
|
||||
if (stripe_index > i)
|
||||
zone_info[i].alloc_offset += map->stripe_size;
|
||||
zone_info[i].alloc_offset += BTRFS_STRIPE_LEN;
|
||||
else if (stripe_index == i)
|
||||
zone_info[i].alloc_offset +=
|
||||
(last_alloc - stripe_offset);
|
||||
zone_info[i].alloc_offset += stripe_offset;
|
||||
}
|
||||
|
||||
if (test_bit(0, active) != test_bit(i, active)) {
|
||||
@@ -1574,6 +1575,8 @@ static int btrfs_load_block_group_raid10(struct btrfs_block_group *bg,
|
||||
u64 last_alloc)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = bg->fs_info;
|
||||
u64 stripe_nr = 0, stripe_offset = 0;
|
||||
u32 stripe_index = 0;
|
||||
|
||||
if ((map->type & BTRFS_BLOCK_GROUP_DATA) && !fs_info->stripe_root) {
|
||||
btrfs_err(fs_info, "zoned: data %s needs raid-stripe-tree",
|
||||
@@ -1581,6 +1584,14 @@ static int btrfs_load_block_group_raid10(struct btrfs_block_group *bg,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (last_alloc) {
|
||||
u32 factor = map->num_stripes / map->sub_stripes;
|
||||
|
||||
stripe_nr = last_alloc >> BTRFS_STRIPE_LEN_SHIFT;
|
||||
stripe_offset = last_alloc & BTRFS_STRIPE_LEN_MASK;
|
||||
stripe_nr = div_u64_rem(stripe_nr, factor, &stripe_index);
|
||||
}
|
||||
|
||||
for (int i = 0; i < map->num_stripes; i++) {
|
||||
if (zone_info[i].alloc_offset == WP_MISSING_DEV)
|
||||
continue;
|
||||
@@ -1594,26 +1605,12 @@ static int btrfs_load_block_group_raid10(struct btrfs_block_group *bg,
|
||||
}
|
||||
|
||||
if (zone_info[i].alloc_offset == WP_CONVENTIONAL) {
|
||||
u64 stripe_nr, full_stripe_nr;
|
||||
u64 stripe_offset;
|
||||
int stripe_index;
|
||||
|
||||
stripe_nr = div64_u64(last_alloc, map->stripe_size);
|
||||
stripe_offset = stripe_nr * map->stripe_size;
|
||||
full_stripe_nr = div_u64(stripe_nr,
|
||||
map->num_stripes / map->sub_stripes);
|
||||
div_u64_rem(stripe_nr,
|
||||
(map->num_stripes / map->sub_stripes),
|
||||
&stripe_index);
|
||||
|
||||
zone_info[i].alloc_offset =
|
||||
full_stripe_nr * map->stripe_size;
|
||||
zone_info[i].alloc_offset = btrfs_stripe_nr_to_offset(stripe_nr);
|
||||
|
||||
if (stripe_index > (i / map->sub_stripes))
|
||||
zone_info[i].alloc_offset += map->stripe_size;
|
||||
zone_info[i].alloc_offset += BTRFS_STRIPE_LEN;
|
||||
else if (stripe_index == (i / map->sub_stripes))
|
||||
zone_info[i].alloc_offset +=
|
||||
(last_alloc - stripe_offset);
|
||||
zone_info[i].alloc_offset += stripe_offset;
|
||||
}
|
||||
|
||||
if ((i % map->sub_stripes) == 0) {
|
||||
@@ -1683,8 +1680,6 @@ int btrfs_load_block_group_zone_info(struct btrfs_block_group *cache, bool new)
|
||||
set_bit(BLOCK_GROUP_FLAG_SEQUENTIAL_ZONE, &cache->runtime_flags);
|
||||
|
||||
if (num_conventional > 0) {
|
||||
/* Zone capacity is always zone size in emulation */
|
||||
cache->zone_capacity = cache->length;
|
||||
ret = calculate_alloc_pointer(cache, &last_alloc, new);
|
||||
if (ret) {
|
||||
btrfs_err(fs_info,
|
||||
@@ -1693,6 +1688,7 @@ int btrfs_load_block_group_zone_info(struct btrfs_block_group *cache, bool new)
|
||||
goto out;
|
||||
} else if (map->num_stripes == num_conventional) {
|
||||
cache->alloc_offset = last_alloc;
|
||||
cache->zone_capacity = cache->length;
|
||||
set_bit(BLOCK_GROUP_FLAG_ZONE_IS_ACTIVE, &cache->runtime_flags);
|
||||
goto out;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user