#undef SCRAMBLE_DELAYED_REFS
-/*
- * Declare a helper function to detect underflow of various space info members
- */
-#define DECLARE_SPACE_INFO_UPDATE(name) \
-static inline void update_##name(struct btrfs_fs_info *fs_info, \
- struct btrfs_space_info *sinfo, \
- s64 bytes) \
-{ \
- lockdep_assert_held(&sinfo->lock); \
- trace_update_##name(fs_info, sinfo, sinfo->name, bytes); \
- if (bytes < 0 && sinfo->name < -bytes) { \
- WARN_ON(1); \
- sinfo->name = 0; \
- return; \
- } \
- sinfo->name += bytes; \
-}
-
-DECLARE_SPACE_INFO_UPDATE(bytes_may_use);
-DECLARE_SPACE_INFO_UPDATE(bytes_pinned);
static int __btrfs_free_extent(struct btrfs_trans_handle *trans,
struct btrfs_delayed_ref_node *node, u64 parent,
data_sinfo->flags, bytes, 1);
return -ENOSPC;
}
- update_bytes_may_use(fs_info, data_sinfo, bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info, data_sinfo, bytes);
trace_btrfs_space_reservation(fs_info, "space_info",
data_sinfo->flags, bytes, 1);
spin_unlock(&data_sinfo->lock);
data_sinfo = fs_info->data_sinfo;
spin_lock(&data_sinfo->lock);
- update_bytes_may_use(fs_info, data_sinfo, -len);
+ btrfs_space_info_update_bytes_may_use(fs_info, data_sinfo, -len);
trace_btrfs_space_reservation(fs_info, "space_info",
data_sinfo->flags, len, 0);
spin_unlock(&data_sinfo->lock);
* If not things get more complicated.
*/
if (used + orig_bytes <= space_info->total_bytes) {
- update_bytes_may_use(fs_info, space_info, orig_bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info, space_info,
+ orig_bytes);
trace_btrfs_space_reservation(fs_info, "space_info",
space_info->flags, orig_bytes, 1);
ret = 0;
} else if (btrfs_can_overcommit(fs_info, space_info, orig_bytes, flush,
system_chunk)) {
- update_bytes_may_use(fs_info, space_info, orig_bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info, space_info,
+ orig_bytes);
trace_btrfs_space_reservation(fs_info, "space_info",
space_info->flags, orig_bytes, 1);
ret = 0;
flush = BTRFS_RESERVE_FLUSH_ALL;
goto again;
}
- update_bytes_may_use(fs_info, space_info, -num_bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info, space_info, -num_bytes);
trace_btrfs_space_reservation(fs_info, "space_info",
space_info->flags, num_bytes, 0);
spin_unlock(&space_info->lock);
ticket->bytes, 1);
list_del_init(&ticket->list);
num_bytes -= ticket->bytes;
- update_bytes_may_use(fs_info, space_info,
- ticket->bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info,
+ space_info, ticket->bytes);
ticket->bytes = 0;
space_info->tickets_id++;
wake_up(&ticket->wait);
trace_btrfs_space_reservation(fs_info, "space_info",
space_info->flags,
num_bytes, 1);
- update_bytes_may_use(fs_info, space_info, num_bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info,
+ space_info, num_bytes);
ticket->bytes -= num_bytes;
num_bytes = 0;
}
num_bytes = min(num_bytes,
block_rsv->size - block_rsv->reserved);
block_rsv->reserved += num_bytes;
- update_bytes_may_use(fs_info, sinfo, num_bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info, sinfo,
+ num_bytes);
trace_btrfs_space_reservation(fs_info, "space_info",
sinfo->flags, num_bytes,
1);
}
} else if (block_rsv->reserved > block_rsv->size) {
num_bytes = block_rsv->reserved - block_rsv->size;
- update_bytes_may_use(fs_info, sinfo, -num_bytes);
+ btrfs_space_info_update_bytes_may_use(fs_info, sinfo,
+ -num_bytes);
trace_btrfs_space_reservation(fs_info, "space_info",
sinfo->flags, num_bytes, 0);
block_rsv->reserved = block_rsv->size;
old_val -= num_bytes;
btrfs_set_block_group_used(&cache->item, old_val);
cache->pinned += num_bytes;
- update_bytes_pinned(info, cache->space_info, num_bytes);
+ btrfs_space_info_update_bytes_pinned(info,
+ cache->space_info, num_bytes);
cache->space_info->bytes_used -= num_bytes;
cache->space_info->disk_used -= num_bytes * factor;
spin_unlock(&cache->lock);
spin_lock(&cache->space_info->lock);
spin_lock(&cache->lock);
cache->pinned += num_bytes;
- update_bytes_pinned(fs_info, cache->space_info, num_bytes);
+ btrfs_space_info_update_bytes_pinned(fs_info, cache->space_info,
+ num_bytes);
if (reserved) {
cache->reserved -= num_bytes;
cache->space_info->bytes_reserved -= num_bytes;
} else {
cache->reserved += num_bytes;
space_info->bytes_reserved += num_bytes;
- update_bytes_may_use(cache->fs_info, space_info, -ram_bytes);
+ btrfs_space_info_update_bytes_may_use(cache->fs_info,
+ space_info, -ram_bytes);
if (delalloc)
cache->delalloc_bytes += num_bytes;
}
spin_lock(&space_info->lock);
spin_lock(&cache->lock);
cache->pinned -= len;
- update_bytes_pinned(fs_info, space_info, -len);
+ btrfs_space_info_update_bytes_pinned(fs_info, space_info, -len);
trace_btrfs_space_reservation(fs_info, "pinned",
space_info->flags, len, 0);
to_add = min(len, global_rsv->size -
global_rsv->reserved);
global_rsv->reserved += to_add;
- update_bytes_may_use(fs_info, space_info,
- to_add);
+ btrfs_space_info_update_bytes_may_use(fs_info,
+ space_info, to_add);
if (global_rsv->reserved >= global_rsv->size)
global_rsv->full = 1;
trace_btrfs_space_reservation(fs_info,
spin_lock(&space_info->lock);
spin_lock(&block_group->lock);
- update_bytes_pinned(fs_info, space_info, -block_group->pinned);
+ btrfs_space_info_update_bytes_pinned(fs_info, space_info,
+ -block_group->pinned);
space_info->bytes_readonly += block_group->pinned;
percpu_counter_add_batch(&space_info->total_bytes_pinned,
-block_group->pinned,