Lines Matching refs:slab_count
2529 WRITE_ONCE(scrubber->slab_count, scrubber->slab_count + 1); in register_slab_for_scrubbing()
2738 WRITE_ONCE(scrubber->slab_count, scrubber->slab_count - 1); in slab_scrubbed()
3002 allocator->slab_count++; in register_slab_with_allocator()
3478 result = vdo_allocate(allocator->slab_count, struct slab_status, __func__, in get_slab_statuses()
3508 allocator->slab_count * depot->slab_config.data_blocks); in vdo_prepare_slabs_for_allocation()
3516 .nr = allocator->slab_count, in vdo_prepare_slabs_for_allocation()
3517 .size = allocator->slab_count, in vdo_prepare_slabs_for_allocation()
3629 READ_ONCE(scrubber->slab_count), in vdo_dump_block_allocator()
3765 static int allocate_slabs(struct slab_depot *depot, slab_count_t slab_count) in allocate_slabs() argument
3772 result = vdo_allocate(slab_count, struct vdo_slab *, in allocate_slabs()
3779 depot->slab_count * sizeof(struct vdo_slab *)); in allocate_slabs()
3784 slab_origin = depot->first_block + (depot->slab_count * slab_size); in allocate_slabs()
3786 for (depot->new_slab_count = depot->slab_count; in allocate_slabs()
3787 depot->new_slab_count < slab_count; in allocate_slabs()
3813 for (i = depot->slab_count; i < depot->new_slab_count; i++) in vdo_abandon_new_slabs()
4058 slab_count_t slab_count; in allocate_components() local
4099 slab_count = vdo_compute_slab_count(depot->first_block, depot->last_block, in allocate_components()
4101 if (thread_config->physical_zone_count > slab_count) { in allocate_components()
4105 slab_count); in allocate_components()
4116 result = allocate_slabs(depot, slab_count); in allocate_components()
4121 for (i = depot->slab_count; i < depot->new_slab_count; i++) { in allocate_components()
4125 WRITE_ONCE(depot->slab_count, depot->slab_count + 1); in allocate_components()
4232 for (i = 0; i < depot->slab_count; i++) in vdo_free_slab_depot()
4281 get_depot_slab_iterator(depot, depot->slab_count - 1, 0, 1); in vdo_allocate_reference_counters()
4311 if (slab_number >= depot->slab_count) in get_slab_number()
4428 return (READ_ONCE(depot->slab_count) * depot->slab_config.data_blocks); in vdo_get_slab_depot_data_blocks()
4654 if ((partition->count >> depot->slab_size_shift) <= depot->slab_count) in vdo_prepare_to_grow_slab_depot()
4668 if (new_slab_count <= depot->slab_count) in vdo_prepare_to_grow_slab_depot()
4700 WRITE_ONCE(depot->slab_count, depot->new_slab_count); in finish_registration()
4716 for (i = depot->slab_count; i < depot->new_slab_count; i++) { in register_new_slabs()
4993 totals.slab_count += allocator->slab_count; in get_block_allocator_statistics()
5060 slab_count_t slab_count = READ_ONCE(depot->slab_count); in vdo_get_slab_depot_statistics() local
5066 unrecovered += READ_ONCE(depot->allocators[zone].scrubber.slab_count); in vdo_get_slab_depot_statistics()
5069 stats->recovery_percentage = (slab_count - unrecovered) * 100 / slab_count; in vdo_get_slab_depot_statistics()
5087 (unsigned int) depot->old_zone_count, READ_ONCE(depot->slab_count), in vdo_dump_slab_depot()