dm zoned: move random and sequential zones into struct dmz_dev

Random and sequential zones should be part of the respective
device structure to make arbitration between devices possible.

Signed-off-by: Hannes Reinecke <hare@suse.de>
Reviewed-by: Damien Le Moal <damien.lemoal@wdc.com>
Signed-off-by: Mike Snitzer <snitzer@redhat.com>
This commit is contained in:
Hannes Reinecke 2020-06-02 13:09:51 +02:00 committed by Mike Snitzer
parent f97809aec5
commit bd82fdabf1
4 changed files with 119 additions and 78 deletions

View File

@ -192,21 +192,12 @@ struct dmz_metadata {
/* Zone allocation management */ /* Zone allocation management */
struct mutex map_lock; struct mutex map_lock;
struct dmz_mblock **map_mblk; struct dmz_mblock **map_mblk;
unsigned int nr_rnd;
atomic_t unmap_nr_rnd;
struct list_head unmap_rnd_list;
struct list_head map_rnd_list;
unsigned int nr_cache; unsigned int nr_cache;
atomic_t unmap_nr_cache; atomic_t unmap_nr_cache;
struct list_head unmap_cache_list; struct list_head unmap_cache_list;
struct list_head map_cache_list; struct list_head map_cache_list;
unsigned int nr_seq;
atomic_t unmap_nr_seq;
struct list_head unmap_seq_list;
struct list_head map_seq_list;
atomic_t nr_reserved_seq_zones; atomic_t nr_reserved_seq_zones;
struct list_head reserved_seq_zones_list; struct list_head reserved_seq_zones_list;
@ -279,14 +270,14 @@ unsigned int dmz_nr_chunks(struct dmz_metadata *zmd)
return zmd->nr_chunks; return zmd->nr_chunks;
} }
unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd) unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd, int idx)
{ {
return zmd->nr_rnd; return zmd->dev[idx].nr_rnd;
} }
unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd) unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd, int idx)
{ {
return atomic_read(&zmd->unmap_nr_rnd); return atomic_read(&zmd->dev[idx].unmap_nr_rnd);
} }
unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd) unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd)
@ -299,14 +290,14 @@ unsigned int dmz_nr_unmap_cache_zones(struct dmz_metadata *zmd)
return atomic_read(&zmd->unmap_nr_cache); return atomic_read(&zmd->unmap_nr_cache);
} }
unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd) unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd, int idx)
{ {
return zmd->nr_seq; return zmd->dev[idx].nr_seq;
} }
unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd) unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd, int idx)
{ {
return atomic_read(&zmd->unmap_nr_seq); return atomic_read(&zmd->dev[idx].unmap_nr_seq);
} }
static struct dm_zone *dmz_get(struct dmz_metadata *zmd, unsigned int zone_id) static struct dm_zone *dmz_get(struct dmz_metadata *zmd, unsigned int zone_id)
@ -1500,6 +1491,14 @@ static int dmz_init_zones(struct dmz_metadata *zmd)
dev->metadata = zmd; dev->metadata = zmd;
zmd->nr_zones += dev->nr_zones; zmd->nr_zones += dev->nr_zones;
atomic_set(&dev->unmap_nr_rnd, 0);
INIT_LIST_HEAD(&dev->unmap_rnd_list);
INIT_LIST_HEAD(&dev->map_rnd_list);
atomic_set(&dev->unmap_nr_seq, 0);
INIT_LIST_HEAD(&dev->unmap_seq_list);
INIT_LIST_HEAD(&dev->map_seq_list);
} }
if (!zmd->nr_zones) { if (!zmd->nr_zones) {
@ -1720,9 +1719,9 @@ static int dmz_load_mapping(struct dmz_metadata *zmd)
if (dmz_is_cache(dzone)) if (dmz_is_cache(dzone))
list_add_tail(&dzone->link, &zmd->map_cache_list); list_add_tail(&dzone->link, &zmd->map_cache_list);
else if (dmz_is_rnd(dzone)) else if (dmz_is_rnd(dzone))
list_add_tail(&dzone->link, &zmd->map_rnd_list); list_add_tail(&dzone->link, &dzone->dev->map_rnd_list);
else else
list_add_tail(&dzone->link, &zmd->map_seq_list); list_add_tail(&dzone->link, &dzone->dev->map_seq_list);
/* Check buffer zone */ /* Check buffer zone */
bzone_id = le32_to_cpu(dmap[e].bzone_id); bzone_id = le32_to_cpu(dmap[e].bzone_id);
@ -1756,7 +1755,7 @@ static int dmz_load_mapping(struct dmz_metadata *zmd)
if (dmz_is_cache(bzone)) if (dmz_is_cache(bzone))
list_add_tail(&bzone->link, &zmd->map_cache_list); list_add_tail(&bzone->link, &zmd->map_cache_list);
else else
list_add_tail(&bzone->link, &zmd->map_rnd_list); list_add_tail(&bzone->link, &bzone->dev->map_rnd_list);
next: next:
chunk++; chunk++;
e++; e++;
@ -1781,9 +1780,9 @@ next:
if (dmz_is_cache(dzone)) if (dmz_is_cache(dzone))
zmd->nr_cache++; zmd->nr_cache++;
else if (dmz_is_rnd(dzone)) else if (dmz_is_rnd(dzone))
zmd->nr_rnd++; dzone->dev->nr_rnd++;
else else
zmd->nr_seq++; dzone->dev->nr_seq++;
if (dmz_is_data(dzone)) { if (dmz_is_data(dzone)) {
/* Already initialized */ /* Already initialized */
@ -1797,16 +1796,18 @@ next:
list_add_tail(&dzone->link, &zmd->unmap_cache_list); list_add_tail(&dzone->link, &zmd->unmap_cache_list);
atomic_inc(&zmd->unmap_nr_cache); atomic_inc(&zmd->unmap_nr_cache);
} else if (dmz_is_rnd(dzone)) { } else if (dmz_is_rnd(dzone)) {
list_add_tail(&dzone->link, &zmd->unmap_rnd_list); list_add_tail(&dzone->link,
atomic_inc(&zmd->unmap_nr_rnd); &dzone->dev->unmap_rnd_list);
atomic_inc(&dzone->dev->unmap_nr_rnd);
} else if (atomic_read(&zmd->nr_reserved_seq_zones) < zmd->nr_reserved_seq) { } else if (atomic_read(&zmd->nr_reserved_seq_zones) < zmd->nr_reserved_seq) {
list_add_tail(&dzone->link, &zmd->reserved_seq_zones_list); list_add_tail(&dzone->link, &zmd->reserved_seq_zones_list);
set_bit(DMZ_RESERVED, &dzone->flags); set_bit(DMZ_RESERVED, &dzone->flags);
atomic_inc(&zmd->nr_reserved_seq_zones); atomic_inc(&zmd->nr_reserved_seq_zones);
zmd->nr_seq--; dzone->dev->nr_seq--;
} else { } else {
list_add_tail(&dzone->link, &zmd->unmap_seq_list); list_add_tail(&dzone->link,
atomic_inc(&zmd->unmap_nr_seq); &dzone->dev->unmap_seq_list);
atomic_inc(&dzone->dev->unmap_nr_seq);
} }
} }
@ -1840,13 +1841,13 @@ static void __dmz_lru_zone(struct dmz_metadata *zmd, struct dm_zone *zone)
list_del_init(&zone->link); list_del_init(&zone->link);
if (dmz_is_seq(zone)) { if (dmz_is_seq(zone)) {
/* LRU rotate sequential zone */ /* LRU rotate sequential zone */
list_add_tail(&zone->link, &zmd->map_seq_list); list_add_tail(&zone->link, &zone->dev->map_seq_list);
} else if (dmz_is_cache(zone)) { } else if (dmz_is_cache(zone)) {
/* LRU rotate cache zone */ /* LRU rotate cache zone */
list_add_tail(&zone->link, &zmd->map_cache_list); list_add_tail(&zone->link, &zmd->map_cache_list);
} else { } else {
/* LRU rotate random zone */ /* LRU rotate random zone */
list_add_tail(&zone->link, &zmd->map_rnd_list); list_add_tail(&zone->link, &zone->dev->map_rnd_list);
} }
} }
@ -1928,14 +1929,24 @@ static struct dm_zone *dmz_get_rnd_zone_for_reclaim(struct dmz_metadata *zmd,
{ {
struct dm_zone *dzone = NULL; struct dm_zone *dzone = NULL;
struct dm_zone *zone; struct dm_zone *zone;
struct list_head *zone_list = &zmd->map_rnd_list; struct list_head *zone_list;
/* If we have cache zones select from the cache zone list */ /* If we have cache zones select from the cache zone list */
if (zmd->nr_cache) { if (zmd->nr_cache) {
zone_list = &zmd->map_cache_list; zone_list = &zmd->map_cache_list;
/* Try to relaim random zones, too, when idle */ /* Try to relaim random zones, too, when idle */
if (idle && list_empty(zone_list)) if (idle && list_empty(zone_list)) {
zone_list = &zmd->map_rnd_list; int i;
for (i = 1; i < zmd->nr_devs; i++) {
zone_list = &zmd->dev[i].map_rnd_list;
if (!list_empty(zone_list))
break;
}
}
} else {
/* Otherwise the random zones are on the first disk */
zone_list = &zmd->dev[0].map_rnd_list;
} }
list_for_each_entry(zone, zone_list, link) { list_for_each_entry(zone, zone_list, link) {
@ -1956,12 +1967,17 @@ static struct dm_zone *dmz_get_rnd_zone_for_reclaim(struct dmz_metadata *zmd,
static struct dm_zone *dmz_get_seq_zone_for_reclaim(struct dmz_metadata *zmd) static struct dm_zone *dmz_get_seq_zone_for_reclaim(struct dmz_metadata *zmd)
{ {
struct dm_zone *zone; struct dm_zone *zone;
int i;
list_for_each_entry(zone, &zmd->map_seq_list, link) { for (i = 0; i < zmd->nr_devs; i++) {
if (!zone->bzone) struct dmz_dev *dev = &zmd->dev[i];
continue;
if (dmz_lock_zone_reclaim(zone)) list_for_each_entry(zone, &dev->map_seq_list, link) {
return zone; if (!zone->bzone)
continue;
if (dmz_lock_zone_reclaim(zone))
return zone;
}
} }
return NULL; return NULL;
@ -2147,7 +2163,7 @@ again:
if (dmz_is_cache(bzone)) if (dmz_is_cache(bzone))
list_add_tail(&bzone->link, &zmd->map_cache_list); list_add_tail(&bzone->link, &zmd->map_cache_list);
else else
list_add_tail(&bzone->link, &zmd->map_rnd_list); list_add_tail(&bzone->link, &bzone->dev->map_rnd_list);
out: out:
dmz_unlock_map(zmd); dmz_unlock_map(zmd);
@ -2162,21 +2178,27 @@ struct dm_zone *dmz_alloc_zone(struct dmz_metadata *zmd, unsigned long flags)
{ {
struct list_head *list; struct list_head *list;
struct dm_zone *zone; struct dm_zone *zone;
unsigned int dev_idx = 0;
again:
if (flags & DMZ_ALLOC_CACHE) if (flags & DMZ_ALLOC_CACHE)
list = &zmd->unmap_cache_list; list = &zmd->unmap_cache_list;
else if (flags & DMZ_ALLOC_RND) else if (flags & DMZ_ALLOC_RND)
list = &zmd->unmap_rnd_list; list = &zmd->dev[dev_idx].unmap_rnd_list;
else else
list = &zmd->unmap_seq_list; list = &zmd->dev[dev_idx].unmap_seq_list;
again:
if (list_empty(list)) { if (list_empty(list)) {
/* /*
* No free zone: return NULL if this is for not reclaim. * No free zone: return NULL if this is for not reclaim.
*/ */
if (!(flags & DMZ_ALLOC_RECLAIM)) if (!(flags & DMZ_ALLOC_RECLAIM))
return NULL; return NULL;
if (dev_idx < zmd->nr_devs) {
dev_idx++;
goto again;
}
/* /*
* Fallback to the reserved sequential zones * Fallback to the reserved sequential zones
*/ */
@ -2195,9 +2217,9 @@ again:
if (dmz_is_cache(zone)) if (dmz_is_cache(zone))
atomic_dec(&zmd->unmap_nr_cache); atomic_dec(&zmd->unmap_nr_cache);
else if (dmz_is_rnd(zone)) else if (dmz_is_rnd(zone))
atomic_dec(&zmd->unmap_nr_rnd); atomic_dec(&zone->dev->unmap_nr_rnd);
else else
atomic_dec(&zmd->unmap_nr_seq); atomic_dec(&zone->dev->unmap_nr_seq);
if (dmz_is_offline(zone)) { if (dmz_is_offline(zone)) {
dmz_zmd_warn(zmd, "Zone %u is offline", zone->id); dmz_zmd_warn(zmd, "Zone %u is offline", zone->id);
@ -2227,14 +2249,14 @@ void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone)
list_add_tail(&zone->link, &zmd->unmap_cache_list); list_add_tail(&zone->link, &zmd->unmap_cache_list);
atomic_inc(&zmd->unmap_nr_cache); atomic_inc(&zmd->unmap_nr_cache);
} else if (dmz_is_rnd(zone)) { } else if (dmz_is_rnd(zone)) {
list_add_tail(&zone->link, &zmd->unmap_rnd_list); list_add_tail(&zone->link, &zone->dev->unmap_rnd_list);
atomic_inc(&zmd->unmap_nr_rnd); atomic_inc(&zone->dev->unmap_nr_rnd);
} else if (dmz_is_reserved(zone)) { } else if (dmz_is_reserved(zone)) {
list_add_tail(&zone->link, &zmd->reserved_seq_zones_list); list_add_tail(&zone->link, &zmd->reserved_seq_zones_list);
atomic_inc(&zmd->nr_reserved_seq_zones); atomic_inc(&zmd->nr_reserved_seq_zones);
} else { } else {
list_add_tail(&zone->link, &zmd->unmap_seq_list); list_add_tail(&zone->link, &zone->dev->unmap_seq_list);
atomic_inc(&zmd->unmap_nr_seq); atomic_inc(&zone->dev->unmap_nr_seq);
} }
wake_up_all(&zmd->free_wq); wake_up_all(&zmd->free_wq);
@ -2254,9 +2276,9 @@ void dmz_map_zone(struct dmz_metadata *zmd, struct dm_zone *dzone,
if (dmz_is_cache(dzone)) if (dmz_is_cache(dzone))
list_add_tail(&dzone->link, &zmd->map_cache_list); list_add_tail(&dzone->link, &zmd->map_cache_list);
else if (dmz_is_rnd(dzone)) else if (dmz_is_rnd(dzone))
list_add_tail(&dzone->link, &zmd->map_rnd_list); list_add_tail(&dzone->link, &dzone->dev->map_rnd_list);
else else
list_add_tail(&dzone->link, &zmd->map_seq_list); list_add_tail(&dzone->link, &dzone->dev->map_seq_list);
} }
/* /*
@ -2824,18 +2846,11 @@ int dmz_ctr_metadata(struct dmz_dev *dev, int num_dev,
INIT_LIST_HEAD(&zmd->mblk_dirty_list); INIT_LIST_HEAD(&zmd->mblk_dirty_list);
mutex_init(&zmd->map_lock); mutex_init(&zmd->map_lock);
atomic_set(&zmd->unmap_nr_rnd, 0);
INIT_LIST_HEAD(&zmd->unmap_rnd_list);
INIT_LIST_HEAD(&zmd->map_rnd_list);
atomic_set(&zmd->unmap_nr_cache, 0); atomic_set(&zmd->unmap_nr_cache, 0);
INIT_LIST_HEAD(&zmd->unmap_cache_list); INIT_LIST_HEAD(&zmd->unmap_cache_list);
INIT_LIST_HEAD(&zmd->map_cache_list); INIT_LIST_HEAD(&zmd->map_cache_list);
atomic_set(&zmd->unmap_nr_seq, 0);
INIT_LIST_HEAD(&zmd->unmap_seq_list);
INIT_LIST_HEAD(&zmd->map_seq_list);
atomic_set(&zmd->nr_reserved_seq_zones, 0); atomic_set(&zmd->nr_reserved_seq_zones, 0);
INIT_LIST_HEAD(&zmd->reserved_seq_zones_list); INIT_LIST_HEAD(&zmd->reserved_seq_zones_list);
@ -2904,10 +2919,14 @@ int dmz_ctr_metadata(struct dmz_dev *dev, int num_dev,
zmd->nr_data_zones, zmd->nr_chunks); zmd->nr_data_zones, zmd->nr_chunks);
dmz_zmd_debug(zmd, " %u cache zones (%u unmapped)", dmz_zmd_debug(zmd, " %u cache zones (%u unmapped)",
zmd->nr_cache, atomic_read(&zmd->unmap_nr_cache)); zmd->nr_cache, atomic_read(&zmd->unmap_nr_cache));
dmz_zmd_debug(zmd, " %u random zones (%u unmapped)", for (i = 0; i < zmd->nr_devs; i++) {
zmd->nr_rnd, atomic_read(&zmd->unmap_nr_rnd)); dmz_zmd_debug(zmd, " %u random zones (%u unmapped)",
dmz_zmd_debug(zmd, " %u sequential zones (%u unmapped)", dmz_nr_rnd_zones(zmd, i),
zmd->nr_seq, atomic_read(&zmd->unmap_nr_seq)); dmz_nr_unmap_rnd_zones(zmd, i));
dmz_zmd_debug(zmd, " %u sequential zones (%u unmapped)",
dmz_nr_seq_zones(zmd, i),
dmz_nr_unmap_seq_zones(zmd, i));
}
dmz_zmd_debug(zmd, " %u reserved sequential data zones", dmz_zmd_debug(zmd, " %u reserved sequential data zones",
zmd->nr_reserved_seq); zmd->nr_reserved_seq);
dmz_zmd_debug(zmd, "Format:"); dmz_zmd_debug(zmd, "Format:");

View File

@ -447,15 +447,14 @@ static unsigned int dmz_reclaim_percentage(struct dmz_reclaim *zrc)
{ {
struct dmz_metadata *zmd = zrc->metadata; struct dmz_metadata *zmd = zrc->metadata;
unsigned int nr_cache = dmz_nr_cache_zones(zmd); unsigned int nr_cache = dmz_nr_cache_zones(zmd);
unsigned int nr_rnd = dmz_nr_rnd_zones(zmd);
unsigned int nr_unmap, nr_zones; unsigned int nr_unmap, nr_zones;
if (nr_cache) { if (nr_cache) {
nr_zones = nr_cache; nr_zones = nr_cache;
nr_unmap = dmz_nr_unmap_cache_zones(zmd); nr_unmap = dmz_nr_unmap_cache_zones(zmd);
} else { } else {
nr_zones = nr_rnd; nr_zones = dmz_nr_rnd_zones(zmd, zrc->dev_idx);
nr_unmap = dmz_nr_unmap_rnd_zones(zmd); nr_unmap = dmz_nr_unmap_rnd_zones(zmd, zrc->dev_idx);
} }
return nr_unmap * 100 / nr_zones; return nr_unmap * 100 / nr_zones;
} }
@ -467,7 +466,7 @@ static bool dmz_should_reclaim(struct dmz_reclaim *zrc, unsigned int p_unmap)
{ {
unsigned int nr_reclaim; unsigned int nr_reclaim;
nr_reclaim = dmz_nr_rnd_zones(zrc->metadata); nr_reclaim = dmz_nr_rnd_zones(zrc->metadata, zrc->dev_idx);
if (dmz_nr_cache_zones(zrc->metadata)) { if (dmz_nr_cache_zones(zrc->metadata)) {
/* /*
@ -528,8 +527,8 @@ static void dmz_reclaim_work(struct work_struct *work)
zrc->kc_throttle.throttle = min(75U, 100U - p_unmap / 2); zrc->kc_throttle.throttle = min(75U, 100U - p_unmap / 2);
} }
nr_unmap_rnd = dmz_nr_unmap_rnd_zones(zmd); nr_unmap_rnd = dmz_nr_unmap_rnd_zones(zmd, zrc->dev_idx);
nr_rnd = dmz_nr_rnd_zones(zmd); nr_rnd = dmz_nr_rnd_zones(zmd, zrc->dev_idx);
DMDEBUG("(%s/%u): Reclaim (%u): %s, %u%% free zones (%u/%u cache %u/%u random)", DMDEBUG("(%s/%u): Reclaim (%u): %s, %u%% free zones (%u/%u cache %u/%u random)",
dmz_metadata_label(zmd), zrc->dev_idx, dmz_metadata_label(zmd), zrc->dev_idx,
@ -537,8 +536,8 @@ static void dmz_reclaim_work(struct work_struct *work)
(dmz_target_idle(zrc) ? "Idle" : "Busy"), (dmz_target_idle(zrc) ? "Idle" : "Busy"),
p_unmap, dmz_nr_unmap_cache_zones(zmd), p_unmap, dmz_nr_unmap_cache_zones(zmd),
dmz_nr_cache_zones(zmd), dmz_nr_cache_zones(zmd),
dmz_nr_unmap_rnd_zones(zmd), dmz_nr_unmap_rnd_zones(zmd, zrc->dev_idx),
dmz_nr_rnd_zones(zmd)); dmz_nr_rnd_zones(zmd, zrc->dev_idx));
ret = dmz_do_reclaim(zrc); ret = dmz_do_reclaim(zrc);
if (ret && ret != -EINTR) { if (ret && ret != -EINTR) {

View File

@ -1075,17 +1075,30 @@ static void dmz_status(struct dm_target *ti, status_type_t type,
ssize_t sz = 0; ssize_t sz = 0;
char buf[BDEVNAME_SIZE]; char buf[BDEVNAME_SIZE];
struct dmz_dev *dev; struct dmz_dev *dev;
int i;
switch (type) { switch (type) {
case STATUSTYPE_INFO: case STATUSTYPE_INFO:
DMEMIT("%u zones %u/%u cache %u/%u random %u/%u sequential", DMEMIT("%u zones %u/%u cache",
dmz_nr_zones(dmz->metadata), dmz_nr_zones(dmz->metadata),
dmz_nr_unmap_cache_zones(dmz->metadata), dmz_nr_unmap_cache_zones(dmz->metadata),
dmz_nr_cache_zones(dmz->metadata), dmz_nr_cache_zones(dmz->metadata));
dmz_nr_unmap_rnd_zones(dmz->metadata), for (i = 0; i < DMZ_MAX_DEVS; i++) {
dmz_nr_rnd_zones(dmz->metadata), if (!dmz->ddev[i])
dmz_nr_unmap_seq_zones(dmz->metadata), continue;
dmz_nr_seq_zones(dmz->metadata)); /*
* For a multi-device setup the first device
* contains only cache zones.
*/
if ((i == 0) &&
(dmz_nr_cache_zones(dmz->metadata) > 0))
continue;
DMEMIT(" %u/%u random %u/%u sequential",
dmz_nr_unmap_rnd_zones(dmz->metadata, i),
dmz_nr_rnd_zones(dmz->metadata, i),
dmz_nr_unmap_seq_zones(dmz->metadata, i),
dmz_nr_seq_zones(dmz->metadata, i));
}
break; break;
case STATUSTYPE_TABLE: case STATUSTYPE_TABLE:
dev = &dmz->dev[0]; dev = &dmz->dev[0];

View File

@ -67,6 +67,16 @@ struct dmz_dev {
unsigned int flags; unsigned int flags;
sector_t zone_nr_sectors; sector_t zone_nr_sectors;
unsigned int nr_rnd;
atomic_t unmap_nr_rnd;
struct list_head unmap_rnd_list;
struct list_head map_rnd_list;
unsigned int nr_seq;
atomic_t unmap_nr_seq;
struct list_head unmap_seq_list;
struct list_head map_seq_list;
}; };
#define dmz_bio_chunk(zmd, bio) ((bio)->bi_iter.bi_sector >> \ #define dmz_bio_chunk(zmd, bio) ((bio)->bi_iter.bi_sector >> \
@ -213,10 +223,10 @@ void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
unsigned int dmz_nr_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_zones(struct dmz_metadata *zmd);
unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd);
unsigned int dmz_nr_unmap_cache_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_unmap_cache_zones(struct dmz_metadata *zmd);
unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd, int idx);
unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd, int idx);
unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd, int idx);
unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd); unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd, int idx);
unsigned int dmz_zone_nr_blocks(struct dmz_metadata *zmd); unsigned int dmz_zone_nr_blocks(struct dmz_metadata *zmd);
unsigned int dmz_zone_nr_blocks_shift(struct dmz_metadata *zmd); unsigned int dmz_zone_nr_blocks_shift(struct dmz_metadata *zmd);
unsigned int dmz_zone_nr_sectors(struct dmz_metadata *zmd); unsigned int dmz_zone_nr_sectors(struct dmz_metadata *zmd);