dm cache policy smq: cleanup free_target_met() and clean_target_met()

Depending on the passed @idle arg, there may be no need to calculate
'nr_free' or 'nr_clean' respectively in free_target_met() and
clean_target_met().

Signed-off-by: Mike Snitzer <snitzer@redhat.com>
This commit is contained in:
Mike Snitzer 2017-05-04 10:04:18 -04:00
parent ce1d64e84d
commit 97dfb20309
1 changed files with 13 additions and 11 deletions

View File

@ -1120,28 +1120,30 @@ static bool clean_target_met(struct smq_policy *mq, bool idle)
* Cache entries may not be populated. So we cannot rely on the
* size of the clean queue.
*/
unsigned nr_clean = from_cblock(mq->cache_size) - q_size(&mq->dirty);
unsigned nr_clean;
if (idle)
if (idle) {
/*
* We'd like to clean everything.
*/
return q_size(&mq->dirty) == 0u;
else
}
nr_clean = from_cblock(mq->cache_size) - q_size(&mq->dirty);
return (nr_clean + btracker_nr_writebacks_queued(mq->bg_work)) >=
percent_to_target(mq, CLEAN_TARGET);
}
static bool free_target_met(struct smq_policy *mq, bool idle)
{
unsigned nr_free = from_cblock(mq->cache_size) -
mq->cache_alloc.nr_allocated;
unsigned nr_free;
if (idle)
if (!idle)
return true;
nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated;
return (nr_free + btracker_nr_demotions_queued(mq->bg_work)) >=
percent_to_target(mq, FREE_TARGET);
else
return true;
}
/*----------------------------------------------------------------*/