dm raid1: use hold framework in do_failures
Use the hold framework in do_failures. This patch doesn't change the bio processing logic, it just simplifies failure handling and avoids periodically polling the failures list. Signed-off-by: Mikulas Patocka <mpatocka@redhat.com> Reviewed-by: Takahiro Yasui <tyasui@redhat.com> Tested-by: Takahiro Yasui <tyasui@redhat.com> Signed-off-by: Alasdair G Kergon <agk@redhat.com>
This commit is contained in:
parent
0478850768
commit
0f398a8403
|
@ -745,20 +745,12 @@ static void do_failures(struct mirror_set *ms, struct bio_list *failures)
|
||||||
{
|
{
|
||||||
struct bio *bio;
|
struct bio *bio;
|
||||||
|
|
||||||
if (!failures->head)
|
if (likely(!failures->head))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (!ms->log_failure) {
|
|
||||||
while ((bio = bio_list_pop(failures))) {
|
|
||||||
ms->in_sync = 0;
|
|
||||||
dm_rh_mark_nosync(ms->rh, bio, bio->bi_size, 0);
|
|
||||||
}
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If the log has failed, unattempted writes are being
|
* If the log has failed, unattempted writes are being
|
||||||
* put on the failures list. We can't issue those writes
|
* put on the holds list. We can't issue those writes
|
||||||
* until a log has been marked, so we must store them.
|
* until a log has been marked, so we must store them.
|
||||||
*
|
*
|
||||||
* If a 'noflush' suspend is in progress, we can requeue
|
* If a 'noflush' suspend is in progress, we can requeue
|
||||||
|
@ -773,23 +765,15 @@ static void do_failures(struct mirror_set *ms, struct bio_list *failures)
|
||||||
* for us to treat them the same and requeue them
|
* for us to treat them the same and requeue them
|
||||||
* as well.
|
* as well.
|
||||||
*/
|
*/
|
||||||
if (dm_noflush_suspending(ms->ti)) {
|
|
||||||
while ((bio = bio_list_pop(failures)))
|
while ((bio = bio_list_pop(failures))) {
|
||||||
bio_endio(bio, DM_ENDIO_REQUEUE);
|
if (ms->log_failure)
|
||||||
return;
|
hold_bio(ms, bio);
|
||||||
|
else {
|
||||||
|
ms->in_sync = 0;
|
||||||
|
dm_rh_mark_nosync(ms->rh, bio, bio->bi_size, 0);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (atomic_read(&ms->suspend)) {
|
|
||||||
while ((bio = bio_list_pop(failures)))
|
|
||||||
bio_endio(bio, -EIO);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
spin_lock_irq(&ms->lock);
|
|
||||||
bio_list_merge(&ms->failures, failures);
|
|
||||||
spin_unlock_irq(&ms->lock);
|
|
||||||
|
|
||||||
delayed_wake(ms);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void trigger_event(struct work_struct *work)
|
static void trigger_event(struct work_struct *work)
|
||||||
|
|
Loading…
Reference in New Issue