block: convert to dequeueing model (easy ones)
plat-omap/mailbox, floppy, viocd, mspro_block, i2o_block and mmc/card/queue are already pretty close to dequeueing model and can be converted with simple changes. Convert them. While at it, * xen-blkfront: !fs check moved downwards to share dequeue call with normal path. * mspro_block: __blk_end_request(..., blk_rq_cur_byte()) converted to __blk_end_request_cur() * mmc/card/queue: loop of __blk_end_request() converted to __blk_end_request_all() [ Impact: dequeue in-flight request ] Signed-off-by: Tejun Heo <tj@kernel.org> Cc: Rusty Russell <rusty@rustcorp.com.au> Cc: Jeremy Fitzhardinge <jeremy@xensource.com> Cc: Stephen Rothwell <sfr@canb.auug.org.au> Cc: Alex Dubov <oakad@yahoo.com> Cc: Markus Lidel <Markus.Lidel@shadowconnect.com> Cc: Pierre Ossman <drzeus@drzeus.cx> Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
This commit is contained in:
parent
fb3ac7f6b8
commit
296b2f6ae6
|
@ -198,6 +198,8 @@ static void mbox_tx_work(struct work_struct *work)
|
||||||
|
|
||||||
spin_lock(q->queue_lock);
|
spin_lock(q->queue_lock);
|
||||||
rq = elv_next_request(q);
|
rq = elv_next_request(q);
|
||||||
|
if (rq)
|
||||||
|
blkdev_dequeue_request(rq);
|
||||||
spin_unlock(q->queue_lock);
|
spin_unlock(q->queue_lock);
|
||||||
|
|
||||||
if (!rq)
|
if (!rq)
|
||||||
|
@ -208,6 +210,9 @@ static void mbox_tx_work(struct work_struct *work)
|
||||||
ret = __mbox_msg_send(mbox, tx_data->msg, tx_data->arg);
|
ret = __mbox_msg_send(mbox, tx_data->msg, tx_data->arg);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
enable_mbox_irq(mbox, IRQ_TX);
|
enable_mbox_irq(mbox, IRQ_TX);
|
||||||
|
spin_lock(q->queue_lock);
|
||||||
|
blk_requeue_request(q, rq);
|
||||||
|
spin_unlock(q->queue_lock);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -238,6 +243,8 @@ static void mbox_rx_work(struct work_struct *work)
|
||||||
while (1) {
|
while (1) {
|
||||||
spin_lock_irqsave(q->queue_lock, flags);
|
spin_lock_irqsave(q->queue_lock, flags);
|
||||||
rq = elv_next_request(q);
|
rq = elv_next_request(q);
|
||||||
|
if (rq)
|
||||||
|
blkdev_dequeue_request(rq);
|
||||||
spin_unlock_irqrestore(q->queue_lock, flags);
|
spin_unlock_irqrestore(q->queue_lock, flags);
|
||||||
if (!rq)
|
if (!rq)
|
||||||
break;
|
break;
|
||||||
|
@ -345,6 +352,8 @@ omap_mbox_read(struct device *dev, struct device_attribute *attr, char *buf)
|
||||||
while (1) {
|
while (1) {
|
||||||
spin_lock_irqsave(q->queue_lock, flags);
|
spin_lock_irqsave(q->queue_lock, flags);
|
||||||
rq = elv_next_request(q);
|
rq = elv_next_request(q);
|
||||||
|
if (rq)
|
||||||
|
blkdev_dequeue_request(rq);
|
||||||
spin_unlock_irqrestore(q->queue_lock, flags);
|
spin_unlock_irqrestore(q->queue_lock, flags);
|
||||||
|
|
||||||
if (!rq)
|
if (!rq)
|
||||||
|
|
|
@ -931,7 +931,7 @@ static inline void unlock_fdc(void)
|
||||||
del_timer(&fd_timeout);
|
del_timer(&fd_timeout);
|
||||||
cont = NULL;
|
cont = NULL;
|
||||||
clear_bit(0, &fdc_busy);
|
clear_bit(0, &fdc_busy);
|
||||||
if (elv_next_request(floppy_queue))
|
if (current_req || elv_next_request(floppy_queue))
|
||||||
do_fd_request(floppy_queue);
|
do_fd_request(floppy_queue);
|
||||||
spin_unlock_irqrestore(&floppy_lock, flags);
|
spin_unlock_irqrestore(&floppy_lock, flags);
|
||||||
wake_up(&fdc_wait);
|
wake_up(&fdc_wait);
|
||||||
|
@ -2913,6 +2913,8 @@ static void redo_fd_request(void)
|
||||||
|
|
||||||
spin_lock_irq(floppy_queue->queue_lock);
|
spin_lock_irq(floppy_queue->queue_lock);
|
||||||
req = elv_next_request(floppy_queue);
|
req = elv_next_request(floppy_queue);
|
||||||
|
if (req)
|
||||||
|
blkdev_dequeue_request(req);
|
||||||
spin_unlock_irq(floppy_queue->queue_lock);
|
spin_unlock_irq(floppy_queue->queue_lock);
|
||||||
if (!req) {
|
if (!req) {
|
||||||
do_floppy = NULL;
|
do_floppy = NULL;
|
||||||
|
|
|
@ -301,22 +301,23 @@ static void do_blkif_request(struct request_queue *rq)
|
||||||
|
|
||||||
while ((req = elv_next_request(rq)) != NULL) {
|
while ((req = elv_next_request(rq)) != NULL) {
|
||||||
info = req->rq_disk->private_data;
|
info = req->rq_disk->private_data;
|
||||||
if (!blk_fs_request(req)) {
|
|
||||||
__blk_end_request_cur(req, -EIO);
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (RING_FULL(&info->ring))
|
if (RING_FULL(&info->ring))
|
||||||
goto wait;
|
goto wait;
|
||||||
|
|
||||||
|
blkdev_dequeue_request(req);
|
||||||
|
|
||||||
|
if (!blk_fs_request(req)) {
|
||||||
|
__blk_end_request_all(req, -EIO);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
pr_debug("do_blk_req %p: cmd %p, sec %lx, "
|
pr_debug("do_blk_req %p: cmd %p, sec %lx, "
|
||||||
"(%u/%u) buffer:%p [%s]\n",
|
"(%u/%u) buffer:%p [%s]\n",
|
||||||
req, req->cmd, (unsigned long)blk_rq_pos(req),
|
req, req->cmd, (unsigned long)blk_rq_pos(req),
|
||||||
blk_rq_cur_sectors(req), blk_rq_sectors(req),
|
blk_rq_cur_sectors(req), blk_rq_sectors(req),
|
||||||
req->buffer, rq_data_dir(req) ? "write" : "read");
|
req->buffer, rq_data_dir(req) ? "write" : "read");
|
||||||
|
|
||||||
|
|
||||||
blkdev_dequeue_request(req);
|
|
||||||
if (blkif_queue_request(req)) {
|
if (blkif_queue_request(req)) {
|
||||||
blk_requeue_request(rq, req);
|
blk_requeue_request(rq, req);
|
||||||
wait:
|
wait:
|
||||||
|
|
|
@ -298,6 +298,8 @@ static void do_viocd_request(struct request_queue *q)
|
||||||
struct request *req;
|
struct request *req;
|
||||||
|
|
||||||
while ((rwreq == 0) && ((req = elv_next_request(q)) != NULL)) {
|
while ((rwreq == 0) && ((req = elv_next_request(q)) != NULL)) {
|
||||||
|
blkdev_dequeue_request(req);
|
||||||
|
|
||||||
if (!blk_fs_request(req))
|
if (!blk_fs_request(req))
|
||||||
__blk_end_request_all(req, -EIO);
|
__blk_end_request_all(req, -EIO);
|
||||||
else if (send_request(req) < 0) {
|
else if (send_request(req) < 0) {
|
||||||
|
|
|
@ -672,8 +672,7 @@ try_again:
|
||||||
msb->req_sg);
|
msb->req_sg);
|
||||||
|
|
||||||
if (!msb->seg_count) {
|
if (!msb->seg_count) {
|
||||||
chunk = __blk_end_request(msb->block_req, -ENOMEM,
|
chunk = __blk_end_request_cur(msb->block_req, -ENOMEM);
|
||||||
blk_rq_cur_bytes(msb->block_req));
|
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -711,6 +710,7 @@ try_again:
|
||||||
dev_dbg(&card->dev, "issue end\n");
|
dev_dbg(&card->dev, "issue end\n");
|
||||||
return -EAGAIN;
|
return -EAGAIN;
|
||||||
}
|
}
|
||||||
|
blkdev_dequeue_request(msb->block_req);
|
||||||
|
|
||||||
dev_dbg(&card->dev, "trying again\n");
|
dev_dbg(&card->dev, "trying again\n");
|
||||||
chunk = 1;
|
chunk = 1;
|
||||||
|
@ -825,8 +825,10 @@ static void mspro_block_submit_req(struct request_queue *q)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (msb->eject) {
|
if (msb->eject) {
|
||||||
while ((req = elv_next_request(q)) != NULL)
|
while ((req = elv_next_request(q)) != NULL) {
|
||||||
|
blkdev_dequeue_request(req);
|
||||||
__blk_end_request_all(req, -ENODEV);
|
__blk_end_request_all(req, -ENODEV);
|
||||||
|
}
|
||||||
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
|
@ -916,8 +916,10 @@ static void i2o_block_request_fn(struct request_queue *q)
|
||||||
blk_stop_queue(q);
|
blk_stop_queue(q);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
} else
|
} else {
|
||||||
__blk_end_request_cur(req, -EIO);
|
blkdev_dequeue_request(req);
|
||||||
|
__blk_end_request_all(req, -EIO);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -54,8 +54,11 @@ static int mmc_queue_thread(void *d)
|
||||||
|
|
||||||
spin_lock_irq(q->queue_lock);
|
spin_lock_irq(q->queue_lock);
|
||||||
set_current_state(TASK_INTERRUPTIBLE);
|
set_current_state(TASK_INTERRUPTIBLE);
|
||||||
if (!blk_queue_plugged(q))
|
if (!blk_queue_plugged(q)) {
|
||||||
req = elv_next_request(q);
|
req = elv_next_request(q);
|
||||||
|
if (req)
|
||||||
|
blkdev_dequeue_request(req);
|
||||||
|
}
|
||||||
mq->req = req;
|
mq->req = req;
|
||||||
spin_unlock_irq(q->queue_lock);
|
spin_unlock_irq(q->queue_lock);
|
||||||
|
|
||||||
|
@ -88,15 +91,12 @@ static void mmc_request(struct request_queue *q)
|
||||||
{
|
{
|
||||||
struct mmc_queue *mq = q->queuedata;
|
struct mmc_queue *mq = q->queuedata;
|
||||||
struct request *req;
|
struct request *req;
|
||||||
int ret;
|
|
||||||
|
|
||||||
if (!mq) {
|
if (!mq) {
|
||||||
printk(KERN_ERR "MMC: killing requests for dead queue\n");
|
printk(KERN_ERR "MMC: killing requests for dead queue\n");
|
||||||
while ((req = elv_next_request(q)) != NULL) {
|
while ((req = elv_next_request(q)) != NULL) {
|
||||||
do {
|
blkdev_dequeue_request(req);
|
||||||
ret = __blk_end_request(req, -EIO,
|
__blk_end_request_all(req, -EIO);
|
||||||
blk_rq_cur_bytes(req));
|
|
||||||
} while (ret);
|
|
||||||
}
|
}
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue