Skip to content

Commit 55cbeea

Browse files
axboegregkh
authored andcommitted
blk-mq: punt failed direct issue to dispatch list
commit c616cbe upstream. After the direct dispatch corruption fix, we permanently disallow direct dispatch of non read/write requests. This works fine off the normal IO path, as they will be retried like any other failed direct dispatch request. But for the blk_insert_cloned_request() that only DM uses to bypass the bottom level scheduler, we always first attempt direct dispatch. For some types of requests, that's now a permanent failure, and no amount of retrying will make that succeed. This results in a livelock. Instead of making special cases for what we can direct issue, and now having to deal with DM solving the livelock while still retaining a BUSY condition feedback loop, always just add a request that has been through ->queue_rq() to the hardware queue dispatch list. These are safe to use as no merging can take place there. Additionally, if requests do have prepped data from drivers, we aren't dependent on them not sharing space in the request structure to safely add them to the IO scheduler lists. This basically reverts ffe81d4 and is based on a patch from Ming, but with the list insert case covered as well. Fixes: ffe81d4 ("blk-mq: fix corruption with direct issue") Cc: stable@vger.kernel.org Suggested-by: Ming Lei <ming.lei@redhat.com> Reported-by: Bart Van Assche <bvanassche@acm.org> Tested-by: Ming Lei <ming.lei@redhat.com> Acked-by: Mike Snitzer <snitzer@redhat.com> Signed-off-by: Jens Axboe <axboe@kernel.dk> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
1 parent 6235c22 commit 55cbeea

1 file changed

Lines changed: 5 additions & 28 deletions

File tree

block/blk-mq.c

Lines changed: 5 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -1698,15 +1698,6 @@ static blk_status_t __blk_mq_issue_directly(struct blk_mq_hw_ctx *hctx,
16981698
break;
16991699
case BLK_STS_RESOURCE:
17001700
case BLK_STS_DEV_RESOURCE:
1701-
/*
1702-
* If direct dispatch fails, we cannot allow any merging on
1703-
* this IO. Drivers (like SCSI) may have set up permanent state
1704-
* for this request, like SG tables and mappings, and if we
1705-
* merge to it later on then we'll still only do IO to the
1706-
* original part.
1707-
*/
1708-
rq->cmd_flags |= REQ_NOMERGE;
1709-
17101701
blk_mq_update_dispatch_busy(hctx, true);
17111702
__blk_mq_requeue_request(rq);
17121703
break;
@@ -1719,18 +1710,6 @@ static blk_status_t __blk_mq_issue_directly(struct blk_mq_hw_ctx *hctx,
17191710
return ret;
17201711
}
17211712

1722-
/*
1723-
* Don't allow direct dispatch of anything but regular reads/writes,
1724-
* as some of the other commands can potentially share request space
1725-
* with data we need for the IO scheduler. If we attempt a direct dispatch
1726-
* on those and fail, we can't safely add it to the scheduler afterwards
1727-
* without potentially overwriting data that the driver has already written.
1728-
*/
1729-
static bool blk_rq_can_direct_dispatch(struct request *rq)
1730-
{
1731-
return req_op(rq) == REQ_OP_READ || req_op(rq) == REQ_OP_WRITE;
1732-
}
1733-
17341713
static blk_status_t __blk_mq_try_issue_directly(struct blk_mq_hw_ctx *hctx,
17351714
struct request *rq,
17361715
blk_qc_t *cookie,
@@ -1752,7 +1731,7 @@ static blk_status_t __blk_mq_try_issue_directly(struct blk_mq_hw_ctx *hctx,
17521731
goto insert;
17531732
}
17541733

1755-
if (!blk_rq_can_direct_dispatch(rq) || (q->elevator && !bypass_insert))
1734+
if (q->elevator && !bypass_insert)
17561735
goto insert;
17571736

17581737
if (!blk_mq_get_dispatch_budget(hctx))
@@ -1768,7 +1747,7 @@ static blk_status_t __blk_mq_try_issue_directly(struct blk_mq_hw_ctx *hctx,
17681747
if (bypass_insert)
17691748
return BLK_STS_RESOURCE;
17701749

1771-
blk_mq_sched_insert_request(rq, false, run_queue, false);
1750+
blk_mq_request_bypass_insert(rq, run_queue);
17721751
return BLK_STS_OK;
17731752
}
17741753

@@ -1784,7 +1763,7 @@ static void blk_mq_try_issue_directly(struct blk_mq_hw_ctx *hctx,
17841763

17851764
ret = __blk_mq_try_issue_directly(hctx, rq, cookie, false);
17861765
if (ret == BLK_STS_RESOURCE || ret == BLK_STS_DEV_RESOURCE)
1787-
blk_mq_sched_insert_request(rq, false, true, false);
1766+
blk_mq_request_bypass_insert(rq, true);
17881767
else if (ret != BLK_STS_OK)
17891768
blk_mq_end_request(rq, ret);
17901769

@@ -1814,15 +1793,13 @@ void blk_mq_try_issue_list_directly(struct blk_mq_hw_ctx *hctx,
18141793
struct request *rq = list_first_entry(list, struct request,
18151794
queuelist);
18161795

1817-
if (!blk_rq_can_direct_dispatch(rq))
1818-
break;
1819-
18201796
list_del_init(&rq->queuelist);
18211797
ret = blk_mq_request_issue_directly(rq);
18221798
if (ret != BLK_STS_OK) {
18231799
if (ret == BLK_STS_RESOURCE ||
18241800
ret == BLK_STS_DEV_RESOURCE) {
1825-
list_add(&rq->queuelist, list);
1801+
blk_mq_request_bypass_insert(rq,
1802+
list_empty(list));
18261803
break;
18271804
}
18281805
blk_mq_end_request(rq, ret);

0 commit comments

Comments
 (0)