mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 04:38:03 +00:00
scsi: core: avoid host-wide host_busy counter for scsi_mq
It isn't necessary to check the host depth in scsi_queue_rq() any more since it has been respected by blk-mq before calling scsi_queue_rq() via getting driver tag. Lots of LUNs may attach to same host and per-host IOPS may reach millions, so we should avoid expensive atomic operations on the host-wide counter in the IO path. This patch implements scsi_host_busy() via blk_mq_tagset_busy_iter() with one scsi command state for reading the count of busy IOs for scsi_mq. It is observed that IOPS is increased by 15% in IO test on scsi_debug (32 LUNs, 32 submit queues, 1024 can_queue, libaio/dio) in a dual-socket system. Cc: Jens Axboe <axboe@kernel.dk> Cc: Ewan D. Milne <emilne@redhat.com> Cc: Omar Sandoval <osandov@fb.com>, Cc: "Martin K. Petersen" <martin.petersen@oracle.com>, Cc: James Bottomley <james.bottomley@hansenpartnership.com>, Cc: Christoph Hellwig <hch@lst.de>, Cc: Kashyap Desai <kashyap.desai@broadcom.com> Cc: Hannes Reinecke <hare@suse.de> Cc: Laurence Oberman <loberman@redhat.com> Cc: Bart Van Assche <bvanassche@acm.org> Link: https://lore.kernel.org/r/20191025065855.6309-1-ming.lei@redhat.com Signed-off-by: Ming Lei <ming.lei@redhat.com> Reviewed-by: Jens Axboe <axboe@kernel.dk> Reviewed-by: Bart Van Assche <bvanassche@acm.org> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
This commit is contained in:
parent
7f674c38a3
commit
6eb045e092
@ -38,6 +38,7 @@
|
||||
#include <scsi/scsi_device.h>
|
||||
#include <scsi/scsi_host.h>
|
||||
#include <scsi/scsi_transport.h>
|
||||
#include <scsi/scsi_cmnd.h>
|
||||
|
||||
#include "scsi_priv.h"
|
||||
#include "scsi_logging.h"
|
||||
@ -554,13 +555,29 @@ struct Scsi_Host *scsi_host_get(struct Scsi_Host *shost)
|
||||
}
|
||||
EXPORT_SYMBOL(scsi_host_get);
|
||||
|
||||
static bool scsi_host_check_in_flight(struct request *rq, void *data,
|
||||
bool reserved)
|
||||
{
|
||||
int *count = data;
|
||||
struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq);
|
||||
|
||||
if (test_bit(SCMD_STATE_INFLIGHT, &cmd->state))
|
||||
(*count)++;
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
/**
|
||||
* scsi_host_busy - Return the host busy counter
|
||||
* @shost: Pointer to Scsi_Host to inc.
|
||||
**/
|
||||
int scsi_host_busy(struct Scsi_Host *shost)
|
||||
{
|
||||
return atomic_read(&shost->host_busy);
|
||||
int cnt = 0;
|
||||
|
||||
blk_mq_tagset_busy_iter(&shost->tag_set,
|
||||
scsi_host_check_in_flight, &cnt);
|
||||
return cnt;
|
||||
}
|
||||
EXPORT_SYMBOL(scsi_host_busy);
|
||||
|
||||
|
@ -186,7 +186,7 @@ void scsi_finish_command(struct scsi_cmnd *cmd)
|
||||
struct scsi_driver *drv;
|
||||
unsigned int good_bytes;
|
||||
|
||||
scsi_device_unbusy(sdev);
|
||||
scsi_device_unbusy(sdev, cmd);
|
||||
|
||||
/*
|
||||
* Clear the flags that say that the device/target/host is no longer
|
||||
|
@ -189,7 +189,7 @@ static void __scsi_queue_insert(struct scsi_cmnd *cmd, int reason, bool unbusy)
|
||||
* active on the host/device.
|
||||
*/
|
||||
if (unbusy)
|
||||
scsi_device_unbusy(device);
|
||||
scsi_device_unbusy(device, cmd);
|
||||
|
||||
/*
|
||||
* Requeue this command. It will go before all other commands
|
||||
@ -321,20 +321,20 @@ static void scsi_init_cmd_errh(struct scsi_cmnd *cmd)
|
||||
}
|
||||
|
||||
/*
|
||||
* Decrement the host_busy counter and wake up the error handler if necessary.
|
||||
* Avoid as follows that the error handler is not woken up if shost->host_busy
|
||||
* == shost->host_failed: use call_rcu() in scsi_eh_scmd_add() in combination
|
||||
* with an RCU read lock in this function to ensure that this function in its
|
||||
* entirety either finishes before scsi_eh_scmd_add() increases the
|
||||
* Wake up the error handler if necessary. Avoid as follows that the error
|
||||
* handler is not woken up if host in-flight requests number ==
|
||||
* shost->host_failed: use call_rcu() in scsi_eh_scmd_add() in combination
|
||||
* with an RCU read lock in this function to ensure that this function in
|
||||
* its entirety either finishes before scsi_eh_scmd_add() increases the
|
||||
* host_failed counter or that it notices the shost state change made by
|
||||
* scsi_eh_scmd_add().
|
||||
*/
|
||||
static void scsi_dec_host_busy(struct Scsi_Host *shost)
|
||||
static void scsi_dec_host_busy(struct Scsi_Host *shost, struct scsi_cmnd *cmd)
|
||||
{
|
||||
unsigned long flags;
|
||||
|
||||
rcu_read_lock();
|
||||
atomic_dec(&shost->host_busy);
|
||||
__clear_bit(SCMD_STATE_INFLIGHT, &cmd->state);
|
||||
if (unlikely(scsi_host_in_recovery(shost))) {
|
||||
spin_lock_irqsave(shost->host_lock, flags);
|
||||
if (shost->host_failed || shost->host_eh_scheduled)
|
||||
@ -344,12 +344,12 @@ static void scsi_dec_host_busy(struct Scsi_Host *shost)
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
void scsi_device_unbusy(struct scsi_device *sdev)
|
||||
void scsi_device_unbusy(struct scsi_device *sdev, struct scsi_cmnd *cmd)
|
||||
{
|
||||
struct Scsi_Host *shost = sdev->host;
|
||||
struct scsi_target *starget = scsi_target(sdev);
|
||||
|
||||
scsi_dec_host_busy(shost);
|
||||
scsi_dec_host_busy(shost, cmd);
|
||||
|
||||
if (starget->can_queue > 0)
|
||||
atomic_dec(&starget->target_busy);
|
||||
@ -430,9 +430,6 @@ static inline bool scsi_target_is_busy(struct scsi_target *starget)
|
||||
|
||||
static inline bool scsi_host_is_busy(struct Scsi_Host *shost)
|
||||
{
|
||||
if (shost->can_queue > 0 &&
|
||||
atomic_read(&shost->host_busy) >= shost->can_queue)
|
||||
return true;
|
||||
if (atomic_read(&shost->host_blocked) > 0)
|
||||
return true;
|
||||
if (shost->host_self_blocked)
|
||||
@ -1139,6 +1136,7 @@ void scsi_init_command(struct scsi_device *dev, struct scsi_cmnd *cmd)
|
||||
unsigned int flags = cmd->flags & SCMD_PRESERVED_FLAGS;
|
||||
unsigned long jiffies_at_alloc;
|
||||
int retries;
|
||||
bool in_flight;
|
||||
|
||||
if (!blk_rq_is_scsi(rq) && !(flags & SCMD_INITIALIZED)) {
|
||||
flags |= SCMD_INITIALIZED;
|
||||
@ -1147,6 +1145,7 @@ void scsi_init_command(struct scsi_device *dev, struct scsi_cmnd *cmd)
|
||||
|
||||
jiffies_at_alloc = cmd->jiffies_at_alloc;
|
||||
retries = cmd->retries;
|
||||
in_flight = test_bit(SCMD_STATE_INFLIGHT, &cmd->state);
|
||||
/* zero out the cmd, except for the embedded scsi_request */
|
||||
memset((char *)cmd + sizeof(cmd->req), 0,
|
||||
sizeof(*cmd) - sizeof(cmd->req) + dev->host->hostt->cmd_size);
|
||||
@ -1158,6 +1157,8 @@ void scsi_init_command(struct scsi_device *dev, struct scsi_cmnd *cmd)
|
||||
INIT_DELAYED_WORK(&cmd->abort_work, scmd_eh_abort_handler);
|
||||
cmd->jiffies_at_alloc = jiffies_at_alloc;
|
||||
cmd->retries = retries;
|
||||
if (in_flight)
|
||||
__set_bit(SCMD_STATE_INFLIGHT, &cmd->state);
|
||||
|
||||
scsi_add_cmd_to_list(cmd);
|
||||
}
|
||||
@ -1367,16 +1368,14 @@ static inline int scsi_target_queue_ready(struct Scsi_Host *shost,
|
||||
*/
|
||||
static inline int scsi_host_queue_ready(struct request_queue *q,
|
||||
struct Scsi_Host *shost,
|
||||
struct scsi_device *sdev)
|
||||
struct scsi_device *sdev,
|
||||
struct scsi_cmnd *cmd)
|
||||
{
|
||||
unsigned int busy;
|
||||
|
||||
if (scsi_host_in_recovery(shost))
|
||||
return 0;
|
||||
|
||||
busy = atomic_inc_return(&shost->host_busy) - 1;
|
||||
if (atomic_read(&shost->host_blocked) > 0) {
|
||||
if (busy)
|
||||
if (scsi_host_busy(shost) > 0)
|
||||
goto starved;
|
||||
|
||||
/*
|
||||
@ -1390,8 +1389,6 @@ static inline int scsi_host_queue_ready(struct request_queue *q,
|
||||
"unblocking host at zero depth\n"));
|
||||
}
|
||||
|
||||
if (shost->can_queue > 0 && busy >= shost->can_queue)
|
||||
goto starved;
|
||||
if (shost->host_self_blocked)
|
||||
goto starved;
|
||||
|
||||
@ -1403,6 +1400,8 @@ static inline int scsi_host_queue_ready(struct request_queue *q,
|
||||
spin_unlock_irq(shost->host_lock);
|
||||
}
|
||||
|
||||
__set_bit(SCMD_STATE_INFLIGHT, &cmd->state);
|
||||
|
||||
return 1;
|
||||
|
||||
starved:
|
||||
@ -1411,7 +1410,7 @@ static inline int scsi_host_queue_ready(struct request_queue *q,
|
||||
list_add_tail(&sdev->starved_entry, &shost->starved_list);
|
||||
spin_unlock_irq(shost->host_lock);
|
||||
out_dec:
|
||||
scsi_dec_host_busy(shost);
|
||||
scsi_dec_host_busy(shost, cmd);
|
||||
return 0;
|
||||
}
|
||||
|
||||
@ -1665,7 +1664,7 @@ static blk_status_t scsi_queue_rq(struct blk_mq_hw_ctx *hctx,
|
||||
ret = BLK_STS_RESOURCE;
|
||||
if (!scsi_target_queue_ready(shost, sdev))
|
||||
goto out_put_budget;
|
||||
if (!scsi_host_queue_ready(q, shost, sdev))
|
||||
if (!scsi_host_queue_ready(q, shost, sdev, cmd))
|
||||
goto out_dec_target_busy;
|
||||
|
||||
if (!(req->rq_flags & RQF_DONTPREP)) {
|
||||
@ -1697,7 +1696,7 @@ static blk_status_t scsi_queue_rq(struct blk_mq_hw_ctx *hctx,
|
||||
return BLK_STS_OK;
|
||||
|
||||
out_dec_host_busy:
|
||||
scsi_dec_host_busy(shost);
|
||||
scsi_dec_host_busy(shost, cmd);
|
||||
out_dec_target_busy:
|
||||
if (scsi_target(sdev)->can_queue > 0)
|
||||
atomic_dec(&scsi_target(sdev)->target_busy);
|
||||
|
@ -87,7 +87,7 @@ int scsi_noretry_cmd(struct scsi_cmnd *scmd);
|
||||
extern void scsi_add_cmd_to_list(struct scsi_cmnd *cmd);
|
||||
extern void scsi_del_cmd_from_list(struct scsi_cmnd *cmd);
|
||||
extern int scsi_maybe_unblock_host(struct scsi_device *sdev);
|
||||
extern void scsi_device_unbusy(struct scsi_device *sdev);
|
||||
extern void scsi_device_unbusy(struct scsi_device *sdev, struct scsi_cmnd *cmd);
|
||||
extern void scsi_queue_insert(struct scsi_cmnd *cmd, int reason);
|
||||
extern void scsi_io_completion(struct scsi_cmnd *, unsigned int);
|
||||
extern void scsi_run_host_queues(struct Scsi_Host *shost);
|
||||
|
@ -63,6 +63,7 @@ struct scsi_pointer {
|
||||
|
||||
/* for scmd->state */
|
||||
#define SCMD_STATE_COMPLETE 0
|
||||
#define SCMD_STATE_INFLIGHT 1
|
||||
|
||||
struct scsi_cmnd {
|
||||
struct scsi_request req;
|
||||
|
@ -345,7 +345,7 @@ struct scsi_host_template {
|
||||
/*
|
||||
* This determines if we will use a non-interrupt driven
|
||||
* or an interrupt driven scheme. It is set to the maximum number
|
||||
* of simultaneous commands a given host adapter will accept.
|
||||
* of simultaneous commands a single hw queue in HBA will accept.
|
||||
*/
|
||||
int can_queue;
|
||||
|
||||
@ -554,7 +554,6 @@ struct Scsi_Host {
|
||||
/* Area to keep a shared tag map */
|
||||
struct blk_mq_tag_set tag_set;
|
||||
|
||||
atomic_t host_busy; /* commands actually active on low-level */
|
||||
atomic_t host_blocked;
|
||||
|
||||
unsigned int host_failed; /* commands that failed.
|
||||
|
Loading…
Reference in New Issue
Block a user