scsi: ufs: core: mcq: Add completion support in poll
Complete CQE requests in poll. Assumption is that several poll completion may happen in different CPUs for the same completion queue. Hence a spin lock protection is added. Co-developed-by: Can Guo <quic_cang@quicinc.com> Signed-off-by: Can Guo <quic_cang@quicinc.com> Signed-off-by: Asutosh Das <quic_asutoshd@quicinc.com> Reviewed-by: Bart Van Assche <bvanassche@acm.org> Reviewed-by: Manivannan Sadhasivam <mani@kernel.org> Reviewed-by: Stanley Chu <stanley.chu@mediatek.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
This commit is contained in:
parent
f87b2c4182
commit
ed975065c3
4 changed files with 24 additions and 0 deletions
|
@ -294,6 +294,18 @@ unsigned long ufshcd_mcq_poll_cqe_nolock(struct ufs_hba *hba,
|
||||||
return completed_reqs;
|
return completed_reqs;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
unsigned long ufshcd_mcq_poll_cqe_lock(struct ufs_hba *hba,
|
||||||
|
struct ufs_hw_queue *hwq)
|
||||||
|
{
|
||||||
|
unsigned long completed_reqs;
|
||||||
|
|
||||||
|
spin_lock(&hwq->cq_lock);
|
||||||
|
completed_reqs = ufshcd_mcq_poll_cqe_nolock(hba, hwq);
|
||||||
|
spin_unlock(&hwq->cq_lock);
|
||||||
|
|
||||||
|
return completed_reqs;
|
||||||
|
}
|
||||||
|
|
||||||
void ufshcd_mcq_make_queues_operational(struct ufs_hba *hba)
|
void ufshcd_mcq_make_queues_operational(struct ufs_hba *hba)
|
||||||
{
|
{
|
||||||
struct ufs_hw_queue *hwq;
|
struct ufs_hw_queue *hwq;
|
||||||
|
@ -390,6 +402,7 @@ int ufshcd_mcq_init(struct ufs_hba *hba)
|
||||||
hwq = &hba->uhq[i];
|
hwq = &hba->uhq[i];
|
||||||
hwq->max_entries = hba->nutrs;
|
hwq->max_entries = hba->nutrs;
|
||||||
spin_lock_init(&hwq->sq_lock);
|
spin_lock_init(&hwq->sq_lock);
|
||||||
|
spin_lock_init(&hwq->cq_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* The very first HW queue serves device commands */
|
/* The very first HW queue serves device commands */
|
||||||
|
|
|
@ -75,6 +75,8 @@ unsigned long ufshcd_mcq_poll_cqe_nolock(struct ufs_hba *hba,
|
||||||
struct ufs_hw_queue *hwq);
|
struct ufs_hw_queue *hwq);
|
||||||
struct ufs_hw_queue *ufshcd_mcq_req_to_hwq(struct ufs_hba *hba,
|
struct ufs_hw_queue *ufshcd_mcq_req_to_hwq(struct ufs_hba *hba,
|
||||||
struct request *req);
|
struct request *req);
|
||||||
|
unsigned long ufshcd_mcq_poll_cqe_lock(struct ufs_hba *hba,
|
||||||
|
struct ufs_hw_queue *hwq);
|
||||||
|
|
||||||
#define UFSHCD_MCQ_IO_QUEUE_OFFSET 1
|
#define UFSHCD_MCQ_IO_QUEUE_OFFSET 1
|
||||||
#define SD_ASCII_STD true
|
#define SD_ASCII_STD true
|
||||||
|
|
|
@ -5461,6 +5461,13 @@ static int ufshcd_poll(struct Scsi_Host *shost, unsigned int queue_num)
|
||||||
struct ufs_hba *hba = shost_priv(shost);
|
struct ufs_hba *hba = shost_priv(shost);
|
||||||
unsigned long completed_reqs, flags;
|
unsigned long completed_reqs, flags;
|
||||||
u32 tr_doorbell;
|
u32 tr_doorbell;
|
||||||
|
struct ufs_hw_queue *hwq;
|
||||||
|
|
||||||
|
if (is_mcq_enabled(hba)) {
|
||||||
|
hwq = &hba->uhq[queue_num + UFSHCD_MCQ_IO_QUEUE_OFFSET];
|
||||||
|
|
||||||
|
return ufshcd_mcq_poll_cqe_lock(hba, hwq);
|
||||||
|
}
|
||||||
|
|
||||||
spin_lock_irqsave(&hba->outstanding_lock, flags);
|
spin_lock_irqsave(&hba->outstanding_lock, flags);
|
||||||
tr_doorbell = ufshcd_readl(hba, REG_UTP_TRANSFER_REQ_DOOR_BELL);
|
tr_doorbell = ufshcd_readl(hba, REG_UTP_TRANSFER_REQ_DOOR_BELL);
|
||||||
|
|
|
@ -1086,6 +1086,7 @@ struct ufs_hba {
|
||||||
* @sq_lock: serialize submission queue access
|
* @sq_lock: serialize submission queue access
|
||||||
* @cq_tail_slot: current slot to which CQ tail pointer is pointing
|
* @cq_tail_slot: current slot to which CQ tail pointer is pointing
|
||||||
* @cq_head_slot: current slot to which CQ head pointer is pointing
|
* @cq_head_slot: current slot to which CQ head pointer is pointing
|
||||||
|
* @cq_lock: Synchronize between multiple polling instances
|
||||||
*/
|
*/
|
||||||
struct ufs_hw_queue {
|
struct ufs_hw_queue {
|
||||||
void __iomem *mcq_sq_head;
|
void __iomem *mcq_sq_head;
|
||||||
|
@ -1103,6 +1104,7 @@ struct ufs_hw_queue {
|
||||||
spinlock_t sq_lock;
|
spinlock_t sq_lock;
|
||||||
u32 cq_tail_slot;
|
u32 cq_tail_slot;
|
||||||
u32 cq_head_slot;
|
u32 cq_head_slot;
|
||||||
|
spinlock_t cq_lock;
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline bool is_mcq_enabled(struct ufs_hba *hba)
|
static inline bool is_mcq_enabled(struct ufs_hba *hba)
|
||||||
|
|
Loading…
Add table
Reference in a new issue