The driver serializes ioctls through a mutex lock but access to the ioctl data buffer is not guarded by the mutex. This results in multiple user threads being able to write to the driver's ioctl buffer simultaneously. Protect the ioctl buffer with the ioctl mutex. Signed-off-by: Sumit Saxena <sumit.saxena@broadcom.com> Signed-off-by: Ranjan Kumar <ranjan.kumar@broadcom.com> Link: https://lore.kernel.org/r/20241110194405.10108-2-ranjan.kumar@broadcom.com Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
3345 lines
97 KiB
C
3345 lines
97 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Driver for Broadcom MPI3 Storage Controllers
|
|
*
|
|
* Copyright (C) 2017-2023 Broadcom Inc.
|
|
* (mailto: mpi3mr-linuxdrv.pdl@broadcom.com)
|
|
*
|
|
*/
|
|
|
|
#include "mpi3mr.h"
|
|
#include <linux/bsg-lib.h>
|
|
#include <uapi/scsi/scsi_bsg_mpi3mr.h>
|
|
|
|
/**
|
|
* mpi3mr_alloc_trace_buffer: Allocate trace buffer
|
|
* @mrioc: Adapter instance reference
|
|
* @trace_size: Trace buffer size
|
|
*
|
|
* Allocate trace buffer
|
|
* Return: 0 on success, non-zero on failure.
|
|
*/
|
|
static int mpi3mr_alloc_trace_buffer(struct mpi3mr_ioc *mrioc, u32 trace_size)
|
|
{
|
|
struct diag_buffer_desc *diag_buffer = &mrioc->diag_buffers[0];
|
|
|
|
diag_buffer->addr = dma_alloc_coherent(&mrioc->pdev->dev,
|
|
trace_size, &diag_buffer->dma_addr, GFP_KERNEL);
|
|
if (diag_buffer->addr) {
|
|
dprint_init(mrioc, "trace diag buffer is allocated successfully\n");
|
|
return 0;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_alloc_diag_bufs - Allocate memory for diag buffers
|
|
* @mrioc: Adapter instance reference
|
|
*
|
|
* This functions checks whether the driver defined buffer sizes
|
|
* are greater than IOCFacts provided controller local buffer
|
|
* sizes and if the driver defined sizes are more then the
|
|
* driver allocates the specific buffer by reading driver page1
|
|
*
|
|
* Return: Nothing.
|
|
*/
|
|
void mpi3mr_alloc_diag_bufs(struct mpi3mr_ioc *mrioc)
|
|
{
|
|
struct diag_buffer_desc *diag_buffer;
|
|
struct mpi3_driver_page1 driver_pg1;
|
|
u32 trace_dec_size, trace_min_size, fw_dec_size, fw_min_size,
|
|
trace_size, fw_size;
|
|
u16 pg_sz = sizeof(driver_pg1);
|
|
int retval = 0;
|
|
bool retry = false;
|
|
|
|
if (mrioc->diag_buffers[0].addr || mrioc->diag_buffers[1].addr)
|
|
return;
|
|
|
|
retval = mpi3mr_cfg_get_driver_pg1(mrioc, &driver_pg1, pg_sz);
|
|
if (retval) {
|
|
ioc_warn(mrioc,
|
|
"%s: driver page 1 read failed, allocating trace\n"
|
|
"and firmware diag buffers of default size\n", __func__);
|
|
trace_size = fw_size = MPI3MR_DEFAULT_HDB_MAX_SZ;
|
|
trace_dec_size = fw_dec_size = MPI3MR_DEFAULT_HDB_DEC_SZ;
|
|
trace_min_size = fw_min_size = MPI3MR_DEFAULT_HDB_MIN_SZ;
|
|
|
|
} else {
|
|
trace_size = driver_pg1.host_diag_trace_max_size * 1024;
|
|
trace_dec_size = driver_pg1.host_diag_trace_decrement_size
|
|
* 1024;
|
|
trace_min_size = driver_pg1.host_diag_trace_min_size * 1024;
|
|
fw_size = driver_pg1.host_diag_fw_max_size * 1024;
|
|
fw_dec_size = driver_pg1.host_diag_fw_decrement_size * 1024;
|
|
fw_min_size = driver_pg1.host_diag_fw_min_size * 1024;
|
|
dprint_init(mrioc,
|
|
"%s:trace diag buffer sizes read from driver\n"
|
|
"page1: maximum size = %dKB, decrement size = %dKB\n"
|
|
", minimum size = %dKB\n", __func__, driver_pg1.host_diag_trace_max_size,
|
|
driver_pg1.host_diag_trace_decrement_size,
|
|
driver_pg1.host_diag_trace_min_size);
|
|
dprint_init(mrioc,
|
|
"%s:firmware diag buffer sizes read from driver\n"
|
|
"page1: maximum size = %dKB, decrement size = %dKB\n"
|
|
", minimum size = %dKB\n", __func__, driver_pg1.host_diag_fw_max_size,
|
|
driver_pg1.host_diag_fw_decrement_size,
|
|
driver_pg1.host_diag_fw_min_size);
|
|
if ((trace_size == 0) && (fw_size == 0))
|
|
return;
|
|
}
|
|
|
|
|
|
retry_trace:
|
|
diag_buffer = &mrioc->diag_buffers[0];
|
|
diag_buffer->type = MPI3_DIAG_BUFFER_TYPE_TRACE;
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_NOT_ALLOCATED;
|
|
if ((mrioc->facts.diag_trace_sz < trace_size) && (trace_size >=
|
|
trace_min_size)) {
|
|
if (!retry)
|
|
dprint_init(mrioc,
|
|
"trying to allocate trace diag buffer of size = %dKB\n",
|
|
trace_size / 1024);
|
|
if (get_order(trace_size) > MAX_PAGE_ORDER ||
|
|
mpi3mr_alloc_trace_buffer(mrioc, trace_size)) {
|
|
retry = true;
|
|
trace_size -= trace_dec_size;
|
|
dprint_init(mrioc, "trace diag buffer allocation failed\n"
|
|
"retrying smaller size %dKB\n", trace_size / 1024);
|
|
goto retry_trace;
|
|
} else
|
|
diag_buffer->size = trace_size;
|
|
}
|
|
|
|
retry = false;
|
|
retry_fw:
|
|
|
|
diag_buffer = &mrioc->diag_buffers[1];
|
|
|
|
diag_buffer->type = MPI3_DIAG_BUFFER_TYPE_FW;
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_NOT_ALLOCATED;
|
|
if ((mrioc->facts.diag_fw_sz < fw_size) && (fw_size >= fw_min_size)) {
|
|
if (get_order(fw_size) <= MAX_PAGE_ORDER) {
|
|
diag_buffer->addr
|
|
= dma_alloc_coherent(&mrioc->pdev->dev, fw_size,
|
|
&diag_buffer->dma_addr,
|
|
GFP_KERNEL);
|
|
}
|
|
if (!retry)
|
|
dprint_init(mrioc,
|
|
"%s:trying to allocate firmware diag buffer of size = %dKB\n",
|
|
__func__, fw_size / 1024);
|
|
if (diag_buffer->addr) {
|
|
dprint_init(mrioc, "%s:firmware diag buffer allocated successfully\n",
|
|
__func__);
|
|
diag_buffer->size = fw_size;
|
|
} else {
|
|
retry = true;
|
|
fw_size -= fw_dec_size;
|
|
dprint_init(mrioc, "%s:trace diag buffer allocation failed,\n"
|
|
"retrying smaller size %dKB\n",
|
|
__func__, fw_size / 1024);
|
|
goto retry_fw;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_issue_diag_buf_post - Send diag buffer post req
|
|
* @mrioc: Adapter instance reference
|
|
* @diag_buffer: Diagnostic buffer descriptor
|
|
*
|
|
* Issue diagnostic buffer post MPI request through admin queue
|
|
* and wait for the completion of it or time out.
|
|
*
|
|
* Return: 0 on success, non-zero on failures.
|
|
*/
|
|
int mpi3mr_issue_diag_buf_post(struct mpi3mr_ioc *mrioc,
|
|
struct diag_buffer_desc *diag_buffer)
|
|
{
|
|
struct mpi3_diag_buffer_post_request diag_buf_post_req;
|
|
u8 prev_status;
|
|
int retval = 0;
|
|
|
|
memset(&diag_buf_post_req, 0, sizeof(diag_buf_post_req));
|
|
mutex_lock(&mrioc->init_cmds.mutex);
|
|
if (mrioc->init_cmds.state & MPI3MR_CMD_PENDING) {
|
|
dprint_bsg_err(mrioc, "%s: command is in use\n", __func__);
|
|
mutex_unlock(&mrioc->init_cmds.mutex);
|
|
return -1;
|
|
}
|
|
mrioc->init_cmds.state = MPI3MR_CMD_PENDING;
|
|
mrioc->init_cmds.is_waiting = 1;
|
|
mrioc->init_cmds.callback = NULL;
|
|
diag_buf_post_req.host_tag = cpu_to_le16(MPI3MR_HOSTTAG_INITCMDS);
|
|
diag_buf_post_req.function = MPI3_FUNCTION_DIAG_BUFFER_POST;
|
|
diag_buf_post_req.type = diag_buffer->type;
|
|
diag_buf_post_req.address = le64_to_cpu(diag_buffer->dma_addr);
|
|
diag_buf_post_req.length = le32_to_cpu(diag_buffer->size);
|
|
|
|
dprint_bsg_info(mrioc, "%s: posting diag buffer type %d\n", __func__,
|
|
diag_buffer->type);
|
|
prev_status = diag_buffer->status;
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED;
|
|
init_completion(&mrioc->init_cmds.done);
|
|
retval = mpi3mr_admin_request_post(mrioc, &diag_buf_post_req,
|
|
sizeof(diag_buf_post_req), 1);
|
|
if (retval) {
|
|
dprint_bsg_err(mrioc, "%s: admin request post failed\n",
|
|
__func__);
|
|
goto out_unlock;
|
|
}
|
|
wait_for_completion_timeout(&mrioc->init_cmds.done,
|
|
(MPI3MR_INTADMCMD_TIMEOUT * HZ));
|
|
if (!(mrioc->init_cmds.state & MPI3MR_CMD_COMPLETE)) {
|
|
mrioc->init_cmds.is_waiting = 0;
|
|
dprint_bsg_err(mrioc, "%s: command timedout\n", __func__);
|
|
mpi3mr_check_rh_fault_ioc(mrioc,
|
|
MPI3MR_RESET_FROM_DIAG_BUFFER_POST_TIMEOUT);
|
|
retval = -1;
|
|
goto out_unlock;
|
|
}
|
|
if ((mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
|
|
!= MPI3_IOCSTATUS_SUCCESS) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: command failed, buffer_type (%d) ioc_status(0x%04x) log_info(0x%08x)\n",
|
|
__func__, diag_buffer->type,
|
|
(mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK),
|
|
mrioc->init_cmds.ioc_loginfo);
|
|
retval = -1;
|
|
goto out_unlock;
|
|
}
|
|
dprint_bsg_info(mrioc, "%s: diag buffer type %d posted successfully\n",
|
|
__func__, diag_buffer->type);
|
|
|
|
out_unlock:
|
|
if (retval)
|
|
diag_buffer->status = prev_status;
|
|
mrioc->init_cmds.state = MPI3MR_CMD_NOTUSED;
|
|
mutex_unlock(&mrioc->init_cmds.mutex);
|
|
return retval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_post_diag_bufs - Post diag buffers to the controller
|
|
* @mrioc: Adapter instance reference
|
|
*
|
|
* This function calls helper function to post both trace and
|
|
* firmware buffers to the controller.
|
|
*
|
|
* Return: None
|
|
*/
|
|
int mpi3mr_post_diag_bufs(struct mpi3mr_ioc *mrioc)
|
|
{
|
|
u8 i;
|
|
struct diag_buffer_desc *diag_buffer;
|
|
|
|
for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
|
|
diag_buffer = &mrioc->diag_buffers[i];
|
|
if (!(diag_buffer->addr))
|
|
continue;
|
|
if (mpi3mr_issue_diag_buf_post(mrioc, diag_buffer))
|
|
return -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_issue_diag_buf_release - Send diag buffer release req
|
|
* @mrioc: Adapter instance reference
|
|
* @diag_buffer: Diagnostic buffer descriptor
|
|
*
|
|
* Issue diagnostic buffer manage MPI request with release
|
|
* action request through admin queue and wait for the
|
|
* completion of it or time out.
|
|
*
|
|
* Return: 0 on success, non-zero on failures.
|
|
*/
|
|
int mpi3mr_issue_diag_buf_release(struct mpi3mr_ioc *mrioc,
|
|
struct diag_buffer_desc *diag_buffer)
|
|
{
|
|
struct mpi3_diag_buffer_manage_request diag_buf_manage_req;
|
|
int retval = 0;
|
|
|
|
if ((diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
|
|
(diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
|
|
return retval;
|
|
|
|
memset(&diag_buf_manage_req, 0, sizeof(diag_buf_manage_req));
|
|
mutex_lock(&mrioc->init_cmds.mutex);
|
|
if (mrioc->init_cmds.state & MPI3MR_CMD_PENDING) {
|
|
dprint_reset(mrioc, "%s: command is in use\n", __func__);
|
|
mutex_unlock(&mrioc->init_cmds.mutex);
|
|
return -1;
|
|
}
|
|
mrioc->init_cmds.state = MPI3MR_CMD_PENDING;
|
|
mrioc->init_cmds.is_waiting = 1;
|
|
mrioc->init_cmds.callback = NULL;
|
|
diag_buf_manage_req.host_tag = cpu_to_le16(MPI3MR_HOSTTAG_INITCMDS);
|
|
diag_buf_manage_req.function = MPI3_FUNCTION_DIAG_BUFFER_MANAGE;
|
|
diag_buf_manage_req.type = diag_buffer->type;
|
|
diag_buf_manage_req.action = MPI3_DIAG_BUFFER_ACTION_RELEASE;
|
|
|
|
|
|
dprint_reset(mrioc, "%s: releasing diag buffer type %d\n", __func__,
|
|
diag_buffer->type);
|
|
init_completion(&mrioc->init_cmds.done);
|
|
retval = mpi3mr_admin_request_post(mrioc, &diag_buf_manage_req,
|
|
sizeof(diag_buf_manage_req), 1);
|
|
if (retval) {
|
|
dprint_reset(mrioc, "%s: admin request post failed\n", __func__);
|
|
mpi3mr_set_trigger_data_in_hdb(diag_buffer,
|
|
MPI3MR_HDB_TRIGGER_TYPE_UNKNOWN, NULL, 1);
|
|
goto out_unlock;
|
|
}
|
|
wait_for_completion_timeout(&mrioc->init_cmds.done,
|
|
(MPI3MR_INTADMCMD_TIMEOUT * HZ));
|
|
if (!(mrioc->init_cmds.state & MPI3MR_CMD_COMPLETE)) {
|
|
mrioc->init_cmds.is_waiting = 0;
|
|
dprint_reset(mrioc, "%s: command timedout\n", __func__);
|
|
mpi3mr_check_rh_fault_ioc(mrioc,
|
|
MPI3MR_RESET_FROM_DIAG_BUFFER_RELEASE_TIMEOUT);
|
|
retval = -1;
|
|
goto out_unlock;
|
|
}
|
|
if ((mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
|
|
!= MPI3_IOCSTATUS_SUCCESS) {
|
|
dprint_reset(mrioc,
|
|
"%s: command failed, buffer_type (%d) ioc_status(0x%04x) log_info(0x%08x)\n",
|
|
__func__, diag_buffer->type,
|
|
(mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK),
|
|
mrioc->init_cmds.ioc_loginfo);
|
|
retval = -1;
|
|
goto out_unlock;
|
|
}
|
|
dprint_reset(mrioc, "%s: diag buffer type %d released successfully\n",
|
|
__func__, diag_buffer->type);
|
|
|
|
out_unlock:
|
|
mrioc->init_cmds.state = MPI3MR_CMD_NOTUSED;
|
|
mutex_unlock(&mrioc->init_cmds.mutex);
|
|
return retval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_process_trigger - Generic HDB Trigger handler
|
|
* @mrioc: Adapter instance reference
|
|
* @trigger_type: Trigger type
|
|
* @trigger_data: Trigger data
|
|
* @trigger_flags: Trigger flags
|
|
*
|
|
* This function checks validity of HDB, triggers and based on
|
|
* trigger information, creates an event to be processed in the
|
|
* firmware event worker thread .
|
|
*
|
|
* This function should be called with trigger spinlock held
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
static void mpi3mr_process_trigger(struct mpi3mr_ioc *mrioc, u8 trigger_type,
|
|
union mpi3mr_trigger_data *trigger_data, u8 trigger_flags)
|
|
{
|
|
struct trigger_event_data event_data;
|
|
struct diag_buffer_desc *trace_hdb = NULL;
|
|
struct diag_buffer_desc *fw_hdb = NULL;
|
|
u64 global_trigger;
|
|
|
|
trace_hdb = mpi3mr_diag_buffer_for_type(mrioc,
|
|
MPI3_DIAG_BUFFER_TYPE_TRACE);
|
|
if (trace_hdb &&
|
|
(trace_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
|
|
(trace_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
|
|
trace_hdb = NULL;
|
|
|
|
fw_hdb = mpi3mr_diag_buffer_for_type(mrioc, MPI3_DIAG_BUFFER_TYPE_FW);
|
|
|
|
if (fw_hdb &&
|
|
(fw_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
|
|
(fw_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
|
|
fw_hdb = NULL;
|
|
|
|
if (mrioc->snapdump_trigger_active || (mrioc->fw_release_trigger_active
|
|
&& mrioc->trace_release_trigger_active) ||
|
|
(!trace_hdb && !fw_hdb) || (!mrioc->driver_pg2) ||
|
|
((trigger_type == MPI3MR_HDB_TRIGGER_TYPE_ELEMENT)
|
|
&& (!mrioc->driver_pg2->num_triggers)))
|
|
return;
|
|
|
|
memset(&event_data, 0, sizeof(event_data));
|
|
event_data.trigger_type = trigger_type;
|
|
memcpy(&event_data.trigger_specific_data, trigger_data,
|
|
sizeof(*trigger_data));
|
|
global_trigger = le64_to_cpu(mrioc->driver_pg2->global_trigger);
|
|
|
|
if (global_trigger & MPI3_DRIVER2_GLOBALTRIGGER_SNAPDUMP_ENABLED) {
|
|
event_data.snapdump = true;
|
|
event_data.trace_hdb = trace_hdb;
|
|
event_data.fw_hdb = fw_hdb;
|
|
mrioc->snapdump_trigger_active = true;
|
|
} else if (trigger_type == MPI3MR_HDB_TRIGGER_TYPE_GLOBAL) {
|
|
if ((trace_hdb) && (global_trigger &
|
|
MPI3_DRIVER2_GLOBALTRIGGER_DIAG_TRACE_RELEASE) &&
|
|
(!mrioc->trace_release_trigger_active)) {
|
|
event_data.trace_hdb = trace_hdb;
|
|
mrioc->trace_release_trigger_active = true;
|
|
}
|
|
if ((fw_hdb) && (global_trigger &
|
|
MPI3_DRIVER2_GLOBALTRIGGER_DIAG_FW_RELEASE) &&
|
|
(!mrioc->fw_release_trigger_active)) {
|
|
event_data.fw_hdb = fw_hdb;
|
|
mrioc->fw_release_trigger_active = true;
|
|
}
|
|
} else if (trigger_type == MPI3MR_HDB_TRIGGER_TYPE_ELEMENT) {
|
|
if ((trace_hdb) && (trigger_flags &
|
|
MPI3_DRIVER2_TRIGGER_FLAGS_DIAG_TRACE_RELEASE) &&
|
|
(!mrioc->trace_release_trigger_active)) {
|
|
event_data.trace_hdb = trace_hdb;
|
|
mrioc->trace_release_trigger_active = true;
|
|
}
|
|
if ((fw_hdb) && (trigger_flags &
|
|
MPI3_DRIVER2_TRIGGER_FLAGS_DIAG_FW_RELEASE) &&
|
|
(!mrioc->fw_release_trigger_active)) {
|
|
event_data.fw_hdb = fw_hdb;
|
|
mrioc->fw_release_trigger_active = true;
|
|
}
|
|
}
|
|
|
|
if (event_data.trace_hdb || event_data.fw_hdb)
|
|
mpi3mr_hdb_trigger_data_event(mrioc, &event_data);
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_global_trigger - Global HDB trigger handler
|
|
* @mrioc: Adapter instance reference
|
|
* @trigger_data: Trigger data
|
|
*
|
|
* This function checks whether the given global trigger is
|
|
* enabled in the driver page 2 and if so calls generic trigger
|
|
* handler to queue event for HDB release.
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_global_trigger(struct mpi3mr_ioc *mrioc, u64 trigger_data)
|
|
{
|
|
unsigned long flags;
|
|
union mpi3mr_trigger_data trigger_specific_data;
|
|
|
|
spin_lock_irqsave(&mrioc->trigger_lock, flags);
|
|
if (le64_to_cpu(mrioc->driver_pg2->global_trigger) & trigger_data) {
|
|
memset(&trigger_specific_data, 0,
|
|
sizeof(trigger_specific_data));
|
|
trigger_specific_data.global = trigger_data;
|
|
mpi3mr_process_trigger(mrioc, MPI3MR_HDB_TRIGGER_TYPE_GLOBAL,
|
|
&trigger_specific_data, 0);
|
|
}
|
|
spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_scsisense_trigger - SCSI sense HDB trigger handler
|
|
* @mrioc: Adapter instance reference
|
|
* @sensekey: Sense Key
|
|
* @asc: Additional Sense Code
|
|
* @ascq: Additional Sense Code Qualifier
|
|
*
|
|
* This function compares SCSI sense trigger values with driver
|
|
* page 2 values and calls generic trigger handler to release
|
|
* HDBs if match found
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_scsisense_trigger(struct mpi3mr_ioc *mrioc, u8 sensekey, u8 asc,
|
|
u8 ascq)
|
|
{
|
|
struct mpi3_driver2_trigger_scsi_sense *scsi_sense_trigger = NULL;
|
|
u64 i = 0;
|
|
unsigned long flags;
|
|
u8 num_triggers, trigger_flags;
|
|
|
|
if (mrioc->scsisense_trigger_present) {
|
|
spin_lock_irqsave(&mrioc->trigger_lock, flags);
|
|
scsi_sense_trigger = (struct mpi3_driver2_trigger_scsi_sense *)
|
|
mrioc->driver_pg2->trigger;
|
|
num_triggers = mrioc->driver_pg2->num_triggers;
|
|
for (i = 0; i < num_triggers; i++, scsi_sense_trigger++) {
|
|
if (scsi_sense_trigger->type !=
|
|
MPI3_DRIVER2_TRIGGER_TYPE_SCSI_SENSE)
|
|
continue;
|
|
if (!(scsi_sense_trigger->sense_key ==
|
|
MPI3_DRIVER2_TRIGGER_SCSI_SENSE_SENSE_KEY_MATCH_ALL
|
|
|| scsi_sense_trigger->sense_key == sensekey))
|
|
continue;
|
|
if (!(scsi_sense_trigger->asc ==
|
|
MPI3_DRIVER2_TRIGGER_SCSI_SENSE_ASC_MATCH_ALL ||
|
|
scsi_sense_trigger->asc == asc))
|
|
continue;
|
|
if (!(scsi_sense_trigger->ascq ==
|
|
MPI3_DRIVER2_TRIGGER_SCSI_SENSE_ASCQ_MATCH_ALL ||
|
|
scsi_sense_trigger->ascq == ascq))
|
|
continue;
|
|
trigger_flags = scsi_sense_trigger->flags;
|
|
mpi3mr_process_trigger(mrioc,
|
|
MPI3MR_HDB_TRIGGER_TYPE_ELEMENT,
|
|
(union mpi3mr_trigger_data *)scsi_sense_trigger,
|
|
trigger_flags);
|
|
break;
|
|
}
|
|
spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_event_trigger - MPI event HDB trigger handler
|
|
* @mrioc: Adapter instance reference
|
|
* @event: MPI Event
|
|
*
|
|
* This function compares event trigger values with driver page
|
|
* 2 values and calls generic trigger handler to release
|
|
* HDBs if match found.
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_event_trigger(struct mpi3mr_ioc *mrioc, u8 event)
|
|
{
|
|
struct mpi3_driver2_trigger_event *event_trigger = NULL;
|
|
u64 i = 0;
|
|
unsigned long flags;
|
|
u8 num_triggers, trigger_flags;
|
|
|
|
if (mrioc->event_trigger_present) {
|
|
spin_lock_irqsave(&mrioc->trigger_lock, flags);
|
|
event_trigger = (struct mpi3_driver2_trigger_event *)
|
|
mrioc->driver_pg2->trigger;
|
|
num_triggers = mrioc->driver_pg2->num_triggers;
|
|
|
|
for (i = 0; i < num_triggers; i++, event_trigger++) {
|
|
if (event_trigger->type !=
|
|
MPI3_DRIVER2_TRIGGER_TYPE_EVENT)
|
|
continue;
|
|
if (event_trigger->event != event)
|
|
continue;
|
|
trigger_flags = event_trigger->flags;
|
|
mpi3mr_process_trigger(mrioc,
|
|
MPI3MR_HDB_TRIGGER_TYPE_ELEMENT,
|
|
(union mpi3mr_trigger_data *)event_trigger,
|
|
trigger_flags);
|
|
break;
|
|
}
|
|
spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_reply_trigger - MPI Reply HDB trigger handler
|
|
* @mrioc: Adapter instance reference
|
|
* @ioc_status: Masked value of IOC Status from MPI Reply
|
|
* @ioc_loginfo: IOC Log Info from MPI Reply
|
|
*
|
|
* This function compares IOC status and IOC log info trigger
|
|
* values with driver page 2 values and calls generic trigger
|
|
* handler to release HDBs if match found.
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_reply_trigger(struct mpi3mr_ioc *mrioc, u16 ioc_status,
|
|
u32 ioc_loginfo)
|
|
{
|
|
struct mpi3_driver2_trigger_reply *reply_trigger = NULL;
|
|
u64 i = 0;
|
|
unsigned long flags;
|
|
u8 num_triggers, trigger_flags;
|
|
|
|
if (mrioc->reply_trigger_present) {
|
|
spin_lock_irqsave(&mrioc->trigger_lock, flags);
|
|
reply_trigger = (struct mpi3_driver2_trigger_reply *)
|
|
mrioc->driver_pg2->trigger;
|
|
num_triggers = mrioc->driver_pg2->num_triggers;
|
|
for (i = 0; i < num_triggers; i++, reply_trigger++) {
|
|
if (reply_trigger->type !=
|
|
MPI3_DRIVER2_TRIGGER_TYPE_REPLY)
|
|
continue;
|
|
if ((le16_to_cpu(reply_trigger->ioc_status) !=
|
|
ioc_status)
|
|
&& (le16_to_cpu(reply_trigger->ioc_status) !=
|
|
MPI3_DRIVER2_TRIGGER_REPLY_IOCSTATUS_MATCH_ALL))
|
|
continue;
|
|
if ((le32_to_cpu(reply_trigger->ioc_log_info) !=
|
|
(le32_to_cpu(reply_trigger->ioc_log_info_mask) &
|
|
ioc_loginfo)))
|
|
continue;
|
|
trigger_flags = reply_trigger->flags;
|
|
mpi3mr_process_trigger(mrioc,
|
|
MPI3MR_HDB_TRIGGER_TYPE_ELEMENT,
|
|
(union mpi3mr_trigger_data *)reply_trigger,
|
|
trigger_flags);
|
|
break;
|
|
}
|
|
spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_get_num_trigger - Gets number of HDB triggers
|
|
* @mrioc: Adapter instance reference
|
|
* @num_triggers: Number of triggers
|
|
* @page_action: Page action
|
|
*
|
|
* This function reads number of triggers by reading driver page
|
|
* 2
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static int mpi3mr_get_num_trigger(struct mpi3mr_ioc *mrioc, u8 *num_triggers,
|
|
u8 page_action)
|
|
{
|
|
struct mpi3_driver_page2 drvr_page2;
|
|
int retval = 0;
|
|
|
|
*num_triggers = 0;
|
|
|
|
retval = mpi3mr_cfg_get_driver_pg2(mrioc, &drvr_page2,
|
|
sizeof(struct mpi3_driver_page2), page_action);
|
|
|
|
if (retval) {
|
|
dprint_init(mrioc, "%s: driver page 2 read failed\n", __func__);
|
|
return retval;
|
|
}
|
|
*num_triggers = drvr_page2.num_triggers;
|
|
return retval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_refresh_trigger - Handler for Refresh trigger BSG
|
|
* @mrioc: Adapter instance reference
|
|
* @page_action: Page action
|
|
*
|
|
* This function caches the driver page 2 in the driver's memory
|
|
* by reading driver page 2 from the controller for a given page
|
|
* type and updates the HDB trigger values
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
int mpi3mr_refresh_trigger(struct mpi3mr_ioc *mrioc, u8 page_action)
|
|
{
|
|
u16 pg_sz = sizeof(struct mpi3_driver_page2);
|
|
struct mpi3_driver_page2 *drvr_page2 = NULL;
|
|
u8 trigger_type, num_triggers;
|
|
int retval;
|
|
int i = 0;
|
|
unsigned long flags;
|
|
|
|
retval = mpi3mr_get_num_trigger(mrioc, &num_triggers, page_action);
|
|
|
|
if (retval)
|
|
goto out;
|
|
|
|
pg_sz = offsetof(struct mpi3_driver_page2, trigger) +
|
|
(num_triggers * sizeof(union mpi3_driver2_trigger_element));
|
|
drvr_page2 = kzalloc(pg_sz, GFP_KERNEL);
|
|
if (!drvr_page2) {
|
|
retval = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
retval = mpi3mr_cfg_get_driver_pg2(mrioc, drvr_page2, pg_sz, page_action);
|
|
if (retval) {
|
|
dprint_init(mrioc, "%s: driver page 2 read failed\n", __func__);
|
|
kfree(drvr_page2);
|
|
goto out;
|
|
}
|
|
spin_lock_irqsave(&mrioc->trigger_lock, flags);
|
|
kfree(mrioc->driver_pg2);
|
|
mrioc->driver_pg2 = drvr_page2;
|
|
mrioc->reply_trigger_present = false;
|
|
mrioc->event_trigger_present = false;
|
|
mrioc->scsisense_trigger_present = false;
|
|
|
|
for (i = 0; (i < mrioc->driver_pg2->num_triggers); i++) {
|
|
trigger_type = mrioc->driver_pg2->trigger[i].event.type;
|
|
switch (trigger_type) {
|
|
case MPI3_DRIVER2_TRIGGER_TYPE_REPLY:
|
|
mrioc->reply_trigger_present = true;
|
|
break;
|
|
case MPI3_DRIVER2_TRIGGER_TYPE_EVENT:
|
|
mrioc->event_trigger_present = true;
|
|
break;
|
|
case MPI3_DRIVER2_TRIGGER_TYPE_SCSI_SENSE:
|
|
mrioc->scsisense_trigger_present = true;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
|
|
out:
|
|
return retval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_release_diag_bufs - Release diag buffers
|
|
* @mrioc: Adapter instance reference
|
|
* @skip_rel_action: Skip release action and set buffer state
|
|
*
|
|
* This function calls helper function to release both trace and
|
|
* firmware buffers from the controller.
|
|
*
|
|
* Return: None
|
|
*/
|
|
void mpi3mr_release_diag_bufs(struct mpi3mr_ioc *mrioc, u8 skip_rel_action)
|
|
{
|
|
u8 i;
|
|
struct diag_buffer_desc *diag_buffer;
|
|
|
|
for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
|
|
diag_buffer = &mrioc->diag_buffers[i];
|
|
if (!(diag_buffer->addr))
|
|
continue;
|
|
if (diag_buffer->status == MPI3MR_HDB_BUFSTATUS_RELEASED)
|
|
continue;
|
|
if (!skip_rel_action)
|
|
mpi3mr_issue_diag_buf_release(mrioc, diag_buffer);
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_RELEASED;
|
|
atomic64_inc(&event_counter);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_set_trigger_data_in_hdb - Updates HDB trigger type and
|
|
* trigger data
|
|
*
|
|
* @hdb: HDB pointer
|
|
* @type: Trigger type
|
|
* @data: Trigger data
|
|
* @force: Trigger overwrite flag
|
|
* @trigger_data: Pointer to trigger data information
|
|
*
|
|
* Updates trigger type and trigger data based on parameter
|
|
* passed to this function
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_set_trigger_data_in_hdb(struct diag_buffer_desc *hdb,
|
|
u8 type, union mpi3mr_trigger_data *trigger_data, bool force)
|
|
{
|
|
if ((!force) && (hdb->trigger_type != MPI3MR_HDB_TRIGGER_TYPE_UNKNOWN))
|
|
return;
|
|
hdb->trigger_type = type;
|
|
if (!trigger_data)
|
|
memset(&hdb->trigger_data, 0, sizeof(*trigger_data));
|
|
else
|
|
memcpy(&hdb->trigger_data, trigger_data, sizeof(*trigger_data));
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_set_trigger_data_in_all_hdb - Updates HDB trigger type
|
|
* and trigger data for all HDB
|
|
*
|
|
* @mrioc: Adapter instance reference
|
|
* @type: Trigger type
|
|
* @data: Trigger data
|
|
* @force: Trigger overwrite flag
|
|
* @trigger_data: Pointer to trigger data information
|
|
*
|
|
* Updates trigger type and trigger data based on parameter
|
|
* passed to this function
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_set_trigger_data_in_all_hdb(struct mpi3mr_ioc *mrioc,
|
|
u8 type, union mpi3mr_trigger_data *trigger_data, bool force)
|
|
{
|
|
struct diag_buffer_desc *hdb = NULL;
|
|
|
|
hdb = mpi3mr_diag_buffer_for_type(mrioc, MPI3_DIAG_BUFFER_TYPE_TRACE);
|
|
if (hdb)
|
|
mpi3mr_set_trigger_data_in_hdb(hdb, type, trigger_data, force);
|
|
hdb = mpi3mr_diag_buffer_for_type(mrioc, MPI3_DIAG_BUFFER_TYPE_FW);
|
|
if (hdb)
|
|
mpi3mr_set_trigger_data_in_hdb(hdb, type, trigger_data, force);
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_hdbstatuschg_evt_th - HDB status change evt tophalf
|
|
* @mrioc: Adapter instance reference
|
|
* @event_reply: event data
|
|
*
|
|
* Modifies the status of the applicable diag buffer descriptors
|
|
*
|
|
* Return: Nothing
|
|
*/
|
|
void mpi3mr_hdbstatuschg_evt_th(struct mpi3mr_ioc *mrioc,
|
|
struct mpi3_event_notification_reply *event_reply)
|
|
{
|
|
struct mpi3_event_data_diag_buffer_status_change *evtdata;
|
|
struct diag_buffer_desc *diag_buffer;
|
|
|
|
evtdata = (struct mpi3_event_data_diag_buffer_status_change *)
|
|
event_reply->event_data;
|
|
|
|
diag_buffer = mpi3mr_diag_buffer_for_type(mrioc, evtdata->type);
|
|
if (!diag_buffer)
|
|
return;
|
|
if ((diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
|
|
(diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
|
|
return;
|
|
switch (evtdata->reason_code) {
|
|
case MPI3_EVENT_DIAG_BUFFER_STATUS_CHANGE_RC_RELEASED:
|
|
{
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_RELEASED;
|
|
mpi3mr_set_trigger_data_in_hdb(diag_buffer,
|
|
MPI3MR_HDB_TRIGGER_TYPE_FW_RELEASED, NULL, 0);
|
|
atomic64_inc(&event_counter);
|
|
break;
|
|
}
|
|
case MPI3_EVENT_DIAG_BUFFER_STATUS_CHANGE_RC_RESUMED:
|
|
{
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED;
|
|
break;
|
|
}
|
|
case MPI3_EVENT_DIAG_BUFFER_STATUS_CHANGE_RC_PAUSED:
|
|
{
|
|
diag_buffer->status = MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED;
|
|
break;
|
|
}
|
|
default:
|
|
dprint_event_th(mrioc, "%s: unknown reason_code(%d)\n",
|
|
__func__, evtdata->reason_code);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_diag_buffer_for_type - returns buffer desc for type
|
|
* @mrioc: Adapter instance reference
|
|
* @buf_type: Diagnostic buffer type
|
|
*
|
|
* Identifies matching diag descriptor from mrioc for given diag
|
|
* buffer type.
|
|
*
|
|
* Return: diag buffer descriptor on success, NULL on failures.
|
|
*/
|
|
|
|
struct diag_buffer_desc *
|
|
mpi3mr_diag_buffer_for_type(struct mpi3mr_ioc *mrioc, u8 buf_type)
|
|
{
|
|
u8 i;
|
|
|
|
for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
|
|
if (mrioc->diag_buffers[i].type == buf_type)
|
|
return &mrioc->diag_buffers[i];
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_pel_abort - sends PEL abort request
|
|
* @mrioc: Adapter instance reference
|
|
*
|
|
* This function sends PEL abort request to the firmware through
|
|
* admin request queue.
|
|
*
|
|
* Return: 0 on success, -1 on failure
|
|
*/
|
|
static int mpi3mr_bsg_pel_abort(struct mpi3mr_ioc *mrioc)
|
|
{
|
|
struct mpi3_pel_req_action_abort pel_abort_req;
|
|
struct mpi3_pel_reply *pel_reply;
|
|
int retval = 0;
|
|
u16 pe_log_status;
|
|
|
|
if (mrioc->reset_in_progress) {
|
|
dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
|
|
return -1;
|
|
}
|
|
if (mrioc->stop_bsgs || mrioc->block_on_pci_err) {
|
|
dprint_bsg_err(mrioc, "%s: bsgs are blocked\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
memset(&pel_abort_req, 0, sizeof(pel_abort_req));
|
|
mutex_lock(&mrioc->pel_abort_cmd.mutex);
|
|
if (mrioc->pel_abort_cmd.state & MPI3MR_CMD_PENDING) {
|
|
dprint_bsg_err(mrioc, "%s: command is in use\n", __func__);
|
|
mutex_unlock(&mrioc->pel_abort_cmd.mutex);
|
|
return -1;
|
|
}
|
|
mrioc->pel_abort_cmd.state = MPI3MR_CMD_PENDING;
|
|
mrioc->pel_abort_cmd.is_waiting = 1;
|
|
mrioc->pel_abort_cmd.callback = NULL;
|
|
pel_abort_req.host_tag = cpu_to_le16(MPI3MR_HOSTTAG_PEL_ABORT);
|
|
pel_abort_req.function = MPI3_FUNCTION_PERSISTENT_EVENT_LOG;
|
|
pel_abort_req.action = MPI3_PEL_ACTION_ABORT;
|
|
pel_abort_req.abort_host_tag = cpu_to_le16(MPI3MR_HOSTTAG_PEL_WAIT);
|
|
|
|
mrioc->pel_abort_requested = 1;
|
|
init_completion(&mrioc->pel_abort_cmd.done);
|
|
retval = mpi3mr_admin_request_post(mrioc, &pel_abort_req,
|
|
sizeof(pel_abort_req), 0);
|
|
if (retval) {
|
|
retval = -1;
|
|
dprint_bsg_err(mrioc, "%s: admin request post failed\n",
|
|
__func__);
|
|
mrioc->pel_abort_requested = 0;
|
|
goto out_unlock;
|
|
}
|
|
|
|
wait_for_completion_timeout(&mrioc->pel_abort_cmd.done,
|
|
(MPI3MR_INTADMCMD_TIMEOUT * HZ));
|
|
if (!(mrioc->pel_abort_cmd.state & MPI3MR_CMD_COMPLETE)) {
|
|
mrioc->pel_abort_cmd.is_waiting = 0;
|
|
dprint_bsg_err(mrioc, "%s: command timedout\n", __func__);
|
|
if (!(mrioc->pel_abort_cmd.state & MPI3MR_CMD_RESET))
|
|
mpi3mr_soft_reset_handler(mrioc,
|
|
MPI3MR_RESET_FROM_PELABORT_TIMEOUT, 1);
|
|
retval = -1;
|
|
goto out_unlock;
|
|
}
|
|
if ((mrioc->pel_abort_cmd.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
|
|
!= MPI3_IOCSTATUS_SUCCESS) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: command failed, ioc_status(0x%04x) log_info(0x%08x)\n",
|
|
__func__, (mrioc->pel_abort_cmd.ioc_status &
|
|
MPI3_IOCSTATUS_STATUS_MASK),
|
|
mrioc->pel_abort_cmd.ioc_loginfo);
|
|
retval = -1;
|
|
goto out_unlock;
|
|
}
|
|
if (mrioc->pel_abort_cmd.state & MPI3MR_CMD_REPLY_VALID) {
|
|
pel_reply = (struct mpi3_pel_reply *)mrioc->pel_abort_cmd.reply;
|
|
pe_log_status = le16_to_cpu(pel_reply->pe_log_status);
|
|
if (pe_log_status != MPI3_PEL_STATUS_SUCCESS) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: command failed, pel_status(0x%04x)\n",
|
|
__func__, pe_log_status);
|
|
retval = -1;
|
|
}
|
|
}
|
|
|
|
out_unlock:
|
|
mrioc->pel_abort_cmd.state = MPI3MR_CMD_NOTUSED;
|
|
mutex_unlock(&mrioc->pel_abort_cmd.mutex);
|
|
return retval;
|
|
}
|
|
/**
|
|
* mpi3mr_bsg_verify_adapter - verify adapter number is valid
|
|
* @ioc_number: Adapter number
|
|
*
|
|
* This function returns the adapter instance pointer of given
|
|
* adapter number. If adapter number does not match with the
|
|
* driver's adapter list, driver returns NULL.
|
|
*
|
|
* Return: adapter instance reference
|
|
*/
|
|
static struct mpi3mr_ioc *mpi3mr_bsg_verify_adapter(int ioc_number)
|
|
{
|
|
struct mpi3mr_ioc *mrioc = NULL;
|
|
|
|
spin_lock(&mrioc_list_lock);
|
|
list_for_each_entry(mrioc, &mrioc_list, list) {
|
|
if (mrioc->id == ioc_number) {
|
|
spin_unlock(&mrioc_list_lock);
|
|
return mrioc;
|
|
}
|
|
}
|
|
spin_unlock(&mrioc_list_lock);
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_refresh_hdb_triggers - Refresh HDB trigger data
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG Job pointer
|
|
*
|
|
* This function reads the controller trigger config page as
|
|
* defined by the input page type and refreshes the driver's
|
|
* local trigger information structures with the controller's
|
|
* config page data.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long
|
|
mpi3mr_bsg_refresh_hdb_triggers(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
struct mpi3mr_bsg_out_refresh_hdb_triggers refresh_triggers;
|
|
uint32_t data_out_sz;
|
|
u8 page_action;
|
|
long rval = -EINVAL;
|
|
|
|
data_out_sz = job->request_payload.payload_len;
|
|
|
|
if (data_out_sz != sizeof(refresh_triggers)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size argument\n",
|
|
__func__);
|
|
return rval;
|
|
}
|
|
|
|
if (mrioc->unrecoverable) {
|
|
dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
|
|
__func__);
|
|
return -EFAULT;
|
|
}
|
|
if (mrioc->reset_in_progress) {
|
|
dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sg_copy_to_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&refresh_triggers, sizeof(refresh_triggers));
|
|
|
|
switch (refresh_triggers.page_type) {
|
|
case MPI3MR_HDB_REFRESH_TYPE_CURRENT:
|
|
page_action = MPI3_CONFIG_ACTION_READ_CURRENT;
|
|
break;
|
|
case MPI3MR_HDB_REFRESH_TYPE_DEFAULT:
|
|
page_action = MPI3_CONFIG_ACTION_READ_DEFAULT;
|
|
break;
|
|
case MPI3MR_HDB_HDB_REFRESH_TYPE_PERSISTENT:
|
|
page_action = MPI3_CONFIG_ACTION_READ_PERSISTENT;
|
|
break;
|
|
default:
|
|
dprint_bsg_err(mrioc,
|
|
"%s: unsupported refresh trigger, page_type %d\n",
|
|
__func__, refresh_triggers.page_type);
|
|
return rval;
|
|
}
|
|
rval = mpi3mr_refresh_trigger(mrioc, page_action);
|
|
|
|
return rval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_upload_hdb - Upload a specific HDB to user space
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG Job pointer
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_bsg_upload_hdb(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
struct mpi3mr_bsg_out_upload_hdb upload_hdb;
|
|
struct diag_buffer_desc *diag_buffer;
|
|
uint32_t data_out_size;
|
|
uint32_t data_in_size;
|
|
|
|
data_out_size = job->request_payload.payload_len;
|
|
data_in_size = job->reply_payload.payload_len;
|
|
|
|
if (data_out_size != sizeof(upload_hdb)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size argument\n",
|
|
__func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
sg_copy_to_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&upload_hdb, sizeof(upload_hdb));
|
|
|
|
if ((!upload_hdb.length) || (data_in_size != upload_hdb.length)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid length argument\n",
|
|
__func__);
|
|
return -EINVAL;
|
|
}
|
|
diag_buffer = mpi3mr_diag_buffer_for_type(mrioc, upload_hdb.buf_type);
|
|
if ((!diag_buffer) || (!diag_buffer->addr)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid buffer type %d\n",
|
|
__func__, upload_hdb.buf_type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if ((diag_buffer->status != MPI3MR_HDB_BUFSTATUS_RELEASED) &&
|
|
(diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED)) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: invalid buffer status %d for type %d\n",
|
|
__func__, diag_buffer->status, upload_hdb.buf_type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if ((upload_hdb.start_offset + upload_hdb.length) > diag_buffer->size) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: invalid start offset %d, length %d for type %d\n",
|
|
__func__, upload_hdb.start_offset, upload_hdb.length,
|
|
upload_hdb.buf_type);
|
|
return -EINVAL;
|
|
}
|
|
sg_copy_from_buffer(job->reply_payload.sg_list,
|
|
job->reply_payload.sg_cnt,
|
|
(diag_buffer->addr + upload_hdb.start_offset),
|
|
data_in_size);
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_repost_hdb - Re-post HDB
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job pointer
|
|
*
|
|
* This function retrieves the HDB descriptor corresponding to a
|
|
* given buffer type and if the HDB is in released status then
|
|
* posts the HDB with the firmware.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_bsg_repost_hdb(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
struct mpi3mr_bsg_out_repost_hdb repost_hdb;
|
|
struct diag_buffer_desc *diag_buffer;
|
|
uint32_t data_out_sz;
|
|
|
|
data_out_sz = job->request_payload.payload_len;
|
|
|
|
if (data_out_sz != sizeof(repost_hdb)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size argument\n",
|
|
__func__);
|
|
return -EINVAL;
|
|
}
|
|
if (mrioc->unrecoverable) {
|
|
dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
|
|
__func__);
|
|
return -EFAULT;
|
|
}
|
|
if (mrioc->reset_in_progress) {
|
|
dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sg_copy_to_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&repost_hdb, sizeof(repost_hdb));
|
|
|
|
diag_buffer = mpi3mr_diag_buffer_for_type(mrioc, repost_hdb.buf_type);
|
|
if ((!diag_buffer) || (!diag_buffer->addr)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid buffer type %d\n",
|
|
__func__, repost_hdb.buf_type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (diag_buffer->status != MPI3MR_HDB_BUFSTATUS_RELEASED) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: invalid buffer status %d for type %d\n",
|
|
__func__, diag_buffer->status, repost_hdb.buf_type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (mpi3mr_issue_diag_buf_post(mrioc, diag_buffer)) {
|
|
dprint_bsg_err(mrioc, "%s: post failed for type %d\n",
|
|
__func__, repost_hdb.buf_type);
|
|
return -EFAULT;
|
|
}
|
|
mpi3mr_set_trigger_data_in_hdb(diag_buffer,
|
|
MPI3MR_HDB_TRIGGER_TYPE_UNKNOWN, NULL, 1);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_query_hdb - Handler for query HDB command
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job pointer
|
|
*
|
|
* This function prepares and copies the host diagnostic buffer
|
|
* entries to the user buffer.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_bsg_query_hdb(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
long rval = 0;
|
|
struct mpi3mr_bsg_in_hdb_status *hbd_status;
|
|
struct mpi3mr_hdb_entry *hbd_status_entry;
|
|
u32 length, min_length;
|
|
u8 i;
|
|
struct diag_buffer_desc *diag_buffer;
|
|
uint32_t data_in_sz = 0;
|
|
|
|
data_in_sz = job->request_payload.payload_len;
|
|
|
|
length = (sizeof(*hbd_status) + ((MPI3MR_MAX_NUM_HDB - 1) *
|
|
sizeof(*hbd_status_entry)));
|
|
hbd_status = kmalloc(length, GFP_KERNEL);
|
|
if (!hbd_status)
|
|
return -ENOMEM;
|
|
hbd_status_entry = &hbd_status->entry[0];
|
|
|
|
hbd_status->num_hdb_types = MPI3MR_MAX_NUM_HDB;
|
|
for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
|
|
diag_buffer = &mrioc->diag_buffers[i];
|
|
hbd_status_entry->buf_type = diag_buffer->type;
|
|
hbd_status_entry->status = diag_buffer->status;
|
|
hbd_status_entry->trigger_type = diag_buffer->trigger_type;
|
|
memcpy(&hbd_status_entry->trigger_data,
|
|
&diag_buffer->trigger_data,
|
|
sizeof(hbd_status_entry->trigger_data));
|
|
hbd_status_entry->size = (diag_buffer->size / 1024);
|
|
hbd_status_entry++;
|
|
}
|
|
hbd_status->element_trigger_format =
|
|
MPI3MR_HDB_QUERY_ELEMENT_TRIGGER_FORMAT_DATA;
|
|
|
|
if (data_in_sz < 4) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size passed\n", __func__);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
min_length = min(data_in_sz, length);
|
|
if (job->request_payload.payload_len >= min_length) {
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
hbd_status, min_length);
|
|
rval = 0;
|
|
}
|
|
out:
|
|
kfree(hbd_status);
|
|
return rval;
|
|
}
|
|
|
|
|
|
/**
|
|
* mpi3mr_enable_logdata - Handler for log data enable
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job reference
|
|
*
|
|
* This function enables log data caching in the driver if not
|
|
* already enabled and return the maximum number of log data
|
|
* entries that can be cached in the driver.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_enable_logdata(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
struct mpi3mr_logdata_enable logdata_enable;
|
|
|
|
if (!mrioc->logdata_buf) {
|
|
mrioc->logdata_entry_sz =
|
|
(mrioc->reply_sz - (sizeof(struct mpi3_event_notification_reply) - 4))
|
|
+ MPI3MR_BSG_LOGDATA_ENTRY_HEADER_SZ;
|
|
mrioc->logdata_buf_idx = 0;
|
|
mrioc->logdata_buf = kcalloc(MPI3MR_BSG_LOGDATA_MAX_ENTRIES,
|
|
mrioc->logdata_entry_sz, GFP_KERNEL);
|
|
|
|
if (!mrioc->logdata_buf)
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memset(&logdata_enable, 0, sizeof(logdata_enable));
|
|
logdata_enable.max_entries =
|
|
MPI3MR_BSG_LOGDATA_MAX_ENTRIES;
|
|
if (job->request_payload.payload_len >= sizeof(logdata_enable)) {
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&logdata_enable, sizeof(logdata_enable));
|
|
return 0;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
/**
|
|
* mpi3mr_get_logdata - Handler for get log data
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job pointer
|
|
* This function copies the log data entries to the user buffer
|
|
* when log caching is enabled in the driver.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_get_logdata(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
u16 num_entries, sz, entry_sz = mrioc->logdata_entry_sz;
|
|
|
|
if ((!mrioc->logdata_buf) || (job->request_payload.payload_len < entry_sz))
|
|
return -EINVAL;
|
|
|
|
num_entries = job->request_payload.payload_len / entry_sz;
|
|
if (num_entries > MPI3MR_BSG_LOGDATA_MAX_ENTRIES)
|
|
num_entries = MPI3MR_BSG_LOGDATA_MAX_ENTRIES;
|
|
sz = num_entries * entry_sz;
|
|
|
|
if (job->request_payload.payload_len >= sz) {
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
mrioc->logdata_buf, sz);
|
|
return 0;
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_pel_enable - Handler for PEL enable driver
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job pointer
|
|
*
|
|
* This function is the handler for PEL enable driver.
|
|
* Validates the application given class and locale and if
|
|
* requires aborts the existing PEL wait request and/or issues
|
|
* new PEL wait request to the firmware and returns.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure.
|
|
*/
|
|
static long mpi3mr_bsg_pel_enable(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
long rval = -EINVAL;
|
|
struct mpi3mr_bsg_out_pel_enable pel_enable;
|
|
u8 issue_pel_wait;
|
|
u8 tmp_class;
|
|
u16 tmp_locale;
|
|
|
|
if (job->request_payload.payload_len != sizeof(pel_enable)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size argument\n",
|
|
__func__);
|
|
return rval;
|
|
}
|
|
|
|
if (mrioc->unrecoverable) {
|
|
dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
|
|
__func__);
|
|
return -EFAULT;
|
|
}
|
|
|
|
if (mrioc->reset_in_progress) {
|
|
dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
if (mrioc->stop_bsgs) {
|
|
dprint_bsg_err(mrioc, "%s: bsgs are blocked\n", __func__);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sg_copy_to_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&pel_enable, sizeof(pel_enable));
|
|
|
|
if (pel_enable.pel_class > MPI3_PEL_CLASS_FAULT) {
|
|
dprint_bsg_err(mrioc, "%s: out of range class %d sent\n",
|
|
__func__, pel_enable.pel_class);
|
|
rval = 0;
|
|
goto out;
|
|
}
|
|
if (!mrioc->pel_enabled)
|
|
issue_pel_wait = 1;
|
|
else {
|
|
if ((mrioc->pel_class <= pel_enable.pel_class) &&
|
|
!((mrioc->pel_locale & pel_enable.pel_locale) ^
|
|
pel_enable.pel_locale)) {
|
|
issue_pel_wait = 0;
|
|
rval = 0;
|
|
} else {
|
|
pel_enable.pel_locale |= mrioc->pel_locale;
|
|
|
|
if (mrioc->pel_class < pel_enable.pel_class)
|
|
pel_enable.pel_class = mrioc->pel_class;
|
|
|
|
rval = mpi3mr_bsg_pel_abort(mrioc);
|
|
if (rval) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: pel_abort failed, status(%ld)\n",
|
|
__func__, rval);
|
|
goto out;
|
|
}
|
|
issue_pel_wait = 1;
|
|
}
|
|
}
|
|
if (issue_pel_wait) {
|
|
tmp_class = mrioc->pel_class;
|
|
tmp_locale = mrioc->pel_locale;
|
|
mrioc->pel_class = pel_enable.pel_class;
|
|
mrioc->pel_locale = pel_enable.pel_locale;
|
|
mrioc->pel_enabled = 1;
|
|
rval = mpi3mr_pel_get_seqnum_post(mrioc, NULL);
|
|
if (rval) {
|
|
mrioc->pel_class = tmp_class;
|
|
mrioc->pel_locale = tmp_locale;
|
|
mrioc->pel_enabled = 0;
|
|
dprint_bsg_err(mrioc,
|
|
"%s: pel get sequence number failed, status(%ld)\n",
|
|
__func__, rval);
|
|
}
|
|
}
|
|
|
|
out:
|
|
return rval;
|
|
}
|
|
/**
|
|
* mpi3mr_get_all_tgt_info - Get all target information
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job reference
|
|
*
|
|
* This function copies the driver managed target devices device
|
|
* handle, persistent ID, bus ID and taret ID to the user
|
|
* provided buffer for the specific controller. This function
|
|
* also provides the number of devices managed by the driver for
|
|
* the specific controller.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_get_all_tgt_info(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
u16 num_devices = 0, i = 0, size;
|
|
unsigned long flags;
|
|
struct mpi3mr_tgt_dev *tgtdev;
|
|
struct mpi3mr_device_map_info *devmap_info = NULL;
|
|
struct mpi3mr_all_tgt_info *alltgt_info = NULL;
|
|
uint32_t min_entrylen = 0, kern_entrylen = 0, usr_entrylen = 0;
|
|
|
|
if (job->request_payload.payload_len < sizeof(u32)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size argument\n",
|
|
__func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
spin_lock_irqsave(&mrioc->tgtdev_lock, flags);
|
|
list_for_each_entry(tgtdev, &mrioc->tgtdev_list, list)
|
|
num_devices++;
|
|
spin_unlock_irqrestore(&mrioc->tgtdev_lock, flags);
|
|
|
|
if ((job->request_payload.payload_len <= sizeof(u64)) ||
|
|
list_empty(&mrioc->tgtdev_list)) {
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&num_devices, sizeof(num_devices));
|
|
return 0;
|
|
}
|
|
|
|
kern_entrylen = num_devices * sizeof(*devmap_info);
|
|
size = sizeof(u64) + kern_entrylen;
|
|
alltgt_info = kzalloc(size, GFP_KERNEL);
|
|
if (!alltgt_info)
|
|
return -ENOMEM;
|
|
|
|
devmap_info = alltgt_info->dmi;
|
|
memset((u8 *)devmap_info, 0xFF, kern_entrylen);
|
|
spin_lock_irqsave(&mrioc->tgtdev_lock, flags);
|
|
list_for_each_entry(tgtdev, &mrioc->tgtdev_list, list) {
|
|
if (i < num_devices) {
|
|
devmap_info[i].handle = tgtdev->dev_handle;
|
|
devmap_info[i].perst_id = tgtdev->perst_id;
|
|
if (tgtdev->host_exposed && tgtdev->starget) {
|
|
devmap_info[i].target_id = tgtdev->starget->id;
|
|
devmap_info[i].bus_id =
|
|
tgtdev->starget->channel;
|
|
}
|
|
i++;
|
|
}
|
|
}
|
|
num_devices = i;
|
|
spin_unlock_irqrestore(&mrioc->tgtdev_lock, flags);
|
|
|
|
alltgt_info->num_devices = num_devices;
|
|
|
|
usr_entrylen = (job->request_payload.payload_len - sizeof(u64)) /
|
|
sizeof(*devmap_info);
|
|
usr_entrylen *= sizeof(*devmap_info);
|
|
min_entrylen = min(usr_entrylen, kern_entrylen);
|
|
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
alltgt_info, (min_entrylen + sizeof(u64)));
|
|
kfree(alltgt_info);
|
|
return 0;
|
|
}
|
|
/**
|
|
* mpi3mr_get_change_count - Get topology change count
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job reference
|
|
*
|
|
* This function copies the toplogy change count provided by the
|
|
* driver in events and cached in the driver to the user
|
|
* provided buffer for the specific controller.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_get_change_count(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
struct mpi3mr_change_count chgcnt;
|
|
|
|
memset(&chgcnt, 0, sizeof(chgcnt));
|
|
chgcnt.change_count = mrioc->change_count;
|
|
if (job->request_payload.payload_len >= sizeof(chgcnt)) {
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&chgcnt, sizeof(chgcnt));
|
|
return 0;
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_adp_reset - Issue controller reset
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job reference
|
|
*
|
|
* This function identifies the user provided reset type and
|
|
* issues approporiate reset to the controller and wait for that
|
|
* to complete and reinitialize the controller and then returns
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_bsg_adp_reset(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
long rval = -EINVAL;
|
|
u8 save_snapdump;
|
|
struct mpi3mr_bsg_adp_reset adpreset;
|
|
|
|
if (job->request_payload.payload_len !=
|
|
sizeof(adpreset)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid size argument\n",
|
|
__func__);
|
|
goto out;
|
|
}
|
|
|
|
if (mrioc->unrecoverable || mrioc->block_on_pci_err)
|
|
return -EINVAL;
|
|
|
|
sg_copy_to_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&adpreset, sizeof(adpreset));
|
|
|
|
switch (adpreset.reset_type) {
|
|
case MPI3MR_BSG_ADPRESET_SOFT:
|
|
save_snapdump = 0;
|
|
break;
|
|
case MPI3MR_BSG_ADPRESET_DIAG_FAULT:
|
|
save_snapdump = 1;
|
|
break;
|
|
default:
|
|
dprint_bsg_err(mrioc, "%s: unknown reset_type(%d)\n",
|
|
__func__, adpreset.reset_type);
|
|
goto out;
|
|
}
|
|
|
|
rval = mpi3mr_soft_reset_handler(mrioc, MPI3MR_RESET_FROM_APP,
|
|
save_snapdump);
|
|
|
|
if (rval)
|
|
dprint_bsg_err(mrioc,
|
|
"%s: reset handler returned error(%ld) for reset type %d\n",
|
|
__func__, rval, adpreset.reset_type);
|
|
out:
|
|
return rval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_populate_adpinfo - Get adapter info command handler
|
|
* @mrioc: Adapter instance reference
|
|
* @job: BSG job reference
|
|
*
|
|
* This function provides adapter information for the given
|
|
* controller
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_bsg_populate_adpinfo(struct mpi3mr_ioc *mrioc,
|
|
struct bsg_job *job)
|
|
{
|
|
enum mpi3mr_iocstate ioc_state;
|
|
struct mpi3mr_bsg_in_adpinfo adpinfo;
|
|
|
|
memset(&adpinfo, 0, sizeof(adpinfo));
|
|
adpinfo.adp_type = MPI3MR_BSG_ADPTYPE_AVGFAMILY;
|
|
adpinfo.pci_dev_id = mrioc->pdev->device;
|
|
adpinfo.pci_dev_hw_rev = mrioc->pdev->revision;
|
|
adpinfo.pci_subsys_dev_id = mrioc->pdev->subsystem_device;
|
|
adpinfo.pci_subsys_ven_id = mrioc->pdev->subsystem_vendor;
|
|
adpinfo.pci_bus = mrioc->pdev->bus->number;
|
|
adpinfo.pci_dev = PCI_SLOT(mrioc->pdev->devfn);
|
|
adpinfo.pci_func = PCI_FUNC(mrioc->pdev->devfn);
|
|
adpinfo.pci_seg_id = pci_domain_nr(mrioc->pdev->bus);
|
|
adpinfo.app_intfc_ver = MPI3MR_IOCTL_VERSION;
|
|
|
|
ioc_state = mpi3mr_get_iocstate(mrioc);
|
|
if (ioc_state == MRIOC_STATE_UNRECOVERABLE)
|
|
adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_UNRECOVERABLE;
|
|
else if ((mrioc->reset_in_progress) || (mrioc->stop_bsgs))
|
|
adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_IN_RESET;
|
|
else if (ioc_state == MRIOC_STATE_FAULT)
|
|
adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_FAULT;
|
|
else
|
|
adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_OPERATIONAL;
|
|
|
|
memcpy((u8 *)&adpinfo.driver_info, (u8 *)&mrioc->driver_info,
|
|
sizeof(adpinfo.driver_info));
|
|
|
|
if (job->request_payload.payload_len >= sizeof(adpinfo)) {
|
|
sg_copy_from_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
&adpinfo, sizeof(adpinfo));
|
|
return 0;
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_process_drv_cmds - Driver Command handler
|
|
* @job: BSG job reference
|
|
*
|
|
* This function is the top level handler for driver commands,
|
|
* this does basic validation of the buffer and identifies the
|
|
* opcode and switches to correct sub handler.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static long mpi3mr_bsg_process_drv_cmds(struct bsg_job *job)
|
|
{
|
|
long rval = -EINVAL;
|
|
struct mpi3mr_ioc *mrioc = NULL;
|
|
struct mpi3mr_bsg_packet *bsg_req = NULL;
|
|
struct mpi3mr_bsg_drv_cmd *drvrcmd = NULL;
|
|
|
|
bsg_req = job->request;
|
|
drvrcmd = &bsg_req->cmd.drvrcmd;
|
|
|
|
mrioc = mpi3mr_bsg_verify_adapter(drvrcmd->mrioc_id);
|
|
if (!mrioc)
|
|
return -ENODEV;
|
|
|
|
if (drvrcmd->opcode == MPI3MR_DRVBSG_OPCODE_ADPINFO) {
|
|
rval = mpi3mr_bsg_populate_adpinfo(mrioc, job);
|
|
return rval;
|
|
}
|
|
|
|
if (mutex_lock_interruptible(&mrioc->bsg_cmds.mutex))
|
|
return -ERESTARTSYS;
|
|
|
|
switch (drvrcmd->opcode) {
|
|
case MPI3MR_DRVBSG_OPCODE_ADPRESET:
|
|
rval = mpi3mr_bsg_adp_reset(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_ALLTGTDEVINFO:
|
|
rval = mpi3mr_get_all_tgt_info(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_GETCHGCNT:
|
|
rval = mpi3mr_get_change_count(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_LOGDATAENABLE:
|
|
rval = mpi3mr_enable_logdata(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_GETLOGDATA:
|
|
rval = mpi3mr_get_logdata(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_PELENABLE:
|
|
rval = mpi3mr_bsg_pel_enable(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_QUERY_HDB:
|
|
rval = mpi3mr_bsg_query_hdb(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_REPOST_HDB:
|
|
rval = mpi3mr_bsg_repost_hdb(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_UPLOAD_HDB:
|
|
rval = mpi3mr_bsg_upload_hdb(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_REFRESH_HDB_TRIGGERS:
|
|
rval = mpi3mr_bsg_refresh_hdb_triggers(mrioc, job);
|
|
break;
|
|
case MPI3MR_DRVBSG_OPCODE_UNKNOWN:
|
|
default:
|
|
pr_err("%s: unsupported driver command opcode %d\n",
|
|
MPI3MR_DRIVER_NAME, drvrcmd->opcode);
|
|
break;
|
|
}
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
return rval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_total_num_ioctl_sges - Count number of SGEs required
|
|
* @drv_bufs: DMA address of the buffers to be placed in sgl
|
|
* @bufcnt: Number of DMA buffers
|
|
*
|
|
* This function returns total number of data SGEs required
|
|
* including zero length SGEs and excluding management request
|
|
* and response buffer for the given list of data buffer
|
|
* descriptors
|
|
*
|
|
* Return: Number of SGE elements needed
|
|
*/
|
|
static inline u16 mpi3mr_total_num_ioctl_sges(struct mpi3mr_buf_map *drv_bufs,
|
|
u8 bufcnt)
|
|
{
|
|
u16 i, sge_count = 0;
|
|
|
|
for (i = 0; i < bufcnt; i++, drv_bufs++) {
|
|
if (drv_bufs->data_dir == DMA_NONE ||
|
|
drv_bufs->kern_buf)
|
|
continue;
|
|
sge_count += drv_bufs->num_dma_desc;
|
|
if (!drv_bufs->num_dma_desc)
|
|
sge_count++;
|
|
}
|
|
return sge_count;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_build_sgl - SGL construction for MPI commands
|
|
* @mrioc: Adapter instance reference
|
|
* @mpi_req: MPI request
|
|
* @sgl_offset: offset to start sgl in the MPI request
|
|
* @drv_bufs: DMA address of the buffers to be placed in sgl
|
|
* @bufcnt: Number of DMA buffers
|
|
* @is_rmc: Does the buffer list has management command buffer
|
|
* @is_rmr: Does the buffer list has management response buffer
|
|
* @num_datasges: Number of data buffers in the list
|
|
*
|
|
* This function places the DMA address of the given buffers in
|
|
* proper format as SGEs in the given MPI request.
|
|
*
|
|
* Return: 0 on success,-1 on failure
|
|
*/
|
|
static int mpi3mr_bsg_build_sgl(struct mpi3mr_ioc *mrioc, u8 *mpi_req,
|
|
u32 sgl_offset, struct mpi3mr_buf_map *drv_bufs,
|
|
u8 bufcnt, u8 is_rmc, u8 is_rmr, u8 num_datasges)
|
|
{
|
|
struct mpi3_request_header *mpi_header =
|
|
(struct mpi3_request_header *)mpi_req;
|
|
u8 *sgl = (mpi_req + sgl_offset), count = 0;
|
|
struct mpi3_mgmt_passthrough_request *rmgmt_req =
|
|
(struct mpi3_mgmt_passthrough_request *)mpi_req;
|
|
struct mpi3mr_buf_map *drv_buf_iter = drv_bufs;
|
|
u8 flag, sgl_flags, sgl_flag_eob, sgl_flags_last, last_chain_sgl_flag;
|
|
u16 available_sges, i, sges_needed;
|
|
u32 sge_element_size = sizeof(struct mpi3_sge_common);
|
|
bool chain_used = false;
|
|
|
|
sgl_flags = MPI3_SGE_FLAGS_ELEMENT_TYPE_SIMPLE |
|
|
MPI3_SGE_FLAGS_DLAS_SYSTEM;
|
|
sgl_flag_eob = sgl_flags | MPI3_SGE_FLAGS_END_OF_BUFFER;
|
|
sgl_flags_last = sgl_flag_eob | MPI3_SGE_FLAGS_END_OF_LIST;
|
|
last_chain_sgl_flag = MPI3_SGE_FLAGS_ELEMENT_TYPE_LAST_CHAIN |
|
|
MPI3_SGE_FLAGS_DLAS_SYSTEM;
|
|
|
|
sges_needed = mpi3mr_total_num_ioctl_sges(drv_bufs, bufcnt);
|
|
|
|
if (is_rmc) {
|
|
mpi3mr_add_sg_single(&rmgmt_req->command_sgl,
|
|
sgl_flags_last, drv_buf_iter->kern_buf_len,
|
|
drv_buf_iter->kern_buf_dma);
|
|
sgl = (u8 *)drv_buf_iter->kern_buf +
|
|
drv_buf_iter->bsg_buf_len;
|
|
available_sges = (drv_buf_iter->kern_buf_len -
|
|
drv_buf_iter->bsg_buf_len) / sge_element_size;
|
|
|
|
if (sges_needed > available_sges)
|
|
return -1;
|
|
|
|
chain_used = true;
|
|
drv_buf_iter++;
|
|
count++;
|
|
if (is_rmr) {
|
|
mpi3mr_add_sg_single(&rmgmt_req->response_sgl,
|
|
sgl_flags_last, drv_buf_iter->kern_buf_len,
|
|
drv_buf_iter->kern_buf_dma);
|
|
drv_buf_iter++;
|
|
count++;
|
|
} else
|
|
mpi3mr_build_zero_len_sge(
|
|
&rmgmt_req->response_sgl);
|
|
if (num_datasges) {
|
|
i = 0;
|
|
goto build_sges;
|
|
}
|
|
} else {
|
|
if (sgl_offset >= MPI3MR_ADMIN_REQ_FRAME_SZ)
|
|
return -1;
|
|
available_sges = (MPI3MR_ADMIN_REQ_FRAME_SZ - sgl_offset) /
|
|
sge_element_size;
|
|
if (!available_sges)
|
|
return -1;
|
|
}
|
|
if (!num_datasges) {
|
|
mpi3mr_build_zero_len_sge(sgl);
|
|
return 0;
|
|
}
|
|
if (mpi_header->function == MPI3_BSG_FUNCTION_SMP_PASSTHROUGH) {
|
|
if ((sges_needed > 2) || (sges_needed > available_sges))
|
|
return -1;
|
|
for (; count < bufcnt; count++, drv_buf_iter++) {
|
|
if (drv_buf_iter->data_dir == DMA_NONE ||
|
|
!drv_buf_iter->num_dma_desc)
|
|
continue;
|
|
mpi3mr_add_sg_single(sgl, sgl_flags_last,
|
|
drv_buf_iter->dma_desc[0].size,
|
|
drv_buf_iter->dma_desc[0].dma_addr);
|
|
sgl += sge_element_size;
|
|
}
|
|
return 0;
|
|
}
|
|
i = 0;
|
|
|
|
build_sges:
|
|
for (; count < bufcnt; count++, drv_buf_iter++) {
|
|
if (drv_buf_iter->data_dir == DMA_NONE)
|
|
continue;
|
|
if (!drv_buf_iter->num_dma_desc) {
|
|
if (chain_used && !available_sges)
|
|
return -1;
|
|
if (!chain_used && (available_sges == 1) &&
|
|
(sges_needed > 1))
|
|
goto setup_chain;
|
|
flag = sgl_flag_eob;
|
|
if (num_datasges == 1)
|
|
flag = sgl_flags_last;
|
|
mpi3mr_add_sg_single(sgl, flag, 0, 0);
|
|
sgl += sge_element_size;
|
|
sges_needed--;
|
|
available_sges--;
|
|
num_datasges--;
|
|
continue;
|
|
}
|
|
for (; i < drv_buf_iter->num_dma_desc; i++) {
|
|
if (chain_used && !available_sges)
|
|
return -1;
|
|
if (!chain_used && (available_sges == 1) &&
|
|
(sges_needed > 1))
|
|
goto setup_chain;
|
|
flag = sgl_flags;
|
|
if (i == (drv_buf_iter->num_dma_desc - 1)) {
|
|
if (num_datasges == 1)
|
|
flag = sgl_flags_last;
|
|
else
|
|
flag = sgl_flag_eob;
|
|
}
|
|
|
|
mpi3mr_add_sg_single(sgl, flag,
|
|
drv_buf_iter->dma_desc[i].size,
|
|
drv_buf_iter->dma_desc[i].dma_addr);
|
|
sgl += sge_element_size;
|
|
available_sges--;
|
|
sges_needed--;
|
|
}
|
|
num_datasges--;
|
|
i = 0;
|
|
}
|
|
return 0;
|
|
|
|
setup_chain:
|
|
available_sges = mrioc->ioctl_chain_sge.size / sge_element_size;
|
|
if (sges_needed > available_sges)
|
|
return -1;
|
|
mpi3mr_add_sg_single(sgl, last_chain_sgl_flag,
|
|
(sges_needed * sge_element_size),
|
|
mrioc->ioctl_chain_sge.dma_addr);
|
|
memset(mrioc->ioctl_chain_sge.addr, 0, mrioc->ioctl_chain_sge.size);
|
|
sgl = (u8 *)mrioc->ioctl_chain_sge.addr;
|
|
chain_used = true;
|
|
goto build_sges;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_get_nvme_data_fmt - returns the NVMe data format
|
|
* @nvme_encap_request: NVMe encapsulated MPI request
|
|
*
|
|
* This function returns the type of the data format specified
|
|
* in user provided NVMe command in NVMe encapsulated request.
|
|
*
|
|
* Return: Data format of the NVMe command (PRP/SGL etc)
|
|
*/
|
|
static unsigned int mpi3mr_get_nvme_data_fmt(
|
|
struct mpi3_nvme_encapsulated_request *nvme_encap_request)
|
|
{
|
|
u8 format = 0;
|
|
|
|
format = ((nvme_encap_request->command[0] & 0xc000) >> 14);
|
|
return format;
|
|
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_build_nvme_sgl - SGL constructor for NVME
|
|
* encapsulated request
|
|
* @mrioc: Adapter instance reference
|
|
* @nvme_encap_request: NVMe encapsulated MPI request
|
|
* @drv_bufs: DMA address of the buffers to be placed in sgl
|
|
* @bufcnt: Number of DMA buffers
|
|
*
|
|
* This function places the DMA address of the given buffers in
|
|
* proper format as SGEs in the given NVMe encapsulated request.
|
|
*
|
|
* Return: 0 on success, -1 on failure
|
|
*/
|
|
static int mpi3mr_build_nvme_sgl(struct mpi3mr_ioc *mrioc,
|
|
struct mpi3_nvme_encapsulated_request *nvme_encap_request,
|
|
struct mpi3mr_buf_map *drv_bufs, u8 bufcnt)
|
|
{
|
|
struct mpi3mr_nvme_pt_sge *nvme_sgl;
|
|
__le64 sgl_dma;
|
|
u8 count;
|
|
size_t length = 0;
|
|
u16 available_sges = 0, i;
|
|
u32 sge_element_size = sizeof(struct mpi3mr_nvme_pt_sge);
|
|
struct mpi3mr_buf_map *drv_buf_iter = drv_bufs;
|
|
u64 sgemod_mask = ((u64)((mrioc->facts.sge_mod_mask) <<
|
|
mrioc->facts.sge_mod_shift) << 32);
|
|
u64 sgemod_val = ((u64)(mrioc->facts.sge_mod_value) <<
|
|
mrioc->facts.sge_mod_shift) << 32;
|
|
u32 size;
|
|
|
|
nvme_sgl = (struct mpi3mr_nvme_pt_sge *)
|
|
((u8 *)(nvme_encap_request->command) + MPI3MR_NVME_CMD_SGL_OFFSET);
|
|
|
|
/*
|
|
* Not all commands require a data transfer. If no data, just return
|
|
* without constructing any sgl.
|
|
*/
|
|
for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
|
|
if (drv_buf_iter->data_dir == DMA_NONE)
|
|
continue;
|
|
length = drv_buf_iter->kern_buf_len;
|
|
break;
|
|
}
|
|
if (!length || !drv_buf_iter->num_dma_desc)
|
|
return 0;
|
|
|
|
if (drv_buf_iter->num_dma_desc == 1) {
|
|
available_sges = 1;
|
|
goto build_sges;
|
|
}
|
|
|
|
sgl_dma = cpu_to_le64(mrioc->ioctl_chain_sge.dma_addr);
|
|
if (sgl_dma & sgemod_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: SGL chain address collides with SGE modifier\n",
|
|
__func__);
|
|
return -1;
|
|
}
|
|
|
|
sgl_dma &= ~sgemod_mask;
|
|
sgl_dma |= sgemod_val;
|
|
|
|
memset(mrioc->ioctl_chain_sge.addr, 0, mrioc->ioctl_chain_sge.size);
|
|
available_sges = mrioc->ioctl_chain_sge.size / sge_element_size;
|
|
if (available_sges < drv_buf_iter->num_dma_desc)
|
|
return -1;
|
|
memset(nvme_sgl, 0, sizeof(struct mpi3mr_nvme_pt_sge));
|
|
nvme_sgl->base_addr = sgl_dma;
|
|
size = drv_buf_iter->num_dma_desc * sizeof(struct mpi3mr_nvme_pt_sge);
|
|
nvme_sgl->length = cpu_to_le32(size);
|
|
nvme_sgl->type = MPI3MR_NVMESGL_LAST_SEGMENT;
|
|
nvme_sgl = (struct mpi3mr_nvme_pt_sge *)mrioc->ioctl_chain_sge.addr;
|
|
|
|
build_sges:
|
|
for (i = 0; i < drv_buf_iter->num_dma_desc; i++) {
|
|
sgl_dma = cpu_to_le64(drv_buf_iter->dma_desc[i].dma_addr);
|
|
if (sgl_dma & sgemod_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: SGL address collides with SGE modifier\n",
|
|
__func__);
|
|
return -1;
|
|
}
|
|
|
|
sgl_dma &= ~sgemod_mask;
|
|
sgl_dma |= sgemod_val;
|
|
|
|
nvme_sgl->base_addr = sgl_dma;
|
|
nvme_sgl->length = cpu_to_le32(drv_buf_iter->dma_desc[i].size);
|
|
nvme_sgl->type = MPI3MR_NVMESGL_DATA_SEGMENT;
|
|
nvme_sgl++;
|
|
available_sges--;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_build_nvme_prp - PRP constructor for NVME
|
|
* encapsulated request
|
|
* @mrioc: Adapter instance reference
|
|
* @nvme_encap_request: NVMe encapsulated MPI request
|
|
* @drv_bufs: DMA address of the buffers to be placed in SGL
|
|
* @bufcnt: Number of DMA buffers
|
|
*
|
|
* This function places the DMA address of the given buffers in
|
|
* proper format as PRP entries in the given NVMe encapsulated
|
|
* request.
|
|
*
|
|
* Return: 0 on success, -1 on failure
|
|
*/
|
|
static int mpi3mr_build_nvme_prp(struct mpi3mr_ioc *mrioc,
|
|
struct mpi3_nvme_encapsulated_request *nvme_encap_request,
|
|
struct mpi3mr_buf_map *drv_bufs, u8 bufcnt)
|
|
{
|
|
int prp_size = MPI3MR_NVME_PRP_SIZE;
|
|
__le64 *prp_entry, *prp1_entry, *prp2_entry;
|
|
__le64 *prp_page;
|
|
dma_addr_t prp_entry_dma, prp_page_dma, dma_addr;
|
|
u32 offset, entry_len, dev_pgsz;
|
|
u32 page_mask_result, page_mask;
|
|
size_t length = 0, desc_len;
|
|
u8 count;
|
|
struct mpi3mr_buf_map *drv_buf_iter = drv_bufs;
|
|
u64 sgemod_mask = ((u64)((mrioc->facts.sge_mod_mask) <<
|
|
mrioc->facts.sge_mod_shift) << 32);
|
|
u64 sgemod_val = ((u64)(mrioc->facts.sge_mod_value) <<
|
|
mrioc->facts.sge_mod_shift) << 32;
|
|
u16 dev_handle = nvme_encap_request->dev_handle;
|
|
struct mpi3mr_tgt_dev *tgtdev;
|
|
u16 desc_count = 0;
|
|
|
|
tgtdev = mpi3mr_get_tgtdev_by_handle(mrioc, dev_handle);
|
|
if (!tgtdev) {
|
|
dprint_bsg_err(mrioc, "%s: invalid device handle 0x%04x\n",
|
|
__func__, dev_handle);
|
|
return -1;
|
|
}
|
|
|
|
if (tgtdev->dev_spec.pcie_inf.pgsz == 0) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: NVMe device page size is zero for handle 0x%04x\n",
|
|
__func__, dev_handle);
|
|
mpi3mr_tgtdev_put(tgtdev);
|
|
return -1;
|
|
}
|
|
|
|
dev_pgsz = 1 << (tgtdev->dev_spec.pcie_inf.pgsz);
|
|
mpi3mr_tgtdev_put(tgtdev);
|
|
page_mask = dev_pgsz - 1;
|
|
|
|
if (dev_pgsz > MPI3MR_IOCTL_SGE_SIZE) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: NVMe device page size(%d) is greater than ioctl data sge size(%d) for handle 0x%04x\n",
|
|
__func__, dev_pgsz, MPI3MR_IOCTL_SGE_SIZE, dev_handle);
|
|
return -1;
|
|
}
|
|
|
|
if (MPI3MR_IOCTL_SGE_SIZE % dev_pgsz) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: ioctl data sge size(%d) is not a multiple of NVMe device page size(%d) for handle 0x%04x\n",
|
|
__func__, MPI3MR_IOCTL_SGE_SIZE, dev_pgsz, dev_handle);
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* Not all commands require a data transfer. If no data, just return
|
|
* without constructing any PRP.
|
|
*/
|
|
for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
|
|
if (drv_buf_iter->data_dir == DMA_NONE)
|
|
continue;
|
|
length = drv_buf_iter->kern_buf_len;
|
|
break;
|
|
}
|
|
|
|
if (!length || !drv_buf_iter->num_dma_desc)
|
|
return 0;
|
|
|
|
for (count = 0; count < drv_buf_iter->num_dma_desc; count++) {
|
|
dma_addr = drv_buf_iter->dma_desc[count].dma_addr;
|
|
if (dma_addr & page_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s:dma_addr %pad is not aligned with page size 0x%x\n",
|
|
__func__, &dma_addr, dev_pgsz);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
dma_addr = drv_buf_iter->dma_desc[0].dma_addr;
|
|
desc_len = drv_buf_iter->dma_desc[0].size;
|
|
|
|
mrioc->prp_sz = 0;
|
|
mrioc->prp_list_virt = dma_alloc_coherent(&mrioc->pdev->dev,
|
|
dev_pgsz, &mrioc->prp_list_dma, GFP_KERNEL);
|
|
|
|
if (!mrioc->prp_list_virt)
|
|
return -1;
|
|
mrioc->prp_sz = dev_pgsz;
|
|
|
|
/*
|
|
* Set pointers to PRP1 and PRP2, which are in the NVMe command.
|
|
* PRP1 is located at a 24 byte offset from the start of the NVMe
|
|
* command. Then set the current PRP entry pointer to PRP1.
|
|
*/
|
|
prp1_entry = (__le64 *)((u8 *)(nvme_encap_request->command) +
|
|
MPI3MR_NVME_CMD_PRP1_OFFSET);
|
|
prp2_entry = (__le64 *)((u8 *)(nvme_encap_request->command) +
|
|
MPI3MR_NVME_CMD_PRP2_OFFSET);
|
|
prp_entry = prp1_entry;
|
|
/*
|
|
* For the PRP entries, use the specially allocated buffer of
|
|
* contiguous memory.
|
|
*/
|
|
prp_page = (__le64 *)mrioc->prp_list_virt;
|
|
prp_page_dma = mrioc->prp_list_dma;
|
|
|
|
/*
|
|
* Check if we are within 1 entry of a page boundary we don't
|
|
* want our first entry to be a PRP List entry.
|
|
*/
|
|
page_mask_result = (uintptr_t)((u8 *)prp_page + prp_size) & page_mask;
|
|
if (!page_mask_result) {
|
|
dprint_bsg_err(mrioc, "%s: PRP page is not page aligned\n",
|
|
__func__);
|
|
goto err_out;
|
|
}
|
|
|
|
/*
|
|
* Set PRP physical pointer, which initially points to the current PRP
|
|
* DMA memory page.
|
|
*/
|
|
prp_entry_dma = prp_page_dma;
|
|
|
|
|
|
/* Loop while the length is not zero. */
|
|
while (length) {
|
|
page_mask_result = (prp_entry_dma + prp_size) & page_mask;
|
|
if (!page_mask_result && (length > dev_pgsz)) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: single PRP page is not sufficient\n",
|
|
__func__);
|
|
goto err_out;
|
|
}
|
|
|
|
/* Need to handle if entry will be part of a page. */
|
|
offset = dma_addr & page_mask;
|
|
entry_len = dev_pgsz - offset;
|
|
|
|
if (prp_entry == prp1_entry) {
|
|
/*
|
|
* Must fill in the first PRP pointer (PRP1) before
|
|
* moving on.
|
|
*/
|
|
*prp1_entry = cpu_to_le64(dma_addr);
|
|
if (*prp1_entry & sgemod_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: PRP1 address collides with SGE modifier\n",
|
|
__func__);
|
|
goto err_out;
|
|
}
|
|
*prp1_entry &= ~sgemod_mask;
|
|
*prp1_entry |= sgemod_val;
|
|
|
|
/*
|
|
* Now point to the second PRP entry within the
|
|
* command (PRP2).
|
|
*/
|
|
prp_entry = prp2_entry;
|
|
} else if (prp_entry == prp2_entry) {
|
|
/*
|
|
* Should the PRP2 entry be a PRP List pointer or just
|
|
* a regular PRP pointer? If there is more than one
|
|
* more page of data, must use a PRP List pointer.
|
|
*/
|
|
if (length > dev_pgsz) {
|
|
/*
|
|
* PRP2 will contain a PRP List pointer because
|
|
* more PRP's are needed with this command. The
|
|
* list will start at the beginning of the
|
|
* contiguous buffer.
|
|
*/
|
|
*prp2_entry = cpu_to_le64(prp_entry_dma);
|
|
if (*prp2_entry & sgemod_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: PRP list address collides with SGE modifier\n",
|
|
__func__);
|
|
goto err_out;
|
|
}
|
|
*prp2_entry &= ~sgemod_mask;
|
|
*prp2_entry |= sgemod_val;
|
|
|
|
/*
|
|
* The next PRP Entry will be the start of the
|
|
* first PRP List.
|
|
*/
|
|
prp_entry = prp_page;
|
|
continue;
|
|
} else {
|
|
/*
|
|
* After this, the PRP Entries are complete.
|
|
* This command uses 2 PRP's and no PRP list.
|
|
*/
|
|
*prp2_entry = cpu_to_le64(dma_addr);
|
|
if (*prp2_entry & sgemod_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: PRP2 collides with SGE modifier\n",
|
|
__func__);
|
|
goto err_out;
|
|
}
|
|
*prp2_entry &= ~sgemod_mask;
|
|
*prp2_entry |= sgemod_val;
|
|
}
|
|
} else {
|
|
/*
|
|
* Put entry in list and bump the addresses.
|
|
*
|
|
* After PRP1 and PRP2 are filled in, this will fill in
|
|
* all remaining PRP entries in a PRP List, one per
|
|
* each time through the loop.
|
|
*/
|
|
*prp_entry = cpu_to_le64(dma_addr);
|
|
if (*prp_entry & sgemod_mask) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: PRP address collides with SGE modifier\n",
|
|
__func__);
|
|
goto err_out;
|
|
}
|
|
*prp_entry &= ~sgemod_mask;
|
|
*prp_entry |= sgemod_val;
|
|
prp_entry++;
|
|
prp_entry_dma += prp_size;
|
|
}
|
|
|
|
/* decrement length accounting for last partial page. */
|
|
if (entry_len >= length) {
|
|
length = 0;
|
|
} else {
|
|
if (entry_len <= desc_len) {
|
|
dma_addr += entry_len;
|
|
desc_len -= entry_len;
|
|
}
|
|
if (!desc_len) {
|
|
if ((++desc_count) >=
|
|
drv_buf_iter->num_dma_desc) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s: Invalid len %zd while building PRP\n",
|
|
__func__, length);
|
|
goto err_out;
|
|
}
|
|
dma_addr =
|
|
drv_buf_iter->dma_desc[desc_count].dma_addr;
|
|
desc_len =
|
|
drv_buf_iter->dma_desc[desc_count].size;
|
|
}
|
|
length -= entry_len;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
err_out:
|
|
if (mrioc->prp_list_virt) {
|
|
dma_free_coherent(&mrioc->pdev->dev, mrioc->prp_sz,
|
|
mrioc->prp_list_virt, mrioc->prp_list_dma);
|
|
mrioc->prp_list_virt = NULL;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_map_data_buffer_dma - build dma descriptors for data
|
|
* buffers
|
|
* @mrioc: Adapter instance reference
|
|
* @drv_buf: buffer map descriptor
|
|
* @desc_count: Number of already consumed dma descriptors
|
|
*
|
|
* This function computes how many pre-allocated DMA descriptors
|
|
* are required for the given data buffer and if those number of
|
|
* descriptors are free, then setup the mapping of the scattered
|
|
* DMA address to the given data buffer, if the data direction
|
|
* of the buffer is DMA_TO_DEVICE then the actual data is copied to
|
|
* the DMA buffers
|
|
*
|
|
* Return: 0 on success, -1 on failure
|
|
*/
|
|
static int mpi3mr_map_data_buffer_dma(struct mpi3mr_ioc *mrioc,
|
|
struct mpi3mr_buf_map *drv_buf,
|
|
u16 desc_count)
|
|
{
|
|
u16 i, needed_desc = drv_buf->kern_buf_len / MPI3MR_IOCTL_SGE_SIZE;
|
|
u32 buf_len = drv_buf->kern_buf_len, copied_len = 0;
|
|
|
|
if (drv_buf->kern_buf_len % MPI3MR_IOCTL_SGE_SIZE)
|
|
needed_desc++;
|
|
if ((needed_desc + desc_count) > MPI3MR_NUM_IOCTL_SGE) {
|
|
dprint_bsg_err(mrioc, "%s: DMA descriptor mapping error %d:%d:%d\n",
|
|
__func__, needed_desc, desc_count, MPI3MR_NUM_IOCTL_SGE);
|
|
return -1;
|
|
}
|
|
drv_buf->dma_desc = kzalloc(sizeof(*drv_buf->dma_desc) * needed_desc,
|
|
GFP_KERNEL);
|
|
if (!drv_buf->dma_desc)
|
|
return -1;
|
|
for (i = 0; i < needed_desc; i++, desc_count++) {
|
|
drv_buf->dma_desc[i].addr = mrioc->ioctl_sge[desc_count].addr;
|
|
drv_buf->dma_desc[i].dma_addr =
|
|
mrioc->ioctl_sge[desc_count].dma_addr;
|
|
if (buf_len < mrioc->ioctl_sge[desc_count].size)
|
|
drv_buf->dma_desc[i].size = buf_len;
|
|
else
|
|
drv_buf->dma_desc[i].size =
|
|
mrioc->ioctl_sge[desc_count].size;
|
|
buf_len -= drv_buf->dma_desc[i].size;
|
|
memset(drv_buf->dma_desc[i].addr, 0,
|
|
mrioc->ioctl_sge[desc_count].size);
|
|
if (drv_buf->data_dir == DMA_TO_DEVICE) {
|
|
memcpy(drv_buf->dma_desc[i].addr,
|
|
drv_buf->bsg_buf + copied_len,
|
|
drv_buf->dma_desc[i].size);
|
|
copied_len += drv_buf->dma_desc[i].size;
|
|
}
|
|
}
|
|
drv_buf->num_dma_desc = needed_desc;
|
|
return 0;
|
|
}
|
|
/**
|
|
* mpi3mr_bsg_process_mpt_cmds - MPI Pass through BSG handler
|
|
* @job: BSG job reference
|
|
*
|
|
* This function is the top level handler for MPI Pass through
|
|
* command, this does basic validation of the input data buffers,
|
|
* identifies the given buffer types and MPI command, allocates
|
|
* DMAable memory for user given buffers, construstcs SGL
|
|
* properly and passes the command to the firmware.
|
|
*
|
|
* Once the MPI command is completed the driver copies the data
|
|
* if any and reply, sense information to user provided buffers.
|
|
* If the command is timed out then issues controller reset
|
|
* prior to returning.
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
|
|
static long mpi3mr_bsg_process_mpt_cmds(struct bsg_job *job)
|
|
{
|
|
long rval = -EINVAL;
|
|
struct mpi3mr_ioc *mrioc = NULL;
|
|
u8 *mpi_req = NULL, *sense_buff_k = NULL;
|
|
u8 mpi_msg_size = 0;
|
|
struct mpi3mr_bsg_packet *bsg_req = NULL;
|
|
struct mpi3mr_bsg_mptcmd *karg;
|
|
struct mpi3mr_buf_entry *buf_entries = NULL;
|
|
struct mpi3mr_buf_map *drv_bufs = NULL, *drv_buf_iter = NULL;
|
|
u8 count, bufcnt = 0, is_rmcb = 0, is_rmrb = 0;
|
|
u8 din_cnt = 0, dout_cnt = 0;
|
|
u8 invalid_be = 0, erb_offset = 0xFF, mpirep_offset = 0xFF;
|
|
u8 block_io = 0, nvme_fmt = 0, resp_code = 0;
|
|
struct mpi3_request_header *mpi_header = NULL;
|
|
struct mpi3_status_reply_descriptor *status_desc;
|
|
struct mpi3_scsi_task_mgmt_request *tm_req;
|
|
u32 erbsz = MPI3MR_SENSE_BUF_SZ, tmplen;
|
|
u16 dev_handle;
|
|
struct mpi3mr_tgt_dev *tgtdev;
|
|
struct mpi3mr_stgt_priv_data *stgt_priv = NULL;
|
|
struct mpi3mr_bsg_in_reply_buf *bsg_reply_buf = NULL;
|
|
u32 din_size = 0, dout_size = 0;
|
|
u8 *din_buf = NULL, *dout_buf = NULL;
|
|
u8 *sgl_iter = NULL, *sgl_din_iter = NULL, *sgl_dout_iter = NULL;
|
|
u16 rmc_size = 0, desc_count = 0;
|
|
|
|
bsg_req = job->request;
|
|
karg = (struct mpi3mr_bsg_mptcmd *)&bsg_req->cmd.mptcmd;
|
|
|
|
mrioc = mpi3mr_bsg_verify_adapter(karg->mrioc_id);
|
|
if (!mrioc)
|
|
return -ENODEV;
|
|
|
|
if (mutex_lock_interruptible(&mrioc->bsg_cmds.mutex))
|
|
return -ERESTARTSYS;
|
|
|
|
if (mrioc->bsg_cmds.state & MPI3MR_CMD_PENDING) {
|
|
dprint_bsg_err(mrioc, "%s: command is in use\n", __func__);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
if (!mrioc->ioctl_sges_allocated) {
|
|
dprint_bsg_err(mrioc, "%s: DMA memory was not allocated\n",
|
|
__func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (karg->timeout < MPI3MR_APP_DEFAULT_TIMEOUT)
|
|
karg->timeout = MPI3MR_APP_DEFAULT_TIMEOUT;
|
|
|
|
mpi_req = kzalloc(MPI3MR_ADMIN_REQ_FRAME_SZ, GFP_KERNEL);
|
|
if (!mpi_req) {
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
return -ENOMEM;
|
|
}
|
|
mpi_header = (struct mpi3_request_header *)mpi_req;
|
|
|
|
bufcnt = karg->buf_entry_list.num_of_entries;
|
|
drv_bufs = kzalloc((sizeof(*drv_bufs) * bufcnt), GFP_KERNEL);
|
|
if (!drv_bufs) {
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
dout_buf = kzalloc(job->request_payload.payload_len,
|
|
GFP_KERNEL);
|
|
if (!dout_buf) {
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
din_buf = kzalloc(job->reply_payload.payload_len,
|
|
GFP_KERNEL);
|
|
if (!din_buf) {
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
sg_copy_to_buffer(job->request_payload.sg_list,
|
|
job->request_payload.sg_cnt,
|
|
dout_buf, job->request_payload.payload_len);
|
|
|
|
buf_entries = karg->buf_entry_list.buf_entry;
|
|
sgl_din_iter = din_buf;
|
|
sgl_dout_iter = dout_buf;
|
|
drv_buf_iter = drv_bufs;
|
|
|
|
for (count = 0; count < bufcnt; count++, buf_entries++, drv_buf_iter++) {
|
|
|
|
switch (buf_entries->buf_type) {
|
|
case MPI3MR_BSG_BUFTYPE_RAIDMGMT_CMD:
|
|
sgl_iter = sgl_dout_iter;
|
|
sgl_dout_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_TO_DEVICE;
|
|
is_rmcb = 1;
|
|
if ((count != 0) || !buf_entries->buf_len)
|
|
invalid_be = 1;
|
|
break;
|
|
case MPI3MR_BSG_BUFTYPE_RAIDMGMT_RESP:
|
|
sgl_iter = sgl_din_iter;
|
|
sgl_din_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_FROM_DEVICE;
|
|
is_rmrb = 1;
|
|
if (count != 1 || !is_rmcb || !buf_entries->buf_len)
|
|
invalid_be = 1;
|
|
break;
|
|
case MPI3MR_BSG_BUFTYPE_DATA_IN:
|
|
sgl_iter = sgl_din_iter;
|
|
sgl_din_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_FROM_DEVICE;
|
|
din_cnt++;
|
|
din_size += buf_entries->buf_len;
|
|
if ((din_cnt > 1) && !is_rmcb)
|
|
invalid_be = 1;
|
|
break;
|
|
case MPI3MR_BSG_BUFTYPE_DATA_OUT:
|
|
sgl_iter = sgl_dout_iter;
|
|
sgl_dout_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_TO_DEVICE;
|
|
dout_cnt++;
|
|
dout_size += buf_entries->buf_len;
|
|
if ((dout_cnt > 1) && !is_rmcb)
|
|
invalid_be = 1;
|
|
break;
|
|
case MPI3MR_BSG_BUFTYPE_MPI_REPLY:
|
|
sgl_iter = sgl_din_iter;
|
|
sgl_din_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_NONE;
|
|
mpirep_offset = count;
|
|
if (!buf_entries->buf_len)
|
|
invalid_be = 1;
|
|
break;
|
|
case MPI3MR_BSG_BUFTYPE_ERR_RESPONSE:
|
|
sgl_iter = sgl_din_iter;
|
|
sgl_din_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_NONE;
|
|
erb_offset = count;
|
|
if (!buf_entries->buf_len)
|
|
invalid_be = 1;
|
|
break;
|
|
case MPI3MR_BSG_BUFTYPE_MPI_REQUEST:
|
|
sgl_iter = sgl_dout_iter;
|
|
sgl_dout_iter += buf_entries->buf_len;
|
|
drv_buf_iter->data_dir = DMA_NONE;
|
|
mpi_msg_size = buf_entries->buf_len;
|
|
if ((!mpi_msg_size || (mpi_msg_size % 4)) ||
|
|
(mpi_msg_size > MPI3MR_ADMIN_REQ_FRAME_SZ)) {
|
|
dprint_bsg_err(mrioc, "%s: invalid MPI message size\n",
|
|
__func__);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
memcpy(mpi_req, sgl_iter, buf_entries->buf_len);
|
|
break;
|
|
default:
|
|
invalid_be = 1;
|
|
break;
|
|
}
|
|
if (invalid_be) {
|
|
dprint_bsg_err(mrioc, "%s: invalid buffer entries passed\n",
|
|
__func__);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
if (sgl_dout_iter > (dout_buf + job->request_payload.payload_len)) {
|
|
dprint_bsg_err(mrioc, "%s: data_out buffer length mismatch\n",
|
|
__func__);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
if (sgl_din_iter > (din_buf + job->reply_payload.payload_len)) {
|
|
dprint_bsg_err(mrioc, "%s: data_in buffer length mismatch\n",
|
|
__func__);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
drv_buf_iter->bsg_buf = sgl_iter;
|
|
drv_buf_iter->bsg_buf_len = buf_entries->buf_len;
|
|
}
|
|
|
|
if (is_rmcb && ((din_size + dout_size) > MPI3MR_MAX_APP_XFER_SIZE)) {
|
|
dprint_bsg_err(mrioc, "%s:%d: invalid data transfer size passed for function 0x%x din_size = %d, dout_size = %d\n",
|
|
__func__, __LINE__, mpi_header->function, din_size,
|
|
dout_size);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
if (din_size > MPI3MR_MAX_APP_XFER_SIZE) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s:%d: invalid data transfer size passed for function 0x%x din_size=%d\n",
|
|
__func__, __LINE__, mpi_header->function, din_size);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
if (dout_size > MPI3MR_MAX_APP_XFER_SIZE) {
|
|
dprint_bsg_err(mrioc,
|
|
"%s:%d: invalid data transfer size passed for function 0x%x dout_size = %d\n",
|
|
__func__, __LINE__, mpi_header->function, dout_size);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
if (mpi_header->function == MPI3_BSG_FUNCTION_SMP_PASSTHROUGH) {
|
|
if (din_size > MPI3MR_IOCTL_SGE_SIZE ||
|
|
dout_size > MPI3MR_IOCTL_SGE_SIZE) {
|
|
dprint_bsg_err(mrioc, "%s:%d: invalid message size passed:%d:%d:%d:%d\n",
|
|
__func__, __LINE__, din_cnt, dout_cnt, din_size,
|
|
dout_size);
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
rval = -EINVAL;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
drv_buf_iter = drv_bufs;
|
|
for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
|
|
if (drv_buf_iter->data_dir == DMA_NONE)
|
|
continue;
|
|
|
|
drv_buf_iter->kern_buf_len = drv_buf_iter->bsg_buf_len;
|
|
if (is_rmcb && !count) {
|
|
drv_buf_iter->kern_buf_len =
|
|
mrioc->ioctl_chain_sge.size;
|
|
drv_buf_iter->kern_buf =
|
|
mrioc->ioctl_chain_sge.addr;
|
|
drv_buf_iter->kern_buf_dma =
|
|
mrioc->ioctl_chain_sge.dma_addr;
|
|
drv_buf_iter->dma_desc = NULL;
|
|
drv_buf_iter->num_dma_desc = 0;
|
|
memset(drv_buf_iter->kern_buf, 0,
|
|
drv_buf_iter->kern_buf_len);
|
|
tmplen = min(drv_buf_iter->kern_buf_len,
|
|
drv_buf_iter->bsg_buf_len);
|
|
rmc_size = tmplen;
|
|
memcpy(drv_buf_iter->kern_buf, drv_buf_iter->bsg_buf, tmplen);
|
|
} else if (is_rmrb && (count == 1)) {
|
|
drv_buf_iter->kern_buf_len =
|
|
mrioc->ioctl_resp_sge.size;
|
|
drv_buf_iter->kern_buf =
|
|
mrioc->ioctl_resp_sge.addr;
|
|
drv_buf_iter->kern_buf_dma =
|
|
mrioc->ioctl_resp_sge.dma_addr;
|
|
drv_buf_iter->dma_desc = NULL;
|
|
drv_buf_iter->num_dma_desc = 0;
|
|
memset(drv_buf_iter->kern_buf, 0,
|
|
drv_buf_iter->kern_buf_len);
|
|
tmplen = min(drv_buf_iter->kern_buf_len,
|
|
drv_buf_iter->bsg_buf_len);
|
|
drv_buf_iter->kern_buf_len = tmplen;
|
|
memset(drv_buf_iter->bsg_buf, 0,
|
|
drv_buf_iter->bsg_buf_len);
|
|
} else {
|
|
if (!drv_buf_iter->kern_buf_len)
|
|
continue;
|
|
if (mpi3mr_map_data_buffer_dma(mrioc, drv_buf_iter, desc_count)) {
|
|
rval = -ENOMEM;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
dprint_bsg_err(mrioc, "%s:%d: mapping data buffers failed\n",
|
|
__func__, __LINE__);
|
|
goto out;
|
|
}
|
|
desc_count += drv_buf_iter->num_dma_desc;
|
|
}
|
|
}
|
|
|
|
if (erb_offset != 0xFF) {
|
|
sense_buff_k = kzalloc(erbsz, GFP_KERNEL);
|
|
if (!sense_buff_k) {
|
|
rval = -ENOMEM;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (mrioc->unrecoverable) {
|
|
dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
|
|
__func__);
|
|
rval = -EFAULT;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
if (mrioc->reset_in_progress) {
|
|
dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
|
|
rval = -EAGAIN;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
if (mrioc->stop_bsgs || mrioc->block_on_pci_err) {
|
|
dprint_bsg_err(mrioc, "%s: bsgs are blocked\n", __func__);
|
|
rval = -EAGAIN;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
|
|
if (mpi_header->function == MPI3_BSG_FUNCTION_NVME_ENCAPSULATED) {
|
|
nvme_fmt = mpi3mr_get_nvme_data_fmt(
|
|
(struct mpi3_nvme_encapsulated_request *)mpi_req);
|
|
if (nvme_fmt == MPI3MR_NVME_DATA_FORMAT_PRP) {
|
|
if (mpi3mr_build_nvme_prp(mrioc,
|
|
(struct mpi3_nvme_encapsulated_request *)mpi_req,
|
|
drv_bufs, bufcnt)) {
|
|
rval = -ENOMEM;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
} else if (nvme_fmt == MPI3MR_NVME_DATA_FORMAT_SGL1 ||
|
|
nvme_fmt == MPI3MR_NVME_DATA_FORMAT_SGL2) {
|
|
if (mpi3mr_build_nvme_sgl(mrioc,
|
|
(struct mpi3_nvme_encapsulated_request *)mpi_req,
|
|
drv_bufs, bufcnt)) {
|
|
rval = -EINVAL;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
} else {
|
|
dprint_bsg_err(mrioc,
|
|
"%s:invalid NVMe command format\n", __func__);
|
|
rval = -EINVAL;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
} else {
|
|
if (mpi3mr_bsg_build_sgl(mrioc, mpi_req, mpi_msg_size,
|
|
drv_bufs, bufcnt, is_rmcb, is_rmrb,
|
|
(dout_cnt + din_cnt))) {
|
|
dprint_bsg_err(mrioc, "%s: sgl build failed\n", __func__);
|
|
rval = -EAGAIN;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (mpi_header->function == MPI3_BSG_FUNCTION_SCSI_TASK_MGMT) {
|
|
tm_req = (struct mpi3_scsi_task_mgmt_request *)mpi_req;
|
|
if (tm_req->task_type !=
|
|
MPI3_SCSITASKMGMT_TASKTYPE_ABORT_TASK) {
|
|
dev_handle = tm_req->dev_handle;
|
|
block_io = 1;
|
|
}
|
|
}
|
|
if (block_io) {
|
|
tgtdev = mpi3mr_get_tgtdev_by_handle(mrioc, dev_handle);
|
|
if (tgtdev && tgtdev->starget && tgtdev->starget->hostdata) {
|
|
stgt_priv = (struct mpi3mr_stgt_priv_data *)
|
|
tgtdev->starget->hostdata;
|
|
atomic_inc(&stgt_priv->block_io);
|
|
mpi3mr_tgtdev_put(tgtdev);
|
|
}
|
|
}
|
|
|
|
mrioc->bsg_cmds.state = MPI3MR_CMD_PENDING;
|
|
mrioc->bsg_cmds.is_waiting = 1;
|
|
mrioc->bsg_cmds.callback = NULL;
|
|
mrioc->bsg_cmds.is_sense = 0;
|
|
mrioc->bsg_cmds.sensebuf = sense_buff_k;
|
|
memset(mrioc->bsg_cmds.reply, 0, mrioc->reply_sz);
|
|
mpi_header->host_tag = cpu_to_le16(MPI3MR_HOSTTAG_BSG_CMDS);
|
|
if (mrioc->logging_level & MPI3_DEBUG_BSG_INFO) {
|
|
dprint_bsg_info(mrioc,
|
|
"%s: posting bsg request to the controller\n", __func__);
|
|
dprint_dump(mpi_req, MPI3MR_ADMIN_REQ_FRAME_SZ,
|
|
"bsg_mpi3_req");
|
|
if (mpi_header->function == MPI3_BSG_FUNCTION_MGMT_PASSTHROUGH) {
|
|
drv_buf_iter = &drv_bufs[0];
|
|
dprint_dump(drv_buf_iter->kern_buf,
|
|
rmc_size, "mpi3_mgmt_req");
|
|
}
|
|
}
|
|
|
|
init_completion(&mrioc->bsg_cmds.done);
|
|
rval = mpi3mr_admin_request_post(mrioc, mpi_req,
|
|
MPI3MR_ADMIN_REQ_FRAME_SZ, 0);
|
|
|
|
|
|
if (rval) {
|
|
mrioc->bsg_cmds.is_waiting = 0;
|
|
dprint_bsg_err(mrioc,
|
|
"%s: posting bsg request is failed\n", __func__);
|
|
rval = -EAGAIN;
|
|
goto out_unlock;
|
|
}
|
|
wait_for_completion_timeout(&mrioc->bsg_cmds.done,
|
|
(karg->timeout * HZ));
|
|
if (block_io && stgt_priv)
|
|
atomic_dec(&stgt_priv->block_io);
|
|
if (!(mrioc->bsg_cmds.state & MPI3MR_CMD_COMPLETE)) {
|
|
mrioc->bsg_cmds.is_waiting = 0;
|
|
rval = -EAGAIN;
|
|
if (mrioc->bsg_cmds.state & MPI3MR_CMD_RESET)
|
|
goto out_unlock;
|
|
if (((mpi_header->function != MPI3_FUNCTION_SCSI_IO) &&
|
|
(mpi_header->function != MPI3_FUNCTION_NVME_ENCAPSULATED))
|
|
|| (mrioc->logging_level & MPI3_DEBUG_BSG_ERROR)) {
|
|
ioc_info(mrioc, "%s: bsg request timedout after %d seconds\n",
|
|
__func__, karg->timeout);
|
|
if (!(mrioc->logging_level & MPI3_DEBUG_BSG_INFO)) {
|
|
dprint_dump(mpi_req, MPI3MR_ADMIN_REQ_FRAME_SZ,
|
|
"bsg_mpi3_req");
|
|
if (mpi_header->function ==
|
|
MPI3_FUNCTION_MGMT_PASSTHROUGH) {
|
|
drv_buf_iter = &drv_bufs[0];
|
|
dprint_dump(drv_buf_iter->kern_buf,
|
|
rmc_size, "mpi3_mgmt_req");
|
|
}
|
|
}
|
|
}
|
|
if ((mpi_header->function == MPI3_BSG_FUNCTION_NVME_ENCAPSULATED) ||
|
|
(mpi_header->function == MPI3_BSG_FUNCTION_SCSI_IO)) {
|
|
dprint_bsg_err(mrioc, "%s: bsg request timedout after %d seconds,\n"
|
|
"issuing target reset to (0x%04x)\n", __func__,
|
|
karg->timeout, mpi_header->function_dependent);
|
|
mpi3mr_issue_tm(mrioc,
|
|
MPI3_SCSITASKMGMT_TASKTYPE_TARGET_RESET,
|
|
mpi_header->function_dependent, 0,
|
|
MPI3MR_HOSTTAG_BLK_TMS, MPI3MR_RESETTM_TIMEOUT,
|
|
&mrioc->host_tm_cmds, &resp_code, NULL);
|
|
}
|
|
if (!(mrioc->bsg_cmds.state & MPI3MR_CMD_COMPLETE) &&
|
|
!(mrioc->bsg_cmds.state & MPI3MR_CMD_RESET))
|
|
mpi3mr_soft_reset_handler(mrioc,
|
|
MPI3MR_RESET_FROM_APP_TIMEOUT, 1);
|
|
goto out_unlock;
|
|
}
|
|
dprint_bsg_info(mrioc, "%s: bsg request is completed\n", __func__);
|
|
|
|
if (mrioc->prp_list_virt) {
|
|
dma_free_coherent(&mrioc->pdev->dev, mrioc->prp_sz,
|
|
mrioc->prp_list_virt, mrioc->prp_list_dma);
|
|
mrioc->prp_list_virt = NULL;
|
|
}
|
|
|
|
if ((mrioc->bsg_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
|
|
!= MPI3_IOCSTATUS_SUCCESS) {
|
|
dprint_bsg_info(mrioc,
|
|
"%s: command failed, ioc_status(0x%04x) log_info(0x%08x)\n",
|
|
__func__,
|
|
(mrioc->bsg_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK),
|
|
mrioc->bsg_cmds.ioc_loginfo);
|
|
}
|
|
|
|
if ((mpirep_offset != 0xFF) &&
|
|
drv_bufs[mpirep_offset].bsg_buf_len) {
|
|
drv_buf_iter = &drv_bufs[mpirep_offset];
|
|
drv_buf_iter->kern_buf_len = (sizeof(*bsg_reply_buf) +
|
|
mrioc->reply_sz);
|
|
bsg_reply_buf = kzalloc(drv_buf_iter->kern_buf_len, GFP_KERNEL);
|
|
|
|
if (!bsg_reply_buf) {
|
|
rval = -ENOMEM;
|
|
goto out_unlock;
|
|
}
|
|
if (mrioc->bsg_cmds.state & MPI3MR_CMD_REPLY_VALID) {
|
|
bsg_reply_buf->mpi_reply_type =
|
|
MPI3MR_BSG_MPI_REPLY_BUFTYPE_ADDRESS;
|
|
memcpy(bsg_reply_buf->reply_buf,
|
|
mrioc->bsg_cmds.reply, mrioc->reply_sz);
|
|
} else {
|
|
bsg_reply_buf->mpi_reply_type =
|
|
MPI3MR_BSG_MPI_REPLY_BUFTYPE_STATUS;
|
|
status_desc = (struct mpi3_status_reply_descriptor *)
|
|
bsg_reply_buf->reply_buf;
|
|
status_desc->ioc_status = mrioc->bsg_cmds.ioc_status;
|
|
status_desc->ioc_log_info = mrioc->bsg_cmds.ioc_loginfo;
|
|
}
|
|
tmplen = min(drv_buf_iter->kern_buf_len,
|
|
drv_buf_iter->bsg_buf_len);
|
|
memcpy(drv_buf_iter->bsg_buf, bsg_reply_buf, tmplen);
|
|
}
|
|
|
|
if (erb_offset != 0xFF && mrioc->bsg_cmds.sensebuf &&
|
|
mrioc->bsg_cmds.is_sense) {
|
|
drv_buf_iter = &drv_bufs[erb_offset];
|
|
tmplen = min(erbsz, drv_buf_iter->bsg_buf_len);
|
|
memcpy(drv_buf_iter->bsg_buf, sense_buff_k, tmplen);
|
|
}
|
|
|
|
drv_buf_iter = drv_bufs;
|
|
for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
|
|
if (drv_buf_iter->data_dir == DMA_NONE)
|
|
continue;
|
|
if ((count == 1) && is_rmrb) {
|
|
memcpy(drv_buf_iter->bsg_buf,
|
|
drv_buf_iter->kern_buf,
|
|
drv_buf_iter->kern_buf_len);
|
|
} else if (drv_buf_iter->data_dir == DMA_FROM_DEVICE) {
|
|
tmplen = 0;
|
|
for (desc_count = 0;
|
|
desc_count < drv_buf_iter->num_dma_desc;
|
|
desc_count++) {
|
|
memcpy(((u8 *)drv_buf_iter->bsg_buf + tmplen),
|
|
drv_buf_iter->dma_desc[desc_count].addr,
|
|
drv_buf_iter->dma_desc[desc_count].size);
|
|
tmplen +=
|
|
drv_buf_iter->dma_desc[desc_count].size;
|
|
}
|
|
}
|
|
}
|
|
|
|
out_unlock:
|
|
if (din_buf) {
|
|
job->reply_payload_rcv_len =
|
|
sg_copy_from_buffer(job->reply_payload.sg_list,
|
|
job->reply_payload.sg_cnt,
|
|
din_buf, job->reply_payload.payload_len);
|
|
}
|
|
mrioc->bsg_cmds.is_sense = 0;
|
|
mrioc->bsg_cmds.sensebuf = NULL;
|
|
mrioc->bsg_cmds.state = MPI3MR_CMD_NOTUSED;
|
|
mutex_unlock(&mrioc->bsg_cmds.mutex);
|
|
out:
|
|
kfree(sense_buff_k);
|
|
kfree(dout_buf);
|
|
kfree(din_buf);
|
|
kfree(mpi_req);
|
|
if (drv_bufs) {
|
|
drv_buf_iter = drv_bufs;
|
|
for (count = 0; count < bufcnt; count++, drv_buf_iter++)
|
|
kfree(drv_buf_iter->dma_desc);
|
|
kfree(drv_bufs);
|
|
}
|
|
kfree(bsg_reply_buf);
|
|
return rval;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_app_save_logdata - Save Log Data events
|
|
* @mrioc: Adapter instance reference
|
|
* @event_data: event data associated with log data event
|
|
* @event_data_size: event data size to copy
|
|
*
|
|
* If log data event caching is enabled by the applicatiobns,
|
|
* then this function saves the log data in the circular queue
|
|
* and Sends async signal SIGIO to indicate there is an async
|
|
* event from the firmware to the event monitoring applications.
|
|
*
|
|
* Return:Nothing
|
|
*/
|
|
void mpi3mr_app_save_logdata(struct mpi3mr_ioc *mrioc, char *event_data,
|
|
u16 event_data_size)
|
|
{
|
|
u32 index = mrioc->logdata_buf_idx, sz;
|
|
struct mpi3mr_logdata_entry *entry;
|
|
|
|
if (!(mrioc->logdata_buf))
|
|
return;
|
|
|
|
entry = (struct mpi3mr_logdata_entry *)
|
|
(mrioc->logdata_buf + (index * mrioc->logdata_entry_sz));
|
|
entry->valid_entry = 1;
|
|
sz = min(mrioc->logdata_entry_sz, event_data_size);
|
|
memcpy(entry->data, event_data, sz);
|
|
mrioc->logdata_buf_idx =
|
|
((++index) % MPI3MR_BSG_LOGDATA_MAX_ENTRIES);
|
|
atomic64_inc(&event_counter);
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_request - bsg request entry point
|
|
* @job: BSG job reference
|
|
*
|
|
* This is driver's entry point for bsg requests
|
|
*
|
|
* Return: 0 on success and proper error codes on failure
|
|
*/
|
|
static int mpi3mr_bsg_request(struct bsg_job *job)
|
|
{
|
|
long rval = -EINVAL;
|
|
unsigned int reply_payload_rcv_len = 0;
|
|
|
|
struct mpi3mr_bsg_packet *bsg_req = job->request;
|
|
|
|
switch (bsg_req->cmd_type) {
|
|
case MPI3MR_DRV_CMD:
|
|
rval = mpi3mr_bsg_process_drv_cmds(job);
|
|
break;
|
|
case MPI3MR_MPT_CMD:
|
|
rval = mpi3mr_bsg_process_mpt_cmds(job);
|
|
break;
|
|
default:
|
|
pr_err("%s: unsupported BSG command(0x%08x)\n",
|
|
MPI3MR_DRIVER_NAME, bsg_req->cmd_type);
|
|
break;
|
|
}
|
|
|
|
bsg_job_done(job, rval, reply_payload_rcv_len);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_exit - de-registration from bsg layer
|
|
* @mrioc: Adapter instance reference
|
|
*
|
|
* This will be called during driver unload and all
|
|
* bsg resources allocated during load will be freed.
|
|
*
|
|
* Return:Nothing
|
|
*/
|
|
void mpi3mr_bsg_exit(struct mpi3mr_ioc *mrioc)
|
|
{
|
|
struct device *bsg_dev = &mrioc->bsg_dev;
|
|
if (!mrioc->bsg_queue)
|
|
return;
|
|
|
|
bsg_remove_queue(mrioc->bsg_queue);
|
|
mrioc->bsg_queue = NULL;
|
|
|
|
device_del(bsg_dev);
|
|
put_device(bsg_dev);
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_node_release -release bsg device node
|
|
* @dev: bsg device node
|
|
*
|
|
* decrements bsg dev parent reference count
|
|
*
|
|
* Return:Nothing
|
|
*/
|
|
static void mpi3mr_bsg_node_release(struct device *dev)
|
|
{
|
|
put_device(dev->parent);
|
|
}
|
|
|
|
/**
|
|
* mpi3mr_bsg_init - registration with bsg layer
|
|
* @mrioc: Adapter instance reference
|
|
*
|
|
* This will be called during driver load and it will
|
|
* register driver with bsg layer
|
|
*
|
|
* Return:Nothing
|
|
*/
|
|
void mpi3mr_bsg_init(struct mpi3mr_ioc *mrioc)
|
|
{
|
|
struct device *bsg_dev = &mrioc->bsg_dev;
|
|
struct device *parent = &mrioc->shost->shost_gendev;
|
|
struct queue_limits lim = {
|
|
.max_hw_sectors = MPI3MR_MAX_APP_XFER_SECTORS,
|
|
.max_segments = MPI3MR_MAX_APP_XFER_SEGMENTS,
|
|
};
|
|
|
|
device_initialize(bsg_dev);
|
|
|
|
bsg_dev->parent = get_device(parent);
|
|
bsg_dev->release = mpi3mr_bsg_node_release;
|
|
|
|
dev_set_name(bsg_dev, "mpi3mrctl%u", mrioc->id);
|
|
|
|
if (device_add(bsg_dev)) {
|
|
ioc_err(mrioc, "%s: bsg device add failed\n",
|
|
dev_name(bsg_dev));
|
|
put_device(bsg_dev);
|
|
return;
|
|
}
|
|
|
|
mrioc->bsg_queue = bsg_setup_queue(bsg_dev, dev_name(bsg_dev), &lim,
|
|
mpi3mr_bsg_request, NULL, 0);
|
|
if (IS_ERR(mrioc->bsg_queue)) {
|
|
ioc_err(mrioc, "%s: bsg registration failed\n",
|
|
dev_name(bsg_dev));
|
|
device_del(bsg_dev);
|
|
put_device(bsg_dev);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* version_fw_show - SysFS callback for firmware version read
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying firmware version
|
|
*/
|
|
static ssize_t
|
|
version_fw_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
struct mpi3mr_compimg_ver *fwver = &mrioc->facts.fw_ver;
|
|
|
|
return sysfs_emit(buf, "%d.%d.%d.%d.%05d-%05d\n",
|
|
fwver->gen_major, fwver->gen_minor, fwver->ph_major,
|
|
fwver->ph_minor, fwver->cust_id, fwver->build_num);
|
|
}
|
|
static DEVICE_ATTR_RO(version_fw);
|
|
|
|
/**
|
|
* fw_queue_depth_show - SysFS callback for firmware max cmds
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying firmware max commands
|
|
*/
|
|
static ssize_t
|
|
fw_queue_depth_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
|
|
return sysfs_emit(buf, "%d\n", mrioc->facts.max_reqs);
|
|
}
|
|
static DEVICE_ATTR_RO(fw_queue_depth);
|
|
|
|
/**
|
|
* op_req_q_count_show - SysFS callback for request queue count
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying request queue count
|
|
*/
|
|
static ssize_t
|
|
op_req_q_count_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
|
|
return sysfs_emit(buf, "%d\n", mrioc->num_op_req_q);
|
|
}
|
|
static DEVICE_ATTR_RO(op_req_q_count);
|
|
|
|
/**
|
|
* reply_queue_count_show - SysFS callback for reply queue count
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying reply queue count
|
|
*/
|
|
static ssize_t
|
|
reply_queue_count_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
|
|
return sysfs_emit(buf, "%d\n", mrioc->num_op_reply_q);
|
|
}
|
|
|
|
static DEVICE_ATTR_RO(reply_queue_count);
|
|
|
|
/**
|
|
* logging_level_show - Show controller debug level
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* A sysfs 'read/write' shost attribute, to show the current
|
|
* debug log level used by the driver for the specific
|
|
* controller.
|
|
*
|
|
* Return: sysfs_emit() return
|
|
*/
|
|
static ssize_t
|
|
logging_level_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
|
|
return sysfs_emit(buf, "%08xh\n", mrioc->logging_level);
|
|
}
|
|
|
|
/**
|
|
* logging_level_store- Change controller debug level
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
* @count: size of the buffer
|
|
*
|
|
* A sysfs 'read/write' shost attribute, to change the current
|
|
* debug log level used by the driver for the specific
|
|
* controller.
|
|
*
|
|
* Return: strlen() return
|
|
*/
|
|
static ssize_t
|
|
logging_level_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
int val = 0;
|
|
|
|
if (kstrtoint(buf, 0, &val) != 0)
|
|
return -EINVAL;
|
|
|
|
mrioc->logging_level = val;
|
|
ioc_info(mrioc, "logging_level=%08xh\n", mrioc->logging_level);
|
|
return strlen(buf);
|
|
}
|
|
static DEVICE_ATTR_RW(logging_level);
|
|
|
|
/**
|
|
* adp_state_show() - SysFS callback for adapter state show
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying adapter state
|
|
*/
|
|
static ssize_t
|
|
adp_state_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
struct mpi3mr_ioc *mrioc = shost_priv(shost);
|
|
enum mpi3mr_iocstate ioc_state;
|
|
uint8_t adp_state;
|
|
|
|
ioc_state = mpi3mr_get_iocstate(mrioc);
|
|
if (ioc_state == MRIOC_STATE_UNRECOVERABLE)
|
|
adp_state = MPI3MR_BSG_ADPSTATE_UNRECOVERABLE;
|
|
else if (mrioc->reset_in_progress || mrioc->stop_bsgs ||
|
|
mrioc->block_on_pci_err)
|
|
adp_state = MPI3MR_BSG_ADPSTATE_IN_RESET;
|
|
else if (ioc_state == MRIOC_STATE_FAULT)
|
|
adp_state = MPI3MR_BSG_ADPSTATE_FAULT;
|
|
else
|
|
adp_state = MPI3MR_BSG_ADPSTATE_OPERATIONAL;
|
|
|
|
return sysfs_emit(buf, "%u\n", adp_state);
|
|
}
|
|
|
|
static DEVICE_ATTR_RO(adp_state);
|
|
|
|
static struct attribute *mpi3mr_host_attrs[] = {
|
|
&dev_attr_version_fw.attr,
|
|
&dev_attr_fw_queue_depth.attr,
|
|
&dev_attr_op_req_q_count.attr,
|
|
&dev_attr_reply_queue_count.attr,
|
|
&dev_attr_logging_level.attr,
|
|
&dev_attr_adp_state.attr,
|
|
NULL,
|
|
};
|
|
|
|
static const struct attribute_group mpi3mr_host_attr_group = {
|
|
.attrs = mpi3mr_host_attrs
|
|
};
|
|
|
|
const struct attribute_group *mpi3mr_host_groups[] = {
|
|
&mpi3mr_host_attr_group,
|
|
NULL,
|
|
};
|
|
|
|
|
|
/*
|
|
* SCSI Device attributes under sysfs
|
|
*/
|
|
|
|
/**
|
|
* sas_address_show - SysFS callback for dev SASaddress display
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying SAS address of the
|
|
* specific SAS/SATA end device.
|
|
*/
|
|
static ssize_t
|
|
sas_address_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct scsi_device *sdev = to_scsi_device(dev);
|
|
struct mpi3mr_sdev_priv_data *sdev_priv_data;
|
|
struct mpi3mr_stgt_priv_data *tgt_priv_data;
|
|
struct mpi3mr_tgt_dev *tgtdev;
|
|
|
|
sdev_priv_data = sdev->hostdata;
|
|
if (!sdev_priv_data)
|
|
return 0;
|
|
|
|
tgt_priv_data = sdev_priv_data->tgt_priv_data;
|
|
if (!tgt_priv_data)
|
|
return 0;
|
|
tgtdev = tgt_priv_data->tgt_dev;
|
|
if (!tgtdev || tgtdev->dev_type != MPI3_DEVICE_DEVFORM_SAS_SATA)
|
|
return 0;
|
|
return sysfs_emit(buf, "0x%016llx\n",
|
|
(unsigned long long)tgtdev->dev_spec.sas_sata_inf.sas_address);
|
|
}
|
|
|
|
static DEVICE_ATTR_RO(sas_address);
|
|
|
|
/**
|
|
* device_handle_show - SysFS callback for device handle display
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying firmware internal
|
|
* device handle of the specific device.
|
|
*/
|
|
static ssize_t
|
|
device_handle_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct scsi_device *sdev = to_scsi_device(dev);
|
|
struct mpi3mr_sdev_priv_data *sdev_priv_data;
|
|
struct mpi3mr_stgt_priv_data *tgt_priv_data;
|
|
struct mpi3mr_tgt_dev *tgtdev;
|
|
|
|
sdev_priv_data = sdev->hostdata;
|
|
if (!sdev_priv_data)
|
|
return 0;
|
|
|
|
tgt_priv_data = sdev_priv_data->tgt_priv_data;
|
|
if (!tgt_priv_data)
|
|
return 0;
|
|
tgtdev = tgt_priv_data->tgt_dev;
|
|
if (!tgtdev)
|
|
return 0;
|
|
return sysfs_emit(buf, "0x%04x\n", tgtdev->dev_handle);
|
|
}
|
|
|
|
static DEVICE_ATTR_RO(device_handle);
|
|
|
|
/**
|
|
* persistent_id_show - SysFS callback for persisten ID display
|
|
* @dev: class device
|
|
* @attr: Device attributes
|
|
* @buf: Buffer to copy
|
|
*
|
|
* Return: sysfs_emit() return after copying persistent ID of the
|
|
* of the specific device.
|
|
*/
|
|
static ssize_t
|
|
persistent_id_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct scsi_device *sdev = to_scsi_device(dev);
|
|
struct mpi3mr_sdev_priv_data *sdev_priv_data;
|
|
struct mpi3mr_stgt_priv_data *tgt_priv_data;
|
|
struct mpi3mr_tgt_dev *tgtdev;
|
|
|
|
sdev_priv_data = sdev->hostdata;
|
|
if (!sdev_priv_data)
|
|
return 0;
|
|
|
|
tgt_priv_data = sdev_priv_data->tgt_priv_data;
|
|
if (!tgt_priv_data)
|
|
return 0;
|
|
tgtdev = tgt_priv_data->tgt_dev;
|
|
if (!tgtdev)
|
|
return 0;
|
|
return sysfs_emit(buf, "%d\n", tgtdev->perst_id);
|
|
}
|
|
static DEVICE_ATTR_RO(persistent_id);
|
|
|
|
/**
|
|
* sas_ncq_prio_supported_show - Indicate if device supports NCQ priority
|
|
* @dev: pointer to embedded device
|
|
* @attr: sas_ncq_prio_supported attribute descriptor
|
|
* @buf: the buffer returned
|
|
*
|
|
* A sysfs 'read-only' sdev attribute, only works with SATA devices
|
|
*/
|
|
static ssize_t
|
|
sas_ncq_prio_supported_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct scsi_device *sdev = to_scsi_device(dev);
|
|
|
|
return sysfs_emit(buf, "%d\n", sas_ata_ncq_prio_supported(sdev));
|
|
}
|
|
static DEVICE_ATTR_RO(sas_ncq_prio_supported);
|
|
|
|
/**
|
|
* sas_ncq_prio_enable_show - send prioritized io commands to device
|
|
* @dev: pointer to embedded device
|
|
* @attr: sas_ncq_prio_enable attribute descriptor
|
|
* @buf: the buffer returned
|
|
*
|
|
* A sysfs 'read/write' sdev attribute, only works with SATA devices
|
|
*/
|
|
static ssize_t
|
|
sas_ncq_prio_enable_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct scsi_device *sdev = to_scsi_device(dev);
|
|
struct mpi3mr_sdev_priv_data *sdev_priv_data = sdev->hostdata;
|
|
|
|
if (!sdev_priv_data)
|
|
return 0;
|
|
|
|
return sysfs_emit(buf, "%d\n", sdev_priv_data->ncq_prio_enable);
|
|
}
|
|
|
|
static ssize_t
|
|
sas_ncq_prio_enable_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct scsi_device *sdev = to_scsi_device(dev);
|
|
struct mpi3mr_sdev_priv_data *sdev_priv_data = sdev->hostdata;
|
|
bool ncq_prio_enable = 0;
|
|
|
|
if (kstrtobool(buf, &ncq_prio_enable))
|
|
return -EINVAL;
|
|
|
|
if (!sas_ata_ncq_prio_supported(sdev))
|
|
return -EINVAL;
|
|
|
|
sdev_priv_data->ncq_prio_enable = ncq_prio_enable;
|
|
|
|
return strlen(buf);
|
|
}
|
|
static DEVICE_ATTR_RW(sas_ncq_prio_enable);
|
|
|
|
static struct attribute *mpi3mr_dev_attrs[] = {
|
|
&dev_attr_sas_address.attr,
|
|
&dev_attr_device_handle.attr,
|
|
&dev_attr_persistent_id.attr,
|
|
&dev_attr_sas_ncq_prio_supported.attr,
|
|
&dev_attr_sas_ncq_prio_enable.attr,
|
|
NULL,
|
|
};
|
|
|
|
static const struct attribute_group mpi3mr_dev_attr_group = {
|
|
.attrs = mpi3mr_dev_attrs
|
|
};
|
|
|
|
const struct attribute_group *mpi3mr_dev_groups[] = {
|
|
&mpi3mr_dev_attr_group,
|
|
NULL,
|
|
};
|