IB/srp: Create an insecure all physical rkey only if needed
The SRP initiator only needs this if the insecure register_always=N performance optimization is enabled, or if FRWR/FMR is not supported in the driver. Do not create an all physical MR unless it is needed to support either of those modes. Default register_always to true so the out of the box configuration does not create an insecure all physical MR. Signed-off-by: Jason Gunthorpe <jgunthorpe@obsidianresearch.com> [bvanassche: reworked and rebased this patch] Signed-off-by: Bart Van Assche <bart.vanassche@sandisk.com> Signed-off-by: Doug Ledford <dledford@redhat.com>
This commit is contained in:
parent
330179f2fa
commit
03f6fb93fd
2 changed files with 25 additions and 19 deletions
|
@ -68,8 +68,8 @@ static unsigned int srp_sg_tablesize;
|
||||||
static unsigned int cmd_sg_entries;
|
static unsigned int cmd_sg_entries;
|
||||||
static unsigned int indirect_sg_entries;
|
static unsigned int indirect_sg_entries;
|
||||||
static bool allow_ext_sg;
|
static bool allow_ext_sg;
|
||||||
static bool prefer_fr;
|
static bool prefer_fr = true;
|
||||||
static bool register_always;
|
static bool register_always = true;
|
||||||
static int topspin_workarounds = 1;
|
static int topspin_workarounds = 1;
|
||||||
|
|
||||||
module_param(srp_sg_tablesize, uint, 0444);
|
module_param(srp_sg_tablesize, uint, 0444);
|
||||||
|
@ -1353,9 +1353,9 @@ static int srp_finish_mapping(struct srp_map_state *state,
|
||||||
if (state->npages == 0)
|
if (state->npages == 0)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (state->npages == 1 && !register_always)
|
if (state->npages == 1 && target->global_mr)
|
||||||
srp_map_desc(state, state->base_dma_addr, state->dma_len,
|
srp_map_desc(state, state->base_dma_addr, state->dma_len,
|
||||||
target->rkey);
|
target->global_mr->rkey);
|
||||||
else
|
else
|
||||||
ret = dev->use_fast_reg ? srp_map_finish_fr(state, ch) :
|
ret = dev->use_fast_reg ? srp_map_finish_fr(state, ch) :
|
||||||
srp_map_finish_fmr(state, ch);
|
srp_map_finish_fmr(state, ch);
|
||||||
|
@ -1442,7 +1442,8 @@ static int srp_map_sg(struct srp_map_state *state, struct srp_rdma_ch *ch,
|
||||||
} else {
|
} else {
|
||||||
for_each_sg(scat, sg, count, i) {
|
for_each_sg(scat, sg, count, i) {
|
||||||
srp_map_desc(state, ib_sg_dma_address(dev->dev, sg),
|
srp_map_desc(state, ib_sg_dma_address(dev->dev, sg),
|
||||||
ib_sg_dma_len(dev->dev, sg), target->rkey);
|
ib_sg_dma_len(dev->dev, sg),
|
||||||
|
target->global_mr->rkey);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1531,7 +1532,7 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch,
|
||||||
fmt = SRP_DATA_DESC_DIRECT;
|
fmt = SRP_DATA_DESC_DIRECT;
|
||||||
len = sizeof (struct srp_cmd) + sizeof (struct srp_direct_buf);
|
len = sizeof (struct srp_cmd) + sizeof (struct srp_direct_buf);
|
||||||
|
|
||||||
if (count == 1 && !register_always) {
|
if (count == 1 && target->global_mr) {
|
||||||
/*
|
/*
|
||||||
* The midlayer only generated a single gather/scatter
|
* The midlayer only generated a single gather/scatter
|
||||||
* entry, or DMA mapping coalesced everything to a
|
* entry, or DMA mapping coalesced everything to a
|
||||||
|
@ -1541,7 +1542,7 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch,
|
||||||
struct srp_direct_buf *buf = (void *) cmd->add_data;
|
struct srp_direct_buf *buf = (void *) cmd->add_data;
|
||||||
|
|
||||||
buf->va = cpu_to_be64(ib_sg_dma_address(ibdev, scat));
|
buf->va = cpu_to_be64(ib_sg_dma_address(ibdev, scat));
|
||||||
buf->key = cpu_to_be32(target->rkey);
|
buf->key = cpu_to_be32(target->global_mr->rkey);
|
||||||
buf->len = cpu_to_be32(ib_sg_dma_len(ibdev, scat));
|
buf->len = cpu_to_be32(ib_sg_dma_len(ibdev, scat));
|
||||||
|
|
||||||
req->nmdesc = 0;
|
req->nmdesc = 0;
|
||||||
|
@ -1595,14 +1596,14 @@ static int srp_map_data(struct scsi_cmnd *scmnd, struct srp_rdma_ch *ch,
|
||||||
memcpy(indirect_hdr->desc_list, req->indirect_desc,
|
memcpy(indirect_hdr->desc_list, req->indirect_desc,
|
||||||
count * sizeof (struct srp_direct_buf));
|
count * sizeof (struct srp_direct_buf));
|
||||||
|
|
||||||
if (register_always && (dev->use_fast_reg || dev->use_fmr)) {
|
if (!target->global_mr) {
|
||||||
ret = srp_map_idb(ch, req, state.gen.next, state.gen.end,
|
ret = srp_map_idb(ch, req, state.gen.next, state.gen.end,
|
||||||
idb_len, &idb_rkey);
|
idb_len, &idb_rkey);
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
req->nmdesc++;
|
req->nmdesc++;
|
||||||
} else {
|
} else {
|
||||||
idb_rkey = target->rkey;
|
idb_rkey = target->global_mr->rkey;
|
||||||
}
|
}
|
||||||
|
|
||||||
indirect_hdr->table_desc.va = cpu_to_be64(req->indirect_dma_addr);
|
indirect_hdr->table_desc.va = cpu_to_be64(req->indirect_dma_addr);
|
||||||
|
@ -3157,7 +3158,7 @@ static ssize_t srp_create_target(struct device *dev,
|
||||||
target->scsi_host = target_host;
|
target->scsi_host = target_host;
|
||||||
target->srp_host = host;
|
target->srp_host = host;
|
||||||
target->lkey = host->srp_dev->pd->local_dma_lkey;
|
target->lkey = host->srp_dev->pd->local_dma_lkey;
|
||||||
target->rkey = host->srp_dev->mr->rkey;
|
target->global_mr = host->srp_dev->global_mr;
|
||||||
target->cmd_sg_cnt = cmd_sg_entries;
|
target->cmd_sg_cnt = cmd_sg_entries;
|
||||||
target->sg_tablesize = indirect_sg_entries ? : cmd_sg_entries;
|
target->sg_tablesize = indirect_sg_entries ? : cmd_sg_entries;
|
||||||
target->allow_ext_sg = allow_ext_sg;
|
target->allow_ext_sg = allow_ext_sg;
|
||||||
|
@ -3447,12 +3448,16 @@ static void srp_add_one(struct ib_device *device)
|
||||||
if (IS_ERR(srp_dev->pd))
|
if (IS_ERR(srp_dev->pd))
|
||||||
goto free_dev;
|
goto free_dev;
|
||||||
|
|
||||||
srp_dev->mr = ib_get_dma_mr(srp_dev->pd,
|
if (!register_always || (!srp_dev->has_fmr && !srp_dev->has_fr)) {
|
||||||
IB_ACCESS_LOCAL_WRITE |
|
srp_dev->global_mr = ib_get_dma_mr(srp_dev->pd,
|
||||||
IB_ACCESS_REMOTE_READ |
|
IB_ACCESS_LOCAL_WRITE |
|
||||||
IB_ACCESS_REMOTE_WRITE);
|
IB_ACCESS_REMOTE_READ |
|
||||||
if (IS_ERR(srp_dev->mr))
|
IB_ACCESS_REMOTE_WRITE);
|
||||||
goto err_pd;
|
if (IS_ERR(srp_dev->global_mr))
|
||||||
|
goto err_pd;
|
||||||
|
} else {
|
||||||
|
srp_dev->global_mr = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
for (p = rdma_start_port(device); p <= rdma_end_port(device); ++p) {
|
for (p = rdma_start_port(device); p <= rdma_end_port(device); ++p) {
|
||||||
host = srp_add_port(srp_dev, p);
|
host = srp_add_port(srp_dev, p);
|
||||||
|
@ -3509,7 +3514,8 @@ static void srp_remove_one(struct ib_device *device, void *client_data)
|
||||||
kfree(host);
|
kfree(host);
|
||||||
}
|
}
|
||||||
|
|
||||||
ib_dereg_mr(srp_dev->mr);
|
if (srp_dev->global_mr)
|
||||||
|
ib_dereg_mr(srp_dev->global_mr);
|
||||||
ib_dealloc_pd(srp_dev->pd);
|
ib_dealloc_pd(srp_dev->pd);
|
||||||
|
|
||||||
kfree(srp_dev);
|
kfree(srp_dev);
|
||||||
|
|
|
@ -95,7 +95,7 @@ struct srp_device {
|
||||||
struct list_head dev_list;
|
struct list_head dev_list;
|
||||||
struct ib_device *dev;
|
struct ib_device *dev;
|
||||||
struct ib_pd *pd;
|
struct ib_pd *pd;
|
||||||
struct ib_mr *mr;
|
struct ib_mr *global_mr;
|
||||||
u64 mr_page_mask;
|
u64 mr_page_mask;
|
||||||
int mr_page_size;
|
int mr_page_size;
|
||||||
int mr_max_size;
|
int mr_max_size;
|
||||||
|
@ -183,10 +183,10 @@ struct srp_target_port {
|
||||||
spinlock_t lock;
|
spinlock_t lock;
|
||||||
|
|
||||||
/* read only in the hot path */
|
/* read only in the hot path */
|
||||||
|
struct ib_mr *global_mr;
|
||||||
struct srp_rdma_ch *ch;
|
struct srp_rdma_ch *ch;
|
||||||
u32 ch_count;
|
u32 ch_count;
|
||||||
u32 lkey;
|
u32 lkey;
|
||||||
u32 rkey;
|
|
||||||
enum srp_target_state state;
|
enum srp_target_state state;
|
||||||
unsigned int max_iu_len;
|
unsigned int max_iu_len;
|
||||||
unsigned int cmd_sg_cnt;
|
unsigned int cmd_sg_cnt;
|
||||||
|
|
Loading…
Add table
Reference in a new issue