sfc: move channel alloc/removal code
Reallocation and copying code is included, as well as some housekeeping code. Other files have been patched up a bit to accommodate the changes. Small code styling fixes included. Signed-off-by: Alexandru-Mihai Maftei <amaftei@solarflare.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
e20ba5b1d1
commit
8397548507
5 changed files with 441 additions and 435 deletions
|
@ -96,13 +96,6 @@ static unsigned int rx_irq_mod_usec = 60;
|
||||||
*/
|
*/
|
||||||
static unsigned int tx_irq_mod_usec = 150;
|
static unsigned int tx_irq_mod_usec = 150;
|
||||||
|
|
||||||
/* This is the first interrupt mode to try out of:
|
|
||||||
* 0 => MSI-X
|
|
||||||
* 1 => MSI
|
|
||||||
* 2 => legacy
|
|
||||||
*/
|
|
||||||
static unsigned int interrupt_mode;
|
|
||||||
|
|
||||||
/* This is the requested number of CPUs to use for Receive-Side Scaling (RSS),
|
/* This is the requested number of CPUs to use for Receive-Side Scaling (RSS),
|
||||||
* i.e. the number of CPUs among which we may distribute simultaneous
|
* i.e. the number of CPUs among which we may distribute simultaneous
|
||||||
* interrupt handling.
|
* interrupt handling.
|
||||||
|
@ -244,349 +237,11 @@ void efx_remove_eventq(struct efx_channel *channel)
|
||||||
*
|
*
|
||||||
*************************************************************************/
|
*************************************************************************/
|
||||||
|
|
||||||
/* Allocate and initialise a channel structure. */
|
|
||||||
struct efx_channel *
|
|
||||||
efx_alloc_channel(struct efx_nic *efx, int i, struct efx_channel *old_channel)
|
|
||||||
{
|
|
||||||
struct efx_channel *channel;
|
|
||||||
struct efx_rx_queue *rx_queue;
|
|
||||||
struct efx_tx_queue *tx_queue;
|
|
||||||
int j;
|
|
||||||
|
|
||||||
channel = kzalloc(sizeof(*channel), GFP_KERNEL);
|
|
||||||
if (!channel)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
channel->efx = efx;
|
|
||||||
channel->channel = i;
|
|
||||||
channel->type = &efx_default_channel_type;
|
|
||||||
|
|
||||||
for (j = 0; j < EFX_TXQ_TYPES; j++) {
|
|
||||||
tx_queue = &channel->tx_queue[j];
|
|
||||||
tx_queue->efx = efx;
|
|
||||||
tx_queue->queue = i * EFX_TXQ_TYPES + j;
|
|
||||||
tx_queue->channel = channel;
|
|
||||||
}
|
|
||||||
|
|
||||||
#ifdef CONFIG_RFS_ACCEL
|
|
||||||
INIT_DELAYED_WORK(&channel->filter_work, efx_filter_rfs_expire);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
rx_queue = &channel->rx_queue;
|
|
||||||
rx_queue->efx = efx;
|
|
||||||
timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0);
|
|
||||||
|
|
||||||
return channel;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Allocate and initialise a channel structure, copying parameters
|
|
||||||
* (but not resources) from an old channel structure.
|
|
||||||
*/
|
|
||||||
struct efx_channel *efx_copy_channel(const struct efx_channel *old_channel)
|
|
||||||
{
|
|
||||||
struct efx_channel *channel;
|
|
||||||
struct efx_rx_queue *rx_queue;
|
|
||||||
struct efx_tx_queue *tx_queue;
|
|
||||||
int j;
|
|
||||||
|
|
||||||
channel = kmalloc(sizeof(*channel), GFP_KERNEL);
|
|
||||||
if (!channel)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
*channel = *old_channel;
|
|
||||||
|
|
||||||
channel->napi_dev = NULL;
|
|
||||||
INIT_HLIST_NODE(&channel->napi_str.napi_hash_node);
|
|
||||||
channel->napi_str.napi_id = 0;
|
|
||||||
channel->napi_str.state = 0;
|
|
||||||
memset(&channel->eventq, 0, sizeof(channel->eventq));
|
|
||||||
|
|
||||||
for (j = 0; j < EFX_TXQ_TYPES; j++) {
|
|
||||||
tx_queue = &channel->tx_queue[j];
|
|
||||||
if (tx_queue->channel)
|
|
||||||
tx_queue->channel = channel;
|
|
||||||
tx_queue->buffer = NULL;
|
|
||||||
memset(&tx_queue->txd, 0, sizeof(tx_queue->txd));
|
|
||||||
}
|
|
||||||
|
|
||||||
rx_queue = &channel->rx_queue;
|
|
||||||
rx_queue->buffer = NULL;
|
|
||||||
memset(&rx_queue->rxd, 0, sizeof(rx_queue->rxd));
|
|
||||||
timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0);
|
|
||||||
#ifdef CONFIG_RFS_ACCEL
|
|
||||||
INIT_DELAYED_WORK(&channel->filter_work, efx_filter_rfs_expire);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
return channel;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int efx_probe_channel(struct efx_channel *channel)
|
|
||||||
{
|
|
||||||
struct efx_tx_queue *tx_queue;
|
|
||||||
struct efx_rx_queue *rx_queue;
|
|
||||||
int rc;
|
|
||||||
|
|
||||||
netif_dbg(channel->efx, probe, channel->efx->net_dev,
|
|
||||||
"creating channel %d\n", channel->channel);
|
|
||||||
|
|
||||||
rc = channel->type->pre_probe(channel);
|
|
||||||
if (rc)
|
|
||||||
goto fail;
|
|
||||||
|
|
||||||
rc = efx_probe_eventq(channel);
|
|
||||||
if (rc)
|
|
||||||
goto fail;
|
|
||||||
|
|
||||||
efx_for_each_channel_tx_queue(tx_queue, channel) {
|
|
||||||
rc = efx_probe_tx_queue(tx_queue);
|
|
||||||
if (rc)
|
|
||||||
goto fail;
|
|
||||||
}
|
|
||||||
|
|
||||||
efx_for_each_channel_rx_queue(rx_queue, channel) {
|
|
||||||
rc = efx_probe_rx_queue(rx_queue);
|
|
||||||
if (rc)
|
|
||||||
goto fail;
|
|
||||||
}
|
|
||||||
|
|
||||||
channel->rx_list = NULL;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
fail:
|
|
||||||
efx_remove_channel(channel);
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_get_channel_name(struct efx_channel *channel, char *buf, size_t len)
|
|
||||||
{
|
|
||||||
struct efx_nic *efx = channel->efx;
|
|
||||||
const char *type;
|
|
||||||
int number;
|
|
||||||
|
|
||||||
number = channel->channel;
|
|
||||||
|
|
||||||
if (number >= efx->xdp_channel_offset &&
|
|
||||||
!WARN_ON_ONCE(!efx->n_xdp_channels)) {
|
|
||||||
type = "-xdp";
|
|
||||||
number -= efx->xdp_channel_offset;
|
|
||||||
} else if (efx->tx_channel_offset == 0) {
|
|
||||||
type = "";
|
|
||||||
} else if (number < efx->tx_channel_offset) {
|
|
||||||
type = "-rx";
|
|
||||||
} else {
|
|
||||||
type = "-tx";
|
|
||||||
number -= efx->tx_channel_offset;
|
|
||||||
}
|
|
||||||
snprintf(buf, len, "%s%s-%d", efx->name, type, number);
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_set_channel_names(struct efx_nic *efx)
|
|
||||||
{
|
|
||||||
struct efx_channel *channel;
|
|
||||||
|
|
||||||
efx_for_each_channel(channel, efx)
|
|
||||||
channel->type->get_name(channel,
|
|
||||||
efx->msi_context[channel->channel].name,
|
|
||||||
sizeof(efx->msi_context[0].name));
|
|
||||||
}
|
|
||||||
|
|
||||||
int efx_probe_channels(struct efx_nic *efx)
|
|
||||||
{
|
|
||||||
struct efx_channel *channel;
|
|
||||||
int rc;
|
|
||||||
|
|
||||||
/* Restart special buffer allocation */
|
|
||||||
efx->next_buffer_table = 0;
|
|
||||||
|
|
||||||
/* Probe channels in reverse, so that any 'extra' channels
|
|
||||||
* use the start of the buffer table. This allows the traffic
|
|
||||||
* channels to be resized without moving them or wasting the
|
|
||||||
* entries before them.
|
|
||||||
*/
|
|
||||||
efx_for_each_channel_rev(channel, efx) {
|
|
||||||
rc = efx_probe_channel(channel);
|
|
||||||
if (rc) {
|
|
||||||
netif_err(efx, probe, efx->net_dev,
|
|
||||||
"failed to create channel %d\n",
|
|
||||||
channel->channel);
|
|
||||||
goto fail;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
efx_set_channel_names(efx);
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
fail:
|
|
||||||
efx_remove_channels(efx);
|
|
||||||
return rc;
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_remove_channel(struct efx_channel *channel)
|
|
||||||
{
|
|
||||||
struct efx_tx_queue *tx_queue;
|
|
||||||
struct efx_rx_queue *rx_queue;
|
|
||||||
|
|
||||||
netif_dbg(channel->efx, drv, channel->efx->net_dev,
|
|
||||||
"destroy chan %d\n", channel->channel);
|
|
||||||
|
|
||||||
efx_for_each_channel_rx_queue(rx_queue, channel)
|
|
||||||
efx_remove_rx_queue(rx_queue);
|
|
||||||
efx_for_each_possible_channel_tx_queue(tx_queue, channel)
|
|
||||||
efx_remove_tx_queue(tx_queue);
|
|
||||||
efx_remove_eventq(channel);
|
|
||||||
channel->type->post_remove(channel);
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_remove_channels(struct efx_nic *efx)
|
|
||||||
{
|
|
||||||
struct efx_channel *channel;
|
|
||||||
|
|
||||||
efx_for_each_channel(channel, efx)
|
|
||||||
efx_remove_channel(channel);
|
|
||||||
|
|
||||||
kfree(efx->xdp_tx_queues);
|
|
||||||
}
|
|
||||||
|
|
||||||
int efx_realloc_channels(struct efx_nic *efx, u32 rxq_entries, u32 txq_entries)
|
|
||||||
{
|
|
||||||
struct efx_channel *other_channel[EFX_MAX_CHANNELS], *channel;
|
|
||||||
u32 old_rxq_entries, old_txq_entries;
|
|
||||||
unsigned i, next_buffer_table = 0;
|
|
||||||
int rc, rc2;
|
|
||||||
|
|
||||||
rc = efx_check_disabled(efx);
|
|
||||||
if (rc)
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
/* Not all channels should be reallocated. We must avoid
|
|
||||||
* reallocating their buffer table entries.
|
|
||||||
*/
|
|
||||||
efx_for_each_channel(channel, efx) {
|
|
||||||
struct efx_rx_queue *rx_queue;
|
|
||||||
struct efx_tx_queue *tx_queue;
|
|
||||||
|
|
||||||
if (channel->type->copy)
|
|
||||||
continue;
|
|
||||||
next_buffer_table = max(next_buffer_table,
|
|
||||||
channel->eventq.index +
|
|
||||||
channel->eventq.entries);
|
|
||||||
efx_for_each_channel_rx_queue(rx_queue, channel)
|
|
||||||
next_buffer_table = max(next_buffer_table,
|
|
||||||
rx_queue->rxd.index +
|
|
||||||
rx_queue->rxd.entries);
|
|
||||||
efx_for_each_channel_tx_queue(tx_queue, channel)
|
|
||||||
next_buffer_table = max(next_buffer_table,
|
|
||||||
tx_queue->txd.index +
|
|
||||||
tx_queue->txd.entries);
|
|
||||||
}
|
|
||||||
|
|
||||||
efx_device_detach_sync(efx);
|
|
||||||
efx_stop_all(efx);
|
|
||||||
efx_soft_disable_interrupts(efx);
|
|
||||||
|
|
||||||
/* Clone channels (where possible) */
|
|
||||||
memset(other_channel, 0, sizeof(other_channel));
|
|
||||||
for (i = 0; i < efx->n_channels; i++) {
|
|
||||||
channel = efx->channel[i];
|
|
||||||
if (channel->type->copy)
|
|
||||||
channel = channel->type->copy(channel);
|
|
||||||
if (!channel) {
|
|
||||||
rc = -ENOMEM;
|
|
||||||
goto out;
|
|
||||||
}
|
|
||||||
other_channel[i] = channel;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Swap entry counts and channel pointers */
|
|
||||||
old_rxq_entries = efx->rxq_entries;
|
|
||||||
old_txq_entries = efx->txq_entries;
|
|
||||||
efx->rxq_entries = rxq_entries;
|
|
||||||
efx->txq_entries = txq_entries;
|
|
||||||
for (i = 0; i < efx->n_channels; i++) {
|
|
||||||
channel = efx->channel[i];
|
|
||||||
efx->channel[i] = other_channel[i];
|
|
||||||
other_channel[i] = channel;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Restart buffer table allocation */
|
|
||||||
efx->next_buffer_table = next_buffer_table;
|
|
||||||
|
|
||||||
for (i = 0; i < efx->n_channels; i++) {
|
|
||||||
channel = efx->channel[i];
|
|
||||||
if (!channel->type->copy)
|
|
||||||
continue;
|
|
||||||
rc = efx_probe_channel(channel);
|
|
||||||
if (rc)
|
|
||||||
goto rollback;
|
|
||||||
efx_init_napi_channel(efx->channel[i]);
|
|
||||||
}
|
|
||||||
|
|
||||||
out:
|
|
||||||
/* Destroy unused channel structures */
|
|
||||||
for (i = 0; i < efx->n_channels; i++) {
|
|
||||||
channel = other_channel[i];
|
|
||||||
if (channel && channel->type->copy) {
|
|
||||||
efx_fini_napi_channel(channel);
|
|
||||||
efx_remove_channel(channel);
|
|
||||||
kfree(channel);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
rc2 = efx_soft_enable_interrupts(efx);
|
|
||||||
if (rc2) {
|
|
||||||
rc = rc ? rc : rc2;
|
|
||||||
netif_err(efx, drv, efx->net_dev,
|
|
||||||
"unable to restart interrupts on channel reallocation\n");
|
|
||||||
efx_schedule_reset(efx, RESET_TYPE_DISABLE);
|
|
||||||
} else {
|
|
||||||
efx_start_all(efx);
|
|
||||||
efx_device_attach_if_not_resetting(efx);
|
|
||||||
}
|
|
||||||
return rc;
|
|
||||||
|
|
||||||
rollback:
|
|
||||||
/* Swap back */
|
|
||||||
efx->rxq_entries = old_rxq_entries;
|
|
||||||
efx->txq_entries = old_txq_entries;
|
|
||||||
for (i = 0; i < efx->n_channels; i++) {
|
|
||||||
channel = efx->channel[i];
|
|
||||||
efx->channel[i] = other_channel[i];
|
|
||||||
other_channel[i] = channel;
|
|
||||||
}
|
|
||||||
goto out;
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue)
|
void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue)
|
||||||
{
|
{
|
||||||
mod_timer(&rx_queue->slow_fill, jiffies + msecs_to_jiffies(10));
|
mod_timer(&rx_queue->slow_fill, jiffies + msecs_to_jiffies(10));
|
||||||
}
|
}
|
||||||
|
|
||||||
bool efx_default_channel_want_txqs(struct efx_channel *channel)
|
|
||||||
{
|
|
||||||
return channel->channel - channel->efx->tx_channel_offset <
|
|
||||||
channel->efx->n_tx_channels;
|
|
||||||
}
|
|
||||||
|
|
||||||
static const struct efx_channel_type efx_default_channel_type = {
|
|
||||||
.pre_probe = efx_channel_dummy_op_int,
|
|
||||||
.post_remove = efx_channel_dummy_op_void,
|
|
||||||
.get_name = efx_get_channel_name,
|
|
||||||
.copy = efx_copy_channel,
|
|
||||||
.want_txqs = efx_default_channel_want_txqs,
|
|
||||||
.keep_eventq = false,
|
|
||||||
.want_pio = true,
|
|
||||||
};
|
|
||||||
|
|
||||||
int efx_channel_dummy_op_int(struct efx_channel *channel)
|
|
||||||
{
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_channel_dummy_op_void(struct efx_channel *channel)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
/**************************************************************************
|
/**************************************************************************
|
||||||
*
|
*
|
||||||
* Port handling
|
* Port handling
|
||||||
|
@ -1100,52 +755,6 @@ void efx_remove_interrupts(struct efx_nic *efx)
|
||||||
efx->legacy_irq = 0;
|
efx->legacy_irq = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
int efx_set_channels(struct efx_nic *efx)
|
|
||||||
{
|
|
||||||
struct efx_channel *channel;
|
|
||||||
struct efx_tx_queue *tx_queue;
|
|
||||||
int xdp_queue_number;
|
|
||||||
|
|
||||||
efx->tx_channel_offset =
|
|
||||||
efx_separate_tx_channels ?
|
|
||||||
efx->n_channels - efx->n_tx_channels : 0;
|
|
||||||
|
|
||||||
if (efx->xdp_tx_queue_count) {
|
|
||||||
EFX_WARN_ON_PARANOID(efx->xdp_tx_queues);
|
|
||||||
|
|
||||||
/* Allocate array for XDP TX queue lookup. */
|
|
||||||
efx->xdp_tx_queues = kcalloc(efx->xdp_tx_queue_count,
|
|
||||||
sizeof(*efx->xdp_tx_queues),
|
|
||||||
GFP_KERNEL);
|
|
||||||
if (!efx->xdp_tx_queues)
|
|
||||||
return -ENOMEM;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* We need to mark which channels really have RX and TX
|
|
||||||
* queues, and adjust the TX queue numbers if we have separate
|
|
||||||
* RX-only and TX-only channels.
|
|
||||||
*/
|
|
||||||
xdp_queue_number = 0;
|
|
||||||
efx_for_each_channel(channel, efx) {
|
|
||||||
if (channel->channel < efx->n_rx_channels)
|
|
||||||
channel->rx_queue.core_index = channel->channel;
|
|
||||||
else
|
|
||||||
channel->rx_queue.core_index = -1;
|
|
||||||
|
|
||||||
efx_for_each_channel_tx_queue(tx_queue, channel) {
|
|
||||||
tx_queue->queue -= (efx->tx_channel_offset *
|
|
||||||
EFX_TXQ_TYPES);
|
|
||||||
|
|
||||||
if (efx_channel_is_xdp_tx(channel) &&
|
|
||||||
xdp_queue_number < efx->xdp_tx_queue_count) {
|
|
||||||
efx->xdp_tx_queues[xdp_queue_number] = tx_queue;
|
|
||||||
xdp_queue_number++;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int efx_probe_nic(struct efx_nic *efx)
|
static int efx_probe_nic(struct efx_nic *efx)
|
||||||
{
|
{
|
||||||
int rc;
|
int rc;
|
||||||
|
@ -2022,31 +1631,6 @@ static const struct pci_device_id efx_pci_table[] = {
|
||||||
*
|
*
|
||||||
**************************************************************************/
|
**************************************************************************/
|
||||||
|
|
||||||
int efx_init_channels(struct efx_nic *efx)
|
|
||||||
{
|
|
||||||
unsigned int i;
|
|
||||||
|
|
||||||
for (i = 0; i < EFX_MAX_CHANNELS; i++) {
|
|
||||||
efx->channel[i] = efx_alloc_channel(efx, i, NULL);
|
|
||||||
if (!efx->channel[i])
|
|
||||||
return -ENOMEM;
|
|
||||||
efx->msi_context[i].efx = efx;
|
|
||||||
efx->msi_context[i].index = i;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Higher numbered interrupt modes are less capable! */
|
|
||||||
if (WARN_ON_ONCE(efx->type->max_interrupt_mode >
|
|
||||||
efx->type->min_interrupt_mode)) {
|
|
||||||
return -EIO;
|
|
||||||
}
|
|
||||||
efx->interrupt_mode = max(efx->type->max_interrupt_mode,
|
|
||||||
interrupt_mode);
|
|
||||||
efx->interrupt_mode = min(efx->type->min_interrupt_mode,
|
|
||||||
interrupt_mode);
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_update_sw_stats(struct efx_nic *efx, u64 *stats)
|
void efx_update_sw_stats(struct efx_nic *efx, u64 *stats)
|
||||||
{
|
{
|
||||||
u64 n_rx_nodesc_trunc = 0;
|
u64 n_rx_nodesc_trunc = 0;
|
||||||
|
@ -2829,10 +2413,6 @@ static struct pci_driver efx_pci_driver = {
|
||||||
*
|
*
|
||||||
*************************************************************************/
|
*************************************************************************/
|
||||||
|
|
||||||
module_param(interrupt_mode, uint, 0444);
|
|
||||||
MODULE_PARM_DESC(interrupt_mode,
|
|
||||||
"Interrupt mode (0=>MSIX 1=>MSI 2=>legacy)");
|
|
||||||
|
|
||||||
static int __init efx_init_module(void)
|
static int __init efx_init_module(void)
|
||||||
{
|
{
|
||||||
int rc;
|
int rc;
|
||||||
|
|
|
@ -203,10 +203,6 @@ static inline bool efx_rss_active(struct efx_rss_context *ctx)
|
||||||
return ctx->context_id != EFX_EF10_RSS_CONTEXT_INVALID;
|
return ctx->context_id != EFX_EF10_RSS_CONTEXT_INVALID;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Channels */
|
|
||||||
int efx_channel_dummy_op_int(struct efx_channel *channel);
|
|
||||||
void efx_channel_dummy_op_void(struct efx_channel *channel);
|
|
||||||
|
|
||||||
/* Ethtool support */
|
/* Ethtool support */
|
||||||
extern const struct ethtool_ops efx_ethtool_ops;
|
extern const struct ethtool_ops efx_ethtool_ops;
|
||||||
|
|
||||||
|
|
|
@ -18,6 +18,16 @@
|
||||||
#include "nic.h"
|
#include "nic.h"
|
||||||
#include "sriov.h"
|
#include "sriov.h"
|
||||||
|
|
||||||
|
/* This is the first interrupt mode to try out of:
|
||||||
|
* 0 => MSI-X
|
||||||
|
* 1 => MSI
|
||||||
|
* 2 => legacy
|
||||||
|
*/
|
||||||
|
static unsigned int interrupt_mode;
|
||||||
|
module_param(interrupt_mode, uint, 0444);
|
||||||
|
MODULE_PARM_DESC(interrupt_mode,
|
||||||
|
"Interrupt mode (0=>MSIX 1=>MSI 2=>legacy)");
|
||||||
|
|
||||||
static unsigned int irq_adapt_low_thresh = 8000;
|
static unsigned int irq_adapt_low_thresh = 8000;
|
||||||
module_param(irq_adapt_low_thresh, uint, 0644);
|
module_param(irq_adapt_low_thresh, uint, 0644);
|
||||||
MODULE_PARM_DESC(irq_adapt_low_thresh,
|
MODULE_PARM_DESC(irq_adapt_low_thresh,
|
||||||
|
@ -33,6 +43,436 @@ MODULE_PARM_DESC(irq_adapt_high_thresh,
|
||||||
*/
|
*/
|
||||||
static int napi_weight = 64;
|
static int napi_weight = 64;
|
||||||
|
|
||||||
|
/***************
|
||||||
|
* Housekeeping
|
||||||
|
***************/
|
||||||
|
|
||||||
|
int efx_channel_dummy_op_int(struct efx_channel *channel)
|
||||||
|
{
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void efx_channel_dummy_op_void(struct efx_channel *channel)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static const struct efx_channel_type efx_default_channel_type = {
|
||||||
|
.pre_probe = efx_channel_dummy_op_int,
|
||||||
|
.post_remove = efx_channel_dummy_op_void,
|
||||||
|
.get_name = efx_get_channel_name,
|
||||||
|
.copy = efx_copy_channel,
|
||||||
|
.want_txqs = efx_default_channel_want_txqs,
|
||||||
|
.keep_eventq = false,
|
||||||
|
.want_pio = true,
|
||||||
|
};
|
||||||
|
|
||||||
|
/**************************************************************************
|
||||||
|
*
|
||||||
|
* Channel handling
|
||||||
|
*
|
||||||
|
*************************************************************************/
|
||||||
|
|
||||||
|
/* Allocate and initialise a channel structure. */
|
||||||
|
struct efx_channel *
|
||||||
|
efx_alloc_channel(struct efx_nic *efx, int i, struct efx_channel *old_channel)
|
||||||
|
{
|
||||||
|
struct efx_rx_queue *rx_queue;
|
||||||
|
struct efx_tx_queue *tx_queue;
|
||||||
|
struct efx_channel *channel;
|
||||||
|
int j;
|
||||||
|
|
||||||
|
channel = kzalloc(sizeof(*channel), GFP_KERNEL);
|
||||||
|
if (!channel)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
channel->efx = efx;
|
||||||
|
channel->channel = i;
|
||||||
|
channel->type = &efx_default_channel_type;
|
||||||
|
|
||||||
|
for (j = 0; j < EFX_TXQ_TYPES; j++) {
|
||||||
|
tx_queue = &channel->tx_queue[j];
|
||||||
|
tx_queue->efx = efx;
|
||||||
|
tx_queue->queue = i * EFX_TXQ_TYPES + j;
|
||||||
|
tx_queue->channel = channel;
|
||||||
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_RFS_ACCEL
|
||||||
|
INIT_DELAYED_WORK(&channel->filter_work, efx_filter_rfs_expire);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
rx_queue = &channel->rx_queue;
|
||||||
|
rx_queue->efx = efx;
|
||||||
|
timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0);
|
||||||
|
|
||||||
|
return channel;
|
||||||
|
}
|
||||||
|
|
||||||
|
int efx_init_channels(struct efx_nic *efx)
|
||||||
|
{
|
||||||
|
unsigned int i;
|
||||||
|
|
||||||
|
for (i = 0; i < EFX_MAX_CHANNELS; i++) {
|
||||||
|
efx->channel[i] = efx_alloc_channel(efx, i, NULL);
|
||||||
|
if (!efx->channel[i])
|
||||||
|
return -ENOMEM;
|
||||||
|
efx->msi_context[i].efx = efx;
|
||||||
|
efx->msi_context[i].index = i;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Higher numbered interrupt modes are less capable! */
|
||||||
|
if (WARN_ON_ONCE(efx->type->max_interrupt_mode >
|
||||||
|
efx->type->min_interrupt_mode)) {
|
||||||
|
return -EIO;
|
||||||
|
}
|
||||||
|
efx->interrupt_mode = max(efx->type->max_interrupt_mode,
|
||||||
|
interrupt_mode);
|
||||||
|
efx->interrupt_mode = min(efx->type->min_interrupt_mode,
|
||||||
|
interrupt_mode);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void efx_fini_channels(struct efx_nic *efx)
|
||||||
|
{
|
||||||
|
unsigned int i;
|
||||||
|
|
||||||
|
for (i = 0; i < EFX_MAX_CHANNELS; i++)
|
||||||
|
if (efx->channel[i]) {
|
||||||
|
kfree(efx->channel[i]);
|
||||||
|
efx->channel[i] = NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Allocate and initialise a channel structure, copying parameters
|
||||||
|
* (but not resources) from an old channel structure.
|
||||||
|
*/
|
||||||
|
struct efx_channel *efx_copy_channel(const struct efx_channel *old_channel)
|
||||||
|
{
|
||||||
|
struct efx_rx_queue *rx_queue;
|
||||||
|
struct efx_tx_queue *tx_queue;
|
||||||
|
struct efx_channel *channel;
|
||||||
|
int j;
|
||||||
|
|
||||||
|
channel = kmalloc(sizeof(*channel), GFP_KERNEL);
|
||||||
|
if (!channel)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
*channel = *old_channel;
|
||||||
|
|
||||||
|
channel->napi_dev = NULL;
|
||||||
|
INIT_HLIST_NODE(&channel->napi_str.napi_hash_node);
|
||||||
|
channel->napi_str.napi_id = 0;
|
||||||
|
channel->napi_str.state = 0;
|
||||||
|
memset(&channel->eventq, 0, sizeof(channel->eventq));
|
||||||
|
|
||||||
|
for (j = 0; j < EFX_TXQ_TYPES; j++) {
|
||||||
|
tx_queue = &channel->tx_queue[j];
|
||||||
|
if (tx_queue->channel)
|
||||||
|
tx_queue->channel = channel;
|
||||||
|
tx_queue->buffer = NULL;
|
||||||
|
memset(&tx_queue->txd, 0, sizeof(tx_queue->txd));
|
||||||
|
}
|
||||||
|
|
||||||
|
rx_queue = &channel->rx_queue;
|
||||||
|
rx_queue->buffer = NULL;
|
||||||
|
memset(&rx_queue->rxd, 0, sizeof(rx_queue->rxd));
|
||||||
|
timer_setup(&rx_queue->slow_fill, efx_rx_slow_fill, 0);
|
||||||
|
#ifdef CONFIG_RFS_ACCEL
|
||||||
|
INIT_DELAYED_WORK(&channel->filter_work, efx_filter_rfs_expire);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
return channel;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int efx_probe_channel(struct efx_channel *channel)
|
||||||
|
{
|
||||||
|
struct efx_tx_queue *tx_queue;
|
||||||
|
struct efx_rx_queue *rx_queue;
|
||||||
|
int rc;
|
||||||
|
|
||||||
|
netif_dbg(channel->efx, probe, channel->efx->net_dev,
|
||||||
|
"creating channel %d\n", channel->channel);
|
||||||
|
|
||||||
|
rc = channel->type->pre_probe(channel);
|
||||||
|
if (rc)
|
||||||
|
goto fail;
|
||||||
|
|
||||||
|
rc = efx_probe_eventq(channel);
|
||||||
|
if (rc)
|
||||||
|
goto fail;
|
||||||
|
|
||||||
|
efx_for_each_channel_tx_queue(tx_queue, channel) {
|
||||||
|
rc = efx_probe_tx_queue(tx_queue);
|
||||||
|
if (rc)
|
||||||
|
goto fail;
|
||||||
|
}
|
||||||
|
|
||||||
|
efx_for_each_channel_rx_queue(rx_queue, channel) {
|
||||||
|
rc = efx_probe_rx_queue(rx_queue);
|
||||||
|
if (rc)
|
||||||
|
goto fail;
|
||||||
|
}
|
||||||
|
|
||||||
|
channel->rx_list = NULL;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
fail:
|
||||||
|
efx_remove_channel(channel);
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
void efx_get_channel_name(struct efx_channel *channel, char *buf, size_t len)
|
||||||
|
{
|
||||||
|
struct efx_nic *efx = channel->efx;
|
||||||
|
const char *type;
|
||||||
|
int number;
|
||||||
|
|
||||||
|
number = channel->channel;
|
||||||
|
|
||||||
|
if (number >= efx->xdp_channel_offset &&
|
||||||
|
!WARN_ON_ONCE(!efx->n_xdp_channels)) {
|
||||||
|
type = "-xdp";
|
||||||
|
number -= efx->xdp_channel_offset;
|
||||||
|
} else if (efx->tx_channel_offset == 0) {
|
||||||
|
type = "";
|
||||||
|
} else if (number < efx->tx_channel_offset) {
|
||||||
|
type = "-rx";
|
||||||
|
} else {
|
||||||
|
type = "-tx";
|
||||||
|
number -= efx->tx_channel_offset;
|
||||||
|
}
|
||||||
|
snprintf(buf, len, "%s%s-%d", efx->name, type, number);
|
||||||
|
}
|
||||||
|
|
||||||
|
void efx_set_channel_names(struct efx_nic *efx)
|
||||||
|
{
|
||||||
|
struct efx_channel *channel;
|
||||||
|
|
||||||
|
efx_for_each_channel(channel, efx)
|
||||||
|
channel->type->get_name(channel,
|
||||||
|
efx->msi_context[channel->channel].name,
|
||||||
|
sizeof(efx->msi_context[0].name));
|
||||||
|
}
|
||||||
|
|
||||||
|
int efx_probe_channels(struct efx_nic *efx)
|
||||||
|
{
|
||||||
|
struct efx_channel *channel;
|
||||||
|
int rc;
|
||||||
|
|
||||||
|
/* Restart special buffer allocation */
|
||||||
|
efx->next_buffer_table = 0;
|
||||||
|
|
||||||
|
/* Probe channels in reverse, so that any 'extra' channels
|
||||||
|
* use the start of the buffer table. This allows the traffic
|
||||||
|
* channels to be resized without moving them or wasting the
|
||||||
|
* entries before them.
|
||||||
|
*/
|
||||||
|
efx_for_each_channel_rev(channel, efx) {
|
||||||
|
rc = efx_probe_channel(channel);
|
||||||
|
if (rc) {
|
||||||
|
netif_err(efx, probe, efx->net_dev,
|
||||||
|
"failed to create channel %d\n",
|
||||||
|
channel->channel);
|
||||||
|
goto fail;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
efx_set_channel_names(efx);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
fail:
|
||||||
|
efx_remove_channels(efx);
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
void efx_remove_channel(struct efx_channel *channel)
|
||||||
|
{
|
||||||
|
struct efx_tx_queue *tx_queue;
|
||||||
|
struct efx_rx_queue *rx_queue;
|
||||||
|
|
||||||
|
netif_dbg(channel->efx, drv, channel->efx->net_dev,
|
||||||
|
"destroy chan %d\n", channel->channel);
|
||||||
|
|
||||||
|
efx_for_each_channel_rx_queue(rx_queue, channel)
|
||||||
|
efx_remove_rx_queue(rx_queue);
|
||||||
|
efx_for_each_possible_channel_tx_queue(tx_queue, channel)
|
||||||
|
efx_remove_tx_queue(tx_queue);
|
||||||
|
efx_remove_eventq(channel);
|
||||||
|
channel->type->post_remove(channel);
|
||||||
|
}
|
||||||
|
|
||||||
|
void efx_remove_channels(struct efx_nic *efx)
|
||||||
|
{
|
||||||
|
struct efx_channel *channel;
|
||||||
|
|
||||||
|
efx_for_each_channel(channel, efx)
|
||||||
|
efx_remove_channel(channel);
|
||||||
|
|
||||||
|
kfree(efx->xdp_tx_queues);
|
||||||
|
}
|
||||||
|
|
||||||
|
int efx_realloc_channels(struct efx_nic *efx, u32 rxq_entries, u32 txq_entries)
|
||||||
|
{
|
||||||
|
struct efx_channel *other_channel[EFX_MAX_CHANNELS], *channel;
|
||||||
|
unsigned int i, next_buffer_table = 0;
|
||||||
|
u32 old_rxq_entries, old_txq_entries;
|
||||||
|
int rc, rc2;
|
||||||
|
|
||||||
|
rc = efx_check_disabled(efx);
|
||||||
|
if (rc)
|
||||||
|
return rc;
|
||||||
|
|
||||||
|
/* Not all channels should be reallocated. We must avoid
|
||||||
|
* reallocating their buffer table entries.
|
||||||
|
*/
|
||||||
|
efx_for_each_channel(channel, efx) {
|
||||||
|
struct efx_rx_queue *rx_queue;
|
||||||
|
struct efx_tx_queue *tx_queue;
|
||||||
|
|
||||||
|
if (channel->type->copy)
|
||||||
|
continue;
|
||||||
|
next_buffer_table = max(next_buffer_table,
|
||||||
|
channel->eventq.index +
|
||||||
|
channel->eventq.entries);
|
||||||
|
efx_for_each_channel_rx_queue(rx_queue, channel)
|
||||||
|
next_buffer_table = max(next_buffer_table,
|
||||||
|
rx_queue->rxd.index +
|
||||||
|
rx_queue->rxd.entries);
|
||||||
|
efx_for_each_channel_tx_queue(tx_queue, channel)
|
||||||
|
next_buffer_table = max(next_buffer_table,
|
||||||
|
tx_queue->txd.index +
|
||||||
|
tx_queue->txd.entries);
|
||||||
|
}
|
||||||
|
|
||||||
|
efx_device_detach_sync(efx);
|
||||||
|
efx_stop_all(efx);
|
||||||
|
efx_soft_disable_interrupts(efx);
|
||||||
|
|
||||||
|
/* Clone channels (where possible) */
|
||||||
|
memset(other_channel, 0, sizeof(other_channel));
|
||||||
|
for (i = 0; i < efx->n_channels; i++) {
|
||||||
|
channel = efx->channel[i];
|
||||||
|
if (channel->type->copy)
|
||||||
|
channel = channel->type->copy(channel);
|
||||||
|
if (!channel) {
|
||||||
|
rc = -ENOMEM;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
other_channel[i] = channel;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Swap entry counts and channel pointers */
|
||||||
|
old_rxq_entries = efx->rxq_entries;
|
||||||
|
old_txq_entries = efx->txq_entries;
|
||||||
|
efx->rxq_entries = rxq_entries;
|
||||||
|
efx->txq_entries = txq_entries;
|
||||||
|
for (i = 0; i < efx->n_channels; i++) {
|
||||||
|
channel = efx->channel[i];
|
||||||
|
efx->channel[i] = other_channel[i];
|
||||||
|
other_channel[i] = channel;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Restart buffer table allocation */
|
||||||
|
efx->next_buffer_table = next_buffer_table;
|
||||||
|
|
||||||
|
for (i = 0; i < efx->n_channels; i++) {
|
||||||
|
channel = efx->channel[i];
|
||||||
|
if (!channel->type->copy)
|
||||||
|
continue;
|
||||||
|
rc = efx_probe_channel(channel);
|
||||||
|
if (rc)
|
||||||
|
goto rollback;
|
||||||
|
efx_init_napi_channel(efx->channel[i]);
|
||||||
|
}
|
||||||
|
|
||||||
|
out:
|
||||||
|
/* Destroy unused channel structures */
|
||||||
|
for (i = 0; i < efx->n_channels; i++) {
|
||||||
|
channel = other_channel[i];
|
||||||
|
if (channel && channel->type->copy) {
|
||||||
|
efx_fini_napi_channel(channel);
|
||||||
|
efx_remove_channel(channel);
|
||||||
|
kfree(channel);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
rc2 = efx_soft_enable_interrupts(efx);
|
||||||
|
if (rc2) {
|
||||||
|
rc = rc ? rc : rc2;
|
||||||
|
netif_err(efx, drv, efx->net_dev,
|
||||||
|
"unable to restart interrupts on channel reallocation\n");
|
||||||
|
efx_schedule_reset(efx, RESET_TYPE_DISABLE);
|
||||||
|
} else {
|
||||||
|
efx_start_all(efx);
|
||||||
|
efx_device_attach_if_not_resetting(efx);
|
||||||
|
}
|
||||||
|
return rc;
|
||||||
|
|
||||||
|
rollback:
|
||||||
|
/* Swap back */
|
||||||
|
efx->rxq_entries = old_rxq_entries;
|
||||||
|
efx->txq_entries = old_txq_entries;
|
||||||
|
for (i = 0; i < efx->n_channels; i++) {
|
||||||
|
channel = efx->channel[i];
|
||||||
|
efx->channel[i] = other_channel[i];
|
||||||
|
other_channel[i] = channel;
|
||||||
|
}
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
int efx_set_channels(struct efx_nic *efx)
|
||||||
|
{
|
||||||
|
struct efx_channel *channel;
|
||||||
|
struct efx_tx_queue *tx_queue;
|
||||||
|
int xdp_queue_number;
|
||||||
|
|
||||||
|
efx->tx_channel_offset =
|
||||||
|
efx_separate_tx_channels ?
|
||||||
|
efx->n_channels - efx->n_tx_channels : 0;
|
||||||
|
|
||||||
|
if (efx->xdp_tx_queue_count) {
|
||||||
|
EFX_WARN_ON_PARANOID(efx->xdp_tx_queues);
|
||||||
|
|
||||||
|
/* Allocate array for XDP TX queue lookup. */
|
||||||
|
efx->xdp_tx_queues = kcalloc(efx->xdp_tx_queue_count,
|
||||||
|
sizeof(*efx->xdp_tx_queues),
|
||||||
|
GFP_KERNEL);
|
||||||
|
if (!efx->xdp_tx_queues)
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* We need to mark which channels really have RX and TX
|
||||||
|
* queues, and adjust the TX queue numbers if we have separate
|
||||||
|
* RX-only and TX-only channels.
|
||||||
|
*/
|
||||||
|
xdp_queue_number = 0;
|
||||||
|
efx_for_each_channel(channel, efx) {
|
||||||
|
if (channel->channel < efx->n_rx_channels)
|
||||||
|
channel->rx_queue.core_index = channel->channel;
|
||||||
|
else
|
||||||
|
channel->rx_queue.core_index = -1;
|
||||||
|
|
||||||
|
efx_for_each_channel_tx_queue(tx_queue, channel) {
|
||||||
|
tx_queue->queue -= (efx->tx_channel_offset *
|
||||||
|
EFX_TXQ_TYPES);
|
||||||
|
|
||||||
|
if (efx_channel_is_xdp_tx(channel) &&
|
||||||
|
xdp_queue_number < efx->xdp_tx_queue_count) {
|
||||||
|
efx->xdp_tx_queues[xdp_queue_number] = tx_queue;
|
||||||
|
xdp_queue_number++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool efx_default_channel_want_txqs(struct efx_channel *channel)
|
||||||
|
{
|
||||||
|
return channel->channel - channel->efx->tx_channel_offset <
|
||||||
|
channel->efx->n_tx_channels;
|
||||||
|
}
|
||||||
|
|
||||||
/*************
|
/*************
|
||||||
* START/STOP
|
* START/STOP
|
||||||
*************/
|
*************/
|
||||||
|
|
|
@ -884,17 +884,6 @@ fail:
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
void efx_fini_channels(struct efx_nic *efx)
|
|
||||||
{
|
|
||||||
unsigned int i;
|
|
||||||
|
|
||||||
for (i = 0; i < EFX_MAX_CHANNELS; i++)
|
|
||||||
if (efx->channel[i]) {
|
|
||||||
kfree(efx->channel[i]);
|
|
||||||
efx->channel[i] = NULL;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
void efx_fini_struct(struct efx_nic *efx)
|
void efx_fini_struct(struct efx_nic *efx)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_RFS_ACCEL
|
#ifdef CONFIG_RFS_ACCEL
|
||||||
|
|
|
@ -7,6 +7,7 @@
|
||||||
#include <linux/module.h>
|
#include <linux/module.h>
|
||||||
#include "net_driver.h"
|
#include "net_driver.h"
|
||||||
#include "efx.h"
|
#include "efx.h"
|
||||||
|
#include "efx_channels.h"
|
||||||
#include "nic.h"
|
#include "nic.h"
|
||||||
#include "io.h"
|
#include "io.h"
|
||||||
#include "mcdi.h"
|
#include "mcdi.h"
|
||||||
|
|
Loading…
Add table
Reference in a new issue