net/mlx5: Forward SRQ resource events

Allow forwarding of SRQ events to mlx5_core interfaces, e.g. mlx5_ib.
Use mlx5_notifier_register/unregister in srq.c in order to allow seamless
transition of srq.c to infiniband subsystem.

Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
This commit is contained in:
Saeed Mahameed 2018-11-26 14:39:07 -08:00
parent 451be51c0b
commit 4e2df04ad2
3 changed files with 14 additions and 30 deletions

View file

@ -44,6 +44,9 @@ static struct mlx5_nb events_nbs_ref[] = {
{.nb.notifier_call = forward_event, .event_type = MLX5_EVENT_TYPE_PATH_MIG_FAILED },
{.nb.notifier_call = forward_event, .event_type = MLX5_EVENT_TYPE_WQ_INVAL_REQ_ERROR },
{.nb.notifier_call = forward_event, .event_type = MLX5_EVENT_TYPE_WQ_ACCESS_ERROR },
/* SRQ events */
{.nb.notifier_call = forward_event, .event_type = MLX5_EVENT_TYPE_SRQ_CATAS_ERROR },
{.nb.notifier_call = forward_event, .event_type = MLX5_EVENT_TYPE_SRQ_RQ_LIMIT },
};
struct mlx5_events {

View file

@ -40,15 +40,21 @@
#include "mlx5_core.h"
#include "lib/eq.h"
static int srq_event_notifier(struct mlx5_srq_table *table,
static int srq_event_notifier(struct notifier_block *nb,
unsigned long type, void *data)
{
struct mlx5_srq_table *table;
struct mlx5_core_dev *dev;
struct mlx5_core_srq *srq;
struct mlx5_priv *priv;
struct mlx5_eqe *eqe;
u32 srqn;
if (type != MLX5_EVENT_TYPE_SRQ_CATAS_ERROR &&
type != MLX5_EVENT_TYPE_SRQ_RQ_LIMIT)
return NOTIFY_DONE;
table = container_of(nb, struct mlx5_srq_table, nb);
priv = container_of(table, struct mlx5_priv, srq_table);
dev = container_of(priv, struct mlx5_core_dev, priv);
@ -77,26 +83,6 @@ static int srq_event_notifier(struct mlx5_srq_table *table,
return NOTIFY_OK;
}
static int catas_err_notifier(struct notifier_block *nb,
unsigned long type, void *data)
{
struct mlx5_srq_table *table;
table = mlx5_nb_cof(nb, struct mlx5_srq_table, catas_err_nb);
/* type == MLX5_EVENT_TYPE_SRQ_CATAS_ERROR */
return srq_event_notifier(table, type, data);
}
static int rq_limit_notifier(struct notifier_block *nb,
unsigned long type, void *data)
{
struct mlx5_srq_table *table;
table = mlx5_nb_cof(nb, struct mlx5_srq_table, rq_limit_nb);
/* type == MLX5_EVENT_TYPE_SRQ_RQ_LIMIT */
return srq_event_notifier(table, type, data);
}
static int get_pas_size(struct mlx5_srq_attr *in)
{
u32 log_page_size = in->log_page_size + 12;
@ -743,17 +729,13 @@ void mlx5_init_srq_table(struct mlx5_core_dev *dev)
spin_lock_init(&table->lock);
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
MLX5_NB_INIT(&table->catas_err_nb, catas_err_notifier, SRQ_CATAS_ERROR);
mlx5_eq_notifier_register(dev, &table->catas_err_nb);
MLX5_NB_INIT(&table->rq_limit_nb, rq_limit_notifier, SRQ_RQ_LIMIT);
mlx5_eq_notifier_register(dev, &table->rq_limit_nb);
table->nb.notifier_call = srq_event_notifier;
mlx5_notifier_register(dev, &table->nb);
}
void mlx5_cleanup_srq_table(struct mlx5_core_dev *dev)
{
struct mlx5_srq_table *table = &dev->priv.srq_table;
mlx5_eq_notifier_unregister(dev, &table->rq_limit_nb);
mlx5_eq_notifier_unregister(dev, &table->catas_err_nb);
mlx5_notifier_unregister(dev, &table->nb);
}

View file

@ -465,8 +465,7 @@ struct mlx5_qp_table {
};
struct mlx5_srq_table {
struct mlx5_nb catas_err_nb;
struct mlx5_nb rq_limit_nb;
struct notifier_block nb;
/* protect radix tree
*/
spinlock_t lock;