1
0
mirror of https://github.com/corundum/corundum.git synced 2025-01-16 08:12:53 +08:00
corundum/modules/mqnic/mqnic_eq.c
Alex Forencich bed12ee774 Consolidate CQs
Signed-off-by: Alex Forencich <alex@alexforencich.com>
2023-07-10 17:52:34 -07:00

251 lines
5.9 KiB
C

// SPDX-License-Identifier: BSD-2-Clause-Views
/*
* Copyright (c) 2019-2023 The Regents of the University of California
*/
#include "mqnic.h"
static int mqnic_eq_int(struct notifier_block *nb, unsigned long action, void *data)
{
struct mqnic_eq *eq = container_of(nb, struct mqnic_eq, irq_nb);
mqnic_process_eq(eq);
mqnic_arm_eq(eq);
return NOTIFY_DONE;
}
struct mqnic_eq *mqnic_create_eq(struct mqnic_if *interface)
{
struct mqnic_eq *eq;
eq = kzalloc(sizeof(*eq), GFP_KERNEL);
if (!eq)
return ERR_PTR(-ENOMEM);
eq->dev = interface->dev;
eq->interface = interface;
eq->eqn = -1;
eq->enabled = 0;
eq->irq_nb.notifier_call = mqnic_eq_int;
eq->hw_addr = NULL;
eq->prod_ptr = 0;
eq->cons_ptr = 0;
spin_lock_init(&eq->table_lock);
INIT_RADIX_TREE(&eq->cq_table, GFP_KERNEL);
return eq;
}
void mqnic_destroy_eq(struct mqnic_eq *eq)
{
mqnic_close_eq(eq);
kfree(eq);
}
int mqnic_open_eq(struct mqnic_eq *eq, struct mqnic_irq *irq, int size)
{
int ret;
if (eq->enabled || eq->hw_addr || eq->buf || !irq)
return -EINVAL;
eq->eqn = mqnic_res_alloc(eq->interface->eq_res);
if (eq->eqn < 0)
return -ENOMEM;
eq->size = roundup_pow_of_two(size);
eq->size_mask = eq->size - 1;
eq->stride = roundup_pow_of_two(MQNIC_EVENT_SIZE);
eq->buf_size = eq->size * eq->stride;
eq->buf = dma_alloc_coherent(eq->dev, eq->buf_size, &eq->buf_dma_addr, GFP_KERNEL);
if (!eq->buf) {
ret = -ENOMEM;
goto fail;
}
// register interrupt
ret = atomic_notifier_chain_register(&irq->nh, &eq->irq_nb);
if (ret)
goto fail;
eq->irq = irq;
eq->hw_addr = mqnic_res_get_addr(eq->interface->eq_res, eq->eqn);
eq->prod_ptr = 0;
eq->cons_ptr = 0;
memset(eq->buf, 1, eq->buf_size);
// deactivate queue
iowrite32(MQNIC_EQ_CMD_SET_ENABLE | 0, eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
// set base address
iowrite32((eq->buf_dma_addr & 0xfffff000),
eq->hw_addr + MQNIC_EQ_BASE_ADDR_VF_REG + 0);
iowrite32(eq->buf_dma_addr >> 32,
eq->hw_addr + MQNIC_EQ_BASE_ADDR_VF_REG + 4);
// set size
iowrite32(MQNIC_EQ_CMD_SET_SIZE | ilog2(eq->size),
eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
// set IRQN
iowrite32(MQNIC_EQ_CMD_SET_IRQN | eq->irq->index,
eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
// set pointers
iowrite32(MQNIC_EQ_CMD_SET_PROD_PTR | (eq->prod_ptr & MQNIC_EQ_PTR_MASK),
eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
iowrite32(MQNIC_EQ_CMD_SET_CONS_PTR | (eq->cons_ptr & MQNIC_EQ_PTR_MASK),
eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
// activate queue
iowrite32(MQNIC_EQ_CMD_SET_ENABLE | 1, eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
eq->enabled = 1;
return 0;
fail:
mqnic_close_eq(eq);
return ret;
}
void mqnic_close_eq(struct mqnic_eq *eq)
{
int ret;
if (eq->hw_addr) {
// deactivate queue
iowrite32(MQNIC_EQ_CMD_SET_ENABLE | 0, eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
}
// unregister interrupt
if (eq->irq)
ret = atomic_notifier_chain_unregister(&eq->irq->nh, &eq->irq_nb);
eq->irq = NULL;
eq->hw_addr = NULL;
if (eq->buf) {
dma_free_coherent(eq->dev, eq->buf_size, eq->buf, eq->buf_dma_addr);
eq->buf = NULL;
eq->buf_dma_addr = 0;
}
mqnic_res_free(eq->interface->eq_res, eq->eqn);
eq->eqn = -1;
eq->enabled = 0;
}
int mqnic_eq_attach_cq(struct mqnic_eq *eq, struct mqnic_cq *cq)
{
int ret;
spin_lock_irq(&eq->table_lock);
ret = radix_tree_insert(&eq->cq_table, cq->cqn, cq);
spin_unlock_irq(&eq->table_lock);
return ret;
}
void mqnic_eq_detach_cq(struct mqnic_eq *eq, struct mqnic_cq *cq)
{
struct mqnic_cq *item;
spin_lock_irq(&eq->table_lock);
item = radix_tree_delete(&eq->cq_table, cq->cqn);
spin_unlock_irq(&eq->table_lock);
if (IS_ERR(item)) {
dev_err(eq->dev, "%s on IF %d EQ %d: radix_tree_delete failed: %ld",
__func__, eq->interface->index, eq->eqn, PTR_ERR(item));
} else if (!item) {
dev_err(eq->dev, "%s on IF %d EQ %d: CQ %d not in table",
__func__, eq->interface->index, eq->eqn, cq->cqn);
} else if (item != cq) {
dev_err(eq->dev, "%s on IF %d EQ %d: entry mismatch when removing CQ %d",
__func__, eq->interface->index, eq->eqn, cq->cqn);
}
}
void mqnic_eq_read_prod_ptr(struct mqnic_eq *eq)
{
eq->prod_ptr += ((ioread32(eq->hw_addr + MQNIC_EQ_PTR_REG) & MQNIC_EQ_PTR_MASK) - eq->prod_ptr) & MQNIC_EQ_PTR_MASK;
}
void mqnic_eq_write_cons_ptr(struct mqnic_eq *eq)
{
iowrite32(MQNIC_EQ_CMD_SET_CONS_PTR | (eq->cons_ptr & MQNIC_EQ_PTR_MASK),
eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
}
void mqnic_arm_eq(struct mqnic_eq *eq)
{
if (!eq->enabled)
return;
iowrite32(MQNIC_EQ_CMD_SET_ARM | 1, eq->hw_addr + MQNIC_EQ_CTRL_STATUS_REG);
}
void mqnic_process_eq(struct mqnic_eq *eq)
{
struct mqnic_if *interface = eq->interface;
struct mqnic_event *event;
struct mqnic_cq *cq;
u32 eq_index;
u32 eq_cons_ptr;
int done = 0;
eq_cons_ptr = eq->cons_ptr;
eq_index = eq_cons_ptr & eq->size_mask;
while (1) {
event = (struct mqnic_event *)(eq->buf + eq_index * eq->stride);
if (!!(event->phase & cpu_to_le32(0x80000000)) == !!(eq_cons_ptr & eq->size))
break;
dma_rmb();
if (event->type == MQNIC_EVENT_TYPE_CPL) {
// completion event
rcu_read_lock();
cq = radix_tree_lookup(&eq->cq_table, le16_to_cpu(event->source));
rcu_read_unlock();
if (likely(cq)) {
if (likely(cq->handler))
cq->handler(cq);
} else {
dev_err(eq->dev, "%s on IF %d EQ %d: unknown event source %d (index %d, type %d)",
__func__, interface->index, eq->eqn, le16_to_cpu(event->source),
eq_index, le16_to_cpu(event->type));
print_hex_dump(KERN_ERR, "", DUMP_PREFIX_NONE, 16, 1,
event, MQNIC_EVENT_SIZE, true);
}
} else {
dev_err(eq->dev, "%s on IF %d EQ %d: unknown event type %d (index %d, source %d)",
__func__, interface->index, eq->eqn, le16_to_cpu(event->type),
eq_index, le16_to_cpu(event->source));
print_hex_dump(KERN_ERR, "", DUMP_PREFIX_NONE, 16, 1,
event, MQNIC_EVENT_SIZE, true);
}
done++;
eq_cons_ptr++;
eq_index = eq_cons_ptr & eq->size_mask;
}
// update EQ consumer pointer
eq->cons_ptr = eq_cons_ptr;
mqnic_eq_write_cons_ptr(eq);
}