// SPDX-License-Identifier: BSD-2-Clause-Views /* * Copyright (c) 2019-2023 The Regents of the University of California */ #include "mqnic.h" struct mqnic_cq *mqnic_create_cq(struct mqnic_if *interface) { struct mqnic_cq *cq; cq = kzalloc(sizeof(*cq), GFP_KERNEL); if (!cq) return ERR_PTR(-ENOMEM); cq->dev = interface->dev; cq->interface = interface; cq->cqn = -1; cq->enabled = 0; cq->hw_addr = NULL; cq->prod_ptr = 0; cq->cons_ptr = 0; return cq; } void mqnic_destroy_cq(struct mqnic_cq *cq) { mqnic_close_cq(cq); kfree(cq); } int mqnic_open_cq(struct mqnic_cq *cq, struct mqnic_eq *eq, int size) { int ret; if (cq->enabled || cq->hw_addr || cq->buf || !eq) return -EINVAL; cq->cqn = mqnic_res_alloc(cq->interface->cq_res); if (cq->cqn < 0) return -ENOMEM; cq->size = roundup_pow_of_two(size); cq->size_mask = cq->size - 1; cq->stride = roundup_pow_of_two(MQNIC_CPL_SIZE); cq->buf_size = cq->size * cq->stride; cq->buf = dma_alloc_coherent(cq->dev, cq->buf_size, &cq->buf_dma_addr, GFP_KERNEL); if (!cq->buf) { ret = -ENOMEM; goto fail; } cq->eq = eq; mqnic_eq_attach_cq(eq, cq); cq->hw_addr = mqnic_res_get_addr(cq->interface->cq_res, cq->cqn); cq->prod_ptr = 0; cq->cons_ptr = 0; memset(cq->buf, 1, cq->buf_size); // deactivate queue iowrite32(MQNIC_CQ_CMD_SET_ENABLE | 0, cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); // set base address iowrite32((cq->buf_dma_addr & 0xfffff000), cq->hw_addr + MQNIC_CQ_BASE_ADDR_VF_REG + 0); iowrite32(cq->buf_dma_addr >> 32, cq->hw_addr + MQNIC_CQ_BASE_ADDR_VF_REG + 4); // set size iowrite32(MQNIC_CQ_CMD_SET_SIZE | ilog2(cq->size), cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); // set EQN iowrite32(MQNIC_CQ_CMD_SET_EQN | cq->eq->eqn, cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); // set pointers iowrite32(MQNIC_CQ_CMD_SET_PROD_PTR | (cq->prod_ptr & MQNIC_CQ_PTR_MASK), cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); iowrite32(MQNIC_CQ_CMD_SET_CONS_PTR | (cq->cons_ptr & MQNIC_CQ_PTR_MASK), cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); // activate queue iowrite32(MQNIC_CQ_CMD_SET_ENABLE | 1, cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); cq->enabled = 1; return 0; fail: mqnic_close_eq(eq); return ret; } void mqnic_close_cq(struct mqnic_cq *cq) { if (cq->hw_addr) { // deactivate queue iowrite32(MQNIC_CQ_CMD_SET_ENABLE | 0, cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); } if (cq->eq) { mqnic_eq_detach_cq(cq->eq, cq); cq->eq = NULL; } cq->hw_addr = NULL; if (cq->buf) { dma_free_coherent(cq->dev, cq->buf_size, cq->buf, cq->buf_dma_addr); cq->buf = NULL; cq->buf_dma_addr = 0; } mqnic_res_free(cq->interface->cq_res, cq->cqn); cq->cqn = -1; cq->enabled = 0; } void mqnic_cq_read_prod_ptr(struct mqnic_cq *cq) { cq->prod_ptr += ((ioread32(cq->hw_addr + MQNIC_CQ_PTR_REG) >> 16) - cq->prod_ptr) & MQNIC_CQ_PTR_MASK; } void mqnic_cq_write_cons_ptr(struct mqnic_cq *cq) { iowrite32(MQNIC_CQ_CMD_SET_CONS_PTR | (cq->cons_ptr & MQNIC_CQ_PTR_MASK), cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); } void mqnic_arm_cq(struct mqnic_cq *cq) { if (!cq->enabled) return; iowrite32(MQNIC_CQ_CMD_SET_ARM | 1, cq->hw_addr + MQNIC_CQ_CTRL_STATUS_REG); }