freebsd-src/sys/dev/dpaa2/dpaa2_io.c
Dmitry Salychev 58983e4b02
dpaa2: Clean up channels in separate tasks
Each channel gets its own DMA resources, cleanup and "bufferpool"
tasks, and a separate cleanup taskqueue to isolate channels operation
as much as possible to avoid various kernel panics under heavy network
load.

As a side-effect of this work, dpaa2_buf structure is simplified and
all of the functions to re-seed those buffers are gathered now in
dpaa2_buf.h and .c files; functions to work with channels are
extracted into dpaa2_channel.h and .c files as well.

Reported by:		dch
Reviewed by:		bz
Approved by:		bz (mentor)
MFC after:		1 week
Differential Revision:	https://reviews.freebsd.org/D41296
2023-08-20 10:17:26 +02:00

604 lines
16 KiB
C
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

/*-
* SPDX-License-Identifier: BSD-2-Clause
*
* Copyright © 2021-2022 Dmitry Salychev
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
#include <sys/cdefs.h>
/*
* QBMan command interface and the DPAA2 I/O (DPIO) driver.
*
* The DPIO object allows configuration of the QBMan software portal with
* optional notification capabilities.
*
* Software portals are used by the driver to communicate with the QBMan. The
* DPIO objects main purpose is to enable the driver to perform I/O enqueue
* and dequeue operations, as well as buffer release and acquire operations
* using QBMan.
*/
#include "opt_rss.h"
#include <sys/param.h>
#include <sys/kernel.h>
#include <sys/bus.h>
#include <sys/rman.h>
#include <sys/module.h>
#include <sys/malloc.h>
#include <sys/lock.h>
#include <sys/mutex.h>
#include <sys/_cpuset.h>
#include <sys/cpuset.h>
#include <sys/taskqueue.h>
#include <sys/smp.h>
#include <vm/vm.h>
#include <machine/bus.h>
#include <machine/resource.h>
#include <dev/pci/pcivar.h>
#ifdef RSS
#include <net/rss_config.h>
#endif
#include "pcib_if.h"
#include "pci_if.h"
#include "dpaa2_mc.h"
#include "dpaa2_mcp.h"
#include "dpaa2_swp.h"
#include "dpaa2_swp_if.h"
#include "dpaa2_cmd_if.h"
#include "dpaa2_io.h"
#include "dpaa2_ni.h"
#include "dpaa2_channel.h"
#define DPIO_IRQ_INDEX 0 /* index of the only DPIO IRQ */
#define DPIO_POLL_MAX 32
/*
* Memory:
* 0: cache-enabled part of the QBMan software portal.
* 1: cache-inhibited part of the QBMan software portal.
* 2: control registers of the QBMan software portal?
*
* Note that MSI should be allocated separately using pseudo-PCI interface.
*/
struct resource_spec dpaa2_io_spec[] = {
/*
* System Memory resources.
*/
#define MEM_RES_NUM (3u)
#define MEM_RID_OFF (0u)
#define MEM_RID(rid) ((rid) + MEM_RID_OFF)
{ SYS_RES_MEMORY, MEM_RID(0), RF_ACTIVE | RF_UNMAPPED },
{ SYS_RES_MEMORY, MEM_RID(1), RF_ACTIVE | RF_UNMAPPED },
{ SYS_RES_MEMORY, MEM_RID(2), RF_ACTIVE | RF_UNMAPPED | RF_OPTIONAL },
/*
* DPMCP resources.
*
* NOTE: MC command portals (MCPs) are used to send commands to, and
* receive responses from, the MC firmware. One portal per DPIO.
*/
#define MCP_RES_NUM (1u)
#define MCP_RID_OFF (MEM_RID_OFF + MEM_RES_NUM)
#define MCP_RID(rid) ((rid) + MCP_RID_OFF)
/* --- */
{ DPAA2_DEV_MCP, MCP_RID(0), RF_ACTIVE | RF_SHAREABLE | RF_OPTIONAL },
/* --- */
RESOURCE_SPEC_END
};
/* Configuration routines. */
static int dpaa2_io_setup_irqs(device_t dev);
static int dpaa2_io_release_irqs(device_t dev);
static int dpaa2_io_setup_msi(struct dpaa2_io_softc *sc);
static int dpaa2_io_release_msi(struct dpaa2_io_softc *sc);
/* Interrupt handlers */
static void dpaa2_io_intr(void *arg);
static int
dpaa2_io_probe(device_t dev)
{
/* DPIO device will be added by a parent resource container itself. */
device_set_desc(dev, "DPAA2 I/O");
return (BUS_PROBE_DEFAULT);
}
static int
dpaa2_io_detach(device_t dev)
{
device_t pdev = device_get_parent(dev);
device_t child = dev;
struct dpaa2_io_softc *sc = device_get_softc(dev);
struct dpaa2_devinfo *rcinfo = device_get_ivars(pdev);
struct dpaa2_devinfo *dinfo = device_get_ivars(dev);
struct dpaa2_cmd cmd;
uint16_t rc_token, io_token;
int error;
DPAA2_CMD_INIT(&cmd);
/* Tear down interrupt handler and release IRQ resources. */
dpaa2_io_release_irqs(dev);
/* Free software portal helper object. */
dpaa2_swp_free_portal(sc->swp);
error = DPAA2_CMD_RC_OPEN(dev, child, &cmd, rcinfo->id, &rc_token);
if (error) {
device_printf(dev, "%s: failed to open DPRC: error=%d\n",
__func__, error);
goto err_exit;
}
error = DPAA2_CMD_IO_OPEN(dev, child, &cmd, dinfo->id, &io_token);
if (error) {
device_printf(dev, "%s: failed to open DPIO: id=%d, error=%d\n",
__func__, dinfo->id, error);
goto close_rc;
}
error = DPAA2_CMD_IO_DISABLE(dev, child, &cmd);
if (error && bootverbose) {
device_printf(dev, "%s: failed to disable DPIO: id=%d, "
"error=%d\n", __func__, dinfo->id, error);
}
(void)DPAA2_CMD_IO_CLOSE(dev, child, &cmd);
(void)DPAA2_CMD_RC_CLOSE(dev, child, DPAA2_CMD_TK(&cmd, rc_token));
/* Unmap memory resources of the portal. */
for (int i = 0; i < MEM_RES_NUM; i++) {
if (sc->res[MEM_RID(i)] == NULL) {
continue;
}
error = bus_unmap_resource(sc->dev, SYS_RES_MEMORY,
sc->res[MEM_RID(i)], &sc->map[MEM_RID(i)]);
if (error && bootverbose) {
device_printf(dev, "%s: failed to unmap memory "
"resource: rid=%d, error=%d\n", __func__, MEM_RID(i),
error);
}
}
/* Release allocated resources. */
bus_release_resources(dev, dpaa2_io_spec, sc->res);
return (0);
close_rc:
(void)DPAA2_CMD_RC_CLOSE(dev, child, DPAA2_CMD_TK(&cmd, rc_token));
err_exit:
return (error);
}
static int
dpaa2_io_attach(device_t dev)
{
device_t pdev = device_get_parent(dev);
device_t child = dev;
device_t mcp_dev;
struct dpaa2_io_softc *sc = device_get_softc(dev);
struct dpaa2_devinfo *rcinfo = device_get_ivars(pdev);
struct dpaa2_devinfo *dinfo = device_get_ivars(dev);
struct dpaa2_devinfo *mcp_dinfo;
struct dpaa2_cmd cmd;
struct resource_map_request req;
struct {
vm_memattr_t memattr;
char *label;
} map_args[MEM_RES_NUM] = {
{ VM_MEMATTR_WRITE_BACK, "cache-enabled part" },
{ VM_MEMATTR_DEVICE, "cache-inhibited part" },
{ VM_MEMATTR_DEVICE, "control registers" }
};
uint16_t rc_token, io_token;
int error;
sc->dev = dev;
sc->swp = NULL;
sc->intr = NULL;
sc->irq_resource = NULL;
/* Allocate resources. */
error = bus_alloc_resources(sc->dev, dpaa2_io_spec, sc->res);
if (error) {
device_printf(dev, "%s: failed to allocate resources: "
"error=%d\n", __func__, error);
return (ENXIO);
}
/* Set allocated MC portal up. */
mcp_dev = (device_t) rman_get_start(sc->res[MCP_RID(0)]);
mcp_dinfo = device_get_ivars(mcp_dev);
dinfo->portal = mcp_dinfo->portal;
/* Map memory resources of the portal. */
for (int i = 0; i < MEM_RES_NUM; i++) {
if (sc->res[MEM_RID(i)] == NULL) {
continue;
}
resource_init_map_request(&req);
req.memattr = map_args[i].memattr;
error = bus_map_resource(sc->dev, SYS_RES_MEMORY,
sc->res[MEM_RID(i)], &req, &sc->map[MEM_RID(i)]);
if (error) {
device_printf(dev, "%s: failed to map %s: error=%d\n",
__func__, map_args[i].label, error);
goto err_exit;
}
}
DPAA2_CMD_INIT(&cmd);
error = DPAA2_CMD_RC_OPEN(dev, child, &cmd, rcinfo->id, &rc_token);
if (error) {
device_printf(dev, "%s: failed to open DPRC: error=%d\n",
__func__, error);
goto err_exit;
}
error = DPAA2_CMD_IO_OPEN(dev, child, &cmd, dinfo->id, &io_token);
if (error) {
device_printf(dev, "%s: failed to open DPIO: id=%d, error=%d\n",
__func__, dinfo->id, error);
goto close_rc;
}
error = DPAA2_CMD_IO_RESET(dev, child, &cmd);
if (error) {
device_printf(dev, "%s: failed to reset DPIO: id=%d, error=%d\n",
__func__, dinfo->id, error);
goto close_io;
}
error = DPAA2_CMD_IO_GET_ATTRIBUTES(dev, child, &cmd, &sc->attr);
if (error) {
device_printf(dev, "%s: failed to get DPIO attributes: id=%d, "
"error=%d\n", __func__, dinfo->id, error);
goto close_io;
}
error = DPAA2_CMD_IO_ENABLE(dev, child, &cmd);
if (error) {
device_printf(dev, "%s: failed to enable DPIO: id=%d, "
"error=%d\n", __func__, dinfo->id, error);
goto close_io;
}
/* Prepare descriptor of the QBMan software portal. */
sc->swp_desc.dpio_dev = dev;
sc->swp_desc.swp_version = sc->attr.swp_version;
sc->swp_desc.swp_clk = sc->attr.swp_clk;
sc->swp_desc.swp_id = sc->attr.swp_id;
sc->swp_desc.has_notif = sc->attr.priors_num ? true : false;
sc->swp_desc.has_8prio = sc->attr.priors_num == 8u ? true : false;
sc->swp_desc.cena_res = sc->res[0];
sc->swp_desc.cena_map = &sc->map[0];
sc->swp_desc.cinh_res = sc->res[1];
sc->swp_desc.cinh_map = &sc->map[1];
/*
* Compute how many 256 QBMAN cycles fit into one ns. This is because
* the interrupt timeout period register needs to be specified in QBMAN
* clock cycles in increments of 256.
*/
sc->swp_desc.swp_cycles_ratio = 256000 /
(sc->swp_desc.swp_clk / 1000000);
/* Initialize QBMan software portal. */
error = dpaa2_swp_init_portal(&sc->swp, &sc->swp_desc, DPAA2_SWP_DEF);
if (error) {
device_printf(dev, "%s: failed to initialize dpaa2_swp: "
"error=%d\n", __func__, error);
goto err_exit;
}
error = dpaa2_io_setup_irqs(dev);
if (error) {
device_printf(dev, "%s: failed to setup IRQs: error=%d\n",
__func__, error);
goto err_exit;
}
if (bootverbose) {
device_printf(dev, "dpio_id=%d, swp_id=%d, chan_mode=%s, "
"notif_priors=%d, swp_version=0x%x\n",
sc->attr.id, sc->attr.swp_id,
sc->attr.chan_mode == DPAA2_IO_LOCAL_CHANNEL
? "local_channel" : "no_channel", sc->attr.priors_num,
sc->attr.swp_version);
}
(void)DPAA2_CMD_IO_CLOSE(dev, child, &cmd);
(void)DPAA2_CMD_RC_CLOSE(dev, child, DPAA2_CMD_TK(&cmd, rc_token));
return (0);
close_io:
(void)DPAA2_CMD_IO_CLOSE(dev, child, DPAA2_CMD_TK(&cmd, io_token));
close_rc:
(void)DPAA2_CMD_RC_CLOSE(dev, child, DPAA2_CMD_TK(&cmd, rc_token));
err_exit:
dpaa2_io_detach(dev);
return (ENXIO);
}
/**
* @brief Enqueue multiple frames to a frame queue using one FQID.
*/
static int
dpaa2_io_enq_multiple_fq(device_t iodev, uint32_t fqid,
struct dpaa2_fd *fd, int frames_n)
{
struct dpaa2_io_softc *sc = device_get_softc(iodev);
struct dpaa2_swp *swp = sc->swp;
struct dpaa2_eq_desc ed;
uint32_t flags = 0;
memset(&ed, 0, sizeof(ed));
/* Setup enqueue descriptor. */
dpaa2_swp_set_ed_norp(&ed, false);
dpaa2_swp_set_ed_fq(&ed, fqid);
return (dpaa2_swp_enq_mult(swp, &ed, fd, &flags, frames_n));
}
/**
* @brief Configure the channel data availability notification (CDAN)
* in a particular WQ channel paired with DPIO.
*/
static int
dpaa2_io_conf_wq_channel(device_t iodev, struct dpaa2_io_notif_ctx *ctx)
{
struct dpaa2_io_softc *sc = device_get_softc(iodev);
/* Enable generation of the CDAN notifications. */
if (ctx->cdan_en) {
return (dpaa2_swp_conf_wq_channel(sc->swp, ctx->fq_chan_id,
DPAA2_WQCHAN_WE_EN | DPAA2_WQCHAN_WE_CTX, ctx->cdan_en,
ctx->qman_ctx));
}
return (0);
}
/**
* @brief Query current configuration/state of the buffer pool.
*/
static int
dpaa2_io_query_bp(device_t iodev, uint16_t bpid, struct dpaa2_bp_conf *conf)
{
struct dpaa2_io_softc *sc = device_get_softc(iodev);
return (dpaa2_swp_query_bp(sc->swp, bpid, conf));
}
/**
* @brief Release one or more buffer pointers to the QBMan buffer pool.
*/
static int
dpaa2_io_release_bufs(device_t iodev, uint16_t bpid, bus_addr_t *buf,
uint32_t buf_num)
{
struct dpaa2_io_softc *sc = device_get_softc(iodev);
return (dpaa2_swp_release_bufs(sc->swp, bpid, buf, buf_num));
}
/**
* @brief Configure DPNI object to generate interrupts.
*/
static int
dpaa2_io_setup_irqs(device_t dev)
{
struct dpaa2_io_softc *sc = device_get_softc(dev);
int error;
/*
* Setup interrupts generated by the software portal.
*/
dpaa2_swp_set_intr_trigger(sc->swp, DPAA2_SWP_INTR_DQRI);
dpaa2_swp_clear_intr_status(sc->swp, 0xFFFFFFFFu);
/* Configure IRQs. */
error = dpaa2_io_setup_msi(sc);
if (error) {
device_printf(dev, "%s: failed to allocate MSI: error=%d\n",
__func__, error);
return (error);
}
if ((sc->irq_resource = bus_alloc_resource_any(dev, SYS_RES_IRQ,
&sc->irq_rid[0], RF_ACTIVE | RF_SHAREABLE)) == NULL) {
device_printf(dev, "%s: failed to allocate IRQ resource\n",
__func__);
return (ENXIO);
}
if (bus_setup_intr(dev, sc->irq_resource, INTR_TYPE_NET | INTR_MPSAFE |
INTR_ENTROPY, NULL, dpaa2_io_intr, sc, &sc->intr)) {
device_printf(dev, "%s: failed to setup IRQ resource\n",
__func__);
return (ENXIO);
}
/* Wrap DPIO ID around number of CPUs/RSS buckets */
#ifdef RSS
sc->cpu = rss_getcpu(sc->attr.id % rss_getnumbuckets());
#else
sc->cpu = sc->attr.id % mp_ncpus;
#endif
CPU_SETOF(sc->cpu, &sc->cpu_mask);
bus_bind_intr(dev, sc->irq_resource, sc->cpu);
/*
* Setup and enable Static Dequeue Command to receive CDANs from
* channel 0.
*/
if (sc->swp_desc.has_notif)
dpaa2_swp_set_push_dequeue(sc->swp, 0, true);
return (0);
}
static int
dpaa2_io_release_irqs(device_t dev)
{
struct dpaa2_io_softc *sc = device_get_softc(dev);
/* Disable receiving CDANs from channel 0. */
if (sc->swp_desc.has_notif)
dpaa2_swp_set_push_dequeue(sc->swp, 0, false);
/* Release IRQ resources. */
if (sc->intr != NULL)
bus_teardown_intr(dev, sc->irq_resource, &sc->intr);
if (sc->irq_resource != NULL)
bus_release_resource(dev, SYS_RES_IRQ, sc->irq_rid[0],
sc->irq_resource);
(void)dpaa2_io_release_msi(device_get_softc(dev));
/* Configure software portal to stop generating interrupts. */
dpaa2_swp_set_intr_trigger(sc->swp, 0);
dpaa2_swp_clear_intr_status(sc->swp, 0xFFFFFFFFu);
return (0);
}
/**
* @brief Allocate MSI interrupts for this DPAA2 I/O object.
*/
static int
dpaa2_io_setup_msi(struct dpaa2_io_softc *sc)
{
int val;
val = pci_msi_count(sc->dev);
if (val < DPAA2_IO_MSI_COUNT)
device_printf(sc->dev, "MSI: actual=%d, expected=%d\n", val,
DPAA2_IO_MSI_COUNT);
val = MIN(val, DPAA2_IO_MSI_COUNT);
if (pci_alloc_msi(sc->dev, &val) != 0)
return (EINVAL);
for (int i = 0; i < val; i++)
sc->irq_rid[i] = i + 1;
return (0);
}
static int
dpaa2_io_release_msi(struct dpaa2_io_softc *sc)
{
int error;
error = pci_release_msi(sc->dev);
if (error) {
device_printf(sc->dev, "%s: failed to release MSI: error=%d/n",
__func__, error);
return (error);
}
return (0);
}
/**
* @brief DPAA2 I/O interrupt handler.
*/
static void
dpaa2_io_intr(void *arg)
{
struct dpaa2_io_softc *sc = (struct dpaa2_io_softc *) arg;
/* struct dpaa2_ni_softc *nisc = NULL; */
struct dpaa2_io_notif_ctx *ctx[DPIO_POLL_MAX];
struct dpaa2_channel *chan;
struct dpaa2_dq dq;
uint32_t idx, status;
uint16_t flags;
int rc, cdan_n = 0;
status = dpaa2_swp_read_intr_status(sc->swp);
if (status == 0) {
return;
}
DPAA2_SWP_LOCK(sc->swp, &flags);
if (flags & DPAA2_SWP_DESTROYED) {
/* Terminate operation if portal is destroyed. */
DPAA2_SWP_UNLOCK(sc->swp);
return;
}
for (int i = 0; i < DPIO_POLL_MAX; i++) {
rc = dpaa2_swp_dqrr_next_locked(sc->swp, &dq, &idx);
if (rc) {
break;
}
if ((dq.common.verb & DPAA2_DQRR_RESULT_MASK) ==
DPAA2_DQRR_RESULT_CDAN) {
ctx[cdan_n++] = (struct dpaa2_io_notif_ctx *) dq.scn.ctx;
} else {
/* TODO: Report unknown DQRR entry. */
}
dpaa2_swp_write_reg(sc->swp, DPAA2_SWP_CINH_DCAP, idx);
}
DPAA2_SWP_UNLOCK(sc->swp);
for (int i = 0; i < cdan_n; i++) {
chan = (struct dpaa2_channel *)ctx[i]->channel;
/* nisc = device_get_softc(chan->ni_dev); */
taskqueue_enqueue(chan->cleanup_tq, &chan->cleanup_task);
}
/* Enable software portal interrupts back */
dpaa2_swp_clear_intr_status(sc->swp, status);
dpaa2_swp_write_reg(sc->swp, DPAA2_SWP_CINH_IIR, 0);
}
static device_method_t dpaa2_io_methods[] = {
/* Device interface */
DEVMETHOD(device_probe, dpaa2_io_probe),
DEVMETHOD(device_attach, dpaa2_io_attach),
DEVMETHOD(device_detach, dpaa2_io_detach),
/* QBMan software portal interface */
DEVMETHOD(dpaa2_swp_enq_multiple_fq, dpaa2_io_enq_multiple_fq),
DEVMETHOD(dpaa2_swp_conf_wq_channel, dpaa2_io_conf_wq_channel),
DEVMETHOD(dpaa2_swp_query_bp, dpaa2_io_query_bp),
DEVMETHOD(dpaa2_swp_release_bufs, dpaa2_io_release_bufs),
DEVMETHOD_END
};
static driver_t dpaa2_io_driver = {
"dpaa2_io",
dpaa2_io_methods,
sizeof(struct dpaa2_io_softc),
};
DRIVER_MODULE(dpaa2_io, dpaa2_rc, dpaa2_io_driver, 0, 0);