/*- * Copyright (c) 2016-2018 Ruslan Bukin * All rights reserved. * * This software was developed by SRI International and the University of * Cambridge Computer Laboratory under DARPA/AFRL contract FA8750-10-C-0237 * ("CTSRD"), as part of the DARPA CRASH research programme. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ /* Ingenic JZ4780 PDMA Controller. */ #include __FBSDID("$FreeBSD$"); #include "opt_platform.h" #include #include #include #include #include #include #include #include #include #include #include #include #ifdef FDT #include #include #include #endif #include #include #include #include "xdma_if.h" #define PDMA_DEBUG #undef PDMA_DEBUG #ifdef PDMA_DEBUG #define dprintf(fmt, ...) printf(fmt, ##__VA_ARGS__) #else #define dprintf(fmt, ...) #endif #define PDMA_DESC_RING_ALIGN 2048 struct pdma_softc { device_t dev; struct resource *res[2]; bus_space_tag_t bst; bus_space_handle_t bsh; void *ih; }; struct pdma_fdt_data { int tx; int rx; int chan; }; struct pdma_channel { struct pdma_fdt_data data; int cur_desc; int used; int index; int flags; #define CHAN_DESCR_RELINK (1 << 0) /* Descriptors */ bus_dma_tag_t desc_tag; bus_dmamap_t desc_map; struct pdma_hwdesc *desc_ring; bus_addr_t desc_ring_paddr; /* xDMA */ xdma_channel_t *xchan; struct xdma_request *req; }; #define PDMA_NCHANNELS 32 struct pdma_channel pdma_channels[PDMA_NCHANNELS]; static struct resource_spec pdma_spec[] = { { SYS_RES_MEMORY, 0, RF_ACTIVE }, { SYS_RES_IRQ, 0, RF_ACTIVE }, { -1, 0 } }; static int pdma_probe(device_t dev); static int pdma_attach(device_t dev); static int pdma_detach(device_t dev); static int chan_start(struct pdma_softc *sc, struct pdma_channel *chan); static void pdma_intr(void *arg) { struct xdma_request *req; xdma_transfer_status_t status; struct pdma_channel *chan; struct pdma_softc *sc; xdma_channel_t *xchan; int pending; int i; sc = arg; pending = READ4(sc, PDMA_DIRQP); /* Ack all the channels. */ WRITE4(sc, PDMA_DIRQP, 0); for (i = 0; i < PDMA_NCHANNELS; i++) { if (pending & (1 << i)) { chan = &pdma_channels[i]; xchan = chan->xchan; req = chan->req; /* TODO: check for AR, HLT error bits here. */ /* Disable channel */ WRITE4(sc, PDMA_DCS(chan->index), 0); if (chan->flags & CHAN_DESCR_RELINK) { /* Enable again */ chan->cur_desc = (chan->cur_desc + 1) % \ req->block_num; chan_start(sc, chan); } status.error = 0; xdma_callback(chan->xchan, &status); } } } static int pdma_probe(device_t dev) { if (!ofw_bus_status_okay(dev)) return (ENXIO); if (!ofw_bus_is_compatible(dev, "ingenic,jz4780-dma")) return (ENXIO); device_set_desc(dev, "Ingenic JZ4780 PDMA Controller"); return (BUS_PROBE_DEFAULT); } static int pdma_attach(device_t dev) { struct pdma_softc *sc; phandle_t xref, node; int err; int reg; sc = device_get_softc(dev); sc->dev = dev; if (bus_alloc_resources(dev, pdma_spec, sc->res)) { device_printf(dev, "could not allocate resources for device\n"); return (ENXIO); } /* Memory interface */ sc->bst = rman_get_bustag(sc->res[0]); sc->bsh = rman_get_bushandle(sc->res[0]); /* Setup interrupt handler */ err = bus_setup_intr(dev, sc->res[1], INTR_TYPE_MISC | INTR_MPSAFE, NULL, pdma_intr, sc, &sc->ih); if (err) { device_printf(dev, "Unable to alloc interrupt resource.\n"); return (ENXIO); } node = ofw_bus_get_node(dev); xref = OF_xref_from_node(node); OF_device_register_xref(xref, dev); reg = READ4(sc, PDMA_DMAC); reg &= ~(DMAC_HLT | DMAC_AR); reg |= (DMAC_DMAE); WRITE4(sc, PDMA_DMAC, reg); WRITE4(sc, PDMA_DMACP, 0); return (0); } static int pdma_detach(device_t dev) { struct pdma_softc *sc; sc = device_get_softc(dev); bus_release_resources(dev, pdma_spec, sc->res); return (0); } static int chan_start(struct pdma_softc *sc, struct pdma_channel *chan) { struct xdma_channel *xchan; xchan = chan->xchan; /* 8 byte descriptor. */ WRITE4(sc, PDMA_DCS(chan->index), DCS_DES8); WRITE4(sc, PDMA_DDA(chan->index), chan->desc_ring_paddr + 8 * 4 * chan->cur_desc); WRITE4(sc, PDMA_DDS, (1 << chan->index)); /* Channel transfer enable. */ WRITE4(sc, PDMA_DCS(chan->index), (DCS_DES8 | DCS_CTE)); return (0); } static int chan_stop(struct pdma_softc *sc, struct pdma_channel *chan) { int timeout; WRITE4(sc, PDMA_DCS(chan->index), 0); timeout = 100; do { if ((READ4(sc, PDMA_DCS(chan->index)) & DCS_CTE) == 0) { break; } } while (timeout--); if (timeout == 0) { device_printf(sc->dev, "%s: Can't stop channel %d\n", __func__, chan->index); } return (0); } static void dwc_get1paddr(void *arg, bus_dma_segment_t *segs, int nsegs, int error) { if (error != 0) return; *(bus_addr_t *)arg = segs[0].ds_addr; } static int pdma_channel_setup_descriptors(device_t dev, struct pdma_channel *chan) { struct pdma_softc *sc; int error; sc = device_get_softc(dev); /* * Set up TX descriptor ring, descriptors, and dma maps. */ error = bus_dma_tag_create( bus_get_dma_tag(sc->dev), /* Parent tag. */ PDMA_DESC_RING_ALIGN, 0, /* alignment, boundary */ BUS_SPACE_MAXADDR_32BIT, /* lowaddr */ BUS_SPACE_MAXADDR, /* highaddr */ NULL, NULL, /* filter, filterarg */ CHAN_DESC_SIZE, 1, /* maxsize, nsegments */ CHAN_DESC_SIZE, /* maxsegsize */ 0, /* flags */ NULL, NULL, /* lockfunc, lockarg */ &chan->desc_tag); if (error != 0) { device_printf(sc->dev, "could not create TX ring DMA tag.\n"); return (-1); } error = bus_dmamem_alloc(chan->desc_tag, (void**)&chan->desc_ring, BUS_DMA_COHERENT | BUS_DMA_WAITOK | BUS_DMA_ZERO, &chan->desc_map); if (error != 0) { device_printf(sc->dev, "could not allocate TX descriptor ring.\n"); return (-1); } error = bus_dmamap_load(chan->desc_tag, chan->desc_map, chan->desc_ring, CHAN_DESC_SIZE, dwc_get1paddr, &chan->desc_ring_paddr, 0); if (error != 0) { device_printf(sc->dev, "could not load TX descriptor ring map.\n"); return (-1); } return (0); } static int pdma_channel_alloc(device_t dev, struct xdma_channel *xchan) { struct pdma_channel *chan; struct pdma_softc *sc; int i; sc = device_get_softc(dev); for (i = 0; i < PDMA_NCHANNELS; i++) { chan = &pdma_channels[i]; if (chan->used == 0) { chan->xchan = xchan; xchan->chan = (void *)chan; chan->used = 1; chan->index = i; pdma_channel_setup_descriptors(dev, chan); return (0); } } return (-1); } static int pdma_channel_free(device_t dev, struct xdma_channel *xchan) { struct pdma_channel *chan; struct pdma_softc *sc; sc = device_get_softc(dev); chan = (struct pdma_channel *)xchan->chan; chan->used = 0; return (0); } static int access_width(struct xdma_request *req, uint32_t *dcm, uint32_t *max_width) { *dcm = 0; *max_width = max(req->src_width, req->dst_width); switch (req->src_width) { case 1: *dcm |= DCM_SP_1; break; case 2: *dcm |= DCM_SP_2; break; case 4: *dcm |= DCM_SP_4; break; default: return (-1); } switch (req->dst_width) { case 1: *dcm |= DCM_DP_1; break; case 2: *dcm |= DCM_DP_2; break; case 4: *dcm |= DCM_DP_4; break; default: return (-1); } switch (*max_width) { case 1: *dcm |= DCM_TSZ_1; break; case 2: *dcm |= DCM_TSZ_2; break; case 4: *dcm |= DCM_TSZ_4; break; default: return (-1); }; return (0); } static int pdma_channel_request(device_t dev, struct xdma_channel *xchan, struct xdma_request *req) { struct pdma_fdt_data *data; struct pdma_channel *chan; struct pdma_hwdesc *desc; xdma_controller_t *xdma; struct pdma_softc *sc; int max_width; uint32_t reg; uint32_t dcm; int i; sc = device_get_softc(dev); dprintf("%s: block_len %d block_num %d\n", __func__, req->block_len, req->block_num); xdma = xchan->xdma; data = (struct pdma_fdt_data *)xdma->data; chan = (struct pdma_channel *)xchan->chan; /* Ensure we are not in operation */ chan_stop(sc, chan); if (req->operation == XDMA_CYCLIC) chan->flags = CHAN_DESCR_RELINK; chan->cur_desc = 0; chan->req = req; for (i = 0; i < req->block_num; i++) { desc = &chan->desc_ring[i]; if (req->direction == XDMA_MEM_TO_DEV) { desc->dsa = req->src_addr + (i * req->block_len); desc->dta = req->dst_addr; desc->drt = data->tx; desc->dcm = DCM_SAI; } else if (req->direction == XDMA_DEV_TO_MEM) { desc->dsa = req->src_addr; desc->dta = req->dst_addr + (i * req->block_len); desc->drt = data->rx; desc->dcm = DCM_DAI; } else if (req->direction == XDMA_MEM_TO_MEM) { desc->dsa = req->src_addr + (i * req->block_len); desc->dta = req->dst_addr + (i * req->block_len); desc->drt = DRT_AUTO; desc->dcm = DCM_SAI | DCM_DAI; } if (access_width(req, &dcm, &max_width) != 0) { device_printf(dev, "%s: can't configure access width\n", __func__); return (-1); } desc->dcm |= dcm | DCM_TIE; desc->dtc = (req->block_len / max_width); /* * TODO: bus dma pre read/write sync here */ /* * PDMA does not provide interrupt after processing each descriptor, * but after processing all the chain only. * As a workaround we do unlink descriptors here, so our chain will * consists of single descriptor only. And then we reconfigure channel * on each interrupt again. */ if ((chan->flags & CHAN_DESCR_RELINK) == 0) { if (i != (req->block_num - 1)) { desc->dcm |= DCM_LINK; reg = ((i + 1) * sizeof(struct pdma_hwdesc)); desc->dtc |= (reg >> 4) << 24; } } } return (0); } static int pdma_channel_control(device_t dev, xdma_channel_t *xchan, int cmd) { struct pdma_channel *chan; struct pdma_softc *sc; sc = device_get_softc(dev); chan = (struct pdma_channel *)xchan->chan; switch (cmd) { case XDMA_CMD_BEGIN: chan_start(sc, chan); break; case XDMA_CMD_TERMINATE: chan_stop(sc, chan); break; case XDMA_CMD_PAUSE: /* TODO: implement me */ return (-1); } return (0); } #ifdef FDT static int pdma_ofw_md_data(device_t dev, pcell_t *cells, int ncells, void **ptr) { struct pdma_fdt_data *data; if (ncells != 3) { return (-1); } data = malloc(sizeof(struct pdma_fdt_data), M_DEVBUF, (M_WAITOK | M_ZERO)); if (data == NULL) { device_printf(dev, "%s: Cant allocate memory\n", __func__); return (-1); } data->tx = cells[0]; data->rx = cells[1]; data->chan = cells[2]; *ptr = data; return (0); } #endif static device_method_t pdma_methods[] = { /* Device interface */ DEVMETHOD(device_probe, pdma_probe), DEVMETHOD(device_attach, pdma_attach), DEVMETHOD(device_detach, pdma_detach), /* xDMA Interface */ DEVMETHOD(xdma_channel_alloc, pdma_channel_alloc), DEVMETHOD(xdma_channel_free, pdma_channel_free), DEVMETHOD(xdma_channel_request, pdma_channel_request), DEVMETHOD(xdma_channel_control, pdma_channel_control), #ifdef FDT DEVMETHOD(xdma_ofw_md_data, pdma_ofw_md_data), #endif DEVMETHOD_END }; static driver_t pdma_driver = { "pdma", pdma_methods, sizeof(struct pdma_softc), }; static devclass_t pdma_devclass; EARLY_DRIVER_MODULE(pdma, simplebus, pdma_driver, pdma_devclass, 0, 0, BUS_PASS_INTERRUPT + BUS_PASS_ORDER_LATE);