|  | // SPDX-License-Identifier: GPL-2.0-or-later | 
|  | /* | 
|  | * DMA driver for STMicroelectronics STi FDMA controller | 
|  | * | 
|  | * Copyright (C) 2014 STMicroelectronics | 
|  | * | 
|  | * Author: Ludovic Barre <Ludovic.barre@st.com> | 
|  | *	   Peter Griffin <peter.griffin@linaro.org> | 
|  | */ | 
|  |  | 
|  | #include <linux/init.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/of_device.h> | 
|  | #include <linux/of_dma.h> | 
|  | #include <linux/platform_device.h> | 
|  | #include <linux/interrupt.h> | 
|  | #include <linux/remoteproc.h> | 
|  |  | 
|  | #include "st_fdma.h" | 
|  |  | 
|  | static inline struct st_fdma_chan *to_st_fdma_chan(struct dma_chan *c) | 
|  | { | 
|  | return container_of(c, struct st_fdma_chan, vchan.chan); | 
|  | } | 
|  |  | 
|  | static struct st_fdma_desc *to_st_fdma_desc(struct virt_dma_desc *vd) | 
|  | { | 
|  | return container_of(vd, struct st_fdma_desc, vdesc); | 
|  | } | 
|  |  | 
|  | static int st_fdma_dreq_get(struct st_fdma_chan *fchan) | 
|  | { | 
|  | struct st_fdma_dev *fdev = fchan->fdev; | 
|  | u32 req_line_cfg = fchan->cfg.req_line; | 
|  | u32 dreq_line; | 
|  | int try = 0; | 
|  |  | 
|  | /* | 
|  | * dreq_mask is shared for n channels of fdma, so all accesses must be | 
|  | * atomic. if the dreq_mask is changed between ffz and set_bit, | 
|  | * we retry | 
|  | */ | 
|  | do { | 
|  | if (fdev->dreq_mask == ~0L) { | 
|  | dev_err(fdev->dev, "No req lines available\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (try || req_line_cfg >= ST_FDMA_NR_DREQS) { | 
|  | dev_err(fdev->dev, "Invalid or used req line\n"); | 
|  | return -EINVAL; | 
|  | } else { | 
|  | dreq_line = req_line_cfg; | 
|  | } | 
|  |  | 
|  | try++; | 
|  | } while (test_and_set_bit(dreq_line, &fdev->dreq_mask)); | 
|  |  | 
|  | dev_dbg(fdev->dev, "get dreq_line:%d mask:%#lx\n", | 
|  | dreq_line, fdev->dreq_mask); | 
|  |  | 
|  | return dreq_line; | 
|  | } | 
|  |  | 
|  | static void st_fdma_dreq_put(struct st_fdma_chan *fchan) | 
|  | { | 
|  | struct st_fdma_dev *fdev = fchan->fdev; | 
|  |  | 
|  | dev_dbg(fdev->dev, "put dreq_line:%#x\n", fchan->dreq_line); | 
|  | clear_bit(fchan->dreq_line, &fdev->dreq_mask); | 
|  | } | 
|  |  | 
|  | static void st_fdma_xfer_desc(struct st_fdma_chan *fchan) | 
|  | { | 
|  | struct virt_dma_desc *vdesc; | 
|  | unsigned long nbytes, ch_cmd, cmd; | 
|  |  | 
|  | vdesc = vchan_next_desc(&fchan->vchan); | 
|  | if (!vdesc) | 
|  | return; | 
|  |  | 
|  | fchan->fdesc = to_st_fdma_desc(vdesc); | 
|  | nbytes = fchan->fdesc->node[0].desc->nbytes; | 
|  | cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id); | 
|  | ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START; | 
|  |  | 
|  | /* start the channel for the descriptor */ | 
|  | fnode_write(fchan, nbytes, FDMA_CNTN_OFST); | 
|  | fchan_write(fchan, ch_cmd, FDMA_CH_CMD_OFST); | 
|  | writel(cmd, | 
|  | fchan->fdev->slim_rproc->peri + FDMA_CMD_SET_OFST); | 
|  |  | 
|  | dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id); | 
|  | } | 
|  |  | 
|  | static void st_fdma_ch_sta_update(struct st_fdma_chan *fchan, | 
|  | unsigned long int_sta) | 
|  | { | 
|  | unsigned long ch_sta, ch_err; | 
|  | int ch_id = fchan->vchan.chan.chan_id; | 
|  | struct st_fdma_dev *fdev = fchan->fdev; | 
|  |  | 
|  | ch_sta = fchan_read(fchan, FDMA_CH_CMD_OFST); | 
|  | ch_err = ch_sta & FDMA_CH_CMD_ERR_MASK; | 
|  | ch_sta &= FDMA_CH_CMD_STA_MASK; | 
|  |  | 
|  | if (int_sta & FDMA_INT_STA_ERR) { | 
|  | dev_warn(fdev->dev, "chan:%d, error:%ld\n", ch_id, ch_err); | 
|  | fchan->status = DMA_ERROR; | 
|  | return; | 
|  | } | 
|  |  | 
|  | switch (ch_sta) { | 
|  | case FDMA_CH_CMD_STA_PAUSED: | 
|  | fchan->status = DMA_PAUSED; | 
|  | break; | 
|  |  | 
|  | case FDMA_CH_CMD_STA_RUNNING: | 
|  | fchan->status = DMA_IN_PROGRESS; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | static irqreturn_t st_fdma_irq_handler(int irq, void *dev_id) | 
|  | { | 
|  | struct st_fdma_dev *fdev = dev_id; | 
|  | irqreturn_t ret = IRQ_NONE; | 
|  | struct st_fdma_chan *fchan = &fdev->chans[0]; | 
|  | unsigned long int_sta, clr; | 
|  |  | 
|  | int_sta = fdma_read(fdev, FDMA_INT_STA_OFST); | 
|  | clr = int_sta; | 
|  |  | 
|  | for (; int_sta != 0 ; int_sta >>= 2, fchan++) { | 
|  | if (!(int_sta & (FDMA_INT_STA_CH | FDMA_INT_STA_ERR))) | 
|  | continue; | 
|  |  | 
|  | spin_lock(&fchan->vchan.lock); | 
|  | st_fdma_ch_sta_update(fchan, int_sta); | 
|  |  | 
|  | if (fchan->fdesc) { | 
|  | if (!fchan->fdesc->iscyclic) { | 
|  | list_del(&fchan->fdesc->vdesc.node); | 
|  | vchan_cookie_complete(&fchan->fdesc->vdesc); | 
|  | fchan->fdesc = NULL; | 
|  | fchan->status = DMA_COMPLETE; | 
|  | } else { | 
|  | vchan_cyclic_callback(&fchan->fdesc->vdesc); | 
|  | } | 
|  |  | 
|  | /* Start the next descriptor (if available) */ | 
|  | if (!fchan->fdesc) | 
|  | st_fdma_xfer_desc(fchan); | 
|  | } | 
|  |  | 
|  | spin_unlock(&fchan->vchan.lock); | 
|  | ret = IRQ_HANDLED; | 
|  | } | 
|  |  | 
|  | fdma_write(fdev, clr, FDMA_INT_CLR_OFST); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static struct dma_chan *st_fdma_of_xlate(struct of_phandle_args *dma_spec, | 
|  | struct of_dma *ofdma) | 
|  | { | 
|  | struct st_fdma_dev *fdev = ofdma->of_dma_data; | 
|  | struct dma_chan *chan; | 
|  | struct st_fdma_chan *fchan; | 
|  | int ret; | 
|  |  | 
|  | if (dma_spec->args_count < 1) | 
|  | return ERR_PTR(-EINVAL); | 
|  |  | 
|  | if (fdev->dma_device.dev->of_node != dma_spec->np) | 
|  | return ERR_PTR(-EINVAL); | 
|  |  | 
|  | ret = rproc_boot(fdev->slim_rproc->rproc); | 
|  | if (ret == -ENOENT) | 
|  | return ERR_PTR(-EPROBE_DEFER); | 
|  | else if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | chan = dma_get_any_slave_channel(&fdev->dma_device); | 
|  | if (!chan) | 
|  | goto err_chan; | 
|  |  | 
|  | fchan = to_st_fdma_chan(chan); | 
|  |  | 
|  | fchan->cfg.of_node = dma_spec->np; | 
|  | fchan->cfg.req_line = dma_spec->args[0]; | 
|  | fchan->cfg.req_ctrl = 0; | 
|  | fchan->cfg.type = ST_FDMA_TYPE_FREE_RUN; | 
|  |  | 
|  | if (dma_spec->args_count > 1) | 
|  | fchan->cfg.req_ctrl = dma_spec->args[1] | 
|  | & FDMA_REQ_CTRL_CFG_MASK; | 
|  |  | 
|  | if (dma_spec->args_count > 2) | 
|  | fchan->cfg.type = dma_spec->args[2]; | 
|  |  | 
|  | if (fchan->cfg.type == ST_FDMA_TYPE_FREE_RUN) { | 
|  | fchan->dreq_line = 0; | 
|  | } else { | 
|  | fchan->dreq_line = st_fdma_dreq_get(fchan); | 
|  | if (IS_ERR_VALUE(fchan->dreq_line)) { | 
|  | chan = ERR_PTR(fchan->dreq_line); | 
|  | goto err_chan; | 
|  | } | 
|  | } | 
|  |  | 
|  | dev_dbg(fdev->dev, "xlate req_line:%d type:%d req_ctrl:%#lx\n", | 
|  | fchan->cfg.req_line, fchan->cfg.type, fchan->cfg.req_ctrl); | 
|  |  | 
|  | return chan; | 
|  |  | 
|  | err_chan: | 
|  | rproc_shutdown(fdev->slim_rproc->rproc); | 
|  | return chan; | 
|  |  | 
|  | } | 
|  |  | 
|  | static void st_fdma_free_desc(struct virt_dma_desc *vdesc) | 
|  | { | 
|  | struct st_fdma_desc *fdesc; | 
|  | int i; | 
|  |  | 
|  | fdesc = to_st_fdma_desc(vdesc); | 
|  | for (i = 0; i < fdesc->n_nodes; i++) | 
|  | dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc, | 
|  | fdesc->node[i].pdesc); | 
|  | kfree(fdesc); | 
|  | } | 
|  |  | 
|  | static struct st_fdma_desc *st_fdma_alloc_desc(struct st_fdma_chan *fchan, | 
|  | int sg_len) | 
|  | { | 
|  | struct st_fdma_desc *fdesc; | 
|  | int i; | 
|  |  | 
|  | fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT); | 
|  | if (!fdesc) | 
|  | return NULL; | 
|  |  | 
|  | fdesc->fchan = fchan; | 
|  | fdesc->n_nodes = sg_len; | 
|  | for (i = 0; i < sg_len; i++) { | 
|  | fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool, | 
|  | GFP_NOWAIT, &fdesc->node[i].pdesc); | 
|  | if (!fdesc->node[i].desc) | 
|  | goto err; | 
|  | } | 
|  | return fdesc; | 
|  |  | 
|  | err: | 
|  | while (--i >= 0) | 
|  | dma_pool_free(fchan->node_pool, fdesc->node[i].desc, | 
|  | fdesc->node[i].pdesc); | 
|  | kfree(fdesc); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static int st_fdma_alloc_chan_res(struct dma_chan *chan) | 
|  | { | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  |  | 
|  | /* Create the dma pool for descriptor allocation */ | 
|  | fchan->node_pool = dma_pool_create(dev_name(&chan->dev->device), | 
|  | fchan->fdev->dev, | 
|  | sizeof(struct st_fdma_hw_node), | 
|  | __alignof__(struct st_fdma_hw_node), | 
|  | 0); | 
|  |  | 
|  | if (!fchan->node_pool) { | 
|  | dev_err(fchan->fdev->dev, "unable to allocate desc pool\n"); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | dev_dbg(fchan->fdev->dev, "alloc ch_id:%d type:%d\n", | 
|  | fchan->vchan.chan.chan_id, fchan->cfg.type); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void st_fdma_free_chan_res(struct dma_chan *chan) | 
|  | { | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  | struct rproc *rproc = fchan->fdev->slim_rproc->rproc; | 
|  | unsigned long flags; | 
|  |  | 
|  | dev_dbg(fchan->fdev->dev, "%s: freeing chan:%d\n", | 
|  | __func__, fchan->vchan.chan.chan_id); | 
|  |  | 
|  | if (fchan->cfg.type != ST_FDMA_TYPE_FREE_RUN) | 
|  | st_fdma_dreq_put(fchan); | 
|  |  | 
|  | spin_lock_irqsave(&fchan->vchan.lock, flags); | 
|  | fchan->fdesc = NULL; | 
|  | spin_unlock_irqrestore(&fchan->vchan.lock, flags); | 
|  |  | 
|  | dma_pool_destroy(fchan->node_pool); | 
|  | fchan->node_pool = NULL; | 
|  | memset(&fchan->cfg, 0, sizeof(struct st_fdma_cfg)); | 
|  |  | 
|  | rproc_shutdown(rproc); | 
|  | } | 
|  |  | 
|  | static struct dma_async_tx_descriptor *st_fdma_prep_dma_memcpy( | 
|  | struct dma_chan *chan,	dma_addr_t dst, dma_addr_t src, | 
|  | size_t len, unsigned long flags) | 
|  | { | 
|  | struct st_fdma_chan *fchan; | 
|  | struct st_fdma_desc *fdesc; | 
|  | struct st_fdma_hw_node *hw_node; | 
|  |  | 
|  | if (!len) | 
|  | return NULL; | 
|  |  | 
|  | fchan = to_st_fdma_chan(chan); | 
|  |  | 
|  | /* We only require a single descriptor */ | 
|  | fdesc = st_fdma_alloc_desc(fchan, 1); | 
|  | if (!fdesc) { | 
|  | dev_err(fchan->fdev->dev, "no memory for desc\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | hw_node = fdesc->node[0].desc; | 
|  | hw_node->next = 0; | 
|  | hw_node->control = FDMA_NODE_CTRL_REQ_MAP_FREE_RUN; | 
|  | hw_node->control |= FDMA_NODE_CTRL_SRC_INCR; | 
|  | hw_node->control |= FDMA_NODE_CTRL_DST_INCR; | 
|  | hw_node->control |= FDMA_NODE_CTRL_INT_EON; | 
|  | hw_node->nbytes = len; | 
|  | hw_node->saddr = src; | 
|  | hw_node->daddr = dst; | 
|  | hw_node->generic.length = len; | 
|  | hw_node->generic.sstride = 0; | 
|  | hw_node->generic.dstride = 0; | 
|  |  | 
|  | return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); | 
|  | } | 
|  |  | 
|  | static int config_reqctrl(struct st_fdma_chan *fchan, | 
|  | enum dma_transfer_direction direction) | 
|  | { | 
|  | u32 maxburst = 0, addr = 0; | 
|  | enum dma_slave_buswidth width; | 
|  | int ch_id = fchan->vchan.chan.chan_id; | 
|  | struct st_fdma_dev *fdev = fchan->fdev; | 
|  |  | 
|  | switch (direction) { | 
|  |  | 
|  | case DMA_DEV_TO_MEM: | 
|  | fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_WNR; | 
|  | maxburst = fchan->scfg.src_maxburst; | 
|  | width = fchan->scfg.src_addr_width; | 
|  | addr = fchan->scfg.src_addr; | 
|  | break; | 
|  |  | 
|  | case DMA_MEM_TO_DEV: | 
|  | fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_WNR; | 
|  | maxburst = fchan->scfg.dst_maxburst; | 
|  | width = fchan->scfg.dst_addr_width; | 
|  | addr = fchan->scfg.dst_addr; | 
|  | break; | 
|  |  | 
|  | default: | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_OPCODE_MASK; | 
|  |  | 
|  | switch (width) { | 
|  |  | 
|  | case DMA_SLAVE_BUSWIDTH_1_BYTE: | 
|  | fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST1; | 
|  | break; | 
|  |  | 
|  | case DMA_SLAVE_BUSWIDTH_2_BYTES: | 
|  | fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST2; | 
|  | break; | 
|  |  | 
|  | case DMA_SLAVE_BUSWIDTH_4_BYTES: | 
|  | fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST4; | 
|  | break; | 
|  |  | 
|  | case DMA_SLAVE_BUSWIDTH_8_BYTES: | 
|  | fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST8; | 
|  | break; | 
|  |  | 
|  | default: | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_NUM_OPS_MASK; | 
|  | fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_NUM_OPS(maxburst-1); | 
|  | dreq_write(fchan, fchan->cfg.req_ctrl, FDMA_REQ_CTRL_OFST); | 
|  |  | 
|  | fchan->cfg.dev_addr = addr; | 
|  | fchan->cfg.dir = direction; | 
|  |  | 
|  | dev_dbg(fdev->dev, "chan:%d config_reqctrl:%#x req_ctrl:%#lx\n", | 
|  | ch_id, addr, fchan->cfg.req_ctrl); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void fill_hw_node(struct st_fdma_hw_node *hw_node, | 
|  | struct st_fdma_chan *fchan, | 
|  | enum dma_transfer_direction direction) | 
|  | { | 
|  | if (direction == DMA_MEM_TO_DEV) { | 
|  | hw_node->control |= FDMA_NODE_CTRL_SRC_INCR; | 
|  | hw_node->control |= FDMA_NODE_CTRL_DST_STATIC; | 
|  | hw_node->daddr = fchan->cfg.dev_addr; | 
|  | } else { | 
|  | hw_node->control |= FDMA_NODE_CTRL_SRC_STATIC; | 
|  | hw_node->control |= FDMA_NODE_CTRL_DST_INCR; | 
|  | hw_node->saddr = fchan->cfg.dev_addr; | 
|  | } | 
|  |  | 
|  | hw_node->generic.sstride = 0; | 
|  | hw_node->generic.dstride = 0; | 
|  | } | 
|  |  | 
|  | static inline struct st_fdma_chan *st_fdma_prep_common(struct dma_chan *chan, | 
|  | size_t len, enum dma_transfer_direction direction) | 
|  | { | 
|  | struct st_fdma_chan *fchan; | 
|  |  | 
|  | if (!chan || !len) | 
|  | return NULL; | 
|  |  | 
|  | fchan = to_st_fdma_chan(chan); | 
|  |  | 
|  | if (!is_slave_direction(direction)) { | 
|  | dev_err(fchan->fdev->dev, "bad direction?\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | return fchan; | 
|  | } | 
|  |  | 
|  | static struct dma_async_tx_descriptor *st_fdma_prep_dma_cyclic( | 
|  | struct dma_chan *chan, dma_addr_t buf_addr, size_t len, | 
|  | size_t period_len, enum dma_transfer_direction direction, | 
|  | unsigned long flags) | 
|  | { | 
|  | struct st_fdma_chan *fchan; | 
|  | struct st_fdma_desc *fdesc; | 
|  | int sg_len, i; | 
|  |  | 
|  | fchan = st_fdma_prep_common(chan, len, direction); | 
|  | if (!fchan) | 
|  | return NULL; | 
|  |  | 
|  | if (!period_len) | 
|  | return NULL; | 
|  |  | 
|  | if (config_reqctrl(fchan, direction)) { | 
|  | dev_err(fchan->fdev->dev, "bad width or direction\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | /* the buffer length must be a multiple of period_len */ | 
|  | if (len % period_len != 0) { | 
|  | dev_err(fchan->fdev->dev, "len is not multiple of period\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | sg_len = len / period_len; | 
|  | fdesc = st_fdma_alloc_desc(fchan, sg_len); | 
|  | if (!fdesc) { | 
|  | dev_err(fchan->fdev->dev, "no memory for desc\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | fdesc->iscyclic = true; | 
|  |  | 
|  | for (i = 0; i < sg_len; i++) { | 
|  | struct st_fdma_hw_node *hw_node = fdesc->node[i].desc; | 
|  |  | 
|  | hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; | 
|  |  | 
|  | hw_node->control = | 
|  | FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line); | 
|  | hw_node->control |= FDMA_NODE_CTRL_INT_EON; | 
|  |  | 
|  | fill_hw_node(hw_node, fchan, direction); | 
|  |  | 
|  | if (direction == DMA_MEM_TO_DEV) | 
|  | hw_node->saddr = buf_addr + (i * period_len); | 
|  | else | 
|  | hw_node->daddr = buf_addr + (i * period_len); | 
|  |  | 
|  | hw_node->nbytes = period_len; | 
|  | hw_node->generic.length = period_len; | 
|  | } | 
|  |  | 
|  | return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); | 
|  | } | 
|  |  | 
|  | static struct dma_async_tx_descriptor *st_fdma_prep_slave_sg( | 
|  | struct dma_chan *chan, struct scatterlist *sgl, | 
|  | unsigned int sg_len, enum dma_transfer_direction direction, | 
|  | unsigned long flags, void *context) | 
|  | { | 
|  | struct st_fdma_chan *fchan; | 
|  | struct st_fdma_desc *fdesc; | 
|  | struct st_fdma_hw_node *hw_node; | 
|  | struct scatterlist *sg; | 
|  | int i; | 
|  |  | 
|  | fchan = st_fdma_prep_common(chan, sg_len, direction); | 
|  | if (!fchan) | 
|  | return NULL; | 
|  |  | 
|  | if (!sgl) | 
|  | return NULL; | 
|  |  | 
|  | fdesc = st_fdma_alloc_desc(fchan, sg_len); | 
|  | if (!fdesc) { | 
|  | dev_err(fchan->fdev->dev, "no memory for desc\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | fdesc->iscyclic = false; | 
|  |  | 
|  | for_each_sg(sgl, sg, sg_len, i) { | 
|  | hw_node = fdesc->node[i].desc; | 
|  |  | 
|  | hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc; | 
|  | hw_node->control = FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line); | 
|  |  | 
|  | fill_hw_node(hw_node, fchan, direction); | 
|  |  | 
|  | if (direction == DMA_MEM_TO_DEV) | 
|  | hw_node->saddr = sg_dma_address(sg); | 
|  | else | 
|  | hw_node->daddr = sg_dma_address(sg); | 
|  |  | 
|  | hw_node->nbytes = sg_dma_len(sg); | 
|  | hw_node->generic.length = sg_dma_len(sg); | 
|  | } | 
|  |  | 
|  | /* interrupt at end of last node */ | 
|  | hw_node->control |= FDMA_NODE_CTRL_INT_EON; | 
|  |  | 
|  | return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags); | 
|  | } | 
|  |  | 
|  | static size_t st_fdma_desc_residue(struct st_fdma_chan *fchan, | 
|  | struct virt_dma_desc *vdesc, | 
|  | bool in_progress) | 
|  | { | 
|  | struct st_fdma_desc *fdesc = fchan->fdesc; | 
|  | size_t residue = 0; | 
|  | dma_addr_t cur_addr = 0; | 
|  | int i; | 
|  |  | 
|  | if (in_progress) { | 
|  | cur_addr = fchan_read(fchan, FDMA_CH_CMD_OFST); | 
|  | cur_addr &= FDMA_CH_CMD_DATA_MASK; | 
|  | } | 
|  |  | 
|  | for (i = fchan->fdesc->n_nodes - 1 ; i >= 0; i--) { | 
|  | if (cur_addr == fdesc->node[i].pdesc) { | 
|  | residue += fnode_read(fchan, FDMA_CNTN_OFST); | 
|  | break; | 
|  | } | 
|  | residue += fdesc->node[i].desc->nbytes; | 
|  | } | 
|  |  | 
|  | return residue; | 
|  | } | 
|  |  | 
|  | static enum dma_status st_fdma_tx_status(struct dma_chan *chan, | 
|  | dma_cookie_t cookie, | 
|  | struct dma_tx_state *txstate) | 
|  | { | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  | struct virt_dma_desc *vd; | 
|  | enum dma_status ret; | 
|  | unsigned long flags; | 
|  |  | 
|  | ret = dma_cookie_status(chan, cookie, txstate); | 
|  | if (ret == DMA_COMPLETE || !txstate) | 
|  | return ret; | 
|  |  | 
|  | spin_lock_irqsave(&fchan->vchan.lock, flags); | 
|  | vd = vchan_find_desc(&fchan->vchan, cookie); | 
|  | if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie) | 
|  | txstate->residue = st_fdma_desc_residue(fchan, vd, true); | 
|  | else if (vd) | 
|  | txstate->residue = st_fdma_desc_residue(fchan, vd, false); | 
|  | else | 
|  | txstate->residue = 0; | 
|  |  | 
|  | spin_unlock_irqrestore(&fchan->vchan.lock, flags); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void st_fdma_issue_pending(struct dma_chan *chan) | 
|  | { | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  | unsigned long flags; | 
|  |  | 
|  | spin_lock_irqsave(&fchan->vchan.lock, flags); | 
|  |  | 
|  | if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc) | 
|  | st_fdma_xfer_desc(fchan); | 
|  |  | 
|  | spin_unlock_irqrestore(&fchan->vchan.lock, flags); | 
|  | } | 
|  |  | 
|  | static int st_fdma_pause(struct dma_chan *chan) | 
|  | { | 
|  | unsigned long flags; | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  | int ch_id = fchan->vchan.chan.chan_id; | 
|  | unsigned long cmd = FDMA_CMD_PAUSE(ch_id); | 
|  |  | 
|  | dev_dbg(fchan->fdev->dev, "pause chan:%d\n", ch_id); | 
|  |  | 
|  | spin_lock_irqsave(&fchan->vchan.lock, flags); | 
|  | if (fchan->fdesc) | 
|  | fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST); | 
|  | spin_unlock_irqrestore(&fchan->vchan.lock, flags); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int st_fdma_resume(struct dma_chan *chan) | 
|  | { | 
|  | unsigned long flags; | 
|  | unsigned long val; | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  | int ch_id = fchan->vchan.chan.chan_id; | 
|  |  | 
|  | dev_dbg(fchan->fdev->dev, "resume chan:%d\n", ch_id); | 
|  |  | 
|  | spin_lock_irqsave(&fchan->vchan.lock, flags); | 
|  | if (fchan->fdesc) { | 
|  | val = fchan_read(fchan, FDMA_CH_CMD_OFST); | 
|  | val &= FDMA_CH_CMD_DATA_MASK; | 
|  | fchan_write(fchan, val, FDMA_CH_CMD_OFST); | 
|  | } | 
|  | spin_unlock_irqrestore(&fchan->vchan.lock, flags); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int st_fdma_terminate_all(struct dma_chan *chan) | 
|  | { | 
|  | unsigned long flags; | 
|  | LIST_HEAD(head); | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  | int ch_id = fchan->vchan.chan.chan_id; | 
|  | unsigned long cmd = FDMA_CMD_PAUSE(ch_id); | 
|  |  | 
|  | dev_dbg(fchan->fdev->dev, "terminate chan:%d\n", ch_id); | 
|  |  | 
|  | spin_lock_irqsave(&fchan->vchan.lock, flags); | 
|  | fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST); | 
|  | fchan->fdesc = NULL; | 
|  | vchan_get_all_descriptors(&fchan->vchan, &head); | 
|  | spin_unlock_irqrestore(&fchan->vchan.lock, flags); | 
|  | vchan_dma_desc_free_list(&fchan->vchan, &head); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int st_fdma_slave_config(struct dma_chan *chan, | 
|  | struct dma_slave_config *slave_cfg) | 
|  | { | 
|  | struct st_fdma_chan *fchan = to_st_fdma_chan(chan); | 
|  |  | 
|  | memcpy(&fchan->scfg, slave_cfg, sizeof(fchan->scfg)); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static const struct st_fdma_driverdata fdma_mpe31_stih407_11 = { | 
|  | .name = "STiH407", | 
|  | .id = 0, | 
|  | }; | 
|  |  | 
|  | static const struct st_fdma_driverdata fdma_mpe31_stih407_12 = { | 
|  | .name = "STiH407", | 
|  | .id = 1, | 
|  | }; | 
|  |  | 
|  | static const struct st_fdma_driverdata fdma_mpe31_stih407_13 = { | 
|  | .name = "STiH407", | 
|  | .id = 2, | 
|  | }; | 
|  |  | 
|  | static const struct of_device_id st_fdma_match[] = { | 
|  | { .compatible = "st,stih407-fdma-mpe31-11" | 
|  | , .data = &fdma_mpe31_stih407_11 }, | 
|  | { .compatible = "st,stih407-fdma-mpe31-12" | 
|  | , .data = &fdma_mpe31_stih407_12 }, | 
|  | { .compatible = "st,stih407-fdma-mpe31-13" | 
|  | , .data = &fdma_mpe31_stih407_13 }, | 
|  | {}, | 
|  | }; | 
|  | MODULE_DEVICE_TABLE(of, st_fdma_match); | 
|  |  | 
|  | static int st_fdma_parse_dt(struct platform_device *pdev, | 
|  | const struct st_fdma_driverdata *drvdata, | 
|  | struct st_fdma_dev *fdev) | 
|  | { | 
|  | snprintf(fdev->fw_name, FW_NAME_SIZE, "fdma_%s_%d.elf", | 
|  | drvdata->name, drvdata->id); | 
|  |  | 
|  | return of_property_read_u32(pdev->dev.of_node, "dma-channels", | 
|  | &fdev->nr_channels); | 
|  | } | 
|  | #define FDMA_DMA_BUSWIDTHS	(BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \ | 
|  | BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \ | 
|  | BIT(DMA_SLAVE_BUSWIDTH_3_BYTES) | \ | 
|  | BIT(DMA_SLAVE_BUSWIDTH_4_BYTES)) | 
|  |  | 
|  | static void st_fdma_free(struct st_fdma_dev *fdev) | 
|  | { | 
|  | struct st_fdma_chan *fchan; | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < fdev->nr_channels; i++) { | 
|  | fchan = &fdev->chans[i]; | 
|  | list_del(&fchan->vchan.chan.device_node); | 
|  | tasklet_kill(&fchan->vchan.task); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int st_fdma_probe(struct platform_device *pdev) | 
|  | { | 
|  | struct st_fdma_dev *fdev; | 
|  | const struct of_device_id *match; | 
|  | struct device_node *np = pdev->dev.of_node; | 
|  | const struct st_fdma_driverdata *drvdata; | 
|  | int ret, i; | 
|  |  | 
|  | match = of_match_device((st_fdma_match), &pdev->dev); | 
|  | if (!match || !match->data) { | 
|  | dev_err(&pdev->dev, "No device match found\n"); | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | drvdata = match->data; | 
|  |  | 
|  | fdev = devm_kzalloc(&pdev->dev, sizeof(*fdev), GFP_KERNEL); | 
|  | if (!fdev) | 
|  | return -ENOMEM; | 
|  |  | 
|  | ret = st_fdma_parse_dt(pdev, drvdata, fdev); | 
|  | if (ret) { | 
|  | dev_err(&pdev->dev, "unable to find platform data\n"); | 
|  | goto err; | 
|  | } | 
|  |  | 
|  | fdev->chans = devm_kcalloc(&pdev->dev, fdev->nr_channels, | 
|  | sizeof(struct st_fdma_chan), GFP_KERNEL); | 
|  | if (!fdev->chans) | 
|  | return -ENOMEM; | 
|  |  | 
|  | fdev->dev = &pdev->dev; | 
|  | fdev->drvdata = drvdata; | 
|  | platform_set_drvdata(pdev, fdev); | 
|  |  | 
|  | fdev->irq = platform_get_irq(pdev, 0); | 
|  | if (fdev->irq < 0) { | 
|  | dev_err(&pdev->dev, "Failed to get irq resource\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ret = devm_request_irq(&pdev->dev, fdev->irq, st_fdma_irq_handler, 0, | 
|  | dev_name(&pdev->dev), fdev); | 
|  | if (ret) { | 
|  | dev_err(&pdev->dev, "Failed to request irq (%d)\n", ret); | 
|  | goto err; | 
|  | } | 
|  |  | 
|  | fdev->slim_rproc = st_slim_rproc_alloc(pdev, fdev->fw_name); | 
|  | if (IS_ERR(fdev->slim_rproc)) { | 
|  | ret = PTR_ERR(fdev->slim_rproc); | 
|  | dev_err(&pdev->dev, "slim_rproc_alloc failed (%d)\n", ret); | 
|  | goto err; | 
|  | } | 
|  |  | 
|  | /* Initialise list of FDMA channels */ | 
|  | INIT_LIST_HEAD(&fdev->dma_device.channels); | 
|  | for (i = 0; i < fdev->nr_channels; i++) { | 
|  | struct st_fdma_chan *fchan = &fdev->chans[i]; | 
|  |  | 
|  | fchan->fdev = fdev; | 
|  | fchan->vchan.desc_free = st_fdma_free_desc; | 
|  | vchan_init(&fchan->vchan, &fdev->dma_device); | 
|  | } | 
|  |  | 
|  | /* Initialise the FDMA dreq (reserve 0 & 31 for FDMA use) */ | 
|  | fdev->dreq_mask = BIT(0) | BIT(31); | 
|  |  | 
|  | dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask); | 
|  | dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask); | 
|  | dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask); | 
|  |  | 
|  | fdev->dma_device.dev = &pdev->dev; | 
|  | fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res; | 
|  | fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res; | 
|  | fdev->dma_device.device_prep_dma_cyclic	= st_fdma_prep_dma_cyclic; | 
|  | fdev->dma_device.device_prep_slave_sg = st_fdma_prep_slave_sg; | 
|  | fdev->dma_device.device_prep_dma_memcpy = st_fdma_prep_dma_memcpy; | 
|  | fdev->dma_device.device_tx_status = st_fdma_tx_status; | 
|  | fdev->dma_device.device_issue_pending = st_fdma_issue_pending; | 
|  | fdev->dma_device.device_terminate_all = st_fdma_terminate_all; | 
|  | fdev->dma_device.device_config = st_fdma_slave_config; | 
|  | fdev->dma_device.device_pause = st_fdma_pause; | 
|  | fdev->dma_device.device_resume = st_fdma_resume; | 
|  |  | 
|  | fdev->dma_device.src_addr_widths = FDMA_DMA_BUSWIDTHS; | 
|  | fdev->dma_device.dst_addr_widths = FDMA_DMA_BUSWIDTHS; | 
|  | fdev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); | 
|  | fdev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST; | 
|  |  | 
|  | ret = dmaenginem_async_device_register(&fdev->dma_device); | 
|  | if (ret) { | 
|  | dev_err(&pdev->dev, | 
|  | "Failed to register DMA device (%d)\n", ret); | 
|  | goto err_rproc; | 
|  | } | 
|  |  | 
|  | ret = of_dma_controller_register(np, st_fdma_of_xlate, fdev); | 
|  | if (ret) { | 
|  | dev_err(&pdev->dev, | 
|  | "Failed to register controller (%d)\n", ret); | 
|  | goto err_rproc; | 
|  | } | 
|  |  | 
|  | dev_info(&pdev->dev, "ST FDMA engine driver, irq:%d\n", fdev->irq); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | err_rproc: | 
|  | st_fdma_free(fdev); | 
|  | st_slim_rproc_put(fdev->slim_rproc); | 
|  | err: | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int st_fdma_remove(struct platform_device *pdev) | 
|  | { | 
|  | struct st_fdma_dev *fdev = platform_get_drvdata(pdev); | 
|  |  | 
|  | devm_free_irq(&pdev->dev, fdev->irq, fdev); | 
|  | st_slim_rproc_put(fdev->slim_rproc); | 
|  | of_dma_controller_free(pdev->dev.of_node); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static struct platform_driver st_fdma_platform_driver = { | 
|  | .driver = { | 
|  | .name = DRIVER_NAME, | 
|  | .of_match_table = st_fdma_match, | 
|  | }, | 
|  | .probe = st_fdma_probe, | 
|  | .remove = st_fdma_remove, | 
|  | }; | 
|  | module_platform_driver(st_fdma_platform_driver); | 
|  |  | 
|  | MODULE_LICENSE("GPL v2"); | 
|  | MODULE_DESCRIPTION("STMicroelectronics FDMA engine driver"); | 
|  | MODULE_AUTHOR("Ludovic.barre <Ludovic.barre@st.com>"); | 
|  | MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>"); | 
|  | MODULE_ALIAS("platform: " DRIVER_NAME); |