2010-05-24 11:28:19 +08:00
|
|
|
/* linux/drivers/dma/pl330.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 2010 Samsung Electronics Co. Ltd.
|
|
|
|
* Jaswinder Singh <jassi.brar@samsung.com>
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License as published by
|
|
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
|
|
* (at your option) any later version.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/io.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/dmaengine.h>
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/amba/bus.h>
|
|
|
|
#include <linux/amba/pl330.h>
|
2011-09-02 08:44:29 +08:00
|
|
|
#include <linux/pm_runtime.h>
|
2011-09-02 08:44:30 +08:00
|
|
|
#include <linux/scatterlist.h>
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
#define NR_DEFAULT_DESC 16
|
|
|
|
|
|
|
|
enum desc_status {
|
|
|
|
/* In the DMAC pool */
|
|
|
|
FREE,
|
|
|
|
/*
|
|
|
|
* Allocted to some channel during prep_xxx
|
|
|
|
* Also may be sitting on the work_list.
|
|
|
|
*/
|
|
|
|
PREP,
|
|
|
|
/*
|
|
|
|
* Sitting on the work_list and already submitted
|
|
|
|
* to the PL330 core. Not more than two descriptors
|
|
|
|
* of a channel can be BUSY at any time.
|
|
|
|
*/
|
|
|
|
BUSY,
|
|
|
|
/*
|
|
|
|
* Sitting on the channel work_list but xfer done
|
|
|
|
* by PL330 core
|
|
|
|
*/
|
|
|
|
DONE,
|
|
|
|
};
|
|
|
|
|
|
|
|
struct dma_pl330_chan {
|
|
|
|
/* Schedule desc completion */
|
|
|
|
struct tasklet_struct task;
|
|
|
|
|
|
|
|
/* DMA-Engine Channel */
|
|
|
|
struct dma_chan chan;
|
|
|
|
|
|
|
|
/* Last completed cookie */
|
|
|
|
dma_cookie_t completed;
|
|
|
|
|
|
|
|
/* List of to be xfered descriptors */
|
|
|
|
struct list_head work_list;
|
|
|
|
|
|
|
|
/* Pointer to the DMAC that manages this channel,
|
|
|
|
* NULL if the channel is available to be acquired.
|
|
|
|
* As the parent, this DMAC also provides descriptors
|
|
|
|
* to the channel.
|
|
|
|
*/
|
|
|
|
struct dma_pl330_dmac *dmac;
|
|
|
|
|
|
|
|
/* To protect channel manipulation */
|
|
|
|
spinlock_t lock;
|
|
|
|
|
|
|
|
/* Token of a hardware channel thread of PL330 DMAC
|
|
|
|
* NULL if the channel is available to be acquired.
|
|
|
|
*/
|
|
|
|
void *pl330_chid;
|
2011-09-02 08:44:30 +08:00
|
|
|
|
|
|
|
/* For D-to-M and M-to-D channels */
|
|
|
|
int burst_sz; /* the peripheral fifo width */
|
2011-09-02 08:44:31 +08:00
|
|
|
int burst_len; /* the number of burst */
|
2011-09-02 08:44:30 +08:00
|
|
|
dma_addr_t fifo_addr;
|
2011-09-02 08:44:33 +08:00
|
|
|
|
|
|
|
/* for cyclic capability */
|
|
|
|
bool cyclic;
|
2010-05-24 11:28:19 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
struct dma_pl330_dmac {
|
|
|
|
struct pl330_info pif;
|
|
|
|
|
|
|
|
/* DMA-Engine Device */
|
|
|
|
struct dma_device ddma;
|
|
|
|
|
|
|
|
/* Pool of descriptors available for the DMAC's channels */
|
|
|
|
struct list_head desc_pool;
|
|
|
|
/* To protect desc_pool manipulation */
|
|
|
|
spinlock_t pool_lock;
|
|
|
|
|
|
|
|
/* Peripheral channels connected to this DMAC */
|
2011-07-26 05:05:04 +08:00
|
|
|
struct dma_pl330_chan *peripherals; /* keep at end */
|
2011-09-02 08:44:29 +08:00
|
|
|
|
|
|
|
struct clk *clk;
|
2010-05-24 11:28:19 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
struct dma_pl330_desc {
|
|
|
|
/* To attach to a queue as child */
|
|
|
|
struct list_head node;
|
|
|
|
|
|
|
|
/* Descriptor for the DMA Engine API */
|
|
|
|
struct dma_async_tx_descriptor txd;
|
|
|
|
|
|
|
|
/* Xfer for PL330 core */
|
|
|
|
struct pl330_xfer px;
|
|
|
|
|
|
|
|
struct pl330_reqcfg rqcfg;
|
|
|
|
struct pl330_req req;
|
|
|
|
|
|
|
|
enum desc_status status;
|
|
|
|
|
|
|
|
/* The channel which currently holds this desc */
|
|
|
|
struct dma_pl330_chan *pchan;
|
|
|
|
};
|
|
|
|
|
|
|
|
static inline struct dma_pl330_chan *
|
|
|
|
to_pchan(struct dma_chan *ch)
|
|
|
|
{
|
|
|
|
if (!ch)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return container_of(ch, struct dma_pl330_chan, chan);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct dma_pl330_desc *
|
|
|
|
to_desc(struct dma_async_tx_descriptor *tx)
|
|
|
|
{
|
|
|
|
return container_of(tx, struct dma_pl330_desc, txd);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void free_desc_list(struct list_head *list)
|
|
|
|
{
|
|
|
|
struct dma_pl330_dmac *pdmac;
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
struct dma_pl330_chan *pch;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
if (list_empty(list))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* Finish off the work list */
|
|
|
|
list_for_each_entry(desc, list, node) {
|
|
|
|
dma_async_tx_callback callback;
|
|
|
|
void *param;
|
|
|
|
|
|
|
|
/* All desc in a list belong to same channel */
|
|
|
|
pch = desc->pchan;
|
|
|
|
callback = desc->txd.callback;
|
|
|
|
param = desc->txd.callback_param;
|
|
|
|
|
|
|
|
if (callback)
|
|
|
|
callback(param);
|
|
|
|
|
|
|
|
desc->pchan = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
pdmac = pch->dmac;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pdmac->pool_lock, flags);
|
|
|
|
list_splice_tail_init(list, &pdmac->desc_pool);
|
|
|
|
spin_unlock_irqrestore(&pdmac->pool_lock, flags);
|
|
|
|
}
|
|
|
|
|
2011-09-02 08:44:33 +08:00
|
|
|
static inline void handle_cyclic_desc_list(struct list_head *list)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
struct dma_pl330_chan *pch;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
if (list_empty(list))
|
|
|
|
return;
|
|
|
|
|
|
|
|
list_for_each_entry(desc, list, node) {
|
|
|
|
dma_async_tx_callback callback;
|
|
|
|
|
|
|
|
/* Change status to reload it */
|
|
|
|
desc->status = PREP;
|
|
|
|
pch = desc->pchan;
|
|
|
|
callback = desc->txd.callback;
|
|
|
|
if (callback)
|
|
|
|
callback(desc->txd.callback_param);
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
|
|
|
list_splice_tail_init(list, &pch->work_list);
|
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
}
|
|
|
|
|
2010-05-24 11:28:19 +08:00
|
|
|
static inline void fill_queue(struct dma_pl330_chan *pch)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
list_for_each_entry(desc, &pch->work_list, node) {
|
|
|
|
|
|
|
|
/* If already submitted */
|
|
|
|
if (desc->status == BUSY)
|
|
|
|
break;
|
|
|
|
|
|
|
|
ret = pl330_submit_req(pch->pl330_chid,
|
|
|
|
&desc->req);
|
|
|
|
if (!ret) {
|
|
|
|
desc->status = BUSY;
|
|
|
|
break;
|
|
|
|
} else if (ret == -EAGAIN) {
|
|
|
|
/* QFull or DMAC Dying */
|
|
|
|
break;
|
|
|
|
} else {
|
|
|
|
/* Unacceptable request */
|
|
|
|
desc->status = DONE;
|
|
|
|
dev_err(pch->dmac->pif.dev, "%s:%d Bad Desc(%d)\n",
|
|
|
|
__func__, __LINE__, desc->txd.cookie);
|
|
|
|
tasklet_schedule(&pch->task);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void pl330_tasklet(unsigned long data)
|
|
|
|
{
|
|
|
|
struct dma_pl330_chan *pch = (struct dma_pl330_chan *)data;
|
|
|
|
struct dma_pl330_desc *desc, *_dt;
|
|
|
|
unsigned long flags;
|
|
|
|
LIST_HEAD(list);
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
|
|
|
|
|
|
|
/* Pick up ripe tomatoes */
|
|
|
|
list_for_each_entry_safe(desc, _dt, &pch->work_list, node)
|
|
|
|
if (desc->status == DONE) {
|
|
|
|
pch->completed = desc->txd.cookie;
|
|
|
|
list_move_tail(&desc->node, &list);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Try to submit a req imm. next to the last completed cookie */
|
|
|
|
fill_queue(pch);
|
|
|
|
|
|
|
|
/* Make sure the PL330 Channel thread is active */
|
|
|
|
pl330_chan_ctrl(pch->pl330_chid, PL330_OP_START);
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
|
2011-09-02 08:44:33 +08:00
|
|
|
if (pch->cyclic)
|
|
|
|
handle_cyclic_desc_list(&list);
|
|
|
|
else
|
|
|
|
free_desc_list(&list);
|
2010-05-24 11:28:19 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void dma_pl330_rqcb(void *token, enum pl330_op_err err)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc = token;
|
|
|
|
struct dma_pl330_chan *pch = desc->pchan;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
/* If desc aborted */
|
|
|
|
if (!pch)
|
|
|
|
return;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
|
|
|
|
|
|
|
desc->status = DONE;
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
|
|
|
|
tasklet_schedule(&pch->task);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pl330_alloc_chan_resources(struct dma_chan *chan)
|
|
|
|
{
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
|
|
|
struct dma_pl330_dmac *pdmac = pch->dmac;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
|
|
|
|
|
|
|
pch->completed = chan->cookie = 1;
|
2011-09-02 08:44:33 +08:00
|
|
|
pch->cyclic = false;
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
pch->pl330_chid = pl330_request_channel(&pdmac->pif);
|
|
|
|
if (!pch->pl330_chid) {
|
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
tasklet_init(&pch->task, pl330_tasklet, (unsigned long) pch);
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pl330_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, unsigned long arg)
|
|
|
|
{
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
2011-09-02 08:44:32 +08:00
|
|
|
struct dma_pl330_desc *desc, *_dt;
|
2010-05-24 11:28:19 +08:00
|
|
|
unsigned long flags;
|
2011-09-02 08:44:31 +08:00
|
|
|
struct dma_pl330_dmac *pdmac = pch->dmac;
|
|
|
|
struct dma_slave_config *slave_config;
|
2011-09-02 08:44:32 +08:00
|
|
|
LIST_HEAD(list);
|
2010-05-24 11:28:19 +08:00
|
|
|
|
2011-09-02 08:44:31 +08:00
|
|
|
switch (cmd) {
|
|
|
|
case DMA_TERMINATE_ALL:
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
2010-05-24 11:28:19 +08:00
|
|
|
|
2011-09-02 08:44:31 +08:00
|
|
|
/* FLUSH the PL330 Channel thread */
|
|
|
|
pl330_chan_ctrl(pch->pl330_chid, PL330_OP_FLUSH);
|
2010-05-24 11:28:19 +08:00
|
|
|
|
2011-09-02 08:44:31 +08:00
|
|
|
/* Mark all desc done */
|
2011-09-02 08:44:32 +08:00
|
|
|
list_for_each_entry_safe(desc, _dt, &pch->work_list , node) {
|
2011-09-02 08:44:31 +08:00
|
|
|
desc->status = DONE;
|
2011-09-02 08:44:32 +08:00
|
|
|
pch->completed = desc->txd.cookie;
|
|
|
|
list_move_tail(&desc->node, &list);
|
|
|
|
}
|
2010-05-24 11:28:19 +08:00
|
|
|
|
2011-09-02 08:44:32 +08:00
|
|
|
list_splice_tail_init(&list, &pdmac->desc_pool);
|
2011-09-02 08:44:31 +08:00
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
break;
|
|
|
|
case DMA_SLAVE_CONFIG:
|
|
|
|
slave_config = (struct dma_slave_config *)arg;
|
|
|
|
|
2011-10-14 01:04:23 +08:00
|
|
|
if (slave_config->direction == DMA_MEM_TO_DEV) {
|
2011-09-02 08:44:31 +08:00
|
|
|
if (slave_config->dst_addr)
|
|
|
|
pch->fifo_addr = slave_config->dst_addr;
|
|
|
|
if (slave_config->dst_addr_width)
|
|
|
|
pch->burst_sz = __ffs(slave_config->dst_addr_width);
|
|
|
|
if (slave_config->dst_maxburst)
|
|
|
|
pch->burst_len = slave_config->dst_maxburst;
|
2011-10-14 01:04:23 +08:00
|
|
|
} else if (slave_config->direction == DMA_DEV_TO_MEM) {
|
2011-09-02 08:44:31 +08:00
|
|
|
if (slave_config->src_addr)
|
|
|
|
pch->fifo_addr = slave_config->src_addr;
|
|
|
|
if (slave_config->src_addr_width)
|
|
|
|
pch->burst_sz = __ffs(slave_config->src_addr_width);
|
|
|
|
if (slave_config->src_maxburst)
|
|
|
|
pch->burst_len = slave_config->src_maxburst;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_err(pch->dmac->pif.dev, "Not supported command.\n");
|
|
|
|
return -ENXIO;
|
|
|
|
}
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void pl330_free_chan_resources(struct dma_chan *chan)
|
|
|
|
{
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
|
|
|
|
|
|
|
tasklet_kill(&pch->task);
|
|
|
|
|
|
|
|
pl330_release_channel(pch->pl330_chid);
|
|
|
|
pch->pl330_chid = NULL;
|
|
|
|
|
2011-09-02 08:44:33 +08:00
|
|
|
if (pch->cyclic)
|
|
|
|
list_splice_tail_init(&pch->work_list, &pch->dmac->desc_pool);
|
|
|
|
|
2010-05-24 11:28:19 +08:00
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum dma_status
|
|
|
|
pl330_tx_status(struct dma_chan *chan, dma_cookie_t cookie,
|
|
|
|
struct dma_tx_state *txstate)
|
|
|
|
{
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
|
|
|
dma_cookie_t last_done, last_used;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
last_done = pch->completed;
|
|
|
|
last_used = chan->cookie;
|
|
|
|
|
|
|
|
ret = dma_async_is_complete(cookie, last_done, last_used);
|
|
|
|
|
|
|
|
dma_set_tx_state(txstate, last_done, last_used, 0);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void pl330_issue_pending(struct dma_chan *chan)
|
|
|
|
{
|
|
|
|
pl330_tasklet((unsigned long) to_pchan(chan));
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We returned the last one of the circular list of descriptor(s)
|
|
|
|
* from prep_xxx, so the argument to submit corresponds to the last
|
|
|
|
* descriptor of the list.
|
|
|
|
*/
|
|
|
|
static dma_cookie_t pl330_tx_submit(struct dma_async_tx_descriptor *tx)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc, *last = to_desc(tx);
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(tx->chan);
|
|
|
|
dma_cookie_t cookie;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pch->lock, flags);
|
|
|
|
|
|
|
|
/* Assign cookies to all nodes */
|
|
|
|
cookie = tx->chan->cookie;
|
|
|
|
|
|
|
|
while (!list_empty(&last->node)) {
|
|
|
|
desc = list_entry(last->node.next, struct dma_pl330_desc, node);
|
|
|
|
|
|
|
|
if (++cookie < 0)
|
|
|
|
cookie = 1;
|
|
|
|
desc->txd.cookie = cookie;
|
|
|
|
|
|
|
|
list_move_tail(&desc->node, &pch->work_list);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (++cookie < 0)
|
|
|
|
cookie = 1;
|
|
|
|
last->txd.cookie = cookie;
|
|
|
|
|
|
|
|
list_add_tail(&last->node, &pch->work_list);
|
|
|
|
|
|
|
|
tx->chan->cookie = cookie;
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pch->lock, flags);
|
|
|
|
|
|
|
|
return cookie;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void _init_desc(struct dma_pl330_desc *desc)
|
|
|
|
{
|
|
|
|
desc->pchan = NULL;
|
|
|
|
desc->req.x = &desc->px;
|
|
|
|
desc->req.token = desc;
|
|
|
|
desc->rqcfg.swap = SWAP_NO;
|
|
|
|
desc->rqcfg.privileged = 0;
|
|
|
|
desc->rqcfg.insnaccess = 0;
|
|
|
|
desc->rqcfg.scctl = SCCTRL0;
|
|
|
|
desc->rqcfg.dcctl = DCCTRL0;
|
|
|
|
desc->req.cfg = &desc->rqcfg;
|
|
|
|
desc->req.xfer_cb = dma_pl330_rqcb;
|
|
|
|
desc->txd.tx_submit = pl330_tx_submit;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&desc->node);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Returns the number of descriptors added to the DMAC pool */
|
|
|
|
int add_desc(struct dma_pl330_dmac *pdmac, gfp_t flg, int count)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
unsigned long flags;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (!pdmac)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
desc = kmalloc(count * sizeof(*desc), flg);
|
|
|
|
if (!desc)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pdmac->pool_lock, flags);
|
|
|
|
|
|
|
|
for (i = 0; i < count; i++) {
|
|
|
|
_init_desc(&desc[i]);
|
|
|
|
list_add_tail(&desc[i].node, &pdmac->desc_pool);
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pdmac->pool_lock, flags);
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct dma_pl330_desc *
|
|
|
|
pluck_desc(struct dma_pl330_dmac *pdmac)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc = NULL;
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
if (!pdmac)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pdmac->pool_lock, flags);
|
|
|
|
|
|
|
|
if (!list_empty(&pdmac->desc_pool)) {
|
|
|
|
desc = list_entry(pdmac->desc_pool.next,
|
|
|
|
struct dma_pl330_desc, node);
|
|
|
|
|
|
|
|
list_del_init(&desc->node);
|
|
|
|
|
|
|
|
desc->status = PREP;
|
|
|
|
desc->txd.callback = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pdmac->pool_lock, flags);
|
|
|
|
|
|
|
|
return desc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct dma_pl330_desc *pl330_get_desc(struct dma_pl330_chan *pch)
|
|
|
|
{
|
|
|
|
struct dma_pl330_dmac *pdmac = pch->dmac;
|
|
|
|
struct dma_pl330_peri *peri = pch->chan.private;
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
|
|
|
|
/* Pluck one desc from the pool of DMAC */
|
|
|
|
desc = pluck_desc(pdmac);
|
|
|
|
|
|
|
|
/* If the DMAC pool is empty, alloc new */
|
|
|
|
if (!desc) {
|
|
|
|
if (!add_desc(pdmac, GFP_ATOMIC, 1))
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* Try again */
|
|
|
|
desc = pluck_desc(pdmac);
|
|
|
|
if (!desc) {
|
|
|
|
dev_err(pch->dmac->pif.dev,
|
|
|
|
"%s:%d ALERT!\n", __func__, __LINE__);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Initialize the descriptor */
|
|
|
|
desc->pchan = pch;
|
|
|
|
desc->txd.cookie = 0;
|
|
|
|
async_tx_ack(&desc->txd);
|
|
|
|
|
2011-07-26 05:05:04 +08:00
|
|
|
if (peri) {
|
|
|
|
desc->req.rqtype = peri->rqtype;
|
2011-09-02 08:44:30 +08:00
|
|
|
desc->req.peri = pch->chan.chan_id;
|
2011-07-26 05:05:04 +08:00
|
|
|
} else {
|
|
|
|
desc->req.rqtype = MEMTOMEM;
|
|
|
|
desc->req.peri = 0;
|
|
|
|
}
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
dma_async_tx_descriptor_init(&desc->txd, &pch->chan);
|
|
|
|
|
|
|
|
return desc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void fill_px(struct pl330_xfer *px,
|
|
|
|
dma_addr_t dst, dma_addr_t src, size_t len)
|
|
|
|
{
|
|
|
|
px->next = NULL;
|
|
|
|
px->bytes = len;
|
|
|
|
px->dst_addr = dst;
|
|
|
|
px->src_addr = src;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct dma_pl330_desc *
|
|
|
|
__pl330_prep_dma_memcpy(struct dma_pl330_chan *pch, dma_addr_t dst,
|
|
|
|
dma_addr_t src, size_t len)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc = pl330_get_desc(pch);
|
|
|
|
|
|
|
|
if (!desc) {
|
|
|
|
dev_err(pch->dmac->pif.dev, "%s:%d Unable to fetch desc\n",
|
|
|
|
__func__, __LINE__);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ideally we should lookout for reqs bigger than
|
|
|
|
* those that can be programmed with 256 bytes of
|
|
|
|
* MC buffer, but considering a req size is seldom
|
|
|
|
* going to be word-unaligned and more than 200MB,
|
|
|
|
* we take it easy.
|
|
|
|
* Also, should the limit is reached we'd rather
|
|
|
|
* have the platform increase MC buffer size than
|
|
|
|
* complicating this API driver.
|
|
|
|
*/
|
|
|
|
fill_px(&desc->px, dst, src, len);
|
|
|
|
|
|
|
|
return desc;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Call after fixing burst size */
|
|
|
|
static inline int get_burst_len(struct dma_pl330_desc *desc, size_t len)
|
|
|
|
{
|
|
|
|
struct dma_pl330_chan *pch = desc->pchan;
|
|
|
|
struct pl330_info *pi = &pch->dmac->pif;
|
|
|
|
int burst_len;
|
|
|
|
|
|
|
|
burst_len = pi->pcfg.data_bus_width / 8;
|
|
|
|
burst_len *= pi->pcfg.data_buf_dep;
|
|
|
|
burst_len >>= desc->rqcfg.brst_size;
|
|
|
|
|
|
|
|
/* src/dst_burst_len can't be more than 16 */
|
|
|
|
if (burst_len > 16)
|
|
|
|
burst_len = 16;
|
|
|
|
|
|
|
|
while (burst_len > 1) {
|
|
|
|
if (!(len % (burst_len << desc->rqcfg.brst_size)))
|
|
|
|
break;
|
|
|
|
burst_len--;
|
|
|
|
}
|
|
|
|
|
|
|
|
return burst_len;
|
|
|
|
}
|
|
|
|
|
2011-09-02 08:44:33 +08:00
|
|
|
static struct dma_async_tx_descriptor *pl330_prep_dma_cyclic(
|
|
|
|
struct dma_chan *chan, dma_addr_t dma_addr, size_t len,
|
2011-10-14 01:04:23 +08:00
|
|
|
size_t period_len, enum dma_transfer_direction direction)
|
2011-09-02 08:44:33 +08:00
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
|
|
|
dma_addr_t dst;
|
|
|
|
dma_addr_t src;
|
|
|
|
|
|
|
|
desc = pl330_get_desc(pch);
|
|
|
|
if (!desc) {
|
|
|
|
dev_err(pch->dmac->pif.dev, "%s:%d Unable to fetch desc\n",
|
|
|
|
__func__, __LINE__);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (direction) {
|
2011-10-14 01:04:23 +08:00
|
|
|
case DMA_MEM_TO_DEV:
|
2011-09-02 08:44:33 +08:00
|
|
|
desc->rqcfg.src_inc = 1;
|
|
|
|
desc->rqcfg.dst_inc = 0;
|
|
|
|
src = dma_addr;
|
|
|
|
dst = pch->fifo_addr;
|
|
|
|
break;
|
2011-10-14 01:04:23 +08:00
|
|
|
case DMA_DEV_TO_MEM:
|
2011-09-02 08:44:33 +08:00
|
|
|
desc->rqcfg.src_inc = 0;
|
|
|
|
desc->rqcfg.dst_inc = 1;
|
|
|
|
src = pch->fifo_addr;
|
|
|
|
dst = dma_addr;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_err(pch->dmac->pif.dev, "%s:%d Invalid dma direction\n",
|
|
|
|
__func__, __LINE__);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
desc->rqcfg.brst_size = pch->burst_sz;
|
|
|
|
desc->rqcfg.brst_len = 1;
|
|
|
|
|
|
|
|
pch->cyclic = true;
|
|
|
|
|
|
|
|
fill_px(&desc->px, dst, src, period_len);
|
|
|
|
|
|
|
|
return &desc->txd;
|
|
|
|
}
|
|
|
|
|
2010-05-24 11:28:19 +08:00
|
|
|
static struct dma_async_tx_descriptor *
|
|
|
|
pl330_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dst,
|
|
|
|
dma_addr_t src, size_t len, unsigned long flags)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *desc;
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
|
|
|
struct dma_pl330_peri *peri = chan->private;
|
|
|
|
struct pl330_info *pi;
|
|
|
|
int burst;
|
|
|
|
|
2011-07-26 05:05:04 +08:00
|
|
|
if (unlikely(!pch || !len))
|
2010-05-24 11:28:19 +08:00
|
|
|
return NULL;
|
|
|
|
|
2011-07-26 05:05:04 +08:00
|
|
|
if (peri && peri->rqtype != MEMTOMEM)
|
2010-05-24 11:28:19 +08:00
|
|
|
return NULL;
|
|
|
|
|
|
|
|
pi = &pch->dmac->pif;
|
|
|
|
|
|
|
|
desc = __pl330_prep_dma_memcpy(pch, dst, src, len);
|
|
|
|
if (!desc)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
desc->rqcfg.src_inc = 1;
|
|
|
|
desc->rqcfg.dst_inc = 1;
|
|
|
|
|
|
|
|
/* Select max possible burst size */
|
|
|
|
burst = pi->pcfg.data_bus_width / 8;
|
|
|
|
|
|
|
|
while (burst > 1) {
|
|
|
|
if (!(len % burst))
|
|
|
|
break;
|
|
|
|
burst /= 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
desc->rqcfg.brst_size = 0;
|
|
|
|
while (burst != (1 << desc->rqcfg.brst_size))
|
|
|
|
desc->rqcfg.brst_size++;
|
|
|
|
|
|
|
|
desc->rqcfg.brst_len = get_burst_len(desc, len);
|
|
|
|
|
|
|
|
desc->txd.flags = flags;
|
|
|
|
|
|
|
|
return &desc->txd;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct dma_async_tx_descriptor *
|
|
|
|
pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
|
2011-10-14 01:04:23 +08:00
|
|
|
unsigned int sg_len, enum dma_transfer_direction direction,
|
2010-05-24 11:28:19 +08:00
|
|
|
unsigned long flg)
|
|
|
|
{
|
|
|
|
struct dma_pl330_desc *first, *desc = NULL;
|
|
|
|
struct dma_pl330_chan *pch = to_pchan(chan);
|
|
|
|
struct dma_pl330_peri *peri = chan->private;
|
|
|
|
struct scatterlist *sg;
|
|
|
|
unsigned long flags;
|
2011-09-02 08:44:30 +08:00
|
|
|
int i;
|
2010-05-24 11:28:19 +08:00
|
|
|
dma_addr_t addr;
|
|
|
|
|
2011-07-26 05:05:04 +08:00
|
|
|
if (unlikely(!pch || !sgl || !sg_len || !peri))
|
2010-05-24 11:28:19 +08:00
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* Make sure the direction is consistent */
|
2011-10-14 01:04:23 +08:00
|
|
|
if ((direction == DMA_MEM_TO_DEV &&
|
2010-05-24 11:28:19 +08:00
|
|
|
peri->rqtype != MEMTODEV) ||
|
2011-10-14 01:04:23 +08:00
|
|
|
(direction == DMA_DEV_TO_MEM &&
|
2010-05-24 11:28:19 +08:00
|
|
|
peri->rqtype != DEVTOMEM)) {
|
|
|
|
dev_err(pch->dmac->pif.dev, "%s:%d Invalid Direction\n",
|
|
|
|
__func__, __LINE__);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2011-09-02 08:44:30 +08:00
|
|
|
addr = pch->fifo_addr;
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
first = NULL;
|
|
|
|
|
|
|
|
for_each_sg(sgl, sg, sg_len, i) {
|
|
|
|
|
|
|
|
desc = pl330_get_desc(pch);
|
|
|
|
if (!desc) {
|
|
|
|
struct dma_pl330_dmac *pdmac = pch->dmac;
|
|
|
|
|
|
|
|
dev_err(pch->dmac->pif.dev,
|
|
|
|
"%s:%d Unable to fetch desc\n",
|
|
|
|
__func__, __LINE__);
|
|
|
|
if (!first)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&pdmac->pool_lock, flags);
|
|
|
|
|
|
|
|
while (!list_empty(&first->node)) {
|
|
|
|
desc = list_entry(first->node.next,
|
|
|
|
struct dma_pl330_desc, node);
|
|
|
|
list_move_tail(&desc->node, &pdmac->desc_pool);
|
|
|
|
}
|
|
|
|
|
|
|
|
list_move_tail(&first->node, &pdmac->desc_pool);
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&pdmac->pool_lock, flags);
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!first)
|
|
|
|
first = desc;
|
|
|
|
else
|
|
|
|
list_add_tail(&desc->node, &first->node);
|
|
|
|
|
2011-10-14 01:04:23 +08:00
|
|
|
if (direction == DMA_MEM_TO_DEV) {
|
2010-05-24 11:28:19 +08:00
|
|
|
desc->rqcfg.src_inc = 1;
|
|
|
|
desc->rqcfg.dst_inc = 0;
|
|
|
|
fill_px(&desc->px,
|
|
|
|
addr, sg_dma_address(sg), sg_dma_len(sg));
|
|
|
|
} else {
|
|
|
|
desc->rqcfg.src_inc = 0;
|
|
|
|
desc->rqcfg.dst_inc = 1;
|
|
|
|
fill_px(&desc->px,
|
|
|
|
sg_dma_address(sg), addr, sg_dma_len(sg));
|
|
|
|
}
|
|
|
|
|
2011-09-02 08:44:30 +08:00
|
|
|
desc->rqcfg.brst_size = pch->burst_sz;
|
2010-05-24 11:28:19 +08:00
|
|
|
desc->rqcfg.brst_len = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Return the last desc in the chain */
|
|
|
|
desc->txd.flags = flg;
|
|
|
|
return &desc->txd;
|
|
|
|
}
|
|
|
|
|
|
|
|
static irqreturn_t pl330_irq_handler(int irq, void *data)
|
|
|
|
{
|
|
|
|
if (pl330_update(data))
|
|
|
|
return IRQ_HANDLED;
|
|
|
|
else
|
|
|
|
return IRQ_NONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __devinit
|
2011-02-19 23:55:00 +08:00
|
|
|
pl330_probe(struct amba_device *adev, const struct amba_id *id)
|
2010-05-24 11:28:19 +08:00
|
|
|
{
|
|
|
|
struct dma_pl330_platdata *pdat;
|
|
|
|
struct dma_pl330_dmac *pdmac;
|
|
|
|
struct dma_pl330_chan *pch;
|
|
|
|
struct pl330_info *pi;
|
|
|
|
struct dma_device *pd;
|
|
|
|
struct resource *res;
|
|
|
|
int i, ret, irq;
|
2011-07-26 05:05:04 +08:00
|
|
|
int num_chan;
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
pdat = adev->dev.platform_data;
|
|
|
|
|
|
|
|
/* Allocate a new DMAC and its Channels */
|
2011-07-26 05:05:04 +08:00
|
|
|
pdmac = kzalloc(sizeof(*pdmac), GFP_KERNEL);
|
2010-05-24 11:28:19 +08:00
|
|
|
if (!pdmac) {
|
|
|
|
dev_err(&adev->dev, "unable to allocate mem\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
pi = &pdmac->pif;
|
|
|
|
pi->dev = &adev->dev;
|
|
|
|
pi->pl330_data = NULL;
|
2011-07-26 05:05:04 +08:00
|
|
|
pi->mcbufsz = pdat ? pdat->mcbuf_sz : 0;
|
2010-05-24 11:28:19 +08:00
|
|
|
|
|
|
|
res = &adev->res;
|
|
|
|
request_mem_region(res->start, resource_size(res), "dma-pl330");
|
|
|
|
|
|
|
|
pi->base = ioremap(res->start, resource_size(res));
|
|
|
|
if (!pi->base) {
|
|
|
|
ret = -ENXIO;
|
|
|
|
goto probe_err1;
|
|
|
|
}
|
|
|
|
|
2011-09-02 08:44:29 +08:00
|
|
|
pdmac->clk = clk_get(&adev->dev, "dma");
|
|
|
|
if (IS_ERR(pdmac->clk)) {
|
|
|
|
dev_err(&adev->dev, "Cannot get operation clock.\n");
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto probe_err1;
|
|
|
|
}
|
|
|
|
|
|
|
|
amba_set_drvdata(adev, pdmac);
|
|
|
|
|
|
|
|
#ifdef CONFIG_PM_RUNTIME
|
|
|
|
/* to use the runtime PM helper functions */
|
|
|
|
pm_runtime_enable(&adev->dev);
|
|
|
|
|
|
|
|
/* enable the power domain */
|
|
|
|
if (pm_runtime_get_sync(&adev->dev)) {
|
|
|
|
dev_err(&adev->dev, "failed to get runtime pm\n");
|
|
|
|
ret = -ENODEV;
|
|
|
|
goto probe_err1;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
/* enable dma clk */
|
|
|
|
clk_enable(pdmac->clk);
|
|
|
|
#endif
|
|
|
|
|
2010-05-24 11:28:19 +08:00
|
|
|
irq = adev->irq[0];
|
|
|
|
ret = request_irq(irq, pl330_irq_handler, 0,
|
|
|
|
dev_name(&adev->dev), pi);
|
|
|
|
if (ret)
|
|
|
|
goto probe_err2;
|
|
|
|
|
|
|
|
ret = pl330_add(pi);
|
|
|
|
if (ret)
|
|
|
|
goto probe_err3;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&pdmac->desc_pool);
|
|
|
|
spin_lock_init(&pdmac->pool_lock);
|
|
|
|
|
|
|
|
/* Create a descriptor pool of default size */
|
|
|
|
if (!add_desc(pdmac, GFP_KERNEL, NR_DEFAULT_DESC))
|
|
|
|
dev_warn(&adev->dev, "unable to allocate desc\n");
|
|
|
|
|
|
|
|
pd = &pdmac->ddma;
|
|
|
|
INIT_LIST_HEAD(&pd->channels);
|
|
|
|
|
|
|
|
/* Initialize channel parameters */
|
2011-11-03 14:48:49 +08:00
|
|
|
num_chan = max(pdat ? (int)pdat->nr_valid_peri : 0,
|
|
|
|
(int)pi->pcfg.num_chan);
|
2011-07-26 05:05:04 +08:00
|
|
|
pdmac->peripherals = kzalloc(num_chan * sizeof(*pch), GFP_KERNEL);
|
2010-05-24 11:28:19 +08:00
|
|
|
|
2011-07-26 05:05:04 +08:00
|
|
|
for (i = 0; i < num_chan; i++) {
|
|
|
|
pch = &pdmac->peripherals[i];
|
|
|
|
if (pdat) {
|
|
|
|
struct dma_pl330_peri *peri = &pdat->peri[i];
|
|
|
|
|
|
|
|
switch (peri->rqtype) {
|
|
|
|
case MEMTOMEM:
|
|
|
|
dma_cap_set(DMA_MEMCPY, pd->cap_mask);
|
|
|
|
break;
|
|
|
|
case MEMTODEV:
|
|
|
|
case DEVTOMEM:
|
|
|
|
dma_cap_set(DMA_SLAVE, pd->cap_mask);
|
2011-09-02 08:44:33 +08:00
|
|
|
dma_cap_set(DMA_CYCLIC, pd->cap_mask);
|
2011-07-26 05:05:04 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_err(&adev->dev, "DEVTODEV Not Supported\n");
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
pch->chan.private = peri;
|
|
|
|
} else {
|
2010-05-24 11:28:19 +08:00
|
|
|
dma_cap_set(DMA_MEMCPY, pd->cap_mask);
|
2011-07-26 05:05:04 +08:00
|
|
|
pch->chan.private = NULL;
|
2010-05-24 11:28:19 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&pch->work_list);
|
|
|
|
spin_lock_init(&pch->lock);
|
|
|
|
pch->pl330_chid = NULL;
|
|
|
|
pch->chan.device = pd;
|
|
|
|
pch->dmac = pdmac;
|
|
|
|
|
|
|
|
/* Add the channel to the DMAC list */
|
|
|
|
list_add_tail(&pch->chan.device_node, &pd->channels);
|
|
|
|
}
|
|
|
|
|
|
|
|
pd->dev = &adev->dev;
|
|
|
|
|
|
|
|
pd->device_alloc_chan_resources = pl330_alloc_chan_resources;
|
|
|
|
pd->device_free_chan_resources = pl330_free_chan_resources;
|
|
|
|
pd->device_prep_dma_memcpy = pl330_prep_dma_memcpy;
|
2011-09-02 08:44:33 +08:00
|
|
|
pd->device_prep_dma_cyclic = pl330_prep_dma_cyclic;
|
2010-05-24 11:28:19 +08:00
|
|
|
pd->device_tx_status = pl330_tx_status;
|
|
|
|
pd->device_prep_slave_sg = pl330_prep_slave_sg;
|
|
|
|
pd->device_control = pl330_control;
|
|
|
|
pd->device_issue_pending = pl330_issue_pending;
|
|
|
|
|
|
|
|
ret = dma_async_device_register(pd);
|
|
|
|
if (ret) {
|
|
|
|
dev_err(&adev->dev, "unable to register DMAC\n");
|
|
|
|
goto probe_err4;
|
|
|
|
}
|
|
|
|
|
|
|
|
dev_info(&adev->dev,
|
|
|
|
"Loaded driver for PL330 DMAC-%d\n", adev->periphid);
|
|
|
|
dev_info(&adev->dev,
|
|
|
|
"\tDBUFF-%ux%ubytes Num_Chans-%u Num_Peri-%u Num_Events-%u\n",
|
|
|
|
pi->pcfg.data_buf_dep,
|
|
|
|
pi->pcfg.data_bus_width / 8, pi->pcfg.num_chan,
|
|
|
|
pi->pcfg.num_peri, pi->pcfg.num_events);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
probe_err4:
|
|
|
|
pl330_del(pi);
|
|
|
|
probe_err3:
|
|
|
|
free_irq(irq, pi);
|
|
|
|
probe_err2:
|
|
|
|
iounmap(pi->base);
|
|
|
|
probe_err1:
|
|
|
|
release_mem_region(res->start, resource_size(res));
|
|
|
|
kfree(pdmac);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int __devexit pl330_remove(struct amba_device *adev)
|
|
|
|
{
|
|
|
|
struct dma_pl330_dmac *pdmac = amba_get_drvdata(adev);
|
|
|
|
struct dma_pl330_chan *pch, *_p;
|
|
|
|
struct pl330_info *pi;
|
|
|
|
struct resource *res;
|
|
|
|
int irq;
|
|
|
|
|
|
|
|
if (!pdmac)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
amba_set_drvdata(adev, NULL);
|
|
|
|
|
|
|
|
/* Idle the DMAC */
|
|
|
|
list_for_each_entry_safe(pch, _p, &pdmac->ddma.channels,
|
|
|
|
chan.device_node) {
|
|
|
|
|
|
|
|
/* Remove the channel */
|
|
|
|
list_del(&pch->chan.device_node);
|
|
|
|
|
|
|
|
/* Flush the channel */
|
|
|
|
pl330_control(&pch->chan, DMA_TERMINATE_ALL, 0);
|
|
|
|
pl330_free_chan_resources(&pch->chan);
|
|
|
|
}
|
|
|
|
|
|
|
|
pi = &pdmac->pif;
|
|
|
|
|
|
|
|
pl330_del(pi);
|
|
|
|
|
|
|
|
irq = adev->irq[0];
|
|
|
|
free_irq(irq, pi);
|
|
|
|
|
|
|
|
iounmap(pi->base);
|
|
|
|
|
|
|
|
res = &adev->res;
|
|
|
|
release_mem_region(res->start, resource_size(res));
|
|
|
|
|
2011-09-02 08:44:29 +08:00
|
|
|
#ifdef CONFIG_PM_RUNTIME
|
|
|
|
pm_runtime_put(&adev->dev);
|
|
|
|
pm_runtime_disable(&adev->dev);
|
|
|
|
#else
|
|
|
|
clk_disable(pdmac->clk);
|
|
|
|
#endif
|
|
|
|
|
2010-05-24 11:28:19 +08:00
|
|
|
kfree(pdmac);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct amba_id pl330_ids[] = {
|
|
|
|
{
|
|
|
|
.id = 0x00041330,
|
|
|
|
.mask = 0x000fffff,
|
|
|
|
},
|
|
|
|
{ 0, 0 },
|
|
|
|
};
|
|
|
|
|
2011-09-02 08:44:29 +08:00
|
|
|
#ifdef CONFIG_PM_RUNTIME
|
|
|
|
static int pl330_runtime_suspend(struct device *dev)
|
|
|
|
{
|
|
|
|
struct dma_pl330_dmac *pdmac = dev_get_drvdata(dev);
|
|
|
|
|
|
|
|
if (!pdmac) {
|
|
|
|
dev_err(dev, "failed to get dmac\n");
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
clk_disable(pdmac->clk);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pl330_runtime_resume(struct device *dev)
|
|
|
|
{
|
|
|
|
struct dma_pl330_dmac *pdmac = dev_get_drvdata(dev);
|
|
|
|
|
|
|
|
if (!pdmac) {
|
|
|
|
dev_err(dev, "failed to get dmac\n");
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
|
|
|
clk_enable(pdmac->clk);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define pl330_runtime_suspend NULL
|
|
|
|
#define pl330_runtime_resume NULL
|
|
|
|
#endif /* CONFIG_PM_RUNTIME */
|
|
|
|
|
|
|
|
static const struct dev_pm_ops pl330_pm_ops = {
|
|
|
|
.runtime_suspend = pl330_runtime_suspend,
|
|
|
|
.runtime_resume = pl330_runtime_resume,
|
|
|
|
};
|
|
|
|
|
2010-05-24 11:28:19 +08:00
|
|
|
static struct amba_driver pl330_driver = {
|
|
|
|
.drv = {
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
.name = "dma-pl330",
|
2011-09-02 08:44:29 +08:00
|
|
|
.pm = &pl330_pm_ops,
|
2010-05-24 11:28:19 +08:00
|
|
|
},
|
|
|
|
.id_table = pl330_ids,
|
|
|
|
.probe = pl330_probe,
|
|
|
|
.remove = pl330_remove,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int __init pl330_init(void)
|
|
|
|
{
|
|
|
|
return amba_driver_register(&pl330_driver);
|
|
|
|
}
|
|
|
|
module_init(pl330_init);
|
|
|
|
|
|
|
|
static void __exit pl330_exit(void)
|
|
|
|
{
|
|
|
|
amba_driver_unregister(&pl330_driver);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
module_exit(pl330_exit);
|
|
|
|
|
|
|
|
MODULE_AUTHOR("Jaswinder Singh <jassi.brar@samsung.com>");
|
|
|
|
MODULE_DESCRIPTION("API Driver for PL330 DMAC");
|
|
|
|
MODULE_LICENSE("GPL");
|