Merge remote-tracking branches 'spi/topic/blackfin', 'spi/topic/cadence', 'spi/topic/dw' and 'spi/topic/err' into spi-next
This commit is contained in:
commit
431959c079
|
@ -51,19 +51,6 @@ config INTEL_MIC_X100_DMA
|
||||||
OS and tools for MIC to use with this driver are available from
|
OS and tools for MIC to use with this driver are available from
|
||||||
<http://software.intel.com/en-us/mic-developer>.
|
<http://software.intel.com/en-us/mic-developer>.
|
||||||
|
|
||||||
config INTEL_MID_DMAC
|
|
||||||
tristate "Intel MID DMA support for Peripheral DMA controllers"
|
|
||||||
depends on PCI && X86
|
|
||||||
select DMA_ENGINE
|
|
||||||
default n
|
|
||||||
help
|
|
||||||
Enable support for the Intel(R) MID DMA engine present
|
|
||||||
in Intel MID chipsets.
|
|
||||||
|
|
||||||
Say Y here if you have such a chipset.
|
|
||||||
|
|
||||||
If unsure, say N.
|
|
||||||
|
|
||||||
config ASYNC_TX_ENABLE_CHANNEL_SWITCH
|
config ASYNC_TX_ENABLE_CHANNEL_SWITCH
|
||||||
bool
|
bool
|
||||||
|
|
||||||
|
|
|
@ -6,7 +6,6 @@ obj-$(CONFIG_DMA_VIRTUAL_CHANNELS) += virt-dma.o
|
||||||
obj-$(CONFIG_DMA_ACPI) += acpi-dma.o
|
obj-$(CONFIG_DMA_ACPI) += acpi-dma.o
|
||||||
obj-$(CONFIG_DMA_OF) += of-dma.o
|
obj-$(CONFIG_DMA_OF) += of-dma.o
|
||||||
|
|
||||||
obj-$(CONFIG_INTEL_MID_DMAC) += intel_mid_dma.o
|
|
||||||
obj-$(CONFIG_DMATEST) += dmatest.o
|
obj-$(CONFIG_DMATEST) += dmatest.o
|
||||||
obj-$(CONFIG_INTEL_IOATDMA) += ioat/
|
obj-$(CONFIG_INTEL_IOATDMA) += ioat/
|
||||||
obj-$(CONFIG_INTEL_IOP_ADMA) += iop-adma.o
|
obj-$(CONFIG_INTEL_IOP_ADMA) += iop-adma.o
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -1,299 +0,0 @@
|
||||||
/*
|
|
||||||
* intel_mid_dma_regs.h - Intel MID DMA Drivers
|
|
||||||
*
|
|
||||||
* Copyright (C) 2008-10 Intel Corp
|
|
||||||
* Author: Vinod Koul <vinod.koul@intel.com>
|
|
||||||
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
||||||
*
|
|
||||||
* This program is free software; you can redistribute it and/or modify
|
|
||||||
* it under the terms of the GNU General Public License as published by
|
|
||||||
* the Free Software Foundation; version 2 of the License.
|
|
||||||
*
|
|
||||||
* This program is distributed in the hope that it will be useful, but
|
|
||||||
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
||||||
* General Public License for more details.
|
|
||||||
*
|
|
||||||
* You should have received a copy of the GNU General Public License along
|
|
||||||
* with this program; if not, write to the Free Software Foundation, Inc.,
|
|
||||||
* 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA.
|
|
||||||
*
|
|
||||||
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
||||||
*
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
#ifndef __INTEL_MID_DMAC_REGS_H__
|
|
||||||
#define __INTEL_MID_DMAC_REGS_H__
|
|
||||||
|
|
||||||
#include <linux/dmaengine.h>
|
|
||||||
#include <linux/dmapool.h>
|
|
||||||
#include <linux/pci_ids.h>
|
|
||||||
|
|
||||||
#define INTEL_MID_DMA_DRIVER_VERSION "1.1.0"
|
|
||||||
|
|
||||||
#define REG_BIT0 0x00000001
|
|
||||||
#define REG_BIT8 0x00000100
|
|
||||||
#define INT_MASK_WE 0x8
|
|
||||||
#define CLEAR_DONE 0xFFFFEFFF
|
|
||||||
#define UNMASK_INTR_REG(chan_num) \
|
|
||||||
((REG_BIT0 << chan_num) | (REG_BIT8 << chan_num))
|
|
||||||
#define MASK_INTR_REG(chan_num) (REG_BIT8 << chan_num)
|
|
||||||
|
|
||||||
#define ENABLE_CHANNEL(chan_num) \
|
|
||||||
((REG_BIT0 << chan_num) | (REG_BIT8 << chan_num))
|
|
||||||
|
|
||||||
#define DISABLE_CHANNEL(chan_num) \
|
|
||||||
(REG_BIT8 << chan_num)
|
|
||||||
|
|
||||||
#define DESCS_PER_CHANNEL 16
|
|
||||||
/*DMA Registers*/
|
|
||||||
/*registers associated with channel programming*/
|
|
||||||
#define DMA_REG_SIZE 0x400
|
|
||||||
#define DMA_CH_SIZE 0x58
|
|
||||||
|
|
||||||
/*CH X REG = (DMA_CH_SIZE)*CH_NO + REG*/
|
|
||||||
#define SAR 0x00 /* Source Address Register*/
|
|
||||||
#define DAR 0x08 /* Destination Address Register*/
|
|
||||||
#define LLP 0x10 /* Linked List Pointer Register*/
|
|
||||||
#define CTL_LOW 0x18 /* Control Register*/
|
|
||||||
#define CTL_HIGH 0x1C /* Control Register*/
|
|
||||||
#define CFG_LOW 0x40 /* Configuration Register Low*/
|
|
||||||
#define CFG_HIGH 0x44 /* Configuration Register high*/
|
|
||||||
|
|
||||||
#define STATUS_TFR 0x2E8
|
|
||||||
#define STATUS_BLOCK 0x2F0
|
|
||||||
#define STATUS_ERR 0x308
|
|
||||||
|
|
||||||
#define RAW_TFR 0x2C0
|
|
||||||
#define RAW_BLOCK 0x2C8
|
|
||||||
#define RAW_ERR 0x2E0
|
|
||||||
|
|
||||||
#define MASK_TFR 0x310
|
|
||||||
#define MASK_BLOCK 0x318
|
|
||||||
#define MASK_SRC_TRAN 0x320
|
|
||||||
#define MASK_DST_TRAN 0x328
|
|
||||||
#define MASK_ERR 0x330
|
|
||||||
|
|
||||||
#define CLEAR_TFR 0x338
|
|
||||||
#define CLEAR_BLOCK 0x340
|
|
||||||
#define CLEAR_SRC_TRAN 0x348
|
|
||||||
#define CLEAR_DST_TRAN 0x350
|
|
||||||
#define CLEAR_ERR 0x358
|
|
||||||
|
|
||||||
#define INTR_STATUS 0x360
|
|
||||||
#define DMA_CFG 0x398
|
|
||||||
#define DMA_CHAN_EN 0x3A0
|
|
||||||
|
|
||||||
/*DMA channel control registers*/
|
|
||||||
union intel_mid_dma_ctl_lo {
|
|
||||||
struct {
|
|
||||||
u32 int_en:1; /*enable or disable interrupts*/
|
|
||||||
/*should be 0*/
|
|
||||||
u32 dst_tr_width:3; /*destination transfer width*/
|
|
||||||
/*usually 32 bits = 010*/
|
|
||||||
u32 src_tr_width:3; /*source transfer width*/
|
|
||||||
/*usually 32 bits = 010*/
|
|
||||||
u32 dinc:2; /*destination address inc/dec*/
|
|
||||||
/*For mem:INC=00, Periphral NoINC=11*/
|
|
||||||
u32 sinc:2; /*source address inc or dec, as above*/
|
|
||||||
u32 dst_msize:3; /*destination burst transaction length*/
|
|
||||||
/*always = 16 ie 011*/
|
|
||||||
u32 src_msize:3; /*source burst transaction length*/
|
|
||||||
/*always = 16 ie 011*/
|
|
||||||
u32 reser1:3;
|
|
||||||
u32 tt_fc:3; /*transfer type and flow controller*/
|
|
||||||
/*M-M = 000
|
|
||||||
P-M = 010
|
|
||||||
M-P = 001*/
|
|
||||||
u32 dms:2; /*destination master select = 0*/
|
|
||||||
u32 sms:2; /*source master select = 0*/
|
|
||||||
u32 llp_dst_en:1; /*enable/disable destination LLP = 0*/
|
|
||||||
u32 llp_src_en:1; /*enable/disable source LLP = 0*/
|
|
||||||
u32 reser2:3;
|
|
||||||
} ctlx;
|
|
||||||
u32 ctl_lo;
|
|
||||||
};
|
|
||||||
|
|
||||||
union intel_mid_dma_ctl_hi {
|
|
||||||
struct {
|
|
||||||
u32 block_ts:12; /*block transfer size*/
|
|
||||||
u32 done:1; /*Done - updated by DMAC*/
|
|
||||||
u32 reser:19; /*configured by DMAC*/
|
|
||||||
} ctlx;
|
|
||||||
u32 ctl_hi;
|
|
||||||
|
|
||||||
};
|
|
||||||
|
|
||||||
/*DMA channel configuration registers*/
|
|
||||||
union intel_mid_dma_cfg_lo {
|
|
||||||
struct {
|
|
||||||
u32 reser1:5;
|
|
||||||
u32 ch_prior:3; /*channel priority = 0*/
|
|
||||||
u32 ch_susp:1; /*channel suspend = 0*/
|
|
||||||
u32 fifo_empty:1; /*FIFO empty or not R bit = 0*/
|
|
||||||
u32 hs_sel_dst:1; /*select HW/SW destn handshaking*/
|
|
||||||
/*HW = 0, SW = 1*/
|
|
||||||
u32 hs_sel_src:1; /*select HW/SW src handshaking*/
|
|
||||||
u32 reser2:6;
|
|
||||||
u32 dst_hs_pol:1; /*dest HS interface polarity*/
|
|
||||||
u32 src_hs_pol:1; /*src HS interface polarity*/
|
|
||||||
u32 max_abrst:10; /*max AMBA burst len = 0 (no sw limit*/
|
|
||||||
u32 reload_src:1; /*auto reload src addr =1 if src is P*/
|
|
||||||
u32 reload_dst:1; /*AR destn addr =1 if dstn is P*/
|
|
||||||
} cfgx;
|
|
||||||
u32 cfg_lo;
|
|
||||||
};
|
|
||||||
|
|
||||||
union intel_mid_dma_cfg_hi {
|
|
||||||
struct {
|
|
||||||
u32 fcmode:1; /*flow control mode = 1*/
|
|
||||||
u32 fifo_mode:1; /*FIFO mode select = 1*/
|
|
||||||
u32 protctl:3; /*protection control = 0*/
|
|
||||||
u32 rsvd:2;
|
|
||||||
u32 src_per:4; /*src hw HS interface*/
|
|
||||||
u32 dst_per:4; /*dstn hw HS interface*/
|
|
||||||
u32 reser2:17;
|
|
||||||
} cfgx;
|
|
||||||
u32 cfg_hi;
|
|
||||||
};
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* struct intel_mid_dma_chan - internal mid representation of a DMA channel
|
|
||||||
* @chan: dma_chan strcture represetation for mid chan
|
|
||||||
* @ch_regs: MMIO register space pointer to channel register
|
|
||||||
* @dma_base: MMIO register space DMA engine base pointer
|
|
||||||
* @ch_id: DMA channel id
|
|
||||||
* @lock: channel spinlock
|
|
||||||
* @active_list: current active descriptors
|
|
||||||
* @queue: current queued up descriptors
|
|
||||||
* @free_list: current free descriptors
|
|
||||||
* @slave: dma slave structure
|
|
||||||
* @descs_allocated: total number of descriptors allocated
|
|
||||||
* @dma: dma device structure pointer
|
|
||||||
* @busy: bool representing if ch is busy (active txn) or not
|
|
||||||
* @in_use: bool representing if ch is in use or not
|
|
||||||
* @raw_tfr: raw trf interrupt received
|
|
||||||
* @raw_block: raw block interrupt received
|
|
||||||
*/
|
|
||||||
struct intel_mid_dma_chan {
|
|
||||||
struct dma_chan chan;
|
|
||||||
void __iomem *ch_regs;
|
|
||||||
void __iomem *dma_base;
|
|
||||||
int ch_id;
|
|
||||||
spinlock_t lock;
|
|
||||||
struct list_head active_list;
|
|
||||||
struct list_head queue;
|
|
||||||
struct list_head free_list;
|
|
||||||
unsigned int descs_allocated;
|
|
||||||
struct middma_device *dma;
|
|
||||||
bool busy;
|
|
||||||
bool in_use;
|
|
||||||
u32 raw_tfr;
|
|
||||||
u32 raw_block;
|
|
||||||
struct intel_mid_dma_slave *mid_slave;
|
|
||||||
};
|
|
||||||
|
|
||||||
static inline struct intel_mid_dma_chan *to_intel_mid_dma_chan(
|
|
||||||
struct dma_chan *chan)
|
|
||||||
{
|
|
||||||
return container_of(chan, struct intel_mid_dma_chan, chan);
|
|
||||||
}
|
|
||||||
|
|
||||||
enum intel_mid_dma_state {
|
|
||||||
RUNNING = 0,
|
|
||||||
SUSPENDED,
|
|
||||||
};
|
|
||||||
/**
|
|
||||||
* struct middma_device - internal representation of a DMA device
|
|
||||||
* @pdev: PCI device
|
|
||||||
* @dma_base: MMIO register space pointer of DMA
|
|
||||||
* @dma_pool: for allocating DMA descriptors
|
|
||||||
* @common: embedded struct dma_device
|
|
||||||
* @tasklet: dma tasklet for processing interrupts
|
|
||||||
* @ch: per channel data
|
|
||||||
* @pci_id: DMA device PCI ID
|
|
||||||
* @intr_mask: Interrupt mask to be used
|
|
||||||
* @mask_reg: MMIO register for periphral mask
|
|
||||||
* @chan_base: Base ch index (read from driver data)
|
|
||||||
* @max_chan: max number of chs supported (from drv_data)
|
|
||||||
* @block_size: Block size of DMA transfer supported (from drv_data)
|
|
||||||
* @pimr_mask: MMIO register addr for periphral interrupt (from drv_data)
|
|
||||||
* @state: dma PM device state
|
|
||||||
*/
|
|
||||||
struct middma_device {
|
|
||||||
struct pci_dev *pdev;
|
|
||||||
void __iomem *dma_base;
|
|
||||||
struct pci_pool *dma_pool;
|
|
||||||
struct dma_device common;
|
|
||||||
struct tasklet_struct tasklet;
|
|
||||||
struct intel_mid_dma_chan ch[MAX_CHAN];
|
|
||||||
unsigned int pci_id;
|
|
||||||
unsigned int intr_mask;
|
|
||||||
void __iomem *mask_reg;
|
|
||||||
int chan_base;
|
|
||||||
int max_chan;
|
|
||||||
int block_size;
|
|
||||||
unsigned int pimr_mask;
|
|
||||||
enum intel_mid_dma_state state;
|
|
||||||
};
|
|
||||||
|
|
||||||
static inline struct middma_device *to_middma_device(struct dma_device *common)
|
|
||||||
{
|
|
||||||
return container_of(common, struct middma_device, common);
|
|
||||||
}
|
|
||||||
|
|
||||||
struct intel_mid_dma_desc {
|
|
||||||
void __iomem *block; /*ch ptr*/
|
|
||||||
struct list_head desc_node;
|
|
||||||
struct dma_async_tx_descriptor txd;
|
|
||||||
size_t len;
|
|
||||||
dma_addr_t sar;
|
|
||||||
dma_addr_t dar;
|
|
||||||
u32 cfg_hi;
|
|
||||||
u32 cfg_lo;
|
|
||||||
u32 ctl_lo;
|
|
||||||
u32 ctl_hi;
|
|
||||||
struct pci_pool *lli_pool;
|
|
||||||
struct intel_mid_dma_lli *lli;
|
|
||||||
dma_addr_t lli_phys;
|
|
||||||
unsigned int lli_length;
|
|
||||||
unsigned int current_lli;
|
|
||||||
dma_addr_t next;
|
|
||||||
enum dma_transfer_direction dirn;
|
|
||||||
enum dma_status status;
|
|
||||||
enum dma_slave_buswidth width; /*width of DMA txn*/
|
|
||||||
enum intel_mid_dma_mode cfg_mode; /*mode configuration*/
|
|
||||||
|
|
||||||
};
|
|
||||||
|
|
||||||
struct intel_mid_dma_lli {
|
|
||||||
dma_addr_t sar;
|
|
||||||
dma_addr_t dar;
|
|
||||||
dma_addr_t llp;
|
|
||||||
u32 ctl_lo;
|
|
||||||
u32 ctl_hi;
|
|
||||||
} __attribute__ ((packed));
|
|
||||||
|
|
||||||
static inline int test_ch_en(void __iomem *dma, u32 ch_no)
|
|
||||||
{
|
|
||||||
u32 en_reg = ioread32(dma + DMA_CHAN_EN);
|
|
||||||
return (en_reg >> ch_no) & 0x1;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline struct intel_mid_dma_desc *to_intel_mid_dma_desc
|
|
||||||
(struct dma_async_tx_descriptor *txd)
|
|
||||||
{
|
|
||||||
return container_of(txd, struct intel_mid_dma_desc, txd);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline struct intel_mid_dma_slave *to_intel_mid_dma_slave
|
|
||||||
(struct dma_slave_config *slave)
|
|
||||||
{
|
|
||||||
return container_of(slave, struct intel_mid_dma_slave, dma_slave);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
int dma_resume(struct device *dev);
|
|
||||||
|
|
||||||
#endif /*__INTEL_MID_DMAC_REGS_H__*/
|
|
|
@ -159,10 +159,9 @@ config SPI_BUTTERFLY
|
||||||
|
|
||||||
config SPI_CADENCE
|
config SPI_CADENCE
|
||||||
tristate "Cadence SPI controller"
|
tristate "Cadence SPI controller"
|
||||||
depends on ARM
|
|
||||||
help
|
help
|
||||||
This selects the Cadence SPI controller master driver
|
This selects the Cadence SPI controller master driver
|
||||||
used by Xilinx Zynq.
|
used by Xilinx Zynq and ZynqMP.
|
||||||
|
|
||||||
config SPI_CLPS711X
|
config SPI_CLPS711X
|
||||||
tristate "CLPS711X host SPI controller"
|
tristate "CLPS711X host SPI controller"
|
||||||
|
@ -632,7 +631,7 @@ config SPI_DW_PCI
|
||||||
|
|
||||||
config SPI_DW_MID_DMA
|
config SPI_DW_MID_DMA
|
||||||
bool "DMA support for DW SPI controller on Intel MID platform"
|
bool "DMA support for DW SPI controller on Intel MID platform"
|
||||||
depends on SPI_DW_PCI && INTEL_MID_DMAC
|
depends on SPI_DW_PCI && DW_DMAC_PCI
|
||||||
|
|
||||||
config SPI_DW_MMIO
|
config SPI_DW_MMIO
|
||||||
tristate "Memory-mapped io interface driver for DW SPI core"
|
tristate "Memory-mapped io interface driver for DW SPI core"
|
||||||
|
|
|
@ -559,7 +559,7 @@ static void bfin_spi_pump_transfers(unsigned long data)
|
||||||
struct spi_transfer *previous = NULL;
|
struct spi_transfer *previous = NULL;
|
||||||
struct bfin_spi_slave_data *chip = NULL;
|
struct bfin_spi_slave_data *chip = NULL;
|
||||||
unsigned int bits_per_word;
|
unsigned int bits_per_word;
|
||||||
u16 cr, cr_width, dma_width, dma_config;
|
u16 cr, cr_width = 0, dma_width, dma_config;
|
||||||
u32 tranf_success = 1;
|
u32 tranf_success = 1;
|
||||||
u8 full_duplex = 0;
|
u8 full_duplex = 0;
|
||||||
|
|
||||||
|
@ -648,7 +648,6 @@ static void bfin_spi_pump_transfers(unsigned long data)
|
||||||
} else if (bits_per_word == 8) {
|
} else if (bits_per_word == 8) {
|
||||||
drv_data->n_bytes = bits_per_word/8;
|
drv_data->n_bytes = bits_per_word/8;
|
||||||
drv_data->len = transfer->len;
|
drv_data->len = transfer->len;
|
||||||
cr_width = 0;
|
|
||||||
drv_data->ops = &bfin_bfin_spi_transfer_ops_u8;
|
drv_data->ops = &bfin_bfin_spi_transfer_ops_u8;
|
||||||
}
|
}
|
||||||
cr = bfin_read(&drv_data->regs->ctl) & ~(BIT_CTL_TIMOD | BIT_CTL_WORDSIZE);
|
cr = bfin_read(&drv_data->regs->ctl) & ~(BIT_CTL_TIMOD | BIT_CTL_WORDSIZE);
|
||||||
|
|
|
@ -23,29 +23,31 @@
|
||||||
#include "spi-dw.h"
|
#include "spi-dw.h"
|
||||||
|
|
||||||
#ifdef CONFIG_SPI_DW_MID_DMA
|
#ifdef CONFIG_SPI_DW_MID_DMA
|
||||||
#include <linux/intel_mid_dma.h>
|
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
|
#include <linux/platform_data/dma-dw.h>
|
||||||
|
|
||||||
#define RX_BUSY 0
|
#define RX_BUSY 0
|
||||||
#define TX_BUSY 1
|
#define TX_BUSY 1
|
||||||
|
|
||||||
struct mid_dma {
|
static struct dw_dma_slave mid_dma_tx = { .dst_id = 1 };
|
||||||
struct intel_mid_dma_slave dmas_tx;
|
static struct dw_dma_slave mid_dma_rx = { .src_id = 0 };
|
||||||
struct intel_mid_dma_slave dmas_rx;
|
|
||||||
};
|
|
||||||
|
|
||||||
static bool mid_spi_dma_chan_filter(struct dma_chan *chan, void *param)
|
static bool mid_spi_dma_chan_filter(struct dma_chan *chan, void *param)
|
||||||
{
|
{
|
||||||
struct dw_spi *dws = param;
|
struct dw_dma_slave *s = param;
|
||||||
|
|
||||||
return dws->dma_dev == chan->device->dev;
|
if (s->dma_dev != chan->device->dev)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
chan->private = s;
|
||||||
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int mid_spi_dma_init(struct dw_spi *dws)
|
static int mid_spi_dma_init(struct dw_spi *dws)
|
||||||
{
|
{
|
||||||
struct mid_dma *dw_dma = dws->dma_priv;
|
|
||||||
struct pci_dev *dma_dev;
|
struct pci_dev *dma_dev;
|
||||||
struct intel_mid_dma_slave *rxs, *txs;
|
struct dw_dma_slave *tx = dws->dma_tx;
|
||||||
|
struct dw_dma_slave *rx = dws->dma_rx;
|
||||||
dma_cap_mask_t mask;
|
dma_cap_mask_t mask;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -56,28 +58,22 @@ static int mid_spi_dma_init(struct dw_spi *dws)
|
||||||
if (!dma_dev)
|
if (!dma_dev)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
|
|
||||||
dws->dma_dev = &dma_dev->dev;
|
|
||||||
|
|
||||||
dma_cap_zero(mask);
|
dma_cap_zero(mask);
|
||||||
dma_cap_set(DMA_SLAVE, mask);
|
dma_cap_set(DMA_SLAVE, mask);
|
||||||
|
|
||||||
/* 1. Init rx channel */
|
/* 1. Init rx channel */
|
||||||
dws->rxchan = dma_request_channel(mask, mid_spi_dma_chan_filter, dws);
|
rx->dma_dev = &dma_dev->dev;
|
||||||
|
dws->rxchan = dma_request_channel(mask, mid_spi_dma_chan_filter, rx);
|
||||||
if (!dws->rxchan)
|
if (!dws->rxchan)
|
||||||
goto err_exit;
|
goto err_exit;
|
||||||
rxs = &dw_dma->dmas_rx;
|
dws->master->dma_rx = dws->rxchan;
|
||||||
rxs->hs_mode = LNW_DMA_HW_HS;
|
|
||||||
rxs->cfg_mode = LNW_DMA_PER_TO_MEM;
|
|
||||||
dws->rxchan->private = rxs;
|
|
||||||
|
|
||||||
/* 2. Init tx channel */
|
/* 2. Init tx channel */
|
||||||
dws->txchan = dma_request_channel(mask, mid_spi_dma_chan_filter, dws);
|
tx->dma_dev = &dma_dev->dev;
|
||||||
|
dws->txchan = dma_request_channel(mask, mid_spi_dma_chan_filter, tx);
|
||||||
if (!dws->txchan)
|
if (!dws->txchan)
|
||||||
goto free_rxchan;
|
goto free_rxchan;
|
||||||
txs = &dw_dma->dmas_tx;
|
dws->master->dma_tx = dws->txchan;
|
||||||
txs->hs_mode = LNW_DMA_HW_HS;
|
|
||||||
txs->cfg_mode = LNW_DMA_MEM_TO_PER;
|
|
||||||
dws->txchan->private = txs;
|
|
||||||
|
|
||||||
dws->dma_inited = 1;
|
dws->dma_inited = 1;
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -100,6 +96,42 @@ static void mid_spi_dma_exit(struct dw_spi *dws)
|
||||||
dma_release_channel(dws->rxchan);
|
dma_release_channel(dws->rxchan);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static irqreturn_t dma_transfer(struct dw_spi *dws)
|
||||||
|
{
|
||||||
|
u16 irq_status = dw_readl(dws, DW_SPI_ISR);
|
||||||
|
|
||||||
|
if (!irq_status)
|
||||||
|
return IRQ_NONE;
|
||||||
|
|
||||||
|
dw_readl(dws, DW_SPI_ICR);
|
||||||
|
spi_reset_chip(dws);
|
||||||
|
|
||||||
|
dev_err(&dws->master->dev, "%s: FIFO overrun/underrun\n", __func__);
|
||||||
|
dws->master->cur_msg->status = -EIO;
|
||||||
|
spi_finalize_current_transfer(dws->master);
|
||||||
|
return IRQ_HANDLED;
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool mid_spi_can_dma(struct spi_master *master, struct spi_device *spi,
|
||||||
|
struct spi_transfer *xfer)
|
||||||
|
{
|
||||||
|
struct dw_spi *dws = spi_master_get_devdata(master);
|
||||||
|
|
||||||
|
if (!dws->dma_inited)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
return xfer->len > dws->fifo_len;
|
||||||
|
}
|
||||||
|
|
||||||
|
static enum dma_slave_buswidth convert_dma_width(u32 dma_width) {
|
||||||
|
if (dma_width == 1)
|
||||||
|
return DMA_SLAVE_BUSWIDTH_1_BYTE;
|
||||||
|
else if (dma_width == 2)
|
||||||
|
return DMA_SLAVE_BUSWIDTH_2_BYTES;
|
||||||
|
|
||||||
|
return DMA_SLAVE_BUSWIDTH_UNDEFINED;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* dws->dma_chan_busy is set before the dma transfer starts, callback for tx
|
* dws->dma_chan_busy is set before the dma transfer starts, callback for tx
|
||||||
* channel will clear a corresponding bit.
|
* channel will clear a corresponding bit.
|
||||||
|
@ -111,33 +143,30 @@ static void dw_spi_dma_tx_done(void *arg)
|
||||||
clear_bit(TX_BUSY, &dws->dma_chan_busy);
|
clear_bit(TX_BUSY, &dws->dma_chan_busy);
|
||||||
if (test_bit(RX_BUSY, &dws->dma_chan_busy))
|
if (test_bit(RX_BUSY, &dws->dma_chan_busy))
|
||||||
return;
|
return;
|
||||||
dw_spi_xfer_done(dws);
|
spi_finalize_current_transfer(dws->master);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct dma_async_tx_descriptor *dw_spi_dma_prepare_tx(struct dw_spi *dws)
|
static struct dma_async_tx_descriptor *dw_spi_dma_prepare_tx(struct dw_spi *dws,
|
||||||
|
struct spi_transfer *xfer)
|
||||||
{
|
{
|
||||||
struct dma_slave_config txconf;
|
struct dma_slave_config txconf;
|
||||||
struct dma_async_tx_descriptor *txdesc;
|
struct dma_async_tx_descriptor *txdesc;
|
||||||
|
|
||||||
if (!dws->tx_dma)
|
if (!xfer->tx_buf)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
txconf.direction = DMA_MEM_TO_DEV;
|
txconf.direction = DMA_MEM_TO_DEV;
|
||||||
txconf.dst_addr = dws->dma_addr;
|
txconf.dst_addr = dws->dma_addr;
|
||||||
txconf.dst_maxburst = LNW_DMA_MSIZE_16;
|
txconf.dst_maxburst = 16;
|
||||||
txconf.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
txconf.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
||||||
txconf.dst_addr_width = dws->dma_width;
|
txconf.dst_addr_width = convert_dma_width(dws->dma_width);
|
||||||
txconf.device_fc = false;
|
txconf.device_fc = false;
|
||||||
|
|
||||||
dmaengine_slave_config(dws->txchan, &txconf);
|
dmaengine_slave_config(dws->txchan, &txconf);
|
||||||
|
|
||||||
memset(&dws->tx_sgl, 0, sizeof(dws->tx_sgl));
|
|
||||||
dws->tx_sgl.dma_address = dws->tx_dma;
|
|
||||||
dws->tx_sgl.length = dws->len;
|
|
||||||
|
|
||||||
txdesc = dmaengine_prep_slave_sg(dws->txchan,
|
txdesc = dmaengine_prep_slave_sg(dws->txchan,
|
||||||
&dws->tx_sgl,
|
xfer->tx_sg.sgl,
|
||||||
1,
|
xfer->tx_sg.nents,
|
||||||
DMA_MEM_TO_DEV,
|
DMA_MEM_TO_DEV,
|
||||||
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
||||||
if (!txdesc)
|
if (!txdesc)
|
||||||
|
@ -160,33 +189,30 @@ static void dw_spi_dma_rx_done(void *arg)
|
||||||
clear_bit(RX_BUSY, &dws->dma_chan_busy);
|
clear_bit(RX_BUSY, &dws->dma_chan_busy);
|
||||||
if (test_bit(TX_BUSY, &dws->dma_chan_busy))
|
if (test_bit(TX_BUSY, &dws->dma_chan_busy))
|
||||||
return;
|
return;
|
||||||
dw_spi_xfer_done(dws);
|
spi_finalize_current_transfer(dws->master);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct dma_async_tx_descriptor *dw_spi_dma_prepare_rx(struct dw_spi *dws)
|
static struct dma_async_tx_descriptor *dw_spi_dma_prepare_rx(struct dw_spi *dws,
|
||||||
|
struct spi_transfer *xfer)
|
||||||
{
|
{
|
||||||
struct dma_slave_config rxconf;
|
struct dma_slave_config rxconf;
|
||||||
struct dma_async_tx_descriptor *rxdesc;
|
struct dma_async_tx_descriptor *rxdesc;
|
||||||
|
|
||||||
if (!dws->rx_dma)
|
if (!xfer->rx_buf)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
rxconf.direction = DMA_DEV_TO_MEM;
|
rxconf.direction = DMA_DEV_TO_MEM;
|
||||||
rxconf.src_addr = dws->dma_addr;
|
rxconf.src_addr = dws->dma_addr;
|
||||||
rxconf.src_maxburst = LNW_DMA_MSIZE_16;
|
rxconf.src_maxburst = 16;
|
||||||
rxconf.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
rxconf.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
||||||
rxconf.src_addr_width = dws->dma_width;
|
rxconf.src_addr_width = convert_dma_width(dws->dma_width);
|
||||||
rxconf.device_fc = false;
|
rxconf.device_fc = false;
|
||||||
|
|
||||||
dmaengine_slave_config(dws->rxchan, &rxconf);
|
dmaengine_slave_config(dws->rxchan, &rxconf);
|
||||||
|
|
||||||
memset(&dws->rx_sgl, 0, sizeof(dws->rx_sgl));
|
|
||||||
dws->rx_sgl.dma_address = dws->rx_dma;
|
|
||||||
dws->rx_sgl.length = dws->len;
|
|
||||||
|
|
||||||
rxdesc = dmaengine_prep_slave_sg(dws->rxchan,
|
rxdesc = dmaengine_prep_slave_sg(dws->rxchan,
|
||||||
&dws->rx_sgl,
|
xfer->rx_sg.sgl,
|
||||||
1,
|
xfer->rx_sg.nents,
|
||||||
DMA_DEV_TO_MEM,
|
DMA_DEV_TO_MEM,
|
||||||
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
||||||
if (!rxdesc)
|
if (!rxdesc)
|
||||||
|
@ -198,37 +224,36 @@ static struct dma_async_tx_descriptor *dw_spi_dma_prepare_rx(struct dw_spi *dws)
|
||||||
return rxdesc;
|
return rxdesc;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dw_spi_dma_setup(struct dw_spi *dws)
|
static int mid_spi_dma_setup(struct dw_spi *dws, struct spi_transfer *xfer)
|
||||||
{
|
{
|
||||||
u16 dma_ctrl = 0;
|
u16 dma_ctrl = 0;
|
||||||
|
|
||||||
spi_enable_chip(dws, 0);
|
dw_writel(dws, DW_SPI_DMARDLR, 0xf);
|
||||||
|
dw_writel(dws, DW_SPI_DMATDLR, 0x10);
|
||||||
|
|
||||||
dw_writew(dws, DW_SPI_DMARDLR, 0xf);
|
if (xfer->tx_buf)
|
||||||
dw_writew(dws, DW_SPI_DMATDLR, 0x10);
|
|
||||||
|
|
||||||
if (dws->tx_dma)
|
|
||||||
dma_ctrl |= SPI_DMA_TDMAE;
|
dma_ctrl |= SPI_DMA_TDMAE;
|
||||||
if (dws->rx_dma)
|
if (xfer->rx_buf)
|
||||||
dma_ctrl |= SPI_DMA_RDMAE;
|
dma_ctrl |= SPI_DMA_RDMAE;
|
||||||
dw_writew(dws, DW_SPI_DMACR, dma_ctrl);
|
dw_writel(dws, DW_SPI_DMACR, dma_ctrl);
|
||||||
|
|
||||||
spi_enable_chip(dws, 1);
|
/* Set the interrupt mask */
|
||||||
|
spi_umask_intr(dws, SPI_INT_TXOI | SPI_INT_RXUI | SPI_INT_RXOI);
|
||||||
|
|
||||||
|
dws->transfer_handler = dma_transfer;
|
||||||
|
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int mid_spi_dma_transfer(struct dw_spi *dws, int cs_change)
|
static int mid_spi_dma_transfer(struct dw_spi *dws, struct spi_transfer *xfer)
|
||||||
{
|
{
|
||||||
struct dma_async_tx_descriptor *txdesc, *rxdesc;
|
struct dma_async_tx_descriptor *txdesc, *rxdesc;
|
||||||
|
|
||||||
/* 1. setup DMA related registers */
|
/* Prepare the TX dma transfer */
|
||||||
if (cs_change)
|
txdesc = dw_spi_dma_prepare_tx(dws, xfer);
|
||||||
dw_spi_dma_setup(dws);
|
|
||||||
|
|
||||||
/* 2. Prepare the TX dma transfer */
|
/* Prepare the RX dma transfer */
|
||||||
txdesc = dw_spi_dma_prepare_tx(dws);
|
rxdesc = dw_spi_dma_prepare_rx(dws, xfer);
|
||||||
|
|
||||||
/* 3. Prepare the RX dma transfer */
|
|
||||||
rxdesc = dw_spi_dma_prepare_rx(dws);
|
|
||||||
|
|
||||||
/* rx must be started before tx due to spi instinct */
|
/* rx must be started before tx due to spi instinct */
|
||||||
if (rxdesc) {
|
if (rxdesc) {
|
||||||
|
@ -246,10 +271,25 @@ static int mid_spi_dma_transfer(struct dw_spi *dws, int cs_change)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void mid_spi_dma_stop(struct dw_spi *dws)
|
||||||
|
{
|
||||||
|
if (test_bit(TX_BUSY, &dws->dma_chan_busy)) {
|
||||||
|
dmaengine_terminate_all(dws->txchan);
|
||||||
|
clear_bit(TX_BUSY, &dws->dma_chan_busy);
|
||||||
|
}
|
||||||
|
if (test_bit(RX_BUSY, &dws->dma_chan_busy)) {
|
||||||
|
dmaengine_terminate_all(dws->rxchan);
|
||||||
|
clear_bit(RX_BUSY, &dws->dma_chan_busy);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static struct dw_spi_dma_ops mid_dma_ops = {
|
static struct dw_spi_dma_ops mid_dma_ops = {
|
||||||
.dma_init = mid_spi_dma_init,
|
.dma_init = mid_spi_dma_init,
|
||||||
.dma_exit = mid_spi_dma_exit,
|
.dma_exit = mid_spi_dma_exit,
|
||||||
|
.dma_setup = mid_spi_dma_setup,
|
||||||
|
.can_dma = mid_spi_can_dma,
|
||||||
.dma_transfer = mid_spi_dma_transfer,
|
.dma_transfer = mid_spi_dma_transfer,
|
||||||
|
.dma_stop = mid_spi_dma_stop,
|
||||||
};
|
};
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -282,9 +322,8 @@ int dw_spi_mid_init(struct dw_spi *dws)
|
||||||
iounmap(clk_reg);
|
iounmap(clk_reg);
|
||||||
|
|
||||||
#ifdef CONFIG_SPI_DW_MID_DMA
|
#ifdef CONFIG_SPI_DW_MID_DMA
|
||||||
dws->dma_priv = kzalloc(sizeof(struct mid_dma), GFP_KERNEL);
|
dws->dma_tx = &mid_dma_tx;
|
||||||
if (!dws->dma_priv)
|
dws->dma_rx = &mid_dma_rx;
|
||||||
return -ENOMEM;
|
|
||||||
dws->dma_ops = &mid_dma_ops;
|
dws->dma_ops = &mid_dma_ops;
|
||||||
#endif
|
#endif
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -28,11 +28,6 @@
|
||||||
#include <linux/debugfs.h>
|
#include <linux/debugfs.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#define START_STATE ((void *)0)
|
|
||||||
#define RUNNING_STATE ((void *)1)
|
|
||||||
#define DONE_STATE ((void *)2)
|
|
||||||
#define ERROR_STATE ((void *)-1)
|
|
||||||
|
|
||||||
/* Slave spi_dev related */
|
/* Slave spi_dev related */
|
||||||
struct chip_data {
|
struct chip_data {
|
||||||
u16 cr0;
|
u16 cr0;
|
||||||
|
@ -143,13 +138,26 @@ static inline void dw_spi_debugfs_remove(struct dw_spi *dws)
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_DEBUG_FS */
|
#endif /* CONFIG_DEBUG_FS */
|
||||||
|
|
||||||
|
static void dw_spi_set_cs(struct spi_device *spi, bool enable)
|
||||||
|
{
|
||||||
|
struct dw_spi *dws = spi_master_get_devdata(spi->master);
|
||||||
|
struct chip_data *chip = spi_get_ctldata(spi);
|
||||||
|
|
||||||
|
/* Chip select logic is inverted from spi_set_cs() */
|
||||||
|
if (chip && chip->cs_control)
|
||||||
|
chip->cs_control(!enable);
|
||||||
|
|
||||||
|
if (!enable)
|
||||||
|
dw_writel(dws, DW_SPI_SER, BIT(spi->chip_select));
|
||||||
|
}
|
||||||
|
|
||||||
/* Return the max entries we can fill into tx fifo */
|
/* Return the max entries we can fill into tx fifo */
|
||||||
static inline u32 tx_max(struct dw_spi *dws)
|
static inline u32 tx_max(struct dw_spi *dws)
|
||||||
{
|
{
|
||||||
u32 tx_left, tx_room, rxtx_gap;
|
u32 tx_left, tx_room, rxtx_gap;
|
||||||
|
|
||||||
tx_left = (dws->tx_end - dws->tx) / dws->n_bytes;
|
tx_left = (dws->tx_end - dws->tx) / dws->n_bytes;
|
||||||
tx_room = dws->fifo_len - dw_readw(dws, DW_SPI_TXFLR);
|
tx_room = dws->fifo_len - dw_readl(dws, DW_SPI_TXFLR);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Another concern is about the tx/rx mismatch, we
|
* Another concern is about the tx/rx mismatch, we
|
||||||
|
@ -170,7 +178,7 @@ static inline u32 rx_max(struct dw_spi *dws)
|
||||||
{
|
{
|
||||||
u32 rx_left = (dws->rx_end - dws->rx) / dws->n_bytes;
|
u32 rx_left = (dws->rx_end - dws->rx) / dws->n_bytes;
|
||||||
|
|
||||||
return min_t(u32, rx_left, dw_readw(dws, DW_SPI_RXFLR));
|
return min_t(u32, rx_left, dw_readl(dws, DW_SPI_RXFLR));
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dw_writer(struct dw_spi *dws)
|
static void dw_writer(struct dw_spi *dws)
|
||||||
|
@ -186,7 +194,7 @@ static void dw_writer(struct dw_spi *dws)
|
||||||
else
|
else
|
||||||
txw = *(u16 *)(dws->tx);
|
txw = *(u16 *)(dws->tx);
|
||||||
}
|
}
|
||||||
dw_writew(dws, DW_SPI_DR, txw);
|
dw_writel(dws, DW_SPI_DR, txw);
|
||||||
dws->tx += dws->n_bytes;
|
dws->tx += dws->n_bytes;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -197,7 +205,7 @@ static void dw_reader(struct dw_spi *dws)
|
||||||
u16 rxw;
|
u16 rxw;
|
||||||
|
|
||||||
while (max--) {
|
while (max--) {
|
||||||
rxw = dw_readw(dws, DW_SPI_DR);
|
rxw = dw_readl(dws, DW_SPI_DR);
|
||||||
/* Care rx only if the transfer's original "rx" is not null */
|
/* Care rx only if the transfer's original "rx" is not null */
|
||||||
if (dws->rx_end - dws->len) {
|
if (dws->rx_end - dws->len) {
|
||||||
if (dws->n_bytes == 1)
|
if (dws->n_bytes == 1)
|
||||||
|
@ -209,103 +217,22 @@ static void dw_reader(struct dw_spi *dws)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void *next_transfer(struct dw_spi *dws)
|
|
||||||
{
|
|
||||||
struct spi_message *msg = dws->cur_msg;
|
|
||||||
struct spi_transfer *trans = dws->cur_transfer;
|
|
||||||
|
|
||||||
/* Move to next transfer */
|
|
||||||
if (trans->transfer_list.next != &msg->transfers) {
|
|
||||||
dws->cur_transfer =
|
|
||||||
list_entry(trans->transfer_list.next,
|
|
||||||
struct spi_transfer,
|
|
||||||
transfer_list);
|
|
||||||
return RUNNING_STATE;
|
|
||||||
}
|
|
||||||
|
|
||||||
return DONE_STATE;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Note: first step is the protocol driver prepares
|
|
||||||
* a dma-capable memory, and this func just need translate
|
|
||||||
* the virt addr to physical
|
|
||||||
*/
|
|
||||||
static int map_dma_buffers(struct dw_spi *dws)
|
|
||||||
{
|
|
||||||
if (!dws->cur_msg->is_dma_mapped
|
|
||||||
|| !dws->dma_inited
|
|
||||||
|| !dws->cur_chip->enable_dma
|
|
||||||
|| !dws->dma_ops)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
if (dws->cur_transfer->tx_dma)
|
|
||||||
dws->tx_dma = dws->cur_transfer->tx_dma;
|
|
||||||
|
|
||||||
if (dws->cur_transfer->rx_dma)
|
|
||||||
dws->rx_dma = dws->cur_transfer->rx_dma;
|
|
||||||
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Caller already set message->status; dma and pio irqs are blocked */
|
|
||||||
static void giveback(struct dw_spi *dws)
|
|
||||||
{
|
|
||||||
struct spi_transfer *last_transfer;
|
|
||||||
struct spi_message *msg;
|
|
||||||
|
|
||||||
msg = dws->cur_msg;
|
|
||||||
dws->cur_msg = NULL;
|
|
||||||
dws->cur_transfer = NULL;
|
|
||||||
dws->prev_chip = dws->cur_chip;
|
|
||||||
dws->cur_chip = NULL;
|
|
||||||
dws->dma_mapped = 0;
|
|
||||||
|
|
||||||
last_transfer = list_last_entry(&msg->transfers, struct spi_transfer,
|
|
||||||
transfer_list);
|
|
||||||
|
|
||||||
if (!last_transfer->cs_change)
|
|
||||||
spi_chip_sel(dws, msg->spi, 0);
|
|
||||||
|
|
||||||
spi_finalize_current_message(dws->master);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void int_error_stop(struct dw_spi *dws, const char *msg)
|
static void int_error_stop(struct dw_spi *dws, const char *msg)
|
||||||
{
|
{
|
||||||
/* Stop the hw */
|
spi_reset_chip(dws);
|
||||||
spi_enable_chip(dws, 0);
|
|
||||||
|
|
||||||
dev_err(&dws->master->dev, "%s\n", msg);
|
dev_err(&dws->master->dev, "%s\n", msg);
|
||||||
dws->cur_msg->state = ERROR_STATE;
|
dws->master->cur_msg->status = -EIO;
|
||||||
tasklet_schedule(&dws->pump_transfers);
|
spi_finalize_current_transfer(dws->master);
|
||||||
}
|
}
|
||||||
|
|
||||||
void dw_spi_xfer_done(struct dw_spi *dws)
|
|
||||||
{
|
|
||||||
/* Update total byte transferred return count actual bytes read */
|
|
||||||
dws->cur_msg->actual_length += dws->len;
|
|
||||||
|
|
||||||
/* Move to next transfer */
|
|
||||||
dws->cur_msg->state = next_transfer(dws);
|
|
||||||
|
|
||||||
/* Handle end of message */
|
|
||||||
if (dws->cur_msg->state == DONE_STATE) {
|
|
||||||
dws->cur_msg->status = 0;
|
|
||||||
giveback(dws);
|
|
||||||
} else
|
|
||||||
tasklet_schedule(&dws->pump_transfers);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL_GPL(dw_spi_xfer_done);
|
|
||||||
|
|
||||||
static irqreturn_t interrupt_transfer(struct dw_spi *dws)
|
static irqreturn_t interrupt_transfer(struct dw_spi *dws)
|
||||||
{
|
{
|
||||||
u16 irq_status = dw_readw(dws, DW_SPI_ISR);
|
u16 irq_status = dw_readl(dws, DW_SPI_ISR);
|
||||||
|
|
||||||
/* Error handling */
|
/* Error handling */
|
||||||
if (irq_status & (SPI_INT_TXOI | SPI_INT_RXOI | SPI_INT_RXUI)) {
|
if (irq_status & (SPI_INT_TXOI | SPI_INT_RXOI | SPI_INT_RXUI)) {
|
||||||
dw_readw(dws, DW_SPI_TXOICR);
|
dw_readl(dws, DW_SPI_ICR);
|
||||||
dw_readw(dws, DW_SPI_RXOICR);
|
|
||||||
dw_readw(dws, DW_SPI_RXUICR);
|
|
||||||
int_error_stop(dws, "interrupt_transfer: fifo overrun/underrun");
|
int_error_stop(dws, "interrupt_transfer: fifo overrun/underrun");
|
||||||
return IRQ_HANDLED;
|
return IRQ_HANDLED;
|
||||||
}
|
}
|
||||||
|
@ -313,7 +240,7 @@ static irqreturn_t interrupt_transfer(struct dw_spi *dws)
|
||||||
dw_reader(dws);
|
dw_reader(dws);
|
||||||
if (dws->rx_end == dws->rx) {
|
if (dws->rx_end == dws->rx) {
|
||||||
spi_mask_intr(dws, SPI_INT_TXEI);
|
spi_mask_intr(dws, SPI_INT_TXEI);
|
||||||
dw_spi_xfer_done(dws);
|
spi_finalize_current_transfer(dws->master);
|
||||||
return IRQ_HANDLED;
|
return IRQ_HANDLED;
|
||||||
}
|
}
|
||||||
if (irq_status & SPI_INT_TXEI) {
|
if (irq_status & SPI_INT_TXEI) {
|
||||||
|
@ -328,13 +255,14 @@ static irqreturn_t interrupt_transfer(struct dw_spi *dws)
|
||||||
|
|
||||||
static irqreturn_t dw_spi_irq(int irq, void *dev_id)
|
static irqreturn_t dw_spi_irq(int irq, void *dev_id)
|
||||||
{
|
{
|
||||||
struct dw_spi *dws = dev_id;
|
struct spi_master *master = dev_id;
|
||||||
u16 irq_status = dw_readw(dws, DW_SPI_ISR) & 0x3f;
|
struct dw_spi *dws = spi_master_get_devdata(master);
|
||||||
|
u16 irq_status = dw_readl(dws, DW_SPI_ISR) & 0x3f;
|
||||||
|
|
||||||
if (!irq_status)
|
if (!irq_status)
|
||||||
return IRQ_NONE;
|
return IRQ_NONE;
|
||||||
|
|
||||||
if (!dws->cur_msg) {
|
if (!master->cur_msg) {
|
||||||
spi_mask_intr(dws, SPI_INT_TXEI);
|
spi_mask_intr(dws, SPI_INT_TXEI);
|
||||||
return IRQ_HANDLED;
|
return IRQ_HANDLED;
|
||||||
}
|
}
|
||||||
|
@ -343,7 +271,7 @@ static irqreturn_t dw_spi_irq(int irq, void *dev_id)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Must be called inside pump_transfers() */
|
/* Must be called inside pump_transfers() */
|
||||||
static void poll_transfer(struct dw_spi *dws)
|
static int poll_transfer(struct dw_spi *dws)
|
||||||
{
|
{
|
||||||
do {
|
do {
|
||||||
dw_writer(dws);
|
dw_writer(dws);
|
||||||
|
@ -351,64 +279,32 @@ static void poll_transfer(struct dw_spi *dws)
|
||||||
cpu_relax();
|
cpu_relax();
|
||||||
} while (dws->rx_end > dws->rx);
|
} while (dws->rx_end > dws->rx);
|
||||||
|
|
||||||
dw_spi_xfer_done(dws);
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void pump_transfers(unsigned long data)
|
static int dw_spi_transfer_one(struct spi_master *master,
|
||||||
|
struct spi_device *spi, struct spi_transfer *transfer)
|
||||||
{
|
{
|
||||||
struct dw_spi *dws = (struct dw_spi *)data;
|
struct dw_spi *dws = spi_master_get_devdata(master);
|
||||||
struct spi_message *message = NULL;
|
struct chip_data *chip = spi_get_ctldata(spi);
|
||||||
struct spi_transfer *transfer = NULL;
|
|
||||||
struct spi_transfer *previous = NULL;
|
|
||||||
struct spi_device *spi = NULL;
|
|
||||||
struct chip_data *chip = NULL;
|
|
||||||
u8 bits = 0;
|
|
||||||
u8 imask = 0;
|
u8 imask = 0;
|
||||||
u8 cs_change = 0;
|
u16 txlevel = 0;
|
||||||
u16 txint_level = 0;
|
|
||||||
u16 clk_div = 0;
|
u16 clk_div = 0;
|
||||||
u32 speed = 0;
|
u32 speed = 0;
|
||||||
u32 cr0 = 0;
|
u32 cr0 = 0;
|
||||||
|
int ret;
|
||||||
|
|
||||||
/* Get current state information */
|
dws->dma_mapped = 0;
|
||||||
message = dws->cur_msg;
|
|
||||||
transfer = dws->cur_transfer;
|
|
||||||
chip = dws->cur_chip;
|
|
||||||
spi = message->spi;
|
|
||||||
|
|
||||||
if (message->state == ERROR_STATE) {
|
|
||||||
message->status = -EIO;
|
|
||||||
goto early_exit;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Handle end of message */
|
|
||||||
if (message->state == DONE_STATE) {
|
|
||||||
message->status = 0;
|
|
||||||
goto early_exit;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Delay if requested at end of transfer */
|
|
||||||
if (message->state == RUNNING_STATE) {
|
|
||||||
previous = list_entry(transfer->transfer_list.prev,
|
|
||||||
struct spi_transfer,
|
|
||||||
transfer_list);
|
|
||||||
if (previous->delay_usecs)
|
|
||||||
udelay(previous->delay_usecs);
|
|
||||||
}
|
|
||||||
|
|
||||||
dws->n_bytes = chip->n_bytes;
|
dws->n_bytes = chip->n_bytes;
|
||||||
dws->dma_width = chip->dma_width;
|
dws->dma_width = chip->dma_width;
|
||||||
dws->cs_control = chip->cs_control;
|
|
||||||
|
|
||||||
dws->rx_dma = transfer->rx_dma;
|
|
||||||
dws->tx_dma = transfer->tx_dma;
|
|
||||||
dws->tx = (void *)transfer->tx_buf;
|
dws->tx = (void *)transfer->tx_buf;
|
||||||
dws->tx_end = dws->tx + transfer->len;
|
dws->tx_end = dws->tx + transfer->len;
|
||||||
dws->rx = transfer->rx_buf;
|
dws->rx = transfer->rx_buf;
|
||||||
dws->rx_end = dws->rx + transfer->len;
|
dws->rx_end = dws->rx + transfer->len;
|
||||||
dws->len = dws->cur_transfer->len;
|
dws->len = transfer->len;
|
||||||
if (chip != dws->prev_chip)
|
|
||||||
cs_change = 1;
|
spi_enable_chip(dws, 0);
|
||||||
|
|
||||||
cr0 = chip->cr0;
|
cr0 = chip->cr0;
|
||||||
|
|
||||||
|
@ -416,32 +312,37 @@ static void pump_transfers(unsigned long data)
|
||||||
if (transfer->speed_hz) {
|
if (transfer->speed_hz) {
|
||||||
speed = chip->speed_hz;
|
speed = chip->speed_hz;
|
||||||
|
|
||||||
if ((transfer->speed_hz != speed) || (!chip->clk_div)) {
|
if ((transfer->speed_hz != speed) || !chip->clk_div) {
|
||||||
speed = transfer->speed_hz;
|
speed = transfer->speed_hz;
|
||||||
|
|
||||||
/* clk_div doesn't support odd number */
|
/* clk_div doesn't support odd number */
|
||||||
clk_div = dws->max_freq / speed;
|
clk_div = (dws->max_freq / speed + 1) & 0xfffe;
|
||||||
clk_div = (clk_div + 1) & 0xfffe;
|
|
||||||
|
|
||||||
chip->speed_hz = speed;
|
chip->speed_hz = speed;
|
||||||
chip->clk_div = clk_div;
|
chip->clk_div = clk_div;
|
||||||
|
|
||||||
|
spi_set_clk(dws, chip->clk_div);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (transfer->bits_per_word) {
|
if (transfer->bits_per_word) {
|
||||||
bits = transfer->bits_per_word;
|
if (transfer->bits_per_word == 8) {
|
||||||
dws->n_bytes = dws->dma_width = bits >> 3;
|
dws->n_bytes = 1;
|
||||||
cr0 = (bits - 1)
|
dws->dma_width = 1;
|
||||||
|
} else if (transfer->bits_per_word == 16) {
|
||||||
|
dws->n_bytes = 2;
|
||||||
|
dws->dma_width = 2;
|
||||||
|
}
|
||||||
|
cr0 = (transfer->bits_per_word - 1)
|
||||||
| (chip->type << SPI_FRF_OFFSET)
|
| (chip->type << SPI_FRF_OFFSET)
|
||||||
| (spi->mode << SPI_MODE_OFFSET)
|
| (spi->mode << SPI_MODE_OFFSET)
|
||||||
| (chip->tmode << SPI_TMOD_OFFSET);
|
| (chip->tmode << SPI_TMOD_OFFSET);
|
||||||
}
|
}
|
||||||
message->state = RUNNING_STATE;
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Adjust transfer mode if necessary. Requires platform dependent
|
* Adjust transfer mode if necessary. Requires platform dependent
|
||||||
* chipselect mechanism.
|
* chipselect mechanism.
|
||||||
*/
|
*/
|
||||||
if (dws->cs_control) {
|
if (chip->cs_control) {
|
||||||
if (dws->rx && dws->tx)
|
if (dws->rx && dws->tx)
|
||||||
chip->tmode = SPI_TMOD_TR;
|
chip->tmode = SPI_TMOD_TR;
|
||||||
else if (dws->rx)
|
else if (dws->rx)
|
||||||
|
@ -453,80 +354,60 @@ static void pump_transfers(unsigned long data)
|
||||||
cr0 |= (chip->tmode << SPI_TMOD_OFFSET);
|
cr0 |= (chip->tmode << SPI_TMOD_OFFSET);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
dw_writel(dws, DW_SPI_CTRL0, cr0);
|
||||||
|
|
||||||
/* Check if current transfer is a DMA transaction */
|
/* Check if current transfer is a DMA transaction */
|
||||||
dws->dma_mapped = map_dma_buffers(dws);
|
if (master->can_dma && master->can_dma(master, spi, transfer))
|
||||||
|
dws->dma_mapped = master->cur_msg_mapped;
|
||||||
|
|
||||||
|
/* For poll mode just disable all interrupts */
|
||||||
|
spi_mask_intr(dws, 0xff);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Interrupt mode
|
* Interrupt mode
|
||||||
* we only need set the TXEI IRQ, as TX/RX always happen syncronizely
|
* we only need set the TXEI IRQ, as TX/RX always happen syncronizely
|
||||||
*/
|
*/
|
||||||
if (!dws->dma_mapped && !chip->poll_mode) {
|
if (dws->dma_mapped) {
|
||||||
int templen = dws->len / dws->n_bytes;
|
ret = dws->dma_ops->dma_setup(dws, transfer);
|
||||||
|
if (ret < 0) {
|
||||||
txint_level = dws->fifo_len / 2;
|
spi_enable_chip(dws, 1);
|
||||||
txint_level = (templen > txint_level) ? txint_level : templen;
|
return ret;
|
||||||
|
}
|
||||||
|
} else if (!chip->poll_mode) {
|
||||||
|
txlevel = min_t(u16, dws->fifo_len / 2, dws->len / dws->n_bytes);
|
||||||
|
dw_writel(dws, DW_SPI_TXFLTR, txlevel);
|
||||||
|
|
||||||
|
/* Set the interrupt mask */
|
||||||
imask |= SPI_INT_TXEI | SPI_INT_TXOI |
|
imask |= SPI_INT_TXEI | SPI_INT_TXOI |
|
||||||
SPI_INT_RXUI | SPI_INT_RXOI;
|
SPI_INT_RXUI | SPI_INT_RXOI;
|
||||||
|
spi_umask_intr(dws, imask);
|
||||||
|
|
||||||
dws->transfer_handler = interrupt_transfer;
|
dws->transfer_handler = interrupt_transfer;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
spi_enable_chip(dws, 1);
|
||||||
* Reprogram registers only if
|
|
||||||
* 1. chip select changes
|
|
||||||
* 2. clk_div is changed
|
|
||||||
* 3. control value changes
|
|
||||||
*/
|
|
||||||
if (dw_readw(dws, DW_SPI_CTRL0) != cr0 || cs_change || clk_div || imask) {
|
|
||||||
spi_enable_chip(dws, 0);
|
|
||||||
|
|
||||||
if (dw_readw(dws, DW_SPI_CTRL0) != cr0)
|
if (dws->dma_mapped) {
|
||||||
dw_writew(dws, DW_SPI_CTRL0, cr0);
|
ret = dws->dma_ops->dma_transfer(dws, transfer);
|
||||||
|
if (ret < 0)
|
||||||
spi_set_clk(dws, clk_div ? clk_div : chip->clk_div);
|
return ret;
|
||||||
spi_chip_sel(dws, spi, 1);
|
|
||||||
|
|
||||||
/* Set the interrupt mask, for poll mode just disable all int */
|
|
||||||
spi_mask_intr(dws, 0xff);
|
|
||||||
if (imask)
|
|
||||||
spi_umask_intr(dws, imask);
|
|
||||||
if (txint_level)
|
|
||||||
dw_writew(dws, DW_SPI_TXFLTR, txint_level);
|
|
||||||
|
|
||||||
spi_enable_chip(dws, 1);
|
|
||||||
if (cs_change)
|
|
||||||
dws->prev_chip = chip;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (dws->dma_mapped)
|
|
||||||
dws->dma_ops->dma_transfer(dws, cs_change);
|
|
||||||
|
|
||||||
if (chip->poll_mode)
|
if (chip->poll_mode)
|
||||||
poll_transfer(dws);
|
return poll_transfer(dws);
|
||||||
|
|
||||||
return;
|
return 1;
|
||||||
|
|
||||||
early_exit:
|
|
||||||
giveback(dws);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int dw_spi_transfer_one_message(struct spi_master *master,
|
static void dw_spi_handle_err(struct spi_master *master,
|
||||||
struct spi_message *msg)
|
struct spi_message *msg)
|
||||||
{
|
{
|
||||||
struct dw_spi *dws = spi_master_get_devdata(master);
|
struct dw_spi *dws = spi_master_get_devdata(master);
|
||||||
|
|
||||||
dws->cur_msg = msg;
|
if (dws->dma_mapped)
|
||||||
/* Initial message state */
|
dws->dma_ops->dma_stop(dws);
|
||||||
dws->cur_msg->state = START_STATE;
|
|
||||||
dws->cur_transfer = list_entry(dws->cur_msg->transfers.next,
|
|
||||||
struct spi_transfer,
|
|
||||||
transfer_list);
|
|
||||||
dws->cur_chip = spi_get_ctldata(dws->cur_msg->spi);
|
|
||||||
|
|
||||||
/* Launch transfers */
|
spi_reset_chip(dws);
|
||||||
tasklet_schedule(&dws->pump_transfers);
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This may be called twice for each spi dev */
|
/* This may be called twice for each spi dev */
|
||||||
|
@ -561,8 +442,6 @@ static int dw_spi_setup(struct spi_device *spi)
|
||||||
|
|
||||||
chip->rx_threshold = 0;
|
chip->rx_threshold = 0;
|
||||||
chip->tx_threshold = 0;
|
chip->tx_threshold = 0;
|
||||||
|
|
||||||
chip->enable_dma = chip_info->enable_dma;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (spi->bits_per_word == 8) {
|
if (spi->bits_per_word == 8) {
|
||||||
|
@ -610,9 +489,7 @@ static void dw_spi_cleanup(struct spi_device *spi)
|
||||||
/* Restart the controller, disable all interrupts, clean rx fifo */
|
/* Restart the controller, disable all interrupts, clean rx fifo */
|
||||||
static void spi_hw_init(struct device *dev, struct dw_spi *dws)
|
static void spi_hw_init(struct device *dev, struct dw_spi *dws)
|
||||||
{
|
{
|
||||||
spi_enable_chip(dws, 0);
|
spi_reset_chip(dws);
|
||||||
spi_mask_intr(dws, 0xff);
|
|
||||||
spi_enable_chip(dws, 1);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Try to detect the FIFO depth if not set by interface driver,
|
* Try to detect the FIFO depth if not set by interface driver,
|
||||||
|
@ -622,11 +499,11 @@ static void spi_hw_init(struct device *dev, struct dw_spi *dws)
|
||||||
u32 fifo;
|
u32 fifo;
|
||||||
|
|
||||||
for (fifo = 1; fifo < 256; fifo++) {
|
for (fifo = 1; fifo < 256; fifo++) {
|
||||||
dw_writew(dws, DW_SPI_TXFLTR, fifo);
|
dw_writel(dws, DW_SPI_TXFLTR, fifo);
|
||||||
if (fifo != dw_readw(dws, DW_SPI_TXFLTR))
|
if (fifo != dw_readl(dws, DW_SPI_TXFLTR))
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
dw_writew(dws, DW_SPI_TXFLTR, 0);
|
dw_writel(dws, DW_SPI_TXFLTR, 0);
|
||||||
|
|
||||||
dws->fifo_len = (fifo == 1) ? 0 : fifo;
|
dws->fifo_len = (fifo == 1) ? 0 : fifo;
|
||||||
dev_dbg(dev, "Detected FIFO size: %u bytes\n", dws->fifo_len);
|
dev_dbg(dev, "Detected FIFO size: %u bytes\n", dws->fifo_len);
|
||||||
|
@ -646,13 +523,12 @@ int dw_spi_add_host(struct device *dev, struct dw_spi *dws)
|
||||||
|
|
||||||
dws->master = master;
|
dws->master = master;
|
||||||
dws->type = SSI_MOTO_SPI;
|
dws->type = SSI_MOTO_SPI;
|
||||||
dws->prev_chip = NULL;
|
|
||||||
dws->dma_inited = 0;
|
dws->dma_inited = 0;
|
||||||
dws->dma_addr = (dma_addr_t)(dws->paddr + 0x60);
|
dws->dma_addr = (dma_addr_t)(dws->paddr + 0x60);
|
||||||
snprintf(dws->name, sizeof(dws->name), "dw_spi%d", dws->bus_num);
|
snprintf(dws->name, sizeof(dws->name), "dw_spi%d", dws->bus_num);
|
||||||
|
|
||||||
ret = devm_request_irq(dev, dws->irq, dw_spi_irq, IRQF_SHARED,
|
ret = devm_request_irq(dev, dws->irq, dw_spi_irq, IRQF_SHARED,
|
||||||
dws->name, dws);
|
dws->name, master);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
dev_err(&master->dev, "can not get IRQ\n");
|
dev_err(&master->dev, "can not get IRQ\n");
|
||||||
goto err_free_master;
|
goto err_free_master;
|
||||||
|
@ -664,7 +540,9 @@ int dw_spi_add_host(struct device *dev, struct dw_spi *dws)
|
||||||
master->num_chipselect = dws->num_cs;
|
master->num_chipselect = dws->num_cs;
|
||||||
master->setup = dw_spi_setup;
|
master->setup = dw_spi_setup;
|
||||||
master->cleanup = dw_spi_cleanup;
|
master->cleanup = dw_spi_cleanup;
|
||||||
master->transfer_one_message = dw_spi_transfer_one_message;
|
master->set_cs = dw_spi_set_cs;
|
||||||
|
master->transfer_one = dw_spi_transfer_one;
|
||||||
|
master->handle_err = dw_spi_handle_err;
|
||||||
master->max_speed_hz = dws->max_freq;
|
master->max_speed_hz = dws->max_freq;
|
||||||
master->dev.of_node = dev->of_node;
|
master->dev.of_node = dev->of_node;
|
||||||
|
|
||||||
|
@ -676,11 +554,11 @@ int dw_spi_add_host(struct device *dev, struct dw_spi *dws)
|
||||||
if (ret) {
|
if (ret) {
|
||||||
dev_warn(dev, "DMA init failed\n");
|
dev_warn(dev, "DMA init failed\n");
|
||||||
dws->dma_inited = 0;
|
dws->dma_inited = 0;
|
||||||
|
} else {
|
||||||
|
master->can_dma = dws->dma_ops->can_dma;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
tasklet_init(&dws->pump_transfers, pump_transfers, (unsigned long)dws);
|
|
||||||
|
|
||||||
spi_master_set_devdata(master, dws);
|
spi_master_set_devdata(master, dws);
|
||||||
ret = devm_spi_register_master(dev, master);
|
ret = devm_spi_register_master(dev, master);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
|
|
|
@ -91,12 +91,15 @@ struct dw_spi;
|
||||||
struct dw_spi_dma_ops {
|
struct dw_spi_dma_ops {
|
||||||
int (*dma_init)(struct dw_spi *dws);
|
int (*dma_init)(struct dw_spi *dws);
|
||||||
void (*dma_exit)(struct dw_spi *dws);
|
void (*dma_exit)(struct dw_spi *dws);
|
||||||
int (*dma_transfer)(struct dw_spi *dws, int cs_change);
|
int (*dma_setup)(struct dw_spi *dws, struct spi_transfer *xfer);
|
||||||
|
bool (*can_dma)(struct spi_master *master, struct spi_device *spi,
|
||||||
|
struct spi_transfer *xfer);
|
||||||
|
int (*dma_transfer)(struct dw_spi *dws, struct spi_transfer *xfer);
|
||||||
|
void (*dma_stop)(struct dw_spi *dws);
|
||||||
};
|
};
|
||||||
|
|
||||||
struct dw_spi {
|
struct dw_spi {
|
||||||
struct spi_master *master;
|
struct spi_master *master;
|
||||||
struct spi_device *cur_dev;
|
|
||||||
enum dw_ssi_type type;
|
enum dw_ssi_type type;
|
||||||
char name[16];
|
char name[16];
|
||||||
|
|
||||||
|
@ -109,41 +112,26 @@ struct dw_spi {
|
||||||
u16 bus_num;
|
u16 bus_num;
|
||||||
u16 num_cs; /* supported slave numbers */
|
u16 num_cs; /* supported slave numbers */
|
||||||
|
|
||||||
/* Message Transfer pump */
|
|
||||||
struct tasklet_struct pump_transfers;
|
|
||||||
|
|
||||||
/* Current message transfer state info */
|
/* Current message transfer state info */
|
||||||
struct spi_message *cur_msg;
|
|
||||||
struct spi_transfer *cur_transfer;
|
|
||||||
struct chip_data *cur_chip;
|
|
||||||
struct chip_data *prev_chip;
|
|
||||||
size_t len;
|
size_t len;
|
||||||
void *tx;
|
void *tx;
|
||||||
void *tx_end;
|
void *tx_end;
|
||||||
void *rx;
|
void *rx;
|
||||||
void *rx_end;
|
void *rx_end;
|
||||||
int dma_mapped;
|
int dma_mapped;
|
||||||
dma_addr_t rx_dma;
|
|
||||||
dma_addr_t tx_dma;
|
|
||||||
size_t rx_map_len;
|
|
||||||
size_t tx_map_len;
|
|
||||||
u8 n_bytes; /* current is a 1/2 bytes op */
|
u8 n_bytes; /* current is a 1/2 bytes op */
|
||||||
u8 max_bits_per_word; /* maxim is 16b */
|
|
||||||
u32 dma_width;
|
u32 dma_width;
|
||||||
irqreturn_t (*transfer_handler)(struct dw_spi *dws);
|
irqreturn_t (*transfer_handler)(struct dw_spi *dws);
|
||||||
void (*cs_control)(u32 command);
|
|
||||||
|
|
||||||
/* Dma info */
|
/* DMA info */
|
||||||
int dma_inited;
|
int dma_inited;
|
||||||
struct dma_chan *txchan;
|
struct dma_chan *txchan;
|
||||||
struct scatterlist tx_sgl;
|
|
||||||
struct dma_chan *rxchan;
|
struct dma_chan *rxchan;
|
||||||
struct scatterlist rx_sgl;
|
|
||||||
unsigned long dma_chan_busy;
|
unsigned long dma_chan_busy;
|
||||||
struct device *dma_dev;
|
|
||||||
dma_addr_t dma_addr; /* phy address of the Data register */
|
dma_addr_t dma_addr; /* phy address of the Data register */
|
||||||
struct dw_spi_dma_ops *dma_ops;
|
struct dw_spi_dma_ops *dma_ops;
|
||||||
void *dma_priv; /* platform relate info */
|
void *dma_tx;
|
||||||
|
void *dma_rx;
|
||||||
|
|
||||||
/* Bus interface info */
|
/* Bus interface info */
|
||||||
void *priv;
|
void *priv;
|
||||||
|
@ -162,16 +150,6 @@ static inline void dw_writel(struct dw_spi *dws, u32 offset, u32 val)
|
||||||
__raw_writel(val, dws->regs + offset);
|
__raw_writel(val, dws->regs + offset);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline u16 dw_readw(struct dw_spi *dws, u32 offset)
|
|
||||||
{
|
|
||||||
return __raw_readw(dws->regs + offset);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline void dw_writew(struct dw_spi *dws, u32 offset, u16 val)
|
|
||||||
{
|
|
||||||
__raw_writew(val, dws->regs + offset);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline void spi_enable_chip(struct dw_spi *dws, int enable)
|
static inline void spi_enable_chip(struct dw_spi *dws, int enable)
|
||||||
{
|
{
|
||||||
dw_writel(dws, DW_SPI_SSIENR, (enable ? 1 : 0));
|
dw_writel(dws, DW_SPI_SSIENR, (enable ? 1 : 0));
|
||||||
|
@ -182,22 +160,6 @@ static inline void spi_set_clk(struct dw_spi *dws, u16 div)
|
||||||
dw_writel(dws, DW_SPI_BAUDR, div);
|
dw_writel(dws, DW_SPI_BAUDR, div);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void spi_chip_sel(struct dw_spi *dws, struct spi_device *spi,
|
|
||||||
int active)
|
|
||||||
{
|
|
||||||
u16 cs = spi->chip_select;
|
|
||||||
int gpio_val = active ? (spi->mode & SPI_CS_HIGH) :
|
|
||||||
!(spi->mode & SPI_CS_HIGH);
|
|
||||||
|
|
||||||
if (dws->cs_control)
|
|
||||||
dws->cs_control(active);
|
|
||||||
if (gpio_is_valid(spi->cs_gpio))
|
|
||||||
gpio_set_value(spi->cs_gpio, gpio_val);
|
|
||||||
|
|
||||||
if (active)
|
|
||||||
dw_writel(dws, DW_SPI_SER, 1 << cs);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Disable IRQ bits */
|
/* Disable IRQ bits */
|
||||||
static inline void spi_mask_intr(struct dw_spi *dws, u32 mask)
|
static inline void spi_mask_intr(struct dw_spi *dws, u32 mask)
|
||||||
{
|
{
|
||||||
|
@ -216,16 +178,27 @@ static inline void spi_umask_intr(struct dw_spi *dws, u32 mask)
|
||||||
dw_writel(dws, DW_SPI_IMR, new_mask);
|
dw_writel(dws, DW_SPI_IMR, new_mask);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This does disable the SPI controller, interrupts, and re-enable the
|
||||||
|
* controller back. Transmit and receive FIFO buffers are cleared when the
|
||||||
|
* device is disabled.
|
||||||
|
*/
|
||||||
|
static inline void spi_reset_chip(struct dw_spi *dws)
|
||||||
|
{
|
||||||
|
spi_enable_chip(dws, 0);
|
||||||
|
spi_mask_intr(dws, 0xff);
|
||||||
|
spi_enable_chip(dws, 1);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Each SPI slave device to work with dw_api controller should
|
* Each SPI slave device to work with dw_api controller should
|
||||||
* has such a structure claiming its working mode (PIO/DMA etc),
|
* has such a structure claiming its working mode (poll or PIO/DMA),
|
||||||
* which can be save in the "controller_data" member of the
|
* which can be save in the "controller_data" member of the
|
||||||
* struct spi_device.
|
* struct spi_device.
|
||||||
*/
|
*/
|
||||||
struct dw_spi_chip {
|
struct dw_spi_chip {
|
||||||
u8 poll_mode; /* 1 for controller polling mode */
|
u8 poll_mode; /* 1 for controller polling mode */
|
||||||
u8 type; /* SPI/SSP/MicroWire */
|
u8 type; /* SPI/SSP/MicroWire */
|
||||||
u8 enable_dma;
|
|
||||||
void (*cs_control)(u32 command);
|
void (*cs_control)(u32 command);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -233,7 +206,6 @@ extern int dw_spi_add_host(struct device *dev, struct dw_spi *dws);
|
||||||
extern void dw_spi_remove_host(struct dw_spi *dws);
|
extern void dw_spi_remove_host(struct dw_spi *dws);
|
||||||
extern int dw_spi_suspend_host(struct dw_spi *dws);
|
extern int dw_spi_suspend_host(struct dw_spi *dws);
|
||||||
extern int dw_spi_resume_host(struct dw_spi *dws);
|
extern int dw_spi_resume_host(struct dw_spi *dws);
|
||||||
extern void dw_spi_xfer_done(struct dw_spi *dws);
|
|
||||||
|
|
||||||
/* platform related setup */
|
/* platform related setup */
|
||||||
extern int dw_spi_mid_init(struct dw_spi *dws); /* Intel MID platforms */
|
extern int dw_spi_mid_init(struct dw_spi *dws); /* Intel MID platforms */
|
||||||
|
|
|
@ -850,7 +850,7 @@ static int spi_transfer_one_message(struct spi_master *master,
|
||||||
if (msg->status == -EINPROGRESS)
|
if (msg->status == -EINPROGRESS)
|
||||||
msg->status = ret;
|
msg->status = ret;
|
||||||
|
|
||||||
if (msg->status)
|
if (msg->status && master->handle_err)
|
||||||
master->handle_err(master, msg);
|
master->handle_err(master, msg);
|
||||||
|
|
||||||
spi_finalize_current_message(master);
|
spi_finalize_current_message(master);
|
||||||
|
|
|
@ -1,76 +0,0 @@
|
||||||
/*
|
|
||||||
* intel_mid_dma.h - Intel MID DMA Drivers
|
|
||||||
*
|
|
||||||
* Copyright (C) 2008-10 Intel Corp
|
|
||||||
* Author: Vinod Koul <vinod.koul@intel.com>
|
|
||||||
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
||||||
*
|
|
||||||
* This program is free software; you can redistribute it and/or modify
|
|
||||||
* it under the terms of the GNU General Public License as published by
|
|
||||||
* the Free Software Foundation; version 2 of the License.
|
|
||||||
*
|
|
||||||
* This program is distributed in the hope that it will be useful, but
|
|
||||||
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
||||||
* General Public License for more details.
|
|
||||||
*
|
|
||||||
* You should have received a copy of the GNU General Public License along
|
|
||||||
* with this program; if not, write to the Free Software Foundation, Inc.,
|
|
||||||
* 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA.
|
|
||||||
*
|
|
||||||
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
||||||
*
|
|
||||||
*
|
|
||||||
*/
|
|
||||||
#ifndef __INTEL_MID_DMA_H__
|
|
||||||
#define __INTEL_MID_DMA_H__
|
|
||||||
|
|
||||||
#include <linux/dmaengine.h>
|
|
||||||
|
|
||||||
#define DMA_PREP_CIRCULAR_LIST (1 << 10)
|
|
||||||
|
|
||||||
/*DMA mode configurations*/
|
|
||||||
enum intel_mid_dma_mode {
|
|
||||||
LNW_DMA_PER_TO_MEM = 0, /*periphral to memory configuration*/
|
|
||||||
LNW_DMA_MEM_TO_PER, /*memory to periphral configuration*/
|
|
||||||
LNW_DMA_MEM_TO_MEM, /*mem to mem confg (testing only)*/
|
|
||||||
};
|
|
||||||
|
|
||||||
/*DMA handshaking*/
|
|
||||||
enum intel_mid_dma_hs_mode {
|
|
||||||
LNW_DMA_HW_HS = 0, /*HW Handshaking only*/
|
|
||||||
LNW_DMA_SW_HS = 1, /*SW Handshaking not recommended*/
|
|
||||||
};
|
|
||||||
|
|
||||||
/*Burst size configuration*/
|
|
||||||
enum intel_mid_dma_msize {
|
|
||||||
LNW_DMA_MSIZE_1 = 0x0,
|
|
||||||
LNW_DMA_MSIZE_4 = 0x1,
|
|
||||||
LNW_DMA_MSIZE_8 = 0x2,
|
|
||||||
LNW_DMA_MSIZE_16 = 0x3,
|
|
||||||
LNW_DMA_MSIZE_32 = 0x4,
|
|
||||||
LNW_DMA_MSIZE_64 = 0x5,
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* struct intel_mid_dma_slave - DMA slave structure
|
|
||||||
*
|
|
||||||
* @dirn: DMA trf direction
|
|
||||||
* @src_width: tx register width
|
|
||||||
* @dst_width: rx register width
|
|
||||||
* @hs_mode: HW/SW handshaking mode
|
|
||||||
* @cfg_mode: DMA data transfer mode (per-per/mem-per/mem-mem)
|
|
||||||
* @src_msize: Source DMA burst size
|
|
||||||
* @dst_msize: Dst DMA burst size
|
|
||||||
* @per_addr: Periphral address
|
|
||||||
* @device_instance: DMA peripheral device instance, we can have multiple
|
|
||||||
* peripheral device connected to single DMAC
|
|
||||||
*/
|
|
||||||
struct intel_mid_dma_slave {
|
|
||||||
enum intel_mid_dma_hs_mode hs_mode; /*handshaking*/
|
|
||||||
enum intel_mid_dma_mode cfg_mode; /*mode configuration*/
|
|
||||||
unsigned int device_instance; /*0, 1 for periphral instance*/
|
|
||||||
struct dma_slave_config dma_slave;
|
|
||||||
};
|
|
||||||
|
|
||||||
#endif /*__INTEL_MID_DMA_H__*/
|
|
|
@ -294,7 +294,7 @@ static inline void spi_unregister_driver(struct spi_driver *sdrv)
|
||||||
* transfer_one_message are mutually exclusive; when both
|
* transfer_one_message are mutually exclusive; when both
|
||||||
* are set, the generic subsystem does not call your
|
* are set, the generic subsystem does not call your
|
||||||
* transfer_one callback.
|
* transfer_one callback.
|
||||||
* @handle_err: the subsystem calls the driver to handle and error that occurs
|
* @handle_err: the subsystem calls the driver to handle an error that occurs
|
||||||
* in the generic implementation of transfer_one_message().
|
* in the generic implementation of transfer_one_message().
|
||||||
* @unprepare_message: undo any work done by prepare_message().
|
* @unprepare_message: undo any work done by prepare_message().
|
||||||
* @cs_gpios: Array of GPIOs to use as chip select lines; one per CS
|
* @cs_gpios: Array of GPIOs to use as chip select lines; one per CS
|
||||||
|
|
Loading…
Reference in New Issue