Revert "genirq: Use irqd_get_trigger_type to compare the trigger type for shared...
[linux/fpc-iii.git] / drivers / dma / mv_xor.h
blob4fa0fe2e2f0896e0f2db8fe84e5d51abb9261fe8
1 /*
2 * Copyright (C) 2007, 2008, Marvell International Ltd.
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
11 * for more details.
13 * You should have received a copy of the GNU General Public License
14 * along with this program; if not, write to the Free Software Foundation,
15 * Inc., 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
18 #ifndef MV_XOR_H
19 #define MV_XOR_H
21 #include <linux/types.h>
22 #include <linux/io.h>
23 #include <linux/dmaengine.h>
24 #include <linux/interrupt.h>
26 #define MV_XOR_POOL_SIZE PAGE_SIZE
27 #define MV_XOR_SLOT_SIZE 64
28 #define MV_XOR_THRESHOLD 1
29 #define MV_XOR_MAX_CHANNELS 2
31 #define MV_XOR_MIN_BYTE_COUNT SZ_128
32 #define MV_XOR_MAX_BYTE_COUNT (SZ_16M - 1)
34 /* Values for the XOR_CONFIG register */
35 #define XOR_OPERATION_MODE_XOR 0
36 #define XOR_OPERATION_MODE_MEMCPY 2
37 #define XOR_DESCRIPTOR_SWAP BIT(14)
38 #define XOR_DESC_SUCCESS 0x40000000
40 #define XOR_DESC_DMA_OWNED BIT(31)
41 #define XOR_DESC_EOD_INT_EN BIT(31)
43 #define XOR_CURR_DESC(chan) (chan->mmr_high_base + 0x10 + (chan->idx * 4))
44 #define XOR_NEXT_DESC(chan) (chan->mmr_high_base + 0x00 + (chan->idx * 4))
45 #define XOR_BYTE_COUNT(chan) (chan->mmr_high_base + 0x20 + (chan->idx * 4))
46 #define XOR_DEST_POINTER(chan) (chan->mmr_high_base + 0xB0 + (chan->idx * 4))
47 #define XOR_BLOCK_SIZE(chan) (chan->mmr_high_base + 0xC0 + (chan->idx * 4))
48 #define XOR_INIT_VALUE_LOW(chan) (chan->mmr_high_base + 0xE0)
49 #define XOR_INIT_VALUE_HIGH(chan) (chan->mmr_high_base + 0xE4)
51 #define XOR_CONFIG(chan) (chan->mmr_base + 0x10 + (chan->idx * 4))
52 #define XOR_ACTIVATION(chan) (chan->mmr_base + 0x20 + (chan->idx * 4))
53 #define XOR_INTR_CAUSE(chan) (chan->mmr_base + 0x30)
54 #define XOR_INTR_MASK(chan) (chan->mmr_base + 0x40)
55 #define XOR_ERROR_CAUSE(chan) (chan->mmr_base + 0x50)
56 #define XOR_ERROR_ADDR(chan) (chan->mmr_base + 0x60)
58 #define XOR_INT_END_OF_DESC BIT(0)
59 #define XOR_INT_END_OF_CHAIN BIT(1)
60 #define XOR_INT_STOPPED BIT(2)
61 #define XOR_INT_PAUSED BIT(3)
62 #define XOR_INT_ERR_DECODE BIT(4)
63 #define XOR_INT_ERR_RDPROT BIT(5)
64 #define XOR_INT_ERR_WRPROT BIT(6)
65 #define XOR_INT_ERR_OWN BIT(7)
66 #define XOR_INT_ERR_PAR BIT(8)
67 #define XOR_INT_ERR_MBUS BIT(9)
69 #define XOR_INTR_ERRORS (XOR_INT_ERR_DECODE | XOR_INT_ERR_RDPROT | \
70 XOR_INT_ERR_WRPROT | XOR_INT_ERR_OWN | \
71 XOR_INT_ERR_PAR | XOR_INT_ERR_MBUS)
73 #define XOR_INTR_MASK_VALUE (XOR_INT_END_OF_DESC | XOR_INT_END_OF_CHAIN | \
74 XOR_INT_STOPPED | XOR_INTR_ERRORS)
76 #define WINDOW_BASE(w) (0x50 + ((w) << 2))
77 #define WINDOW_SIZE(w) (0x70 + ((w) << 2))
78 #define WINDOW_REMAP_HIGH(w) (0x90 + ((w) << 2))
79 #define WINDOW_BAR_ENABLE(chan) (0x40 + ((chan) << 2))
80 #define WINDOW_OVERRIDE_CTRL(chan) (0xA0 + ((chan) << 2))
82 struct mv_xor_device {
83 void __iomem *xor_base;
84 void __iomem *xor_high_base;
85 struct clk *clk;
86 struct mv_xor_chan *channels[MV_XOR_MAX_CHANNELS];
89 /**
90 * struct mv_xor_chan - internal representation of a XOR channel
91 * @pending: allows batching of hardware operations
92 * @lock: serializes enqueue/dequeue operations to the descriptors pool
93 * @mmr_base: memory mapped register base
94 * @idx: the index of the xor channel
95 * @chain: device chain view of the descriptors
96 * @completed_slots: slots completed by HW but still need to be acked
97 * @device: parent device
98 * @common: common dmaengine channel object members
99 * @last_used: place holder for allocation to continue from where it left off
100 * @all_slots: complete domain of slots usable by the channel
101 * @slots_allocated: records the actual size of the descriptor slot pool
102 * @irq_tasklet: bottom half where mv_xor_slot_cleanup runs
104 struct mv_xor_chan {
105 int pending;
106 spinlock_t lock; /* protects the descriptor slot pool */
107 void __iomem *mmr_base;
108 void __iomem *mmr_high_base;
109 unsigned int idx;
110 int irq;
111 enum dma_transaction_type current_type;
112 struct list_head chain;
113 struct list_head completed_slots;
114 dma_addr_t dma_desc_pool;
115 void *dma_desc_pool_virt;
116 size_t pool_size;
117 struct dma_device dmadev;
118 struct dma_chan dmachan;
119 struct mv_xor_desc_slot *last_used;
120 struct list_head all_slots;
121 int slots_allocated;
122 struct tasklet_struct irq_tasklet;
123 char dummy_src[MV_XOR_MIN_BYTE_COUNT];
124 char dummy_dst[MV_XOR_MIN_BYTE_COUNT];
125 dma_addr_t dummy_src_addr, dummy_dst_addr;
129 * struct mv_xor_desc_slot - software descriptor
130 * @slot_node: node on the mv_xor_chan.all_slots list
131 * @chain_node: node on the mv_xor_chan.chain list
132 * @completed_node: node on the mv_xor_chan.completed_slots list
133 * @hw_desc: virtual address of the hardware descriptor chain
134 * @phys: hardware address of the hardware descriptor chain
135 * @slot_used: slot in use or not
136 * @idx: pool index
137 * @tx_list: list of slots that make up a multi-descriptor transaction
138 * @async_tx: support for the async_tx api
140 struct mv_xor_desc_slot {
141 struct list_head slot_node;
142 struct list_head chain_node;
143 struct list_head completed_node;
144 enum dma_transaction_type type;
145 void *hw_desc;
146 u16 slot_used;
147 u16 idx;
148 struct dma_async_tx_descriptor async_tx;
152 * This structure describes XOR descriptor size 64bytes. The
153 * mv_phy_src_idx() macro must be used when indexing the values of the
154 * phy_src_addr[] array. This is due to the fact that the 'descriptor
155 * swap' feature, used on big endian systems, swaps descriptors data
156 * within blocks of 8 bytes. So two consecutive values of the
157 * phy_src_addr[] array are actually swapped in big-endian, which
158 * explains the different mv_phy_src_idx() implementation.
160 #if defined(__LITTLE_ENDIAN)
161 struct mv_xor_desc {
162 u32 status; /* descriptor execution status */
163 u32 crc32_result; /* result of CRC-32 calculation */
164 u32 desc_command; /* type of operation to be carried out */
165 u32 phy_next_desc; /* next descriptor address pointer */
166 u32 byte_count; /* size of src/dst blocks in bytes */
167 u32 phy_dest_addr; /* destination block address */
168 u32 phy_src_addr[8]; /* source block addresses */
169 u32 reserved0;
170 u32 reserved1;
172 #define mv_phy_src_idx(src_idx) (src_idx)
173 #else
174 struct mv_xor_desc {
175 u32 crc32_result; /* result of CRC-32 calculation */
176 u32 status; /* descriptor execution status */
177 u32 phy_next_desc; /* next descriptor address pointer */
178 u32 desc_command; /* type of operation to be carried out */
179 u32 phy_dest_addr; /* destination block address */
180 u32 byte_count; /* size of src/dst blocks in bytes */
181 u32 phy_src_addr[8]; /* source block addresses */
182 u32 reserved1;
183 u32 reserved0;
185 #define mv_phy_src_idx(src_idx) (src_idx ^ 1)
186 #endif
188 #define to_mv_sw_desc(addr_hw_desc) \
189 container_of(addr_hw_desc, struct mv_xor_desc_slot, hw_desc)
191 #define mv_hw_desc_slot_idx(hw_desc, idx) \
192 ((void *)(((unsigned long)hw_desc) + ((idx) << 5)))
194 #endif