2 * Copyright (C) ST-Ericsson SA 2007-2010
3 * Author: Per Forlin <per.forlin@stericsson.com> for ST-Ericsson
4 * Author: Jonas Aaberg <jonas.aberg@stericsson.com> for ST-Ericsson
5 * License terms: GNU General Public License (GPL) version 2
8 #include <linux/kernel.h>
9 #include <linux/platform_data/dma-ste-dma40.h>
11 #include "ste_dma40_ll.h"
13 u8
d40_width_to_bits(enum dma_slave_buswidth width
)
15 if (width
== DMA_SLAVE_BUSWIDTH_1_BYTE
)
16 return STEDMA40_ESIZE_8_BIT
;
17 else if (width
== DMA_SLAVE_BUSWIDTH_2_BYTES
)
18 return STEDMA40_ESIZE_16_BIT
;
19 else if (width
== DMA_SLAVE_BUSWIDTH_8_BYTES
)
20 return STEDMA40_ESIZE_64_BIT
;
22 return STEDMA40_ESIZE_32_BIT
;
25 /* Sets up proper LCSP1 and LCSP3 register for a logical channel */
26 void d40_log_cfg(struct stedma40_chan_cfg
*cfg
,
27 u32
*lcsp1
, u32
*lcsp3
)
32 /* src is mem? -> increase address pos */
33 if (cfg
->dir
== DMA_MEM_TO_DEV
||
34 cfg
->dir
== DMA_MEM_TO_MEM
)
35 l1
|= BIT(D40_MEM_LCSP1_SCFG_INCR_POS
);
37 /* dst is mem? -> increase address pos */
38 if (cfg
->dir
== DMA_DEV_TO_MEM
||
39 cfg
->dir
== DMA_MEM_TO_MEM
)
40 l3
|= BIT(D40_MEM_LCSP3_DCFG_INCR_POS
);
42 /* src is hw? -> master port 1 */
43 if (cfg
->dir
== DMA_DEV_TO_MEM
||
44 cfg
->dir
== DMA_DEV_TO_DEV
)
45 l1
|= BIT(D40_MEM_LCSP1_SCFG_MST_POS
);
47 /* dst is hw? -> master port 1 */
48 if (cfg
->dir
== DMA_MEM_TO_DEV
||
49 cfg
->dir
== DMA_DEV_TO_DEV
)
50 l3
|= BIT(D40_MEM_LCSP3_DCFG_MST_POS
);
52 l3
|= BIT(D40_MEM_LCSP3_DCFG_EIM_POS
);
53 l3
|= cfg
->dst_info
.psize
<< D40_MEM_LCSP3_DCFG_PSIZE_POS
;
54 l3
|= d40_width_to_bits(cfg
->dst_info
.data_width
)
55 << D40_MEM_LCSP3_DCFG_ESIZE_POS
;
57 l1
|= BIT(D40_MEM_LCSP1_SCFG_EIM_POS
);
58 l1
|= cfg
->src_info
.psize
<< D40_MEM_LCSP1_SCFG_PSIZE_POS
;
59 l1
|= d40_width_to_bits(cfg
->src_info
.data_width
)
60 << D40_MEM_LCSP1_SCFG_ESIZE_POS
;
67 void d40_phy_cfg(struct stedma40_chan_cfg
*cfg
, u32
*src_cfg
, u32
*dst_cfg
)
72 if ((cfg
->dir
== DMA_DEV_TO_MEM
) ||
73 (cfg
->dir
== DMA_DEV_TO_DEV
)) {
74 /* Set master port to 1 */
75 src
|= BIT(D40_SREG_CFG_MST_POS
);
76 src
|= D40_TYPE_TO_EVENT(cfg
->dev_type
);
78 if (cfg
->src_info
.flow_ctrl
== STEDMA40_NO_FLOW_CTRL
)
79 src
|= BIT(D40_SREG_CFG_PHY_TM_POS
);
81 src
|= 3 << D40_SREG_CFG_PHY_TM_POS
;
83 if ((cfg
->dir
== DMA_MEM_TO_DEV
) ||
84 (cfg
->dir
== DMA_DEV_TO_DEV
)) {
85 /* Set master port to 1 */
86 dst
|= BIT(D40_SREG_CFG_MST_POS
);
87 dst
|= D40_TYPE_TO_EVENT(cfg
->dev_type
);
89 if (cfg
->dst_info
.flow_ctrl
== STEDMA40_NO_FLOW_CTRL
)
90 dst
|= BIT(D40_SREG_CFG_PHY_TM_POS
);
92 dst
|= 3 << D40_SREG_CFG_PHY_TM_POS
;
94 /* Interrupt on end of transfer for destination */
95 dst
|= BIT(D40_SREG_CFG_TIM_POS
);
97 /* Generate interrupt on error */
98 src
|= BIT(D40_SREG_CFG_EIM_POS
);
99 dst
|= BIT(D40_SREG_CFG_EIM_POS
);
102 if (cfg
->src_info
.psize
!= STEDMA40_PSIZE_PHY_1
) {
103 src
|= BIT(D40_SREG_CFG_PHY_PEN_POS
);
104 src
|= cfg
->src_info
.psize
<< D40_SREG_CFG_PSIZE_POS
;
106 if (cfg
->dst_info
.psize
!= STEDMA40_PSIZE_PHY_1
) {
107 dst
|= BIT(D40_SREG_CFG_PHY_PEN_POS
);
108 dst
|= cfg
->dst_info
.psize
<< D40_SREG_CFG_PSIZE_POS
;
112 src
|= d40_width_to_bits(cfg
->src_info
.data_width
)
113 << D40_SREG_CFG_ESIZE_POS
;
114 dst
|= d40_width_to_bits(cfg
->dst_info
.data_width
)
115 << D40_SREG_CFG_ESIZE_POS
;
117 /* Set the priority bit to high for the physical channel */
118 if (cfg
->high_priority
) {
119 src
|= BIT(D40_SREG_CFG_PRI_POS
);
120 dst
|= BIT(D40_SREG_CFG_PRI_POS
);
123 if (cfg
->src_info
.big_endian
)
124 src
|= BIT(D40_SREG_CFG_LBE_POS
);
125 if (cfg
->dst_info
.big_endian
)
126 dst
|= BIT(D40_SREG_CFG_LBE_POS
);
132 static int d40_phy_fill_lli(struct d40_phy_lli
*lli
,
137 struct stedma40_half_channel_info
*info
,
140 bool addr_inc
= flags
& LLI_ADDR_INC
;
141 bool term_int
= flags
& LLI_TERM_INT
;
142 unsigned int data_width
= info
->data_width
;
143 int psize
= info
->psize
;
146 if (psize
== STEDMA40_PSIZE_PHY_1
)
149 num_elems
= 2 << psize
;
151 /* Must be aligned */
152 if (!IS_ALIGNED(data
, data_width
))
155 /* Transfer size can't be smaller than (num_elms * elem_size) */
156 if (data_size
< num_elems
* data_width
)
159 /* The number of elements. IE now many chunks */
160 lli
->reg_elt
= (data_size
/ data_width
) << D40_SREG_ELEM_PHY_ECNT_POS
;
163 * Distance to next element sized entry.
164 * Usually the size of the element unless you want gaps.
167 lli
->reg_elt
|= data_width
<< D40_SREG_ELEM_PHY_EIDX_POS
;
169 /* Where the data is */
171 lli
->reg_cfg
= reg_cfg
;
173 /* If this scatter list entry is the last one, no next link */
175 lli
->reg_lnk
= BIT(D40_SREG_LNK_PHY_TCP_POS
);
177 lli
->reg_lnk
= next_lli
;
179 /* Set/clear interrupt generation on this link item.*/
181 lli
->reg_cfg
|= BIT(D40_SREG_CFG_TIM_POS
);
183 lli
->reg_cfg
&= ~BIT(D40_SREG_CFG_TIM_POS
);
186 * Post link - D40_SREG_LNK_PHY_PRE_POS = 0
187 * Relink happens after transfer completion.
193 static int d40_seg_size(int size
, int data_width1
, int data_width2
)
195 u32 max_w
= max(data_width1
, data_width2
);
196 u32 min_w
= min(data_width1
, data_width2
);
197 u32 seg_max
= ALIGN(STEDMA40_MAX_SEG_SIZE
* min_w
, max_w
);
199 if (seg_max
> STEDMA40_MAX_SEG_SIZE
)
205 if (size
<= 2 * seg_max
)
206 return ALIGN(size
/ 2, max_w
);
211 static struct d40_phy_lli
*
212 d40_phy_buf_to_lli(struct d40_phy_lli
*lli
, dma_addr_t addr
, u32 size
,
213 dma_addr_t lli_phys
, dma_addr_t first_phys
, u32 reg_cfg
,
214 struct stedma40_half_channel_info
*info
,
215 struct stedma40_half_channel_info
*otherinfo
,
218 bool lastlink
= flags
& LLI_LAST_LINK
;
219 bool addr_inc
= flags
& LLI_ADDR_INC
;
220 bool term_int
= flags
& LLI_TERM_INT
;
221 bool cyclic
= flags
& LLI_CYCLIC
;
223 dma_addr_t next
= lli_phys
;
224 int size_rest
= size
;
228 * This piece may be split up based on d40_seg_size(); we only want the
229 * term int on the last part.
232 flags
&= ~LLI_TERM_INT
;
235 size_seg
= d40_seg_size(size_rest
, info
->data_width
,
236 otherinfo
->data_width
);
237 size_rest
-= size_seg
;
239 if (size_rest
== 0 && term_int
)
240 flags
|= LLI_TERM_INT
;
242 if (size_rest
== 0 && lastlink
)
243 next
= cyclic
? first_phys
: 0;
245 next
= ALIGN(next
+ sizeof(struct d40_phy_lli
),
248 err
= d40_phy_fill_lli(lli
, addr
, size_seg
, next
,
249 reg_cfg
, info
, flags
);
265 int d40_phy_sg_to_lli(struct scatterlist
*sg
,
268 struct d40_phy_lli
*lli_sg
,
271 struct stedma40_half_channel_info
*info
,
272 struct stedma40_half_channel_info
*otherinfo
,
277 struct scatterlist
*current_sg
= sg
;
278 struct d40_phy_lli
*lli
= lli_sg
;
279 dma_addr_t l_phys
= lli_phys
;
282 flags
|= LLI_ADDR_INC
;
284 for_each_sg(sg
, current_sg
, sg_len
, i
) {
285 dma_addr_t sg_addr
= sg_dma_address(current_sg
);
286 unsigned int len
= sg_dma_len(current_sg
);
287 dma_addr_t dst
= target
?: sg_addr
;
289 total_size
+= sg_dma_len(current_sg
);
292 flags
|= LLI_TERM_INT
| LLI_LAST_LINK
;
294 l_phys
= ALIGN(lli_phys
+ (lli
- lli_sg
) *
295 sizeof(struct d40_phy_lli
), D40_LLI_ALIGN
);
297 lli
= d40_phy_buf_to_lli(lli
, dst
, len
, l_phys
, lli_phys
,
298 reg_cfg
, info
, otherinfo
, flags
);
308 /* DMA logical lli operations */
310 static void d40_log_lli_link(struct d40_log_lli
*lli_dst
,
311 struct d40_log_lli
*lli_src
,
312 int next
, unsigned int flags
)
314 bool interrupt
= flags
& LLI_TERM_INT
;
318 if (next
!= -EINVAL
) {
324 lli_dst
->lcsp13
|= D40_MEM_LCSP1_SCFG_TIM_MASK
;
325 lli_dst
->lcsp13
|= D40_MEM_LCSP3_DTCP_MASK
;
328 lli_src
->lcsp13
= (lli_src
->lcsp13
& ~D40_MEM_LCSP1_SLOS_MASK
) |
329 (slos
<< D40_MEM_LCSP1_SLOS_POS
);
331 lli_dst
->lcsp13
= (lli_dst
->lcsp13
& ~D40_MEM_LCSP1_SLOS_MASK
) |
332 (dlos
<< D40_MEM_LCSP1_SLOS_POS
);
335 void d40_log_lli_lcpa_write(struct d40_log_lli_full
*lcpa
,
336 struct d40_log_lli
*lli_dst
,
337 struct d40_log_lli
*lli_src
,
338 int next
, unsigned int flags
)
340 d40_log_lli_link(lli_dst
, lli_src
, next
, flags
);
342 writel_relaxed(lli_src
->lcsp02
, &lcpa
[0].lcsp0
);
343 writel_relaxed(lli_src
->lcsp13
, &lcpa
[0].lcsp1
);
344 writel_relaxed(lli_dst
->lcsp02
, &lcpa
[0].lcsp2
);
345 writel_relaxed(lli_dst
->lcsp13
, &lcpa
[0].lcsp3
);
348 void d40_log_lli_lcla_write(struct d40_log_lli
*lcla
,
349 struct d40_log_lli
*lli_dst
,
350 struct d40_log_lli
*lli_src
,
351 int next
, unsigned int flags
)
353 d40_log_lli_link(lli_dst
, lli_src
, next
, flags
);
355 writel_relaxed(lli_src
->lcsp02
, &lcla
[0].lcsp02
);
356 writel_relaxed(lli_src
->lcsp13
, &lcla
[0].lcsp13
);
357 writel_relaxed(lli_dst
->lcsp02
, &lcla
[1].lcsp02
);
358 writel_relaxed(lli_dst
->lcsp13
, &lcla
[1].lcsp13
);
361 static void d40_log_fill_lli(struct d40_log_lli
*lli
,
362 dma_addr_t data
, u32 data_size
,
367 bool addr_inc
= flags
& LLI_ADDR_INC
;
369 lli
->lcsp13
= reg_cfg
;
371 /* The number of elements to transfer */
372 lli
->lcsp02
= ((data_size
/ data_width
) <<
373 D40_MEM_LCSP0_ECNT_POS
) & D40_MEM_LCSP0_ECNT_MASK
;
375 BUG_ON((data_size
/ data_width
) > STEDMA40_MAX_SEG_SIZE
);
377 /* 16 LSBs address of the current element */
378 lli
->lcsp02
|= data
& D40_MEM_LCSP0_SPTR_MASK
;
379 /* 16 MSBs address of the current element */
380 lli
->lcsp13
|= data
& D40_MEM_LCSP1_SPTR_MASK
;
383 lli
->lcsp13
|= D40_MEM_LCSP1_SCFG_INCR_MASK
;
387 static struct d40_log_lli
*d40_log_buf_to_lli(struct d40_log_lli
*lli_sg
,
390 u32 lcsp13
, /* src or dst*/
395 bool addr_inc
= flags
& LLI_ADDR_INC
;
396 struct d40_log_lli
*lli
= lli_sg
;
397 int size_rest
= size
;
401 size_seg
= d40_seg_size(size_rest
, data_width1
, data_width2
);
402 size_rest
-= size_seg
;
404 d40_log_fill_lli(lli
,
417 int d40_log_sg_to_lli(struct scatterlist
*sg
,
420 struct d40_log_lli
*lli_sg
,
421 u32 lcsp13
, /* src or dst*/
422 u32 data_width1
, u32 data_width2
)
425 struct scatterlist
*current_sg
= sg
;
427 struct d40_log_lli
*lli
= lli_sg
;
428 unsigned long flags
= 0;
431 flags
|= LLI_ADDR_INC
;
433 for_each_sg(sg
, current_sg
, sg_len
, i
) {
434 dma_addr_t sg_addr
= sg_dma_address(current_sg
);
435 unsigned int len
= sg_dma_len(current_sg
);
436 dma_addr_t addr
= dev_addr
?: sg_addr
;
438 total_size
+= sg_dma_len(current_sg
);
440 lli
= d40_log_buf_to_lli(lli
, addr
, len
,