1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (c) 2019 HiSilicon Limited. */
3 #include <linux/align.h>
4 #include <linux/dma-mapping.h>
5 #include <linux/hisi_acc_qm.h>
6 #include <linux/module.h>
7 #include <linux/slab.h>
9 #define HISI_ACC_SGL_SGE_NR_MIN 1
10 #define HISI_ACC_SGL_NR_MAX 256
11 #define HISI_ACC_SGL_ALIGN_SIZE 64
12 #define HISI_ACC_MEM_BLOCK_NR 5
23 /* use default sgl head size 64B */
24 struct hisi_acc_hw_sgl
{
26 __le16 entry_sum_in_chain
;
27 __le16 entry_sum_in_sgl
;
28 __le16 entry_length_in_sgl
;
31 struct hisi_acc_hw_sgl
*next
;
32 struct acc_hw_sge sge_entries
[];
35 struct hisi_acc_sgl_pool
{
37 struct hisi_acc_hw_sgl
*sgl
;
40 } mem_block
[HISI_ACC_MEM_BLOCK_NR
];
41 u32 sgl_num_per_block
;
49 * hisi_acc_create_sgl_pool() - Create a hw sgl pool.
50 * @dev: The device which hw sgl pool belongs to.
51 * @count: Count of hisi_acc_hw_sgl in pool.
52 * @sge_nr: The count of sge in hw_sgl
54 * This function creates a hw sgl pool, after this user can get hw sgl memory
57 struct hisi_acc_sgl_pool
*hisi_acc_create_sgl_pool(struct device
*dev
,
58 u32 count
, u32 sge_nr
)
60 u32 sgl_size
, block_size
, sgl_num_per_block
, block_num
, remain_sgl
;
61 struct hisi_acc_sgl_pool
*pool
;
62 struct mem_block
*block
;
65 if (!dev
|| !count
|| !sge_nr
|| sge_nr
> HISI_ACC_SGL_SGE_NR_MAX
)
66 return ERR_PTR(-EINVAL
);
68 sgl_size
= ALIGN(sizeof(struct acc_hw_sge
) * sge_nr
+
69 sizeof(struct hisi_acc_hw_sgl
),
70 HISI_ACC_SGL_ALIGN_SIZE
);
73 * the pool may allocate a block of memory of size PAGE_SIZE * 2^MAX_PAGE_ORDER,
74 * block size may exceed 2^31 on ia64, so the max of block size is 2^31
76 block_size
= 1 << (PAGE_SHIFT
+ MAX_PAGE_ORDER
< 32 ?
77 PAGE_SHIFT
+ MAX_PAGE_ORDER
: 31);
78 sgl_num_per_block
= block_size
/ sgl_size
;
79 block_num
= count
/ sgl_num_per_block
;
80 remain_sgl
= count
% sgl_num_per_block
;
82 if ((!remain_sgl
&& block_num
> HISI_ACC_MEM_BLOCK_NR
) ||
83 (remain_sgl
> 0 && block_num
> HISI_ACC_MEM_BLOCK_NR
- 1))
84 return ERR_PTR(-EINVAL
);
86 pool
= kzalloc(sizeof(*pool
), GFP_KERNEL
);
88 return ERR_PTR(-ENOMEM
);
89 block
= pool
->mem_block
;
91 for (i
= 0; i
< block_num
; i
++) {
92 block
[i
].sgl
= dma_alloc_coherent(dev
, block_size
,
96 dev_err(dev
, "Fail to allocate hw SG buffer!\n");
100 block
[i
].size
= block_size
;
103 if (remain_sgl
> 0) {
104 block
[i
].sgl
= dma_alloc_coherent(dev
, remain_sgl
* sgl_size
,
108 dev_err(dev
, "Fail to allocate remained hw SG buffer!\n");
112 block
[i
].size
= remain_sgl
* sgl_size
;
115 pool
->sgl_num_per_block
= sgl_num_per_block
;
116 pool
->block_num
= remain_sgl
? block_num
+ 1 : block_num
;
118 pool
->sgl_size
= sgl_size
;
119 pool
->sge_nr
= sge_nr
;
124 for (j
= 0; j
< i
; j
++)
125 dma_free_coherent(dev
, block_size
, block
[j
].sgl
,
128 kfree_sensitive(pool
);
129 return ERR_PTR(-ENOMEM
);
131 EXPORT_SYMBOL_GPL(hisi_acc_create_sgl_pool
);
134 * hisi_acc_free_sgl_pool() - Free a hw sgl pool.
135 * @dev: The device which hw sgl pool belongs to.
136 * @pool: Pointer of pool.
138 * This function frees memory of a hw sgl pool.
140 void hisi_acc_free_sgl_pool(struct device
*dev
, struct hisi_acc_sgl_pool
*pool
)
142 struct mem_block
*block
;
148 block
= pool
->mem_block
;
150 for (i
= 0; i
< pool
->block_num
; i
++)
151 dma_free_coherent(dev
, block
[i
].size
, block
[i
].sgl
,
156 EXPORT_SYMBOL_GPL(hisi_acc_free_sgl_pool
);
158 static struct hisi_acc_hw_sgl
*acc_get_sgl(struct hisi_acc_sgl_pool
*pool
,
159 u32 index
, dma_addr_t
*hw_sgl_dma
)
161 struct mem_block
*block
;
162 u32 block_index
, offset
;
164 block
= pool
->mem_block
;
165 block_index
= index
/ pool
->sgl_num_per_block
;
166 offset
= index
% pool
->sgl_num_per_block
;
168 *hw_sgl_dma
= block
[block_index
].sgl_dma
+ pool
->sgl_size
* offset
;
169 return (void *)block
[block_index
].sgl
+ pool
->sgl_size
* offset
;
172 static void sg_map_to_hw_sg(struct scatterlist
*sgl
,
173 struct acc_hw_sge
*hw_sge
)
175 hw_sge
->buf
= sg_dma_address(sgl
);
176 hw_sge
->len
= cpu_to_le32(sg_dma_len(sgl
));
177 hw_sge
->page_ctrl
= sg_virt(sgl
);
180 static void inc_hw_sgl_sge(struct hisi_acc_hw_sgl
*hw_sgl
)
182 u16 var
= le16_to_cpu(hw_sgl
->entry_sum_in_sgl
);
185 hw_sgl
->entry_sum_in_sgl
= cpu_to_le16(var
);
188 static void update_hw_sgl_sum_sge(struct hisi_acc_hw_sgl
*hw_sgl
, u16 sum
)
190 hw_sgl
->entry_sum_in_chain
= cpu_to_le16(sum
);
193 static void clear_hw_sgl_sge(struct hisi_acc_hw_sgl
*hw_sgl
)
195 struct acc_hw_sge
*hw_sge
= hw_sgl
->sge_entries
;
196 u16 entry_sum
= le16_to_cpu(hw_sgl
->entry_sum_in_sgl
);
199 for (i
= 0; i
< entry_sum
; i
++) {
200 hw_sge
[i
].page_ctrl
= NULL
;
207 * hisi_acc_sg_buf_map_to_hw_sgl - Map a scatterlist to a hw sgl.
208 * @dev: The device which hw sgl belongs to.
209 * @sgl: Scatterlist which will be mapped to hw sgl.
210 * @pool: Pool which hw sgl memory will be allocated in.
211 * @index: Index of hisi_acc_hw_sgl in pool.
212 * @hw_sgl_dma: The dma address of allocated hw sgl.
214 * This function builds hw sgl according input sgl, user can use hw_sgl_dma
215 * as src/dst in its BD. Only support single hw sgl currently.
217 struct hisi_acc_hw_sgl
*
218 hisi_acc_sg_buf_map_to_hw_sgl(struct device
*dev
,
219 struct scatterlist
*sgl
,
220 struct hisi_acc_sgl_pool
*pool
,
221 u32 index
, dma_addr_t
*hw_sgl_dma
)
223 struct hisi_acc_hw_sgl
*curr_hw_sgl
;
224 unsigned int i
, sg_n_mapped
;
225 dma_addr_t curr_sgl_dma
= 0;
226 struct acc_hw_sge
*curr_hw_sge
;
227 struct scatterlist
*sg
;
230 if (!dev
|| !sgl
|| !pool
|| !hw_sgl_dma
|| index
>= pool
->count
)
231 return ERR_PTR(-EINVAL
);
233 sg_n
= sg_nents(sgl
);
235 sg_n_mapped
= dma_map_sg(dev
, sgl
, sg_n
, DMA_BIDIRECTIONAL
);
237 dev_err(dev
, "DMA mapping for SG error!\n");
238 return ERR_PTR(-EINVAL
);
241 if (sg_n_mapped
> pool
->sge_nr
) {
242 dev_err(dev
, "the number of entries in input scatterlist is bigger than SGL pool setting.\n");
247 curr_hw_sgl
= acc_get_sgl(pool
, index
, &curr_sgl_dma
);
248 if (IS_ERR(curr_hw_sgl
)) {
249 dev_err(dev
, "Get SGL error!\n");
253 curr_hw_sgl
->entry_length_in_sgl
= cpu_to_le16(pool
->sge_nr
);
254 curr_hw_sge
= curr_hw_sgl
->sge_entries
;
256 for_each_sg(sgl
, sg
, sg_n_mapped
, i
) {
257 sg_map_to_hw_sg(sg
, curr_hw_sge
);
258 inc_hw_sgl_sge(curr_hw_sgl
);
262 update_hw_sgl_sum_sge(curr_hw_sgl
, pool
->sge_nr
);
263 *hw_sgl_dma
= curr_sgl_dma
;
268 dma_unmap_sg(dev
, sgl
, sg_n
, DMA_BIDIRECTIONAL
);
272 EXPORT_SYMBOL_GPL(hisi_acc_sg_buf_map_to_hw_sgl
);
275 * hisi_acc_sg_buf_unmap() - Unmap allocated hw sgl.
276 * @dev: The device which hw sgl belongs to.
277 * @sgl: Related scatterlist.
278 * @hw_sgl: Virtual address of hw sgl.
280 * This function unmaps allocated hw sgl.
282 void hisi_acc_sg_buf_unmap(struct device
*dev
, struct scatterlist
*sgl
,
283 struct hisi_acc_hw_sgl
*hw_sgl
)
285 if (!dev
|| !sgl
|| !hw_sgl
)
288 dma_unmap_sg(dev
, sgl
, sg_nents(sgl
), DMA_BIDIRECTIONAL
);
289 clear_hw_sgl_sge(hw_sgl
);
290 hw_sgl
->entry_sum_in_chain
= 0;
291 hw_sgl
->entry_sum_in_sgl
= 0;
292 hw_sgl
->entry_length_in_sgl
= 0;
294 EXPORT_SYMBOL_GPL(hisi_acc_sg_buf_unmap
);