1// SPDX-License-Identifier: GPL-2.0
2/* Copyright (c) 2019 HiSilicon Limited. */
3#include <linux/dma-mapping.h>
4#include <linux/module.h>
5#include <linux/slab.h>
6#include "qm.h"
7
8#define HISI_ACC_SGL_SGE_NR_MIN		1
9#define HISI_ACC_SGL_NR_MAX		256
10#define HISI_ACC_SGL_ALIGN_SIZE		64
11#define HISI_ACC_MEM_BLOCK_NR		5
12
13struct acc_hw_sge {
14	dma_addr_t buf;
15	void *page_ctrl;
16	__le32 len;
17	__le32 pad;
18	__le32 pad0;
19	__le32 pad1;
20};
21
22/* use default sgl head size 64B */
23struct hisi_acc_hw_sgl {
24	dma_addr_t next_dma;
25	__le16 entry_sum_in_chain;
26	__le16 entry_sum_in_sgl;
27	__le16 entry_length_in_sgl;
28	__le16 pad0;
29	__le64 pad1[5];
30	struct hisi_acc_hw_sgl *next;
31	struct acc_hw_sge sge_entries[];
32} __aligned(1);
33
34struct hisi_acc_sgl_pool {
35	struct mem_block {
36		struct hisi_acc_hw_sgl *sgl;
37		dma_addr_t sgl_dma;
38		size_t size;
39	} mem_block[HISI_ACC_MEM_BLOCK_NR];
40	u32 sgl_num_per_block;
41	u32 block_num;
42	u32 count;
43	u32 sge_nr;
44	size_t sgl_size;
45};
46
47/**
48 * hisi_acc_create_sgl_pool() - Create a hw sgl pool.
49 * @dev: The device which hw sgl pool belongs to.
50 * @count: Count of hisi_acc_hw_sgl in pool.
51 * @sge_nr: The count of sge in hw_sgl
52 *
53 * This function creates a hw sgl pool, after this user can get hw sgl memory
54 * from it.
55 */
56struct hisi_acc_sgl_pool *hisi_acc_create_sgl_pool(struct device *dev,
57						   u32 count, u32 sge_nr)
58{
59	u32 sgl_size, block_size, sgl_num_per_block, block_num, remain_sgl = 0;
60	struct hisi_acc_sgl_pool *pool;
61	struct mem_block *block;
62	u32 i, j;
63
64	if (!dev || !count || !sge_nr || sge_nr > HISI_ACC_SGL_SGE_NR_MAX)
65		return ERR_PTR(-EINVAL);
66
67	sgl_size = sizeof(struct acc_hw_sge) * sge_nr +
68		   sizeof(struct hisi_acc_hw_sgl);
69	block_size = 1 << (PAGE_SHIFT + MAX_ORDER <= 32 ?
70			   PAGE_SHIFT + MAX_ORDER - 1 : 31);
71	sgl_num_per_block = block_size / sgl_size;
72	block_num = count / sgl_num_per_block;
73	remain_sgl = count % sgl_num_per_block;
74
75	if ((!remain_sgl && block_num > HISI_ACC_MEM_BLOCK_NR) ||
76	    (remain_sgl > 0 && block_num > HISI_ACC_MEM_BLOCK_NR - 1))
77		return ERR_PTR(-EINVAL);
78
79	pool = kzalloc(sizeof(*pool), GFP_KERNEL);
80	if (!pool)
81		return ERR_PTR(-ENOMEM);
82	block = pool->mem_block;
83
84	for (i = 0; i < block_num; i++) {
85		block[i].sgl = dma_alloc_coherent(dev, block_size,
86						  &block[i].sgl_dma,
87						  GFP_KERNEL);
88		if (!block[i].sgl)
89			goto err_free_mem;
90
91		block[i].size = block_size;
92	}
93
94	if (remain_sgl > 0) {
95		block[i].sgl = dma_alloc_coherent(dev, remain_sgl * sgl_size,
96						  &block[i].sgl_dma,
97						  GFP_KERNEL);
98		if (!block[i].sgl)
99			goto err_free_mem;
100
101		block[i].size = remain_sgl * sgl_size;
102	}
103
104	pool->sgl_num_per_block = sgl_num_per_block;
105	pool->block_num = remain_sgl ? block_num + 1 : block_num;
106	pool->count = count;
107	pool->sgl_size = sgl_size;
108	pool->sge_nr = sge_nr;
109
110	return pool;
111
112err_free_mem:
113	for (j = 0; j < i; j++) {
114		dma_free_coherent(dev, block_size, block[j].sgl,
115				  block[j].sgl_dma);
116	}
117	kfree_sensitive(pool);
118	return ERR_PTR(-ENOMEM);
119}
120EXPORT_SYMBOL_GPL(hisi_acc_create_sgl_pool);
121
122/**
123 * hisi_acc_free_sgl_pool() - Free a hw sgl pool.
124 * @dev: The device which hw sgl pool belongs to.
125 * @pool: Pointer of pool.
126 *
127 * This function frees memory of a hw sgl pool.
128 */
129void hisi_acc_free_sgl_pool(struct device *dev, struct hisi_acc_sgl_pool *pool)
130{
131	struct mem_block *block;
132	int i;
133
134	if (!dev || !pool)
135		return;
136
137	block = pool->mem_block;
138
139	for (i = 0; i < pool->block_num; i++)
140		dma_free_coherent(dev, block[i].size, block[i].sgl,
141				  block[i].sgl_dma);
142
143	kfree(pool);
144}
145EXPORT_SYMBOL_GPL(hisi_acc_free_sgl_pool);
146
147static struct hisi_acc_hw_sgl *acc_get_sgl(struct hisi_acc_sgl_pool *pool,
148					   u32 index, dma_addr_t *hw_sgl_dma)
149{
150	struct mem_block *block;
151	u32 block_index, offset;
152
153	if (!pool || !hw_sgl_dma || index >= pool->count)
154		return ERR_PTR(-EINVAL);
155
156	block = pool->mem_block;
157	block_index = index / pool->sgl_num_per_block;
158	offset = index % pool->sgl_num_per_block;
159
160	*hw_sgl_dma = block[block_index].sgl_dma + pool->sgl_size * offset;
161	return (void *)block[block_index].sgl + pool->sgl_size * offset;
162}
163
164static void sg_map_to_hw_sg(struct scatterlist *sgl,
165			    struct acc_hw_sge *hw_sge)
166{
167	hw_sge->buf = sg_dma_address(sgl);
168	hw_sge->len = cpu_to_le32(sg_dma_len(sgl));
169}
170
171static void inc_hw_sgl_sge(struct hisi_acc_hw_sgl *hw_sgl)
172{
173	u16 var = le16_to_cpu(hw_sgl->entry_sum_in_sgl);
174
175	var++;
176	hw_sgl->entry_sum_in_sgl = cpu_to_le16(var);
177}
178
179static void update_hw_sgl_sum_sge(struct hisi_acc_hw_sgl *hw_sgl, u16 sum)
180{
181	hw_sgl->entry_sum_in_chain = cpu_to_le16(sum);
182}
183
184/**
185 * hisi_acc_sg_buf_map_to_hw_sgl - Map a scatterlist to a hw sgl.
186 * @dev: The device which hw sgl belongs to.
187 * @sgl: Scatterlist which will be mapped to hw sgl.
188 * @pool: Pool which hw sgl memory will be allocated in.
189 * @index: Index of hisi_acc_hw_sgl in pool.
190 * @hw_sgl_dma: The dma address of allocated hw sgl.
191 *
192 * This function builds hw sgl according input sgl, user can use hw_sgl_dma
193 * as src/dst in its BD. Only support single hw sgl currently.
194 */
195struct hisi_acc_hw_sgl *
196hisi_acc_sg_buf_map_to_hw_sgl(struct device *dev,
197			      struct scatterlist *sgl,
198			      struct hisi_acc_sgl_pool *pool,
199			      u32 index, dma_addr_t *hw_sgl_dma)
200{
201	struct hisi_acc_hw_sgl *curr_hw_sgl;
202	dma_addr_t curr_sgl_dma = 0;
203	struct acc_hw_sge *curr_hw_sge;
204	struct scatterlist *sg;
205	int i, sg_n, sg_n_mapped;
206
207	if (!dev || !sgl || !pool || !hw_sgl_dma)
208		return ERR_PTR(-EINVAL);
209
210	sg_n = sg_nents(sgl);
211
212	sg_n_mapped = dma_map_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL);
213	if (!sg_n_mapped)
214		return ERR_PTR(-EINVAL);
215
216	if (sg_n_mapped > pool->sge_nr) {
217		dma_unmap_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL);
218		return ERR_PTR(-EINVAL);
219	}
220
221	curr_hw_sgl = acc_get_sgl(pool, index, &curr_sgl_dma);
222	if (IS_ERR(curr_hw_sgl)) {
223		dma_unmap_sg(dev, sgl, sg_n, DMA_BIDIRECTIONAL);
224		return ERR_PTR(-ENOMEM);
225
226	}
227	curr_hw_sgl->entry_length_in_sgl = cpu_to_le16(pool->sge_nr);
228	curr_hw_sge = curr_hw_sgl->sge_entries;
229
230	for_each_sg(sgl, sg, sg_n_mapped, i) {
231		sg_map_to_hw_sg(sg, curr_hw_sge);
232		inc_hw_sgl_sge(curr_hw_sgl);
233		curr_hw_sge++;
234	}
235
236	update_hw_sgl_sum_sge(curr_hw_sgl, pool->sge_nr);
237	*hw_sgl_dma = curr_sgl_dma;
238
239	return curr_hw_sgl;
240}
241EXPORT_SYMBOL_GPL(hisi_acc_sg_buf_map_to_hw_sgl);
242
243/**
244 * hisi_acc_sg_buf_unmap() - Unmap allocated hw sgl.
245 * @dev: The device which hw sgl belongs to.
246 * @sgl: Related scatterlist.
247 * @hw_sgl: Virtual address of hw sgl.
248 * @hw_sgl_dma: DMA address of hw sgl.
249 * @pool: Pool which hw sgl is allocated in.
250 *
251 * This function unmaps allocated hw sgl.
252 */
253void hisi_acc_sg_buf_unmap(struct device *dev, struct scatterlist *sgl,
254			   struct hisi_acc_hw_sgl *hw_sgl)
255{
256	if (!dev || !sgl || !hw_sgl)
257		return;
258
259	dma_unmap_sg(dev, sgl, sg_nents(sgl), DMA_BIDIRECTIONAL);
260
261	hw_sgl->entry_sum_in_chain = 0;
262	hw_sgl->entry_sum_in_sgl = 0;
263	hw_sgl->entry_length_in_sgl = 0;
264}
265EXPORT_SYMBOL_GPL(hisi_acc_sg_buf_unmap);
266