1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Generic on-chip SRAM allocation driver
5 * Copyright (C) 2012 Philipp Zabel, Pengutronix
9 #include <linux/delay.h>
10 #include <linux/genalloc.h>
12 #include <linux/list_sort.h>
13 #include <linux/of_address.h>
14 #include <linux/of_device.h>
15 #include <linux/platform_device.h>
16 #include <linux/regmap.h>
17 #include <linux/slab.h>
18 #include <linux/mfd/syscon.h>
19 #include <soc/at91/atmel-secumod.h>
23 #define SRAM_GRANULARITY 32
25 static ssize_t
sram_read(struct file
*filp
, struct kobject
*kobj
,
26 struct bin_attribute
*attr
,
27 char *buf
, loff_t pos
, size_t count
)
29 struct sram_partition
*part
;
31 part
= container_of(attr
, struct sram_partition
, battr
);
33 mutex_lock(&part
->lock
);
34 memcpy_fromio(buf
, part
->base
+ pos
, count
);
35 mutex_unlock(&part
->lock
);
40 static ssize_t
sram_write(struct file
*filp
, struct kobject
*kobj
,
41 struct bin_attribute
*attr
,
42 char *buf
, loff_t pos
, size_t count
)
44 struct sram_partition
*part
;
46 part
= container_of(attr
, struct sram_partition
, battr
);
48 mutex_lock(&part
->lock
);
49 memcpy_toio(part
->base
+ pos
, buf
, count
);
50 mutex_unlock(&part
->lock
);
55 static int sram_add_pool(struct sram_dev
*sram
, struct sram_reserve
*block
,
56 phys_addr_t start
, struct sram_partition
*part
)
60 part
->pool
= devm_gen_pool_create(sram
->dev
, ilog2(SRAM_GRANULARITY
),
61 NUMA_NO_NODE
, block
->label
);
62 if (IS_ERR(part
->pool
))
63 return PTR_ERR(part
->pool
);
65 ret
= gen_pool_add_virt(part
->pool
, (unsigned long)part
->base
, start
,
66 block
->size
, NUMA_NO_NODE
);
68 dev_err(sram
->dev
, "failed to register subpool: %d\n", ret
);
75 static int sram_add_export(struct sram_dev
*sram
, struct sram_reserve
*block
,
76 phys_addr_t start
, struct sram_partition
*part
)
78 sysfs_bin_attr_init(&part
->battr
);
79 part
->battr
.attr
.name
= devm_kasprintf(sram
->dev
, GFP_KERNEL
,
81 (unsigned long long)start
);
82 if (!part
->battr
.attr
.name
)
85 part
->battr
.attr
.mode
= S_IRUSR
| S_IWUSR
;
86 part
->battr
.read
= sram_read
;
87 part
->battr
.write
= sram_write
;
88 part
->battr
.size
= block
->size
;
90 return device_create_bin_file(sram
->dev
, &part
->battr
);
93 static int sram_add_partition(struct sram_dev
*sram
, struct sram_reserve
*block
,
97 struct sram_partition
*part
= &sram
->partition
[sram
->partitions
];
99 mutex_init(&part
->lock
);
100 part
->base
= sram
->virt_base
+ block
->start
;
103 ret
= sram_add_pool(sram
, block
, start
, part
);
108 ret
= sram_add_export(sram
, block
, start
, part
);
112 if (block
->protect_exec
) {
113 ret
= sram_check_protect_exec(sram
, block
, part
);
117 ret
= sram_add_pool(sram
, block
, start
, part
);
121 sram_add_protect_exec(part
);
129 static void sram_free_partitions(struct sram_dev
*sram
)
131 struct sram_partition
*part
;
133 if (!sram
->partitions
)
136 part
= &sram
->partition
[sram
->partitions
- 1];
137 for (; sram
->partitions
; sram
->partitions
--, part
--) {
138 if (part
->battr
.size
)
139 device_remove_bin_file(sram
->dev
, &part
->battr
);
142 gen_pool_avail(part
->pool
) < gen_pool_size(part
->pool
))
143 dev_err(sram
->dev
, "removed pool while SRAM allocated\n");
147 static int sram_reserve_cmp(void *priv
, struct list_head
*a
,
150 struct sram_reserve
*ra
= list_entry(a
, struct sram_reserve
, list
);
151 struct sram_reserve
*rb
= list_entry(b
, struct sram_reserve
, list
);
153 return ra
->start
- rb
->start
;
156 static int sram_reserve_regions(struct sram_dev
*sram
, struct resource
*res
)
158 struct device_node
*np
= sram
->dev
->of_node
, *child
;
159 unsigned long size
, cur_start
, cur_size
;
160 struct sram_reserve
*rblocks
, *block
;
161 struct list_head reserve_list
;
162 unsigned int nblocks
, exports
= 0;
166 INIT_LIST_HEAD(&reserve_list
);
168 size
= resource_size(res
);
171 * We need an additional block to mark the end of the memory region
172 * after the reserved blocks from the dt are processed.
174 nblocks
= (np
) ? of_get_available_child_count(np
) + 1 : 1;
175 rblocks
= kcalloc(nblocks
, sizeof(*rblocks
), GFP_KERNEL
);
180 for_each_available_child_of_node(np
, child
) {
181 struct resource child_res
;
183 ret
= of_address_to_resource(child
, 0, &child_res
);
186 "could not get address for node %pOF\n",
191 if (child_res
.start
< res
->start
|| child_res
.end
> res
->end
) {
193 "reserved block %pOF outside the sram area\n",
199 block
->start
= child_res
.start
- res
->start
;
200 block
->size
= resource_size(&child_res
);
201 list_add_tail(&block
->list
, &reserve_list
);
203 if (of_find_property(child
, "export", NULL
))
204 block
->export
= true;
206 if (of_find_property(child
, "pool", NULL
))
209 if (of_find_property(child
, "protect-exec", NULL
))
210 block
->protect_exec
= true;
212 if ((block
->export
|| block
->pool
|| block
->protect_exec
) &&
217 ret
= of_property_read_string(child
, "label", &label
);
218 if (ret
&& ret
!= -EINVAL
) {
220 "%pOF has invalid label name\n",
227 block
->label
= devm_kstrdup(sram
->dev
,
234 dev_dbg(sram
->dev
, "found %sblock '%s' 0x%x-0x%x\n",
235 block
->export
? "exported " : "", block
->label
,
236 block
->start
, block
->start
+ block
->size
);
238 dev_dbg(sram
->dev
, "found reserved block 0x%x-0x%x\n",
239 block
->start
, block
->start
+ block
->size
);
246 /* the last chunk marks the end of the region */
247 rblocks
[nblocks
- 1].start
= size
;
248 rblocks
[nblocks
- 1].size
= 0;
249 list_add_tail(&rblocks
[nblocks
- 1].list
, &reserve_list
);
251 list_sort(NULL
, &reserve_list
, sram_reserve_cmp
);
254 sram
->partition
= devm_kcalloc(sram
->dev
,
255 exports
, sizeof(*sram
->partition
),
257 if (!sram
->partition
) {
264 list_for_each_entry(block
, &reserve_list
, list
) {
265 /* can only happen if sections overlap */
266 if (block
->start
< cur_start
) {
268 "block at 0x%x starts after current offset 0x%lx\n",
269 block
->start
, cur_start
);
271 sram_free_partitions(sram
);
275 if ((block
->export
|| block
->pool
|| block
->protect_exec
) &&
277 ret
= sram_add_partition(sram
, block
,
278 res
->start
+ block
->start
);
280 sram_free_partitions(sram
);
285 /* current start is in a reserved block, so continue after it */
286 if (block
->start
== cur_start
) {
287 cur_start
= block
->start
+ block
->size
;
292 * allocate the space between the current starting
293 * address and the following reserved block, or the
296 cur_size
= block
->start
- cur_start
;
298 dev_dbg(sram
->dev
, "adding chunk 0x%lx-0x%lx\n",
299 cur_start
, cur_start
+ cur_size
);
301 ret
= gen_pool_add_virt(sram
->pool
,
302 (unsigned long)sram
->virt_base
+ cur_start
,
303 res
->start
+ cur_start
, cur_size
, -1);
305 sram_free_partitions(sram
);
309 /* next allocation after this reserved block */
310 cur_start
= block
->start
+ block
->size
;
320 static int atmel_securam_wait(void)
322 struct regmap
*regmap
;
325 regmap
= syscon_regmap_lookup_by_compatible("atmel,sama5d2-secumod");
329 return regmap_read_poll_timeout(regmap
, AT91_SECUMOD_RAMRDY
, val
,
330 val
& AT91_SECUMOD_RAMRDY_READY
,
334 static const struct of_device_id sram_dt_ids
[] = {
335 { .compatible
= "mmio-sram" },
336 { .compatible
= "atmel,sama5d2-securam", .data
= atmel_securam_wait
},
340 static int sram_probe(struct platform_device
*pdev
)
342 struct sram_dev
*sram
;
344 int (*init_func
)(void);
346 sram
= devm_kzalloc(&pdev
->dev
, sizeof(*sram
), GFP_KERNEL
);
350 sram
->dev
= &pdev
->dev
;
352 if (of_property_read_bool(pdev
->dev
.of_node
, "no-memory-wc"))
353 sram
->virt_base
= devm_platform_ioremap_resource(pdev
, 0);
355 sram
->virt_base
= devm_platform_ioremap_resource_wc(pdev
, 0);
356 if (IS_ERR(sram
->virt_base
)) {
357 dev_err(&pdev
->dev
, "could not map SRAM registers\n");
358 return PTR_ERR(sram
->virt_base
);
361 sram
->pool
= devm_gen_pool_create(sram
->dev
, ilog2(SRAM_GRANULARITY
),
363 if (IS_ERR(sram
->pool
))
364 return PTR_ERR(sram
->pool
);
366 sram
->clk
= devm_clk_get(sram
->dev
, NULL
);
367 if (IS_ERR(sram
->clk
))
370 clk_prepare_enable(sram
->clk
);
372 ret
= sram_reserve_regions(sram
,
373 platform_get_resource(pdev
, IORESOURCE_MEM
, 0));
375 goto err_disable_clk
;
377 platform_set_drvdata(pdev
, sram
);
379 init_func
= of_device_get_match_data(&pdev
->dev
);
383 goto err_free_partitions
;
386 dev_dbg(sram
->dev
, "SRAM pool: %zu KiB @ 0x%p\n",
387 gen_pool_size(sram
->pool
) / 1024, sram
->virt_base
);
392 sram_free_partitions(sram
);
395 clk_disable_unprepare(sram
->clk
);
400 static int sram_remove(struct platform_device
*pdev
)
402 struct sram_dev
*sram
= platform_get_drvdata(pdev
);
404 sram_free_partitions(sram
);
406 if (gen_pool_avail(sram
->pool
) < gen_pool_size(sram
->pool
))
407 dev_err(sram
->dev
, "removed while SRAM allocated\n");
410 clk_disable_unprepare(sram
->clk
);
415 static struct platform_driver sram_driver
= {
418 .of_match_table
= sram_dt_ids
,
421 .remove
= sram_remove
,
424 static int __init
sram_init(void)
426 return platform_driver_register(&sram_driver
);
429 postcore_initcall(sram_init
);