Lines Matching defs:block

55 static int sram_add_pool(struct sram_dev *sram, struct sram_reserve *block,
61 NUMA_NO_NODE, block->label);
66 block->size, NUMA_NO_NODE);
75 static int sram_add_export(struct sram_dev *sram, struct sram_reserve *block,
88 part->battr.size = block->size;
93 static int sram_add_partition(struct sram_dev *sram, struct sram_reserve *block,
105 virt_base = devm_ioremap_resource(sram->dev, &block->res);
107 virt_base = devm_ioremap_resource_wc(sram->dev, &block->res);
110 dev_err(sram->dev, "could not map SRAM at %pr\n", &block->res);
116 part->base = sram->virt_base + block->start;
119 if (block->pool) {
120 ret = sram_add_pool(sram, block, start, part);
124 if (block->export) {
125 ret = sram_add_export(sram, block, start, part);
129 if (block->protect_exec) {
130 ret = sram_check_protect_exec(sram, block, part);
134 ret = sram_add_pool(sram, block, start, part);
177 struct sram_reserve *rblocks, *block;
188 * We need an additional block to mark the end of the memory region
196 block = &rblocks[0];
210 "reserved block %pOF outside the sram area\n",
216 block->start = child_res.start - res->start;
217 block->size = resource_size(&child_res);
218 block->res = child_res;
219 list_add_tail(&block->list, &reserve_list);
221 block->export = of_property_read_bool(child, "export");
222 block->pool = of_property_read_bool(child, "pool");
223 block->protect_exec = of_property_read_bool(child, "protect-exec");
225 if ((block->export || block->pool || block->protect_exec) &&
226 block->size) {
238 block->label = devm_kasprintf(sram->dev, GFP_KERNEL,
241 block->label = devm_kstrdup(sram->dev,
243 if (!block->label) {
249 block->export ? "exported " : "", block->label,
250 block->start, block->start + block->size);
252 dev_dbg(sram->dev, "found reserved block 0x%x-0x%x\n",
253 block->start, block->start + block->size);
256 block++;
278 list_for_each_entry(block, &reserve_list, list) {
280 if (block->start < cur_start) {
282 "block at 0x%x starts after current offset 0x%lx\n",
283 block->start, cur_start);
289 if ((block->export || block->pool || block->protect_exec) &&
290 block->size) {
291 ret = sram_add_partition(sram, block,
292 res->start + block->start);
299 /* current start is in a reserved block, so continue after it */
300 if (block->start == cur_start) {
301 cur_start = block->start + block->size;
307 * address and the following reserved block, or the
310 cur_size = block->start - cur_start;
325 /* next allocation after this reserved block */
326 cur_start = block->start + block->size;