Lines Matching defs:prot

193 		unsigned long end, pgprot_t prot, struct page **pages, int *nr,
213 set_pte_at(&init_mm, addr, pte, mk_pte(page, prot));
221 unsigned long end, pgprot_t prot, struct page **pages, int *nr,
232 if (vmap_pte_range(pmd, addr, next, prot, pages, nr, mask))
239 unsigned long end, pgprot_t prot, struct page **pages, int *nr,
250 if (vmap_pmd_range(pud, addr, next, prot, pages, nr, mask))
257 unsigned long end, pgprot_t prot, struct page **pages, int *nr,
268 if (vmap_pud_range(p4d, addr, next, prot, pages, nr, mask))
278 * @prot: page protection flags to use
293 pgprot_t prot, struct page **pages)
309 err = vmap_p4d_range(pgd, addr, next, prot, pages, &nr, &mask);
320 int map_kernel_range(unsigned long start, unsigned long size, pgprot_t prot,
325 ret = map_kernel_range_noflush(start, size, prot, pages);
2376 * @prot: page protection for the mapping
2387 unsigned long flags, pgprot_t prot)
2402 if (map_kernel_range((unsigned long)area->addr, size, pgprot_nx(prot),
2419 pgprot_t prot;
2429 *pte = pte_mkspecial(pfn_pte(data->pfns[data->idx++], data->prot));
2437 * @prot: page protection for the mapping
2442 void *vmap_pfn(unsigned long *pfns, unsigned int count, pgprot_t prot)
2444 struct vmap_pfn_data data = { .pfns = pfns, .prot = pgprot_nx(prot) };
2466 pgprot_t prot, int node)
2517 prot, pages) < 0)
2537 * @prot: protection mask for the allocated pages
2544 * kernel virtual space, using a pagetable protection of @prot.
2550 pgprot_t prot, unsigned long vm_flags, int node,
2566 addr = __vmalloc_area_node(area, gfp_mask, prot, node);