Lines Matching defs:mapping

208 	struct address_space *mapping;
215 mapping = page_mapping(p);
216 if (mapping == NULL || mapping->host == NULL)
219 dev = mapping->host->i_sb->s_dev;
428 * page->mapping are sufficient for mapping the page back to its
454 * so "tk->size_shift == 0" effectively checks no mapping on
459 * has a mapping for the page.
640 struct address_space *mapping = folio->mapping;
643 i_mmap_lock_read(mapping);
651 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff,
665 i_mmap_unlock_read(mapping);
680 struct address_space *mapping, pgoff_t pgoff,
686 i_mmap_lock_read(mapping);
693 vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
699 i_mmap_unlock_read(mapping);
709 if (!folio->mapping)
843 * process has multiple entries mapping the error page.
930 struct address_space *mapping)
934 if (mapping->a_ops->error_remove_page) {
936 int err = mapping->a_ops->error_remove_page(mapping, p);
1017 struct address_space *mapping;
1038 mapping = page_mapping(p);
1039 if (!mapping) {
1051 extra_pins = shmem_mapping(mapping);
1058 ret = truncate_error_page(p, page_to_pfn(p), mapping);
1075 struct address_space *mapping = page_mapping(p);
1079 if (mapping) {
1082 * who check the mapping.
1114 mapping_set_error(mapping, -EIO);
1186 struct address_space *mapping;
1189 mapping = page_mapping(hpage);
1190 if (mapping) {
1191 res = truncate_error_page(hpage, page_to_pfn(p), mapping);
1554 struct address_space *mapping;
1587 mapping = page_mapping(hpage);
1588 if (!(flags & MF_MUST_KILL) && !PageDirty(hpage) && mapping &&
1589 mapping_can_writeback(mapping)) {
1614 mapping = hugetlb_page_mapping_lock_write(hpage);
1615 if (mapping) {
1617 i_mmap_unlock_write(mapping);
1619 pr_info("%#lx: could not lock mapping for mapped huge page\n", pfn);
1691 struct address_space *mapping, pgoff_t index, int flags)
1702 * Unmap the largest mapping to avoid breaking up device-dax
1704 * mapping being torn down is communicated in siginfo, see
1709 unmap_mapping_range(mapping, start, size, 0);
1736 * also prevents changes to the mapping of this pfn until
1776 unmap_and_kill(&to_kill, pfn, folio->mapping, folio->index, flags);
1785 * @mapping: address_space of the file in use
1790 int mf_dax_kill_procs(struct address_space *mapping, pgoff_t index,
1802 cookie = dax_lock_mapping_entry(mapping, index, &page);
1810 collect_procs_fsdax(page, mapping, index, &to_kill);
1811 unmap_and_kill(&to_kill, page_to_pfn(page), mapping,
1814 dax_unlock_mapping_entry(mapping, index, cookie);
2358 if (PageLRU(p) && !PageSwapCache(p) && p->mapping == NULL) {
2553 unpoison_pr_info("Unpoison: the hwpoison page has non-NULL mapping %#lx\n",