Lines Matching refs:mapping

139 file_ra_state_init(struct file_ra_state *ra, struct address_space *mapping)
141 ra->ra_pages = inode_to_bdi(mapping->host)->ra_pages;
148 const struct address_space_operations *aops = rac->mapping->a_ops;
208 struct address_space *mapping = ractl->mapping;
210 gfp_t gfp_mask = readahead_gfp_mask(mapping);
220 * filesystems already specify __GFP_NOFS in their mapping's
225 filemap_invalidate_lock_shared(mapping);
230 struct folio *folio = xa_load(&mapping->i_pages, index + i);
250 if (filemap_add_folio(mapping, folio, index + i,
270 filemap_invalidate_unlock_shared(mapping);
284 struct inode *inode = ractl->mapping->host;
309 struct address_space *mapping = ractl->mapping;
311 struct backing_dev_info *bdi = inode_to_bdi(mapping->host);
314 if (unlikely(!mapping->a_ops->read_folio && !mapping->a_ops->readahead))
418 static pgoff_t count_history_pages(struct address_space *mapping,
424 head = page_cache_prev_miss(mapping, index - 1, max);
433 static int try_context_readahead(struct address_space *mapping,
441 size = count_history_pages(mapping, index, max);
475 err = filemap_add_folio(ractl->mapping, folio, index, gfp);
489 struct address_space *mapping = ractl->mapping;
491 pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT;
494 gfp_t gfp = readahead_gfp_mask(mapping);
496 if (!mapping_large_folio_support(mapping) || ra->size < 4)
509 filemap_invalidate_lock_shared(mapping);
536 filemap_invalidate_unlock_shared(mapping);
555 struct backing_dev_info *bdi = inode_to_bdi(ractl->mapping->host);
599 start = page_cache_next_miss(ractl->mapping, index + 1,
633 if (try_context_readahead(ractl->mapping, ra, index, req_size,
781 struct address_space *mapping = ractl->mapping;
784 gfp_t gfp_mask = readahead_gfp_mask(mapping);
791 struct folio *folio = xa_load(&mapping->i_pages, index);
799 if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) {
818 struct folio *folio = xa_load(&mapping->i_pages, index);
826 if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) {