Lines Matching defs:page

161  * chunk contains 64 4kB-page, so this chunk contain 64 stripes). For
212 struct page *meta_page; /* store meta block */
843 static void r5l_append_payload_page(struct r5l_log *log, struct page *page)
855 if (!bio_add_page(io->current_bio, page, PAGE_SIZE, 0))
942 r5l_append_payload_page(log, sh->dev[i].page);
949 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page);
950 r5l_append_payload_page(log, sh->dev[sh->qd_idx].page);
955 r5l_append_payload_page(log, sh->dev[sh->pd_idx].page);
1024 addr = kmap_atomic(sh->dev[i].page);
1606 struct page *meta_page; /* current meta */
1615 * read ahead page pool (ra_pool)
1617 * read every page with sync_page_io(). The read ahead page pool
1621 struct page *ra_pool[R5L_RECOVERY_PAGE_POOL_SIZE];
1623 sector_t pool_offset; /* offset of first page in the pool */
1631 struct page *page;
1636 page = alloc_page(GFP_KERNEL);
1638 if (!page)
1640 ctx->ra_pool[ctx->total_pages] = page;
1697 * try read a page from the read ahead page pool, if the page is not in the
1702 struct page *page,
1717 memcpy(page_address(page),
1727 struct page *page = ctx->meta_page;
1732 ret = r5l_recovery_read_page(log, ctx, page, ctx->pos);
1736 mb = page_address(page);
1760 struct page *page,
1765 mb = page_address(page);
1777 struct page *page;
1780 page = alloc_page(GFP_KERNEL);
1781 if (!page)
1783 r5l_recovery_create_empty_meta_block(log, page, pos, seq);
1784 mb = page_address(page);
1787 if (!sync_page_io(log->rdev, pos, PAGE_SIZE, page, REQ_OP_WRITE |
1789 __free_page(page);
1792 __free_page(page);
1816 r5l_recovery_read_page(log, ctx, sh->dev[dd_idx].page, log_offset);
1835 r5l_recovery_read_page(log, ctx, sh->dev[sh->pd_idx].page, log_offset);
1842 log, ctx, sh->dev[sh->qd_idx].page,
1897 sh->dev[disk_index].page, REQ_OP_WRITE,
1907 sh->dev[disk_index].page, REQ_OP_WRITE,
1979 struct page *page,
1985 r5l_recovery_read_page(log, ctx, page, log_offset);
1986 addr = kmap_atomic(page);
1994 * if there is mismatch for any data page, we drop all data in the mata block
2005 struct page *page;
2009 page = alloc_page(GFP_KERNEL);
2010 if (!page)
2019 log, ctx, page, log_offset,
2024 log, ctx, page, log_offset,
2029 log, ctx, page,
2053 put_page(page);
2057 put_page(page);
2066 * -EAGAIN for checksum mismatch of data page
2351 struct page *page;
2354 page = alloc_page(GFP_KERNEL);
2355 if (!page) {
2370 r5l_recovery_create_empty_meta_block(log, page,
2372 mb = page_address(page);
2388 addr = kmap_atomic(dev->page);
2394 dev->page, REQ_OP_WRITE, false);
2405 sync_page_io(log->rdev, ctx->pos, PAGE_SIZE, page,
2415 __free_page(page);
2529 static ssize_t r5c_journal_mode_show(struct mddev *mddev, char *page)
2545 page, PAGE_SIZE, "[%s] %s\n",
2551 page, PAGE_SIZE, "%s [%s]\n",
2597 const char *page, size_t length)
2606 if (page[len - 1] == '\n')
2611 !strncmp(page, r5c_journal_mode_str[mode], len))
2767 if (sh->dev[i].page != sh->dev[i].orig_page) {
2768 struct page *p = sh->dev[i].orig_page;
2770 sh->dev[i].orig_page = sh->dev[i].page;
2791 if (dev->orig_page != dev->page)
2901 addr = kmap_atomic(sh->dev[i].page);
2960 struct page *page;
2970 page = alloc_page(GFP_KERNEL);
2971 if (!page)
2974 if (!sync_page_io(rdev, cp, PAGE_SIZE, page, REQ_OP_READ, false)) {
2978 mb = page_address(page);
3016 __free_page(page);
3028 __free_page(page);