Lines Matching defs:mode

86 int isolate_movable_page(struct page *page, isolate_mode_t mode)
129 if (!mapping->a_ops->isolate_page(page, mode))
702 enum migrate_mode mode, int extra_count)
713 if (mode != MIGRATE_SYNC_NO_COPY)
728 enum migrate_mode mode)
730 return migrate_page_extra(mapping, newpage, page, mode, 0);
737 enum migrate_mode mode)
742 if (mode != MIGRATE_ASYNC) {
774 struct page *newpage, struct page *page, enum migrate_mode mode,
782 return migrate_page(mapping, newpage, page, mode);
790 if (!buffer_migrate_lock_buffers(head, mode))
833 if (mode != MIGRATE_SYNC_NO_COPY)
858 struct page *newpage, struct page *page, enum migrate_mode mode)
860 return __buffer_migrate_page(mapping, newpage, page, mode, false);
871 struct page *newpage, struct page *page, enum migrate_mode mode)
873 return __buffer_migrate_page(mapping, newpage, page, mode, true);
922 struct page *newpage, struct page *page, enum migrate_mode mode)
926 switch (mode) {
942 return mode == MIGRATE_SYNC ? -EAGAIN : -EBUSY;
944 return migrate_page(mapping, newpage, page, mode);
959 enum migrate_mode mode)
972 rc = migrate_page(mapping, newpage, page, mode);
982 page, mode);
985 page, mode);
999 page, mode);
1039 int force, enum migrate_mode mode)
1047 if (!force || mode == MIGRATE_ASYNC)
1076 switch (mode) {
1118 rc = move_to_new_page(newpage, page, mode);
1149 rc = move_to_new_page(newpage, page, mode);
1189 int force, enum migrate_mode mode,
1215 rc = __unmap_and_move(page, newpage, force, mode);
1295 enum migrate_mode mode, int reason)
1322 switch (mode) {
1356 * semaphore in write mode here and set TTU_RMAP_LOCKED
1375 rc = move_to_new_page(new_hpage, hpage, mode);
1422 * @mode: The migration mode that specifies the constraints for
1435 enum migrate_mode mode, int reason)
1472 pass > 2, mode, reason);
1475 private, page, pass > 2, mode,
1549 nr_thp_failed, nr_thp_split, mode, reason);
2109 * dirty pages in MIGRATE_ASYNC mode which is a waste of cycles.
2811 * in read mode (hence no one can unmap the range being migrated).