Lines Matching refs:va_end
456 return (va->va_end - va->va_start);
503 else if (addr >= va->va_end)
551 if (va->va_start < tmp_va->va_end &&
552 va->va_end <= tmp_va->va_start)
554 else if (va->va_end > tmp_va->va_start &&
555 va->va_start >= tmp_va->va_end)
559 va->va_start, va->va_end, tmp_va->va_start, tmp_va->va_end);
661 * when VA size is modified by changing its va_start/va_end. Or
776 if (sibling->va_start == va->va_end) {
797 if (sibling->va_end == va->va_start) {
808 sibling->va_end = va->va_end;
846 return (nva_start_addr + size <= va->va_end);
964 nva_start_addr + size > va->va_end)
969 if (va->va_end == nva_start_addr + size)
973 } else if (va->va_end == nva_start_addr + size) {
1016 va->va_end = nva_start_addr;
1061 lva->va_end = nva_start_addr;
1221 va->va_end = addr + size;
1231 BUG_ON(va->va_end > vend);
1346 if (va->va_end > end)
1347 end = va->va_end;
1355 unsigned long nr = (va->va_end - va->va_start) >> PAGE_SHIFT;
1357 unsigned long orig_end = va->va_end;
1372 va->va_start, va->va_end);
1419 nr_lazy = atomic_long_add_return((va->va_end - va->va_start) >>
1434 flush_cache_vunmap(va->va_start, va->va_end);
1435 unmap_kernel_range_noflush(va->va_start, va->va_end - va->va_start);
1437 flush_tlb_kernel_range(va->va_start, va->va_end);
1823 (va->va_end - va->va_start));
1944 free->va_end = busy->va_start;
1952 vmap_start = busy->va_end;
1959 free->va_end = vmap_end;
1998 va->va_end = va->va_start + tmp->size;
2032 vm->size = va->va_end - va->va_start;
3111 * i.e. va->va_start < addr && va->va_end < addr or NULL
3127 if (tmp->va_end >= addr)
3157 addr = min((*va)->va_end & ~(align - 1), vmalloc_end);
3271 if (base + end > va->va_end) {
3324 va->va_end = start + size;
3360 orig_end = vas[area]->va_end;
3365 va->va_start, va->va_end);
3410 orig_end = vas[area]->va_end;
3415 va->va_start, va->va_end);
3501 (void *)va->va_start, (void *)va->va_end,
3502 va->va_end - va->va_start);
3519 (void *)va->va_start, (void *)va->va_end,
3520 va->va_end - va->va_start);