Lines Matching refs:vbo
774 u64 vbo;
786 vbo = (u64)from * rs;
787 for (; from < to; from++, vbo += rs) {
790 err = ntfs_get_bh(sbi, run, vbo, rs, &nb);
1128 u64 vbo, const void *buf, size_t bytes, int sync)
1132 u32 off = vbo & sbi->cluster_mask;
1133 CLST lcn, clen, vcn = vbo >> cluster_bits, vcn_next;
1176 const struct runs_tree *run, u64 vbo)
1183 if (!run_lookup_entry(run, vbo >> cluster_bits, &lcn, NULL, NULL))
1186 lbo = ((u64)lcn << cluster_bits) + (vbo & sbi->cluster_mask);
1192 u64 vbo, void *buf, u32 bytes, struct ntfs_buffers *nb)
1198 u32 off = vbo & sbi->cluster_mask;
1200 CLST vcn_next, vcn = vbo >> cluster_bits;
1208 if (vbo > MFT_REC_VOL * sbi->record_size) {
1214 lbo = vbo + sbi->mft.lbo;
1309 int ntfs_read_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo,
1313 int err = ntfs_read_run_nb(sbi, run, vbo, rhdr, bytes, nb);
1320 int ntfs_get_bh(struct ntfs_sb_info *sbi, const struct runs_tree *run, u64 vbo,
1327 CLST vcn_next, vcn = vbo >> cluster_bits;
1341 off = vbo & sbi->cluster_mask;
1498 struct page **pages, u32 nr_pages, u64 vbo, u32 bytes,
1518 /* Align vbo and bytes to be 512 bytes aligned. */
1519 lbo = (vbo + bytes + 511) & ~511ull;
1520 vbo = vbo & ~511ull;
1521 bytes = lbo - vbo;
1523 vcn = vbo >> cluster_bits;
1528 off = vbo & sbi->cluster_mask;
1545 off = vbo & (PAGE_SIZE - 1);
1554 vbo += add;
1667 u64 vbo, u64 *lbo, u64 *bytes)
1673 if (!run_lookup_entry(run, vbo >> cluster_bits, &lcn, &len, NULL))
1676 off = vbo & sbi->cluster_mask;