Lines Matching defs:r_cpos
890 if (le64_to_cpu(rec->r_cpos) +
893 else if (le64_to_cpu(rec->r_cpos) > cpos)
904 ret_rec->r_cpos = cpu_to_le64(cpos);
907 le64_to_cpu(rec->r_cpos) < cpos + len)
909 cpu_to_le32(le64_to_cpu(rec->r_cpos) - cpos);
1155 (le64_to_cpu(rb->rf_records.rl_recs[index].r_cpos) +
1157 le64_to_cpu(rb->rf_records.rl_recs[index + 1].r_cpos)))
1375 u32 r_cpos = ocfs2_get_ref_rec_low_cpos(r);
1377 if (l_cpos > r_cpos)
1379 if (l_cpos < r_cpos)
1387 u64 l_cpos = le64_to_cpu(l->r_cpos);
1388 u64 r_cpos = le64_to_cpu(r->r_cpos);
1390 if (l_cpos > r_cpos)
1392 if (l_cpos < r_cpos)
1671 new_cpos = le64_to_cpu(rec->r_cpos) & OCFS2_32BIT_POS_MASK;
1725 /* change the r_cpos in the leaf block. */
1753 u64 cpos = le64_to_cpu(rec->r_cpos);
1791 (unsigned long long)le64_to_cpu(rec->r_cpos),
1844 trace_ocfs2_split_refcount_rec(le64_to_cpu(orig_rec->r_cpos),
1847 le64_to_cpu(split_rec->r_cpos),
1857 (split_rec->r_cpos == orig_rec->r_cpos ||
1858 le64_to_cpu(split_rec->r_cpos) +
1860 le64_to_cpu(orig_rec->r_cpos) + le32_to_cpu(orig_rec->r_clusters)))
1870 (split_rec->r_cpos != orig_rec->r_cpos &&
1871 le64_to_cpu(split_rec->r_cpos) +
1873 le64_to_cpu(orig_rec->r_cpos) + le32_to_cpu(orig_rec->r_clusters)))
1880 u64 cpos = le64_to_cpu(orig_rec->r_cpos);
1925 len = (le64_to_cpu(orig_rec->r_cpos) +
1927 (le64_to_cpu(split_rec->r_cpos) +
1938 le64_add_cpu(&tail_rec->r_cpos,
1950 * orig_rec above, so the check for r_cpos is faked.
1952 if (split_rec->r_cpos != orig_rec->r_cpos && tail_rec != orig_rec) {
1953 len = le64_to_cpu(split_rec->r_cpos) -
1954 le64_to_cpu(orig_rec->r_cpos);
1965 (unsigned long long)le64_to_cpu(split_rec->r_cpos),
2017 if (rec.r_refcount && le64_to_cpu(rec.r_cpos) == cpos &&
2033 (unsigned long long)le64_to_cpu(rec.r_cpos),
2045 le64_to_cpu(rec.r_cpos) + set_len) - cpos;
2046 rec.r_cpos = cpu_to_le64(cpos);
2051 (unsigned long long)le64_to_cpu(rec.r_cpos),
2179 BUG_ON(cpos < le64_to_cpu(rec->r_cpos));
2181 le64_to_cpu(rec->r_cpos) + le32_to_cpu(rec->r_clusters));
2187 if (cpos == le64_to_cpu(rec->r_cpos) &&
2193 split.r_cpos = cpu_to_le64(cpos);
2254 r_len = min((u64)(cpos + len), le64_to_cpu(rec.r_cpos) +
2418 (unsigned long long)le64_to_cpu(rec.r_cpos),
2422 len = min((u64)cpos + clusters, le64_to_cpu(rec.r_cpos) +
2447 cpos != le64_to_cpu(rec.r_cpos))
2451 if (cpos + clusters < le64_to_cpu(rec.r_cpos) +
3256 le64_to_cpu(rec.r_cpos) +