Lines Matching defs:range
147 size += range_len(&dev_dax->ranges[i].range);
373 struct range *range = &dev_dax->ranges[i].range;
377 dev_dbg(&dev_dax->dev, "delete range[%d]: %#llx:%#llx\n", i,
378 (unsigned long long)range->start,
379 (unsigned long long)range->end);
381 __release_region(&dax_region->res, range->start, range_len(range));
560 struct range *range, int target_node, unsigned int align,
575 if (!IS_ALIGNED(range->start, align)
576 || !IS_ALIGNED(range_len(range), align))
591 .start = range->start,
592 .end = range->end,
670 rc = sprintf(buf, "%#llx\n", dax_range->range.start);
686 rc = sprintf(buf, "%#llx\n", dax_range->range.end);
814 pgoff += PHYS_PFN(range_len(&ranges[i].range));
818 .range = {
824 dev_dbg(dev, "alloc range[%d]: %pa:%pa\n", dev_dax->nr_range - 1,
847 struct range *range = &dax_range->range;
856 rc = adjust_resource(res, range->start, size);
860 *range = (struct range) {
861 .start = range->start,
862 .end = range->start + size - 1,
865 dev_dbg(dev, "%s range[%d]: %#llx:%#llx\n", is_shrink ? "shrink" : "extend",
866 last_range, (unsigned long long) range->start,
867 (unsigned long long) range->end);
902 struct range *range = &dev_dax->ranges[i].range;
907 shrink = min_t(u64, to_shrink, range_len(range));
908 if (shrink >= range_len(range)) {
920 && res->start == range->start) {
928 return adjust_dev_dax_range(dev_dax, adjust, range_len(range)
948 if (last->range.start != res->start || last->range.end != res->end)
1064 static ssize_t range_parse(const char *opt, size_t len, struct range *range)
1082 range->start = addr;
1087 range->end = addr;
1100 struct range r;
1146 size_t len = range_len(&dev_dax->ranges[i].range);
1149 dev_dbg(dev, "%s: align %u invalid for range %d\n",
1223 start = dev_dax->ranges[0].range.start;
1395 /* register mapping device for the initial allocation range */
1396 if (dev_dax->nr_range && range_len(&dev_dax->ranges[0].range)) {