Lines Matching defs:nd_pfn
21 struct nd_pfn *nd_pfn = to_nd_pfn(dev);
24 nd_detach_ndns(&nd_pfn->dev, &nd_pfn->ndns);
25 ida_simple_remove(&nd_region->pfn_ida, nd_pfn->id);
26 kfree(nd_pfn->uuid);
27 kfree(nd_pfn);
30 struct nd_pfn *to_nd_pfn(struct device *dev)
32 struct nd_pfn *nd_pfn = container_of(dev, struct nd_pfn, dev);
35 return nd_pfn;
42 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
44 switch (nd_pfn->mode) {
57 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
69 nd_pfn->mode = PFN_MODE_PMEM;
72 nd_pfn->mode = PFN_MODE_RAM;
75 nd_pfn->mode = PFN_MODE_NONE;
91 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
93 return sprintf(buf, "%ld\n", nd_pfn->align);
124 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
130 rc = nd_size_select_store(dev, buf, &nd_pfn->align,
144 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
146 if (nd_pfn->uuid)
147 return sprintf(buf, "%pUb\n", nd_pfn->uuid);
154 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
158 rc = nd_uuid_store(dev, &nd_pfn->uuid, buf, len);
170 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
174 rc = sprintf(buf, "%s\n", nd_pfn->ndns
175 ? dev_name(&nd_pfn->ndns->dev) : "");
183 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
188 rc = nd_namespace_store(dev, &nd_pfn->ndns, buf, len);
201 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
206 struct nd_pfn_sb *pfn_sb = nd_pfn->pfn_sb;
208 struct nd_namespace_common *ndns = nd_pfn->ndns;
227 struct nd_pfn *nd_pfn = to_nd_pfn_safe(dev);
232 struct nd_pfn_sb *pfn_sb = nd_pfn->pfn_sb;
234 struct nd_namespace_common *ndns = nd_pfn->ndns;
285 .name = "nd_pfn",
298 struct device *nd_pfn_devinit(struct nd_pfn *nd_pfn,
303 if (!nd_pfn)
306 nd_pfn->mode = PFN_MODE_NONE;
307 nd_pfn->align = nd_pfn_default_alignment();
308 dev = &nd_pfn->dev;
309 device_initialize(&nd_pfn->dev);
310 lockdep_set_class(&nd_pfn->dev.mutex, &nvdimm_pfn_key);
311 if (ndns && !__nd_attach_ndns(&nd_pfn->dev, ndns, &nd_pfn->ndns)) {
320 static struct nd_pfn *nd_pfn_alloc(struct nd_region *nd_region)
322 struct nd_pfn *nd_pfn;
325 nd_pfn = kzalloc(sizeof(*nd_pfn), GFP_KERNEL);
326 if (!nd_pfn)
329 nd_pfn->id = ida_simple_get(&nd_region->pfn_ida, 0, 0, GFP_KERNEL);
330 if (nd_pfn->id < 0) {
331 kfree(nd_pfn);
335 dev = &nd_pfn->dev;
336 dev_set_name(dev, "pfn%d.%d", nd_region->id, nd_pfn->id);
340 return nd_pfn;
345 struct nd_pfn *nd_pfn;
351 nd_pfn = nd_pfn_alloc(nd_region);
352 dev = nd_pfn_devinit(nd_pfn, NULL);
364 static int nd_pfn_clear_memmap_errors(struct nd_pfn *nd_pfn)
366 struct nd_region *nd_region = to_nd_region(nd_pfn->dev.parent);
367 struct nd_namespace_common *ndns = nd_pfn->ndns;
369 struct nd_pfn_sb *pfn_sb = nd_pfn->pfn_sb;
374 if (nd_pfn->mode != PFN_MODE_PMEM)
385 devm_namespace_disable(&nd_pfn->dev, ndns);
386 rc = devm_namespace_enable(&nd_pfn->dev, ndns, le64_to_cpu(pfn_sb->dataoff));
397 dev_dbg(&nd_pfn->dev, "meta: %x badblocks at %llx\n",
415 dev_err(&nd_pfn->dev,
443 * @nd_pfn: fsdax namespace runtime state / properties
450 int nd_pfn_validate(struct nd_pfn *nd_pfn, const char *sig)
458 struct nd_pfn_sb *pfn_sb = nd_pfn->pfn_sb;
459 struct nd_namespace_common *ndns = nd_pfn->ndns;
465 if (!is_memory(nd_pfn->dev.parent))
514 dev_err(&nd_pfn->dev,
522 dev_err(&nd_pfn->dev,
535 dev_err(&nd_pfn->dev, "init failed, alignment mismatch: "
536 "%ld:%ld\n", nd_pfn->align, align);
540 if (!nd_pfn->uuid) {
546 nd_pfn->uuid = kmemdup(pfn_sb->uuid, 16, GFP_KERNEL);
547 if (!nd_pfn->uuid)
549 nd_pfn->align = align;
550 nd_pfn->mode = mode;
556 if (memcmp(nd_pfn->uuid, pfn_sb->uuid, 16) != 0)
565 if (nd_pfn->align != align || nd_pfn->mode != mode) {
566 dev_err(&nd_pfn->dev,
568 dev_dbg(&nd_pfn->dev, "align: %lx:%lx mode: %d:%d\n",
569 nd_pfn->align, align, nd_pfn->mode,
576 dev_err(&nd_pfn->dev, "alignment: %lx exceeds capacity %llx\n",
591 dev_err(&nd_pfn->dev, "pfn array size exceeds capacity of %s\n",
598 dev_err(&nd_pfn->dev,
605 dev_err(&nd_pfn->dev, "resource start misaligned\n");
610 dev_err(&nd_pfn->dev, "resource end misaligned\n");
615 dev_err(&nd_pfn->dev, "bad offset with small namespace\n");
625 struct nd_pfn *nd_pfn;
642 nd_pfn = nd_pfn_alloc(nd_region);
643 pfn_dev = nd_pfn_devinit(nd_pfn, ndns);
648 nd_pfn = to_nd_pfn(pfn_dev);
649 nd_pfn->pfn_sb = pfn_sb;
650 rc = nd_pfn_validate(nd_pfn, PFN_SIG);
653 nd_detach_ndns(pfn_dev, &nd_pfn->ndns);
682 static int __nvdimm_setup_pfn(struct nd_pfn *nd_pfn, struct dev_pagemap *pgmap)
686 struct nd_pfn_sb *pfn_sb = nd_pfn->pfn_sb;
691 struct nd_namespace_common *ndns = nd_pfn->ndns;
706 if (nd_pfn->mode == PFN_MODE_RAM) {
709 nd_pfn->npfns = le64_to_cpu(pfn_sb->npfns);
710 } else if (nd_pfn->mode == PFN_MODE_PMEM) {
711 nd_pfn->npfns = PHYS_PFN((range_len(range) - offset));
712 if (le64_to_cpu(nd_pfn->pfn_sb->npfns) > nd_pfn->npfns)
713 dev_info(&nd_pfn->dev,
715 le64_to_cpu(nd_pfn->pfn_sb->npfns),
716 nd_pfn->npfns);
727 static int nd_pfn_init(struct nd_pfn *nd_pfn)
729 struct nd_namespace_common *ndns = nd_pfn->ndns;
741 pfn_sb = devm_kmalloc(&nd_pfn->dev, sizeof(*pfn_sb), GFP_KERNEL);
745 nd_pfn->pfn_sb = pfn_sb;
746 if (is_nd_dax(&nd_pfn->dev))
751 rc = nd_pfn_validate(nd_pfn, sig);
753 return nd_pfn_clear_memmap_errors(nd_pfn);
760 nd_region = to_nd_region(nd_pfn->dev.parent);
762 dev_info(&nd_pfn->dev,
771 align = max(nd_pfn->align, memremap_compat_align());
779 dev_err(&nd_pfn->dev, "%s: start %pa misaligned to %#lx\n",
785 if (nd_pfn->mode == PFN_MODE_PMEM) {
807 dev_err(&nd_pfn->dev,
813 } else if (nd_pfn->mode == PFN_MODE_RAM)
820 dev_err(&nd_pfn->dev, "%s unable to satisfy requested alignment\n",
826 pfn_sb->mode = cpu_to_le32(nd_pfn->mode);
830 memcpy(pfn_sb->uuid, nd_pfn->uuid, 16);
835 pfn_sb->align = cpu_to_le32(nd_pfn->align);
844 rc = nd_pfn_clear_memmap_errors(nd_pfn);
852 * Determine the effective resource range and vmem_altmap from an nd_pfn
855 int nvdimm_setup_pfn(struct nd_pfn *nd_pfn, struct dev_pagemap *pgmap)
859 if (!nd_pfn->uuid || !nd_pfn->ndns)
862 rc = nd_pfn_init(nd_pfn);
867 return __nvdimm_setup_pfn(nd_pfn, pgmap);