Lines Matching refs:ndd

298 	struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
305 for_each_dpa_resource(ndd, res)
314 nd_dbg_dpa(nd_region, ndd, res, "delete %d\n", rc);
315 nvdimm_free_dpa(ndd, res);
323 nd_dbg_dpa(nd_region, ndd, res, "shrink %d\n", rc);
362 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
367 res = nvdimm_allocate_dpa(ndd, label_id, nd_mapping->start, n);
371 nd_dbg_dpa(nd_region, ndd, res, "init %d\n", rc);
379 * @ndd: dimm device data for debug
392 static void space_valid(struct nd_region *nd_region, struct nvdimm_drvdata *ndd,
437 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
442 for_each_dpa_resource(ndd, res)
451 for_each_dpa_resource(ndd, res) {
468 space_valid(nd_region, ndd, label_id, NULL, next, exist,
479 space_valid(nd_region, ndd, label_id, res, next, exist,
490 space_valid(nd_region, ndd, label_id, res, next, exist,
534 new_res = nvdimm_allocate_dpa(ndd, label_id,
549 nd_dbg_dpa(nd_region, ndd, new_res, "%s(%d) %d\n",
577 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
583 for_each_dpa_resource(ndd, res) {
593 nvdimm_free_dpa(ndd, next);
595 nd_dbg_dpa(nd_region, ndd, res, "merge %d\n", rc);
645 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
648 for_each_dpa_resource_safe(ndd, res, _res)
650 nvdimm_free_dpa(ndd, res);
705 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
709 if (!ndd) {
717 for_each_dpa_resource(ndd, res)
749 struct nvdimm_drvdata *ndd;
786 ndd = to_ndd(nd_mapping);
792 if (!ndd)
795 allocated += nvdimm_allocated_dpa(ndd, &label_id);
993 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
997 for_each_dpa_resource(ndd, res)
1010 nsl_get_uuid(ndd, nd_label, &uuid);
1012 nsl_get_flags(ndd, nd_label));
1154 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1157 for_each_dpa_resource(ndd, res)
1175 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1182 if (!ndd) {
1187 nsindex = to_namespace_index(ndd, ndd->ns_current);
1563 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1573 position = nsl_get_position(ndd, nd_label);
1575 if (!nsl_validate_isetcookie(ndd, nd_label, cookie))
1578 if (!nsl_uuid_equal(ndd, nd_label, uuid))
1581 if (!nsl_validate_type_guid(ndd, nd_label,
1586 dev_dbg(ndd->dev, "duplicate entry for uuid\n");
1590 if (!nsl_validate_nlabel(nd_region, ndd, nd_label))
1612 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1622 if (nsl_uuid_equal(ndd, nd_label, pmem_id))
1638 pmem_start = nsl_get_dpa(ndd, nd_label);
1639 pmem_end = pmem_start + nsl_get_rawsize(ndd, nd_label);
1645 dev_name(ndd->dev),
1646 nsl_uuid_raw(ndd, nd_label));
1666 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1668 to_namespace_index(ndd, ndd->ns_current);
1685 if (!nsl_validate_isetcookie(ndd, nd_label, cookie)) {
1687 nsl_uuid_raw(ndd, nd_label));
1688 if (!nsl_validate_isetcookie(ndd, nd_label, altcookie))
1692 nsl_uuid_raw(ndd, nd_label));
1708 nsl_get_uuid(ndd, nd_label, &uuid);
1725 nvdimm_name(nvdimm), nsl_uuid_raw(ndd, nd_label));
1735 nsl_get_uuid(ndd, nd_label, &uuid);
1743 struct nvdimm_drvdata *ndd;
1755 ndd = to_ndd(nd_mapping);
1756 size += nsl_get_rawsize(ndd, label0);
1757 if (nsl_get_position(ndd, label0) != 0)
1760 nspm->alt_name = kmemdup(nsl_ref_name(ndd, label0),
1762 nsl_get_uuid(ndd, label0, &uuid);
1764 nspm->lbasize = nsl_get_lbasize(ndd, label0);
1766 nsl_get_claim_class(ndd, label0);
1890 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1901 if (!nsl_uuid_equal(ndd, nd_label, uuid))
1933 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
1945 if (nsl_get_dpa(ndd, nd_label) < nd_mapping->start ||
1946 nsl_get_dpa(ndd, nd_label) > map_end)
2071 struct nvdimm_drvdata *ndd = nd_mapping->ndd;
2078 put_ndd(ndd);
2079 nd_mapping->ndd = NULL;
2080 if (ndd)
2091 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping);
2100 if (!ndd) {
2115 nd_mapping->ndd = ndd;
2117 get_ndd(ndd);
2119 count = nd_label_active_count(ndd);
2120 dev_dbg(ndd->dev, "count: %d\n", count);
2129 label = nd_label_active(ndd, j);