Lines Matching defs:range
50 u64 (*map)(__be32 *addr, const __be32 *range,
70 static u64 of_bus_default_map(__be32 *addr, const __be32 *range,
75 cp = of_read_number(range, na);
76 s = of_read_number(range + na + pna, ns);
166 static u64 of_bus_pci_map(__be32 *addr, const __be32 *range, int na, int ns,
173 rf = of_bus_pci_get_flags(range);
180 cp = of_read_number(range + 1, na - 1);
181 s = of_read_number(range + na + pna, ns);
258 * @range: the PCI range that describes the resource
259 * @np: device node where the range belongs to
261 * reflect the values contained in the range.
263 * Returns EINVAL if the range cannot be converted to resource.
265 * Note that if the range is an IO range, the resource will be converted
267 * if the range cannot be matched to any host bridge IO space (our case here).
268 * To guard against that we try to register the IO range first.
271 int of_pci_range_to_resource(struct of_pci_range *range,
275 res->flags = range->flags;
281 err = pci_register_io_range(&np->fwnode, range->cpu_addr,
282 range->size);
285 port = pci_address_to_pio(range->cpu_addr);
293 upper_32_bits(range->cpu_addr)) {
298 res->start = range->cpu_addr;
300 res->end = res->start + range->size - 1;
329 static u64 of_bus_isa_map(__be32 *addr, const __be32 *range, int na, int ns,
335 if ((addr[0] ^ range[0]) & cpu_to_be32(1))
339 cp = of_read_number(range + 1, na - 1);
340 s = of_read_number(range + na + pna, ns);
724 parser->range = of_get_property(node, name, &rlen);
725 if (parser->range == NULL)
728 parser->end = parser->range + rlen / sizeof(__be32);
749 struct of_pci_range *range)
756 if (!range)
759 if (!parser->range || parser->range + np > parser->end)
762 range->flags = parser->bus->get_flags(parser->range);
768 range->bus_addr = of_read_number(parser->range + busflag_na, na - busflag_na);
771 range->cpu_addr = of_translate_dma_address(parser->node,
772 parser->range + na);
774 range->cpu_addr = of_translate_address(parser->node,
775 parser->range + na);
776 range->size = of_read_number(parser->range + parser->pna + na, ns);
778 parser->range += np;
781 while (parser->range + np <= parser->end) {
785 flags = parser->bus->get_flags(parser->range);
786 bus_addr = of_read_number(parser->range + busflag_na, na - busflag_na);
789 parser->range + na);
792 parser->range + na);
793 size = of_read_number(parser->range + parser->pna + na, ns);
795 if (flags != range->flags)
797 if (bus_addr != range->bus_addr + range->size ||
798 cpu_addr != range->cpu_addr + range->size)
801 range->size += size;
802 parser->range += np;
805 return range;
823 /* memory-mapped I/O range */
887 * @np: the device whose io range will be mapped
888 * @index: index of the io range
906 * @device: the device whose io range will be mapped
907 * @index: index of the io range
943 * of_dma_get_range - Get DMA range info and put it into a map array
944 * @np: device node to get DMA range info
945 * @map: dma range structure to return
964 struct of_range range;
993 for_each_of_range(&parser, &range) {
994 if (range.cpu_addr == OF_BAD_ADDR) {
996 range.bus_addr, node);
1019 for_each_of_range(&parser, &range) {
1021 range.bus_addr, range.cpu_addr, range.size);
1022 if (range.cpu_addr == OF_BAD_ADDR)
1024 r->cpu_start = range.cpu_addr;
1025 r->dma_start = range.bus_addr;
1026 r->size = range.size;
1027 r->offset = range.cpu_addr - range.bus_addr;
1050 struct of_range range;
1061 for_each_of_range(&parser, &range)
1062 if (range.cpu_addr + range.size > cpu_end)
1063 cpu_end = range.cpu_addr + range.size - 1;