Lines Matching refs:tbl

48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl,
52 tbl->it_blocksize = 16;
53 tbl->it_base = (unsigned long)tce_mem;
54 tbl->it_page_shift = page_shift;
55 tbl->it_offset = dma_offset >> tbl->it_page_shift;
56 tbl->it_index = 0;
57 tbl->it_size = tce_size >> 3;
58 tbl->it_busno = 0;
59 tbl->it_type = TCE_PCI;
83 static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc)
85 __be64 *tmp = user ? tbl->it_userspace : (__be64 *) tbl->it_base;
86 int level = tbl->it_indirect_levels;
87 const long shift = ilog2(tbl->it_level_size);
88 unsigned long mask = (tbl->it_level_size - 1) << (level * shift);
100 tmp2 = pnv_alloc_tce_level(tbl->it_nid,
101 ilog2(tbl->it_level_size) + 3);
110 ilog2(tbl->it_level_size) + 3, 1);
124 int pnv_tce_build(struct iommu_table *tbl, long index, long npages,
129 u64 rpn = __pa(uaddr) >> tbl->it_page_shift;
137 ((rpn + i) << tbl->it_page_shift);
138 unsigned long idx = index - tbl->it_offset + i;
140 *(pnv_tce(tbl, false, idx, true)) = cpu_to_be64(newtce);
147 int pnv_tce_xchg(struct iommu_table *tbl, long index,
153 unsigned long idx = index - tbl->it_offset;
156 BUG_ON(*hpa & ~IOMMU_PAGE_MASK(tbl));
159 ptce = pnv_tce(tbl, false, idx, false);
167 ptce = pnv_tce(tbl, false, idx, alloc);
182 __be64 *pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc)
184 if (WARN_ON_ONCE(!tbl->it_userspace))
187 return pnv_tce(tbl, true, index - tbl->it_offset, alloc);
191 void pnv_tce_free(struct iommu_table *tbl, long index, long npages)
196 unsigned long idx = index - tbl->it_offset + i;
197 __be64 *ptce = pnv_tce(tbl, false, idx, false);
203 i |= tbl->it_level_size - 1;
207 unsigned long pnv_tce_get(struct iommu_table *tbl, long index)
209 __be64 *ptce = pnv_tce(tbl, false, index - tbl->it_offset, false);
241 void pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl)
243 const unsigned long size = tbl->it_indirect_levels ?
244 tbl->it_level_size : tbl->it_size;
246 if (!tbl->it_size)
249 pnv_pci_ioda2_table_do_free_pages((__be64 *)tbl->it_base, size,
250 tbl->it_indirect_levels);
251 if (tbl->it_userspace) {
252 pnv_pci_ioda2_table_do_free_pages(tbl->it_userspace, size,
253 tbl->it_indirect_levels);
293 bool alloc_userspace_copy, struct iommu_table *tbl)
348 pnv_pci_setup_iommu_table(tbl, addr, tce_table_size, bus_offset,
350 tbl->it_level_size = 1ULL << (level_shift - 3);
351 tbl->it_indirect_levels = levels - 1;
352 tbl->it_userspace = uas;
353 tbl->it_nid = nid;
356 window_size, tce_table_size, bus_offset, tbl->it_base,
357 tbl->it_userspace, 1, levels);
371 void pnv_pci_unlink_table_and_group(struct iommu_table *tbl,
378 if (!tbl || !table_group)
385 list_for_each_entry_rcu(tgl, &tbl->it_group_list, next) {
401 if (table_group->tables[i] == tbl) {
402 iommu_tce_table_put(tbl);
412 struct iommu_table *tbl,
417 if (WARN_ON(!tbl || !table_group))
426 list_add_rcu(&tgl->next, &tbl->it_group_list);
428 table_group->tables[num] = iommu_tce_table_get(tbl);