Lines Matching refs:dte
177 static inline phys_addr_t rk_dte_pt_address(u32 dte)
179 return (phys_addr_t)dte & RK_DTE_PT_ADDRESS_MASK;
182 static inline bool rk_dte_is_pt_valid(u32 dte)
184 return dte & RK_DTE_PT_VALID;
480 u32 dte;
496 dte = *dte_addr;
498 if (!rk_dte_is_pt_valid(dte))
501 pte_addr_phys = rk_dte_pt_address(dte) + (pte_index * 4);
514 dev_err(iommu->dev, "mmu_dte_addr: %pa dte@%pa: %#08x valid: %u pte@%pa: %#08x valid: %u page@%pa flags: %#03x\n",
515 &mmu_dte_addr_phys, &dte_addr_phys, dte,
516 rk_dte_is_pt_valid(dte), &pte_addr_phys, pte,
595 u32 dte, pte;
600 dte = rk_domain->dt[rk_iova_dte_index(iova)];
601 if (!rk_dte_is_pt_valid(dte))
604 pt_phys = rk_dte_pt_address(dte);
659 u32 dte_index, dte;
667 dte = *dte_addr;
668 if (rk_dte_is_pt_valid(dte))
682 dte = rk_mk_dte(pt_dma);
683 *dte_addr = dte;
689 pt_phys = rk_dte_pt_address(dte);
740 * mapped cachelines holding stale values for its dte and pte.
742 * dte or pte shared with an existing mapping.
777 * aligned, we will always only be mapping from a single dte here.
804 u32 dte;
815 * aligned, we will always only be unmapping from a single dte here.
817 dte = rk_domain->dt[rk_iova_dte_index(iova)];
819 if (!rk_dte_is_pt_valid(dte)) {
824 pt_phys = rk_dte_pt_address(dte);
1038 u32 dte = rk_domain->dt[i];
1039 if (rk_dte_is_pt_valid(dte)) {
1040 phys_addr_t pt_phys = rk_dte_pt_address(dte);