Home
last modified time | relevance | path

Searched refs:tlbie (Results 1 - 25 of 28) sorted by relevance

12

/kernel/linux/linux-5.10/arch/powerpc/mm/book3s64/
H A Dhash_native.c180 asm volatile(ASM_FTR_IFCLR("tlbie %0,0", PPC_TLBIE(%1,%0), %2) in ___tlbie()
199 asm volatile(ASM_FTR_IFCLR("tlbie %0,1", PPC_TLBIE(%1,%0), %2) in ___tlbie()
223 * re-order the tlbie in fixup_tlbie_vpn()
233 /* Need the extra ptesync to ensure we don't reorder tlbie*/ in fixup_tlbie_vpn()
298 static inline void tlbie(unsigned long vpn, int psize, int apsize, in tlbie() function
487 tlbie(vpn, bpsize, apsize, ssize, local); in native_hpte_updatepp()
569 tlbie(vpn, psize, psize, ssize, 0); in native_hpte_updateboltedpp()
599 tlbie(vpn, psize, psize, ssize, 0); in native_hpte_removebolted()
637 tlbie(vpn, bpsize, apsize, ssize, local); in native_hpte_invalidate()
696 * We need to do tlb invalidate for all the address, tlbie in native_hugepage_invalidate()
[all...]
/kernel/linux/linux-6.6/arch/powerpc/mm/book3s64/
H A Dhash_native.c99 asm volatile(ASM_FTR_IFCLR("tlbie %0,0", PPC_TLBIE(%1,%0), %2) in ___tlbie()
118 asm volatile(ASM_FTR_IFCLR("tlbie %0,1", PPC_TLBIE(%1,%0), %2) in ___tlbie()
142 * re-order the tlbie in fixup_tlbie_vpn()
152 /* Need the extra ptesync to ensure we don't reorder tlbie*/ in fixup_tlbie_vpn()
217 static inline void tlbie(unsigned long vpn, int psize, int apsize, in tlbie() function
425 tlbie(vpn, bpsize, apsize, ssize, local); in native_hpte_updatepp()
512 tlbie(vpn, psize, psize, ssize, 0); in native_hpte_updateboltedpp()
547 tlbie(vpn, psize, psize, ssize, 0); in native_hpte_removebolted()
589 tlbie(vpn, bpsize, apsize, ssize, local); in native_hpte_invalidate()
646 * We need to do tlb invalidate for all the address, tlbie in native_hugepage_invalidate()
[all...]
/kernel/linux/linux-5.10/arch/powerpc/mm/book3s32/
H A Dhash_low.S351 tlbie r4
448 * Between the tlbie above and updating the hash table entry below,
456 * PTE in their TLB. So we don't need to bother with another tlbie here,
458 * address. :-) The tlbie above is mainly to make sure that this CPU comes
610 tlbie r4 /* in hw tlb too */
658 tlbie r3
666 tlbie r3
697 0: tlbie r4
/kernel/linux/linux-5.10/arch/powerpc/kernel/
H A Dhead_8xx.S195 tlbie addr; \
197 tlbie addr; \
331 tlbie r4
357 tlbie r4
610 tlbie r0
631 sync /* wait for tlbia/tlbie to finish */
H A Dswsusp_32.S364 tlbie r4
/kernel/linux/linux-6.6/arch/powerpc/kernel/
H A Dhead_8xx.S185 tlbie tmp; \
187 tlbie tmp
317 tlbie r12
342 tlbie r4
584 tlbie r0
601 sync /* wait for tlbia/tlbie to finish */
H A Dswsusp_32.S366 tlbie r4
/kernel/linux/linux-5.10/arch/powerpc/kvm/
H A Dbook3s_pr_papr.c114 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_remove()
202 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_bulk_remove()
247 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_protect()
H A Dbook3s_32_mmu.c404 mmu->tlbie = kvmppc_mmu_book3s_32_tlbie; in kvmppc_mmu_book3s_32_init()
H A Dbook3s_64_mmu.c537 dprintk("KVM MMU: tlbie(0x%lx)\n", va); in kvmppc_mmu_book3s_64_tlbie()
540 * The tlbie instruction changed behaviour starting with in kvmppc_mmu_book3s_64_tlbie()
665 mmu->tlbie = kvmppc_mmu_book3s_64_tlbie; in kvmppc_mmu_book3s_64_init()
H A Dbook3s_emulate.c355 vcpu->arch.mmu.tlbie(vcpu, addr, large); in kvmppc_core_emulate_op_pr()
/kernel/linux/linux-6.6/arch/powerpc/kvm/
H A Dbook3s_pr_papr.c114 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_remove()
202 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_bulk_remove()
247 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_protect()
H A Dbook3s_32_mmu.c404 mmu->tlbie = kvmppc_mmu_book3s_32_tlbie; in kvmppc_mmu_book3s_32_init()
H A Dbook3s_64_mmu.c536 dprintk("KVM MMU: tlbie(0x%lx)\n", va); in kvmppc_mmu_book3s_64_tlbie()
539 * The tlbie instruction changed behaviour starting with in kvmppc_mmu_book3s_64_tlbie()
664 mmu->tlbie = kvmppc_mmu_book3s_64_tlbie; in kvmppc_mmu_book3s_64_init()
H A Dbook3s_emulate.c351 vcpu->arch.mmu.tlbie(vcpu, addr, large); in kvmppc_core_emulate_op_pr()
/kernel/linux/linux-5.10/arch/powerpc/include/asm/
H A Dtrace.h188 TRACE_EVENT(tlbie,
H A Dkvm_host.h412 void (*tlbie)(struct kvm_vcpu *vcpu, ulong addr, bool large); member
H A Dppc_asm.h421 * the 403GCX and 405GP tlbia IS defined and tlbie is not.
433 0: tlbie r4; \
/kernel/linux/linux-6.6/arch/powerpc/mm/book3s32/
H A Dhash_low.S347 tlbie r4
417 * Between the tlbie above and updating the hash table entry below,
425 * PTE in their TLB. So we don't need to bother with another tlbie here,
427 * address. :-) The tlbie above is mainly to make sure that this CPU comes
573 tlbie r4 /* in hw tlb too */
/kernel/linux/linux-6.6/arch/powerpc/include/asm/
H A Dtrace.h291 TRACE_EVENT(tlbie,
H A Dkvm_host.h403 void (*tlbie)(struct kvm_vcpu *vcpu, ulong addr, bool large); member
H A Dppc_asm.h478 * the 403GCX and 405GP tlbia IS defined and tlbie is not.
490 0: tlbie r4; \
/kernel/linux/linux-5.10/arch/powerpc/platforms/powermac/
H A Dsleep.S395 tlbie r4
/kernel/linux/linux-6.6/arch/powerpc/platforms/powermac/
H A Dsleep.S395 tlbie r4
/kernel/linux/linux-5.10/tools/testing/selftests/powerpc/primitives/asm/
H A Dppc_asm.h421 * the 403GCX and 405GP tlbia IS defined and tlbie is not.
433 0: tlbie r4; \

Completed in 22 milliseconds

12