/kernel/linux/linux-5.10/arch/powerpc/sysdev/xive/ |
H A D | spapr.c | 287 unsigned long qpage, in plpar_int_set_queue_config() 292 pr_devel("H_INT_SET_QUEUE_CONFIG flags=%lx target=%lx priority=%lx qpage=%lx qsize=%lx\n", in plpar_int_set_queue_config() 293 flags, target, priority, qpage, qsize); in plpar_int_set_queue_config() 297 priority, qpage, qsize); in plpar_int_set_queue_config() 301 pr_err("H_INT_SET_QUEUE_CONFIG cpu=%ld prio=%ld qpage=%lx returned %ld\n", in plpar_int_set_queue_config() 302 target, priority, qpage, rc); in plpar_int_set_queue_config() 465 __be32 *qpage, u32 order) in xive_spapr_configure_queue() 474 if (WARN_ON(!qpage)) in xive_spapr_configure_queue() 476 qpage_phys = __pa(qpage); in xive_spapr_configure_queue() 507 q->qpage in xive_spapr_configure_queue() 284 plpar_int_set_queue_config(unsigned long flags, unsigned long target, unsigned long priority, unsigned long qpage, unsigned long qsize) plpar_int_set_queue_config() argument 464 xive_spapr_configure_queue(u32 target, struct xive_q *q, u8 prio, __be32 *qpage, u32 order) xive_spapr_configure_queue() argument 520 __be32 *qpage; xive_spapr_setup_queue() local [all...] |
H A D | native.c | 133 __be32 *qpage, u32 order, bool can_escalate) in xive_native_configure_queue() 142 if (WARN_ON(!qpage)) in xive_native_configure_queue() 144 qpage_phys = __pa(qpage); in xive_native_configure_queue() 186 * q->qpage is set due to how it manages IPI EOIs in xive_native_configure_queue() 189 q->qpage = qpage; in xive_native_configure_queue() 220 __be32 *qpage; in xive_native_setup_queue() local 222 qpage = xive_queue_page_alloc(cpu, xive_queue_shift); in xive_native_setup_queue() 223 if (IS_ERR(qpage)) in xive_native_setup_queue() 224 return PTR_ERR(qpage); in xive_native_setup_queue() 132 xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, __be32 *qpage, u32 order, bool can_escalate) xive_native_configure_queue() argument 764 __be64 qpage; xive_native_get_queue_info() local [all...] |
H A D | common.c | 86 if (!q->qpage) in xive_read_eq() 88 cur = be32_to_cpup(q->qpage + q->idx); in xive_read_eq() 232 if (!q->qpage) in xive_dump_eq() 235 i0 = be32_to_cpup(q->qpage + idx); in xive_dump_eq() 237 i1 = be32_to_cpup(q->qpage + idx); in xive_dump_eq() 1331 if (xc->queue[xive_irq_priority].qpage) in xive_cleanup_cpu_queues() 1340 if (!xc->queue[xive_irq_priority].qpage) in xive_setup_cpu_queues() 1556 __be32 *qpage; in xive_queue_page_alloc() local 1562 qpage = (__be32 *)page_address(pages); in xive_queue_page_alloc() 1563 memset(qpage, in xive_queue_page_alloc() [all...] |
/kernel/linux/linux-6.6/arch/powerpc/sysdev/xive/ |
H A D | spapr.c | 301 unsigned long qpage, in plpar_int_set_queue_config() 306 pr_debug("H_INT_SET_QUEUE_CONFIG flags=0x%lx target=%ld priority=0x%lx qpage=0x%lx qsize=0x%lx\n", in plpar_int_set_queue_config() 307 flags, target, priority, qpage, qsize); in plpar_int_set_queue_config() 311 priority, qpage, qsize); in plpar_int_set_queue_config() 315 pr_err("H_INT_SET_QUEUE_CONFIG cpu=%ld prio=%ld qpage=0x%lx returned %ld\n", in plpar_int_set_queue_config() 316 target, priority, qpage, rc); in plpar_int_set_queue_config() 479 __be32 *qpage, u32 order) in xive_spapr_configure_queue() 488 if (WARN_ON(!qpage)) in xive_spapr_configure_queue() 490 qpage_phys = __pa(qpage); in xive_spapr_configure_queue() 521 q->qpage in xive_spapr_configure_queue() 298 plpar_int_set_queue_config(unsigned long flags, unsigned long target, unsigned long priority, unsigned long qpage, unsigned long qsize) plpar_int_set_queue_config() argument 478 xive_spapr_configure_queue(u32 target, struct xive_q *q, u8 prio, __be32 *qpage, u32 order) xive_spapr_configure_queue() argument 534 __be32 *qpage; xive_spapr_setup_queue() local [all...] |
H A D | native.c | 132 __be32 *qpage, u32 order, bool can_escalate) in xive_native_configure_queue() 141 if (WARN_ON(!qpage)) in xive_native_configure_queue() 143 qpage_phys = __pa(qpage); in xive_native_configure_queue() 185 * q->qpage is set due to how it manages IPI EOIs in xive_native_configure_queue() 188 q->qpage = qpage; in xive_native_configure_queue() 219 __be32 *qpage; in xive_native_setup_queue() local 221 qpage = xive_queue_page_alloc(cpu, xive_queue_shift); in xive_native_setup_queue() 222 if (IS_ERR(qpage)) in xive_native_setup_queue() 223 return PTR_ERR(qpage); in xive_native_setup_queue() 131 xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, __be32 *qpage, u32 order, bool can_escalate) xive_native_configure_queue() argument 788 __be64 qpage; xive_native_get_queue_info() local [all...] |
H A D | common.c | 107 if (!q->qpage) in xive_read_eq() 109 cur = be32_to_cpup(q->qpage + q->idx); in xive_read_eq() 260 if (!q->qpage) in xive_dump_eq() 263 i0 = be32_to_cpup(q->qpage + idx); in xive_dump_eq() 265 i1 = be32_to_cpup(q->qpage + idx); in xive_dump_eq() 1475 if (xc->queue[xive_irq_priority].qpage) in xive_cleanup_cpu_queues() 1484 if (!xc->queue[xive_irq_priority].qpage) in xive_setup_cpu_queues() 1696 __be32 *qpage; in xive_queue_page_alloc() local 1702 qpage = (__be32 *)page_address(pages); in xive_queue_page_alloc() 1703 memset(qpage, in xive_queue_page_alloc() [all...] |
/kernel/linux/linux-5.10/arch/powerpc/kvm/ |
H A D | book3s_xive_template.c | 125 __be32 *qpage; in GLUE() local 150 * since qpage can be set concurrently and we don't want in GLUE() 153 qpage = READ_ONCE(q->qpage); in GLUE() 158 * non-queueing priority (ie, qpage = 0). in GLUE() 160 hirq = __xive_read_eq(qpage, q->msk, &idx, &toggle); in GLUE() 176 if (hirq == XICS_IPI || (prio == 0 && !qpage)) { in GLUE() 388 __be32 *qpage; in GLUE() local 393 qpage = READ_ONCE(q->qpage); in GLUE() [all...] |
H A D | book3s_xive.h | 259 static inline u32 __xive_read_eq(__be32 *qpage, u32 msk, u32 *idx, u32 *toggle) in __xive_read_eq() argument 263 if (!qpage) in __xive_read_eq() 265 cur = be32_to_cpup(qpage + *idx); in __xive_read_eq()
|
H A D | book3s_xive.c | 250 void *qpage; in xive_provision_queue() local 253 if (WARN_ON(q->qpage)) in xive_provision_queue() 257 qpage = (__be32 *)__get_free_pages(GFP_KERNEL, xive->q_page_order); in xive_provision_queue() 258 if (!qpage) { in xive_provision_queue() 263 memset(qpage, 0, 1 << xive->q_order); in xive_provision_queue() 266 * Reconfigure the queue. This will set q->qpage only once the in xive_provision_queue() 269 * qpage being non-NULL, and instead will only EOI when we receive in xive_provision_queue() 272 rc = xive_native_configure_queue(xc->vp_id, q, prio, qpage, in xive_provision_queue() 345 if (WARN_ON(!q->qpage)) in xive_try_pick_queue() 1194 if (q->qpage) { in kvmppc_xive_cleanup_vcpu() [all...] |
H A D | book3s_xive_native.c | 53 if (q->qpage) { in kvmppc_xive_native_cleanup_queue() 54 put_page(virt_to_page(q->qpage)); in kvmppc_xive_native_cleanup_queue() 55 q->qpage = NULL; in kvmppc_xive_native_cleanup_queue() 60 u8 prio, __be32 *qpage, in kvmppc_xive_native_configure_queue() 64 __be32 *qpage_prev = q->qpage; in kvmppc_xive_native_configure_queue() 66 rc = xive_native_configure_queue(vp_id, q, prio, qpage, order, in kvmppc_xive_native_configure_queue() 747 if (!q->qpage) in kvmppc_xive_native_get_queue_config() 901 if (!q->qpage) in kvmppc_xive_native_vcpu_eq_sync() 59 kvmppc_xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, __be32 *qpage, u32 order, bool can_escalate) kvmppc_xive_native_configure_queue() argument
|
/kernel/linux/linux-6.6/arch/powerpc/kvm/ |
H A D | book3s_xive.c | 143 __be32 *qpage; in xive_vm_scan_interrupts() local 168 * since qpage can be set concurrently and we don't want in xive_vm_scan_interrupts() 171 qpage = READ_ONCE(q->qpage); in xive_vm_scan_interrupts() 176 * non-queueing priority (ie, qpage = 0). in xive_vm_scan_interrupts() 178 hirq = __xive_read_eq(qpage, q->msk, &idx, &toggle); in xive_vm_scan_interrupts() 194 if (hirq == XICS_IPI || (prio == 0 && !qpage)) { in xive_vm_scan_interrupts() 402 __be32 *qpage; in xive_vm_scan_for_rerouted_irqs() local 407 qpage = READ_ONCE(q->qpage); in xive_vm_scan_for_rerouted_irqs() 941 void *qpage; xive_provision_queue() local [all...] |
H A D | book3s_xive.h | 273 static inline u32 __xive_read_eq(__be32 *qpage, u32 msk, u32 *idx, u32 *toggle) in __xive_read_eq() argument 277 if (!qpage) in __xive_read_eq() 279 cur = be32_to_cpup(qpage + *idx); in __xive_read_eq()
|
H A D | book3s_xive_native.c | 50 if (q->qpage) { in kvmppc_xive_native_cleanup_queue() 51 put_page(virt_to_page(q->qpage)); in kvmppc_xive_native_cleanup_queue() 52 q->qpage = NULL; in kvmppc_xive_native_cleanup_queue() 57 u8 prio, __be32 *qpage, in kvmppc_xive_native_configure_queue() 61 __be32 *qpage_prev = q->qpage; in kvmppc_xive_native_configure_queue() 63 rc = xive_native_configure_queue(vp_id, q, prio, qpage, order, in kvmppc_xive_native_configure_queue() 749 if (!q->qpage) in kvmppc_xive_native_get_queue_config() 903 if (!q->qpage) in kvmppc_xive_native_vcpu_eq_sync() 56 kvmppc_xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, __be32 *qpage, u32 order, bool can_escalate) kvmppc_xive_native_configure_queue() argument
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | xive.h | 75 __be32 *qpage; member 117 __be32 *qpage, u32 order, bool can_escalate);
|
H A D | opal.h | 262 uint64_t qpage,
|
/kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
H A D | xive.h | 75 __be32 *qpage; member 119 __be32 *qpage, u32 order, bool can_escalate);
|
H A D | opal.h | 256 uint64_t qpage,
|