Home
last modified time | relevance | path

Searched refs:rhp (Results 1 - 25 of 69) sorted by relevance

123

/kernel/linux/linux-5.10/include/trace/events/
H A Drcu.h485 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen),
487 TP_ARGS(rcuname, rhp, qlen),
491 __field(void *, rhp)
498 __entry->rhp = rhp;
499 __entry->func = rhp->func;
503 TP_printk("%s rhp=%p func=%ps %ld",
504 __entry->rcuname, __entry->rhp, __entry->func,
518 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset,
521 TP_ARGS(rcuname, rhp, offse
[all...]
/kernel/linux/linux-5.10/kernel/rcu/
H A Drcu_segcblist.c28 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp) in rcu_cblist_enqueue() argument
30 *rclp->tail = rhp; in rcu_cblist_enqueue()
31 rclp->tail = &rhp->next; in rcu_cblist_enqueue()
37 * any contents of the first. If rhp is non-NULL, enqueue it as the sole
45 struct rcu_head *rhp) in rcu_cblist_flush_enqueue()
53 if (!rhp) { in rcu_cblist_flush_enqueue()
56 rhp->next = NULL; in rcu_cblist_flush_enqueue()
57 srclp->head = rhp; in rcu_cblist_flush_enqueue()
58 srclp->tail = &rhp->next; in rcu_cblist_flush_enqueue()
69 struct rcu_head *rhp; in rcu_cblist_dequeue() local
43 rcu_cblist_flush_enqueue(struct rcu_cblist *drclp, struct rcu_cblist *srclp, struct rcu_head *rhp) rcu_cblist_flush_enqueue() argument
244 rcu_segcblist_enqueue(struct rcu_segcblist *rsclp, struct rcu_head *rhp) rcu_segcblist_enqueue() argument
264 rcu_segcblist_entrain(struct rcu_segcblist *rsclp, struct rcu_head *rhp) rcu_segcblist_entrain() argument
[all...]
H A Dsrcutiny.c116 struct rcu_head *rhp; in srcu_drive_gp() local
139 rhp = lh; in srcu_drive_gp()
142 rhp->func(rhp); in srcu_drive_gp()
178 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument
183 rhp->func = func; in call_srcu()
184 rhp->next = NULL; in call_srcu()
186 *ssp->srcu_cb_tail = rhp; in call_srcu()
187 ssp->srcu_cb_tail = &rhp->next; in call_srcu()
H A Drcu_segcblist.h19 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp);
22 struct rcu_head *rhp);
88 struct rcu_head *rhp);
90 struct rcu_head *rhp);
H A Dtasks.h152 static void call_rcu_tasks_generic(struct rcu_head *rhp, rcu_callback_t func, in call_rcu_tasks_generic() argument
158 rhp->next = NULL; in call_rcu_tasks_generic()
159 rhp->func = func; in call_rcu_tasks_generic()
162 WRITE_ONCE(*rtp->cbs_tail, rhp); in call_rcu_tasks_generic()
163 rtp->cbs_tail = &rhp->next; in call_rcu_tasks_generic()
509 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func);
514 * @rhp: structure to be used for queueing the RCU updates.
530 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) in call_rcu_tasks() argument
532 call_rcu_tasks_generic(rhp, func, &rcu_tasks); in call_rcu_tasks()
656 void call_rcu_tasks_rude(struct rcu_head *rhp, rcu_callback_
678 call_rcu_tasks_rude(struct rcu_head *rhp, rcu_callback_t func) call_rcu_tasks_rude() argument
1171 call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func) call_rcu_tasks_trace() argument
[all...]
H A Dupdate.c467 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument
471 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read()
475 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \
534 struct early_boot_kfree_rcu *rhp; in early_boot_test_call_rcu() local
539 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in early_boot_test_call_rcu()
540 if (!WARN_ON_ONCE(!rhp)) in early_boot_test_call_rcu()
541 kfree_rcu(rhp, rh); in early_boot_test_call_rcu()
H A Dsrcutree.c809 static void srcu_leak_callback(struct rcu_head *rhp) in srcu_leak_callback() argument
817 struct rcu_head *rhp, bool do_norm) in srcu_gp_start_if_needed()
830 if (rhp) in srcu_gp_start_if_needed()
831 rcu_segcblist_enqueue(&sdp->srcu_cblist, rhp); in srcu_gp_start_if_needed()
881 static void __call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in __call_srcu() argument
884 if (debug_rcu_head_queue(rhp)) { in __call_srcu()
886 WRITE_ONCE(rhp->func, srcu_leak_callback); in __call_srcu()
890 rhp->func = func; in __call_srcu()
891 (void)srcu_gp_start_if_needed(ssp, rhp, do_norm); in __call_srcu()
897 * @rhp
816 srcu_gp_start_if_needed(struct srcu_struct *ssp, struct rcu_head *rhp, bool do_norm) srcu_gp_start_if_needed() argument
911 call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, rcu_callback_t func) call_srcu() argument
1079 srcu_barrier_cb(struct rcu_head *rhp) srcu_barrier_cb() argument
1235 struct rcu_head *rhp; srcu_invoke_callbacks() local
[all...]
H A Drcuscale.c348 static void rcu_scale_async_cb(struct rcu_head *rhp) in rcu_scale_async_cb() argument
351 kfree(rhp); in rcu_scale_async_cb()
363 struct rcu_head *rhp = NULL; in rcu_scale_writer() local
403 if (!rhp) in rcu_scale_writer()
404 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_scale_writer()
405 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_scale_writer()
407 cur_ops->async(rhp, rcu_scale_async_cb); in rcu_scale_writer()
408 rhp = NULL; in rcu_scale_writer()
413 kfree(rhp); /* Becaus in rcu_scale_writer()
[all...]
H A Dsync.c43 static void rcu_sync_func(struct rcu_head *rhp);
52 * @rhp: Pointer to rcu_head in rcu_sync structure to use for synchronization
73 static void rcu_sync_func(struct rcu_head *rhp) in rcu_sync_func() argument
75 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head); in rcu_sync_func()
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/
H A Dmem.c388 return xa_insert_irq(&mhp->rhp->mrs, mmid, mhp, GFP_KERNEL); in finish_mem_reg()
391 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument
397 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem()
410 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem()
419 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl()
432 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local
440 rhp = php->rhp; in c4iw_get_dma_mr()
458 mhp->rhp = rhp; in c4iw_get_dma_mr()
498 struct c4iw_dev *rhp; c4iw_reg_user_mr() local
601 struct c4iw_dev *rhp; c4iw_alloc_mr() local
702 struct c4iw_dev *rhp; c4iw_dereg_mr() local
729 c4iw_invalidate_mr(struct c4iw_dev *rhp, u32 rkey) c4iw_invalidate_mr() argument
[all...]
H A Dprovider.c65 struct c4iw_dev *rhp; in c4iw_dealloc_ucontext() local
69 rhp = to_c4iw_dev(ucontext->ibucontext.device); in c4iw_dealloc_ucontext()
73 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext()
81 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local
87 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext()
93 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext()
114 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext()
196 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local
200 rhp = php->rhp; in c4iw_deallocate_pd()
214 struct c4iw_dev *rhp; c4iw_allocate_pd() local
[all...]
H A Dqp.c720 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl()
744 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl()
807 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); in build_tpte_memreg()
916 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db()
918 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db()
921 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_sq_db()
925 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db()
933 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db()
935 if (qhp->rhp->db_state == NORMAL) in ring_kernel_rq_db()
938 add_to_fc_list(&qhp->rhp in ring_kernel_rq_db()
1085 struct c4iw_dev *rhp; c4iw_post_send() local
1692 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) rdma_fini() argument
1749 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) rdma_init() argument
1832 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) c4iw_modify_qp() argument
2072 struct c4iw_dev *rhp; c4iw_destroy_qp() local
2113 struct c4iw_dev *rhp; c4iw_create_qp() local
2359 struct c4iw_dev *rhp; c4iw_ib_modify_qp() local
2666 struct c4iw_dev *rhp; c4iw_create_srq() local
2798 struct c4iw_dev *rhp; c4iw_destroy_srq() local
[all...]
H A Dcq.c351 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq()
803 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); in __c4iw_poll_cq_one()
844 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one()
933 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one()
984 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); in c4iw_destroy_cq()
990 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq()
1003 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); in c4iw_create_cq() local
1020 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
1056 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq()
1078 ret = create_cq(&rhp in c4iw_create_cq()
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/
H A Dmem.c404 return xa_insert_irq(&mhp->rhp->mrs, mmid, mhp, GFP_KERNEL); in finish_mem_reg()
407 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument
413 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem()
426 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem()
435 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl()
448 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local
456 rhp = php->rhp; in c4iw_get_dma_mr()
474 mhp->rhp = rhp; in c4iw_get_dma_mr()
514 struct c4iw_dev *rhp; c4iw_reg_user_mr() local
617 struct c4iw_dev *rhp; c4iw_alloc_mw() local
667 struct c4iw_dev *rhp; c4iw_dealloc_mw() local
685 struct c4iw_dev *rhp; c4iw_alloc_mr() local
786 struct c4iw_dev *rhp; c4iw_dereg_mr() local
813 c4iw_invalidate_mr(struct c4iw_dev *rhp, u32 rkey) c4iw_invalidate_mr() argument
[all...]
H A Dprovider.c64 struct c4iw_dev *rhp; in c4iw_dealloc_ucontext() local
68 rhp = to_c4iw_dev(ucontext->ibucontext.device); in c4iw_dealloc_ucontext()
72 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext()
80 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local
86 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext()
92 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext()
113 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext()
195 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local
199 rhp = php->rhp; in c4iw_deallocate_pd()
213 struct c4iw_dev *rhp; c4iw_allocate_pd() local
[all...]
H A Dqp.c720 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl()
744 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl()
807 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); in build_tpte_memreg()
916 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db()
918 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db()
921 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_sq_db()
925 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db()
933 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db()
935 if (qhp->rhp->db_state == NORMAL) in ring_kernel_rq_db()
938 add_to_fc_list(&qhp->rhp in ring_kernel_rq_db()
1085 struct c4iw_dev *rhp; c4iw_post_send() local
1692 rdma_fini(struct c4iw_dev *rhp, struct c4iw_qp *qhp, struct c4iw_ep *ep) rdma_fini() argument
1749 rdma_init(struct c4iw_dev *rhp, struct c4iw_qp *qhp) rdma_init() argument
1832 c4iw_modify_qp(struct c4iw_dev *rhp, struct c4iw_qp *qhp, enum c4iw_qp_attr_mask mask, struct c4iw_qp_attributes *attrs, int internal) c4iw_modify_qp() argument
2073 struct c4iw_dev *rhp; c4iw_destroy_qp() local
2115 struct c4iw_dev *rhp; c4iw_create_qp() local
2371 struct c4iw_dev *rhp; c4iw_ib_modify_qp() local
2675 struct c4iw_dev *rhp; c4iw_create_srq() local
2804 struct c4iw_dev *rhp; c4iw_destroy_srq() local
[all...]
H A Dcq.c351 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq()
803 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); in __c4iw_poll_cq_one()
844 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one()
933 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one()
978 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); in c4iw_destroy_cq()
984 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq()
997 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); in c4iw_create_cq() local
1014 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
1050 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq()
1072 ret = create_cq(&rhp in c4iw_create_cq()
[all...]
/kernel/linux/linux-6.6/include/trace/events/
H A Drcu.h514 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen),
516 TP_ARGS(rcuname, rhp, qlen),
520 __field(void *, rhp)
527 __entry->rhp = rhp;
528 __entry->func = rhp->func;
532 TP_printk("%s rhp=%p func=%ps %ld",
533 __entry->rcuname, __entry->rhp, __entry->func,
573 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset,
576 TP_ARGS(rcuname, rhp, offse
[all...]
/kernel/linux/linux-6.6/kernel/rcu/
H A Drcu_segcblist.c28 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp) in rcu_cblist_enqueue() argument
30 *rclp->tail = rhp; in rcu_cblist_enqueue()
31 rclp->tail = &rhp->next; in rcu_cblist_enqueue()
37 * any contents of the first. If rhp is non-NULL, enqueue it as the sole
45 struct rcu_head *rhp) in rcu_cblist_flush_enqueue()
53 if (!rhp) { in rcu_cblist_flush_enqueue()
56 rhp->next = NULL; in rcu_cblist_flush_enqueue()
57 srclp->head = rhp; in rcu_cblist_flush_enqueue()
58 srclp->tail = &rhp->next; in rcu_cblist_flush_enqueue()
69 struct rcu_head *rhp; in rcu_cblist_dequeue() local
43 rcu_cblist_flush_enqueue(struct rcu_cblist *drclp, struct rcu_cblist *srclp, struct rcu_head *rhp) rcu_cblist_flush_enqueue() argument
340 rcu_segcblist_enqueue(struct rcu_segcblist *rsclp, struct rcu_head *rhp) rcu_segcblist_enqueue() argument
360 rcu_segcblist_entrain(struct rcu_segcblist *rsclp, struct rcu_head *rhp) rcu_segcblist_entrain() argument
[all...]
H A Dsrcutiny.c116 struct rcu_head *rhp; in srcu_drive_gp() local
139 rhp = lh; in srcu_drive_gp()
142 rhp->func(rhp); in srcu_drive_gp()
178 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument
183 rhp->func = func; in call_srcu()
184 rhp->next = NULL; in call_srcu()
186 *ssp->srcu_cb_tail = rhp; in call_srcu()
187 ssp->srcu_cb_tail = &rhp->next; in call_srcu()
H A Dupdate.c515 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument
519 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read()
523 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \
603 struct early_boot_kfree_rcu *rhp; in early_boot_test_call_rcu() local
610 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in early_boot_test_call_rcu()
611 if (!WARN_ON_ONCE(!rhp)) in early_boot_test_call_rcu()
612 kfree_rcu(rhp, rh); in early_boot_test_call_rcu()
H A Dtasks.h322 static void call_rcu_tasks_generic(struct rcu_head *rhp, rcu_callback_t func, in call_rcu_tasks_generic() argument
334 rhp->next = NULL; in call_rcu_tasks_generic()
335 rhp->func = func; in call_rcu_tasks_generic()
365 rcu_segcblist_enqueue(&rtpcp->cblist, rhp); in call_rcu_tasks_generic()
384 static void rcu_barrier_tasks_generic_cb(struct rcu_head *rhp) in rcu_barrier_tasks_generic_cb() argument
389 rtpcp = container_of(rhp, struct rcu_tasks_percpu, barrier_q_head); in rcu_barrier_tasks_generic_cb()
515 struct rcu_head *rhp; in rcu_tasks_invoke_cbs() local
540 for (rhp = rcu_cblist_dequeue(&rcl); rhp; rhp in rcu_tasks_invoke_cbs()
1071 call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) call_rcu_tasks() argument
1235 call_rcu_tasks_rude(struct rcu_head *rhp, rcu_callback_t func) call_rcu_tasks_rude() argument
1867 call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func) call_rcu_tasks_trace() argument
1996 test_rcu_tasks_callback(struct rcu_head *rhp) test_rcu_tasks_callback() argument
[all...]
H A Drcu_segcblist.h24 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp);
27 struct rcu_head *rhp);
139 struct rcu_head *rhp);
141 struct rcu_head *rhp);
H A Drcutorture.c1707 static void rcu_torture_timer_cb(struct rcu_head *rhp) in rcu_torture_timer_cb() argument
1709 kfree(rhp); in rcu_torture_timer_cb()
2060 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); in rcu_torture_timer() local
2062 if (rhp) in rcu_torture_timer()
2063 cur_ops->call(rhp, rcu_torture_timer_cb); in rcu_torture_timer()
2298 struct rcu_head *rhp; in rcu_torture_mem_dump_obj() local
2305 rhp = kmem_cache_alloc(kcp, GFP_KERNEL); in rcu_torture_mem_dump_obj()
2306 if (WARN_ON_ONCE(!rhp)) { in rcu_torture_mem_dump_obj()
2310 pr_alert("mem_dump_obj() slab test: rcu_torture_stats = %px, &rhp in rcu_torture_mem_dump_obj()
2509 rcu_torture_fwd_prog_cb(struct rcu_head *rhp) rcu_torture_fwd_prog_cb() argument
2581 rcu_torture_fwd_cb_cr(struct rcu_head *rhp) rcu_torture_fwd_cb_cr() argument
3017 struct rcu_head *rhp = rcu_void; rcu_torture_barrier1cb() local
3423 rcu_torture_leak_cb(struct rcu_head *rhp) rcu_torture_leak_cb() argument
3427 rcu_torture_err_cb(struct rcu_head *rhp) rcu_torture_err_cb() argument
3452 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); rcu_test_debug_objects() local
[all...]
H A Drcuscale.c442 static void rcu_scale_async_cb(struct rcu_head *rhp) in rcu_scale_async_cb() argument
445 kfree(rhp); in rcu_scale_async_cb()
458 struct rcu_head *rhp = NULL; in rcu_scale_writer() local
503 if (!rhp) in rcu_scale_writer()
504 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_scale_writer()
505 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_scale_writer()
507 cur_ops->async(rhp, rcu_scale_async_cb); in rcu_scale_writer()
508 rhp = NULL; in rcu_scale_writer()
513 kfree(rhp); /* Becaus in rcu_scale_writer()
[all...]

Completed in 25 milliseconds

123