/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | intel_runtime_pm.c | 101 stacks = krealloc(rpm->debug.owners, in track_intel_runtime_pm_wakeref() 106 rpm->debug.owners = stacks; in track_intel_runtime_pm_wakeref() 130 if (rpm->debug.owners[n] == stack) { in untrack_intel_runtime_pm_wakeref() 131 memmove(rpm->debug.owners + n, in untrack_intel_runtime_pm_wakeref() 132 rpm->debug.owners + n + 1, in untrack_intel_runtime_pm_wakeref() 197 sort(dbg->owners, dbg->count, sizeof(*dbg->owners), cmphandle, NULL); in __print_intel_runtime_pm_wakeref() 200 depot_stack_handle_t stack = dbg->owners[i]; in __print_intel_runtime_pm_wakeref() 204 while (i + 1 < dbg->count && dbg->owners[i + 1] == stack) in __print_intel_runtime_pm_wakeref() 219 debug->owners in __untrack_all_wakerefs() [all...] |
H A D | intel_runtime_pm.h | 71 depot_stack_handle_t *owners; member
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
H A D | intel_runtime_pm.c | 92 stacks = krealloc(rpm->debug.owners, in track_intel_runtime_pm_wakeref() 97 rpm->debug.owners = stacks; in track_intel_runtime_pm_wakeref() 121 if (rpm->debug.owners[n] == stack) { in untrack_intel_runtime_pm_wakeref() 122 memmove(rpm->debug.owners + n, in untrack_intel_runtime_pm_wakeref() 123 rpm->debug.owners + n + 1, in untrack_intel_runtime_pm_wakeref() 188 sort(dbg->owners, dbg->count, sizeof(*dbg->owners), cmphandle, NULL); in __print_intel_runtime_pm_wakeref() 191 depot_stack_handle_t stack = dbg->owners[i]; in __print_intel_runtime_pm_wakeref() 195 while (i + 1 < dbg->count && dbg->owners[i + 1] == stack) in __print_intel_runtime_pm_wakeref() 210 debug->owners in __untrack_all_wakerefs() [all...] |
H A D | intel_runtime_pm.h | 86 depot_stack_handle_t *owners; member
|
/kernel/linux/linux-5.10/kernel/ |
H A D | audit_tree.c | 37 } owners[]; member 68 * tree.chunks anchors chunk.owners[].list hash_lock 133 if (chunk->owners[i].owner) in free_chunk() 134 put_tree(chunk->owners[i].owner); in free_chunk() 193 chunk = kzalloc(struct_size(chunk, owners, count), GFP_KERNEL); in alloc_chunk() 202 INIT_LIST_HEAD(&chunk->owners[i].list); in alloc_chunk() 203 chunk->owners[i].index = i; in alloc_chunk() 265 if (chunk->owners[n].owner == tree) in audit_tree_match() 276 return container_of(p, struct audit_chunk, owners[0]); in find_chunk() 303 if (!old->owners[ in replace_chunk() [all...] |
/kernel/linux/linux-6.6/kernel/ |
H A D | audit_tree.c | 37 } owners[]; member 68 * tree.chunks anchors chunk.owners[].list hash_lock 133 if (chunk->owners[i].owner) in free_chunk() 134 put_tree(chunk->owners[i].owner); in free_chunk() 193 chunk = kzalloc(struct_size(chunk, owners, count), GFP_KERNEL); in alloc_chunk() 202 INIT_LIST_HEAD(&chunk->owners[i].list); in alloc_chunk() 203 chunk->owners[i].index = i; in alloc_chunk() 265 if (chunk->owners[n].owner == tree) in audit_tree_match() 276 return container_of(p, struct audit_chunk, owners[0]); in find_chunk() 303 if (!old->owners[ in replace_chunk() [all...] |
/kernel/linux/linux-6.6/arch/riscv/kvm/ |
H A D | aia.c | 23 struct kvm_vcpu *owners[BITS_PER_LONG]; member 42 if (hgctrl->owners[i] == owner) { in aia_find_hgei() 407 hgctrl->owners[ret] = owner; in kvm_riscv_aia_alloc_hgei() 434 hgctrl->owners[hgei] = NULL; in kvm_riscv_aia_free_hgei() 469 if (hgctrl->owners[i]) in hgei_interrupt() 470 kvm_vcpu_kick(hgctrl->owners[i]); in hgei_interrupt() 570 vcpu = hgctrl->owners[i]; in kvm_riscv_aia_disable()
|
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | inet_hashtables.c | 79 INIT_HLIST_HEAD(&tb->owners); in inet_bind_bucket_create() 90 if (hlist_empty(&tb->owners)) { in inet_bind_bucket_destroy() 119 INIT_HLIST_HEAD(&tb->owners); in inet_bind2_bucket_init() 142 if (hlist_empty(&tb->owners) && hlist_empty(&tb->deathrow)) { in inet_bind2_bucket_destroy() 172 sk_add_bind_node(sk, &tb->owners); in inet_bind_hash() 174 sk_add_bind2_node(sk, &tb2->owners); in inet_bind_hash() 962 sk_add_bind2_node(sk, &tb2->owners); in __inet_bhash2_update_saddr() 1063 WARN_ON(hlist_empty(&tb->owners)); in __inet_hash_connect()
|
H A D | inet_connection_sock.c | 217 sk_for_each_bound_bhash2(sk2, &tb2->owners) { in inet_bhash2_conflict() 253 * in tb->owners and tb2->owners list belong in inet_csk_bind_conflict() 260 sk_for_each_bound(sk2, &tb->owners) in inet_csk_bind_conflict() 435 * owners list. in sk_reuseport_match() 458 if (hlist_empty(&tb->owners)) { in inet_csk_update_fastreuse() 550 if (!hlist_empty(&tb->owners)) { in inet_csk_get_port()
|
H A D | inet_timewait_sock.c | 136 inet_twsk_add_bind_node(tw, &tw->tw_tb->owners); in inet_twsk_hashdance()
|
/kernel/linux/linux-5.10/arch/x86/events/amd/ |
H A D | core.c | 397 if (cmpxchg(nb->owners + i, event, NULL) == event) in __amd_put_nb_event_constraints() 466 old = cmpxchg(nb->owners + idx, NULL, event); in __amd_get_nb_event_constraints() 467 else if (nb->owners[idx] == event) in __amd_get_nb_event_constraints() 478 cmpxchg(nb->owners + new, event, NULL); in __amd_get_nb_event_constraints()
|
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | inet_hashtables.c | 87 INIT_HLIST_HEAD(&tb->owners); in inet_bind_bucket_create() 98 if (hlist_empty(&tb->owners)) { in inet_bind_bucket_destroy() 108 sk_add_bind_node(sk, &tb->owners); in inet_bind_hash() 803 WARN_ON(hlist_empty(&tb->owners)); in __inet_hash_connect()
|
H A D | inet_timewait_sock.c | 118 inet_twsk_add_bind_node(tw, &tw->tw_tb->owners); in inet_twsk_hashdance()
|
H A D | inet_connection_sock.c | 146 * in tb->owners list belong to the same net - the in inet_csk_bind_conflict() 150 sk_for_each_bound(sk2, &tb->owners) { in inet_csk_bind_conflict() 287 * owners list. in sk_reuseport_match() 310 if (hlist_empty(&tb->owners)) { in inet_csk_update_fastreuse() 392 if (!hlist_empty(&tb->owners)) { in inet_csk_get_port()
|
/kernel/linux/linux-6.6/arch/x86/events/amd/ |
H A D | core.c | 407 if (cmpxchg(nb->owners + i, event, NULL) == event) in __amd_put_nb_event_constraints() 476 old = cmpxchg(nb->owners + idx, NULL, event); in __amd_get_nb_event_constraints() 477 else if (nb->owners[idx] == event) in __amd_get_nb_event_constraints() 488 cmpxchg(nb->owners + new, event, NULL); in __amd_get_nb_event_constraints()
|
/kernel/linux/linux-6.6/include/net/ |
H A D | inet_hashtables.h | 64 * sk->sk_reuse set, we don't even have to walk the owners list at all, 91 struct hlist_head owners; member 110 struct hlist_head owners; member 111 /* bhash has twsk in owners, but bhash2 has twsk in
|
/kernel/linux/linux-5.10/arch/xtensa/kernel/ |
H A D | coprocessor.S | 153 movi a0, coprocessor_owner # list of owners
|
/kernel/linux/linux-5.10/include/net/ |
H A D | inet_hashtables.h | 63 * sk->sk_reuse set, we don't even have to walk the owners list at all, 90 struct hlist_head owners; member
|
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | cma.c | 218 struct hlist_head owners; member 1652 hlist_for_each_entry(id_priv, &bind_list->owners, node) { in cma_find_listener() 1827 if (hlist_empty(&bind_list->owners)) { in cma_release_port() 3512 hlist_add_head(&id_priv->node, &bind_list->owners); in cma_bind_port() 3551 hlist_for_each_entry(cur_id, &bind_list->owners, node) { in cma_port_is_unique() 3642 hlist_for_each_entry(cur_id, &bind_list->owners, node) { in cma_check_port()
|
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | cma.c | 227 struct hlist_head owners; member 1799 hlist_for_each_entry(id_priv, &bind_list->owners, node) { in cma_find_listener() 1975 if (hlist_empty(&bind_list->owners)) { in cma_release_port() 3643 hlist_add_head(&id_priv->node, &bind_list->owners); in cma_bind_port() 3682 hlist_for_each_entry(cur_id, &bind_list->owners, node) { in cma_port_is_unique() 3773 hlist_for_each_entry(cur_id, &bind_list->owners, node) { in cma_check_port()
|
/kernel/linux/linux-5.10/arch/x86/events/ |
H A D | perf_event.h | 111 struct perf_event *owners[X86_PMC_IDX_MAX]; member
|
/kernel/linux/linux-6.6/arch/x86/events/ |
H A D | perf_event.h | 116 struct perf_event *owners[X86_PMC_IDX_MAX]; member
|