/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | i915_active.c | 33 struct i915_active *ref; member 81 struct i915_active *ref = addr; in active_debug_hint() local 83 return (void *)ref->active ?: (void *)ref->retire ?: (void *)ref; in active_debug_hint() 91 static void debug_active_init(struct i915_active *ref) in debug_active_init() argument 93 debug_object_init(ref, &active_debug_desc); in debug_active_init() 96 static void debug_active_activate(struct i915_active *ref) in debug_active_activate() argument 98 lockdep_assert_held(&ref->tree_lock); in debug_active_activate() 99 debug_object_activate(ref, in debug_active_activate() 102 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument 109 debug_active_fini(struct i915_active *ref) debug_active_fini() argument 114 debug_active_assert(struct i915_active *ref) debug_active_assert() argument 121 debug_active_init(struct i915_active *ref) debug_active_init() argument 122 debug_active_activate(struct i915_active *ref) debug_active_activate() argument 123 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument 124 debug_active_fini(struct i915_active *ref) debug_active_fini() argument 125 debug_active_assert(struct i915_active *ref) debug_active_assert() argument 130 __active_retire(struct i915_active *ref) __active_retire() argument 184 struct i915_active *ref = container_of(wrk, typeof(*ref), work); active_work() local 194 active_retire(struct i915_active *ref) active_retire() argument 237 __active_lookup(struct i915_active *ref, u64 idx) __active_lookup() argument 296 active_instance(struct i915_active *ref, u64 idx) active_instance() argument 345 __i915_active_init(struct i915_active *ref, int (*active)(struct i915_active *ref), void (*retire)(struct i915_active *ref), struct lock_class_key *mkey, struct lock_class_key *wkey) __i915_active_init() argument 375 ____active_del_barrier(struct i915_active *ref, struct active_node *node, struct intel_engine_cs *engine) ____active_del_barrier() argument 418 __active_del_barrier(struct i915_active *ref, struct active_node *node) __active_del_barrier() argument 424 replace_barrier(struct i915_active *ref, struct i915_active_fence *active) replace_barrier() argument 437 i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence) i915_active_ref() argument 472 __i915_active_set_fence(struct i915_active *ref, struct i915_active_fence *active, struct dma_fence *fence) __i915_active_set_fence() argument 491 __active_fence(struct i915_active *ref, u64 idx) __active_fence() argument 507 __i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence) __i915_active_ref() argument 514 i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f) i915_active_set_exclusive() argument 520 i915_active_acquire_if_busy(struct i915_active *ref) i915_active_acquire_if_busy() argument 526 __i915_active_activate(struct i915_active *ref) __i915_active_activate() argument 534 i915_active_acquire(struct i915_active *ref) i915_active_acquire() argument 561 i915_active_acquire_for_context(struct i915_active *ref, u64 idx) i915_active_acquire_for_context() argument 579 i915_active_release(struct i915_active *ref) i915_active_release() argument 615 flush_lazy_signals(struct i915_active *ref) flush_lazy_signals() argument 632 __i915_active_wait(struct i915_active *ref, int state) __i915_active_wait() argument 682 struct i915_active *ref; global() member 699 __await_barrier(struct i915_active *ref, struct i915_sw_fence *fence) __await_barrier() argument 722 await_active(struct i915_active *ref, unsigned int flags, int (*fn)(void *arg, struct dma_fence *fence), void *arg, struct i915_sw_fence *barrier) await_active() argument 769 i915_request_await_active(struct i915_request *rq, struct i915_active *ref, unsigned int flags) i915_request_await_active() argument 782 i915_sw_fence_await_active(struct i915_sw_fence *fence, struct i915_active *ref, unsigned int flags) i915_sw_fence_await_active() argument 789 i915_active_fini(struct i915_active *ref) i915_active_fini() argument 805 reuse_idle_barrier(struct i915_active *ref, u64 idx) reuse_idle_barrier() argument 888 i915_active_acquire_preallocate_barrier(struct i915_active *ref, struct intel_engine_cs *engine) i915_active_acquire_preallocate_barrier() argument 970 i915_active_acquire_barrier(struct i915_active *ref) i915_active_acquire_barrier() argument 1172 struct kref ref; global() member 1175 i915_active_get(struct i915_active *ref) i915_active_get() argument 1183 auto_release(struct kref *ref) auto_release() argument 1191 i915_active_put(struct i915_active *ref) i915_active_put() argument 1198 auto_active(struct i915_active *ref) auto_active() argument 1205 auto_retire(struct i915_active *ref) auto_retire() argument [all...] |
H A D | i915_active.h | 152 void __i915_active_init(struct i915_active *ref, 153 int (*active)(struct i915_active *ref), 154 void (*retire)(struct i915_active *ref), 159 #define i915_active_init(ref, active, retire) do { \ 163 __i915_active_init(ref, active, retire, &__mkey, &__wkey); \ 167 __i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence); 168 int i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence); 171 i915_active_add_request(struct i915_active *ref, struct i915_request *rq) in i915_active_add_request() argument 173 return i915_active_ref(ref, in i915_active_add_request() 179 i915_active_set_exclusive(struct i915_active *ref, struc 181 i915_active_has_exclusive(struct i915_active *ref) i915_active_has_exclusive() argument 187 i915_active_wait(struct i915_active *ref) i915_active_wait() argument 208 __i915_active_acquire(struct i915_active *ref) __i915_active_acquire() argument 215 i915_active_is_idle(const struct i915_active *ref) i915_active_is_idle() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
H A D | i915_active.c | 29 struct i915_active *ref; member 77 struct i915_active *ref = addr; in active_debug_hint() local 79 return (void *)ref->active ?: (void *)ref->retire ?: (void *)ref; in active_debug_hint() 87 static void debug_active_init(struct i915_active *ref) in debug_active_init() argument 89 debug_object_init(ref, &active_debug_desc); in debug_active_init() 92 static void debug_active_activate(struct i915_active *ref) in debug_active_activate() argument 94 lockdep_assert_held(&ref->tree_lock); in debug_active_activate() 95 debug_object_activate(ref, in debug_active_activate() 98 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument 105 debug_active_fini(struct i915_active *ref) debug_active_fini() argument 110 debug_active_assert(struct i915_active *ref) debug_active_assert() argument 117 debug_active_init(struct i915_active *ref) debug_active_init() argument 118 debug_active_activate(struct i915_active *ref) debug_active_activate() argument 119 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument 120 debug_active_fini(struct i915_active *ref) debug_active_fini() argument 121 debug_active_assert(struct i915_active *ref) debug_active_assert() argument 126 __active_retire(struct i915_active *ref) __active_retire() argument 179 struct i915_active *ref = container_of(wrk, typeof(*ref), work); active_work() local 189 active_retire(struct i915_active *ref) active_retire() argument 232 __active_lookup(struct i915_active *ref, u64 idx) __active_lookup() argument 289 active_instance(struct i915_active *ref, u64 idx) active_instance() argument 338 __i915_active_init(struct i915_active *ref, int (*active)(struct i915_active *ref), void (*retire)(struct i915_active *ref), unsigned long flags, struct lock_class_key *mkey, struct lock_class_key *wkey) __i915_active_init() argument 365 ____active_del_barrier(struct i915_active *ref, struct active_node *node, struct intel_engine_cs *engine) ____active_del_barrier() argument 408 __active_del_barrier(struct i915_active *ref, struct active_node *node) __active_del_barrier() argument 414 replace_barrier(struct i915_active *ref, struct i915_active_fence *active) replace_barrier() argument 427 i915_active_add_request(struct i915_active *ref, struct i915_request *rq) i915_active_add_request() argument 464 __i915_active_set_fence(struct i915_active *ref, struct i915_active_fence *active, struct dma_fence *fence) __i915_active_set_fence() argument 483 i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f) i915_active_set_exclusive() argument 489 i915_active_acquire_if_busy(struct i915_active *ref) i915_active_acquire_if_busy() argument 495 __i915_active_activate(struct i915_active *ref) __i915_active_activate() argument 503 i915_active_acquire(struct i915_active *ref) i915_active_acquire() argument 530 i915_active_acquire_for_context(struct i915_active *ref, u64 idx) i915_active_acquire_for_context() argument 548 i915_active_release(struct i915_active *ref) i915_active_release() argument 584 flush_lazy_signals(struct i915_active *ref) flush_lazy_signals() argument 601 __i915_active_wait(struct i915_active *ref, int state) __i915_active_wait() argument 651 struct i915_active *ref; global() member 668 __await_barrier(struct i915_active *ref, struct i915_sw_fence *fence) __await_barrier() argument 691 await_active(struct i915_active *ref, unsigned int flags, int (*fn)(void *arg, struct dma_fence *fence), void *arg, struct i915_sw_fence *barrier) await_active() argument 738 i915_request_await_active(struct i915_request *rq, struct i915_active *ref, unsigned int flags) i915_request_await_active() argument 751 i915_sw_fence_await_active(struct i915_sw_fence *fence, struct i915_active *ref, unsigned int flags) i915_sw_fence_await_active() argument 758 i915_active_fini(struct i915_active *ref) i915_active_fini() argument 774 reuse_idle_barrier(struct i915_active *ref, u64 idx) reuse_idle_barrier() argument 857 i915_active_acquire_preallocate_barrier(struct i915_active *ref, struct intel_engine_cs *engine) i915_active_acquire_preallocate_barrier() argument 939 i915_active_acquire_barrier(struct i915_active *ref) i915_active_acquire_barrier() argument 1141 struct kref ref; global() member 1144 i915_active_get(struct i915_active *ref) i915_active_get() argument 1152 auto_release(struct kref *ref) auto_release() argument 1160 i915_active_put(struct i915_active *ref) i915_active_put() argument 1167 auto_active(struct i915_active *ref) auto_active() argument 1173 auto_retire(struct i915_active *ref) auto_retire() argument [all...] |
H A D | i915_active.h | 152 void __i915_active_init(struct i915_active *ref, 153 int (*active)(struct i915_active *ref), 154 void (*retire)(struct i915_active *ref), 160 #define i915_active_init(ref, active, retire, flags) do { \ 164 __i915_active_init(ref, active, retire, flags, &__mkey, &__wkey); \ 167 int i915_active_add_request(struct i915_active *ref, struct i915_request *rq); 170 i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f); 172 int __i915_active_wait(struct i915_active *ref, int state); 173 static inline int i915_active_wait(struct i915_active *ref) in i915_active_wait() argument 175 return __i915_active_wait(ref, TASK_INTERRUPTIBL in i915_active_wait() 194 __i915_active_acquire(struct i915_active *ref) __i915_active_acquire() argument 201 i915_active_is_idle(const struct i915_active *ref) i915_active_is_idle() argument [all...] |
/kernel/linux/linux-5.10/lib/ |
H A D | percpu-refcount.c | 12 * don't try to detect the ref hitting 0 - which means that get/put can just 23 * the ref hitting 0 on every put - this would require global synchronization 27 * the ref can't hit 0 before the user drops the initial ref, so as long as we 28 * convert to non percpu mode before the initial ref is dropped everything 41 static unsigned long __percpu *percpu_count_ptr(struct percpu_ref *ref) in percpu_count_ptr() argument 44 (ref->percpu_count_ptr & ~__PERCPU_REF_ATOMIC_DEAD); in percpu_count_ptr() 49 * @ref: percpu_ref to initialize 54 * Initializes @ref. @ref start 62 percpu_ref_init(struct percpu_ref *ref, percpu_ref_func_t *release, unsigned int flags, gfp_t gfp) percpu_ref_init() argument 107 __percpu_ref_exit(struct percpu_ref *ref) __percpu_ref_exit() argument 129 percpu_ref_exit(struct percpu_ref *ref) percpu_ref_exit() argument 153 struct percpu_ref *ref = data->ref; percpu_ref_call_confirm_rcu() local 170 struct percpu_ref *ref = data->ref; percpu_ref_switch_to_atomic_rcu() local 203 percpu_ref_noop_confirm_switch(struct percpu_ref *ref) percpu_ref_noop_confirm_switch() argument 207 __percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_to_atomic() argument 230 __percpu_ref_switch_to_percpu(struct percpu_ref *ref) __percpu_ref_switch_to_percpu() argument 258 __percpu_ref_switch_mode(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_mode() argument 299 percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) percpu_ref_switch_to_atomic() argument 321 percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref) percpu_ref_switch_to_atomic_sync() argument 346 percpu_ref_switch_to_percpu(struct percpu_ref *ref) percpu_ref_switch_to_percpu() argument 376 percpu_ref_kill_and_confirm(struct percpu_ref *ref, percpu_ref_func_t *confirm_kill) percpu_ref_kill_and_confirm() argument 403 percpu_ref_is_zero(struct percpu_ref *ref) percpu_ref_is_zero() argument 434 percpu_ref_reinit(struct percpu_ref *ref) percpu_ref_reinit() argument 456 percpu_ref_resurrect(struct percpu_ref *ref) percpu_ref_resurrect() argument [all...] |
/kernel/linux/linux-6.6/lib/ |
H A D | percpu-refcount.c | 13 * don't try to detect the ref hitting 0 - which means that get/put can just 24 * the ref hitting 0 on every put - this would require global synchronization 28 * the ref can't hit 0 before the user drops the initial ref, so as long as we 29 * convert to non percpu mode before the initial ref is dropped everything 42 static unsigned long __percpu *percpu_count_ptr(struct percpu_ref *ref) in percpu_count_ptr() argument 45 (ref->percpu_count_ptr & ~__PERCPU_REF_ATOMIC_DEAD); in percpu_count_ptr() 50 * @ref: percpu_ref to initialize 55 * Initializes @ref. @ref start 63 percpu_ref_init(struct percpu_ref *ref, percpu_ref_func_t *release, unsigned int flags, gfp_t gfp) percpu_ref_init() argument 108 __percpu_ref_exit(struct percpu_ref *ref) __percpu_ref_exit() argument 130 percpu_ref_exit(struct percpu_ref *ref) percpu_ref_exit() argument 154 struct percpu_ref *ref = data->ref; percpu_ref_call_confirm_rcu() local 171 struct percpu_ref *ref = data->ref; percpu_ref_switch_to_atomic_rcu() local 209 percpu_ref_noop_confirm_switch(struct percpu_ref *ref) percpu_ref_noop_confirm_switch() argument 213 __percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_to_atomic() argument 237 __percpu_ref_switch_to_percpu(struct percpu_ref *ref) __percpu_ref_switch_to_percpu() argument 265 __percpu_ref_switch_mode(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_mode() argument 306 percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) percpu_ref_switch_to_atomic() argument 328 percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref) percpu_ref_switch_to_atomic_sync() argument 353 percpu_ref_switch_to_percpu(struct percpu_ref *ref) percpu_ref_switch_to_percpu() argument 383 percpu_ref_kill_and_confirm(struct percpu_ref *ref, percpu_ref_func_t *confirm_kill) percpu_ref_kill_and_confirm() argument 410 percpu_ref_is_zero(struct percpu_ref *ref) percpu_ref_is_zero() argument 441 percpu_ref_reinit(struct percpu_ref *ref) percpu_ref_reinit() argument 463 percpu_ref_resurrect(struct percpu_ref *ref) percpu_ref_resurrect() argument [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | percpu-refcount.h | 21 * puts the ref back in single atomic_t mode, collecting the per cpu refs and 22 * issuing the appropriate barriers, and then marks the ref as shutting down so 23 * that percpu_ref_put() will check for the ref hitting 0. After it returns, 24 * it's safe to drop the initial ref. 35 * and it's then safe to drop the initial ref with percpu_ref_put(). 74 * Start w/ ref == 1 in atomic mode. Can be switched to percpu 76 * with this flag, the ref will stay in atomic mode until 83 * Start dead w/ ref == 0 in atomic mode. Must be revived with 102 struct percpu_ref *ref; member 107 * The low bit of the pointer indicates whether the ref i 147 percpu_ref_kill(struct percpu_ref *ref) percpu_ref_kill() argument 158 __ref_is_percpu(struct percpu_ref *ref, unsigned long __percpu **percpu_countp) __ref_is_percpu() argument 198 percpu_ref_get_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_get_many() argument 220 percpu_ref_get(struct percpu_ref *ref) percpu_ref_get() argument 235 percpu_ref_tryget_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_tryget_many() argument 264 percpu_ref_tryget(struct percpu_ref *ref) percpu_ref_tryget() argument 275 percpu_ref_tryget_live_rcu(struct percpu_ref *ref) percpu_ref_tryget_live_rcu() argument 306 percpu_ref_tryget_live(struct percpu_ref *ref) percpu_ref_tryget_live() argument 326 percpu_ref_put_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_put_many() argument 349 percpu_ref_put(struct percpu_ref *ref) percpu_ref_put() argument 363 percpu_ref_is_dying(struct percpu_ref *ref) percpu_ref_is_dying() argument [all...] |
H A D | rcuref.h | 21 * @ref: Pointer to the reference count 24 static inline void rcuref_init(rcuref_t *ref, unsigned int cnt) in rcuref_init() argument 26 atomic_set(&ref->refcnt, cnt - 1); in rcuref_init() 31 * @ref: Pointer to the reference count 35 static inline unsigned int rcuref_read(rcuref_t *ref) in rcuref_read() argument 37 unsigned int c = atomic_read(&ref->refcnt); in rcuref_read() 43 extern __must_check bool rcuref_get_slowpath(rcuref_t *ref); 47 * @ref: Pointer to the reference count 61 static inline __must_check bool rcuref_get(rcuref_t *ref) in rcuref_get() argument 67 if (likely(!atomic_add_negative_relaxed(1, &ref in rcuref_get() 79 __rcuref_put(rcuref_t *ref) __rcuref_put() argument 119 rcuref_put_rcusafe(rcuref_t *ref) rcuref_put_rcusafe() argument 145 rcuref_put(rcuref_t *ref) rcuref_put() argument [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | percpu-refcount.h | 21 * puts the ref back in single atomic_t mode, collecting the per cpu refs and 22 * issuing the appropriate barriers, and then marks the ref as shutting down so 23 * that percpu_ref_put() will check for the ref hitting 0. After it returns, 24 * it's safe to drop the initial ref. 35 * and it's then safe to drop the initial ref with percpu_ref_put(). 74 * Start w/ ref == 1 in atomic mode. Can be switched to percpu 76 * with this flag, the ref will stay in atomic mode until 83 * Start dead w/ ref == 0 in atomic mode. Must be revived with 102 struct percpu_ref *ref; member 107 * The low bit of the pointer indicates whether the ref i 147 percpu_ref_kill(struct percpu_ref *ref) percpu_ref_kill() argument 158 __ref_is_percpu(struct percpu_ref *ref, unsigned long __percpu **percpu_countp) __ref_is_percpu() argument 198 percpu_ref_get_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_get_many() argument 220 percpu_ref_get(struct percpu_ref *ref) percpu_ref_get() argument 235 percpu_ref_tryget_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_tryget_many() argument 264 percpu_ref_tryget(struct percpu_ref *ref) percpu_ref_tryget() argument 284 percpu_ref_tryget_live(struct percpu_ref *ref) percpu_ref_tryget_live() argument 313 percpu_ref_put_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_put_many() argument 336 percpu_ref_put(struct percpu_ref *ref) percpu_ref_put() argument 350 percpu_ref_is_dying(struct percpu_ref *ref) percpu_ref_is_dying() argument [all...] |
/kernel/linux/linux-5.10/fs/jffs2/ |
H A D | xattr.c | 424 * verify_xattr_ref(c, ref) 427 * save_xattr_ref(c, ref) 432 * delete_xattr_ref(c, ref) 444 static int verify_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) in verify_xattr_ref() argument 454 if (ref_flags(ref->node) != REF_UNCHECKED) in verify_xattr_ref() 456 offset = ref_offset(ref->node); in verify_xattr_ref() 482 ref->ino = je32_to_cpu(rr.ino); in verify_xattr_ref() 483 ref->xid = je32_to_cpu(rr.xid); in verify_xattr_ref() 484 ref->xseqno = je32_to_cpu(rr.xseqno); in verify_xattr_ref() 485 if (ref in verify_xattr_ref() 506 save_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) save_xattr_ref() argument 554 struct jffs2_xattr_ref *ref; create_xattr_ref() local 576 delete_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) delete_xattr_ref() argument 600 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_delete_inode() local 618 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_free_inode() local 640 struct jffs2_xattr_ref *ref, *cmp, **pref, **pcmp; check_xattr_ref_inode() local 743 struct jffs2_xattr_ref *ref, *_ref; jffs2_clear_xattr_subsystem() local 777 struct jffs2_xattr_ref *ref, *_ref; jffs2_build_xattr_subsystem() local 972 struct jffs2_xattr_ref *ref, **pref; jffs2_listxattr() local 1043 struct jffs2_xattr_ref *ref, **pref; do_jffs2_getxattr() local 1105 struct jffs2_xattr_ref *ref, *newref, **pref; do_jffs2_setxattr() local 1260 jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref, struct jffs2_raw_node_ref *raw) jffs2_garbage_collect_xattr_ref() argument 1337 jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) jffs2_release_xattr_ref() argument [all...] |
H A D | malloc.c | 204 struct jffs2_raw_node_ref **p, *ref; in jffs2_prealloc_raw_node_refs() local 210 ref = *p; in jffs2_prealloc_raw_node_refs() 215 if (ref && ref->flash_offset != REF_EMPTY_NODE) in jffs2_prealloc_raw_node_refs() 216 ref++; in jffs2_prealloc_raw_node_refs() 219 if (!ref) { in jffs2_prealloc_raw_node_refs() 221 ref = *p = jffs2_alloc_refblock(); in jffs2_prealloc_raw_node_refs() 222 if (!ref) in jffs2_prealloc_raw_node_refs() 225 if (ref->flash_offset == REF_LINK_NODE) { in jffs2_prealloc_raw_node_refs() 226 p = &ref in jffs2_prealloc_raw_node_refs() 299 struct jffs2_xattr_ref *ref; jffs2_alloc_xattr_ref() local 310 jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref) jffs2_free_xattr_ref() argument [all...] |
/kernel/linux/linux-6.6/fs/jffs2/ |
H A D | xattr.c | 424 * verify_xattr_ref(c, ref) 427 * save_xattr_ref(c, ref) 432 * delete_xattr_ref(c, ref) 444 static int verify_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) in verify_xattr_ref() argument 454 if (ref_flags(ref->node) != REF_UNCHECKED) in verify_xattr_ref() 456 offset = ref_offset(ref->node); in verify_xattr_ref() 482 ref->ino = je32_to_cpu(rr.ino); in verify_xattr_ref() 483 ref->xid = je32_to_cpu(rr.xid); in verify_xattr_ref() 484 ref->xseqno = je32_to_cpu(rr.xseqno); in verify_xattr_ref() 485 if (ref in verify_xattr_ref() 506 save_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) save_xattr_ref() argument 554 struct jffs2_xattr_ref *ref; create_xattr_ref() local 576 delete_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) delete_xattr_ref() argument 600 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_delete_inode() local 618 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_free_inode() local 640 struct jffs2_xattr_ref *ref, *cmp, **pref, **pcmp; check_xattr_ref_inode() local 743 struct jffs2_xattr_ref *ref, *_ref; jffs2_clear_xattr_subsystem() local 777 struct jffs2_xattr_ref *ref, *_ref; jffs2_build_xattr_subsystem() local 972 struct jffs2_xattr_ref *ref, **pref; jffs2_listxattr() local 1042 struct jffs2_xattr_ref *ref, **pref; do_jffs2_getxattr() local 1104 struct jffs2_xattr_ref *ref, *newref, **pref; do_jffs2_setxattr() local 1259 jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref, struct jffs2_raw_node_ref *raw) jffs2_garbage_collect_xattr_ref() argument 1336 jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) jffs2_release_xattr_ref() argument [all...] |
H A D | malloc.c | 204 struct jffs2_raw_node_ref **p, *ref; in jffs2_prealloc_raw_node_refs() local 210 ref = *p; in jffs2_prealloc_raw_node_refs() 215 if (ref && ref->flash_offset != REF_EMPTY_NODE) in jffs2_prealloc_raw_node_refs() 216 ref++; in jffs2_prealloc_raw_node_refs() 219 if (!ref) { in jffs2_prealloc_raw_node_refs() 221 ref = *p = jffs2_alloc_refblock(); in jffs2_prealloc_raw_node_refs() 222 if (!ref) in jffs2_prealloc_raw_node_refs() 225 if (ref->flash_offset == REF_LINK_NODE) { in jffs2_prealloc_raw_node_refs() 226 p = &ref in jffs2_prealloc_raw_node_refs() 299 struct jffs2_xattr_ref *ref; jffs2_alloc_xattr_ref() local 310 jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref) jffs2_free_xattr_ref() argument [all...] |
/kernel/linux/linux-5.10/drivers/clk/mxs/ |
H A D | clk-ref.c | 33 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_enable() local 35 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + CLR); in clk_ref_enable() 42 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_disable() local 44 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + SET); in clk_ref_disable() 50 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_recalc_rate() local 52 u8 frac = (readl_relaxed(ref->reg) >> (ref->idx * 8)) & 0x3f; in clk_ref_recalc_rate() 86 struct clk_ref *ref in clk_ref_set_rate() local 124 struct clk_ref *ref; mxs_clk_ref() local [all...] |
/kernel/linux/linux-6.6/drivers/clk/mxs/ |
H A D | clk-ref.c | 33 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_enable() local 35 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + CLR); in clk_ref_enable() 42 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_disable() local 44 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + SET); in clk_ref_disable() 50 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_recalc_rate() local 52 u8 frac = (readl_relaxed(ref->reg) >> (ref->idx * 8)) & 0x3f; in clk_ref_recalc_rate() 86 struct clk_ref *ref in clk_ref_set_rate() local 124 struct clk_ref *ref; mxs_clk_ref() local [all...] |
/kernel/linux/linux-6.6/fs/smb/client/ |
H A D | dfs.h | 25 struct dfs_ref *ref; member 31 #define ref_walk_cur(w) ((w)->ref) 56 static inline void __ref_walk_free(struct dfs_ref *ref) in __ref_walk_free() argument 58 kfree(ref->path); in __ref_walk_free() 59 kfree(ref->full_path); in __ref_walk_free() 60 dfs_cache_free_tgts(&ref->tl); in __ref_walk_free() 61 memset(ref, 0, sizeof(*ref)); in __ref_walk_free() 66 struct dfs_ref *ref = ref_walk_start(rw); in ref_walk_free() local 68 for (; ref < in ref_walk_free() 75 struct dfs_ref *ref = ref_walk_cur(rw) + 1; ref_walk_advance() local 88 struct dfs_ref *ref = ref_walk_cur(rw); ref_walk_next_tgt() local 131 dfs_get_referral(struct cifs_mount_ctx *mnt_ctx, const char *path, struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl) dfs_get_referral() argument [all...] |
/kernel/linux/linux-5.10/kernel/dma/ |
H A D | debug.c | 293 struct dma_debug_entry *ref, in __hash_bucket_find() 300 if (!match(ref, entry)) in __hash_bucket_find() 315 entry->size == ref->size ? ++match_lvl : 0; in __hash_bucket_find() 316 entry->type == ref->type ? ++match_lvl : 0; in __hash_bucket_find() 317 entry->direction == ref->direction ? ++match_lvl : 0; in __hash_bucket_find() 318 entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0; in __hash_bucket_find() 343 struct dma_debug_entry *ref) in bucket_find_exact() 345 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact() 349 struct dma_debug_entry *ref, in bucket_find_contain() 353 unsigned int max_range = dma_get_max_seg_size(ref in bucket_find_contain() 292 __hash_bucket_find(struct hash_bucket *bucket, struct dma_debug_entry *ref, match_fn match) __hash_bucket_find() argument 342 bucket_find_exact(struct hash_bucket *bucket, struct dma_debug_entry *ref) bucket_find_exact() argument 348 bucket_find_contain(struct hash_bucket **bucket, struct dma_debug_entry *ref, unsigned long *flags) bucket_find_contain() argument 956 check_unmap(struct dma_debug_entry *ref) check_unmap() argument 1097 check_sync(struct device *dev, struct dma_debug_entry *ref, bool to_cpu) check_sync() argument 1251 struct dma_debug_entry ref; debug_dma_mapping_error() local 1290 struct dma_debug_entry ref = { debug_dma_unmap_page() local 1340 get_nr_mapped_entries(struct device *dev, struct dma_debug_entry *ref) get_nr_mapped_entries() argument 1370 struct dma_debug_entry ref = { debug_dma_unmap_sg() local 1428 struct dma_debug_entry ref = { debug_dma_free_coherent() local 1479 struct dma_debug_entry ref = { debug_dma_unmap_resource() local 1496 struct dma_debug_entry ref; debug_dma_sync_single_for_cpu() local 1515 struct dma_debug_entry ref; debug_dma_sync_single_for_device() local 1541 struct dma_debug_entry ref = { debug_dma_sync_sg_for_cpu() local 1573 struct dma_debug_entry ref = { debug_dma_sync_sg_for_device() local [all...] |
/kernel/linux/linux-6.6/kernel/dma/ |
H A D | debug.c | 294 struct dma_debug_entry *ref, in __hash_bucket_find() 301 if (!match(ref, entry)) in __hash_bucket_find() 316 entry->size == ref->size ? ++match_lvl : 0; in __hash_bucket_find() 317 entry->type == ref->type ? ++match_lvl : 0; in __hash_bucket_find() 318 entry->direction == ref->direction ? ++match_lvl : 0; in __hash_bucket_find() 319 entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0; in __hash_bucket_find() 344 struct dma_debug_entry *ref) in bucket_find_exact() 346 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact() 350 struct dma_debug_entry *ref, in bucket_find_contain() 354 struct dma_debug_entry *entry, index = *ref; in bucket_find_contain() 293 __hash_bucket_find(struct hash_bucket *bucket, struct dma_debug_entry *ref, match_fn match) __hash_bucket_find() argument 343 bucket_find_exact(struct hash_bucket *bucket, struct dma_debug_entry *ref) bucket_find_exact() argument 349 bucket_find_contain(struct hash_bucket **bucket, struct dma_debug_entry *ref, unsigned long *flags) bucket_find_contain() argument 960 check_unmap(struct dma_debug_entry *ref) check_unmap() argument 1091 check_sync(struct device *dev, struct dma_debug_entry *ref, bool to_cpu) check_sync() argument 1246 struct dma_debug_entry ref; debug_dma_mapping_error() local 1285 struct dma_debug_entry ref = { debug_dma_unmap_page() local 1336 get_nr_mapped_entries(struct device *dev, struct dma_debug_entry *ref) get_nr_mapped_entries() argument 1366 struct dma_debug_entry ref = { debug_dma_unmap_sg() local 1425 struct dma_debug_entry ref = { debug_dma_free_coherent() local 1477 struct dma_debug_entry ref = { debug_dma_unmap_resource() local 1494 struct dma_debug_entry ref; debug_dma_sync_single_for_cpu() local 1513 struct dma_debug_entry ref; debug_dma_sync_single_for_device() local 1539 struct dma_debug_entry ref = { debug_dma_sync_sg_for_cpu() local 1571 struct dma_debug_entry ref = { debug_dma_sync_sg_for_device() local [all...] |
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | ref-verify.c | 12 #include "delayed-ref.h" 13 #include "ref-verify.h" 46 * back to the delayed ref action. We hold the ref we are changing in the 54 struct ref_entry ref; member 62 * to it as well as all of the ref actions that have occurred to it. We never 165 struct ref_entry *ref) in insert_ref_entry() 175 cmp = comp_refs(entry, ref); in insert_ref_entry() 184 rb_link_node(&ref->node, parent_node, p); in insert_ref_entry() 185 rb_insert_color(&ref in insert_ref_entry() 164 insert_ref_entry(struct rb_root *root, struct ref_entry *ref) insert_ref_entry() argument 238 struct ref_entry *ref; free_block_entry() local 317 struct ref_entry *ref = NULL, *exist; add_tree_block() local 362 struct ref_entry *ref; add_shared_data_ref() local 392 struct ref_entry *ref; add_extent_data_ref() local 626 struct ref_entry *ref; dump_block_entry() local 665 struct ref_entry *ref = NULL, *exist; btrfs_ref_tree_mod() local [all...] |
H A D | delayed-ref.c | 11 #include "delayed-ref.h" 57 * Release a ref head's reservation. 62 * Drops the delayed ref head's count from the delayed refs rsv and free any 287 /* insert a new ref to head ref rbtree */ 367 * Find a head entry based on bytenr. This returns the delayed ref head if it 427 struct btrfs_delayed_ref_node *ref) in drop_delayed_ref() 430 rb_erase_cached(&ref->ref_node, &head->ref_tree); in drop_delayed_ref() 431 RB_CLEAR_NODE(&ref->ref_node); in drop_delayed_ref() 432 if (!list_empty(&ref in drop_delayed_ref() 425 drop_delayed_ref(struct btrfs_delayed_ref_root *delayed_refs, struct btrfs_delayed_ref_head *head, struct btrfs_delayed_ref_node *ref) drop_delayed_ref() argument 438 merge_ref(struct btrfs_delayed_ref_root *delayed_refs, struct btrfs_delayed_ref_head *head, struct btrfs_delayed_ref_node *ref, u64 seq) merge_ref() argument 488 struct btrfs_delayed_ref_node *ref; btrfs_merge_delayed_refs() local 587 insert_delayed_ref(struct btrfs_delayed_ref_root *root, struct btrfs_delayed_ref_head *href, struct btrfs_delayed_ref_node *ref) insert_delayed_ref() argument 864 init_delayed_ref_common(struct btrfs_fs_info *fs_info, struct btrfs_delayed_ref_node *ref, u64 bytenr, u64 num_bytes, u64 ref_root, int action, u8 ref_type) init_delayed_ref_common() argument 898 struct btrfs_delayed_tree_ref *ref; btrfs_add_delayed_tree_ref() local 991 struct btrfs_delayed_data_ref *ref; btrfs_add_delayed_data_ref() local [all...] |
/kernel/linux/linux-5.10/arch/powerpc/perf/ |
H A D | imc-pmu.c | 354 struct imc_pmu_ref *ref; in ppc_nest_imc_cpu_offline() local 405 ref = get_nest_pmu_ref(cpu); in ppc_nest_imc_cpu_offline() 406 if (!ref) in ppc_nest_imc_cpu_offline() 409 ref->refc = 0; in ppc_nest_imc_cpu_offline() 455 struct imc_pmu_ref *ref; in nest_imc_counters_release() local 468 ref = get_nest_pmu_ref(event->cpu); in nest_imc_counters_release() 469 if (!ref) in nest_imc_counters_release() 473 spin_lock(&ref->lock); in nest_imc_counters_release() 474 if (ref->refc == 0) { in nest_imc_counters_release() 480 * function set the ref in nest_imc_counters_release() 510 struct imc_pmu_ref *ref; nest_imc_event_init() local 670 struct imc_pmu_ref *ref; ppc_core_imc_cpu_offline() local 757 struct imc_pmu_ref *ref; core_imc_counters_release() local 813 struct imc_pmu_ref *ref; core_imc_event_init() local 1118 struct imc_pmu_ref *ref; thread_imc_event_add() local 1160 struct imc_pmu_ref *ref; thread_imc_event_del() local 1353 struct imc_pmu_ref *ref = NULL; trace_imc_event_add() local 1403 struct imc_pmu_ref *ref = NULL; trace_imc_event_del() local [all...] |
/kernel/linux/linux-6.6/arch/powerpc/perf/ |
H A D | imc-pmu.c | 361 struct imc_pmu_ref *ref; in ppc_nest_imc_cpu_offline() local 412 ref = get_nest_pmu_ref(cpu); in ppc_nest_imc_cpu_offline() 413 if (!ref) in ppc_nest_imc_cpu_offline() 416 ref->refc = 0; in ppc_nest_imc_cpu_offline() 462 struct imc_pmu_ref *ref; in nest_imc_counters_release() local 475 ref = get_nest_pmu_ref(event->cpu); in nest_imc_counters_release() 476 if (!ref) in nest_imc_counters_release() 480 spin_lock(&ref->lock); in nest_imc_counters_release() 481 if (ref->refc == 0) { in nest_imc_counters_release() 487 * function set the ref in nest_imc_counters_release() 517 struct imc_pmu_ref *ref; nest_imc_event_init() local 677 struct imc_pmu_ref *ref; ppc_core_imc_cpu_offline() local 764 struct imc_pmu_ref *ref; core_imc_counters_release() local 820 struct imc_pmu_ref *ref; core_imc_event_init() local 1125 struct imc_pmu_ref *ref; thread_imc_event_add() local 1167 struct imc_pmu_ref *ref; thread_imc_event_del() local 1360 struct imc_pmu_ref *ref = NULL; trace_imc_event_add() local 1410 struct imc_pmu_ref *ref = NULL; trace_imc_event_del() local [all...] |
/kernel/linux/linux-5.10/drivers/xen/ |
H A D | grant-table.c | 117 void (*update_entry)(grant_ref_t ref, domid_t domid, 127 int (*end_foreign_access_ref)(grant_ref_t ref, int readonly); 135 unsigned long (*end_foreign_transfer_ref)(grant_ref_t ref); 139 unsigned long (*read_frame)(grant_ref_t ref); 169 int ref, rc = 0; in get_free_entries() local 180 ref = head = gnttab_free_head; in get_free_entries() 189 return ref; in get_free_entries() 218 static void put_free_entry(grant_ref_t ref) in put_free_entry() argument 222 gnttab_entry(ref) = gnttab_free_head; in put_free_entry() 223 gnttab_free_head = ref; in put_free_entry() 240 gnttab_update_entry_v1(grant_ref_t ref, domid_t domid, unsigned long frame, unsigned flags) gnttab_update_entry_v1() argument 249 gnttab_update_entry_v2(grant_ref_t ref, domid_t domid, unsigned long frame, unsigned int flags) gnttab_update_entry_v2() argument 261 gnttab_grant_foreign_access_ref(grant_ref_t ref, domid_t domid, unsigned long frame, int readonly) gnttab_grant_foreign_access_ref() argument 272 int ref; gnttab_grant_foreign_access() local 284 gnttab_end_foreign_access_ref_v1(grant_ref_t ref, int readonly) gnttab_end_foreign_access_ref_v1() argument 300 gnttab_end_foreign_access_ref_v2(grant_ref_t ref, int readonly) gnttab_end_foreign_access_ref_v2() argument 323 _gnttab_end_foreign_access_ref(grant_ref_t ref, int readonly) _gnttab_end_foreign_access_ref() argument 328 gnttab_end_foreign_access_ref(grant_ref_t ref, int readonly) gnttab_end_foreign_access_ref() argument 337 gnttab_read_frame_v1(grant_ref_t ref) gnttab_read_frame_v1() argument 342 gnttab_read_frame_v2(grant_ref_t ref) gnttab_read_frame_v2() argument 349 grant_ref_t ref; global() member 400 gnttab_add_deferred(grant_ref_t ref, bool readonly, struct page *page) gnttab_add_deferred() argument 435 gnttab_try_end_foreign_access(grant_ref_t ref) gnttab_try_end_foreign_access() argument 446 gnttab_end_foreign_access(grant_ref_t ref, int readonly, unsigned long page) gnttab_end_foreign_access() argument 460 int ref; gnttab_grant_foreign_transfer() local 471 gnttab_grant_foreign_transfer_ref(grant_ref_t ref, domid_t domid, unsigned long pfn) gnttab_grant_foreign_transfer_ref() argument 478 gnttab_end_foreign_transfer_ref_v1(grant_ref_t ref) gnttab_end_foreign_transfer_ref_v1() argument 509 gnttab_end_foreign_transfer_ref_v2(grant_ref_t ref) gnttab_end_foreign_transfer_ref_v2() argument 540 gnttab_end_foreign_transfer_ref(grant_ref_t ref) gnttab_end_foreign_transfer_ref() argument 546 gnttab_end_foreign_transfer(grant_ref_t ref) gnttab_end_foreign_transfer() argument 554 gnttab_free_grant_reference(grant_ref_t ref) gnttab_free_grant_reference() argument 562 grant_ref_t ref; gnttab_free_grant_references() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/ |
H A D | i915_active.c | 19 struct kref ref; member 25 kref_get(&active->ref); in __live_get() 34 static void __live_release(struct kref *ref) in __live_release() argument 36 struct live_active *active = container_of(ref, typeof(*active), ref); in __live_release() 43 kref_put(&active->ref, __live_release); in __live_put() 70 kref_init(&active->ref); in __live_alloc() 115 pr_err("Failed to track active ref!\n"); in __live_active_setup() 278 void i915_active_print(struct i915_active *ref, struct drm_printer *m) in i915_active_print() argument 280 drm_printf(m, "active %ps:%ps\n", ref in i915_active_print() 314 active_flush(struct i915_active *ref, struct i915_active_fence *active) active_flush() argument 331 i915_active_unlock_wait(struct i915_active *ref) i915_active_unlock_wait() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/selftests/ |
H A D | i915_active.c | 20 struct kref ref; member 26 kref_get(&active->ref); in __live_get() 35 static void __live_release(struct kref *ref) in __live_release() argument 37 struct live_active *active = container_of(ref, typeof(*active), ref); in __live_release() 44 kref_put(&active->ref, __live_release); in __live_put() 71 kref_init(&active->ref); in __live_alloc() 116 pr_err("Failed to track active ref!\n"); in __live_active_setup() 279 void i915_active_print(struct i915_active *ref, struct drm_printer *m) in i915_active_print() argument 281 drm_printf(m, "active %ps:%ps\n", ref in i915_active_print() 315 active_flush(struct i915_active *ref, struct i915_active_fence *active) active_flush() argument 332 i915_active_unlock_wait(struct i915_active *ref) i915_active_unlock_wait() argument [all...] |