Home
last modified time | relevance | path

Searched refs:ref (Results 1 - 25 of 1461) sorted by relevance

12345678910>>...59

/kernel/linux/linux-5.10/drivers/gpu/drm/i915/
H A Di915_active.c33 struct i915_active *ref; member
81 struct i915_active *ref = addr; in active_debug_hint() local
83 return (void *)ref->active ?: (void *)ref->retire ?: (void *)ref; in active_debug_hint()
91 static void debug_active_init(struct i915_active *ref) in debug_active_init() argument
93 debug_object_init(ref, &active_debug_desc); in debug_active_init()
96 static void debug_active_activate(struct i915_active *ref) in debug_active_activate() argument
98 lockdep_assert_held(&ref->tree_lock); in debug_active_activate()
99 debug_object_activate(ref, in debug_active_activate()
102 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument
109 debug_active_fini(struct i915_active *ref) debug_active_fini() argument
114 debug_active_assert(struct i915_active *ref) debug_active_assert() argument
121 debug_active_init(struct i915_active *ref) debug_active_init() argument
122 debug_active_activate(struct i915_active *ref) debug_active_activate() argument
123 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument
124 debug_active_fini(struct i915_active *ref) debug_active_fini() argument
125 debug_active_assert(struct i915_active *ref) debug_active_assert() argument
130 __active_retire(struct i915_active *ref) __active_retire() argument
184 struct i915_active *ref = container_of(wrk, typeof(*ref), work); active_work() local
194 active_retire(struct i915_active *ref) active_retire() argument
237 __active_lookup(struct i915_active *ref, u64 idx) __active_lookup() argument
296 active_instance(struct i915_active *ref, u64 idx) active_instance() argument
345 __i915_active_init(struct i915_active *ref, int (*active)(struct i915_active *ref), void (*retire)(struct i915_active *ref), struct lock_class_key *mkey, struct lock_class_key *wkey) __i915_active_init() argument
375 ____active_del_barrier(struct i915_active *ref, struct active_node *node, struct intel_engine_cs *engine) ____active_del_barrier() argument
418 __active_del_barrier(struct i915_active *ref, struct active_node *node) __active_del_barrier() argument
424 replace_barrier(struct i915_active *ref, struct i915_active_fence *active) replace_barrier() argument
437 i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence) i915_active_ref() argument
472 __i915_active_set_fence(struct i915_active *ref, struct i915_active_fence *active, struct dma_fence *fence) __i915_active_set_fence() argument
491 __active_fence(struct i915_active *ref, u64 idx) __active_fence() argument
507 __i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence) __i915_active_ref() argument
514 i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f) i915_active_set_exclusive() argument
520 i915_active_acquire_if_busy(struct i915_active *ref) i915_active_acquire_if_busy() argument
526 __i915_active_activate(struct i915_active *ref) __i915_active_activate() argument
534 i915_active_acquire(struct i915_active *ref) i915_active_acquire() argument
561 i915_active_acquire_for_context(struct i915_active *ref, u64 idx) i915_active_acquire_for_context() argument
579 i915_active_release(struct i915_active *ref) i915_active_release() argument
615 flush_lazy_signals(struct i915_active *ref) flush_lazy_signals() argument
632 __i915_active_wait(struct i915_active *ref, int state) __i915_active_wait() argument
682 struct i915_active *ref; global() member
699 __await_barrier(struct i915_active *ref, struct i915_sw_fence *fence) __await_barrier() argument
722 await_active(struct i915_active *ref, unsigned int flags, int (*fn)(void *arg, struct dma_fence *fence), void *arg, struct i915_sw_fence *barrier) await_active() argument
769 i915_request_await_active(struct i915_request *rq, struct i915_active *ref, unsigned int flags) i915_request_await_active() argument
782 i915_sw_fence_await_active(struct i915_sw_fence *fence, struct i915_active *ref, unsigned int flags) i915_sw_fence_await_active() argument
789 i915_active_fini(struct i915_active *ref) i915_active_fini() argument
805 reuse_idle_barrier(struct i915_active *ref, u64 idx) reuse_idle_barrier() argument
888 i915_active_acquire_preallocate_barrier(struct i915_active *ref, struct intel_engine_cs *engine) i915_active_acquire_preallocate_barrier() argument
970 i915_active_acquire_barrier(struct i915_active *ref) i915_active_acquire_barrier() argument
1172 struct kref ref; global() member
1175 i915_active_get(struct i915_active *ref) i915_active_get() argument
1183 auto_release(struct kref *ref) auto_release() argument
1191 i915_active_put(struct i915_active *ref) i915_active_put() argument
1198 auto_active(struct i915_active *ref) auto_active() argument
1205 auto_retire(struct i915_active *ref) auto_retire() argument
[all...]
H A Di915_active.h152 void __i915_active_init(struct i915_active *ref,
153 int (*active)(struct i915_active *ref),
154 void (*retire)(struct i915_active *ref),
159 #define i915_active_init(ref, active, retire) do { \
163 __i915_active_init(ref, active, retire, &__mkey, &__wkey); \
167 __i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence);
168 int i915_active_ref(struct i915_active *ref, u64 idx, struct dma_fence *fence);
171 i915_active_add_request(struct i915_active *ref, struct i915_request *rq) in i915_active_add_request() argument
173 return i915_active_ref(ref, in i915_active_add_request()
179 i915_active_set_exclusive(struct i915_active *ref, struc
181 i915_active_has_exclusive(struct i915_active *ref) i915_active_has_exclusive() argument
187 i915_active_wait(struct i915_active *ref) i915_active_wait() argument
208 __i915_active_acquire(struct i915_active *ref) __i915_active_acquire() argument
215 i915_active_is_idle(const struct i915_active *ref) i915_active_is_idle() argument
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/
H A Di915_active.c29 struct i915_active *ref; member
77 struct i915_active *ref = addr; in active_debug_hint() local
79 return (void *)ref->active ?: (void *)ref->retire ?: (void *)ref; in active_debug_hint()
87 static void debug_active_init(struct i915_active *ref) in debug_active_init() argument
89 debug_object_init(ref, &active_debug_desc); in debug_active_init()
92 static void debug_active_activate(struct i915_active *ref) in debug_active_activate() argument
94 lockdep_assert_held(&ref->tree_lock); in debug_active_activate()
95 debug_object_activate(ref, in debug_active_activate()
98 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument
105 debug_active_fini(struct i915_active *ref) debug_active_fini() argument
110 debug_active_assert(struct i915_active *ref) debug_active_assert() argument
117 debug_active_init(struct i915_active *ref) debug_active_init() argument
118 debug_active_activate(struct i915_active *ref) debug_active_activate() argument
119 debug_active_deactivate(struct i915_active *ref) debug_active_deactivate() argument
120 debug_active_fini(struct i915_active *ref) debug_active_fini() argument
121 debug_active_assert(struct i915_active *ref) debug_active_assert() argument
126 __active_retire(struct i915_active *ref) __active_retire() argument
179 struct i915_active *ref = container_of(wrk, typeof(*ref), work); active_work() local
189 active_retire(struct i915_active *ref) active_retire() argument
232 __active_lookup(struct i915_active *ref, u64 idx) __active_lookup() argument
289 active_instance(struct i915_active *ref, u64 idx) active_instance() argument
338 __i915_active_init(struct i915_active *ref, int (*active)(struct i915_active *ref), void (*retire)(struct i915_active *ref), unsigned long flags, struct lock_class_key *mkey, struct lock_class_key *wkey) __i915_active_init() argument
365 ____active_del_barrier(struct i915_active *ref, struct active_node *node, struct intel_engine_cs *engine) ____active_del_barrier() argument
408 __active_del_barrier(struct i915_active *ref, struct active_node *node) __active_del_barrier() argument
414 replace_barrier(struct i915_active *ref, struct i915_active_fence *active) replace_barrier() argument
427 i915_active_add_request(struct i915_active *ref, struct i915_request *rq) i915_active_add_request() argument
464 __i915_active_set_fence(struct i915_active *ref, struct i915_active_fence *active, struct dma_fence *fence) __i915_active_set_fence() argument
483 i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f) i915_active_set_exclusive() argument
489 i915_active_acquire_if_busy(struct i915_active *ref) i915_active_acquire_if_busy() argument
495 __i915_active_activate(struct i915_active *ref) __i915_active_activate() argument
503 i915_active_acquire(struct i915_active *ref) i915_active_acquire() argument
530 i915_active_acquire_for_context(struct i915_active *ref, u64 idx) i915_active_acquire_for_context() argument
548 i915_active_release(struct i915_active *ref) i915_active_release() argument
584 flush_lazy_signals(struct i915_active *ref) flush_lazy_signals() argument
601 __i915_active_wait(struct i915_active *ref, int state) __i915_active_wait() argument
651 struct i915_active *ref; global() member
668 __await_barrier(struct i915_active *ref, struct i915_sw_fence *fence) __await_barrier() argument
691 await_active(struct i915_active *ref, unsigned int flags, int (*fn)(void *arg, struct dma_fence *fence), void *arg, struct i915_sw_fence *barrier) await_active() argument
738 i915_request_await_active(struct i915_request *rq, struct i915_active *ref, unsigned int flags) i915_request_await_active() argument
751 i915_sw_fence_await_active(struct i915_sw_fence *fence, struct i915_active *ref, unsigned int flags) i915_sw_fence_await_active() argument
758 i915_active_fini(struct i915_active *ref) i915_active_fini() argument
774 reuse_idle_barrier(struct i915_active *ref, u64 idx) reuse_idle_barrier() argument
857 i915_active_acquire_preallocate_barrier(struct i915_active *ref, struct intel_engine_cs *engine) i915_active_acquire_preallocate_barrier() argument
939 i915_active_acquire_barrier(struct i915_active *ref) i915_active_acquire_barrier() argument
1141 struct kref ref; global() member
1144 i915_active_get(struct i915_active *ref) i915_active_get() argument
1152 auto_release(struct kref *ref) auto_release() argument
1160 i915_active_put(struct i915_active *ref) i915_active_put() argument
1167 auto_active(struct i915_active *ref) auto_active() argument
1173 auto_retire(struct i915_active *ref) auto_retire() argument
[all...]
H A Di915_active.h152 void __i915_active_init(struct i915_active *ref,
153 int (*active)(struct i915_active *ref),
154 void (*retire)(struct i915_active *ref),
160 #define i915_active_init(ref, active, retire, flags) do { \
164 __i915_active_init(ref, active, retire, flags, &__mkey, &__wkey); \
167 int i915_active_add_request(struct i915_active *ref, struct i915_request *rq);
170 i915_active_set_exclusive(struct i915_active *ref, struct dma_fence *f);
172 int __i915_active_wait(struct i915_active *ref, int state);
173 static inline int i915_active_wait(struct i915_active *ref) in i915_active_wait() argument
175 return __i915_active_wait(ref, TASK_INTERRUPTIBL in i915_active_wait()
194 __i915_active_acquire(struct i915_active *ref) __i915_active_acquire() argument
201 i915_active_is_idle(const struct i915_active *ref) i915_active_is_idle() argument
[all...]
/kernel/linux/linux-5.10/lib/
H A Dpercpu-refcount.c12 * don't try to detect the ref hitting 0 - which means that get/put can just
23 * the ref hitting 0 on every put - this would require global synchronization
27 * the ref can't hit 0 before the user drops the initial ref, so as long as we
28 * convert to non percpu mode before the initial ref is dropped everything
41 static unsigned long __percpu *percpu_count_ptr(struct percpu_ref *ref) in percpu_count_ptr() argument
44 (ref->percpu_count_ptr & ~__PERCPU_REF_ATOMIC_DEAD); in percpu_count_ptr()
49 * @ref: percpu_ref to initialize
54 * Initializes @ref. @ref start
62 percpu_ref_init(struct percpu_ref *ref, percpu_ref_func_t *release, unsigned int flags, gfp_t gfp) percpu_ref_init() argument
107 __percpu_ref_exit(struct percpu_ref *ref) __percpu_ref_exit() argument
129 percpu_ref_exit(struct percpu_ref *ref) percpu_ref_exit() argument
153 struct percpu_ref *ref = data->ref; percpu_ref_call_confirm_rcu() local
170 struct percpu_ref *ref = data->ref; percpu_ref_switch_to_atomic_rcu() local
203 percpu_ref_noop_confirm_switch(struct percpu_ref *ref) percpu_ref_noop_confirm_switch() argument
207 __percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_to_atomic() argument
230 __percpu_ref_switch_to_percpu(struct percpu_ref *ref) __percpu_ref_switch_to_percpu() argument
258 __percpu_ref_switch_mode(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_mode() argument
299 percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) percpu_ref_switch_to_atomic() argument
321 percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref) percpu_ref_switch_to_atomic_sync() argument
346 percpu_ref_switch_to_percpu(struct percpu_ref *ref) percpu_ref_switch_to_percpu() argument
376 percpu_ref_kill_and_confirm(struct percpu_ref *ref, percpu_ref_func_t *confirm_kill) percpu_ref_kill_and_confirm() argument
403 percpu_ref_is_zero(struct percpu_ref *ref) percpu_ref_is_zero() argument
434 percpu_ref_reinit(struct percpu_ref *ref) percpu_ref_reinit() argument
456 percpu_ref_resurrect(struct percpu_ref *ref) percpu_ref_resurrect() argument
[all...]
/kernel/linux/linux-6.6/lib/
H A Dpercpu-refcount.c13 * don't try to detect the ref hitting 0 - which means that get/put can just
24 * the ref hitting 0 on every put - this would require global synchronization
28 * the ref can't hit 0 before the user drops the initial ref, so as long as we
29 * convert to non percpu mode before the initial ref is dropped everything
42 static unsigned long __percpu *percpu_count_ptr(struct percpu_ref *ref) in percpu_count_ptr() argument
45 (ref->percpu_count_ptr & ~__PERCPU_REF_ATOMIC_DEAD); in percpu_count_ptr()
50 * @ref: percpu_ref to initialize
55 * Initializes @ref. @ref start
63 percpu_ref_init(struct percpu_ref *ref, percpu_ref_func_t *release, unsigned int flags, gfp_t gfp) percpu_ref_init() argument
108 __percpu_ref_exit(struct percpu_ref *ref) __percpu_ref_exit() argument
130 percpu_ref_exit(struct percpu_ref *ref) percpu_ref_exit() argument
154 struct percpu_ref *ref = data->ref; percpu_ref_call_confirm_rcu() local
171 struct percpu_ref *ref = data->ref; percpu_ref_switch_to_atomic_rcu() local
209 percpu_ref_noop_confirm_switch(struct percpu_ref *ref) percpu_ref_noop_confirm_switch() argument
213 __percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_to_atomic() argument
237 __percpu_ref_switch_to_percpu(struct percpu_ref *ref) __percpu_ref_switch_to_percpu() argument
265 __percpu_ref_switch_mode(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) __percpu_ref_switch_mode() argument
306 percpu_ref_switch_to_atomic(struct percpu_ref *ref, percpu_ref_func_t *confirm_switch) percpu_ref_switch_to_atomic() argument
328 percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref) percpu_ref_switch_to_atomic_sync() argument
353 percpu_ref_switch_to_percpu(struct percpu_ref *ref) percpu_ref_switch_to_percpu() argument
383 percpu_ref_kill_and_confirm(struct percpu_ref *ref, percpu_ref_func_t *confirm_kill) percpu_ref_kill_and_confirm() argument
410 percpu_ref_is_zero(struct percpu_ref *ref) percpu_ref_is_zero() argument
441 percpu_ref_reinit(struct percpu_ref *ref) percpu_ref_reinit() argument
463 percpu_ref_resurrect(struct percpu_ref *ref) percpu_ref_resurrect() argument
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dpercpu-refcount.h21 * puts the ref back in single atomic_t mode, collecting the per cpu refs and
22 * issuing the appropriate barriers, and then marks the ref as shutting down so
23 * that percpu_ref_put() will check for the ref hitting 0. After it returns,
24 * it's safe to drop the initial ref.
35 * and it's then safe to drop the initial ref with percpu_ref_put().
74 * Start w/ ref == 1 in atomic mode. Can be switched to percpu
76 * with this flag, the ref will stay in atomic mode until
83 * Start dead w/ ref == 0 in atomic mode. Must be revived with
102 struct percpu_ref *ref; member
107 * The low bit of the pointer indicates whether the ref i
147 percpu_ref_kill(struct percpu_ref *ref) percpu_ref_kill() argument
158 __ref_is_percpu(struct percpu_ref *ref, unsigned long __percpu **percpu_countp) __ref_is_percpu() argument
198 percpu_ref_get_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_get_many() argument
220 percpu_ref_get(struct percpu_ref *ref) percpu_ref_get() argument
235 percpu_ref_tryget_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_tryget_many() argument
264 percpu_ref_tryget(struct percpu_ref *ref) percpu_ref_tryget() argument
275 percpu_ref_tryget_live_rcu(struct percpu_ref *ref) percpu_ref_tryget_live_rcu() argument
306 percpu_ref_tryget_live(struct percpu_ref *ref) percpu_ref_tryget_live() argument
326 percpu_ref_put_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_put_many() argument
349 percpu_ref_put(struct percpu_ref *ref) percpu_ref_put() argument
363 percpu_ref_is_dying(struct percpu_ref *ref) percpu_ref_is_dying() argument
[all...]
H A Drcuref.h21 * @ref: Pointer to the reference count
24 static inline void rcuref_init(rcuref_t *ref, unsigned int cnt) in rcuref_init() argument
26 atomic_set(&ref->refcnt, cnt - 1); in rcuref_init()
31 * @ref: Pointer to the reference count
35 static inline unsigned int rcuref_read(rcuref_t *ref) in rcuref_read() argument
37 unsigned int c = atomic_read(&ref->refcnt); in rcuref_read()
43 extern __must_check bool rcuref_get_slowpath(rcuref_t *ref);
47 * @ref: Pointer to the reference count
61 static inline __must_check bool rcuref_get(rcuref_t *ref) in rcuref_get() argument
67 if (likely(!atomic_add_negative_relaxed(1, &ref in rcuref_get()
79 __rcuref_put(rcuref_t *ref) __rcuref_put() argument
119 rcuref_put_rcusafe(rcuref_t *ref) rcuref_put_rcusafe() argument
145 rcuref_put(rcuref_t *ref) rcuref_put() argument
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Dpercpu-refcount.h21 * puts the ref back in single atomic_t mode, collecting the per cpu refs and
22 * issuing the appropriate barriers, and then marks the ref as shutting down so
23 * that percpu_ref_put() will check for the ref hitting 0. After it returns,
24 * it's safe to drop the initial ref.
35 * and it's then safe to drop the initial ref with percpu_ref_put().
74 * Start w/ ref == 1 in atomic mode. Can be switched to percpu
76 * with this flag, the ref will stay in atomic mode until
83 * Start dead w/ ref == 0 in atomic mode. Must be revived with
102 struct percpu_ref *ref; member
107 * The low bit of the pointer indicates whether the ref i
147 percpu_ref_kill(struct percpu_ref *ref) percpu_ref_kill() argument
158 __ref_is_percpu(struct percpu_ref *ref, unsigned long __percpu **percpu_countp) __ref_is_percpu() argument
198 percpu_ref_get_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_get_many() argument
220 percpu_ref_get(struct percpu_ref *ref) percpu_ref_get() argument
235 percpu_ref_tryget_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_tryget_many() argument
264 percpu_ref_tryget(struct percpu_ref *ref) percpu_ref_tryget() argument
284 percpu_ref_tryget_live(struct percpu_ref *ref) percpu_ref_tryget_live() argument
313 percpu_ref_put_many(struct percpu_ref *ref, unsigned long nr) percpu_ref_put_many() argument
336 percpu_ref_put(struct percpu_ref *ref) percpu_ref_put() argument
350 percpu_ref_is_dying(struct percpu_ref *ref) percpu_ref_is_dying() argument
[all...]
/kernel/linux/linux-5.10/fs/jffs2/
H A Dxattr.c424 * verify_xattr_ref(c, ref)
427 * save_xattr_ref(c, ref)
432 * delete_xattr_ref(c, ref)
444 static int verify_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) in verify_xattr_ref() argument
454 if (ref_flags(ref->node) != REF_UNCHECKED) in verify_xattr_ref()
456 offset = ref_offset(ref->node); in verify_xattr_ref()
482 ref->ino = je32_to_cpu(rr.ino); in verify_xattr_ref()
483 ref->xid = je32_to_cpu(rr.xid); in verify_xattr_ref()
484 ref->xseqno = je32_to_cpu(rr.xseqno); in verify_xattr_ref()
485 if (ref in verify_xattr_ref()
506 save_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) save_xattr_ref() argument
554 struct jffs2_xattr_ref *ref; create_xattr_ref() local
576 delete_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) delete_xattr_ref() argument
600 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_delete_inode() local
618 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_free_inode() local
640 struct jffs2_xattr_ref *ref, *cmp, **pref, **pcmp; check_xattr_ref_inode() local
743 struct jffs2_xattr_ref *ref, *_ref; jffs2_clear_xattr_subsystem() local
777 struct jffs2_xattr_ref *ref, *_ref; jffs2_build_xattr_subsystem() local
972 struct jffs2_xattr_ref *ref, **pref; jffs2_listxattr() local
1043 struct jffs2_xattr_ref *ref, **pref; do_jffs2_getxattr() local
1105 struct jffs2_xattr_ref *ref, *newref, **pref; do_jffs2_setxattr() local
1260 jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref, struct jffs2_raw_node_ref *raw) jffs2_garbage_collect_xattr_ref() argument
1337 jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) jffs2_release_xattr_ref() argument
[all...]
H A Dmalloc.c204 struct jffs2_raw_node_ref **p, *ref; in jffs2_prealloc_raw_node_refs() local
210 ref = *p; in jffs2_prealloc_raw_node_refs()
215 if (ref && ref->flash_offset != REF_EMPTY_NODE) in jffs2_prealloc_raw_node_refs()
216 ref++; in jffs2_prealloc_raw_node_refs()
219 if (!ref) { in jffs2_prealloc_raw_node_refs()
221 ref = *p = jffs2_alloc_refblock(); in jffs2_prealloc_raw_node_refs()
222 if (!ref) in jffs2_prealloc_raw_node_refs()
225 if (ref->flash_offset == REF_LINK_NODE) { in jffs2_prealloc_raw_node_refs()
226 p = &ref in jffs2_prealloc_raw_node_refs()
299 struct jffs2_xattr_ref *ref; jffs2_alloc_xattr_ref() local
310 jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref) jffs2_free_xattr_ref() argument
[all...]
/kernel/linux/linux-6.6/fs/jffs2/
H A Dxattr.c424 * verify_xattr_ref(c, ref)
427 * save_xattr_ref(c, ref)
432 * delete_xattr_ref(c, ref)
444 static int verify_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) in verify_xattr_ref() argument
454 if (ref_flags(ref->node) != REF_UNCHECKED) in verify_xattr_ref()
456 offset = ref_offset(ref->node); in verify_xattr_ref()
482 ref->ino = je32_to_cpu(rr.ino); in verify_xattr_ref()
483 ref->xid = je32_to_cpu(rr.xid); in verify_xattr_ref()
484 ref->xseqno = je32_to_cpu(rr.xseqno); in verify_xattr_ref()
485 if (ref in verify_xattr_ref()
506 save_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) save_xattr_ref() argument
554 struct jffs2_xattr_ref *ref; create_xattr_ref() local
576 delete_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) delete_xattr_ref() argument
600 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_delete_inode() local
618 struct jffs2_xattr_ref *ref, *_ref; jffs2_xattr_free_inode() local
640 struct jffs2_xattr_ref *ref, *cmp, **pref, **pcmp; check_xattr_ref_inode() local
743 struct jffs2_xattr_ref *ref, *_ref; jffs2_clear_xattr_subsystem() local
777 struct jffs2_xattr_ref *ref, *_ref; jffs2_build_xattr_subsystem() local
972 struct jffs2_xattr_ref *ref, **pref; jffs2_listxattr() local
1042 struct jffs2_xattr_ref *ref, **pref; do_jffs2_getxattr() local
1104 struct jffs2_xattr_ref *ref, *newref, **pref; do_jffs2_setxattr() local
1259 jffs2_garbage_collect_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref, struct jffs2_raw_node_ref *raw) jffs2_garbage_collect_xattr_ref() argument
1336 jffs2_release_xattr_ref(struct jffs2_sb_info *c, struct jffs2_xattr_ref *ref) jffs2_release_xattr_ref() argument
[all...]
H A Dmalloc.c204 struct jffs2_raw_node_ref **p, *ref; in jffs2_prealloc_raw_node_refs() local
210 ref = *p; in jffs2_prealloc_raw_node_refs()
215 if (ref && ref->flash_offset != REF_EMPTY_NODE) in jffs2_prealloc_raw_node_refs()
216 ref++; in jffs2_prealloc_raw_node_refs()
219 if (!ref) { in jffs2_prealloc_raw_node_refs()
221 ref = *p = jffs2_alloc_refblock(); in jffs2_prealloc_raw_node_refs()
222 if (!ref) in jffs2_prealloc_raw_node_refs()
225 if (ref->flash_offset == REF_LINK_NODE) { in jffs2_prealloc_raw_node_refs()
226 p = &ref in jffs2_prealloc_raw_node_refs()
299 struct jffs2_xattr_ref *ref; jffs2_alloc_xattr_ref() local
310 jffs2_free_xattr_ref(struct jffs2_xattr_ref *ref) jffs2_free_xattr_ref() argument
[all...]
/kernel/linux/linux-5.10/drivers/clk/mxs/
H A Dclk-ref.c33 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_enable() local
35 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + CLR); in clk_ref_enable()
42 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_disable() local
44 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + SET); in clk_ref_disable()
50 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_recalc_rate() local
52 u8 frac = (readl_relaxed(ref->reg) >> (ref->idx * 8)) & 0x3f; in clk_ref_recalc_rate()
86 struct clk_ref *ref in clk_ref_set_rate() local
124 struct clk_ref *ref; mxs_clk_ref() local
[all...]
/kernel/linux/linux-6.6/drivers/clk/mxs/
H A Dclk-ref.c33 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_enable() local
35 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + CLR); in clk_ref_enable()
42 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_disable() local
44 writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + SET); in clk_ref_disable()
50 struct clk_ref *ref = to_clk_ref(hw); in clk_ref_recalc_rate() local
52 u8 frac = (readl_relaxed(ref->reg) >> (ref->idx * 8)) & 0x3f; in clk_ref_recalc_rate()
86 struct clk_ref *ref in clk_ref_set_rate() local
124 struct clk_ref *ref; mxs_clk_ref() local
[all...]
/kernel/linux/linux-6.6/fs/smb/client/
H A Ddfs.h25 struct dfs_ref *ref; member
31 #define ref_walk_cur(w) ((w)->ref)
56 static inline void __ref_walk_free(struct dfs_ref *ref) in __ref_walk_free() argument
58 kfree(ref->path); in __ref_walk_free()
59 kfree(ref->full_path); in __ref_walk_free()
60 dfs_cache_free_tgts(&ref->tl); in __ref_walk_free()
61 memset(ref, 0, sizeof(*ref)); in __ref_walk_free()
66 struct dfs_ref *ref = ref_walk_start(rw); in ref_walk_free() local
68 for (; ref < in ref_walk_free()
75 struct dfs_ref *ref = ref_walk_cur(rw) + 1; ref_walk_advance() local
88 struct dfs_ref *ref = ref_walk_cur(rw); ref_walk_next_tgt() local
131 dfs_get_referral(struct cifs_mount_ctx *mnt_ctx, const char *path, struct dfs_info3_param *ref, struct dfs_cache_tgt_list *tl) dfs_get_referral() argument
[all...]
/kernel/linux/linux-5.10/kernel/dma/
H A Ddebug.c293 struct dma_debug_entry *ref, in __hash_bucket_find()
300 if (!match(ref, entry)) in __hash_bucket_find()
315 entry->size == ref->size ? ++match_lvl : 0; in __hash_bucket_find()
316 entry->type == ref->type ? ++match_lvl : 0; in __hash_bucket_find()
317 entry->direction == ref->direction ? ++match_lvl : 0; in __hash_bucket_find()
318 entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0; in __hash_bucket_find()
343 struct dma_debug_entry *ref) in bucket_find_exact()
345 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact()
349 struct dma_debug_entry *ref, in bucket_find_contain()
353 unsigned int max_range = dma_get_max_seg_size(ref in bucket_find_contain()
292 __hash_bucket_find(struct hash_bucket *bucket, struct dma_debug_entry *ref, match_fn match) __hash_bucket_find() argument
342 bucket_find_exact(struct hash_bucket *bucket, struct dma_debug_entry *ref) bucket_find_exact() argument
348 bucket_find_contain(struct hash_bucket **bucket, struct dma_debug_entry *ref, unsigned long *flags) bucket_find_contain() argument
956 check_unmap(struct dma_debug_entry *ref) check_unmap() argument
1097 check_sync(struct device *dev, struct dma_debug_entry *ref, bool to_cpu) check_sync() argument
1251 struct dma_debug_entry ref; debug_dma_mapping_error() local
1290 struct dma_debug_entry ref = { debug_dma_unmap_page() local
1340 get_nr_mapped_entries(struct device *dev, struct dma_debug_entry *ref) get_nr_mapped_entries() argument
1370 struct dma_debug_entry ref = { debug_dma_unmap_sg() local
1428 struct dma_debug_entry ref = { debug_dma_free_coherent() local
1479 struct dma_debug_entry ref = { debug_dma_unmap_resource() local
1496 struct dma_debug_entry ref; debug_dma_sync_single_for_cpu() local
1515 struct dma_debug_entry ref; debug_dma_sync_single_for_device() local
1541 struct dma_debug_entry ref = { debug_dma_sync_sg_for_cpu() local
1573 struct dma_debug_entry ref = { debug_dma_sync_sg_for_device() local
[all...]
/kernel/linux/linux-6.6/kernel/dma/
H A Ddebug.c294 struct dma_debug_entry *ref, in __hash_bucket_find()
301 if (!match(ref, entry)) in __hash_bucket_find()
316 entry->size == ref->size ? ++match_lvl : 0; in __hash_bucket_find()
317 entry->type == ref->type ? ++match_lvl : 0; in __hash_bucket_find()
318 entry->direction == ref->direction ? ++match_lvl : 0; in __hash_bucket_find()
319 entry->sg_call_ents == ref->sg_call_ents ? ++match_lvl : 0; in __hash_bucket_find()
344 struct dma_debug_entry *ref) in bucket_find_exact()
346 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact()
350 struct dma_debug_entry *ref, in bucket_find_contain()
354 struct dma_debug_entry *entry, index = *ref; in bucket_find_contain()
293 __hash_bucket_find(struct hash_bucket *bucket, struct dma_debug_entry *ref, match_fn match) __hash_bucket_find() argument
343 bucket_find_exact(struct hash_bucket *bucket, struct dma_debug_entry *ref) bucket_find_exact() argument
349 bucket_find_contain(struct hash_bucket **bucket, struct dma_debug_entry *ref, unsigned long *flags) bucket_find_contain() argument
960 check_unmap(struct dma_debug_entry *ref) check_unmap() argument
1091 check_sync(struct device *dev, struct dma_debug_entry *ref, bool to_cpu) check_sync() argument
1246 struct dma_debug_entry ref; debug_dma_mapping_error() local
1285 struct dma_debug_entry ref = { debug_dma_unmap_page() local
1336 get_nr_mapped_entries(struct device *dev, struct dma_debug_entry *ref) get_nr_mapped_entries() argument
1366 struct dma_debug_entry ref = { debug_dma_unmap_sg() local
1425 struct dma_debug_entry ref = { debug_dma_free_coherent() local
1477 struct dma_debug_entry ref = { debug_dma_unmap_resource() local
1494 struct dma_debug_entry ref; debug_dma_sync_single_for_cpu() local
1513 struct dma_debug_entry ref; debug_dma_sync_single_for_device() local
1539 struct dma_debug_entry ref = { debug_dma_sync_sg_for_cpu() local
1571 struct dma_debug_entry ref = { debug_dma_sync_sg_for_device() local
[all...]
/kernel/linux/linux-6.6/fs/btrfs/
H A Dref-verify.c12 #include "delayed-ref.h"
13 #include "ref-verify.h"
46 * back to the delayed ref action. We hold the ref we are changing in the
54 struct ref_entry ref; member
62 * to it as well as all of the ref actions that have occurred to it. We never
165 struct ref_entry *ref) in insert_ref_entry()
175 cmp = comp_refs(entry, ref); in insert_ref_entry()
184 rb_link_node(&ref->node, parent_node, p); in insert_ref_entry()
185 rb_insert_color(&ref in insert_ref_entry()
164 insert_ref_entry(struct rb_root *root, struct ref_entry *ref) insert_ref_entry() argument
238 struct ref_entry *ref; free_block_entry() local
317 struct ref_entry *ref = NULL, *exist; add_tree_block() local
362 struct ref_entry *ref; add_shared_data_ref() local
392 struct ref_entry *ref; add_extent_data_ref() local
626 struct ref_entry *ref; dump_block_entry() local
665 struct ref_entry *ref = NULL, *exist; btrfs_ref_tree_mod() local
[all...]
H A Ddelayed-ref.c11 #include "delayed-ref.h"
57 * Release a ref head's reservation.
62 * Drops the delayed ref head's count from the delayed refs rsv and free any
287 /* insert a new ref to head ref rbtree */
367 * Find a head entry based on bytenr. This returns the delayed ref head if it
427 struct btrfs_delayed_ref_node *ref) in drop_delayed_ref()
430 rb_erase_cached(&ref->ref_node, &head->ref_tree); in drop_delayed_ref()
431 RB_CLEAR_NODE(&ref->ref_node); in drop_delayed_ref()
432 if (!list_empty(&ref in drop_delayed_ref()
425 drop_delayed_ref(struct btrfs_delayed_ref_root *delayed_refs, struct btrfs_delayed_ref_head *head, struct btrfs_delayed_ref_node *ref) drop_delayed_ref() argument
438 merge_ref(struct btrfs_delayed_ref_root *delayed_refs, struct btrfs_delayed_ref_head *head, struct btrfs_delayed_ref_node *ref, u64 seq) merge_ref() argument
488 struct btrfs_delayed_ref_node *ref; btrfs_merge_delayed_refs() local
587 insert_delayed_ref(struct btrfs_delayed_ref_root *root, struct btrfs_delayed_ref_head *href, struct btrfs_delayed_ref_node *ref) insert_delayed_ref() argument
864 init_delayed_ref_common(struct btrfs_fs_info *fs_info, struct btrfs_delayed_ref_node *ref, u64 bytenr, u64 num_bytes, u64 ref_root, int action, u8 ref_type) init_delayed_ref_common() argument
898 struct btrfs_delayed_tree_ref *ref; btrfs_add_delayed_tree_ref() local
991 struct btrfs_delayed_data_ref *ref; btrfs_add_delayed_data_ref() local
[all...]
/kernel/linux/linux-5.10/arch/powerpc/perf/
H A Dimc-pmu.c354 struct imc_pmu_ref *ref; in ppc_nest_imc_cpu_offline() local
405 ref = get_nest_pmu_ref(cpu); in ppc_nest_imc_cpu_offline()
406 if (!ref) in ppc_nest_imc_cpu_offline()
409 ref->refc = 0; in ppc_nest_imc_cpu_offline()
455 struct imc_pmu_ref *ref; in nest_imc_counters_release() local
468 ref = get_nest_pmu_ref(event->cpu); in nest_imc_counters_release()
469 if (!ref) in nest_imc_counters_release()
473 spin_lock(&ref->lock); in nest_imc_counters_release()
474 if (ref->refc == 0) { in nest_imc_counters_release()
480 * function set the ref in nest_imc_counters_release()
510 struct imc_pmu_ref *ref; nest_imc_event_init() local
670 struct imc_pmu_ref *ref; ppc_core_imc_cpu_offline() local
757 struct imc_pmu_ref *ref; core_imc_counters_release() local
813 struct imc_pmu_ref *ref; core_imc_event_init() local
1118 struct imc_pmu_ref *ref; thread_imc_event_add() local
1160 struct imc_pmu_ref *ref; thread_imc_event_del() local
1353 struct imc_pmu_ref *ref = NULL; trace_imc_event_add() local
1403 struct imc_pmu_ref *ref = NULL; trace_imc_event_del() local
[all...]
/kernel/linux/linux-6.6/arch/powerpc/perf/
H A Dimc-pmu.c361 struct imc_pmu_ref *ref; in ppc_nest_imc_cpu_offline() local
412 ref = get_nest_pmu_ref(cpu); in ppc_nest_imc_cpu_offline()
413 if (!ref) in ppc_nest_imc_cpu_offline()
416 ref->refc = 0; in ppc_nest_imc_cpu_offline()
462 struct imc_pmu_ref *ref; in nest_imc_counters_release() local
475 ref = get_nest_pmu_ref(event->cpu); in nest_imc_counters_release()
476 if (!ref) in nest_imc_counters_release()
480 spin_lock(&ref->lock); in nest_imc_counters_release()
481 if (ref->refc == 0) { in nest_imc_counters_release()
487 * function set the ref in nest_imc_counters_release()
517 struct imc_pmu_ref *ref; nest_imc_event_init() local
677 struct imc_pmu_ref *ref; ppc_core_imc_cpu_offline() local
764 struct imc_pmu_ref *ref; core_imc_counters_release() local
820 struct imc_pmu_ref *ref; core_imc_event_init() local
1125 struct imc_pmu_ref *ref; thread_imc_event_add() local
1167 struct imc_pmu_ref *ref; thread_imc_event_del() local
1360 struct imc_pmu_ref *ref = NULL; trace_imc_event_add() local
1410 struct imc_pmu_ref *ref = NULL; trace_imc_event_del() local
[all...]
/kernel/linux/linux-5.10/drivers/xen/
H A Dgrant-table.c117 void (*update_entry)(grant_ref_t ref, domid_t domid,
127 int (*end_foreign_access_ref)(grant_ref_t ref, int readonly);
135 unsigned long (*end_foreign_transfer_ref)(grant_ref_t ref);
139 unsigned long (*read_frame)(grant_ref_t ref);
169 int ref, rc = 0; in get_free_entries() local
180 ref = head = gnttab_free_head; in get_free_entries()
189 return ref; in get_free_entries()
218 static void put_free_entry(grant_ref_t ref) in put_free_entry() argument
222 gnttab_entry(ref) = gnttab_free_head; in put_free_entry()
223 gnttab_free_head = ref; in put_free_entry()
240 gnttab_update_entry_v1(grant_ref_t ref, domid_t domid, unsigned long frame, unsigned flags) gnttab_update_entry_v1() argument
249 gnttab_update_entry_v2(grant_ref_t ref, domid_t domid, unsigned long frame, unsigned int flags) gnttab_update_entry_v2() argument
261 gnttab_grant_foreign_access_ref(grant_ref_t ref, domid_t domid, unsigned long frame, int readonly) gnttab_grant_foreign_access_ref() argument
272 int ref; gnttab_grant_foreign_access() local
284 gnttab_end_foreign_access_ref_v1(grant_ref_t ref, int readonly) gnttab_end_foreign_access_ref_v1() argument
300 gnttab_end_foreign_access_ref_v2(grant_ref_t ref, int readonly) gnttab_end_foreign_access_ref_v2() argument
323 _gnttab_end_foreign_access_ref(grant_ref_t ref, int readonly) _gnttab_end_foreign_access_ref() argument
328 gnttab_end_foreign_access_ref(grant_ref_t ref, int readonly) gnttab_end_foreign_access_ref() argument
337 gnttab_read_frame_v1(grant_ref_t ref) gnttab_read_frame_v1() argument
342 gnttab_read_frame_v2(grant_ref_t ref) gnttab_read_frame_v2() argument
349 grant_ref_t ref; global() member
400 gnttab_add_deferred(grant_ref_t ref, bool readonly, struct page *page) gnttab_add_deferred() argument
435 gnttab_try_end_foreign_access(grant_ref_t ref) gnttab_try_end_foreign_access() argument
446 gnttab_end_foreign_access(grant_ref_t ref, int readonly, unsigned long page) gnttab_end_foreign_access() argument
460 int ref; gnttab_grant_foreign_transfer() local
471 gnttab_grant_foreign_transfer_ref(grant_ref_t ref, domid_t domid, unsigned long pfn) gnttab_grant_foreign_transfer_ref() argument
478 gnttab_end_foreign_transfer_ref_v1(grant_ref_t ref) gnttab_end_foreign_transfer_ref_v1() argument
509 gnttab_end_foreign_transfer_ref_v2(grant_ref_t ref) gnttab_end_foreign_transfer_ref_v2() argument
540 gnttab_end_foreign_transfer_ref(grant_ref_t ref) gnttab_end_foreign_transfer_ref() argument
546 gnttab_end_foreign_transfer(grant_ref_t ref) gnttab_end_foreign_transfer() argument
554 gnttab_free_grant_reference(grant_ref_t ref) gnttab_free_grant_reference() argument
562 grant_ref_t ref; gnttab_free_grant_references() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/
H A Di915_active.c19 struct kref ref; member
25 kref_get(&active->ref); in __live_get()
34 static void __live_release(struct kref *ref) in __live_release() argument
36 struct live_active *active = container_of(ref, typeof(*active), ref); in __live_release()
43 kref_put(&active->ref, __live_release); in __live_put()
70 kref_init(&active->ref); in __live_alloc()
115 pr_err("Failed to track active ref!\n"); in __live_active_setup()
278 void i915_active_print(struct i915_active *ref, struct drm_printer *m) in i915_active_print() argument
280 drm_printf(m, "active %ps:%ps\n", ref in i915_active_print()
314 active_flush(struct i915_active *ref, struct i915_active_fence *active) active_flush() argument
331 i915_active_unlock_wait(struct i915_active *ref) i915_active_unlock_wait() argument
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/selftests/
H A Di915_active.c20 struct kref ref; member
26 kref_get(&active->ref); in __live_get()
35 static void __live_release(struct kref *ref) in __live_release() argument
37 struct live_active *active = container_of(ref, typeof(*active), ref); in __live_release()
44 kref_put(&active->ref, __live_release); in __live_put()
71 kref_init(&active->ref); in __live_alloc()
116 pr_err("Failed to track active ref!\n"); in __live_active_setup()
279 void i915_active_print(struct i915_active *ref, struct drm_printer *m) in i915_active_print() argument
281 drm_printf(m, "active %ps:%ps\n", ref in i915_active_print()
315 active_flush(struct i915_active *ref, struct i915_active_fence *active) active_flush() argument
332 i915_active_unlock_wait(struct i915_active *ref) i915_active_unlock_wait() argument
[all...]

Completed in 19 milliseconds

12345678910>>...59