/kernel/linux/linux-5.10/fs/ |
H A D | eventfd.c | 34 wait_queue_head_t wqh; member 38 * wakeup is performed on "wqh". A read(2) will return the "count" 63 spin_lock_irqsave(&ctx->wqh.lock, flags); in eventfd_signal_mask() 68 if (waitqueue_active(&ctx->wqh)) in eventfd_signal_mask() 69 wake_up_locked_poll(&ctx->wqh, EPOLLIN | mask); in eventfd_signal_mask() 71 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in eventfd_signal_mask() 127 wake_up_poll(&ctx->wqh, EPOLLHUP); in eventfd_release() 138 poll_wait(file, &ctx->wqh, wait); in eventfd_poll() 141 * All writes to ctx->count occur within ctx->wqh.lock. This read in eventfd_poll() 142 * can be done outside ctx->wqh in eventfd_poll() [all...] |
H A D | timerfd.c | 38 wait_queue_head_t wqh; member 67 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_triggered() 70 wake_up_locked_poll(&ctx->wqh, EPOLLIN); in timerfd_triggered() 71 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_triggered() 107 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 111 wake_up_locked_poll(&ctx->wqh, EPOLLIN); in timerfd_clock_was_set() 113 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 239 poll_wait(file, &ctx->wqh, wait); in timerfd_poll() 241 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_poll() 244 spin_unlock_irqrestore(&ctx->wqh in timerfd_poll() [all...] |
H A D | fs-writeback.c | 1345 wait_queue_head_t *wqh; variable 1347 wqh = bit_waitqueue(&inode->i_state, __I_SYNC); 1350 __wait_on_bit(wqh, &wq, bit_wait, 1375 wait_queue_head_t *wqh = bit_waitqueue(&inode->i_state, __I_SYNC); variable 1378 prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE); 1383 finish_wait(wqh, &wait);
|
/kernel/linux/linux-6.6/fs/ |
H A D | eventfd.c | 32 wait_queue_head_t wqh; member 36 * wakeup is performed on "wqh". If EFD_SEMAPHORE flag was not 61 spin_lock_irqsave(&ctx->wqh.lock, flags); in eventfd_signal_mask() 66 if (waitqueue_active(&ctx->wqh)) in eventfd_signal_mask() 67 wake_up_locked_poll(&ctx->wqh, EPOLLIN | mask); in eventfd_signal_mask() 69 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in eventfd_signal_mask() 125 wake_up_poll(&ctx->wqh, EPOLLHUP); in eventfd_release() 136 poll_wait(file, &ctx->wqh, wait); in eventfd_poll() 139 * All writes to ctx->count occur within ctx->wqh.lock. This read in eventfd_poll() 140 * can be done outside ctx->wqh in eventfd_poll() [all...] |
H A D | timerfd.c | 38 wait_queue_head_t wqh; member 67 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_triggered() 70 wake_up_locked_poll(&ctx->wqh, EPOLLIN); in timerfd_triggered() 71 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_triggered() 107 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 111 wake_up_locked_poll(&ctx->wqh, EPOLLIN); in timerfd_clock_was_set() 113 spin_unlock_irqrestore(&ctx->wqh.lock, flags); in timerfd_clock_was_set() 255 poll_wait(file, &ctx->wqh, wait); in timerfd_poll() 257 spin_lock_irqsave(&ctx->wqh.lock, flags); in timerfd_poll() 260 spin_unlock_irqrestore(&ctx->wqh in timerfd_poll() [all...] |
H A D | fs-writeback.c | 1489 wait_queue_head_t *wqh; variable 1491 wqh = bit_waitqueue(&inode->i_state, __I_SYNC); 1494 __wait_on_bit(wqh, &wq, bit_wait, 1519 wait_queue_head_t *wqh = bit_waitqueue(&inode->i_state, __I_SYNC); variable 1522 prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE); 1527 finish_wait(wqh, &wait);
|
/kernel/linux/linux-6.6/kernel/locking/ |
H A D | rtmutex.c | 520 static __always_inline void rt_mutex_wake_q_add_task(struct rt_wake_q_head *wqh, in rt_mutex_wake_q_add_task() argument 526 WARN_ON_ONCE(wqh->rtlock_task); in rt_mutex_wake_q_add_task() 528 wqh->rtlock_task = task; in rt_mutex_wake_q_add_task() 530 wake_q_add(&wqh->head, task); in rt_mutex_wake_q_add_task() 534 static __always_inline void rt_mutex_wake_q_add(struct rt_wake_q_head *wqh, in rt_mutex_wake_q_add() argument 537 rt_mutex_wake_q_add_task(wqh, w->task, w->wake_state); in rt_mutex_wake_q_add() 540 static __always_inline void rt_mutex_wake_up_q(struct rt_wake_q_head *wqh) in rt_mutex_wake_up_q() argument 542 if (IS_ENABLED(CONFIG_PREEMPT_RT) && wqh->rtlock_task) { in rt_mutex_wake_up_q() 543 wake_up_state(wqh->rtlock_task, TASK_RTLOCK_WAIT); in rt_mutex_wake_up_q() 544 put_task_struct(wqh in rt_mutex_wake_up_q() 1288 mark_wakeup_next_waiter(struct rt_wake_q_head *wqh, struct rt_mutex_base *lock) mark_wakeup_next_waiter() argument [all...] |
H A D | rtmutex_api.c | 163 * @wqh: The wake queue head from which to get the next lock waiter 166 struct rt_wake_q_head *wqh) in __rt_mutex_futex_unlock() 183 mark_wakeup_next_waiter(wqh, lock); in __rt_mutex_futex_unlock() 190 DEFINE_RT_WAKE_Q(wqh); in rt_mutex_futex_unlock() 195 postunlock = __rt_mutex_futex_unlock(lock, &wqh); in rt_mutex_futex_unlock() 199 rt_mutex_postunlock(&wqh); in rt_mutex_futex_unlock() 479 void __sched rt_mutex_postunlock(struct rt_wake_q_head *wqh) in rt_mutex_postunlock() argument 481 rt_mutex_wake_up_q(wqh); in rt_mutex_postunlock() 165 __rt_mutex_futex_unlock(struct rt_mutex_base *lock, struct rt_wake_q_head *wqh) __rt_mutex_futex_unlock() argument
|
H A D | rwbase_rt.c | 145 DEFINE_RT_WAKE_Q(wqh); in __rwbase_read_unlock() 156 rt_mutex_wake_q_add_task(&wqh, owner, state); in __rwbase_read_unlock() 161 rt_mutex_wake_up_q(&wqh); in __rwbase_read_unlock()
|
H A D | rtmutex_common.h | 101 struct rt_wake_q_head *wqh); 103 extern void rt_mutex_postunlock(struct rt_wake_q_head *wqh);
|
/kernel/linux/linux-6.6/arch/powerpc/platforms/pseries/ |
H A D | rtas-work-area.c | 44 struct wait_queue_head wqh; member 49 .wqh = __WAIT_QUEUE_HEAD_INITIALIZER(rwa_state.wqh), 103 wait_event(rwa_state.wqh, in __rtas_work_area_alloc() 123 wake_up(&rwa_state.wqh); in rtas_work_area_free()
|
/kernel/linux/linux-5.10/mm/ |
H A D | backing-dev.c | 936 wait_queue_head_t *wqh = &congestion_wqh[sync]; in clear_bdi_congested() local 943 if (waitqueue_active(wqh)) in clear_bdi_congested() 944 wake_up(wqh); in clear_bdi_congested() 972 wait_queue_head_t *wqh = &congestion_wqh[sync]; in congestion_wait() local 974 prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE); in congestion_wait() 976 finish_wait(wqh, &wait); in congestion_wait() 1003 wait_queue_head_t *wqh = &congestion_wqh[sync]; in wait_iff_congested() local 1021 prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE); in wait_iff_congested() 1023 finish_wait(wqh, &wait); in wait_iff_congested()
|
/kernel/linux/linux-5.10/drivers/vfio/ |
H A D | virqfd.c | 79 wait_queue_head_t *wqh, poll_table *pt) in virqfd_ptable_queue_proc() 82 add_wait_queue(wqh, &virqfd->wait); in virqfd_ptable_queue_proc() 78 virqfd_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, poll_table *pt) virqfd_ptable_queue_proc() argument
|
/kernel/linux/linux-6.6/drivers/vfio/ |
H A D | virqfd.c | 79 wait_queue_head_t *wqh, poll_table *pt) in virqfd_ptable_queue_proc() 82 add_wait_queue(wqh, &virqfd->wait); in virqfd_ptable_queue_proc() 78 virqfd_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, poll_table *pt) virqfd_ptable_queue_proc() argument
|
/kernel/linux/linux-6.6/drivers/virt/acrn/ |
H A D | irqfd.c | 75 /* Called with wqh->lock held and interrupts disabled */ 90 /* Do shutdown work in thread to hold wqh->lock */ in hsm_irqfd_wakeup() 96 static void hsm_irqfd_poll_func(struct file *file, wait_queue_head_t *wqh, in hsm_irqfd_poll_func() argument 102 add_wait_queue(wqh, &irqfd->wait); in hsm_irqfd_poll_func()
|
/kernel/linux/linux-6.6/kernel/ |
H A D | seccomp.c | 214 * @wqh: A wait queue for poll if a notifier is in use. 236 wait_queue_head_t wqh; member 534 if (waitqueue_active(&orig->wqh)) in __seccomp_filter_orphan() 535 wake_up_poll(&orig->wqh, EPOLLHUP); in __seccomp_filter_orphan() 689 init_waitqueue_head(&sfilter->wqh); in seccomp_prepare_filter() 1130 wake_up_poll_on_current_cpu(&match->wqh, EPOLLIN | EPOLLRDNORM); in seccomp_do_user_notification() 1132 wake_up_poll(&match->wqh, EPOLLIN | EPOLLRDNORM); in seccomp_do_user_notification() 1483 ret = prepare_to_wait_event(&filter->wqh, &wait, TASK_INTERRUPTIBLE); in recv_wait_event() 1493 finish_wait(&filter->wqh, &wait); in recv_wait_event() 1540 wake_up_poll(&filter->wqh, EPOLLOU in seccomp_notify_recv() [all...] |
/kernel/linux/linux-5.10/kernel/ |
H A D | seccomp.c | 167 * @wqh: A wait queue for poll if a notifier is in use. 187 wait_queue_head_t wqh; member 436 if (waitqueue_active(&orig->wqh)) in __seccomp_filter_orphan() 437 wake_up_poll(&orig->wqh, EPOLLHUP); in __seccomp_filter_orphan() 579 init_waitqueue_head(&sfilter->wqh); in seccomp_prepare_filter() 866 wake_up_poll(&match->wqh, EPOLLIN | EPOLLRDNORM); in seccomp_do_user_notification() 1227 wake_up_poll(&filter->wqh, EPOLLOUT | EPOLLWRNORM); in seccomp_notify_recv() 1451 poll_wait(file, &filter->wqh, poll_tab); in seccomp_notify_poll()
|
/kernel/linux/linux-6.6/kernel/futex/ |
H A D | pi.c | 618 DEFINE_RT_WAKE_Q(wqh); in wake_futex_pi() 671 postunlock = __rt_mutex_futex_unlock(&pi_state->pi_mutex, &wqh); in wake_futex_pi() 678 rt_mutex_postunlock(&wqh); in wake_futex_pi()
|
/kernel/linux/linux-5.10/kernel/locking/ |
H A D | rtmutex_common.h | 155 struct wake_q_head *wqh);
|
/kernel/linux/linux-6.6/mm/ |
H A D | internal.h | 226 wait_queue_head_t *wqh; in wake_throttle_isolated() local 228 wqh = &pgdat->reclaim_wait[VMSCAN_THROTTLE_ISOLATED]; in wake_throttle_isolated() 229 if (waitqueue_active(wqh)) in wake_throttle_isolated() 230 wake_up(wqh); in wake_throttle_isolated()
|
/kernel/linux/linux-5.10/drivers/vhost/ |
H A D | vhost.c | 154 static void vhost_poll_func(struct file *file, wait_queue_head_t *wqh, in vhost_poll_func() argument 160 poll->wqh = wqh; in vhost_poll_func() 161 add_wait_queue(wqh, &poll->wait); in vhost_poll_func() 196 poll->wqh = NULL; in vhost_poll_init() 208 if (poll->wqh) in vhost_poll_start() 227 if (poll->wqh) { in vhost_poll_stop() 228 remove_wait_queue(poll->wqh, &poll->wait); in vhost_poll_stop() 229 poll->wqh = NULL; in vhost_poll_stop()
|
H A D | vhost.h | 32 wait_queue_head_t *wqh; member
|
/kernel/linux/linux-6.6/drivers/vhost/ |
H A D | vhost.c | 154 static void vhost_poll_func(struct file *file, wait_queue_head_t *wqh, in vhost_poll_func() argument 160 poll->wqh = wqh; in vhost_poll_func() 161 add_wait_queue(wqh, &poll->wait); in vhost_poll_func() 197 poll->wqh = NULL; in vhost_poll_init() 210 if (poll->wqh) in vhost_poll_start() 229 if (poll->wqh) { in vhost_poll_stop() 230 remove_wait_queue(poll->wqh, &poll->wait); in vhost_poll_stop() 231 poll->wqh = NULL; in vhost_poll_stop()
|
H A D | vhost.h | 43 wait_queue_head_t *wqh; member
|
/kernel/linux/linux-5.10/virt/kvm/ |
H A D | eventfd.c | 182 * Called with wqh->lock held and interrupts disabled 222 * other side is required to acquire wqh->lock, which we hold in irqfd_wakeup() 234 irqfd_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh, in irqfd_ptable_queue_proc() argument 239 add_wait_queue(wqh, &irqfd->wait); in irqfd_ptable_queue_proc()
|