xref: /third_party/libuv/src/unix/posix-poll.c (revision e66f31c5)
1/* Copyright libuv project contributors. All rights reserved.
2 *
3 * Permission is hereby granted, free of charge, to any person obtaining a copy
4 * of this software and associated documentation files (the "Software"), to
5 * deal in the Software without restriction, including without limitation the
6 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
7 * sell copies of the Software, and to permit persons to whom the Software is
8 * furnished to do so, subject to the following conditions:
9 *
10 * The above copyright notice and this permission notice shall be included in
11 * all copies or substantial portions of the Software.
12 *
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
18 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
19 * IN THE SOFTWARE.
20 */
21
22#include "uv.h"
23#include "internal.h"
24
25/* POSIX defines poll() as a portable way to wait on file descriptors.
26 * Here we maintain a dynamically sized array of file descriptors and
27 * events to pass as the first argument to poll().
28 */
29
30#include <assert.h>
31#include <stddef.h>
32#include <stdint.h>
33#include <errno.h>
34#include <unistd.h>
35
36int uv__platform_loop_init(uv_loop_t* loop) {
37  loop->poll_fds = NULL;
38  loop->poll_fds_used = 0;
39  loop->poll_fds_size = 0;
40  loop->poll_fds_iterating = 0;
41  return 0;
42}
43
44void uv__platform_loop_delete(uv_loop_t* loop) {
45  uv__free(loop->poll_fds);
46  loop->poll_fds = NULL;
47}
48
49int uv__io_fork(uv_loop_t* loop) {
50  uv__platform_loop_delete(loop);
51  return uv__platform_loop_init(loop);
52}
53
54/* Allocate or dynamically resize our poll fds array.  */
55static void uv__pollfds_maybe_resize(uv_loop_t* loop) {
56  size_t i;
57  size_t n;
58  struct pollfd* p;
59
60  if (loop->poll_fds_used < loop->poll_fds_size)
61    return;
62
63  n = loop->poll_fds_size ? loop->poll_fds_size * 2 : 64;
64  p = uv__reallocf(loop->poll_fds, n * sizeof(*loop->poll_fds));
65  if (p == NULL)
66    abort();
67
68  loop->poll_fds = p;
69  for (i = loop->poll_fds_size; i < n; i++) {
70    loop->poll_fds[i].fd = -1;
71    loop->poll_fds[i].events = 0;
72    loop->poll_fds[i].revents = 0;
73  }
74  loop->poll_fds_size = n;
75}
76
77/* Primitive swap operation on poll fds array elements.  */
78static void uv__pollfds_swap(uv_loop_t* loop, size_t l, size_t r) {
79  struct pollfd pfd;
80  pfd = loop->poll_fds[l];
81  loop->poll_fds[l] = loop->poll_fds[r];
82  loop->poll_fds[r] = pfd;
83}
84
85/* Add a watcher's fd to our poll fds array with its pending events.  */
86static void uv__pollfds_add(uv_loop_t* loop, uv__io_t* w) {
87  size_t i;
88  struct pollfd* pe;
89
90  /* If the fd is already in the set just update its events.  */
91  assert(!loop->poll_fds_iterating);
92  for (i = 0; i < loop->poll_fds_used; ++i) {
93    if (loop->poll_fds[i].fd == w->fd) {
94      loop->poll_fds[i].events = w->pevents;
95      return;
96    }
97  }
98
99  /* Otherwise, allocate a new slot in the set for the fd.  */
100  uv__pollfds_maybe_resize(loop);
101  pe = &loop->poll_fds[loop->poll_fds_used++];
102  pe->fd = w->fd;
103  pe->events = w->pevents;
104}
105
106/* Remove a watcher's fd from our poll fds array.  */
107static void uv__pollfds_del(uv_loop_t* loop, int fd) {
108  size_t i;
109  assert(!loop->poll_fds_iterating);
110  for (i = 0; i < loop->poll_fds_used;) {
111    if (loop->poll_fds[i].fd == fd) {
112      /* swap to last position and remove */
113      --loop->poll_fds_used;
114      uv__pollfds_swap(loop, i, loop->poll_fds_used);
115      loop->poll_fds[loop->poll_fds_used].fd = -1;
116      loop->poll_fds[loop->poll_fds_used].events = 0;
117      loop->poll_fds[loop->poll_fds_used].revents = 0;
118      /* This method is called with an fd of -1 to purge the invalidated fds,
119       * so we may possibly have multiples to remove.
120       */
121      if (-1 != fd)
122        return;
123    } else {
124      /* We must only increment the loop counter when the fds do not match.
125       * Otherwise, when we are purging an invalidated fd, the value just
126       * swapped here from the previous end of the array will be skipped.
127       */
128       ++i;
129    }
130  }
131}
132
133
134void uv__io_poll(uv_loop_t* loop, int timeout) {
135  uv__loop_internal_fields_t* lfields;
136  sigset_t* pset;
137  sigset_t set;
138  uint64_t time_base;
139  uint64_t time_diff;
140  struct uv__queue* q;
141  uv__io_t* w;
142  size_t i;
143  unsigned int nevents;
144  int nfds;
145  int have_signals;
146  struct pollfd* pe;
147  int fd;
148  int user_timeout;
149  int reset_timeout;
150
151  if (loop->nfds == 0) {
152    assert(uv__queue_empty(&loop->watcher_queue));
153    return;
154  }
155
156  lfields = uv__get_internal_fields(loop);
157
158  /* Take queued watchers and add their fds to our poll fds array.  */
159  while (!uv__queue_empty(&loop->watcher_queue)) {
160    q = uv__queue_head(&loop->watcher_queue);
161    uv__queue_remove(q);
162    uv__queue_init(q);
163
164    w = uv__queue_data(q, uv__io_t, watcher_queue);
165    assert(w->pevents != 0);
166    assert(w->fd >= 0);
167    assert(w->fd < (int) loop->nwatchers);
168
169    uv__pollfds_add(loop, w);
170
171    w->events = w->pevents;
172  }
173
174  /* Prepare a set of signals to block around poll(), if any.  */
175  pset = NULL;
176  if (loop->flags & UV_LOOP_BLOCK_SIGPROF) {
177    pset = &set;
178    sigemptyset(pset);
179    sigaddset(pset, SIGPROF);
180  }
181
182  assert(timeout >= -1);
183  time_base = loop->time;
184
185  if (lfields->flags & UV_METRICS_IDLE_TIME) {
186    reset_timeout = 1;
187    user_timeout = timeout;
188    timeout = 0;
189  } else {
190    reset_timeout = 0;
191  }
192
193  /* Loop calls to poll() and processing of results.  If we get some
194   * results from poll() but they turn out not to be interesting to
195   * our caller then we need to loop around and poll() again.
196   */
197  for (;;) {
198    /* Only need to set the provider_entry_time if timeout != 0. The function
199     * will return early if the loop isn't configured with UV_METRICS_IDLE_TIME.
200     */
201    if (timeout != 0)
202      uv__metrics_set_provider_entry_time(loop);
203
204    /* Store the current timeout in a location that's globally accessible so
205     * other locations like uv__work_done() can determine whether the queue
206     * of events in the callback were waiting when poll was called.
207     */
208    lfields->current_timeout = timeout;
209
210    if (pset != NULL)
211      if (pthread_sigmask(SIG_BLOCK, pset, NULL))
212        abort();
213    nfds = poll(loop->poll_fds, (nfds_t)loop->poll_fds_used, timeout);
214    if (pset != NULL)
215      if (pthread_sigmask(SIG_UNBLOCK, pset, NULL))
216        abort();
217
218    /* Update loop->time unconditionally. It's tempting to skip the update when
219     * timeout == 0 (i.e. non-blocking poll) but there is no guarantee that the
220     * operating system didn't reschedule our process while in the syscall.
221     */
222    SAVE_ERRNO(uv__update_time(loop));
223
224    if (nfds == 0) {
225      if (reset_timeout != 0) {
226        timeout = user_timeout;
227        reset_timeout = 0;
228        if (timeout == -1)
229          continue;
230        if (timeout > 0)
231          goto update_timeout;
232      }
233
234      assert(timeout != -1);
235      return;
236    }
237
238    if (nfds == -1) {
239      if (errno != EINTR)
240        abort();
241
242      if (reset_timeout != 0) {
243        timeout = user_timeout;
244        reset_timeout = 0;
245      }
246
247      if (timeout == -1)
248        continue;
249
250      if (timeout == 0)
251        return;
252
253      /* Interrupted by a signal. Update timeout and poll again. */
254      goto update_timeout;
255    }
256
257    /* Tell uv__platform_invalidate_fd not to manipulate our array
258     * while we are iterating over it.
259     */
260    loop->poll_fds_iterating = 1;
261
262    /* Initialize a count of events that we care about.  */
263    nevents = 0;
264    have_signals = 0;
265
266    /* Loop over the entire poll fds array looking for returned events.  */
267    for (i = 0; i < loop->poll_fds_used; i++) {
268      pe = loop->poll_fds + i;
269      fd = pe->fd;
270
271      /* Skip invalidated events, see uv__platform_invalidate_fd.  */
272      if (fd == -1)
273        continue;
274
275      assert(fd >= 0);
276      assert((unsigned) fd < loop->nwatchers);
277
278      w = loop->watchers[fd];
279
280      if (w == NULL) {
281        /* File descriptor that we've stopped watching, ignore.  */
282        uv__platform_invalidate_fd(loop, fd);
283        continue;
284      }
285
286      /* Filter out events that user has not requested us to watch
287       * (e.g. POLLNVAL).
288       */
289      pe->revents &= w->pevents | POLLERR | POLLHUP;
290
291      if (pe->revents != 0) {
292        /* Run signal watchers last.  */
293        if (w == &loop->signal_io_watcher) {
294          have_signals = 1;
295        } else {
296          uv__metrics_update_idle_time(loop);
297          w->cb(loop, w, pe->revents);
298        }
299
300        nevents++;
301      }
302    }
303
304    uv__metrics_inc_events(loop, nevents);
305    if (reset_timeout != 0) {
306      timeout = user_timeout;
307      reset_timeout = 0;
308      uv__metrics_inc_events_waiting(loop, nevents);
309    }
310
311    if (have_signals != 0) {
312      uv__metrics_update_idle_time(loop);
313      loop->signal_io_watcher.cb(loop, &loop->signal_io_watcher, POLLIN);
314    }
315
316    loop->poll_fds_iterating = 0;
317
318    /* Purge invalidated fds from our poll fds array.  */
319    uv__pollfds_del(loop, -1);
320
321    if (have_signals != 0)
322      return;  /* Event loop should cycle now so don't poll again. */
323
324    if (nevents != 0)
325      return;
326
327    if (timeout == 0)
328      return;
329
330    if (timeout == -1)
331      continue;
332
333update_timeout:
334    assert(timeout > 0);
335
336    time_diff = loop->time - time_base;
337    if (time_diff >= (uint64_t) timeout)
338      return;
339
340    timeout -= time_diff;
341  }
342}
343
344/* Remove the given fd from our poll fds array because no one
345 * is interested in its events anymore.
346 */
347void uv__platform_invalidate_fd(uv_loop_t* loop, int fd) {
348  size_t i;
349
350  assert(fd >= 0);
351
352  if (loop->poll_fds_iterating) {
353    /* uv__io_poll is currently iterating.  Just invalidate fd.  */
354    for (i = 0; i < loop->poll_fds_used; i++)
355      if (loop->poll_fds[i].fd == fd) {
356        loop->poll_fds[i].fd = -1;
357        loop->poll_fds[i].events = 0;
358        loop->poll_fds[i].revents = 0;
359      }
360  } else {
361    /* uv__io_poll is not iterating.  Delete fd from the set.  */
362    uv__pollfds_del(loop, fd);
363  }
364}
365
366/* Check whether the given fd is supported by poll().  */
367int uv__io_check_fd(uv_loop_t* loop, int fd) {
368  struct pollfd p[1];
369  int rv;
370
371  p[0].fd = fd;
372  p[0].events = POLLIN;
373
374  do
375    rv = poll(p, 1, 0);
376  while (rv == -1 && (errno == EINTR || errno == EAGAIN));
377
378  if (rv == -1)
379    return UV__ERR(errno);
380
381  if (p[0].revents & POLLNVAL)
382    return UV_EINVAL;
383
384  return 0;
385}
386