Lines Matching defs:events
184 __poll_t events;
217 * this is the underlying eventfd context to deliver events to.
439 * events from being lost.
969 * from aio_get_req() (the we're out of events case). It must be
1187 * events fetched
1214 * we also see the events that were stored up to the tail.
1323 * of available events. May fail with -ENOMEM if insufficient kernel
1689 struct poll_table_struct pt = { ._key = req->events };
1694 mask = vfs_poll(req->file, &pt) & req->events;
1757 if (mask && !(mask & req->events))
1765 * the events, so inline completion isn't possible.
1863 /* reject any unknown events outside the normal event mask. */
1871 req->events = demangle_poll(iocb->aio_buf) | EPOLLERR | EPOLLHUP;
1879 apt.pt._key = req->events;
1888 mask = vfs_poll(req->file, &apt.pt) & req->events;
2183 struct io_event __user *events,
2199 ret = read_events(ioctx, min_nr, nr, events, until);
2207 * Attempts to read at least min_nr events and up to nr events from
2209 * it succeeds, the number of read events is returned. May fail with
2214 * before sufficient events are available, where timeout == NULL
2223 struct io_event __user *, events,
2232 ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &ts : NULL);
2249 struct io_event __user *, events,
2268 ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &ts : NULL);
2284 struct io_event __user *, events,
2304 ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &ts : NULL);
2321 struct io_event __user *, events,
2330 ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL);
2351 struct io_event __user *, events,
2370 ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL);
2386 struct io_event __user *, events,
2405 ret = do_io_getevents(ctx_id, min_nr, nr, events, timeout ? &t : NULL);