Lines Matching refs:pt
39 struct lws_context_per_thread *pt = &context->pt[tsi];
45 for (n = 0; n < (int)pt->fds_count; n++) {
46 lws_sockfd_type fd = pt->fds[n].fd;
48 if (!pt->fds[n].revents)
51 m = lws_service_fd_tsi(context, &pt->fds[n], tsi);
60 if (m && pt->fds[n].fd != fd)
64 lws_service_do_ripe_rxflow(pt);
76 struct lws_context_per_thread *pt;
96 pt = &context->pt[tsi];
97 vpt = (volatile struct lws_context_per_thread *)pt;
109 if (!pt->service_tid_detected && context->vhost_list) {
110 lws_fakewsi_def_plwsa(pt);
114 pt->service_tid = context->vhost_list->protocols[0].callback(
118 pt->service_tid_detected = 1;
121 lws_pt_lock(pt, __func__);
125 us = __lws_sul_service_ripe(pt->pt_sul_owner, LWS_COUNT_PT_SUL_OWNERS, us);
135 lws_pt_unlock(pt);
152 n = poll(pt->fds, pt->fds_count, (int)timeout_us /* ms now */ );
167 * lws_foreign_thread_pollfd list on the pt. Drain the list
172 lws_pt_lock(pt, __func__);
194 lws_pt_unlock(pt);
200 m |= !!pt->ws.rx_draining_ext_list;
204 if (pt->context->tls_ops &&
205 pt->context->tls_ops->fake_POLLIN_for_buffered)
206 m |= pt->context->tls_ops->fake_POLLIN_for_buffered(pt);
214 lws_service_do_ripe_rxflow(pt);
224 if (pt->destroy_self) {
225 lws_context_destroy(pt->context);