| /* |
| * libwebsockets - small server side websockets and web server implementation |
| * |
| * Copyright (C) 2010 - 2019 Andy Green <[email protected]> |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a copy |
| * of this software and associated documentation files (the "Software"), to |
| * deal in the Software without restriction, including without limitation the |
| * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or |
| * sell copies of the Software, and to permit persons to whom the Software is |
| * furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
| * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
| * IN THE SOFTWARE. |
| */ |
| |
| #if !defined(_GNU_SOURCE) |
| #define _GNU_SOURCE |
| #endif |
| #include "private-lib-core.h" |
| |
| int |
| lws_poll_listen_fd(struct lws_pollfd *fd) |
| { |
| return poll(fd, 1, 0); |
| } |
| |
| int |
| _lws_plat_service_forced_tsi(struct lws_context *context, int tsi) |
| { |
| struct lws_context_per_thread *pt = &context->pt[tsi]; |
| int m, n, r; |
| |
| r = lws_service_flag_pending(context, tsi); |
| |
| /* any socket with events to service? */ |
| for (n = 0; n < (int)pt->fds_count; n++) { |
| lws_sockfd_type fd = pt->fds[n].fd; |
| |
| if (!pt->fds[n].revents) |
| continue; |
| |
| m = lws_service_fd_tsi(context, &pt->fds[n], tsi); |
| if (m < 0) { |
| lwsl_err("%s: lws_service_fd_tsi returned %d\n", |
| __func__, m); |
| return -1; |
| } |
| |
| /* if something closed, retry this slot since may have been |
| * swapped with end fd */ |
| if (m && pt->fds[n].fd != fd) |
| n--; |
| } |
| |
| lws_service_do_ripe_rxflow(pt); |
| |
| return r; |
| } |
| |
| #define LWS_POLL_WAIT_LIMIT 2000000000 |
| |
| int |
| _lws_plat_service_tsi(struct lws_context *context, int timeout_ms, int tsi) |
| { |
| volatile struct lws_foreign_thread_pollfd *ftp, *next; |
| volatile struct lws_context_per_thread *vpt; |
| struct lws_context_per_thread *pt; |
| lws_usec_t timeout_us, us; |
| #if defined(LWS_WITH_SYS_METRICS) |
| lws_usec_t a, b; |
| #endif |
| int n; |
| #if (defined(LWS_ROLE_WS) && !defined(LWS_WITHOUT_EXTENSIONS)) || defined(LWS_WITH_TLS) |
| int m; |
| #endif |
| |
| /* stay dead once we are dead */ |
| |
| if (!context) |
| return 1; |
| |
| #if defined(LWS_WITH_SYS_METRICS) |
| b = |
| #endif |
| us = lws_now_usecs(); |
| |
| pt = &context->pt[tsi]; |
| vpt = (volatile struct lws_context_per_thread *)pt; |
| |
| if (timeout_ms < 0) |
| timeout_ms = 0; |
| else |
| /* force a default timeout of 23 days */ |
| timeout_ms = LWS_POLL_WAIT_LIMIT; |
| timeout_us = ((lws_usec_t)timeout_ms) * LWS_US_PER_MS; |
| |
| if (context->event_loop_ops->run_pt) |
| context->event_loop_ops->run_pt(context, tsi); |
| |
| if (!pt->service_tid_detected && context->vhost_list) { |
| lws_fakewsi_def_plwsa(pt); |
| |
| lws_fakewsi_prep_plwsa_ctx(context); |
| |
| pt->service_tid = context->vhost_list->protocols[0].callback( |
| (struct lws *)plwsa, |
| LWS_CALLBACK_GET_THREAD_ID, |
| NULL, NULL, 0); |
| pt->service_tid_detected = 1; |
| } |
| |
| lws_pt_lock(pt, __func__); |
| /* |
| * service ripe scheduled events, and limit wait to next expected one |
| */ |
| us = __lws_sul_service_ripe(pt->pt_sul_owner, LWS_COUNT_PT_SUL_OWNERS, us); |
| if (us && us < timeout_us) |
| /* |
| * If something wants zero wait, that's OK, but if the next sul |
| * coming ripe is an interval less than our wait resolution, |
| * bump it to be the wait resolution. |
| */ |
| timeout_us = us < context->us_wait_resolution ? |
| context->us_wait_resolution : us; |
| |
| lws_pt_unlock(pt); |
| |
| /* |
| * is there anybody with pending stuff that needs service forcing? |
| */ |
| if (!lws_service_adjust_timeout(context, 1, tsi)) |
| timeout_us = 0; |
| |
| /* ensure we don't wrap at 2^31 with poll()'s signed int ms */ |
| |
| timeout_us /= LWS_US_PER_MS; /* ms now */ |
| |
| #if defined(LWS_WITH_SYS_METRICS) |
| a = lws_now_usecs() - b; |
| #endif |
| vpt->inside_poll = 1; |
| lws_memory_barrier(); |
| n = poll(pt->fds, pt->fds_count, (int)timeout_us /* ms now */ ); |
| vpt->inside_poll = 0; |
| lws_memory_barrier(); |
| |
| #if defined(LWS_WITH_SYS_METRICS) |
| b = lws_now_usecs(); |
| #endif |
| /* Collision will be rare and brief. Spin until it completes */ |
| while (vpt->foreign_spinlock) |
| ; |
| |
| /* |
| * At this point we are not inside a foreign thread pollfd |
| * change, and we have marked ourselves as outside the poll() |
| * wait. So we are the only guys that can modify the |
| * lws_foreign_thread_pollfd list on the pt. Drain the list |
| * and apply the changes to the affected pollfds in the correct |
| * order. |
| */ |
| |
| lws_pt_lock(pt, __func__); |
| |
| ftp = vpt->foreign_pfd_list; |
| //lwsl_notice("cleared list %p\n", ftp); |
| while (ftp) { |
| struct lws *wsi; |
| struct lws_pollfd *pfd; |
| |
| next = ftp->next; |
| pfd = &vpt->fds[ftp->fd_index]; |
| if (lws_socket_is_valid(pfd->fd)) { |
| wsi = wsi_from_fd(context, pfd->fd); |
| if (wsi) |
| __lws_change_pollfd(wsi, ftp->_and, |
| ftp->_or); |
| } |
| lws_free((void *)ftp); |
| ftp = next; |
| } |
| vpt->foreign_pfd_list = NULL; |
| lws_memory_barrier(); |
| |
| lws_pt_unlock(pt); |
| |
| #if (defined(LWS_ROLE_WS) && !defined(LWS_WITHOUT_EXTENSIONS)) || defined(LWS_WITH_TLS) |
| m = 0; |
| #endif |
| #if defined(LWS_ROLE_WS) && !defined(LWS_WITHOUT_EXTENSIONS) |
| m |= !!pt->ws.rx_draining_ext_list; |
| #endif |
| |
| #if defined(LWS_WITH_TLS) |
| if (pt->context->tls_ops && |
| pt->context->tls_ops->fake_POLLIN_for_buffered) |
| m |= pt->context->tls_ops->fake_POLLIN_for_buffered(pt); |
| #endif |
| |
| if ( |
| #if (defined(LWS_ROLE_WS) && !defined(LWS_WITHOUT_EXTENSIONS)) || defined(LWS_WITH_TLS) |
| !m && |
| #endif |
| !n) /* nothing to do */ |
| lws_service_do_ripe_rxflow(pt); |
| else |
| if (_lws_plat_service_forced_tsi(context, tsi) < 0) |
| return -1; |
| |
| #if defined(LWS_WITH_SYS_METRICS) |
| lws_metric_event(context->mt_service, METRES_GO, |
| (u_mt_t) (a + (lws_now_usecs() - b))); |
| #endif |
| |
| if (pt->destroy_self) { |
| lws_context_destroy(pt->context); |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| int |
| lws_plat_service(struct lws_context *context, int timeout_ms) |
| { |
| return _lws_plat_service_tsi(context, timeout_ms, 0); |
| } |