/* * libwebsockets - small server side websockets and web server implementation * * Copyright (C) 2010 - 2021 Andy Green * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to * deal in the Software without restriction, including without limitation the * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or * sell copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include "private-lib-core.h" void lws_tls_session_vh_destroy(struct lws_vhost *vh); const struct lws_role_ops *available_roles[] = { #if defined(LWS_ROLE_H2) &role_ops_h2, #endif #if defined(LWS_ROLE_H1) &role_ops_h1, #endif #if defined(LWS_ROLE_WS) &role_ops_ws, #endif #if defined(LWS_ROLE_DBUS) &role_ops_dbus, #endif #if defined(LWS_ROLE_RAW_PROXY) &role_ops_raw_proxy, #endif #if defined(LWS_ROLE_MQTT) && defined(LWS_WITH_CLIENT) &role_ops_mqtt, #endif #if defined(LWS_WITH_NETLINK) &role_ops_netlink, #endif NULL }; #if defined(LWS_WITH_ABSTRACT) const struct lws_protocols *available_abstract_protocols[] = { #if defined(LWS_ROLE_RAW) &protocol_abs_client_raw_skt, #endif NULL }; #endif #if defined(LWS_WITH_SECURE_STREAMS) const struct lws_protocols *available_secstream_protocols[] = { #if defined(LWS_ROLE_H1) &protocol_secstream_h1, #endif #if defined(LWS_ROLE_H2) &protocol_secstream_h2, #endif #if defined(LWS_ROLE_WS) &protocol_secstream_ws, #endif #if defined(LWS_ROLE_MQTT) &protocol_secstream_mqtt, #endif &protocol_secstream_raw, NULL }; #endif static const char * const mount_protocols[] = { "http://", "https://", "file://", "cgi://", ">http://", ">https://", "callback://" }; const struct lws_role_ops * lws_role_by_name(const char *name) { LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) if (!strcmp(ar->name, name)) return ar; LWS_FOR_EVERY_AVAILABLE_ROLE_END; if (!strcmp(name, role_ops_raw_skt.name)) return &role_ops_raw_skt; #if defined(LWS_ROLE_RAW_FILE) if (!strcmp(name, role_ops_raw_file.name)) return &role_ops_raw_file; #endif return NULL; } int lws_role_call_alpn_negotiated(struct lws *wsi, const char *alpn) { #if defined(LWS_WITH_TLS) if (!alpn) return 0; #if !defined(LWS_ESP_PLATFORM) lwsl_info("%s: '%s'\n", __func__, alpn); #endif LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) if (ar->alpn && !strcmp(ar->alpn, alpn) && lws_rops_fidx(ar, LWS_ROPS_alpn_negotiated)) { #if defined(LWS_WITH_SERVER) lws_metrics_tag_wsi_add(wsi, "upg", ar->name); #endif return (lws_rops_func_fidx(ar, LWS_ROPS_alpn_negotiated)). alpn_negotiated(wsi, alpn); } LWS_FOR_EVERY_AVAILABLE_ROLE_END; #endif return 0; } int lws_role_call_adoption_bind(struct lws *wsi, int type, const char *prot) { int n; /* * if the vhost is told to bind accepted sockets to a given role, * then look it up by name and try to bind to the specific role. */ if (lws_check_opt(wsi->a.vhost->options, LWS_SERVER_OPTION_ADOPT_APPLY_LISTEN_ACCEPT_CONFIG) && wsi->a.vhost->listen_accept_role) { const struct lws_role_ops *role = lws_role_by_name(wsi->a.vhost->listen_accept_role); if (!prot) prot = wsi->a.vhost->listen_accept_protocol; if (!role) lwsl_err("%s: can't find role '%s'\n", __func__, wsi->a.vhost->listen_accept_role); if (!strcmp(wsi->a.vhost->listen_accept_role, "raw-proxy")) type |= LWS_ADOPT_FLAG_RAW_PROXY; if (role && lws_rops_fidx(role, LWS_ROPS_adoption_bind)) { n = (lws_rops_func_fidx(role, LWS_ROPS_adoption_bind)). adoption_bind(wsi, type, prot); if (n < 0) return -1; if (n) /* did the bind */ return 0; } if (type & _LWS_ADOPT_FINISH) { lwsl_debug("%s: leaving bound to role %s\n", __func__, wsi->role_ops->name); return 0; } lwsl_warn("%s: adoption bind to role '%s', " "protocol '%s', type 0x%x, failed\n", __func__, wsi->a.vhost->listen_accept_role, prot, type); } /* * Otherwise ask each of the roles in order of preference if they * want to bind to this accepted socket */ LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) if (lws_rops_fidx(ar, LWS_ROPS_adoption_bind) && (lws_rops_func_fidx(ar, LWS_ROPS_adoption_bind)). adoption_bind(wsi, type, prot)) return 0; LWS_FOR_EVERY_AVAILABLE_ROLE_END; /* fall back to raw socket role if, eg, h1 not configured */ if (lws_rops_fidx(&role_ops_raw_skt, LWS_ROPS_adoption_bind) && (lws_rops_func_fidx(&role_ops_raw_skt, LWS_ROPS_adoption_bind)). adoption_bind(wsi, type, prot)) return 0; #if defined(LWS_ROLE_RAW_FILE) lwsl_notice("%s: falling back to raw file role bind\n", __func__); /* fall back to raw file role if, eg, h1 not configured */ if (lws_rops_fidx(&role_ops_raw_file, LWS_ROPS_adoption_bind) && (lws_rops_func_fidx(&role_ops_raw_file, LWS_ROPS_adoption_bind)). adoption_bind(wsi, type, prot)) return 0; #endif return 1; } #if defined(LWS_WITH_CLIENT) int lws_role_call_client_bind(struct lws *wsi, const struct lws_client_connect_info *i) { LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) if (lws_rops_fidx(ar, LWS_ROPS_client_bind)) { int m = (lws_rops_func_fidx(ar, LWS_ROPS_client_bind)). client_bind(wsi, i); if (m < 0) return m; if (m) return 0; } LWS_FOR_EVERY_AVAILABLE_ROLE_END; /* fall back to raw socket role if, eg, h1 not configured */ if (lws_rops_fidx(&role_ops_raw_skt, LWS_ROPS_client_bind) && (lws_rops_func_fidx(&role_ops_raw_skt, LWS_ROPS_client_bind)). client_bind(wsi, i)) return 0; return 1; } #endif void * lws_protocol_vh_priv_zalloc(struct lws_vhost *vhost, const struct lws_protocols *prot, int size) { int n = 0; if (!vhost || !prot) return NULL; /* allocate the vh priv array only on demand */ if (!vhost->protocol_vh_privs) { vhost->protocol_vh_privs = (void **)lws_zalloc( (size_t)vhost->count_protocols * sizeof(void *), "protocol_vh_privs"); if (!vhost->protocol_vh_privs) return NULL; } while (n < vhost->count_protocols && &vhost->protocols[n] != prot) n++; if (n == vhost->count_protocols) { n = 0; while (n < vhost->count_protocols && strcmp(vhost->protocols[n].name, prot->name)) n++; if (n == vhost->count_protocols) return NULL; } vhost->protocol_vh_privs[n] = lws_zalloc((size_t)size, "vh priv"); return vhost->protocol_vh_privs[n]; } void * lws_protocol_vh_priv_get(struct lws_vhost *vhost, const struct lws_protocols *prot) { int n = 0; if (!vhost || !vhost->protocol_vh_privs || !prot) return NULL; while (n < vhost->count_protocols && &vhost->protocols[n] != prot) n++; if (n == vhost->count_protocols) { n = 0; while (n < vhost->count_protocols && strcmp(vhost->protocols[n].name, prot->name)) n++; if (n == vhost->count_protocols) { lwsl_err("%s: unknown protocol %p\n", __func__, prot); return NULL; } } return vhost->protocol_vh_privs[n]; } void * lws_vhd_find_by_pvo(struct lws_context *cx, const char *protname, const char *pvo_name, const char *pvo_value) { struct lws_vhost *vh; int n; /* let's go through all the vhosts */ vh = cx->vhost_list; while (vh) { if (vh->protocol_vh_privs) { for (n = 0; n < vh->count_protocols; n++) { const struct lws_protocol_vhost_options *pv; if (strcmp(vh->protocols[n].name, protname)) continue; /* this vh has an instance of the required protocol */ pv = lws_pvo_search(vh->pvo, protname); if (!pv) continue; pv = lws_pvo_search(pv->options, pvo_name); if (!pv) continue; /* ... he also has a pvo of the right name... */ if (!strcmp(pv->value, pvo_value)) /* * ... yes, the pvo has the right value too, * return a pointer to this vhost-protocol * private alloc (ie, its "vhd") */ return vh->protocol_vh_privs[n]; } } else lwsl_notice("%s: no privs yet on %s\n", __func__, lws_vh_tag(vh)); vh = vh->vhost_next; } return NULL; } const struct lws_protocol_vhost_options * lws_vhost_protocol_options(struct lws_vhost *vh, const char *name) { const struct lws_protocol_vhost_options *pvo = vh->pvo; if (!name) return NULL; while (pvo) { if (!strcmp(pvo->name, name)) return pvo; pvo = pvo->next; } return NULL; } int lws_protocol_init_vhost(struct lws_vhost *vh, int *any) { const struct lws_protocol_vhost_options *pvo, *pvo1; lws_fakewsi_def_plwsa(&vh->context->pt[0]); int n; lws_fakewsi_prep_plwsa_ctx(vh->context); plwsa->vhost = vh; /* initialize supported protocols on this vhost */ for (n = 0; n < vh->count_protocols; n++) { plwsa->protocol = &vh->protocols[n]; if (!vh->protocols[n].name) continue; pvo = lws_vhost_protocol_options(vh, vh->protocols[n].name); if (pvo) { /* * linked list of options specific to * vh + protocol */ pvo1 = pvo; pvo = pvo1->options; while (pvo) { lwsl_debug( " vhost \"%s\", " "protocol \"%s\", " "option \"%s\"\n", vh->name, vh->protocols[n].name, pvo->name); if (!strcmp(pvo->name, "default")) { lwsl_info("Setting default " "protocol for vh %s to %s\n", vh->name, vh->protocols[n].name); vh->default_protocol_index = (unsigned char)n; } if (!strcmp(pvo->name, "raw")) { lwsl_info("Setting raw " "protocol for vh %s to %s\n", vh->name, vh->protocols[n].name); vh->raw_protocol_index = (unsigned char)n; } pvo = pvo->next; } } else lwsl_debug("%s: not instantiating %s.%s\n", __func__, vh->name, vh->protocols[n].name); #if defined(LWS_WITH_TLS) if (any) *any |= !!vh->tls.ssl_ctx; #endif plwsa->vhost = vh; plwsa->protocol = &vh->protocols[n]; pvo = lws_vhost_protocol_options(vh, vh->protocols[n].name); /* * inform all the protocols that they are doing their * one-time initialization if they want to. * * NOTE the fakewsi is garbage, except the key pointers that are * prepared in case the protocol handler wants to touch them */ if (pvo || !vh->pvo) { lwsl_info("%s: init %s.%s\n", __func__, vh->name, vh->protocols[n].name); if (vh->protocols[n].callback((struct lws *)plwsa, LWS_CALLBACK_PROTOCOL_INIT, NULL, (void *)(pvo ? pvo->options : NULL), 0)) { if (vh->protocol_vh_privs && vh->protocol_vh_privs[n]) { lws_free(vh->protocol_vh_privs[n]); vh->protocol_vh_privs[n] = NULL; } lwsl_err("%s: protocol %s failed init\n", __func__, vh->protocols[n].name); return 1; } } } vh->created_vhost_protocols = 1; return 0; } /* * inform every vhost that hasn't already done it, that * his protocols are initializing */ int lws_protocol_init(struct lws_context *context) { struct lws_vhost *vh = context->vhost_list; int any = 0, r = 0; if (context->doing_protocol_init) return 0; context->doing_protocol_init = 1; lwsl_info("%s\n", __func__); while (vh) { /* only do the protocol init once for a given vhost */ if (vh->created_vhost_protocols || (lws_check_opt(vh->options, LWS_SERVER_OPTION_SKIP_PROTOCOL_INIT))) goto next; if (lws_protocol_init_vhost(vh, &any)) { lwsl_warn("%s: init vhost %s failed\n", __func__, vh->name); r = -1; } next: vh = vh->vhost_next; } context->doing_protocol_init = 0; if (r) lwsl_warn("%s: some protocols did not init\n", __func__); if (!context->protocol_init_done) { context->protocol_init_done = 1; lws_finalize_startup(context); return 0; } #if defined(LWS_WITH_SERVER) if (any) { lws_tls_check_all_cert_lifetimes(context); } #endif return 0; } /* list of supported protocols and callbacks */ static const struct lws_protocols protocols_dummy[] = { /* first protocol must always be HTTP handler */ { "http-only", /* name */ lws_callback_http_dummy, /* callback */ 0, /* per_session_data_size */ 0, /* rx_buffer_size */ 0, /* id */ NULL, /* user */ 0 /* tx_packet_size */ }, /* * the other protocols are provided by lws plugins */ { NULL, NULL, 0, 0, 0, NULL, 0} /* terminator */ }; #ifdef LWS_PLAT_OPTEE #undef LWS_HAVE_GETENV #endif struct lws_vhost * lws_create_vhost(struct lws_context *context, const struct lws_context_creation_info *info) { struct lws_vhost *vh, **vh1 = &context->vhost_list; const struct lws_http_mount *mounts; const struct lws_protocols *pcols = info->protocols; #ifdef LWS_WITH_PLUGINS struct lws_plugin *plugin = context->plugin_list; #endif struct lws_protocols *lwsp; int m, f = !info->pvo, fx = 0, abs_pcol_count = 0, sec_pcol_count = 0; char buf[96]; char *p; #if defined(LWS_WITH_SYS_ASYNC_DNS) extern struct lws_protocols lws_async_dns_protocol; #endif int n; if (lws_fi(&info->fic, "vh_create_oom")) vh = NULL; else vh = lws_zalloc(sizeof(*vh) #if defined(LWS_WITH_EVENT_LIBS) + context->event_loop_ops->evlib_size_vh #endif , __func__); if (!vh) goto early_bail; #if defined(LWS_WITH_EVENT_LIBS) vh->evlib_vh = (void *)&vh[1]; #endif #if LWS_MAX_SMP > 1 lws_mutex_refcount_init(&vh->mr); #endif if (!pcols && !info->pprotocols) pcols = &protocols_dummy[0]; vh->context = context; if (!info->vhost_name) vh->name = "default"; else vh->name = info->vhost_name; { char *end = buf + sizeof(buf) - 1; p = buf; p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), "%s", vh->name); if (info->iface) p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), "|%s", info->iface); if (info->port && !(info->port & 0xffff)) p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), "|%u", info->port); } __lws_lc_tag(&context->lcg[LWSLCG_VHOST], &vh->lc, "%s|%s|%d", buf, info->iface ? info->iface : "", info->port); #if defined(LWS_WITH_SYS_FAULT_INJECTION) vh->fic.name = "vh"; if (info->fic.fi_owner.count) /* * This moves all the lws_fi_t from info->fi to the vhost fi, * leaving it empty */ lws_fi_import(&vh->fic, &info->fic); lws_fi_inherit_copy(&vh->fic, &context->fic, "vh", vh->name); if (lws_fi(&vh->fic, "vh_create_oom")) goto bail; #endif #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) vh->http.error_document_404 = info->error_document_404; #endif if (lws_check_opt(info->options, LWS_SERVER_OPTION_ONLY_RAW)) lwsl_info("%s set to only support RAW\n", vh->name); vh->iface = info->iface; #if !defined(LWS_PLAT_FREERTOS) && !defined(OPTEE_TA) && !defined(WIN32) vh->bind_iface = info->bind_iface; #endif #if defined(LWS_WITH_CLIENT) if (info->connect_timeout_secs) vh->connect_timeout_secs = (int)info->connect_timeout_secs; else vh->connect_timeout_secs = 20; #endif /* apply the context default lws_retry */ if (info->retry_and_idle_policy) vh->retry_policy = info->retry_and_idle_policy; else vh->retry_policy = &context->default_retry; /* * let's figure out how many protocols the user is handing us, using the * old or new way depending on what he gave us */ if (!pcols) for (vh->count_protocols = 0; info->pprotocols[vh->count_protocols]; vh->count_protocols++) ; else for (vh->count_protocols = 0; pcols[vh->count_protocols].callback; vh->count_protocols++) ; vh->options = info->options; vh->pvo = info->pvo; vh->headers = info->headers; vh->user = info->user; vh->finalize = info->finalize; vh->finalize_arg = info->finalize_arg; vh->listen_accept_role = info->listen_accept_role; vh->listen_accept_protocol = info->listen_accept_protocol; vh->unix_socket_perms = info->unix_socket_perms; LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) if (lws_rops_fidx(ar, LWS_ROPS_init_vhost) && (lws_rops_func_fidx(ar, LWS_ROPS_init_vhost)).init_vhost(vh, info)) return NULL; LWS_FOR_EVERY_AVAILABLE_ROLE_END; if (info->keepalive_timeout) vh->keepalive_timeout = info->keepalive_timeout; else vh->keepalive_timeout = 5; if (info->timeout_secs_ah_idle) vh->timeout_secs_ah_idle = (int)info->timeout_secs_ah_idle; else vh->timeout_secs_ah_idle = 10; #if defined(LWS_WITH_TLS) vh->tls.alpn = info->alpn; vh->tls.ssl_info_event_mask = info->ssl_info_event_mask; if (info->ecdh_curve) lws_strncpy(vh->tls.ecdh_curve, info->ecdh_curve, sizeof(vh->tls.ecdh_curve)); /* carefully allocate and take a copy of cert + key paths if present */ n = 0; if (info->ssl_cert_filepath) n += (int)strlen(info->ssl_cert_filepath) + 1; if (info->ssl_private_key_filepath) n += (int)strlen(info->ssl_private_key_filepath) + 1; if (n) { vh->tls.key_path = vh->tls.alloc_cert_path = lws_malloc((unsigned int)n, "vh paths"); if (info->ssl_cert_filepath) { n = (int)strlen(info->ssl_cert_filepath) + 1; memcpy(vh->tls.alloc_cert_path, info->ssl_cert_filepath, (unsigned int)n); vh->tls.key_path += n; } if (info->ssl_private_key_filepath) memcpy(vh->tls.key_path, info->ssl_private_key_filepath, strlen(info->ssl_private_key_filepath) + 1); } #endif #if defined(LWS_WITH_HTTP_PROXY) && defined(LWS_ROLE_WS) fx = 1; #endif #if defined(LWS_WITH_ABSTRACT) abs_pcol_count = (int)LWS_ARRAY_SIZE(available_abstract_protocols) - 1; #endif #if defined(LWS_WITH_SECURE_STREAMS) sec_pcol_count = (int)LWS_ARRAY_SIZE(available_secstream_protocols) - 1; #endif /* * give the vhost a unified list of protocols including: * * - internal, async_dns if enabled (first vhost only) * - internal, abstracted ones * - the ones that came from plugins * - his user protocols */ if (lws_fi(&vh->fic, "vh_create_pcols_oom")) lwsp = NULL; else lwsp = lws_zalloc(sizeof(struct lws_protocols) * ((unsigned int)vh->count_protocols + (unsigned int)abs_pcol_count + (unsigned int)sec_pcol_count + (unsigned int)context->plugin_protocol_count + (unsigned int)fx + 1), "vh plugin table"); if (!lwsp) { lwsl_err("OOM\n"); goto bail; } /* * 1: user protocols (from pprotocols or protocols) */ m = vh->count_protocols; if (!pcols) { for (n = 0; n < m; n++) memcpy(&lwsp[n], info->pprotocols[n], sizeof(lwsp[0])); } else memcpy(lwsp, pcols, sizeof(struct lws_protocols) * (unsigned int)m); /* * 2: abstract protocols */ #if defined(LWS_WITH_ABSTRACT) for (n = 0; n < abs_pcol_count; n++) { memcpy(&lwsp[m++], available_abstract_protocols[n], sizeof(*lwsp)); vh->count_protocols++; } #endif /* * 3: async dns protocol (first vhost only) */ #if defined(LWS_WITH_SYS_ASYNC_DNS) if (!context->vhost_list) { memcpy(&lwsp[m++], &lws_async_dns_protocol, sizeof(struct lws_protocols)); vh->count_protocols++; } #endif #if defined(LWS_WITH_SECURE_STREAMS) for (n = 0; n < sec_pcol_count; n++) { memcpy(&lwsp[m++], available_secstream_protocols[n], sizeof(*lwsp)); vh->count_protocols++; } #endif /* * 3: For compatibility, all protocols enabled on vhost if only * the default vhost exists. Otherwise only vhosts who ask * for a protocol get it enabled. */ if (context->options & LWS_SERVER_OPTION_EXPLICIT_VHOSTS) f = 0; (void)f; #ifdef LWS_WITH_PLUGINS if (plugin) { while (plugin) { const lws_plugin_protocol_t *plpr = (const lws_plugin_protocol_t *)plugin->hdr; for (n = 0; n < plpr->count_protocols; n++) { /* * for compatibility's sake, no pvo implies * allow all protocols */ if (f || lws_vhost_protocol_options(vh, plpr->protocols[n].name)) { memcpy(&lwsp[m], &plpr->protocols[n], sizeof(struct lws_protocols)); m++; vh->count_protocols++; } } plugin = plugin->list; } } #endif #if defined(LWS_WITH_HTTP_PROXY) && defined(LWS_ROLE_WS) memcpy(&lwsp[m++], &lws_ws_proxy, sizeof(*lwsp)); vh->count_protocols++; #endif vh->protocols = lwsp; vh->allocated_vhost_protocols = 1; vh->same_vh_protocol_owner = (struct lws_dll2_owner *) lws_zalloc(sizeof(struct lws_dll2_owner) * (unsigned int)vh->count_protocols, "same vh list"); #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) vh->http.mount_list = info->mounts; #endif #if defined(LWS_WITH_SYS_METRICS) && defined(LWS_WITH_SERVER) { char *end = buf + sizeof(buf) - 1; p = buf; p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), "vh.%s", vh->name); if (info->iface) p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), ".%s", info->iface); if (info->port && !(info->port & 0xffff)) p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), ".%u", info->port); p += lws_snprintf(p, lws_ptr_diff_size_t(end, p), ".rx"); vh->mt_traffic_rx = lws_metric_create(context, 0, buf); p[-2] = 't'; vh->mt_traffic_tx = lws_metric_create(context, 0, buf); } #endif #ifdef LWS_WITH_UNIX_SOCK if (LWS_UNIX_SOCK_ENABLED(vh)) { lwsl_info("Creating Vhost '%s' path \"%s\", %d protocols\n", vh->name, vh->iface, vh->count_protocols); } else #endif { switch(info->port) { case CONTEXT_PORT_NO_LISTEN: strcpy(buf, "(serving disabled)"); break; case CONTEXT_PORT_NO_LISTEN_SERVER: strcpy(buf, "(no listener)"); break; default: lws_snprintf(buf, sizeof(buf), "port %u", info->port); break; } lwsl_info("Creating Vhost '%s' %s, %d protocols, IPv6 %s\n", vh->name, buf, vh->count_protocols, LWS_IPV6_ENABLED(vh) ? "on" : "off"); } mounts = info->mounts; while (mounts) { (void)mount_protocols[0]; lwsl_info(" mounting %s%s to %s\n", mount_protocols[mounts->origin_protocol], mounts->origin ? mounts->origin : "none", mounts->mountpoint); mounts = mounts->mount_next; } vh->listen_port = info->port; #if defined(LWS_WITH_SOCKS5) vh->socks_proxy_port = 0; vh->socks_proxy_address[0] = '\0'; #endif #if defined(LWS_WITH_CLIENT) && defined(LWS_CLIENT_HTTP_PROXYING) /* either use proxy from info, or try get it from env var */ #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) vh->http.http_proxy_port = 0; vh->http.http_proxy_address[0] = '\0'; /* http proxy */ if (info->http_proxy_address) { /* override for backwards compatibility */ if (info->http_proxy_port) vh->http.http_proxy_port = info->http_proxy_port; lws_set_proxy(vh, info->http_proxy_address); } else #endif { #ifdef LWS_HAVE_GETENV #if defined(__COVERITY__) p = NULL; #else p = getenv("http_proxy"); /* coverity[tainted_scalar] */ if (p) { lws_strncpy(buf, p, sizeof(buf)); lws_set_proxy(vh, buf); } #endif #endif } #endif #if defined(LWS_WITH_SOCKS5) lws_socks5c_ads_server(vh, info); #endif vh->ka_time = info->ka_time; vh->ka_interval = info->ka_interval; vh->ka_probes = info->ka_probes; if (vh->options & LWS_SERVER_OPTION_STS) lwsl_notice(" STS enabled\n"); #ifdef LWS_WITH_ACCESS_LOG if (info->log_filepath) { if (lws_fi(&vh->fic, "vh_create_access_log_open_fail")) vh->log_fd = (int)LWS_INVALID_FILE; else vh->log_fd = lws_open(info->log_filepath, O_CREAT | O_APPEND | O_RDWR, 0600); if (vh->log_fd == (int)LWS_INVALID_FILE) { lwsl_err("unable to open log filepath %s\n", info->log_filepath); goto bail; } #ifndef WIN32 if (context->uid != (uid_t)-1) if (chown(info->log_filepath, context->uid, context->gid) == -1) lwsl_err("unable to chown log file %s\n", info->log_filepath); #endif } else vh->log_fd = (int)LWS_INVALID_FILE; #endif if (lws_fi(&vh->fic, "vh_create_ssl_srv") || lws_context_init_server_ssl(info, vh)) { lwsl_err("%s: lws_context_init_server_ssl failed\n", __func__); goto bail1; } if (lws_fi(&vh->fic, "vh_create_ssl_cli") || lws_context_init_client_ssl(info, vh)) { lwsl_err("%s: lws_context_init_client_ssl failed\n", __func__); goto bail1; } #if defined(LWS_WITH_SERVER) lws_context_lock(context, __func__); if (lws_fi(&vh->fic, "vh_create_srv_init")) n = -1; else n = _lws_vhost_init_server(info, vh); lws_context_unlock(context); if (n < 0) { lwsl_err("init server failed\n"); goto bail1; } #endif #if defined(LWS_WITH_SYS_ASYNC_DNS) n = !!context->vhost_list; #endif while (1) { if (!(*vh1)) { *vh1 = vh; break; } vh1 = &(*vh1)->vhost_next; }; #if defined(LWS_WITH_SYS_ASYNC_DNS) if (!n) lws_async_dns_init(context); #endif /* for the case we are adding a vhost much later, after server init */ if (context->protocol_init_done) if (lws_fi(&vh->fic, "vh_create_protocol_init") || lws_protocol_init(context)) { lwsl_err("%s: lws_protocol_init failed\n", __func__); goto bail1; } return vh; bail1: lws_vhost_destroy(vh); return NULL; bail: __lws_lc_untag(&vh->lc); lws_fi_destroy(&vh->fic); lws_free(vh); early_bail: lws_fi_destroy(&info->fic); return NULL; } int lws_init_vhost_client_ssl(const struct lws_context_creation_info *info, struct lws_vhost *vhost) { struct lws_context_creation_info i; memcpy(&i, info, sizeof(i)); i.port = CONTEXT_PORT_NO_LISTEN; return lws_context_init_client_ssl(&i, vhost); } void lws_cancel_service_pt(struct lws *wsi) { lws_plat_pipe_signal(wsi->a.context, wsi->tsi); } void lws_cancel_service(struct lws_context *context) { struct lws_context_per_thread *pt = &context->pt[0]; short m; if (context->service_no_longer_possible) return; lwsl_debug("%s\n", __func__); for (m = 0; m < context->count_threads; m++) { if (pt->pipe_wsi) lws_plat_pipe_signal(pt->context, m); pt++; } } int __lws_create_event_pipes(struct lws_context *context) { struct lws_context_per_thread *pt; struct lws *wsi; int n; /* * Create the pt event pipes... these are unique in that they are * not bound to a vhost or protocol (both are NULL) */ #if LWS_MAX_SMP > 1 for (n = 0; n < context->count_threads; n++) { #else n = 0; { #endif pt = &context->pt[n]; if (pt->pipe_wsi) return 0; wsi = __lws_wsi_create_with_role(context, n, &role_ops_pipe); if (!wsi) return 1; __lws_lc_tag(&context->lcg[LWSLCG_WSI], &wsi->lc, "pipe"); wsi->event_pipe = 1; pt->pipe_wsi = wsi; if (!lws_plat_pipe_create(wsi)) { /* * platform code returns 0 if it actually created pipes * and initialized pt->dummy_pipe_fds[]. If it used * some other mechanism outside of signaling in the * normal event loop, we skip treating the pipe as * related to dummy_pipe_fds[], adding it to the fds, * etc. */ wsi->desc.sockfd = context->pt[n].dummy_pipe_fds[0]; lwsl_debug("event pipe fd %d\n", wsi->desc.sockfd); if (lws_wsi_inject_to_loop(pt, wsi)) goto bail; } } return 0; bail: return 1; } void lws_destroy_event_pipe(struct lws *wsi) { int n; lwsl_info("%s\n", __func__); n = lws_wsi_extract_from_loop(wsi); lws_plat_pipe_close(wsi); if (!n) lws_free(wsi); } /* * Start close process for any wsi bound to this vhost that belong to the * service thread we are called from. Because of async event lib close, or * protocol staged close on wsi, latency with pts joining in closing their * wsi on the vhost, this may take some time. * * When the wsi count bound to the vhost (from all pts) drops to zero, the * vhost destruction will be finalized. */ void __lws_vhost_destroy_pt_wsi_dieback_start(struct lws_vhost *vh) { #if LWS_MAX_SMP > 1 /* calling pt thread has done its wsi dieback */ int tsi = lws_pthread_self_to_tsi(vh->context); #else int tsi = 0; #endif struct lws_context *ctx = vh->context; struct lws_context_per_thread *pt = &ctx->pt[tsi]; unsigned int n; #if LWS_MAX_SMP > 1 if (vh->close_flow_vs_tsi[lws_pthread_self_to_tsi(vh->context)]) /* this pt has already done its bit */ return; #endif #if defined(LWS_WITH_CLIENT) /* * destroy any wsi that are associated with us but have no socket * (and will otherwise be missed for destruction) */ lws_start_foreach_dll_safe(struct lws_dll2 *, d, d1, vh->vh_awaiting_socket_owner.head) { struct lws *w = lws_container_of(d, struct lws, vh_awaiting_socket); if (w->tsi == tsi) { lwsl_debug("%s: closing aso\n", __func__); lws_close_free_wsi(w, LWS_CLOSE_STATUS_NOSTATUS, "awaiting skt"); } } lws_end_foreach_dll_safe(d, d1); #endif /* * Close any wsi on this pt bound to the vhost */ n = 0; while (n < pt->fds_count) { struct lws *wsi = wsi_from_fd(ctx, pt->fds[n].fd); if (wsi && wsi->tsi == tsi && wsi->a.vhost == vh) { lwsl_debug("%s: pt %d: closing wsi %p: role %s\n", __func__, tsi, wsi, wsi->role_ops->name); lws_wsi_close(wsi, LWS_TO_KILL_ASYNC); if (pt->pipe_wsi == wsi) pt->pipe_wsi = NULL; } n++; } #if LWS_MAX_SMP > 1 /* calling pt thread has done its wsi dieback */ vh->close_flow_vs_tsi[lws_pthread_self_to_tsi(vh->context)] = 1; #endif } /* * Mark the vhost as being destroyed, so things trying to use it abort. * * Dispose of the listen socket. */ void lws_vhost_destroy1(struct lws_vhost *vh) { struct lws_context *context = vh->context; lwsl_info("%s\n", __func__); lws_context_lock(context, "vhost destroy 1"); /* ---------- context { */ if (vh->being_destroyed) goto out; lws_vhost_lock(vh); /* -------------- vh { */ #if defined(LWS_WITH_TLS_SESSIONS) && defined(LWS_WITH_TLS) lws_tls_session_vh_destroy(vh); #endif vh->being_destroyed = 1; lws_dll2_add_tail(&vh->vh_being_destroyed_list, &context->owner_vh_being_destroyed); #if defined(LWS_WITH_NETWORK) /* * PHASE 1: take down or reassign any listen wsi * * Are there other vhosts that are piggybacking on our listen socket? * If so we need to hand the listen socket off to one of the others * so it will remain open. * * If not, close the listen socket now. * * Either way the listen socket response to the vhost close is * immediately performed. */ if (vh->lserv_wsi) { lws_start_foreach_ll(struct lws_vhost *, v, context->vhost_list) { if (v != vh && !v->being_destroyed && v->listen_port == vh->listen_port && ((!v->iface && !vh->iface) || (v->iface && vh->iface && !strcmp(v->iface, vh->iface)))) { /* * this can only be a listen wsi, which is * restricted... it has no protocol or other * bindings or states. So we can simply * swap it to a vhost that has the same * iface + port, but is not closing. */ lwsl_notice("%s: listen skt migrate %s -> %s\n", __func__, lws_vh_tag(vh), lws_vh_tag(v)); assert(v->lserv_wsi == NULL); v->lserv_wsi = vh->lserv_wsi; if (v->lserv_wsi) { /* req cx + vh lock */ /* * If the vhost sees it's being destroyed and * in the unbind the number of wsis bound to * it falls to zero, it will destroy the * vhost opportunistically before we can * complete the transfer. Add a fake wsi * bind temporarily to disallow this... */ v->count_bound_wsi++; __lws_vhost_unbind_wsi(vh->lserv_wsi); lws_vhost_bind_wsi(v, v->lserv_wsi); /* * ... remove the fake wsi bind */ v->count_bound_wsi--; vh->lserv_wsi = NULL; } break; } } lws_end_foreach_ll(v, vhost_next); if (vh->lserv_wsi) { /* * we didn't pass it off to another vhost on the same * listen port... let's close it next time around the * event loop without waiting for the logical destroy * of the vhost itself */ lws_set_timeout(vh->lserv_wsi, 1, LWS_TO_KILL_ASYNC); vh->lserv_wsi = NULL; } } #endif lws_vhost_unlock(vh); /* } vh -------------- */ out: lws_context_unlock(context); /* --------------------------- context { */ } #if defined(LWS_WITH_ABSTRACT) static int destroy_ais(struct lws_dll2 *d, void *user) { lws_abs_t *ai = lws_container_of(d, lws_abs_t, abstract_instances); lws_abs_destroy_instance(&ai); return 0; } #endif /* * Either start close or destroy any wsi on the vhost that belong to this pt, * if SMP mark the vh that we have done it for * * Must not have lock on vh */ void __lws_vhost_destroy2(struct lws_vhost *vh) { const struct lws_protocols *protocol = NULL; struct lws_context *context = vh->context; struct lws wsi; int n; vh->being_destroyed = 0; // lwsl_info("%s: %s\n", __func__, vh->name); #if defined(LWS_WITH_DEPRECATED_THINGS) /* * destroy any pending timed events */ while (vh->timed_vh_protocol_list) __lws_timed_callback_remove(vh, vh->timed_vh_protocol_list); #endif /* * remove ourselves from the defer binding list */ lws_start_foreach_llp(struct lws_vhost **, pv, vh->context->no_listener_vhost_list) { if (*pv == vh) { lwsl_debug("deferred iface: removing vh %s\n", (*pv)->name); *pv = vh->no_listener_vhost_list; vh->no_listener_vhost_list = NULL; break; } } lws_end_foreach_llp(pv, no_listener_vhost_list); /* * let the protocols destroy the per-vhost protocol objects */ memset(&wsi, 0, sizeof(wsi)); wsi.a.context = vh->context; wsi.a.vhost = vh; /* not a real bound wsi */ protocol = vh->protocols; if (protocol && vh->created_vhost_protocols) { n = 0; while (n < vh->count_protocols) { wsi.a.protocol = protocol; lwsl_debug("%s: protocol destroy\n", __func__); if (protocol->callback) protocol->callback(&wsi, LWS_CALLBACK_PROTOCOL_DESTROY, NULL, NULL, 0); protocol++; n++; } } /* * remove vhost from context list of vhosts */ lws_start_foreach_llp(struct lws_vhost **, pv, context->vhost_list) { if (*pv == vh) { *pv = vh->vhost_next; break; } } lws_end_foreach_llp(pv, vhost_next); /* add ourselves to the pending destruction list */ if (vh->context->vhost_pending_destruction_list != vh) { vh->vhost_next = vh->context->vhost_pending_destruction_list; vh->context->vhost_pending_destruction_list = vh; } //lwsl_debug("%s: do dfl '%s'\n", __func__, vh->name); /* remove ourselves from the pending destruction list */ lws_start_foreach_llp(struct lws_vhost **, pv, context->vhost_pending_destruction_list) { if ((*pv) == vh) { *pv = (*pv)->vhost_next; break; } } lws_end_foreach_llp(pv, vhost_next); /* * Free all the allocations associated with the vhost */ protocol = vh->protocols; if (protocol) { n = 0; while (n < vh->count_protocols) { if (vh->protocol_vh_privs && vh->protocol_vh_privs[n]) { lws_free(vh->protocol_vh_privs[n]); vh->protocol_vh_privs[n] = NULL; } protocol++; n++; } } if (vh->protocol_vh_privs) lws_free(vh->protocol_vh_privs); lws_ssl_SSL_CTX_destroy(vh); lws_free(vh->same_vh_protocol_owner); if ( #if defined(LWS_WITH_PLUGINS) context->plugin_list || #endif (context->options & LWS_SERVER_OPTION_EXPLICIT_VHOSTS) || vh->allocated_vhost_protocols) lws_free((void *)vh->protocols); #if defined(LWS_WITH_NETWORK) LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) if (lws_rops_fidx(ar, LWS_ROPS_destroy_vhost)) lws_rops_func_fidx(ar, LWS_ROPS_destroy_vhost). destroy_vhost(vh); LWS_FOR_EVERY_AVAILABLE_ROLE_END; #endif #ifdef LWS_WITH_ACCESS_LOG if (vh->log_fd != (int)LWS_INVALID_FILE) close(vh->log_fd); #endif #if defined (LWS_WITH_TLS) lws_free_set_NULL(vh->tls.alloc_cert_path); #endif #if LWS_MAX_SMP > 1 lws_mutex_refcount_destroy(&context->mr); #endif #if defined(LWS_WITH_UNIX_SOCK) if (LWS_UNIX_SOCK_ENABLED(vh)) { n = unlink(vh->iface); if (n) lwsl_info("Closing unix socket %s: errno %d\n", vh->iface, errno); } #endif /* * although async event callbacks may still come for wsi handles with * pending close in the case of asycn event library like libuv, * they do not refer to the vhost. So it's safe to free. */ if (vh->finalize) vh->finalize(vh, vh->finalize_arg); #if defined(LWS_WITH_ABSTRACT) /* * abstract instances */ lws_dll2_foreach_safe(&vh->abstract_instances_owner, NULL, destroy_ais); #endif #if defined(LWS_WITH_SERVER) && defined(LWS_WITH_SYS_METRICS) lws_metric_destroy(&vh->mt_traffic_rx, 0); lws_metric_destroy(&vh->mt_traffic_tx, 0); #endif lws_dll2_remove(&vh->vh_being_destroyed_list); #if defined(LWS_WITH_SYS_FAULT_INJECTION) lws_fi_destroy(&vh->fic); #endif __lws_lc_untag(&vh->lc); memset(vh, 0, sizeof(*vh)); lws_free(vh); } /* * Starts the vhost destroy process * * Vhosts are not simple to deal with because they are an abstraction that * crosses SMP thread boundaries, a wsi on any pt can bind to any vhost. If we * want another pt to do something to its wsis safely, we have to asynchronously * ask it to do it. * * In addition, with event libs, closing any handles (which are bound to vhosts * in their wsi) can happens asynchronously, so we can't just linearly do some * cleanup flow and free it in one step. * * The vhost destroy is cut into two pieces: * * 1) dispose of the listen socket, either by passing it on to another vhost * that was already sharing it, or just closing it. * * If any wsi bound to the vhost, mark the vhost as in the process of being * destroyed, triggering each pt to close all wsi bound to the vhost next * time around the event loop. Call lws_cancel_service() so all the pts wake * to deal with this without long poll waits making delays. * * 2) When the number of wsis bound to the vhost reaches zero, do the final * vhost destroy flow, this can be triggered from any pt. */ void lws_vhost_destroy(struct lws_vhost *vh) { struct lws_context *context = vh->context; lws_context_lock(context, __func__); /* ------ context { */ /* dispose of the listen socket one way or another */ lws_vhost_destroy1(vh); /* start async closure of all wsi on this pt thread attached to vh */ __lws_vhost_destroy_pt_wsi_dieback_start(vh); lwsl_info("%s: count_bound_wsi %d\n", __func__, vh->count_bound_wsi); /* if there are none, finalize now since no further chance */ if (!vh->count_bound_wsi) { __lws_vhost_destroy2(vh); goto out; } /* * We have some wsi bound to this vhost, we have to wait for these to * complete close and unbind before progressing the vhost removal. * * When the last bound wsi on this vh is destroyed we will auto-call * __lws_vhost_destroy2() to finalize vh destruction */ #if LWS_MAX_SMP > 1 /* alert other pts they also need to do dieback flow for their wsi */ lws_cancel_service(context); #endif out: lws_context_unlock(context); /* } context ------------------- */ } void * lws_vhost_user(struct lws_vhost *vhost) { return vhost->user; } int lws_get_vhost_listen_port(struct lws_vhost *vhost) { return vhost->listen_port; } #if defined(LWS_WITH_SERVER) void lws_context_deprecate(struct lws_context *context, lws_reload_func cb) { struct lws_vhost *vh = context->vhost_list, *vh1; /* * "deprecation" means disable the context from accepting any new * connections and free up listen sockets to be used by a replacement * context. * * Otherwise the deprecated context remains operational, until its * number of connected sockets falls to zero, when it is deleted. */ /* for each vhost, close his listen socket */ while (vh) { struct lws *wsi = vh->lserv_wsi; if (wsi) { wsi->socket_is_permanently_unusable = 1; lws_close_free_wsi(wsi, LWS_CLOSE_STATUS_NOSTATUS, "ctx deprecate"); wsi->a.context->deprecation_pending_listen_close_count++; /* * other vhosts can share the listen port, they * point to the same wsi. So zap those too. */ vh1 = context->vhost_list; while (vh1) { if (vh1->lserv_wsi == wsi) vh1->lserv_wsi = NULL; vh1 = vh1->vhost_next; } } vh = vh->vhost_next; } context->deprecated = 1; context->deprecation_cb = cb; } #endif #if defined(LWS_WITH_NETWORK) struct lws_vhost * lws_get_vhost_by_name(struct lws_context *context, const char *name) { lws_start_foreach_ll(struct lws_vhost *, v, context->vhost_list) { if (!v->being_destroyed && !strcmp(v->name, name)) return v; } lws_end_foreach_ll(v, vhost_next); return NULL; } #if defined(LWS_WITH_CLIENT) /* * This is the logic checking to see if the new connection wsi should have a * pipelining or muxing relationship with an existing "active connection" to * the same endpoint under the same conditions. * * This was originally in the client code but since the list is held on the * vhost (to ensure the same client tls ctx is involved) it's cleaner in vhost.c * * ACTIVE_CONNS_QUEUED: We're queued on an active connection, set *nwsi to that * ACTIVE_CONNS_MUXED: We are joining an active mux conn *nwsi as a child * ACTIVE_CONNS_SOLO: There's no existing conn to join either way */ int lws_vhost_active_conns(struct lws *wsi, struct lws **nwsi, const char *adsin) { #if defined(LWS_WITH_TLS) const char *my_alpn = lws_wsi_client_stash_item(wsi, CIS_ALPN, _WSI_TOKEN_CLIENT_ALPN); #endif #if defined(LWS_WITH_TLS) char newconn_cannot_use_h1 = 0; if ((wsi->tls.use_ssl & LCCSCF_USE_SSL) && my_alpn && !strstr(my_alpn, "http/1.1")) /* * new guy wants to use tls, he specifies the alpn and he does * not list h1 as a choice ==> he can't bind to existing h1 */ newconn_cannot_use_h1 = 1; #endif if (!lws_dll2_is_detached(&wsi->dll2_cli_txn_queue)) { struct lws *w = lws_container_of( wsi->dll2_cli_txn_queue.owner, struct lws, dll2_cli_txn_queue_owner); *nwsi = w; return ACTIVE_CONNS_QUEUED; } #if defined(LWS_ROLE_H2) || defined(LWS_ROLE_MQTT) if (wsi->mux.parent_wsi) { /* * We already decided... */ *nwsi = wsi->mux.parent_wsi; return ACTIVE_CONNS_MUXED; } #endif lws_context_lock(wsi->a.context, __func__); /* -------------- cx { */ lws_vhost_lock(wsi->a.vhost); /* ----------------------------------- { */ lws_start_foreach_dll_safe(struct lws_dll2 *, d, d1, wsi->a.vhost->dll_cli_active_conns_owner.head) { struct lws *w = lws_container_of(d, struct lws, dll_cli_active_conns); lwsl_debug("%s: check %s %s %s %s %d %d\n", __func__, lws_wsi_tag(wsi), lws_wsi_tag(w), adsin, w->cli_hostname_copy, wsi->c_port, w->c_port); if (w != wsi && /* * "same internet protocol"... this is a bit tricky, * since h2 start out as h1, and may stay at h1. * * But an idle h1 connection cannot be used by a connection * request that doesn't have http/1.1 in its alpn list... */ (w->role_ops == wsi->role_ops || (lwsi_role_http(w) && lwsi_role_http(wsi))) && /* ... same role, or at least both some kind of http */ w->cli_hostname_copy && !strcmp(adsin, w->cli_hostname_copy) && /* same endpoint hostname */ #if defined(LWS_WITH_TLS) !(newconn_cannot_use_h1 && w->role_ops == &role_ops_h1) && /* if we can't use h1, old guy must not be h1 */ (wsi->tls.use_ssl & LCCSCF_USE_SSL) == (w->tls.use_ssl & LCCSCF_USE_SSL) && /* must both agree on tls use or not */ #endif wsi->c_port == w->c_port) { /* same endpoint port */ /* * There's already an active connection. * * The server may have told the existing active * connection that it doesn't support pipelining... */ if (w->keepalive_rejected) { lwsl_notice("defeating pipelining due to no " "keepalive on server\n"); goto solo; } #if defined(LWS_WITH_HTTP2) /* * h2: if in usable state already: just use it without * going through the queue */ if (w->client_h2_alpn && w->client_mux_migrated && (lwsi_state(w) == LRS_H2_WAITING_TO_SEND_HEADERS || lwsi_state(w) == LRS_ESTABLISHED || lwsi_state(w) == LRS_IDLING)) { lwsl_notice("%s: just join h2 directly 0x%x\n", __func__, lwsi_state(w)); if (lwsi_state(w) == LRS_IDLING) { // lwsi_set_state(w, LRS_ESTABLISHED); _lws_generic_transaction_completed_active_conn(&w, 0); } //lwsi_set_state(w, LRS_H1C_ISSUE_HANDSHAKE2); wsi->client_h2_alpn = 1; lws_wsi_h2_adopt(w, wsi); lws_vhost_unlock(wsi->a.vhost); /* } ---------- */ lws_context_unlock(wsi->a.context); /* -------------- cx { */ *nwsi = w; return ACTIVE_CONNS_MUXED; } #endif #if defined(LWS_ROLE_MQTT) /* * MQTT: if in usable state already: just use it without * going through the queue */ if (lwsi_role_mqtt(wsi) && w->client_mux_migrated && lwsi_state(w) == LRS_ESTABLISHED) { if (lws_wsi_mqtt_adopt(w, wsi)) { lwsl_notice("%s: join mqtt directly\n", __func__); lws_dll2_remove(&wsi->dll2_cli_txn_queue); wsi->client_mux_substream = 1; lws_vhost_unlock(wsi->a.vhost); /* } ---------- */ lws_context_unlock(wsi->a.context); /* -------------- cx { */ return ACTIVE_CONNS_MUXED; } } #endif /* * If the connection is viable but not yet in a usable * state, let's attach ourselves to it and wait for it * to get there or fail. */ lwsl_notice("%s: apply %s to txn queue on %s state 0x%lx\n", __func__, lws_wsi_tag(wsi), lws_wsi_tag(w), (unsigned long)w->wsistate); /* * ...let's add ourselves to his transaction queue... * we are adding ourselves at the TAIL */ lws_dll2_add_tail(&wsi->dll2_cli_txn_queue, &w->dll2_cli_txn_queue_owner); if (lwsi_state(w) == LRS_IDLING) { // lwsi_set_state(w, LRS_ESTABLISHED); _lws_generic_transaction_completed_active_conn(&w, 0); } /* * For eg, h1 next we'd pipeline our headers out on him, * and wait for our turn at client transaction_complete * to take over parsing the rx. */ lws_vhost_unlock(wsi->a.vhost); /* } ---------- */ lws_context_unlock(wsi->a.context); /* -------------- cx { */ *nwsi = w; return ACTIVE_CONNS_QUEUED; } } lws_end_foreach_dll_safe(d, d1); solo: lws_vhost_unlock(wsi->a.vhost); /* } ---------------------------------- */ lws_context_unlock(wsi->a.context); /* -------------- cx { */ /* there is nobody already connected in the same way */ return ACTIVE_CONNS_SOLO; } #endif #endif const char * lws_vh_tag(struct lws_vhost *vh) { return lws_lc_tag(&vh->lc); }