/* * libwebsockets - small server side websockets and web server implementation * * Copyright (C) 2010 - 2020 Andy Green * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to * deal in the Software without restriction, including without limitation the * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or * sell copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include "private-lib-core.h" #ifndef LWS_BUILD_HASH #define LWS_BUILD_HASH "unknown-build-hash" #endif static const char *library_version = LWS_LIBRARY_VERSION " " LWS_BUILD_HASH; #if defined(LWS_WITH_NETWORK) /* in ms */ static uint32_t default_backoff_table[] = { 1000, 3000, 9000, 17000 }; #endif /** * lws_get_library_version: get version and git hash library built from * * returns a const char * to a string like "1.1 178d78c" * representing the library version followed by the git head hash it * was built from */ const char * lws_get_library_version(void) { return library_version; } #if defined(LWS_WITH_STATS) static void lws_sul_stats_cb(lws_sorted_usec_list_t *sul) { struct lws_context_per_thread *pt = lws_container_of(sul, struct lws_context_per_thread, sul_stats); lws_stats_log_dump(pt->context); __lws_sul_insert(&pt->pt_sul_owner, &pt->sul_stats, 10 * LWS_US_PER_SEC); } #endif #if defined(LWS_WITH_PEER_LIMITS) static void lws_sul_peer_limits_cb(lws_sorted_usec_list_t *sul) { struct lws_context_per_thread *pt = lws_container_of(sul, struct lws_context_per_thread, sul_peer_limits); lws_peer_cull_peer_wait_list(pt->context); __lws_sul_insert(&pt->pt_sul_owner, &pt->sul_peer_limits, 10 * LWS_US_PER_SEC); } #endif #if defined(LWS_WITH_NETWORK) #if defined(_DEBUG) static const char * system_state_names[] = { "undef", "CONTEXT_CREATED", "INITIALIZED", "IFACE_COLDPLUG", "DHCP", "TIME_VALID", "POLICY_VALID", "REGISTERED", "AUTH1", "AUTH2", "OPERATIONAL", "POLICY_INVALID" }; #endif /* * Handle provoking protocol init when we pass through the right system state */ static int lws_state_notify_protocol_init(struct lws_state_manager *mgr, struct lws_state_notify_link *link, int current, int target) { struct lws_context *context = lws_container_of(mgr, struct lws_context, mgr_system); int n; /* * Deal with any attachments that were waiting for the right state * to come along */ for (n = 0; n < context->count_threads; n++) lws_system_do_attach(&context->pt[n]); #if defined(LWS_WITH_SYS_DHCP_CLIENT) if (target == LWS_SYSTATE_DHCP) { /* * Don't let it past here until at least one iface has been * configured for operation with DHCP */ if (!lws_dhcpc_status(context, NULL)) return 1; } #endif #if defined(LWS_WITH_SECURE_STREAMS_SYS_AUTH_API_AMAZON_COM) /* * Skip this if we are running something without the policy for it */ if (target == LWS_SYSTATE_AUTH1 && context->pss_policies && !lws_system_blob_get_size(lws_system_get_blob(context, LWS_SYSBLOB_TYPE_AUTH, 0))) { lwsl_info("%s: AUTH1 state triggering api.amazon.com auth\n", __func__); /* * Start trying to acquire it if it's not already in progress * returns nonzero if we determine it's not needed */ if (!lws_ss_sys_auth_api_amazon_com(context)) return 1; } #endif #if defined(LWS_WITH_SECURE_STREAMS) /* * Skip this if we are running something without the policy for it */ if (target == LWS_SYSTATE_POLICY_VALID && context->pss_policies && !context->policy_updated) { /* * Start trying to acquire it if it's not already in progress * returns nonzero if we determine it's not needed */ if (!lws_ss_sys_fetch_policy(context)) return 1; } #endif /* protocol part */ if (context->protocol_init_done) return 0; if (target != LWS_SYSTATE_POLICY_VALID) return 0; lwsl_info("%s: doing protocol init on POLICY_VALID\n", __func__); lws_protocol_init(context); return 0; } static void lws_context_creation_completion_cb(lws_sorted_usec_list_t *sul) { struct lws_context *context = lws_container_of(sul, struct lws_context, sul_system_state); /* if nothing is there to intercept anything, go all the way */ lws_state_transition_steps(&context->mgr_system, LWS_SYSTATE_OPERATIONAL); } #endif struct lws_context * lws_create_context(const struct lws_context_creation_info *info) { struct lws_context *context = NULL; #if defined(LWS_WITH_FILE_OPS) struct lws_plat_file_ops *prev; #endif #ifndef LWS_NO_DAEMONIZE pid_t pid_daemon = get_daemonize_pid(); #endif #if defined(LWS_WITH_NETWORK) int n, count_threads = 1; uint8_t *u; #endif #if defined(__ANDROID__) struct rlimit rt; #endif size_t s1 = 4096, size = sizeof(struct lws_context); int lpf = info->fd_limit_per_thread; if (lpf) { lpf+= 2; #if defined(LWS_WITH_SYS_ASYNC_DNS) lpf++; #endif #if defined(LWS_WITH_SYS_NTPCLIENT) lpf++; #endif #if defined(LWS_WITH_SYS_DHCP_CLIENT) lpf++; #endif } lwsl_info("Initial logging level %d\n", log_level); lwsl_info("Libwebsockets version: %s\n", library_version); #ifdef LWS_WITH_IPV6 if (!lws_check_opt(info->options, LWS_SERVER_OPTION_DISABLE_IPV6)) lwsl_info("IPV6 compiled in and enabled\n"); else lwsl_info("IPV6 compiled in but disabled\n"); #else lwsl_info("IPV6 not compiled in\n"); #endif lwsl_info(" LWS_DEF_HEADER_LEN : %u\n", LWS_DEF_HEADER_LEN); lwsl_info(" LWS_MAX_SMP : %u\n", LWS_MAX_SMP); lwsl_info(" sizeof (*info) : %ld\n", (long)sizeof(*info)); #if defined(LWS_WITH_STATS) lwsl_info(" LWS_WITH_STATS : on\n"); #endif lwsl_info(" SYSTEM_RANDOM_FILEPATH: '%s'\n", SYSTEM_RANDOM_FILEPATH); #if defined(LWS_WITH_HTTP2) lwsl_info(" HTTP2 support : available\n"); #else lwsl_info(" HTTP2 support : not configured\n"); #endif if (lws_plat_context_early_init()) return NULL; #if defined(LWS_WITH_NETWORK) if (info->count_threads) count_threads = info->count_threads; if (count_threads > LWS_MAX_SMP) count_threads = LWS_MAX_SMP; if (info->pt_serv_buf_size) s1 = info->pt_serv_buf_size; /* pt fakewsi and the pt serv buf allocations ride after the context */ size += count_threads * (s1 + sizeof(struct lws)); #endif context = lws_zalloc(size, "context"); if (!context) { lwsl_err("No memory for websocket context\n"); return NULL; } context->uid = info->uid; context->gid = info->gid; context->username = info->username; context->groupname = info->groupname; context->system_ops = info->system_ops; context->pt_serv_buf_size = (unsigned int)s1; context->udp_loss_sim_tx_pc = info->udp_loss_sim_tx_pc; context->udp_loss_sim_rx_pc = info->udp_loss_sim_rx_pc; if (context->udp_loss_sim_tx_pc || context->udp_loss_sim_rx_pc) lwsl_warn("%s: simulating udp loss tx: %d%%, rx: %d%%\n", __func__, context->udp_loss_sim_tx_pc, context->udp_loss_sim_rx_pc); #if defined(LWS_WITH_SECURE_STREAMS_PROXY_API) context->ss_proxy_bind = info->ss_proxy_bind; context->ss_proxy_port = info->ss_proxy_port; context->ss_proxy_address = info->ss_proxy_address; lwsl_notice("%s: using ss proxy bind '%s', port %d, ads '%s'\n", __func__, context->ss_proxy_bind, context->ss_proxy_port, context->ss_proxy_address); #endif #if defined(LWS_WITH_NETWORK) context->count_threads = count_threads; #if defined(LWS_WITH_DETAILED_LATENCY) context->detailed_latency_cb = info->detailed_latency_cb; context->detailed_latency_filepath = info->detailed_latency_filepath; context->latencies_fd = -1; #endif #if defined(LWS_WITHOUT_EXTENSIONS) if (info->extensions) lwsl_warn("%s: LWS_WITHOUT_EXTENSIONS but extensions ptr set\n", __func__); #endif #endif #if defined(LWS_WITH_SECURE_STREAMS) context->pss_policies_json = info->pss_policies_json; context->pss_plugins = info->pss_plugins; #endif /* if he gave us names, set the uid / gid */ if (lws_plat_drop_app_privileges(context, 0)) goto bail; lwsl_info("context created\n"); #if defined(LWS_WITH_TLS) && defined(LWS_WITH_NETWORK) #if defined(LWS_WITH_MBEDTLS) context->tls_ops = &tls_ops_mbedtls; #else context->tls_ops = &tls_ops_openssl; #endif #endif #if LWS_MAX_SMP > 1 lws_mutex_refcount_init(&context->mr); #endif #if defined(LWS_PLAT_FREERTOS) #if defined(LWS_AMAZON_RTOS) context->last_free_heap = xPortGetFreeHeapSize(); #else context->last_free_heap = esp_get_free_heap_size(); #endif #endif #if defined(LWS_WITH_FILE_OPS) /* default to just the platform fops implementation */ context->fops_platform.LWS_FOP_OPEN = _lws_plat_file_open; context->fops_platform.LWS_FOP_CLOSE = _lws_plat_file_close; context->fops_platform.LWS_FOP_SEEK_CUR = _lws_plat_file_seek_cur; context->fops_platform.LWS_FOP_READ = _lws_plat_file_read; context->fops_platform.LWS_FOP_WRITE = _lws_plat_file_write; context->fops_platform.fi[0].sig = NULL; /* * arrange a linear linked-list of fops starting from context->fops * * platform fops * [ -> fops_zip (copied into context so .next settable) ] * [ -> info->fops ] */ context->fops = &context->fops_platform; prev = (struct lws_plat_file_ops *)context->fops; #if defined(LWS_WITH_ZIP_FOPS) /* make a soft copy so we can set .next */ context->fops_zip = fops_zip; prev->next = &context->fops_zip; prev = (struct lws_plat_file_ops *)prev->next; #endif /* if user provided fops, tack them on the end of the list */ if (info->fops) prev->next = info->fops; #endif #if defined(LWS_WITH_SERVER) context->reject_service_keywords = info->reject_service_keywords; #endif if (info->external_baggage_free_on_destroy) context->external_baggage_free_on_destroy = info->external_baggage_free_on_destroy; #if defined(LWS_WITH_NETWORK) context->time_up = lws_now_usecs(); #endif context->pcontext_finalize = info->pcontext; context->simultaneous_ssl_restriction = info->simultaneous_ssl_restriction; context->options = info->options; #ifndef LWS_NO_DAEMONIZE if (pid_daemon) { context->started_with_parent = pid_daemon; lwsl_info(" Started with daemon pid %u\n", (unsigned int)pid_daemon); } #endif #if defined(__ANDROID__) n = getrlimit(RLIMIT_NOFILE, &rt); if (n == -1) { lwsl_err("Get RLIMIT_NOFILE failed!\n"); return NULL; } context->max_fds = rt.rlim_cur; #else #if defined(WIN32) || defined(_WIN32) || defined(LWS_AMAZON_RTOS) context->max_fds = getdtablesize(); #else context->max_fds = sysconf(_SC_OPEN_MAX); #endif if (context->max_fds < 0) { lwsl_err("%s: problem getting process max files\n", __func__); return NULL; } #endif /* * deal with any max_fds override, if it's reducing (setting it to * more than ulimit -n is meaningless). The platform init will * figure out what if this is something it can deal with. */ if (info->fd_limit_per_thread) { int mf = lpf * context->count_threads; if (mf < context->max_fds) { context->max_fds_unrelated_to_ulimit = 1; context->max_fds = mf; } } context->token_limits = info->token_limits; #if defined(LWS_WITH_NETWORK) /* * set the context event loops ops struct * * after this, all event_loop actions use the generic ops */ #if defined(LWS_WITH_POLL) context->event_loop_ops = &event_loop_ops_poll; #endif if (lws_check_opt(context->options, LWS_SERVER_OPTION_LIBUV)) #if defined(LWS_WITH_LIBUV) context->event_loop_ops = &event_loop_ops_uv; #else goto fail_event_libs; #endif if (lws_check_opt(context->options, LWS_SERVER_OPTION_LIBEV)) #if defined(LWS_WITH_LIBEV) context->event_loop_ops = &event_loop_ops_ev; #else goto fail_event_libs; #endif if (lws_check_opt(context->options, LWS_SERVER_OPTION_LIBEVENT)) #if defined(LWS_WITH_LIBEVENT) context->event_loop_ops = &event_loop_ops_event; #else goto fail_event_libs; #endif if (lws_check_opt(context->options, LWS_SERVER_OPTION_GLIB)) #if defined(LWS_WITH_GLIB) context->event_loop_ops = &event_loop_ops_glib; #else goto fail_event_libs; #endif if (!context->event_loop_ops) goto fail_event_libs; lwsl_info("Using event loop: %s\n", context->event_loop_ops->name); #endif #if defined(LWS_WITH_TLS) && defined(LWS_WITH_NETWORK) time(&context->tls.last_cert_check_s); if (info->alpn) context->tls.alpn_default = info->alpn; else { char *p = context->tls.alpn_discovered, first = 1; LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) { if (ar->alpn) { if (!first) *p++ = ','; p += lws_snprintf(p, context->tls.alpn_discovered + sizeof(context->tls.alpn_discovered) - 2 - p, "%s", ar->alpn); first = 0; } } LWS_FOR_EVERY_AVAILABLE_ROLE_END; context->tls.alpn_default = context->tls.alpn_discovered; } lwsl_info("Default ALPN advertisment: %s\n", context->tls.alpn_default); #endif if (info->timeout_secs) context->timeout_secs = info->timeout_secs; else context->timeout_secs = AWAITING_TIMEOUT; context->ws_ping_pong_interval = info->ws_ping_pong_interval; lwsl_info(" default timeout (secs): %u\n", context->timeout_secs); if (info->max_http_header_data) context->max_http_header_data = info->max_http_header_data; else if (info->max_http_header_data2) context->max_http_header_data = info->max_http_header_data2; else context->max_http_header_data = LWS_DEF_HEADER_LEN; if (info->max_http_header_pool) context->max_http_header_pool = info->max_http_header_pool; else if (info->max_http_header_pool2) context->max_http_header_pool = info->max_http_header_pool2; else context->max_http_header_pool = context->max_fds; if (info->fd_limit_per_thread) context->fd_limit_per_thread = lpf; else if (context->count_threads) context->fd_limit_per_thread = context->max_fds / context->count_threads; #if defined(LWS_WITH_NETWORK) context->default_retry.retry_ms_table = default_backoff_table; context->default_retry.conceal_count = context->default_retry.retry_ms_table_count = LWS_ARRAY_SIZE(default_backoff_table); context->default_retry.jitter_percent = 20; context->default_retry.secs_since_valid_ping = 300; context->default_retry.secs_since_valid_hangup = 310; if (info->retry_and_idle_policy && info->retry_and_idle_policy->secs_since_valid_ping) { context->default_retry.secs_since_valid_ping = info->retry_and_idle_policy->secs_since_valid_ping; context->default_retry.secs_since_valid_hangup = info->retry_and_idle_policy->secs_since_valid_hangup; } /* * Allocate the per-thread storage for scratchpad buffers, * and header data pool */ u = (uint8_t *)&context[1]; for (n = 0; n < context->count_threads; n++) { context->pt[n].serv_buf = u; u += context->pt_serv_buf_size; context->pt[n].context = context; context->pt[n].tid = n; /* * We overallocated for a fakewsi (can't compose it in the * pt because size isn't known at that time). point to it * and zero it down. Fakewsis are needed to make callbacks work * when the source of the callback is not actually from a wsi * context. */ context->pt[n].fake_wsi = (struct lws *)u; u += sizeof(struct lws); memset(context->pt[n].fake_wsi, 0, sizeof(struct lws)); #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) context->pt[n].http.ah_list = NULL; context->pt[n].http.ah_pool_length = 0; #endif lws_pt_mutex_init(&context->pt[n]); #if defined(LWS_WITH_SEQUENCER) lws_seq_pt_init(&context->pt[n]); #endif LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) { if (ar->pt_init_destroy) ar->pt_init_destroy(context, info, &context->pt[n], 0); } LWS_FOR_EVERY_AVAILABLE_ROLE_END; #if defined(LWS_WITH_CGI) role_ops_cgi.pt_init_destroy(context, info, &context->pt[n], 0); #endif } lwsl_info(" Threads: %d each %d fds\n", context->count_threads, context->fd_limit_per_thread); if (!info->ka_interval && info->ka_time > 0) { lwsl_err("info->ka_interval can't be 0 if ka_time used\n"); return NULL; } #if defined(LWS_WITH_PEER_LIMITS) /* scale the peer hash table according to the max fds for the process, * so that the max list depth averages 16. Eg, 1024 fd -> 64, * 102400 fd -> 6400 */ context->pl_hash_elements = (context->count_threads * context->fd_limit_per_thread) / 16; context->pl_hash_table = lws_zalloc(sizeof(struct lws_peer *) * context->pl_hash_elements, "peer limits hash table"); context->ip_limit_ah = info->ip_limit_ah; context->ip_limit_wsi = info->ip_limit_wsi; #endif lwsl_info(" mem: context: %5lu B (%ld ctx + (%ld thr x %d))\n", (long)sizeof(struct lws_context) + (context->count_threads * context->pt_serv_buf_size), (long)sizeof(struct lws_context), (long)context->count_threads, context->pt_serv_buf_size); #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) lwsl_info(" mem: http hdr size: (%u + %lu), max count %u\n", context->max_http_header_data, (long)sizeof(struct allocated_headers), context->max_http_header_pool); #endif /* * fds table contains pollfd structs for as many pollfds as we can * handle... spread across as many service threads as we have going */ n = sizeof(struct lws_pollfd) * context->count_threads * context->fd_limit_per_thread; context->pt[0].fds = lws_zalloc(n, "fds table"); if (context->pt[0].fds == NULL) { lwsl_err("OOM allocating %d fds\n", context->max_fds); goto bail; } lwsl_info(" mem: pollfd map: %5u B\n", n); #endif #if defined(LWS_WITH_SERVER) if (info->server_string) { context->server_string = info->server_string; context->server_string_len = (short) strlen(context->server_string); } #endif #if LWS_MAX_SMP > 1 /* each thread serves his own chunk of fds */ for (n = 1; n < (int)context->count_threads; n++) context->pt[n].fds = context->pt[n - 1].fds + context->fd_limit_per_thread; #endif if (lws_plat_init(context, info)) goto bail; #if defined(LWS_WITH_NETWORK) if (context->event_loop_ops->init_context) if (context->event_loop_ops->init_context(context, info)) goto bail; if (context->event_loop_ops->init_pt) for (n = 0; n < context->count_threads; n++) { void *lp = NULL; if (info->foreign_loops) lp = info->foreign_loops[n]; if (context->event_loop_ops->init_pt(context, lp, n)) goto bail; } if (lws_create_event_pipes(context)) goto bail; #endif lws_context_init_ssl_library(info); context->user_space = info->user; #if defined(LWS_WITH_SERVER) strcpy(context->canonical_hostname, "unknown"); #if defined(LWS_WITH_NETWORK) lws_server_get_canonical_hostname(context, info); #endif #endif #if defined(LWS_WITH_STATS) context->pt[0].sul_stats.cb = lws_sul_stats_cb; __lws_sul_insert(&context->pt[0].pt_sul_owner, &context->pt[0].sul_stats, 10 * LWS_US_PER_SEC); #endif #if defined(LWS_WITH_PEER_LIMITS) context->pt[0].sul_peer_limits.cb = lws_sul_peer_limits_cb; __lws_sul_insert(&context->pt[0].pt_sul_owner, &context->pt[0].sul_peer_limits, 10 * LWS_US_PER_SEC); #endif #if defined(LWS_HAVE_SYS_CAPABILITY_H) && defined(LWS_HAVE_LIBCAP) memcpy(context->caps, info->caps, sizeof(context->caps)); context->count_caps = info->count_caps; #endif #if defined(LWS_WITH_NETWORK) #if defined(LWS_WITH_SYS_ASYNC_DNS) || defined(LWS_WITH_SYS_NTPCLIENT) || \ defined(LWS_WITH_SYS_DHCP_CLIENT) { /* * system vhost */ struct lws_context_creation_info ii; const struct lws_protocols *pp[4]; struct lws_vhost *vh; #if defined(LWS_WITH_SYS_ASYNC_DNS) extern const struct lws_protocols lws_async_dns_protocol; #endif #if defined(LWS_WITH_SYS_NTPCLIENT) extern const struct lws_protocols lws_system_protocol_ntpc; #endif #if defined(LWS_WITH_SYS_DHCP_CLIENT) extern const struct lws_protocols lws_system_protocol_dhcpc; #endif n = 0; #if defined(LWS_WITH_SYS_ASYNC_DNS) pp[n++] = &lws_async_dns_protocol; #endif #if defined(LWS_WITH_SYS_NTPCLIENT) pp[n++] = &lws_system_protocol_ntpc; #endif #if defined(LWS_WITH_SYS_DHCP_CLIENT) pp[n++] = &lws_system_protocol_dhcpc; #endif pp[n] = NULL; memset(&ii, 0, sizeof(ii)); ii.vhost_name = "system"; ii.pprotocols = pp; vh = lws_create_vhost(context, &ii); if (!vh) { lwsl_err("%s: failed to create system vhost\n", __func__); goto bail; } context->vhost_system = vh; if (lws_protocol_init_vhost(vh, NULL)) { lwsl_err("%s: failed to init system vhost\n", __func__); goto bail; } #if defined(LWS_WITH_SYS_ASYNC_DNS) if (lws_async_dns_init(context)) goto bail; #endif } #endif /* * init the lws_state mgr for the system state */ #if defined(_DEBUG) context->mgr_system.state_names = system_state_names; #endif context->mgr_system.name = "system"; context->mgr_system.state = LWS_SYSTATE_CONTEXT_CREATED; context->mgr_system.parent = context; context->protocols_notify.name = "prot_init"; context->protocols_notify.notify_cb = lws_state_notify_protocol_init; lws_state_reg_notifier(&context->mgr_system, &context->protocols_notify); /* * insert user notifiers here so they can participate with vetoing us * trying to jump straight to operational, or at least observe us * reaching 'operational', before we returned from context creation. */ lws_state_reg_notifier_list(&context->mgr_system, info->register_notifier_list); /* * if he's not saying he'll make his own vhosts later then act * compatibly and make a default vhost using the data in the info */ if (!lws_check_opt(info->options, LWS_SERVER_OPTION_EXPLICIT_VHOSTS)) if (!lws_create_vhost(context, info)) { lwsl_err("Failed to create default vhost\n"); #if defined(LWS_WITH_PEER_LIMITS) lws_free_set_NULL(context->pl_hash_table); #endif goto fail_clean_pipes; } #if defined(LWS_WITH_SECURE_STREAMS) if (context->pss_policies_json) { /* * You must create your context with the explicit vhosts flag * in order to use secure streams */ assert(lws_check_opt(info->options, LWS_SERVER_OPTION_EXPLICIT_VHOSTS)); if (lws_ss_policy_parse_begin(context)) goto bail; n = lws_ss_policy_parse(context, (uint8_t *)context->pss_policies_json, strlen(context->pss_policies_json)); if (n != LEJP_CONTINUE && n < 0) goto bail; if (lws_ss_policy_set(context, "hardcoded")) { lwsl_err("%s: policy set failed\n", __func__); goto bail; } } else lws_create_vhost(context, info); #endif lws_context_init_extensions(info, context); lwsl_info(" mem: per-conn: %5lu bytes + protocol rx buf\n", (unsigned long)sizeof(struct lws)); /* * drop any root privs for this process * to listen on port < 1023 we would have needed root, but now we are * listening, we don't want the power for anything else */ if (!lws_check_opt(info->options, LWS_SERVER_OPTION_EXPLICIT_VHOSTS)) if (lws_plat_drop_app_privileges(context, 1)) goto bail; /* * We want to move on the syste, state as far as it can go towards * OPERATIONAL now. But we have to return from here first so the user * code that called us can set its copy of context, which it may be * relying on to perform operations triggered by the state change. * * We set up a sul to come back immediately and do the state change. */ lws_sul_schedule(context, 0, &context->sul_system_state, lws_context_creation_completion_cb, 1); /* expedite post-context init (eg, protocols) */ lws_cancel_service(context); #endif return context; #if defined(LWS_WITH_NETWORK) fail_clean_pipes: for (n = 0; n < context->count_threads; n++) lws_destroy_event_pipe(context->pt[n].pipe_wsi); lws_free_set_NULL(context->pt[0].fds); lws_plat_context_late_destroy(context); lws_free_set_NULL(context); return NULL; #endif bail: lws_context_destroy(context); return NULL; #if defined(LWS_WITH_NETWORK) fail_event_libs: lwsl_err("Requested event library support not configured, available:\n"); { extern const struct lws_event_loop_ops *available_event_libs[]; const struct lws_event_loop_ops **elops = available_event_libs; while (*elops) { lwsl_err(" - %s\n", (*elops)->name); elops++; } } #endif lws_free(context); return NULL; } int lws_context_is_deprecated(struct lws_context *context) { return context->deprecated; } /* * When using an event loop, the context destruction is in three separate * parts. This is to cover both internal and foreign event loops cleanly. * * - lws_context_destroy() simply starts a soft close of all wsi and * related allocations. The event loop continues. * * As the closes complete in the event loop, reference counting is used * to determine when everything is closed. It then calls * lws_context_destroy2(). * * - lws_context_destroy2() cleans up the rest of the higher-level logical * lws pieces like vhosts. If the loop was foreign, it then proceeds to * lws_context_destroy3(). If it the loop is internal, it stops the * internal loops and waits for lws_context_destroy() to be called again * outside the event loop (since we cannot destroy the loop from * within the loop). That will cause lws_context_destroy3() to run * directly. * * - lws_context_destroy3() destroys any internal event loops and then * destroys the context itself, setting what was info.pcontext to NULL. */ /* * destroy the actual context itself */ static void lws_context_destroy3(struct lws_context *context) { struct lws_context **pcontext_finalize = context->pcontext_finalize; int n; #if defined(LWS_WITH_NETWORK) lwsl_debug("%s\n", __func__); for (n = 0; n < context->count_threads; n++) { struct lws_context_per_thread *pt = &context->pt[n]; (void)pt; #if defined(LWS_WITH_SEQUENCER) lws_seq_destroy_all_on_pt(pt); #endif LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) { if (ar->pt_init_destroy) ar->pt_init_destroy(context, NULL, pt, 1); } LWS_FOR_EVERY_AVAILABLE_ROLE_END; #if defined(LWS_WITH_CGI) role_ops_cgi.pt_init_destroy(context, NULL, pt, 1); #endif if (context->event_loop_ops->destroy_pt) context->event_loop_ops->destroy_pt(context, n); #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) while (pt->http.ah_list) _lws_destroy_ah(pt, pt->http.ah_list); #endif } #if defined(LWS_WITH_SYS_ASYNC_DNS) lws_async_dns_deinit(&context->async_dns); #endif #if defined(LWS_WITH_SYS_DHCP_CLIENT) lws_dhcpc_remove(context, NULL); #endif if (context->pt[0].fds) lws_free_set_NULL(context->pt[0].fds); #endif lws_context_deinit_ssl_library(context); #if defined(LWS_WITH_DETAILED_LATENCIES) if (context->latencies_fd != -1) compatible_close(context->latencies_fd); #endif for (n = 0; n < LWS_SYSBLOB_TYPE_COUNT; n++) lws_system_blob_destroy( lws_system_get_blob(context, n, 0)); lws_free(context); lwsl_info("%s: ctx %p freed\n", __func__, context); if (pcontext_finalize) *pcontext_finalize = NULL; } /* * really start destroying things */ void lws_context_destroy2(struct lws_context *context) { #if defined(LWS_WITH_NETWORK) struct lws_vhost *vh = NULL, *vh1; int n; #endif #if defined(LWS_WITH_PEER_LIMITS) uint32_t nu; #endif lwsl_info("%s: ctx %p\n", __func__, context); lws_context_lock(context, "context destroy 2"); /* ------ context { */ context->being_destroyed2 = 1; #if defined(LWS_WITH_NETWORK) /* * We're going to trash things like vhost-protocols * So we need to finish dealing with wsi close that * might make callbacks first */ for (n = 0; n < context->count_threads; n++) { struct lws_context_per_thread *pt = &context->pt[n]; (void)pt; #if defined(LWS_WITH_SECURE_STREAMS) lws_dll2_foreach_safe(&pt->ss_owner, NULL, lws_ss_destroy_dll); if (context->ac_policy) lwsac_free(&context->ac_policy); #endif #if defined(LWS_WITH_SECURE_STREAMS_PROXY_API) lws_dll2_foreach_safe(&pt->ss_client_owner, NULL, lws_sspc_destroy_dll); #endif #if defined(LWS_WITH_SEQUENCER) lws_seq_destroy_all_on_pt(pt); #endif LWS_FOR_EVERY_AVAILABLE_ROLE_START(ar) { if (ar->pt_init_destroy) ar->pt_init_destroy(context, NULL, pt, 1); } LWS_FOR_EVERY_AVAILABLE_ROLE_END; #if defined(LWS_WITH_CGI) role_ops_cgi.pt_init_destroy(context, NULL, pt, 1); #endif if (context->event_loop_ops->destroy_pt) context->event_loop_ops->destroy_pt(context, n); #if defined(LWS_ROLE_H1) || defined(LWS_ROLE_H2) while (pt->http.ah_list) _lws_destroy_ah(pt, pt->http.ah_list); #endif } /* * free all the per-vhost allocations */ vh = context->vhost_list; while (vh) { vh1 = vh->vhost_next; __lws_vhost_destroy2(vh); vh = vh1; } lwsl_debug("%p: post vh listl\n", __func__); /* remove ourselves from the pending destruction list */ while (context->vhost_pending_destruction_list) /* removes itself from list */ __lws_vhost_destroy2(context->vhost_pending_destruction_list); #endif lwsl_debug("%p: post pdl\n", __func__); lws_stats_log_dump(context); #if defined(LWS_WITH_NETWORK) lws_ssl_context_destroy(context); #endif lws_plat_context_late_destroy(context); #if defined(LWS_WITH_PEER_LIMITS) for (nu = 0; nu < context->pl_hash_elements; nu++) { lws_start_foreach_llp(struct lws_peer **, peer, context->pl_hash_table[nu]) { struct lws_peer *df = *peer; *peer = df->next; lws_free(df); continue; } lws_end_foreach_llp(peer, next); } lws_free(context->pl_hash_table); #endif lwsl_debug("%p: baggage\n", __func__); if (context->external_baggage_free_on_destroy) free(context->external_baggage_free_on_destroy); #if defined(LWS_WITH_NETWORK) lws_check_deferred_free(context, 0, 1); #endif #if LWS_MAX_SMP > 1 lws_mutex_refcount_destroy(&context->mr); #endif #if defined(LWS_WITH_NETWORK) if (context->event_loop_ops->destroy_context2) if (context->event_loop_ops->destroy_context2(context)) { lws_context_unlock(context); /* } context ----------- */ context->finalize_destroy_after_internal_loops_stopped = 1; return; } lwsl_debug("%p: post dc2\n", __func__); if (!context->pt[0].event_loop_foreign) { int n; for (n = 0; n < context->count_threads; n++) if (context->pt[n].inside_service) { lwsl_debug("%p: bailing as inside service\n", __func__); lws_context_unlock(context); /* } context --- */ return; } } #endif lws_context_unlock(context); /* } context ------------------- */ lws_context_destroy3(context); } #if defined(LWS_WITH_NETWORK) static void lws_pt_destroy(struct lws_context_per_thread *pt) { volatile struct lws_foreign_thread_pollfd *ftp, *next; volatile struct lws_context_per_thread *vpt; assert(!pt->is_destroyed); pt->destroy_self = 0; vpt = (volatile struct lws_context_per_thread *)pt; ftp = vpt->foreign_pfd_list; while (ftp) { next = ftp->next; lws_free((void *)ftp); ftp = next; } vpt->foreign_pfd_list = NULL; if (pt->pipe_wsi) lws_destroy_event_pipe(pt->pipe_wsi); pt->pipe_wsi = NULL; while (pt->fds_count) { struct lws *wsi = wsi_from_fd(pt->context, pt->fds[0].fd); if (!wsi) break; lws_close_free_wsi(wsi, LWS_CLOSE_STATUS_NOSTATUS_CONTEXT_DESTROY, "ctx destroy" /* no protocol close */); } lws_pt_mutex_destroy(pt); pt->is_destroyed = 1; lwsl_info("%s: pt destroyed\n", __func__); } #endif /* * Begin the context takedown */ void lws_context_destroy(struct lws_context *context) { #if defined(LWS_WITH_NETWORK) struct lws_vhost *vh = NULL; int m, deferred_pt = 0; #endif if (!context || context->inside_context_destroy) return; context->inside_context_destroy = 1; #if defined(LWS_WITH_NETWORK) if (context->finalize_destroy_after_internal_loops_stopped) { if (context->event_loop_ops->destroy_context2) context->event_loop_ops->destroy_context2(context); lws_context_destroy3(context); /* context is invalid, no need to reset inside flag */ return; } #endif if (context->being_destroyed1) { if (!context->being_destroyed2) { lws_context_destroy2(context); return; } lwsl_info("%s: ctx %p: already being destroyed\n", __func__, context); lws_context_destroy3(context); /* context is invalid, no need to reset inside flag */ return; } lwsl_info("%s: ctx %p\n", __func__, context); context->being_destroyed = 1; #if defined(LWS_WITH_NETWORK) lws_state_transition(&context->mgr_system, LWS_SYSTATE_POLICY_INVALID); m = context->count_threads; while (m--) { struct lws_context_per_thread *pt = &context->pt[m]; if (pt->is_destroyed) continue; if (pt->inside_lws_service) { pt->destroy_self = 1; deferred_pt = 1; continue; } lws_pt_destroy(pt); } if (deferred_pt) { lwsl_info("%s: waiting for deferred pt close\n", __func__); lws_cancel_service(context); goto out; } context->being_destroyed1 = 1; context->requested_kill = 1; /* * inform all the protocols that they are done and will have no more * callbacks. * * We can't free things until after the event loop shuts down. */ if (context->protocol_init_done) vh = context->vhost_list; while (vh) { struct lws_vhost *vhn = vh->vhost_next; lws_vhost_destroy1(vh); vh = vhn; } #endif lws_plat_context_early_destroy(context); #if defined(LWS_WITH_NETWORK) /* * We face two different needs depending if foreign loop or not. * * 1) If foreign loop, we really want to advance the destroy_context() * past here, and block only for libuv-style async close completion. * * 2a) If poll, and we exited by ourselves and are calling a final * destroy_context() outside of any service already, we want to * advance all the way in one step. * * 2b) If poll, and we are reacting to a SIGINT, service thread(s) may * be in poll wait or servicing. We can't advance the * destroy_context() to the point it's freeing things; we have to * leave that for the final destroy_context() after the service * thread(s) are finished calling for service. */ if (context->event_loop_ops->destroy_context1) { context->event_loop_ops->destroy_context1(context); goto out; } #endif #if defined(LWS_PLAT_FREERTOS) #if defined(LWS_AMAZON_RTOS) context->last_free_heap = xPortGetFreeHeapSize(); #else context->last_free_heap = esp_get_free_heap_size(); #endif #endif context->inside_context_destroy = 0; lws_context_destroy2(context); return; #if defined(LWS_WITH_NETWORK) out: context->inside_context_destroy = 0; #endif } struct lws_context * lws_system_context_from_system_mgr(lws_state_manager_t *mgr) { #if defined(LWS_WITH_NETWORK) return lws_container_of(mgr, struct lws_context, mgr_system); #else return NULL; #endif }