fa33e0f62d
License is LGPLv2.1 + static linking exception, version is 2.4.1
878 lines
23 KiB
C
878 lines
23 KiB
C
/*
|
|
* libwebsockets - small server side websockets and web server implementation
|
|
*
|
|
* Copyright (C) 2010-2017 Andy Green <andy@warmcat.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation:
|
|
* version 2.1 of the License.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
|
|
* MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "private-libwebsockets.h"
|
|
|
|
static int
|
|
lws_0405_frame_mask_generate(struct lws *wsi)
|
|
{
|
|
#if 0
|
|
wsi->u.ws.mask[0] = 0;
|
|
wsi->u.ws.mask[1] = 0;
|
|
wsi->u.ws.mask[2] = 0;
|
|
wsi->u.ws.mask[3] = 0;
|
|
#else
|
|
int n;
|
|
/* fetch the per-frame nonce */
|
|
|
|
n = lws_get_random(lws_get_context(wsi), wsi->u.ws.mask, 4);
|
|
if (n != 4) {
|
|
lwsl_parser("Unable to read from random device %s %d\n",
|
|
SYSTEM_RANDOM_FILEPATH, n);
|
|
return 1;
|
|
}
|
|
#endif
|
|
/* start masking from first byte of masking key buffer */
|
|
wsi->u.ws.mask_idx = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* notice this returns number of bytes consumed, or -1
|
|
*/
|
|
int lws_issue_raw(struct lws *wsi, unsigned char *buf, size_t len)
|
|
{
|
|
struct lws_context *context = lws_get_context(wsi);
|
|
struct lws_context_per_thread *pt = &wsi->context->pt[(int)wsi->tsi];
|
|
size_t real_len = len;
|
|
unsigned int n;
|
|
int m;
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_API_WRITE, 1);
|
|
|
|
if (!len)
|
|
return 0;
|
|
/* just ignore sends after we cleared the truncation buffer */
|
|
if (wsi->state == LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE &&
|
|
!wsi->trunc_len)
|
|
return len;
|
|
|
|
if (wsi->trunc_len && (buf < wsi->trunc_alloc ||
|
|
buf > (wsi->trunc_alloc + wsi->trunc_len + wsi->trunc_offset))) {
|
|
char dump[20];
|
|
strncpy(dump, (char *)buf, sizeof(dump) - 1);
|
|
dump[sizeof(dump) - 1] = '\0';
|
|
#if defined(LWS_WITH_ESP8266)
|
|
lwsl_err("****** %p: Sending new %lu (%s), pending truncated ...\n",
|
|
wsi, (unsigned long)len, dump);
|
|
#else
|
|
lwsl_err("****** %p: Sending new %lu (%s), pending truncated ...\n"
|
|
" It's illegal to do an lws_write outside of\n"
|
|
" the writable callback: fix your code\n",
|
|
wsi, (unsigned long)len, dump);
|
|
#endif
|
|
assert(0);
|
|
|
|
return -1;
|
|
}
|
|
|
|
m = lws_ext_cb_active(wsi, LWS_EXT_CB_PACKET_TX_DO_SEND, &buf, len);
|
|
if (m < 0)
|
|
return -1;
|
|
if (m) /* handled */ {
|
|
n = m;
|
|
goto handle_truncated_send;
|
|
}
|
|
|
|
if (!wsi->http2_substream && !lws_socket_is_valid(wsi->desc.sockfd))
|
|
lwsl_warn("** error invalid sock but expected to send\n");
|
|
|
|
/* limit sending */
|
|
if (wsi->protocol->tx_packet_size)
|
|
n = wsi->protocol->tx_packet_size;
|
|
else {
|
|
n = wsi->protocol->rx_buffer_size;
|
|
if (!n)
|
|
n = context->pt_serv_buf_size;
|
|
}
|
|
n += LWS_PRE + 4;
|
|
if (n > len)
|
|
n = len;
|
|
#if defined(LWS_WITH_ESP8266)
|
|
if (wsi->pending_send_completion) {
|
|
n = 0;
|
|
goto handle_truncated_send;
|
|
}
|
|
#endif
|
|
|
|
/* nope, send it on the socket directly */
|
|
lws_latency_pre(context, wsi);
|
|
n = lws_ssl_capable_write(wsi, buf, n);
|
|
lws_latency(context, wsi, "send lws_issue_raw", n, n == len);
|
|
|
|
switch (n) {
|
|
case LWS_SSL_CAPABLE_ERROR:
|
|
/* we're going to close, let close know sends aren't possible */
|
|
wsi->socket_is_permanently_unusable = 1;
|
|
return -1;
|
|
case LWS_SSL_CAPABLE_MORE_SERVICE:
|
|
/* nothing got sent, not fatal, retry the whole thing later */
|
|
n = 0;
|
|
break;
|
|
}
|
|
|
|
handle_truncated_send:
|
|
/*
|
|
* we were already handling a truncated send?
|
|
*/
|
|
if (wsi->trunc_len) {
|
|
lwsl_info("%p partial adv %d (vs %ld)\n", wsi, n, (long)real_len);
|
|
wsi->trunc_offset += n;
|
|
wsi->trunc_len -= n;
|
|
|
|
if (!wsi->trunc_len) {
|
|
lwsl_info("***** %p partial send completed\n", wsi);
|
|
/* done with it, but don't free it */
|
|
n = real_len;
|
|
if (wsi->state == LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE) {
|
|
lwsl_info("***** %p signalling to close now\n", wsi);
|
|
return -1; /* retry closing now */
|
|
}
|
|
}
|
|
/* always callback on writeable */
|
|
lws_callback_on_writable(wsi);
|
|
|
|
return n;
|
|
}
|
|
|
|
if ((unsigned int)n == real_len)
|
|
/* what we just sent went out cleanly */
|
|
return n;
|
|
|
|
/*
|
|
* Newly truncated send. Buffer the remainder (it will get
|
|
* first priority next time the socket is writable)
|
|
*/
|
|
lwsl_debug("%p new partial sent %d from %lu total\n", wsi, n,
|
|
(unsigned long)real_len);
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_WRITE_PARTIALS, 1);
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_B_PARTIALS_ACCEPTED_PARTS, n);
|
|
|
|
/*
|
|
* - if we still have a suitable malloc lying around, use it
|
|
* - or, if too small, reallocate it
|
|
* - or, if no buffer, create it
|
|
*/
|
|
if (!wsi->trunc_alloc || real_len - n > wsi->trunc_alloc_len) {
|
|
lws_free(wsi->trunc_alloc);
|
|
|
|
wsi->trunc_alloc_len = real_len - n;
|
|
wsi->trunc_alloc = lws_malloc(real_len - n, "truncated send alloc");
|
|
if (!wsi->trunc_alloc) {
|
|
lwsl_err("truncated send: unable to malloc %lu\n",
|
|
(unsigned long)(real_len - n));
|
|
return -1;
|
|
}
|
|
}
|
|
wsi->trunc_offset = 0;
|
|
wsi->trunc_len = real_len - n;
|
|
memcpy(wsi->trunc_alloc, buf + n, real_len - n);
|
|
|
|
/* since something buffered, force it to get another chance to send */
|
|
lws_callback_on_writable(wsi);
|
|
|
|
return real_len;
|
|
}
|
|
|
|
LWS_VISIBLE int lws_write(struct lws *wsi, unsigned char *buf, size_t len,
|
|
enum lws_write_protocol wp)
|
|
{
|
|
struct lws_context_per_thread *pt = &wsi->context->pt[(int)wsi->tsi];
|
|
int masked7 = (wsi->mode == LWSCM_WS_CLIENT);
|
|
unsigned char is_masked_bit = 0;
|
|
unsigned char *dropmask = NULL;
|
|
struct lws_tokens eff_buf;
|
|
size_t orig_len = len;
|
|
int pre = 0, n;
|
|
|
|
if (wsi->parent_carries_io) {
|
|
struct lws_write_passthru pas;
|
|
|
|
pas.buf = buf;
|
|
pas.len = len;
|
|
pas.wp = wp;
|
|
pas.wsi = wsi;
|
|
|
|
if (wsi->parent->protocol->callback(wsi->parent,
|
|
LWS_CALLBACK_CHILD_WRITE_VIA_PARENT,
|
|
wsi->parent->user_space,
|
|
(void *)&pas, 0))
|
|
return 1;
|
|
|
|
return len;
|
|
}
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_API_LWS_WRITE, 1);
|
|
|
|
if ((int)len < 0) {
|
|
lwsl_err("%s: suspicious len int %d, ulong %lu\n", __func__,
|
|
(int)len, (unsigned long)len);
|
|
return -1;
|
|
}
|
|
|
|
lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_B_WRITE, len);
|
|
|
|
#ifdef LWS_WITH_ACCESS_LOG
|
|
wsi->access_log.sent += len;
|
|
#endif
|
|
if (wsi->vhost)
|
|
wsi->vhost->conn_stats.tx += len;
|
|
|
|
if (wsi->state == LWSS_ESTABLISHED && wsi->u.ws.tx_draining_ext) {
|
|
/* remove us from the list */
|
|
struct lws **w = &pt->tx_draining_ext_list;
|
|
|
|
wsi->u.ws.tx_draining_ext = 0;
|
|
/* remove us from context draining ext list */
|
|
while (*w) {
|
|
if (*w == wsi) {
|
|
*w = wsi->u.ws.tx_draining_ext_list;
|
|
break;
|
|
}
|
|
w = &((*w)->u.ws.tx_draining_ext_list);
|
|
}
|
|
wsi->u.ws.tx_draining_ext_list = NULL;
|
|
wp = (wsi->u.ws.tx_draining_stashed_wp & 0xc0) |
|
|
LWS_WRITE_CONTINUATION;
|
|
|
|
lwsl_ext("FORCED draining wp to 0x%02X\n", wp);
|
|
}
|
|
|
|
lws_restart_ws_ping_pong_timer(wsi);
|
|
|
|
if ((wp & 0x1f) == LWS_WRITE_HTTP ||
|
|
(wp & 0x1f) == LWS_WRITE_HTTP_FINAL ||
|
|
(wp & 0x1f) == LWS_WRITE_HTTP_HEADERS_CONTINUATION ||
|
|
(wp & 0x1f) == LWS_WRITE_HTTP_HEADERS)
|
|
goto send_raw;
|
|
|
|
/* if not in a state to send stuff, then just send nothing */
|
|
|
|
if (wsi->state != LWSS_ESTABLISHED &&
|
|
((wsi->state != LWSS_RETURNED_CLOSE_ALREADY &&
|
|
wsi->state != LWSS_AWAITING_CLOSE_ACK) ||
|
|
wp != LWS_WRITE_CLOSE))
|
|
return 0;
|
|
|
|
/* if we are continuing a frame that already had its header done */
|
|
|
|
if (wsi->u.ws.inside_frame) {
|
|
lwsl_debug("INSIDE FRAME\n");
|
|
goto do_more_inside_frame;
|
|
}
|
|
|
|
wsi->u.ws.clean_buffer = 1;
|
|
|
|
/*
|
|
* give a chance to the extensions to modify payload
|
|
* the extension may decide to produce unlimited payload erratically
|
|
* (eg, compression extension), so we require only that if he produces
|
|
* something, it will be a complete fragment of the length known at
|
|
* the time (just the fragment length known), and if he has
|
|
* more we will come back next time he is writeable and allow him to
|
|
* produce more fragments until he's drained.
|
|
*
|
|
* This allows what is sent each time it is writeable to be limited to
|
|
* a size that can be sent without partial sends or blocking, allows
|
|
* interleaving of control frames and other connection service.
|
|
*/
|
|
eff_buf.token = (char *)buf;
|
|
eff_buf.token_len = len;
|
|
|
|
switch ((int)wp) {
|
|
case LWS_WRITE_PING:
|
|
case LWS_WRITE_PONG:
|
|
case LWS_WRITE_CLOSE:
|
|
break;
|
|
default:
|
|
lwsl_debug("LWS_EXT_CB_PAYLOAD_TX\n");
|
|
n = lws_ext_cb_active(wsi, LWS_EXT_CB_PAYLOAD_TX, &eff_buf, wp);
|
|
if (n < 0)
|
|
return -1;
|
|
|
|
if (n && eff_buf.token_len) {
|
|
lwsl_debug("drain len %d\n", (int)eff_buf.token_len);
|
|
/* extension requires further draining */
|
|
wsi->u.ws.tx_draining_ext = 1;
|
|
wsi->u.ws.tx_draining_ext_list = pt->tx_draining_ext_list;
|
|
pt->tx_draining_ext_list = wsi;
|
|
/* we must come back to do more */
|
|
lws_callback_on_writable(wsi);
|
|
/*
|
|
* keep a copy of the write type for the overall
|
|
* action that has provoked generation of these
|
|
* fragments, so the last guy can use its FIN state.
|
|
*/
|
|
wsi->u.ws.tx_draining_stashed_wp = wp;
|
|
/* this is definitely not actually the last fragment
|
|
* because the extension asserted he has more coming
|
|
* So make sure this intermediate one doesn't go out
|
|
* with a FIN.
|
|
*/
|
|
wp |= LWS_WRITE_NO_FIN;
|
|
}
|
|
|
|
if (eff_buf.token_len && wsi->u.ws.stashed_write_pending) {
|
|
wsi->u.ws.stashed_write_pending = 0;
|
|
wp = (wp &0xc0) | (int)wsi->u.ws.stashed_write_type;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* an extension did something we need to keep... for example, if
|
|
* compression extension, it has already updated its state according
|
|
* to this being issued
|
|
*/
|
|
if ((char *)buf != eff_buf.token) {
|
|
/*
|
|
* ext might eat it, but not have anything to issue yet.
|
|
* In that case we have to follow his lead, but stash and
|
|
* replace the write type that was lost here the first time.
|
|
*/
|
|
if (len && !eff_buf.token_len) {
|
|
if (!wsi->u.ws.stashed_write_pending)
|
|
wsi->u.ws.stashed_write_type = (char)wp & 0x3f;
|
|
wsi->u.ws.stashed_write_pending = 1;
|
|
return len;
|
|
}
|
|
/*
|
|
* extension recreated it:
|
|
* need to buffer this if not all sent
|
|
*/
|
|
wsi->u.ws.clean_buffer = 0;
|
|
}
|
|
|
|
buf = (unsigned char *)eff_buf.token;
|
|
len = eff_buf.token_len;
|
|
|
|
if (!buf) {
|
|
lwsl_err("null buf (%d)\n", (int)len);
|
|
return -1;
|
|
}
|
|
|
|
switch (wsi->ietf_spec_revision) {
|
|
case 13:
|
|
if (masked7) {
|
|
pre += 4;
|
|
dropmask = &buf[0 - pre];
|
|
is_masked_bit = 0x80;
|
|
}
|
|
|
|
switch (wp & 0xf) {
|
|
case LWS_WRITE_TEXT:
|
|
n = LWSWSOPC_TEXT_FRAME;
|
|
break;
|
|
case LWS_WRITE_BINARY:
|
|
n = LWSWSOPC_BINARY_FRAME;
|
|
break;
|
|
case LWS_WRITE_CONTINUATION:
|
|
n = LWSWSOPC_CONTINUATION;
|
|
break;
|
|
|
|
case LWS_WRITE_CLOSE:
|
|
n = LWSWSOPC_CLOSE;
|
|
break;
|
|
case LWS_WRITE_PING:
|
|
n = LWSWSOPC_PING;
|
|
break;
|
|
case LWS_WRITE_PONG:
|
|
n = LWSWSOPC_PONG;
|
|
break;
|
|
default:
|
|
lwsl_warn("lws_write: unknown write opc / wp\n");
|
|
return -1;
|
|
}
|
|
|
|
if (!(wp & LWS_WRITE_NO_FIN))
|
|
n |= 1 << 7;
|
|
|
|
if (len < 126) {
|
|
pre += 2;
|
|
buf[-pre] = n;
|
|
buf[-pre + 1] = (unsigned char)(len | is_masked_bit);
|
|
} else {
|
|
if (len < 65536) {
|
|
pre += 4;
|
|
buf[-pre] = n;
|
|
buf[-pre + 1] = 126 | is_masked_bit;
|
|
buf[-pre + 2] = (unsigned char)(len >> 8);
|
|
buf[-pre + 3] = (unsigned char)len;
|
|
} else {
|
|
pre += 10;
|
|
buf[-pre] = n;
|
|
buf[-pre + 1] = 127 | is_masked_bit;
|
|
#if defined __LP64__
|
|
buf[-pre + 2] = (len >> 56) & 0x7f;
|
|
buf[-pre + 3] = len >> 48;
|
|
buf[-pre + 4] = len >> 40;
|
|
buf[-pre + 5] = len >> 32;
|
|
#else
|
|
buf[-pre + 2] = 0;
|
|
buf[-pre + 3] = 0;
|
|
buf[-pre + 4] = 0;
|
|
buf[-pre + 5] = 0;
|
|
#endif
|
|
buf[-pre + 6] = (unsigned char)(len >> 24);
|
|
buf[-pre + 7] = (unsigned char)(len >> 16);
|
|
buf[-pre + 8] = (unsigned char)(len >> 8);
|
|
buf[-pre + 9] = (unsigned char)len;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
do_more_inside_frame:
|
|
|
|
/*
|
|
* Deal with masking if we are in client -> server direction and
|
|
* the wp demands it
|
|
*/
|
|
|
|
if (masked7) {
|
|
if (!wsi->u.ws.inside_frame)
|
|
if (lws_0405_frame_mask_generate(wsi)) {
|
|
lwsl_err("frame mask generation failed\n");
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* in v7, just mask the payload
|
|
*/
|
|
if (dropmask) { /* never set if already inside frame */
|
|
for (n = 4; n < (int)len + 4; n++)
|
|
dropmask[n] = dropmask[n] ^ wsi->u.ws.mask[
|
|
(wsi->u.ws.mask_idx++) & 3];
|
|
|
|
/* copy the frame nonce into place */
|
|
memcpy(dropmask, wsi->u.ws.mask, 4);
|
|
}
|
|
}
|
|
|
|
send_raw:
|
|
switch ((int)(wp & 0x1f)) {
|
|
case LWS_WRITE_CLOSE:
|
|
/* lwsl_hexdump(&buf[-pre], len); */
|
|
case LWS_WRITE_HTTP:
|
|
case LWS_WRITE_HTTP_FINAL:
|
|
case LWS_WRITE_HTTP_HEADERS:
|
|
case LWS_WRITE_HTTP_HEADERS_CONTINUATION:
|
|
case LWS_WRITE_PONG:
|
|
case LWS_WRITE_PING:
|
|
#ifdef LWS_WITH_HTTP2
|
|
if (wsi->mode == LWSCM_HTTP2_SERVING) {
|
|
unsigned char flags = 0;
|
|
|
|
n = LWS_H2_FRAME_TYPE_DATA;
|
|
if ((wp & 0x1f) == LWS_WRITE_HTTP_HEADERS) {
|
|
n = LWS_H2_FRAME_TYPE_HEADERS;
|
|
if (!(wp & LWS_WRITE_NO_FIN))
|
|
flags = LWS_H2_FLAG_END_HEADERS;
|
|
if (wsi->u.h2.send_END_STREAM || (wp & LWS_WRITE_H2_STREAM_END)) {
|
|
flags |= LWS_H2_FLAG_END_STREAM;
|
|
wsi->u.h2.send_END_STREAM = 1;
|
|
}
|
|
}
|
|
|
|
if ((wp & 0x1f) == LWS_WRITE_HTTP_HEADERS_CONTINUATION) {
|
|
n = LWS_H2_FRAME_TYPE_CONTINUATION;
|
|
if (!(wp & LWS_WRITE_NO_FIN))
|
|
flags = LWS_H2_FLAG_END_HEADERS;
|
|
if (wsi->u.h2.send_END_STREAM || (wp & LWS_WRITE_H2_STREAM_END)) {
|
|
flags |= LWS_H2_FLAG_END_STREAM;
|
|
wsi->u.h2.send_END_STREAM = 1;
|
|
}
|
|
}
|
|
|
|
if (((wp & 0x1f) == LWS_WRITE_HTTP ||
|
|
(wp & 0x1f) == LWS_WRITE_HTTP_FINAL) &&
|
|
wsi->u.http.tx_content_length) {
|
|
wsi->u.http.tx_content_remain -= len;
|
|
lwsl_info("%s: content_remain = %llu\n", __func__,
|
|
(unsigned long long)wsi->u.http.tx_content_remain);
|
|
if (!wsi->u.http.tx_content_remain) {
|
|
lwsl_info("%s: selecting final write mode\n", __func__);
|
|
wp = LWS_WRITE_HTTP_FINAL;
|
|
}
|
|
}
|
|
|
|
if ((wp & 0x1f) == LWS_WRITE_HTTP_FINAL || (wp & LWS_WRITE_H2_STREAM_END)) {
|
|
//lws_get_network_wsi(wsi)->u.h2.END_STREAM) {
|
|
lwsl_info("%s: setting END_STREAM\n", __func__);
|
|
flags |= LWS_H2_FLAG_END_STREAM;
|
|
wsi->u.h2.send_END_STREAM = 1;
|
|
}
|
|
|
|
return lws_h2_frame_write(wsi, n, flags,
|
|
wsi->u.h2.my_sid, len, buf);
|
|
}
|
|
#endif
|
|
return lws_issue_raw(wsi, (unsigned char *)buf - pre, len + pre);
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* give any active extensions a chance to munge the buffer
|
|
* before send. We pass in a pointer to an lws_tokens struct
|
|
* prepared with the default buffer and content length that's in
|
|
* there. Rather than rewrite the default buffer, extensions
|
|
* that expect to grow the buffer can adapt .token to
|
|
* point to their own per-connection buffer in the extension
|
|
* user allocation. By default with no extensions or no
|
|
* extension callback handling, just the normal input buffer is
|
|
* used then so it is efficient.
|
|
*
|
|
* callback returns 1 in case it wants to spill more buffers
|
|
*
|
|
* This takes care of holding the buffer if send is incomplete, ie,
|
|
* if wsi->u.ws.clean_buffer is 0 (meaning an extension meddled with
|
|
* the buffer). If wsi->u.ws.clean_buffer is 1, it will instead
|
|
* return to the user code how much OF THE USER BUFFER was consumed.
|
|
*/
|
|
|
|
n = lws_issue_raw_ext_access(wsi, buf - pre, len + pre);
|
|
wsi->u.ws.inside_frame = 1;
|
|
if (n <= 0)
|
|
return n;
|
|
|
|
if (n == (int)len + pre) {
|
|
/* everything in the buffer was handled (or rebuffered...) */
|
|
wsi->u.ws.inside_frame = 0;
|
|
return orig_len;
|
|
}
|
|
|
|
/*
|
|
* it is how many bytes of user buffer got sent... may be < orig_len
|
|
* in which case callback when writable has already been arranged
|
|
* and user code can call lws_write() again with the rest
|
|
* later.
|
|
*/
|
|
|
|
return n - pre;
|
|
}
|
|
|
|
LWS_VISIBLE int lws_serve_http_file_fragment(struct lws *wsi)
|
|
{
|
|
struct lws_context *context = wsi->context;
|
|
struct lws_context_per_thread *pt = &context->pt[(int)wsi->tsi];
|
|
struct lws_process_html_args args;
|
|
lws_filepos_t amount, poss;
|
|
unsigned char *p, *pstart;
|
|
#if defined(LWS_WITH_RANGES)
|
|
unsigned char finished = 0;
|
|
#endif
|
|
int n, m;
|
|
|
|
lwsl_debug("wsi->http2_substream %d\n", wsi->http2_substream);
|
|
|
|
while (!lws_send_pipe_choked(wsi)) {
|
|
|
|
if (wsi->trunc_len) {
|
|
if (lws_issue_raw(wsi, wsi->trunc_alloc +
|
|
wsi->trunc_offset,
|
|
wsi->trunc_len) < 0) {
|
|
lwsl_info("%s: closing\n", __func__);
|
|
goto file_had_it;
|
|
}
|
|
continue;
|
|
}
|
|
|
|
if (wsi->u.http.filepos == wsi->u.http.filelen)
|
|
goto all_sent;
|
|
|
|
n = 0;
|
|
|
|
pstart = pt->serv_buf + LWS_H2_FRAME_HEADER_LENGTH;
|
|
|
|
p = pstart;
|
|
|
|
#if defined(LWS_WITH_RANGES)
|
|
if (wsi->u.http.range.count_ranges && !wsi->u.http.range.inside) {
|
|
|
|
lwsl_notice("%s: doing range start %llu\n", __func__, wsi->u.http.range.start);
|
|
|
|
if ((long long)lws_vfs_file_seek_cur(wsi->u.http.fop_fd,
|
|
wsi->u.http.range.start -
|
|
wsi->u.http.filepos) < 0)
|
|
goto file_had_it;
|
|
|
|
wsi->u.http.filepos = wsi->u.http.range.start;
|
|
|
|
if (wsi->u.http.range.count_ranges > 1) {
|
|
n = lws_snprintf((char *)p, context->pt_serv_buf_size - LWS_H2_FRAME_HEADER_LENGTH,
|
|
"_lws\x0d\x0a"
|
|
"Content-Type: %s\x0d\x0a"
|
|
"Content-Range: bytes %llu-%llu/%llu\x0d\x0a"
|
|
"\x0d\x0a",
|
|
wsi->u.http.multipart_content_type,
|
|
wsi->u.http.range.start,
|
|
wsi->u.http.range.end,
|
|
wsi->u.http.range.extent);
|
|
p += n;
|
|
}
|
|
|
|
wsi->u.http.range.budget = wsi->u.http.range.end -
|
|
wsi->u.http.range.start + 1;
|
|
wsi->u.http.range.inside = 1;
|
|
}
|
|
#endif
|
|
|
|
poss = context->pt_serv_buf_size - n - LWS_H2_FRAME_HEADER_LENGTH;
|
|
|
|
/*
|
|
* if there is a hint about how much we will do well to send at one time,
|
|
* restrict ourselves to only trying to send that.
|
|
*/
|
|
if (wsi->protocol->tx_packet_size &&
|
|
poss > wsi->protocol->tx_packet_size)
|
|
poss = wsi->protocol->tx_packet_size;
|
|
|
|
#if defined(LWS_WITH_HTTP2)
|
|
m = lws_h2_tx_cr_get(wsi);
|
|
if (!m) {
|
|
lwsl_info("%s: came here with no tx credit", __func__);
|
|
return 0;
|
|
}
|
|
if (m < poss)
|
|
poss = m;
|
|
/*
|
|
* consumption of the actual payload amount sent will be handled
|
|
* when the http2 data frame is sent
|
|
*/
|
|
#endif
|
|
|
|
#if defined(LWS_WITH_RANGES)
|
|
if (wsi->u.http.range.count_ranges) {
|
|
if (wsi->u.http.range.count_ranges > 1)
|
|
poss -= 7; /* allow for final boundary */
|
|
if (poss > wsi->u.http.range.budget)
|
|
poss = wsi->u.http.range.budget;
|
|
}
|
|
#endif
|
|
if (wsi->sending_chunked) {
|
|
/* we need to drop the chunk size in here */
|
|
p += 10;
|
|
/* allow for the chunk to grow by 128 in translation */
|
|
poss -= 10 + 128;
|
|
}
|
|
|
|
if (lws_vfs_file_read(wsi->u.http.fop_fd, &amount, p, poss) < 0)
|
|
goto file_had_it; /* caller will close */
|
|
|
|
if (wsi->sending_chunked)
|
|
n = (int)amount;
|
|
else
|
|
n = (p - pstart) + (int)amount;
|
|
|
|
lwsl_debug("%s: sending %d\n", __func__, n);
|
|
|
|
if (n) {
|
|
lws_set_timeout(wsi, PENDING_TIMEOUT_HTTP_CONTENT,
|
|
context->timeout_secs);
|
|
|
|
if (wsi->sending_chunked) {
|
|
args.p = (char *)p;
|
|
args.len = n;
|
|
args.max_len = (unsigned int)poss + 128;
|
|
args.final = wsi->u.http.filepos + n ==
|
|
wsi->u.http.filelen;
|
|
if (user_callback_handle_rxflow(
|
|
wsi->vhost->protocols[(int)wsi->protocol_interpret_idx].callback, wsi,
|
|
LWS_CALLBACK_PROCESS_HTML,
|
|
wsi->user_space, &args, 0) < 0)
|
|
goto file_had_it;
|
|
n = args.len;
|
|
p = (unsigned char *)args.p;
|
|
} else
|
|
p = pstart;
|
|
|
|
#if defined(LWS_WITH_RANGES)
|
|
if (wsi->u.http.range.send_ctr + 1 ==
|
|
wsi->u.http.range.count_ranges && // last range
|
|
wsi->u.http.range.count_ranges > 1 && // was 2+ ranges (ie, multipart)
|
|
wsi->u.http.range.budget - amount == 0) {// final part
|
|
n += lws_snprintf((char *)pstart + n, 6,
|
|
"_lws\x0d\x0a"); // append trailing boundary
|
|
lwsl_debug("added trailing boundary\n");
|
|
}
|
|
#endif
|
|
m = lws_write(wsi, p, n,
|
|
wsi->u.http.filepos == wsi->u.http.filelen ?
|
|
LWS_WRITE_HTTP_FINAL :
|
|
LWS_WRITE_HTTP
|
|
);
|
|
if (m < 0)
|
|
goto file_had_it;
|
|
|
|
wsi->u.http.filepos += amount;
|
|
|
|
#if defined(LWS_WITH_RANGES)
|
|
if (wsi->u.http.range.count_ranges >= 1) {
|
|
wsi->u.http.range.budget -= amount;
|
|
if (wsi->u.http.range.budget == 0) {
|
|
lwsl_notice("range budget exhausted\n");
|
|
wsi->u.http.range.inside = 0;
|
|
wsi->u.http.range.send_ctr++;
|
|
|
|
if (lws_ranges_next(&wsi->u.http.range) < 1) {
|
|
finished = 1;
|
|
goto all_sent;
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
if (m != n) {
|
|
/* adjust for what was not sent */
|
|
if (lws_vfs_file_seek_cur(wsi->u.http.fop_fd,
|
|
m - n) ==
|
|
(unsigned long)-1)
|
|
goto file_had_it;
|
|
}
|
|
}
|
|
|
|
all_sent:
|
|
if ((!wsi->trunc_len && wsi->u.http.filepos >= wsi->u.http.filelen)
|
|
#if defined(LWS_WITH_RANGES)
|
|
|| finished)
|
|
#else
|
|
)
|
|
#endif
|
|
{
|
|
wsi->state = LWSS_HTTP;
|
|
/* we might be in keepalive, so close it off here */
|
|
lws_vfs_file_close(&wsi->u.http.fop_fd);
|
|
|
|
lwsl_debug("file completed\n");
|
|
|
|
if (wsi->protocol->callback &&
|
|
user_callback_handle_rxflow(wsi->protocol->callback,
|
|
wsi, LWS_CALLBACK_HTTP_FILE_COMPLETION,
|
|
wsi->user_space, NULL,
|
|
0) < 0) {
|
|
/*
|
|
* For http/1.x, the choices from
|
|
* transaction_completed are either
|
|
* 0 to use the connection for pipelined
|
|
* or nonzero to hang it up.
|
|
*
|
|
* However for http/2. while we are
|
|
* still interested in hanging up the
|
|
* nwsi if there was a network-level
|
|
* fatal error, simply completing the
|
|
* transaction is a matter of the stream
|
|
* state, not the root connection at the
|
|
* network level
|
|
*/
|
|
if (wsi->http2_substream)
|
|
return 1;
|
|
else
|
|
return -1;
|
|
}
|
|
|
|
return 1; /* >0 indicates completed */
|
|
}
|
|
}
|
|
|
|
lws_callback_on_writable(wsi);
|
|
|
|
return 0; /* indicates further processing must be done */
|
|
|
|
file_had_it:
|
|
lws_vfs_file_close(&wsi->u.http.fop_fd);
|
|
|
|
return -1;
|
|
}
|
|
|
|
#if LWS_POSIX
|
|
LWS_VISIBLE int
|
|
lws_ssl_capable_read_no_ssl(struct lws *wsi, unsigned char *buf, int len)
|
|
{
|
|
struct lws_context *context = wsi->context;
|
|
struct lws_context_per_thread *pt = &context->pt[(int)wsi->tsi];
|
|
int n;
|
|
|
|
lws_stats_atomic_bump(context, pt, LWSSTATS_C_API_READ, 1);
|
|
|
|
n = recv(wsi->desc.sockfd, (char *)buf, len, 0);
|
|
if (n >= 0) {
|
|
if (wsi->vhost)
|
|
wsi->vhost->conn_stats.rx += n;
|
|
lws_stats_atomic_bump(context, pt, LWSSTATS_B_READ, n);
|
|
lws_restart_ws_ping_pong_timer(wsi);
|
|
return n;
|
|
}
|
|
#if LWS_POSIX
|
|
if (LWS_ERRNO == LWS_EAGAIN ||
|
|
LWS_ERRNO == LWS_EWOULDBLOCK ||
|
|
LWS_ERRNO == LWS_EINTR)
|
|
return LWS_SSL_CAPABLE_MORE_SERVICE;
|
|
#endif
|
|
lwsl_notice("error on reading from skt : %d\n", LWS_ERRNO);
|
|
return LWS_SSL_CAPABLE_ERROR;
|
|
}
|
|
|
|
LWS_VISIBLE int
|
|
lws_ssl_capable_write_no_ssl(struct lws *wsi, unsigned char *buf, int len)
|
|
{
|
|
int n = 0;
|
|
|
|
#if LWS_POSIX
|
|
n = send(wsi->desc.sockfd, (char *)buf, len, MSG_NOSIGNAL);
|
|
// lwsl_info("%s: sent len %d result %d", __func__, len, n);
|
|
if (n >= 0)
|
|
return n;
|
|
|
|
if (LWS_ERRNO == LWS_EAGAIN ||
|
|
LWS_ERRNO == LWS_EWOULDBLOCK ||
|
|
LWS_ERRNO == LWS_EINTR) {
|
|
if (LWS_ERRNO == LWS_EWOULDBLOCK) {
|
|
lws_set_blocking_send(wsi);
|
|
}
|
|
|
|
return LWS_SSL_CAPABLE_MORE_SERVICE;
|
|
}
|
|
#else
|
|
(void)n;
|
|
(void)wsi;
|
|
(void)buf;
|
|
(void)len;
|
|
// !!!
|
|
#endif
|
|
|
|
lwsl_debug("ERROR writing len %d to skt fd %d err %d / errno %d\n",
|
|
len, wsi->desc.sockfd, n, LWS_ERRNO);
|
|
return LWS_SSL_CAPABLE_ERROR;
|
|
}
|
|
#endif
|
|
LWS_VISIBLE int
|
|
lws_ssl_pending_no_ssl(struct lws *wsi)
|
|
{
|
|
(void)wsi;
|
|
#if defined(LWS_WITH_ESP32)
|
|
return 100;
|
|
#else
|
|
return 0;
|
|
#endif
|
|
}
|