hio/lib/http-thr.c

1027 lines
33 KiB
C
Raw Normal View History

2020-05-25 08:04:30 +00:00
/*
Copyright (c) 2016-2020 Chung, Hyung-Hwan. All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in the
documentation and/or other materials provided with the distribution.
THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR
2022-06-11 05:32:01 +00:00
IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
2020-05-25 08:04:30 +00:00
OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
2020-05-25 08:04:30 +00:00
#include "http-prv.h"
2021-07-22 07:30:20 +00:00
#include <hio-thr.h>
#include <hio-fmt.h>
#include <hio-chr.h>
2020-05-25 08:04:30 +00:00
#include <pthread.h>
#define THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH
enum thr_res_mode_t
{
THR_RES_MODE_CHUNKED,
THR_RES_MODE_CLOSE,
THR_RES_MODE_LENGTH
};
typedef enum thr_res_mode_t thr_res_mode_t;
#define THR_PENDING_IO_THRESHOLD 5
#define THR_OVER_READ_FROM_CLIENT (1 << 0)
#define THR_OVER_READ_FROM_PEER (1 << 1)
#define THR_OVER_WRITE_TO_CLIENT (1 << 2)
#define THR_OVER_WRITE_TO_PEER (1 << 3)
#define THR_OVER_ALL (THR_OVER_READ_FROM_CLIENT | THR_OVER_READ_FROM_PEER | THR_OVER_WRITE_TO_CLIENT | THR_OVER_WRITE_TO_PEER)
struct thr_func_start_t
{
2023-01-11 15:33:52 +00:00
hio_t* hio; /* for faster and safer access in case htts has been already destroyed */
hio_svc_htts_t* htts;
2021-07-22 07:30:20 +00:00
hio_svc_htts_thr_func_t thr_func;
void* thr_ctx;
2021-07-22 07:30:20 +00:00
hio_svc_htts_thr_func_info_t tfi;
};
typedef struct thr_func_start_t thr_func_start_t;
struct thr_task_t
2020-05-25 08:04:30 +00:00
{
HIO_SVC_HTTS_TASK_HEADER;
2020-05-25 08:04:30 +00:00
hio_svc_htts_task_on_kill_t on_kill; /* user-provided on_kill callback */
int options;
2021-07-22 07:30:20 +00:00
hio_oow_t num_pending_writes_to_client;
hio_oow_t num_pending_writes_to_peer;
hio_dev_thr_t* peer;
hio_htrd_t* peer_htrd;
unsigned int over: 4; /* must be large enough to accomodate THR_OVER_ALL */
unsigned int req_content_length_unlimited: 1;
unsigned int ever_attempted_to_write_to_client: 1;
unsigned int client_eof_detected: 1;
unsigned int client_disconnected: 1;
2020-05-26 14:58:55 +00:00
unsigned int client_htrd_recbs_changed: 1;
2021-07-22 07:30:20 +00:00
hio_oow_t req_content_length; /* client request content length */
thr_res_mode_t res_mode_to_cli;
2021-07-22 07:30:20 +00:00
hio_dev_sck_on_read_t client_org_on_read;
hio_dev_sck_on_write_t client_org_on_write;
hio_dev_sck_on_disconnect_t client_org_on_disconnect;
hio_htrd_recbs_t client_htrd_org_recbs;
};
typedef struct thr_task_t thr_task_t;
2020-05-25 08:04:30 +00:00
struct thr_peer_xtn_t
{
thr_task_t* task;
};
typedef struct thr_peer_xtn_t thr_peer_xtn_t;
2023-03-04 16:13:19 +00:00
static void thr_task_halt_participating_devices (thr_task_t* thr)
{
HIO_DEBUG4 (thr->htts->hio, "HTTS(%p) - Halting participating devices in thr task %p(csck=%p,peer=%p)\n", thr->htts, thr, thr->task_csck, thr->peer);
if (thr->task_csck) hio_dev_sck_halt (thr->task_csck);
/* check for peer as it may not have been started */
2023-03-04 16:13:19 +00:00
if (thr->peer) hio_dev_thr_halt (thr->peer);
}
static int thr_write_to_client (thr_task_t* thr, const void* data, hio_iolen_t dlen)
{
if (thr->task_csck)
{
2023-03-04 16:13:19 +00:00
thr->ever_attempted_to_write_to_client = 1;
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_client++;
if (hio_dev_sck_write(thr->task_csck, data, dlen, HIO_NULL, HIO_NULL) <= -1)
{
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_client--;
return -1;
}
if (thr->num_pending_writes_to_client > THR_PENDING_IO_THRESHOLD)
{
2023-03-04 16:13:19 +00:00
if (hio_dev_thr_read(thr->peer, 0) <= -1) return -1;
}
}
return 0;
}
static int thr_writev_to_client (thr_task_t* thr, hio_iovec_t* iov, hio_iolen_t iovcnt)
{
if (thr->task_csck)
{
2023-03-04 16:13:19 +00:00
thr->ever_attempted_to_write_to_client = 1;
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_client++;
if (hio_dev_sck_writev(thr->task_csck, iov, iovcnt, HIO_NULL, HIO_NULL) <= -1)
{
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_client--;
return -1;
}
if (thr->num_pending_writes_to_client > THR_PENDING_IO_THRESHOLD)
{
2023-03-04 16:13:19 +00:00
if (hio_dev_thr_read(thr->peer, 0) <= -1) return -1;
}
}
return 0;
}
static int thr_send_final_status_to_client (thr_task_t* thr, int status_code, int force_close)
{
2023-03-04 16:13:19 +00:00
hio_svc_htts_cli_t* cli = thr->task_client;
if (!cli) return 0; /* client disconnected probably */
if (hio_svc_htts_task_buildfinalres(thr, status_code, HIO_NULL, HIO_NULL, force_close) <= -1) return -1;
return (thr_write_to_client(thr, HIO_BECS_PTR(cli->sbuf), HIO_BECS_LEN(cli->sbuf)) <= -1 ||
(force_close && thr_write_to_client(thr, HIO_NULL, 0) <= -1))? -1: 0;
}
static int thr_write_last_chunk_to_client (thr_task_t* thr)
{
2023-03-04 16:13:19 +00:00
if (!thr->ever_attempted_to_write_to_client)
{
if (thr_send_final_status_to_client(thr, HIO_HTTP_STATUS_INTERNAL_SERVER_ERROR, 0) <= -1) return -1;
}
else
{
if (thr->res_mode_to_cli == THR_RES_MODE_CHUNKED &&
thr_write_to_client(thr, "0\r\n\r\n", 5) <= -1) return -1;
}
if (!thr->task_keep_client_alive && thr_write_to_client(thr, HIO_NULL, 0) <= -1) return -1;
return 0;
}
static int thr_write_to_peer (thr_task_t* thr, const void* data, hio_iolen_t dlen)
{
2023-03-04 16:13:19 +00:00
if (thr->peer)
{
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_peer++;
if (hio_dev_thr_write(thr->peer, data, dlen, HIO_NULL) <= -1)
{
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_peer--;
return -1;
}
/* TODO: check if it's already finished or something.. */
if (thr->num_pending_writes_to_peer > THR_PENDING_IO_THRESHOLD)
{
if (thr->task_csck && hio_dev_sck_read(thr->task_csck, 0) <= -1) return -1;
}
}
return 0;
}
2023-03-04 16:13:19 +00:00
static HIO_INLINE void thr_task_mark_over (thr_task_t* thr, int over_bits)
{
2023-03-04 16:13:19 +00:00
hio_svc_htts_t* htts = thr->htts;
hio_t* hio = htts->hio;
unsigned int old_over;
2023-03-04 16:13:19 +00:00
old_over = thr->over;
thr->over |= over_bits;
HIO_DEBUG8 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - old_over=%x | new-bits=%x => over=%x\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer, (int)old_over, (int)over_bits, (int)thr->over);
if (!(old_over & THR_OVER_READ_FROM_CLIENT) && (thr->over & THR_OVER_READ_FROM_CLIENT))
{
if (thr->task_csck && hio_dev_sck_read(thr->task_csck, 0) <= -1)
{
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - halting client for failure to disable input watching\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
hio_dev_sck_halt (thr->task_csck);
}
}
if (!(old_over & THR_OVER_READ_FROM_PEER) && (thr->over & THR_OVER_READ_FROM_PEER))
{
2023-03-04 16:13:19 +00:00
if (thr->peer && hio_dev_thr_read(thr->peer, 0) <= -1)
{
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - halting peer for failure to disable input watching\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
2023-03-04 16:13:19 +00:00
hio_dev_thr_halt (thr->peer);
}
}
if (old_over != THR_OVER_ALL && thr->over == THR_OVER_ALL)
{
/* ready to stop */
2023-03-04 16:13:19 +00:00
if (thr->peer)
{
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - halting peer as it is unneeded\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
2023-03-04 16:13:19 +00:00
hio_dev_thr_halt (thr->peer);
}
if (thr->task_csck)
{
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr->task_client != HIO_NULL);
if (thr->task_keep_client_alive && !thr->client_eof_detected)
{
/* how to arrange to delete this thr_task object and put the socket back to the normal waiting state??? */
2023-03-04 16:13:19 +00:00
HIO_ASSERT (thr->htts->hio, thr->task_client->task == (hio_svc_htts_task_t*)thr);
HIO_SVC_HTTS_TASK_UNREF (thr->task_client->task);
/* IMPORTANT: thr_task must not be accessed from here down as it could have been destroyed */
}
else
{
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - halting client for no keep-alive\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
hio_dev_sck_shutdown (thr->task_csck, HIO_DEV_SCK_SHUTDOWN_WRITE);
hio_dev_sck_halt (thr->task_csck);
}
}
}
}
static void thr_task_on_kill (hio_svc_htts_task_t* task)
2020-05-25 08:04:30 +00:00
{
2023-03-04 16:13:19 +00:00
thr_task_t* thr = (thr_task_t*)task;
hio_t* hio = thr->htts->hio;
2020-07-08 05:24:40 +00:00
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - killing the task\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
2023-03-04 16:13:19 +00:00
if (thr->on_kill) thr->on_kill (task);
2023-03-04 16:13:19 +00:00
if (thr->peer)
{
2023-03-04 16:13:19 +00:00
thr_peer_xtn_t* peer_xtn = hio_dev_thr_getxtn(thr->peer);
if (peer_xtn->task)
{
2023-03-04 16:13:19 +00:00
/* peer_xtn->task may not be NULL if the resource is killed regardless of the reference count.
* anyway, don't use HIO_SVC_HTTS_TASK_UNREF (peer_xtn->task) because the resource itself
* is already being killed. */
2023-03-04 16:13:19 +00:00
peer_xtn->task = HIO_NULL;
}
2023-03-04 16:13:19 +00:00
hio_dev_thr_kill (thr->peer);
thr->peer = HIO_NULL;
}
2023-03-04 16:13:19 +00:00
if (thr->peer_htrd)
{
2023-03-04 16:13:19 +00:00
thr_peer_xtn_t* peer_xtn = hio_htrd_getxtn(thr->peer_htrd);
if (peer_xtn->task) peer_xtn->task = HIO_NULL; // no HIO_SVC_HTTS_TASK_UNREF() for the same reason above
2023-03-04 16:13:19 +00:00
hio_htrd_close (thr->peer_htrd);
thr->peer_htrd = HIO_NULL;
}
if (thr->task_csck)
{
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr->task_client != HIO_NULL);
/* restore callbacks */
if (thr->client_org_on_read) thr->task_csck->on_read = thr->client_org_on_read;
if (thr->client_org_on_write) thr->task_csck->on_write = thr->client_org_on_write;
if (thr->client_org_on_disconnect) thr->task_csck->on_disconnect = thr->client_org_on_disconnect;
2023-03-04 16:13:19 +00:00
if (thr->client_htrd_recbs_changed) hio_htrd_setrecbs (thr->task_client->htrd, &thr->client_htrd_org_recbs);
if (!thr->task_keep_client_alive || hio_dev_sck_read(thr->task_csck, 1) <= -1)
{
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - halting client for failure to enable input watching\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
hio_dev_sck_halt (thr->task_csck);
}
}
2023-03-04 16:13:19 +00:00
thr->client_org_on_read = HIO_NULL;
thr->client_org_on_write = HIO_NULL;
thr->client_org_on_disconnect = HIO_NULL;
thr->client_htrd_recbs_changed = 0;
2023-03-04 16:13:19 +00:00
if (thr->task_next) HIO_SVC_HTTS_TASKL_UNLINK_TASK (thr); /* detach from the htts service only if it's attached */
HIO_DEBUG5 (hio, "HTTS(%p) - thr(t=%p,c=%p[%d],p=%p) - killed the task\n", thr->htts, thr, thr->task_client, (thr->task_csck? thr->task_csck->hnd: -1), thr->peer);
}
2023-03-04 16:08:53 +00:00
static void thr_peer_on_close (hio_dev_thr_t* peer, hio_dev_thr_sid_t sid)
{
2023-03-04 16:08:53 +00:00
hio_t* hio = peer->hio;
thr_peer_xtn_t* peer_xtn = (thr_peer_xtn_t*)hio_dev_thr_getxtn(peer);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = peer_xtn->task;
2023-03-04 16:13:19 +00:00
if (!thr) return; /* thr task already gone */
switch (sid)
{
2021-07-22 07:30:20 +00:00
case HIO_DEV_THR_MASTER:
2023-03-04 16:13:19 +00:00
HIO_DEBUG2 (hio, "HTTS(%p) - peer %p closing master\n", thr->htts, peer);
thr->peer = HIO_NULL; /* clear this peer from the state */
2023-03-04 16:08:53 +00:00
HIO_ASSERT (hio, peer_xtn->task != HIO_NULL);
HIO_SVC_HTTS_TASK_UNREF (peer_xtn->task);
2023-03-04 16:13:19 +00:00
if (thr->peer_htrd)
{
/* once this peer device is closed, peer's htrd is also never used.
* it's safe to detach the extra information attached on the htrd object. */
2023-03-04 16:13:19 +00:00
peer_xtn = hio_htrd_getxtn(thr->peer_htrd);
2023-03-04 16:08:53 +00:00
HIO_ASSERT (hio, peer_xtn->task != HIO_NULL);
HIO_SVC_HTTS_TASK_UNREF (peer_xtn->task);
}
break;
2021-07-22 07:30:20 +00:00
case HIO_DEV_THR_OUT:
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr->peer == peer);
HIO_DEBUG3 (hio, "HTTS(%p) - peer %p closing slave[%d]\n", thr->htts, peer, sid);
if (!(thr->over & THR_OVER_READ_FROM_PEER))
{
if (thr_write_last_chunk_to_client(thr) <= -1)
2023-03-04 16:13:19 +00:00
thr_task_halt_participating_devices (thr);
else
thr_task_mark_over (thr, THR_OVER_READ_FROM_PEER);
}
break;
2021-07-22 07:30:20 +00:00
case HIO_DEV_THR_IN:
thr_task_mark_over (thr, THR_OVER_WRITE_TO_PEER);
break;
default:
2023-03-04 16:13:19 +00:00
HIO_DEBUG3 (hio, "HTTS(%p) - peer %p closing slave[%d]\n", thr->htts, peer, sid);
/* do nothing */
break;
}
}
2023-03-04 16:08:53 +00:00
static int thr_peer_on_read (hio_dev_thr_t* peer, const void* data, hio_iolen_t dlen)
{
2023-03-04 16:08:53 +00:00
hio_t* hio = peer->hio;
thr_peer_xtn_t* peer_xtn = (thr_peer_xtn_t*)hio_dev_thr_getxtn(peer);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = peer_xtn->task;
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr != HIO_NULL);
if (dlen <= -1)
{
2023-03-04 16:13:19 +00:00
HIO_DEBUG2 (hio, "HTTPS(%p) - read error from peer %p\n", thr->htts, peer);
goto oops;
}
if (dlen == 0)
{
2023-03-04 16:13:19 +00:00
HIO_DEBUG2 (hio, "HTTPS(%p) - EOF from peer %p\n", thr->htts, peer);
if (!(thr->over & THR_OVER_READ_FROM_PEER))
{
int n;
/* the thr script could be misbehaviing.
* it still has to read more but EOF is read.
* otherwise client_peer_htrd_poke() should have been called */
n = thr_write_last_chunk_to_client(thr);
thr_task_mark_over (thr, THR_OVER_READ_FROM_PEER);
if (n <= -1) goto oops;
}
}
else
{
2021-07-22 07:30:20 +00:00
hio_oow_t rem;
HIO_ASSERT (hio, !(thr->over & THR_OVER_READ_FROM_PEER));
2023-03-04 16:13:19 +00:00
if (hio_htrd_feed(thr->peer_htrd, data, dlen, &rem) <= -1)
{
2023-03-04 16:13:19 +00:00
HIO_DEBUG2 (hio, "HTTPS(%p) - unable to feed peer htrd - peer %p\n", thr->htts, peer);
2023-03-04 16:13:19 +00:00
if (!thr->ever_attempted_to_write_to_client &&
!(thr->over & THR_OVER_WRITE_TO_CLIENT))
{
thr_send_final_status_to_client (thr, HIO_HTTP_STATUS_INTERNAL_SERVER_ERROR, 1); /* don't care about error because it jumps to oops below anyway */
}
goto oops;
}
if (rem > 0)
{
/* If the script specifies Content-Length and produces longer data, it will come here */
/*printf ("AAAAAAAAAAAAAAAAAa EEEEEXcessive DATA..................\n");*/
/* TODO: or drop this request?? */
}
}
return 0;
oops:
2023-03-04 16:13:19 +00:00
thr_task_halt_participating_devices (thr);
return 0;
}
static int peer_capture_response_header (hio_htre_t* req, const hio_bch_t* key, const hio_htre_hdrval_t* val, void* ctx)
{
return hio_svc_htts_task_addreshdrs((thr_task_t*)ctx, key, val);
}
2021-07-22 07:30:20 +00:00
static int thr_peer_htrd_peek (hio_htrd_t* htrd, hio_htre_t* req)
{
#if 0
2023-03-04 16:13:19 +00:00
thr_peer_xtn_t* peer_xtn = hio_htrd_getxtn(htrd);
thr_task_t* thr = peer_xtn->task;
hio_svc_htts_cli_t* cli = thr->task_client;
2021-07-22 07:30:20 +00:00
hio_bch_t dtbuf[64];
int status_code = HIO_HTTP_STATUS_OK;
const hio_bch_t* status_desc = HIO_NULL;
if (req->attr.content_length)
{
// TOOD: remove content_length if content_length is negative or not numeric.
thr->res_mode_to_cli = THR_RES_MODE_LENGTH;
}
if (req->attr.status) hio_parse_http_status_header_value(req->attr.status, &status_code, &status_desc);
2021-07-22 07:30:20 +00:00
hio_svc_htts_fmtgmtime (cli->htts, HIO_NULL, dtbuf, HIO_COUNTOF(dtbuf));
2021-07-22 07:30:20 +00:00
if (hio_becs_fmt(cli->sbuf, "HTTP/%d.%d %d %hs\r\nServer: %hs\r\nDate: %hs\r\n",
2023-03-04 16:13:19 +00:00
thr->task_req_version.major, thr->task_req_version.minor,
status_code, (status_desc? status_desc: hio_http_status_to_bcstr(status_code)),
2021-07-22 07:30:20 +00:00
cli->htts->server_name, dtbuf) == (hio_oow_t)-1) return -1;
2021-07-22 07:30:20 +00:00
if (hio_htre_walkheaders(req, thr_peer_capture_response_header, cli) <= -1) return -1;
2023-03-04 16:13:19 +00:00
switch (thr->res_mode_to_cli)
{
case THR_RES_MODE_CHUNKED:
2021-07-22 07:30:20 +00:00
if (hio_becs_cat(cli->sbuf, "Transfer-Encoding: chunked\r\n") == (hio_oow_t)-1) return -1;
/*if (hio_becs_cat(cli->sbuf, "Connection: keep-alive\r\n") == (hio_oow_t)-1) return -1;*/
break;
case THR_RES_MODE_CLOSE:
2021-07-22 07:30:20 +00:00
if (hio_becs_cat(cli->sbuf, "Connection: close\r\n") == (hio_oow_t)-1) return -1;
break;
case THR_RES_MODE_LENGTH:
if (hio_becs_cat(cli->sbuf, (thr->task_keep_client_alive? "Connection: keep-alive\r\n": "Connection: close\r\n")) == (hio_oow_t)-1) return -1;
}
2021-07-22 07:30:20 +00:00
if (hio_becs_cat(cli->sbuf, "\r\n") == (hio_oow_t)-1) return -1;
thr->task_status_code = status_code;
return thr_write_to_client(thr, HIO_BECS_PTR(cli->sbuf), HIO_BECS_LEN(cli->sbuf));
#else
thr_peer_xtn_t* peer = hio_htrd_getxtn(htrd);
thr_task_t* thr = peer->task;
hio_svc_htts_cli_t* cli = thr->task_client;
int status_code = HIO_HTTP_STATUS_OK;
const hio_bch_t* status_desc = HIO_NULL;
int chunked;
if (HIO_UNLIKELY(!cli))
{
/* client disconnected or not connectd */
return 0;
}
// TOOD: remove content_length if content_length is negative or not numeric.
if (req->attr.content_length) thr->res_mode_to_cli = THR_RES_MODE_LENGTH;
if (req->attr.status) hio_parse_http_status_header_value(req->attr.status, &status_code, &status_desc);
chunked = thr->task_keep_client_alive && !req->attr.content_length;
if (hio_svc_htts_task_startreshdr(thr, status_code, status_desc, chunked) <= -1 ||
hio_htre_walkheaders(req, peer_capture_response_header, thr) <= -1 ||
hio_svc_htts_task_endreshdr(thr) <= -1) return -1;
thr->task_status_code = status_code;
return thr_write_to_client(thr, HIO_BECS_PTR(cli->sbuf), HIO_BECS_LEN(cli->sbuf));
#endif
}
2021-07-22 07:30:20 +00:00
static int thr_peer_htrd_poke (hio_htrd_t* htrd, hio_htre_t* req)
{
/* client request got completed */
2023-03-04 16:13:19 +00:00
thr_peer_xtn_t* peer_xtn = hio_htrd_getxtn(htrd);
thr_task_t* thr = peer_xtn->task;
if (thr_write_last_chunk_to_client(thr) <= -1) return -1;
thr_task_mark_over (thr, THR_OVER_READ_FROM_PEER);
return 0;
}
2021-07-22 07:30:20 +00:00
static int thr_peer_htrd_push_content (hio_htrd_t* htrd, hio_htre_t* req, const hio_bch_t* data, hio_oow_t dlen)
{
2023-03-04 16:13:19 +00:00
thr_peer_xtn_t* peer_xtn = hio_htrd_getxtn(htrd);
thr_task_t* thr = peer_xtn->task;
2023-03-04 16:13:19 +00:00
HIO_ASSERT (thr->htts->hio, htrd == thr->peer_htrd);
2023-03-04 16:13:19 +00:00
switch (thr->res_mode_to_cli)
{
case THR_RES_MODE_CHUNKED:
{
2021-07-22 07:30:20 +00:00
hio_iovec_t iov[3];
hio_bch_t lbuf[16];
hio_oow_t llen;
2021-07-22 07:30:20 +00:00
/* hio_fmt_uintmax_to_bcstr() null-terminates the output. only HIO_COUNTOF(lbuf) - 1
* is enough to hold '\r' and '\n' at the back without '\0'. */
2021-07-22 07:30:20 +00:00
llen = hio_fmt_uintmax_to_bcstr(lbuf, HIO_COUNTOF(lbuf) - 1, dlen, 16 | HIO_FMT_UINTMAX_UPPERCASE, 0, '\0', HIO_NULL);
lbuf[llen++] = '\r';
lbuf[llen++] = '\n';
iov[0].iov_ptr = lbuf;
iov[0].iov_len = llen;
iov[1].iov_ptr = (void*)data;
iov[1].iov_len = dlen;
iov[2].iov_ptr = "\r\n";
iov[2].iov_len = 2;
if (thr_writev_to_client(thr, iov, HIO_COUNTOF(iov)) <= -1) goto oops;
break;
}
case THR_RES_MODE_CLOSE:
case THR_RES_MODE_LENGTH:
if (thr_write_to_client(thr, data, dlen) <= -1) goto oops;
break;
}
if (thr->num_pending_writes_to_client > THR_PENDING_IO_THRESHOLD)
{
2023-03-04 16:13:19 +00:00
if (hio_dev_thr_read(thr->peer, 0) <= -1) goto oops;
}
return 0;
oops:
return -1;
}
2021-07-22 07:30:20 +00:00
static hio_htrd_recbs_t thr_peer_htrd_recbs =
{
thr_peer_htrd_peek,
thr_peer_htrd_poke,
thr_peer_htrd_push_content
};
2021-07-22 07:30:20 +00:00
static int thr_client_htrd_poke (hio_htrd_t* htrd, hio_htre_t* req)
{
/* client request got completed */
2021-07-22 07:30:20 +00:00
hio_svc_htts_cli_htrd_xtn_t* htrdxtn = (hio_svc_htts_cli_htrd_xtn_t*)hio_htrd_getxtn(htrd);
hio_dev_sck_t* sck = htrdxtn->sck;
hio_svc_htts_cli_t* cli = hio_dev_sck_getxtn(sck);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = (thr_task_t*)cli->task;
/* indicate EOF to the client peer */
if (thr_write_to_peer(thr, HIO_NULL, 0) <= -1) return -1;
thr_task_mark_over (thr, THR_OVER_READ_FROM_CLIENT);
return 0;
2020-05-25 08:04:30 +00:00
}
2021-07-22 07:30:20 +00:00
static int thr_client_htrd_push_content (hio_htrd_t* htrd, hio_htre_t* req, const hio_bch_t* data, hio_oow_t dlen)
2020-05-25 08:04:30 +00:00
{
2021-07-22 07:30:20 +00:00
hio_svc_htts_cli_htrd_xtn_t* htrdxtn = (hio_svc_htts_cli_htrd_xtn_t*)hio_htrd_getxtn(htrd);
hio_dev_sck_t* sck = htrdxtn->sck;
hio_svc_htts_cli_t* cli = hio_dev_sck_getxtn(sck);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = (thr_task_t*)cli->task;
2020-05-25 08:04:30 +00:00
2021-07-22 07:30:20 +00:00
HIO_ASSERT (sck->hio, cli->sck == sck);
return thr_write_to_peer(thr, data, dlen);
2020-05-25 08:04:30 +00:00
}
2021-07-22 07:30:20 +00:00
static hio_htrd_recbs_t thr_client_htrd_recbs =
2020-05-25 08:04:30 +00:00
{
2021-07-22 07:30:20 +00:00
HIO_NULL,
thr_client_htrd_poke,
thr_client_htrd_push_content
};
2023-03-04 16:08:53 +00:00
static int thr_peer_on_write (hio_dev_thr_t* peer, hio_iolen_t wrlen, void* wrctx)
{
2023-03-04 16:08:53 +00:00
hio_t* hio = peer->hio;
thr_peer_xtn_t* peer_xtn = (thr_peer_xtn_t*)hio_dev_thr_getxtn(peer);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = peer_xtn->task;
2023-03-04 16:13:19 +00:00
if (!thr) return 0; /* there is nothing i can do. the thr_task is being cleared or has been cleared already. */
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr->peer == peer);
if (wrlen <= -1)
{
2023-03-04 16:13:19 +00:00
HIO_DEBUG2 (hio, "HTTS(%p) - unable to write to peer %p\n", thr->htts, peer);
goto oops;
}
else if (wrlen == 0)
{
/* indicated EOF */
/* do nothing here as i didn't incremented num_pending_writes_to_peer when making the write request */
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_peer--;
HIO_ASSERT (hio, thr->num_pending_writes_to_peer == 0);
HIO_DEBUG2 (hio, "HTTS(%p) - indicated EOF to peer %p\n", thr->htts, peer);
/* indicated EOF to the peer side. i need no more data from the client side.
* i don't need to enable input watching in the client side either */
thr_task_mark_over (thr, THR_OVER_WRITE_TO_PEER);
}
else
{
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr->num_pending_writes_to_peer > 0);
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_peer--;
if (thr->num_pending_writes_to_peer == THR_PENDING_IO_THRESHOLD)
{
if (!(thr->over & THR_OVER_READ_FROM_CLIENT) &&
hio_dev_sck_read(thr->task_csck, 1) <= -1) goto oops;
}
if ((thr->over & THR_OVER_READ_FROM_CLIENT) && thr->num_pending_writes_to_peer <= 0)
{
thr_task_mark_over (thr, THR_OVER_WRITE_TO_PEER);
}
}
return 0;
oops:
2023-03-04 16:13:19 +00:00
thr_task_halt_participating_devices (thr);
return 0;
}
2021-07-22 07:30:20 +00:00
static void thr_client_on_disconnect (hio_dev_sck_t* sck)
{
2021-07-22 07:30:20 +00:00
hio_svc_htts_cli_t* cli = hio_dev_sck_getxtn(sck);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = (thr_task_t*)cli->task;
hio_svc_htts_t* htts = thr->htts;
hio_t* hio = sck->hio;
HIO_ASSERT (hio, sck = thr->task_csck);
2023-03-04 16:13:19 +00:00
HIO_DEBUG4 (hio, "HTTS(%p) - thr(t=%p,c=%p,csck=%p) - client socket disconnect notified\n", htts, thr, cli, sck);
2023-03-04 16:13:19 +00:00
thr->client_disconnected = 1;
thr->task_csck = HIO_NULL;
2023-03-04 16:13:19 +00:00
thr->task_client = HIO_NULL;
if (thr->client_org_on_disconnect)
{
2023-03-04 16:13:19 +00:00
thr->client_org_on_disconnect (sck);
/* this original callback destroys the associated resource.
* thr_task must not be accessed from here down */
}
2023-03-04 16:13:19 +00:00
HIO_DEBUG4 (hio, "HTTS(%p) - thr(t=%p,c=%p,csck=%p) - client socket disconnect handled\n", htts, thr, cli, sck);
/* Note: after this callback, the actual device pointed to by 'sck' will be freed in the main loop. */
}
2021-07-22 07:30:20 +00:00
static int thr_client_on_read (hio_dev_sck_t* sck, const void* buf, hio_iolen_t len, const hio_skad_t* srcaddr)
{
2021-07-22 07:30:20 +00:00
hio_t* hio = sck->hio;
hio_svc_htts_cli_t* cli = hio_dev_sck_getxtn(sck);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = (thr_task_t*)cli->task;
2021-07-22 07:30:20 +00:00
HIO_ASSERT (hio, sck == cli->sck);
if (len <= -1)
{
/* read error */
2021-07-22 07:30:20 +00:00
HIO_DEBUG2 (cli->htts->hio, "HTTPS(%p) - read error on client %p(%d)\n", sck, (int)sck->hnd);
goto oops;
}
2023-03-04 16:13:19 +00:00
if (!thr->peer)
{
/* the peer is gone */
goto oops; /* do what? just return 0? */
}
if (len == 0)
{
/* EOF on the client side. arrange to close */
2023-03-04 16:13:19 +00:00
HIO_DEBUG3 (hio, "HTTPS(%p) - EOF from client %p(hnd=%d)\n", thr->htts, sck, (int)sck->hnd);
thr->client_eof_detected = 1;
if (!(thr->over & THR_OVER_READ_FROM_CLIENT)) /* if this is true, EOF is received without thr_client_htrd_poke() */
{
int n;
n = thr_write_to_peer(thr, HIO_NULL, 0);
thr_task_mark_over (thr, THR_OVER_READ_FROM_CLIENT);
if (n <= -1) goto oops;
}
}
else
{
2021-07-22 07:30:20 +00:00
hio_oow_t rem;
HIO_ASSERT (hio, !(thr->over & THR_OVER_READ_FROM_CLIENT));
2021-07-22 07:30:20 +00:00
if (hio_htrd_feed(cli->htrd, buf, len, &rem) <= -1) goto oops;
if (rem > 0)
{
/* TODO store this to client buffer. once the current resource is completed, arrange to call on_read() with it */
2021-07-22 07:30:20 +00:00
HIO_DEBUG3 (hio, "HTTPS(%p) - excessive data after contents by thr client %p(%d)\n", sck->hio, sck, (int)sck->hnd);
}
}
return 0;
oops:
2023-03-04 16:13:19 +00:00
thr_task_halt_participating_devices (thr);
return 0;
}
2021-07-22 07:30:20 +00:00
static int thr_client_on_write (hio_dev_sck_t* sck, hio_iolen_t wrlen, void* wrctx, const hio_skad_t* dstaddr)
{
2021-07-22 07:30:20 +00:00
hio_t* hio = sck->hio;
hio_svc_htts_cli_t* cli = hio_dev_sck_getxtn(sck);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = (thr_task_t*)cli->task;
if (wrlen <= -1)
{
2021-07-22 07:30:20 +00:00
HIO_DEBUG3 (hio, "HTTPS(%p) - unable to write to client %p(%d)\n", sck->hio, sck, (int)sck->hnd);
goto oops;
}
if (wrlen == 0)
{
/* if the connect is keep-alive, this part may not be called */
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_client--;
HIO_ASSERT (hio, thr->num_pending_writes_to_client == 0);
HIO_DEBUG3 (hio, "HTTS(%p) - indicated EOF to client %p(%d)\n", thr->htts, sck, (int)sck->hnd);
/* since EOF has been indicated to the client, it must not write to the client any further.
* this also means that i don't need any data from the peer side either.
* i don't need to enable input watching on the peer side */
thr_task_mark_over (thr, THR_OVER_WRITE_TO_CLIENT);
}
else
{
2023-03-04 16:13:19 +00:00
HIO_ASSERT (hio, thr->num_pending_writes_to_client > 0);
2023-03-04 16:13:19 +00:00
thr->num_pending_writes_to_client--;
if (thr->peer && thr->num_pending_writes_to_client == THR_PENDING_IO_THRESHOLD)
{
if (!(thr->over & THR_OVER_READ_FROM_PEER) &&
2023-03-04 16:13:19 +00:00
hio_dev_thr_read(thr->peer, 1) <= -1) goto oops;
}
if ((thr->over & THR_OVER_READ_FROM_PEER) && thr->num_pending_writes_to_client <= 0)
{
thr_task_mark_over (thr, THR_OVER_WRITE_TO_CLIENT);
}
}
return 0;
oops:
2023-03-04 16:13:19 +00:00
thr_task_halt_participating_devices (thr);
return 0;
}
static void free_thr_start_info (void* ctx)
{
2023-01-11 15:33:52 +00:00
/* this function is a thread cleanup handler.
* it can get invoked after htts is destroyed by hio_svc_htts_stop() because
* hio_dev_thr_kill() pushes back the job using hio_addcfmb() and the
* actual cfmb clean-up is performed after the service stop.
* it is not realiable to use tfs->htts or tfs->htts->hio. use tfs->hio only here.
==3845396== Invalid read of size 8
==3845396== at 0x40A7D5: free_thr_start_info (http-thr.c:804)
==3845396== by 0x40A7D5: thr_func (http-thr.c:815)
==3845396== by 0x41AE46: run_thr_func (thr.c:127)
==3845396== by 0x4A132A4: start_thread (in /usr/lib64/libpthread-2.33.so)
==3845396== by 0x4B2B322: clone (in /usr/lib64/libc-2.33.so)
==3845396== Address 0x4c38b00 is 0 bytes inside a block of size 464 free'd
==3845396== at 0x48430E4: free (vg_replace_malloc.c:872)
==3845396== by 0x4091EE: hio_svc_htts_stop (http-svr.c:555)
==3845396== by 0x40F5BE: hio_fini (hio.c:185)
==3845396== by 0x40F848: hio_close (hio.c:101)
==3845396== by 0x402CB4: main (webs.c:511)
==3845396== Block was alloc'd at
==3845396== at 0x484086F: malloc (vg_replace_malloc.c:381)
==3845396== by 0x412873: hio_callocmem (hio.c:2019)
==3845396== by 0x40978E: hio_svc_htts_start (http-svr.c:350)
==3845396== by 0x403900: webs_start (webs.c:385)
==3845396== by 0x402C6C: main (webs.c:498)
*/
thr_func_start_t* tfs = (thr_func_start_t*)ctx;
2023-01-11 15:33:52 +00:00
hio_t* hio = tfs->hio;
if (tfs->tfi.req_path) hio_freemem (hio, tfs->tfi.req_path);
if (tfs->tfi.req_param) hio_freemem (hio, tfs->tfi.req_param);
hio_freemem (hio, tfs);
}
2021-07-22 07:30:20 +00:00
static void thr_func (hio_t* hio, hio_dev_thr_iopair_t* iop, void* ctx)
{
thr_func_start_t* tfs = (thr_func_start_t*)ctx;
pthread_cleanup_push (free_thr_start_info, tfs);
tfs->thr_func (tfs->htts, iop, &tfs->tfi, tfs->thr_ctx);
pthread_cleanup_pop (1);
}
2021-07-22 07:30:20 +00:00
static int thr_capture_request_header (hio_htre_t* req, const hio_bch_t* key, const hio_htre_hdrval_t* val, void* ctx)
{
thr_func_start_t* tfs = (thr_func_start_t*)ctx;
2021-07-22 07:30:20 +00:00
if (hio_comp_bcstr(key, "X-HTTP-Method-Override", 1) == 0)
{
2021-07-22 07:30:20 +00:00
tfs->tfi.req_x_http_method_override = hio_bchars_to_http_method(val->ptr, val->len); /* don't care about multiple values */
}
#if 0
2021-07-22 07:30:20 +00:00
if (hio_comp_bcstr(key, "Connection", 1) != 0 &&
hio_comp_bcstr(key, "Transfer-Encoding", 1) != 0 &&
hio_comp_bcstr(key, "Content-Length", 1) != 0 &&
hio_comp_bcstr(key, "Expect", 1) != 0)
{
do
{
/* TODO: ... */
val = val->next;
}
while (val);
}
#endif
return 0;
}
int hio_svc_htts_dothr (hio_svc_htts_t* htts, hio_dev_sck_t* csck, hio_htre_t* req, hio_svc_htts_thr_func_t func, void* ctx, int options, hio_svc_htts_task_on_kill_t on_kill)
{
2021-07-22 07:30:20 +00:00
hio_t* hio = htts->hio;
hio_svc_htts_cli_t* cli = hio_dev_sck_getxtn(csck);
2023-03-04 16:13:19 +00:00
thr_task_t* thr = HIO_NULL;
thr_peer_xtn_t* peer_xtn;
2021-07-22 07:30:20 +00:00
hio_dev_thr_make_t mi;
thr_func_start_t* tfs;
int have_content;
/* ensure that you call this function before any contents is received */
2021-07-22 07:30:20 +00:00
HIO_ASSERT (hio, hio_htre_getcontentlen(req) == 0);
2021-07-22 07:30:20 +00:00
tfs = hio_callocmem(hio, HIO_SIZEOF(*tfs));
if (!tfs) goto oops;
2023-01-11 15:33:52 +00:00
tfs->hio = hio;
tfs->htts = htts;
tfs->thr_func = func;
tfs->thr_ctx = ctx;
2021-07-22 07:30:20 +00:00
tfs->tfi.req_method = hio_htre_getqmethodtype(req);
tfs->tfi.req_version = *hio_htre_getversion(req);
tfs->tfi.req_path = hio_dupbcstr(hio, hio_htre_getqpath(req), HIO_NULL);
if (!tfs->tfi.req_path) goto oops;
2021-07-22 07:30:20 +00:00
if (hio_htre_getqparam(req))
{
2021-07-22 07:30:20 +00:00
tfs->tfi.req_param = hio_dupbcstr(hio, hio_htre_getqparam(req), HIO_NULL);
if (!tfs->tfi.req_param) goto oops;
}
tfs->tfi.req_x_http_method_override = -1;
2021-07-22 07:30:20 +00:00
if (hio_htre_walkheaders(req, thr_capture_request_header, tfs) <= -1) goto oops;
tfs->tfi.server_addr = cli->sck->localaddr;
tfs->tfi.client_addr = cli->sck->remoteaddr;
2021-07-22 07:30:20 +00:00
HIO_MEMSET (&mi, 0, HIO_SIZEOF(mi));
mi.thr_func = thr_func;
mi.thr_ctx = tfs;
mi.on_read = thr_peer_on_read;
mi.on_write = thr_peer_on_write;
mi.on_close = thr_peer_on_close;
2020-05-25 08:04:30 +00:00
thr = (thr_task_t*)hio_svc_htts_task_make(htts, HIO_SIZEOF(*thr), thr_task_on_kill, req, csck);
2023-03-04 16:13:19 +00:00
if (HIO_UNLIKELY(!thr)) goto oops;
2023-03-04 16:13:19 +00:00
thr->on_kill = on_kill;
thr->options = options;
2020-05-25 08:04:30 +00:00
2023-03-04 16:13:19 +00:00
/*thr->num_pending_writes_to_client = 0;
thr->num_pending_writes_to_peer = 0;*/
thr->req_content_length_unlimited = hio_htre_getreqcontentlen(req, &thr->req_content_length);
2023-03-04 16:13:19 +00:00
thr->client_org_on_read = csck->on_read;
thr->client_org_on_write = csck->on_write;
thr->client_org_on_disconnect = csck->on_disconnect;
csck->on_read = thr_client_on_read;
csck->on_write = thr_client_on_write;
csck->on_disconnect = thr_client_on_disconnect;
/* attach the thr task to the client socket via the task field in the extended space of the socket */
HIO_ASSERT (hio, cli->task == HIO_NULL);
2023-03-04 16:13:19 +00:00
HIO_SVC_HTTS_TASK_REF ((hio_svc_htts_task_t*)thr, cli->task);
2023-03-04 16:13:19 +00:00
thr->peer = hio_dev_thr_make(hio, HIO_SIZEOF(*peer_xtn), &mi);
if (HIO_UNLIKELY(!thr->peer))
{
/* no need to detach the attached task here because that is handled
* in the kill/disconnect callbacks of relevant devices */
2021-07-22 07:30:20 +00:00
HIO_DEBUG3 (hio, "HTTS(%p) - failed to create thread for %p(%d)\n", htts, csck, (int)csck->hnd);
goto oops;
2020-07-27 12:10:55 +00:00
}
2021-07-22 07:30:20 +00:00
tfs = HIO_NULL; /* mark that tfs is delegated to the thread */
/* attach the thr task to the peer thread device */
2023-03-04 16:13:19 +00:00
peer_xtn = hio_dev_thr_getxtn(thr->peer);
HIO_SVC_HTTS_TASK_REF (thr, peer_xtn->task);
2023-03-04 16:13:19 +00:00
thr->peer_htrd = hio_htrd_open(hio, HIO_SIZEOF(*peer_xtn));
if (HIO_UNLIKELY(!thr->peer_htrd)) goto oops;
hio_htrd_setoption (thr->peer_htrd, HIO_HTRD_SKIP_INITIAL_LINE | HIO_HTRD_RESPONSE);
hio_htrd_setrecbs (thr->peer_htrd, &thr_peer_htrd_recbs);
2020-05-25 08:04:30 +00:00
/* attach the thr task to the htrd parser set on the peer thread device */
2023-03-04 16:13:19 +00:00
peer_xtn = hio_htrd_getxtn(thr->peer_htrd);
HIO_SVC_HTTS_TASK_REF (thr, peer_xtn->task);
#if !defined(THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH)
2023-03-04 16:13:19 +00:00
if (thr->req_content_length_unlimited)
{
/* Transfer-Encoding is chunked. no content-length is known in advance. */
/* option 1. buffer contents. if it gets too large, send 413 Request Entity Too Large.
* option 2. send 411 Length Required immediately
* option 3. set Content-Length to -1 and use EOF to indicate the end of content [Non-Standard] */
if (thr_send_final_status_to_client(thr, HIO_HTTP_STATUS_LENGTH_REQUIRED, 1) <= -1) goto oops;
}
#endif
2021-07-22 07:30:20 +00:00
if (req->flags & HIO_HTRE_ATTR_EXPECT100)
{
/* TODO: Expect: 100-continue? who should handle this? thr? or the http server? */
/* CAN I LET the thr SCRIPT handle this? */
if (!(options & HIO_SVC_HTTS_THR_NO_100_CONTINUE) &&
hio_comp_http_version_numbers(&req->version, 1, 1) >= 0 &&
2023-03-04 16:13:19 +00:00
(thr->req_content_length_unlimited || thr->req_content_length > 0))
{
/*
* Don't send 100 Continue if http verions is lower than 1.1
* [RFC7231]
* A server that receives a 100-continue expectation in an HTTP/1.0
* request MUST ignore that expectation.
*
* Don't send 100 Continue if expected content lenth is 0.
* [RFC7231]
* A server MAY omit sending a 100 (Continue) response if it has
* already received some or all of the message body for the
* corresponding request, or if the framing indicates that there is
* no message body.
*/
2021-07-22 07:30:20 +00:00
hio_bch_t msgbuf[64];
hio_oow_t msglen;
2023-03-04 16:13:19 +00:00
msglen = hio_fmttobcstr(hio, msgbuf, HIO_COUNTOF(msgbuf), "HTTP/%d.%d %d %hs\r\n\r\n", thr->task_req_version.major, thr->task_req_version.minor, HIO_HTTP_STATUS_CONTINUE, hio_http_status_to_bcstr(HIO_HTTP_STATUS_CONTINUE));
if (thr_write_to_client(thr, msgbuf, msglen) <= -1) goto oops;
2023-03-04 16:13:19 +00:00
thr->ever_attempted_to_write_to_client = 0; /* reset this as it's polluted for 100 continue */
}
}
2021-07-22 07:30:20 +00:00
else if (req->flags & HIO_HTRE_ATTR_EXPECT)
{
/* 417 Expectation Failed */
thr_send_final_status_to_client(thr, HIO_HTTP_STATUS_EXPECTATION_FAILED, 1);
goto oops;
}
#if defined(THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH)
2023-03-04 16:13:19 +00:00
have_content = thr->req_content_length > 0 || thr->req_content_length_unlimited;
#else
2023-03-04 16:13:19 +00:00
have_content = thr->req_content_length > 0;
#endif
if (have_content)
{
/* change the callbacks to subscribe to contents to be uploaded */
2023-03-04 16:13:19 +00:00
thr->client_htrd_org_recbs = *hio_htrd_getrecbs(thr->task_client->htrd);
thr_client_htrd_recbs.peek = thr->client_htrd_org_recbs.peek;
hio_htrd_setrecbs (thr->task_client->htrd, &thr_client_htrd_recbs);
thr->client_htrd_recbs_changed = 1;
}
else
{
/* no content to be uploaded from the client */
/* indicate EOF to the peer and disable input wathching from the client */
if (thr_write_to_peer(thr, HIO_NULL, 0) <= -1) goto oops;
thr_task_mark_over (thr, THR_OVER_READ_FROM_CLIENT | THR_OVER_WRITE_TO_PEER);
}
2020-05-25 08:04:30 +00:00
thr->res_mode_to_cli = thr->task_keep_client_alive? THR_RES_MODE_CHUNKED: THR_RES_MODE_CLOSE;
/* the mode still can get switched from THR_RES_MODE_CHUNKED to THR_RES_MODE_LENGTH
if the thread function emits Content-Length */
2020-05-25 08:04:30 +00:00
/* TODO: store current input watching state and use it when destroying the thr_task data */
if (hio_dev_sck_read(csck, !(thr->over & THR_OVER_READ_FROM_CLIENT)) <= -1) goto oops;
2023-03-04 16:13:19 +00:00
HIO_SVC_HTTS_TASKL_APPEND_TASK (&htts->task, (hio_svc_htts_task_t*)thr);
2020-05-25 08:04:30 +00:00
return 0;
oops:
2021-07-22 07:30:20 +00:00
HIO_DEBUG2 (hio, "HTTS(%p) - FAILURE in dothr - socket(%p)\n", htts, csck);
if (tfs) free_thr_start_info (tfs);
2023-03-04 16:13:19 +00:00
if (thr) thr_task_halt_participating_devices (thr);
2020-05-25 08:04:30 +00:00
return -1;
}