2020-05-25 08:04:30 +00:00
/*
* $ Id $
*
Copyright ( c ) 2016 - 2020 Chung , Hyung - Hwan . All rights reserved .
Redistribution and use in source and binary forms , with or without
modification , are permitted provided that the following conditions
are met :
1. Redistributions of source code must retain the above copyright
notice , this list of conditions and the following disclaimer .
2. Redistributions in binary form must reproduce the above copyright
notice , this list of conditions and the following disclaimer in the
documentation and / or other materials provided with the distribution .
THIS SOFTWARE IS PROVIDED BY THE AUTHOR " AS IS " AND ANY EXPRESS OR
IMPLIED WARRANTIES , INCLUDING , BUT NOT LIMITED TO , THE IMPLIED WAfRRANTIES
OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED .
IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT , INDIRECT ,
INCIDENTAL , SPECIAL , EXEMPLARY , OR CONSEQUENTIAL DAMAGES ( INCLUDING , BUT
NOT LIMITED TO , PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES ; LOSS OF USE ,
DATA , OR PROFITS ; OR BUSINESS INTERRUPTION ) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY , WHETHER IN CONTRACT , STRICT LIABILITY , OR TORT
( INCLUDING NEGLIGENCE OR OTHERWISE ) ARISING IN ANY WAY OUT OF THE USE OF
THIS SOFTWARE , EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE .
*/
# include "http-prv.h"
# include <mio-thr.h>
# include <mio-fmt.h>
# include <mio-chr.h>
# include <pthread.h>
2020-05-26 01:13:34 +00:00
# define THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH
enum thr_state_res_mode_t
{
THR_STATE_RES_MODE_CHUNKED ,
THR_STATE_RES_MODE_CLOSE ,
THR_STATE_RES_MODE_LENGTH
} ;
typedef enum thr_state_res_mode_t thr_state_res_mode_t ;
# define THR_STATE_PENDING_IO_THRESHOLD 5
# define THR_STATE_OVER_READ_FROM_CLIENT (1 << 0)
# define THR_STATE_OVER_READ_FROM_PEER (1 << 1)
# define THR_STATE_OVER_WRITE_TO_CLIENT (1 << 2)
# define THR_STATE_OVER_WRITE_TO_PEER (1 << 3)
# define THR_STATE_OVER_ALL (THR_STATE_OVER_READ_FROM_CLIENT | THR_STATE_OVER_READ_FROM_PEER | THR_STATE_OVER_WRITE_TO_CLIENT | THR_STATE_OVER_WRITE_TO_PEER)
struct thr_func_start_t
{
mio_t * mio ;
mio_svc_htts_thr_func_t thr_func ;
void * thr_ctx ;
mio_svc_htts_thr_func_info_t tfi ;
} ;
typedef struct thr_func_start_t thr_func_start_t ;
2020-05-25 08:04:30 +00:00
struct thr_state_t
{
MIO_SVC_HTTS_RSRC_HEADER ;
2020-05-26 01:13:34 +00:00
mio_oow_t num_pending_writes_to_client ;
mio_oow_t num_pending_writes_to_peer ;
mio_dev_thr_t * peer ;
mio_htrd_t * peer_htrd ;
mio_svc_htts_cli_t * client ;
mio_http_version_t req_version ; /* client request */
unsigned int over : 4 ; /* must be large enough to accomodate THR_STATE_OVER_ALL */
unsigned int keep_alive : 1 ;
unsigned int req_content_length_unlimited : 1 ;
unsigned int ever_attempted_to_write_to_client : 1 ;
unsigned int client_disconnected : 1 ;
2020-05-26 14:58:55 +00:00
unsigned int client_htrd_recbs_changed : 1 ;
2020-05-26 01:13:34 +00:00
mio_oow_t req_content_length ; /* client request content length */
thr_state_res_mode_t res_mode_to_cli ;
mio_dev_sck_on_read_t client_org_on_read ;
mio_dev_sck_on_write_t client_org_on_write ;
mio_dev_sck_on_disconnect_t client_org_on_disconnect ;
mio_htrd_recbs_t client_htrd_org_recbs ;
} ;
2020-07-31 15:07:28 +00:00
2020-05-25 08:04:30 +00:00
typedef struct thr_state_t thr_state_t ;
2020-05-26 01:13:34 +00:00
struct thr_peer_xtn_t
{
thr_state_t * state ;
} ;
typedef struct thr_peer_xtn_t thr_peer_xtn_t ;
static void thr_state_halt_participating_devices ( thr_state_t * thr_state )
{
MIO_ASSERT ( thr_state - > client - > htts - > mio , thr_state - > client ! = MIO_NULL ) ;
MIO_ASSERT ( thr_state - > client - > htts - > mio , thr_state - > client - > sck ! = MIO_NULL ) ;
MIO_DEBUG4 ( thr_state - > client - > htts - > mio , " HTTS(%p) - Halting participating devices in thr state %p(client=%p,peer=%p) \n " , thr_state - > client - > htts , thr_state , thr_state - > client - > sck , thr_state - > peer ) ;
mio_dev_sck_halt ( thr_state - > client - > sck ) ;
/* check for peer as it may not have been started */
if ( thr_state - > peer ) mio_dev_thr_halt ( thr_state - > peer ) ;
}
static int thr_state_write_to_client ( thr_state_t * thr_state , const void * data , mio_iolen_t dlen )
{
thr_state - > ever_attempted_to_write_to_client = 1 ;
thr_state - > num_pending_writes_to_client + + ;
if ( mio_dev_sck_write ( thr_state - > client - > sck , data , dlen , MIO_NULL , MIO_NULL ) < = - 1 )
{
thr_state - > num_pending_writes_to_client - - ;
return - 1 ;
}
if ( thr_state - > num_pending_writes_to_client > THR_STATE_PENDING_IO_THRESHOLD )
{
if ( mio_dev_thr_read ( thr_state - > peer , 0 ) < = - 1 ) return - 1 ;
}
return 0 ;
}
static int thr_state_writev_to_client ( thr_state_t * thr_state , mio_iovec_t * iov , mio_iolen_t iovcnt )
{
thr_state - > ever_attempted_to_write_to_client = 1 ;
thr_state - > num_pending_writes_to_client + + ;
if ( mio_dev_sck_writev ( thr_state - > client - > sck , iov , iovcnt , MIO_NULL , MIO_NULL ) < = - 1 )
{
thr_state - > num_pending_writes_to_client - - ;
return - 1 ;
}
if ( thr_state - > num_pending_writes_to_client > THR_STATE_PENDING_IO_THRESHOLD )
{
if ( mio_dev_thr_read ( thr_state - > peer , 0 ) < = - 1 ) return - 1 ;
}
return 0 ;
}
static int thr_state_send_final_status_to_client ( thr_state_t * thr_state , int status_code , int force_close )
{
mio_svc_htts_cli_t * cli = thr_state - > client ;
mio_bch_t dtbuf [ 64 ] ;
mio_svc_htts_fmtgmtime ( cli - > htts , MIO_NULL , dtbuf , MIO_COUNTOF ( dtbuf ) ) ;
if ( ! force_close ) force_close = ! thr_state - > keep_alive ;
if ( mio_becs_fmt ( cli - > sbuf , " HTTP/%d.%d %d %hs \r \n Server: %hs \r \n Date: %s \r \n Connection: %hs \r \n Content-Length: 0 \r \n \r \n " ,
thr_state - > req_version . major , thr_state - > req_version . minor ,
status_code , mio_http_status_to_bcstr ( status_code ) ,
cli - > htts - > server_name , dtbuf ,
( force_close ? " close " : " keep-alive " ) ) = = ( mio_oow_t ) - 1 ) return - 1 ;
return ( thr_state_write_to_client ( thr_state , MIO_BECS_PTR ( cli - > sbuf ) , MIO_BECS_LEN ( cli - > sbuf ) ) < = - 1 | |
( force_close & & thr_state_write_to_client ( thr_state , MIO_NULL , 0 ) < = - 1 ) ) ? - 1 : 0 ;
}
static int thr_state_write_last_chunk_to_client ( thr_state_t * thr_state )
{
if ( ! thr_state - > ever_attempted_to_write_to_client )
{
if ( thr_state_send_final_status_to_client ( thr_state , 500 , 0 ) < = - 1 ) return - 1 ;
}
else
{
if ( thr_state - > res_mode_to_cli = = THR_STATE_RES_MODE_CHUNKED & &
thr_state_write_to_client ( thr_state , " 0 \r \n \r \n " , 5 ) < = - 1 ) return - 1 ;
}
if ( ! thr_state - > keep_alive & & thr_state_write_to_client ( thr_state , MIO_NULL , 0 ) < = - 1 ) return - 1 ;
return 0 ;
}
static int thr_state_write_to_peer ( thr_state_t * thr_state , const void * data , mio_iolen_t dlen )
{
thr_state - > num_pending_writes_to_peer + + ;
if ( mio_dev_thr_write ( thr_state - > peer , data , dlen , MIO_NULL ) < = - 1 )
{
thr_state - > num_pending_writes_to_peer - - ;
return - 1 ;
}
/* TODO: check if it's already finished or something.. */
if ( thr_state - > num_pending_writes_to_peer > THR_STATE_PENDING_IO_THRESHOLD )
{
if ( mio_dev_sck_read ( thr_state - > client - > sck , 0 ) < = - 1 ) return - 1 ;
}
return 0 ;
}
static MIO_INLINE void thr_state_mark_over ( thr_state_t * thr_state , int over_bits )
{
unsigned int old_over ;
old_over = thr_state - > over ;
thr_state - > over | = over_bits ;
MIO_DEBUG5 ( thr_state - > htts - > mio , " HTTS(%p) - client=%p peer=%p new-bits=%x over=%x \n " , thr_state - > htts , thr_state - > client - > sck , thr_state - > peer , ( int ) over_bits , ( int ) thr_state - > over ) ;
if ( ! ( old_over & THR_STATE_OVER_READ_FROM_CLIENT ) & & ( thr_state - > over & THR_STATE_OVER_READ_FROM_CLIENT ) )
{
if ( mio_dev_sck_read ( thr_state - > client - > sck , 0 ) < = - 1 )
{
MIO_DEBUG2 ( thr_state - > htts - > mio , " HTTS(%p) - halting client(%p) for failure to disable input watching \n " , thr_state - > htts , thr_state - > client - > sck ) ;
mio_dev_sck_halt ( thr_state - > client - > sck ) ;
}
}
if ( ! ( old_over & THR_STATE_OVER_READ_FROM_PEER ) & & ( thr_state - > over & THR_STATE_OVER_READ_FROM_PEER ) )
{
if ( thr_state - > peer & & mio_dev_thr_read ( thr_state - > peer , 0 ) < = - 1 )
{
MIO_DEBUG2 ( thr_state - > htts - > mio , " HTTS(%p) - halting peer(%p) for failure to disable input watching \n " , thr_state - > htts , thr_state - > peer ) ;
mio_dev_thr_halt ( thr_state - > peer ) ;
}
}
if ( old_over ! = THR_STATE_OVER_ALL & & thr_state - > over = = THR_STATE_OVER_ALL )
{
/* ready to stop */
if ( thr_state - > peer )
{
MIO_DEBUG2 ( thr_state - > htts - > mio , " HTTS(%p) - halting peer(%p) as it is unneeded \n " , thr_state - > htts , thr_state - > peer ) ;
mio_dev_thr_halt ( thr_state - > peer ) ;
}
if ( thr_state - > keep_alive )
{
/* how to arrange to delete this thr_state object and put the socket back to the normal waiting state??? */
MIO_ASSERT ( thr_state - > htts - > mio , thr_state - > client - > rsrc = = ( mio_svc_htts_rsrc_t * ) thr_state ) ;
MIO_SVC_HTTS_RSRC_DETACH ( thr_state - > client - > rsrc ) ;
/* thr_state must not be access from here down as it could have been destroyed */
}
else
{
MIO_DEBUG2 ( thr_state - > htts - > mio , " HTTS(%p) - halting client(%p) for no keep-alive \n " , thr_state - > htts , thr_state - > client - > sck ) ;
mio_dev_sck_shutdown ( thr_state - > client - > sck , MIO_DEV_SCK_SHUTDOWN_WRITE ) ;
mio_dev_sck_halt ( thr_state - > client - > sck ) ;
}
}
}
2020-05-25 08:04:30 +00:00
static void thr_state_on_kill ( thr_state_t * thr_state )
{
2020-07-08 05:24:40 +00:00
mio_t * mio = thr_state - > htts - > mio ;
MIO_DEBUG2 ( mio , " HTTS(%p) - killing thr_state client(%p) \n " , thr_state - > htts , thr_state - > client - > sck ) ;
2020-05-26 01:13:34 +00:00
if ( thr_state - > peer )
{
thr_peer_xtn_t * thr_peer = mio_dev_thr_getxtn ( thr_state - > peer ) ;
thr_peer - > state = MIO_NULL ; /* thr_peer->state many not be NULL if the resource is killed regardless of the reference count */
mio_dev_thr_kill ( thr_state - > peer ) ;
thr_state - > peer = MIO_NULL ;
}
if ( thr_state - > peer_htrd )
{
thr_peer_xtn_t * thr_peer = mio_htrd_getxtn ( thr_state - > peer_htrd ) ;
thr_peer - > state = MIO_NULL ; /* thr_peer->state many not be NULL if the resource is killed regardless of the reference count */
mio_htrd_close ( thr_state - > peer_htrd ) ;
thr_state - > peer_htrd = MIO_NULL ;
}
if ( thr_state - > client_org_on_read )
{
thr_state - > client - > sck - > on_read = thr_state - > client_org_on_read ;
thr_state - > client_org_on_read = MIO_NULL ;
}
if ( thr_state - > client_org_on_write )
{
thr_state - > client - > sck - > on_write = thr_state - > client_org_on_write ;
thr_state - > client_org_on_write = MIO_NULL ;
}
if ( thr_state - > client_org_on_disconnect )
{
thr_state - > client - > sck - > on_disconnect = thr_state - > client_org_on_disconnect ;
thr_state - > client_org_on_disconnect = MIO_NULL ;
}
2020-05-26 14:58:55 +00:00
if ( thr_state - > client_htrd_recbs_changed )
{
/* restore the callbacks */
mio_htrd_setrecbs ( thr_state - > client - > htrd , & thr_state - > client_htrd_org_recbs ) ;
}
2020-05-26 01:13:34 +00:00
if ( ! thr_state - > client_disconnected )
{
2020-07-08 05:24:40 +00:00
/*printf ("ENABLING INPUT WATCHING on CLIENT %p. \n", thr_state->client->sck);*/
2020-05-26 01:13:34 +00:00
if ( ! thr_state - > keep_alive | | mio_dev_sck_read ( thr_state - > client - > sck , 1 ) < = - 1 )
{
2020-07-08 05:24:40 +00:00
MIO_DEBUG2 ( mio , " HTTS(%p) - halting client(%p) for failure to enable input watching \n " , thr_state - > htts , thr_state - > client - > sck ) ;
2020-05-26 01:13:34 +00:00
mio_dev_sck_halt ( thr_state - > client - > sck ) ;
}
}
2020-07-08 05:24:40 +00:00
/*printf ("**** THR_STATE_ON_KILL DONE\n");*/
2020-05-26 01:13:34 +00:00
}
static void thr_peer_on_close ( mio_dev_thr_t * thr , mio_dev_thr_sid_t sid )
{
mio_t * mio = thr - > mio ;
thr_peer_xtn_t * thr_peer = ( thr_peer_xtn_t * ) mio_dev_thr_getxtn ( thr ) ;
thr_state_t * thr_state = thr_peer - > state ;
if ( ! thr_state ) return ; /* thr state already gone */
switch ( sid )
{
case MIO_DEV_THR_MASTER :
MIO_DEBUG2 ( mio , " HTTS(%p) - peer %p closing master \n " , thr_state - > client - > htts , thr ) ;
thr_state - > peer = MIO_NULL ; /* clear this peer from the state */
MIO_ASSERT ( mio , thr_peer - > state ! = MIO_NULL ) ;
MIO_SVC_HTTS_RSRC_DETACH ( thr_peer - > state ) ;
if ( thr_state - > peer_htrd )
{
/* once this peer device is closed, peer's htrd is also never used.
* it ' s safe to detach the extra information attached on the htrd object . */
thr_peer = mio_htrd_getxtn ( thr_state - > peer_htrd ) ;
MIO_ASSERT ( mio , thr_peer - > state ! = MIO_NULL ) ;
MIO_SVC_HTTS_RSRC_DETACH ( thr_peer - > state ) ;
}
break ;
case MIO_DEV_THR_OUT :
MIO_ASSERT ( mio , thr_state - > peer = = thr ) ;
MIO_DEBUG3 ( mio , " HTTS(%p) - peer %p closing slave[%d] \n " , thr_state - > client - > htts , thr , sid ) ;
if ( ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_PEER ) )
{
if ( thr_state_write_last_chunk_to_client ( thr_state ) < = - 1 )
thr_state_halt_participating_devices ( thr_state ) ;
else
thr_state_mark_over ( thr_state , THR_STATE_OVER_READ_FROM_PEER ) ;
}
break ;
case MIO_DEV_THR_IN :
thr_state_mark_over ( thr_state , THR_STATE_OVER_WRITE_TO_PEER ) ;
break ;
default :
MIO_DEBUG3 ( mio , " HTTS(%p) - peer %p closing slave[%d] \n " , thr_state - > client - > htts , thr , sid ) ;
/* do nothing */
break ;
}
}
static int thr_peer_on_read ( mio_dev_thr_t * thr , const void * data , mio_iolen_t dlen )
{
mio_t * mio = thr - > mio ;
thr_peer_xtn_t * thr_peer = ( thr_peer_xtn_t * ) mio_dev_thr_getxtn ( thr ) ;
thr_state_t * thr_state = thr_peer - > state ;
MIO_ASSERT ( mio , thr_state ! = MIO_NULL ) ;
if ( dlen < = - 1 )
{
MIO_DEBUG2 ( mio , " HTTPS(%p) - read error from peer %p \n " , thr_state - > client - > htts , thr ) ;
goto oops ;
}
if ( dlen = = 0 )
{
MIO_DEBUG2 ( mio , " HTTPS(%p) - EOF from peer %p \n " , thr_state - > client - > htts , thr ) ;
if ( ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_PEER ) )
{
/* the thr script could be misbehaviing.
* it still has to read more but EOF is read .
* otherwise client_peer_htrd_poke ( ) should have been called */
if ( thr_state_write_last_chunk_to_client ( thr_state ) < = - 1 ) goto oops ;
thr_state_mark_over ( thr_state , THR_STATE_OVER_READ_FROM_PEER ) ;
}
}
else
{
mio_oow_t rem ;
MIO_ASSERT ( mio , ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_PEER ) ) ;
if ( mio_htrd_feed ( thr_state - > peer_htrd , data , dlen , & rem ) < = - 1 )
{
2020-07-16 10:46:17 +00:00
MIO_DEBUG2 ( mio , " HTTPS(%p) - unable to feed peer htrd - peer %p \n " , thr_state - > htts , thr ) ;
2020-05-26 01:13:34 +00:00
if ( ! thr_state - > ever_attempted_to_write_to_client & &
! ( thr_state - > over & THR_STATE_OVER_WRITE_TO_CLIENT ) )
{
thr_state_send_final_status_to_client ( thr_state , 500 , 1 ) ; /* don't care about error because it jumps to oops below anyway */
}
goto oops ;
}
if ( rem > 0 )
{
/* If the script specifies Content-Length and produces longer data, it will come here */
printf ( " AAAAAAAAAAAAAAAAAa EEEEEXcessive DATA.................. \n " ) ;
/* TODO: or drop this request?? */
}
}
return 0 ;
oops :
thr_state_halt_participating_devices ( thr_state ) ;
return 0 ;
}
static int thr_peer_capture_response_header ( mio_htre_t * req , const mio_bch_t * key , const mio_htre_hdrval_t * val , void * ctx )
{
mio_svc_htts_cli_t * cli = ( mio_svc_htts_cli_t * ) ctx ;
/* capture a header except Status, Connection, Transfer-Encoding, and Server */
if ( mio_comp_bcstr ( key , " Status " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Connection " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Transfer-Encoding " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Server " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Date " , 1 ) ! = 0 )
{
do
{
if ( mio_becs_cat ( cli - > sbuf , key ) = = ( mio_oow_t ) - 1 | |
mio_becs_cat ( cli - > sbuf , " : " ) = = ( mio_oow_t ) - 1 | |
mio_becs_cat ( cli - > sbuf , val - > ptr ) = = ( mio_oow_t ) - 1 | |
mio_becs_cat ( cli - > sbuf , " \r \n " ) = = ( mio_oow_t ) - 1 )
{
return - 1 ;
}
val = val - > next ;
}
while ( val ) ;
}
return 0 ;
}
static int thr_peer_htrd_peek ( mio_htrd_t * htrd , mio_htre_t * req )
{
thr_peer_xtn_t * thr_peer = mio_htrd_getxtn ( htrd ) ;
thr_state_t * thr_state = thr_peer - > state ;
mio_svc_htts_cli_t * cli = thr_state - > client ;
mio_bch_t dtbuf [ 64 ] ;
int status_code = 200 ;
if ( req - > attr . content_length )
{
// TOOD: remove content_length if content_length is negative or not numeric.
thr_state - > res_mode_to_cli = THR_STATE_RES_MODE_LENGTH ;
}
if ( req - > attr . status )
{
int is_sober ;
const mio_bch_t * endptr ;
mio_intmax_t v ;
2021-07-12 14:40:51 +00:00
v = mio_bchars_to_intmax ( req - > attr . status , mio_count_bcstr ( req - > attr . status ) , MIO_BCHARS_TO_INTMAX_MAKE_OPTION ( 0 , 0 , 0 , 10 ) , & endptr , & is_sober ) ;
2020-05-26 01:13:34 +00:00
if ( * endptr = = ' \0 ' & & is_sober & & v > 0 & & v < = MIO_TYPE_MAX ( int ) ) status_code = v ;
}
mio_svc_htts_fmtgmtime ( cli - > htts , MIO_NULL , dtbuf , MIO_COUNTOF ( dtbuf ) ) ;
if ( mio_becs_fmt ( cli - > sbuf , " HTTP/%d.%d %d %hs \r \n Server: %hs \r \n Date: %hs \r \n " ,
thr_state - > req_version . major , thr_state - > req_version . minor ,
status_code , mio_http_status_to_bcstr ( status_code ) ,
cli - > htts - > server_name , dtbuf ) = = ( mio_oow_t ) - 1 ) return - 1 ;
if ( mio_htre_walkheaders ( req , thr_peer_capture_response_header , cli ) < = - 1 ) return - 1 ;
switch ( thr_state - > res_mode_to_cli )
{
case THR_STATE_RES_MODE_CHUNKED :
if ( mio_becs_cat ( cli - > sbuf , " Transfer-Encoding: chunked \r \n " ) = = ( mio_oow_t ) - 1 ) return - 1 ;
/*if (mio_becs_cat(cli->sbuf, "Connection: keep-alive\r\n") == (mio_oow_t)-1) return -1;*/
break ;
case THR_STATE_RES_MODE_CLOSE :
if ( mio_becs_cat ( cli - > sbuf , " Connection: close \r \n " ) = = ( mio_oow_t ) - 1 ) return - 1 ;
break ;
case THR_STATE_RES_MODE_LENGTH :
if ( mio_becs_cat ( cli - > sbuf , ( thr_state - > keep_alive ? " Connection: keep-alive \r \n " : " Connection: close \r \n " ) ) = = ( mio_oow_t ) - 1 ) return - 1 ;
}
if ( mio_becs_cat ( cli - > sbuf , " \r \n " ) = = ( mio_oow_t ) - 1 ) return - 1 ;
return thr_state_write_to_client ( thr_state , MIO_BECS_PTR ( cli - > sbuf ) , MIO_BECS_LEN ( cli - > sbuf ) ) ;
}
static int thr_peer_htrd_poke ( mio_htrd_t * htrd , mio_htre_t * req )
{
/* client request got completed */
thr_peer_xtn_t * thr_peer = mio_htrd_getxtn ( htrd ) ;
thr_state_t * thr_state = thr_peer - > state ;
if ( thr_state_write_last_chunk_to_client ( thr_state ) < = - 1 ) return - 1 ;
thr_state_mark_over ( thr_state , THR_STATE_OVER_READ_FROM_PEER ) ;
return 0 ;
}
static int thr_peer_htrd_push_content ( mio_htrd_t * htrd , mio_htre_t * req , const mio_bch_t * data , mio_oow_t dlen )
{
thr_peer_xtn_t * thr_peer = mio_htrd_getxtn ( htrd ) ;
thr_state_t * thr_state = thr_peer - > state ;
MIO_ASSERT ( thr_state - > client - > htts - > mio , htrd = = thr_state - > peer_htrd ) ;
switch ( thr_state - > res_mode_to_cli )
{
case THR_STATE_RES_MODE_CHUNKED :
{
mio_iovec_t iov [ 3 ] ;
mio_bch_t lbuf [ 16 ] ;
mio_oow_t llen ;
/* mio_fmt_uintmax_to_bcstr() null-terminates the output. only MIO_COUNTOF(lbuf) - 1
* is enough to hold ' \r ' and ' \n ' at the back without ' \0 ' . */
llen = mio_fmt_uintmax_to_bcstr ( lbuf , MIO_COUNTOF ( lbuf ) - 1 , dlen , 16 | MIO_FMT_UINTMAX_UPPERCASE , 0 , ' \0 ' , MIO_NULL ) ;
lbuf [ llen + + ] = ' \r ' ;
lbuf [ llen + + ] = ' \n ' ;
iov [ 0 ] . iov_ptr = lbuf ;
iov [ 0 ] . iov_len = llen ;
iov [ 1 ] . iov_ptr = ( void * ) data ;
iov [ 1 ] . iov_len = dlen ;
iov [ 2 ] . iov_ptr = " \r \n " ;
iov [ 2 ] . iov_len = 2 ;
2020-07-16 10:46:17 +00:00
if ( thr_state_writev_to_client ( thr_state , iov , MIO_COUNTOF ( iov ) ) < = - 1 )
{
goto oops ;
}
2020-05-26 01:13:34 +00:00
break ;
}
case THR_STATE_RES_MODE_CLOSE :
case THR_STATE_RES_MODE_LENGTH :
2020-07-16 10:46:17 +00:00
if ( thr_state_write_to_client ( thr_state , data , dlen ) < = - 1 )
{
goto oops ;
}
2020-05-26 01:13:34 +00:00
break ;
}
if ( thr_state - > num_pending_writes_to_client > THR_STATE_PENDING_IO_THRESHOLD )
{
2020-07-16 10:46:17 +00:00
if ( mio_dev_thr_read ( thr_state - > peer , 0 ) < = - 1 )
{
goto oops ;
}
2020-05-26 01:13:34 +00:00
}
return 0 ;
oops :
return - 1 ;
}
static mio_htrd_recbs_t thr_peer_htrd_recbs =
{
thr_peer_htrd_peek ,
thr_peer_htrd_poke ,
thr_peer_htrd_push_content
} ;
static int thr_client_htrd_poke ( mio_htrd_t * htrd , mio_htre_t * req )
{
/* client request got completed */
mio_svc_htts_cli_htrd_xtn_t * htrdxtn = ( mio_svc_htts_cli_htrd_xtn_t * ) mio_htrd_getxtn ( htrd ) ;
mio_dev_sck_t * sck = htrdxtn - > sck ;
mio_svc_htts_cli_t * cli = mio_dev_sck_getxtn ( sck ) ;
thr_state_t * thr_state = ( thr_state_t * ) cli - > rsrc ;
/* indicate EOF to the client peer */
if ( thr_state_write_to_peer ( thr_state , MIO_NULL , 0 ) < = - 1 ) return - 1 ;
thr_state_mark_over ( thr_state , THR_STATE_OVER_READ_FROM_CLIENT ) ;
return 0 ;
2020-05-25 08:04:30 +00:00
}
2020-05-26 01:13:34 +00:00
static int thr_client_htrd_push_content ( mio_htrd_t * htrd , mio_htre_t * req , const mio_bch_t * data , mio_oow_t dlen )
2020-05-25 08:04:30 +00:00
{
2020-05-26 01:13:34 +00:00
mio_svc_htts_cli_htrd_xtn_t * htrdxtn = ( mio_svc_htts_cli_htrd_xtn_t * ) mio_htrd_getxtn ( htrd ) ;
mio_dev_sck_t * sck = htrdxtn - > sck ;
mio_svc_htts_cli_t * cli = mio_dev_sck_getxtn ( sck ) ;
thr_state_t * thr_state = ( thr_state_t * ) cli - > rsrc ;
2020-05-25 08:04:30 +00:00
2020-05-26 01:13:34 +00:00
MIO_ASSERT ( sck - > mio , cli - > sck = = sck ) ;
return thr_state_write_to_peer ( thr_state , data , dlen ) ;
2020-05-25 08:04:30 +00:00
}
2020-05-26 01:13:34 +00:00
static mio_htrd_recbs_t thr_client_htrd_recbs =
2020-05-25 08:04:30 +00:00
{
2020-05-26 01:13:34 +00:00
MIO_NULL ,
thr_client_htrd_poke ,
thr_client_htrd_push_content
} ;
static int thr_peer_on_write ( mio_dev_thr_t * thr , mio_iolen_t wrlen , void * wrctx )
{
mio_t * mio = thr - > mio ;
thr_peer_xtn_t * thr_peer = ( thr_peer_xtn_t * ) mio_dev_thr_getxtn ( thr ) ;
thr_state_t * thr_state = thr_peer - > state ;
if ( thr_state = = MIO_NULL ) return 0 ; /* there is nothing i can do. the thr_state is being cleared or has been cleared already. */
MIO_ASSERT ( mio , thr_state - > peer = = thr ) ;
if ( wrlen < = - 1 )
{
MIO_DEBUG2 ( mio , " HTTS(%p) - unable to write to peer %p \n " , thr_state - > client - > htts , thr ) ;
goto oops ;
}
else if ( wrlen = = 0 )
{
/* indicated EOF */
/* do nothing here as i didn't incremented num_pending_writes_to_peer when making the write request */
thr_state - > num_pending_writes_to_peer - - ;
MIO_ASSERT ( mio , thr_state - > num_pending_writes_to_peer = = 0 ) ;
MIO_DEBUG2 ( mio , " HTTS(%p) - indicated EOF to peer %p \n " , thr_state - > client - > htts , thr ) ;
/* indicated EOF to the peer side. i need no more data from the client side.
* i don ' t need to enable input watching in the client side either */
thr_state_mark_over ( thr_state , THR_STATE_OVER_WRITE_TO_PEER ) ;
}
else
{
MIO_ASSERT ( mio , thr_state - > num_pending_writes_to_peer > 0 ) ;
thr_state - > num_pending_writes_to_peer - - ;
if ( thr_state - > num_pending_writes_to_peer = = THR_STATE_PENDING_IO_THRESHOLD )
{
if ( ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_CLIENT ) & &
mio_dev_sck_read ( thr_state - > client - > sck , 1 ) < = - 1 ) goto oops ;
}
if ( ( thr_state - > over & THR_STATE_OVER_READ_FROM_CLIENT ) & & thr_state - > num_pending_writes_to_peer < = 0 )
{
thr_state_mark_over ( thr_state , THR_STATE_OVER_WRITE_TO_PEER ) ;
}
}
return 0 ;
oops :
thr_state_halt_participating_devices ( thr_state ) ;
return 0 ;
}
static void thr_client_on_disconnect ( mio_dev_sck_t * sck )
{
mio_svc_htts_cli_t * cli = mio_dev_sck_getxtn ( sck ) ;
thr_state_t * thr_state = ( thr_state_t * ) cli - > rsrc ;
thr_state - > client_disconnected = 1 ;
thr_state - > client_org_on_disconnect ( sck ) ;
}
static int thr_client_on_read ( mio_dev_sck_t * sck , const void * buf , mio_iolen_t len , const mio_skad_t * srcaddr )
{
mio_t * mio = sck - > mio ;
mio_svc_htts_cli_t * cli = mio_dev_sck_getxtn ( sck ) ;
thr_state_t * thr_state = ( thr_state_t * ) cli - > rsrc ;
MIO_ASSERT ( mio , sck = = cli - > sck ) ;
if ( len < = - 1 )
{
/* read error */
MIO_DEBUG2 ( cli - > htts - > mio , " HTTPS(%p) - read error on client %p(%d) \n " , sck , ( int ) sck - > hnd ) ;
goto oops ;
}
if ( ! thr_state - > peer )
{
/* the peer is gone */
goto oops ; /* do what? just return 0? */
}
if ( len = = 0 )
{
/* EOF on the client side. arrange to close */
MIO_DEBUG3 ( mio , " HTTPS(%p) - EOF from client %p(hnd=%d) \n " , thr_state - > client - > htts , sck , ( int ) sck - > hnd ) ;
if ( ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_CLIENT ) ) /* if this is true, EOF is received without thr_client_htrd_poke() */
{
if ( thr_state_write_to_peer ( thr_state , MIO_NULL , 0 ) < = - 1 ) goto oops ;
thr_state_mark_over ( thr_state , THR_STATE_OVER_READ_FROM_CLIENT ) ;
}
}
else
{
mio_oow_t rem ;
MIO_ASSERT ( mio , ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_CLIENT ) ) ;
if ( mio_htrd_feed ( cli - > htrd , buf , len , & rem ) < = - 1 ) goto oops ;
if ( rem > 0 )
{
/* TODO store this to client buffer. once the current resource is completed, arrange to call on_read() with it */
2020-07-15 10:30:12 +00:00
MIO_DEBUG3 ( mio , " HTTPS(%p) - excessive data after contents by thr client %p(%d) \n " , sck - > mio , sck , ( int ) sck - > hnd ) ;
2020-05-26 01:13:34 +00:00
}
}
return 0 ;
oops :
thr_state_halt_participating_devices ( thr_state ) ;
return 0 ;
}
static int thr_client_on_write ( mio_dev_sck_t * sck , mio_iolen_t wrlen , void * wrctx , const mio_skad_t * dstaddr )
{
mio_t * mio = sck - > mio ;
mio_svc_htts_cli_t * cli = mio_dev_sck_getxtn ( sck ) ;
thr_state_t * thr_state = ( thr_state_t * ) cli - > rsrc ;
if ( wrlen < = - 1 )
{
MIO_DEBUG3 ( mio , " HTTPS(%p) - unable to write to client %p(%d) \n " , sck - > mio , sck , ( int ) sck - > hnd ) ;
goto oops ;
}
if ( wrlen = = 0 )
{
/* if the connect is keep-alive, this part may not be called */
thr_state - > num_pending_writes_to_client - - ;
MIO_ASSERT ( mio , thr_state - > num_pending_writes_to_client = = 0 ) ;
MIO_DEBUG3 ( mio , " HTTS(%p) - indicated EOF to client %p(%d) \n " , thr_state - > client - > htts , sck , ( int ) sck - > hnd ) ;
/* since EOF has been indicated to the client, it must not write to the client any further.
* this also means that i don ' t need any data from the peer side either .
* i don ' t need to enable input watching on the peer side */
thr_state_mark_over ( thr_state , THR_STATE_OVER_WRITE_TO_CLIENT ) ;
}
else
{
MIO_ASSERT ( mio , thr_state - > num_pending_writes_to_client > 0 ) ;
thr_state - > num_pending_writes_to_client - - ;
if ( thr_state - > peer & & thr_state - > num_pending_writes_to_client = = THR_STATE_PENDING_IO_THRESHOLD )
{
if ( ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_PEER ) & &
mio_dev_thr_read ( thr_state - > peer , 1 ) < = - 1 ) goto oops ;
}
if ( ( thr_state - > over & THR_STATE_OVER_READ_FROM_PEER ) & & thr_state - > num_pending_writes_to_client < = 0 )
{
thr_state_mark_over ( thr_state , THR_STATE_OVER_WRITE_TO_CLIENT ) ;
}
}
return 0 ;
oops :
thr_state_halt_participating_devices ( thr_state ) ;
return 0 ;
}
static void free_thr_start_info ( void * ctx )
{
thr_func_start_t * tfs = ( thr_func_start_t * ) ctx ;
if ( tfs - > tfi . req_path ) mio_freemem ( tfs - > mio , tfs - > tfi . req_path ) ;
if ( tfs - > tfi . req_param ) mio_freemem ( tfs - > mio , tfs - > tfi . req_param ) ;
mio_freemem ( tfs - > mio , tfs ) ;
}
static void thr_func ( mio_t * mio , mio_dev_thr_iopair_t * iop , void * ctx )
{
thr_func_start_t * tfs = ( thr_func_start_t * ) ctx ;
pthread_cleanup_push ( free_thr_start_info , tfs ) ;
tfs - > thr_func ( mio , iop , & tfs - > tfi , tfs - > thr_ctx ) ;
pthread_cleanup_pop ( 1 ) ;
}
2020-06-08 18:11:36 +00:00
static int thr_capture_request_header ( mio_htre_t * req , const mio_bch_t * key , const mio_htre_hdrval_t * val , void * ctx )
{
thr_func_start_t * tfs = ( thr_func_start_t * ) ctx ;
if ( mio_comp_bcstr ( key , " X-HTTP-Method-Override " , 1 ) = = 0 )
{
tfs - > tfi . req_x_http_method_override = mio_bchars_to_http_method ( val - > ptr , val - > len ) ; /* don't care about multiple values */
}
#if 0
if ( mio_comp_bcstr ( key , " Connection " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Transfer-Encoding " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Content-Length " , 1 ) ! = 0 & &
mio_comp_bcstr ( key , " Expect " , 1 ) ! = 0 )
{
do
{
/* TODO: ... */
val = val - > next ;
}
while ( val ) ;
}
# endif
return 0 ;
}
2020-05-26 01:13:34 +00:00
int mio_svc_htts_dothr ( mio_svc_htts_t * htts , mio_dev_sck_t * csck , mio_htre_t * req , mio_svc_htts_thr_func_t func , void * ctx )
{
mio_t * mio = htts - > mio ;
mio_svc_htts_cli_t * cli = mio_dev_sck_getxtn ( csck ) ;
2020-05-25 08:04:30 +00:00
thr_state_t * thr_state = MIO_NULL ;
2020-05-26 01:13:34 +00:00
thr_peer_xtn_t * thr_peer ;
mio_dev_thr_make_t mi ;
thr_func_start_t * tfs ;
/* ensure that you call this function before any contents is received */
MIO_ASSERT ( mio , mio_htre_getcontentlen ( req ) = = 0 ) ;
tfs = mio_callocmem ( mio , MIO_SIZEOF ( * tfs ) ) ;
if ( ! tfs ) goto oops ;
tfs - > mio = mio ;
tfs - > thr_func = func ;
tfs - > thr_ctx = ctx ;
tfs - > tfi . req_method = mio_htre_getqmethodtype ( req ) ;
tfs - > tfi . req_version = * mio_htre_getversion ( req ) ;
tfs - > tfi . req_path = mio_dupbcstr ( mio , mio_htre_getqpath ( req ) , MIO_NULL ) ;
if ( ! tfs - > tfi . req_path ) goto oops ;
if ( mio_htre_getqparam ( req ) )
{
tfs - > tfi . req_param = mio_dupbcstr ( mio , mio_htre_getqparam ( req ) , MIO_NULL ) ;
if ( ! tfs - > tfi . req_param ) goto oops ;
}
2020-06-08 18:11:36 +00:00
tfs - > tfi . req_x_http_method_override = - 1 ;
2020-07-27 10:30:16 +00:00
if ( mio_htre_walkheaders ( req , thr_capture_request_header , tfs ) < = - 1 ) goto oops ;
2020-06-08 18:11:36 +00:00
2020-05-26 01:13:34 +00:00
tfs - > tfi . server_addr = cli - > sck - > localaddr ;
tfs - > tfi . client_addr = cli - > sck - > remoteaddr ;
MIO_MEMSET ( & mi , 0 , MIO_SIZEOF ( mi ) ) ;
mi . thr_func = thr_func ;
mi . thr_ctx = tfs ;
mi . on_read = thr_peer_on_read ;
mi . on_write = thr_peer_on_write ;
mi . on_close = thr_peer_on_close ;
2020-05-25 08:04:30 +00:00
thr_state = ( thr_state_t * ) mio_svc_htts_rsrc_make ( htts , MIO_SIZEOF ( * thr_state ) , thr_state_on_kill ) ;
if ( MIO_UNLIKELY ( ! thr_state ) ) goto oops ;
2020-05-26 01:13:34 +00:00
thr_state - > client = cli ;
/*thr_state->num_pending_writes_to_client = 0;
thr_state - > num_pending_writes_to_peer = 0 ; */
thr_state - > req_version = * mio_htre_getversion ( req ) ;
2020-05-26 13:15:25 +00:00
thr_state - > req_content_length_unlimited = mio_htre_getreqcontentlen ( req , & thr_state - > req_content_length ) ;
2020-05-26 01:13:34 +00:00
thr_state - > client_org_on_read = csck - > on_read ;
thr_state - > client_org_on_write = csck - > on_write ;
thr_state - > client_org_on_disconnect = csck - > on_disconnect ;
csck - > on_read = thr_client_on_read ;
csck - > on_write = thr_client_on_write ;
csck - > on_disconnect = thr_client_on_disconnect ;
MIO_ASSERT ( mio , cli - > rsrc = = MIO_NULL ) ;
MIO_SVC_HTTS_RSRC_ATTACH ( thr_state , cli - > rsrc ) ;
thr_state - > peer = mio_dev_thr_make ( mio , MIO_SIZEOF ( * thr_peer ) , & mi ) ;
2020-07-27 12:10:55 +00:00
if ( MIO_UNLIKELY ( ! thr_state - > peer ) )
{
MIO_DEBUG3 ( mio , " HTTS(%p) - failed to create thread for %p(%d) \n " , htts , csck , ( int ) csck - > hnd ) ;
goto oops ;
}
2020-05-26 01:13:34 +00:00
tfs = MIO_NULL ; /* mark that tfs is delegated to the thread */
thr_peer = mio_dev_thr_getxtn ( thr_state - > peer ) ;
MIO_SVC_HTTS_RSRC_ATTACH ( thr_state , thr_peer - > state ) ;
thr_state - > peer_htrd = mio_htrd_open ( mio , MIO_SIZEOF ( * thr_peer ) ) ;
if ( MIO_UNLIKELY ( ! thr_state - > peer_htrd ) ) goto oops ;
mio_htrd_setopt ( thr_state - > peer_htrd , MIO_HTRD_SKIPINITIALLINE | MIO_HTRD_RESPONSE ) ;
mio_htrd_setrecbs ( thr_state - > peer_htrd , & thr_peer_htrd_recbs ) ;
2020-05-25 08:04:30 +00:00
2020-05-26 01:13:34 +00:00
thr_peer = mio_htrd_getxtn ( thr_state - > peer_htrd ) ;
MIO_SVC_HTTS_RSRC_ATTACH ( thr_state , thr_peer - > state ) ;
# if !defined(THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH)
if ( thr_state - > req_content_length_unlimited )
{
/* Transfer-Encoding is chunked. no content-length is known in advance. */
/* option 1. buffer contents. if it gets too large, send 413 Request Entity Too Large.
* option 2. send 411 Length Required immediately
* option 3. set Content - Length to - 1 and use EOF to indicate the end of content [ Non - Standard ] */
if ( thr_state_send_final_status_to_client ( thr_state , 411 , 1 ) < = - 1 ) goto oops ;
}
# endif
if ( req - > flags & MIO_HTRE_ATTR_EXPECT100 )
{
/* TODO: Expect: 100-continue? who should handle this? thr? or the http server? */
/* CAN I LET the thr SCRIPT handle this? */
if ( mio_comp_http_version_numbers ( & req - > version , 1 , 1 ) > = 0 & &
( thr_state - > req_content_length_unlimited | | thr_state - > req_content_length > 0 ) )
{
/*
* Don ' t send 100 Continue if http verions is lower than 1.1
* [ RFC7231 ]
* A server that receives a 100 - continue expectation in an HTTP / 1.0
* request MUST ignore that expectation .
*
* Don ' t send 100 Continue if expected content lenth is 0.
* [ RFC7231 ]
* A server MAY omit sending a 100 ( Continue ) response if it has
* already received some or all of the message body for the
* corresponding request , or if the framing indicates that there is
* no message body .
*/
mio_bch_t msgbuf [ 64 ] ;
mio_oow_t msglen ;
msglen = mio_fmttobcstr ( mio , msgbuf , MIO_COUNTOF ( msgbuf ) , " HTTP/%d.%d 100 Continue \r \n \r \n " , thr_state - > req_version . major , thr_state - > req_version . minor ) ;
if ( thr_state_write_to_client ( thr_state , msgbuf , msglen ) < = - 1 ) goto oops ;
thr_state - > ever_attempted_to_write_to_client = 0 ; /* reset this as it's polluted for 100 continue */
}
}
else if ( req - > flags & MIO_HTRE_ATTR_EXPECT )
{
/* 417 Expectation Failed */
thr_state_send_final_status_to_client ( thr_state , 417 , 1 ) ;
goto oops ;
}
# if defined(THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH)
if ( thr_state - > req_content_length_unlimited )
{
/* change the callbacks to subscribe to contents to be uploaded */
thr_state - > client_htrd_org_recbs = * mio_htrd_getrecbs ( thr_state - > client - > htrd ) ;
thr_client_htrd_recbs . peek = thr_state - > client_htrd_org_recbs . peek ;
mio_htrd_setrecbs ( thr_state - > client - > htrd , & thr_client_htrd_recbs ) ;
2020-05-26 14:58:55 +00:00
thr_state - > client_htrd_recbs_changed = 1 ;
2020-05-26 01:13:34 +00:00
}
else
{
# endif
if ( thr_state - > req_content_length > 0 )
{
/* change the callbacks to subscribe to contents to be uploaded */
thr_state - > client_htrd_org_recbs = * mio_htrd_getrecbs ( thr_state - > client - > htrd ) ;
thr_client_htrd_recbs . peek = thr_state - > client_htrd_org_recbs . peek ;
mio_htrd_setrecbs ( thr_state - > client - > htrd , & thr_client_htrd_recbs ) ;
2020-05-26 14:58:55 +00:00
thr_state - > client_htrd_recbs_changed = 1 ;
2020-05-26 01:13:34 +00:00
}
else
{
/* no content to be uploaded from the client */
/* indicate EOF to the peer and disable input wathching from the client */
if ( thr_state_write_to_peer ( thr_state , MIO_NULL , 0 ) < = - 1 ) goto oops ;
thr_state_mark_over ( thr_state , THR_STATE_OVER_READ_FROM_CLIENT | THR_STATE_OVER_WRITE_TO_PEER ) ;
}
# if defined(THR_ALLOW_UNLIMITED_REQ_CONTENT_LENGTH)
}
2020-05-25 08:04:30 +00:00
# endif
2020-05-26 01:13:34 +00:00
/* this may change later if Content-Length is included in the thr output */
if ( req - > flags & MIO_HTRE_ATTR_KEEPALIVE )
2020-05-25 08:04:30 +00:00
{
2020-05-26 01:13:34 +00:00
thr_state - > keep_alive = 1 ;
thr_state - > res_mode_to_cli = THR_STATE_RES_MODE_CHUNKED ;
/* the mode still can get switched to THR_STATE_RES_MODE_LENGTH if the thr script emits Content-Length */
}
else
{
thr_state - > keep_alive = 0 ;
thr_state - > res_mode_to_cli = THR_STATE_RES_MODE_CLOSE ;
2020-05-25 08:04:30 +00:00
}
2020-05-26 01:13:34 +00:00
/* TODO: store current input watching state and use it when destroying the thr_state data */
if ( mio_dev_sck_read ( csck , ! ( thr_state - > over & THR_STATE_OVER_READ_FROM_CLIENT ) ) < = - 1 ) goto oops ;
2020-05-25 08:04:30 +00:00
return 0 ;
oops :
2020-05-26 01:13:34 +00:00
MIO_DEBUG2 ( mio , " HTTS(%p) - FAILURE in dothr - socket(%p) \n " , htts , csck ) ;
if ( tfs ) free_thr_start_info ( tfs ) ;
if ( thr_state ) thr_state_halt_participating_devices ( thr_state ) ;
2020-05-25 08:04:30 +00:00
return - 1 ;
}