2023-11-26 15:13:28 +00:00
/*
Copyright ( c ) 2016 - 2018 Chung , Hyung - Hwan . All rights reserved .
Redistribution and use in source and binary forms , with or without
modification , are permitted provided that the following conditions
are met :
1. Redistributions of source code must retain the above copyright
notice , this list of conditions and the following disclaimer .
2. Redistributions in binary form must reproduce the above copyright
notice , this list of conditions and the following disclaimer in the
documentation and / or other materials provided with the distribution .
THIS SOFTWARE IS PROVIDED BY THE AUTHOR " AS IS " AND ANY EXPRESS OR
IMPLIED WARRANTIES , INCLUDING , BUT NOT LIMITED TO , THE IMPLIED WARRANTIES
OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED .
IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT , INDIRECT ,
INCIDENTAL , SPECIAL , EXEMPLARY , OR CONSEQUENTIAL DAMAGES ( INCLUDING , BUT
NOT LIMITED TO , PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES ; LOSS OF USE ,
DATA , OR PROFITS ; OR BUSINESS INTERRUPTION ) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY , WHETHER IN CONTRACT , STRICT LIABILITY , OR TORT
( INCLUDING NEGLIGENCE OR OTHERWISE ) ARISING IN ANY WAY OUT OF THE USE OF
THIS SOFTWARE , EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE .
*/
2024-04-14 09:33:15 +00:00
# include "hcl-x.h"
2023-11-26 15:13:28 +00:00
# include "hcl-prv.h"
# include "hcl-tmr.h"
# include "hcl-xutl.h"
# include <stdio.h>
# include <string.h>
# include <errno.h>
# define HCL_SERVER_TOKEN_NAME_ALIGN 64
# define HCL_SERVER_WID_MAP_ALIGN 512
2024-04-20 03:02:22 +00:00
# define HCL_XPROTO_REPLY_BUF_SIZE 1300
2023-11-26 15:13:28 +00:00
# if defined(_WIN32)
# include <windows.h>
# include <tchar.h>
# elif defined(__OS2__)
# define INCL_DOSMODULEMGR
# define INCL_DOSPROCESS
# define INCL_DOSERRORS
# include <os2.h>
2023-12-09 05:01:03 +00:00
# elif defined(__DOS__)
2023-11-26 15:13:28 +00:00
# include <dos.h>
# include <time.h>
2023-12-09 05:01:03 +00:00
# include <signal.h>
2023-11-26 15:13:28 +00:00
# elif defined(macintosh)
# include <Timer.h>
# else
# if defined(HAVE_TIME_H)
# include <time.h>
# endif
# if defined(HAVE_SYS_TIME_H)
# include <sys / time.h>
# endif
# if defined(HAVE_SIGNAL_H)
# include <signal.h>
# endif
# if defined(HAVE_SYS_MMAN_H)
# include <sys / mman.h>
# endif
# if defined(HAVE_SYS_UIO_H)
# include <sys / uio.h>
# endif
# if defined(HAVE_SYS_EPOLL_H)
# include <sys / epoll.h>
# endif
# include <unistd.h>
# include <fcntl.h>
# include <sys / types.h>
# include <sys / socket.h>
# include <netinet / in.h>
# include <pthread.h>
# include <poll.h>
# endif
struct bb_t
{
char buf [ 1024 ] ;
hcl_oow_t pos ;
hcl_oow_t len ;
int fd ;
hcl_bch_t * fn ;
} ;
typedef struct bb_t bb_t ;
2024-04-21 07:29:55 +00:00
struct proto_xtn_t
{
hcl_server_worker_t * worker ;
} ;
typedef struct proto_xtn_t proto_xtn_t ;
2023-11-26 15:13:28 +00:00
struct worker_hcl_xtn_t
{
2024-04-20 03:02:22 +00:00
hcl_server_worker_t * worker ;
2023-11-26 15:13:28 +00:00
int vm_running ;
} ;
typedef struct worker_hcl_xtn_t worker_hcl_xtn_t ;
struct server_hcl_xtn_t
{
hcl_server_t * server ;
} ;
typedef struct server_hcl_xtn_t server_hcl_xtn_t ;
/* ---------------------------------- */
2024-04-20 03:02:22 +00:00
enum hcl_xproto_rcv_state_t
2023-11-26 15:13:28 +00:00
{
2024-04-20 12:57:05 +00:00
HCL_XPROTO_RCV_HDR ,
HCL_XPROTO_RCV_PLD
2023-11-26 15:13:28 +00:00
} ;
2024-04-20 03:02:22 +00:00
typedef enum hcl_xproto_rcv_state_t hcl_xproto_rcv_state_t ;
2023-11-26 15:13:28 +00:00
2024-04-20 03:02:22 +00:00
struct hcl_xproto_t
2023-11-26 15:13:28 +00:00
{
2024-04-21 07:29:55 +00:00
hcl_oow_t _instsize ;
hcl_mmgr_t * _mmgr ;
2023-11-26 15:13:28 +00:00
hcl_tmr_index_t exec_runtime_event_index ;
struct
{
2024-04-20 03:02:22 +00:00
hcl_xproto_rcv_state_t state ;
2024-04-14 09:33:15 +00:00
hcl_oow_t len_needed ;
unsigned int eof : 1 ;
2023-11-26 15:13:28 +00:00
hcl_oow_t len ;
2024-04-20 12:57:05 +00:00
hcl_uint8_t buf [ HCL_XPKT_MAX_PLD_LEN ] ;
2023-11-26 15:13:28 +00:00
2024-04-20 12:57:05 +00:00
/* normalize header of hcl_xpkt_hdr_t with combined bits into separate placeholders */
struct
{
hcl_uint8_t id ;
hcl_uint8_t type ;
hcl_uint16_t len ; /* this is wider than the len field of hcl_xpkt_hdr_t */
} hdr ;
2024-04-14 09:33:15 +00:00
} rcv ;
2023-11-26 15:13:28 +00:00
struct
{
2024-04-14 17:23:55 +00:00
} snd ;
2023-11-26 15:13:28 +00:00
} ;
2024-04-20 03:02:22 +00:00
/* ---------------------------------- */
2023-11-26 15:13:28 +00:00
enum hcl_server_worker_state_t
{
HCL_SERVER_WORKER_STATE_DEAD = 0 ,
HCL_SERVER_WORKER_STATE_ALIVE = 1 ,
HCL_SERVER_WORKER_STATE_ZOMBIE = 2 /* the worker is not chained in the server's client list */
} ;
typedef enum hcl_server_worker_state_t hcl_server_worker_state_t ;
enum hcl_server_worker_opstate_t
{
HCL_SERVER_WORKER_OPSTATE_IDLE = 0 ,
HCL_SERVER_WORKER_OPSTATE_ERROR = 1 ,
HCL_SERVER_WORKER_OPSTATE_WAIT = 2 ,
HCL_SERVER_WORKER_OPSTATE_READ = 3 ,
HCL_SERVER_WORKER_OPSTATE_COMPILE = 4 ,
HCL_SERVER_WORKER_OPSTATE_EXECUTE = 5
} ;
typedef enum hcl_server_worker_opstate_t hcl_server_worker_opstate_t ;
struct hcl_server_worker_t
{
pthread_t thr ;
hcl_oow_t wid ;
int sck ;
hcl_sckaddr_t peeraddr ;
int claimed ;
hcl_ntime_t alloc_time ;
hcl_server_worker_state_t state ;
hcl_server_worker_opstate_t opstate ;
2024-04-20 03:02:22 +00:00
hcl_xproto_t * proto ;
hcl_t * hcl ;
2023-11-26 15:13:28 +00:00
hcl_server_t * server ;
hcl_server_worker_t * prev_worker ;
hcl_server_worker_t * next_worker ;
} ;
struct hcl_server_wid_map_data_t
{
int used ;
union
{
hcl_server_worker_t * worker ;
hcl_oow_t next ;
} u ;
} ;
typedef struct hcl_server_wid_map_data_t hcl_server_wid_map_data_t ;
typedef struct hcl_server_listener_t hcl_server_listener_t ;
struct hcl_server_listener_t
{
int sck ;
hcl_sckaddr_t sckaddr ;
hcl_server_listener_t * next_listener ;
} ;
struct hcl_server_t
{
hcl_oow_t _instsize ;
hcl_mmgr_t * _mmgr ;
hcl_cmgr_t * _cmgr ;
hcl_server_prim_t prim ;
/* [NOTE]
* this dummy_hcl is used when the main thread requires logging mostly .
* as there is no explicit locking when calling HCL_LOG ( ) functions ,
* the code must ensure that the logging functions are called in the
* context of the main server thraed only . error message setting is
* also performed in the main thread context for the same reason .
*
* however , you may have noticed mixed use of HCL_ASSERT with dummy_hcl
* in both the server thread context and the client thread contexts .
* it should be ok as assertion is only for debugging and it ' s operation
* is thread safe . */
hcl_t * dummy_hcl ;
hcl_tmr_t * tmr ;
hcl_errnum_t errnum ;
struct
{
hcl_ooch_t buf [ HCL_ERRMSG_CAPA ] ;
hcl_oow_t len ;
} errmsg ;
int stopreq ;
struct
{
hcl_bitmask_t trait ;
hcl_bitmask_t logmask ;
hcl_oow_t worker_stack_size ;
hcl_oow_t worker_max_count ;
hcl_ntime_t worker_idle_timeout ;
hcl_oow_t actor_heap_size ;
hcl_ntime_t actor_max_runtime ;
hcl_ooch_t script_include_path [ HCL_PATH_MAX + 1 ] ;
void * module_inctx ;
} cfg ;
struct
{
int ep_fd ;
struct epoll_event ev_buf [ 128 ] ;
hcl_server_listener_t * head ;
hcl_oow_t count ;
} listener ;
struct
{
hcl_server_worker_t * head ;
hcl_server_worker_t * tail ;
hcl_oow_t count ;
2024-04-14 09:33:15 +00:00
} worker_list [ 2 ] ; /* DEAD and ALIVE oly. ZOMBIEs are not chained here */
2023-11-26 15:13:28 +00:00
struct
{
hcl_server_wid_map_data_t * ptr ;
hcl_oow_t capa ;
hcl_oow_t free_first ;
hcl_oow_t free_last ;
} wid_map ; /* worker's id map */
int mux_pipe [ 2 ] ; /* pipe to break the blocking multiplexer in the main server loop */
pthread_mutex_t worker_mutex ;
pthread_mutex_t tmr_mutex ;
pthread_mutex_t log_mutex ;
} ;
2024-04-14 17:23:55 +00:00
/* ========================================================================= */
2024-04-20 03:02:22 +00:00
static int send_stdout_bytes ( hcl_xproto_t * proto , const hcl_bch_t * data , hcl_oow_t len ) ;
2024-04-14 17:23:55 +00:00
# if defined(HCL_OOCH_IS_UCH)
2024-04-20 03:02:22 +00:00
static int send_stdout_chars ( hcl_xproto_t * proto , const hcl_ooch_t * data , hcl_oow_t len ) ;
2024-04-14 17:23:55 +00:00
# else
# define send_stdout_chars(proto,data,len) send_stdout_bytes(proto,data,len)
# endif
2023-11-26 15:13:28 +00:00
/* ========================================================================= */
static const hcl_bch_t * get_base_name ( const hcl_bch_t * path )
{
const hcl_bch_t * p , * last = HCL_NULL ;
for ( p = path ; * p ! = ' \0 ' ; p + + )
{
if ( HCL_IS_PATH_SEP ( * p ) ) last = p ;
}
return ( last = = HCL_NULL ) ? path : ( last + 1 ) ;
}
static HCL_INLINE int open_read_stream ( hcl_t * hcl , hcl_io_cciarg_t * arg )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
bb_t * bb = HCL_NULL ;
hcl_server_t * server ;
2024-04-20 03:02:22 +00:00
server = xtn - > worker - > server ;
2023-11-26 15:13:28 +00:00
if ( arg - > includer )
{
/* includee */
/* TOOD: Do i need to skip prepending the include path if the included path is an absolute path?
* it may be good for security if i don ' t skip it . we can lock the included files in a given directory */
hcl_oow_t ucslen , bcslen , parlen ;
const hcl_bch_t * fn , * fb ;
# if defined(HCL_OOCH_IS_UCH)
if ( hcl_convootobcstr ( hcl , arg - > name , & ucslen , HCL_NULL , & bcslen ) < = - 1 ) goto oops ;
# else
bcslen = hcl_count_bcstr ( arg - > name ) ;
# endif
fn = ( ( bb_t * ) arg - > includer - > handle ) - > fn ;
if ( fn [ 0 ] = = ' \0 ' & & server - > cfg . script_include_path [ 0 ] ! = ' \0 ' )
{
# if defined(HCL_OOCH_IS_UCH)
if ( hcl_convootobcstr ( hcl , server - > cfg . script_include_path , & ucslen , HCL_NULL , & parlen ) < = - 1 ) goto oops ;
# else
parlen = hcl_count_bcstr ( server - > cfg . script_include_path ) ;
# endif
}
else
{
fb = get_base_name ( fn ) ;
parlen = fb - fn ;
}
bb = ( bb_t * ) hcl_callocmem ( hcl , HCL_SIZEOF ( * bb ) + ( HCL_SIZEOF ( hcl_bch_t ) * ( parlen + bcslen + 2 ) ) ) ;
if ( ! bb ) goto oops ;
bb - > fn = ( hcl_bch_t * ) ( bb + 1 ) ;
if ( fn [ 0 ] = = ' \0 ' & & server - > cfg . script_include_path [ 0 ] ! = ' \0 ' )
{
# if defined(HCL_OOCH_IS_UCH)
hcl_convootobcstr ( hcl , server - > cfg . script_include_path , & ucslen , bb - > fn , & parlen ) ;
# else
hcl_copy_bchars ( bb - > fn , server - > cfg . script_include_path , parlen ) ;
# endif
if ( ! HCL_IS_PATH_SEP ( bb - > fn [ parlen ] ) ) bb - > fn [ parlen + + ] = HCL_DFL_PATH_SEP ; /* +2 was used in hcl_callocmem() for this (+1 for this, +1 for '\0' */
}
else
{
hcl_copy_bchars ( bb - > fn , fn , parlen ) ;
}
# if defined(HCL_OOCH_IS_UCH)
hcl_convootobcstr ( hcl , arg - > name , & ucslen , & bb - > fn [ parlen ] , & bcslen ) ;
# else
hcl_copy_bcstr ( & bb - > fn [ parlen ] , bcslen + 1 , arg - > name ) ;
# endif
bb - > fd = open ( bb - > fn , O_RDONLY , 0 ) ;
if ( bb - > fd < = - 1 )
{
hcl_seterrnum ( hcl , HCL_EIOERR ) ;
goto oops ;
}
}
else
{
/* main stream */
hcl_oow_t pathlen = 0 ;
bb = ( bb_t * ) hcl_callocmem ( hcl , HCL_SIZEOF ( * bb ) + ( HCL_SIZEOF ( hcl_bch_t ) * ( pathlen + 1 ) ) ) ;
if ( ! bb ) goto oops ;
/* copy ane empty string as a main stream's name */
bb - > fn = ( hcl_bch_t * ) ( bb + 1 ) ;
hcl_copy_bcstr ( bb - > fn , pathlen + 1 , " " ) ;
2024-04-20 03:02:22 +00:00
bb - > fd = xtn - > worker - > sck ;
2023-11-26 15:13:28 +00:00
}
HCL_ASSERT ( hcl , bb - > fd > = 0 ) ;
arg - > handle = bb ;
return 0 ;
oops :
if ( bb )
{
2024-04-20 03:02:22 +00:00
if ( bb - > fd > = 0 & & bb - > fd ! = xtn - > worker - > sck ) close ( bb - > fd ) ;
2023-11-26 15:13:28 +00:00
hcl_freemem ( hcl , bb ) ;
}
return - 1 ;
}
static HCL_INLINE int close_read_stream ( hcl_t * hcl , hcl_io_cciarg_t * arg )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
bb_t * bb ;
bb = ( bb_t * ) arg - > handle ;
HCL_ASSERT ( hcl , bb ! = HCL_NULL & & bb - > fd > = 0 ) ;
2024-04-20 03:02:22 +00:00
if ( bb - > fd ! = xtn - > worker - > sck ) close ( bb - > fd ) ;
2023-11-26 15:13:28 +00:00
hcl_freemem ( hcl , bb ) ;
arg - > handle = HCL_NULL ;
return 0 ;
}
static HCL_INLINE int read_read_stream ( hcl_t * hcl , hcl_io_cciarg_t * arg )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
bb_t * bb ;
hcl_oow_t bcslen , ucslen , remlen ;
hcl_server_worker_t * worker ;
ssize_t x ;
int y ;
bb = ( bb_t * ) arg - > handle ;
HCL_ASSERT ( hcl , bb ! = HCL_NULL & & bb - > fd > = 0 ) ;
2024-04-20 03:02:22 +00:00
worker = xtn - > worker ;
2023-11-26 15:13:28 +00:00
start_over :
if ( arg - > includer )
{
/* includee */
if ( HCL_UNLIKELY ( worker - > server - > stopreq ) )
{
hcl_seterrbfmt ( hcl , HCL_EGENERIC , " stop requested " ) ;
return - 1 ;
}
x = read ( bb - > fd , & bb - > buf [ bb - > len ] , HCL_COUNTOF ( bb - > buf ) - bb - > len ) ;
if ( x < = - 1 )
{
if ( errno = = EINTR ) goto start_over ;
hcl_seterrwithsyserr ( hcl , 0 , errno ) ;
return - 1 ;
}
bb - > len + = x ;
}
else
{
/* main stream */
hcl_server_t * server ;
HCL_ASSERT ( hcl , bb - > fd = = worker - > sck ) ;
server = worker - > server ;
while ( 1 )
{
int n ;
struct pollfd pfd ;
int tmout , actual_tmout ;
if ( HCL_UNLIKELY ( server - > stopreq ) )
{
hcl_seterrbfmt ( hcl , HCL_EGENERIC , " stop requested " ) ;
return - 1 ;
}
tmout = HCL_SECNSEC_TO_MSEC ( server - > cfg . worker_idle_timeout . sec , server - > cfg . worker_idle_timeout . nsec ) ;
actual_tmout = ( tmout < = 0 ) ? 10000 : tmout ;
pfd . fd = bb - > fd ;
pfd . events = POLLIN | POLLERR ;
n = poll ( & pfd , 1 , actual_tmout ) ;
if ( n < = - 1 )
{
if ( errno = = EINTR ) goto start_over ;
hcl_seterrwithsyserr ( hcl , 0 , errno ) ;
return - 1 ;
}
else if ( n > = 1 ) break ;
/* timed out - no activity on the pfd */
if ( tmout > 0 )
{
hcl_seterrbfmt ( hcl , HCL_EGENERIC , " no activity on the worker socket %d " , bb - > fd ) ;
return - 1 ;
}
}
x = recv ( bb - > fd , & bb - > buf [ bb - > len ] , HCL_COUNTOF ( bb - > buf ) - bb - > len , 0 ) ;
if ( x < = - 1 )
{
if ( errno = = EINTR ) goto start_over ;
hcl_seterrwithsyserr ( hcl , 0 , errno ) ;
return - 1 ;
}
bb - > len + = x ;
}
# if defined(HCL_OOCH_IS_UCH)
bcslen = bb - > len ;
2024-04-14 09:33:15 +00:00
ucslen = HCL_COUNTOF ( arg - > buf . c ) ;
y = hcl_convbtooochars ( hcl , bb - > buf , & bcslen , arg - > buf . c , & ucslen ) ;
2023-11-26 15:13:28 +00:00
if ( y < = - 1 & & ucslen < = 0 )
{
if ( y = = - 3 & & x ! = 0 ) goto start_over ; /* incomplete sequence and not EOF yet */
return - 1 ;
}
/* if ucslen is greater than 0, i see that some characters have been
* converted properly */
# else
2024-04-14 09:33:15 +00:00
bcslen = ( bb - > len < HCL_COUNTOF ( arg - > buf . b ) ) ? bb - > len : HCL_COUNTOF ( arg - > buf . b ) ;
2023-11-26 15:13:28 +00:00
ucslen = bcslen ;
2024-04-14 09:33:15 +00:00
hcl_copy_bchars ( arg - > buf . b , bb - > buf , bcslen ) ;
2023-11-26 15:13:28 +00:00
# endif
remlen = bb - > len - bcslen ;
if ( remlen > 0 ) HCL_MEMMOVE ( bb - > buf , & bb - > buf [ bcslen ] , remlen ) ;
bb - > len = remlen ;
arg - > xlen = ucslen ;
return 0 ;
}
static int read_handler ( hcl_t * hcl , hcl_io_cmd_t cmd , void * arg )
{
switch ( cmd )
{
case HCL_IO_OPEN :
return open_read_stream ( hcl , ( hcl_io_cciarg_t * ) arg ) ;
case HCL_IO_CLOSE :
return close_read_stream ( hcl , ( hcl_io_cciarg_t * ) arg ) ;
case HCL_IO_READ :
return read_read_stream ( hcl , ( hcl_io_cciarg_t * ) arg ) ;
default :
hcl_seterrnum ( hcl , HCL_EINTERN ) ;
return - 1 ;
}
}
static int scan_handler ( hcl_t * hcl , hcl_io_cmd_t cmd , void * arg )
{
switch ( cmd )
{
case HCL_IO_OPEN :
return 0 ;
case HCL_IO_CLOSE :
return 0 ;
case HCL_IO_READ :
2024-04-14 09:33:15 +00:00
#if 0
2023-11-26 15:13:28 +00:00
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
hcl_io_udiarg_t * inarg = ( hcl_io_udiarg_t * ) arg ;
// what if it writes a request to require more input??
2024-04-20 03:02:22 +00:00
if ( hcl_xproto_handle_incoming ( xtn - > proto ) < = - 1 )
2023-11-26 15:13:28 +00:00
{
}
}
2024-04-14 09:33:15 +00:00
# else
/* TODO: read from the input buffer or pipe*/
# endif
2023-11-26 15:13:28 +00:00
default :
hcl_seterrnum ( hcl , HCL_EINTERN ) ;
return - 1 ;
}
}
static int print_handler ( hcl_t * hcl , hcl_io_cmd_t cmd , void * arg )
{
switch ( cmd )
{
case HCL_IO_OPEN :
2024-04-14 17:23:55 +00:00
printf ( " IO OPEN SOMETHING........... \n " ) ;
2023-11-26 15:13:28 +00:00
return 0 ;
case HCL_IO_CLOSE :
2024-04-14 17:23:55 +00:00
printf ( " IO CLOSE SOMETHING........... \n " ) ;
2023-11-26 15:13:28 +00:00
return 0 ;
case HCL_IO_WRITE :
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
hcl_io_udoarg_t * outarg = ( hcl_io_udoarg_t * ) arg ;
2024-04-14 09:33:15 +00:00
printf ( " IO WRITE SOMETHING........... \n " ) ;
2024-04-20 03:02:22 +00:00
if ( send_stdout_chars ( xtn - > worker - > proto , outarg - > ptr , outarg - > len ) < = - 1 )
2023-11-26 15:13:28 +00:00
{
/* TODO: change error code and message. propagage the errormessage from proto */
hcl_seterrbfmt ( hcl , HCL_EIOERR , " failed to write message via proto " ) ;
/* writing failure on the socket is a critical failure.
* execution must get aborted */
hcl_abort ( hcl ) ;
return - 1 ;
}
2024-04-14 17:23:55 +00:00
2023-11-26 15:13:28 +00:00
outarg - > xlen = outarg - > len ;
return 0 ;
}
case HCL_IO_WRITE_BYTES :
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
hcl_io_udoarg_t * outarg = ( hcl_io_udoarg_t * ) arg ;
2024-04-14 09:33:15 +00:00
printf ( " IO WRITE SOMETHING BYTES........... \n " ) ;
2024-04-20 03:02:22 +00:00
if ( send_stdout_bytes ( xtn - > worker - > proto , outarg - > ptr , outarg - > len ) < = - 1 )
2023-11-26 15:13:28 +00:00
{
/* TODO: change error code and message. propagage the errormessage from proto */
hcl_seterrbfmt ( hcl , HCL_EIOERR , " failed to write message via proto " ) ;
/* writing failure on the socket is a critical failure.
* execution must get aborted */
hcl_abort ( hcl ) ;
return - 1 ;
}
outarg - > xlen = outarg - > len ;
return 0 ;
}
default :
hcl_seterrnum ( hcl , HCL_EINTERN ) ;
return - 1 ;
}
}
/* ========================================================================= */
2024-04-14 09:33:15 +00:00
static void server_log_write ( hcl_t * hcl , hcl_bitmask_t mask , const hcl_ooch_t * msg , hcl_oow_t len )
2023-11-26 15:13:28 +00:00
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
hcl_server_t * server ;
2024-04-20 03:02:22 +00:00
server = xtn - > worker - > server ;
2023-11-26 15:13:28 +00:00
pthread_mutex_lock ( & server - > log_mutex ) ;
2024-04-20 03:02:22 +00:00
server - > prim . log_write ( server , xtn - > worker - > wid , mask , msg , len ) ;
2023-11-26 15:13:28 +00:00
pthread_mutex_unlock ( & server - > log_mutex ) ;
}
2024-04-14 09:33:15 +00:00
static void server_log_write_for_dummy ( hcl_t * hcl , hcl_bitmask_t mask , const hcl_ooch_t * msg , hcl_oow_t len )
2023-11-26 15:13:28 +00:00
{
server_hcl_xtn_t * xtn = ( server_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
hcl_server_t * server ;
server = xtn - > server ;
pthread_mutex_lock ( & server - > log_mutex ) ;
server - > prim . log_write ( server , HCL_SERVER_WID_INVALID , mask , msg , len ) ;
pthread_mutex_unlock ( & server - > log_mutex ) ;
}
/* ========================================================================= */
static int vm_startup ( hcl_t * hcl )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
xtn - > vm_running = 1 ;
return 0 ;
}
static void vm_cleanup ( hcl_t * hcl )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
xtn - > vm_running = 0 ;
}
static void vm_checkbc ( hcl_t * hcl , hcl_oob_t bcode )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
2024-04-20 03:02:22 +00:00
if ( xtn - > worker - > server - > stopreq ) hcl_abort ( hcl ) ;
2023-11-26 15:13:28 +00:00
}
/*
static void gc_hcl ( hcl_t * hcl )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
}
static void fini_hcl ( hcl_t * hcl )
{
worker_hcl_xtn_t * xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
}
*/
/* ========================================================================= */
# define SERVER_LOGMASK_INFO (HCL_LOG_INFO | HCL_LOG_APP)
# define SERVER_LOGMASK_ERROR (HCL_LOG_ERROR | HCL_LOG_APP)
2024-04-14 09:33:15 +00:00
static int on_fed_cnode ( hcl_t * hcl , hcl_cnode_t * obj )
{
2024-04-18 14:06:28 +00:00
/*worker_hcl_xtn_t* xtn = (worker_hcl_xtn_t*)hcl_getxtn(hcl);*/
2024-04-20 03:02:22 +00:00
/*hcl_xproto_t* proto = xtn->proto;*/
2024-04-14 09:33:15 +00:00
int flags = 0 ;
printf ( " on_fed_cnode...... \n " ) ;
/* the compile error must not break the input loop.
* this function returns 0 to go on despite a compile - time error .
*
* if a single line or continued lines contain multiple expressions ,
* execution is delayed until the last expression is compiled . */
2024-04-14 17:23:55 +00:00
printf ( " COMPILING hcl_copingll...... \n " ) ;
2024-04-14 09:33:15 +00:00
if ( hcl_compile ( hcl , obj , flags ) < = - 1 )
{
2024-04-14 17:23:55 +00:00
hcl_logbfmt ( hcl , HCL_LOG_STDERR , " COMPILER ERROR - %js \n " , hcl_geterrmsg ( hcl ) ) ;
2024-04-14 09:33:15 +00:00
#if 0
print_error ( hcl , " failed to compile " ) ;
xtn - > feed . pos = xtn - > feed . len ; /* arrange to discard the rest of the line */
show_prompt ( hcl , 0 ) ;
# endif
}
return 0 ;
}
2023-11-26 15:13:28 +00:00
static void exec_runtime_handler ( hcl_tmr_t * tmr , const hcl_ntime_t * now , hcl_tmr_event_t * evt )
{
/* [NOTE] this handler is executed in the main server thread
* when it calls hcl_tmr_fire ( ) . */
2024-04-20 03:02:22 +00:00
hcl_xproto_t * proto ;
hcl_server_worker_t * worker ;
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
2024-04-20 03:02:22 +00:00
proto = ( hcl_xproto_t * ) evt - > ctx ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2023-11-26 15:13:28 +00:00
2024-04-20 03:02:22 +00:00
/* TODO: can we use worker->hcl for logging before abort?? */
2024-04-21 07:29:55 +00:00
HCL_LOG1 ( worker - > server - > dummy_hcl , SERVER_LOGMASK_INFO , " Aborting script execution for max_actor_runtime exceeded [%zu] \n " , worker - > wid ) ;
2024-04-20 03:02:22 +00:00
hcl_abort ( worker - > hcl ) ;
2023-11-26 15:13:28 +00:00
}
static void exec_runtime_updater ( hcl_tmr_t * tmr , hcl_tmr_index_t old_index , hcl_tmr_index_t new_index , hcl_tmr_event_t * evt )
{
/* [NOTE] this handler is executed in the main server thread
* when it calls hcl_tmr_fire ( ) */
2024-04-20 03:02:22 +00:00
hcl_xproto_t * proto ;
hcl_server_worker_t * worker ;
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
2024-04-20 03:02:22 +00:00
proto = ( hcl_xproto_t * ) evt - > ctx ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-20 03:02:22 +00:00
HCL_ASSERT ( worker - > hcl , proto - > exec_runtime_event_index = = old_index ) ;
2023-11-26 15:13:28 +00:00
/* the event is being removed by hcl_tmr_fire() or by hcl_tmr_delete()
* if new_index is HCL_TMR_INVALID_INDEX . it ' s being updated if not . */
proto - > exec_runtime_event_index = new_index ;
}
2024-04-20 03:02:22 +00:00
static int insert_exec_timer ( hcl_xproto_t * proto , const hcl_ntime_t * tmout )
2023-11-26 15:13:28 +00:00
{
/* [NOTE] this is executed in the worker thread */
hcl_tmr_event_t event ;
hcl_tmr_index_t index ;
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
2024-04-20 03:02:22 +00:00
hcl_server_worker_t * worker ;
2023-11-26 15:13:28 +00:00
hcl_server_t * server ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-20 03:02:22 +00:00
server = worker - > server ;
2023-11-26 15:13:28 +00:00
2024-04-20 03:02:22 +00:00
HCL_ASSERT ( worker - > hcl , proto - > exec_runtime_event_index = = HCL_TMR_INVALID_INDEX ) ;
2023-11-26 15:13:28 +00:00
HCL_MEMSET ( & event , 0 , HCL_SIZEOF ( event ) ) ;
event . ctx = proto ;
2024-04-20 03:02:22 +00:00
worker - > hcl - > vmprim . vm_gettime ( worker - > hcl , & event . when ) ;
2023-11-26 15:13:28 +00:00
HCL_ADD_NTIME ( & event . when , & event . when , tmout ) ;
event . handler = exec_runtime_handler ;
event . updater = exec_runtime_updater ;
pthread_mutex_lock ( & server - > tmr_mutex ) ;
index = hcl_tmr_insert ( server - > tmr , & event ) ;
proto - > exec_runtime_event_index = index ;
if ( index ! = HCL_TMR_INVALID_INDEX )
{
/* inform the server of timer event change */
write ( server - > mux_pipe [ 1 ] , " X " , 1 ) ; /* don't care even if it fails */
}
pthread_mutex_unlock ( & server - > tmr_mutex ) ;
return ( index = = HCL_TMR_INVALID_INDEX ) ? - 1 : 0 ;
}
2024-04-20 03:02:22 +00:00
static void delete_exec_timer ( hcl_xproto_t * proto )
2023-11-26 15:13:28 +00:00
{
/* [NOTE] this is executed in the worker thread. if the event has been fired
* in the server thread , proto - > exec_runtime_event_index should be
* HCL_TMR_INVALID_INDEX as set by exec_runtime_handler */
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
2024-04-20 03:02:22 +00:00
hcl_server_worker_t * worker ;
2023-11-26 15:13:28 +00:00
hcl_server_t * server ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-20 03:02:22 +00:00
server = worker - > server ;
2023-11-26 15:13:28 +00:00
pthread_mutex_lock ( & server - > tmr_mutex ) ;
if ( proto - > exec_runtime_event_index ! = HCL_TMR_INVALID_INDEX )
{
/* the event has not been fired yet. let's delete it
* if it has been fired , the index it shall be HCL_TMR_INVALID_INDEX already */
hcl_tmr_delete ( server - > tmr , proto - > exec_runtime_event_index ) ;
2024-04-20 03:02:22 +00:00
HCL_ASSERT ( worker - > hcl , proto - > exec_runtime_event_index = = HCL_TMR_INVALID_INDEX ) ;
2023-11-26 15:13:28 +00:00
/*proto->exec_runtime_event_index = HCL_TMR_INVALID_INDEX; */
}
pthread_mutex_unlock ( & server - > tmr_mutex ) ;
}
2024-04-20 03:02:22 +00:00
static int execute_script ( hcl_xproto_t * proto , const hcl_bch_t * trigger )
2023-11-26 15:13:28 +00:00
{
hcl_oop_t obj ;
const hcl_ooch_t * failmsg = HCL_NULL ;
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
2024-04-20 03:02:22 +00:00
hcl_server_worker_t * worker ;
2023-11-26 15:13:28 +00:00
hcl_server_t * server ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-20 03:02:22 +00:00
server = worker - > server ;
2023-11-26 15:13:28 +00:00
2024-04-14 09:33:15 +00:00
#if 0
2024-04-20 03:02:22 +00:00
hcl_xproto_start_reply ( proto ) ;
2024-04-14 09:33:15 +00:00
# endif
2023-11-26 15:13:28 +00:00
if ( server - > cfg . actor_max_runtime . sec < = 0 & & server - > cfg . actor_max_runtime . sec < = 0 )
{
2024-04-20 03:02:22 +00:00
obj = hcl_execute ( worker - > hcl ) ;
if ( ! obj ) failmsg = hcl_geterrmsg ( worker - > hcl ) ;
2023-11-26 15:13:28 +00:00
}
else
{
if ( insert_exec_timer ( proto , & server - > cfg . actor_max_runtime ) < = - 1 )
{
2024-04-20 03:02:22 +00:00
HCL_LOG0 ( worker - > hcl , SERVER_LOGMASK_ERROR , " Cannot start execution timer \n " ) ;
hcl_seterrbfmt ( worker - > hcl , HCL_ESYSMEM , " cannot start execution timer " ) ; /* i do this just to compose the error message */
failmsg = hcl_geterrmsg ( worker - > hcl ) ;
2023-11-26 15:13:28 +00:00
}
else
{
2024-04-20 03:02:22 +00:00
obj = hcl_execute ( worker - > hcl ) ;
if ( ! obj ) failmsg = hcl_geterrmsg ( worker - > hcl ) ;
2023-11-26 15:13:28 +00:00
delete_exec_timer ( proto ) ;
}
}
2024-04-14 09:33:15 +00:00
#if 0
2024-04-20 03:02:22 +00:00
if ( hcl_xproto_end_reply ( proto , failmsg ) < = - 1 )
2023-11-26 15:13:28 +00:00
{
2024-04-20 03:02:22 +00:00
HCL_LOG1 ( worker - > hcl , SERVER_LOGMASK_ERROR , " Cannot finalize reply for %hs \n " , trigger ) ;
2023-11-26 15:13:28 +00:00
return - 1 ;
}
2024-04-14 09:33:15 +00:00
# endif
2023-11-26 15:13:28 +00:00
return 0 ;
}
2024-04-20 03:02:22 +00:00
static void send_error_message ( hcl_xproto_t * proto , const hcl_ooch_t * errmsg )
2023-11-26 15:13:28 +00:00
{
2024-04-14 09:33:15 +00:00
#if 0
2024-04-20 03:02:22 +00:00
hcl_xproto_start_reply ( proto ) ;
if ( hcl_xproto_end_reply ( proto , errmsg ) < = - 1 )
2023-11-26 15:13:28 +00:00
{
HCL_LOG1 ( proto - > hcl , SERVER_LOGMASK_ERROR , " Unable to send error message - %s \n " , errmsg ) ;
}
2024-04-14 09:33:15 +00:00
# endif
2023-11-26 15:13:28 +00:00
}
static void reformat_synerr ( hcl_t * hcl )
{
hcl_synerr_t synerr ;
const hcl_ooch_t * orgmsg ;
static hcl_ooch_t nullstr [ ] = { ' \0 ' } ;
hcl_getsynerr ( hcl , & synerr ) ;
orgmsg = hcl_backuperrmsg ( hcl ) ;
hcl_seterrbfmt (
hcl , HCL_ESYNERR ,
" %js%hs%.*js at %js%hsline %zu column %zu " ,
orgmsg ,
( synerr . tgt . len > 0 ? " near " : " " ) ,
synerr . tgt . len , synerr . tgt . val ,
( synerr . loc . file ? synerr . loc . file : nullstr ) ,
( synerr . loc . file ? " " : " " ) ,
synerr . loc . line , synerr . loc . colm
) ;
}
2024-04-20 03:02:22 +00:00
static void send_proto_hcl_error ( hcl_xproto_t * proto )
2023-11-26 15:13:28 +00:00
{
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
hcl_server_worker_t * worker ;
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-20 03:02:22 +00:00
if ( HCL_ERRNUM ( worker - > hcl ) = = HCL_ESYNERR ) reformat_synerr ( worker - > hcl ) ;
send_error_message ( proto , hcl_geterrmsg ( worker - > hcl ) ) ;
2023-11-26 15:13:28 +00:00
}
2024-04-20 03:02:22 +00:00
static void show_server_workers ( hcl_xproto_t * proto )
2023-11-26 15:13:28 +00:00
{
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
hcl_server_worker_t * worker , * w ;
2023-11-26 15:13:28 +00:00
hcl_server_t * server ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
server = worker - > server ;
2023-11-26 15:13:28 +00:00
pthread_mutex_lock ( & server - > worker_mutex ) ;
for ( w = server - > worker_list [ HCL_SERVER_WORKER_STATE_ALIVE ] . head ; w ; w = w - > next_worker )
{
/* TODO: implement this better... */
2024-04-21 07:29:55 +00:00
hcl_prbfmt ( worker - > hcl , " %zu %d %d \n " , w - > wid , w - > sck , 1000 ) ;
2023-11-26 15:13:28 +00:00
}
pthread_mutex_unlock ( & server - > worker_mutex ) ;
}
2024-04-20 03:02:22 +00:00
static int kill_server_worker ( hcl_xproto_t * proto , hcl_oow_t wid )
2023-11-26 15:13:28 +00:00
{
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
hcl_server_worker_t * worker ;
2023-11-26 15:13:28 +00:00
hcl_server_t * server ;
int xret = 0 ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
server = worker - > server ;
2023-11-26 15:13:28 +00:00
pthread_mutex_lock ( & server - > worker_mutex ) ;
if ( wid > = server - > wid_map . capa )
{
hcl_server_seterrnum ( server , HCL_ENOENT ) ;
xret = - 1 ;
}
else
{
hcl_server_worker_t * worker ;
if ( ! server - > wid_map . ptr [ wid ] . used )
{
hcl_server_seterrnum ( server , HCL_ENOENT ) ;
xret = - 1 ;
}
else
{
worker = server - > wid_map . ptr [ wid ] . u . worker ;
if ( ! worker )
{
hcl_server_seterrnum ( server , HCL_ENOENT ) ;
xret = - 1 ;
}
else
{
if ( worker - > sck ) shutdown ( worker - > sck , SHUT_RDWR ) ;
2024-04-20 03:02:22 +00:00
if ( worker - > hcl ) hcl_abort ( worker - > hcl ) ;
2023-11-26 15:13:28 +00:00
}
}
}
pthread_mutex_unlock ( & server - > worker_mutex ) ;
return xret ;
}
2024-04-20 03:02:22 +00:00
static int handle_packet ( hcl_xproto_t * proto , hcl_xpkt_type_t type , void * data , hcl_oow_t len )
2023-11-26 15:13:28 +00:00
{
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
hcl_server_worker_t * worker ;
hcl_t * hcl ;
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
hcl = worker - > hcl ;
2023-11-26 15:13:28 +00:00
2024-04-20 12:57:05 +00:00
printf ( " HANDLE PACKET TYPE => %d \n " , type ) ;
2024-04-18 14:06:28 +00:00
switch ( type )
2024-04-14 09:33:15 +00:00
{
2024-04-20 12:57:05 +00:00
case HCL_XPKT_CODE :
2024-04-18 14:06:28 +00:00
printf ( " FEEDING [%.*s] \n " , ( int ) len , data ) ;
if ( hcl_feedbchars ( hcl , data , len ) < = - 1 )
{
/* TODO: backup error message...and create a new message */
goto oops ;
}
break ;
2023-11-26 15:13:28 +00:00
2024-04-18 14:06:28 +00:00
case HCL_XPKT_EXECUTE :
2023-11-26 15:13:28 +00:00
{
2024-04-18 14:06:28 +00:00
hcl_oop_t retv ;
printf ( " EXECUTING hcl_executing...... \n " ) ;
hcl_decode ( hcl , hcl_getcode ( hcl ) , 0 , hcl_getbclen ( hcl ) ) ;
retv = hcl_execute ( hcl ) ;
hcl_flushudio ( hcl ) ;
hcl_clearcode ( hcl ) ;
if ( ! retv )
2024-04-14 09:33:15 +00:00
{
2024-04-18 14:06:28 +00:00
/* TODO: backup error message...and create a new message */
goto oops ;
2024-04-14 09:33:15 +00:00
}
2024-04-18 14:06:28 +00:00
break ;
2023-11-26 15:13:28 +00:00
}
2024-04-14 09:33:15 +00:00
2024-04-18 14:06:28 +00:00
case HCL_XPKT_STDIN :
/* store ... push stdin pipe... */
/*if (hcl_feedstdin() <= -1) */
break ;
2024-04-14 09:33:15 +00:00
2024-04-18 14:06:28 +00:00
case HCL_XPKT_LIST_WORKERS :
break ;
2023-11-26 15:13:28 +00:00
2024-04-18 14:06:28 +00:00
case HCL_XPKT_KILL_WORKER :
break ;
2024-04-20 03:02:22 +00:00
2024-04-18 14:06:28 +00:00
case HCL_XPKT_DISCONNECT :
return 0 ; /* disconnect received */
2023-11-26 15:13:28 +00:00
2024-04-18 14:06:28 +00:00
default :
/* unknown packet type */
/* TODO: proper error message */
goto oops ;
2023-11-26 15:13:28 +00:00
}
2024-04-18 14:06:28 +00:00
return 1 ;
2023-11-26 15:13:28 +00:00
2024-04-18 14:06:28 +00:00
oops :
return - 1 ;
2023-11-26 15:13:28 +00:00
}
2024-04-14 17:23:55 +00:00
static int send_iov ( int sck , struct iovec * iov , int count )
{
int index = 0 ;
while ( 1 )
{
ssize_t nwritten ;
struct msghdr msg ;
memset ( & msg , 0 , HCL_SIZEOF ( msg ) ) ;
msg . msg_iov = ( struct iovec * ) & iov [ index ] ;
msg . msg_iovlen = count - index ;
nwritten = sendmsg ( sck , & msg , 0 ) ;
if ( nwritten < = - 1 )
{
/* error occurred inside the worker thread shouldn't affect the error information
* in the server object . so here , i just log a message */
fprintf ( stderr , " Unable to sendmsg on %d - %s \n " , sck , strerror ( errno ) ) ;
return - 1 ;
}
while ( index < count & & ( size_t ) nwritten > = iov [ index ] . iov_len )
nwritten - = iov [ index + + ] . iov_len ;
if ( index = = count ) break ;
iov [ index ] . iov_base = ( void * ) ( ( hcl_uint8_t * ) iov [ index ] . iov_base + nwritten ) ;
iov [ index ] . iov_len - = nwritten ;
}
return 0 ;
}
2024-04-20 03:02:22 +00:00
static int send_stdout_bytes ( hcl_xproto_t * proto , const hcl_bch_t * data , hcl_oow_t len )
2024-04-14 17:23:55 +00:00
{
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
hcl_server_worker_t * worker ;
2024-04-14 17:23:55 +00:00
hcl_xpkt_hdr_t hdr ;
struct iovec iov [ 2 ] ;
const hcl_bch_t * ptr , * cur , * end ;
2024-04-20 12:57:05 +00:00
hcl_uint16_t seglen ;
2024-04-14 17:23:55 +00:00
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-14 17:23:55 +00:00
ptr = cur = data ;
end = data + len ;
2024-04-18 04:11:44 +00:00
printf ( " SENDING BYTES [%.*s] \n " , ( int ) len , data ) ;
2024-04-14 17:23:55 +00:00
while ( ptr < end )
{
2024-04-20 12:57:05 +00:00
while ( cur ! = end & & cur - ptr < HCL_XPKT_MAX_PLD_LEN ) cur + + ;
seglen = cur - ptr ;
2024-04-14 17:23:55 +00:00
hdr . id = 1 ; /* TODO: */
2024-04-20 12:57:05 +00:00
hdr . type = HCL_XPKT_STDOUT | ( ( ( seglen > > 8 ) & 0x0F ) < < 4 ) ;
hdr . len = seglen & 0xFF ;
2024-04-14 17:23:55 +00:00
iov [ 0 ] . iov_base = & hdr ;
iov [ 0 ] . iov_len = HCL_SIZEOF ( hdr ) ;
iov [ 1 ] . iov_base = ptr ;
2024-04-20 12:57:05 +00:00
iov [ 1 ] . iov_len = seglen ;
2024-04-14 17:23:55 +00:00
2024-04-21 07:29:55 +00:00
if ( send_iov ( worker - > sck , iov , 2 ) < = - 1 )
2024-04-14 17:23:55 +00:00
{
/* TODO: error message */
return - 1 ;
}
ptr = cur ;
}
return 0 ;
}
# if defined(HCL_OOCH_IS_UCH)
2024-04-20 03:02:22 +00:00
static int send_stdout_chars ( hcl_xproto_t * proto , const hcl_ooch_t * data , hcl_oow_t len )
2024-04-14 17:23:55 +00:00
{
2024-04-21 07:29:55 +00:00
proto_xtn_t * prtxtn ;
hcl_server_worker_t * worker ;
2024-04-14 17:23:55 +00:00
const hcl_ooch_t * ptr , * end ;
hcl_bch_t tmp [ 256 ] ;
hcl_oow_t tln , pln ;
int n ;
2024-04-21 07:29:55 +00:00
prtxtn = ( proto_xtn_t * ) hcl_xproto_getxtn ( proto ) ;
worker = prtxtn - > worker ;
2024-04-14 17:23:55 +00:00
ptr = data ;
end = data + len ;
while ( ptr < end )
{
pln = end - ptr ;
tln = HCL_COUNTOF ( tmp ) ;
2024-04-20 03:02:22 +00:00
n = hcl_convutobchars ( worker - > hcl , ptr , & pln , tmp , & tln ) ;
2024-04-14 17:23:55 +00:00
if ( n < = - 1 & & n ! = - 2 ) return - 1 ;
if ( send_stdout_bytes ( proto , tmp , tln ) < = - 1 ) return - 1 ;
ptr + = pln ;
}
return 0 ;
}
# endif
2024-04-20 03:02:22 +00:00
/* ========================================================================= */
2024-04-21 07:29:55 +00:00
hcl_xproto_t * hcl_xproto_open ( hcl_mmgr_t * mmgr , hcl_oow_t xtnsize )
2024-04-20 03:02:22 +00:00
{
hcl_xproto_t * proto ;
2024-04-21 07:29:55 +00:00
proto = ( hcl_xproto_t * ) HCL_MMGR_ALLOC ( mmgr , HCL_SIZEOF ( * proto ) + xtnsize ) ;
2024-04-20 03:02:22 +00:00
if ( HCL_UNLIKELY ( ! proto ) ) return HCL_NULL ;
HCL_MEMSET ( proto , 0 , HCL_SIZEOF ( * proto ) ) ;
2024-04-21 07:29:55 +00:00
proto - > _instsize = HCL_SIZEOF ( * proto ) ;
proto - > _mmgr = mmgr ;
2024-04-20 03:02:22 +00:00
proto - > exec_runtime_event_index = HCL_TMR_INVALID_INDEX ;
2024-04-20 12:57:05 +00:00
proto - > rcv . state = HCL_XPROTO_RCV_HDR ;
proto - > rcv . len_needed = HCL_XPKT_HDR_LEN ;
2024-04-20 03:02:22 +00:00
proto - > rcv . eof = 0 ;
return proto ;
}
void hcl_xproto_close ( hcl_xproto_t * proto )
{
2024-04-21 07:29:55 +00:00
HCL_MMGR_FREE ( proto - > _mmgr , proto ) ;
}
void * hcl_xproto_getxtn ( hcl_xproto_t * proto )
{
return ( proto + 1 ) ;
}
int hcl_xprpto_feed ( hcl_xproto_t * proto , const void * data , hcl_oow_t len )
{
}
hcl_uint8_t * hcl_xproto_getbuf ( hcl_xproto_t * proto , hcl_oow_t * capa )
{
* capa = HCL_COUNTOF ( proto - > rcv . buf ) - proto - > rcv . len ;
return & proto - > rcv . buf [ proto - > rcv . len ] ;
}
void hcl_xproto_seteof ( hcl_xproto_t * proto , int v )
{
proto - > rcv . eof = v ;
}
void hcl_xproto_advbuf ( hcl_xproto_t * proto , hcl_oow_t inc )
{
proto - > rcv . len + = inc ;
2024-04-20 03:02:22 +00:00
}
int hcl_xproto_ready ( hcl_xproto_t * proto )
{
/* has it received suffient data for processing? */
return proto - > rcv . len > = proto - > rcv . len_needed ;
}
2024-04-21 07:29:55 +00:00
int hcl_xproto_process ( hcl_xproto_t * proto )
2024-04-20 03:02:22 +00:00
{
int n ;
2024-04-20 12:57:05 +00:00
hcl_xpkt_hdr_t * hdr ;
2024-04-20 03:02:22 +00:00
switch ( proto - > rcv . state )
{
2024-04-20 12:57:05 +00:00
case HCL_XPROTO_RCV_HDR :
if ( proto - > rcv . len < HCL_XPKT_HDR_LEN ) goto carry_on ; /* need more data */
2024-04-20 03:02:22 +00:00
2024-04-20 12:57:05 +00:00
hdr = ( hcl_xpkt_hdr_t * ) proto - > rcv . buf ;
proto - > rcv . hdr . id = hdr - > id ;
proto - > rcv . hdr . type = hdr - > type & 0x0F ;
proto - > rcv . hdr . len = ( hcl_uint16_t ) hdr - > len | ( ( hcl_uint16_t ) ( hdr - > type > > 4 ) < < 8 ) ;
2024-04-20 03:02:22 +00:00
/* consume the header */
2024-04-20 12:57:05 +00:00
HCL_MEMMOVE ( proto - > rcv . buf , & proto - > rcv . buf [ HCL_XPKT_HDR_LEN ] , proto - > rcv . len - HCL_XPKT_HDR_LEN ) ;
proto - > rcv . len - = HCL_XPKT_HDR_LEN ;
2024-04-20 03:02:22 +00:00
/* switch to the payload mode */
if ( proto - > rcv . hdr . len > 0 )
{
2024-04-20 12:57:05 +00:00
proto - > rcv . state = HCL_XPROTO_RCV_PLD ;
2024-04-20 03:02:22 +00:00
proto - > rcv . len_needed = proto - > rcv . hdr . len ;
}
else
{
/* take shortcut */
/* TODO: convert handle_packet as call back */
n = handle_packet ( proto , proto - > rcv . hdr . type , proto - > rcv . buf , proto - > rcv . hdr . len ) ;
if ( n < = - 1 ) goto fail_with_errmsg ;
if ( n = = 0 ) return 0 ;
}
break ;
2024-04-20 12:57:05 +00:00
case HCL_XPROTO_RCV_PLD :
2024-04-20 03:02:22 +00:00
if ( proto - > rcv . len < proto - > rcv . hdr . len ) goto carry_on ; /* need more payload data */
/* TODO: convert handle_packet as call back */
n = handle_packet ( proto , proto - > rcv . hdr . type , proto - > rcv . buf , proto - > rcv . hdr . len ) ;
/* TODO: minimize the use of HCL_MEMOVE... use the buffer */
/* switch to the header mode */
if ( proto - > rcv . hdr . len > 0 )
{
HCL_MEMMOVE ( proto - > rcv . buf , & proto - > rcv . buf [ proto - > rcv . hdr . len ] , proto - > rcv . len - proto - > rcv . hdr . len ) ;
proto - > rcv . len - = proto - > rcv . hdr . len ;
}
2024-04-20 12:57:05 +00:00
proto - > rcv . state = HCL_XPROTO_RCV_HDR ;
proto - > rcv . len_needed = HCL_XPKT_HDR_LEN ;
2024-04-20 03:02:22 +00:00
if ( n < = - 1 ) goto fail_with_errmsg ;
if ( n = = 0 ) return 0 ;
break ;
default :
/*
hcl_seterrbfmt ( hcl , HCL_EINTERN , " invalid request state %d " , ( int ) proto - > rcv . state ) ;
*/
/* TODO: call back */
goto fail_with_errmsg ;
}
carry_on :
return 1 ;
fail_with_errmsg :
// TODO: proper error handling
//send_proto_hcl_error (proto);
//HCL_LOG1 (hcl, SERVER_LOGMASK_ERROR, "Unable to compile .SCRIPT contents - %js\n", hcl_geterrmsg(worker->hcl));
return - 1 ;
}
2024-04-14 17:23:55 +00:00
2023-11-26 15:13:28 +00:00
/* ========================================================================= */
hcl_server_t * hcl_server_open ( hcl_mmgr_t * mmgr , hcl_oow_t xtnsize , hcl_server_prim_t * prim , hcl_errnum_t * errnum )
{
hcl_server_t * server = HCL_NULL ;
hcl_t * hcl = HCL_NULL ;
hcl_tmr_t * tmr = HCL_NULL ;
server_hcl_xtn_t * xtn ;
int pfd [ 2 ] , fcv ;
hcl_bitmask_t trait ;
server = ( hcl_server_t * ) HCL_MMGR_ALLOC ( mmgr , HCL_SIZEOF ( * server ) + xtnsize ) ;
if ( ! server )
{
if ( errnum ) * errnum = HCL_ESYSMEM ;
return HCL_NULL ;
}
hcl = hcl_openstdwithmmgr ( mmgr , HCL_SIZEOF ( * xtn ) , errnum ) ;
if ( ! hcl ) goto oops ;
/* replace the vmprim.log_write function */
2024-04-14 09:33:15 +00:00
hcl - > vmprim . log_write = server_log_write_for_dummy ;
2023-11-26 15:13:28 +00:00
tmr = hcl_tmr_open ( hcl , 0 , 1024 ) ; /* TOOD: make the timer's default size configurable */
if ( ! tmr )
{
if ( errnum ) * errnum = HCL_ESYSMEM ;
goto oops ;
}
if ( pipe ( pfd ) < = - 1 )
{
if ( errnum ) * errnum = hcl - > vmprim . syserrstrb ( hcl , 0 , errno , HCL_NULL , 0 ) ;
goto oops ;
}
# if defined(O_NONBLOCK) || defined(O_CLOEXEC)
fcv = fcntl ( pfd [ 0 ] , F_GETFD , 0 ) ;
if ( fcv > = 0 )
{
# if defined(O_NONBLOCK)
fcv | = O_NONBLOCK ;
# endif
# if defined(O_CLOEXEC)
fcv | = O_CLOEXEC ;
# endif
fcntl ( pfd [ 0 ] , F_SETFD , fcv ) ;
}
fcv = fcntl ( pfd [ 1 ] , F_GETFD , 0 ) ;
if ( fcv > = 0 )
{
# if defined(O_NONBLOCK)
fcv | = O_NONBLOCK ;
# endif
# if defined(O_CLOEXEC)
fcv | = O_CLOEXEC ;
# endif
fcntl ( pfd [ 1 ] , F_SETFD , fcv ) ;
}
# endif
xtn = ( server_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
xtn - > server = server ;
HCL_MEMSET ( server , 0 , HCL_SIZEOF ( * server ) + xtnsize ) ;
server - > _instsize = HCL_SIZEOF ( * server ) ;
server - > _mmgr = mmgr ;
server - > _cmgr = hcl_get_utf8_cmgr ( ) ;
server - > prim = * prim ;
server - > dummy_hcl = hcl ;
server - > tmr = tmr ;
server - > cfg . logmask = ~ ( hcl_bitmask_t ) 0 ;
server - > cfg . worker_stack_size = 512000UL ;
server - > cfg . actor_heap_size = 512000UL ;
HCL_INIT_NTIME ( & server - > cfg . worker_idle_timeout , 0 , 0 ) ;
HCL_INIT_NTIME ( & server - > cfg . actor_max_runtime , 0 , 0 ) ;
server - > mux_pipe [ 0 ] = pfd [ 0 ] ;
server - > mux_pipe [ 1 ] = pfd [ 1 ] ;
server - > wid_map . free_first = HCL_SERVER_WID_INVALID ;
server - > wid_map . free_last = HCL_SERVER_WID_INVALID ;
server - > listener . ep_fd = - 1 ;
pthread_mutex_init ( & server - > worker_mutex , HCL_NULL ) ;
pthread_mutex_init ( & server - > tmr_mutex , HCL_NULL ) ;
pthread_mutex_init ( & server - > log_mutex , HCL_NULL ) ;
/* the dummy hcl is used for this server to perform primitive operations
* such as getting system time or logging . so the heap size doesn ' t
* need to be changed from the tiny value set above . */
hcl_setoption ( server - > dummy_hcl , HCL_LOG_MASK , & server - > cfg . logmask ) ;
hcl_setcmgr ( server - > dummy_hcl , hcl_server_getcmgr ( server ) ) ;
hcl_getoption ( server - > dummy_hcl , HCL_TRAIT , & trait ) ;
# if defined(HCL_BUILD_DEBUG)
if ( server - > cfg . trait & HCL_SERVER_TRAIT_DEBUG_GC ) trait | = HCL_TRAIT_DEBUG_GC ;
if ( server - > cfg . trait & HCL_SERVER_TRAIT_DEBUG_BIGINT ) trait | = HCL_TRAIT_DEBUG_BIGINT ;
# endif
hcl_setoption ( server - > dummy_hcl , HCL_TRAIT , & trait ) ;
return server ;
oops :
/* NOTE: pipe should be closed if jump to here is made after pipe() above */
if ( tmr ) hcl_tmr_close ( tmr ) ;
if ( hcl ) hcl_close ( hcl ) ;
if ( server ) HCL_MMGR_FREE ( mmgr , server ) ;
return HCL_NULL ;
}
void hcl_server_close ( hcl_server_t * server )
{
HCL_ASSERT ( server - > dummy_hcl , server - > listener . head = = HCL_NULL ) ;
HCL_ASSERT ( server - > dummy_hcl , server - > listener . count = = 0 ) ;
HCL_ASSERT ( server - > dummy_hcl , server - > listener . ep_fd = = - 1 ) ;
if ( server - > wid_map . ptr )
{
hcl_server_freemem ( server , server - > wid_map . ptr ) ;
server - > wid_map . capa = 0 ;
server - > wid_map . free_first = HCL_SERVER_WID_INVALID ;
server - > wid_map . free_last = HCL_SERVER_WID_INVALID ;
}
pthread_mutex_destroy ( & server - > log_mutex ) ;
pthread_mutex_destroy ( & server - > tmr_mutex ) ;
pthread_mutex_destroy ( & server - > worker_mutex ) ;
close ( server - > mux_pipe [ 0 ] ) ;
close ( server - > mux_pipe [ 1 ] ) ;
hcl_tmr_close ( server - > tmr ) ;
hcl_close ( server - > dummy_hcl ) ;
HCL_MMGR_FREE ( server - > _mmgr , server ) ;
}
static HCL_INLINE int prepare_to_acquire_wid ( hcl_server_t * server )
{
hcl_oow_t new_capa ;
hcl_oow_t i , j ;
hcl_server_wid_map_data_t * tmp ;
HCL_ASSERT ( server - > dummy_hcl , server - > wid_map . free_first = = HCL_SERVER_WID_INVALID ) ;
HCL_ASSERT ( server - > dummy_hcl , server - > wid_map . free_last = = HCL_SERVER_WID_INVALID ) ;
new_capa = HCL_ALIGN_POW2 ( server - > wid_map . capa + 1 , HCL_SERVER_WID_MAP_ALIGN ) ;
if ( new_capa > HCL_SERVER_WID_MAX )
{
if ( server - > wid_map . capa > = HCL_SERVER_WID_MAX )
{
hcl_server_seterrnum ( server , HCL_EFLOOD ) ;
return - 1 ;
}
new_capa = HCL_SERVER_WID_MAX ;
}
tmp = ( hcl_server_wid_map_data_t * ) hcl_server_reallocmem ( server , server - > wid_map . ptr , HCL_SIZEOF ( * tmp ) * new_capa ) ;
if ( ! tmp ) return - 1 ;
server - > wid_map . free_first = server - > wid_map . capa ;
for ( i = server - > wid_map . capa , j = server - > wid_map . capa + 1 ; j < new_capa ; i + + , j + + )
{
tmp [ i ] . used = 0 ;
tmp [ i ] . u . next = j ;
}
tmp [ i ] . used = 0 ;
tmp [ i ] . u . next = HCL_SERVER_WID_INVALID ;
server - > wid_map . free_last = i ;
server - > wid_map . ptr = tmp ;
server - > wid_map . capa = new_capa ;
return 0 ;
}
static HCL_INLINE void acquire_wid ( hcl_server_t * server , hcl_server_worker_t * worker )
{
hcl_oow_t wid ;
wid = server - > wid_map . free_first ;
worker - > wid = wid ;
server - > wid_map . free_first = server - > wid_map . ptr [ wid ] . u . next ;
if ( server - > wid_map . free_first = = HCL_SERVER_WID_INVALID ) server - > wid_map . free_last = HCL_SERVER_WID_INVALID ;
server - > wid_map . ptr [ wid ] . used = 1 ;
server - > wid_map . ptr [ wid ] . u . worker = worker ;
}
static HCL_INLINE void release_wid ( hcl_server_t * server , hcl_server_worker_t * worker )
{
hcl_oow_t wid ;
wid = worker - > wid ;
HCL_ASSERT ( server - > dummy_hcl , wid < server - > wid_map . capa & & wid ! = HCL_SERVER_WID_INVALID ) ;
server - > wid_map . ptr [ wid ] . used = 0 ;
server - > wid_map . ptr [ wid ] . u . next = HCL_SERVER_WID_INVALID ;
if ( server - > wid_map . free_last = = HCL_SERVER_WID_INVALID )
{
HCL_ASSERT ( server - > dummy_hcl , server - > wid_map . free_first < = HCL_SERVER_WID_INVALID ) ;
server - > wid_map . free_first = wid ;
}
else
{
server - > wid_map . ptr [ server - > wid_map . free_last ] . u . next = wid ;
}
server - > wid_map . free_last = wid ;
worker - > wid = HCL_SERVER_WID_INVALID ;
}
static hcl_server_worker_t * alloc_worker ( hcl_server_t * server , int cli_sck , const hcl_sckaddr_t * peeraddr )
{
hcl_server_worker_t * worker ;
worker = ( hcl_server_worker_t * ) hcl_server_allocmem ( server , HCL_SIZEOF ( * worker ) ) ;
if ( ! worker ) return HCL_NULL ;
HCL_MEMSET ( worker , 0 , HCL_SIZEOF ( * worker ) ) ;
worker - > state = HCL_SERVER_WORKER_STATE_ZOMBIE ;
worker - > opstate = HCL_SERVER_WORKER_OPSTATE_IDLE ;
worker - > sck = cli_sck ;
worker - > peeraddr = * peeraddr ;
worker - > server = server ;
server - > dummy_hcl - > vmprim . vm_gettime ( server - > dummy_hcl , & worker - > alloc_time ) ; /* TODO: the callback may return monotonic time. find a way to guarantee it is realtime??? */
if ( server - > wid_map . free_first = = HCL_SERVER_WID_INVALID & & prepare_to_acquire_wid ( server ) < = - 1 )
{
hcl_server_freemem ( server , worker ) ;
return HCL_NULL ;
}
acquire_wid ( server , worker ) ;
return worker ;
}
2024-04-20 03:02:22 +00:00
static void fini_worker_socket ( hcl_server_worker_t * worker )
2023-11-26 15:13:28 +00:00
{
if ( worker - > sck > = 0 )
{
2024-04-20 03:02:22 +00:00
if ( worker - > hcl )
2023-11-26 15:13:28 +00:00
{
2024-04-20 03:02:22 +00:00
HCL_LOG2 ( worker - > hcl , SERVER_LOGMASK_INFO , " Closing worker socket %d [%zu] \n " , worker - > sck , worker - > wid ) ;
2023-11-26 15:13:28 +00:00
}
else
{
/* this should be in the main server thread. i use dummy_hcl for logging */
HCL_LOG2 ( worker - > server - > dummy_hcl , SERVER_LOGMASK_INFO , " Closing worker socket %d [%zu] \n " , worker - > sck , worker - > wid ) ;
}
close ( worker - > sck ) ;
worker - > sck = - 1 ;
}
}
static void free_worker ( hcl_server_worker_t * worker )
{
2024-04-20 03:02:22 +00:00
fini_worker_socket ( worker ) ;
2023-11-26 15:13:28 +00:00
2024-04-20 03:02:22 +00:00
if ( worker - > hcl )
2023-11-26 15:13:28 +00:00
{
2024-04-20 03:02:22 +00:00
HCL_LOG1 ( worker - > hcl , SERVER_LOGMASK_INFO , " Killing worker [%zu] \n " , worker - > wid ) ;
2023-11-26 15:13:28 +00:00
}
else
{
/* this should be in the main server thread. i use dummy_hcl for logging */
HCL_LOG1 ( worker - > server - > dummy_hcl , SERVER_LOGMASK_INFO , " Killing worker [%zu] \n " , worker - > wid ) ;
}
release_wid ( worker - > server , worker ) ;
hcl_server_freemem ( worker - > server , worker ) ;
}
static void add_worker_to_server ( hcl_server_t * server , hcl_server_worker_state_t wstate , hcl_server_worker_t * worker )
{
HCL_ASSERT ( server - > dummy_hcl , worker - > server = = server ) ;
if ( server - > worker_list [ wstate ] . tail )
{
server - > worker_list [ wstate ] . tail - > next_worker = worker ;
worker - > prev_worker = server - > worker_list [ wstate ] . tail ;
server - > worker_list [ wstate ] . tail = worker ;
worker - > next_worker = HCL_NULL ;
}
else
{
server - > worker_list [ wstate ] . tail = worker ;
server - > worker_list [ wstate ] . head = worker ;
worker - > prev_worker = HCL_NULL ;
worker - > next_worker = HCL_NULL ;
}
server - > worker_list [ wstate ] . count + + ;
worker - > state = wstate ;
}
static void zap_worker_in_server ( hcl_server_t * server , hcl_server_worker_t * worker )
{
hcl_server_worker_state_t wstate ;
HCL_ASSERT ( server - > dummy_hcl , worker - > server = = server ) ;
wstate = worker - > state ;
if ( worker - > prev_worker ) worker - > prev_worker - > next_worker = worker - > next_worker ;
else server - > worker_list [ wstate ] . head = worker - > next_worker ;
if ( worker - > next_worker ) worker - > next_worker - > prev_worker = worker - > prev_worker ;
else server - > worker_list [ wstate ] . tail = worker - > prev_worker ;
HCL_ASSERT ( server - > dummy_hcl , server - > worker_list [ wstate ] . count > 0 ) ;
server - > worker_list [ wstate ] . count - - ;
worker - > state = HCL_SERVER_WORKER_STATE_ZOMBIE ;
worker - > prev_worker = HCL_NULL ;
worker - > next_worker = HCL_NULL ;
}
2024-04-14 17:23:55 +00:00
static int worker_step ( hcl_server_worker_t * worker )
{
2024-04-20 03:02:22 +00:00
hcl_xproto_t * proto = worker - > proto ;
2024-04-14 17:23:55 +00:00
hcl_server_t * server = worker - > server ;
2024-04-20 03:02:22 +00:00
hcl_t * hcl = worker - > hcl ;
2024-04-14 17:23:55 +00:00
struct pollfd pfd ;
int tmout , actual_tmout ;
ssize_t x ;
int n ;
//HCL_ASSERT (hcl, proto->rcv.len < proto->rcv.len_needed);
if ( HCL_UNLIKELY ( proto - > rcv . eof ) )
{
// TODO: may not be an error if writable needs to be checked...
hcl_seterrbfmt ( hcl , HCL_EGENERIC , " connection closed " ) ;
return - 1 ;
}
tmout = HCL_SECNSEC_TO_MSEC ( server - > cfg . worker_idle_timeout . sec , server - > cfg . worker_idle_timeout . nsec ) ;
actual_tmout = ( tmout < = 0 ) ? 10000 : tmout ;
pfd . fd = worker - > sck ;
pfd . events = 0 ;
if ( proto - > rcv . len < proto - > rcv . len_needed ) pfd . events | = POLLIN ;
//if (proto->snd.len > 0) pfd.events |= POLLOUT;
if ( pfd . events )
{
n = poll ( & pfd , 1 , actual_tmout ) ;
if ( n < = - 1 )
{
if ( errno = = EINTR ) return 0 ;
hcl_seterrwithsyserr ( hcl , 0 , errno ) ;
return - 1 ;
}
else if ( n = = 0 )
{
/* timed out - no activity on the pfd */
if ( tmout > 0 )
{
/* timeout explicity set. no activity for that duration. considered idle */
hcl_seterrbfmt ( hcl , HCL_EGENERIC , " no activity on the worker socket %d " , worker - > sck ) ;
return - 1 ;
}
2024-04-18 14:06:28 +00:00
goto carry_on ;
2024-04-14 17:23:55 +00:00
}
if ( pfd . revents & POLLERR )
{
hcl_seterrbfmt ( hcl , HCL_EGENERIC , " error condition detected on workder socket %d " , worker - > sck ) ;
return - 1 ;
}
if ( pfd . revents & POLLOUT )
{
}
if ( pfd . revents & POLLIN )
{
2024-04-21 07:29:55 +00:00
hcl_oow_t bcap ;
hcl_uint8_t * bptr ;
bptr = hcl_xproto_getbuf ( proto , & bcap ) ; ;
x = recv ( worker - > sck , bptr , bcap , 0 ) ;
2024-04-14 17:23:55 +00:00
if ( x < = - 1 )
{
2024-04-18 14:06:28 +00:00
if ( errno = = EINTR ) goto carry_on ; /* didn't read read */
2024-04-14 17:23:55 +00:00
hcl_seterrwithsyserr ( hcl , 0 , errno ) ;
return - 1 ;
}
2024-04-21 07:29:55 +00:00
if ( x = = 0 ) hcl_xproto_seteof ( proto , 1 ) ;
hcl_xproto_advbuf ( proto , x ) ;
2024-04-14 17:23:55 +00:00
}
}
/* the receiver buffer has enough data */
2024-04-20 03:02:22 +00:00
while ( hcl_xproto_ready ( worker - > proto ) )
2024-04-14 17:23:55 +00:00
{
2024-04-20 03:02:22 +00:00
if ( ( n = hcl_xproto_process ( worker - > proto ) ) < = - 1 )
2024-04-14 17:23:55 +00:00
{
/* TODO: proper error message */
return - 1 ;
}
2024-04-18 14:06:28 +00:00
if ( n = = 0 )
{
/* TODO: chceck if there is remaining data in the buffer...?? */
return 0 ; /* tell the caller to break the step loop */
}
2024-04-14 17:23:55 +00:00
}
2024-04-18 14:06:28 +00:00
carry_on :
return 1 ; /* carry on */
2024-04-14 17:23:55 +00:00
}
2024-04-20 03:02:22 +00:00
static int init_worker_hcl ( hcl_server_worker_t * worker )
{
hcl_server_t * server = worker - > server ;
hcl_t * hcl ;
worker_hcl_xtn_t * xtn ;
hcl_bitmask_t trait ;
hcl_cb_t hclcb ;
hcl = hcl_openstdwithmmgr ( hcl_server_getmmgr ( server ) , HCL_SIZEOF ( * xtn ) , HCL_NULL ) ;
if ( HCL_UNLIKELY ( ! hcl ) ) goto oops ;
/* replace the vmprim.log_write function */
hcl - > vmprim . log_write = server_log_write ;
xtn = ( worker_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
xtn - > worker = worker ;
hcl_setoption ( hcl , HCL_MOD_INCTX , & server - > cfg . module_inctx ) ;
hcl_setoption ( hcl , HCL_LOG_MASK , & server - > cfg . logmask ) ;
hcl_setcmgr ( hcl , hcl_server_getcmgr ( server ) ) ;
hcl_getoption ( hcl , HCL_TRAIT , & trait ) ;
# if defined(HCL_BUILD_DEBUG)
if ( server - > cfg . trait & HCL_SERVER_TRAIT_DEBUG_GC ) trait | = HCL_TRAIT_DEBUG_GC ;
if ( server - > cfg . trait & HCL_SERVER_TRAIT_DEBUG_BIGINT ) trait | = HCL_TRAIT_DEBUG_BIGINT ;
# endif
trait | = HCL_TRAIT_LANG_ENABLE_BLOCK ;
trait | = HCL_TRAIT_LANG_ENABLE_EOL ;
hcl_setoption ( hcl , HCL_TRAIT , & trait ) ;
HCL_MEMSET ( & hclcb , 0 , HCL_SIZEOF ( hclcb ) ) ;
/*hclcb.fini = fini_hcl;
hclcb . gc = gc_hcl ; */
hclcb . vm_startup = vm_startup ;
hclcb . vm_cleanup = vm_cleanup ;
hclcb . vm_checkbc = vm_checkbc ;
hcl_regcb ( hcl , & hclcb ) ;
if ( hcl_ignite ( hcl , server - > cfg . actor_heap_size ) < = - 1 ) goto oops ;
if ( hcl_addbuiltinprims ( hcl ) < = - 1 ) goto oops ;
if ( hcl_attachccio ( hcl , read_handler ) < = - 1 ) goto oops ;
if ( hcl_attachudio ( hcl , scan_handler , print_handler ) < = - 1 ) goto oops ;
if ( hcl_beginfeed ( hcl , on_fed_cnode ) < = - 1 ) goto oops ;
worker - > hcl = hcl ;
return 0 ;
oops :
if ( hcl ) hcl_close ( hcl ) ;
return - 1 ;
}
static void fini_worker_hcl ( hcl_server_worker_t * worker )
{
if ( HCL_LIKELY ( worker - > hcl ) )
{
hcl_endfeed ( worker - > hcl ) ;
hcl_close ( worker - > hcl ) ;
worker - > hcl = HCL_NULL ;
}
}
static int init_worker_proto ( hcl_server_worker_t * worker )
{
hcl_xproto_t * proto ;
2024-04-21 07:29:55 +00:00
proto_xtn_t * xtn ;
2024-04-20 03:02:22 +00:00
2024-04-21 07:29:55 +00:00
proto = hcl_xproto_open ( hcl_server_getmmgr ( worker - > server ) , HCL_SIZEOF ( * xtn ) ) ;
2024-04-20 03:02:22 +00:00
if ( HCL_UNLIKELY ( ! proto ) ) return - 1 ;
2024-04-21 07:29:55 +00:00
xtn = hcl_xproto_getxtn ( proto ) ;
xtn - > worker = worker ;
2024-04-20 03:02:22 +00:00
worker - > proto = proto ;
return 0 ;
}
static void fini_worker_proto ( hcl_server_worker_t * worker )
{
if ( HCL_LIKELY ( worker - > proto ) )
{
hcl_xproto_close ( worker - > proto ) ;
worker - > proto = HCL_NULL ;
}
}
2023-11-26 15:13:28 +00:00
static void * worker_main ( void * ctx )
{
hcl_server_worker_t * worker = ( hcl_server_worker_t * ) ctx ;
hcl_server_t * server = worker - > server ;
sigset_t set ;
2024-04-20 03:02:22 +00:00
int n ;
2023-11-26 15:13:28 +00:00
sigfillset ( & set ) ;
pthread_sigmask ( SIG_BLOCK , & set , HCL_NULL ) ;
worker - > thr = pthread_self ( ) ;
2024-04-20 03:02:22 +00:00
n = init_worker_hcl ( worker ) ;
if ( HCL_UNLIKELY ( n < = - 1 ) )
2023-11-26 15:13:28 +00:00
{
2024-04-20 03:02:22 +00:00
/* TODO: capture error ... */
return HCL_NULL ;
}
n = init_worker_proto ( worker ) ;
if ( HCL_UNLIKELY ( n < = - 1 ) )
{
fini_worker_hcl ( worker ) ;
2023-11-26 15:13:28 +00:00
return HCL_NULL ;
}
pthread_mutex_lock ( & server - > worker_mutex ) ;
add_worker_to_server ( server , HCL_SERVER_WORKER_STATE_ALIVE , worker ) ;
pthread_mutex_unlock ( & server - > worker_mutex ) ;
2024-04-14 17:23:55 +00:00
/* the worker loop */
2023-11-26 15:13:28 +00:00
while ( ! server - > stopreq )
{
2024-04-18 14:06:28 +00:00
int n ;
2023-11-26 15:13:28 +00:00
worker - > opstate = HCL_SERVER_WORKER_OPSTATE_WAIT ;
2024-04-14 09:33:15 +00:00
2024-04-18 14:06:28 +00:00
if ( ( n = worker_step ( worker ) ) < = 0 )
2023-11-26 15:13:28 +00:00
{
2024-04-18 14:06:28 +00:00
worker - > opstate = ( n < = - 1 ) ? HCL_SERVER_WORKER_OPSTATE_ERROR : HCL_SERVER_WORKER_OPSTATE_IDLE ;
2023-11-26 15:13:28 +00:00
break ;
}
}
2024-04-20 03:02:22 +00:00
hcl_xproto_close ( worker - > proto ) ;
2023-11-26 15:13:28 +00:00
worker - > proto = HCL_NULL ;
2024-04-20 03:02:22 +00:00
fini_worker_hcl ( worker ) ;
2023-11-26 15:13:28 +00:00
pthread_mutex_lock ( & server - > worker_mutex ) ;
2024-04-20 03:02:22 +00:00
fini_worker_socket ( worker ) ;
2023-11-26 15:13:28 +00:00
if ( ! worker - > claimed )
{
zap_worker_in_server ( server , worker ) ;
add_worker_to_server ( server , HCL_SERVER_WORKER_STATE_DEAD , worker ) ;
}
pthread_mutex_unlock ( & server - > worker_mutex ) ;
return HCL_NULL ;
}
static void purge_all_workers ( hcl_server_t * server , hcl_server_worker_state_t wstate )
{
hcl_server_worker_t * worker ;
while ( 1 )
{
pthread_mutex_lock ( & server - > worker_mutex ) ;
worker = server - > worker_list [ wstate ] . head ;
if ( worker )
{
zap_worker_in_server ( server , worker ) ;
worker - > claimed = 1 ;
if ( worker - > sck > = 0 ) shutdown ( worker - > sck , SHUT_RDWR ) ;
}
pthread_mutex_unlock ( & server - > worker_mutex ) ;
if ( ! worker ) break ;
pthread_join ( worker - > thr , HCL_NULL ) ;
free_worker ( worker ) ;
}
}
void hcl_server_logbfmt ( hcl_server_t * server , hcl_bitmask_t mask , const hcl_bch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_logbfmtv ( server - > dummy_hcl , mask , fmt , ap ) ;
va_end ( ap ) ;
}
void hcl_server_logufmt ( hcl_server_t * server , hcl_bitmask_t mask , const hcl_uch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_logufmtv ( server - > dummy_hcl , mask , fmt , ap ) ;
va_end ( ap ) ;
}
static void set_err_with_syserr ( hcl_server_t * server , int syserr_type , int syserr_code , const char * bfmt , . . . )
{
hcl_t * hcl = server - > dummy_hcl ;
hcl_errnum_t errnum ;
hcl_oow_t tmplen , tmplen2 ;
va_list ap ;
static hcl_bch_t b_dash [ ] = { ' ' , ' - ' , ' ' , ' \0 ' } ;
static hcl_uch_t u_dash [ ] = { ' ' , ' - ' , ' ' , ' \0 ' } ;
if ( hcl - > shuterr ) return ;
if ( hcl - > vmprim . syserrstrb )
{
errnum = hcl - > vmprim . syserrstrb ( hcl , syserr_type , syserr_code , hcl - > errmsg . tmpbuf . bch , HCL_COUNTOF ( hcl - > errmsg . tmpbuf . bch ) ) ;
va_start ( ap , bfmt ) ;
hcl_seterrbfmtv ( hcl , errnum , bfmt , ap ) ;
va_end ( ap ) ;
# if defined(HCL_OOCH_IS_UCH)
hcl - > errmsg . len + = hcl_copy_ucstr ( & hcl - > errmsg . buf [ hcl - > errmsg . len ] , HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len , u_dash ) ;
tmplen2 = HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len ;
hcl_convbtoucstr ( hcl , hcl - > errmsg . tmpbuf . bch , & tmplen , & hcl - > errmsg . buf [ hcl - > errmsg . len ] , & tmplen2 ) ;
hcl - > errmsg . len + = tmplen2 ; /* ignore conversion errors */
# else
hcl - > errmsg . len + = hcl_copy_bcstr ( & hcl - > errmsg . buf [ hcl - > errmsg . len ] , HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len , b_dash ) ;
hcl - > errmsg . len + = hcl_copy_bcstr ( & hcl - > errmsg . buf [ hcl - > errmsg . len ] , HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len , hcl - > errmsg . tmpbuf . bch ) ;
# endif
}
else
{
HCL_ASSERT ( hcl , hcl - > vmprim . syserrstru ! = HCL_NULL ) ;
errnum = hcl - > vmprim . syserrstru ( hcl , syserr_type , syserr_code , hcl - > errmsg . tmpbuf . uch , HCL_COUNTOF ( hcl - > errmsg . tmpbuf . uch ) ) ;
va_start ( ap , bfmt ) ;
hcl_seterrbfmtv ( hcl , errnum , bfmt , ap ) ;
va_end ( ap ) ;
# if defined(HCL_OOCH_IS_UCH)
hcl - > errmsg . len + = hcl_copy_ucstr ( & hcl - > errmsg . buf [ hcl - > errmsg . len ] , HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len , u_dash ) ;
hcl - > errmsg . len + = hcl_copy_ucstr ( & hcl - > errmsg . buf [ hcl - > errmsg . len ] , HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len , hcl - > errmsg . tmpbuf . uch ) ;
# else
hcl - > errmsg . len + = hcl_copy_bcstr ( & hcl - > errmsg . buf [ hcl - > errmsg . len ] , HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len , b_dash ) ;
tmplen2 = HCL_COUNTOF ( hcl - > errmsg . buf ) - hcl - > errmsg . len ;
hcl_convutobcstr ( hcl , hcl - > errmsg . tmpbuf . uch , & tmplen , & hcl - > errmsg . buf [ hcl - > errmsg . len ] , & tmplen2 ) ;
hcl - > errmsg . len + = tmplen2 ; /* ignore conversion errors */
# endif
}
server - > errnum = errnum ;
hcl_copy_oochars ( server - > errmsg . buf , server - > dummy_hcl - > errmsg . buf , HCL_COUNTOF ( server - > errmsg . buf ) ) ;
server - > errmsg . len = server - > dummy_hcl - > errmsg . len ;
}
static void free_all_listeners ( hcl_server_t * server )
{
hcl_server_listener_t * lp ;
struct epoll_event dummy_ev ;
epoll_ctl ( server - > listener . ep_fd , EPOLL_CTL_DEL , server - > mux_pipe [ 0 ] , & dummy_ev ) ;
while ( server - > listener . head )
{
lp = server - > listener . head ;
server - > listener . head = lp - > next_listener ;
server - > listener . count - - ;
epoll_ctl ( server - > listener . ep_fd , EPOLL_CTL_DEL , lp - > sck , & dummy_ev ) ;
close ( lp - > sck ) ;
hcl_server_freemem ( server , lp ) ;
}
HCL_ASSERT ( server - > dummy_hcl , server - > listener . ep_fd > = 0 ) ;
close ( server - > listener . ep_fd ) ;
server - > listener . ep_fd = - 1 ;
}
static int setup_listeners ( hcl_server_t * server , const hcl_bch_t * addrs )
{
const hcl_bch_t * addr_ptr , * comma ;
int ep_fd , fcv ;
struct epoll_event ev ;
ep_fd = epoll_create ( 1024 ) ;
if ( ep_fd < = - 1 )
{
set_err_with_syserr ( server , 0 , errno , " unable to create multiplexer " ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
return - 1 ;
}
# if defined(O_CLOEXEC)
fcv = fcntl ( ep_fd , F_GETFD , 0 ) ;
if ( fcv > = 0 ) fcntl ( ep_fd , F_SETFD , fcv | O_CLOEXEC ) ;
# endif
HCL_MEMSET ( & ev , 0 , HCL_SIZEOF ( ev ) ) ;
ev . events = EPOLLIN | EPOLLHUP | EPOLLERR ;
ev . data . fd = server - > mux_pipe [ 0 ] ;
if ( epoll_ctl ( ep_fd , EPOLL_CTL_ADD , server - > mux_pipe [ 0 ] , & ev ) < = - 1 )
{
set_err_with_syserr ( server , 0 , errno , " unable to register pipe %d to multiplexer " , server - > mux_pipe [ 0 ] ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
close ( ep_fd ) ;
return - 1 ;
}
server - > listener . ep_fd = ep_fd ;
addr_ptr = addrs ;
while ( 1 )
{
hcl_sckaddr_t srv_addr ;
int srv_fd , sck_fam , optval ;
hcl_scklen_t srv_len ;
hcl_oow_t addr_len ;
hcl_server_listener_t * listener ;
comma = hcl_find_bchar_in_bcstr ( addr_ptr , ' , ' ) ;
addr_len = comma ? comma - addr_ptr : hcl_count_bcstr ( addr_ptr ) ;
/* [NOTE] no whitespaces are allowed before and after a comma */
sck_fam = hcl_bchars_to_sckaddr ( addr_ptr , addr_len , & srv_addr , & srv_len ) ;
if ( sck_fam < = - 1 )
{
hcl_server_seterrbfmt ( server , HCL_EINVAL , " unable to convert address - %.*hs " , addr_len , addr_ptr ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
goto next_segment ;
}
srv_fd = socket ( sck_fam , SOCK_STREAM , 0 ) ;
if ( srv_fd < = - 1 )
{
set_err_with_syserr ( server , 0 , errno , " unable to open server socket for %.*hs " , addr_len , addr_ptr ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
goto next_segment ;
}
optval = 1 ;
setsockopt ( srv_fd , SOL_SOCKET , SO_REUSEADDR , & optval , HCL_SIZEOF ( int ) ) ;
# if defined(O_NONBLOCK) || defined(O_CLOEXEC)
fcv = fcntl ( srv_fd , F_GETFD , 0 ) ;
if ( fcv > = 0 )
{
# if defined(O_NONBLOCK)
fcv | = O_NONBLOCK ;
# endif
# if defined(O_CLOEXEC)
fcv | = O_CLOEXEC ;
# endif
fcntl ( srv_fd , F_SETFL , fcv ) ;
}
# endif
if ( bind ( srv_fd , ( struct sockaddr * ) & srv_addr , srv_len ) = = - 1 )
{
set_err_with_syserr ( server , 0 , errno , " unable to bind server socket %d for %.*hs " , srv_fd , addr_len , addr_ptr ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
close ( srv_fd ) ;
goto next_segment ;
}
if ( listen ( srv_fd , 128 ) < = - 1 )
{
set_err_with_syserr ( server , 0 , errno , " unable to listen on server socket %d for %.*hs " , srv_fd , addr_len , addr_ptr ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
close ( srv_fd ) ;
goto next_segment ;
}
HCL_MEMSET ( & ev , 0 , HCL_SIZEOF ( ev ) ) ;
ev . events = EPOLLIN | EPOLLHUP | EPOLLERR ;
ev . data . fd = srv_fd ;
if ( epoll_ctl ( ep_fd , EPOLL_CTL_ADD , srv_fd , & ev ) < = - 1 )
{
set_err_with_syserr ( server , 0 , errno , " unable to register server socket %d to multiplexer for %.*hs " , srv_fd , addr_len , addr_ptr ) ;
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " %js \n " , hcl_server_geterrmsg ( server ) ) ;
close ( srv_fd ) ;
goto next_segment ;
}
listener = ( hcl_server_listener_t * ) hcl_server_allocmem ( server , HCL_SIZEOF ( * listener ) ) ;
if ( ! listener )
{
close ( srv_fd ) ;
goto next_segment ;
}
HCL_MEMSET ( listener , 0 , HCL_SIZEOF ( * listener ) ) ;
listener - > sck = srv_fd ;
listener - > sckaddr = srv_addr ;
listener - > next_listener = server - > listener . head ;
server - > listener . head = listener ;
server - > listener . count + + ;
next_segment :
if ( ! comma ) break ;
addr_ptr = comma + 1 ;
}
if ( ! server - > listener . head )
{
/* no valid server has been configured */
hcl_server_seterrbfmt ( server , HCL_EINVAL , " unable to set up listeners with %hs " , addrs ) ;
free_all_listeners ( server ) ;
return - 1 ;
}
return 0 ;
}
int hcl_server_start ( hcl_server_t * server , const hcl_bch_t * addrs )
{
int xret = 0 , fcv ;
pthread_attr_t thr_attr ;
if ( setup_listeners ( server , addrs ) < = - 1 ) return - 1 ;
pthread_attr_init ( & thr_attr ) ;
pthread_attr_setstacksize ( & thr_attr , server - > cfg . worker_stack_size ) ;
server - > stopreq = 0 ;
while ( ! server - > stopreq )
{
hcl_sckaddr_t cli_addr ;
hcl_scklen_t cli_len ;
int cli_fd ;
pthread_t thr ;
hcl_ntime_t tmout ;
hcl_server_worker_t * worker ;
int n ;
pthread_mutex_lock ( & server - > tmr_mutex ) ;
n = hcl_tmr_gettmout ( server - > tmr , HCL_NULL , & tmout ) ;
pthread_mutex_unlock ( & server - > tmr_mutex ) ;
if ( n < = - 1 ) HCL_INIT_NTIME ( & tmout , 10 , 0 ) ;
n = epoll_wait ( server - > listener . ep_fd , server - > listener . ev_buf , HCL_COUNTOF ( server - > listener . ev_buf ) , HCL_SECNSEC_TO_MSEC ( tmout . sec , tmout . nsec ) ) ;
purge_all_workers ( server , HCL_SERVER_WORKER_STATE_DEAD ) ;
if ( n < = - 1 )
{
if ( server - > stopreq ) break ; /* normal termination requested */
if ( errno = = EINTR ) continue ; /* interrupted but not termination requested */
set_err_with_syserr ( server , 0 , errno , " unable to poll for events in server " ) ;
xret = - 1 ;
break ;
}
pthread_mutex_lock ( & server - > tmr_mutex ) ;
hcl_tmr_fire ( server - > tmr , HCL_NULL , HCL_NULL ) ;
pthread_mutex_unlock ( & server - > tmr_mutex ) ;
while ( n > 0 )
{
struct epoll_event * evp ;
- - n ;
evp = & server - > listener . ev_buf [ n ] ;
if ( ! evp - > events /*& (POLLIN | POLLHUP | POLLERR) */ ) continue ;
if ( evp - > data . fd = = server - > mux_pipe [ 0 ] )
{
char tmp [ 128 ] ;
while ( read ( server - > mux_pipe [ 0 ] , tmp , HCL_SIZEOF ( tmp ) ) > 0 ) /* nothing */ ;
}
else
{
/* the reset should be the listener's socket */
cli_len = HCL_SIZEOF ( cli_addr ) ;
cli_fd = accept ( evp - > data . fd , ( struct sockaddr * ) & cli_addr , & cli_len ) ;
if ( cli_fd = = - 1 )
{
if ( server - > stopreq ) break ; /* normal termination requested */
if ( errno = = EINTR ) continue ; /* interrupted but no termination requested */
# if defined(EWOULDBLOCK) && defined(EAGAIN) && (EWOULDBLOCK != EAGAIN)
if ( errno = = EWOULDBLOCK | | errno = = EAGAIN ) continue ;
# elif defined(EWOULDBLOCK)
if ( errno = = EWOULDBLOCK ) continue ;
# elif defined(EAGAIN)
if ( errno = = EAGAIN ) continue ;
# endif
set_err_with_syserr ( server , 0 , errno , " unable to accept worker on server socket %d " , evp - > data . fd ) ;
xret = - 1 ;
break ;
}
# if defined(O_NONBLOCK) || defined(O_CLOEXEC)
fcv = fcntl ( cli_fd , F_GETFD , 0 ) ;
if ( fcv > = 0 )
{
# if defined(O_NONBLOCK)
fcv & = ~ O_NONBLOCK ; // force the accepted socket to be blocking
# endif
# if defined(O_CLOEXEC)
fcv | = O_CLOEXEC ;
# endif
fcntl ( cli_fd , F_SETFD , fcv ) ;
}
# endif
if ( server - > cfg . worker_max_count > 0 )
{
int flood ;
pthread_mutex_lock ( & server - > worker_mutex ) ;
flood = ( server - > worker_list [ HCL_SERVER_WORKER_STATE_ALIVE ] . count > = server - > cfg . worker_max_count ) ;
pthread_mutex_unlock ( & server - > worker_mutex ) ;
if ( flood )
{
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " Not accepting connection for too many workers - socket %d \n " , cli_fd ) ;
goto drop_connection ;
}
}
worker = alloc_worker ( server , cli_fd , & cli_addr ) ;
if ( ! worker )
{
HCL_LOG1 ( server - > dummy_hcl , SERVER_LOGMASK_ERROR , " Unable to accomodate worker - socket %d \n " , cli_fd ) ;
drop_connection :
close ( cli_fd ) ;
}
else
{
HCL_LOG2 ( server - > dummy_hcl , SERVER_LOGMASK_INFO , " Accomodated worker [%zu] - socket %d \n " , worker - > wid , cli_fd ) ;
if ( pthread_create ( & thr , & thr_attr , worker_main , worker ) ! = 0 )
{
free_worker ( worker ) ;
}
}
}
}
}
purge_all_workers ( server , HCL_SERVER_WORKER_STATE_ALIVE ) ;
purge_all_workers ( server , HCL_SERVER_WORKER_STATE_DEAD ) ;
pthread_attr_destroy ( & thr_attr ) ;
free_all_listeners ( server ) ;
return xret ;
}
void hcl_server_stop ( hcl_server_t * server )
{
server - > stopreq = 1 ;
write ( server - > mux_pipe [ 1 ] , " Q " , 1 ) ; /* don't care about failure */
}
int hcl_server_setoption ( hcl_server_t * server , hcl_server_option_t id , const void * value )
{
switch ( id )
{
case HCL_SERVER_TRAIT :
server - > cfg . trait = * ( const hcl_bitmask_t * ) value ;
if ( server - > dummy_hcl )
{
/* setting this affects the dummy hcl immediately.
* existing hcl instances inside worker threads won ' t get
* affected . new hcl instances to be created later
* is supposed to use the new value */
hcl_bitmask_t trait ;
hcl_getoption ( server - > dummy_hcl , HCL_TRAIT , & trait ) ;
# if defined(HCL_BUILD_DEBUG)
if ( server - > cfg . trait & HCL_SERVER_TRAIT_DEBUG_GC ) trait | = HCL_TRAIT_DEBUG_GC ;
if ( server - > cfg . trait & HCL_SERVER_TRAIT_DEBUG_BIGINT ) trait | = HCL_TRAIT_DEBUG_BIGINT ;
# endif
hcl_setoption ( server - > dummy_hcl , HCL_TRAIT , & trait ) ;
}
return 0 ;
case HCL_SERVER_LOG_MASK :
server - > cfg . logmask = * ( const hcl_bitmask_t * ) value ;
if ( server - > dummy_hcl )
{
/* setting this affects the dummy hcl immediately.
* existing hcl instances inside worker threads won ' t get
* affected . new hcl instances to be created later
* is supposed to use the new value */
hcl_setoption ( server - > dummy_hcl , HCL_LOG_MASK , value ) ;
}
return 0 ;
case HCL_SERVER_WORKER_MAX_COUNT :
server - > cfg . worker_max_count = * ( hcl_oow_t * ) value ;
return 0 ;
case HCL_SERVER_WORKER_STACK_SIZE :
server - > cfg . worker_stack_size = * ( hcl_oow_t * ) value ;
return 0 ;
case HCL_SERVER_WORKER_IDLE_TIMEOUT :
server - > cfg . worker_idle_timeout = * ( hcl_ntime_t * ) value ;
return 0 ;
case HCL_SERVER_ACTOR_HEAP_SIZE :
server - > cfg . actor_heap_size = * ( hcl_oow_t * ) value ;
return 0 ;
case HCL_SERVER_ACTOR_MAX_RUNTIME :
server - > cfg . actor_max_runtime = * ( hcl_ntime_t * ) value ;
return 0 ;
case HCL_SERVER_SCRIPT_INCLUDE_PATH :
hcl_copy_oocstr ( server - > cfg . script_include_path , HCL_COUNTOF ( server - > cfg . script_include_path ) , ( const hcl_ooch_t * ) value ) ;
return 0 ;
case HCL_SERVER_MODULE_INCTX :
server - > cfg . module_inctx = * ( void * * ) value ;
return 0 ;
}
hcl_server_seterrnum ( server , HCL_EINVAL ) ;
return - 1 ;
}
int hcl_server_getoption ( hcl_server_t * server , hcl_server_option_t id , void * value )
{
switch ( id )
{
case HCL_SERVER_TRAIT :
* ( hcl_bitmask_t * ) value = server - > cfg . trait ;
return 0 ;
case HCL_SERVER_LOG_MASK :
* ( hcl_bitmask_t * ) value = server - > cfg . logmask ;
return 0 ;
case HCL_SERVER_WORKER_MAX_COUNT :
* ( hcl_oow_t * ) value = server - > cfg . worker_max_count ;
return 0 ;
case HCL_SERVER_WORKER_STACK_SIZE :
* ( hcl_oow_t * ) value = server - > cfg . worker_stack_size ;
return 0 ;
case HCL_SERVER_WORKER_IDLE_TIMEOUT :
* ( hcl_ntime_t * ) value = server - > cfg . worker_idle_timeout ;
return 0 ;
case HCL_SERVER_ACTOR_HEAP_SIZE :
* ( hcl_oow_t * ) value = server - > cfg . actor_heap_size ;
return 0 ;
case HCL_SERVER_ACTOR_MAX_RUNTIME :
* ( hcl_ntime_t * ) value = server - > cfg . actor_max_runtime ;
return 0 ;
case HCL_SERVER_SCRIPT_INCLUDE_PATH :
* ( hcl_ooch_t * * ) value = server - > cfg . script_include_path ;
return 0 ;
case HCL_SERVER_MODULE_INCTX :
* ( void * * ) value = server - > cfg . module_inctx ;
return 0 ;
} ;
hcl_server_seterrnum ( server , HCL_EINVAL ) ;
return - 1 ;
}
void * hcl_server_getxtn ( hcl_server_t * server )
{
return ( void * ) ( ( hcl_uint8_t * ) server + server - > _instsize ) ;
}
hcl_mmgr_t * hcl_server_getmmgr ( hcl_server_t * server )
{
return server - > _mmgr ;
}
hcl_cmgr_t * hcl_server_getcmgr ( hcl_server_t * server )
{
return server - > _cmgr ;
}
void hcl_server_setcmgr ( hcl_server_t * server , hcl_cmgr_t * cmgr )
{
server - > _cmgr = cmgr ;
}
hcl_errnum_t hcl_server_geterrnum ( hcl_server_t * server )
{
return server - > errnum ;
}
const hcl_ooch_t * hcl_server_geterrstr ( hcl_server_t * server )
{
return hcl_errnum_to_errstr ( server - > errnum ) ;
}
const hcl_ooch_t * hcl_server_geterrmsg ( hcl_server_t * server )
{
if ( server - > errmsg . len < = 0 ) return hcl_errnum_to_errstr ( server - > errnum ) ;
return server - > errmsg . buf ;
}
void hcl_server_seterrnum ( hcl_server_t * server , hcl_errnum_t errnum )
{
/*if (server->shuterr) return; */
server - > errnum = errnum ;
server - > errmsg . len = 0 ;
}
void hcl_server_seterrbfmt ( hcl_server_t * server , hcl_errnum_t errnum , const hcl_bch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_seterrbfmtv ( server - > dummy_hcl , errnum , fmt , ap ) ;
va_end ( ap ) ;
HCL_ASSERT ( server - > dummy_hcl , HCL_COUNTOF ( server - > errmsg . buf ) = = HCL_COUNTOF ( server - > dummy_hcl - > errmsg . buf ) ) ;
server - > errnum = errnum ;
hcl_copy_oochars ( server - > errmsg . buf , server - > dummy_hcl - > errmsg . buf , HCL_COUNTOF ( server - > errmsg . buf ) ) ;
server - > errmsg . len = server - > dummy_hcl - > errmsg . len ;
}
void hcl_server_seterrufmt ( hcl_server_t * server , hcl_errnum_t errnum , const hcl_uch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_seterrufmtv ( server - > dummy_hcl , errnum , fmt , ap ) ;
va_end ( ap ) ;
HCL_ASSERT ( server - > dummy_hcl , HCL_COUNTOF ( server - > errmsg . buf ) = = HCL_COUNTOF ( server - > dummy_hcl - > errmsg . buf ) ) ;
server - > errnum = errnum ;
server - > errnum = errnum ;
hcl_copy_oochars ( server - > errmsg . buf , server - > dummy_hcl - > errmsg . buf , HCL_COUNTOF ( server - > errmsg . buf ) ) ;
server - > errmsg . len = server - > dummy_hcl - > errmsg . len ;
}
void * hcl_server_allocmem ( hcl_server_t * server , hcl_oow_t size )
{
void * ptr ;
ptr = HCL_MMGR_ALLOC ( server - > _mmgr , size ) ;
if ( ! ptr ) hcl_server_seterrnum ( server , HCL_ESYSMEM ) ;
return ptr ;
}
void * hcl_server_callocmem ( hcl_server_t * server , hcl_oow_t size )
{
void * ptr ;
ptr = HCL_MMGR_ALLOC ( server - > _mmgr , size ) ;
if ( ! ptr ) hcl_server_seterrnum ( server , HCL_ESYSMEM ) ;
else HCL_MEMSET ( ptr , 0 , size ) ;
return ptr ;
}
void * hcl_server_reallocmem ( hcl_server_t * server , void * ptr , hcl_oow_t size )
{
ptr = HCL_MMGR_REALLOC ( server - > _mmgr , ptr , size ) ;
if ( ! ptr ) hcl_server_seterrnum ( server , HCL_ESYSMEM ) ;
return ptr ;
}
void hcl_server_freemem ( hcl_server_t * server , void * ptr )
{
HCL_MMGR_FREE ( server - > _mmgr , ptr ) ;
}
2024-04-14 09:33:15 +00:00
/* ------------------------------------------------------------------------ */
#if 0
hcl_client_proto_t * hcl_client_proto_open ( hcl_mmgr_t * mmgr , hcl_oow_t xtnsize , hcl_errnum_t * errnum )
{
}
void hcl_client_proto_close ( hcl_client_proto_t * proto )
{
}
int hcl_client_proto_send_code ( hcl_client_proto_t * client )
{
}
int hcl_client_proto_handle_response ( hcl_client_proto_t * proto )
{
}
# endif
/* ------------------------------------------------------------------------ */
struct client_hcl_xtn_t
{
hcl_client_t * client ;
} ;
typedef struct client_hcl_xtn_t client_hcl_xtn_t ;
struct hcl_client_t
{
hcl_oow_t _instsize ;
hcl_mmgr_t * _mmgr ;
hcl_cmgr_t * _cmgr ;
hcl_client_prim_t prim ;
hcl_t * dummy_hcl ;
hcl_errnum_t errnum ;
struct
{
hcl_ooch_t buf [ HCL_ERRMSG_CAPA ] ;
hcl_oow_t len ;
} errmsg ;
struct
{
hcl_bitmask_t trait ;
hcl_bitmask_t logmask ;
} cfg ;
} ;
/* ========================================================================= */
static void client_log_write_for_dummy ( hcl_t * hcl , hcl_bitmask_t mask , const hcl_ooch_t * msg , hcl_oow_t len )
{
client_hcl_xtn_t * xtn = ( client_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
hcl_client_t * client ;
client = xtn - > client ;
client - > prim . log_write ( client , mask , msg , len ) ;
}
hcl_client_t * hcl_client_open ( hcl_mmgr_t * mmgr , hcl_oow_t xtnsize , hcl_client_prim_t * prim , hcl_errnum_t * errnum )
{
hcl_client_t * client ;
hcl_t * hcl ;
client_hcl_xtn_t * xtn ;
client = ( hcl_client_t * ) HCL_MMGR_ALLOC ( mmgr , HCL_SIZEOF ( * client ) + xtnsize ) ;
if ( ! client )
{
if ( errnum ) * errnum = HCL_ESYSMEM ;
return HCL_NULL ;
}
hcl = hcl_openstdwithmmgr ( mmgr , HCL_SIZEOF ( * xtn ) , errnum ) ;
if ( ! hcl )
{
HCL_MMGR_FREE ( mmgr , client ) ;
return HCL_NULL ;
}
/* replace the vmprim.log_write function */
hcl - > vmprim . log_write = client_log_write_for_dummy ;
xtn = ( client_hcl_xtn_t * ) hcl_getxtn ( hcl ) ;
xtn - > client = client ;
HCL_MEMSET ( client , 0 , HCL_SIZEOF ( * client ) + xtnsize ) ;
client - > _instsize = HCL_SIZEOF ( * client ) ;
client - > _mmgr = mmgr ;
client - > _cmgr = hcl_get_utf8_cmgr ( ) ;
client - > prim = * prim ;
client - > dummy_hcl = hcl ;
client - > cfg . logmask = ~ ( hcl_bitmask_t ) 0 ;
/* the dummy hcl is used for this client to perform primitive operations
* such as getting system time or logging . so the heap size doesn ' t
* need to be changed from the tiny value set above . */
hcl_setoption ( client - > dummy_hcl , HCL_LOG_MASK , & client - > cfg . logmask ) ;
hcl_setcmgr ( client - > dummy_hcl , client - > _cmgr ) ;
return client ;
}
void hcl_client_close ( hcl_client_t * client )
{
hcl_close ( client - > dummy_hcl ) ;
HCL_MMGR_FREE ( client - > _mmgr , client ) ;
}
int hcl_client_setoption ( hcl_client_t * client , hcl_client_option_t id , const void * value )
{
switch ( id )
{
case HCL_CLIENT_TRAIT :
client - > cfg . trait = * ( const hcl_bitmask_t * ) value ;
return 0 ;
case HCL_CLIENT_LOG_MASK :
client - > cfg . logmask = * ( const hcl_bitmask_t * ) value ;
if ( client - > dummy_hcl )
{
/* setting this affects the dummy hcl immediately.
* existing hcl instances inside worker threads won ' t get
* affected . new hcl instances to be created later
* is supposed to use the new value */
hcl_setoption ( client - > dummy_hcl , HCL_LOG_MASK , value ) ;
}
return 0 ;
}
hcl_client_seterrnum ( client , HCL_EINVAL ) ;
return - 1 ;
}
int hcl_client_getoption ( hcl_client_t * client , hcl_client_option_t id , void * value )
{
switch ( id )
{
case HCL_CLIENT_TRAIT :
* ( hcl_bitmask_t * ) value = client - > cfg . trait ;
return 0 ;
case HCL_CLIENT_LOG_MASK :
* ( hcl_bitmask_t * ) value = client - > cfg . logmask ;
return 0 ;
} ;
hcl_client_seterrnum ( client , HCL_EINVAL ) ;
return - 1 ;
}
void * hcl_client_getxtn ( hcl_client_t * client )
{
return ( void * ) ( ( hcl_uint8_t * ) client + client - > _instsize ) ;
}
hcl_mmgr_t * hcl_client_getmmgr ( hcl_client_t * client )
{
return client - > _mmgr ;
}
hcl_cmgr_t * hcl_client_getcmgr ( hcl_client_t * client )
{
return client - > _cmgr ;
}
void hcl_client_setcmgr ( hcl_client_t * client , hcl_cmgr_t * cmgr )
{
client - > _cmgr = cmgr ;
}
hcl_errnum_t hcl_client_geterrnum ( hcl_client_t * client )
{
return client - > errnum ;
}
const hcl_ooch_t * hcl_client_geterrstr ( hcl_client_t * client )
{
return hcl_errnum_to_errstr ( client - > errnum ) ;
}
const hcl_ooch_t * hcl_client_geterrmsg ( hcl_client_t * client )
{
if ( client - > errmsg . len < = 0 ) return hcl_errnum_to_errstr ( client - > errnum ) ;
return client - > errmsg . buf ;
}
void hcl_client_seterrnum ( hcl_client_t * client , hcl_errnum_t errnum )
{
/*if (client->shuterr) return; */
client - > errnum = errnum ;
client - > errmsg . len = 0 ;
}
void hcl_client_seterrbfmt ( hcl_client_t * client , hcl_errnum_t errnum , const hcl_bch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_seterrbfmtv ( client - > dummy_hcl , errnum , fmt , ap ) ;
va_end ( ap ) ;
HCL_ASSERT ( client - > dummy_hcl , HCL_COUNTOF ( client - > errmsg . buf ) = = HCL_COUNTOF ( client - > dummy_hcl - > errmsg . buf ) ) ;
client - > errnum = errnum ;
hcl_copy_oochars ( client - > errmsg . buf , client - > dummy_hcl - > errmsg . buf , HCL_COUNTOF ( client - > errmsg . buf ) ) ;
client - > errmsg . len = client - > dummy_hcl - > errmsg . len ;
}
void hcl_client_seterrufmt ( hcl_client_t * client , hcl_errnum_t errnum , const hcl_uch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_seterrufmtv ( client - > dummy_hcl , errnum , fmt , ap ) ;
va_end ( ap ) ;
HCL_ASSERT ( client - > dummy_hcl , HCL_COUNTOF ( client - > errmsg . buf ) = = HCL_COUNTOF ( client - > dummy_hcl - > errmsg . buf ) ) ;
client - > errnum = errnum ;
hcl_copy_oochars ( client - > errmsg . buf , client - > dummy_hcl - > errmsg . buf , HCL_COUNTOF ( client - > errmsg . buf ) ) ;
client - > errmsg . len = client - > dummy_hcl - > errmsg . len ;
}
/* ========================================================================= */
void hcl_client_logbfmt ( hcl_client_t * client , hcl_bitmask_t mask , const hcl_bch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_logbfmtv ( client - > dummy_hcl , mask , fmt , ap ) ;
va_end ( ap ) ;
}
void hcl_client_logufmt ( hcl_client_t * client , hcl_bitmask_t mask , const hcl_uch_t * fmt , . . . )
{
va_list ap ;
va_start ( ap , fmt ) ;
hcl_logufmtv ( client - > dummy_hcl , mask , fmt , ap ) ;
va_end ( ap ) ;
}
/* ========================================================================= */
void * hcl_client_allocmem ( hcl_client_t * client , hcl_oow_t size )
{
void * ptr ;
ptr = HCL_MMGR_ALLOC ( client - > _mmgr , size ) ;
if ( ! ptr ) hcl_client_seterrnum ( client , HCL_ESYSMEM ) ;
return ptr ;
}
void * hcl_client_callocmem ( hcl_client_t * client , hcl_oow_t size )
{
void * ptr ;
ptr = HCL_MMGR_ALLOC ( client - > _mmgr , size ) ;
if ( ! ptr ) hcl_client_seterrnum ( client , HCL_ESYSMEM ) ;
else HCL_MEMSET ( ptr , 0 , size ) ;
return ptr ;
}
void * hcl_client_reallocmem ( hcl_client_t * client , void * ptr , hcl_oow_t size )
{
ptr = HCL_MMGR_REALLOC ( client - > _mmgr , ptr , size ) ;
if ( ! ptr ) hcl_client_seterrnum ( client , HCL_ESYSMEM ) ;
return ptr ;
}
void hcl_client_freemem ( hcl_client_t * client , void * ptr )
{
HCL_MMGR_FREE ( client - > _mmgr , ptr ) ;
}
2024-04-20 03:02:22 +00:00