/* * $Id$ * Copyright (c) 2016-2018 Chung, Hyung-Hwan. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "hcl-prv.h" #define PROC_STATE_RUNNING 3 #define PROC_STATE_WAITING 2 #define PROC_STATE_RUNNABLE 1 #define PROC_STATE_SUSPENDED 0 #define PROC_STATE_TERMINATED -1 static HCL_INLINE const char* proc_state_to_string (int state) { static const hcl_bch_t* str[] = { "TERMINATED", "SUSPENDED", "RUNNABLE", "WAITING", "RUNNING" }; return str[state + 1]; } #define PROC_MAP_INC 64 #define SEM_LIST_INC 256 #define SEM_HEAP_INC 256 #define SEM_LIST_MAX (SEM_LIST_INC * 1000) #define SEM_HEAP_MAX (SEM_HEAP_INC * 1000) #define SEM_HEAP_PARENT(x) (((x) - 1) / 2) #define SEM_HEAP_LEFT(x) ((x) * 2 + 1) #define SEM_HEAP_RIGHT(x) ((x) * 2 + 2) #define SEM_HEAP_EARLIER_THAN(stx,x,y) ( \ (HCL_OOP_TO_SMOOI((x)->heap_ftime_sec) < HCL_OOP_TO_SMOOI((y)->heap_ftime_sec)) || \ (HCL_OOP_TO_SMOOI((x)->heap_ftime_sec) == HCL_OOP_TO_SMOOI((y)->heap_ftime_sec) && HCL_OOP_TO_SMOOI((x)->heap_ftime_nsec) < HCL_OOP_TO_SMOOI((y)->heap_ftime_nsec)) \ ) #define LOAD_IP(hcl, v_ctx) ((hcl)->ip = HCL_OOP_TO_SMOOI((v_ctx)->ip)) #define STORE_IP(hcl, v_ctx) ((v_ctx)->ip = HCL_SMOOI_TO_OOP((hcl)->ip)) #define LOAD_SP(hcl, v_ctx) ((hcl)->sp = HCL_OOP_TO_SMOOI((v_ctx)->sp)) #define STORE_SP(hcl, v_ctx) ((v_ctx)->sp = HCL_SMOOI_TO_OOP((hcl)->sp)) #define LOAD_ACTIVE_IP(hcl) LOAD_IP(hcl, (hcl)->active_context) #define STORE_ACTIVE_IP(hcl) STORE_IP(hcl, (hcl)->active_context) #define LOAD_ACTIVE_SP(hcl) LOAD_SP(hcl, (hcl)->processor->active) #define STORE_ACTIVE_SP(hcl) STORE_SP(hcl, (hcl)->processor->active) #define SWITCH_ACTIVE_CONTEXT(hcl,v_ctx) \ do \ { \ STORE_ACTIVE_IP (hcl); \ (hcl)->active_context = (v_ctx); \ (hcl)->active_function = (hcl)->active_context->origin->receiver_or_base; \ (hcl)->active_code = HCL_FUNCTION_GET_CODE_BYTE((hcl)->active_function); \ LOAD_ACTIVE_IP (hcl); \ (hcl)->processor->active->current_context = (hcl)->active_context; \ } while (0) /*#define FETCH_BYTE_CODE(hcl) ((hcl)->code.bc.arr->slot[(hcl)->ip++])*/ #define FETCH_BYTE_CODE(hcl) ((hcl)->active_code[(hcl)->ip++]) #define FETCH_BYTE_CODE_TO(hcl, v_oow) (v_oow = FETCH_BYTE_CODE(hcl)) #if (HCL_HCL_CODE_LONG_PARAM_SIZE == 2) # define FETCH_PARAM_CODE_TO(hcl, v_oow) \ do { \ v_oow = FETCH_BYTE_CODE(hcl); \ v_oow = (v_oow << 8) | FETCH_BYTE_CODE(hcl); \ } while (0) #else # define FETCH_PARAM_CODE_TO(hcl, v_oow) (v_oow = FETCH_BYTE_CODE(hcl)) #endif #if defined(HCL_DEBUG_VM_EXEC) # define LOG_MASK_INST (HCL_LOG_IC | HCL_LOG_MNEMONIC | HCL_LOG_INFO) # define LOG_INST_0(hcl,fmt) HCL_LOG1(hcl, LOG_MASK_INST, "%010zd " fmt "\n", fetched_instruction_pointer) # define LOG_INST_1(hcl,fmt,a1) HCL_LOG2(hcl, LOG_MASK_INST, "%010zd " fmt "\n",fetched_instruction_pointer, a1) # define LOG_INST_2(hcl,fmt,a1,a2) HCL_LOG3(hcl, LOG_MASK_INST, "%010zd " fmt "\n", fetched_instruction_pointer, a1, a2) # define LOG_INST_3(hcl,fmt,a1,a2,a3) HCL_LOG4(hcl, LOG_MASK_INST, "%010zd " fmt "\n", fetched_instruction_pointer, a1, a2, a3) # define LOG_INST_4(hcl,fmt,a1,a2,a3,a4) HCL_LOG5(hcl, LOG_MASK_INST, "%010zd " fmt "\n", fetched_instruction_pointer, a1, a2, a3, a4) #else # define LOG_INST_0(hcl,fmt) # define LOG_INST_1(hcl,fmt,a1) # define LOG_INST_2(hcl,fmt,a1,a2) # define LOG_INST_3(hcl,fmt,a1,a2,a3) # define LOG_INST_3(hcl,fmt,a1,a2,a3,a4) #endif static int vm_startup (hcl_t* hcl) { hcl_cb_t* cb; HCL_DEBUG1 (hcl, "VM started up at IP %zd\n", hcl->ip); for (cb = hcl->cblist; cb; cb = cb->next) { if (cb->vm_startup && cb->vm_startup(hcl) <= -1) { for (cb = cb->prev; cb; cb = cb->prev) { if (cb->vm_cleanup) cb->vm_cleanup (hcl); } return -1; } } hcl->vmprim.gettime (hcl, &hcl->exec_start_time); /* raw time. no adjustment */ return 0; } static void vm_cleanup (hcl_t* hcl) { hcl_cb_t* cb; hcl->vmprim.gettime (hcl, &hcl->exec_end_time); /* raw time. no adjustment */ for (cb = hcl->cblist; cb; cb = cb->next) { if (cb->vm_cleanup) cb->vm_cleanup(hcl); } HCL_DEBUG1 (hcl, "VM cleaned up at IP %zd\n", hcl->ip); } static void vm_checkbc (hcl_t* hcl, hcl_oob_t bcode) { hcl_cb_t* cb; for (cb = hcl->cblist; cb; cb = cb->next) { if (cb->vm_checkbc) cb->vm_checkbc(hcl, bcode); } } /* ------------------------------------------------------------------------- */ static HCL_INLINE hcl_oop_t make_context (hcl_t* hcl, hcl_ooi_t ntmprs) { HCL_ASSERT (hcl, ntmprs >= 0); return hcl_allocoopobj(hcl, HCL_BRAND_CONTEXT, HCL_CONTEXT_NAMED_INSTVARS + (hcl_oow_t)ntmprs); } static HCL_INLINE hcl_oop_function_t make_function (hcl_t* hcl, hcl_oow_t lfsize, const hcl_oob_t* bptr, hcl_oow_t blen) { /* the literal frame is placed in the variable part. * the byte code is placed in the trailer space */ return hcl_allocoopobjwithtrailer(hcl, HCL_BRAND_FUNCTION, HCL_FUNCTION_NAMED_INSTVARS + lfsize, bptr, blen); } static HCL_INLINE void fill_function_data (hcl_t* hcl, hcl_oop_function_t func, hcl_ooi_t nargs, hcl_ooi_t ntmprs, hcl_oop_context_t homectx, const hcl_oop_t* lfptr, hcl_oow_t lfsize) { /* Although this function could be integrated into make_function(), * this function has been separated from make_function() to make GC handling simpler */ hcl_oow_t i; HCL_ASSERT (hcl, nargs >= 0 && nargs <= HCL_SMOOI_MAX); HCL_ASSERT (hcl, ntmprs >= 0 && ntmprs <= HCL_SMOOI_MAX); HCL_ASSERT (hcl, nargs <= ntmprs); /* copy literal frames */ HCL_ASSERT (hcl, lfsize <= HCL_OBJ_GET_SIZE(func) - HCL_FUNCTION_NAMED_INSTVARS); for (i = 0; i < lfsize; i++) { func->literal_frame[i] = lfptr[i]; HCL_DEBUG2 (hcl, "literal frame %d => %O\n", (int)i, lfptr[i]); } /* initialize other fields */ func->home = homectx; func->nargs = HCL_SMOOI_TO_OOP(nargs); func->ntmprs = HCL_SMOOI_TO_OOP(ntmprs); } static HCL_INLINE hcl_oop_block_t make_block (hcl_t* hcl) { /* create a base block used for creation of a block context */ return hcl_allocoopobj(hcl, HCL_BRAND_BLOCK, HCL_BLOCK_NAMED_INSTVARS); } static HCL_INLINE void fill_block_data (hcl_t* hcl, hcl_oop_block_t blk, hcl_ooi_t nargs, hcl_ooi_t ntmprs, hcl_ooi_t ip, hcl_oop_context_t homectx) { HCL_ASSERT (hcl, nargs >= 0 && nargs <= HCL_SMOOI_MAX); HCL_ASSERT (hcl, ntmprs >= 0 && ntmprs <= HCL_SMOOI_MAX); HCL_ASSERT (hcl, nargs <= ntmprs); HCL_ASSERT (hcl, ip >= 0 && nargs <= HCL_SMOOI_MAX); blk->home = homectx; blk->nargs = HCL_SMOOI_TO_OOP(nargs); blk->ntmprs = HCL_SMOOI_TO_OOP(ntmprs); blk->ip = HCL_SMOOI_TO_OOP(ip); } static HCL_INLINE int prepare_to_alloc_pid (hcl_t* hcl) { hcl_oow_t new_capa; hcl_ooi_t i, j; hcl_oop_t* tmp; HCL_ASSERT (hcl, hcl->proc_map_free_first <= -1); HCL_ASSERT (hcl, hcl->proc_map_free_last <= -1); new_capa = hcl->proc_map_capa + PROC_MAP_INC; if (new_capa > HCL_SMOOI_MAX) { if (hcl->proc_map_capa >= HCL_SMOOI_MAX) { #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_FATAL, "Processor - too many processes\n"); #endif hcl_seterrbfmt (hcl, HCL_EPFULL, "maximum number(%zd) of processes reached", HCL_SMOOI_MAX); return -1; } new_capa = HCL_SMOOI_MAX; } tmp = (hcl_oop_t*)hcl_reallocmem (hcl, hcl->proc_map, HCL_SIZEOF(hcl_oop_t) * new_capa); if (!tmp) return -1; hcl->proc_map_free_first = hcl->proc_map_capa; for (i = hcl->proc_map_capa, j = hcl->proc_map_capa + 1; j < new_capa; i++, j++) { tmp[i] = HCL_SMOOI_TO_OOP(j); } tmp[i] = HCL_SMOOI_TO_OOP(-1); hcl->proc_map_free_last = i; hcl->proc_map = tmp; hcl->proc_map_capa = new_capa; return 0; } static HCL_INLINE void alloc_pid (hcl_t* hcl, hcl_oop_process_t proc) { hcl_ooi_t pid; pid = hcl->proc_map_free_first; proc->id = HCL_SMOOI_TO_OOP(pid); HCL_ASSERT (hcl, HCL_OOP_IS_SMOOI(hcl->proc_map[pid])); hcl->proc_map_free_first = HCL_OOP_TO_SMOOI(hcl->proc_map[pid]); if (hcl->proc_map_free_first <= -1) hcl->proc_map_free_last = -1; hcl->proc_map[pid] = (hcl_oop_t)proc; } static HCL_INLINE void free_pid (hcl_t* hcl, hcl_oop_process_t proc) { hcl_ooi_t pid; pid = HCL_OOP_TO_SMOOI(proc->id); HCL_ASSERT (hcl, pid < hcl->proc_map_capa); hcl->proc_map[pid] = HCL_SMOOI_TO_OOP(-1); if (hcl->proc_map_free_last <= -1) { HCL_ASSERT (hcl, hcl->proc_map_free_first <= -1); hcl->proc_map_free_first = pid; } else { hcl->proc_map[hcl->proc_map_free_last] = HCL_SMOOI_TO_OOP(pid); } hcl->proc_map_free_last = pid; } static hcl_oop_process_t make_process (hcl_t* hcl, hcl_oop_context_t c) { hcl_oop_process_t proc; hcl_oow_t stksize; if (hcl->proc_map_free_first <= -1 && prepare_to_alloc_pid(hcl) <= -1) return HCL_NULL; stksize = hcl->option.dfl_procstk_size; if (stksize > HCL_TYPE_MAX(hcl_oow_t) - HCL_PROCESS_NAMED_INSTVARS) stksize = HCL_TYPE_MAX(hcl_oow_t) - HCL_PROCESS_NAMED_INSTVARS; hcl_pushtmp (hcl, (hcl_oop_t*)&c); proc = (hcl_oop_process_t)hcl_allocoopobj(hcl, HCL_BRAND_PROCESS, HCL_PROCESS_NAMED_INSTVARS + stksize); hcl_poptmp (hcl); if (HCL_UNLIKELY(!proc)) return HCL_NULL; /* assign a process id to the process */ alloc_pid (hcl, proc); proc->state = HCL_SMOOI_TO_OOP(PROC_STATE_SUSPENDED); proc->initial_context = c; proc->current_context = c; proc->sp = HCL_SMOOI_TO_OOP(-1); HCL_ASSERT (hcl, (hcl_oop_t)c->sender == hcl->_nil); #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG2 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] **CREATED**->%hs\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state))); #endif return proc; } static HCL_INLINE void sleep_active_process (hcl_t* hcl, int state) { #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - put process %O context %O ip=%zd to sleep\n", hcl->processor->active, hcl->active_context, hcl->ip); #endif STORE_ACTIVE_SP(hcl); #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->%hs in sleep_active_process\n", HCL_OOP_TO_SMOOI(hcl->processor->active->id), proc_state_to_string(HCL_OOP_TO_SMOOI(hcl->processor->active->state)), proc_state_to_string(state)); #endif /* store the current active context to the current process. * it is the suspended context of the process to be suspended */ HCL_ASSERT (hcl, hcl->processor->active != hcl->nil_process); hcl->processor->active->current_context = hcl->active_context; hcl->processor->active->state = HCL_SMOOI_TO_OOP(state); } static HCL_INLINE void wake_new_process (hcl_t* hcl, hcl_oop_process_t proc) { #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG2 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->RUNNING in wake_process\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state))); #endif /* activate the given process */ proc->state = HCL_SMOOI_TO_OOP(PROC_STATE_RUNNING); hcl->processor->active = proc; LOAD_ACTIVE_SP(hcl); /* activate the suspended context of the new process */ SWITCH_ACTIVE_CONTEXT (hcl, proc->current_context); #if defined(HCL_DEBUG_VM_PROCESSOR) && (HCL_DEBUG_VM_PROCESSOR >= 2) HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - woke up process[%zd] context %O ip=%zd\n", HCL_OOP_TO_SMOOI(hcl->processor->active->id), hcl->active_context, hcl->ip); #endif } static void switch_to_process (hcl_t* hcl, hcl_oop_process_t proc, int new_state_for_old_active) { /* the new process must not be the currently active process */ HCL_ASSERT (hcl, hcl->processor->active != proc); /* the new process must be in the runnable state */ HCL_ASSERT (hcl, proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNABLE) || proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_WAITING)); sleep_active_process (hcl, new_state_for_old_active); wake_new_process (hcl, proc); hcl->proc_switched = 1; } static HCL_INLINE hcl_oop_process_t find_next_runnable_process (hcl_t* hcl) { hcl_oop_process_t npr; HCL_ASSERT (hcl, hcl->processor->active->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNING)); npr = hcl->processor->active->next; if ((hcl_oop_t)npr == hcl->_nil) npr = hcl->processor->runnable_head; return npr; } static HCL_INLINE void switch_to_next_runnable_process (hcl_t* hcl) { hcl_oop_process_t nrp; nrp = find_next_runnable_process (hcl); if (nrp != hcl->processor->active) switch_to_process (hcl, nrp, PROC_STATE_RUNNABLE); } static HCL_INLINE int chain_into_processor (hcl_t* hcl, hcl_oop_process_t proc, int new_state) { /* the process is not scheduled at all. * link it to the processor's process list. */ hcl_ooi_t tally; HCL_ASSERT (hcl, (hcl_oop_t)proc->prev == hcl->_nil); HCL_ASSERT (hcl, (hcl_oop_t)proc->next == hcl->_nil); HCL_ASSERT (hcl, proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_SUSPENDED)); #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->%hs in chain_into_processor\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state)), proc_state_to_string(new_state)); #endif tally = HCL_OOP_TO_SMOOI(hcl->processor->tally); HCL_ASSERT (hcl, tally >= 0); if (tally >= HCL_SMOOI_MAX) { #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_FATAL, "Processor - too many process\n"); #endif hcl_seterrnum (hcl, HCL_EPFULL); hcl_seterrbfmt (hcl, HCL_EPFULL, "maximum number(%zd) of processes reached", HCL_SMOOI_MAX); return -1; } /* append to the runnable list */ if (tally > 0) { proc->prev = hcl->processor->runnable_tail; hcl->processor->runnable_tail->next = proc; } else { hcl->processor->runnable_head = proc; } hcl->processor->runnable_tail = proc; proc->state = HCL_SMOOI_TO_OOP(new_state); tally++; hcl->processor->tally = HCL_SMOOI_TO_OOP(tally); return 0; } static HCL_INLINE void unchain_from_processor (hcl_t* hcl, hcl_oop_process_t proc, int new_state) { hcl_ooi_t tally; /* the processor's process chain must be composed of running/runnable * processes only */ HCL_ASSERT (hcl, proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNING) || proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNABLE)); tally = HCL_OOP_TO_SMOOI(hcl->processor->tally); HCL_ASSERT (hcl, tally > 0); if ((hcl_oop_t)proc->prev != hcl->_nil) proc->prev->next = proc->next; else hcl->processor->runnable_head = proc->next; if ((hcl_oop_t)proc->next != hcl->_nil) proc->next->prev = proc->prev; else hcl->processor->runnable_tail = proc->prev; #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->%hs in unchain_from_processor\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state)), proc_state_to_string(HCL_OOP_TO_SMOOI(new_state))); #endif proc->prev = (hcl_oop_process_t)hcl->_nil; proc->next = (hcl_oop_process_t)hcl->_nil; proc->state = HCL_SMOOI_TO_OOP(new_state); tally--; if (tally == 0) hcl->processor->active = hcl->nil_process; hcl->processor->tally = HCL_SMOOI_TO_OOP(tally); } static HCL_INLINE void chain_into_semaphore (hcl_t* hcl, hcl_oop_process_t proc, hcl_oop_semaphore_t sem) { /* append a process to the process list of a semaphore*/ HCL_ASSERT (hcl, (hcl_oop_t)proc->sem == hcl->_nil); HCL_ASSERT (hcl, (hcl_oop_t)proc->prev == hcl->_nil); HCL_ASSERT (hcl, (hcl_oop_t)proc->next == hcl->_nil); if ((hcl_oop_t)sem->waiting_head == hcl->_nil) { HCL_ASSERT (hcl, (hcl_oop_t)sem->waiting_tail == hcl->_nil); sem->waiting_head = proc; } else { proc->prev = sem->waiting_tail; sem->waiting_tail->next = proc; } sem->waiting_tail = proc; proc->sem = sem; } static HCL_INLINE void unchain_from_semaphore (hcl_t* hcl, hcl_oop_process_t proc) { hcl_oop_semaphore_t sem; HCL_ASSERT (hcl, (hcl_oop_t)proc->sem != hcl->_nil); sem = proc->sem; if ((hcl_oop_t)proc->prev != hcl->_nil) proc->prev->next = proc->next; else sem->waiting_head = proc->next; if ((hcl_oop_t)proc->next != hcl->_nil) proc->next->prev = proc->prev; else sem->waiting_tail = proc->prev; proc->prev = (hcl_oop_process_t)hcl->_nil; proc->next = (hcl_oop_process_t)hcl->_nil; proc->sem = (hcl_oop_semaphore_t)hcl->_nil; } static void terminate_process (hcl_t* hcl, hcl_oop_process_t proc) { if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNING) || proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNABLE)) { /* RUNNING/RUNNABLE ---> TERMINATED */ #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG2 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->TERMINATED in terminate_process\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state))); #endif if (proc == hcl->processor->active) { hcl_oop_process_t nrp; nrp = find_next_runnable_process (hcl); unchain_from_processor (hcl, proc, PROC_STATE_TERMINATED); proc->sp = HCL_SMOOI_TO_OOP(-1); /* invalidate the process stack */ proc->current_context = proc->initial_context; /* not needed but just in case */ /* a runnable or running process must not be chanined to the * process list of a semaphore */ HCL_ASSERT (hcl, (hcl_oop_t)proc->sem == hcl->_nil); if (nrp == proc) { /* no runnable process after termination */ HCL_ASSERT (hcl, hcl->processor->active == hcl->nil_process); HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "No runnable process after process termination\n"); } else { switch_to_process (hcl, nrp, PROC_STATE_TERMINATED); } } else { unchain_from_processor (hcl, proc, PROC_STATE_TERMINATED); proc->sp = HCL_SMOOI_TO_OOP(-1); /* invalidate the process stack */ } /* when terminated, clear it from the pid table and set the process id to a negative number */ free_pid (hcl, proc); } else if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_SUSPENDED)) { /* SUSPENDED ---> TERMINATED */ #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG2 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->TERMINATED in terminate_process\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state))); #endif proc->state = HCL_SMOOI_TO_OOP(PROC_STATE_TERMINATED); proc->sp = HCL_SMOOI_TO_OOP(-1); /* invalidate the proce stack */ if ((hcl_oop_t)proc->sem != hcl->_nil) { unchain_from_semaphore (hcl, proc); } /* when terminated, clear it from the pid table and set the process id to a negative number */ free_pid (hcl, proc); } else if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_WAITING)) { /* WAITING ---> TERMINATED */ /* TODO: */ } } static void resume_process (hcl_t* hcl, hcl_oop_process_t proc) { if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_SUSPENDED)) { /* SUSPENED ---> RUNNING */ HCL_ASSERT (hcl, (hcl_oop_t)proc->prev == hcl->_nil); HCL_ASSERT (hcl, (hcl_oop_t)proc->next == hcl->_nil); #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG2 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process[%zd] %hs->RUNNABLE in resume_process\n", HCL_OOP_TO_SMOOI(proc->id), proc_state_to_string(HCL_OOP_TO_SMOOI(proc->state))); #endif chain_into_processor (hcl, proc, PROC_STATE_RUNNABLE); /* TODO: error check */ /*proc->current_context = proc->initial_context;*/ /* don't switch to this process. just set the state to RUNNING */ } #if 0 else if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNABLE)) { /* RUNNABLE ---> RUNNING */ /* TODO: should i allow this? */ HCL_ASSERT (hcl, hcl->processor->active != proc); switch_to_process (hcl, proc, PROC_STATE_RUNNABLE); } #endif } static void suspend_process (hcl_t* hcl, hcl_oop_process_t proc) { if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNING) || proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNABLE)) { /* RUNNING/RUNNABLE ---> SUSPENDED */ #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG1 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process %O RUNNING/RUNNABLE->SUSPENDED\n", proc); #endif if (proc == hcl->processor->active) { hcl_oop_process_t nrp; nrp = find_next_runnable_process (hcl); if (nrp == proc) { /* no runnable process after suspension */ sleep_active_process (hcl, PROC_STATE_RUNNABLE); unchain_from_processor (hcl, proc, PROC_STATE_SUSPENDED); /* the last running/runnable process has been unchained * from the processor and set to SUSPENDED. the active * process must be the nil process */ HCL_ASSERT (hcl, hcl->processor->active == hcl->nil_process); } else { /* keep the unchained process at the runnable state for * the immediate call to switch_to_process() below */ unchain_from_processor (hcl, proc, PROC_STATE_RUNNABLE); /* unchain_from_processor() leaves the active process * untouched unless the unchained process is the last * running/runnable process. so calling switch_to_process() * which expects the active process to be valid is safe */ HCL_ASSERT (hcl, hcl->processor->active != hcl->nil_process); switch_to_process (hcl, nrp, PROC_STATE_SUSPENDED); } } else { unchain_from_processor (hcl, proc, PROC_STATE_SUSPENDED); } } } static void yield_process (hcl_t* hcl, hcl_oop_process_t proc) { if (proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNING)) { /* RUNNING --> RUNNABLE */ hcl_oop_process_t nrp; HCL_ASSERT (hcl, proc == hcl->processor->active); nrp = find_next_runnable_process (hcl); /* if there are more than 1 runnable processes, the next * runnable process must be different from proc */ if (nrp != proc) { #if defined(HCL_DEBUG_VM_PROCESSOR) HCL_LOG1 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "Processor - process %O RUNNING->RUNNABLE\n", proc); #endif switch_to_process (hcl, nrp, PROC_STATE_RUNNABLE); } } } static int async_signal_semaphore (hcl_t* hcl, hcl_oop_semaphore_t sem) { #if 0 if (hcl->sem_list_count >= SEM_LIST_MAX) { hcl_seterrnum (hcl, HCL_ESLFULL); return -1; } if (hcl->sem_list_count >= hcl->sem_list_capa) { hcl_oow_t new_capa; hcl_oop_semaphore_t* tmp; new_capa = hcl->sem_list_capa + SEM_LIST_INC; /* TODO: overflow check.. */ tmp = hcl_reallocmem (hcl, hcl->sem_list, HCL_SIZEOF(hcl_oop_semaphore_t) * new_capa); if (!tmp) return -1; hcl->sem_list = tmp; hcl->sem_list_capa = new_capa; } hcl->sem_list[hcl->sem_list_count] = sem; hcl->sem_list_count++; #endif return 0; } static hcl_oop_process_t signal_semaphore (hcl_t* hcl, hcl_oop_semaphore_t sem) { hcl_oop_process_t proc; hcl_ooi_t count; if ((hcl_oop_t)sem->waiting_head == hcl->_nil) { /* no process is waiting on this semaphore */ count = HCL_OOP_TO_SMOOI(sem->count); count++; sem->count = HCL_SMOOI_TO_OOP(count); /* no process has been resumed */ return (hcl_oop_process_t)hcl->_nil; } else { proc = sem->waiting_head; /* [NOTE] no GC must occur as 'proc' isn't protected with hcl_pushtmp(). */ unchain_from_semaphore (hcl, proc); resume_process (hcl, proc); /* TODO: error check */ /* return the resumed process */ return proc; } } static void await_semaphore (hcl_t* hcl, hcl_oop_semaphore_t sem) { hcl_oop_process_t proc; hcl_ooi_t count; count = HCL_OOP_TO_SMOOI(sem->count); if (count > 0) { /* it's already signalled */ count--; sem->count = HCL_SMOOI_TO_OOP(count); } else { /* not signaled. need to wait */ proc = hcl->processor->active; /* suspend the active process */ suspend_process (hcl, proc); /* link the suspended process to the semaphore's process list */ chain_into_semaphore (hcl, proc, sem); HCL_ASSERT (hcl, sem->waiting_tail == proc); HCL_ASSERT (hcl, hcl->processor->active != proc); } } static void sift_up_sem_heap (hcl_t* hcl, hcl_ooi_t index) { if (index > 0) { hcl_ooi_t parent; hcl_oop_semaphore_t sem, parsem; parent = SEM_HEAP_PARENT(index); sem = hcl->sem_heap[index]; parsem = hcl->sem_heap[parent]; if (SEM_HEAP_EARLIER_THAN(hcl, sem, parsem)) { do { /* move down the parent to the current position */ parsem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap[index] = parsem; /* traverse up */ index = parent; if (index <= 0) break; parent = SEM_HEAP_PARENT(parent); parsem = hcl->sem_heap[parent]; } while (SEM_HEAP_EARLIER_THAN(hcl, sem, parsem)); sem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap[index] = sem; } } } static void sift_down_sem_heap (hcl_t* hcl, hcl_ooi_t index) { hcl_ooi_t base = hcl->sem_heap_count / 2; if (index < base) /* at least 1 child is under the 'index' position */ { hcl_ooi_t left, right, child; hcl_oop_semaphore_t sem, chisem; sem = hcl->sem_heap[index]; do { left = SEM_HEAP_LEFT(index); right = SEM_HEAP_RIGHT(index); if (right < hcl->sem_heap_count && SEM_HEAP_EARLIER_THAN(hcl, hcl->sem_heap[left], hcl->sem_heap[right])) { child = right; } else { child = left; } chisem = hcl->sem_heap[child]; if (SEM_HEAP_EARLIER_THAN(hcl, sem, chisem)) break; chisem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap[index ] = chisem; index = child; } while (index < base); sem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap[index] = sem; } } static int add_to_sem_heap (hcl_t* hcl, hcl_oop_semaphore_t sem) { hcl_ooi_t index; #if 0 if (hcl->sem_heap_count >= SEM_HEAP_MAX) { hcl_seterrnum (hcl, HCL_ESHFULL); return -1; } if (hcl->sem_heap_count >= hcl->sem_heap_capa) { hcl_oow_t new_capa; hcl_oop_semaphore_t* tmp; /* no overflow check when calculating the new capacity * owing to SEM_HEAP_MAX check above */ new_capa = hcl->sem_heap_capa + SEM_HEAP_INC; tmp = hcl_reallocmem (hcl, hcl->sem_heap, HCL_SIZEOF(hcl_oop_semaphore_t) * new_capa); if (!tmp) return -1; hcl->sem_heap = tmp; hcl->sem_heap_capa = new_capa; } HCL_ASSERT (hcl, hcl->sem_heap_count <= HCL_SMOOI_MAX); index = hcl->sem_heap_count; hcl->sem_heap[index] = sem; sem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap_count++; sift_up_sem_heap (hcl, index); #endif return 0; } static void delete_from_sem_heap (hcl_t* hcl, hcl_ooi_t index) { hcl_oop_semaphore_t sem, lastsem; sem = hcl->sem_heap[index]; sem->heap_index = HCL_SMOOI_TO_OOP(-1); hcl->sem_heap_count--; if (hcl->sem_heap_count > 0 && index != hcl->sem_heap_count) { /* move the last item to the deletion position */ lastsem = hcl->sem_heap[hcl->sem_heap_count]; lastsem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap[index] = lastsem; if (SEM_HEAP_EARLIER_THAN(hcl, lastsem, sem)) sift_up_sem_heap (hcl, index); else sift_down_sem_heap (hcl, index); } } static void update_sem_heap (hcl_t* hcl, hcl_ooi_t index, hcl_oop_semaphore_t newsem) { hcl_oop_semaphore_t sem; sem = hcl->sem_heap[index]; sem->heap_index = HCL_SMOOI_TO_OOP(-1); newsem->heap_index = HCL_SMOOI_TO_OOP(index); hcl->sem_heap[index] = newsem; if (SEM_HEAP_EARLIER_THAN(hcl, newsem, sem)) sift_up_sem_heap (hcl, index); else sift_down_sem_heap (hcl, index); } /* ------------------------------------------------------------------------- */ static int __activate_block (hcl_t* hcl, hcl_oop_block_t rcv_blk, hcl_ooi_t nargs, hcl_oop_context_t* pnewctx) { /* prepare a new block context for activation. * the receiver must be a block context which becomes the base * for a new block context. */ hcl_oop_context_t blkctx; hcl_ooi_t local_ntmprs, i; /* TODO: find a better way to support a reentrant block context. */ /* | sum | * sum := [ :n | (n < 2) ifTrue: [1] ifFalse: [ n + (sum value: (n - 1))] ]. * (sum value: 10). * * For the code above, sum is a block context and it is sent value: inside * itself. Let me simply clone a block context to allow reentrancy like this * while the block context is active */ /* the receiver must be a block context */ HCL_ASSERT (hcl, HCL_IS_BLOCK(hcl, rcv_blk)); if (HCL_OOP_TO_SMOOI(rcv_blk->nargs) != nargs) { HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_ERROR, "Error - wrong number of arguments to a block %O - expecting %zd, got %zd\n", rcv_blk, HCL_OOP_TO_SMOOI(rcv_blk->nargs), nargs); hcl_seterrnum (hcl, HCL_ECALLARG); return -1; } local_ntmprs = HCL_OOP_TO_SMOOI(rcv_blk->ntmprs); HCL_ASSERT (hcl, local_ntmprs >= nargs); /* create a new block context to clone rcv_blk */ hcl_pushtmp (hcl, (hcl_oop_t*)&rcv_blk); blkctx = (hcl_oop_context_t)make_context(hcl, local_ntmprs); hcl_poptmp (hcl); if (!blkctx) return -1; #if 0 /* shallow-copy the named part including home, origin, etc. */ for (i = 0; i < HCL_CONTEXT_NAMED_INSTVARS; i++) { ((hcl_oop_oop_t)blkctx)->slot[i] = ((hcl_oop_oop_t)rcv_blk)->slot[i]; } #else blkctx->ip = rcv_blk->ip; blkctx->ntmprs = rcv_blk->ntmprs; blkctx->nargs = rcv_blk->nargs; blkctx->receiver_or_base = (hcl_oop_t)rcv_blk; blkctx->home = rcv_blk->home; /* blkctx->origin = rcv_blk->origin; */ blkctx->origin = rcv_blk->home->origin; #endif /* TODO: check the stack size of a block context to see if it's large enough to hold arguments */ /* copy the arguments to the stack */ for (i = 0; i < nargs; i++) { blkctx->slot[i] = HCL_STACK_GETARG(hcl, nargs, i); } HCL_STACK_POPS (hcl, nargs + 1); /* pop arguments and receiver */ HCL_ASSERT (hcl, (rcv_blk == hcl->initial_context && (hcl_oop_t)blkctx->home == hcl->_nil) || (hcl_oop_t)blkctx->home != hcl->_nil); blkctx->sp = HCL_SMOOI_TO_OOP(-1); /* not important at all */ blkctx->sender = hcl->active_context; *pnewctx = blkctx; return 0; } static HCL_INLINE int activate_block (hcl_t* hcl, hcl_ooi_t nargs) { int x; hcl_oop_block_t rcv; hcl_oop_context_t newctx; rcv = (hcl_oop_block_t)HCL_STACK_GETRCV(hcl, nargs); HCL_ASSERT (hcl, HCL_IS_BLOCK(hcl, rcv)); x = __activate_block(hcl, rcv, nargs, &newctx); if (HCL_UNLIKELY(x <= -1)) return -1; SWITCH_ACTIVE_CONTEXT (hcl, newctx); return 0; } /* ------------------------------------------------------------------------- */ static int __activate_function (hcl_t* hcl, hcl_oop_function_t rcv_func, hcl_ooi_t nargs, hcl_oop_context_t* pnewctx) { /* prepare a new block context for activation. * the receiver must be a block context which becomes the base * for a new block context. */ hcl_oop_context_t functx; hcl_ooi_t local_ntmprs, i; /* * (defun sum (x) * (if (< x 2) 1 * else (+ x (sum (- x 1))))) * (printf ">>>> %d\n" (sum 10)) */ /* the receiver must be a function */ HCL_ASSERT (hcl, HCL_IS_FUNCTION(hcl, rcv_func)); if (HCL_OOP_TO_SMOOI(rcv_func->nargs) != nargs) { HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_ERROR, "Error - wrong number of arguments to a function %O - expecting %zd, got %zd\n", rcv_func, HCL_OOP_TO_SMOOI(rcv_func->nargs), nargs); hcl_seterrnum (hcl, HCL_ECALLARG); return -1; } local_ntmprs = HCL_OOP_TO_SMOOI(rcv_func->ntmprs); HCL_ASSERT (hcl, local_ntmprs >= nargs); /* create a new block context to clone rcv_func */ hcl_pushtmp (hcl, (hcl_oop_t*)&rcv_func); functx = (hcl_oop_context_t)make_context(hcl, local_ntmprs); hcl_poptmp (hcl); if (!functx) return -1; functx->ip = HCL_SMOOI_TO_OOP(0); functx->ntmprs = rcv_func->ntmprs; functx->nargs = rcv_func->nargs; functx->receiver_or_base = (hcl_oop_t)rcv_func; functx->home = rcv_func->home; functx->origin = functx; /* the origin of the context over a function should be itself */ /* TODO: check the stack size of a block context to see if it's large enough to hold arguments */ /* copy the arguments to the stack */ for (i = 0; i < nargs; i++) { functx->slot[i] = HCL_STACK_GETARG(hcl, nargs, i); } HCL_STACK_POPS (hcl, nargs + 1); /* pop arguments and receiver */ HCL_ASSERT (hcl, (hcl_oop_t)functx->home != hcl->_nil); functx->sp = HCL_SMOOI_TO_OOP(-1); /* not important at all */ functx->sender = hcl->active_context; *pnewctx = functx; return 0; } static HCL_INLINE int activate_function (hcl_t* hcl, hcl_ooi_t nargs) { int x; hcl_oop_function_t rcv; hcl_oop_context_t newctx; rcv = (hcl_oop_function_t)HCL_STACK_GETRCV(hcl, nargs); HCL_ASSERT (hcl, HCL_IS_FUNCTION(hcl, rcv)); x = __activate_function(hcl, rcv, nargs, &newctx); if (HCL_UNLIKELY(x <= -1)) return -1; SWITCH_ACTIVE_CONTEXT (hcl, newctx); return 0; } /* ------------------------------------------------------------------------- */ static HCL_INLINE int call_primitive (hcl_t* hcl, hcl_ooi_t nargs) { hcl_oop_word_t rcv; rcv = (hcl_oop_word_t)HCL_STACK_GETRCV(hcl, nargs); HCL_ASSERT (hcl, HCL_IS_PRIM(hcl, rcv)); HCL_ASSERT (hcl, HCL_OBJ_GET_SIZE(rcv) == 4); if (nargs < rcv->slot[1] && nargs > rcv->slot[2]) { /* TODO: include a primitive name... */ HCL_LOG3 (hcl, HCL_LOG_IC | HCL_LOG_ERROR, "Error - wrong number of arguments to a primitive - expecting %zd-%zd, got %zd\n", rcv->slot[1], rcv->slot[2], nargs); hcl_seterrnum (hcl, HCL_ECALLARG); return -1; } return ((hcl_pfimpl_t)rcv->slot[0]) (hcl, (hcl_mod_t*)rcv->slot[3], nargs); } #if 0 /* EXPERIMENTAL CODE INTEGRATING EXTERNAL COMMANDS */ #include #include #include #include #include #include #include #include extern char **environ; #define _PATH_DEFPATH "/usr/bin:/bin" static int is_regular_executable_file_by_me(const char *path) { struct stat st; if (stat(path, &st) == -1) return 0; return S_ISREG(st.st_mode) && access(path, X_OK) == 0; //? use eaccess instead?? } static char* find_exec (hcl_t* hcl, const char *name) { size_t lp, ln; char buf[PATH_MAX]; const char *bp, *path, *p; bp = buf; /* Get the path we're searching. */ if (!(path = getenv("PATH"))) path = _PATH_DEFPATH; ln = strlen(name); do { /* Find the end of this path element. */ for (p = path; *path != 0 && *path != ':'; path++) ; /* * It's a SHELL path -- double, leading and trailing colons * mean the current directory. */ if (p == path) { p = "."; lp = 1; } else { lp = path - p; } /* * If the path is too long complain. This is a possible * security issue; given a way to make the path too long * the user may execute the wrong program. */ if (lp + ln + 2 > sizeof(buf)) continue; memcpy(buf, p, lp); buf[lp] = '/'; memcpy(buf + lp + 1, name, ln); buf[lp + ln + 1] = '\0'; if (is_regular_executable_file_by_me(bp)) return strdup(bp); } while (*path++ == ':'); /* Otherwise, *path was NUL */ done: hcl_seterrbfmt (hcl, HCL_ENOENT, "callable %hs not found", name); return HCL_NULL; } static HCL_INLINE int exec_syscmd (hcl_t* hcl, hcl_ooi_t nargs) { hcl_oop_word_t rcv; hcl_bch_t* cmd = HCL_NULL; hcl_bch_t* xcmd = HCL_NULL; rcv = (hcl_oop_word_t)HCL_STACK_GETRCV(hcl, nargs); /*HCL_ASSERT (hcl, HCL_IS_STRING(hcl, rcv) || HCL_IS_SYMBOL(hcl, rcv));*/ HCL_ASSERT (hcl, HCL_OBJ_IS_CHAR_POINTER(rcv)); if (HCL_OBJ_GET_SIZE(rcv) == 0 || hcl_count_oocstr(HCL_OBJ_GET_CHAR_SLOT(rcv)) != HCL_OBJ_GET_SIZE(rcv)) { /* '\0' is contained in the middle */ hcl_seterrbfmt (hcl, HCL_EINVAL, "invalid callable %O", rcv); goto oops; } cmd = hcl_dupootobcstr(hcl, HCL_OBJ_GET_CHAR_SLOT(rcv), HCL_NULL); if (!cmd) goto oops; if (hcl_find_bchar_in_bcstr(cmd, '/')) { if (!is_regular_executable_file_by_me(cmd)) { hcl_seterrbfmt (hcl, HCL_ECALL, "cannot execute %O", rcv); goto oops; } xcmd = cmd; } else { xcmd = find_exec(hcl, cmd); if (!xcmd) goto oops; } { /* TODO: make it a callback ... */ pid_t pid; int status; pid = fork(); if (pid == -1) goto oops; /* TODO: set a new process group / session leader??? */ if (pid == 0) { hcl_bch_t** argv; hcl_ooi_t i; /* TODO: close file descriptors??? */ argv = (hcl_bch_t**)hcl_allocmem(hcl, (nargs + 2) * HCL_SIZEOF(*argv)); if (argv) { argv[0] = cmd; HCL_DEBUG1 (hcl, "NARG %d\n", (int)nargs); for (i = 0; i < nargs;) { hcl_oop_t ta = HCL_STACK_GETARG(hcl, nargs, i); /* TODO: check if an argument is a string or a symbol */ if (HCL_OOP_IS_SMOOI(ta)) { /* TODO: rewrite this part */ hcl_bch_t tmp[64]; snprintf (tmp, sizeof(tmp), "%ld", (long int)HCL_OOP_TO_SMOOI(ta)); argv[++i] = hcl_dupbchars(hcl, tmp, strlen(tmp)); } else { argv[++i] = hcl_dupootobchars(hcl, HCL_OBJ_GET_CHAR_SLOT(ta), HCL_OBJ_GET_SIZE(ta), HCL_NULL); } HCL_DEBUG2 (hcl, "ARG %d -> %hs\n", (int)i - 1, argv[i]); } argv[nargs + 1] = HCL_NULL; execvp (xcmd, argv); } if (cmd) hcl_freemem (hcl, cmd); if (xcmd && xcmd != cmd) hcl_freemem (hcl, xcmd); _exit (255); } waitpid (pid, &status, 0); /* TOOD: enhance this waiting */ HCL_STACK_SETRET (hcl, nargs, HCL_SMOOI_TO_OOP(WEXITSTATUS(status))); } hcl_freemem (hcl, cmd); if (xcmd != cmd) hcl_freemem (hcl, xcmd); return 0; oops: if (cmd) hcl_freemem (hcl, cmd); if (xcmd && xcmd != cmd) hcl_freemem (hcl, xcmd); return -1; } #endif /* ------------------------------------------------------------------------- */ static hcl_oop_process_t start_initial_process (hcl_t* hcl, hcl_oop_context_t ctx) { hcl_oop_process_t proc; /* there must be no active process when this function is called */ HCL_ASSERT (hcl, hcl->processor->tally == HCL_SMOOI_TO_OOP(0)); HCL_ASSERT (hcl, hcl->processor->active == hcl->nil_process); proc = make_process(hcl, ctx); if (HCL_UNLIKELY(!proc)) return HCL_NULL; /* skip RUNNABLE and go to RUNNING */ if (chain_into_processor(hcl, proc, PROC_STATE_RUNNING) <= -1) return HCL_NULL; hcl->processor->active = proc; /* do something that resume_process() would do with less overhead */ HCL_ASSERT (hcl, (hcl_oop_t)proc->current_context != hcl->_nil); HCL_ASSERT (hcl, proc->current_context == proc->initial_context); SWITCH_ACTIVE_CONTEXT (hcl, proc->current_context); return proc; } static int start_initial_process_and_context (hcl_t* hcl, hcl_ooi_t initial_ip) { hcl_oop_context_t ctx; hcl_oop_process_t proc; /* create the initial context over the initial function */ ctx = (hcl_oop_context_t)make_context(hcl, 0); /* no temporary variables */ if (!ctx) return -1; hcl->ip = initial_ip; hcl->sp = -1; ctx->ip = HCL_SMOOI_TO_OOP(initial_ip); ctx->sp = HCL_SMOOI_TO_OOP(-1); /* pointer to -1 below the bottom */ ctx->nargs = HCL_SMOOI_TO_OOP(0); ctx->ntmprs = HCL_SMOOI_TO_OOP(0); ctx->origin = ctx; /* the origin of the initial context is itself as this is created over the initial function */ ctx->home = hcl->initial_function->home; /* this should be nil */ ctx->sender = (hcl_oop_context_t)hcl->_nil; ctx->receiver_or_base = hcl->initial_function; HCL_ASSERT (hcl, (hcl_oop_t)ctx->home == hcl->_nil); /* [NOTE] * the sender field of the initial context is nil. * especially, the fact that the sender field is nil is used by * the main execution loop for breaking out of the loop */ HCL_ASSERT (hcl, hcl->active_context == HCL_NULL); /* hcl_gc() uses hcl->processor when hcl->active_context * is not NULL. at this poinst, hcl->processor should point to * an instance of ProcessScheduler. */ HCL_ASSERT (hcl, (hcl_oop_t)hcl->processor != hcl->_nil); HCL_ASSERT (hcl, hcl->processor->tally == HCL_SMOOI_TO_OOP(0)); /* start_initial_process() calls the SWITCH_ACTIVE_CONTEXT() macro. * the macro assumes a non-null value in hcl->active_context. * let's force set active_context to ctx directly. */ hcl->active_context = ctx; hcl_pushtmp (hcl, (hcl_oop_t*)&ctx); proc = start_initial_process(hcl, ctx); hcl_poptmp (hcl); if (HCL_UNLIKELY(!proc)) return -1; HCL_STACK_PUSH (hcl, (hcl_oop_t)ctx); STORE_ACTIVE_SP (hcl); /* hcl->active_context->sp = HCL_SMOOI_TO_OOP(hcl->sp) */ HCL_ASSERT (hcl, proc == hcl->processor->active); hcl->initial_context = proc->initial_context; HCL_ASSERT (hcl, hcl->initial_context == hcl->active_context); return 0; } /* ------------------------------------------------------------------------- */ static HCL_INLINE int do_return (hcl_t* hcl, hcl_oop_t return_value) { #if 0 (defun x(a) (defun y(k) (return-from-home (+ k k k))) (+ a a a) (y 40) ; this should act like (return (y 40)) ) (x 2) (y 10) ; this should return from x but x it should end up with dead return... #endif hcl_oop_context_t ctx; // TODO: home could be null... HCL_DEBUG4 (hcl, "do_return >>>>>>>>>> %d active_context %p active_context->home %p home->ip %O\n", HCL_OOP_TO_SMOOI(hcl->active_context->home->ip), hcl->active_context, hcl->active_context->home, hcl->active_context->home->ip); if (hcl->active_context->home->ip == HCL_SMOOI_TO_OOP(-1)) { HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_ERROR, "Error - cannot return from dead context\n"); hcl_seterrbfmt (hcl, HCL_EINTERN, "unable to return from dead context"); /* TODO: can i make this error catchable at the hcl level? */ return -1; } #if 0 ctx = hcl->active_context; while ((hcl_oop_t)ctx != hcl->_nil) { if (ctx == hcl->active_context->origin) goto non_local_return_ok; ctx = ctx->sender; } HCL_ASSERT (hcl, hcl->active_context->origin->ip == HCL_SMOOI_TO_OOP(-1)); HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_ERROR, "Error - cannot return from dead context\n"); hcl_seterrbfmt (hcl, HCL_EINTERN, "unable to return from dead context"); /* TODO: can i make this error catchable at the hcl level? */ return -1; non_local_return_ok: #endif HCL_DEBUG1 (hcl, "NON LOCAL RETURN XXXXXXXXXXXXXXXXXXXXXXXXXXX from active context %p\n", hcl->active_context); if (hcl->active_context == hcl->processor->active->initial_context) { hcl->active_context->ip = HCL_SMOOI_TO_OOP(-1); /* mark that this context has returned */ hcl->active_context->home->ip = HCL_SMOOI_TO_OOP(-1); /* mark that this context has returned */ terminate_process (hcl, hcl->processor->active); } else { #if 0 //hcl->active_context->origin->ip = HCL_SMOOI_TO_OOP(-1); /* restore the stack pointer */ //hcl->sp = HCL_OOP_TO_SMOOI(hcl->active_context->origin->sp); //SWITCH_ACTIVE_CONTEXT (hcl, hcl->active_context->origin->sender); #else hcl->ip = -1; hcl->active_context->home->ip = HCL_SMOOI_TO_OOP(-1); /* mark that this context has returned */ SWITCH_ACTIVE_CONTEXT (hcl, hcl->active_context->home->sender); #endif /* push the return value to the stack of the new active context */ HCL_STACK_PUSH (hcl, return_value); } return 0; } /* ------------------------------------------------------------------------- */ static int execute (hcl_t* hcl) { hcl_oob_t bcode; hcl_oow_t b1, b2; hcl_oop_t return_value; #if defined(HCL_PROFILE_VM) hcl_uintmax_t inst_counter = 0; #endif #if defined(HCL_DEBUG_VM_EXEC) hcl_ooi_t fetched_instruction_pointer; #endif HCL_ASSERT (hcl, hcl->active_context != HCL_NULL); hcl->abort_req = 0; if (vm_startup(hcl) <= -1) return -1; hcl->proc_switched = 0; while (1) { #if defined(ENABLE_MULTI_PROCS) /* i don't think i will ever implement this in HCL. * but let's keep the code here for some while */ if (hcl->sem_heap_count > 0) { hcl_ntime_t ft, now; hcl->vmprim.gettime (hcl, &now); do { HCL_ASSERT (hcl, HCL_OOP_IS_SMOOI(hcl->sem_heap[0]->heap_ftime_sec)); HCL_ASSERT (hcl, HCL_OOP_IS_SMOOI(hcl->sem_heap[0]->heap_ftime_nsec)); HCL_INIT_NTIME (&ft, HCL_OOP_TO_SMOOI(hcl->sem_heap[0]->heap_ftime_sec), HCL_OOP_TO_SMOOI(hcl->sem_heap[0]->heap_ftime_nsec) ); if (HCL_CMP_NTIME(&ft, (hcl_ntime_t*)&now) <= 0) { hcl_oop_process_t proc; /* waited long enough. signal the semaphore */ proc = signal_semaphore (hcl, hcl->sem_heap[0]); /* [NOTE] no hcl_pushtmp() on proc. no GC must occur * in the following line until it's used for * wake_new_process() below. */ delete_from_sem_heap (hcl, 0); /* if no process is waiting on the semaphore, * signal_semaphore() returns hcl->_nil. */ if (hcl->processor->active == hcl->nil_process && (hcl_oop_t)proc != hcl->_nil) { /* this is the only runnable process. * switch the process to the running state. * it uses wake_new_process() instead of * switch_to_process() as there is no running * process at this moment */ HCL_ASSERT (hcl, proc->state == HCL_SMOOI_TO_OOP(PROC_STATE_RUNNABLE)); HCL_ASSERT (hcl, proc == hcl->processor->runnable_head); wake_new_process (hcl, proc); hcl->proc_switched = 1; } } else if (hcl->processor->active == hcl->nil_process) { HCL_SUB_NTIME (&ft, &ft, (hcl_ntime_t*)&now); hcl->vmprim.sleep (hcl, &ft); /* TODO: change this to i/o multiplexer??? */ hcl->vmprim.gettime (hcl, &now); } else { break; } } while (hcl->sem_heap_count > 0); } #endif if (hcl->processor->active == hcl->nil_process) { /* no more waiting semaphore and no more process */ HCL_ASSERT (hcl, hcl->processor->tally = HCL_SMOOI_TO_OOP(0)); HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_DEBUG, "No more runnable process\n"); #if 0 if (there is semaphore awaited.... ) { /* DO SOMETHING */ } #endif break; } #if defined(ENABLE_MULTI_PROCS) while (hcl->sem_list_count > 0) { /* handle async signals */ --hcl->sem_list_count; signal_semaphore (hcl, hcl->sem_list[hcl->sem_list_count]); } /* TODO: implement different process switching scheme - time-slice or clock based??? */ #if defined(HCL_EXTERNAL_PROCESS_SWITCH) if (!hcl->proc_switched && hcl->switch_proc) { switch_to_next_runnable_process (hcl); } hcl->switch_proc = 0; #else if (!hcl->proc_switched) { switch_to_next_runnable_process (hcl); } #endif hcl->proc_switched = 0; #endif if (HCL_UNLIKELY(hcl->ip >= HCL_FUNCTION_GET_CODE_SIZE(hcl->active_function))) { HCL_DEBUG1 (hcl, "Stopping execution as IP reached the end of bytecode(%zu)\n", hcl->code.bc.len); return_value = hcl->_nil; goto handle_return; } #if defined(HCL_DEBUG_VM_EXEC) fetched_instruction_pointer = hcl->ip; #endif FETCH_BYTE_CODE_TO (hcl, bcode); /*while (bcode == HCL_CODE_NOOP) FETCH_BYTE_CODE_TO (hcl, bcode);*/ if (hcl->vm_checkbc_cb_count) vm_checkbc (hcl, bcode); if (HCL_UNLIKELY(hcl->abort_req)) { /* i place this abortion check after vm_checkbc() * to honor hcl_abort() if called in the callback, */ HCL_DEBUG0 (hcl, "Stopping execution for abortion request\n"); return_value = hcl->_nil; goto handle_return; } #if defined(HCL_PROFILE_VM) inst_counter++; #endif switch (bcode) { /* ------------------------------------------------- */ #if 0 case HCL_CODE_PUSH_INSTVAR_X: FETCH_PARAM_CODE_TO (hcl, b1); goto push_instvar; case HCL_CODE_PUSH_INSTVAR_0: case HCL_CODE_PUSH_INSTVAR_1: case HCL_CODE_PUSH_INSTVAR_2: case HCL_CODE_PUSH_INSTVAR_3: case HCL_CODE_PUSH_INSTVAR_4: case HCL_CODE_PUSH_INSTVAR_5: case HCL_CODE_PUSH_INSTVAR_6: case HCL_CODE_PUSH_INSTVAR_7: b1 = bcode & 0x7; /* low 3 bits */ push_instvar: LOG_INST_1 (hcl, "push_instvar %zu", b1); HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_TYPE(hcl->active_context->origin->receiver_or_base) == HCL_OBJ_TYPE_OOP); HCL_STACK_PUSH (hcl, ((hcl_oop_oop_t)hcl->active_context->origin->receiver_or_base)->slot[b1]); break; /* ------------------------------------------------- */ case HCL_CODE_STORE_INTO_INSTVAR_X: FETCH_PARAM_CODE_TO (hcl, b1); goto store_instvar; case HCL_CODE_STORE_INTO_INSTVAR_0: case HCL_CODE_STORE_INTO_INSTVAR_1: case HCL_CODE_STORE_INTO_INSTVAR_2: case HCL_CODE_STORE_INTO_INSTVAR_3: case HCL_CODE_STORE_INTO_INSTVAR_4: case HCL_CODE_STORE_INTO_INSTVAR_5: case HCL_CODE_STORE_INTO_INSTVAR_6: case HCL_CODE_STORE_INTO_INSTVAR_7: b1 = bcode & 0x7; /* low 3 bits */ store_instvar: LOG_INST_1 (hcl, "store_into_instvar %zu", b1); HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_TYPE(hcl->active_context->receiver_or_base) == HCL_OBJ_TYPE_OOP); ((hcl_oop_oop_t)hcl->active_context->origin->receiver_or_base)->slot[b1] = HCL_STACK_GETTOP(hcl); break; /* ------------------------------------------------- */ case HCL_CODE_POP_INTO_INSTVAR_X: FETCH_PARAM_CODE_TO (hcl, b1); goto pop_into_instvar; case HCL_CODE_POP_INTO_INSTVAR_0: case HCL_CODE_POP_INTO_INSTVAR_1: case HCL_CODE_POP_INTO_INSTVAR_2: case HCL_CODE_POP_INTO_INSTVAR_3: case HCL_CODE_POP_INTO_INSTVAR_4: case HCL_CODE_POP_INTO_INSTVAR_5: case HCL_CODE_POP_INTO_INSTVAR_6: case HCL_CODE_POP_INTO_INSTVAR_7: b1 = bcode & 0x7; /* low 3 bits */ pop_into_instvar: LOG_INST_1 (hcl, "pop_into_instvar %zu", b1); HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_TYPE(hcl->active_context->receiver_or_base) == HCL_OBJ_TYPE_OOP); ((hcl_oop_oop_t)hcl->active_context->origin->receiver_or_base)->slot[b1] = HCL_STACK_GETTOP(hcl); HCL_STACK_POP (hcl); break; #endif /* ------------------------------------------------- */ case HCL_CODE_PUSH_TEMPVAR_X: case HCL_CODE_STORE_INTO_TEMPVAR_X: case HCL_CODE_POP_INTO_TEMPVAR_X: FETCH_PARAM_CODE_TO (hcl, b1); goto handle_tempvar; case HCL_CODE_PUSH_TEMPVAR_0: case HCL_CODE_PUSH_TEMPVAR_1: case HCL_CODE_PUSH_TEMPVAR_2: case HCL_CODE_PUSH_TEMPVAR_3: case HCL_CODE_PUSH_TEMPVAR_4: case HCL_CODE_PUSH_TEMPVAR_5: case HCL_CODE_PUSH_TEMPVAR_6: case HCL_CODE_PUSH_TEMPVAR_7: case HCL_CODE_STORE_INTO_TEMPVAR_0: case HCL_CODE_STORE_INTO_TEMPVAR_1: case HCL_CODE_STORE_INTO_TEMPVAR_2: case HCL_CODE_STORE_INTO_TEMPVAR_3: case HCL_CODE_STORE_INTO_TEMPVAR_4: case HCL_CODE_STORE_INTO_TEMPVAR_5: case HCL_CODE_STORE_INTO_TEMPVAR_6: case HCL_CODE_STORE_INTO_TEMPVAR_7: case HCL_CODE_POP_INTO_TEMPVAR_0: case HCL_CODE_POP_INTO_TEMPVAR_1: case HCL_CODE_POP_INTO_TEMPVAR_2: case HCL_CODE_POP_INTO_TEMPVAR_3: case HCL_CODE_POP_INTO_TEMPVAR_4: case HCL_CODE_POP_INTO_TEMPVAR_5: case HCL_CODE_POP_INTO_TEMPVAR_6: case HCL_CODE_POP_INTO_TEMPVAR_7: { hcl_oop_context_t ctx; hcl_ooi_t bx; b1 = bcode & 0x7; /* low 3 bits */ handle_tempvar: /* when CTXTEMPVAR inststructions are used, the above * instructions are used only for temporary access * outside a block. i can assume that the temporary * variable index is pointing to one of temporaries * in the relevant method context */ ctx = hcl->active_context->origin; bx = b1; HCL_ASSERT (hcl, HCL_IS_CONTEXT(hcl, ctx)); if ((bcode >> 4) & 1) { /* push - bit 4 on */ LOG_INST_1 (hcl, "push_tempvar %zu", b1); HCL_STACK_PUSH (hcl, ctx->slot[bx]); } else { /* store or pop - bit 5 off */ ctx->slot[bx] = HCL_STACK_GETTOP(hcl); if ((bcode >> 3) & 1) { /* pop - bit 3 on */ LOG_INST_1 (hcl, "pop_into_tempvar %zu", b1); HCL_STACK_POP (hcl); } else { LOG_INST_1 (hcl, "store_into_tempvar %zu", b1); } } break; } /* ------------------------------------------------- */ case HCL_CODE_PUSH_LITERAL_X2: FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); #if (HCL_HCL_CODE_LONG_PARAM_SIZE == 2) b1 = (b1 << 16) | b2; #else b1 = (b1 << 8) | b2; #endif goto push_literal; case HCL_CODE_PUSH_LITERAL_X: FETCH_PARAM_CODE_TO (hcl, b1); goto push_literal; case HCL_CODE_PUSH_LITERAL_0: case HCL_CODE_PUSH_LITERAL_1: case HCL_CODE_PUSH_LITERAL_2: case HCL_CODE_PUSH_LITERAL_3: case HCL_CODE_PUSH_LITERAL_4: case HCL_CODE_PUSH_LITERAL_5: case HCL_CODE_PUSH_LITERAL_6: case HCL_CODE_PUSH_LITERAL_7: b1 = bcode & 0x7; /* low 3 bits */ push_literal: LOG_INST_1 (hcl, "push_literal @%zu", b1); /*HCL_STACK_PUSH (hcl, hcl->code.lit.arr->slot[b1]);*/ HCL_STACK_PUSH (hcl, hcl->active_function->literal_frame[b1]); break; /* ------------------------------------------------- */ case HCL_CODE_PUSH_OBJECT_X: case HCL_CODE_STORE_INTO_OBJECT_X: case HCL_CODE_POP_INTO_OBJECT_X: FETCH_PARAM_CODE_TO (hcl, b1); goto handle_object; case HCL_CODE_PUSH_OBJECT_0: case HCL_CODE_PUSH_OBJECT_1: case HCL_CODE_PUSH_OBJECT_2: case HCL_CODE_PUSH_OBJECT_3: case HCL_CODE_STORE_INTO_OBJECT_0: case HCL_CODE_STORE_INTO_OBJECT_1: case HCL_CODE_STORE_INTO_OBJECT_2: case HCL_CODE_STORE_INTO_OBJECT_3: case HCL_CODE_POP_INTO_OBJECT_0: case HCL_CODE_POP_INTO_OBJECT_1: case HCL_CODE_POP_INTO_OBJECT_2: case HCL_CODE_POP_INTO_OBJECT_3: { hcl_oop_cons_t ass; b1 = bcode & 0x3; /* low 2 bits */ handle_object: /*ass = hcl->code.lit.arr->slot[b1];*/ ass = (hcl_oop_cons_t)hcl->active_function->literal_frame[b1]; HCL_ASSERT (hcl, HCL_IS_CONS(hcl, ass)); if ((bcode >> 3) & 1) { /* store or pop */ ass->cdr = HCL_STACK_GETTOP(hcl); if ((bcode >> 2) & 1) { /* pop */ LOG_INST_1 (hcl, "pop_into_object @%zu", b1); HCL_STACK_POP (hcl); } else { LOG_INST_1 (hcl, "store_into_object @%zu", b1); } } else { /* push */ LOG_INST_1 (hcl, "push_object @%zu", b1); HCL_STACK_PUSH (hcl, ass->cdr); } break; } /* -------------------------------------------------------- */ case HCL_CODE_JUMP_FORWARD_X: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump_forward %zu", b1); hcl->ip += b1; break; case HCL_CODE_JUMP_FORWARD_0: case HCL_CODE_JUMP_FORWARD_1: case HCL_CODE_JUMP_FORWARD_2: case HCL_CODE_JUMP_FORWARD_3: LOG_INST_1 (hcl, "jump_forward %zu", (hcl_oow_t)(bcode & 0x3)); hcl->ip += (bcode & 0x3); /* low 2 bits */ break; case HCL_CODE_JUMP_BACKWARD_X: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump_backward %zu", b1); hcl->ip -= b1; break; case HCL_CODE_JUMP_BACKWARD_0: case HCL_CODE_JUMP_BACKWARD_1: case HCL_CODE_JUMP_BACKWARD_2: case HCL_CODE_JUMP_BACKWARD_3: LOG_INST_1 (hcl, "jump_backward %zu", (hcl_oow_t)(bcode & 0x3)); hcl->ip -= (bcode & 0x3); /* low 2 bits */ break; case HCL_CODE_JUMP_FORWARD_IF_TRUE: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump_forward_if_true %zu", b1); /*if (HCL_STACK_GETTOP(hcl) == hcl->_true) hcl->ip += b1; TODO: _true or not _false?*/ if (HCL_STACK_GETTOP(hcl) != hcl->_false) hcl->ip += b1; break; case HCL_CODE_JUMP2_FORWARD_IF_TRUE: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump2_forward_if_true %zu", b1); /*if (HCL_STACK_GETTOP(hcl) == hcl->_true) hcl->ip += MAX_CODE_JUMP + b1;*/ if (HCL_STACK_GETTOP(hcl) != hcl->_false) hcl->ip += MAX_CODE_JUMP + b1; break; case HCL_CODE_JUMP_FORWARD_IF_FALSE: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump_forward_if_false %zu", b1); if (HCL_STACK_GETTOP(hcl) == hcl->_false) hcl->ip += b1; break; case HCL_CODE_JUMP2_FORWARD_IF_FALSE: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump2_forward_if_false %zu", b1); if (HCL_STACK_GETTOP(hcl) == hcl->_false) hcl->ip += MAX_CODE_JUMP + b1; break; case HCL_CODE_JUMP2_FORWARD: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump2_forward %zu", b1); hcl->ip += MAX_CODE_JUMP + b1; break; case HCL_CODE_JUMP2_BACKWARD: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "jump2_backward %zu", b1); hcl->ip -= MAX_CODE_JUMP + b1; break; /* -------------------------------------------------------- */ case HCL_CODE_CALL_X: FETCH_PARAM_CODE_TO (hcl, b1); goto handle_call; case HCL_CODE_CALL_0: case HCL_CODE_CALL_1: case HCL_CODE_CALL_2: case HCL_CODE_CALL_3: { hcl_oop_t rcv; b1 = bcode & 0x3; /* low 2 bits */ handle_call: LOG_INST_1 (hcl, "call %zu", b1); rcv = HCL_STACK_GETRCV(hcl, b1); if (HCL_OOP_IS_POINTER(rcv)) { switch (HCL_OBJ_GET_FLAGS_BRAND(rcv)) { case HCL_BRAND_FUNCTION: if (activate_function(hcl, b1) <= -1) goto oops; break; case HCL_BRAND_BLOCK: if (activate_block(hcl, b1) <= -1) goto oops; break; case HCL_BRAND_PRIM: if (call_primitive(hcl, b1) <= -1) goto oops; break; default: goto cannot_call; } } else { cannot_call: /* run time error */ hcl_seterrbfmt (hcl, HCL_ECALL, "cannot call %O", rcv); goto oops; } break; } /* -------------------------------------------------------- */ case HCL_CODE_PUSH_CTXTEMPVAR_X: case HCL_CODE_STORE_INTO_CTXTEMPVAR_X: case HCL_CODE_POP_INTO_CTXTEMPVAR_X: FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); goto handle_ctxtempvar; case HCL_CODE_PUSH_CTXTEMPVAR_0: case HCL_CODE_PUSH_CTXTEMPVAR_1: case HCL_CODE_PUSH_CTXTEMPVAR_2: case HCL_CODE_PUSH_CTXTEMPVAR_3: case HCL_CODE_STORE_INTO_CTXTEMPVAR_0: case HCL_CODE_STORE_INTO_CTXTEMPVAR_1: case HCL_CODE_STORE_INTO_CTXTEMPVAR_2: case HCL_CODE_STORE_INTO_CTXTEMPVAR_3: case HCL_CODE_POP_INTO_CTXTEMPVAR_0: case HCL_CODE_POP_INTO_CTXTEMPVAR_1: case HCL_CODE_POP_INTO_CTXTEMPVAR_2: case HCL_CODE_POP_INTO_CTXTEMPVAR_3: { hcl_ooi_t i; hcl_oop_context_t ctx; b1 = bcode & 0x3; /* low 2 bits */ FETCH_BYTE_CODE_TO (hcl, b2); handle_ctxtempvar: ctx = hcl->active_context; HCL_ASSERT (hcl, (hcl_oop_t)ctx != hcl->_nil); for (i = 0; i < b1; i++) { ctx = (hcl_oop_context_t)ctx->home; /* the initial context has nil in the home field. * the loop must not reach beyond the initial context */ HCL_ASSERT (hcl, (hcl_oop_t)ctx != hcl->_nil); } if ((bcode >> 3) & 1) { /* store or pop */ ctx->slot[b2] = HCL_STACK_GETTOP(hcl); if ((bcode >> 2) & 1) { /* pop */ HCL_STACK_POP (hcl); LOG_INST_2 (hcl, "pop_into_ctxtempvar %zu %zu", b1, b2); } else { LOG_INST_2 (hcl, "store_into_ctxtempvar %zu %zu", b1, b2); } } else { /* push */ HCL_STACK_PUSH (hcl, ctx->slot[b2]); LOG_INST_2 (hcl, "push_ctxtempvar %zu %zu", b1, b2); } break; } /* -------------------------------------------------------- */ case HCL_CODE_PUSH_OBJVAR_X: case HCL_CODE_STORE_INTO_OBJVAR_X: case HCL_CODE_POP_INTO_OBJVAR_X: FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); goto handle_objvar; case HCL_CODE_PUSH_OBJVAR_0: case HCL_CODE_PUSH_OBJVAR_1: case HCL_CODE_PUSH_OBJVAR_2: case HCL_CODE_PUSH_OBJVAR_3: case HCL_CODE_STORE_INTO_OBJVAR_0: case HCL_CODE_STORE_INTO_OBJVAR_1: case HCL_CODE_STORE_INTO_OBJVAR_2: case HCL_CODE_STORE_INTO_OBJVAR_3: case HCL_CODE_POP_INTO_OBJVAR_0: case HCL_CODE_POP_INTO_OBJVAR_1: case HCL_CODE_POP_INTO_OBJVAR_2: case HCL_CODE_POP_INTO_OBJVAR_3: { hcl_oop_oop_t t; /* b1 -> variable index to the object indicated by b2. * b2 -> object index stored in the literal frame. */ b1 = bcode & 0x3; /* low 2 bits */ FETCH_BYTE_CODE_TO (hcl, b2); handle_objvar: /*t = hcl->code.lit.arr->slot[b2];*/ t = (hcl_oop_oop_t)hcl->active_function->literal_frame[b2]; HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_TYPE(t) == HCL_OBJ_TYPE_OOP); HCL_ASSERT (hcl, b1 < HCL_OBJ_GET_SIZE(t)); if ((bcode >> 3) & 1) { /* store or pop */ t->slot[b1] = HCL_STACK_GETTOP(hcl); if ((bcode >> 2) & 1) { /* pop */ HCL_STACK_POP (hcl); LOG_INST_2 (hcl, "pop_into_objvar %zu %zu", b1, b2); } else { LOG_INST_2 (hcl, "store_into_objvar %zu %zu", b1, b2); } } else { /* push */ LOG_INST_2 (hcl, "push_objvar %zu %zu", b1, b2); HCL_STACK_PUSH (hcl, t->slot[b1]); } break; } /* -------------------------------------------------------- */ #if 0 case HCL_CODE_SEND_MESSAGE_X: case HCL_CODE_SEND_MESSAGE_TO_SUPER_X: /* b1 -> number of arguments * b2 -> selector index stored in the literal frame */ FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); goto handle_send_message; case HCL_CODE_SEND_MESSAGE_0: case HCL_CODE_SEND_MESSAGE_1: case HCL_CODE_SEND_MESSAGE_2: case HCL_CODE_SEND_MESSAGE_3: case HCL_CODE_SEND_MESSAGE_TO_SUPER_0: case HCL_CODE_SEND_MESSAGE_TO_SUPER_1: case HCL_CODE_SEND_MESSAGE_TO_SUPER_2: case HCL_CODE_SEND_MESSAGE_TO_SUPER_3: { hcl_oop_char_t selector; b1 = bcode & 0x3; /* low 2 bits */ FETCH_BYTE_CODE_TO (hcl, b2); handle_send_message: /* get the selector from the literal frame */ selector = (hcl_oop_char_t)hcl->active_method->slot[b2]; LOG_INST_3 (hcl, "send_message%hs %zu @%zu", (((bcode >> 2) & 1)? "_to_super": ""), b1, b2); if (send_message (hcl, selector, ((bcode >> 2) & 1), b1) <= -1) goto oops; break; /* CMD_SEND_MESSAGE */ } #endif /* -------------------------------------------------------- */ case HCL_CODE_PUSH_RECEIVER: /* push self or super */ LOG_INST_0 (hcl, "push_receiver"); HCL_STACK_PUSH (hcl, hcl->active_context->origin->receiver_or_base); break; case HCL_CODE_PUSH_NIL: LOG_INST_0 (hcl, "push_nil"); HCL_STACK_PUSH (hcl, hcl->_nil); break; case HCL_CODE_PUSH_TRUE: LOG_INST_0 (hcl, "push_true"); HCL_STACK_PUSH (hcl, hcl->_true); break; case HCL_CODE_PUSH_FALSE: LOG_INST_0 (hcl, "push_false"); HCL_STACK_PUSH (hcl, hcl->_false); break; case HCL_CODE_PUSH_CONTEXT: LOG_INST_0 (hcl, "push_context"); HCL_STACK_PUSH (hcl, (hcl_oop_t)hcl->active_context); break; case HCL_CODE_PUSH_PROCESS: LOG_INST_0 (hcl, "push_process"); HCL_STACK_PUSH (hcl, (hcl_oop_t)hcl->processor->active); break; case HCL_CODE_PUSH_NEGONE: LOG_INST_0 (hcl, "push_negone"); HCL_STACK_PUSH (hcl, HCL_SMOOI_TO_OOP(-1)); break; case HCL_CODE_PUSH_ZERO: LOG_INST_0 (hcl, "push_zero"); HCL_STACK_PUSH (hcl, HCL_SMOOI_TO_OOP(0)); break; case HCL_CODE_PUSH_ONE: LOG_INST_0 (hcl, "push_one"); HCL_STACK_PUSH (hcl, HCL_SMOOI_TO_OOP(1)); break; case HCL_CODE_PUSH_TWO: LOG_INST_0 (hcl, "push_two"); HCL_STACK_PUSH (hcl, HCL_SMOOI_TO_OOP(2)); break; case HCL_CODE_PUSH_INTLIT: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "push_intlit %zu", b1); HCL_STACK_PUSH (hcl, HCL_SMOOI_TO_OOP(b1)); break; case HCL_CODE_PUSH_NEGINTLIT: { hcl_ooi_t num; FETCH_PARAM_CODE_TO (hcl, b1); num = b1; LOG_INST_1 (hcl, "push_negintlit %zu", b1); HCL_STACK_PUSH (hcl, HCL_SMOOI_TO_OOP(-num)); break; } case HCL_CODE_PUSH_CHARLIT: FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "push_charlit %zu", b1); HCL_STACK_PUSH (hcl, HCL_CHAR_TO_OOP(b1)); break; /* -------------------------------------------------------- */ case HCL_CODE_MAKE_ARRAY: { hcl_oop_t t; FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "make_array %zu", b1); /* create an empty array */ t = hcl_makearray (hcl, b1, 0); if (!t) goto oops; HCL_STACK_PUSH (hcl, t); /* push the array created */ break; } case HCL_CODE_POP_INTO_ARRAY: { hcl_oop_t t1, t2; FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "pop_into_array %zu", b1); t1 = HCL_STACK_GETTOP(hcl); /* value to store */ HCL_STACK_POP (hcl); t2 = HCL_STACK_GETTOP(hcl); /* array */ ((hcl_oop_oop_t)t2)->slot[b1] = t1; break; } case HCL_CODE_MAKE_BYTEARRAY: { hcl_oop_t t; FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "make_bytearray %zu", b1); /* create an empty array */ t = hcl_makebytearray (hcl, HCL_NULL, b1); if (!t) goto oops; HCL_STACK_PUSH (hcl, t); /* push the byte array created */ break; } case HCL_CODE_POP_INTO_BYTEARRAY: { hcl_oop_t t1, t2; hcl_ooi_t bv; FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "pop_into_bytearray %zu", b1); t1 = HCL_STACK_GETTOP(hcl); /* value to store */ if (!HCL_OOP_IS_SMOOI(t1) || (bv = HCL_OOP_TO_SMOOI(t1)) < 0 || bv > 255) { hcl_seterrbfmt (hcl, HCL_ERANGE, "not a byte or out of byte range - %O", t1); goto oops; } HCL_STACK_POP (hcl); t2 = HCL_STACK_GETTOP(hcl); /* array */ ((hcl_oop_byte_t)t2)->slot[b1] = bv; break; } case HCL_CODE_MAKE_DIC: { hcl_oop_t t; FETCH_PARAM_CODE_TO (hcl, b1); LOG_INST_1 (hcl, "make_dic %zu", b1); t = (hcl_oop_t)hcl_makedic (hcl, b1 + 10); if (!t) goto oops; HCL_STACK_PUSH (hcl, t); break; } case HCL_CODE_POP_INTO_DIC: { hcl_oop_t t1, t2, t3; LOG_INST_0 (hcl, "pop_into_dic"); t1 = HCL_STACK_GETTOP(hcl); /* value */ HCL_STACK_POP (hcl); t2 = HCL_STACK_GETTOP(hcl); /* key */ HCL_STACK_POP (hcl); t3 = HCL_STACK_GETTOP(hcl); /* dictionary */ if (!hcl_putatdic (hcl, (hcl_oop_dic_t)t3, t2, t1)) goto oops; break; } /* -------------------------------------------------------- */ case HCL_CODE_DUP_STACKTOP: { hcl_oop_t t; LOG_INST_0 (hcl, "dup_stacktop"); HCL_ASSERT (hcl, !HCL_STACK_ISEMPTY(hcl)); t = HCL_STACK_GETTOP(hcl); HCL_STACK_PUSH (hcl, t); break; } case HCL_CODE_POP_STACKTOP: LOG_INST_0 (hcl, "pop_stacktop"); HCL_ASSERT (hcl, !HCL_STACK_ISEMPTY(hcl)); /* at the top level, the value is just popped off the stack * after evaluation of an expression. so it's likely the * return value of the last expression unless explicit * returning is performed */ hcl->last_retv = HCL_STACK_GETTOP(hcl); HCL_STACK_POP (hcl); break; case HCL_CODE_RETURN_STACKTOP: LOG_INST_0 (hcl, "return_stacktop"); return_value = HCL_STACK_GETTOP(hcl); HCL_STACK_POP (hcl); goto handle_return; case HCL_CODE_RETURN_RECEIVER: LOG_INST_0 (hcl, "return_receiver"); return_value = hcl->active_context->origin->receiver_or_base; handle_return: hcl->last_retv = return_value; #if 1 if (do_return(hcl, return_value) <= -1) goto oops; #else //if (hcl->active_context->origin == hcl->processor->active->initial_context->origin) if (hcl->active_context == hcl->processor->active->initial_context) { terminate_process (hcl, hcl->processor->active); } else { if (hcl->active_context->origin == hcl->active_context) { /* it is a method context... */ /* set the instruction pointer to an invalid value. * this is stored into the current method context * before context switching in SWITCH_ACTIVE_CONTEXT and * marks a dead context */ hcl->ip = -1; } else { hcl_oop_context_t ctx; /* method return from within a block(including a non-local return) */ /* // HCL_ASSERT (hcl, HCL_CLASSOF(hcl, hcl->active_context) == hcl->_block_context); */ ctx = hcl->active_context; while ((hcl_oop_t)ctx != hcl->_nil) { if (ctx == hcl->active_context->origin) goto non_local_return_ok; ctx = ctx->sender; } /* cannot return from a method that has returned already */ /* // HCL_ASSERT (hcl, HCL_CLASSOF(hcl, hcl->active_context->origin) == hcl->_method_context); */ HCL_ASSERT (hcl, hcl->active_context->origin->ip == HCL_SMOOI_TO_OOP(-1)); HCL_LOG0 (hcl, HCL_LOG_IC | HCL_LOG_ERROR, "Error - cannot return from dead context\n"); hcl_seterrbfmt (hcl, HCL_EINTERN, "unable to return from dead context"); /* TODO: can i make this error catchable at the hcl level? */ goto oops; non_local_return_ok: /*HCL_DEBUG2 (hcl, "NON_LOCAL RETURN OK TO... %p %p\n", hcl->active_context->origin, hcl->active_context->origin->sender);*/ hcl->active_context->origin->ip = HCL_SMOOI_TO_OOP(-1); } /* restore the stack pointer */ hcl->sp = HCL_OOP_TO_SMOOI(hcl->active_context->origin->sp); SWITCH_ACTIVE_CONTEXT (hcl, hcl->active_context->origin->sender); /* push the return value to the stack of the new active context */ HCL_STACK_PUSH (hcl, return_value); if (hcl->active_context == hcl->initial_context) { /* the new active context is the fake initial context. * this context can't get executed further. */ HCL_ASSERT (hcl, HCL_IS_CONTEXT(hcl, hcl->active_context)); HCL_ASSERT (hcl, (hcl_oop_t)hcl->active_context->sender == hcl->_nil); //HCL_ASSERT (hcl, hcl->active_context->receiver_or_base == hcl->_nil); HCL_ASSERT (hcl, hcl->active_context == hcl->processor->active->initial_context); HCL_ASSERT (hcl, hcl->active_context->origin == hcl->processor->active->initial_context->origin); HCL_ASSERT (hcl, hcl->active_context->origin == hcl->active_context); /* NOTE: this condition is true for the processified block context also. * hcl->active_context->origin == hcl->processor->active->initial_context->origin * however, the check here is done after context switching and the * processified block check has been done against the context before switching */ /* the stack contains the final return value so the stack pointer must be 0. */ HCL_ASSERT (hcl, hcl->sp == 0); if (hcl->option.trait & HCL_TRAIT_AWAIT_PROCS) terminate_process (hcl, hcl->processor->active); else goto done; /* TODO: store the return value to the VM register. * the caller to hcl_execute() can fetch it to return it to the system */ } } #endif break; case HCL_CODE_RETURN_FROM_BLOCK: LOG_INST_0 (hcl, "return_from_block"); HCL_ASSERT(hcl, HCL_IS_CONTEXT(hcl, hcl->active_context)); hcl->last_retv = HCL_STACK_GETTOP(hcl); HCL_DEBUG1 (hcl, "RETURNING(return_from_block) FROM active_context %p\n", hcl->active_context); if (hcl->active_context == hcl->processor->active->initial_context) { /* the active context to return from is an initial context of * the active process. let's terminate the process. * the initial context has been forged over the initial function * in start_initial_process_and_context() */ HCL_ASSERT (hcl, (hcl_oop_t)hcl->active_context->sender == hcl->_nil); hcl->active_context->ip = HCL_SMOOI_TO_OOP(-1); /* mark that this context has returned */ terminate_process (hcl, hcl->processor->active); } else { hcl->ip = -1; /* this will be saved to hcl->active_context->ip in SWITCH_ACTIVE_CONTEXT */ /* it is a normal block return as the active block context * is not the initial context of a process */ SWITCH_ACTIVE_CONTEXT (hcl, (hcl_oop_context_t)hcl->active_context->sender); } break; case HCL_CODE_MAKE_FUNCTION: { hcl_oop_function_t func; hcl_oow_t b3, b4, i, j; hcl_oow_t joff; /* b1 - number of block arguments * b2 - number of block temporaries * b3 - literal frame base * b4 - literal frame size */ FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); FETCH_PARAM_CODE_TO (hcl, b3); FETCH_PARAM_CODE_TO (hcl, b4); LOG_INST_4 (hcl, "make_function %zu %zu %zu %zu", b1, b2, b3, b4); HCL_ASSERT (hcl, b1 >= 0); HCL_ASSERT (hcl, b2 >= b1); /* the MAKE_FUNCTION instruction is followed by the long JUMP_FORWARD_X instruction. * i can decode the instruction and get the size of instructions * of the block context */ HCL_ASSERT (hcl, hcl->active_code[hcl->ip] == HCL_CODE_JUMP_FORWARD_X); joff = hcl->active_code[hcl->ip + 1]; #if (HCL_HCL_CODE_LONG_PARAM_SIZE == 2) joff = (joff << 8) | hcl->active_code[hcl->ip + 2]; #endif /* copy the byte codes from the active context to the new context */ #if (HCL_HCL_CODE_LONG_PARAM_SIZE == 2) func = make_function(hcl, b4, &hcl->active_code[hcl->ip + 3], joff); #else func = make_function(hcl, b4, &hcl->active_code[hcl->ip + 2], joff); #endif if (HCL_UNLIKELY(!func)) goto oops; fill_function_data (hcl, func, b1, b2, hcl->active_context, &hcl->active_function->literal_frame[b3], b4); /* push the new function to the stack of the active context */ HCL_STACK_PUSH (hcl, (hcl_oop_t)func); break; } case HCL_CODE_MAKE_BLOCK: { #if 0 hcl_oop_context_t blkctx; /* b1 - number of block arguments * b2 - number of block temporaries */ FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); LOG_INST_2 (hcl, "make_block %zu %zu", b1, b2); HCL_ASSERT (hcl, b1 >= 0); HCL_ASSERT (hcl, b2 >= b1); /* the block context object created here is used as a base * object for block context activation. activate_context() * clones a block context and activates the cloned context. * this base block context is created with no temporaries * for this reason */ blkctx = (hcl_oop_context_t)make_context(hcl, 0); if (HCL_UNLIKELY(!blkctx)) goto oops; /* the long forward jump instruction has the format of * 11000100 KKKKKKKK or 11000100 KKKKKKKK KKKKKKKK * depending on HCL_HCL_CODE_LONG_PARAM_SIZE. change 'ip' to point to * the instruction after the jump. */ blkctx->ip = HCL_SMOOI_TO_OOP(hcl->ip + HCL_HCL_CODE_LONG_PARAM_SIZE + 1); /* stack pointer below the bottom. this base block context * has an empty stack anyway. */ blkctx->sp = HCL_SMOOI_TO_OOP(-1); /* the number of arguments */ blkctx->nargs = HCL_SMOOI_TO_OOP(b1); /* the number of temporaries including arguments */ blkctx->ntmprs = HCL_SMOOI_TO_OOP(b2); /* no source for a base block context. */ blkctx->receiver_or_base = hcl->_nil; /* set the home context where it's defined */ blkctx->home = (hcl_oop_t)hcl->active_context; blkctx->origin = hcl->active_context->origin; /* push the new block context to the stack of the active context */ HCL_STACK_PUSH (hcl, (hcl_oop_t)blkctx); #else hcl_oop_block_t blkobj; /* b1 - number of block arguments * b2 - number of block temporaries */ FETCH_PARAM_CODE_TO (hcl, b1); FETCH_PARAM_CODE_TO (hcl, b2); LOG_INST_2 (hcl, "make_block %zu %zu", b1, b2); HCL_ASSERT (hcl, b1 >= 0); HCL_ASSERT (hcl, b2 >= b1); blkobj = make_block(hcl); if (HCL_UNLIKELY(!blkobj)) goto oops; /* the long forward jump instruction has the format of * 11000100 KKKKKKKK or 11000100 KKKKKKKK KKKKKKKK * depending on HCL_HCL_CODE_LONG_PARAM_SIZE. change 'ip' to point to * the instruction after the jump. */ fill_block_data (hcl, blkobj, b1, b2, hcl->ip + HCL_HCL_CODE_LONG_PARAM_SIZE + 1, hcl->active_context); /* push the new block context to the stack of the active context */ HCL_STACK_PUSH (hcl, (hcl_oop_t)blkobj); #endif break; } case HCL_CODE_NOOP: /* do nothing */ LOG_INST_0 (hcl, "noop"); break; default: HCL_LOG1 (hcl, HCL_LOG_IC | HCL_LOG_FATAL, "Fatal error - unknown byte code 0x%zx\n", bcode); hcl_seterrnum (hcl, HCL_EINTERN); goto oops; } } done: vm_cleanup (hcl); #if defined(HCL_PROFILE_VM) HCL_LOG1 (hcl, HCL_LOG_IC | HCL_LOG_INFO, "TOTAL INST COUTNER = %zu\n", inst_counter); #endif return 0; oops: /* TODO: anything to do here? */ if (hcl->processor->active != hcl->nil_process) { HCL_LOG1 (hcl, HCL_LOG_IC | HCL_LOG_INFO, "TERMINATING ACTIVE PROCESS %zd for execution error\n", HCL_OOP_TO_SMOOI(hcl->processor->active->id)); terminate_process (hcl, hcl->processor->active); } return -1; } hcl_oop_t hcl_execute (hcl_t* hcl) { ////////////////////////////////////////////////////////////////////////////////////////////// hcl_oop_function_t func; int n; hcl_bitmask_t log_default_type_mask; HCL_ASSERT (hcl, hcl->code.bc.len < HCL_SMOOI_MAX); /* asserted by the compiler */ log_default_type_mask = hcl->log.default_type_mask; hcl->log.default_type_mask |= HCL_LOG_VM; HCL_ASSERT (hcl, hcl->initial_context == HCL_NULL); HCL_ASSERT (hcl, hcl->active_context == HCL_NULL); /* the code generated doesn't cater for its use as an initial funtion. * mutate the generated code so that the intiail function can break * out of the execution loop in execute() smoothly */ HCL_ASSERT (hcl, hcl->code.bc.ptr[hcl->code.bc.len - 1] == HCL_CODE_POP_STACKTOP); #if 1 /* append RETURN_FROM_BLOCK if (hcl_emitbyteinstruction(hcl, HCL_CODE_RETURN_FROM_BLOCK) <= -1) return -1;*/ /* substitute RETURN_FROM_BLOCK for POP_STACKTOP) */ hcl->code.bc.ptr[hcl->code.bc.len - 1] = HCL_CODE_RETURN_FROM_BLOCK; #else /* substitute RETURN_STACKTOP for POP_STACKTOP) */ hcl->code.bc.ptr[hcl->code.bc.len - 1] = HCL_CODE_RETURN_STACKTOP; #endif func = make_function(hcl, hcl->code.lit.len, hcl->code.bc.ptr, hcl->code.bc.len); if (HCL_UNLIKELY(!func)) return HCL_NULL; /* pass nil for the home context of the initial function */ fill_function_data (hcl, func, 0, 0, (hcl_oop_context_t)hcl->_nil, hcl->code.lit.arr->slot, hcl->code.lit.len); hcl->initial_function = func; ////////////////////////////////////////////////////////////////////////////////////////////// hcl->last_retv = hcl->_nil; n = start_initial_process_and_context(hcl, 0); if (n >= 0) { n = execute(hcl); HCL_INFO1 (hcl, "RETURNED VALUE - %O\n", hcl->last_retv); } /* TODO: reset processor fields. set processor->tally to zero. processor->active to nil_process... */ hcl->initial_context = HCL_NULL; hcl->active_context = HCL_NULL; hcl->log.default_type_mask = log_default_type_mask; return (n <= -1)? HCL_NULL: hcl->last_retv; } void hcl_abort (hcl_t* hcl) { hcl->abort_req = 1; }