545 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			545 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * $Id$
 | |
|  *
 | |
|     Copyright (c) 2016-2018 Chung, Hyung-Hwan. All rights reserved.
 | |
| 
 | |
|     Redistribution and use in source and binary forms, with or without
 | |
|     modification, are permitted provided that the following conditions
 | |
|     are met:
 | |
|     1. Redistributions of source code must retain the above copyright
 | |
|        notice, this list of conditions and the following disclaimer.
 | |
|     2. Redistributions in binary form must reproduce the above copyright
 | |
|        notice, this list of conditions and the following disclaimer in the
 | |
|        documentation and/or other materials provided with the distribution.
 | |
| 
 | |
|     THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR
 | |
|     IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 | |
|     OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 | |
|     IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
 | |
|     INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
 | |
|     NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 | |
|     DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 | |
|     THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 | |
|     (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
 | |
|     THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 | |
|  */
 | |
| 
 | |
| #include "hcl-prv.h"
 | |
| 
 | |
| static struct 
 | |
| {
 | |
| 	hcl_oow_t  len;
 | |
| 	hcl_ooch_t ptr[10];
 | |
| 	int syncode;
 | |
| 	hcl_oow_t  offset;
 | |
| } syminfo[] =
 | |
| {
 | |
| 	{  3, { 'a','n','d' },                 HCL_SYNCODE_AND,     HCL_OFFSETOF(hcl_t,_and)  },
 | |
| 	{  5, { 'b','r','e','a','k' },         HCL_SYNCODE_BREAK,   HCL_OFFSETOF(hcl_t,_break)  },
 | |
| 	{  5, { 'd','e','f','u','n' },         HCL_SYNCODE_DEFUN,   HCL_OFFSETOF(hcl_t,_defun)  },
 | |
| 	{  2, { 'd','o' },                     HCL_SYNCODE_DO,      HCL_OFFSETOF(hcl_t,_do)  },
 | |
| 	{  4, { 'e','l','i','f' },             HCL_SYNCODE_ELIF,    HCL_OFFSETOF(hcl_t,_elif)   },
 | |
| 	{  4, { 'e','l','s','e' },             HCL_SYNCODE_ELSE,    HCL_OFFSETOF(hcl_t,_else)   },
 | |
| 	{  2, { 'i','f' },                     HCL_SYNCODE_IF,      HCL_OFFSETOF(hcl_t,_if)     },
 | |
| 	{  6, { 'l','a','m','b','d','a' },     HCL_SYNCODE_LAMBDA,  HCL_OFFSETOF(hcl_t,_lambda) },
 | |
| 	{  2, { 'o','r' },                     HCL_SYNCODE_OR,      HCL_OFFSETOF(hcl_t,_or)  },
 | |
| 	{  6, { 'r','e','t','u','r','n'},      HCL_SYNCODE_RETURN,  HCL_OFFSETOF(hcl_t,_return) },
 | |
| 	{  3, { 's','e','t' },                 HCL_SYNCODE_SET,     HCL_OFFSETOF(hcl_t,_set)    },
 | |
| 	{  5, { 'u','n','t','i','l' },         HCL_SYNCODE_UNTIL,   HCL_OFFSETOF(hcl_t,_until)  },
 | |
| 	{  5, { 'w','h','i','l','e' },         HCL_SYNCODE_WHILE,   HCL_OFFSETOF(hcl_t,_while)  }
 | |
| };
 | |
| 
 | |
| /* ========================================================================= */
 | |
| 
 | |
| static void compact_symbol_table (hcl_t* hcl, hcl_oop_t _nil)
 | |
| {
 | |
| 	hcl_oop_char_t symbol;
 | |
| 	hcl_oow_t i, x, y, z;
 | |
| 	hcl_oow_t bucket_size, index;
 | |
| 	hcl_ooi_t tally;
 | |
| 
 | |
| #if defined(HCL_SUPPORT_GC_DURING_IGNITION)
 | |
| 	if (!hcl->symtab) return; /* symbol table has not been created */
 | |
| #endif
 | |
| 
 | |
| 	/* the symbol table doesn't allow more data items than HCL_SMOOI_MAX.
 | |
| 	 * so hcl->symtab->tally must always be a small integer */
 | |
| 	HCL_ASSERT (hcl, HCL_OOP_IS_SMOOI(hcl->symtab->tally));
 | |
| 	tally = HCL_OOP_TO_SMOOI(hcl->symtab->tally);
 | |
| 	HCL_ASSERT (hcl, tally >= 0); /* it must not be less than 0 */
 | |
| 	if (tally <= 0) return;
 | |
| 
 | |
| 	/* NOTE: in theory, the bucket size can be greater than HCL_SMOOI_MAX
 | |
| 	 * as it is an internal header field and is of an unsigned type */
 | |
| 	bucket_size = HCL_OBJ_GET_SIZE(hcl->symtab->bucket);
 | |
| 
 | |
| 	for (index = 0; index < bucket_size; )
 | |
| 	{
 | |
| 		if (HCL_OBJ_GET_FLAGS_MOVED(hcl->symtab->bucket->slot[index]))
 | |
| 		{
 | |
| 			index++;
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		HCL_ASSERT (hcl, hcl->symtab->bucket->slot[index] != _nil);
 | |
| 
 | |
| 		for (i = 0, x = index, y = index; i < bucket_size; i++)
 | |
| 		{
 | |
| 			y = (y + 1) % bucket_size;
 | |
| 
 | |
| 			/* done if the slot at the current hash index is _nil */
 | |
| 			if (hcl->symtab->bucket->slot[y] == _nil) break;
 | |
| 
 | |
| 			/* get the natural hash index for the data in the slot 
 | |
| 			 * at the current hash index */
 | |
| 			symbol = (hcl_oop_char_t)hcl->symtab->bucket->slot[y];
 | |
| 
 | |
| 			HCL_ASSERT (hcl, HCL_IS_SYMBOL(hcl, symbol));
 | |
| 
 | |
| 			z = hcl_hash_oochars(symbol->slot, HCL_OBJ_GET_SIZE(symbol)) % bucket_size;
 | |
| 
 | |
| 			/* move an element if necessary */
 | |
| 			if ((y > x && (z <= x || z > y)) ||
 | |
| 			    (y < x && (z <= x && z > y)))
 | |
| 			{
 | |
| 				hcl->symtab->bucket->slot[x] = hcl->symtab->bucket->slot[y];
 | |
| 				x = y;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		hcl->symtab->bucket->slot[x] = _nil;
 | |
| 		tally--;
 | |
| 	}
 | |
| 
 | |
| 	HCL_ASSERT (hcl, tally >= 0);
 | |
| 	HCL_ASSERT (hcl, tally <= HCL_SMOOI_MAX);
 | |
| 	hcl->symtab->tally = HCL_SMOOI_TO_OOP(tally);
 | |
| }
 | |
| 
 | |
| 
 | |
| static HCL_INLINE hcl_oow_t get_payload_bytes (hcl_t* hcl, hcl_oop_t oop)
 | |
| {
 | |
| 	hcl_oow_t nbytes_aligned;
 | |
| 
 | |
| #if defined(HCL_USE_OBJECT_TRAILER)
 | |
| 	if (HCL_OBJ_GET_FLAGS_TRAILER(oop))
 | |
| 	{
 | |
| 		hcl_oow_t nbytes;
 | |
| 
 | |
| 		/* only an OOP object can have the trailer. 
 | |
| 		 *
 | |
| 		 * | _flags    |
 | |
| 		 * | _size     |  <-- if it's 3
 | |
| 		 * | _class    |
 | |
| 		 * |   X       |
 | |
| 		 * |   X       |
 | |
| 		 * |   X       |
 | |
| 		 * |   Y       | <-- it may exist if EXTRA is set in _flags.
 | |
| 		 * |   Z       | <-- if TRAILER is set, it is the number of bytes in the trailer
 | |
| 		 * |  |  |  |  | 
 | |
| 		 */
 | |
| 		HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_TYPE(oop) == HCL_OBJ_TYPE_OOP);
 | |
| 		HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_UNIT(oop) == HCL_SIZEOF(hcl_oow_t));
 | |
| 		HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_EXTRA(oop) == 0); /* no 'extra' for an OOP object */
 | |
| 
 | |
| 		nbytes = HCL_OBJ_BYTESOF(oop) + HCL_SIZEOF(hcl_oow_t) + \
 | |
| 		         (hcl_oow_t)((hcl_oop_oop_t)oop)->slot[HCL_OBJ_GET_SIZE(oop)];
 | |
| 		nbytes_aligned = HCL_ALIGN (nbytes, HCL_SIZEOF(hcl_oop_t));
 | |
| 	}
 | |
| 	else
 | |
| 	{
 | |
| #endif
 | |
| 		/* calculate the payload size in bytes */
 | |
| 		nbytes_aligned = HCL_ALIGN (HCL_OBJ_BYTESOF(oop), HCL_SIZEOF(hcl_oop_t));
 | |
| #if defined(HCL_USE_OBJECT_TRAILER)
 | |
| 	}
 | |
| #endif
 | |
| 
 | |
| 	return nbytes_aligned;
 | |
| }
 | |
| 
 | |
| hcl_oop_t hcl_moveoop (hcl_t* hcl, hcl_oop_t oop)
 | |
| {
 | |
| #if defined(HCL_SUPPORT_GC_DURING_IGNITION)
 | |
| 	if (!oop) return oop;
 | |
| #endif
 | |
| 
 | |
| 	if (!HCL_OOP_IS_POINTER(oop)) return oop;
 | |
| 	if (HCL_OBJ_GET_FLAGS_NGC(oop)) return oop; /* non-GC object */
 | |
| 
 | |
| 	if (HCL_OBJ_GET_FLAGS_MOVED(oop))
 | |
| 	{
 | |
| 		/* this object has migrated to the new heap. 
 | |
| 		 * the class field has been updated to the new object
 | |
| 		 * in the 'else' block below. i can simply return it
 | |
| 		 * without further migration. */
 | |
| 		return HCL_OBJ_GET_CLASS(oop);
 | |
| 	}
 | |
| 	else
 | |
| 	{
 | |
| 		hcl_oow_t nbytes_aligned;
 | |
| 		hcl_oop_t tmp;
 | |
| 
 | |
| 		nbytes_aligned = get_payload_bytes (hcl, oop);
 | |
| 
 | |
| 		/* allocate space in the new heap */
 | |
| 		tmp = (hcl_oop_t)hcl_allocheapmem (hcl, hcl->newheap, HCL_SIZEOF(hcl_obj_t) + nbytes_aligned);
 | |
| 
 | |
| 		/* allocation here must not fail because
 | |
| 		 * i'm allocating the new space in a new heap for 
 | |
| 		 * moving an existing object in the current heap. 
 | |
| 		 *
 | |
| 		 * assuming the new heap is as large as the old heap,
 | |
| 		 * and garbage collection doesn't allocate more objects
 | |
| 		 * than in the old heap, it must not fail. */
 | |
| 		HCL_ASSERT (hcl, tmp != HCL_NULL); 
 | |
| 
 | |
| 		/* copy the payload to the new object */
 | |
| 		HCL_MEMCPY (tmp, oop, HCL_SIZEOF(hcl_obj_t) + nbytes_aligned);
 | |
| 
 | |
| 		/* mark the old object that it has migrated to the new heap */
 | |
| 		HCL_OBJ_SET_FLAGS_MOVED(oop, 1);
 | |
| 
 | |
| 		/* let the class field of the old object point to the new 
 | |
| 		 * object allocated in the new heap. it is returned in 
 | |
| 		 * the 'if' block at the top of this function. */
 | |
| 		HCL_OBJ_SET_CLASS (oop, tmp);
 | |
| 
 | |
| 		/* return the new object */
 | |
| 		return tmp;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static hcl_uint8_t* scan_new_heap (hcl_t* hcl, hcl_uint8_t* ptr)
 | |
| {
 | |
| 	while (ptr < hcl->newheap->ptr)
 | |
| 	{
 | |
| 		hcl_oow_t i;
 | |
| 		hcl_oow_t nbytes_aligned;
 | |
| 		hcl_oop_t oop;
 | |
| 
 | |
| 		oop = (hcl_oop_t)ptr;
 | |
| 
 | |
| 	#if defined(HCL_USE_OBJECT_TRAILER)
 | |
| 		if (HCL_OBJ_GET_FLAGS_TRAILER(oop))
 | |
| 		{
 | |
| 			hcl_oow_t nbytes;
 | |
| 
 | |
| 			HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_TYPE(oop) == HCL_OBJ_TYPE_OOP);
 | |
| 			HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_UNIT(oop) == HCL_SIZEOF(hcl_oow_t));
 | |
| 			HCL_ASSERT (hcl, HCL_OBJ_GET_FLAGS_EXTRA(oop) == 0); /* no 'extra' for an OOP object */
 | |
| 
 | |
| 			nbytes = HCL_OBJ_BYTESOF(oop) + HCL_SIZEOF(hcl_oow_t) + \
 | |
| 			         (hcl_oow_t)((hcl_oop_oop_t)oop)->slot[HCL_OBJ_GET_SIZE(oop)];
 | |
| 			nbytes_aligned = HCL_ALIGN (nbytes, HCL_SIZEOF(hcl_oop_t));
 | |
| 		}
 | |
| 		else
 | |
| 		{
 | |
| 	#endif
 | |
| 			nbytes_aligned = HCL_ALIGN (HCL_OBJ_BYTESOF(oop), HCL_SIZEOF(hcl_oop_t));
 | |
| 	#if defined(HCL_USE_OBJECT_TRAILER)
 | |
| 		}
 | |
| 	#endif
 | |
| 
 | |
| 		HCL_OBJ_SET_CLASS (oop, hcl_moveoop(hcl, HCL_OBJ_GET_CLASS(oop)));
 | |
| 		if (HCL_OBJ_GET_FLAGS_TYPE(oop) == HCL_OBJ_TYPE_OOP)
 | |
| 		{
 | |
| 			hcl_oop_oop_t xtmp;
 | |
| 			hcl_oow_t size;
 | |
| 
 | |
| 			if (HCL_OBJ_GET_FLAGS_BRAND(oop) == HCL_BRAND_PROCESS)
 | |
| 			{
 | |
| 				/* the stack in a process object doesn't need to be 
 | |
| 				 * scanned in full. the slots above the stack pointer 
 | |
| 				 * are garbages. */
 | |
| 				size = HCL_PROCESS_NAMED_INSTVARS +
 | |
| 				       HCL_OOP_TO_SMOOI(((hcl_oop_process_t)oop)->sp) + 1;
 | |
| 				HCL_ASSERT (hcl, size <= HCL_OBJ_GET_SIZE(oop));
 | |
| 			}
 | |
| 			else
 | |
| 			{
 | |
| 				size = HCL_OBJ_GET_SIZE(oop);
 | |
| 			}
 | |
| 
 | |
| 			xtmp = (hcl_oop_oop_t)oop;
 | |
| 			for (i = 0; i < size; i++)
 | |
| 			{
 | |
| 				if (HCL_OOP_IS_POINTER(xtmp->slot[i]))
 | |
| 					xtmp->slot[i] = hcl_moveoop (hcl, xtmp->slot[i]);
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		ptr = ptr + HCL_SIZEOF(hcl_obj_t) + nbytes_aligned;
 | |
| 	}
 | |
| 
 | |
| 	/* return the pointer to the beginning of the free space in the heap */
 | |
| 	return ptr; 
 | |
| }
 | |
| 
 | |
| void hcl_gc (hcl_t* hcl)
 | |
| {
 | |
| 	/* 
 | |
| 	 * move a referenced object to the new heap.
 | |
| 	 * inspect the fields of the moved object in the new heap.
 | |
| 	 * move objects pointed to by the fields to the new heap.
 | |
| 	 * finally perform some tricky symbol table clean-up.
 | |
| 	 */
 | |
| 	hcl_uint8_t* ptr;
 | |
| 	hcl_heap_t* tmp;
 | |
| 	hcl_oop_t old_nil;
 | |
| 	hcl_oow_t i;
 | |
| 	hcl_cb_t* cb;
 | |
| 
 | |
| 	if (hcl->active_context)
 | |
| 	{
 | |
| 		HCL_ASSERT (hcl, (hcl_oop_t)hcl->processor != hcl->_nil);
 | |
| 		HCL_ASSERT (hcl, (hcl_oop_t)hcl->processor->active != hcl->_nil);
 | |
| 		/* store the stack pointer to the active process */
 | |
| 		hcl->processor->active->sp = HCL_SMOOI_TO_OOP(hcl->sp);
 | |
| 
 | |
| 		/* store the instruction pointer to the active context */
 | |
| 		hcl->active_context->ip = HCL_SMOOI_TO_OOP(hcl->ip);
 | |
| 	}
 | |
| 
 | |
| 	HCL_LOG4 (hcl, HCL_LOG_GC | HCL_LOG_INFO, 
 | |
| 		"Starting GC curheap base %p ptr %p newheap base %p ptr %p\n",
 | |
| 		hcl->curheap->base, hcl->curheap->ptr, hcl->newheap->base, hcl->newheap->ptr); 
 | |
| 
 | |
| 	/* TODO: allocate common objects like _nil and the root dictionary 
 | |
| 	 *       in the permanant heap.  minimize moving around */
 | |
| 	old_nil = hcl->_nil;
 | |
| 
 | |
| 	/* move _nil and the root object table */
 | |
| 	hcl->_nil               = hcl_moveoop(hcl, hcl->_nil);
 | |
| 	hcl->_true              = hcl_moveoop(hcl, hcl->_true);
 | |
| 	hcl->_false             = hcl_moveoop(hcl, hcl->_false);
 | |
| 
 | |
| 	for (i = 0; i < HCL_COUNTOF(syminfo); i++)
 | |
| 	{
 | |
| 		hcl_oop_t tmp;
 | |
| 		tmp = *(hcl_oop_t*)((hcl_uint8_t*)hcl + syminfo[i].offset);
 | |
| 		tmp = hcl_moveoop(hcl, tmp);
 | |
| 		*(hcl_oop_t*)((hcl_uint8_t*)hcl + syminfo[i].offset) = tmp;
 | |
| 	}
 | |
| 
 | |
| 	hcl->sysdic = (hcl_oop_dic_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->sysdic);
 | |
| 	hcl->processor = (hcl_oop_process_scheduler_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->processor);
 | |
| 	hcl->nil_process = (hcl_oop_process_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->nil_process);
 | |
| 
 | |
| 	for (i = 0; i < hcl->code.lit.len; i++)
 | |
| 	{
 | |
| 		/* the literal array ia a NGC object. but the literal objects 
 | |
| 		 * pointed by the elements of this array must be gabage-collected. */
 | |
| 		((hcl_oop_oop_t)hcl->code.lit.arr)->slot[i] =
 | |
| 			hcl_moveoop(hcl, ((hcl_oop_oop_t)hcl->code.lit.arr)->slot[i]);
 | |
| 	}
 | |
| 
 | |
| 	hcl->p.e = hcl_moveoop (hcl, hcl->p.e);
 | |
| 
 | |
| 	for (i = 0; i < hcl->sem_list_count; i++)
 | |
| 	{
 | |
| 		hcl->sem_list[i] = (hcl_oop_semaphore_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->sem_list[i]);
 | |
| 	}
 | |
| 
 | |
| 	for (i = 0; i < hcl->sem_heap_count; i++)
 | |
| 	{
 | |
| 		hcl->sem_heap[i] = (hcl_oop_semaphore_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->sem_heap[i]);
 | |
| 	}
 | |
| 
 | |
| 	for (i = 0; i < hcl->tmp_count; i++)
 | |
| 	{
 | |
| 		*hcl->tmp_stack[i] = hcl_moveoop(hcl, *hcl->tmp_stack[i]);
 | |
| 	}
 | |
| 
 | |
| 	if (hcl->initial_context)
 | |
| 		hcl->initial_context = (hcl_oop_context_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->initial_context);
 | |
| 	if (hcl->active_context)
 | |
| 		hcl->active_context = (hcl_oop_context_t)hcl_moveoop(hcl, (hcl_oop_t)hcl->active_context);
 | |
| 
 | |
| 	if (hcl->last_retv) hcl->last_retv = hcl_moveoop(hcl, hcl->last_retv);
 | |
| 
 | |
| 	for (cb = hcl->cblist; cb; cb = cb->next)
 | |
| 	{
 | |
| 		if (cb->gc) cb->gc (hcl);
 | |
| 	}
 | |
| 
 | |
| 	/* scan the new heap to move referenced objects */
 | |
| 	ptr = (hcl_uint8_t*) HCL_ALIGN ((hcl_uintptr_t)hcl->newheap->base, HCL_SIZEOF(hcl_oop_t));
 | |
| 	ptr = scan_new_heap (hcl, ptr);
 | |
| 
 | |
| 	/* traverse the symbol table for unreferenced symbols.
 | |
| 	 * if the symbol has not moved to the new heap, the symbol
 | |
| 	 * is not referenced by any other objects than the symbol 
 | |
| 	 * table itself */
 | |
| 	compact_symbol_table (hcl, old_nil);
 | |
| 
 | |
| 	/* move the symbol table itself */
 | |
| 	hcl->symtab = (hcl_oop_dic_t)hcl_moveoop (hcl, (hcl_oop_t)hcl->symtab);
 | |
| 
 | |
| 	/* scan the new heap again from the end position of
 | |
| 	 * the previous scan to move referenced objects by 
 | |
| 	 * the symbol table. */
 | |
| 	ptr = scan_new_heap (hcl, ptr);
 | |
| 
 | |
| 	/* the contents of the current heap is not needed any more.
 | |
| 	 * reset the upper bound to the base. don't forget to align the heap
 | |
| 	 * pointer to the OOP size. See hcl_makeheap() also */
 | |
| 	hcl->curheap->ptr = (hcl_uint8_t*)HCL_ALIGN(((hcl_uintptr_t)hcl->curheap->base), HCL_SIZEOF(hcl_oop_t));
 | |
| 
 | |
| 	/* swap the current heap and old heap */
 | |
| 	tmp = hcl->curheap;
 | |
| 	hcl->curheap = hcl->newheap;
 | |
| 	hcl->newheap = tmp;
 | |
| 
 | |
| 
 | |
| /*
 | |
| 	if (hcl->symtab && HCL_LOG_ENABLED(hcl, HCL_LOG_GC | HCL_LOG_DEBUG))
 | |
| 	{
 | |
| 		hcl_oow_t index;
 | |
| 		hcl_oop_oop_t buc;
 | |
| 		HCL_LOG0 (hcl, HCL_LOG_GC | HCL_LOG_DEBUG, "--------- SURVIVING SYMBOLS IN GC ----------\n");
 | |
| 		buc = (hcl_oop_oop_t) hcl->symtab->bucket;
 | |
| 		for (index = 0; index < HCL_OBJ_GET_SIZE(buc); index++)
 | |
| 		{
 | |
| 			if ((hcl_oop_t)buc->slot[index] != hcl->_nil) 
 | |
| 			{
 | |
| 				HCL_LOG1 (hcl, HCL_LOG_GC | HCL_LOG_DEBUG, "\t%O\n", buc->slot[index]);
 | |
| 			}
 | |
| 		}
 | |
| 		HCL_LOG0 (hcl, HCL_LOG_GC | HCL_LOG_DEBUG, "--------------------------------------------\n");
 | |
| 	}
 | |
| */
 | |
| 
 | |
| /* TODO: include some gc statstics like number of live objects, gc performance, etc */
 | |
| 	HCL_LOG4 (hcl, HCL_LOG_GC | HCL_LOG_INFO, 
 | |
| 		"Finished GC curheap base %p ptr %p newheap base %p ptr %p\n",
 | |
| 		hcl->curheap->base, hcl->curheap->ptr, hcl->newheap->base, hcl->newheap->ptr); 
 | |
| }
 | |
| 
 | |
| 
 | |
| void hcl_pushtmp (hcl_t* hcl, hcl_oop_t* oop_ptr)
 | |
| {
 | |
| 	/* if you have too many temporaries pushed, something must be wrong.
 | |
| 	 * change your code not to exceede the stack limit */
 | |
| 	HCL_ASSERT (hcl, hcl->tmp_count < HCL_COUNTOF(hcl->tmp_stack));
 | |
| 	hcl->tmp_stack[hcl->tmp_count++] = oop_ptr;
 | |
| }
 | |
| 
 | |
| void hcl_poptmp (hcl_t* hcl)
 | |
| {
 | |
| 	HCL_ASSERT (hcl, hcl->tmp_count > 0);
 | |
| 	hcl->tmp_count--;
 | |
| }
 | |
| 
 | |
| void hcl_poptmps (hcl_t* hcl, hcl_oow_t count)
 | |
| {
 | |
| 	HCL_ASSERT (hcl, hcl->tmp_count >= count);
 | |
| 	hcl->tmp_count -= count;
 | |
| }
 | |
| 
 | |
| 
 | |
| hcl_oop_t hcl_shallowcopy (hcl_t* hcl, hcl_oop_t oop)
 | |
| {
 | |
| 	if (HCL_OOP_IS_POINTER(oop) && HCL_OBJ_GET_FLAGS_BRAND(oop) != HCL_BRAND_SYMBOL)
 | |
| 	{
 | |
| 		hcl_oop_t z;
 | |
| 		hcl_oow_t total_bytes;
 | |
| 
 | |
| 		total_bytes = HCL_SIZEOF(hcl_obj_t) + get_payload_bytes(hcl, oop);
 | |
| 
 | |
| 		hcl_pushtmp (hcl, &oop);
 | |
| 		z = (hcl_oop_t)hcl_allocbytes (hcl, total_bytes);
 | |
| 		hcl_poptmp(hcl);
 | |
| 
 | |
| 		HCL_MEMCPY (z, oop, total_bytes);
 | |
| 		return z;
 | |
| 	}
 | |
| 
 | |
| 	return oop;
 | |
| }
 | |
| 
 | |
| /* ========================================================================= */
 | |
| 
 | |
| 
 | |
| int hcl_ignite (hcl_t* hcl)
 | |
| {
 | |
| 	hcl_oow_t i;
 | |
| 
 | |
| 	if (!hcl->_nil) 
 | |
| 	{
 | |
| 		hcl->_nil = hcl_makenil (hcl);
 | |
| 		if (!hcl->_nil) return -1;
 | |
| 	}
 | |
| 
 | |
| 	if (!hcl->_true) 
 | |
| 	{
 | |
| 		hcl->_true = hcl_maketrue (hcl);
 | |
| 		if (!hcl->_true) return -1;
 | |
| 	}
 | |
| 	if (!hcl->_false)
 | |
| 	{
 | |
| 		hcl->_false = hcl_makefalse (hcl);
 | |
| 		if (!hcl->_false) return -1;
 | |
| 	}
 | |
| 
 | |
| 
 | |
| 	if (!hcl->symtab) 
 | |
| 	{
 | |
| 		hcl->symtab = (hcl_oop_dic_t)hcl_makedic (hcl, hcl->option.dfl_symtab_size);
 | |
| 		if (!hcl->symtab) return -1;
 | |
| 	}
 | |
| 
 | |
| 	if (!hcl->sysdic)
 | |
| 	{
 | |
| 		hcl->sysdic = (hcl_oop_dic_t)hcl_makedic (hcl, hcl->option.dfl_sysdic_size);
 | |
| 		if (!hcl->sysdic) return -1;
 | |
| 	}
 | |
| 
 | |
| 	/* symbol table available now. symbols can be created */
 | |
| 	for (i = 0; i < HCL_COUNTOF(syminfo); i++)
 | |
| 	{
 | |
| 		hcl_oop_t tmp;
 | |
| 
 | |
| 		tmp = hcl_makesymbol (hcl, syminfo[i].ptr, syminfo[i].len);
 | |
| 		if (!tmp) return -1;
 | |
| 
 | |
| 		HCL_OBJ_SET_FLAGS_SYNCODE (tmp, syminfo[i].syncode);
 | |
| 		*(hcl_oop_t*)((hcl_uint8_t*)hcl + syminfo[i].offset) = tmp;
 | |
| 	}
 | |
| 
 | |
| 	if (!hcl->nil_process)
 | |
| 	{
 | |
| 		/* Create a nil process used to simplify nil check in GC.
 | |
| 		 * only accessible by VM. not exported via the global dictionary. */
 | |
| 		hcl->nil_process = (hcl_oop_process_t)hcl_allocoopobj(hcl, HCL_BRAND_PROCESS, HCL_PROCESS_NAMED_INSTVARS);
 | |
| 		if (!hcl->nil_process) return -1;
 | |
| 		hcl->nil_process->sp = HCL_SMOOI_TO_OOP(-1);
 | |
| 	}
 | |
| 
 | |
| 	if (!hcl->processor)
 | |
| 	{
 | |
| 		hcl->processor = (hcl_oop_process_scheduler_t)hcl_allocoopobj(hcl, HCL_BRAND_PROCESS_SCHEDULER, HCL_PROCESS_SCHEDULER_NAMED_INSTVARS);
 | |
| 		if (!hcl->processor) return -1;
 | |
| 		hcl->processor->tally = HCL_SMOOI_TO_OOP(0);
 | |
| 		hcl->processor->active = hcl->nil_process;
 | |
| 	}
 | |
| 
 | |
| 	if (!hcl->code.bc.arr)
 | |
| 	{
 | |
| 		hcl->code.bc.arr = (hcl_oop_byte_t)hcl_makengcbytearray(hcl, HCL_NULL, HCL_BC_BUFFER_INIT); /* TODO: set a proper intial size */
 | |
| 		if (!hcl->code.bc.arr) return -1;
 | |
| 		HCL_ASSERT (hcl, hcl->code.bc.len == 0);
 | |
| 	}
 | |
| 
 | |
| 	if (!hcl->code.lit.arr)
 | |
| 	{
 | |
| 		hcl->code.lit.arr = (hcl_oop_oop_t)hcl_makengcarray(hcl, HCL_LIT_BUFFER_INIT); /* TOOD: set a proper initial size */
 | |
| 		if (!hcl->code.lit.arr) return -1;
 | |
| 		HCL_ASSERT (hcl, hcl->code.lit.len == 0);
 | |
| 	}
 | |
| 
 | |
| 	hcl->p.e = hcl->_nil;
 | |
| 
 | |
| 	return 0;
 | |
| }
 |