version 1.1.1.1, 1999/12/03 07:39:09 |
version 1.3, 2000/12/01 09:26:10 |
|
|
/* |
/* |
* Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers |
* Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers |
* Copyright (c) 1991-1994 by Xerox Corporation. All rights reserved. |
* Copyright (c) 1991-1996 by Xerox Corporation. All rights reserved. |
* Copyright (c) 1998 by Silicon Graphics. All rights reserved. |
* Copyright (c) 1998 by Silicon Graphics. All rights reserved. |
|
* Copyright (c) 1999 by Hewlett-Packard Company. All rights reserved. |
* |
* |
* THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED |
* THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED |
* OR IMPLIED. ANY USE IS AT YOUR OWN RISK. |
* OR IMPLIED. ANY USE IS AT YOUR OWN RISK. |
Line 64 word GC_gc_no = 0; |
|
Line 65 word GC_gc_no = 0; |
|
int GC_incremental = 0; /* By default, stop the world. */ |
int GC_incremental = 0; /* By default, stop the world. */ |
#endif |
#endif |
|
|
int GC_full_freq = 4; /* Every 5th collection is a full */ |
int GC_full_freq = 19; /* Every 20th collection is a full */ |
/* collection. */ |
/* collection, whether we need it */ |
|
/* or not. */ |
|
|
|
GC_bool GC_need_full_gc = FALSE; |
|
/* Need full GC do to heap growth. */ |
|
|
|
word GC_used_heap_size_after_full = 0; |
|
|
char * GC_copyright[] = |
char * GC_copyright[] = |
{"Copyright 1988,1989 Hans-J. Boehm and Alan J. Demers ", |
{"Copyright 1988,1989 Hans-J. Boehm and Alan J. Demers ", |
"Copyright (c) 1991-1995 by Xerox Corporation. All rights reserved. ", |
"Copyright (c) 1991-1995 by Xerox Corporation. All rights reserved. ", |
Line 213 GC_bool GC_should_collect() |
|
Line 220 GC_bool GC_should_collect() |
|
return(GC_adj_words_allocd() >= min_words_allocd()); |
return(GC_adj_words_allocd() >= min_words_allocd()); |
} |
} |
|
|
|
|
void GC_notify_full_gc() |
void GC_notify_full_gc() |
{ |
{ |
if (GC_start_call_back != (void (*)())0) { |
if (GC_start_call_back != (void (*)())0) { |
Line 220 void GC_notify_full_gc() |
|
Line 228 void GC_notify_full_gc() |
|
} |
} |
} |
} |
|
|
|
GC_bool GC_is_full_gc = FALSE; |
|
|
/* |
/* |
* Initiate a garbage collection if appropriate. |
* Initiate a garbage collection if appropriate. |
* Choose judiciously |
* Choose judiciously |
Line 229 void GC_notify_full_gc() |
|
Line 239 void GC_notify_full_gc() |
|
void GC_maybe_gc() |
void GC_maybe_gc() |
{ |
{ |
static int n_partial_gcs = 0; |
static int n_partial_gcs = 0; |
GC_bool is_full_gc = FALSE; |
|
|
|
if (GC_should_collect()) { |
if (GC_should_collect()) { |
if (!GC_incremental) { |
if (!GC_incremental) { |
Line 237 void GC_maybe_gc() |
|
Line 246 void GC_maybe_gc() |
|
GC_gcollect_inner(); |
GC_gcollect_inner(); |
n_partial_gcs = 0; |
n_partial_gcs = 0; |
return; |
return; |
} else if (n_partial_gcs >= GC_full_freq) { |
} else if (GC_need_full_gc || n_partial_gcs >= GC_full_freq) { |
# ifdef PRINTSTATS |
# ifdef PRINTSTATS |
GC_printf2( |
GC_printf2( |
"***>Full mark for collection %lu after %ld allocd bytes\n", |
"***>Full mark for collection %lu after %ld allocd bytes\n", |
Line 249 void GC_maybe_gc() |
|
Line 258 void GC_maybe_gc() |
|
GC_clear_marks(); |
GC_clear_marks(); |
n_partial_gcs = 0; |
n_partial_gcs = 0; |
GC_notify_full_gc(); |
GC_notify_full_gc(); |
is_full_gc = TRUE; |
GC_is_full_gc = TRUE; |
} else { |
} else { |
n_partial_gcs++; |
n_partial_gcs++; |
} |
} |
Line 263 void GC_maybe_gc() |
|
Line 272 void GC_maybe_gc() |
|
# endif |
# endif |
GC_finish_collection(); |
GC_finish_collection(); |
} else { |
} else { |
if (!is_full_gc) { |
if (!GC_is_full_gc) { |
/* Count this as the first attempt */ |
/* Count this as the first attempt */ |
GC_n_attempts++; |
GC_n_attempts++; |
} |
} |
Line 300 GC_stop_func stop_func; |
|
Line 309 GC_stop_func stop_func; |
|
/* Make sure all blocks have been reclaimed, so sweep routines */ |
/* Make sure all blocks have been reclaimed, so sweep routines */ |
/* don't see cleared mark bits. */ |
/* don't see cleared mark bits. */ |
/* If we're guaranteed to finish, then this is unnecessary. */ |
/* If we're guaranteed to finish, then this is unnecessary. */ |
if (stop_func != GC_never_stop_func |
/* In the find_leak case, we have to finish to guarantee that */ |
|
/* previously unmarked objects are not reported as leaks. */ |
|
if ((GC_find_leak || stop_func != GC_never_stop_func) |
&& !GC_reclaim_all(stop_func, FALSE)) { |
&& !GC_reclaim_all(stop_func, FALSE)) { |
/* Aborted. So far everything is still consistent. */ |
/* Aborted. So far everything is still consistent. */ |
return(FALSE); |
return(FALSE); |
Line 310 GC_stop_func stop_func; |
|
Line 321 GC_stop_func stop_func; |
|
# ifdef SAVE_CALL_CHAIN |
# ifdef SAVE_CALL_CHAIN |
GC_save_callers(GC_last_stack); |
GC_save_callers(GC_last_stack); |
# endif |
# endif |
|
GC_is_full_gc = TRUE; |
if (!GC_stopped_mark(stop_func)) { |
if (!GC_stopped_mark(stop_func)) { |
if (!GC_incremental) { |
if (!GC_incremental) { |
/* We're partially done and have no way to complete or use */ |
/* We're partially done and have no way to complete or use */ |
Line 481 void GC_finish_collection() |
|
Line 493 void GC_finish_collection() |
|
# ifdef GATHERSTATS |
# ifdef GATHERSTATS |
GC_mem_found = 0; |
GC_mem_found = 0; |
# endif |
# endif |
# ifdef FIND_LEAK |
if (GC_find_leak) { |
/* Mark all objects on the free list. All objects should be */ |
/* Mark all objects on the free list. All objects should be */ |
/* marked when we're done. */ |
/* marked when we're done. */ |
{ |
{ |
Line 504 void GC_finish_collection() |
|
Line 516 void GC_finish_collection() |
|
} |
} |
} |
} |
} |
} |
/* Check that everything is marked */ |
|
GC_start_reclaim(TRUE); |
GC_start_reclaim(TRUE); |
# else |
/* The above just checks; it doesn't really reclaim anything. */ |
|
} |
|
|
GC_finalize(); |
GC_finalize(); |
# ifdef STUBBORN_ALLOC |
# ifdef STUBBORN_ALLOC |
GC_clean_changing_list(); |
GC_clean_changing_list(); |
# endif |
# endif |
|
|
# ifdef PRINTTIMES |
# ifdef PRINTTIMES |
GET_TIME(finalize_time); |
GET_TIME(finalize_time); |
# endif |
# endif |
|
|
/* Clear free list mark bits, in case they got accidentally marked */ |
/* Clear free list mark bits, in case they got accidentally marked */ |
/* Note: HBLKPTR(p) == pointer to head of block containing *p */ |
/* Note: HBLKPTR(p) == pointer to head of block containing *p */ |
/* Also subtract memory remaining from GC_mem_found count. */ |
/* (or GC_find_leak is set and they were intentionally marked.) */ |
/* Note that composite objects on free list are cleared. */ |
/* Also subtract memory remaining from GC_mem_found count. */ |
/* Thus accidentally marking a free list is not a problem; only */ |
/* Note that composite objects on free list are cleared. */ |
/* objects on the list itself will be marked, and that's fixed here. */ |
/* Thus accidentally marking a free list is not a problem; only */ |
|
/* objects on the list itself will be marked, and that's fixed here. */ |
{ |
{ |
register word size; /* current object size */ |
register word size; /* current object size */ |
register ptr_t p; /* pointer to current object */ |
register ptr_t p; /* pointer to current object */ |
Line 548 void GC_finish_collection() |
|
Line 561 void GC_finish_collection() |
|
} |
} |
|
|
|
|
# ifdef PRINTSTATS |
# ifdef PRINTSTATS |
GC_printf1("Bytes recovered before sweep - f.l. count = %ld\n", |
GC_printf1("Bytes recovered before sweep - f.l. count = %ld\n", |
(long)WORDS_TO_BYTES(GC_mem_found)); |
(long)WORDS_TO_BYTES(GC_mem_found)); |
# endif |
# endif |
|
|
/* Reconstruct free lists to contain everything not marked */ |
/* Reconstruct free lists to contain everything not marked */ |
GC_start_reclaim(FALSE); |
GC_start_reclaim(FALSE); |
|
if (GC_is_full_gc) { |
# endif /* !FIND_LEAK */ |
GC_used_heap_size_after_full = USED_HEAP_SIZE; |
|
GC_need_full_gc = FALSE; |
|
} else { |
|
GC_need_full_gc = |
|
BYTES_TO_WORDS(USED_HEAP_SIZE - GC_used_heap_size_after_full) |
|
> min_words_allocd(); |
|
} |
|
|
# ifdef PRINTSTATS |
# ifdef PRINTSTATS |
GC_printf2( |
GC_printf2( |
Line 573 void GC_finish_collection() |
|
Line 591 void GC_finish_collection() |
|
# endif |
# endif |
|
|
GC_n_attempts = 0; |
GC_n_attempts = 0; |
|
GC_is_full_gc = FALSE; |
/* Reset or increment counters for next cycle */ |
/* Reset or increment counters for next cycle */ |
GC_words_allocd_before_gc += GC_words_allocd; |
GC_words_allocd_before_gc += GC_words_allocd; |
GC_non_gc_bytes_at_gc = GC_non_gc_bytes; |
GC_non_gc_bytes_at_gc = GC_non_gc_bytes; |
|
|
if (GC_n_heap_sects >= MAX_HEAP_SECTS) { |
if (GC_n_heap_sects >= MAX_HEAP_SECTS) { |
ABORT("Too many heap sections: Increase MAXHINCR or MAX_HEAP_SECTS"); |
ABORT("Too many heap sections: Increase MAXHINCR or MAX_HEAP_SECTS"); |
} |
} |
if (!GC_install_header(p)) { |
phdr = GC_install_header(p); |
|
if (0 == phdr) { |
/* This is extremely unlikely. Can't add it. This will */ |
/* This is extremely unlikely. Can't add it. This will */ |
/* almost certainly result in a 0 return from the allocator, */ |
/* almost certainly result in a 0 return from the allocator, */ |
/* which is entirely appropriate. */ |
/* which is entirely appropriate. */ |
|
|
GC_heap_sects[GC_n_heap_sects].hs_bytes = bytes; |
GC_heap_sects[GC_n_heap_sects].hs_bytes = bytes; |
GC_n_heap_sects++; |
GC_n_heap_sects++; |
words = BYTES_TO_WORDS(bytes - HDR_BYTES); |
words = BYTES_TO_WORDS(bytes - HDR_BYTES); |
phdr = HDR(p); |
|
phdr -> hb_sz = words; |
phdr -> hb_sz = words; |
phdr -> hb_map = (char *)1; /* A value != GC_invalid_map */ |
phdr -> hb_map = (char *)1; /* A value != GC_invalid_map */ |
phdr -> hb_flags = 0; |
phdr -> hb_flags = 0; |
|
|
} |
} |
} |
} |
|
|
#ifdef PRESERVE_LAST |
|
|
|
GC_bool GC_protect_last_block = FALSE; |
|
|
|
GC_bool GC_in_last_heap_sect(p) |
|
ptr_t p; |
|
{ |
|
struct HeapSect * last_heap_sect; |
|
ptr_t start; |
|
ptr_t end; |
|
|
|
if (!GC_protect_last_block) return FALSE; |
|
last_heap_sect = &(GC_heap_sects[GC_n_heap_sects-1]); |
|
start = last_heap_sect -> hs_start; |
|
if (p < start) return FALSE; |
|
end = start + last_heap_sect -> hs_bytes; |
|
if (p >= end) return FALSE; |
|
return TRUE; |
|
} |
|
#endif |
|
|
|
# if !defined(NO_DEBUGGING) |
# if !defined(NO_DEBUGGING) |
void GC_print_heap_sects() |
void GC_print_heap_sects() |
{ |
{ |
|
|
LOCK(); |
LOCK(); |
if (!GC_is_initialized) GC_init_inner(); |
if (!GC_is_initialized) GC_init_inner(); |
result = (int)GC_expand_hp_inner(divHBLKSZ((word)bytes)); |
result = (int)GC_expand_hp_inner(divHBLKSZ((word)bytes)); |
# ifdef PRESERVE_LAST |
if (result) GC_requested_heapsize += bytes; |
if (result) GC_protect_last_block = FALSE; |
|
# endif |
|
UNLOCK(); |
UNLOCK(); |
ENABLE_SIGNALS(); |
ENABLE_SIGNALS(); |
return(result); |
return(result); |
Line 839 GC_bool GC_collect_or_expand(needed_blocks, ignore_off |
|
Line 835 GC_bool GC_collect_or_expand(needed_blocks, ignore_off |
|
word needed_blocks; |
word needed_blocks; |
GC_bool ignore_off_page; |
GC_bool ignore_off_page; |
{ |
{ |
if (!GC_incremental && !GC_dont_gc && GC_should_collect()) { |
if (!GC_incremental && !GC_dont_gc && |
|
(GC_dont_expand && GC_words_allocd > 0 || GC_should_collect())) { |
GC_notify_full_gc(); |
GC_notify_full_gc(); |
GC_gcollect_inner(); |
GC_gcollect_inner(); |
} else { |
} else { |
Line 877 GC_bool ignore_off_page; |
|
Line 874 GC_bool ignore_off_page; |
|
GC_printf0("Memory available again ...\n"); |
GC_printf0("Memory available again ...\n"); |
} |
} |
# endif |
# endif |
# ifdef PRESERVE_LAST |
|
if (needed_blocks > 1) GC_protect_last_block = TRUE; |
|
/* We were forced to expand the heap as the result */ |
|
/* of a large block allocation. Avoid breaking up */ |
|
/* new block into small pieces. */ |
|
# endif |
|
} |
} |
} |
} |
return(TRUE); |
return(TRUE); |