X-Git-Url: http://repo.macrolet.net/gitweb/?a=blobdiff_plain;f=src%2Fruntime%2Fgencgc.c;h=a5f1dc07196cc81ab2284850681dce13ac50e171;hb=9e58342e3ebb8bbaaa648cd345232b08279f3a0c;hp=27ad418732f8b2abf4e81566e0cd974fcd467662;hpb=a7a9b1029e8b9e45a5b66d62e161cc476cb7b60c;p=sbcl.git diff --git a/src/runtime/gencgc.c b/src/runtime/gencgc.c index 27ad418..a5f1dc0 100644 --- a/src/runtime/gencgc.c +++ b/src/runtime/gencgc.c @@ -57,7 +57,7 @@ #endif /* forward declarations */ -page_index_t gc_find_freeish_pages(long *restart_page_ptr, long nbytes, +page_index_t gc_find_freeish_pages(page_index_t *restart_page_ptr, long nbytes, int page_type_flag); @@ -149,8 +149,8 @@ boolean gencgc_partial_pickup = 0; */ /* the total bytes allocated. These are seen by Lisp DYNAMIC-USAGE. */ -unsigned long bytes_allocated = 0; -unsigned long auto_gc_trigger = 0; +os_vm_size_t bytes_allocated = 0; +os_vm_size_t auto_gc_trigger = 0; /* the source and destination generations. These are set before a GC starts * scavenging. */ @@ -246,7 +246,7 @@ find_page_index(void *addr) } static size_t -npage_bytes(long npages) +npage_bytes(page_index_t npages) { gc_assert(npages>=0); return ((unsigned long)npages)*GENCGC_CARD_BYTES; @@ -285,13 +285,13 @@ struct generation { page_index_t alloc_large_unboxed_start_page; /* the bytes allocated to this generation */ - unsigned long bytes_allocated; + os_vm_size_t bytes_allocated; /* the number of bytes at which to trigger a GC */ - unsigned long gc_trigger; + os_vm_size_t gc_trigger; /* to calculate a new level for gc_trigger */ - unsigned long bytes_consed_between_gc; + os_vm_size_t bytes_consed_between_gc; /* the number of GCs since the last raise */ int num_gc; @@ -305,7 +305,7 @@ struct generation { * objects are added from a GC of a younger generation. Dividing by * the bytes_allocated will give the average age of the memory in * this generation since its last GC. */ - unsigned long cum_sum_bytes_allocated; + os_vm_size_t cum_sum_bytes_allocated; /* a minimum average memory age before a GC will occur helps * prevent a GC when a large number of new live objects have been @@ -363,11 +363,10 @@ unsigned long gencgc_alloc_granularity = GENCGC_ALLOC_GRANULARITY; /* Count the number of pages which are write-protected within the * given generation. */ -static long +static page_index_t count_write_protect_generation_pages(generation_index_t generation) { - page_index_t i; - unsigned long count = 0; + page_index_t i, count = 0; for (i = 0; i < last_free_page; i++) if (page_allocated_p(i) @@ -378,11 +377,11 @@ count_write_protect_generation_pages(generation_index_t generation) } /* Count the number of pages within the given generation. */ -static long +static page_index_t count_generation_pages(generation_index_t generation) { page_index_t i; - long count = 0; + page_index_t count = 0; for (i = 0; i < last_free_page; i++) if (page_allocated_p(i) @@ -392,11 +391,11 @@ count_generation_pages(generation_index_t generation) } #if QSHOW -static long +static page_index_t count_dont_move_pages(void) { page_index_t i; - long count = 0; + page_index_t count = 0; for (i = 0; i < last_free_page; i++) { if (page_allocated_p(i) && (page_table[i].dont_move != 0)) { @@ -457,11 +456,11 @@ write_generation_stats(FILE *file) for (i = 0; i < SCRATCH_GENERATION; i++) { page_index_t j; - long boxed_cnt = 0; - long unboxed_cnt = 0; - long large_boxed_cnt = 0; - long large_unboxed_cnt = 0; - long pinned_cnt=0; + page_index_t boxed_cnt = 0; + page_index_t unboxed_cnt = 0; + page_index_t large_boxed_cnt = 0; + page_index_t large_unboxed_cnt = 0; + page_index_t pinned_cnt=0; for (j = 0; j < last_free_page; j++) if (page_table[j].gen == i) { @@ -488,17 +487,19 @@ write_generation_stats(FILE *file) gc_assert(generations[i].bytes_allocated == count_generation_bytes_allocated(i)); fprintf(file, - " %1d: %5ld %5ld %5ld %5ld %5ld %5ld %5ld %5ld %5ld %8ld %5ld %8ld %4ld %3d %7.4f\n", + " %1d: %5ld %5ld %5ld %5ld", i, generations[i].alloc_start_page, generations[i].alloc_unboxed_start_page, generations[i].alloc_large_start_page, - generations[i].alloc_large_unboxed_start_page, - boxed_cnt, - unboxed_cnt, - large_boxed_cnt, - large_unboxed_cnt, - pinned_cnt, + generations[i].alloc_large_unboxed_start_page); + fprintf(file, + " %5"PAGE_INDEX_FMT" %5"PAGE_INDEX_FMT" %5"PAGE_INDEX_FMT + " %5"PAGE_INDEX_FMT" %5"PAGE_INDEX_FMT, + boxed_cnt, unboxed_cnt, large_boxed_cnt, + large_unboxed_cnt, pinned_cnt); + fprintf(file + " %8ld %5ld %8ld %4ld %3d %7.4f\n", generations[i].bytes_allocated, (npage_bytes(count_generation_pages(i)) - generations[i].bytes_allocated), @@ -507,7 +508,7 @@ write_generation_stats(FILE *file) generations[i].num_gc, generation_average_age(i)); } - fprintf(file," Total bytes allocated = %lu\n", bytes_allocated); + fprintf(file," Total bytes allocated = %lu\n", (unsigned long)bytes_allocated); fprintf(file," Dynamic-space-size bytes = %lu\n", (unsigned long)dynamic_space_size); fpu_restore(fpu_state); @@ -587,7 +588,7 @@ void fast_bzero(void*, size_t); /* in -assem.S */ * OS. Generally done after a large GC. */ void zero_pages_with_mmap(page_index_t start, page_index_t end) { - int i; + page_index_t i; void *addr = page_address(start), *new_addr; size_t length = npage_bytes(1+end-start); @@ -2536,7 +2537,7 @@ static void scavenge_generations(generation_index_t from, generation_index_t to) { page_index_t i; - int num_wp = 0; + page_index_t num_wp = 0; #define SC_GEN_CK 0 #if SC_GEN_CK @@ -2826,16 +2827,19 @@ scavenge_newspace_generation(generation_index_t generation) record_new_objects = 0; #if SC_NS_GEN_CK - /* Check that none of the write_protected pages in this generation - * have been written to. */ - for (i = 0; i < page_table_pages; i++) { - if (page_allocated_p(i) - && (page_table[i].bytes_used != 0) - && (page_table[i].gen == generation) - && (page_table[i].write_protected_cleared != 0) - && (page_table[i].dont_move == 0)) { - lose("write protected page %d written to in scavenge_newspace_generation\ngeneration=%d dont_move=%d\n", - i, generation, page_table[i].dont_move); + { + page_index_t i; + /* Check that none of the write_protected pages in this generation + * have been written to. */ + for (i = 0; i < page_table_pages; i++) { + if (page_allocated_p(i) + && (page_table[i].bytes_used != 0) + && (page_table[i].gen == generation) + && (page_table[i].write_protected_cleared != 0) + && (page_table[i].dont_move == 0)) { + lose("write protected page %d written to in scavenge_newspace_generation\ngeneration=%d dont_move=%d\n", + i, generation, page_table[i].dont_move); + } } } #endif @@ -3686,8 +3690,8 @@ garbage_collect_generation(generation_index_t generation, int raise) /* As a check re-scavenge the newspace once; no new objects should * be found. */ { - long old_bytes_allocated = bytes_allocated; - long bytes_allocated; + os_vm_size_t old_bytes_allocated = bytes_allocated; + os_vm_size_t bytes_allocated; /* Start with a full scavenge. */ scavenge_newspace_generation_one_scan(new_space); @@ -4068,6 +4072,12 @@ gc_init(void) page_table_pages = dynamic_space_size/GENCGC_CARD_BYTES; gc_assert(dynamic_space_size == npage_bytes(page_table_pages)); + /* Default nursery size to 5% of the total dynamic space size, + * min 1Mb. */ + bytes_consed_between_gcs = dynamic_space_size/(os_vm_size_t)20; + if (bytes_consed_between_gcs < (1024*1024)) + bytes_consed_between_gcs = 1024*1024; + /* The page_table must be allocated using "calloc" to initialize * the page structures correctly. There used to be a separate * initialization loop (now commented out; see below) but that was @@ -4130,7 +4140,7 @@ gc_init(void) generations[i].num_gc = 0; generations[i].cum_sum_bytes_allocated = 0; /* the tune-able parameters */ - generations[i].bytes_consed_between_gc = 2000000; + generations[i].bytes_consed_between_gc = bytes_consed_between_gcs; generations[i].number_of_gcs_before_promotion = 1; generations[i].minimum_age_before_gc = 0.75; }