void gc_free_heap(void);
-inline int find_page_index(void *);
-inline void *page_address(int);
+inline long find_page_index(void *);
+inline void *page_address(long);
int gencgc_handle_wp_violation(void *);
-lispobj *search_dynamic_space(lispobj *);
\f
struct page {
* than the actual bytes used for pages within the current
* allocation regions. It should be 0 for all unallocated pages (not
* hard to achieve). */
- int bytes_used;
+ long bytes_used;
- /* It is important to know the offset to the first object in the
- * page. Currently it's only important to know if an object starts
- * at the beginning of the page in which case the offset would be 0. */
- int first_object_offset;
+ /* The name of this field is not well-chosen for its actual use.
+ * This is the offset from the start of the page to the start
+ * of the alloc_region which contains/contained it. It's negative or 0
+ */
+ long first_object_offset;
};
/* values for the page.allocated field */
\f
/* the number of pages needed for the dynamic space - rounding up */
#define NUM_PAGES ((DYNAMIC_SPACE_SIZE+PAGE_BYTES-1)/PAGE_BYTES)
+
extern struct page page_table[NUM_PAGES];
\f
+/* forward declarations */
void sniff_code_object(struct code *code, unsigned displacement);
void gencgc_apply_code_fixups(struct code *old_code, struct code *new_code);
-int update_x86_dynamic_space_free_pointer(void);
+long update_x86_dynamic_space_free_pointer(void);
void gc_alloc_update_page_tables(int unboxed,
struct alloc_region *alloc_region);
+void gc_alloc_update_all_page_tables(void);
+void gc_set_region_empty(struct alloc_region *region);
+
/*
* predicates
*/
static inline int
-space_matches_p(lispobj obj, int space)
+space_matches_p(lispobj obj, long space)
{
- int page_index=(void*)obj - (void *)DYNAMIC_SPACE_START;
+ long page_index=(void*)obj - (void *)DYNAMIC_SPACE_START;
return ((page_index >= 0)
&& ((page_index =
- ((unsigned int)page_index)/PAGE_BYTES) < NUM_PAGES)
+ ((unsigned long)page_index)/PAGE_BYTES) < NUM_PAGES)
&& (page_table[page_index].gen == space));
}