X-Git-Url: http://repo.macrolet.net/gitweb/?a=blobdiff_plain;f=src%2Fruntime%2Fthread.c;h=5b8ec4ac475f9b4c8ed26a6f38fa97ee411423af;hb=9a19ce460a70a6c1de36095b3e2621116b91cc80;hp=5ed27c7e17d5a2d2d55f88f202e72723cafd664d;hpb=b6537fc9d37ad800f8faba89ebbde7fdf8910d2a;p=sbcl.git diff --git a/src/runtime/thread.c b/src/runtime/thread.c index 5ed27c7..5b8ec4a 100644 --- a/src/runtime/thread.c +++ b/src/runtime/thread.c @@ -25,6 +25,12 @@ #include #endif +#ifdef LISP_FEATURE_MACH_EXCEPTION_HANDLER +#include +#include +#include +#endif + #include "runtime.h" #include "validate.h" /* for CONTROL_STACK_SIZE etc */ #include "alloc.h" @@ -49,6 +55,7 @@ #if defined(LISP_FEATURE_DARWIN) && defined(LISP_FEATURE_SB_THREAD) #define QUEUE_FREEABLE_THREAD_STACKS +#define LOCK_CREATE_THREAD #endif #ifdef LISP_FEATURE_FREEBSD @@ -63,7 +70,7 @@ struct freeable_stack { struct freeable_stack *next; #endif os_thread_t os_thread; - os_vm_address_t stack; + os_vm_address_t os_address; }; @@ -84,6 +91,9 @@ pthread_mutex_t all_threads_lock = PTHREAD_MUTEX_INITIALIZER; #ifdef LOCK_CREATE_THREAD static pthread_mutex_t create_thread_lock = PTHREAD_MUTEX_INITIALIZER; #endif +#ifdef LISP_FEATURE_GCC_TLS +__thread struct thread *current_thread; +#endif #endif #if defined(LISP_FEATURE_X86) || defined(LISP_FEATURE_X86_64) @@ -137,41 +147,37 @@ initial_thread_trampoline(struct thread *th) #define THREAD_STRUCT_SIZE (THREAD_CONTROL_STACK_SIZE + BINDING_STACK_SIZE + \ ALIEN_STACK_SIZE + dynamic_values_bytes + \ - 32 * SIGSTKSZ) + 32 * SIGSTKSZ + \ + BACKEND_PAGE_SIZE) #ifdef LISP_FEATURE_SB_THREAD #ifdef QUEUE_FREEABLE_THREAD_STACKS +static void queue_freeable_thread_stack(struct thread *thread_to_be_cleaned_up) { + struct freeable_stack *new_freeable_stack = 0; if (thread_to_be_cleaned_up) { + /* FIXME: os_validate is mmap -- for small things like these + * malloc would probably perform better. */ + new_freeable_stack = (struct freeable_stack *) + os_validate(0, sizeof(struct freeable_stack)); + new_freeable_stack->next = NULL; + new_freeable_stack->os_thread = thread_to_be_cleaned_up->os_thread; + new_freeable_stack->os_address = thread_to_be_cleaned_up->os_address; pthread_mutex_lock(&freeable_stack_lock); if (freeable_stack_queue) { - struct freeable_stack *new_freeable_stack = 0, *next; + struct freeable_stack *next; next = freeable_stack_queue; while (next->next) { next = next->next; } - new_freeable_stack = (struct freeable_stack *) - os_validate(0, sizeof(struct freeable_stack)); - new_freeable_stack->next = NULL; - new_freeable_stack->os_thread = thread_to_be_cleaned_up->os_thread; - new_freeable_stack->stack = (os_vm_address_t) - thread_to_be_cleaned_up->control_stack_start; next->next = new_freeable_stack; - freeable_stack_count++; } else { - struct freeable_stack *new_freeable_stack = 0; - new_freeable_stack = (struct freeable_stack *) - os_validate(0, sizeof(struct freeable_stack)); - new_freeable_stack->next = NULL; - new_freeable_stack->os_thread = thread_to_be_cleaned_up->os_thread; - new_freeable_stack->stack = (os_vm_address_t) - thread_to_be_cleaned_up->control_stack_start; freeable_stack_queue = new_freeable_stack; - freeable_stack_count++; } + freeable_stack_count++; pthread_mutex_unlock(&freeable_stack_lock); } } @@ -188,7 +194,7 @@ free_freeable_stacks() { freeable_stack_count--; gc_assert(pthread_join(old->os_thread, NULL) == 0); FSHOW((stderr, "freeing thread %x stack\n", old->os_thread)); - os_invalidate(old->stack, THREAD_STRUCT_SIZE); + os_invalidate(old->os_address, THREAD_STRUCT_SIZE); os_invalidate((os_vm_address_t)old, sizeof(struct freeable_stack)); pthread_mutex_unlock(&freeable_stack_lock); } @@ -206,7 +212,7 @@ cleanup_thread(void *arg) gc_assert(pthread_join(freeable->os_thread, NULL) == 0); FSHOW((stderr, "/cleaner thread(%p): free stack %p\n", self, freeable->stack)); - os_invalidate(freeable->stack, THREAD_STRUCT_SIZE); + os_invalidate(freeable->os_address, THREAD_STRUCT_SIZE); free(freeable); pthread_detach(self); @@ -225,14 +231,14 @@ create_cleanup_thread(struct thread *thread_to_be_cleaned_up) malloc(sizeof(struct freeable_stack)); gc_assert(freeable != NULL); freeable->os_thread = thread_to_be_cleaned_up->os_thread; - freeable->stack = - (os_vm_address_t) thread_to_be_cleaned_up->control_stack_start; + freeable->os_address = + (os_vm_address_t) thread_to_be_cleaned_up->os_address; result = pthread_create(&thread, NULL, cleanup_thread, freeable); gc_assert(result == 0); sched_yield(); } } - + #else static void free_thread_stack_later(struct thread *thread_to_be_cleaned_up) @@ -242,8 +248,8 @@ free_thread_stack_later(struct thread *thread_to_be_cleaned_up) new_freeable_stack = (struct freeable_stack *) os_validate(0, sizeof(struct freeable_stack)); new_freeable_stack->os_thread = thread_to_be_cleaned_up->os_thread; - new_freeable_stack->stack = (os_vm_address_t) - thread_to_be_cleaned_up->control_stack_start; + new_freeable_stack->os_address = (os_vm_address_t) + thread_to_be_cleaned_up->os_address; } new_freeable_stack = (struct freeable_stack *) swap_lispobjs((lispobj *)(void *)&freeable_stack, @@ -254,7 +260,7 @@ free_thread_stack_later(struct thread *thread_to_be_cleaned_up) * exists and the stack can be safely freed. This is sadly not * mandated by the pthread spec. */ gc_assert(pthread_join(new_freeable_stack->os_thread, NULL) == 0); - os_invalidate(new_freeable_stack->stack, THREAD_STRUCT_SIZE); + os_invalidate(new_freeable_stack->os_address, THREAD_STRUCT_SIZE); os_invalidate((os_vm_address_t) new_freeable_stack, sizeof(struct freeable_stack)); } @@ -312,6 +318,17 @@ new_thread_trampoline(struct thread *th) os_invalidate((os_vm_address_t)th->interrupt_data, (sizeof (struct interrupt_data))); +#ifdef LISP_FEATURE_MACH_EXCEPTION_HANDLER + FSHOW((stderr, "Deallocating mach port %x\n", THREAD_STRUCT_TO_EXCEPTION_PORT(th))); + mach_port_move_member(mach_task_self(), + THREAD_STRUCT_TO_EXCEPTION_PORT(th), + MACH_PORT_NULL); + mach_port_deallocate(mach_task_self(), + THREAD_STRUCT_TO_EXCEPTION_PORT(th)); + mach_port_destroy(mach_task_self(), + THREAD_STRUCT_TO_EXCEPTION_PORT(th)); +#endif + #ifdef QUEUE_FREEABLE_THREAD_STACKS queue_freeable_thread_stack(th); #elif defined(CREATE_CLEANUP_THREAD) @@ -332,7 +349,7 @@ free_thread_struct(struct thread *th) if (th->interrupt_data) os_invalidate((os_vm_address_t) th->interrupt_data, (sizeof (struct interrupt_data))); - os_invalidate((os_vm_address_t) th->control_stack_start, + os_invalidate((os_vm_address_t) th->os_address, THREAD_STRUCT_SIZE); } @@ -346,18 +363,32 @@ create_thread_struct(lispobj initial_function) { union per_thread_data *per_thread; struct thread *th=0; /* subdue gcc */ void *spaces=0; + void *aligned_spaces=0; #ifdef LISP_FEATURE_SB_THREAD int i; #endif - /* may as well allocate all the spaces at once: it saves us from +#ifdef CREATE_CLEANUP_THREAD + /* Give a chance for cleanup threads to run. */ + sched_yield(); +#endif + /* May as well allocate all the spaces at once: it saves us from * having to decide what to do if only some of the allocations - * succeed */ + * succeed. SPACES must be page-aligned, since the GC expects the + * control stack to start at a page boundary. We can't rely on the + * alignment passed from os_validate, since that might assume the + * current (e.g. 4k) pagesize, while we calculate with the biggest + * (e.g. 64k) pagesize allowed by the ABI. */ spaces=os_validate(0, THREAD_STRUCT_SIZE); if(!spaces) - return NULL; + return NULL; + /* Aligning up is safe as THREAD_STRUCT_SIZE has BACKEND_PAGE_SIZE + * padding. */ + aligned_spaces = (void *)((((unsigned long)(char *)spaces) + + BACKEND_PAGE_SIZE - 1) + & ~(unsigned long)(BACKEND_PAGE_SIZE - 1)); per_thread=(union per_thread_data *) - (spaces+ + (aligned_spaces+ THREAD_CONTROL_STACK_SIZE+ BINDING_STACK_SIZE+ ALIEN_STACK_SIZE); @@ -392,7 +423,8 @@ create_thread_struct(lispobj initial_function) { #endif th=&per_thread->thread; - th->control_stack_start = spaces; + th->os_address = spaces; + th->control_stack_start = aligned_spaces; th->binding_stack_start= (lispobj*)((void*)th->control_stack_start+THREAD_CONTROL_STACK_SIZE); th->control_stack_end = th->binding_stack_start; @@ -439,6 +471,7 @@ create_thread_struct(lispobj initial_function) { bind_variable(FREE_INTERRUPT_CONTEXT_INDEX,make_fixnum(0),th); bind_variable(INTERRUPT_PENDING, NIL,th); bind_variable(INTERRUPTS_ENABLED,T,th); + bind_variable(ALLOW_WITH_INTERRUPTS,T,th); bind_variable(GC_PENDING,NIL,th); #ifdef LISP_FEATURE_SB_THREAD bind_variable(STOP_FOR_GC_PENDING,NIL,th); @@ -457,9 +490,20 @@ create_thread_struct(lispobj initial_function) { return th; } +#ifdef LISP_FEATURE_MACH_EXCEPTION_HANDLER +mach_port_t setup_mach_exception_handling_thread(); +kern_return_t mach_thread_init(mach_port_t thread_exception_port); + +#endif + void create_initial_thread(lispobj initial_function) { struct thread *th=create_thread_struct(initial_function); if(th) { +#ifdef LISP_FEATURE_MACH_EXCEPTION_HANDLER + kern_return_t ret; + + setup_mach_exception_handling_thread(); +#endif initial_thread_trampoline(th); /* no return */ } else lose("can't create initial thread\n"); } @@ -516,6 +560,7 @@ boolean create_os_thread(struct thread *th,os_thread_t *kid_tid) } r=0; } + #ifdef QUEUE_FREEABLE_THREAD_STACKS free_freeable_stacks(); #endif