X-Git-Url: http://repo.macrolet.net/gitweb/?a=blobdiff_plain;f=src%2Fruntime%2Finterrupt.c;h=5164c303675a4480f71568e170258ea44e42f3da;hb=94ea2b2082deaa0331dfb66fa6af6ca12dd8dc83;hp=1578aca75ce55bf456467002ccfb4f8eeeee2b94;hpb=3f757cc9b3d6f14600365b7c0dd7d213269d7242;p=sbcl.git diff --git a/src/runtime/interrupt.c b/src/runtime/interrupt.c index 1578aca..5164c30 100644 --- a/src/runtime/interrupt.c +++ b/src/runtime/interrupt.c @@ -47,6 +47,7 @@ #include #include #include +#include #include "sbcl.h" #include "runtime.h" @@ -63,6 +64,7 @@ #include "interr.h" #include "genesis/fdefn.h" #include "genesis/simple-fun.h" +#include "genesis/cons.h" @@ -73,19 +75,6 @@ static void store_signal_data_for_later (struct interrupt_data *data, os_context_t *context); boolean interrupt_maybe_gc_int(int signal, siginfo_t *info, void *v_context); -extern volatile lispobj all_threads_lock; - -/* - * This is a workaround for some slightly silly Linux/GNU Libc - * behaviour: glibc defines sigset_t to support 1024 signals, which is - * more than the kernel. This is usually not a problem, but becomes - * one when we want to save a signal mask from a ucontext, and restore - * it later into another ucontext: the ucontext is allocated on the - * stack by the kernel, so copying a libc-sized sigset_t into it will - * overflow and cause other data on the stack to be corrupted */ - -#define REAL_SIGSET_SIZE_BYTES ((NSIG/8)) - void sigaddset_blockable(sigset_t *s) { sigaddset(s, SIGHUP); @@ -111,6 +100,34 @@ void sigaddset_blockable(sigset_t *s) #endif } +static sigset_t blockable_sigset; + +inline static void check_blockables_blocked_or_lose() +{ + /* Get the current sigmask, by blocking the empty set. */ + sigset_t empty,current; + int i; + sigemptyset(&empty); + thread_sigmask(SIG_BLOCK, &empty, ¤t); + for(i=0;iinterrupt_data; - /* FIXME I'm not altogether sure this is appropriate if we're - * here as the result of a pseudo-atomic */ - SetSymbolValue(INTERRUPT_PENDING, NIL,thread); - - /* restore the saved signal mask from the original signal (the - * one that interrupted us during the critical section) into the - * os_context for the signal we're currently in the handler for. - * This should ensure that when we return from the handler the - * blocked signals are unblocked */ - - memcpy(os_context_sigmask_addr(context), &data->pending_mask, - REAL_SIGSET_SIZE_BYTES); - - sigemptyset(&data->pending_mask); - /* This will break on sparc linux: the deferred handler really wants - * to be called with a void_context */ - run_deferred_handler(data,(void *)context); + + /* Pseudo atomic may trigger several times for a single interrupt, + * and while without-interrupts should not, a false trigger by + * pseudo-atomic may eat a pending handler even from + * without-interrupts. */ + if (data->pending_handler) { + + /* If we're here as the result of a pseudo-atomic as opposed + * to WITHOUT-INTERRUPTS, then INTERRUPT_PENDING is already + * NIL, because maybe_defer_handler sets + * PSEUDO_ATOMIC_INTERRUPTED only if interrupts are enabled.*/ + SetSymbolValue(INTERRUPT_PENDING, NIL,thread); + + /* restore the saved signal mask from the original signal (the + * one that interrupted us during the critical section) into the + * os_context for the signal we're currently in the handler for. + * This should ensure that when we return from the handler the + * blocked signals are unblocked */ + sigcopyset(os_context_sigmask_addr(context), &data->pending_mask); + + sigemptyset(&data->pending_mask); + /* This will break on sparc linux: the deferred handler really wants + * to be called with a void_context */ + run_deferred_handler(data,(void *)context); + } } /* @@ -348,6 +383,8 @@ interrupt_handle_now(int signal, siginfo_t *info, void *void_context) boolean were_in_lisp; #endif union interrupt_handler handler; + check_blockables_blocked_or_lose(); + check_interrupts_enabled_or_lose(context); #ifdef LISP_FEATURE_LINUX /* Under Linux on some architectures, we appear to have to restore @@ -398,7 +435,7 @@ interrupt_handle_now(int signal, siginfo_t *info, void *void_context) lispobj info_sap,context_sap = alloc_sap(context); info_sap = alloc_sap(info); /* Allow signals again. */ - sigprocmask(SIG_SETMASK, os_context_sigmask_addr(context), 0); + thread_sigmask(SIG_SETMASK, os_context_sigmask_addr(context), 0); #ifdef QSHOW_SIGNALS SHOW("calling Lisp-level handler"); @@ -415,7 +452,7 @@ interrupt_handle_now(int signal, siginfo_t *info, void *void_context) #endif /* Allow signals again. */ - sigprocmask(SIG_SETMASK, os_context_sigmask_addr(context), 0); + thread_sigmask(SIG_SETMASK, os_context_sigmask_addr(context), 0); (*handler.c)(signal, info, void_context); } @@ -442,9 +479,14 @@ interrupt_handle_now(int signal, siginfo_t *info, void *void_context) void run_deferred_handler(struct interrupt_data *data, void *v_context) { - (*(data->pending_handler)) - (data->pending_signal,&(data->pending_info), v_context); + /* The pending_handler may enable interrupts (see + * interrupt_maybe_gc_int) and then another interrupt may hit, + * overwrite interrupt_data, so reset the pending handler before + * calling it. Trust the handler to finish with the siginfo before + * enabling interrupts. */ + void (*pending_handler) (int, siginfo_t*, void*)=data->pending_handler; data->pending_handler=0; + (*pending_handler)(data->pending_signal,&(data->pending_info), v_context); } boolean @@ -452,9 +494,23 @@ maybe_defer_handler(void *handler, struct interrupt_data *data, int signal, siginfo_t *info, os_context_t *context) { struct thread *thread=arch_os_get_current_thread(); + + check_blockables_blocked_or_lose(); + + if (SymbolValue(INTERRUPT_PENDING,thread) != NIL) + lose("interrupt already pending"); + /* If interrupts are disabled then INTERRUPT_PENDING is set and + * not PSEDUO_ATOMIC_INTERRUPTED. This is important for a pseudo + * atomic section inside a without-interrupts. + */ if (SymbolValue(INTERRUPTS_ENABLED,thread) == NIL) { store_signal_data_for_later(data,handler,signal,info,context); SetSymbolValue(INTERRUPT_PENDING, T,thread); +#ifdef QSHOW_SIGNALS + FSHOW((stderr, + "/maybe_defer_handler(%x,%d),thread=%ld: deferred\n", + (unsigned int)handler,signal,thread->os_thread)); +#endif return 1; } /* a slightly confusing test. arch_pseudo_atomic_atomic() doesn't @@ -467,15 +523,31 @@ maybe_defer_handler(void *handler, struct interrupt_data *data, arch_pseudo_atomic_atomic(context)) { store_signal_data_for_later(data,handler,signal,info,context); arch_set_pseudo_atomic_interrupted(context); +#ifdef QSHOW_SIGNALS + FSHOW((stderr, + "/maybe_defer_handler(%x,%d),thread=%ld: deferred(PA)\n", + (unsigned int)handler,signal,thread->os_thread)); +#endif return 1; } +#ifdef QSHOW_SIGNALS + FSHOW((stderr, + "/maybe_defer_handler(%x,%d),thread=%ld: not deferred\n", + (unsigned int)handler,signal,thread->os_thread)); +#endif return 0; } + static void store_signal_data_for_later (struct interrupt_data *data, void *handler, int signal, siginfo_t *info, os_context_t *context) { + if (data->pending_handler) + lose("tried to overwrite pending interrupt handler %x with %x\n", + data->pending_handler, handler); + if (!handler) + lose("tried to defer null interrupt handler\n"); data->pending_handler = handler; data->pending_signal = signal; if(info) @@ -486,22 +558,11 @@ store_signal_data_for_later (struct interrupt_data *data, void *handler, * run_deferred_handler happens. Then the usually-blocked * signals are added to the mask in the context so that we are * running with blocked signals when the handler returns */ - sigemptyset(&(data->pending_mask)); - memcpy(&(data->pending_mask), - os_context_sigmask_addr(context), - REAL_SIGSET_SIZE_BYTES); + sigcopyset(&(data->pending_mask),os_context_sigmask_addr(context)); sigaddset_blockable(os_context_sigmask_addr(context)); - } else { - /* this is also called from gencgc alloc(), in which case - * there has been no signal and is therefore no context. */ - sigset_t new; - sigemptyset(&new); - sigaddset_blockable(&new); - sigprocmask(SIG_BLOCK,&new,&(data->pending_mask)); } } - static void maybe_now_maybe_later(int signal, siginfo_t *info, void *void_context) { @@ -521,37 +582,77 @@ maybe_now_maybe_later(int signal, siginfo_t *info, void *void_context) #endif } +static void +low_level_interrupt_handle_now(int signal, siginfo_t *info, void *void_context) +{ + os_context_t *context = (os_context_t*)void_context; + struct thread *thread=arch_os_get_current_thread(); + +#ifdef LISP_FEATURE_LINUX + os_restore_fp_control(context); +#endif + check_blockables_blocked_or_lose(); + check_interrupts_enabled_or_lose(context); + (*thread->interrupt_data->interrupt_low_level_handlers[signal]) + (signal, info, void_context); +#ifdef LISP_FEATURE_DARWIN + /* Work around G5 bug */ + DARWIN_FIX_CONTEXT(context); +#endif +} + +static void +low_level_maybe_now_maybe_later(int signal, siginfo_t *info, void *void_context) +{ + os_context_t *context = arch_os_get_context(&void_context); + struct thread *thread=arch_os_get_current_thread(); + struct interrupt_data *data=thread->interrupt_data; +#ifdef LISP_FEATURE_LINUX + os_restore_fp_control(context); +#endif + if(maybe_defer_handler(low_level_interrupt_handle_now,data, + signal,info,context)) + return; + low_level_interrupt_handle_now(signal, info, context); +#ifdef LISP_FEATURE_DARWIN + /* Work around G5 bug */ + DARWIN_FIX_CONTEXT(context); +#endif +} + #ifdef LISP_FEATURE_SB_THREAD void sig_stop_for_gc_handler(int signal, siginfo_t *info, void *void_context) { os_context_t *context = arch_os_get_context(&void_context); struct thread *thread=arch_os_get_current_thread(); - struct interrupt_data *data=thread->interrupt_data; sigset_t ss; int i; - if(maybe_defer_handler(sig_stop_for_gc_handler,data, - signal,info,context)) { - return; - } /* need the context stored so it can have registers scavenged */ fake_foreign_function_call(context); sigemptyset(&ss); for(i=1;istate=STATE_STOPPED; + if(thread->state!=STATE_RUNNING) { + lose("sig_stop_for_gc_handler: wrong thread state: %ld\n", + fixnum_value(thread->state)); + } + thread->state=STATE_SUSPENDED; sigemptyset(&ss); sigaddset(&ss,SIG_STOP_FOR_GC); sigwaitinfo(&ss,0); + if(thread->state!=STATE_SUSPENDED) { + lose("sig_stop_for_gc_handler: wrong thread state on wakeup: %ld\n", + fixnum_value(thread->state)); + } + thread->state=STATE_RUNNING; undo_fake_foreign_function_call(context); } @@ -591,6 +692,10 @@ gc_trigger_hit(int signal, siginfo_t *info, os_context_t *context) * previously */ +#if (defined(LISP_FEATURE_X86) || defined(LISP_FEATURE_X86_64)) +int *context_eflags_addr(os_context_t *context); +#endif + extern lispobj call_into_lisp(lispobj fun, lispobj *args, int nargs); extern void post_signal_tramp(void); void arrange_return_to_lisp_function(os_context_t *context, lispobj function) @@ -602,6 +707,9 @@ void arrange_return_to_lisp_function(os_context_t *context, lispobj function) /* Build a stack frame showing `interrupted' so that the * user's backtrace makes (as much) sense (as usual) */ + + /* FIXME: what about restoring fp state? */ + /* FIXME: what about restoring errno? */ #ifdef LISP_FEATURE_X86 /* Suppose the existence of some function that saved all * registers, called call_into_lisp, then restored GP registers and @@ -609,13 +717,15 @@ void arrange_return_to_lisp_function(os_context_t *context, lispobj function) push ebp mov ebp esp - pushad + pushfl + pushal push $0 push $0 pushl {address of function to call} call 0x8058db0 addl $12,%esp - popa + popal + popfl leave ret @@ -632,44 +742,47 @@ void arrange_return_to_lisp_function(os_context_t *context, lispobj function) u32 *sp=(u32 *)*os_context_register_addr(context,reg_ESP); - *(sp-14) = post_signal_tramp; /* return address for call_into_lisp */ - *(sp-13) = function; /* args for call_into_lisp : function*/ - *(sp-12) = 0; /* arg array */ - *(sp-11) = 0; /* no. args */ + *(sp-15) = post_signal_tramp; /* return address for call_into_lisp */ + *(sp-14) = function; /* args for call_into_lisp : function*/ + *(sp-13) = 0; /* arg array */ + *(sp-12) = 0; /* no. args */ /* this order matches that used in POPAD */ - *(sp-10)=*os_context_register_addr(context,reg_EDI); - *(sp-9)=*os_context_register_addr(context,reg_ESI); - - *(sp-8)=*os_context_register_addr(context,reg_ESP)-8; - *(sp-7)=0; - *(sp-6)=*os_context_register_addr(context,reg_EBX); - - *(sp-5)=*os_context_register_addr(context,reg_EDX); - *(sp-4)=*os_context_register_addr(context,reg_ECX); - *(sp-3)=*os_context_register_addr(context,reg_EAX); + *(sp-11)=*os_context_register_addr(context,reg_EDI); + *(sp-10)=*os_context_register_addr(context,reg_ESI); + + *(sp-9)=*os_context_register_addr(context,reg_ESP)-8; + /* POPAD ignores the value of ESP: */ + *(sp-8)=0; + *(sp-7)=*os_context_register_addr(context,reg_EBX); + + *(sp-6)=*os_context_register_addr(context,reg_EDX); + *(sp-5)=*os_context_register_addr(context,reg_ECX); + *(sp-4)=*os_context_register_addr(context,reg_EAX); + *(sp-3)=*context_eflags_addr(context); *(sp-2)=*os_context_register_addr(context,reg_EBP); *(sp-1)=*os_context_pc_addr(context); #elif defined(LISP_FEATURE_X86_64) u64 *sp=(u64 *)*os_context_register_addr(context,reg_RSP); - *(sp-19) = post_signal_tramp; /* return address for call_into_lisp */ - - *(sp-18)=*os_context_register_addr(context,reg_R15); - *(sp-17)=*os_context_register_addr(context,reg_R14); - *(sp-16)=*os_context_register_addr(context,reg_R13); - *(sp-15)=*os_context_register_addr(context,reg_R12); - *(sp-14)=*os_context_register_addr(context,reg_R11); - *(sp-13)=*os_context_register_addr(context,reg_R10); - *(sp-12)=*os_context_register_addr(context,reg_R9); - *(sp-11)=*os_context_register_addr(context,reg_R8); - *(sp-10)=*os_context_register_addr(context,reg_RDI); - *(sp-9)=*os_context_register_addr(context,reg_RSI); - *(sp-8)=*os_context_register_addr(context,reg_RSP)-16; - *(sp-7)=0; - *(sp-6)=*os_context_register_addr(context,reg_RBX); - *(sp-5)=*os_context_register_addr(context,reg_RDX); - *(sp-4)=*os_context_register_addr(context,reg_RCX); - *(sp-3)=*os_context_register_addr(context,reg_RAX); + *(sp-20) = post_signal_tramp; /* return address for call_into_lisp */ + + *(sp-19)=*os_context_register_addr(context,reg_R15); + *(sp-18)=*os_context_register_addr(context,reg_R14); + *(sp-17)=*os_context_register_addr(context,reg_R13); + *(sp-16)=*os_context_register_addr(context,reg_R12); + *(sp-15)=*os_context_register_addr(context,reg_R11); + *(sp-14)=*os_context_register_addr(context,reg_R10); + *(sp-13)=*os_context_register_addr(context,reg_R9); + *(sp-12)=*os_context_register_addr(context,reg_R8); + *(sp-11)=*os_context_register_addr(context,reg_RDI); + *(sp-10)=*os_context_register_addr(context,reg_RSI); + *(sp-9)=*os_context_register_addr(context,reg_RSP)-16; + *(sp-8)=0; + *(sp-7)=*os_context_register_addr(context,reg_RBX); + *(sp-6)=*os_context_register_addr(context,reg_RDX); + *(sp-5)=*os_context_register_addr(context,reg_RCX); + *(sp-4)=*os_context_register_addr(context,reg_RAX); + *(sp-3)=*context_eflags_addr(context); *(sp-2)=*os_context_register_addr(context,reg_RBP); *(sp-1)=*os_context_pc_addr(context); @@ -686,15 +799,15 @@ void arrange_return_to_lisp_function(os_context_t *context, lispobj function) *os_context_register_addr(context,reg_ECX) = 0; *os_context_register_addr(context,reg_EBP) = sp-2; #ifdef __NetBSD__ - *os_context_register_addr(context,reg_UESP) = sp-14; + *os_context_register_addr(context,reg_UESP) = sp-15; #else - *os_context_register_addr(context,reg_ESP) = sp-14; + *os_context_register_addr(context,reg_ESP) = sp-15; #endif #elif defined(LISP_FEATURE_X86_64) *os_context_pc_addr(context) = call_into_lisp; *os_context_register_addr(context,reg_RCX) = 0; *os_context_register_addr(context,reg_RBP) = sp-2; - *os_context_register_addr(context,reg_RSP) = sp-19; + *os_context_register_addr(context,reg_RSP) = sp-20; #else /* this much of the calling convention is common to all non-x86 ports */ @@ -718,20 +831,25 @@ void arrange_return_to_lisp_function(os_context_t *context, lispobj function) void interrupt_thread_handler(int num, siginfo_t *info, void *v_context) { os_context_t *context = (os_context_t*)arch_os_get_context(&v_context); + /* The order of interrupt execution is peculiar. If thread A + * interrupts thread B with I1, I2 and B for some reason recieves + * I1 when FUN2 is already on the list, then it is FUN2 that gets + * to run first. But when FUN2 is run SIG_INTERRUPT_THREAD is + * enabled again and I2 hits pretty soon in FUN2 and run + * FUN1. This is of course just one scenario, and the order of + * thread interrupt execution is undefined. */ struct thread *th=arch_os_get_current_thread(); - struct interrupt_data *data= - th ? th->interrupt_data : global_interrupt_data; - if(maybe_defer_handler(interrupt_thread_handler,data,num,info,context)){ - return ; - } - arrange_return_to_lisp_function(context,info->si_value.sival_int); + struct cons *c; + if (th->state != STATE_RUNNING) + lose("interrupt_thread_handler: thread %ld in wrong state: %d\n", + th->os_thread,fixnum_value(th->state)); + get_spinlock(&th->interrupt_fun_lock,(long)th); + c=((struct cons *)native_pointer(th->interrupt_fun)); + arrange_return_to_lisp_function(context,c->car); + th->interrupt_fun=(lispobj *)(c->cdr); + release_spinlock(&th->interrupt_fun_lock); } -void thread_exit_handler(int num, siginfo_t *info, void *v_context) -{ /* called when a child thread exits */ - mark_dead_threads(); -} - #endif /* KLUDGE: Theoretically the approach we use for undefined alien @@ -755,8 +873,8 @@ boolean handle_guard_page_triggered(os_context_t *context,void *addr){ * protection so the error handler has some headroom, protect the * previous page so that we can catch returns from the guard page * and restore it. */ - protect_control_stack_guard_page(th->pid,0); - protect_control_stack_return_guard_page(th->pid,1); + protect_control_stack_guard_page(th,0); + protect_control_stack_return_guard_page(th,1); arrange_return_to_lisp_function (context, SymbolFunction(CONTROL_STACK_EXHAUSTED_ERROR)); @@ -768,8 +886,8 @@ boolean handle_guard_page_triggered(os_context_t *context,void *addr){ * unprotect this one. This works even if we somehow missed * the return-guard-page, and hit it on our way to new * exhaustion instead. */ - protect_control_stack_guard_page(th->pid,1); - protect_control_stack_return_guard_page(th->pid,0); + protect_control_stack_guard_page(th,1); + protect_control_stack_return_guard_page(th,0); return 1; } else if (addr >= undefined_alien_address && @@ -796,12 +914,13 @@ interrupt_maybe_gc(int signal, siginfo_t *info, void *void_context) struct interrupt_data *data= th ? th->interrupt_data : global_interrupt_data; - if(!foreign_function_call_active && gc_trigger_hit(signal, info, context)){ - clear_auto_gc_trigger(); - if(!maybe_defer_handler - (interrupt_maybe_gc_int,data,signal,info,void_context)) - interrupt_maybe_gc_int(signal,info,void_context); - return 1; + if(!data->pending_handler && !foreign_function_call_active && + gc_trigger_hit(signal, info, context)){ + clear_auto_gc_trigger(); + if(!maybe_defer_handler(interrupt_maybe_gc_int, + data,signal,info,void_context)) + interrupt_maybe_gc_int(signal,info,void_context); + return 1; } return 0; } @@ -812,19 +931,27 @@ interrupt_maybe_gc(int signal, siginfo_t *info, void *void_context) boolean interrupt_maybe_gc_int(int signal, siginfo_t *info, void *void_context) { - sigset_t new; os_context_t *context=(os_context_t *) void_context; + + check_blockables_blocked_or_lose(); fake_foreign_function_call(context); + /* SUB-GC may return without GCing if *GC-INHIBIT* is set, in * which case we will be running with no gc trigger barrier * thing for a while. But it shouldn't be long until the end - * of WITHOUT-GCING. */ + * of WITHOUT-GCING. + * + * FIXME: It would be good to protect the end of dynamic space + * and signal a storage condition from there. + */ + + /* restore the signal mask from the interrupted context before + * calling into Lisp */ + if (context) + thread_sigmask(SIG_SETMASK, os_context_sigmask_addr(context), 0); - sigemptyset(&new); - sigaddset_blockable(&new); - /* enable signals before calling into Lisp */ - sigprocmask(SIG_UNBLOCK,&new,0); funcall0(SymbolFunction(SUB_GC)); + undo_fake_foreign_function_call(context); return 1; } @@ -849,7 +976,11 @@ undoably_install_low_level_interrupt_handler (int signal, lose("bad signal number %d", signal); } - sa.sa_sigaction = handler; + if (sigismember(&blockable_sigset,signal)) + sa.sa_sigaction = low_level_maybe_now_maybe_later; + else + sa.sa_sigaction = handler; + sigemptyset(&sa.sa_mask); sigaddset_blockable(&sa.sa_mask); sa.sa_flags = SA_SIGINFO | SA_RESTART; @@ -882,13 +1013,13 @@ install_handler(int signal, void handler(int, siginfo_t*, void*)) sigemptyset(&new); sigaddset(&new, signal); - sigprocmask(SIG_BLOCK, &new, &old); + thread_sigmask(SIG_BLOCK, &new, &old); sigemptyset(&new); sigaddset_blockable(&new); - FSHOW((stderr, "/data->interrupt_low_level_handlers[signal]=%d\n", - data->interrupt_low_level_handlers[signal])); + FSHOW((stderr, "/data->interrupt_low_level_handlers[signal]=%x\n", + (unsigned int)data->interrupt_low_level_handlers[signal])); if (data->interrupt_low_level_handlers[signal]==0) { if (ARE_SAME_HANDLER(handler, SIG_DFL) || ARE_SAME_HANDLER(handler, SIG_IGN)) { @@ -908,7 +1039,7 @@ install_handler(int signal, void handler(int, siginfo_t*, void*)) oldhandler = data->interrupt_handlers[signal]; data->interrupt_handlers[signal].c = handler; - sigprocmask(SIG_SETMASK, &old, 0); + thread_sigmask(SIG_SETMASK, &old, 0); FSHOW((stderr, "/leaving POSIX install_handler(%d, ..)\n", signal)); @@ -920,6 +1051,9 @@ interrupt_init() { int i; SHOW("entering interrupt_init()"); + sigemptyset(&blockable_sigset); + sigaddset_blockable(&blockable_sigset); + global_interrupt_data=calloc(sizeof(struct interrupt_data), 1); /* Set up high level handler information. */