#include <sys/file.h>
#include <sys/types.h>
#include <unistd.h>
+#include <errno.h>
#include "./signal.h"
#include "os.h"
u32 local_ldt_copy[LDT_ENTRIES*LDT_ENTRY_SIZE/sizeof(u32)];
-/* XXX this could be conditionally compiled based on some
- * "debug-friendly" flag. But it doesn't really make stuff slower,
- * just the runtime gets fractionally larger */
+/* This is never actually called, but it's great for calling from gdb when
+ * users have thread-related problems that maintainers can't duplicate */
void debug_get_ldt()
{
- int n=__modify_ldt (0, local_ldt_copy, sizeof local_ldt_copy);
+ int n=modify_ldt (0, local_ldt_copy, sizeof local_ldt_copy);
printf("%d bytes in ldt: print/x local_ldt_copy\n", n);
}
+lispobj modify_ldt_lock; /* protect all calls to modify_ldt */
+
int arch_os_thread_init(struct thread *thread) {
stack_t sigstack;
#ifdef LISP_FEATURE_SB_THREAD
1, 0, 0, /* index, address, length filled in later */
1, MODIFY_LDT_CONTENTS_DATA, 0, 0, 0, 1
};
+ int n;
+ get_spinlock(&modify_ldt_lock,thread);
+ n=modify_ldt(0,local_ldt_copy,sizeof local_ldt_copy);
/* get next free ldt entry */
- int n=__modify_ldt(0,local_ldt_copy,sizeof local_ldt_copy);
+
if(n) {
u32 *p;
for(n=0,p=local_ldt_copy;*p;p+=LDT_ENTRY_SIZE/sizeof(u32))
ldt_entry.base_addr=(unsigned long) thread;
ldt_entry.limit=dynamic_values_bytes;
ldt_entry.limit_in_pages=0;
- if (__modify_ldt (1, &ldt_entry, sizeof (ldt_entry)) != 0)
+ if (modify_ldt (1, &ldt_entry, sizeof (ldt_entry)) != 0) {
+ modify_ldt_lock=0;
/* modify_ldt call failed: something magical is not happening */
return -1;
- __asm__ __volatile__ ("movw %w0, %%gs" : : "q"
+ }
+ __asm__ __volatile__ ("movw %w0, %%fs" : : "q"
((n << 3) /* selector number */
+ (1 << 2) /* TI set = LDT */
+ 3)); /* privilege level */
thread->tls_cookie=n;
+ modify_ldt_lock=0;
+
if(n<0) return 0;
#endif
#ifdef LISP_FEATURE_C_STACK_IS_CONTROL_STACK
#ifdef LISP_FEATURE_SB_THREAD
register struct thread *me=0;
if(all_threads)
- __asm__ ("movl %%gs:%c1,%0" : "=r" (me)
+ __asm__ __volatile__ ("movl %%fs:%c1,%0" : "=r" (me)
: "i" (offsetof (struct thread,this)));
return me;
#else
return all_threads;
#endif
}
+struct thread *debug_get_fs() {
+ register u32 fs;
+ __asm__ __volatile__ ("movl %%fs,%0" : "=r" (fs) : );
+ return fs;
+}
/* free any arch/os-specific resources used by thread, which is now
* defunct. Not called on live threads
};
ldt_entry.entry_number=thread->tls_cookie;
- if (__modify_ldt (1, &ldt_entry, sizeof (ldt_entry)) != 0)
+ get_spinlock(&modify_ldt_lock,thread);
+ if (modify_ldt (1, &ldt_entry, sizeof (ldt_entry)) != 0) {
+ modify_ldt_lock=0;
/* modify_ldt call failed: something magical is not happening */
return 0;
+ }
+ modify_ldt_lock=0;
return 1;
}
os_context_register_addr(os_context_t *context, int offset)
{
switch(offset) {
- case 0: return &context->uc_mcontext.gregs[11]; /* EAX */
- case 2: return &context->uc_mcontext.gregs[10]; /* ECX */
- case 4: return &context->uc_mcontext.gregs[9]; /* EDX */
- case 6: return &context->uc_mcontext.gregs[8]; /* EBX */
- case 8: return &context->uc_mcontext.gregs[7]; /* ESP */
- case 10: return &context->uc_mcontext.gregs[6]; /* EBP */
- case 12: return &context->uc_mcontext.gregs[5]; /* ESI */
- case 14: return &context->uc_mcontext.gregs[4]; /* EDI */
+ case reg_EAX: return &context->uc_mcontext.gregs[11];
+ case reg_ECX: return &context->uc_mcontext.gregs[10];
+ case reg_EDX: return &context->uc_mcontext.gregs[9];
+ case reg_EBX: return &context->uc_mcontext.gregs[8];
+ case reg_ESP: return &context->uc_mcontext.gregs[7];
+ case reg_EBP: return &context->uc_mcontext.gregs[6];
+ case reg_ESI: return &context->uc_mcontext.gregs[5];
+ case reg_EDI: return &context->uc_mcontext.gregs[4];
default: return 0;
}
return &context->uc_mcontext.gregs[offset];