X-Git-Url: http://repo.macrolet.net/gitweb/?a=blobdiff_plain;f=src%2Fcode%2Fcold-init.lisp;h=2b7f3294fe6745f3cab81711bb0cf92af3fcf740;hb=422b88abf96f4842a3d0999cd3b80d96f5a153d6;hp=02bdfd96d61151e2256b54915f2154dd13e9cda4;hpb=ec6d4bd97d9adc6f4003747d8ca92fad7766ccfd;p=sbcl.git diff --git a/src/code/cold-init.lisp b/src/code/cold-init.lisp index 02bdfd9..2b7f329 100644 --- a/src/code/cold-init.lisp +++ b/src/code/cold-init.lisp @@ -204,15 +204,9 @@ ;; FIXME: This list of modes should be defined in one place and ;; explicitly shared between here and REINIT. - ;; - ;; FIXME: In CMU CL, this is done "here" (i.e. in the analogous - ;; lispinit.lisp code) for every processor architecture. But Daniel - ;; Barlow's Alpha patches suppress it for Alpha. Why the difference? - #!+alpha - (set-floating-point-modes :traps '(:overflow - #!-x86 :underflow - :invalid - :divide-by-zero)) + + ;; Why was this marked #!+alpha? CMUCL does it here on all architectures + (set-floating-point-modes :traps '(:overflow :invalid :divide-by-zero)) (show-and-call !class-finalize) @@ -278,18 +272,12 @@ instead (which is another name for the same thing).")) (signal-cold-init-or-reinit) (gc-reinit) (setf (sb!alien:extern-alien "internal_errors_enabled" boolean) t) - (set-floating-point-modes :traps - '(:overflow - :invalid - :divide-by-zero - ;; PRINT seems not to like x86 NPX - ;; denormal floats like - ;; LEAST-NEGATIVE-SINGLE-FLOAT, so - ;; the :UNDERFLOW exceptions are - ;; disabled by default. Joe User can - ;; explicitly enable them if - ;; desired. - #!-x86 :underflow)) + ;; PRINT seems not to like x86 NPX denormal floats like + ;; LEAST-NEGATIVE-SINGLE-FLOAT, so the :UNDERFLOW exceptions are + ;; disabled by default. Joe User can explicitly enable them if + ;; desired. + (set-floating-point-modes :traps '(:overflow :invalid :divide-by-zero)) + ;; Clear pseudo atomic in case this core wasn't compiled with ;; support. ;;