Files
rtems/cpukit/score/cpu/m68k/cpu_asm.S
Sebastian Huber d19cce29dc score: Per-CPU thread dispatch disable level
Use a per-CPU thread dispatch disable level.  So instead of one global
thread dispatch disable level we have now one instance per processor.
This is a major performance improvement for SMP.  On non-SMP
configurations this may simplifiy the interrupt entry/exit code.

The giant lock is still present, but it is now decoupled from the thread
dispatching in _Thread_Dispatch(), _Thread_Handler(),
_Thread_Restart_self() and the interrupt entry/exit.   Access to the
giant lock is now available via _Giant_Acquire() and _Giant_Release().
The giant lock is still implicitly acquired via
_Thread_Dispatch_decrement_disable_level().

The giant lock is only acquired for high-level operations in interrupt
handlers (e.g. release of a semaphore, sending of an event).

As a side-effect this change fixes the lost thread dispatch necessary
indication bug in _Thread_Dispatch().

A per-CPU thread dispatch disable level greatly simplifies the SMP
support for the interrupt entry/exit code since no spin locks have to be
acquired in this area.  It is only necessary to get the current
processor index and use this to calculate the address of the own per-CPU
control.  This reduces the interrupt latency considerably.

All elements for the interrupt entry/exit code are now part of the
Per_CPU_Control structure: thread dispatch disable level, ISR nest level
and thread dispatch necessary.  Nothing else is required (except CPU
port specific stuff like on SPARC).
2013-08-09 23:02:38 +02:00

383 lines
13 KiB
ArmAsm

/* cpu_asm.s
*
* This file contains all assembly code for the MC68020 implementation
* of RTEMS.
*
* COPYRIGHT (c) 1989-2008.
* On-Line Applications Research Corporation (OAR).
*
* The license and distribution terms for this file may be
* found in the file LICENSE in this distribution or at
* http://www.rtems.com/license/LICENSE.
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include <rtems/asm.h>
#include <rtems/score/percpu.h>
/* void _CPU_Context_switch( run_context, heir_context )
*
* This routine performs a normal non-FP context.
*/
.align 4
.global SYM (_CPU_Context_switch)
.set RUNCONTEXT_ARG, 4 | save context argument
.set HEIRCONTEXT_ARG, 8 | restore context argument
SYM (_CPU_Context_switch):
moval a7@(RUNCONTEXT_ARG),a0| a0 = running thread context
movw sr,d1 | d1 = status register
movml d1-d7/a2-a7,a0@ | save context
moval a7@(HEIRCONTEXT_ARG),a0| a0 = heir thread context
#if defined( __mcoldfire__ ) && ( M68K_HAS_FPU == 1 )
moveb a0@(13*4),d0 | get context specific DF bit info in d0
btstb #4,d0 | test context specific DF bit info
beq fpu_on | branch if FPU needs to be switched on
fpu_off: movl _CPU_cacr_shadow,d0 | get content of _CPU_cacr_shadow in d0
btstl #4,d0 | test DF bit info in d0
bne restore | branch if FPU is already switched off
bsetl #4,d0 | set DF bit in d0
bra cacr_set | branch to set the new FPU setting in cacr and _CPU_cacr_shadow
fpu_on: movl _CPU_cacr_shadow,d0 | get content of _CPU_cacr_shadow in d1
btstl #4,d0 | test context specific DF bit info
beq restore | branch if FPU is already switched on
bclrl #4,d0 | clear DF bit info in d0
cacr_set: movew sr,d1 | get content of sr in d1
oril #0x00000700,d1 | mask d1
movew d1,sr | disable all interrupts
movl d0,_CPU_cacr_shadow | move _CPU_cacr_shadow to d1
movec d0,cacr | enable FPU in cacr
#endif
restore: movml a0@,d1-d7/a2-a7 | restore context
movw d1,sr | restore status register
rts
.global SYM (_CPU_Context_Restart_self)
.set CONTEXT_ARG, 4 | context arg
#if defined( __mcoldfire__ ) && ( M68K_HAS_FPU == 1 )
/* XXX _CPU_Context_switch maintains FPU context -- do we have to restore
* that, too??
*/
#warning "_CPU_Context_Restart_self restoring FPU context not implemented"
#endif
SYM(_CPU_Context_Restart_self):
moval a7@(CONTEXT_ARG),a0
bra restore
/*
* Floating point context save and restore.
*
* The code for the MC68881 or MC68882 is based upon the code shown on pages
* 6-38 of the MC68881/68882 Users Manual (rev 1). CPU_FP_CONTEXT_SIZE is
* higher than expected to account for the -1 pushed at end of this sequence.
*/
#if ( CPU_HARDWARE_FP == TRUE )
.set FPCONTEXT_ARG, 4 | save FP context argument
.align 4
.global SYM (_CPU_Context_save_fp)
SYM (_CPU_Context_save_fp):
/* Get context save area pointer argument from the stack */
moval a7@(FPCONTEXT_ARG), a1
moval a1@, a0
#if defined( __mcoldfire__ )
/* Move MACSR to data register and disable rounding */
movel macsr, d0
clrl d1
movl d1, macsr
/* Save MACSR and ACC0 */
movl acc0, d1
moveml d0-d1, a0@(0)
/* Save ACC1 and ACC2 */
movl acc1, d0
movl acc2, d1
moveml d0-d1, a0@(8)
/* Save ACC3 and ACCEXT01 */
movl acc3, d0
movl accext01, d1
moveml d0-d1, a0@(16)
/* Save ACCEXT23 and MASK */
movl accext23, d0
movl mask, d1
moveml d0-d1, a0@(24)
#if ( M68K_HAS_FPU == 1 )
/* Save FP state */
fsave a0@(32)
/* Save FP instruction address */
fmovel fpi, a0@(48)
/* Save FP data */
fmovem fp0-fp7, a0@(52)
#endif
#else
#if defined( __mc68060__ )
lea a0@(-M68K_FP_STATE_SIZE), a0
fsave a0@ | save 68060 state frame
#else
fsave a0@- | save 68881/68882 state frame
#endif
tstb a0@ | check for a null frame
beq.b nosv | Yes, skip save of user model
fmovem fp0-fp7, a0@- | save data registers (fp0-fp7)
fmovem fpc/fps/fpi, a0@- | and save control registers
movl #-1, a0@- | place not-null flag on stack
nosv:
movl a0, a1@ | save pointer to saved context
#endif
/* Return */
rts
.align 4
.global SYM (_CPU_Context_restore_fp)
SYM (_CPU_Context_restore_fp):
/* Get context save area pointer argument from the stack */
moval a7@(FPCONTEXT_ARG), a1
moval a1@, a0
#if defined( __mcoldfire__ )
#if ( M68K_HAS_FPU == 1 )
/* Restore FP data */
fmovem a0@(52), fp0-fp7
/* Restore FP instruction address */
fmovel a0@(48), fpi
/* Restore FP state */
frestore a0@(32)
#endif
/* Disable rounding */
clrl d0
movl d0, macsr
/* Restore MASK and ACCEXT23 */
moveml a0@(24), d0-d1
movl d0, mask
movl d1, accext23
/* Restore ACCEXT01 and ACC3 */
moveml a0@(16), d0-d1
movl d0, accext01
movl d1, acc3
/* Restore ACC2 and ACC1 */
moveml a0@(8), d0-d1
movl d0, acc2
movl d1, acc1
/* Restore ACC0 and MACSR */
moveml a0@(0), d0-d1
movl d0, acc0
movl d1, macsr
#else
tstb a0@ | Null context frame?
beq.b norst | Yes, skip fp restore
addql #4, a0 | throwaway non-null flag
fmovem a0@+, fpc/fps/fpi | restore control registers
fmovem a0@+, fp0-fp7 | restore data regs (fp0-fp7)
norst:
#if defined( __mc68060__ )
frestore a0@ | restore 68060 state frame
lea a0@(M68K_FP_STATE_SIZE), a0
#else
frestore a0@+ | restore 68881/68882 state frame
#endif
movl a0, a1@ | save pointer to saved context
#endif
/* Return */
rts
#endif
/*void _ISR_Handler()
*
* This routine provides the RTEMS interrupt management.
*
* NOTE:
* Upon entry, the master stack will contain an interrupt stack frame
* back to the interrupted thread and the interrupt stack will contain
* a throwaway interrupt stack frame. If dispatching is enabled, and this
* is the outer most interrupt, and a context switch is necessary or
* the current thread has pending signals, then set up the master stack to
* transfer control to the interrupt dispatcher.
*/
#if ( defined(__mcoldfire__) )
.set SR_OFFSET, 2 | Status register offset
.set PC_OFFSET, 4 | Program Counter offset
.set FVO_OFFSET, 0 | Format/vector offset
#elif ( M68K_HAS_VBR == 1)
.set SR_OFFSET, 0 | Status register offset
.set PC_OFFSET, 2 | Program Counter offset
.set FVO_OFFSET, 6 | Format/vector offset
#else
.set SR_OFFSET, 2 | Status register offset
.set PC_OFFSET, 4 | Program Counter offset
.set FVO_OFFSET, 0 | Format/vector offset placed in the stack
#endif /* M68K_HAS_VBR */
.set SAVED, 16 | space for saved registers
.align 4
.global SYM (_ISR_Handler)
SYM (_ISR_Handler):
| disable multitasking
addql #1,THREAD_DISPATCH_DISABLE_LEVEL
#if ( !defined(__mcoldfire__) )
moveml d0-d1/a0-a1,a7@- | save d0-d1,a0-a1
#else
lea a7@(-SAVED),a7
movm.l d0-d1/a0-a1,a7@ | save d0-d1,a0-a1
#endif
movew a7@(SAVED+FVO_OFFSET),d0 | d0 = F/VO
andl #0x03fc,d0 | d0 = vector offset in vbr
#if ( CPU_HAS_SOFTWARE_INTERRUPT_STACK == 1 )
| Make a0 point just above interrupt stack
movel INTERRUPT_STACK_HIGH,a0
cmpl INTERRUPT_STACK_LOW,a7 | stack below interrupt stack?
bcs.b 1f | yes, switch to interrupt stack
cmpl a0,a7 | stack above interrupt stack?
bcs.b 2f | no, do not switch stacks
1:
movel a7,a1 | copy task stack pointer
movel a0,a7 | switch to interrupt stack
movel a1,a7@- | store task stack pointer
| on interrupt stack
2:
#endif /* CPU_HAS_SOFTWARE_INTERRUPT_STACK == 1 */
addql #1,ISR_NEST_LEVEL | one nest level deeper
movel SYM (_ISR_Vector_table),a0 | a0= base of RTEMS table
#if ( M68K_HAS_PREINDEXING == 1 )
movel (a0,d0:w:1),a0 | a0 = address of user routine
#else
addal d0,a0 | a0 = address of vector
movel (a0),a0 | a0 = address of user routine
#endif
lsrl #2,d0 | d0 = vector number
movel d0,a7@- | push vector number
jbsr a0@ | invoke the user ISR
addql #4,a7 | remove vector number
subql #1,ISR_NEST_LEVEL | Reduce interrupt-nesting count
#if ( CPU_HAS_SOFTWARE_INTERRUPT_STACK == 1 )
movel INTERRUPT_STACK_HIGH,a0
subql #4,a0
cmpl a0,a7 | At top of interrupt stack?
bne.b 1f | No, do not restore task stack pointer
movel (a7),a7 | Restore task stack pointer
1:
#endif /* CPU_HAS_SOFTWARE_INTERRUPT_STACK == 1 */
subql #1,THREAD_DISPATCH_DISABLE_LEVEL
| unnest multitasking
bne.b exit | If dispatch disabled, exit
#if ( M68K_HAS_SEPARATE_STACKS == 1 )
movew #0xf000,d0 | isolate format nibble
andw a7@(SAVED+FVO_OFFSET),d0 | get F/VO
cmpiw #0x1000,d0 | is it a throwaway isf?
bne.b exit | NOT outer level, so branch
#else
/*
* If we have a CPU which allows a higher-priority interrupt to preempt a
* lower priority handler before the lower-priority handler can increment
* _Thread_Dispatch_disable_level then we must check the PC on the stack to
* see if it is _ISR_Handler. If it is we have the case of nesting interrupts
* without the dispatch level being incremented.
*/
#if ( !defined(__mcoldfire__) && !__mc68060__ )
cmpl #_ISR_Handler,a7@(SAVED+PC_OFFSET)
beq.b exit
#endif
#endif
tstb DISPATCH_NEEDED
| Is thread switch necessary?
beq.b exit | No, then exit
bframe:
| If sent, will be processed
#if ( M68K_HAS_SEPARATE_STACKS == 1 )
movec msp,a0 | a0 = master stack pointer
movew #0,a0@- | push format word
movel #SYM(_ISR_Dispatch),a0@- | push return addr
movew a0@(6),a0@- | push saved sr
movec a0,msp | set master stack pointer
#else
jsr SYM (_Thread_Dispatch) | Perform context switch
#endif
#if ( !defined(__mcoldfire__) )
exit: moveml a7@+,d0-d1/a0-a1 | restore d0-d1,a0-a1
#else
exit: moveml a7@,d0-d1/a0-a1 | restore d0-d1,a0-a1
lea a7@(SAVED),a7
#endif
#if ( M68K_HAS_VBR == 0 )
addql #2,a7 | pop format/id
#endif /* M68K_HAS_VBR */
rte | return to thread
| OR _Isr_dispatch
/*void _ISR_Dispatch()
*
* Entry point from the outermost interrupt service routine exit.
* The current stack is the supervisor mode stack if this processor
* has separate stacks.
*
* 1. save all registers not preserved across C calls.
* 2. invoke the _Thread_Dispatch routine to switch tasks
* or a signal to the currently executing task.
* 3. restore all registers not preserved across C calls.
* 4. return from interrupt
*/
.global SYM (_ISR_Dispatch)
SYM (_ISR_Dispatch):
#if ( !defined(__mcoldfire__) )
movml d0-d1/a0-a1,a7@-
jsr SYM (_Thread_Dispatch)
movml a7@+,d0-d1/a0-a1
#else
lea a7@(-SAVED),a7
movml d0-d1/a0-a1,a7@
jsr SYM (_Thread_Dispatch)
movml a7@,d0-d1/a0-a1
lea a7@(SAVED),a7
#endif
#if ( M68K_HAS_VBR == 0 )
addql #2,a7 | pop format/id
#endif /* M68K_HAS_VBR */
rte