forked from Imagelibrary/rtems
Use a per-CPU thread dispatch disable level. So instead of one global thread dispatch disable level we have now one instance per processor. This is a major performance improvement for SMP. On non-SMP configurations this may simplifiy the interrupt entry/exit code. The giant lock is still present, but it is now decoupled from the thread dispatching in _Thread_Dispatch(), _Thread_Handler(), _Thread_Restart_self() and the interrupt entry/exit. Access to the giant lock is now available via _Giant_Acquire() and _Giant_Release(). The giant lock is still implicitly acquired via _Thread_Dispatch_decrement_disable_level(). The giant lock is only acquired for high-level operations in interrupt handlers (e.g. release of a semaphore, sending of an event). As a side-effect this change fixes the lost thread dispatch necessary indication bug in _Thread_Dispatch(). A per-CPU thread dispatch disable level greatly simplifies the SMP support for the interrupt entry/exit code since no spin locks have to be acquired in this area. It is only necessary to get the current processor index and use this to calculate the address of the own per-CPU control. This reduces the interrupt latency considerably. All elements for the interrupt entry/exit code are now part of the Per_CPU_Control structure: thread dispatch disable level, ISR nest level and thread dispatch necessary. Nothing else is required (except CPU port specific stuff like on SPARC).
179 lines
5.1 KiB
C
179 lines
5.1 KiB
C
/**
|
|
* @file
|
|
*
|
|
* @brief Dispatch Thread
|
|
* @ingroup ScoreThread
|
|
*/
|
|
|
|
/*
|
|
* COPYRIGHT (c) 1989-2009.
|
|
* On-Line Applications Research Corporation (OAR).
|
|
*
|
|
* The license and distribution terms for this file may be
|
|
* found in the file LICENSE in this distribution or at
|
|
* http://www.rtems.com/license/LICENSE.
|
|
*/
|
|
|
|
#if HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include <rtems/score/threaddispatch.h>
|
|
#include <rtems/score/apiext.h>
|
|
#include <rtems/score/assert.h>
|
|
#include <rtems/score/isr.h>
|
|
#include <rtems/score/threadimpl.h>
|
|
#include <rtems/score/todimpl.h>
|
|
#include <rtems/score/userextimpl.h>
|
|
#include <rtems/score/wkspace.h>
|
|
|
|
void _Thread_Dispatch( void )
|
|
{
|
|
Per_CPU_Control *per_cpu;
|
|
Thread_Control *executing;
|
|
Thread_Control *heir;
|
|
ISR_Level level;
|
|
|
|
#if defined( RTEMS_SMP )
|
|
_ISR_Disable( level );
|
|
#endif
|
|
|
|
per_cpu = _Per_CPU_Get();
|
|
_Assert( per_cpu->thread_dispatch_disable_level == 0 );
|
|
per_cpu->thread_dispatch_disable_level = 1;
|
|
|
|
#if defined( RTEMS_SMP )
|
|
_ISR_Enable( level );
|
|
#endif
|
|
|
|
/*
|
|
* Now determine if we need to perform a dispatch on the current CPU.
|
|
*/
|
|
executing = per_cpu->executing;
|
|
_Per_CPU_ISR_disable_and_acquire( per_cpu, level );
|
|
#if defined( RTEMS_SMP )
|
|
/*
|
|
* On SMP the complete context switch must be atomic with respect to one
|
|
* processor. The scheduler must obtain the per-CPU lock to check if a
|
|
* thread is executing and to update the heir. This ensures that a thread
|
|
* cannot execute on more than one processor at a time. See also
|
|
* _Thread_Handler() since _Context_switch() may branch to this function.
|
|
*/
|
|
if ( per_cpu->dispatch_necessary ) {
|
|
#else
|
|
while ( per_cpu->dispatch_necessary ) {
|
|
#endif
|
|
heir = per_cpu->heir;
|
|
per_cpu->dispatch_necessary = false;
|
|
per_cpu->executing = heir;
|
|
#if defined( RTEMS_SMP )
|
|
executing->is_executing = false;
|
|
heir->is_executing = true;
|
|
#endif
|
|
|
|
/*
|
|
* When the heir and executing are the same, then we are being
|
|
* requested to do the post switch dispatching. This is normally
|
|
* done to dispatch signals.
|
|
*/
|
|
if ( heir == executing )
|
|
goto post_switch;
|
|
|
|
/*
|
|
* Since heir and executing are not the same, we need to do a real
|
|
* context switch.
|
|
*/
|
|
#if __RTEMS_ADA__
|
|
executing->rtems_ada_self = rtems_ada_self;
|
|
rtems_ada_self = heir->rtems_ada_self;
|
|
#endif
|
|
if ( heir->budget_algorithm == THREAD_CPU_BUDGET_ALGORITHM_RESET_TIMESLICE )
|
|
heir->cpu_time_budget = _Thread_Ticks_per_timeslice;
|
|
|
|
#if !defined( RTEMS_SMP )
|
|
_ISR_Enable( level );
|
|
#endif
|
|
|
|
#ifndef __RTEMS_USE_TICKS_FOR_STATISTICS__
|
|
_Thread_Update_cpu_time_used(
|
|
executing,
|
|
&per_cpu->time_of_last_context_switch
|
|
);
|
|
#else
|
|
{
|
|
_TOD_Get_uptime( &per_cpu->time_of_last_context_switch );
|
|
heir->cpu_time_used++;
|
|
}
|
|
#endif
|
|
|
|
#if !defined(__DYNAMIC_REENT__)
|
|
/*
|
|
* Switch libc's task specific data.
|
|
*/
|
|
if ( _Thread_libc_reent ) {
|
|
executing->libc_reent = *_Thread_libc_reent;
|
|
*_Thread_libc_reent = heir->libc_reent;
|
|
}
|
|
#endif
|
|
|
|
_User_extensions_Thread_switch( executing, heir );
|
|
|
|
/*
|
|
* If the CPU has hardware floating point, then we must address saving
|
|
* and restoring it as part of the context switch.
|
|
*
|
|
* The second conditional compilation section selects the algorithm used
|
|
* to context switch between floating point tasks. The deferred algorithm
|
|
* can be significantly better in a system with few floating point tasks
|
|
* because it reduces the total number of save and restore FP context
|
|
* operations. However, this algorithm can not be used on all CPUs due
|
|
* to unpredictable use of FP registers by some compilers for integer
|
|
* operations.
|
|
*/
|
|
|
|
#if ( CPU_HARDWARE_FP == TRUE ) || ( CPU_SOFTWARE_FP == TRUE )
|
|
#if ( CPU_USE_DEFERRED_FP_SWITCH != TRUE )
|
|
if ( executing->fp_context != NULL )
|
|
_Context_Save_fp( &executing->fp_context );
|
|
#endif
|
|
#endif
|
|
|
|
_Context_Switch( &executing->Registers, &heir->Registers );
|
|
|
|
#if ( CPU_HARDWARE_FP == TRUE ) || ( CPU_SOFTWARE_FP == TRUE )
|
|
#if ( CPU_USE_DEFERRED_FP_SWITCH == TRUE )
|
|
if ( (executing->fp_context != NULL) &&
|
|
!_Thread_Is_allocated_fp( executing ) ) {
|
|
if ( _Thread_Allocated_fp != NULL )
|
|
_Context_Save_fp( &_Thread_Allocated_fp->fp_context );
|
|
_Context_Restore_fp( &executing->fp_context );
|
|
_Thread_Allocated_fp = executing;
|
|
}
|
|
#else
|
|
if ( executing->fp_context != NULL )
|
|
_Context_Restore_fp( &executing->fp_context );
|
|
#endif
|
|
#endif
|
|
|
|
/*
|
|
* We have to obtain these values again after the context switch since the
|
|
* heir thread may have migrated from another processor. Values from the
|
|
* stack or non-volatile registers reflect the old execution environment.
|
|
*/
|
|
per_cpu = _Per_CPU_Get();
|
|
executing = per_cpu->executing;
|
|
|
|
#if !defined( RTEMS_SMP )
|
|
_ISR_Disable( level );
|
|
#endif
|
|
}
|
|
|
|
post_switch:
|
|
_Assert( per_cpu->thread_dispatch_disable_level == 1 );
|
|
per_cpu->thread_dispatch_disable_level = 0;
|
|
|
|
_Per_CPU_Release_and_ISR_enable( per_cpu, level );
|
|
|
|
_API_extensions_Run_post_switch( executing );
|
|
}
|