forked from Imagelibrary/rtems
2009-09-06 Sebastian Huber <Sebastian.Huber@embedded-brains.de>
* libcsupport/src/free.c, libmisc/stackchk/check.c, rtems/include/rtems/rtems/region.h, rtems/src/regioncreate.c, rtems/src/regionextend.c, rtems/src/regiongetinfo.c, rtems/src/regiongetsegment.c, rtems/src/regiongetsegmentsize.c, rtems/src/regionresizesegment.c, score/src/pheapallocate.c, score/src/pheapallocatealigned.c, score/src/pheapextend.c, score/src/pheapfree.c, score/src/pheapgetblocksize.c, score/src/pheapgetfreeinfo.c, score/src/pheapgetinfo.c, score/src/pheapgetsize.c, score/src/pheapinit.c, score/src/pheapresizeblock.c, score/src/pheapwalk.c: Update for heap API changes. * score/include/rtems/score/apimutex.h, score/include/rtems/score/object.h: Documentation. * score/include/rtems/score/heap.h, score/include/rtems/score/protectedheap.h, score/inline/rtems/score/heap.inl, score/src/heap.c, score/src/heapallocate.c, score/src/heapallocatealigned.c, score/src/heapextend.c, score/src/heapfree.c, score/src/heapgetfreeinfo.c, score/src/heapgetinfo.c, score/src/heapresizeblock.c, score/src/heapsizeofuserarea.c, score/src/heapwalk.c: Overall cleanup. Added boundary constraint to allocation function. More changes follow.
This commit is contained in:
@@ -1,3 +1,28 @@
|
||||
2009-09-06 Sebastian Huber <Sebastian.Huber@embedded-brains.de>
|
||||
|
||||
* libcsupport/src/free.c, libmisc/stackchk/check.c,
|
||||
rtems/include/rtems/rtems/region.h, rtems/src/regioncreate.c,
|
||||
rtems/src/regionextend.c, rtems/src/regiongetinfo.c,
|
||||
rtems/src/regiongetsegment.c, rtems/src/regiongetsegmentsize.c,
|
||||
rtems/src/regionresizesegment.c, score/src/pheapallocate.c,
|
||||
score/src/pheapallocatealigned.c, score/src/pheapextend.c,
|
||||
score/src/pheapfree.c, score/src/pheapgetblocksize.c,
|
||||
score/src/pheapgetfreeinfo.c, score/src/pheapgetinfo.c,
|
||||
score/src/pheapgetsize.c, score/src/pheapinit.c,
|
||||
score/src/pheapresizeblock.c, score/src/pheapwalk.c:
|
||||
Update for heap API changes.
|
||||
* score/include/rtems/score/apimutex.h,
|
||||
score/include/rtems/score/object.h: Documentation.
|
||||
* score/include/rtems/score/heap.h,
|
||||
score/include/rtems/score/protectedheap.h,
|
||||
score/inline/rtems/score/heap.inl, score/src/heap.c,
|
||||
score/src/heapallocate.c, score/src/heapallocatealigned.c,
|
||||
score/src/heapextend.c, score/src/heapfree.c,
|
||||
score/src/heapgetfreeinfo.c, score/src/heapgetinfo.c,
|
||||
score/src/heapresizeblock.c, score/src/heapsizeofuserarea.c,
|
||||
score/src/heapwalk.c: Overall cleanup. Added boundary constraint to
|
||||
allocation function. More changes follow.
|
||||
|
||||
2009-09-04 Sebastian Huber <Sebastian.Huber@embedded-brains.de>
|
||||
|
||||
* rtems/src/taskmode.c, sapi/src/exshutdown.c,
|
||||
|
||||
@@ -59,8 +59,8 @@ void free(
|
||||
if ( !_Protected_heap_Free( RTEMS_Malloc_Heap, ptr ) ) {
|
||||
printk( "Program heap: free of bad pointer %p -- range %p - %p \n",
|
||||
ptr,
|
||||
RTEMS_Malloc_Heap->begin,
|
||||
RTEMS_Malloc_Heap->end
|
||||
RTEMS_Malloc_Heap->area_begin,
|
||||
RTEMS_Malloc_Heap->area_end
|
||||
);
|
||||
}
|
||||
|
||||
|
||||
@@ -92,7 +92,7 @@ static inline bool Stack_check_Frame_pointer_in_range(
|
||||
|
||||
#else
|
||||
#define Stack_check_Get_pattern_area( _the_stack ) \
|
||||
((Stack_check_Control *) ((char *)(_the_stack)->area + HEAP_LAST_BLOCK_OVERHEAD))
|
||||
((Stack_check_Control *) ((char *)(_the_stack)->area + HEAP_BLOCK_HEADER_SIZE))
|
||||
|
||||
#define Stack_check_Calculate_used( _low, _size, _high_water) \
|
||||
( ((char *)(_low) + (_size)) - (char *)(_high_water) )
|
||||
|
||||
@@ -69,9 +69,9 @@ typedef struct {
|
||||
Objects_Control Object;
|
||||
Thread_queue_Control Wait_queue; /* waiting threads */
|
||||
void *starting_address; /* physical start addr */
|
||||
intptr_t length; /* physical length(bytes) */
|
||||
uint32_t page_size; /* in bytes */
|
||||
intptr_t maximum_segment_size; /* in bytes */
|
||||
uintptr_t length; /* physical length(bytes) */
|
||||
uintptr_t page_size; /* in bytes */
|
||||
uintptr_t maximum_segment_size; /* in bytes */
|
||||
rtems_attribute attribute_set;
|
||||
uint32_t number_of_used_blocks; /* blocks allocated */
|
||||
Heap_Control Memory;
|
||||
@@ -104,9 +104,9 @@ void _Region_Manager_initialization(void);
|
||||
rtems_status_code rtems_region_create(
|
||||
rtems_name name,
|
||||
void *starting_address,
|
||||
intptr_t length,
|
||||
uint32_t page_size,
|
||||
rtems_attribute attribute_set,
|
||||
uintptr_t length,
|
||||
uintptr_t page_size,
|
||||
rtems_attribute attribute_set,
|
||||
Objects_Id *id
|
||||
);
|
||||
|
||||
@@ -121,7 +121,7 @@ rtems_status_code rtems_region_create(
|
||||
rtems_status_code rtems_region_extend(
|
||||
Objects_Id id,
|
||||
void *starting_address,
|
||||
intptr_t length
|
||||
uintptr_t length
|
||||
);
|
||||
|
||||
/**
|
||||
@@ -184,10 +184,10 @@ rtems_status_code rtems_region_delete(
|
||||
*/
|
||||
rtems_status_code rtems_region_get_segment(
|
||||
Objects_Id id,
|
||||
intptr_t size,
|
||||
uintptr_t size,
|
||||
rtems_option option_set,
|
||||
rtems_interval timeout,
|
||||
void **segment
|
||||
void **segment
|
||||
);
|
||||
|
||||
/**
|
||||
@@ -199,7 +199,7 @@ rtems_status_code rtems_region_get_segment(
|
||||
rtems_status_code rtems_region_get_segment_size(
|
||||
Objects_Id id,
|
||||
void *segment,
|
||||
intptr_t *size
|
||||
uintptr_t *size
|
||||
);
|
||||
|
||||
/**
|
||||
@@ -241,8 +241,8 @@ rtems_status_code rtems_region_return_segment(
|
||||
rtems_status_code rtems_region_resize_segment(
|
||||
Objects_Id id,
|
||||
void *segment,
|
||||
intptr_t size,
|
||||
intptr_t *old_size
|
||||
uintptr_t size,
|
||||
uintptr_t *old_size
|
||||
);
|
||||
|
||||
#ifndef __RTEMS_APPLICATION__
|
||||
|
||||
@@ -50,8 +50,8 @@
|
||||
rtems_status_code rtems_region_create(
|
||||
rtems_name name,
|
||||
void *starting_address,
|
||||
intptr_t length,
|
||||
uint32_t page_size,
|
||||
uintptr_t length,
|
||||
uintptr_t page_size,
|
||||
rtems_attribute attribute_set,
|
||||
Objects_Id *id
|
||||
)
|
||||
|
||||
@@ -45,10 +45,10 @@
|
||||
rtems_status_code rtems_region_extend(
|
||||
Objects_Id id,
|
||||
void *starting_address,
|
||||
intptr_t length
|
||||
uintptr_t length
|
||||
)
|
||||
{
|
||||
intptr_t amount_extended;
|
||||
uintptr_t amount_extended;
|
||||
Heap_Extend_status heap_status;
|
||||
Objects_Locations location;
|
||||
rtems_status_code return_status;
|
||||
|
||||
@@ -60,11 +60,8 @@ rtems_status_code rtems_region_get_information(
|
||||
switch ( location ) {
|
||||
|
||||
case OBJECTS_LOCAL:
|
||||
if ( _Heap_Get_information( &the_region->Memory, the_info ) !=
|
||||
HEAP_GET_INFORMATION_SUCCESSFUL )
|
||||
return_status = RTEMS_INVALID_ADDRESS;
|
||||
else
|
||||
return_status = RTEMS_SUCCESSFUL;
|
||||
_Heap_Get_information( &the_region->Memory, the_info );
|
||||
return_status = RTEMS_SUCCESSFUL;
|
||||
break;
|
||||
|
||||
#if defined(RTEMS_MULTIPROCESSING)
|
||||
|
||||
@@ -47,7 +47,7 @@
|
||||
|
||||
rtems_status_code rtems_region_get_segment(
|
||||
Objects_Id id,
|
||||
intptr_t size,
|
||||
uintptr_t size,
|
||||
rtems_option option_set,
|
||||
rtems_interval timeout,
|
||||
void **segment
|
||||
|
||||
@@ -45,7 +45,7 @@
|
||||
rtems_status_code rtems_region_get_segment_size(
|
||||
Objects_Id id,
|
||||
void *segment,
|
||||
intptr_t *size
|
||||
uintptr_t *size
|
||||
)
|
||||
{
|
||||
Objects_Locations location;
|
||||
|
||||
@@ -46,13 +46,13 @@
|
||||
rtems_status_code rtems_region_resize_segment(
|
||||
Objects_Id id,
|
||||
void *segment,
|
||||
intptr_t size,
|
||||
intptr_t *old_size
|
||||
uintptr_t size,
|
||||
uintptr_t *old_size
|
||||
)
|
||||
{
|
||||
intptr_t avail_size;
|
||||
uintptr_t avail_size;
|
||||
Objects_Locations location;
|
||||
intptr_t osize;
|
||||
uintptr_t osize;
|
||||
rtems_status_code return_status = RTEMS_INTERNAL_ERROR;
|
||||
Heap_Resize_status status;
|
||||
register Region_Control *the_region;
|
||||
|
||||
@@ -1,9 +1,9 @@
|
||||
/**
|
||||
* @file rtems/score/apimutex.h
|
||||
* @file
|
||||
*
|
||||
* This include file contains all the constants and structures associated
|
||||
* with the API Mutex Handler. This handler is used by API level
|
||||
* routines to manage mutual exclusion.
|
||||
* @ingroup ScoreAPIMutex
|
||||
*
|
||||
* @brief API Mutex Handler API.
|
||||
*/
|
||||
|
||||
/*
|
||||
@@ -20,115 +20,97 @@
|
||||
#ifndef _RTEMS_SCORE_APIMUTEX_H
|
||||
#define _RTEMS_SCORE_APIMUTEX_H
|
||||
|
||||
/**
|
||||
* @defgroup ScoreAPIMutex API Mutex Handler
|
||||
*
|
||||
* This handler encapsulates functionality which provides mutexes to be used
|
||||
* in the implementation of API functionality.
|
||||
*/
|
||||
/**@{*/
|
||||
|
||||
#ifdef __cplusplus
|
||||
extern "C" {
|
||||
#endif
|
||||
|
||||
/**
|
||||
* @defgroup ScoreAPIMutex API Mutex Handler
|
||||
*
|
||||
* @ingroup Score
|
||||
*
|
||||
* @brief Provides routines to ensure mutual exclusion on API level.
|
||||
*
|
||||
* @{
|
||||
*/
|
||||
|
||||
#include <rtems/score/coremutex.h>
|
||||
#include <rtems/score/isr.h>
|
||||
#include <rtems/score/object.h>
|
||||
|
||||
/**
|
||||
* The following defines the control block used to manage each API mutex.
|
||||
* An API Mutex is an aggregration of an Object and a SuperCore Mutex.
|
||||
* @brief Control block used to manage each API mutex.
|
||||
*/
|
||||
typedef struct {
|
||||
/** This field allows each API Mutex to be a full-fledged RTEMS object. */
|
||||
Objects_Control Object;
|
||||
/** This field contains the SuperCore mutex information. */
|
||||
CORE_mutex_Control Mutex;
|
||||
} API_Mutex_Control;
|
||||
/**
|
||||
* @brief Allows each API Mutex to be a full-fledged RTEMS object.
|
||||
*/
|
||||
Objects_Control Object;
|
||||
|
||||
/**
|
||||
* Contains the SuperCore mutex information.
|
||||
*/
|
||||
CORE_mutex_Control Mutex;
|
||||
} API_Mutex_Control;
|
||||
|
||||
/**
|
||||
* The following variable is the information control block used to manage
|
||||
* this class of objects.
|
||||
* @brief Information control block used to manage this class of objects.
|
||||
*/
|
||||
SCORE_EXTERN Objects_Information _API_Mutex_Information;
|
||||
SCORE_EXTERN Objects_Information _API_Mutex_Information;
|
||||
|
||||
/**
|
||||
* This routine performs the initialization necessary for this handler.
|
||||
* @brief Performs the initialization necessary for this handler.
|
||||
*
|
||||
* @param[in] maximum_mutexes is the maximum number of API mutexes
|
||||
* that may exist at any time
|
||||
* The value @a maximum_mutexes is the maximum number of API mutexes that may
|
||||
* exist at any time.
|
||||
*/
|
||||
void _API_Mutex_Initialization(
|
||||
uint32_t maximum_mutexes
|
||||
);
|
||||
void _API_Mutex_Initialization( uint32_t maximum_mutexes );
|
||||
|
||||
/**
|
||||
* This routine allocates an API mutex from the inactive set.
|
||||
*
|
||||
* @param[out] the_mutex will contain the allocated mutex.
|
||||
* @brief Allocates an API mutex from the inactive set and returns it in
|
||||
* @a mutex.
|
||||
*/
|
||||
void _API_Mutex_Allocate(
|
||||
API_Mutex_Control **the_mutex
|
||||
);
|
||||
void _API_Mutex_Allocate( API_Mutex_Control **mutex );
|
||||
|
||||
/**
|
||||
* This routine acquires the specified API mutex.
|
||||
*
|
||||
* @param[in] the_mutex is the mutex to acquire.
|
||||
* @brief Acquires the specified API mutex @a mutex.
|
||||
*/
|
||||
void _API_Mutex_Lock(
|
||||
API_Mutex_Control *the_mutex
|
||||
);
|
||||
void _API_Mutex_Lock( API_Mutex_Control *mutex );
|
||||
|
||||
/**
|
||||
* This routine releases the specified API mutex.
|
||||
*
|
||||
* @param[in] the_mutex is the mutex to release.
|
||||
* @brief Releases the specified API mutex @a mutex.
|
||||
*/
|
||||
void _API_Mutex_Unlock(
|
||||
API_Mutex_Control *the_mutex
|
||||
);
|
||||
void _API_Mutex_Unlock( API_Mutex_Control *mutex );
|
||||
|
||||
/** @} */
|
||||
|
||||
/**
|
||||
* This variable points to the API Mutex instance that is used
|
||||
* to protect all memory allocation and deallocation in RTEMS.
|
||||
* @defgroup ScoreAllocatorMutex RTEMS Allocator Mutex
|
||||
*
|
||||
* @note When the APIs all use this for allocation and deallocation
|
||||
* protection, then this possibly should be renamed and moved to a
|
||||
* higher level in the hierarchy.
|
||||
* @ingroup ScoreAPIMutex
|
||||
*
|
||||
* @brief Protection for all memory allocations and deallocations in RTEMS.
|
||||
*
|
||||
* When the APIs all use this for allocation and deallocation protection, then
|
||||
* this possibly should be renamed and moved to a higher level in the
|
||||
* hierarchy.
|
||||
*
|
||||
* @{
|
||||
*/
|
||||
SCORE_EXTERN API_Mutex_Control *_RTEMS_Allocator_Mutex;
|
||||
|
||||
/**
|
||||
* This macro locks the RTEMS Allocation Mutex.
|
||||
*
|
||||
* @see _RTEMS_Allocator_Mutex
|
||||
*/
|
||||
SCORE_EXTERN API_Mutex_Control *_RTEMS_Allocator_Mutex;
|
||||
|
||||
#define _RTEMS_Lock_allocator() \
|
||||
_API_Mutex_Lock( _RTEMS_Allocator_Mutex )
|
||||
|
||||
/**
|
||||
* This macro unlocks the RTEMS Allocation Mutex.
|
||||
*
|
||||
* @see _RTEMS_Allocator_Mutex
|
||||
*/
|
||||
#define _RTEMS_Unlock_allocator() \
|
||||
_API_Mutex_Unlock( _RTEMS_Allocator_Mutex )
|
||||
|
||||
/*
|
||||
* There are no inlines for this handler.
|
||||
*/
|
||||
|
||||
#ifndef __RTEMS_APPLICATION__
|
||||
/* #include <rtems/score/apimutex.inl> */
|
||||
#endif
|
||||
/** @} */
|
||||
|
||||
#ifdef __cplusplus
|
||||
}
|
||||
#endif
|
||||
|
||||
/*!@}*/
|
||||
|
||||
#endif
|
||||
/* end of include file */
|
||||
|
||||
@@ -1,7 +1,9 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* Heap Handler API.
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler API.
|
||||
*/
|
||||
|
||||
/*
|
||||
@@ -25,7 +27,9 @@ extern "C" {
|
||||
/**
|
||||
* @defgroup ScoreHeap Heap Handler
|
||||
*
|
||||
* The Heap Handler provides a heap.
|
||||
* @ingroup Score
|
||||
*
|
||||
* @brief The Heap Handler provides a heap.
|
||||
*
|
||||
* A heap is a doubly linked list of variable size blocks which are allocated
|
||||
* using the first fit method. Garbage collection is performed each time a
|
||||
@@ -33,17 +37,22 @@ extern "C" {
|
||||
* information for both allocated and free blocks is contained in the heap
|
||||
* area. A heap control structure contains control information for the heap.
|
||||
*
|
||||
* FIXME: The alignment routines could be made faster should we require only
|
||||
* powers of two to be supported both for 'page_size' and for 'alignment'
|
||||
* arguments. However, both workspace and malloc heaps are initialized with
|
||||
* CPU_HEAP_ALIGNMENT as 'page_size', and while all the BSPs seem to use
|
||||
* The alignment routines could be made faster should we require only powers of
|
||||
* two to be supported both for page size, alignment and boundary arguments.
|
||||
* However, both workspace and malloc heaps are initialized with
|
||||
* CPU_HEAP_ALIGNMENT as page size, and while all the BSPs seem to use
|
||||
* CPU_ALIGNMENT (that is power of two) as CPU_HEAP_ALIGNMENT, for whatever
|
||||
* reason CPU_HEAP_ALIGNMENT is only required to be multiple of CPU_ALIGNMENT
|
||||
* and explicitly not required to be a power of two.
|
||||
*
|
||||
* There are two kinds of blocks. One sort describes a free block from which
|
||||
* we can allocate memory. The other blocks are used and contain allocated
|
||||
* memory. The free blocks are accessible via a list of free blocks.
|
||||
* we can allocate memory. The other blocks are used and provide an allocated
|
||||
* memory area. The free blocks are accessible via a list of free blocks.
|
||||
*
|
||||
* Blocks or areas cover a continuous set of memory addresses. They have a
|
||||
* begin and end address. The end address is not part of the set. The size of
|
||||
* a block or area equals the distance between the begin and end address in
|
||||
* units of bytes.
|
||||
*
|
||||
* Free blocks look like:
|
||||
* <table>
|
||||
@@ -83,7 +92,10 @@ extern "C" {
|
||||
* <table>
|
||||
* <tr><th>Label</th><th colspan=2>Content</th></tr>
|
||||
* <tr><td>heap->begin</td><td colspan=2>heap area begin address</td></tr>
|
||||
* <tr><td>first_block->prev_size</td><td colspan=2>arbitrary value</td></tr>
|
||||
* <tr>
|
||||
* <td>first_block->prev_size</td>
|
||||
* <td colspan=2>page size (the value is arbitrary)</td>
|
||||
* </tr>
|
||||
* <tr>
|
||||
* <td>first_block->size</td>
|
||||
* <td colspan=2>size available for allocation
|
||||
@@ -100,7 +112,7 @@ extern "C" {
|
||||
* </tr>
|
||||
* <tr>
|
||||
* <td>second_block->size</td>
|
||||
* <td colspan=2>arbitrary size | @c HEAP_PREV_BLOCK_FREE</td>
|
||||
* <td colspan=2>page size (the value is arbitrary)</td>
|
||||
* </tr>
|
||||
* <tr><td>heap->end</td><td colspan=2>heap area end address</td></tr>
|
||||
* </table>
|
||||
@@ -108,6 +120,23 @@ extern "C" {
|
||||
* @{
|
||||
*/
|
||||
|
||||
/**
|
||||
* @brief See also @ref Heap_Block.size_and_flag.
|
||||
*/
|
||||
#define HEAP_PREV_BLOCK_USED ((uintptr_t) 1)
|
||||
|
||||
/**
|
||||
* @brief Offset from the block begin up to the block size field
|
||||
* (@ref Heap_Block.size_and_flag).
|
||||
*/
|
||||
#define HEAP_BLOCK_SIZE_OFFSET sizeof(uintptr_t)
|
||||
|
||||
/**
|
||||
* @brief The block header consists of the two size fields
|
||||
* (@ref Heap_Block.prev_size and @ref Heap_Block.size_and_flag).
|
||||
*/
|
||||
#define HEAP_BLOCK_HEADER_SIZE (sizeof(uintptr_t) * 2)
|
||||
|
||||
/**
|
||||
* @brief Description for free or used blocks.
|
||||
*/
|
||||
@@ -119,6 +148,11 @@ typedef struct Heap_Block {
|
||||
* This field is only valid if the previous block is free. This case is
|
||||
* indicated by a cleared @c HEAP_PREV_BLOCK_USED flag in the
|
||||
* @a size_and_flag field of the current block.
|
||||
*
|
||||
* In a used block only the @a size_and_flag field needs to be valid. The
|
||||
* @a prev_size field of the current block is maintained by the previous
|
||||
* block. The current block can use the @a prev_size field in the next block
|
||||
* for allocation.
|
||||
*/
|
||||
uintptr_t prev_size;
|
||||
|
||||
@@ -157,86 +191,119 @@ typedef struct Heap_Block {
|
||||
struct Heap_Block *prev;
|
||||
} Heap_Block;
|
||||
|
||||
#define HEAP_PREV_BLOCK_USED ((uintptr_t) 1)
|
||||
|
||||
#define HEAP_PREV_BLOCK_FREE ((uintptr_t) 0)
|
||||
|
||||
/**
|
||||
* @brief Offset from the block begin up to the block size field.
|
||||
*/
|
||||
#define HEAP_BLOCK_SIZE_OFFSET (sizeof(uintptr_t))
|
||||
|
||||
/**
|
||||
* @brief Offset from the block begin up to the allocated area begin.
|
||||
*/
|
||||
#define HEAP_BLOCK_ALLOC_AREA_OFFSET (sizeof(uintptr_t) * 2)
|
||||
|
||||
#define HEAP_BLOCK_USED_OVERHEAD (sizeof(uintptr_t) * 2)
|
||||
|
||||
#define HEAP_LAST_BLOCK_OVERHEAD HEAP_BLOCK_ALLOC_AREA_OFFSET
|
||||
|
||||
/**
|
||||
* Run-time heap statistics.
|
||||
* @brief Run-time heap statistics.
|
||||
*
|
||||
* @note (double)searches/allocs gives mean number of searches per alloc while
|
||||
* max_search gives maximum number of searches ever performed on a
|
||||
* single call to alloc.
|
||||
*
|
||||
* @note the statistics is always gathered. I believe the imposed overhead is
|
||||
* rather small. Feel free to make it compile-time option if you think
|
||||
* the overhead is too high for your application.
|
||||
* The value @a searches / @a allocs gives the mean number of searches per
|
||||
* allocation, while @a max_search gives maximum number of searches ever
|
||||
* performed on a single allocation call.
|
||||
*/
|
||||
typedef struct {
|
||||
/** instance number of this heap */
|
||||
/**
|
||||
* @brief Instance number of this heap.
|
||||
*/
|
||||
uint32_t instance;
|
||||
/** the size of the memory for heap */
|
||||
|
||||
/**
|
||||
* @brief The size of the memory for heap.
|
||||
*/
|
||||
uintptr_t size;
|
||||
/** current free size */
|
||||
|
||||
/**
|
||||
* @brief Current free size.
|
||||
*/
|
||||
uintptr_t free_size;
|
||||
/** minimum free size ever */
|
||||
|
||||
/**
|
||||
* @brief Minimum free size ever.
|
||||
*/
|
||||
uintptr_t min_free_size;
|
||||
/** current number of free blocks */
|
||||
|
||||
/**
|
||||
* @brief Current number of free blocks.
|
||||
*/
|
||||
uint32_t free_blocks;
|
||||
/** maximum number of free blocks ever */
|
||||
|
||||
/**
|
||||
* @brief Maximum number of free blocks ever.
|
||||
*/
|
||||
uint32_t max_free_blocks;
|
||||
/** current number of used blocks */
|
||||
|
||||
/**
|
||||
* @brief Current number of used blocks.
|
||||
*/
|
||||
uint32_t used_blocks;
|
||||
/** maximum number of blocks searched ever */
|
||||
|
||||
/**
|
||||
* @brief Maximum number of blocks searched ever.
|
||||
*/
|
||||
uint32_t max_search;
|
||||
/** total number of successful calls to alloc */
|
||||
|
||||
/**
|
||||
* @brief Total number of successful calls to alloc.
|
||||
*/
|
||||
uint32_t allocs;
|
||||
/** total number of searches ever */
|
||||
|
||||
/**
|
||||
* @brief Total number of searches ever.
|
||||
*/
|
||||
uint32_t searches;
|
||||
/** total number of suceessful calls to free */
|
||||
|
||||
/**
|
||||
* @brief Total number of suceessful calls to free.
|
||||
*/
|
||||
uint32_t frees;
|
||||
/** total number of successful resizes */
|
||||
|
||||
/**
|
||||
* @brief Total number of successful resizes.
|
||||
*/
|
||||
uint32_t resizes;
|
||||
} Heap_Statistics;
|
||||
|
||||
/**
|
||||
* Control block used to manage each heap.
|
||||
* @brief Control block used to manage a heap.
|
||||
*/
|
||||
typedef struct {
|
||||
/** head and tail of circular list of free blocks */
|
||||
Heap_Block free_list;
|
||||
/** allocation unit and alignment */
|
||||
uintptr_t page_size;
|
||||
/** minimum block size aligned on page_size */
|
||||
uintptr_t min_block_size;
|
||||
/** first address of memory for the heap */
|
||||
uintptr_t begin;
|
||||
/** first address past end of memory for the heap */
|
||||
uintptr_t end;
|
||||
/** first valid block address in the heap */
|
||||
Heap_Block *start;
|
||||
/** last valid block address in the heap */
|
||||
Heap_Block *final;
|
||||
/** run-time statistics */
|
||||
uintptr_t area_begin;
|
||||
uintptr_t area_end;
|
||||
Heap_Block *first_block;
|
||||
Heap_Block *last_block;
|
||||
Heap_Statistics stats;
|
||||
} Heap_Control;
|
||||
|
||||
/**
|
||||
* Status codes for _Heap_Extend
|
||||
* @brief Information about blocks.
|
||||
*/
|
||||
typedef struct {
|
||||
/**
|
||||
* @brief Number of blocks of this type.
|
||||
*/
|
||||
uint32_t number;
|
||||
|
||||
/**
|
||||
* @brief Largest block of this type.
|
||||
*/
|
||||
uint32_t largest;
|
||||
|
||||
/**
|
||||
* @brief Total size of the blocks of this type.
|
||||
*/
|
||||
uint32_t total;
|
||||
} Heap_Information;
|
||||
|
||||
/**
|
||||
* @brief Information block returned by _Heap_Get_information().
|
||||
*/
|
||||
typedef struct {
|
||||
Heap_Information Free;
|
||||
Heap_Information Used;
|
||||
} Heap_Information_block;
|
||||
|
||||
/**
|
||||
* @brief See _Heap_Extend().
|
||||
*/
|
||||
typedef enum {
|
||||
HEAP_EXTEND_SUCCESSFUL,
|
||||
@@ -245,7 +312,7 @@ typedef enum {
|
||||
} Heap_Extend_status;
|
||||
|
||||
/**
|
||||
* Status codes for _Heap_Resize_block
|
||||
* @brief See _Heap_Resize_block().
|
||||
*/
|
||||
typedef enum {
|
||||
HEAP_RESIZE_SUCCESSFUL,
|
||||
@@ -254,40 +321,8 @@ typedef enum {
|
||||
} Heap_Resize_status;
|
||||
|
||||
/**
|
||||
* Status codes for _Heap_Get_information
|
||||
*/
|
||||
typedef enum {
|
||||
HEAP_GET_INFORMATION_SUCCESSFUL = 0,
|
||||
HEAP_GET_INFORMATION_BLOCK_ERROR
|
||||
} Heap_Get_information_status;
|
||||
|
||||
/**
|
||||
* Information block returned by the Heap routines used to
|
||||
* obtain heap information. This information is returned about
|
||||
* either free or used blocks.
|
||||
*/
|
||||
typedef struct {
|
||||
/** Number of blocks of this type. */
|
||||
uint32_t number;
|
||||
/** Largest blocks of this type. */
|
||||
uint32_t largest;
|
||||
/** Total size of the blocks of this type. */
|
||||
uint32_t total;
|
||||
} Heap_Information;
|
||||
|
||||
/**
|
||||
* Information block returned by _Heap_Get_information
|
||||
*/
|
||||
typedef struct {
|
||||
/** This field is information on the used blocks in the heap. */
|
||||
Heap_Information Free;
|
||||
/** This field is information on the used blocks in the heap. */
|
||||
Heap_Information Used;
|
||||
} Heap_Information_block;
|
||||
|
||||
/**
|
||||
* Initializes the @a heap control block to manage the area starting at
|
||||
* @a area_begin of @a area_size bytes.
|
||||
* @brief Initializes the heap control block @a heap to manage the area
|
||||
* starting at @a area_begin of size @a area_size bytes.
|
||||
*
|
||||
* Blocks of memory are allocated from the heap in multiples of @a page_size
|
||||
* byte units. If the @a page_size is equal to zero or is not multiple of
|
||||
@@ -303,16 +338,13 @@ uintptr_t _Heap_Initialize(
|
||||
);
|
||||
|
||||
/**
|
||||
* This routine grows @a heap memory area using the size bytes which
|
||||
* begin at @a starting_address.
|
||||
* @brief Extends the memory area of the heap @a heap using the memory area
|
||||
* starting at @a area_begin of size @a area_size bytes.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the memory
|
||||
* to add to the heap
|
||||
* @param[in] size is the size in bytes of the memory area to add
|
||||
* @param[in] amount_extended points to a user area to return the
|
||||
* @return a status indicating success or the reason for failure
|
||||
* @return *size filled in with the amount of memory added to the heap
|
||||
* The extended space available for allocation will be returned in
|
||||
* @a amount_extended.
|
||||
*
|
||||
* The memory area must start at the end of the currently used memory area.
|
||||
*/
|
||||
Heap_Extend_status _Heap_Extend(
|
||||
Heap_Control *heap,
|
||||
@@ -322,139 +354,115 @@ Heap_Extend_status _Heap_Extend(
|
||||
);
|
||||
|
||||
/**
|
||||
* This function attempts to allocate a block of @a size bytes from
|
||||
* @a heap. If insufficient memory is free in @a heap to allocate
|
||||
* a block of the requested size, then NULL is returned.
|
||||
* @brief Allocates a memory area of size @a size bytes.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] size is the amount of memory to allocate in bytes
|
||||
* @return NULL if unsuccessful and a pointer to the block if successful
|
||||
*/
|
||||
void *_Heap_Allocate( Heap_Control *heap, uintptr_t size );
|
||||
|
||||
/**
|
||||
* This function attempts to allocate a memory block of @a size bytes from
|
||||
* @a heap so that the start of the user memory is aligned on the
|
||||
* @a alignment boundary. If @a alignment is 0, it is set to CPU_ALIGNMENT.
|
||||
* Any other value of @a alignment is taken "as is", i.e., even odd
|
||||
* alignments are possible.
|
||||
* Returns pointer to the start of the memory block if success, NULL if
|
||||
* failure.
|
||||
* If the alignment parameter @a alignment is not equal to zero, the allocated
|
||||
* memory area will begin at an address aligned by this value.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] size is the amount of memory to allocate in bytes
|
||||
* @param[in] alignment the required alignment
|
||||
* @return NULL if unsuccessful and a pointer to the block if successful
|
||||
* If the boundary parameter @a boundary is not equal to zero, the allocated
|
||||
* memory area will fulfill a boundary constraint. The boudnary value
|
||||
* specifies the set of addresses which are aligned by the boundary value. The
|
||||
* interior of the allocated memory area will not contain an element of this
|
||||
* set. The begin or end address of the area may be a member of the set.
|
||||
*
|
||||
* A size value of zero will return a unique address which may be freed with
|
||||
* _Heap_Free().
|
||||
*
|
||||
* Returns a pointer to the begin of the allocated memory area, or @c NULL if
|
||||
* no memory is available or the parameters are inconsistent.
|
||||
*/
|
||||
void *_Heap_Allocate_aligned(
|
||||
void *_Heap_Allocate_aligned_with_boundary(
|
||||
Heap_Control *heap,
|
||||
uintptr_t size,
|
||||
uintptr_t alignment
|
||||
uintptr_t alignment,
|
||||
uintptr_t boundary
|
||||
);
|
||||
|
||||
#define _Heap_Allocate_aligned( heap, size, alignment ) \
|
||||
_Heap_Allocate_aligned_with_boundary( heap, size, alignment, 0 )
|
||||
|
||||
#define _Heap_Allocate( heap, size ) \
|
||||
_Heap_Allocate_aligned_with_boundary( heap, size, 0, 0 )
|
||||
|
||||
/**
|
||||
* @brief Frees the allocated memory area starting at @a addr in the heap
|
||||
* @a heap.
|
||||
*
|
||||
* Inappropriate values for @a addr may corrupt the heap.
|
||||
*
|
||||
* Returns @c true in case of success, and @c false otherwise.
|
||||
*/
|
||||
bool _Heap_Free( Heap_Control *heap, void *addr );
|
||||
|
||||
/**
|
||||
* @brief Walks the heap @a heap to verify its integrity.
|
||||
*
|
||||
* If @a dump is @c true, then diagnostic messages will be printed to standard
|
||||
* output. In this case @a source is used to mark the output lines.
|
||||
*
|
||||
* Returns @c true if no errors occured, and @c false if the heap is corrupt.
|
||||
*/
|
||||
bool _Heap_Walk(
|
||||
Heap_Control *heap,
|
||||
int source,
|
||||
bool dump
|
||||
);
|
||||
|
||||
/**
|
||||
* This function sets @a size to the size of the block of allocatable area
|
||||
* which begins at @a starting_address. The size returned in @a *size could
|
||||
* be greater than the size requested for allocation.
|
||||
* Returns true if the @a starting_address is in the heap, and false
|
||||
* otherwise.
|
||||
* @brief Returns information about used and free blocks for the heap @a heap
|
||||
* in @a info.
|
||||
*/
|
||||
void _Heap_Get_information(
|
||||
Heap_Control *heap,
|
||||
Heap_Information_block *info
|
||||
);
|
||||
|
||||
/**
|
||||
* @brief Returns information about free blocks for the heap @a heap in
|
||||
* @a info.
|
||||
*/
|
||||
void _Heap_Get_free_information(
|
||||
Heap_Control *heap,
|
||||
Heap_Information *info
|
||||
);
|
||||
|
||||
/**
|
||||
* @brief Returns the size of the allocatable memory area starting at @a addr
|
||||
* in @a size.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the user block
|
||||
* to obtain the size of
|
||||
* @param[in] size points to a user area to return the size in
|
||||
* @return true if successfully able to determine the size, false otherwise
|
||||
* @return *size filled in with the size of the user area for this block
|
||||
* The size value may be greater than the initially requested size in
|
||||
* _Heap_Allocate_aligned_with_boundary().
|
||||
*
|
||||
* Inappropriate values for @a addr will not corrupt the heap, but may yield
|
||||
* invalid size values.
|
||||
*
|
||||
* Returns @a true if successful, and @c false otherwise.
|
||||
*/
|
||||
bool _Heap_Size_of_alloc_area(
|
||||
Heap_Control *heap,
|
||||
void *area_begin,
|
||||
void *addr,
|
||||
uintptr_t *size
|
||||
);
|
||||
|
||||
/**
|
||||
* This function tries to resize in place the block that is pointed to by the
|
||||
* @a starting_address to the new @a size.
|
||||
* @brief Resizes the block of the allocated memory area starting at @a addr.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the user block
|
||||
* to be resized
|
||||
* @param[in] size is the new size
|
||||
* @param[in] old_mem_size points to a user area to return the size of the
|
||||
* user memory area of the block before resizing.
|
||||
* @param[in] avail_mem_size points to a user area to return the size of
|
||||
* the user memory area of the free block that has been enlarged or
|
||||
* created due to resizing, 0 if none.
|
||||
* @return HEAP_RESIZE_SUCCESSFUL if successfully able to resize the block,
|
||||
* HEAP_RESIZE_UNSATISFIED if the block can't be resized in place,
|
||||
* HEAP_RESIZE_FATAL_ERROR if failure
|
||||
* @return *old_mem_size filled in with the size of the user memory area of
|
||||
* the block before resizing.
|
||||
* @return *avail_mem_size filled in with the size of the user memory area
|
||||
* of the free block that has been enlarged or created due to
|
||||
* resizing, 0 if none.
|
||||
* The new memory area will have a size of at least @a size bytes. A resize
|
||||
* may be impossible and depends on the current heap usage.
|
||||
*
|
||||
* The size available for allocation in the current block before the resize
|
||||
* will be returned in @a old_size. The size available for allocation in
|
||||
* the resized block will be returned in @a new_size. If the resize was not
|
||||
* successful, then a value of zero will be returned in @a new_size.
|
||||
*
|
||||
* Inappropriate values for @a addr may corrupt the heap.
|
||||
*/
|
||||
Heap_Resize_status _Heap_Resize_block(
|
||||
Heap_Control *heap,
|
||||
void *starting_address,
|
||||
uintptr_t size,
|
||||
uintptr_t *old_mem_size,
|
||||
uintptr_t *avail_mem_size
|
||||
);
|
||||
|
||||
/**
|
||||
* This routine returns the block of memory which begins
|
||||
* at @a alloc_area_begin to @a heap. Any coalescing which is
|
||||
* possible with the freeing of this routine is performed.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] start_address is the starting address of the user block
|
||||
* to free
|
||||
* @return true if successfully freed, false otherwise
|
||||
*/
|
||||
bool _Heap_Free( Heap_Control *heap, void *alloc_area_begin );
|
||||
|
||||
/**
|
||||
* This routine walks the heap to verify its integrity.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] source is a user specified integer which may be used to
|
||||
* indicate where in the application this was invoked from
|
||||
* @param[in] do_dump is set to true if errors should be printed
|
||||
* @return true if the test passed fine, false otherwise.
|
||||
*/
|
||||
bool _Heap_Walk(
|
||||
Heap_Control *heap,
|
||||
int source,
|
||||
bool do_dump
|
||||
);
|
||||
|
||||
/**
|
||||
* This routine walks the heap and tots up the free and allocated
|
||||
* sizes.
|
||||
*
|
||||
* @param[in] heap pointer to heap header
|
||||
* @param[in] the_info pointer to a status information area
|
||||
* @return *the_info is filled with status information
|
||||
* @return 0=success, otherwise heap is corrupt.
|
||||
*/
|
||||
Heap_Get_information_status _Heap_Get_information(
|
||||
Heap_Control *heap,
|
||||
Heap_Information_block *the_info
|
||||
);
|
||||
|
||||
/**
|
||||
* This heap routine returns information about the free blocks
|
||||
* in the specified heap.
|
||||
*
|
||||
* @param[in] heap pointer to heap header.
|
||||
* @param[in] info pointer to the free block information.
|
||||
*
|
||||
* @return free block information filled in.
|
||||
*/
|
||||
void _Heap_Get_free_information(
|
||||
Heap_Control *heap,
|
||||
Heap_Information *info
|
||||
void *addr,
|
||||
uintptr_t size,
|
||||
uintptr_t *old_size,
|
||||
uintptr_t *new_size
|
||||
);
|
||||
|
||||
#if !defined(__RTEMS_APPLICATION__)
|
||||
@@ -462,36 +470,20 @@ void _Heap_Get_free_information(
|
||||
#include <rtems/score/heap.inl>
|
||||
|
||||
/**
|
||||
* @brief Returns the minimal block size for a block which may contain an area
|
||||
* of size @a alloc_size for allocation, or zero in case of an overflow.
|
||||
* @brief Allocates the memory area starting at @a alloc_begin of size
|
||||
* @a alloc_size bytes in the block @a block.
|
||||
*
|
||||
* Uses the heap values @a page_size and @a min_block_size.
|
||||
* The block may be split up into multiple blocks.
|
||||
*
|
||||
* Inappropriate values for @a alloc_begin or @a alloc_size may corrupt the
|
||||
* heap.
|
||||
*
|
||||
* Returns the block containing the allocated memory area.
|
||||
*/
|
||||
uintptr_t _Heap_Calc_block_size(
|
||||
uintptr_t alloc_size,
|
||||
uintptr_t page_size,
|
||||
uintptr_t min_block_size
|
||||
);
|
||||
|
||||
/**
|
||||
* This method allocates a block of size @a alloc_size from @a the_block
|
||||
* belonging to @a heap. Split @a the_block if possible, otherwise
|
||||
* allocate it entirely. When split, make the lower part used, and leave
|
||||
* the upper part free.
|
||||
*
|
||||
* This is an internal routines used by _Heap_Allocate() and
|
||||
* _Heap_Allocate_aligned(). Refer to 'heap.c' for details.
|
||||
*
|
||||
* @param[in] heap is the heap to operate upon
|
||||
* @param[in] the_block is the block to allocates the requested size from
|
||||
* @param[in] alloc_size is the requested number of bytes to take out of
|
||||
* the block
|
||||
*
|
||||
* @return This methods returns the size of the allocated block.
|
||||
*/
|
||||
uintptr_t _Heap_Block_allocate(
|
||||
Heap_Block *_Heap_Block_allocate(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block,
|
||||
uintptr_t alloc_begin,
|
||||
uintptr_t alloc_size
|
||||
);
|
||||
|
||||
|
||||
@@ -29,6 +29,12 @@
|
||||
extern "C" {
|
||||
#endif
|
||||
|
||||
/**
|
||||
* @defgroup Score SuperCore
|
||||
*
|
||||
* @brief Provides services for all APIs.
|
||||
*/
|
||||
|
||||
/**
|
||||
* The following type defines the control block used to manage
|
||||
* object names.
|
||||
|
||||
@@ -1,9 +1,12 @@
|
||||
/**
|
||||
* @file rtems/score/protectedheap.h
|
||||
* @file
|
||||
*
|
||||
* This include file contains the information pertaining to the
|
||||
* Protected Heap Handler.
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler API.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -20,210 +23,125 @@
|
||||
#include <rtems/score/heap.h>
|
||||
#include <rtems/score/apimutex.h>
|
||||
|
||||
/**
|
||||
* @defgroup ScoreProtHeap Protected Heap Handler
|
||||
*
|
||||
* This handler encapsulates functionality which provides the foundation
|
||||
* Protected Heap services.
|
||||
*
|
||||
* It is a simple wrapper for the help with the addition of the
|
||||
* allocation mutex being used for protection.
|
||||
*/
|
||||
/**@{*/
|
||||
|
||||
#ifdef __cplusplus
|
||||
extern "C" {
|
||||
#endif
|
||||
|
||||
/**
|
||||
* This routine initializes @a the_heap record to manage the
|
||||
* contiguous heap of @a size bytes which starts at @a starting_address.
|
||||
* Blocks of memory are allocated from the heap in multiples of
|
||||
* @a page_size byte units. If @a page_size is 0 or is not multiple of
|
||||
* CPU_ALIGNMENT, it's aligned up to the nearest CPU_ALIGNMENT boundary.
|
||||
* @defgroup ScoreProtHeap Protected Heap Handler
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the memory for
|
||||
* the heap
|
||||
* @param[in] size is the size in bytes of the memory area for the heap
|
||||
* @param[in] page_size is the size in bytes of the allocation unit
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @return This method returns the maximum memory available. If
|
||||
* unsuccessful, 0 will be returned.
|
||||
* @brief Provides protected heap services.
|
||||
*
|
||||
* The @ref ScoreAllocatorMutex is used to protect the heap accesses.
|
||||
*
|
||||
* @{
|
||||
*/
|
||||
static inline uint32_t _Protected_heap_Initialize(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t size,
|
||||
uint32_t page_size
|
||||
|
||||
/**
|
||||
* @brief See _Heap_Initialize().
|
||||
*/
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Protected_heap_Initialize(
|
||||
Heap_Control *heap,
|
||||
void *area_begin,
|
||||
uintptr_t area_size,
|
||||
uintptr_t page_size
|
||||
)
|
||||
{
|
||||
return _Heap_Initialize( the_heap, starting_address, size, page_size );
|
||||
return _Heap_Initialize( heap, area_begin, area_size, page_size );
|
||||
}
|
||||
|
||||
/**
|
||||
* This routine grows @a the_heap memory area using the size bytes which
|
||||
* begin at @a starting_address.
|
||||
* @brief See _Heap_Extend().
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the memory
|
||||
* to add to the heap
|
||||
* @param[in] size is the size in bytes of the memory area to add
|
||||
* @return a status indicating success or the reason for failure
|
||||
* Returns @a true in case of success, and @a false otherwise.
|
||||
*/
|
||||
bool _Protected_heap_Extend(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t size
|
||||
Heap_Control *heap,
|
||||
void *area_begin,
|
||||
uintptr_t area_size
|
||||
);
|
||||
|
||||
/**
|
||||
* This function attempts to allocate a block of @a size bytes from
|
||||
* @a the_heap. If insufficient memory is free in @a the_heap to allocate
|
||||
* a block of the requested size, then NULL is returned.
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] size is the amount of memory to allocate in bytes
|
||||
* @return NULL if unsuccessful and a pointer to the block if successful
|
||||
* @brief See _Heap_Allocate_aligned_with_boundary().
|
||||
*/
|
||||
void *_Protected_heap_Allocate(
|
||||
Heap_Control *the_heap,
|
||||
intptr_t size
|
||||
Heap_Control *heap,
|
||||
uintptr_t size
|
||||
);
|
||||
|
||||
/**
|
||||
* This function attempts to allocate a memory block of @a size bytes from
|
||||
* @a the_heap so that the start of the user memory is aligned on the
|
||||
* @a alignment boundary. If @a alignment is 0, it is set to CPU_ALIGNMENT.
|
||||
* Any other value of @a alignment is taken "as is", i.e., even odd
|
||||
* alignments are possible.
|
||||
* Returns pointer to the start of the memory block if success, NULL if
|
||||
* failure.
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] size is the amount of memory to allocate in bytes
|
||||
* @param[in] alignment the required alignment
|
||||
* @return NULL if unsuccessful and a pointer to the block if successful
|
||||
* @brief See _Heap_Allocate_aligned_with_boundary().
|
||||
*/
|
||||
void *_Protected_heap_Allocate_aligned(
|
||||
Heap_Control *the_heap,
|
||||
intptr_t size,
|
||||
uint32_t alignment
|
||||
Heap_Control *heap,
|
||||
uintptr_t size,
|
||||
uintptr_t alignment
|
||||
);
|
||||
|
||||
/**
|
||||
* This function sets @a *size to the size of the block of user memory
|
||||
* which begins at @a starting_address. The size returned in @a *size could
|
||||
* be greater than the size requested for allocation.
|
||||
* Returns true if the @a starting_address is in the heap, and false
|
||||
* otherwise.
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the user block
|
||||
* to obtain the size of
|
||||
* @param[in] size points to a user area to return the size in
|
||||
* @return true if successfully able to determine the size, false otherwise
|
||||
* @return *size filled in with the size of the user area for this block
|
||||
* @brief See _Heap_Size_of_alloc_area().
|
||||
*/
|
||||
bool _Protected_heap_Get_block_size(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t *size
|
||||
Heap_Control *heap,
|
||||
void *addr,
|
||||
uintptr_t *size
|
||||
);
|
||||
|
||||
/**
|
||||
* This function tries to resize in place the block that is pointed to by the
|
||||
* @a starting_address to the new @a size.
|
||||
* @brief See _Heap_Resize_block().
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] starting_address is the starting address of the user block
|
||||
* to be resized
|
||||
* @param[in] size is the new size
|
||||
*
|
||||
* @return true if successfully able to resize the block.
|
||||
* false if the block can't be resized in place.
|
||||
* Returns @a true in case of success, and @a false otherwise.
|
||||
*/
|
||||
bool _Protected_heap_Resize_block(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t size
|
||||
Heap_Control *heap,
|
||||
void *addr,
|
||||
uintptr_t size
|
||||
);
|
||||
|
||||
/**
|
||||
* This routine returns the block of memory which begins
|
||||
* at @a starting_address to @a the_heap. Any coalescing which is
|
||||
* possible with the freeing of this routine is performed.
|
||||
* @brief See _Heap_Free().
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] start_address is the starting address of the user block
|
||||
* to free
|
||||
* @return true if successfully freed, false otherwise
|
||||
* Returns @a true in case of success, and @a false otherwise.
|
||||
*/
|
||||
bool _Protected_heap_Free(
|
||||
Heap_Control *the_heap,
|
||||
void *start_address
|
||||
);
|
||||
bool _Protected_heap_Free( Heap_Control *heap, void *addr );
|
||||
|
||||
/**
|
||||
* This routine walks the heap to verify its integrity.
|
||||
*
|
||||
* @param[in] the_heap is the heap to operate upon
|
||||
* @param[in] source is a user specified integer which may be used to
|
||||
* indicate where in the application this was invoked from
|
||||
* @param[in] do_dump is set to true if errors should be printed
|
||||
* @return true if the test passed fine, false otherwise.
|
||||
* @brief See _Heap_Walk().
|
||||
*/
|
||||
bool _Protected_heap_Walk(
|
||||
Heap_Control *the_heap,
|
||||
int source,
|
||||
bool do_dump
|
||||
);
|
||||
bool _Protected_heap_Walk( Heap_Control *heap, int source, bool dump );
|
||||
|
||||
/**
|
||||
* This routine walks the heap and tots up the free and allocated
|
||||
* sizes.
|
||||
* @brief See _Heap_Get_information().
|
||||
*
|
||||
* @param[in] the_heap pointer to heap header
|
||||
* @param[in] the_info pointer to a status information area
|
||||
*
|
||||
* @return true if successfully able to return information
|
||||
* Returns @a true in case of success, and @a false otherwise.
|
||||
*/
|
||||
bool _Protected_heap_Get_information(
|
||||
Heap_Control *the_heap,
|
||||
Heap_Information_block *the_info
|
||||
Heap_Control *heap,
|
||||
Heap_Information_block *info
|
||||
);
|
||||
|
||||
/**
|
||||
* This heap routine returns information about the free blocks
|
||||
* in the specified heap.
|
||||
* @brief See _Heap_Get_free_information().
|
||||
*
|
||||
* @param[in] the_heap pointer to heap header.
|
||||
* @param[in] info pointer to the free block information.
|
||||
*
|
||||
* @return free block information filled in.
|
||||
* Returns @a true in case of success, and @a false otherwise.
|
||||
*/
|
||||
bool _Protected_heap_Get_free_information(
|
||||
Heap_Control *the_heap,
|
||||
Heap_Information *info
|
||||
Heap_Control *heap,
|
||||
Heap_Information *info
|
||||
);
|
||||
|
||||
/**
|
||||
* This function returns the maximum size of the protected heap.
|
||||
*
|
||||
* @param[in] the_heap points to the heap being operated upon
|
||||
*
|
||||
* @return This method returns the total amount of memory
|
||||
* allocated to the heap.
|
||||
* @brief See _Heap_Get_size().
|
||||
*/
|
||||
uint32_t _Protected_heap_Get_size(
|
||||
Heap_Control *the_heap
|
||||
);
|
||||
uintptr_t _Protected_heap_Get_size( Heap_Control *heap );
|
||||
|
||||
/** @} */
|
||||
|
||||
#ifdef __cplusplus
|
||||
}
|
||||
#endif
|
||||
|
||||
/**@}*/
|
||||
|
||||
#endif
|
||||
/* end of include file */
|
||||
|
||||
@@ -1,8 +1,9 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @brief Static inline implementations of the inlined routines from the heap
|
||||
* handler.
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler API.
|
||||
*/
|
||||
|
||||
/*
|
||||
@@ -41,17 +42,17 @@ RTEMS_INLINE_ROUTINE Heap_Block *_Heap_Free_list_tail( Heap_Control *heap )
|
||||
return &heap->free_list;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE Heap_Block *_Heap_First_free_block( Heap_Control *heap )
|
||||
RTEMS_INLINE_ROUTINE Heap_Block *_Heap_Free_list_first( Heap_Control *heap )
|
||||
{
|
||||
return _Heap_Free_list_head(heap)->next;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE Heap_Block *_Heap_Last_free_block( Heap_Control *heap )
|
||||
RTEMS_INLINE_ROUTINE Heap_Block *_Heap_Free_list_last( Heap_Control *heap )
|
||||
{
|
||||
return _Heap_Free_list_tail(heap)->prev;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE void _Heap_Block_remove_from_free_list( Heap_Block *block )
|
||||
RTEMS_INLINE_ROUTINE void _Heap_Free_list_remove( Heap_Block *block )
|
||||
{
|
||||
Heap_Block *next = block->next;
|
||||
Heap_Block *prev = block->prev;
|
||||
@@ -60,7 +61,7 @@ RTEMS_INLINE_ROUTINE void _Heap_Block_remove_from_free_list( Heap_Block *block )
|
||||
next->prev = prev;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE void _Heap_Block_replace_in_free_list(
|
||||
RTEMS_INLINE_ROUTINE void _Heap_Free_list_replace(
|
||||
Heap_Block *old_block,
|
||||
Heap_Block *new_block
|
||||
)
|
||||
@@ -75,16 +76,16 @@ RTEMS_INLINE_ROUTINE void _Heap_Block_replace_in_free_list(
|
||||
prev->next = new_block;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE void _Heap_Block_insert_after(
|
||||
Heap_Block *prev_block,
|
||||
RTEMS_INLINE_ROUTINE void _Heap_Free_list_insert_after(
|
||||
Heap_Block *block_before,
|
||||
Heap_Block *new_block
|
||||
)
|
||||
{
|
||||
Heap_Block *next = prev_block->next;
|
||||
Heap_Block *next = block_before->next;
|
||||
|
||||
new_block->next = next;
|
||||
new_block->prev = prev_block;
|
||||
prev_block->next = new_block;
|
||||
new_block->prev = block_before;
|
||||
block_before->next = new_block;
|
||||
next->prev = new_block;
|
||||
}
|
||||
|
||||
@@ -122,60 +123,64 @@ RTEMS_INLINE_ROUTINE uintptr_t _Heap_Align_down(
|
||||
* @brief Returns the block which is @a offset away from @a block.
|
||||
*/
|
||||
RTEMS_INLINE_ROUTINE Heap_Block *_Heap_Block_at(
|
||||
const Heap_Block *block,
|
||||
Heap_Block *block,
|
||||
uintptr_t offset
|
||||
)
|
||||
{
|
||||
return (Heap_Block *) ((uintptr_t) block + offset);
|
||||
}
|
||||
|
||||
/**
|
||||
* @brief Returns the begin of the allocatable area of @a block.
|
||||
*/
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Alloc_area_of_block(
|
||||
Heap_Block *block
|
||||
const Heap_Block *block
|
||||
)
|
||||
{
|
||||
return (uintptr_t) block + HEAP_BLOCK_ALLOC_AREA_OFFSET;
|
||||
return (uintptr_t) block + HEAP_BLOCK_HEADER_SIZE;
|
||||
}
|
||||
|
||||
/**
|
||||
* @brief Returns the block associated with the allocatable area starting at
|
||||
* @a alloc_area_begin inside a heap with a page size of @a page_size.
|
||||
*/
|
||||
RTEMS_INLINE_ROUTINE Heap_Block *_Heap_Block_of_alloc_area(
|
||||
uintptr_t alloc_area_begin,
|
||||
uintptr_t alloc_begin,
|
||||
uintptr_t page_size
|
||||
)
|
||||
{
|
||||
return (Heap_Block *) (_Heap_Align_down( alloc_area_begin, page_size )
|
||||
- HEAP_BLOCK_ALLOC_AREA_OFFSET);
|
||||
return (Heap_Block *) (_Heap_Align_down( alloc_begin, page_size )
|
||||
- HEAP_BLOCK_HEADER_SIZE);
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE bool _Heap_Is_prev_used( Heap_Block *block )
|
||||
RTEMS_INLINE_ROUTINE bool _Heap_Is_prev_used( const Heap_Block *block )
|
||||
{
|
||||
return block->size_and_flag & HEAP_PREV_BLOCK_USED;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Block_size( Heap_Block *block )
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Block_size( const Heap_Block *block )
|
||||
{
|
||||
return block->size_and_flag & ~HEAP_PREV_BLOCK_USED;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE bool _Heap_Is_block_in_heap(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block
|
||||
const Heap_Control *heap,
|
||||
const Heap_Block *block
|
||||
)
|
||||
{
|
||||
return _Addresses_Is_in_range( block, heap->start, heap->final );
|
||||
return (uintptr_t) block >= (uintptr_t) heap->first_block
|
||||
&& (uintptr_t) block <= (uintptr_t) heap->last_block;
|
||||
}
|
||||
|
||||
/**
|
||||
* @brief Returns the maximum size of the heap.
|
||||
* @brief Returns the heap area size.
|
||||
*/
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Get_size( Heap_Control *heap )
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Get_size( const Heap_Control *heap )
|
||||
{
|
||||
return (uintptr_t) heap->end - (uintptr_t) heap->begin;
|
||||
return heap->area_end - heap->area_begin;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Max( uintptr_t a, uintptr_t b )
|
||||
{
|
||||
return a > b ? a : b;
|
||||
}
|
||||
|
||||
RTEMS_INLINE_ROUTINE uintptr_t _Heap_Min( uintptr_t a, uintptr_t b )
|
||||
{
|
||||
return a < b ? a : b;
|
||||
}
|
||||
|
||||
/** @} */
|
||||
|
||||
@@ -1,9 +1,17 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2009.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
* Copyright (c) 2009 embedded brains GmbH.
|
||||
*
|
||||
* The license and distribution terms for this file may be
|
||||
* found in the file LICENSE in this distribution or at
|
||||
* http://www.rtems.com/license/LICENSE.
|
||||
@@ -18,6 +26,10 @@
|
||||
#include <rtems/system.h>
|
||||
#include <rtems/score/heap.h>
|
||||
|
||||
#if CPU_ALIGNMENT == 0 || CPU_ALIGNMENT % 4 != 0
|
||||
#error "invalid CPU_ALIGNMENT value"
|
||||
#endif
|
||||
|
||||
static uint32_t instance = 0;
|
||||
|
||||
/*PAGE
|
||||
@@ -113,16 +125,19 @@ static uint32_t instance = 0;
|
||||
|
||||
uintptr_t _Heap_Initialize(
|
||||
Heap_Control *heap,
|
||||
void *area_begin,
|
||||
uintptr_t area_size,
|
||||
void *heap_area_begin_ptr,
|
||||
uintptr_t heap_area_size,
|
||||
uintptr_t page_size
|
||||
)
|
||||
{
|
||||
Heap_Statistics * const stats = &heap->stats;
|
||||
uintptr_t heap_area_begin = (uintptr_t) area_begin;
|
||||
uintptr_t heap_area_end = heap_area_begin + area_size;
|
||||
uintptr_t alloc_area_begin = heap_area_begin + HEAP_BLOCK_ALLOC_AREA_OFFSET;
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
uintptr_t const heap_area_begin = (uintptr_t) heap_area_begin_ptr;
|
||||
uintptr_t const heap_area_end = heap_area_begin + heap_area_size;
|
||||
uintptr_t alloc_area_begin = heap_area_begin + HEAP_BLOCK_HEADER_SIZE;
|
||||
uintptr_t alloc_area_size = 0;
|
||||
uintptr_t first_block_begin = 0;
|
||||
uintptr_t first_block_size = 0;
|
||||
uintptr_t min_block_size = 0;
|
||||
uintptr_t overhead = 0;
|
||||
Heap_Block *first_block = NULL;
|
||||
Heap_Block *second_block = NULL;
|
||||
@@ -132,47 +147,50 @@ uintptr_t _Heap_Initialize(
|
||||
} else {
|
||||
page_size = _Heap_Align_up( page_size, CPU_ALIGNMENT );
|
||||
}
|
||||
|
||||
heap->min_block_size = _Heap_Align_up( sizeof( Heap_Block ), page_size );
|
||||
min_block_size = _Heap_Align_up( sizeof( Heap_Block ), page_size );
|
||||
|
||||
alloc_area_begin = _Heap_Align_up( alloc_area_begin, page_size );
|
||||
overhead = HEAP_LAST_BLOCK_OVERHEAD
|
||||
+ (alloc_area_begin - HEAP_BLOCK_ALLOC_AREA_OFFSET - heap_area_begin);
|
||||
alloc_area_size = _Heap_Align_down ( area_size - overhead, page_size );
|
||||
first_block_begin = alloc_area_begin - HEAP_BLOCK_HEADER_SIZE;
|
||||
overhead = HEAP_BLOCK_HEADER_SIZE + (first_block_begin - heap_area_begin);
|
||||
first_block_size = heap_area_size - overhead;
|
||||
first_block_size = _Heap_Align_down ( first_block_size, page_size );
|
||||
alloc_area_size = first_block_size - HEAP_BLOCK_HEADER_SIZE;
|
||||
|
||||
if (
|
||||
heap_area_end < heap_area_begin
|
||||
|| area_size < overhead
|
||||
|| alloc_area_size == 0
|
||||
|| heap_area_size <= overhead
|
||||
|| first_block_size < min_block_size
|
||||
) {
|
||||
/* Invalid area or area too small */
|
||||
return 0;
|
||||
}
|
||||
|
||||
heap->page_size = page_size;
|
||||
heap->begin = heap_area_begin;
|
||||
heap->end = heap_area_end;
|
||||
|
||||
/* First block */
|
||||
first_block = _Heap_Block_of_alloc_area( alloc_area_begin, page_size );
|
||||
first_block = (Heap_Block *) first_block_begin;
|
||||
first_block->prev_size = page_size;
|
||||
first_block->size_and_flag = alloc_area_size | HEAP_PREV_BLOCK_USED;
|
||||
first_block->size_and_flag = first_block_size | HEAP_PREV_BLOCK_USED;
|
||||
first_block->next = _Heap_Free_list_tail( heap );
|
||||
first_block->prev = _Heap_Free_list_head( heap );
|
||||
_Heap_Free_list_head( heap )->next = first_block;
|
||||
_Heap_Free_list_tail( heap )->prev = first_block;
|
||||
heap->start = first_block;
|
||||
|
||||
/* Second and last block */
|
||||
second_block = _Heap_Block_at( first_block, alloc_area_size );
|
||||
second_block->prev_size = alloc_area_size;
|
||||
second_block->size_and_flag = page_size | HEAP_PREV_BLOCK_FREE;
|
||||
heap->final = second_block;
|
||||
second_block = _Heap_Block_at( first_block, first_block_size );
|
||||
second_block->prev_size = first_block_size;
|
||||
second_block->size_and_flag = page_size;
|
||||
|
||||
/* Heap control */
|
||||
heap->page_size = page_size;
|
||||
heap->min_block_size = min_block_size;
|
||||
heap->area_begin = heap_area_begin;
|
||||
heap->area_end = heap_area_end;
|
||||
heap->first_block = first_block;
|
||||
heap->last_block = second_block;
|
||||
_Heap_Free_list_head( heap )->next = first_block;
|
||||
_Heap_Free_list_tail( heap )->prev = first_block;
|
||||
|
||||
/* Statistics */
|
||||
stats->size = area_size;
|
||||
stats->free_size = alloc_area_size;
|
||||
stats->min_free_size = alloc_area_size;
|
||||
stats->size = heap_area_size;
|
||||
stats->free_size = first_block_size;
|
||||
stats->min_free_size = first_block_size;
|
||||
stats->free_blocks = 1;
|
||||
stats->max_free_blocks = 1;
|
||||
stats->used_blocks = 0;
|
||||
@@ -183,9 +201,9 @@ uintptr_t _Heap_Initialize(
|
||||
stats->resizes = 0;
|
||||
stats->instance = instance++;
|
||||
|
||||
_HAssert( _Heap_Is_aligned( CPU_ALIGNMENT, 4 ));
|
||||
_HAssert( _Heap_Is_aligned( heap->page_size, CPU_ALIGNMENT ));
|
||||
_HAssert( _Heap_Is_aligned( heap->min_block_size, page_size ));
|
||||
_HAssert( _Heap_Is_aligned( CPU_ALIGNMENT, 4 ) );
|
||||
_HAssert( _Heap_Is_aligned( heap->page_size, CPU_ALIGNMENT ) );
|
||||
_HAssert( _Heap_Is_aligned( heap->min_block_size, page_size ) );
|
||||
_HAssert(
|
||||
_Heap_Is_aligned( _Heap_Alloc_area_of_block( first_block ), page_size )
|
||||
);
|
||||
@@ -193,72 +211,142 @@ uintptr_t _Heap_Initialize(
|
||||
_Heap_Is_aligned( _Heap_Alloc_area_of_block( second_block ), page_size )
|
||||
);
|
||||
|
||||
if ( !_Heap_Walk( heap, 0, false ) ) {
|
||||
_Heap_Walk( heap, 0, true );
|
||||
}
|
||||
|
||||
return alloc_area_size;
|
||||
}
|
||||
|
||||
uintptr_t _Heap_Calc_block_size(
|
||||
uintptr_t alloc_size,
|
||||
uintptr_t page_size,
|
||||
uintptr_t min_block_size)
|
||||
{
|
||||
uintptr_t block_size =
|
||||
_Heap_Align_up( alloc_size + HEAP_BLOCK_USED_OVERHEAD, page_size );
|
||||
|
||||
if (block_size < min_block_size) {
|
||||
block_size = min_block_size;
|
||||
}
|
||||
|
||||
if (block_size > alloc_size) {
|
||||
return block_size;
|
||||
} else {
|
||||
/* Integer overflow occured */
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
uintptr_t _Heap_Block_allocate(
|
||||
static Heap_Block *_Heap_Block_split(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block,
|
||||
uintptr_t alloc_size
|
||||
)
|
||||
{
|
||||
Heap_Statistics * const stats = &heap->stats;
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const min_block_size = heap->min_block_size;
|
||||
uintptr_t const min_alloc_size = min_block_size - HEAP_BLOCK_HEADER_SIZE;
|
||||
|
||||
uintptr_t const block_size = _Heap_Block_size( block );
|
||||
uintptr_t const unused_size = block_size - alloc_size;
|
||||
Heap_Block *next_block = _Heap_Block_at( block, block_size );
|
||||
|
||||
_HAssert( _Heap_Is_aligned( block_size, heap->page_size ));
|
||||
_HAssert( _Heap_Is_aligned( alloc_size, heap->page_size ));
|
||||
_HAssert( alloc_size <= block_size );
|
||||
_HAssert( _Heap_Is_prev_used( block ));
|
||||
uintptr_t const used_size =
|
||||
_Heap_Max( alloc_size, min_alloc_size ) + HEAP_BLOCK_HEADER_SIZE;
|
||||
uintptr_t const used_block_size = _Heap_Align_up( used_size, page_size );
|
||||
|
||||
if (unused_size >= heap->min_block_size) {
|
||||
/*
|
||||
* Split the block so that the upper part is still free, and the lower part
|
||||
* becomes used. This is slightly less optimal than leaving the lower part
|
||||
* free as it requires replacing block in the free blocks list, but it
|
||||
* makes it possible to reuse this code in the _Heap_Resize_block().
|
||||
*/
|
||||
Heap_Block *new_block = _Heap_Block_at( block, alloc_size );
|
||||
block->size_and_flag = alloc_size | HEAP_PREV_BLOCK_USED;
|
||||
new_block->size_and_flag = unused_size | HEAP_PREV_BLOCK_USED;
|
||||
next_block->prev_size = unused_size;
|
||||
_Heap_Block_replace_in_free_list( block, new_block );
|
||||
uintptr_t const free_size = block_size + HEAP_BLOCK_SIZE_OFFSET - used_size;
|
||||
uintptr_t const free_size_limit = min_block_size + HEAP_BLOCK_SIZE_OFFSET;
|
||||
|
||||
Heap_Block *const next_block = _Heap_Block_at( block, block_size );
|
||||
|
||||
_HAssert( used_size <= block_size + HEAP_BLOCK_SIZE_OFFSET );
|
||||
_HAssert( used_size + free_size == block_size + HEAP_BLOCK_SIZE_OFFSET );
|
||||
|
||||
if ( free_size >= free_size_limit ) {
|
||||
uintptr_t const free_block_size = block_size - used_block_size;
|
||||
Heap_Block *const free_block = _Heap_Block_at( block, used_block_size );
|
||||
|
||||
_HAssert( used_block_size + free_block_size == block_size );
|
||||
|
||||
block->size_and_flag = used_block_size
|
||||
| (block->size_and_flag & HEAP_PREV_BLOCK_USED);
|
||||
free_block->size_and_flag = free_block_size | HEAP_PREV_BLOCK_USED;
|
||||
next_block->prev_size = free_block_size;
|
||||
|
||||
return free_block;
|
||||
} else {
|
||||
next_block->size_and_flag |= HEAP_PREV_BLOCK_USED;
|
||||
alloc_size = block_size;
|
||||
_Heap_Block_remove_from_free_list( block );
|
||||
|
||||
return NULL;
|
||||
}
|
||||
}
|
||||
|
||||
static Heap_Block *_Heap_Block_allocate_from_begin(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block,
|
||||
uintptr_t alloc_size
|
||||
)
|
||||
{
|
||||
Heap_Block *const free_block = _Heap_Block_split( heap, block, alloc_size );
|
||||
|
||||
if ( free_block != NULL ) {
|
||||
_Heap_Free_list_replace( block, free_block );
|
||||
} else {
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
|
||||
_Heap_Free_list_remove( block );
|
||||
|
||||
/* Statistics */
|
||||
--stats->free_blocks;
|
||||
}
|
||||
|
||||
return block;
|
||||
}
|
||||
|
||||
static Heap_Block *_Heap_Block_allocate_from_end(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block,
|
||||
uintptr_t alloc_begin,
|
||||
uintptr_t alloc_size
|
||||
)
|
||||
{
|
||||
uintptr_t const block_begin = (uintptr_t) block;
|
||||
uintptr_t block_size = _Heap_Block_size( block );
|
||||
uintptr_t block_end = block_begin + block_size;
|
||||
|
||||
Heap_Block *const new_block =
|
||||
_Heap_Block_of_alloc_area( alloc_begin, heap->page_size );
|
||||
uintptr_t const new_block_begin = (uintptr_t) new_block;
|
||||
uintptr_t const new_block_size = block_end - new_block_begin;
|
||||
|
||||
Heap_Block *free_block = NULL;
|
||||
|
||||
block_end = new_block_begin;
|
||||
block_size = block_end - block_begin;
|
||||
|
||||
_HAssert( block_size >= heap->min_block_size );
|
||||
_HAssert( new_block_size >= heap->min_block_size );
|
||||
|
||||
block->size_and_flag = block_size | HEAP_PREV_BLOCK_USED;
|
||||
new_block->prev_size = block_size;
|
||||
new_block->size_and_flag = new_block_size;
|
||||
|
||||
free_block = _Heap_Block_split( heap, new_block, alloc_size );
|
||||
if ( free_block != NULL ) {
|
||||
_Heap_Free_list_insert_after( block, free_block );
|
||||
}
|
||||
|
||||
return new_block;
|
||||
}
|
||||
|
||||
Heap_Block *_Heap_Block_allocate(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block,
|
||||
uintptr_t alloc_begin,
|
||||
uintptr_t alloc_size
|
||||
)
|
||||
{
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
uintptr_t const alloc_area_begin = _Heap_Alloc_area_of_block( block );
|
||||
uintptr_t const alloc_area_offset = alloc_begin - alloc_area_begin;
|
||||
|
||||
_HAssert( _Heap_Is_prev_used( block ) );
|
||||
_HAssert( alloc_area_begin <= alloc_begin );
|
||||
|
||||
if ( alloc_area_offset < heap->page_size ) {
|
||||
alloc_size += alloc_area_offset;
|
||||
|
||||
block = _Heap_Block_allocate_from_begin( heap, block, alloc_size );
|
||||
} else {
|
||||
block = _Heap_Block_allocate_from_end( heap, block, alloc_begin, alloc_size );
|
||||
}
|
||||
|
||||
/* Statistics */
|
||||
++stats->used_blocks;
|
||||
stats->free_size -= alloc_size;
|
||||
if(stats->min_free_size > stats->free_size) {
|
||||
stats->free_size -= _Heap_Block_size( block );
|
||||
if ( stats->min_free_size > stats->free_size ) {
|
||||
stats->min_free_size = stats->free_size;
|
||||
}
|
||||
|
||||
return alloc_size;
|
||||
return block;
|
||||
}
|
||||
|
||||
@@ -1,9 +1,17 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-1999.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
* Copyright (c) 2009 embedded brains GmbH.
|
||||
*
|
||||
* The license and distribution terms for this file may be
|
||||
* found in the file LICENSE in this distribution or at
|
||||
* http://www.rtems.com/license/LICENSE.
|
||||
@@ -19,48 +27,204 @@
|
||||
#include <rtems/score/sysstate.h>
|
||||
#include <rtems/score/heap.h>
|
||||
|
||||
void *_Heap_Allocate( Heap_Control *heap, uintptr_t size )
|
||||
#ifdef RTEMS_HEAP_DEBUG
|
||||
static void _Heap_Check_allocation(
|
||||
const Heap_Control *heap,
|
||||
const Heap_Block *block,
|
||||
uintptr_t alloc_begin,
|
||||
uintptr_t alloc_size,
|
||||
uintptr_t alignment,
|
||||
uintptr_t boundary
|
||||
)
|
||||
{
|
||||
uintptr_t const min_block_size = heap->min_block_size;
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
|
||||
uintptr_t const block_begin = (uintptr_t) block;
|
||||
uintptr_t const block_size = _Heap_Block_size( block );
|
||||
uintptr_t const block_end = block_begin + block_size;
|
||||
|
||||
uintptr_t const alloc_end = alloc_begin + alloc_size;
|
||||
|
||||
uintptr_t const alloc_area_begin = _Heap_Alloc_area_of_block( block );
|
||||
uintptr_t const alloc_area_offset = alloc_begin - alloc_area_begin;
|
||||
uintptr_t const alloc_area_size = alloc_area_offset + alloc_size;
|
||||
|
||||
_HAssert( block_size >= min_block_size );
|
||||
_HAssert( block_begin < block_end );
|
||||
_HAssert(
|
||||
_Heap_Is_aligned( block_begin + HEAP_BLOCK_HEADER_SIZE, page_size )
|
||||
);
|
||||
_HAssert(
|
||||
_Heap_Is_aligned( block_size, page_size )
|
||||
);
|
||||
|
||||
_HAssert( alloc_end <= block_end + HEAP_BLOCK_SIZE_OFFSET );
|
||||
_HAssert( alloc_area_begin == block_begin + HEAP_BLOCK_HEADER_SIZE);
|
||||
_HAssert( alloc_area_offset < page_size );
|
||||
|
||||
_HAssert( _Heap_Is_aligned( alloc_area_begin, page_size ) );
|
||||
if ( alignment == 0 ) {
|
||||
_HAssert( alloc_begin == alloc_area_begin );
|
||||
} else {
|
||||
_HAssert( _Heap_Is_aligned( alloc_begin, alignment ) );
|
||||
}
|
||||
|
||||
if ( boundary != 0 ) {
|
||||
uintptr_t boundary_line = _Heap_Align_down( alloc_end, boundary );
|
||||
|
||||
_HAssert( alloc_size <= boundary );
|
||||
_HAssert( boundary_line <= alloc_begin || alloc_end <= boundary_line );
|
||||
}
|
||||
}
|
||||
#else
|
||||
#define _Heap_Check_allocation( h, b, ab, as, ag, bd ) ((void) 0)
|
||||
#endif
|
||||
|
||||
static uintptr_t _Heap_Check_block(
|
||||
const Heap_Control *heap,
|
||||
const Heap_Block *block,
|
||||
uintptr_t alloc_size,
|
||||
uintptr_t alignment,
|
||||
uintptr_t boundary
|
||||
)
|
||||
{
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const min_block_size = heap->min_block_size;
|
||||
|
||||
uintptr_t const block_begin = (uintptr_t) block;
|
||||
uintptr_t const block_size = _Heap_Block_size( block );
|
||||
uintptr_t const block_end = block_begin + block_size;
|
||||
|
||||
uintptr_t const alloc_begin_floor = _Heap_Alloc_area_of_block( block );
|
||||
uintptr_t const alloc_begin_ceiling = block_end - min_block_size
|
||||
+ HEAP_BLOCK_HEADER_SIZE + page_size - 1;
|
||||
|
||||
uintptr_t alloc_end = block_end + HEAP_BLOCK_SIZE_OFFSET;
|
||||
uintptr_t alloc_begin = alloc_end - alloc_size;
|
||||
|
||||
alloc_begin = _Heap_Align_down( alloc_begin, alignment );
|
||||
|
||||
/* Ensure that the we have a valid new block at the end */
|
||||
if ( alloc_begin > alloc_begin_ceiling ) {
|
||||
alloc_begin = _Heap_Align_down( alloc_begin_ceiling, alignment );
|
||||
}
|
||||
|
||||
alloc_end = alloc_begin + alloc_size;
|
||||
|
||||
/* Ensure boundary constaint */
|
||||
if ( boundary != 0 ) {
|
||||
uintptr_t const boundary_floor = alloc_begin_floor + alloc_size;
|
||||
uintptr_t boundary_line = _Heap_Align_down( alloc_end, boundary );
|
||||
|
||||
while ( alloc_begin < boundary_line && boundary_line < alloc_end ) {
|
||||
if ( boundary_line < boundary_floor ) {
|
||||
return 0;
|
||||
}
|
||||
alloc_begin = boundary_line - alloc_size;
|
||||
alloc_begin = _Heap_Align_down( alloc_begin, alignment );
|
||||
alloc_end = alloc_begin + alloc_size;
|
||||
boundary_line = _Heap_Align_down( alloc_end, boundary );
|
||||
}
|
||||
}
|
||||
|
||||
/* Ensure that the we have a valid new block at the beginning */
|
||||
if ( alloc_begin >= alloc_begin_floor ) {
|
||||
uintptr_t const alloc_block_begin =
|
||||
(uintptr_t) _Heap_Block_of_alloc_area( alloc_begin, page_size );
|
||||
uintptr_t const free_size = alloc_block_begin - block_begin;
|
||||
|
||||
if ( free_size >= min_block_size || free_size == 0 ) {
|
||||
return alloc_begin;
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void *_Heap_Allocate_aligned_with_boundary(
|
||||
Heap_Control *heap,
|
||||
uintptr_t alloc_size,
|
||||
uintptr_t alignment,
|
||||
uintptr_t boundary
|
||||
)
|
||||
{
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
Heap_Block * const tail = _Heap_Free_list_tail( heap );
|
||||
Heap_Block *block = _Heap_First_free_block( heap );
|
||||
Heap_Block *const free_list_tail = _Heap_Free_list_tail( heap );
|
||||
Heap_Block *block = _Heap_Free_list_first( heap );
|
||||
uintptr_t const block_size_floor = alloc_size + HEAP_BLOCK_HEADER_SIZE
|
||||
- HEAP_BLOCK_SIZE_OFFSET;
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t alloc_begin = 0;
|
||||
uint32_t search_count = 0;
|
||||
void *alloc_area_begin_ptr = NULL;
|
||||
|
||||
size = _Heap_Calc_block_size( size, heap->page_size, heap->min_block_size );
|
||||
if( size == 0 ) {
|
||||
if ( block_size_floor < alloc_size ) {
|
||||
/* Integer overflow occured */
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* Find large enough free block.
|
||||
*
|
||||
* Do not bother to mask out the HEAP_PREV_BLOCK_USED bit as it will not
|
||||
* change the result of the size comparison.
|
||||
*/
|
||||
while (block != tail && block->size_and_flag < size) {
|
||||
_HAssert( _Heap_Is_prev_used( block ));
|
||||
if ( boundary != 0 ) {
|
||||
if ( boundary < alloc_size ) {
|
||||
return NULL;
|
||||
}
|
||||
|
||||
block = block->next;
|
||||
++search_count;
|
||||
if ( alignment == 0 ) {
|
||||
alignment = page_size;
|
||||
}
|
||||
}
|
||||
|
||||
if (block != tail) {
|
||||
_Heap_Block_allocate( heap, block, size );
|
||||
|
||||
alloc_area_begin_ptr = (void *) _Heap_Alloc_area_of_block( block );
|
||||
|
||||
_HAssert( _Heap_Is_aligned( (uintptr_t) alloc_area_begin_ptr, heap->page_size ));
|
||||
while ( block != free_list_tail ) {
|
||||
_HAssert( _Heap_Is_prev_used( block ) );
|
||||
|
||||
/* Statistics */
|
||||
++search_count;
|
||||
|
||||
/*
|
||||
* The HEAP_PREV_BLOCK_USED flag is always set in the block size_and_flag
|
||||
* field. Thus the value is about one unit larger than the real block
|
||||
* size. The greater than operator takes this into account.
|
||||
*/
|
||||
if ( block->size_and_flag > block_size_floor ) {
|
||||
if ( alignment == 0 ) {
|
||||
alloc_begin = _Heap_Alloc_area_of_block( block );
|
||||
} else {
|
||||
alloc_begin = _Heap_Check_block(
|
||||
heap,
|
||||
block,
|
||||
alloc_size,
|
||||
alignment,
|
||||
boundary
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
if ( alloc_begin != 0 ) {
|
||||
break;
|
||||
}
|
||||
|
||||
block = block->next;
|
||||
}
|
||||
|
||||
if ( alloc_begin != 0 ) {
|
||||
block = _Heap_Block_allocate( heap, block, alloc_begin, alloc_size );
|
||||
|
||||
_Heap_Check_allocation(
|
||||
heap,
|
||||
block,
|
||||
alloc_begin,
|
||||
alloc_size,
|
||||
alignment,
|
||||
boundary
|
||||
);
|
||||
|
||||
/* Statistics */
|
||||
++stats->allocs;
|
||||
stats->searches += search_count;
|
||||
}
|
||||
|
||||
/* Statistics */
|
||||
if (stats->max_search < search_count) {
|
||||
if ( stats->max_search < search_count ) {
|
||||
stats->max_search = search_count;
|
||||
}
|
||||
|
||||
return alloc_area_begin_ptr;
|
||||
return (void *) alloc_begin;
|
||||
}
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
#if 0
|
||||
/*
|
||||
* Heap Handler
|
||||
*
|
||||
@@ -31,10 +32,10 @@ check_result(
|
||||
)
|
||||
{
|
||||
uintptr_t const user_area = _Heap_Alloc_area_of_block(the_block);
|
||||
uintptr_t const block_end = the_block
|
||||
uintptr_t const block_end = (uintptr_t) the_block
|
||||
+ _Heap_Block_size(the_block) + HEAP_BLOCK_SIZE_OFFSET;
|
||||
uintptr_t const user_end = aligned_user_addr + size;
|
||||
uintptr_t const heap_start = (uintptr_t) the_heap->start + HEAP_LAST_BLOCK_OVERHEAD;
|
||||
uintptr_t const heap_start = (uintptr_t) the_heap->start + HEAP_BLOCK_HEADER_SIZE;
|
||||
uintptr_t const heap_end = (uintptr_t) the_heap->final
|
||||
+ HEAP_BLOCK_SIZE_OFFSET;
|
||||
uintptr_t const page_size = the_heap->page_size;
|
||||
@@ -97,7 +98,7 @@ Heap_Block *block_allocate(
|
||||
/* Don't split the block as remainder is either zero or too small to be
|
||||
used as a separate free block. Change 'alloc_size' to the size of the
|
||||
block and remove the block from the list of free blocks. */
|
||||
_Heap_Block_remove_from_free_list(the_block);
|
||||
_Heap_Free_list_remove(the_block);
|
||||
alloc_size = block_size;
|
||||
stats->free_blocks -= 1;
|
||||
}
|
||||
@@ -157,7 +158,7 @@ void *_Heap_Allocate_aligned(
|
||||
|
||||
/* Find large enough free block that satisfies the alignment requirements. */
|
||||
|
||||
for (the_block = _Heap_First_free_block(the_heap), search_count = 0;
|
||||
for (the_block = _Heap_Free_list_first(the_heap), search_count = 0;
|
||||
the_block != tail;
|
||||
the_block = the_block->next, ++search_count)
|
||||
{
|
||||
@@ -220,7 +221,7 @@ void *_Heap_Allocate_aligned(
|
||||
/* The block is indeed acceptable: calculate the size of the block
|
||||
to be allocated and perform allocation. */
|
||||
uintptr_t const alloc_size =
|
||||
block_end - user_addr + HEAP_BLOCK_ALLOC_AREA_OFFSET;
|
||||
block_end - user_addr + HEAP_BLOCK_HEADER_SIZE;
|
||||
|
||||
_HAssert(_Heap_Is_aligned(aligned_user_addr, alignment));
|
||||
|
||||
@@ -244,3 +245,4 @@ void *_Heap_Allocate_aligned(
|
||||
|
||||
return user_ptr;
|
||||
}
|
||||
#endif
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-1999.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -28,11 +34,11 @@ Heap_Extend_status _Heap_Extend(
|
||||
{
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
uintptr_t const area_begin = (uintptr_t) area_begin_ptr;
|
||||
uintptr_t const heap_area_begin = heap->begin;
|
||||
uintptr_t const heap_area_end = heap->end;
|
||||
uintptr_t const heap_area_begin = heap->area_begin;
|
||||
uintptr_t const heap_area_end = heap->area_end;
|
||||
uintptr_t const new_heap_area_end = heap_area_end + area_size;
|
||||
uintptr_t extend_size = 0;
|
||||
Heap_Block *const old_final = heap->final;
|
||||
Heap_Block *const old_final = heap->last_block;
|
||||
Heap_Block *new_final = NULL;
|
||||
|
||||
/*
|
||||
@@ -60,10 +66,10 @@ Heap_Extend_status _Heap_Extend(
|
||||
* block and free it.
|
||||
*/
|
||||
|
||||
heap->end = new_heap_area_end;
|
||||
heap->area_end = new_heap_area_end;
|
||||
|
||||
extend_size = new_heap_area_end
|
||||
- (uintptr_t) old_final - HEAP_LAST_BLOCK_OVERHEAD;
|
||||
- (uintptr_t) old_final - HEAP_BLOCK_HEADER_SIZE;
|
||||
extend_size = _Heap_Align_down( extend_size, heap->page_size );
|
||||
|
||||
*amount_extended = extend_size;
|
||||
@@ -74,7 +80,7 @@ Heap_Extend_status _Heap_Extend(
|
||||
new_final = _Heap_Block_at( old_final, extend_size );
|
||||
new_final->size_and_flag = heap->page_size | HEAP_PREV_BLOCK_USED;
|
||||
|
||||
heap->final = new_final;
|
||||
heap->last_block = new_final;
|
||||
|
||||
stats->size += area_size;
|
||||
++stats->used_blocks;
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -19,26 +25,18 @@
|
||||
#include <rtems/score/sysstate.h>
|
||||
#include <rtems/score/heap.h>
|
||||
|
||||
bool _Heap_Free( Heap_Control *heap, void *alloc_area_begin_ptr )
|
||||
bool _Heap_Free( Heap_Control *heap, void *alloc_begin_ptr )
|
||||
{
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
uintptr_t alloc_area_begin = (uintptr_t) alloc_area_begin_ptr;
|
||||
uintptr_t alloc_begin = (uintptr_t) alloc_begin_ptr;
|
||||
Heap_Block *block =
|
||||
_Heap_Block_of_alloc_area( alloc_area_begin, heap->page_size );
|
||||
_Heap_Block_of_alloc_area( alloc_begin, heap->page_size );
|
||||
Heap_Block *next_block = NULL;
|
||||
uintptr_t block_size = 0;
|
||||
uintptr_t next_block_size = 0;
|
||||
bool next_is_free = false;
|
||||
|
||||
if (
|
||||
!_Addresses_Is_in_range( alloc_area_begin_ptr, heap->start, heap->final)
|
||||
) {
|
||||
_HAssert( alloc_area_begin_ptr != NULL );
|
||||
return false;
|
||||
}
|
||||
|
||||
if ( !_Heap_Is_block_in_heap( heap, block ) ) {
|
||||
_HAssert( false );
|
||||
return false;
|
||||
}
|
||||
|
||||
@@ -56,7 +54,7 @@ bool _Heap_Free( Heap_Control *heap, void *alloc_area_begin_ptr )
|
||||
}
|
||||
|
||||
next_block_size = _Heap_Block_size( next_block );
|
||||
next_is_free = next_block != heap->final
|
||||
next_is_free = next_block != heap->last_block
|
||||
&& !_Heap_Is_prev_used( _Heap_Block_at( next_block, next_block_size ));
|
||||
|
||||
if ( !_Heap_Is_prev_used( block ) ) {
|
||||
@@ -77,7 +75,7 @@ bool _Heap_Free( Heap_Control *heap, void *alloc_area_begin_ptr )
|
||||
|
||||
if ( next_is_free ) { /* coalesce both */
|
||||
uintptr_t const size = block_size + prev_size + next_block_size;
|
||||
_Heap_Block_remove_from_free_list( next_block );
|
||||
_Heap_Free_list_remove( next_block );
|
||||
stats->free_blocks -= 1;
|
||||
prev_block->size_and_flag = size | HEAP_PREV_BLOCK_USED;
|
||||
next_block = _Heap_Block_at( prev_block, size );
|
||||
@@ -91,14 +89,14 @@ bool _Heap_Free( Heap_Control *heap, void *alloc_area_begin_ptr )
|
||||
}
|
||||
} else if ( next_is_free ) { /* coalesce next */
|
||||
uintptr_t const size = block_size + next_block_size;
|
||||
_Heap_Block_replace_in_free_list( next_block, block );
|
||||
_Heap_Free_list_replace( next_block, block );
|
||||
block->size_and_flag = size | HEAP_PREV_BLOCK_USED;
|
||||
next_block = _Heap_Block_at( block, size );
|
||||
next_block->prev_size = size;
|
||||
} else { /* no coalesce */
|
||||
/* Add 'block' to the head of the free blocks list as it tends to
|
||||
produce less fragmentation than adding to the tail. */
|
||||
_Heap_Block_insert_after( _Heap_Free_list_head( heap), block );
|
||||
_Heap_Free_list_insert_after( _Heap_Free_list_head( heap), block );
|
||||
block->size_and_flag = block_size | HEAP_PREV_BLOCK_USED;
|
||||
next_block->size_and_flag &= ~HEAP_PREV_BLOCK_USED;
|
||||
next_block->prev_size = block_size;
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2004.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -19,21 +25,6 @@
|
||||
#include <rtems/score/sysstate.h>
|
||||
#include <rtems/score/heap.h>
|
||||
|
||||
/*PAGE
|
||||
*
|
||||
* _Heap_Get_free_information
|
||||
*
|
||||
* This heap routine returns information about the free blocks
|
||||
* in the specified heap.
|
||||
*
|
||||
* Input parameters:
|
||||
* the_heap - pointer to heap header.
|
||||
* info - pointer to the free block information.
|
||||
*
|
||||
* Output parameters:
|
||||
* returns - free block information filled in.
|
||||
*/
|
||||
|
||||
void _Heap_Get_free_information(
|
||||
Heap_Control *the_heap,
|
||||
Heap_Information *info
|
||||
@@ -46,7 +37,7 @@ void _Heap_Get_free_information(
|
||||
info->largest = 0;
|
||||
info->total = 0;
|
||||
|
||||
for(the_block = _Heap_First_free_block(the_heap);
|
||||
for(the_block = _Heap_Free_list_first(the_heap);
|
||||
the_block != tail;
|
||||
the_block = the_block->next)
|
||||
{
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2009.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -19,27 +25,13 @@
|
||||
#include <rtems/score/sysstate.h>
|
||||
#include <rtems/score/heap.h>
|
||||
|
||||
/*
|
||||
* _Heap_Get_information
|
||||
*
|
||||
* This kernel routine walks the heap and tots up the free and allocated
|
||||
* sizes. Derived from _Heap_Walk.
|
||||
*
|
||||
* Input parameters:
|
||||
* the_heap - pointer to heap header
|
||||
* the_info - pointer for information to be returned
|
||||
*
|
||||
* Output parameters:
|
||||
* *the_info - contains information about heap
|
||||
* return 0=success, otherwise heap is corrupt.
|
||||
*/
|
||||
Heap_Get_information_status _Heap_Get_information(
|
||||
void _Heap_Get_information(
|
||||
Heap_Control *the_heap,
|
||||
Heap_Information_block *the_info
|
||||
)
|
||||
{
|
||||
Heap_Block *the_block = the_heap->start;
|
||||
Heap_Block *const end = the_heap->final;
|
||||
Heap_Block *the_block = the_heap->first_block;
|
||||
Heap_Block *const end = the_heap->last_block;
|
||||
|
||||
_HAssert(the_block->prev_size == the_heap->page_size);
|
||||
_HAssert(_Heap_Is_prev_used(the_block));
|
||||
@@ -52,7 +44,7 @@ Heap_Get_information_status _Heap_Get_information(
|
||||
the_info->Used.largest = 0;
|
||||
|
||||
while ( the_block != end ) {
|
||||
uint32_t const the_size = _Heap_Block_size(the_block);
|
||||
uintptr_t const the_size = _Heap_Block_size(the_block);
|
||||
Heap_Block *const next_block = _Heap_Block_at(the_block, the_size);
|
||||
Heap_Information *info;
|
||||
|
||||
@@ -74,7 +66,5 @@ Heap_Get_information_status _Heap_Get_information(
|
||||
* "used" as client never allocated it. Make 'Used.total' contain this
|
||||
* blocks' overhead though.
|
||||
*/
|
||||
the_info->Used.total += HEAP_LAST_BLOCK_OVERHEAD;
|
||||
|
||||
return HEAP_GET_INFORMATION_SUCCESSFUL;
|
||||
the_info->Used.total += HEAP_BLOCK_HEADER_SIZE;
|
||||
}
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-1999.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -21,129 +27,163 @@
|
||||
|
||||
Heap_Resize_status _Heap_Resize_block(
|
||||
Heap_Control *heap,
|
||||
void *alloc_area_begin_ptr,
|
||||
uintptr_t size,
|
||||
uintptr_t *old_mem_size,
|
||||
uintptr_t *avail_mem_size
|
||||
void *alloc_begin_ptr,
|
||||
uintptr_t new_alloc_size,
|
||||
uintptr_t *old_size,
|
||||
uintptr_t *new_size
|
||||
)
|
||||
{
|
||||
uintptr_t alloc_area_begin = (uintptr_t) alloc_area_begin_ptr;
|
||||
Heap_Block *block;
|
||||
Heap_Block *next_block;
|
||||
uintptr_t next_block_size;
|
||||
bool next_is_used;
|
||||
Heap_Block *next_next_block;
|
||||
uintptr_t old_block_size;
|
||||
uintptr_t old_user_size;
|
||||
uintptr_t prev_used_flag;
|
||||
Heap_Statistics *const stats = &heap->stats;
|
||||
uintptr_t const min_block_size = heap->min_block_size;
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const alloc_begin = (uintptr_t) alloc_begin_ptr;
|
||||
Heap_Block *const block = _Heap_Block_of_alloc_area( alloc_begin, page_size );
|
||||
Heap_Block *next_block = NULL;
|
||||
Heap_Block *next_next_block = NULL;
|
||||
uintptr_t block_size = 0;
|
||||
uintptr_t block_end = 0;
|
||||
uintptr_t next_block_size = 0;
|
||||
bool next_block_is_used = false;;
|
||||
uintptr_t alloc_size = 0;
|
||||
uintptr_t prev_block_used_flag = 0;
|
||||
|
||||
*old_mem_size = 0;
|
||||
*avail_mem_size = 0;
|
||||
*old_size = 0;
|
||||
*new_size = 0;
|
||||
|
||||
block = _Heap_Block_of_alloc_area(alloc_area_begin, heap->page_size);
|
||||
_HAssert(_Heap_Is_block_in_heap(heap, block));
|
||||
if (!_Heap_Is_block_in_heap(heap, block))
|
||||
if ( !_Heap_Is_block_in_heap( heap, block ) ) {
|
||||
return HEAP_RESIZE_FATAL_ERROR;
|
||||
}
|
||||
|
||||
prev_used_flag = block->size_and_flag & HEAP_PREV_BLOCK_USED;
|
||||
old_block_size = _Heap_Block_size(block);
|
||||
next_block = _Heap_Block_at(block, old_block_size);
|
||||
block_size = _Heap_Block_size( block );
|
||||
block_end = (uintptr_t) block + block_size;
|
||||
prev_block_used_flag = block->size_and_flag & HEAP_PREV_BLOCK_USED;
|
||||
next_block = _Heap_Block_at( block, block_size );
|
||||
|
||||
_HAssert(_Heap_Is_block_in_heap(heap, next_block));
|
||||
_HAssert(_Heap_Is_prev_used(next_block));
|
||||
if ( !_Heap_Is_block_in_heap(heap, next_block) ||
|
||||
!_Heap_Is_prev_used(next_block))
|
||||
return HEAP_RESIZE_FATAL_ERROR;
|
||||
_HAssert( _Heap_Is_block_in_heap( heap, next_block ) );
|
||||
_HAssert( _Heap_Is_prev_used( next_block ) );
|
||||
|
||||
next_block_size = _Heap_Block_size(next_block);
|
||||
next_next_block = _Heap_Block_at(next_block, next_block_size);
|
||||
next_is_used = (next_block == heap->final) ||
|
||||
_Heap_Is_prev_used(next_next_block);
|
||||
next_block_size = _Heap_Block_size( next_block );
|
||||
next_next_block = _Heap_Block_at( next_block, next_block_size );
|
||||
|
||||
/* See _Heap_Size_of_alloc_area() source for explanations */
|
||||
old_user_size = (uintptr_t) next_block - alloc_area_begin
|
||||
+ HEAP_BLOCK_SIZE_OFFSET;
|
||||
_HAssert(
|
||||
next_block == heap->last_block
|
||||
|| _Heap_Is_block_in_heap( heap, next_next_block )
|
||||
);
|
||||
|
||||
*old_mem_size = old_user_size;
|
||||
next_block_is_used = next_block == heap->last_block
|
||||
|| _Heap_Is_prev_used( next_next_block );
|
||||
|
||||
if (size > old_user_size) {
|
||||
/* Need to extend the block: allocate part of the next block and then
|
||||
merge 'block' and allocated block together. */
|
||||
if (next_is_used) /* Next block is in use, -- no way to extend */
|
||||
alloc_size = block_end - alloc_begin + HEAP_BLOCK_SIZE_OFFSET;
|
||||
|
||||
*old_size = alloc_size;
|
||||
|
||||
if ( new_alloc_size > alloc_size ) {
|
||||
/*
|
||||
* Need to extend the block: allocate part of the next block and then
|
||||
* merge the blocks.
|
||||
*/
|
||||
if ( next_block_is_used ) {
|
||||
return HEAP_RESIZE_UNSATISFIED;
|
||||
else {
|
||||
} else {
|
||||
uintptr_t add_block_size =
|
||||
_Heap_Align_up(size - old_user_size, page_size);
|
||||
if (add_block_size < min_block_size)
|
||||
_Heap_Align_up( new_alloc_size - alloc_size, page_size );
|
||||
|
||||
if ( add_block_size < min_block_size ) {
|
||||
add_block_size = min_block_size;
|
||||
if (add_block_size > next_block_size)
|
||||
return HEAP_RESIZE_UNSATISFIED; /* Next block is too small or none. */
|
||||
add_block_size =
|
||||
_Heap_Block_allocate(heap, next_block, add_block_size);
|
||||
/* Merge two subsequent blocks */
|
||||
block->size_and_flag = (old_block_size + add_block_size) | prev_used_flag;
|
||||
}
|
||||
|
||||
if ( add_block_size > next_block_size ) {
|
||||
return HEAP_RESIZE_UNSATISFIED;
|
||||
}
|
||||
|
||||
next_block = _Heap_Block_allocate(
|
||||
heap,
|
||||
next_block,
|
||||
_Heap_Alloc_area_of_block( next_block ),
|
||||
add_block_size - HEAP_BLOCK_HEADER_SIZE
|
||||
);
|
||||
|
||||
/* Merge the blocks */
|
||||
block->size_and_flag = ( block_size + _Heap_Block_size( next_block ) )
|
||||
| prev_block_used_flag;
|
||||
|
||||
/* Statistics */
|
||||
--stats->used_blocks;
|
||||
}
|
||||
} else {
|
||||
|
||||
/* Calculate how much memory we could free */
|
||||
uintptr_t free_block_size =
|
||||
_Heap_Align_down(old_user_size - size, page_size);
|
||||
_Heap_Align_down( alloc_size - new_alloc_size, page_size );
|
||||
|
||||
if (free_block_size > 0) {
|
||||
if ( free_block_size > 0 ) {
|
||||
/*
|
||||
* To free some memory the block should be shortened so that it can can
|
||||
* hold 'new_alloc_size' user bytes and still remain not shorter than
|
||||
* 'min_block_size'.
|
||||
*/
|
||||
uintptr_t new_block_size = block_size - free_block_size;
|
||||
|
||||
/* To free some memory the block should be shortened so that it can
|
||||
can hold 'size' user bytes and still remain not shorter than
|
||||
'min_block_size'. */
|
||||
if ( new_block_size < min_block_size ) {
|
||||
uintptr_t const delta = min_block_size - new_block_size;
|
||||
|
||||
uintptr_t new_block_size = old_block_size - free_block_size;
|
||||
_HAssert( free_block_size >= delta );
|
||||
|
||||
if (new_block_size < min_block_size) {
|
||||
uintptr_t delta = min_block_size - new_block_size;
|
||||
_HAssert(free_block_size >= delta);
|
||||
free_block_size -= delta;
|
||||
if (free_block_size == 0) {
|
||||
|
||||
if ( free_block_size == 0 ) {
|
||||
/* Statistics */
|
||||
++stats->resizes;
|
||||
|
||||
return HEAP_RESIZE_SUCCESSFUL;
|
||||
}
|
||||
|
||||
new_block_size += delta;
|
||||
}
|
||||
|
||||
_HAssert(new_block_size >= min_block_size);
|
||||
_HAssert(new_block_size + free_block_size == old_block_size);
|
||||
_HAssert(_Heap_Is_aligned(new_block_size, page_size));
|
||||
_HAssert(_Heap_Is_aligned(free_block_size, page_size));
|
||||
_HAssert( new_block_size >= min_block_size );
|
||||
_HAssert( new_block_size + free_block_size == block_size );
|
||||
_HAssert( _Heap_Is_aligned( new_block_size, page_size ) );
|
||||
_HAssert( _Heap_Is_aligned( free_block_size, page_size ) );
|
||||
|
||||
if (!next_is_used) {
|
||||
/* Extend the next block to the low addresses by 'free_block_size' */
|
||||
if ( !next_block_is_used ) {
|
||||
/* Extend the next block */
|
||||
Heap_Block *const new_next_block =
|
||||
_Heap_Block_at(block, new_block_size);
|
||||
_Heap_Block_at( block, new_block_size );
|
||||
uintptr_t const new_next_block_size =
|
||||
next_block_size + free_block_size;
|
||||
_HAssert(_Heap_Is_block_in_heap(heap, next_next_block));
|
||||
block->size_and_flag = new_block_size | prev_used_flag;
|
||||
new_next_block->size_and_flag = new_next_block_size | HEAP_PREV_BLOCK_USED;
|
||||
next_next_block->prev_size = new_next_block_size;
|
||||
_Heap_Block_replace_in_free_list(next_block, new_next_block);
|
||||
heap->stats.free_size += free_block_size;
|
||||
*avail_mem_size = new_next_block_size - HEAP_BLOCK_USED_OVERHEAD;
|
||||
|
||||
} else if (free_block_size >= min_block_size) {
|
||||
/* Split the block into 2 used parts, then free the second one. */
|
||||
block->size_and_flag = new_block_size | prev_used_flag;
|
||||
next_block = _Heap_Block_at(block, new_block_size);
|
||||
_HAssert( _Heap_Is_block_in_heap( heap, next_next_block ) );
|
||||
|
||||
block->size_and_flag = new_block_size | prev_block_used_flag;
|
||||
new_next_block->size_and_flag =
|
||||
new_next_block_size | HEAP_PREV_BLOCK_USED;
|
||||
next_next_block->prev_size = new_next_block_size;
|
||||
|
||||
_Heap_Free_list_replace( next_block, new_next_block );
|
||||
|
||||
*new_size = new_next_block_size - HEAP_BLOCK_SIZE_OFFSET;
|
||||
|
||||
/* Statistics */
|
||||
stats->free_size += free_block_size;
|
||||
} else if ( free_block_size >= min_block_size ) {
|
||||
/* Split the block into two used parts, then free the second one */
|
||||
block->size_and_flag = new_block_size | prev_block_used_flag;
|
||||
next_block = _Heap_Block_at( block, new_block_size );
|
||||
next_block->size_and_flag = free_block_size | HEAP_PREV_BLOCK_USED;
|
||||
|
||||
_Heap_Free( heap, (void *) _Heap_Alloc_area_of_block( next_block ) );
|
||||
|
||||
*new_size = free_block_size - HEAP_BLOCK_SIZE_OFFSET;
|
||||
|
||||
/* Statistics */
|
||||
++stats->used_blocks; /* We have created used block */
|
||||
--stats->frees; /* Don't count next call in stats */
|
||||
_Heap_Free(heap, (void *) _Heap_Alloc_area_of_block(next_block));
|
||||
*avail_mem_size = free_block_size - HEAP_BLOCK_USED_OVERHEAD;
|
||||
--stats->frees; /* Do not count next call in stats */
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* Statistics */
|
||||
++stats->resizes;
|
||||
|
||||
return HEAP_RESIZE_SUCCESSFUL;
|
||||
}
|
||||
|
||||
@@ -1,6 +1,12 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-1999.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -21,24 +27,16 @@
|
||||
|
||||
bool _Heap_Size_of_alloc_area(
|
||||
Heap_Control *heap,
|
||||
void *alloc_area_begin_ptr,
|
||||
uintptr_t *size
|
||||
void *alloc_begin_ptr,
|
||||
uintptr_t *alloc_size
|
||||
)
|
||||
{
|
||||
uintptr_t alloc_area_begin = (uintptr_t) alloc_area_begin_ptr;
|
||||
Heap_Block *block =
|
||||
_Heap_Block_of_alloc_area( alloc_area_begin, heap->page_size );
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const alloc_begin = (uintptr_t) alloc_begin_ptr;
|
||||
Heap_Block *block = _Heap_Block_of_alloc_area( alloc_begin, page_size );
|
||||
Heap_Block *next_block = NULL;
|
||||
uintptr_t block_size = 0;
|
||||
|
||||
if (
|
||||
!_Addresses_Is_in_range( alloc_area_begin_ptr, heap->start, heap->final )
|
||||
) {
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
_HAssert(_Heap_Is_block_in_heap( heap, block ));
|
||||
if ( !_Heap_Is_block_in_heap( heap, block ) ) {
|
||||
return false;
|
||||
}
|
||||
@@ -46,26 +44,14 @@ bool _Heap_Size_of_alloc_area(
|
||||
block_size = _Heap_Block_size( block );
|
||||
next_block = _Heap_Block_at( block, block_size );
|
||||
|
||||
_HAssert( _Heap_Is_block_in_heap( heap, next_block ));
|
||||
_HAssert( _Heap_Is_prev_used( next_block ));
|
||||
if (
|
||||
!_Heap_Is_block_in_heap( heap, next_block ) ||
|
||||
!_Heap_Is_prev_used( next_block )
|
||||
!_Heap_Is_block_in_heap( heap, next_block )
|
||||
|| !_Heap_Is_prev_used( next_block )
|
||||
) {
|
||||
return false;
|
||||
}
|
||||
|
||||
/*
|
||||
* 'alloc_area_begin' could be greater than 'block' address plus
|
||||
* HEAP_BLOCK_ALLOC_AREA_OFFSET as _Heap_Allocate_aligned() may produce such
|
||||
* user pointers. To get rid of this offset we calculate user size as
|
||||
* difference between the end of 'block' (='next_block') and
|
||||
* 'alloc_area_begin' and then add correction equal to the offset of the
|
||||
* 'size' field of the 'Heap_Block' structure. The correction is due to the
|
||||
* fact that 'prev_size' field of the next block is actually used as user
|
||||
* accessible area of 'block'.
|
||||
*/
|
||||
*size = (uintptr_t) next_block - alloc_area_begin + HEAP_BLOCK_SIZE_OFFSET;
|
||||
*alloc_size = (uintptr_t) next_block + HEAP_BLOCK_SIZE_OFFSET - alloc_begin;
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
@@ -1,8 +1,14 @@
|
||||
/*
|
||||
* Heap Handler
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
* @ingroup ScoreHeap
|
||||
*
|
||||
* @brief Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT ( c ) 1989-2007.
|
||||
* On-Line Applications Research Corporation ( OAR ).
|
||||
*
|
||||
* The license and distribution terms for this file may be
|
||||
* found in the file LICENSE in this distribution or at
|
||||
@@ -22,197 +28,446 @@
|
||||
#include <rtems/score/interr.h>
|
||||
#include <rtems/bspIo.h>
|
||||
|
||||
#if defined(__GNUC__)
|
||||
#define DO_NOT_INLINE __attribute__((__noinline__))
|
||||
#else
|
||||
#define DO_NOT_INLINE
|
||||
#endif
|
||||
/*
|
||||
* Helper to avoid introducing even more branches and paths in this
|
||||
* code to do coverage analysis on.
|
||||
*
|
||||
* We do not want this inlined.
|
||||
*/
|
||||
static void hw_nl(
|
||||
int error,
|
||||
bool do_dump
|
||||
) DO_NOT_INLINE;
|
||||
static void _Heap_Walk_printk( int source, bool dump, bool error, const char *fmt, ... )
|
||||
{
|
||||
if ( dump ) {
|
||||
va_list ap;
|
||||
|
||||
/*PAGE
|
||||
*
|
||||
* _Heap_Walk
|
||||
*
|
||||
* This kernel routine walks the heap and verifies its correctness.
|
||||
*
|
||||
* Input parameters:
|
||||
* the_heap - pointer to heap header
|
||||
* source - a numeric indicator of the invoker of this routine
|
||||
* do_dump - when true print the information
|
||||
*
|
||||
* Output parameters: NONE
|
||||
*/
|
||||
if ( error ) {
|
||||
printk( "FAIL[%d]: ", source );
|
||||
} else {
|
||||
printk( "PASS[%d]: ", source );
|
||||
}
|
||||
|
||||
va_start( ap, fmt );
|
||||
vprintk( fmt, ap );
|
||||
va_end( ap );
|
||||
}
|
||||
}
|
||||
|
||||
static bool _Heap_Walk_check_free_list(
|
||||
int source,
|
||||
bool dump,
|
||||
Heap_Control *heap
|
||||
)
|
||||
{
|
||||
const Heap_Block *const free_list_tail = _Heap_Free_list_tail( heap );
|
||||
const Heap_Block *const first_free_block = _Heap_Free_list_first( heap );
|
||||
const Heap_Block *free_block = first_free_block;
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const loop_limit =
|
||||
((uintptr_t) heap->last_block - (uintptr_t) heap->first_block)
|
||||
/ heap->min_block_size;
|
||||
uintptr_t loop_counter = 0;
|
||||
|
||||
while ( free_block != free_list_tail && loop_counter < loop_limit ) {
|
||||
if ( !_Heap_Is_block_in_heap( heap, free_block ) ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"free block 0x%08x: not in heap\n",
|
||||
free_block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if (
|
||||
!_Heap_Is_aligned( _Heap_Alloc_area_of_block( free_block ), page_size )
|
||||
) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"free block 0x%08x: alloc area not page aligned\n",
|
||||
free_block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
++loop_counter;
|
||||
|
||||
free_block = free_block->next;
|
||||
}
|
||||
|
||||
if ( loop_counter >= loop_limit ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"free list contains a loop\n"
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
static bool _Heap_Walk_is_in_free_list(
|
||||
Heap_Control *heap,
|
||||
Heap_Block *block
|
||||
)
|
||||
{
|
||||
const Heap_Block *const free_list_tail = _Heap_Free_list_tail( heap );
|
||||
const Heap_Block *free_block = _Heap_Free_list_first( heap );
|
||||
|
||||
while ( free_block != free_list_tail ) {
|
||||
if ( free_block == block ) {
|
||||
return true;
|
||||
}
|
||||
free_block = free_block->next;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool _Heap_Walk_check_control(
|
||||
int source,
|
||||
bool dump,
|
||||
Heap_Control *heap
|
||||
)
|
||||
{
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const min_block_size = heap->min_block_size;
|
||||
Heap_Block *const free_list_tail = _Heap_Free_list_tail( heap );
|
||||
Heap_Block *const free_list_head = _Heap_Free_list_head( heap );
|
||||
Heap_Block *const first_free_block = _Heap_Free_list_first( heap );
|
||||
Heap_Block *const last_free_block = _Heap_Free_list_last( heap );
|
||||
Heap_Block *const first_block = heap->first_block;
|
||||
Heap_Block *const last_block = heap->last_block;
|
||||
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
false,
|
||||
"page size %u, min block size %u\n"
|
||||
"\tarea begin 0x%08x, area end 0x%08x\n"
|
||||
"\tfirst block 0x%08x, last block 0x%08x\n"
|
||||
"\tfirst free 0x%08x, last free 0x%08x\n",
|
||||
page_size, min_block_size,
|
||||
heap->area_begin, heap->area_end,
|
||||
first_block, last_block,
|
||||
first_free_block, last_free_block
|
||||
);
|
||||
|
||||
if ( page_size == 0 ) {
|
||||
_Heap_Walk_printk( source, dump, true, "page size is zero\n" );
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if ( !_Addresses_Is_aligned( (void *) page_size ) ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"page size %u not CPU aligned\n",
|
||||
page_size
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if ( !_Heap_Is_aligned( min_block_size, page_size ) ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"min block size %u not page aligned\n",
|
||||
min_block_size
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if (
|
||||
first_free_block != free_list_head
|
||||
&& !_Addresses_Is_aligned( first_free_block )
|
||||
) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"first free block: 0x%08x not CPU aligned\n",
|
||||
first_free_block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if (
|
||||
last_free_block != free_list_tail
|
||||
&& !_Addresses_Is_aligned( last_free_block )
|
||||
) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"last free block: 0x%08x not CPU aligned\n",
|
||||
last_free_block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if (
|
||||
!_Heap_Is_aligned( _Heap_Alloc_area_of_block( first_block ), page_size )
|
||||
) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"first block: 0x%08x not page aligned\n",
|
||||
first_block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if (
|
||||
!_Heap_Is_aligned( _Heap_Alloc_area_of_block( last_block ), page_size )
|
||||
) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"last block: 0x%08x not page aligned\n",
|
||||
last_block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if ( !_Heap_Is_prev_used( first_block ) ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"first block: HEAP_PREV_BLOCK_USED is cleared\n"
|
||||
);
|
||||
}
|
||||
|
||||
if ( first_block->prev_size != page_size ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
true,
|
||||
"first block: prev size %u != page size %u\n",
|
||||
first_block->prev_size,
|
||||
page_size
|
||||
);
|
||||
}
|
||||
|
||||
return _Heap_Walk_check_free_list( source, dump, heap );
|
||||
}
|
||||
|
||||
bool _Heap_Walk(
|
||||
Heap_Control *the_heap,
|
||||
int source,
|
||||
bool do_dump
|
||||
Heap_Control *heap,
|
||||
int source,
|
||||
bool dump
|
||||
)
|
||||
{
|
||||
Heap_Block *the_block = the_heap->start;
|
||||
Heap_Block *const end = the_heap->final;
|
||||
Heap_Block *const tail = _Heap_Free_list_tail(the_heap);
|
||||
int error = 0;
|
||||
int passes = 0;
|
||||
uintptr_t const page_size = heap->page_size;
|
||||
uintptr_t const min_block_size = heap->min_block_size;
|
||||
Heap_Block *const free_list_tail = _Heap_Free_list_tail( heap );
|
||||
Heap_Block *const free_list_head = _Heap_Free_list_head( heap );
|
||||
Heap_Block *const first_free_block = _Heap_Free_list_first( heap );
|
||||
Heap_Block *const last_free_block = _Heap_Free_list_last( heap );
|
||||
Heap_Block *const last_block = heap->last_block;
|
||||
Heap_Block *block = heap->first_block;
|
||||
bool error = false;
|
||||
|
||||
/* FIXME: Why is this disabled? */
|
||||
do_dump = false;
|
||||
|
||||
/* FIXME: Why is this disabled? */
|
||||
/*
|
||||
* We don't want to allow walking the heap until we have
|
||||
* transferred control to the user task so we watch the
|
||||
* system state.
|
||||
*/
|
||||
|
||||
/*
|
||||
if ( !_System_state_Is_up( _System_state_Get() ) )
|
||||
if ( !_System_state_Is_up( _System_state_Get() ) ) {
|
||||
return true;
|
||||
*/
|
||||
|
||||
/* FIXME: Reason for this? */
|
||||
if (source < 0)
|
||||
source = (int) the_heap->stats.instance;
|
||||
|
||||
if (do_dump)
|
||||
printk("\nPASS: %d start %p final %p first %p last %p begin %p end %p\n",
|
||||
source, the_block, end,
|
||||
_Heap_First_free_block(the_heap), _Heap_Last_free_block(the_heap),
|
||||
the_heap->begin, the_heap->end);
|
||||
|
||||
/*
|
||||
* Handle the 1st block
|
||||
*/
|
||||
|
||||
if (!_Heap_Is_prev_used(the_block)) {
|
||||
printk("PASS: %d !HEAP_PREV_BLOCK_USED flag of 1st block isn't set\n", source);
|
||||
error = 1;
|
||||
}
|
||||
|
||||
if (the_block->prev_size != the_heap->page_size) {
|
||||
printk("PASS: %d !prev_size of 1st block isn't page_size\n", source);
|
||||
error = 1;
|
||||
if ( !_Heap_Walk_check_control( source, dump, heap ) ) {
|
||||
return false;
|
||||
}
|
||||
|
||||
while ( the_block != end ) {
|
||||
uint32_t const the_size = _Heap_Block_size(the_block);
|
||||
Heap_Block *const next_block = _Heap_Block_at(the_block, the_size);
|
||||
bool prev_used = _Heap_Is_prev_used(the_block);
|
||||
while ( block != last_block && _Addresses_Is_aligned( block ) ) {
|
||||
uintptr_t const block_begin = (uintptr_t) block;
|
||||
uintptr_t const block_size = _Heap_Block_size( block );
|
||||
bool const prev_used = _Heap_Is_prev_used( block );
|
||||
Heap_Block *const next_block = _Heap_Block_at( block, block_size );
|
||||
uintptr_t const next_block_begin = (uintptr_t) next_block;
|
||||
|
||||
if (do_dump) {
|
||||
printk("PASS: %d block %p size %d(%c)",
|
||||
source, the_block, the_size, (prev_used ? 'U' : 'F'));
|
||||
if (prev_used)
|
||||
printk(" prev_size %d", the_block->prev_size);
|
||||
else
|
||||
printk(" (prev_size) %d", the_block->prev_size);
|
||||
if ( prev_used ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: size %u\n",
|
||||
block,
|
||||
block_size
|
||||
);
|
||||
} else {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: size %u, prev_size %u\n",
|
||||
block,
|
||||
block_size,
|
||||
block->prev_size
|
||||
);
|
||||
}
|
||||
|
||||
|
||||
if (!_Addresses_Is_aligned(next_block) ) {
|
||||
printk("PASS: %d next_block %p is not aligned\n", source, next_block);
|
||||
error = 1;
|
||||
if (
|
||||
!_Heap_Is_aligned( block_begin + HEAP_BLOCK_HEADER_SIZE, page_size )
|
||||
) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: not page (%u) aligned\n",
|
||||
block,
|
||||
page_size
|
||||
);
|
||||
break;
|
||||
}
|
||||
|
||||
if (!_Heap_Is_prev_used(next_block)) {
|
||||
if (do_dump)
|
||||
printk( " prev %p next %p", the_block->prev, the_block->next);
|
||||
if (_Heap_Block_size(the_block) != next_block->prev_size) {
|
||||
if (do_dump) printk("\n");
|
||||
printk("PASS: %d !front and back sizes don't match", source);
|
||||
error = 1;
|
||||
}
|
||||
if (!prev_used) {
|
||||
|
||||
hw_nl(do_dump, error);
|
||||
printk("PASS: %d !two consecutive blocks are free", source);
|
||||
error = 1;
|
||||
}
|
||||
|
||||
{ /* Check if 'the_block' is in the free block list */
|
||||
Heap_Block* block = _Heap_First_free_block(the_heap);
|
||||
if (!_Addresses_Is_aligned(block) ) {
|
||||
printk(
|
||||
"PASS: %d first free block %p is not aligned\n", source, block);
|
||||
error = 1;
|
||||
break;
|
||||
}
|
||||
while(block != the_block && block != tail) {
|
||||
if (!_Addresses_Is_aligned(block) ) {
|
||||
printk(
|
||||
"PASS: %d a free block %p is not aligned\n", source, block);
|
||||
error = 1;
|
||||
break;
|
||||
}
|
||||
if (!_Heap_Is_block_in_heap(the_heap, block)) {
|
||||
printk("PASS: %d a free block %p is not in heap\n", source, block);
|
||||
error = 1;
|
||||
break;
|
||||
}
|
||||
block = block->next;
|
||||
}
|
||||
if (block != the_block) {
|
||||
hw_nl(do_dump, error);
|
||||
printk("PASS: %d !the_block not in the free list", source);
|
||||
error = 1;
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
hw_nl(do_dump, error);
|
||||
|
||||
if (the_size < the_heap->min_block_size) {
|
||||
printk("PASS: %d !block size is too small\n", source);
|
||||
error = 1;
|
||||
if ( !_Heap_Is_aligned( block_size, page_size ) ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: block size %u not page (%u) aligned\n",
|
||||
block,
|
||||
block_size,
|
||||
page_size
|
||||
);
|
||||
break;
|
||||
}
|
||||
if (!_Heap_Is_aligned( the_size, the_heap->page_size)) {
|
||||
printk("PASS: %d !block size is misaligned\n", source);
|
||||
error = 1;
|
||||
|
||||
if ( block_size < min_block_size ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: size %u < min block size %u\n",
|
||||
block,
|
||||
block_size,
|
||||
min_block_size
|
||||
);
|
||||
break;
|
||||
}
|
||||
|
||||
if (++passes > (do_dump ? 10 : 0) && error)
|
||||
if ( next_block_begin <= block_begin ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: next block 0x%08x is not a successor\n",
|
||||
block,
|
||||
next_block
|
||||
);
|
||||
break;
|
||||
}
|
||||
|
||||
the_block = next_block;
|
||||
if ( !_Heap_Is_prev_used( next_block ) ) {
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: prev 0x%08x%s, next 0x%08x%s\n",
|
||||
block,
|
||||
block->prev,
|
||||
block->prev == first_free_block ?
|
||||
" (= first)"
|
||||
: (block->prev == free_list_head ? " (= head)" : ""),
|
||||
block->next,
|
||||
block->next == last_free_block ?
|
||||
" (= last)"
|
||||
: (block->next == free_list_tail ? " (= tail)" : "")
|
||||
);
|
||||
|
||||
if ( block_size != next_block->prev_size ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: size %u != size %u (in next block 0x%08x)\n",
|
||||
block,
|
||||
block_size,
|
||||
next_block->prev_size,
|
||||
next_block
|
||||
);
|
||||
}
|
||||
|
||||
if ( !prev_used ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: two consecutive blocks are free\n",
|
||||
block
|
||||
);
|
||||
}
|
||||
|
||||
if ( !_Heap_Walk_is_in_free_list( heap, block ) ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: free block not in free list\n",
|
||||
block
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
block = next_block;
|
||||
}
|
||||
|
||||
if (the_block != end) {
|
||||
printk("PASS: %d !last block address isn't equal to 'final' %p %p\n",
|
||||
source, the_block, end);
|
||||
error = 1;
|
||||
if ( !_Addresses_Is_aligned( block ) ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"block 0x%08x: not CPU aligned\n",
|
||||
block
|
||||
);
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
if (_Heap_Block_size(the_block) != the_heap->page_size) {
|
||||
printk("PASS: %d !last block's size isn't page_size (%d != %d)\n", source,
|
||||
_Heap_Block_size(the_block), the_heap->page_size);
|
||||
error = 1;
|
||||
if ( block == last_block ) {
|
||||
uintptr_t const block_size = _Heap_Block_size( block );
|
||||
|
||||
if ( block_size != page_size ) {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"last block 0x%08x: size %u != page size %u\n",
|
||||
block,
|
||||
block_size,
|
||||
page_size
|
||||
);
|
||||
}
|
||||
} else {
|
||||
error = true;
|
||||
_Heap_Walk_printk(
|
||||
source,
|
||||
dump,
|
||||
error,
|
||||
"last block 0x%08x != last block 0x%08x\n",
|
||||
block,
|
||||
last_block
|
||||
);
|
||||
}
|
||||
|
||||
if (do_dump && error)
|
||||
_Internal_error_Occurred( INTERNAL_ERROR_CORE, true, 0xffff0000 );
|
||||
|
||||
return error;
|
||||
|
||||
}
|
||||
|
||||
/*
|
||||
* This method exists to simplify branch paths in the generated code above.
|
||||
*/
|
||||
static void hw_nl(
|
||||
int error,
|
||||
bool do_dump
|
||||
)
|
||||
{
|
||||
if (do_dump || error) printk("\n");
|
||||
return !error;
|
||||
}
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -18,7 +26,7 @@
|
||||
|
||||
void *_Protected_heap_Allocate(
|
||||
Heap_Control *the_heap,
|
||||
intptr_t size
|
||||
uintptr_t size
|
||||
)
|
||||
{
|
||||
void *p;
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -18,8 +26,8 @@
|
||||
|
||||
void *_Protected_heap_Allocate_aligned(
|
||||
Heap_Control *the_heap,
|
||||
intptr_t size,
|
||||
uint32_t alignment
|
||||
uintptr_t size,
|
||||
uintptr_t alignment
|
||||
)
|
||||
{
|
||||
void *p;
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -19,11 +27,11 @@
|
||||
bool _Protected_heap_Extend(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t size
|
||||
uintptr_t size
|
||||
)
|
||||
{
|
||||
Heap_Extend_status status;
|
||||
intptr_t amount_extended;
|
||||
uintptr_t amount_extended;
|
||||
|
||||
_RTEMS_Lock_allocator();
|
||||
status = _Heap_Extend(the_heap, starting_address, size, &amount_extended);
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -19,10 +27,10 @@
|
||||
bool _Protected_heap_Get_block_size(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t *size
|
||||
uintptr_t *size
|
||||
)
|
||||
{
|
||||
bool status;
|
||||
bool status;
|
||||
|
||||
_RTEMS_Lock_allocator();
|
||||
status = _Heap_Size_of_alloc_area( the_heap, starting_address, size );
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -21,8 +29,6 @@ bool _Protected_heap_Get_information(
|
||||
Heap_Information_block *the_info
|
||||
)
|
||||
{
|
||||
Heap_Get_information_status status;
|
||||
|
||||
if ( !the_heap )
|
||||
return false;
|
||||
|
||||
@@ -30,11 +36,8 @@ bool _Protected_heap_Get_information(
|
||||
return false;
|
||||
|
||||
_RTEMS_Lock_allocator();
|
||||
status = _Heap_Get_information( the_heap, the_info );
|
||||
_Heap_Get_information( the_heap, the_info );
|
||||
_RTEMS_Unlock_allocator();
|
||||
|
||||
if ( status == HEAP_GET_INFORMATION_SUCCESSFUL )
|
||||
return true;
|
||||
|
||||
return false;
|
||||
return true;
|
||||
}
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2009.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -16,7 +24,7 @@
|
||||
#include <rtems/system.h>
|
||||
#include <rtems/score/protectedheap.h>
|
||||
|
||||
uint32_t _Protected_heap_Get_size(
|
||||
uintptr_t _Protected_heap_Get_size(
|
||||
Heap_Control *the_heap
|
||||
)
|
||||
{
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
@@ -19,12 +27,12 @@
|
||||
bool _Protected_heap_Resize_block(
|
||||
Heap_Control *the_heap,
|
||||
void *starting_address,
|
||||
intptr_t size
|
||||
uintptr_t size
|
||||
)
|
||||
{
|
||||
Heap_Resize_status status;
|
||||
intptr_t old_mem_size;
|
||||
intptr_t avail_mem_size;
|
||||
uintptr_t old_mem_size;
|
||||
uintptr_t avail_mem_size;
|
||||
|
||||
_RTEMS_Lock_allocator();
|
||||
status = _Heap_Resize_block(
|
||||
|
||||
@@ -1,4 +1,12 @@
|
||||
/**
|
||||
* @file
|
||||
*
|
||||
* @ingroup ScoreProtHeap
|
||||
*
|
||||
* @brief Protected Heap Handler implementation.
|
||||
*/
|
||||
|
||||
/*
|
||||
* COPYRIGHT (c) 1989-2007.
|
||||
* On-Line Applications Research Corporation (OAR).
|
||||
*
|
||||
|
||||
Reference in New Issue
Block a user