rockbox/firmware/test/memory/memory-slab.c
Alan Korr c25510f944 Now memory-page and memory-misc compile fine (others are in stage-development)
Conventions :
* Public headers :
    memory.h,config.h,defines.h,inlines.h,types.h,functions.h

* Private headers :
    memory-page.h,memory-slab.h (here you can find
    prototypes functions or structures we want to share
    only between memory-page.c, memory-slab.c,
    memory-block.c, memory-misc.c).

* Public or private codes in :
    memory-page.c,memory-slab.c,memory-block.c,memory-misc.c


git-svn-id: svn://svn.rockbox.org/rockbox/trunk@126 a1c6a512-1295-4272-9138-f99709370657
2002-04-17 15:00:28 +00:00

409 lines
12 KiB
C

/***************************************************************************
* __________ __ ___.
* Open \______ \ ____ ____ | | _\_ |__ _______ ___
* Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
* Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
* Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
* \/ \/ \/ \/ \/
* $Id$
*
* Copyright (C) 2002 by Alan Korr
*
* All files in this archive are subject to the GNU General Public License.
* See the file COPYING in the source tree root for full license agreement.
*
* This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
* KIND, either express or implied.
*
****************************************************************************/
#if 0
#include <memory.h>
#include "memory-page.h"
#include "memory-slab.h"
///////////////////////////////////////////////////////////////////////////////
// MEMORY SLAB :
////////////////
//
//
static inline struct memory_slab *__memory_push_slab (struct memory_slab *head,struct memory_slab *node)
{
node->less = head;
if (head)
{
node->more = head->more;
head->more = node;
}
else
node->more = 0;
return node;
}
static inline struct memory_slab *__memory_pop_slab (struct memory_slab *head,struct memory_slab *node)
{
if (head)
head->more = node->more;
return node->more;
}
static inline struct memory_slab *__memory_move_slab (struct memory_slab **from,struct memory_slab **to)
{
struct memory_slab *head = *from;
*from = (*from)->more;
if (*from)
(*from)->less = head->less;
head->less = 0;
head->more = (*to);
if (*to)
(*to)->prev = head;
*to = head;
return head;
}
//
///////////////////////////////////////////////////////////////////////////////
///////////////////////////////////////////////////////////////////////////////
// MEMORY CACHE :
/////////////////
//
//
static struct memory_cache *cache_tree;
static inline int __memory_get_order (unsigned size)
{
int order = 0;
size = (size + sizeof(struct memory_free_block) - 1) & - sizeof(struct memory_free_block);
while (size > 0)
{
++order; size <<= 1;
}
return order;
}
static inline struct memory_slab *__memory_get_slab (struct memory_cache *cache,void *address)
{
#ifdef TEST
return (struct memory_slab *)((((unsigned)address + cache->page_size) & -cache->page_size) - sizeof (struct memory_slab));
#else
return (struct memory_slab *)((free_page + (((unsigned)address - free_page + cache->page_size) & -cache->page_size) - sizeof (struct memory_slab)));
#endif
}
static struct memory_cache *__memory_splay_cache (struct memory_cache *root,unsigned int left)
{
struct memory_cache *down;
struct memory_cache *less;
struct memory_cache *more;
struct memory_cache head;
head.less =
head.more = 0;
less =
more = &head;
while (1)
{
if (left < root->left)
{
if ((down = root->less))
{
if (left < down->left)
{
root->less = down->more;
down->more = root;
root = down;
if (!root->less)
break;
}
more->less = root;
more = root;
root = root->less;
continue;
}
break;
}
if (root->left < left)
{
if ((down = root->more))
{
if (root->left < left)
{
root->more = down->less;
down->less = root;
root = down;
if (!root->more)
break;
}
less->more = root;
less = root;
root = root->more;
continue;
}
}
break;
}
less->more = root->less;
more->less = root->more;
root->less = head.more;
root->more = head.less;
return root;
}
static inline struct memory_cache *__memory_insert_cache (struct memory_cache *root,struct memory_cache *node)
{
node->less =
node->more =
node->same = 0;
if (root)
{
if (node->left == ((root = __memory_splay_cache (root,node))->left))
{
node->less = root.less;
node->more = root.more;
node->same = root;
root->less = node;
}
else if (node < root)
{
node->less = root->less;
node->more = root;
root->less = 0;
}
else
{
node->less = root;
node->more = root->more;
node->more = 0;
}
}
return node;
}
static inline struct memory_cache *__memory_remove_cache (struct memory_cache *root,struct memory_cache *node)
{
if (root)
{
root = __memory_splay_cache (root,node);
if (root != node)
{
node->less->same = node->same;
if (node->same)
node->same->less = node->less;
return root;
}
if (root->less)
{
node = __memory_splay_page (root->less,node);
node->more = root->more;
}
else
node = root->more;
}
return root;
}
static inline struct memory_cache *__memory_move_cache (struct memory_cache *root,struct memory_cache *node,int delta)
{
if ((root = __memory_remove_cache (root,node)))
{
node->left += delta;
root = __memory_insert_cache (root,node);
}
return root;
}
//
/////////////////////
// PUBLIC FUNCTIONS :
/////////////////////
//
// - memory_grow_cache : allocate a new slab for a cache
// - memory_shrink_cache : release free slabs from a cache
// - memory_create_cache : create a new cache of size-fixed blocks
// - memory_destroy_cache : destroy the cache and release all the slabs
// - memory_cache_allocate : allocate a block from the cache
// - memory_cache_release : release a block in the cache
//
struct memory_slab *memory_grow_cache (struct memory_cache *cache)
{
struct memory_slab *slab;
unsigned int page;
if (cache)
{
page = (unsigned int)memory_allocate_page (cache->page_order);
if (page)
{
struct memory_free_block *block,**link;
slab = (struct memory_slab *)(page + cache->page_size - sizeof (struct memory_slab));
slab->free = 0;
slab->left = 0;
link = &slab->free;
for ((unsigned int)block = page;
(unsigned int)block + cache->size < (unsigned int)slab;
(unsigned int)block += cache->size)
{
*link = block;
link = &block->link;
++slab->free;
}
*link = 0;
cache->blocks_per_slab = slab->free;
cache->reap = __memory_push_slab (cache->reap,slab);
cache_tree = __memory_move_cache (cache_tree,cache,+1);
return slab;
}
}
return MEMORY_RETURN_FAILURE;
}
static int __memory_shrink_cache (struct memory_cache *cache,int all,int move)
{
struct memory_slab *slab;
unsigned int slabs = 0;
if (cache)
{
while ((slab = cache->reap))
{
++slabs;
cache->reap = __memory_pop_slab (cache->reap,slab);
memory_release_page ((void *)slab);
if (all)
continue;
if (move)
cache_tree = __memory_move_cache (cache_tree,cache,-slabs);
return MEMORY_RETURN_SUCCESS;
}
}
return MEMORY_RETURN_FAILURE;
}
int memory_shrink_cache (struct memory_cache *cache,int all)
{
return __memory_shrink_cache (cache,all,1 /* move cache in cache_tree */);
}
struct memory_cache *memory_create_cache (unsigned int size,int align,int flags)
{
struct memory_cache *cache;
unsigned int waste = 0,blocks_per_page;
int page_order;
unsigned int page_size;
unsigned int original_size = size;
// Align size on 'align' bytes ('align' should equal 1<<n)
// if 'align' is inferior to 4, 32-bit word alignment is done by default.
size = (align > 4) ? ((size + align - 1) & -align) : ((size + sizeof (int) - 1) & -sizeof (int));
if (!(cache = memory_cache_allocate (&cache_cache))
return MEMORY_RETURN_FAILURE;
cache->flags =
cache->left = 0;
cache->used =
cache->free =
cache->reap = 0;
cache->original_size = original_size;
cache->size = size;
page_size = 0;
page_order = MEMORY_PAGE_MINIMAL_SIZE;;
// Trying to determine what is the best number of pages per slab
for (;; ++order,(page_size <<= 1))
{
if (page_order >= MEMORY_MAXIMUM_PAGE_ORDER_PER_SLAB)
{
memory_cache_release (&cache_cache,cache);
return MEMORY_RETURN_FAILURE;
}
waste = page_size;
waste -= sizeof (struct memory_slab);
blocks_per_slab = waste / size;
waste -= block_per_slab * size;
if (blocks_per_slab < MEMORY_MINIMUM_BLOCKS_PER_SLAB)
{
++page_order; page_size <<= 1;
continue;
}
// below 3% of lost space is correct
if ((waste << 16) / page_size) < 1967)
break;
++page_order; page_size <<= 1;
}
cache->page_size = page_size;
cache->page_order = page_order;
cache_tree = __memory_insert_cache (cache_tree,cache);
return cache;
}
int memory_destroy_cache (struct memory_cache *cache)
{
/* FIX ME : this function shouldn't be called if there are still used blocks */
if (cache && !cache->free && !cache->used)
{
cache_tree = __memory_remove_cache (cache_tree,cache);
if (__memory_shrink_cache (cache,1 /* release all free slabs */,0 /* don't move in cache_tree */))
return memory_cache_release (&cache_cache,cache);
}
return MEMORY_RETURN_FAILURE;
}
void *memory_cache_allocate (struct memory_cache *cache)
{
if (cache)
{
do
{
struct memory_slab *slab;
if ((slab = cache->free))
{
if (slab->left > 0)
{
ok: struct memory_free_block *block = slab->free;
slab->free = block->link;
if (--slab->left == 0)
__memory_move_slab (&cache->free,&cache->used);
return block;
}
}
if (cache->reap)
{
slab = __memory_move_slab (&cache->reap,&cache->free);
cache_tree = __memory_move_cache (cache_tree,cache,-1);
goto ok;
}
}
while (__memory_grow_cache (cache));
}
return MEMORY_RETURN_FAILURE;
}
int memory_cache_release (struct memory_cache *cache,void *address)
{
struct memory_slab *slab = __memory_get_slab (cache,address);
((struct memory_free_block *)address)->link = slab->free;
slab->free = (struct memory_free_block *)address;
if (slab->left++ == 0)
__memory_move_slab (&cache->used,&cache->free);
else if (slab->left == cache->blocks_per_slab)
{
__memory_move_slab (&cache->free,&cache->reap);
cache_tree = __memory_move_cache (cache_tree,cache,+1);
}
return MEMORY_RETURN_SUCCESS;
}
//
///////////////////////////////////////////////////////////////////////////////
#endif