Buflib: Stop caching the first unallocated block. It has little benefit but is complicated to keep up-to-date.
git-svn-id: svn://svn.rockbox.org/rockbox/trunk@30487 a1c6a512-1295-4272-9138-f99709370657
This commit is contained in:
parent
0dcbc6cd5d
commit
f7cff8bd69
2 changed files with 42 additions and 44 deletions
|
@ -89,8 +89,9 @@
|
||||||
#define BDEBUGF(...) do { } while(0)
|
#define BDEBUGF(...) do { } while(0)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
static union buflib_data*
|
static union buflib_data* find_first_free(struct buflib_context *ctx);
|
||||||
find_block_before(struct buflib_context *ctx, union buflib_data* block,
|
static union buflib_data* find_block_before(struct buflib_context *ctx,
|
||||||
|
union buflib_data* block,
|
||||||
bool is_free);
|
bool is_free);
|
||||||
/* Initialize buffer manager */
|
/* Initialize buffer manager */
|
||||||
void
|
void
|
||||||
|
@ -105,7 +106,6 @@ buflib_init(struct buflib_context *ctx, void *buf, size_t size)
|
||||||
ctx->handle_table = bd_buf + size;
|
ctx->handle_table = bd_buf + size;
|
||||||
ctx->last_handle = bd_buf + size;
|
ctx->last_handle = bd_buf + size;
|
||||||
ctx->first_free_handle = bd_buf + size - 1;
|
ctx->first_free_handle = bd_buf + size - 1;
|
||||||
ctx->first_free_block = bd_buf;
|
|
||||||
ctx->buf_start = bd_buf;
|
ctx->buf_start = bd_buf;
|
||||||
/* A marker is needed for the end of allocated data, to make sure that it
|
/* A marker is needed for the end of allocated data, to make sure that it
|
||||||
* does not collide with the handle table, and to detect end-of-buffer.
|
* does not collide with the handle table, and to detect end-of-buffer.
|
||||||
|
@ -226,11 +226,12 @@ static bool
|
||||||
buflib_compact(struct buflib_context *ctx)
|
buflib_compact(struct buflib_context *ctx)
|
||||||
{
|
{
|
||||||
BDEBUGF("%s(): Compacting!\n", __func__);
|
BDEBUGF("%s(): Compacting!\n", __func__);
|
||||||
union buflib_data *block;
|
union buflib_data *block,
|
||||||
|
*first_free = find_first_free(ctx);
|
||||||
int shift = 0, len;
|
int shift = 0, len;
|
||||||
/* Store the results of attempting to shrink the handle table */
|
/* Store the results of attempting to shrink the handle table */
|
||||||
bool ret = handle_table_shrink(ctx);
|
bool ret = handle_table_shrink(ctx);
|
||||||
for(block = ctx->first_free_block; block < ctx->alloc_end; block += len)
|
for(block = first_free; block < ctx->alloc_end; block += len)
|
||||||
{
|
{
|
||||||
len = block->val;
|
len = block->val;
|
||||||
/* This block is free, add its length to the shift value */
|
/* This block is free, add its length to the shift value */
|
||||||
|
@ -241,41 +242,41 @@ buflib_compact(struct buflib_context *ctx)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
/* attempt to fill any hole */
|
/* attempt to fill any hole */
|
||||||
if (-ctx->first_free_block->val > block->val)
|
if (-first_free->val >= block->val)
|
||||||
{
|
{
|
||||||
intptr_t size = ctx->first_free_block->val;
|
intptr_t size = -first_free->val;
|
||||||
union buflib_data* next_block = block + block->val;
|
union buflib_data* next_block = block + block->val;
|
||||||
if (move_block(ctx, block, ctx->first_free_block - block))
|
if (move_block(ctx, block, first_free - block))
|
||||||
{
|
{
|
||||||
/* moving was successful. Mark the next block as the new
|
/* moving was successful. Move alloc_end down if necessary */
|
||||||
* first_free_block and merge it with the free space
|
|
||||||
* that the move created */
|
|
||||||
if (ctx->alloc_end == next_block)
|
if (ctx->alloc_end == next_block)
|
||||||
ctx->alloc_end = block;
|
ctx->alloc_end = block;
|
||||||
ctx->first_free_block += block->val;
|
/* Mark the block behind the just moved as free
|
||||||
ctx->first_free_block->val = size + block->val;
|
* be careful to not overwrite an existing block */
|
||||||
|
if (size != block->val)
|
||||||
|
{
|
||||||
|
first_free += block->val;
|
||||||
|
first_free->val = block->val - size; /* negative */
|
||||||
|
}
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* attempt move the allocation by shift */
|
/* attempt move the allocation by shift */
|
||||||
if (shift)
|
if (shift)
|
||||||
{
|
{
|
||||||
/* failing to move creates a hole, therefore mark this
|
/* failing to move creates a hole,
|
||||||
* block as not allocated anymore and move first_free_block up */
|
* therefore mark this block as not allocated */
|
||||||
|
union buflib_data* target_block = block + shift;
|
||||||
if (!move_block(ctx, block, shift))
|
if (!move_block(ctx, block, shift))
|
||||||
{
|
{
|
||||||
union buflib_data* hole = block + shift;
|
target_block->val = shift; /* this is a hole */
|
||||||
hole->val = shift;
|
|
||||||
if (ctx->first_free_block > hole)
|
|
||||||
ctx->first_free_block = hole;
|
|
||||||
shift = 0;
|
shift = 0;
|
||||||
}
|
}
|
||||||
/* if move was successful, the just moved block is now
|
else
|
||||||
* possibly in place of the first free one, so move this thing up */
|
{ /* need to update the next free block, since the above hole
|
||||||
else if (ctx->first_free_block == block+shift)
|
* handling might make shift 0 before alloc_end is reached */
|
||||||
{
|
union buflib_data* new_free = target_block + target_block->val;
|
||||||
ctx->first_free_block += ctx->first_free_block->val;
|
new_free->val = shift;
|
||||||
ctx->first_free_block->val = shift;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -283,9 +284,6 @@ buflib_compact(struct buflib_context *ctx)
|
||||||
* been freed.
|
* been freed.
|
||||||
*/
|
*/
|
||||||
ctx->alloc_end += shift;
|
ctx->alloc_end += shift;
|
||||||
/* only move first_free_block up if it wasn't already by a hole */
|
|
||||||
if (ctx->first_free_block > ctx->alloc_end)
|
|
||||||
ctx->first_free_block = ctx->alloc_end;
|
|
||||||
ctx->compact = true;
|
ctx->compact = true;
|
||||||
return ret || shift;
|
return ret || shift;
|
||||||
}
|
}
|
||||||
|
@ -345,7 +343,6 @@ buflib_buffer_shift(struct buflib_context *ctx, int shift)
|
||||||
for (handle = ctx->last_handle; handle < ctx->handle_table; handle++)
|
for (handle = ctx->last_handle; handle < ctx->handle_table; handle++)
|
||||||
if (handle->alloc)
|
if (handle->alloc)
|
||||||
handle->alloc += shift;
|
handle->alloc += shift;
|
||||||
ctx->first_free_block += shift;
|
|
||||||
ctx->buf_start += shift;
|
ctx->buf_start += shift;
|
||||||
ctx->alloc_end += shift;
|
ctx->alloc_end += shift;
|
||||||
}
|
}
|
||||||
|
@ -443,7 +440,7 @@ buffer_alloc:
|
||||||
/* need to re-evaluate last before the loop because the last allocation
|
/* need to re-evaluate last before the loop because the last allocation
|
||||||
* possibly made room in its front to fit this, so last would be wrong */
|
* possibly made room in its front to fit this, so last would be wrong */
|
||||||
last = false;
|
last = false;
|
||||||
for (block = ctx->first_free_block;;block += block_len)
|
for (block = find_first_free(ctx);;block += block_len)
|
||||||
{
|
{
|
||||||
/* If the last used block extends all the way to the handle table, the
|
/* If the last used block extends all the way to the handle table, the
|
||||||
* block "after" it doesn't have a header. Because of this, it's easier
|
* block "after" it doesn't have a header. Because of this, it's easier
|
||||||
|
@ -499,8 +496,6 @@ buffer_alloc:
|
||||||
/* If we have just taken the first free block, the next allocation search
|
/* If we have just taken the first free block, the next allocation search
|
||||||
* can save some time by starting after this block.
|
* can save some time by starting after this block.
|
||||||
*/
|
*/
|
||||||
if (block == ctx->first_free_block)
|
|
||||||
ctx->first_free_block += size;
|
|
||||||
block += size;
|
block += size;
|
||||||
/* alloc_end must be kept current if we're taking the last block. */
|
/* alloc_end must be kept current if we're taking the last block. */
|
||||||
if (last)
|
if (last)
|
||||||
|
@ -512,6 +507,20 @@ buffer_alloc:
|
||||||
return ctx->handle_table - handle;
|
return ctx->handle_table - handle;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static union buflib_data*
|
||||||
|
find_first_free(struct buflib_context *ctx)
|
||||||
|
{
|
||||||
|
union buflib_data* ret = ctx->buf_start;
|
||||||
|
while(ret < ctx->alloc_end)
|
||||||
|
{
|
||||||
|
if (ret->val < 0)
|
||||||
|
break;
|
||||||
|
ret += ret->val;
|
||||||
|
}
|
||||||
|
/* ret is now either a free block or the same as alloc_end, both is fine */
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
/* Finds the free block before block, and returns NULL if it's not free */
|
/* Finds the free block before block, and returns NULL if it's not free */
|
||||||
static union buflib_data*
|
static union buflib_data*
|
||||||
find_block_before(struct buflib_context *ctx, union buflib_data* block,
|
find_block_before(struct buflib_context *ctx, union buflib_data* block,
|
||||||
|
@ -577,11 +586,6 @@ buflib_free(struct buflib_context *ctx, int handle_num)
|
||||||
}
|
}
|
||||||
handle_free(ctx, handle);
|
handle_free(ctx, handle);
|
||||||
handle->alloc = NULL;
|
handle->alloc = NULL;
|
||||||
/* If this block is before first_free_block, it becomes the new starting
|
|
||||||
* point for free-block search.
|
|
||||||
*/
|
|
||||||
if (block < ctx->first_free_block)
|
|
||||||
ctx->first_free_block = block;
|
|
||||||
|
|
||||||
return 0; /* unconditionally */
|
return 0; /* unconditionally */
|
||||||
}
|
}
|
||||||
|
@ -668,8 +672,6 @@ buflib_shrink(struct buflib_context* ctx, int handle, void* new_start, size_t ne
|
||||||
union buflib_data *free_before = find_block_before(ctx, block, true);
|
union buflib_data *free_before = find_block_before(ctx, block, true);
|
||||||
if (free_before)
|
if (free_before)
|
||||||
free_before->val += block->val;
|
free_before->val += block->val;
|
||||||
else if (ctx->first_free_block > block)
|
|
||||||
ctx->first_free_block = block;
|
|
||||||
|
|
||||||
/* We didn't handle size changes yet, assign block to the new one
|
/* We didn't handle size changes yet, assign block to the new one
|
||||||
* the code below the wants block whether it changed or not */
|
* the code below the wants block whether it changed or not */
|
||||||
|
@ -690,9 +692,6 @@ buflib_shrink(struct buflib_context* ctx, int handle, void* new_start, size_t ne
|
||||||
/* must be negative to indicate being unallocated */
|
/* must be negative to indicate being unallocated */
|
||||||
new_next_block->val = new_next_block - old_next_block;
|
new_next_block->val = new_next_block - old_next_block;
|
||||||
}
|
}
|
||||||
/* update first_free_block for the newly created free space */
|
|
||||||
if (ctx->first_free_block > new_next_block)
|
|
||||||
ctx->first_free_block = new_next_block;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return true;
|
return true;
|
||||||
|
|
|
@ -47,7 +47,6 @@ struct buflib_context
|
||||||
union buflib_data *handle_table;
|
union buflib_data *handle_table;
|
||||||
union buflib_data *first_free_handle;
|
union buflib_data *first_free_handle;
|
||||||
union buflib_data *last_handle;
|
union buflib_data *last_handle;
|
||||||
union buflib_data *first_free_block;
|
|
||||||
union buflib_data *buf_start;
|
union buflib_data *buf_start;
|
||||||
union buflib_data *alloc_end;
|
union buflib_data *alloc_end;
|
||||||
bool compact;
|
bool compact;
|
||||||
|
|
Loading…
Reference in a new issue