2005-02-13 18:55:14 +00:00
|
|
|
/***************************************************************************
|
2002-04-22 12:07:34 +00:00
|
|
|
* __________ __ ___.
|
|
|
|
* Open \______ \ ____ ____ | | _\_ |__ _______ ___
|
|
|
|
* Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
|
|
|
|
* Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
|
|
|
|
* Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
|
|
|
|
* \/ \/ \/ \/ \/
|
|
|
|
* $Id$
|
|
|
|
*
|
|
|
|
* Copyright (C) 2002 by Ulf Ralberg
|
|
|
|
*
|
|
|
|
* All files in this archive are subject to the GNU General Public License.
|
|
|
|
* See the file COPYING in the source tree root for full license agreement.
|
|
|
|
*
|
|
|
|
* This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
|
|
|
|
* KIND, either express or implied.
|
|
|
|
*
|
|
|
|
****************************************************************************/
|
2004-10-15 02:13:43 +00:00
|
|
|
#include "config.h"
|
2003-02-14 09:44:34 +00:00
|
|
|
#include <stdbool.h>
|
2002-04-22 12:07:34 +00:00
|
|
|
#include "thread.h"
|
2002-07-15 22:21:18 +00:00
|
|
|
#include "panic.h"
|
2005-08-26 22:52:31 +00:00
|
|
|
#include "system.h"
|
2003-02-14 09:44:34 +00:00
|
|
|
#include "kernel.h"
|
2004-10-15 02:13:43 +00:00
|
|
|
#include "cpu.h"
|
2006-09-16 16:18:11 +00:00
|
|
|
#include "string.h"
|
2006-09-23 14:38:04 +00:00
|
|
|
#ifdef RB_PROFILE
|
|
|
|
#include <profile.h>
|
|
|
|
#endif
|
2002-04-22 12:07:34 +00:00
|
|
|
|
2005-01-24 13:32:52 +00:00
|
|
|
#define DEADBEEF ((unsigned int)0xdeadbeef)
|
|
|
|
/* Cast to the the machine int type, whose size could be < 4. */
|
2005-01-24 13:18:27 +00:00
|
|
|
|
2006-09-02 07:56:52 +00:00
|
|
|
struct core_entry cores[NUM_CORES] IBSS_ATTR;
|
2006-09-16 16:18:11 +00:00
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
|
|
|
static unsigned short highest_priority IBSS_ATTR;
|
|
|
|
#endif
|
2006-10-15 11:57:52 +00:00
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
2006-11-11 05:33:24 +00:00
|
|
|
static int boosted_threads IBSS_ATTR;
|
2006-10-15 11:57:52 +00:00
|
|
|
#endif
|
2006-09-16 16:18:11 +00:00
|
|
|
|
|
|
|
/* Define to enable additional checks for blocking violations etc. */
|
2006-11-11 05:33:24 +00:00
|
|
|
#define THREAD_EXTRA_CHECKS
|
2005-01-24 13:18:27 +00:00
|
|
|
|
2004-08-03 19:22:56 +00:00
|
|
|
static const char main_thread_name[] = "main";
|
2002-07-15 22:21:18 +00:00
|
|
|
|
|
|
|
extern int stackbegin[];
|
|
|
|
extern int stackend[];
|
2002-04-22 12:07:34 +00:00
|
|
|
|
2006-08-21 17:35:35 +00:00
|
|
|
#ifdef CPU_PP
|
|
|
|
#ifndef BOOTLOADER
|
|
|
|
extern int cop_stackbegin[];
|
|
|
|
extern int cop_stackend[];
|
|
|
|
#else
|
2006-11-11 05:33:24 +00:00
|
|
|
/* The coprocessor stack is not set up in the bootloader code, but the threading
|
|
|
|
* is. No threads are run on the coprocessor, so set up some dummy stack */
|
2006-08-21 17:35:35 +00:00
|
|
|
int *cop_stackbegin = stackbegin;
|
|
|
|
int *cop_stackend = stackend;
|
|
|
|
#endif
|
|
|
|
#endif
|
|
|
|
|
2007-03-04 20:06:41 +00:00
|
|
|
#if (NUM_CORES > 1)
|
|
|
|
bool IDATA_ATTR kernel_running_on_cop = false;
|
|
|
|
#endif
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
/* Conserve IRAM
|
|
|
|
static void add_to_list(struct thread_entry **list,
|
|
|
|
struct thread_entry *thread) ICODE_ATTR;
|
|
|
|
static void remove_from_list(struct thread_entry **list,
|
|
|
|
struct thread_entry *thread) ICODE_ATTR;
|
|
|
|
*/
|
|
|
|
|
|
|
|
void switch_thread(bool save_context, struct thread_entry **blocked_list)
|
|
|
|
ICODE_ATTR;
|
|
|
|
|
2005-05-23 19:48:55 +00:00
|
|
|
static inline void store_context(void* addr) __attribute__ ((always_inline));
|
2006-11-11 05:33:24 +00:00
|
|
|
static inline void load_context(const void* addr)
|
|
|
|
__attribute__ ((always_inline));
|
2002-08-01 08:14:56 +00:00
|
|
|
|
2005-12-10 19:51:56 +00:00
|
|
|
#if defined(CPU_ARM)
|
2006-08-21 17:35:35 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
2005-11-13 23:47:38 +00:00
|
|
|
* Store non-volatile context.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
2005-11-07 23:07:19 +00:00
|
|
|
static inline void store_context(void* addr)
|
|
|
|
{
|
2005-11-13 23:47:38 +00:00
|
|
|
asm volatile(
|
2006-01-31 14:48:10 +00:00
|
|
|
"stmia %0, { r4-r11, sp, lr }\n"
|
2005-11-13 23:47:38 +00:00
|
|
|
: : "r" (addr)
|
|
|
|
);
|
2005-11-07 23:07:19 +00:00
|
|
|
}
|
|
|
|
|
2005-11-13 23:47:38 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
|
|
|
* Load non-volatile context.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
2005-11-07 23:07:19 +00:00
|
|
|
static inline void load_context(const void* addr)
|
|
|
|
{
|
2005-11-13 23:47:38 +00:00
|
|
|
asm volatile(
|
2006-01-31 14:48:10 +00:00
|
|
|
"ldmia %0, { r4-r11, sp, lr }\n" /* load regs r4 to r14 from context */
|
|
|
|
"ldr r0, [%0, #40] \n" /* load start pointer */
|
2005-12-10 19:51:56 +00:00
|
|
|
"mov r1, #0 \n"
|
|
|
|
"cmp r0, r1 \n" /* check for NULL */
|
2006-01-31 14:48:10 +00:00
|
|
|
"strne r1, [%0, #40] \n" /* if it's NULL, we're already running */
|
2005-12-10 19:51:56 +00:00
|
|
|
"movne pc, r0 \n" /* not already running, so jump to start */
|
2005-11-13 23:47:38 +00:00
|
|
|
: : "r" (addr) : "r0", "r1"
|
|
|
|
);
|
2005-11-07 23:07:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(CPU_COLDFIRE)
|
2006-08-21 17:35:35 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
2004-10-15 02:13:43 +00:00
|
|
|
* Store non-volatile context.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
|
|
|
static inline void store_context(void* addr)
|
|
|
|
{
|
2005-06-10 23:05:15 +00:00
|
|
|
asm volatile (
|
2005-09-01 20:06:38 +00:00
|
|
|
"move.l %%macsr,%%d0 \n"
|
|
|
|
"movem.l %%d0/%%d2-%%d7/%%a2-%%a7,(%0) \n"
|
|
|
|
: : "a" (addr) : "d0" /* only! */
|
2005-06-10 23:05:15 +00:00
|
|
|
);
|
2004-10-15 02:13:43 +00:00
|
|
|
}
|
|
|
|
|
2005-09-01 20:06:38 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
2004-10-15 02:13:43 +00:00
|
|
|
* Load non-volatile context.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
|
|
|
static inline void load_context(const void* addr)
|
|
|
|
{
|
2005-06-10 23:05:15 +00:00
|
|
|
asm volatile (
|
2005-09-01 20:06:38 +00:00
|
|
|
"movem.l (%0),%%d0/%%d2-%%d7/%%a2-%%a7 \n" /* Load context */
|
|
|
|
"move.l %%d0,%%macsr \n"
|
|
|
|
"move.l (52,%0),%%d0 \n" /* Get start address */
|
2007-03-09 08:03:18 +00:00
|
|
|
"beq.b 1f \n" /* NULL -> already running */
|
2005-09-01 20:06:38 +00:00
|
|
|
"clr.l (52,%0) \n" /* Clear start address.. */
|
|
|
|
"move.l %%d0,%0 \n"
|
|
|
|
"jmp (%0) \n" /* ..and start the thread */
|
2007-03-09 08:03:18 +00:00
|
|
|
"1: \n"
|
2005-06-10 23:05:15 +00:00
|
|
|
: : "a" (addr) : "d0" /* only! */
|
|
|
|
);
|
2004-10-15 02:13:43 +00:00
|
|
|
}
|
|
|
|
|
2007-02-25 21:43:10 +00:00
|
|
|
/* Set EMAC unit to fractional mode with saturation for each new thread,
|
|
|
|
since that's what'll be the most useful for most things which the dsp
|
|
|
|
will do. Codecs should still initialize their preferred modes
|
|
|
|
explicitly. */
|
|
|
|
#define THREAD_CPU_INIT(core, thread) \
|
|
|
|
({ (thread)->context.macsr = EMAC_FRACTIONAL | EMAC_SATURATE; })
|
|
|
|
|
2004-10-15 02:13:43 +00:00
|
|
|
#elif CONFIG_CPU == SH7034
|
2005-09-01 20:06:38 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
2002-04-22 12:07:34 +00:00
|
|
|
* Store non-volatile context.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
2002-06-25 12:04:23 +00:00
|
|
|
static inline void store_context(void* addr)
|
2002-04-22 12:07:34 +00:00
|
|
|
{
|
2005-06-10 23:05:15 +00:00
|
|
|
asm volatile (
|
|
|
|
"add #36,%0 \n"
|
|
|
|
"sts.l pr, @-%0 \n"
|
|
|
|
"mov.l r15,@-%0 \n"
|
|
|
|
"mov.l r14,@-%0 \n"
|
|
|
|
"mov.l r13,@-%0 \n"
|
|
|
|
"mov.l r12,@-%0 \n"
|
|
|
|
"mov.l r11,@-%0 \n"
|
|
|
|
"mov.l r10,@-%0 \n"
|
|
|
|
"mov.l r9, @-%0 \n"
|
|
|
|
"mov.l r8, @-%0 \n"
|
|
|
|
: : "r" (addr)
|
|
|
|
);
|
2002-04-22 12:07:34 +00:00
|
|
|
}
|
|
|
|
|
2005-06-10 23:05:15 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
2002-04-22 12:07:34 +00:00
|
|
|
* Load non-volatile context.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
2004-08-16 23:37:23 +00:00
|
|
|
static inline void load_context(const void* addr)
|
2002-04-22 12:07:34 +00:00
|
|
|
{
|
2005-06-10 23:05:15 +00:00
|
|
|
asm volatile (
|
|
|
|
"mov.l @%0+,r8 \n"
|
|
|
|
"mov.l @%0+,r9 \n"
|
|
|
|
"mov.l @%0+,r10 \n"
|
|
|
|
"mov.l @%0+,r11 \n"
|
|
|
|
"mov.l @%0+,r12 \n"
|
|
|
|
"mov.l @%0+,r13 \n"
|
|
|
|
"mov.l @%0+,r14 \n"
|
|
|
|
"mov.l @%0+,r15 \n"
|
|
|
|
"lds.l @%0+,pr \n"
|
|
|
|
"mov.l @%0,r0 \n" /* Get start address */
|
|
|
|
"tst r0,r0 \n"
|
|
|
|
"bt .running \n" /* NULL -> already running */
|
|
|
|
"lds r0,pr \n"
|
|
|
|
"mov #0,r0 \n"
|
|
|
|
"rts \n" /* Start the thread */
|
|
|
|
"mov.l r0,@%0 \n" /* Clear start address */
|
|
|
|
".running: \n"
|
|
|
|
: : "r" (addr) : "r0" /* only! */
|
|
|
|
);
|
2002-04-22 12:07:34 +00:00
|
|
|
}
|
2005-06-10 23:05:15 +00:00
|
|
|
|
2004-10-15 02:13:43 +00:00
|
|
|
#endif
|
2002-04-22 12:07:34 +00:00
|
|
|
|
2007-02-25 21:43:10 +00:00
|
|
|
#ifndef THREAD_CPU_INIT
|
|
|
|
/* No cpu specific init - make empty */
|
|
|
|
#define THREAD_CPU_INIT(core, thread)
|
|
|
|
#endif
|
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
static void add_to_list(struct thread_entry **list, struct thread_entry *thread)
|
2002-04-22 12:07:34 +00:00
|
|
|
{
|
2006-09-16 16:18:11 +00:00
|
|
|
if (*list == NULL)
|
|
|
|
{
|
|
|
|
thread->next = thread;
|
|
|
|
thread->prev = thread;
|
|
|
|
*list = thread;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/* Insert last */
|
|
|
|
thread->next = *list;
|
|
|
|
thread->prev = (*list)->prev;
|
|
|
|
thread->prev->next = thread;
|
|
|
|
(*list)->prev = thread;
|
|
|
|
|
|
|
|
/* Insert next
|
|
|
|
thread->next = (*list)->next;
|
|
|
|
thread->prev = *list;
|
|
|
|
thread->next->prev = thread;
|
|
|
|
(*list)->next = thread;
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
}
|
2002-04-22 12:07:34 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
static void remove_from_list(struct thread_entry **list,
|
|
|
|
struct thread_entry *thread)
|
|
|
|
{
|
|
|
|
if (list != NULL)
|
|
|
|
{
|
|
|
|
if (thread == thread->next)
|
|
|
|
{
|
|
|
|
*list = NULL;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (thread == *list)
|
|
|
|
*list = thread->next;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Fix links to jump over the removed entry. */
|
|
|
|
thread->prev->next = thread->next;
|
|
|
|
thread->next->prev = thread->prev;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Compiler trick: Don't declare as static to prevent putting
|
|
|
|
* function in IRAM. */
|
|
|
|
void check_sleepers(void)
|
|
|
|
{
|
|
|
|
struct thread_entry *current, *next;
|
|
|
|
|
|
|
|
/* Check sleeping threads. */
|
|
|
|
current = cores[CURRENT_CORE].sleeping;
|
|
|
|
if (current == NULL)
|
|
|
|
return ;
|
|
|
|
|
|
|
|
for (;;)
|
|
|
|
{
|
|
|
|
next = current->next;
|
|
|
|
|
|
|
|
if ((unsigned)current_tick >= GET_STATE_ARG(current->statearg))
|
|
|
|
{
|
|
|
|
/* Sleep timeout has been reached so bring the thread
|
|
|
|
* back to life again. */
|
|
|
|
remove_from_list(&cores[CURRENT_CORE].sleeping, current);
|
|
|
|
add_to_list(&cores[CURRENT_CORE].running, current);
|
2006-11-11 05:33:24 +00:00
|
|
|
current->statearg = 0;
|
2006-09-16 16:18:11 +00:00
|
|
|
|
|
|
|
/* If there is no more processes in the list, break the loop. */
|
|
|
|
if (cores[CURRENT_CORE].sleeping == NULL)
|
|
|
|
break;
|
|
|
|
|
|
|
|
current = next;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
current = next;
|
|
|
|
|
|
|
|
/* Break the loop once we have walked through the list of all
|
|
|
|
* sleeping processes. */
|
|
|
|
if (current == cores[CURRENT_CORE].sleeping)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-03-26 03:24:36 +00:00
|
|
|
/* Safely finish waking all threads potentialy woken by interrupts -
|
|
|
|
* statearg already zeroed in wakeup_thread. */
|
|
|
|
static void wake_list_awaken(void)
|
|
|
|
{
|
|
|
|
int oldlevel = set_irq_level(HIGHEST_IRQ_LEVEL);
|
|
|
|
|
|
|
|
/* No need for another check in the IRQ lock since IRQs are allowed
|
|
|
|
only to add threads to the waking list. They won't be adding more
|
|
|
|
until we're done here though. */
|
|
|
|
|
|
|
|
struct thread_entry *waking = cores[CURRENT_CORE].waking;
|
|
|
|
struct thread_entry *running = cores[CURRENT_CORE].running;
|
|
|
|
|
|
|
|
if (running != NULL)
|
|
|
|
{
|
|
|
|
/* Place waking threads at the end of the running list. */
|
|
|
|
struct thread_entry *tmp;
|
|
|
|
waking->prev->next = running;
|
|
|
|
running->prev->next = waking;
|
|
|
|
tmp = running->prev;
|
|
|
|
running->prev = waking->prev;
|
|
|
|
waking->prev = tmp;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
/* Just transfer the list as-is - just came out of a core
|
|
|
|
* sleep. */
|
|
|
|
cores[CURRENT_CORE].running = waking;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Done with waking list */
|
|
|
|
cores[CURRENT_CORE].waking = NULL;
|
|
|
|
set_irq_level(oldlevel);
|
|
|
|
}
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
static inline void sleep_core(void)
|
|
|
|
{
|
|
|
|
static long last_tick = 0;
|
2006-12-29 23:42:43 +00:00
|
|
|
#if CONFIG_CPU == S3C2440
|
|
|
|
int i;
|
|
|
|
#endif
|
2007-03-26 03:24:36 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
for (;;)
|
2003-02-14 09:44:34 +00:00
|
|
|
{
|
2007-03-26 03:24:36 +00:00
|
|
|
/* We want to do these ASAP as it may change the decision to sleep
|
|
|
|
the core or the core has woken because an interrupt occurred
|
|
|
|
and posted a message to a queue. */
|
|
|
|
if (cores[CURRENT_CORE].waking != NULL)
|
|
|
|
wake_list_awaken();
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
if (last_tick != current_tick)
|
|
|
|
{
|
|
|
|
check_sleepers();
|
|
|
|
last_tick = current_tick;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* We must sleep until there is at least one process in the list
|
|
|
|
* of running processes. */
|
|
|
|
if (cores[CURRENT_CORE].running != NULL)
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* Enter sleep mode to reduce power usage, woken up on interrupt */
|
2005-07-18 12:40:29 +00:00
|
|
|
#ifdef CPU_COLDFIRE
|
2004-10-15 02:13:43 +00:00
|
|
|
asm volatile ("stop #0x2000");
|
2005-06-10 23:05:15 +00:00
|
|
|
#elif CONFIG_CPU == SH7034
|
2005-08-26 22:52:31 +00:00
|
|
|
and_b(0x7F, &SBYCR);
|
2005-06-10 23:05:15 +00:00
|
|
|
asm volatile ("sleep");
|
2007-03-04 20:06:41 +00:00
|
|
|
#elif defined (CPU_PP)
|
2006-02-08 21:30:35 +00:00
|
|
|
/* This should sleep the CPU. It appears to wake by itself on
|
|
|
|
interrupts */
|
2007-03-04 20:06:41 +00:00
|
|
|
if (CURRENT_CORE == CPU)
|
|
|
|
CPU_CTL = PROC_SLEEP;
|
|
|
|
else
|
|
|
|
COP_CTL = PROC_SLEEP;
|
2006-08-12 08:01:54 +00:00
|
|
|
#elif CONFIG_CPU == S3C2440
|
2006-12-29 23:29:04 +00:00
|
|
|
CLKCON |= (1 << 2); /* set IDLE bit */
|
|
|
|
for(i=0; i<10; i++); /* wait for IDLE */
|
|
|
|
CLKCON &= ~(1 << 2); /* reset IDLE bit when wake up */
|
2004-10-15 02:13:43 +00:00
|
|
|
#endif
|
2003-02-14 09:44:34 +00:00
|
|
|
}
|
2006-09-16 16:18:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef RB_PROFILE
|
|
|
|
static int get_threadnum(struct thread_entry *thread)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < MAXTHREADS; i++)
|
|
|
|
{
|
|
|
|
if (&cores[CURRENT_CORE].threads[i] == thread)
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
|
|
|
|
return -1;
|
|
|
|
}
|
2006-09-23 14:38:04 +00:00
|
|
|
|
|
|
|
void profile_thread(void) {
|
|
|
|
profstart(get_threadnum(cores[CURRENT_CORE].running));
|
|
|
|
}
|
2004-10-15 02:13:43 +00:00
|
|
|
#endif
|
2005-06-10 23:05:15 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
/* Compiler trick: Don't declare as static to prevent putting
|
|
|
|
* function in IRAM. */
|
|
|
|
void change_thread_state(struct thread_entry **blocked_list)
|
|
|
|
{
|
|
|
|
struct thread_entry *old;
|
2006-11-11 05:33:24 +00:00
|
|
|
unsigned long new_state;
|
2006-09-16 16:18:11 +00:00
|
|
|
|
|
|
|
/* Remove the thread from the list of running threads. */
|
|
|
|
old = cores[CURRENT_CORE].running;
|
2006-11-11 05:33:24 +00:00
|
|
|
new_state = GET_STATE(old->statearg);
|
|
|
|
|
|
|
|
/* Check if a thread state change has been requested. */
|
|
|
|
if (new_state)
|
|
|
|
{
|
|
|
|
/* Change running thread state and switch to next thread. */
|
|
|
|
remove_from_list(&cores[CURRENT_CORE].running, old);
|
|
|
|
|
|
|
|
/* And put the thread into a new list of inactive threads. */
|
|
|
|
if (new_state == STATE_BLOCKED)
|
|
|
|
add_to_list(blocked_list, old);
|
|
|
|
else
|
|
|
|
add_to_list(&cores[CURRENT_CORE].sleeping, old);
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
2006-11-11 05:33:24 +00:00
|
|
|
/* Reset priorities */
|
|
|
|
if (old->priority == highest_priority)
|
|
|
|
highest_priority = 100;
|
2006-09-16 16:18:11 +00:00
|
|
|
#endif
|
2006-11-11 05:33:24 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
/* Switch to the next running thread. */
|
|
|
|
cores[CURRENT_CORE].running = old->next;
|
2006-09-16 16:18:11 +00:00
|
|
|
}
|
2005-06-10 23:05:15 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
|
|
|
* Switch thread in round robin fashion.
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
|
|
|
void switch_thread(bool save_context, struct thread_entry **blocked_list)
|
|
|
|
{
|
2006-01-18 20:54:13 +00:00
|
|
|
#ifdef RB_PROFILE
|
2006-09-16 16:18:11 +00:00
|
|
|
profile_thread_stopped(get_threadnum(cores[CURRENT_CORE].running));
|
|
|
|
#endif
|
|
|
|
unsigned int *stackptr;
|
|
|
|
|
|
|
|
#ifdef SIMULATOR
|
|
|
|
/* Do nothing */
|
|
|
|
#else
|
2007-03-26 03:24:36 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
/* Begin task switching by saving our current context so that we can
|
|
|
|
* restore the state of the current thread later to the point prior
|
|
|
|
* to this call. */
|
|
|
|
if (save_context)
|
|
|
|
{
|
|
|
|
store_context(&cores[CURRENT_CORE].running->context);
|
|
|
|
|
|
|
|
/* Check if the current thread stack is overflown */
|
|
|
|
stackptr = cores[CURRENT_CORE].running->stack;
|
|
|
|
if(stackptr[0] != DEADBEEF)
|
2006-11-11 05:33:24 +00:00
|
|
|
panicf("Stkov %s", cores[CURRENT_CORE].running->name);
|
|
|
|
|
|
|
|
/* Rearrange thread lists as needed */
|
|
|
|
change_thread_state(blocked_list);
|
2006-12-16 18:35:12 +00:00
|
|
|
|
|
|
|
#ifdef HAVE_EXTENDED_MESSAGING_AND_NAME
|
|
|
|
/* This has to be done after the scheduler is finished with the
|
|
|
|
blocked_list pointer so that an IRQ can't kill us by attempting
|
|
|
|
a wake but before attempting any core sleep. */
|
2007-03-09 08:03:18 +00:00
|
|
|
if (cores[CURRENT_CORE].switch_to_irq_level != STAY_IRQ_LEVEL)
|
2006-12-16 18:35:12 +00:00
|
|
|
{
|
2007-03-09 08:03:18 +00:00
|
|
|
int level = cores[CURRENT_CORE].switch_to_irq_level;
|
|
|
|
cores[CURRENT_CORE].switch_to_irq_level = STAY_IRQ_LEVEL;
|
2006-12-16 18:35:12 +00:00
|
|
|
set_irq_level(level);
|
|
|
|
}
|
|
|
|
#endif
|
2006-09-16 16:18:11 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Go through the list of sleeping task to check if we need to wake up
|
|
|
|
* any of them due to timeout. Also puts core into sleep state until
|
|
|
|
* there is at least one running process again. */
|
|
|
|
sleep_core();
|
|
|
|
|
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
|
|
|
/* Select the new task based on priorities and the last time a process
|
|
|
|
* got CPU time. */
|
|
|
|
for (;;)
|
|
|
|
{
|
|
|
|
int priority = cores[CURRENT_CORE].running->priority;
|
2007-03-09 08:03:18 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
if (priority < highest_priority)
|
|
|
|
highest_priority = priority;
|
2007-03-09 08:03:18 +00:00
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
if (priority == highest_priority ||
|
|
|
|
(current_tick - cores[CURRENT_CORE].running->last_run >
|
2007-03-09 08:03:18 +00:00
|
|
|
priority * 8) ||
|
|
|
|
cores[CURRENT_CORE].running->priority_x != 0)
|
2006-09-16 16:18:11 +00:00
|
|
|
break;
|
2006-11-11 05:33:24 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
cores[CURRENT_CORE].running = cores[CURRENT_CORE].running->next;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Reset the value of thread's last running time to the current time. */
|
|
|
|
cores[CURRENT_CORE].running->last_run = current_tick;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif
|
|
|
|
/* And finally give control to the next thread. */
|
|
|
|
load_context(&cores[CURRENT_CORE].running->context);
|
|
|
|
|
|
|
|
#ifdef RB_PROFILE
|
|
|
|
profile_thread_started(get_threadnum(cores[CURRENT_CORE].running));
|
2006-01-18 20:54:13 +00:00
|
|
|
#endif
|
2002-04-22 12:07:34 +00:00
|
|
|
}
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
void sleep_thread(int ticks)
|
2003-02-14 09:44:34 +00:00
|
|
|
{
|
2006-11-11 05:33:24 +00:00
|
|
|
struct thread_entry *current;
|
|
|
|
|
|
|
|
current = cores[CURRENT_CORE].running;
|
|
|
|
|
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
|
|
|
if (STATE_IS_BOOSTED(current->statearg)) {
|
|
|
|
boosted_threads--;
|
|
|
|
if (!boosted_threads)
|
2006-11-11 05:53:32 +00:00
|
|
|
{
|
2006-11-11 05:33:24 +00:00
|
|
|
cpu_boost(false);
|
2006-11-11 05:53:32 +00:00
|
|
|
}
|
2006-11-11 05:33:24 +00:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
/* Set the thread's new state and timeout and finally force a task switch
|
|
|
|
* so that scheduler removes thread from the list of running processes
|
|
|
|
* and puts it in list of sleeping tasks. */
|
2006-11-11 05:33:24 +00:00
|
|
|
SET_STATE(current->statearg, STATE_SLEEPING, current_tick + ticks + 1);
|
2006-09-16 16:18:11 +00:00
|
|
|
switch_thread(true, NULL);
|
2006-11-11 05:33:24 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void block_thread(struct thread_entry **list)
|
|
|
|
{
|
|
|
|
struct thread_entry *current;
|
|
|
|
/* Get the entry for the current running thread. */
|
|
|
|
current = cores[CURRENT_CORE].running;
|
|
|
|
|
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
|
|
|
/* Keep the boosted state over indefinite block calls, because
|
|
|
|
* we are waiting until the earliest time that someone else
|
|
|
|
* completes an action */
|
|
|
|
unsigned long boost_flag = STATE_IS_BOOSTED(current->statearg);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef THREAD_EXTRA_CHECKS
|
|
|
|
/* We are not allowed to mix blocking types in one queue. */
|
|
|
|
if (*list && GET_STATE((*list)->statearg) == STATE_BLOCKED_W_TMO)
|
|
|
|
panicf("Blocking violation B->*T");
|
|
|
|
#endif
|
2006-09-16 16:18:11 +00:00
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
/* Set the state to blocked and ask the scheduler to switch tasks,
|
|
|
|
* this takes us off of the run queue until we are explicitly woken */
|
|
|
|
SET_STATE(current->statearg, STATE_BLOCKED, 0);
|
2006-12-16 18:35:12 +00:00
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
switch_thread(true, list);
|
|
|
|
|
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
|
|
|
/* Reset only the boosted flag to indicate we are up and running again. */
|
|
|
|
current->statearg = boost_flag;
|
|
|
|
#else
|
2006-09-16 16:18:11 +00:00
|
|
|
/* Clear all flags to indicate we are up and running again. */
|
2006-11-11 05:33:24 +00:00
|
|
|
current->statearg = 0;
|
|
|
|
#endif
|
2003-02-14 09:44:34 +00:00
|
|
|
}
|
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
void block_thread_w_tmo(struct thread_entry **list, int timeout)
|
2003-02-14 09:44:34 +00:00
|
|
|
{
|
2006-09-16 16:18:11 +00:00
|
|
|
struct thread_entry *current;
|
|
|
|
/* Get the entry for the current running thread. */
|
|
|
|
current = cores[CURRENT_CORE].running;
|
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
|
|
|
/* A block with a timeout is a sleep situation, whatever we are waiting
|
|
|
|
* for _may or may not_ happen, regardless of boost state, (user input
|
|
|
|
* for instance), so this thread no longer needs to boost */
|
|
|
|
if (STATE_IS_BOOSTED(current->statearg)) {
|
|
|
|
boosted_threads--;
|
|
|
|
if (!boosted_threads)
|
2006-11-11 05:53:32 +00:00
|
|
|
{
|
2006-11-11 05:33:24 +00:00
|
|
|
cpu_boost(false);
|
2006-11-11 05:53:32 +00:00
|
|
|
}
|
2006-11-11 05:33:24 +00:00
|
|
|
}
|
2006-09-16 16:18:11 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef THREAD_EXTRA_CHECKS
|
2006-11-11 05:33:24 +00:00
|
|
|
/* We can store only one thread to the "list" if thread is used
|
|
|
|
* in other list (such as core's list for sleeping tasks). */
|
|
|
|
if (*list)
|
|
|
|
panicf("Blocking violation T->*B");
|
2006-09-16 16:18:11 +00:00
|
|
|
#endif
|
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
/* Set the state to blocked with the specified timeout */
|
|
|
|
SET_STATE(current->statearg, STATE_BLOCKED_W_TMO, current_tick + timeout);
|
|
|
|
|
|
|
|
/* Set the "list" for explicit wakeup */
|
|
|
|
*list = current;
|
|
|
|
|
|
|
|
/* Now force a task switch and block until we have been woken up
|
|
|
|
* by another thread or timeout is reached. */
|
|
|
|
switch_thread(true, NULL);
|
|
|
|
|
|
|
|
/* It is now safe for another thread to block on this "list" */
|
|
|
|
*list = NULL;
|
2003-02-14 09:44:34 +00:00
|
|
|
}
|
|
|
|
|
2006-12-16 18:35:12 +00:00
|
|
|
#if defined(HAVE_EXTENDED_MESSAGING_AND_NAME) && !defined(SIMULATOR)
|
|
|
|
void set_irq_level_and_block_thread(struct thread_entry **list, int level)
|
|
|
|
{
|
2007-03-09 08:03:18 +00:00
|
|
|
cores[CURRENT_CORE].switch_to_irq_level = level;
|
2006-12-16 18:35:12 +00:00
|
|
|
block_thread(list);
|
|
|
|
}
|
|
|
|
|
|
|
|
#if 0
|
|
|
|
void set_irq_level_and_block_thread_w_tmo(struct thread_entry **list,
|
|
|
|
int timeout, int level)
|
|
|
|
{
|
2007-03-09 08:03:18 +00:00
|
|
|
cores[CURRENT_CORE].switch_to_irq_level = level;
|
2006-12-16 18:35:12 +00:00
|
|
|
block_thread_w_tmo(list, timeout);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif /* HAVE_EXTENDED_MESSAGING_AND_NAME */
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
void wakeup_thread(struct thread_entry **list)
|
|
|
|
{
|
|
|
|
struct thread_entry *thread;
|
|
|
|
|
|
|
|
/* Check if there is a blocked thread at all. */
|
|
|
|
if (*list == NULL)
|
|
|
|
return ;
|
|
|
|
|
|
|
|
/* Wake up the last thread first. */
|
|
|
|
thread = *list;
|
|
|
|
|
|
|
|
/* Determine thread's current state. */
|
|
|
|
switch (GET_STATE(thread->statearg))
|
|
|
|
{
|
|
|
|
case STATE_BLOCKED:
|
|
|
|
/* Remove thread from the list of blocked threads and add it
|
2007-03-26 03:24:36 +00:00
|
|
|
* to the scheduler's list of running processes. List removal
|
|
|
|
* is safe since each object maintains it's own list of
|
|
|
|
* sleepers and queues protect against reentrancy. */
|
2006-09-16 16:18:11 +00:00
|
|
|
remove_from_list(list, thread);
|
2007-03-26 03:24:36 +00:00
|
|
|
add_to_list(cores[CURRENT_CORE].wakeup_list, thread);
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
case STATE_BLOCKED_W_TMO:
|
|
|
|
/* Just remove the timeout to cause scheduler to immediately
|
|
|
|
* wake up the thread. */
|
2006-11-11 05:33:24 +00:00
|
|
|
thread->statearg = 0;
|
2006-09-16 16:18:11 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
/* Nothing to do. Thread has already been woken up
|
|
|
|
* or it's state is not blocked or blocked with timeout. */
|
|
|
|
return ;
|
|
|
|
}
|
|
|
|
}
|
2006-08-21 17:35:35 +00:00
|
|
|
|
2007-03-26 03:24:36 +00:00
|
|
|
/* Like wakeup_thread but safe against IRQ corruption when IRQs are disabled
|
|
|
|
before calling. */
|
|
|
|
void wakeup_thread_irq_safe(struct thread_entry **list)
|
|
|
|
{
|
|
|
|
struct core_entry *core = &cores[CURRENT_CORE];
|
|
|
|
/* Switch wakeup lists and call wakeup_thread */
|
|
|
|
core->wakeup_list = &core->waking;
|
|
|
|
wakeup_thread(list);
|
|
|
|
/* Switch back to normal running list */
|
|
|
|
core->wakeup_list = &core->running;
|
|
|
|
}
|
|
|
|
|
2006-08-21 17:35:35 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
2007-03-04 20:06:41 +00:00
|
|
|
* Create a thread
|
|
|
|
* If using a dual core architecture, specify which core to start the thread
|
|
|
|
* on, and whether to fall back to the other core if it can't be created
|
|
|
|
* Return ID if context area could be allocated, else NULL.
|
2002-04-22 12:07:34 +00:00
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
2006-09-16 16:18:11 +00:00
|
|
|
struct thread_entry*
|
|
|
|
create_thread(void (*function)(void), void* stack, int stack_size,
|
2007-03-04 20:06:41 +00:00
|
|
|
const char *name IF_PRIO(, int priority)
|
|
|
|
IF_COP(, unsigned int core, bool fallback))
|
2002-04-22 12:07:34 +00:00
|
|
|
{
|
2005-06-10 23:05:15 +00:00
|
|
|
unsigned int i;
|
|
|
|
unsigned int stacklen;
|
|
|
|
unsigned int *stackptr;
|
2006-09-16 16:18:11 +00:00
|
|
|
int n;
|
2005-06-10 23:05:15 +00:00
|
|
|
struct regs *regs;
|
2006-09-02 07:56:52 +00:00
|
|
|
struct thread_entry *thread;
|
2005-06-10 23:05:15 +00:00
|
|
|
|
2007-03-04 20:06:41 +00:00
|
|
|
/*****
|
|
|
|
* Ugly code alert!
|
|
|
|
* To prevent ifdef hell while keeping the binary size down, we define
|
|
|
|
* core here if it hasn't been passed as a parameter
|
|
|
|
*****/
|
|
|
|
#if NUM_CORES == 1
|
|
|
|
#define core CPU
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if NUM_CORES > 1
|
|
|
|
/* If the kernel hasn't initialised on the COP (most likely due to an old
|
|
|
|
* bootloader) then refuse to start threads on the COP
|
|
|
|
*/
|
|
|
|
if((core == COP) && !kernel_running_on_cop)
|
|
|
|
{
|
|
|
|
if (fallback)
|
|
|
|
return create_thread(function, stack, stack_size, name
|
|
|
|
IF_PRIO(, priority) IF_COP(, CPU, false));
|
|
|
|
else
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
for (n = 0; n < MAXTHREADS; n++)
|
|
|
|
{
|
|
|
|
if (cores[core].threads[n].name == NULL)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (n == MAXTHREADS)
|
2007-03-04 20:06:41 +00:00
|
|
|
{
|
|
|
|
#if NUM_CORES > 1
|
|
|
|
if (fallback)
|
|
|
|
return create_thread(function, stack, stack_size, name
|
|
|
|
IF_PRIO(, priority) IF_COP(, 1 - core, fallback));
|
|
|
|
else
|
|
|
|
#endif
|
|
|
|
return NULL;
|
|
|
|
}
|
2006-09-16 16:18:11 +00:00
|
|
|
|
2005-06-10 23:05:15 +00:00
|
|
|
/* Munge the stack to make it easy to spot stack overflows */
|
|
|
|
stacklen = stack_size / sizeof(int);
|
|
|
|
stackptr = stack;
|
|
|
|
for(i = 0;i < stacklen;i++)
|
|
|
|
{
|
|
|
|
stackptr[i] = DEADBEEF;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Store interesting information */
|
2006-09-16 16:18:11 +00:00
|
|
|
thread = &cores[core].threads[n];
|
2006-09-02 07:56:52 +00:00
|
|
|
thread->name = name;
|
|
|
|
thread->stack = stack;
|
|
|
|
thread->stack_size = stack_size;
|
2006-09-16 16:18:11 +00:00
|
|
|
thread->statearg = 0;
|
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
2007-03-09 08:03:18 +00:00
|
|
|
thread->priority_x = 0;
|
2006-09-16 16:18:11 +00:00
|
|
|
thread->priority = priority;
|
|
|
|
highest_priority = 100;
|
|
|
|
#endif
|
|
|
|
add_to_list(&cores[core].running, thread);
|
|
|
|
|
2006-09-02 07:56:52 +00:00
|
|
|
regs = &thread->context;
|
2005-06-10 23:05:15 +00:00
|
|
|
/* Align stack to an even 32 bit boundary */
|
|
|
|
regs->sp = (void*)(((unsigned int)stack + stack_size) & ~3);
|
|
|
|
regs->start = (void*)function;
|
2004-06-22 11:32:36 +00:00
|
|
|
|
2007-02-25 21:43:10 +00:00
|
|
|
/* Do any CPU specific inits after initializing common items
|
|
|
|
to have access to valid data */
|
|
|
|
THREAD_CPU_INIT(core, thread);
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
return thread;
|
2007-03-04 20:06:41 +00:00
|
|
|
#if NUM_CORES == 1
|
|
|
|
#undef core
|
|
|
|
#endif
|
2004-04-30 20:23:04 +00:00
|
|
|
}
|
|
|
|
|
2006-10-15 11:57:52 +00:00
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
|
|
|
void trigger_cpu_boost(void)
|
|
|
|
{
|
2006-11-11 05:33:24 +00:00
|
|
|
if (!STATE_IS_BOOSTED(cores[CURRENT_CORE].running->statearg))
|
2006-10-15 11:57:52 +00:00
|
|
|
{
|
2006-11-11 05:33:24 +00:00
|
|
|
SET_BOOST_STATE(cores[CURRENT_CORE].running->statearg);
|
|
|
|
if (!boosted_threads)
|
2006-11-11 05:53:32 +00:00
|
|
|
{
|
2006-11-11 05:33:24 +00:00
|
|
|
cpu_boost(true);
|
2006-11-11 05:53:32 +00:00
|
|
|
}
|
2006-11-11 05:33:24 +00:00
|
|
|
boosted_threads++;
|
2006-10-15 11:57:52 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2006-08-21 17:35:35 +00:00
|
|
|
/*---------------------------------------------------------------------------
|
|
|
|
* Remove a thread on the current core from the scheduler.
|
2004-04-30 20:23:04 +00:00
|
|
|
* Parameter is the ID as returned from create_thread().
|
|
|
|
*---------------------------------------------------------------------------
|
|
|
|
*/
|
2006-09-16 16:18:11 +00:00
|
|
|
void remove_thread(struct thread_entry *thread)
|
2006-08-21 17:35:35 +00:00
|
|
|
{
|
2006-09-16 16:18:11 +00:00
|
|
|
if (thread == NULL)
|
|
|
|
thread = cores[CURRENT_CORE].running;
|
|
|
|
|
|
|
|
/* Free the entry by removing thread name. */
|
|
|
|
thread->name = NULL;
|
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
|
|
|
highest_priority = 100;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
if (thread == cores[CURRENT_CORE].running)
|
|
|
|
{
|
|
|
|
remove_from_list(&cores[CURRENT_CORE].running, thread);
|
|
|
|
switch_thread(false, NULL);
|
|
|
|
return ;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (thread == cores[CURRENT_CORE].sleeping)
|
|
|
|
remove_from_list(&cores[CURRENT_CORE].sleeping, thread);
|
2006-09-16 18:30:10 +00:00
|
|
|
else
|
|
|
|
remove_from_list(NULL, thread);
|
2006-08-21 17:35:35 +00:00
|
|
|
}
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
2006-10-15 11:57:52 +00:00
|
|
|
int thread_set_priority(struct thread_entry *thread, int priority)
|
2004-04-30 20:23:04 +00:00
|
|
|
{
|
2006-10-15 11:57:52 +00:00
|
|
|
int old_priority;
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
if (thread == NULL)
|
|
|
|
thread = cores[CURRENT_CORE].running;
|
2007-03-09 08:03:18 +00:00
|
|
|
|
2006-10-15 11:57:52 +00:00
|
|
|
old_priority = thread->priority;
|
2006-09-16 16:18:11 +00:00
|
|
|
thread->priority = priority;
|
|
|
|
highest_priority = 100;
|
2006-10-15 11:57:52 +00:00
|
|
|
|
|
|
|
return old_priority;
|
2002-04-22 12:07:34 +00:00
|
|
|
}
|
2006-11-06 18:07:30 +00:00
|
|
|
|
|
|
|
int thread_get_priority(struct thread_entry *thread)
|
|
|
|
{
|
|
|
|
if (thread == NULL)
|
|
|
|
thread = cores[CURRENT_CORE].running;
|
|
|
|
|
|
|
|
return thread->priority;
|
|
|
|
}
|
2007-03-09 08:03:18 +00:00
|
|
|
|
|
|
|
void priority_yield(void)
|
|
|
|
{
|
|
|
|
struct thread_entry *thread = cores[CURRENT_CORE].running;
|
|
|
|
thread->priority_x = 1;
|
|
|
|
switch_thread(true, NULL);
|
|
|
|
thread->priority_x = 0;
|
|
|
|
}
|
|
|
|
#endif /* HAVE_PRIORITY_SCHEDULING */
|
2002-06-07 14:56:10 +00:00
|
|
|
|
2007-03-06 20:32:13 +00:00
|
|
|
struct thread_entry * thread_get_current(void)
|
|
|
|
{
|
|
|
|
return cores[CURRENT_CORE].running;
|
|
|
|
}
|
|
|
|
|
2002-06-07 14:56:10 +00:00
|
|
|
void init_threads(void)
|
|
|
|
{
|
2006-08-21 17:35:35 +00:00
|
|
|
unsigned int core = CURRENT_CORE;
|
|
|
|
|
2007-03-04 20:06:41 +00:00
|
|
|
if (core == CPU)
|
|
|
|
memset(cores, 0, sizeof cores);
|
2006-09-16 16:18:11 +00:00
|
|
|
cores[core].sleeping = NULL;
|
|
|
|
cores[core].running = NULL;
|
2007-03-26 03:24:36 +00:00
|
|
|
cores[core].waking = NULL;
|
|
|
|
cores[core].wakeup_list = &cores[core].running;
|
2007-03-09 08:03:18 +00:00
|
|
|
#ifdef HAVE_EXTENDED_MESSAGING_AND_NAME
|
|
|
|
cores[core].switch_to_irq_level = STAY_IRQ_LEVEL;
|
|
|
|
#endif
|
2006-09-02 07:56:52 +00:00
|
|
|
cores[core].threads[0].name = main_thread_name;
|
2006-09-16 16:18:11 +00:00
|
|
|
cores[core].threads[0].statearg = 0;
|
|
|
|
#ifdef HAVE_PRIORITY_SCHEDULING
|
|
|
|
cores[core].threads[0].priority = PRIORITY_USER_INTERFACE;
|
2007-03-09 08:03:18 +00:00
|
|
|
cores[core].threads[0].priority_x = 0;
|
2006-09-16 16:18:11 +00:00
|
|
|
highest_priority = 100;
|
2006-10-15 11:57:52 +00:00
|
|
|
#endif
|
|
|
|
#ifdef HAVE_SCHEDULER_BOOSTCTRL
|
2006-11-11 05:33:24 +00:00
|
|
|
boosted_threads = 0;
|
2006-09-16 16:18:11 +00:00
|
|
|
#endif
|
|
|
|
add_to_list(&cores[core].running, &cores[core].threads[0]);
|
|
|
|
|
2006-11-11 05:33:24 +00:00
|
|
|
/* In multiple core setups, each core has a different stack. There is
|
|
|
|
* probably a much better way to do this. */
|
2006-09-02 07:56:52 +00:00
|
|
|
if (core == CPU)
|
2006-08-21 17:35:35 +00:00
|
|
|
{
|
2006-09-02 07:56:52 +00:00
|
|
|
cores[CPU].threads[0].stack = stackbegin;
|
|
|
|
cores[CPU].threads[0].stack_size = (int)stackend - (int)stackbegin;
|
2006-08-21 17:35:35 +00:00
|
|
|
} else {
|
|
|
|
#if NUM_CORES > 1 /* This code path will not be run on single core targets */
|
2006-09-02 07:56:52 +00:00
|
|
|
cores[COP].threads[0].stack = cop_stackbegin;
|
2006-11-11 05:33:24 +00:00
|
|
|
cores[COP].threads[0].stack_size =
|
|
|
|
(int)cop_stackend - (int)cop_stackbegin;
|
2006-08-21 17:35:35 +00:00
|
|
|
#endif
|
|
|
|
}
|
2006-09-02 07:56:52 +00:00
|
|
|
cores[core].threads[0].context.start = 0; /* thread 0 already running */
|
2007-03-04 20:06:41 +00:00
|
|
|
#if NUM_CORES > 1
|
|
|
|
if(core == COP)
|
|
|
|
kernel_running_on_cop = true; /* can we use context.start for this? */
|
|
|
|
#endif
|
2002-07-15 22:21:18 +00:00
|
|
|
}
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
int thread_stack_usage(const struct thread_entry *thread)
|
2002-07-15 22:21:18 +00:00
|
|
|
{
|
2005-06-10 23:05:15 +00:00
|
|
|
unsigned int i;
|
2006-09-16 16:18:11 +00:00
|
|
|
unsigned int *stackptr = thread->stack;
|
2005-06-10 23:05:15 +00:00
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
for (i = 0;i < thread->stack_size/sizeof(int);i++)
|
2005-06-10 23:05:15 +00:00
|
|
|
{
|
2006-09-02 07:56:52 +00:00
|
|
|
if (stackptr[i] != DEADBEEF)
|
2005-06-10 23:05:15 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2006-09-16 16:18:11 +00:00
|
|
|
return ((thread->stack_size - i * sizeof(int)) * 100) /
|
|
|
|
thread->stack_size;
|
|
|
|
}
|
|
|
|
|
|
|
|
int thread_get_status(const struct thread_entry *thread)
|
|
|
|
{
|
|
|
|
return GET_STATE(thread->statearg);
|
2006-08-21 17:35:35 +00:00
|
|
|
}
|