28dec004c0
* Add wildcards to various sections placements a la *(".text") => "*(.text*)" * Remove hacky bits from those linker scripts (no problem encountered testing) * Change section for asm functions from .<section> to .<section>.<function> so that -ffunction-sections works for those asm file too. git-svn-id: svn://svn.rockbox.org/rockbox/trunk@31337 a1c6a512-1295-4272-9138-f99709370657
424 lines
15 KiB
ArmAsm
424 lines
15 KiB
ArmAsm
/***************************************************************************
|
|
* __________ __ ___.
|
|
* Open \______ \ ____ ____ | | _\_ |__ _______ ___
|
|
* Source | _// _ \_/ ___\| |/ /| __ \ / _ \ \/ /
|
|
* Jukebox | | ( <_> ) \___| < | \_\ ( <_> > < <
|
|
* Firmware |____|_ /\____/ \___ >__|_ \|___ /\____/__/\_ \
|
|
* \/ \/ \/ \/ \/
|
|
* $Id$
|
|
*
|
|
* Copyright (C) 2006,2007 by Greg White
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version 2
|
|
* of the License, or (at your option) any later version.
|
|
*
|
|
* This software is distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY
|
|
* KIND, either express or implied.
|
|
*
|
|
****************************************************************************/
|
|
#define ASM
|
|
#include "config.h"
|
|
#include "cpu.h"
|
|
|
|
/* Used by ARMv4 & ARMv5 CPUs with cp15 register and MMU */
|
|
|
|
#if CONFIG_CPU == TCC7801 || CONFIG_CPU == AT91SAM9260
|
|
/* MMU present but unused */
|
|
#define HAVE_TEST_AND_CLEAN_CACHE
|
|
|
|
#elif CONFIG_CPU == DM320 || CONFIG_CPU == AS3525v2 || CONFIG_CPU == S5L8702
|
|
#define USE_MMU
|
|
#define HAVE_TEST_AND_CLEAN_CACHE
|
|
|
|
#elif CONFIG_CPU == AS3525
|
|
#define USE_MMU
|
|
#define CACHE_SIZE 8
|
|
|
|
#elif CONFIG_CPU == S3C2440
|
|
#define USE_MMU
|
|
#define CACHE_SIZE 16
|
|
|
|
#elif CONFIG_CPU == S5L8700 || CONFIG_CPU == S5L8701
|
|
/* MMU not present */
|
|
#define CACHE_SIZE 4
|
|
|
|
#elif CONFIG_CPU == IMX233
|
|
#define USE_MMU
|
|
#define CACHE_SIZE 16
|
|
|
|
#else
|
|
#error Cache settings unknown for this CPU !
|
|
|
|
#endif /* CPU specific configuration */
|
|
|
|
@ Index format: 31:26 = index, N:5 = segment, remainder = SBZ
|
|
@ assume 64-way set associative separate I/D caches
|
|
@ CACHE_SIZE = N (kB) = N*2^10 B
|
|
@ number of lines = N*2^(10-CACHEALIGN_BITS)
|
|
@ Index bits = 6
|
|
@ Segment loops = N*2^(10-CACHEALIGN_BITS-6) = N*2^(4-CACHEALIGN_BITS)
|
|
@ Segment loops = N/2^(CACHEALIGN_BITS - 4)
|
|
@ Segment loops = N/(1<<(CACHEALIGN_BITS - 4))
|
|
|
|
#ifdef CACHE_SIZE
|
|
|
|
#if CACHEALIGN_BITS == 4
|
|
#define INDEX_STEPS CACHE_SIZE
|
|
#elif CACHEALIGN_BITS == 5
|
|
#define INDEX_STEPS (CACHE_SIZE/2)
|
|
#endif /* CACHEALIGN_BITS */
|
|
|
|
@ assume 64-way set associative separate I/D caches (log2(64) == 6)
|
|
@ Index format: 31:26 = index, M:N = segment, remainder = SBZ
|
|
@ Segment bits = log2(cache size in bytes / cache line size in byte) - Index bits (== 6)
|
|
@ N = CACHEALIGN_BITS
|
|
|
|
#endif /* CACHE_SIZE */
|
|
|
|
|
|
#ifdef USE_MMU
|
|
|
|
/** MMU setup **/
|
|
|
|
/*
|
|
* void ttb_init(void);
|
|
*/
|
|
.section .text.ttb_init, "ax", %progbits
|
|
.align 2
|
|
.global ttb_init
|
|
.type ttb_init, %function
|
|
ttb_init:
|
|
ldr r0, =TTB_BASE_ADDR @
|
|
mvn r1, #0 @
|
|
mcr p15, 0, r0, c2, c0, 0 @ Set the TTB base address
|
|
mcr p15, 0, r1, c3, c0, 0 @ Set all domains to manager status
|
|
bx lr @
|
|
.size ttb_init, .-ttb_init
|
|
|
|
/*
|
|
* void map_section(unsigned int pa, unsigned int va, int mb, int flags);
|
|
*/
|
|
.section .text.map_section, "ax", %progbits
|
|
.align 2
|
|
.global map_section
|
|
.type map_section, %function
|
|
map_section:
|
|
@ align to 1MB
|
|
@ pa &= (-1 << 20);
|
|
mov r0, r0, lsr #20
|
|
mov r0, r0, lsl #20
|
|
|
|
@ pa |= (flags | 0x412);
|
|
@ bit breakdown:
|
|
@ 10: superuser - r/w, user - no access
|
|
@ 4: should be "1"
|
|
@ 3,2: Cache flags (flags (r3))
|
|
@ 1: Section signature
|
|
orr r0, r0, r3
|
|
orr r0, r0, #0x410
|
|
orr r0, r0, #0x2
|
|
|
|
@ unsigned int* ttbPtr = TTB_BASE + (va >> 20);
|
|
@ sections are 1MB size
|
|
mov r1, r1, lsr #20
|
|
ldr r3, =TTB_BASE_ADDR
|
|
add r1, r3, r1, lsl #0x2
|
|
|
|
@ Add MB to pa, flags are already present in pa, but addition
|
|
@ should not effect them
|
|
@
|
|
@ for( ; mb>0; mb--, pa += (1 << 20))
|
|
@ {
|
|
@ *(ttbPtr++) = pa;
|
|
@ }
|
|
cmp r2, #0
|
|
bxle lr
|
|
mov r3, #0x0
|
|
1: @ loop
|
|
str r0, [r1], #4
|
|
add r0, r0, #0x100000
|
|
add r3, r3, #0x1
|
|
cmp r2, r3
|
|
bne 1b @ loop
|
|
bx lr
|
|
.size map_section, .-map_section
|
|
|
|
/*
|
|
* void enable_mmu(void);
|
|
*/
|
|
.section .text.enable_mmu, "ax", %progbits
|
|
.align 2
|
|
.global enable_mmu
|
|
.type enable_mmu, %function
|
|
enable_mmu:
|
|
mov r0, #0 @
|
|
mcr p15, 0, r0, c8, c7, 0 @ invalidate TLB
|
|
mcr p15, 0, r0, c7, c7,0 @ invalidate both i and dcache
|
|
mrc p15, 0, r0, c1, c0, 0 @
|
|
orr r0, r0, #1 @ enable mmu bit, i and dcache
|
|
orr r0, r0, #1<<2 @ enable dcache
|
|
orr r0, r0, #1<<12 @ enable icache
|
|
mcr p15, 0, r0, c1, c0, 0 @
|
|
nop @
|
|
nop @
|
|
nop @
|
|
nop @
|
|
bx lr @
|
|
.size enable_mmu, .-enable_mmu
|
|
.ltorg
|
|
|
|
#endif /* USE_MMU */
|
|
|
|
|
|
|
|
/** Cache coherency **/
|
|
|
|
/*
|
|
* Write DCache back to RAM for the given range and remove cache lines
|
|
* from DCache afterwards
|
|
* void commit_discard_dcache_range(const void *base, unsigned int size);
|
|
*/
|
|
.section .text.commit_discard_dcache_range, "ax", %progbits
|
|
.align 2
|
|
.global commit_discard_dcache_range
|
|
.type commit_discard_dcache_range, %function
|
|
.global invalidate_dcache_range @ Alias, deprecated
|
|
|
|
@ MVA format: 31:5 = Modified virtual address, 4:0 = SBZ
|
|
commit_discard_dcache_range:
|
|
invalidate_dcache_range:
|
|
add r1, r0, r1 @ size -> end
|
|
cmp r1, r0 @ end <= start?
|
|
bxls lr @
|
|
bic r0, r0, #31 @ Align start to cache line (down)
|
|
1: @ inv_start @
|
|
mcr p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
add r0, r0, #32 @
|
|
cmp r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
bhi 1b @ inv_start @
|
|
mov r0, #0 @
|
|
mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
|
|
bx lr @
|
|
.size commit_discard_dcache_range, .-commit_discard_dcache_range
|
|
|
|
/*
|
|
* Write DCache back to RAM for the given range
|
|
* void commit_dcache_range(const void *base, unsigned int size);
|
|
*/
|
|
.section .text.commit_dcache_range, "ax", %progbits
|
|
.align 2
|
|
.global commit_dcache_range
|
|
.type commit_dcache_range, %function
|
|
.global clean_dcache_range @ Alias, deprecated
|
|
|
|
@ MVA format: 31:5 = Modified virtual address, 4:0 = SBZ
|
|
commit_dcache_range:
|
|
clean_dcache_range:
|
|
add r1, r0, r1 @ size -> end
|
|
cmp r1, r0 @ end <= start?
|
|
bxls lr @
|
|
bic r0, r0, #31 @ Align start to cache line (down)
|
|
1: @ clean_start @
|
|
mcr p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
add r0, r0, #32 @
|
|
cmp r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c10, 1 @ Clean line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
bhi 1b @clean_start @
|
|
mov r0, #0 @
|
|
mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
|
|
bx lr @
|
|
.size commit_dcache_range, .-commit_dcache_range
|
|
|
|
/*
|
|
* Remove cache lines for the given range from DCache
|
|
* will *NOT* do write back except for buffer edges not on a line boundary
|
|
* void discard_dcache_range(const void *base, unsigned int size);
|
|
*/
|
|
.section .text.discard_dcache_range, "ax", %progbits
|
|
.align 2
|
|
.global discard_dcache_range
|
|
.type discard_dcache_range, %function
|
|
.global dump_dcache_range @ Alias, deprecated
|
|
|
|
@ MVA format: 31:5 = Modified virtual address, 4:0 = SBZ
|
|
discard_dcache_range:
|
|
dump_dcache_range:
|
|
add r1, r0, r1 @ size -> end
|
|
cmp r1, r0 @ end <= start?
|
|
bxls lr @
|
|
tst r0, #31 @ Check first line for bits set
|
|
bicne r0, r0, #31 @ Clear low five bits (down)
|
|
mcrne p15, 0, r0, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
@ if not cache aligned
|
|
addne r0, r0, #32 @ Move to the next cache line
|
|
@
|
|
tst r1, #31 @ Check last line for bits set
|
|
bicne r1, r1, #31 @ Clear low five bits (down)
|
|
mcrne p15, 0, r1, c7, c14, 1 @ Clean and invalidate line by MVA
|
|
@ if not cache aligned
|
|
cmp r1, r0 @ end <= start now?
|
|
1: @ discard_start @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
mcrhi p15, 0, r0, c7, c6, 1 @ Invalidate line by MVA
|
|
addhi r0, r0, #32 @
|
|
cmphi r1, r0 @
|
|
bhi 1b @ discard_start @
|
|
mov r0, #0 @
|
|
mcr p15, 0, r0, c7, c10, 4 @ Drain write buffer
|
|
bx lr @
|
|
.size discard_dcache_range, .-discard_dcache_range
|
|
|
|
/*
|
|
* Write entire DCache back to RAM
|
|
* void commit_dcache(void);
|
|
*/
|
|
.section .text.commit_dcache, "ax", %progbits
|
|
.align 2
|
|
.global commit_dcache
|
|
.type commit_dcache, %function
|
|
.global cpucache_commit @ Alias
|
|
.global clean_dcache @ Alias, deprecated
|
|
.global cpucache_flush @ Alias, deprecated
|
|
|
|
commit_dcache:
|
|
cpucache_commit:
|
|
clean_dcache:
|
|
cpucache_flush:
|
|
#ifdef HAVE_TEST_AND_CLEAN_CACHE
|
|
mrc p15, 0, r15, c7, c10, 3 @ test and clean dcache
|
|
bne commit_dcache
|
|
mov r1, #0
|
|
#else
|
|
mov r1, #0x00000000 @
|
|
1: @ commit_start @
|
|
mcr p15, 0, r1, c7, c10, 2 @ Clean entry by index
|
|
add r0, r1, #(1<<CACHEALIGN_BITS)
|
|
mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
|
|
.rept INDEX_STEPS - 2 /* 2 steps already executed */
|
|
add r0, r0, #(1<<CACHEALIGN_BITS)
|
|
mcr p15, 0, r0, c7, c10, 2 @ Clean entry by index
|
|
.endr
|
|
adds r1, r1, #0x04000000 @ will wrap to zero at loop end
|
|
bne 1b @ commit_start @
|
|
#endif /* HAVE_TEST_AND_CLEAN_CACHE */
|
|
mcr p15, 0, r1, c7, c10, 4 @ Drain write buffer
|
|
bx lr @
|
|
.size commit_dcache, .-commit_dcache
|
|
|
|
/*
|
|
* Commit and discard entire DCache, will do writeback
|
|
* void commit_discard_dcache(void);
|
|
*/
|
|
.section .icode.commit_discard_dcache, "ax", %progbits
|
|
.align 2
|
|
.global commit_discard_dcache
|
|
.type commit_discard_dcache, %function
|
|
.global invalidate_dcache @ Alias, deprecated
|
|
|
|
commit_discard_dcache:
|
|
invalidate_dcache:
|
|
#ifdef HAVE_TEST_AND_CLEAN_CACHE
|
|
mrc p15, 0, r15, c7, c14, 3 @ test, clean and invalidate dcache
|
|
bne commit_discard_dcache
|
|
mov r1, #0
|
|
#else
|
|
mov r1, #0x00000000 @
|
|
1: @ inv_start @
|
|
mcr p15, 0, r1, c7, c14, 2 @ Clean and invalidate entry by index
|
|
add r0, r1, #(1<<CACHEALIGN_BITS)
|
|
mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
|
|
.rept INDEX_STEPS - 2 /* 2 steps already executed */
|
|
add r0, r0, #(1<<CACHEALIGN_BITS)
|
|
mcr p15, 0, r0, c7, c14, 2 @ Clean and invalidate entry by index
|
|
.endr
|
|
adds r1, r1, #0x04000000 @ will wrap to zero at loop end
|
|
bne 1b @ inv_start @
|
|
#endif /* HAVE_TEST_AND_CLEAN_CACHE */
|
|
mcr p15, 0, r1, c7, c10, 4 @ Drain write buffer
|
|
bx lr @
|
|
.size commit_discard_dcache, .-commit_discard_dcache
|
|
|
|
/*
|
|
* Discards the entire ICache, and commit+discards the entire DCache
|
|
* void commit_discard_idcache(void);
|
|
*/
|
|
.section .icode.commit_discard_idcache, "ax", %progbits
|
|
.align 2
|
|
.global commit_discard_idcache
|
|
.type commit_discard_idcache, %function
|
|
.global cpucache_commit_discard @ Alias
|
|
.global invalidate_idcache @ Alias, deprecated
|
|
.global cpucache_invalidate @ Alias, deprecated
|
|
|
|
commit_discard_idcache:
|
|
cpucache_commit_discard:
|
|
invalidate_idcache:
|
|
cpucache_invalidate:
|
|
mov r2, lr @ save lr to r1, call uses r0 only
|
|
bl commit_discard_dcache @ commit and discard entire DCache
|
|
mcr p15, 0, r1, c7, c5, 0 @ Invalidate ICache (r1=0 from call)
|
|
bx r2
|
|
.size commit_discard_idcache, .-commit_discard_idcache
|