blob: ee4c133547e029e43b790e2de9b56164a529a609 [file] [log] [blame]
/**
* @file
* @brief Provide Architecture level Initialization
*
* FileName: arch/arm/mach-omap/omap3_core.S
*
* This provides OMAP3 Architecture initialization. Among these,
* @li OMAP ROM Code is located in SRAM, we can piggy back on
* the same addresses
* @li If clock initialization is required, call the same.
* @li Setup a temporary SRAM stack which is necessary to call C
* functions.
* @li Call architecture initialization function a_init
*
* arch_init_lowlevel is enabled if CONFIG_ARCH_HAS_LOWLEVEL_INIT is defined
* cpy_clk_code is called if CONFIG_OMAP3_COPY_CLOCK_SRAM is defined
*/
/*
* (C) Copyright 2006-2008
* Texas Instruments, <www.ti.com>
* Nishanth Menon <x0nishan@ti.com>
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License as
* published by the Free Software Foundation; either version 2 of
* the License, or (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston,
* MA 02111-1307 USA
*/
#include <config.h>
#include <mach/silicon.h>
#include <mach/wdt.h>
#include <mach/clocks.h>
#ifdef CONFIG_ARCH_HAS_LOWLEVEL_INIT
.globl arch_init_lowlevel
arch_init_lowlevel:
/* Copy vectors to mask ROM indirect addr */
mov r0, pc /* Store the current pc address */
sub r0, r0, #8 /* Reduce offset */
ldr r1, arch_start /* Load the link address for arch_int */
ldr r2, barebox_start /* load the link address of start_init*/
sub r1, r1, r2 /* get the offset */
/* subtract the offset from PC of arch=Current start */
sub r0, r0, r1
mov r2, #OMAP_SRAM_INTVECT_COPYSIZE /* r2 <- size to copy */
add r2, r0, r2 /* r2 <- source end address */
ldr r1, SRAM_INTVECT /* build vect addr */
next:
ldmia r0!, {r3-r10} /* copy from source address [r0] */
stmia r1!, {r3-r10} /* copy to target address [r1] */
cmp r0, r2 /* until source end address [r2] */
ble next /* loop until equal */
#ifdef CONFIG_OMAP3_COPY_CLOCK_SRAM
/* No need to copy/exec the clock code - DPLL adjust already done
* in Perip/NAND/oneNAND Boot.
* Put dpll adjust code behind vectors. r1 has address to copy to
*/
bl cpy_clk_code
#endif
ldr r1, SRAM_INTVECT /* build vect addr */
/* Read the interrupt vector base address */
mrc p15, #0, r0, c12, c0, #0
/* Clear the vector base 4:0 is reserved. */
and r0, r0, #0xF
/* Store the SRAM_INTVECT address */
orr r0, r0, r1
/* Store the new vector address */
mcr p15, #0, r0, c12, c0, #0
/* Setup a temporary stack so that we can call C functions */
ldr sp, SRAM_STACK
str ip, [sp] /* stash old link register */
str lr, [sp] /* stash current link register */
mov ip, lr /* save link reg across call */
bl a_init /* Architecture init */
ldr lr, [sp] /* restore current link register */
ldr ip, [sp] /* restore save ip */
/* Invalidate all Dcaches */
#ifndef CONFIG_CPU_V7_DCACHE_SKIP
/* If Arch specific ROM code SMI handling does not exist */
mrc p15, 1, r0, c0, c0, 1 /* read clidr */
ands r3, r0, #0x7000000 /* extract loc from clidr */
mov r3, r3, lsr #23 /* left align loc bit field */
beq finished_inval /* if loc is 0, then no need to clean */
mov r10, #0 /* start clean at cache level 0 */
inval_loop1:
add r2, r10, r10, lsr #1 /* work out 3x current cache level */
mov r1, r0, lsr r2 /* extract cache type bits from clidr */
and r1, r1, # 7 /* mask of the bits for current cache only */
cmp r1, #2 /* see what cache we have at this level */
blt skip_inval /* skip if no cache, or just i-cache */
mcr p15, 2, r10, c0, c0, 0 /* select current cache level in cssr */
isb /* isb to sych the new cssr&csidr */
mrc p15, 1, r1, c0, c0, 0 /* read the new csidr */
and r2, r1, #7 /* extract the length of the cache lines */
add r2, r2, #4 /* add 4 (line length offset) */
ldr r4, =0x3ff
ands r4, r4, r1, lsr #3 /* find maximum number on the way size*/
clz r5, r4 /* find bit position of way size increment */
ldr r7, =0x7fff
ands r7, r7, r1, lsr #13 /* extract max number of the index size */
inval_loop2:
mov r9, r4 /* create working copy of max way size */
inval_loop3:
orr r11, r10, r9, lsl r5 /* factor way and cache number into r11*/
orr r11, r11, r7, lsl r2 /* factor index number into r11 */
mcr p15, 0, r11, c7, c6, 2 /* invalidate by set/way */
subs r9, r9, #1 /* decrement the way */
bge inval_loop3
subs r7, r7, #1 /* decrement the index */
bge inval_loop2
skip_inval:
add r10, r10, #2 /* increment cache number */
cmp r3, r10
bgt inval_loop1
finished_inval:
mov r10, #0 /* swith back to cache level 0 */
mcr p15, 2, r10, c0, c0, 0 /* select current cache level in cssr */
isb
#endif /* CONFIG_CPU_V7_DCACHE_SKIP */
/* back to arch calling code */
mov pc, lr
arch_start:
.word arch_init_lowlevel
barebox_start:
.word exception_vectors
SRAM_INTVECT:
.word OMAP_SRAM_INTVECT
SRAM_STACK:
.word OMAP_SRAM_STACK
#endif /* CONFIG_ARCH_HAS_LOWLEVEL_INIT */