/** * @file * @brief Provide Architecture level Initialization * * FileName: arch/arm/mach-omap/omap3_core.S * * This provides OMAP3 Architecture initialization. Among these, * @li OMAP ROM Code is located in SRAM, we can piggy back on * the same addresses * @li If clock initialization is required, call the same. * @li Setup a temporary SRAM stack which is necessary to call C * functions. * @li Call architecture initialization function a_init * * arch_init_lowlevel is enabled if CONFIG_ARCH_HAS_LOWLEVEL_INIT is defined * cpy_clk_code is called if CONFIG_OMAP3_COPY_CLOCK_SRAM is defined */ /* * (C) Copyright 2006-2008 * Texas Instruments, * Nishanth Menon * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License as * published by the Free Software Foundation; either version 2 of * the License, or (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, * MA 02111-1307 USA */ #include #include #include #include #ifdef CONFIG_ARCH_HAS_LOWLEVEL_INIT .globl arch_init_lowlevel arch_init_lowlevel: /* Copy vectors to mask ROM indirect addr */ mov r0, pc /* Store the current pc address */ sub r0, r0, #8 /* Reduce offset */ ldr r1, arch_start /* Load the link address for arch_int */ ldr r2, barebox_start /* load the link address of start_init*/ sub r1, r1, r2 /* get the offset */ /* subtract the offset from PC of arch=Current start */ sub r0, r0, r1 mov r2, #OMAP_SRAM_INTVECT_COPYSIZE /* r2 <- size to copy */ add r2, r0, r2 /* r2 <- source end address */ ldr r1, SRAM_INTVECT /* build vect addr */ next: ldmia r0!, {r3-r10} /* copy from source address [r0] */ stmia r1!, {r3-r10} /* copy to target address [r1] */ cmp r0, r2 /* until source end address [r2] */ ble next /* loop until equal */ #ifdef CONFIG_OMAP3_COPY_CLOCK_SRAM /* No need to copy/exec the clock code - DPLL adjust already done * in Perip/NAND/oneNAND Boot. * Put dpll adjust code behind vectors. r1 has address to copy to */ bl cpy_clk_code #endif ldr r1, SRAM_INTVECT /* build vect addr */ /* Read the interrupt vector base address */ mrc p15, #0, r0, c12, c0, #0 /* Clear the vector base 4:0 is reserved. */ and r0, r0, #0xF /* Store the SRAM_INTVECT address */ orr r0, r0, r1 /* Store the new vector address */ mcr p15, #0, r0, c12, c0, #0 /* Setup a temporary stack so that we can call C functions */ ldr sp, SRAM_STACK str ip, [sp] /* stash old link register */ str lr, [sp] /* stash current link register */ mov ip, lr /* save link reg across call */ bl a_init /* Architecture init */ ldr lr, [sp] /* restore current link register */ ldr ip, [sp] /* restore save ip */ /* Invalidate all Dcaches */ #ifndef CONFIG_CPU_V7_DCACHE_SKIP /* If Arch specific ROM code SMI handling does not exist */ mrc p15, 1, r0, c0, c0, 1 /* read clidr */ ands r3, r0, #0x7000000 /* extract loc from clidr */ mov r3, r3, lsr #23 /* left align loc bit field */ beq finished_inval /* if loc is 0, then no need to clean */ mov r10, #0 /* start clean at cache level 0 */ inval_loop1: add r2, r10, r10, lsr #1 /* work out 3x current cache level */ mov r1, r0, lsr r2 /* extract cache type bits from clidr */ and r1, r1, # 7 /* mask of the bits for current cache only */ cmp r1, #2 /* see what cache we have at this level */ blt skip_inval /* skip if no cache, or just i-cache */ mcr p15, 2, r10, c0, c0, 0 /* select current cache level in cssr */ isb /* isb to sych the new cssr&csidr */ mrc p15, 1, r1, c0, c0, 0 /* read the new csidr */ and r2, r1, #7 /* extract the length of the cache lines */ add r2, r2, #4 /* add 4 (line length offset) */ ldr r4, =0x3ff ands r4, r4, r1, lsr #3 /* find maximum number on the way size*/ clz r5, r4 /* find bit position of way size increment */ ldr r7, =0x7fff ands r7, r7, r1, lsr #13 /* extract max number of the index size */ inval_loop2: mov r9, r4 /* create working copy of max way size */ inval_loop3: orr r11, r10, r9, lsl r5 /* factor way and cache number into r11*/ orr r11, r11, r7, lsl r2 /* factor index number into r11 */ mcr p15, 0, r11, c7, c6, 2 /* invalidate by set/way */ subs r9, r9, #1 /* decrement the way */ bge inval_loop3 subs r7, r7, #1 /* decrement the index */ bge inval_loop2 skip_inval: add r10, r10, #2 /* increment cache number */ cmp r3, r10 bgt inval_loop1 finished_inval: mov r10, #0 /* swith back to cache level 0 */ mcr p15, 2, r10, c0, c0, 0 /* select current cache level in cssr */ isb #endif /* CONFIG_CPU_V7_DCACHE_SKIP */ /* back to arch calling code */ mov pc, lr arch_start: .word arch_init_lowlevel barebox_start: .word exception_vectors SRAM_INTVECT: .word OMAP_SRAM_INTVECT SRAM_STACK: .word OMAP_SRAM_STACK #endif /* CONFIG_ARCH_HAS_LOWLEVEL_INIT */