diff --git a/arch/arm/cpu/cache-armv7.S b/arch/arm/cpu/cache-armv7.S index 13542d9..5bdf7e4 100644 --- a/arch/arm/cpu/cache-armv7.S +++ b/arch/arm/cpu/cache-armv7.S @@ -34,7 +34,10 @@ .section .text.v7_mmu_cache_off ENTRY(v7_mmu_cache_off) - stmfd sp!, {r0-r7, r9-r11} + /* although 'r12' is an eabi scratch register which does + not need to be restored, save it to ensure an 8-byte + stack alignment */ + stmfd sp!, {r4-r12, lr} mrc p15, 0, r0, c1, c0 #ifdef CONFIG_MMU bic r0, r0, #0x000d @@ -42,7 +45,6 @@ bic r0, r0, #0x000c #endif mcr p15, 0, r0, c1, c0 @ turn MMU and cache off - mov r12, lr bl v7_mmu_cache_flush mov r0, #0 #ifdef CONFIG_MMU @@ -51,35 +53,33 @@ mcr p15, 0, r0, c7, c5, 6 @ invalidate BTC mcr p15, 0, r0, c7, c10, 4 @ DSB mcr p15, 0, r0, c7, c5, 4 @ ISB - ldmfd sp!, {r0-r7, r9-r11} - mov pc, r12 + ldmfd sp!, {r4-r12, pc} ENDPROC(v7_mmu_cache_off) .section .text.v7_mmu_cache_flush ENTRY(v7_mmu_cache_flush) - stmfd sp!, {r10, lr} - mrc p15, 0, r10, c0, c1, 5 @ read ID_MMFR1 - tst r10, #0xf << 16 @ hierarchical cache (ARMv7) - mov r10, #0 + mrc p15, 0, r12, c0, c1, 5 @ read ID_MMFR1 + tst r12, #0xf << 16 @ hierarchical cache (ARMv7) + mov r12, #0 beq hierarchical - mcr p15, 0, r10, c7, c14, 0 @ clean+invalidate D + mcr p15, 0, r12, c7, c14, 0 @ clean+invalidate D b iflush hierarchical: - mcr p15, 0, r10, c7, c10, 5 @ DMB stmfd sp!, {r0-r7, r9-r11} + mcr p15, 0, r12, c7, c10, 5 @ DMB mrc p15, 1, r0, c0, c0, 1 @ read clidr ands r3, r0, #0x7000000 @ extract loc from clidr mov r3, r3, lsr #23 @ left align loc bit field beq finished @ if loc is 0, then no need to clean - mov r10, #0 @ start clean at cache level 0 + mov r12, #0 @ start clean at cache level 0 loop1: - add r2, r10, r10, lsr #1 @ work out 3x current cache level + add r2, r12, r12, lsr #1 @ work out 3x current cache level mov r1, r0, lsr r2 @ extract cache type bits from clidr and r1, r1, #7 @ mask of the bits for current cache only cmp r1, #2 @ see what cache we have at this level blt skip @ skip if no cache, or just i-cache - mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr - mcr p15, 0, r10, c7, c5, 4 @ isb to sych the new cssr&csidr + mcr p15, 2, r12, c0, c0, 0 @ select current cache level in cssr + mcr p15, 0, r12, c7, c5, 4 @ isb to sych the new cssr&csidr mrc p15, 1, r1, c0, c0, 0 @ read the new csidr and r2, r1, #7 @ extract the length of the cache lines add r2, r2, #4 @ add 4 (line length offset) @@ -91,10 +91,10 @@ loop2: mov r9, r4 @ create working copy of max way size loop3: -ARM( orr r11, r10, r9, lsl r5 ) @ factor way and cache number into r11 +ARM( orr r11, r12, r9, lsl r5 ) @ factor way and cache number into r11 ARM( orr r11, r11, r7, lsl r2 ) @ factor index number into r11 THUMB( lsl r6, r9, r5 ) -THUMB( orr r11, r10, r6 ) @ factor way and cache number into r11 +THUMB( orr r11, r12, r6 ) @ factor way and cache number into r11 THUMB( lsl r6, r7, r2 ) THUMB( orr r11, r11, r6 ) @ factor index number into r11 mcr p15, 0, r11, c7, c14, 2 @ clean & invalidate by set/way @@ -103,19 +103,19 @@ subs r7, r7, #1 @ decrement the index bge loop2 skip: - add r10, r10, #2 @ increment cache number - cmp r3, r10 + add r12, r12, #2 @ increment cache number + cmp r3, r12 bgt loop1 finished: ldmfd sp!, {r0-r7, r9-r11} - mov r10, #0 @ switch back to cache level 0 - mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr + mov r12, #0 @ switch back to cache level 0 + mcr p15, 2, r12, c0, c0, 0 @ select current cache level in cssr iflush: - mcr p15, 0, r10, c7, c10, 4 @ DSB - mcr p15, 0, r10, c7, c5, 0 @ invalidate I+BTB - mcr p15, 0, r10, c7, c10, 4 @ DSB - mcr p15, 0, r10, c7, c5, 4 @ ISB - ldmfd sp!, {r10, pc} + mcr p15, 0, r12, c7, c10, 4 @ DSB + mcr p15, 0, r12, c7, c5, 0 @ invalidate I+BTB + mcr p15, 0, r12, c7, c10, 4 @ DSB + mcr p15, 0, r12, c7, c5, 4 @ ISB + mov pc, lr ENDPROC(v7_mmu_cache_flush) /*