2ef9481e66
This patch removes all self references and fixes references to files in the now defunct arch/ppc64 tree. I think this accomplises everything wanted, though there might be a few references I missed. Signed-off-by: Jon Mason <jdmason@us.ibm.com> Signed-off-by: Paul Mackerras <paulus@samba.org>
294 lines
6.3 KiB
ArmAsm
294 lines
6.3 KiB
ArmAsm
/*
|
|
* Useful bootup functions, which are more easily done in asm than C.
|
|
*
|
|
* NOTE: Be very very careful about the registers you use here.
|
|
* We don't follow any ABI calling convention among the
|
|
* assembler functions that call each other, especially early
|
|
* in the initialization. Please preserve at least r3 and r4
|
|
* for these early functions, as they often contain information
|
|
* passed from boot roms into the C decompress function.
|
|
*
|
|
* Author: Tom Rini
|
|
* trini@mvista.com
|
|
* Derived from arch/ppc/boot/prep/head.S (Cort Dougan, many others).
|
|
*
|
|
* 2001-2004 (c) MontaVista, Software, Inc. This file is licensed under
|
|
* the terms of the GNU General Public License version 2. This program
|
|
* is licensed "as is" without any warranty of any kind, whether express
|
|
* or implied.
|
|
*/
|
|
|
|
#include <asm/processor.h>
|
|
#include <asm/cache.h>
|
|
#include <asm/ppc_asm.h>
|
|
|
|
|
|
.text
|
|
|
|
#ifdef CONFIG_6xx
|
|
.globl disable_6xx_mmu
|
|
disable_6xx_mmu:
|
|
/* Establish default MSR value, exception prefix 0xFFF.
|
|
* If necessary, this function must fix up the LR if we
|
|
* return to a different address space once the MMU is
|
|
* disabled.
|
|
*/
|
|
li r8,MSR_IP|MSR_FP
|
|
mtmsr r8
|
|
isync
|
|
|
|
/* Test for a 601 */
|
|
mfpvr r10
|
|
srwi r10,r10,16
|
|
cmpwi 0,r10,1 /* 601 ? */
|
|
beq .clearbats_601
|
|
|
|
/* Clear BATs */
|
|
li r8,0
|
|
mtspr SPRN_DBAT0U,r8
|
|
mtspr SPRN_DBAT0L,r8
|
|
mtspr SPRN_DBAT1U,r8
|
|
mtspr SPRN_DBAT1L,r8
|
|
mtspr SPRN_DBAT2U,r8
|
|
mtspr SPRN_DBAT2L,r8
|
|
mtspr SPRN_DBAT3U,r8
|
|
mtspr SPRN_DBAT3L,r8
|
|
.clearbats_601:
|
|
mtspr SPRN_IBAT0U,r8
|
|
mtspr SPRN_IBAT0L,r8
|
|
mtspr SPRN_IBAT1U,r8
|
|
mtspr SPRN_IBAT1L,r8
|
|
mtspr SPRN_IBAT2U,r8
|
|
mtspr SPRN_IBAT2L,r8
|
|
mtspr SPRN_IBAT3U,r8
|
|
mtspr SPRN_IBAT3L,r8
|
|
isync
|
|
sync
|
|
sync
|
|
|
|
/* Set segment registers */
|
|
li r8,16 /* load up segment register values */
|
|
mtctr r8 /* for context 0 */
|
|
lis r8,0x2000 /* Ku = 1, VSID = 0 */
|
|
li r10,0
|
|
3: mtsrin r8,r10
|
|
addi r8,r8,0x111 /* increment VSID */
|
|
addis r10,r10,0x1000 /* address of next segment */
|
|
bdnz 3b
|
|
blr
|
|
|
|
.globl disable_6xx_l1cache
|
|
disable_6xx_l1cache:
|
|
/* Enable, invalidate and then disable the L1 icache/dcache. */
|
|
li r8,0
|
|
ori r8,r8,(HID0_ICE|HID0_DCE|HID0_ICFI|HID0_DCI)
|
|
mfspr r11,SPRN_HID0
|
|
or r11,r11,r8
|
|
andc r10,r11,r8
|
|
isync
|
|
mtspr SPRN_HID0,r8
|
|
sync
|
|
isync
|
|
mtspr SPRN_HID0,r10
|
|
sync
|
|
isync
|
|
blr
|
|
#endif
|
|
|
|
.globl _setup_L2CR
|
|
_setup_L2CR:
|
|
/*
|
|
* We should be skipping this section on CPUs where this results in an
|
|
* illegal instruction. If not, please send trini@kernel.crashing.org
|
|
* the PVR of your CPU.
|
|
*/
|
|
/* Invalidate/disable L2 cache */
|
|
sync
|
|
isync
|
|
mfspr r8,SPRN_L2CR
|
|
rlwinm r8,r8,0,1,31
|
|
oris r8,r8,L2CR_L2I@h
|
|
sync
|
|
isync
|
|
mtspr SPRN_L2CR,r8
|
|
sync
|
|
isync
|
|
|
|
/* Wait for the invalidation to complete */
|
|
mfspr r8,SPRN_PVR
|
|
srwi r8,r8,16
|
|
cmplwi cr0,r8,0x8000 /* 7450 */
|
|
cmplwi cr1,r8,0x8001 /* 7455 */
|
|
cmplwi cr2,r8,0x8002 /* 7457 */
|
|
cror 4*cr0+eq,4*cr0+eq,4*cr1+eq /* Now test if any are true. */
|
|
cror 4*cr0+eq,4*cr0+eq,4*cr2+eq
|
|
bne 2f
|
|
|
|
1: mfspr r8,SPRN_L2CR /* On 745x, poll L2I bit (bit 10) */
|
|
rlwinm. r9,r8,0,10,10
|
|
bne 1b
|
|
b 3f
|
|
|
|
2: mfspr r8,SPRN_L2CR /* On 75x & 74[01]0, poll L2IP bit (bit 31) */
|
|
rlwinm. r9,r8,0,31,31
|
|
bne 2b
|
|
|
|
3: rlwinm r8,r8,0,11,9 /* Turn off L2I bit */
|
|
sync
|
|
isync
|
|
mtspr SPRN_L2CR,r8
|
|
sync
|
|
isync
|
|
blr
|
|
|
|
.globl _setup_L3CR
|
|
_setup_L3CR:
|
|
/* Invalidate/disable L3 cache */
|
|
sync
|
|
isync
|
|
mfspr r8,SPRN_L3CR
|
|
rlwinm r8,r8,0,1,31
|
|
ori r8,r8,L3CR_L3I@l
|
|
sync
|
|
isync
|
|
mtspr SPRN_L3CR,r8
|
|
sync
|
|
isync
|
|
|
|
/* Wait for the invalidation to complete */
|
|
1: mfspr r8,SPRN_L3CR
|
|
rlwinm. r9,r8,0,21,21
|
|
bne 1b
|
|
|
|
rlwinm r8,r8,0,22,20 /* Turn off L3I bit */
|
|
sync
|
|
isync
|
|
mtspr SPRN_L3CR,r8
|
|
sync
|
|
isync
|
|
blr
|
|
|
|
|
|
/* udelay (on non-601 processors) needs to know the period of the
|
|
* timebase in nanoseconds. This used to be hardcoded to be 60ns
|
|
* (period of 66MHz/4). Now a variable is used that is initialized to
|
|
* 60 for backward compatibility, but it can be overridden as necessary
|
|
* with code something like this:
|
|
* extern unsigned long timebase_period_ns;
|
|
* timebase_period_ns = 1000000000 / bd->bi_tbfreq;
|
|
*/
|
|
.data
|
|
.globl timebase_period_ns
|
|
timebase_period_ns:
|
|
.long 60
|
|
|
|
.text
|
|
/*
|
|
* Delay for a number of microseconds
|
|
*/
|
|
.globl udelay
|
|
udelay:
|
|
mfspr r4,SPRN_PVR
|
|
srwi r4,r4,16
|
|
cmpwi 0,r4,1 /* 601 ? */
|
|
bne .udelay_not_601
|
|
00: li r0,86 /* Instructions / microsecond? */
|
|
mtctr r0
|
|
10: addi r0,r0,0 /* NOP */
|
|
bdnz 10b
|
|
subic. r3,r3,1
|
|
bne 00b
|
|
blr
|
|
|
|
.udelay_not_601:
|
|
mulli r4,r3,1000 /* nanoseconds */
|
|
/* Change r4 to be the number of ticks using:
|
|
* (nanoseconds + (timebase_period_ns - 1 )) / timebase_period_ns
|
|
* timebase_period_ns defaults to 60 (16.6MHz) */
|
|
lis r5,timebase_period_ns@ha
|
|
lwz r5,timebase_period_ns@l(r5)
|
|
add r4,r4,r5
|
|
addi r4,r4,-1
|
|
divw r4,r4,r5 /* BUS ticks */
|
|
1: mftbu r5
|
|
mftb r6
|
|
mftbu r7
|
|
cmpw 0,r5,r7
|
|
bne 1b /* Get [synced] base time */
|
|
addc r9,r6,r4 /* Compute end time */
|
|
addze r8,r5
|
|
2: mftbu r5
|
|
cmpw 0,r5,r8
|
|
blt 2b
|
|
bgt 3f
|
|
mftb r6
|
|
cmpw 0,r6,r9
|
|
blt 2b
|
|
3: blr
|
|
|
|
.section ".relocate_code","xa"
|
|
/*
|
|
* Flush and enable instruction cache
|
|
* First, flush the data cache in case it was enabled and may be
|
|
* holding instructions for copy back.
|
|
*/
|
|
.globl flush_instruction_cache
|
|
flush_instruction_cache:
|
|
mflr r6
|
|
bl flush_data_cache
|
|
|
|
#ifdef CONFIG_8xx
|
|
lis r3, IDC_INVALL@h
|
|
mtspr SPRN_IC_CST, r3
|
|
lis r3, IDC_ENABLE@h
|
|
mtspr SPRN_IC_CST, r3
|
|
lis r3, IDC_DISABLE@h
|
|
mtspr SPRN_DC_CST, r3
|
|
#elif CONFIG_4xx
|
|
lis r3,start@h # r9 = &_start
|
|
lis r4,_etext@ha
|
|
addi r4,r4,_etext@l # r8 = &_etext
|
|
1: dcbf r0,r3 # Flush the data cache
|
|
icbi r0,r3 # Invalidate the instruction cache
|
|
addi r3,r3,0x10 # Increment by one cache line
|
|
cmplw cr0,r3,r4 # Are we at the end yet?
|
|
blt 1b # No, keep flushing and invalidating
|
|
#else
|
|
/* Enable, invalidate and then disable the L1 icache/dcache. */
|
|
li r3,0
|
|
ori r3,r3,(HID0_ICE|HID0_DCE|HID0_ICFI|HID0_DCI)
|
|
mfspr r4,SPRN_HID0
|
|
or r5,r4,r3
|
|
isync
|
|
mtspr SPRN_HID0,r5
|
|
sync
|
|
isync
|
|
ori r5,r4,HID0_ICE /* Enable cache */
|
|
mtspr SPRN_HID0,r5
|
|
sync
|
|
isync
|
|
#endif
|
|
mtlr r6
|
|
blr
|
|
|
|
#define NUM_CACHE_LINES 128*8
|
|
#define cache_flush_buffer 0x1000
|
|
|
|
/*
|
|
* Flush data cache
|
|
* Do this by just reading lots of stuff into the cache.
|
|
*/
|
|
.globl flush_data_cache
|
|
flush_data_cache:
|
|
lis r3,cache_flush_buffer@h
|
|
ori r3,r3,cache_flush_buffer@l
|
|
li r4,NUM_CACHE_LINES
|
|
mtctr r4
|
|
00: lwz r4,0(r3)
|
|
addi r3,r3,L1_CACHE_BYTES /* Next line, please */
|
|
bdnz 00b
|
|
10: blr
|
|
|
|
.previous
|
|
|