mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-16 16:54:20 +08:00
6ab3d5624e
Signed-off-by: Jörn Engel <joern@wohnheim.fh-wedel.de> Signed-off-by: Adrian Bunk <bunk@stusta.de>
359 lines
7.3 KiB
ArmAsm
359 lines
7.3 KiB
ArmAsm
/*
|
|
* This file contains low-level cache management functions
|
|
* used for sleep and CPU speed changes on Apple machines.
|
|
* (In fact the only thing that is Apple-specific is that we assume
|
|
* that we can read from ROM at physical address 0xfff00000.)
|
|
*
|
|
* Copyright (C) 2004 Paul Mackerras (paulus@samba.org) and
|
|
* Benjamin Herrenschmidt (benh@kernel.crashing.org)
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*
|
|
*/
|
|
|
|
#include <asm/processor.h>
|
|
#include <asm/ppc_asm.h>
|
|
#include <asm/cputable.h>
|
|
|
|
/*
|
|
* Flush and disable all data caches (dL1, L2, L3). This is used
|
|
* when going to sleep, when doing a PMU based cpufreq transition,
|
|
* or when "offlining" a CPU on SMP machines. This code is over
|
|
* paranoid, but I've had enough issues with various CPU revs and
|
|
* bugs that I decided it was worth beeing over cautious
|
|
*/
|
|
|
|
_GLOBAL(flush_disable_caches)
|
|
#ifndef CONFIG_6xx
|
|
blr
|
|
#else
|
|
BEGIN_FTR_SECTION
|
|
b flush_disable_745x
|
|
END_FTR_SECTION_IFSET(CPU_FTR_SPEC7450)
|
|
BEGIN_FTR_SECTION
|
|
b flush_disable_75x
|
|
END_FTR_SECTION_IFSET(CPU_FTR_L2CR)
|
|
b __flush_disable_L1
|
|
|
|
/* This is the code for G3 and 74[01]0 */
|
|
flush_disable_75x:
|
|
mflr r10
|
|
|
|
/* Turn off EE and DR in MSR */
|
|
mfmsr r11
|
|
rlwinm r0,r11,0,~MSR_EE
|
|
rlwinm r0,r0,0,~MSR_DR
|
|
sync
|
|
mtmsr r0
|
|
isync
|
|
|
|
/* Stop DST streams */
|
|
BEGIN_FTR_SECTION
|
|
DSSALL
|
|
sync
|
|
END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
|
|
|
|
/* Stop DPM */
|
|
mfspr r8,SPRN_HID0 /* Save SPRN_HID0 in r8 */
|
|
rlwinm r4,r8,0,12,10 /* Turn off HID0[DPM] */
|
|
sync
|
|
mtspr SPRN_HID0,r4 /* Disable DPM */
|
|
sync
|
|
|
|
/* Disp-flush L1. We have a weird problem here that I never
|
|
* totally figured out. On 750FX, using the ROM for the flush
|
|
* results in a non-working flush. We use that workaround for
|
|
* now until I finally understand what's going on. --BenH
|
|
*/
|
|
|
|
/* ROM base by default */
|
|
lis r4,0xfff0
|
|
mfpvr r3
|
|
srwi r3,r3,16
|
|
cmplwi cr0,r3,0x7000
|
|
bne+ 1f
|
|
/* RAM base on 750FX */
|
|
li r4,0
|
|
1: li r4,0x4000
|
|
mtctr r4
|
|
1: lwz r0,0(r4)
|
|
addi r4,r4,32
|
|
bdnz 1b
|
|
sync
|
|
isync
|
|
|
|
/* Disable / invalidate / enable L1 data */
|
|
mfspr r3,SPRN_HID0
|
|
rlwinm r3,r3,0,~(HID0_DCE | HID0_ICE)
|
|
mtspr SPRN_HID0,r3
|
|
sync
|
|
isync
|
|
ori r3,r3,(HID0_DCE|HID0_DCI|HID0_ICE|HID0_ICFI)
|
|
sync
|
|
isync
|
|
mtspr SPRN_HID0,r3
|
|
xori r3,r3,(HID0_DCI|HID0_ICFI)
|
|
mtspr SPRN_HID0,r3
|
|
sync
|
|
|
|
/* Get the current enable bit of the L2CR into r4 */
|
|
mfspr r5,SPRN_L2CR
|
|
/* Set to data-only (pre-745x bit) */
|
|
oris r3,r5,L2CR_L2DO@h
|
|
b 2f
|
|
/* When disabling L2, code must be in L1 */
|
|
.balign 32
|
|
1: mtspr SPRN_L2CR,r3
|
|
3: sync
|
|
isync
|
|
b 1f
|
|
2: b 3f
|
|
3: sync
|
|
isync
|
|
b 1b
|
|
1: /* disp-flush L2. The interesting thing here is that the L2 can be
|
|
* up to 2Mb ... so using the ROM, we'll end up wrapping back to memory
|
|
* but that is probbaly fine. We disp-flush over 4Mb to be safe
|
|
*/
|
|
lis r4,2
|
|
mtctr r4
|
|
lis r4,0xfff0
|
|
1: lwz r0,0(r4)
|
|
addi r4,r4,32
|
|
bdnz 1b
|
|
sync
|
|
isync
|
|
lis r4,2
|
|
mtctr r4
|
|
lis r4,0xfff0
|
|
1: dcbf 0,r4
|
|
addi r4,r4,32
|
|
bdnz 1b
|
|
sync
|
|
isync
|
|
|
|
/* now disable L2 */
|
|
rlwinm r5,r5,0,~L2CR_L2E
|
|
b 2f
|
|
/* When disabling L2, code must be in L1 */
|
|
.balign 32
|
|
1: mtspr SPRN_L2CR,r5
|
|
3: sync
|
|
isync
|
|
b 1f
|
|
2: b 3f
|
|
3: sync
|
|
isync
|
|
b 1b
|
|
1: sync
|
|
isync
|
|
/* Invalidate L2. This is pre-745x, we clear the L2I bit ourselves */
|
|
oris r4,r5,L2CR_L2I@h
|
|
mtspr SPRN_L2CR,r4
|
|
sync
|
|
isync
|
|
|
|
/* Wait for the invalidation to complete */
|
|
1: mfspr r3,SPRN_L2CR
|
|
rlwinm. r0,r3,0,31,31
|
|
bne 1b
|
|
|
|
/* Clear L2I */
|
|
xoris r4,r4,L2CR_L2I@h
|
|
sync
|
|
mtspr SPRN_L2CR,r4
|
|
sync
|
|
|
|
/* now disable the L1 data cache */
|
|
mfspr r0,SPRN_HID0
|
|
rlwinm r0,r0,0,~(HID0_DCE|HID0_ICE)
|
|
mtspr SPRN_HID0,r0
|
|
sync
|
|
isync
|
|
|
|
/* Restore HID0[DPM] to whatever it was before */
|
|
sync
|
|
mfspr r0,SPRN_HID0
|
|
rlwimi r0,r8,0,11,11 /* Turn back HID0[DPM] */
|
|
mtspr SPRN_HID0,r0
|
|
sync
|
|
|
|
/* restore DR and EE */
|
|
sync
|
|
mtmsr r11
|
|
isync
|
|
|
|
mtlr r10
|
|
blr
|
|
|
|
/* This code is for 745x processors */
|
|
flush_disable_745x:
|
|
/* Turn off EE and DR in MSR */
|
|
mfmsr r11
|
|
rlwinm r0,r11,0,~MSR_EE
|
|
rlwinm r0,r0,0,~MSR_DR
|
|
sync
|
|
mtmsr r0
|
|
isync
|
|
|
|
/* Stop prefetch streams */
|
|
DSSALL
|
|
sync
|
|
|
|
/* Disable L2 prefetching */
|
|
mfspr r0,SPRN_MSSCR0
|
|
rlwinm r0,r0,0,0,29
|
|
mtspr SPRN_MSSCR0,r0
|
|
sync
|
|
isync
|
|
lis r4,0
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
dcbf 0,r4
|
|
|
|
/* Due to a bug with the HW flush on some CPU revs, we occasionally
|
|
* experience data corruption. I'm adding a displacement flush along
|
|
* with a dcbf loop over a few Mb to "help". The problem isn't totally
|
|
* fixed by this in theory, but at least, in practice, I couldn't reproduce
|
|
* it even with a big hammer...
|
|
*/
|
|
|
|
lis r4,0x0002
|
|
mtctr r4
|
|
li r4,0
|
|
1:
|
|
lwz r0,0(r4)
|
|
addi r4,r4,32 /* Go to start of next cache line */
|
|
bdnz 1b
|
|
isync
|
|
|
|
/* Now, flush the first 4MB of memory */
|
|
lis r4,0x0002
|
|
mtctr r4
|
|
li r4,0
|
|
sync
|
|
1:
|
|
dcbf 0,r4
|
|
addi r4,r4,32 /* Go to start of next cache line */
|
|
bdnz 1b
|
|
|
|
/* Flush and disable the L1 data cache */
|
|
mfspr r6,SPRN_LDSTCR
|
|
lis r3,0xfff0 /* read from ROM for displacement flush */
|
|
li r4,0xfe /* start with only way 0 unlocked */
|
|
li r5,128 /* 128 lines in each way */
|
|
1: mtctr r5
|
|
rlwimi r6,r4,0,24,31
|
|
mtspr SPRN_LDSTCR,r6
|
|
sync
|
|
isync
|
|
2: lwz r0,0(r3) /* touch each cache line */
|
|
addi r3,r3,32
|
|
bdnz 2b
|
|
rlwinm r4,r4,1,24,30 /* move on to the next way */
|
|
ori r4,r4,1
|
|
cmpwi r4,0xff /* all done? */
|
|
bne 1b
|
|
/* now unlock the L1 data cache */
|
|
li r4,0
|
|
rlwimi r6,r4,0,24,31
|
|
sync
|
|
mtspr SPRN_LDSTCR,r6
|
|
sync
|
|
isync
|
|
|
|
/* Flush the L2 cache using the hardware assist */
|
|
mfspr r3,SPRN_L2CR
|
|
cmpwi r3,0 /* check if it is enabled first */
|
|
bge 4f
|
|
oris r0,r3,(L2CR_L2IO_745x|L2CR_L2DO_745x)@h
|
|
b 2f
|
|
/* When disabling/locking L2, code must be in L1 */
|
|
.balign 32
|
|
1: mtspr SPRN_L2CR,r0 /* lock the L2 cache */
|
|
3: sync
|
|
isync
|
|
b 1f
|
|
2: b 3f
|
|
3: sync
|
|
isync
|
|
b 1b
|
|
1: sync
|
|
isync
|
|
ori r0,r3,L2CR_L2HWF_745x
|
|
sync
|
|
mtspr SPRN_L2CR,r0 /* set the hardware flush bit */
|
|
3: mfspr r0,SPRN_L2CR /* wait for it to go to 0 */
|
|
andi. r0,r0,L2CR_L2HWF_745x
|
|
bne 3b
|
|
sync
|
|
rlwinm r3,r3,0,~L2CR_L2E
|
|
b 2f
|
|
/* When disabling L2, code must be in L1 */
|
|
.balign 32
|
|
1: mtspr SPRN_L2CR,r3 /* disable the L2 cache */
|
|
3: sync
|
|
isync
|
|
b 1f
|
|
2: b 3f
|
|
3: sync
|
|
isync
|
|
b 1b
|
|
1: sync
|
|
isync
|
|
oris r4,r3,L2CR_L2I@h
|
|
mtspr SPRN_L2CR,r4
|
|
sync
|
|
isync
|
|
1: mfspr r4,SPRN_L2CR
|
|
andis. r0,r4,L2CR_L2I@h
|
|
bne 1b
|
|
sync
|
|
|
|
BEGIN_FTR_SECTION
|
|
/* Flush the L3 cache using the hardware assist */
|
|
4: mfspr r3,SPRN_L3CR
|
|
cmpwi r3,0 /* check if it is enabled */
|
|
bge 6f
|
|
oris r0,r3,L3CR_L3IO@h
|
|
ori r0,r0,L3CR_L3DO
|
|
sync
|
|
mtspr SPRN_L3CR,r0 /* lock the L3 cache */
|
|
sync
|
|
isync
|
|
ori r0,r0,L3CR_L3HWF
|
|
sync
|
|
mtspr SPRN_L3CR,r0 /* set the hardware flush bit */
|
|
5: mfspr r0,SPRN_L3CR /* wait for it to go to zero */
|
|
andi. r0,r0,L3CR_L3HWF
|
|
bne 5b
|
|
rlwinm r3,r3,0,~L3CR_L3E
|
|
sync
|
|
mtspr SPRN_L3CR,r3 /* disable the L3 cache */
|
|
sync
|
|
ori r4,r3,L3CR_L3I
|
|
mtspr SPRN_L3CR,r4
|
|
1: mfspr r4,SPRN_L3CR
|
|
andi. r0,r4,L3CR_L3I
|
|
bne 1b
|
|
sync
|
|
END_FTR_SECTION_IFSET(CPU_FTR_L3CR)
|
|
|
|
6: mfspr r0,SPRN_HID0 /* now disable the L1 data cache */
|
|
rlwinm r0,r0,0,~HID0_DCE
|
|
mtspr SPRN_HID0,r0
|
|
sync
|
|
isync
|
|
mtmsr r11 /* restore DR and EE */
|
|
isync
|
|
blr
|
|
#endif /* CONFIG_6xx */
|