android_kernel_xiaomi_sm8350/arch/ppc64/mm
Benjamin Herrenschmidt 67b108131d [PATCH] ppc64: Fix huge pages MMU mapping bug
Current kernel has a couple of sneaky bugs in the ppc64 hugetlb code that
cause huge pages to be potentially left stale in the hash table and TLBs
(improperly invalidated), with all the nasty consequences that can have.

One is that we forgot to set the "secondary" bit in the hash PTEs when
hashing a huge page in the secondary bucket (fortunately very rare).

The other one is on non-LPAR machines (like Apple G5s), flush_hash_range()
which is used to flush a batch of PTEs simply did not work for huge pages.
Historically, our huge page code didn't batch, but this was changed without
fixing this routine.  This patch fixes both.

Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-09-23 13:35:36 -07:00
..
fault.c [PATCH] ppc64: build fix 2005-09-17 11:50:01 -07:00
hash_low.S kbuild: m68k,parisc,ppc,ppc64,s390,xtensa use generic asm-offsets.h support 2005-09-09 20:57:26 +02:00
hash_native.c [PATCH] ppc64: Fix huge pages MMU mapping bug 2005-09-23 13:35:36 -07:00
hash_utils.c
hugetlbpage.c [PATCH] ppc64: Fix huge pages MMU mapping bug 2005-09-23 13:35:36 -07:00
imalloc.c
init.c [PATCH] ppc64: Fix oops for !CONFIG_NUMA 2005-09-09 22:11:34 +10:00
Makefile
mmap.c
numa.c
slb_low.S kbuild: m68k,parisc,ppc,ppc64,s390,xtensa use generic asm-offsets.h support 2005-09-09 20:57:26 +02:00
slb.c
stab.c
tlb.c