init, x86: Move mem_encrypt_init() into arch_cpu_finalize_init()
commit 439e17576eb47f26b78c5bbc72e344d4206d2327 upstream Invoke the X86ism mem_encrypt_init() from X86 arch_cpu_finalize_init() and remove the weak fallback from the core code. No functional change. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Link: https://lore.kernel.org/r/20230613224545.670360645@linutronix.de Signed-off-by: Daniel Sneddon <daniel.sneddon@linux.intel.com> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:
parent
7aa2cec22e
commit
95356fff6f
@ -77,6 +77,8 @@ early_set_memory_decrypted(unsigned long vaddr, unsigned long size) { return 0;
|
|||||||
static inline int __init
|
static inline int __init
|
||||||
early_set_memory_encrypted(unsigned long vaddr, unsigned long size) { return 0; }
|
early_set_memory_encrypted(unsigned long vaddr, unsigned long size) { return 0; }
|
||||||
|
|
||||||
|
static inline void mem_encrypt_init(void) { }
|
||||||
|
|
||||||
#define __bss_decrypted
|
#define __bss_decrypted
|
||||||
|
|
||||||
#endif /* CONFIG_AMD_MEM_ENCRYPT */
|
#endif /* CONFIG_AMD_MEM_ENCRYPT */
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
#include <linux/init.h>
|
#include <linux/init.h>
|
||||||
#include <linux/kprobes.h>
|
#include <linux/kprobes.h>
|
||||||
#include <linux/kgdb.h>
|
#include <linux/kgdb.h>
|
||||||
|
#include <linux/mem_encrypt.h>
|
||||||
#include <linux/smp.h>
|
#include <linux/smp.h>
|
||||||
#include <linux/cpu.h>
|
#include <linux/cpu.h>
|
||||||
#include <linux/io.h>
|
#include <linux/io.h>
|
||||||
@ -2209,4 +2210,14 @@ void __init arch_cpu_finalize_init(void)
|
|||||||
} else {
|
} else {
|
||||||
fpu__init_check_bugs();
|
fpu__init_check_bugs();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This needs to be called before any devices perform DMA
|
||||||
|
* operations that might use the SWIOTLB bounce buffers. It will
|
||||||
|
* mark the bounce buffers as decrypted so that their usage will
|
||||||
|
* not cause "plain-text" data to be decrypted when accessed. It
|
||||||
|
* must be called after late_time_init() so that Hyper-V x86/x64
|
||||||
|
* hypercalls work when the SWIOTLB bounce buffers are decrypted.
|
||||||
|
*/
|
||||||
|
mem_encrypt_init();
|
||||||
}
|
}
|
||||||
|
11
init/main.c
11
init/main.c
@ -93,7 +93,6 @@
|
|||||||
#include <linux/cache.h>
|
#include <linux/cache.h>
|
||||||
#include <linux/rodata_test.h>
|
#include <linux/rodata_test.h>
|
||||||
#include <linux/jump_label.h>
|
#include <linux/jump_label.h>
|
||||||
#include <linux/mem_encrypt.h>
|
|
||||||
|
|
||||||
#include <asm/io.h>
|
#include <asm/io.h>
|
||||||
#include <asm/setup.h>
|
#include <asm/setup.h>
|
||||||
@ -503,8 +502,6 @@ void __init __weak thread_stack_cache_init(void)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
void __init __weak mem_encrypt_init(void) { }
|
|
||||||
|
|
||||||
void __init __weak poking_init(void) { }
|
void __init __weak poking_init(void) { }
|
||||||
|
|
||||||
void __init __weak pgtable_cache_init(void) { }
|
void __init __weak pgtable_cache_init(void) { }
|
||||||
@ -720,14 +717,6 @@ asmlinkage __visible void __init start_kernel(void)
|
|||||||
*/
|
*/
|
||||||
locking_selftest();
|
locking_selftest();
|
||||||
|
|
||||||
/*
|
|
||||||
* This needs to be called before any devices perform DMA
|
|
||||||
* operations that might use the SWIOTLB bounce buffers. It will
|
|
||||||
* mark the bounce buffers as decrypted so that their usage will
|
|
||||||
* not cause "plain-text" data to be decrypted when accessed.
|
|
||||||
*/
|
|
||||||
mem_encrypt_init();
|
|
||||||
|
|
||||||
#ifdef CONFIG_BLK_DEV_INITRD
|
#ifdef CONFIG_BLK_DEV_INITRD
|
||||||
if (initrd_start && !initrd_below_start_ok &&
|
if (initrd_start && !initrd_below_start_ok &&
|
||||||
page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) {
|
page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) {
|
||||||
|
Loading…
Reference in New Issue
Block a user