init, x86: Move mem_encrypt_init() into arch_cpu_finalize_init()
commit 439e17576e
upstream
Invoke the X86ism mem_encrypt_init() from X86 arch_cpu_finalize_init() and
remove the weak fallback from the core code.
No functional change.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Link: https://lore.kernel.org/r/20230613224545.670360645@linutronix.de
Signed-off-by: Daniel Sneddon <daniel.sneddon@linux.intel.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:
parent
b90a399294
commit
ef54e26b2e
|
@ -64,6 +64,8 @@ static inline void __init sme_early_init(void) { }
|
||||||
static inline void __init sme_encrypt_kernel(struct boot_params *bp) { }
|
static inline void __init sme_encrypt_kernel(struct boot_params *bp) { }
|
||||||
static inline void __init sme_enable(struct boot_params *bp) { }
|
static inline void __init sme_enable(struct boot_params *bp) { }
|
||||||
|
|
||||||
|
static inline void mem_encrypt_init(void) { }
|
||||||
|
|
||||||
#endif /* CONFIG_AMD_MEM_ENCRYPT */
|
#endif /* CONFIG_AMD_MEM_ENCRYPT */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -14,6 +14,7 @@
|
||||||
#include <linux/init.h>
|
#include <linux/init.h>
|
||||||
#include <linux/kprobes.h>
|
#include <linux/kprobes.h>
|
||||||
#include <linux/kgdb.h>
|
#include <linux/kgdb.h>
|
||||||
|
#include <linux/mem_encrypt.h>
|
||||||
#include <linux/smp.h>
|
#include <linux/smp.h>
|
||||||
#include <linux/cpu.h>
|
#include <linux/cpu.h>
|
||||||
#include <linux/io.h>
|
#include <linux/io.h>
|
||||||
|
@ -2049,4 +2050,14 @@ void __init arch_cpu_finalize_init(void)
|
||||||
} else {
|
} else {
|
||||||
fpu__init_check_bugs();
|
fpu__init_check_bugs();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This needs to be called before any devices perform DMA
|
||||||
|
* operations that might use the SWIOTLB bounce buffers. It will
|
||||||
|
* mark the bounce buffers as decrypted so that their usage will
|
||||||
|
* not cause "plain-text" data to be decrypted when accessed. It
|
||||||
|
* must be called after late_time_init() so that Hyper-V x86/x64
|
||||||
|
* hypercalls work when the SWIOTLB bounce buffers are decrypted.
|
||||||
|
*/
|
||||||
|
mem_encrypt_init();
|
||||||
}
|
}
|
||||||
|
|
10
init/main.c
10
init/main.c
|
@ -485,8 +485,6 @@ void __init __weak thread_stack_cache_init(void)
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
void __init __weak mem_encrypt_init(void) { }
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Set up kernel memory allocators
|
* Set up kernel memory allocators
|
||||||
*/
|
*/
|
||||||
|
@ -648,14 +646,6 @@ asmlinkage __visible void __init start_kernel(void)
|
||||||
*/
|
*/
|
||||||
locking_selftest();
|
locking_selftest();
|
||||||
|
|
||||||
/*
|
|
||||||
* This needs to be called before any devices perform DMA
|
|
||||||
* operations that might use the SWIOTLB bounce buffers. It will
|
|
||||||
* mark the bounce buffers as decrypted so that their usage will
|
|
||||||
* not cause "plain-text" data to be decrypted when accessed.
|
|
||||||
*/
|
|
||||||
mem_encrypt_init();
|
|
||||||
|
|
||||||
#ifdef CONFIG_BLK_DEV_INITRD
|
#ifdef CONFIG_BLK_DEV_INITRD
|
||||||
if (initrd_start && !initrd_below_start_ok &&
|
if (initrd_start && !initrd_below_start_ok &&
|
||||||
page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) {
|
page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) {
|
||||||
|
|
Loading…
Reference in New Issue