1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3  *
4  *	Trampoline.S	Derived from Setup.S by Linus Torvalds
5  *
6  *	4 Jan 1997 Michael Chastain: changed to gnu as.
7  *	15 Sept 2005 Eric Biederman: 64bit PIC support
8  *
9  *	Entry: CS:IP point to the start of our code, we are
10  *	in real mode with no stack, but the rest of the
11  *	trampoline page to make our stack and everything else
12  *	is a mystery.
13  *
14  *	On entry to trampoline_start, the processor is in real mode
15  *	with 16-bit addressing and 16-bit data.  CS has some value
16  *	and IP is zero.  Thus, data addresses need to be absolute
17  *	(no relocation) and are taken with regard to r_base.
18  *
19  *	With the addition of trampoline_level4_pgt this code can
20  *	now enter a 64bit kernel that lives at arbitrary 64bit
21  *	physical addresses.
22  *
23  *	If you work on this file, check the object module with objdump
24  *	--full-contents --reloc to make sure there are no relocation
25  *	entries.
26  */
27 
28 #include <linux/linkage.h>
29 #include <asm/pgtable_types.h>
30 #include <asm/page_types.h>
31 #include <asm/msr.h>
32 #include <asm/segment.h>
33 #include <asm/processor-flags.h>
34 #include <asm/realmode.h>
35 #include "realmode.h"
36 
37 	.text
38 	.code16
39 
40 .macro LOCK_AND_LOAD_REALMODE_ESP lock_pa=0
41 	/*
42 	 * Make sure only one CPU fiddles with the realmode stack
43 	 */
44 .Llock_rm\@:
45 	.if \lock_pa
46         lock btsl       $0, pa_tr_lock
47 	.else
48         lock btsl       $0, tr_lock
49 	.endif
50         jnc             2f
51         pause
52         jmp             .Llock_rm\@
53 2:
54 	# Setup stack
55 	movl	$rm_stack_end, %esp
56 .endm
57 
58 	.balign	PAGE_SIZE
59 SYM_CODE_START(trampoline_start)
60 	cli			# We should be safe anyway
61 	wbinvd
62 
63 	LJMPW_RM(1f)
64 1:
65 	mov	%cs, %ax	# Code and data in the same place
66 	mov	%ax, %ds
67 	mov	%ax, %es
68 	mov	%ax, %ss
69 
70 	LOCK_AND_LOAD_REALMODE_ESP
71 
72 	call	verify_cpu		# Verify the cpu supports long mode
73 	testl   %eax, %eax		# Check for return code
74 	jnz	no_longmode
75 
76 .Lswitch_to_protected:
77 	/*
78 	 * GDT tables in non default location kernel can be beyond 16MB and
79 	 * lgdt will not be able to load the address as in real mode default
80 	 * operand size is 16bit. Use lgdtl instead to force operand size
81 	 * to 32 bit.
82 	 */
83 
84 	lidtl	tr_idt	# load idt with 0, 0
85 	lgdtl	tr_gdt	# load gdt with whatever is appropriate
86 
87 	movw	$__KERNEL_DS, %dx	# Data segment descriptor
88 
89 	# Enable protected mode
90 	movl	$(CR0_STATE & ~X86_CR0_PG), %eax
91 	movl	%eax, %cr0		# into protected mode
92 
93 	# flush prefetch and jump to startup_32
94 	ljmpl	$__KERNEL32_CS, $pa_startup_32
95 
96 no_longmode:
97 	hlt
98 	jmp no_longmode
99 SYM_CODE_END(trampoline_start)
100 
101 #ifdef CONFIG_AMD_MEM_ENCRYPT
102 /* SEV-ES supports non-zero IP for entry points - no alignment needed */
103 SYM_CODE_START(sev_es_trampoline_start)
104 	cli			# We should be safe anyway
105 
106 	LJMPW_RM(1f)
107 1:
108 	mov	%cs, %ax	# Code and data in the same place
109 	mov	%ax, %ds
110 	mov	%ax, %es
111 	mov	%ax, %ss
112 
113 	LOCK_AND_LOAD_REALMODE_ESP
114 
115 	jmp	.Lswitch_to_protected
116 SYM_CODE_END(sev_es_trampoline_start)
117 #endif	/* CONFIG_AMD_MEM_ENCRYPT */
118 
119 #include "../kernel/verify_cpu.S"
120 
121 	.section ".text32","ax"
122 	.code32
123 	.balign 4
124 SYM_CODE_START(startup_32)
125 	movl	%edx, %ss
126 	addl	$pa_real_mode_base, %esp
127 	movl	%edx, %ds
128 	movl	%edx, %es
129 	movl	%edx, %fs
130 	movl	%edx, %gs
131 
132 	/*
133 	 * Check for memory encryption support. This is a safety net in
134 	 * case BIOS hasn't done the necessary step of setting the bit in
135 	 * the MSR for this AP. If SME is active and we've gotten this far
136 	 * then it is safe for us to set the MSR bit and continue. If we
137 	 * don't we'll eventually crash trying to execute encrypted
138 	 * instructions.
139 	 */
140 	btl	$TH_FLAGS_SME_ACTIVE_BIT, pa_tr_flags
141 	jnc	.Ldone
142 	movl	$MSR_AMD64_SYSCFG, %ecx
143 	rdmsr
144 	bts	$MSR_AMD64_SYSCFG_MEM_ENCRYPT_BIT, %eax
145 	jc	.Ldone
146 
147 	/*
148 	 * Memory encryption is enabled but the SME enable bit for this
149 	 * CPU has has not been set.  It is safe to set it, so do so.
150 	 */
151 	wrmsr
152 .Ldone:
153 
154 	movl	pa_tr_cr4, %eax
155 	movl	%eax, %cr4		# Enable PAE mode
156 
157 	# Setup trampoline 4 level pagetables
158 	movl	$pa_trampoline_pgd, %eax
159 	movl	%eax, %cr3
160 
161 	# Set up EFER
162 	movl	$MSR_EFER, %ecx
163 	rdmsr
164 	/*
165 	 * Skip writing to EFER if the register already has desired
166 	 * value (to avoid #VE for the TDX guest).
167 	 */
168 	cmp	pa_tr_efer, %eax
169 	jne	.Lwrite_efer
170 	cmp	pa_tr_efer + 4, %edx
171 	je	.Ldone_efer
172 .Lwrite_efer:
173 	movl	pa_tr_efer, %eax
174 	movl	pa_tr_efer + 4, %edx
175 	wrmsr
176 
177 .Ldone_efer:
178 	# Enable paging and in turn activate Long Mode.
179 	movl	$CR0_STATE, %eax
180 	movl	%eax, %cr0
181 
182 	/*
183 	 * At this point we're in long mode but in 32bit compatibility mode
184 	 * with EFER.LME = 1, CS.L = 0, CS.D = 1 (and in turn
185 	 * EFER.LMA = 1). Now we want to jump in 64bit mode, to do that we use
186 	 * the new gdt/idt that has __KERNEL_CS with CS.L = 1.
187 	 */
188 	ljmpl	$__KERNEL_CS, $pa_startup_64
189 SYM_CODE_END(startup_32)
190 
191 SYM_CODE_START(pa_trampoline_compat)
192 	/*
193 	 * In compatibility mode.  Prep ESP and DX for startup_32, then disable
194 	 * paging and complete the switch to legacy 32-bit mode.
195 	 */
196 	LOCK_AND_LOAD_REALMODE_ESP lock_pa=1
197 	movw	$__KERNEL_DS, %dx
198 
199 	movl	$(CR0_STATE & ~X86_CR0_PG), %eax
200 	movl	%eax, %cr0
201 	ljmpl   $__KERNEL32_CS, $pa_startup_32
202 SYM_CODE_END(pa_trampoline_compat)
203 
204 	.section ".text64","ax"
205 	.code64
206 	.balign 4
207 SYM_CODE_START(startup_64)
208 	# Now jump into the kernel using virtual addresses
209 	jmpq	*tr_start(%rip)
210 SYM_CODE_END(startup_64)
211 
212 SYM_CODE_START(trampoline_start64)
213 	/*
214 	 * APs start here on a direct transfer from 64-bit BIOS with identity
215 	 * mapped page tables.  Load the kernel's GDT in order to gear down to
216 	 * 32-bit mode (to handle 4-level vs. 5-level paging), and to (re)load
217 	 * segment registers.  Load the zero IDT so any fault triggers a
218 	 * shutdown instead of jumping back into BIOS.
219 	 */
220 	lidt	tr_idt(%rip)
221 	lgdt	tr_gdt64(%rip)
222 
223 	ljmpl	*tr_compat(%rip)
224 SYM_CODE_END(trampoline_start64)
225 
226 	.section ".rodata","a"
227 	# Duplicate the global descriptor table
228 	# so the kernel can live anywhere
229 	.balign	16
230 SYM_DATA_START(tr_gdt)
231 	.short	tr_gdt_end - tr_gdt - 1	# gdt limit
232 	.long	pa_tr_gdt
233 	.short	0
234 	.quad	0x00cf9b000000ffff	# __KERNEL32_CS
235 	.quad	0x00af9b000000ffff	# __KERNEL_CS
236 	.quad	0x00cf93000000ffff	# __KERNEL_DS
237 SYM_DATA_END_LABEL(tr_gdt, SYM_L_LOCAL, tr_gdt_end)
238 
239 SYM_DATA_START(tr_gdt64)
240 	.short	tr_gdt_end - tr_gdt - 1	# gdt limit
241 	.long	pa_tr_gdt
242 	.long	0
243 SYM_DATA_END(tr_gdt64)
244 
245 SYM_DATA_START(tr_compat)
246 	.long	pa_trampoline_compat
247 	.short	__KERNEL32_CS
248 SYM_DATA_END(tr_compat)
249 
250 	.bss
251 	.balign	PAGE_SIZE
252 SYM_DATA(trampoline_pgd, .space PAGE_SIZE)
253 
254 	.balign	8
255 SYM_DATA_START(trampoline_header)
256 	SYM_DATA_LOCAL(tr_start,	.space 8)
257 	SYM_DATA(tr_efer,		.space 8)
258 	SYM_DATA(tr_cr4,		.space 4)
259 	SYM_DATA(tr_flags,		.space 4)
260 	SYM_DATA(tr_lock,		.space 4)
261 SYM_DATA_END(trampoline_header)
262 
263 #include "trampoline_common.S"
264