hc
2024-12-19 9370bb92b2d16684ee45cf24e879c93c509162da
kernel/arch/arm64/kernel/hyp-stub.S
....@@ -1,27 +1,16 @@
1
+/* SPDX-License-Identifier: GPL-2.0-only */
12 /*
23 * Hypervisor stub
34 *
45 * Copyright (C) 2012 ARM Ltd.
56 * Author: Marc Zyngier <marc.zyngier@arm.com>
6
- *
7
- * This program is free software; you can redistribute it and/or modify
8
- * it under the terms of the GNU General Public License version 2 as
9
- * published by the Free Software Foundation.
10
- *
11
- * This program is distributed in the hope that it will be useful,
12
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
13
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14
- * GNU General Public License for more details.
15
- *
16
- * You should have received a copy of the GNU General Public License
17
- * along with this program. If not, see <http://www.gnu.org/licenses/>.
187 */
198
209 #include <linux/init.h>
2110 #include <linux/linkage.h>
22
-#include <linux/irqchip/arm-gic-v3.h>
2311
2412 #include <asm/assembler.h>
13
+#include <asm/el2_setup.h>
2514 #include <asm/kvm_arm.h>
2615 #include <asm/kvm_asm.h>
2716 #include <asm/ptrace.h>
....@@ -32,7 +21,7 @@
3221
3322 .align 11
3423
35
-ENTRY(__hyp_stub_vectors)
24
+SYM_CODE_START(__hyp_stub_vectors)
3625 ventry el2_sync_invalid // Synchronous EL2t
3726 ventry el2_irq_invalid // IRQ EL2t
3827 ventry el2_fiq_invalid // FIQ EL2t
....@@ -52,15 +41,18 @@
5241 ventry el1_irq_invalid // IRQ 32-bit EL1
5342 ventry el1_fiq_invalid // FIQ 32-bit EL1
5443 ventry el1_error_invalid // Error 32-bit EL1
55
-ENDPROC(__hyp_stub_vectors)
44
+SYM_CODE_END(__hyp_stub_vectors)
5645
5746 .align 11
5847
59
-el1_sync:
48
+SYM_CODE_START_LOCAL(el1_sync)
6049 cmp x0, #HVC_SET_VECTORS
61
- b.ne 2f
50
+ b.ne 1f
6251 msr vbar_el2, x1
6352 b 9f
53
+
54
+1: cmp x0, #HVC_VHE_RESTART
55
+ b.eq mutate_to_vhe
6456
6557 2: cmp x0, #HVC_SOFT_RESTART
6658 b.ne 3f
....@@ -74,17 +66,114 @@
7466 beq 9f // Nothing to reset!
7567
7668 /* Someone called kvm_call_hyp() against the hyp-stub... */
77
- ldr x0, =HVC_STUB_ERR
69
+ mov_q x0, HVC_STUB_ERR
7870 eret
7971
8072 9: mov x0, xzr
8173 eret
82
-ENDPROC(el1_sync)
74
+SYM_CODE_END(el1_sync)
75
+
76
+// nVHE? No way! Give me the real thing!
77
+SYM_CODE_START_LOCAL(mutate_to_vhe)
78
+ // Sanity check: MMU *must* be off
79
+ mrs x1, sctlr_el2
80
+ tbnz x1, #0, 1f
81
+
82
+ // Needs to be VHE capable, obviously
83
+ mrs x1, id_aa64mmfr1_el1
84
+ ubfx x1, x1, #ID_AA64MMFR1_VHE_SHIFT, #4
85
+ cbz x1, 1f
86
+
87
+ // Check whether VHE is disabled from the command line
88
+ adr_l x1, id_aa64mmfr1_override
89
+ ldr x2, [x1, FTR_OVR_VAL_OFFSET]
90
+ ldr x1, [x1, FTR_OVR_MASK_OFFSET]
91
+ ubfx x2, x2, #ID_AA64MMFR1_VHE_SHIFT, #4
92
+ ubfx x1, x1, #ID_AA64MMFR1_VHE_SHIFT, #4
93
+ cmp x1, xzr
94
+ and x2, x2, x1
95
+ csinv x2, x2, xzr, ne
96
+ cbnz x2, 2f
97
+
98
+1: mov_q x0, HVC_STUB_ERR
99
+ eret
100
+2:
101
+ // Engage the VHE magic!
102
+ mov_q x0, HCR_HOST_VHE_FLAGS
103
+ msr hcr_el2, x0
104
+ isb
105
+
106
+ // Use the EL1 allocated stack, per-cpu offset
107
+ mrs x0, sp_el1
108
+ mov sp, x0
109
+ mrs x0, tpidr_el1
110
+ msr tpidr_el2, x0
111
+
112
+ // FP configuration, vectors
113
+ mrs_s x0, SYS_CPACR_EL12
114
+ msr cpacr_el1, x0
115
+ mrs_s x0, SYS_VBAR_EL12
116
+ msr vbar_el1, x0
117
+
118
+ // Use EL2 translations for SPE & TRBE and disable access from EL1
119
+ mrs x0, mdcr_el2
120
+ bic x0, x0, #(MDCR_EL2_E2PB_MASK << MDCR_EL2_E2PB_SHIFT)
121
+ bic x0, x0, #(MDCR_EL2_E2TB_MASK << MDCR_EL2_E2TB_SHIFT)
122
+ msr mdcr_el2, x0
123
+
124
+ // Transfer the MM state from EL1 to EL2
125
+ mrs_s x0, SYS_TCR_EL12
126
+ msr tcr_el1, x0
127
+ mrs_s x0, SYS_TTBR0_EL12
128
+ msr ttbr0_el1, x0
129
+ mrs_s x0, SYS_TTBR1_EL12
130
+ msr ttbr1_el1, x0
131
+ mrs_s x0, SYS_MAIR_EL12
132
+ msr mair_el1, x0
133
+ isb
134
+
135
+ // Hack the exception return to stay at EL2
136
+ mrs x0, spsr_el1
137
+ and x0, x0, #~PSR_MODE_MASK
138
+ mov x1, #PSR_MODE_EL2h
139
+ orr x0, x0, x1
140
+ msr spsr_el1, x0
141
+
142
+ b enter_vhe
143
+SYM_CODE_END(mutate_to_vhe)
144
+
145
+ // At the point where we reach enter_vhe(), we run with
146
+ // the MMU off (which is enforced by mutate_to_vhe()).
147
+ // We thus need to be in the idmap, or everything will
148
+ // explode when enabling the MMU.
149
+
150
+ .pushsection .idmap.text, "ax"
151
+
152
+SYM_CODE_START_LOCAL(enter_vhe)
153
+ // Invalidate TLBs before enabling the MMU
154
+ tlbi vmalle1
155
+ dsb nsh
156
+ isb
157
+
158
+ // Enable the EL2 S1 MMU, as set up from EL1
159
+ mrs_s x0, SYS_SCTLR_EL12
160
+ set_sctlr_el1 x0
161
+
162
+ // Disable the EL1 S1 MMU for a good measure
163
+ mov_q x0, INIT_SCTLR_EL1_MMU_OFF
164
+ msr_s SYS_SCTLR_EL12, x0
165
+
166
+ mov x0, xzr
167
+
168
+ eret
169
+SYM_CODE_END(enter_vhe)
170
+
171
+ .popsection
83172
84173 .macro invalid_vector label
85
-\label:
174
+SYM_CODE_START_LOCAL(\label)
86175 b \label
87
-ENDPROC(\label)
176
+SYM_CODE_END(\label)
88177 .endm
89178
90179 invalid_vector el2_sync_invalid
....@@ -95,6 +184,8 @@
95184 invalid_vector el1_irq_invalid
96185 invalid_vector el1_fiq_invalid
97186 invalid_vector el1_error_invalid
187
+
188
+ .popsection
98189
99190 /*
100191 * __hyp_set_vectors: Call this after boot to set the initial hypervisor
....@@ -117,15 +208,39 @@
117208 * initialisation entry point.
118209 */
119210
120
-ENTRY(__hyp_set_vectors)
211
+SYM_FUNC_START(__hyp_set_vectors)
121212 mov x1, x0
122213 mov x0, #HVC_SET_VECTORS
123214 hvc #0
124215 ret
125
-ENDPROC(__hyp_set_vectors)
216
+SYM_FUNC_END(__hyp_set_vectors)
126217
127
-ENTRY(__hyp_reset_vectors)
218
+SYM_FUNC_START(__hyp_reset_vectors)
128219 mov x0, #HVC_RESET_VECTORS
129220 hvc #0
130221 ret
131
-ENDPROC(__hyp_reset_vectors)
222
+SYM_FUNC_END(__hyp_reset_vectors)
223
+
224
+/*
225
+ * Entry point to switch to VHE if deemed capable
226
+ */
227
+SYM_FUNC_START(switch_to_vhe)
228
+#ifdef CONFIG_ARM64_VHE
229
+ // Need to have booted at EL2
230
+ adr_l x1, __boot_cpu_mode
231
+ ldr w0, [x1]
232
+ cmp w0, #BOOT_CPU_MODE_EL2
233
+ b.ne 1f
234
+
235
+ // and still be at EL1
236
+ mrs x0, CurrentEL
237
+ cmp x0, #CurrentEL_EL1
238
+ b.ne 1f
239
+
240
+ // Turn the world upside down
241
+ mov x0, #HVC_VHE_RESTART
242
+ hvc #0
243
+1:
244
+#endif
245
+ ret
246
+SYM_FUNC_END(switch_to_vhe)