blob: 04dbac6bdf4db972c275fd7fbb9a885e2eb68a63 [file] [log] [blame]
Christoffer Dall749cf76c2013-01-20 18:28:06 -05001/*
2 * Copyright (C) 2012 - Virtual Open Systems and Columbia University
3 * Author: Christoffer Dall <c.dall@virtualopensystems.com>
4 *
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License, version 2, as
7 * published by the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write to the Free Software
16 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
17 */
18
Christoffer Dall5b3e5e52013-01-20 18:28:09 -050019#include <linux/mm.h>
20#include <linux/kvm_host.h>
21#include <asm/kvm_arm.h>
Christoffer Dall749cf76c2013-01-20 18:28:06 -050022#include <asm/kvm_emulate.h>
Marc Zyngierc5997562012-12-08 18:13:18 +000023#include <asm/opcodes.h>
Christoffer Dall5b3e5e52013-01-20 18:28:09 -050024#include <trace/events/kvm.h>
25
26#include "trace.h"
Christoffer Dall749cf76c2013-01-20 18:28:06 -050027
28#define VCPU_NR_MODES 6
29#define VCPU_REG_OFFSET_USR 0
30#define VCPU_REG_OFFSET_FIQ 1
31#define VCPU_REG_OFFSET_IRQ 2
32#define VCPU_REG_OFFSET_SVC 3
33#define VCPU_REG_OFFSET_ABT 4
34#define VCPU_REG_OFFSET_UND 5
35#define REG_OFFSET(_reg) \
36 (offsetof(struct kvm_regs, _reg) / sizeof(u32))
37
38#define USR_REG_OFFSET(_num) REG_OFFSET(usr_regs.uregs[_num])
39
40static const unsigned long vcpu_reg_offsets[VCPU_NR_MODES][15] = {
41 /* USR/SYS Registers */
42 [VCPU_REG_OFFSET_USR] = {
43 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
44 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
45 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
46 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
47 USR_REG_OFFSET(12), USR_REG_OFFSET(13), USR_REG_OFFSET(14),
48 },
49
50 /* FIQ Registers */
51 [VCPU_REG_OFFSET_FIQ] = {
52 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
53 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
54 USR_REG_OFFSET(6), USR_REG_OFFSET(7),
55 REG_OFFSET(fiq_regs[0]), /* r8 */
56 REG_OFFSET(fiq_regs[1]), /* r9 */
57 REG_OFFSET(fiq_regs[2]), /* r10 */
58 REG_OFFSET(fiq_regs[3]), /* r11 */
59 REG_OFFSET(fiq_regs[4]), /* r12 */
60 REG_OFFSET(fiq_regs[5]), /* r13 */
61 REG_OFFSET(fiq_regs[6]), /* r14 */
62 },
63
64 /* IRQ Registers */
65 [VCPU_REG_OFFSET_IRQ] = {
66 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
67 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
68 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
69 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
70 USR_REG_OFFSET(12),
71 REG_OFFSET(irq_regs[0]), /* r13 */
72 REG_OFFSET(irq_regs[1]), /* r14 */
73 },
74
75 /* SVC Registers */
76 [VCPU_REG_OFFSET_SVC] = {
77 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
78 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
79 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
80 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
81 USR_REG_OFFSET(12),
82 REG_OFFSET(svc_regs[0]), /* r13 */
83 REG_OFFSET(svc_regs[1]), /* r14 */
84 },
85
86 /* ABT Registers */
87 [VCPU_REG_OFFSET_ABT] = {
88 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
89 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
90 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
91 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
92 USR_REG_OFFSET(12),
93 REG_OFFSET(abt_regs[0]), /* r13 */
94 REG_OFFSET(abt_regs[1]), /* r14 */
95 },
96
97 /* UND Registers */
98 [VCPU_REG_OFFSET_UND] = {
99 USR_REG_OFFSET(0), USR_REG_OFFSET(1), USR_REG_OFFSET(2),
100 USR_REG_OFFSET(3), USR_REG_OFFSET(4), USR_REG_OFFSET(5),
101 USR_REG_OFFSET(6), USR_REG_OFFSET(7), USR_REG_OFFSET(8),
102 USR_REG_OFFSET(9), USR_REG_OFFSET(10), USR_REG_OFFSET(11),
103 USR_REG_OFFSET(12),
104 REG_OFFSET(und_regs[0]), /* r13 */
105 REG_OFFSET(und_regs[1]), /* r14 */
106 },
107};
108
109/*
110 * Return a pointer to the register number valid in the current mode of
111 * the virtual CPU.
112 */
Marc Zyngierdb730d82012-10-03 11:17:02 +0100113unsigned long *vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num)
Christoffer Dall749cf76c2013-01-20 18:28:06 -0500114{
Marc Zyngierdb730d82012-10-03 11:17:02 +0100115 unsigned long *reg_array = (unsigned long *)&vcpu->arch.regs;
116 unsigned long mode = *vcpu_cpsr(vcpu) & MODE_MASK;
Christoffer Dall749cf76c2013-01-20 18:28:06 -0500117
118 switch (mode) {
119 case USR_MODE...SVC_MODE:
120 mode &= ~MODE32_BIT; /* 0 ... 3 */
121 break;
122
123 case ABT_MODE:
124 mode = VCPU_REG_OFFSET_ABT;
125 break;
126
127 case UND_MODE:
128 mode = VCPU_REG_OFFSET_UND;
129 break;
130
131 case SYSTEM_MODE:
132 mode = VCPU_REG_OFFSET_USR;
133 break;
134
135 default:
136 BUG();
137 }
138
139 return reg_array + vcpu_reg_offsets[mode][reg_num];
140}
141
142/*
143 * Return the SPSR for the current mode of the virtual CPU.
144 */
Marc Zyngierdb730d82012-10-03 11:17:02 +0100145unsigned long *vcpu_spsr(struct kvm_vcpu *vcpu)
Christoffer Dall749cf76c2013-01-20 18:28:06 -0500146{
Marc Zyngierdb730d82012-10-03 11:17:02 +0100147 unsigned long mode = *vcpu_cpsr(vcpu) & MODE_MASK;
Christoffer Dall749cf76c2013-01-20 18:28:06 -0500148 switch (mode) {
149 case SVC_MODE:
150 return &vcpu->arch.regs.KVM_ARM_SVC_spsr;
151 case ABT_MODE:
152 return &vcpu->arch.regs.KVM_ARM_ABT_spsr;
153 case UND_MODE:
154 return &vcpu->arch.regs.KVM_ARM_UND_spsr;
155 case IRQ_MODE:
156 return &vcpu->arch.regs.KVM_ARM_IRQ_spsr;
157 case FIQ_MODE:
158 return &vcpu->arch.regs.KVM_ARM_FIQ_spsr;
159 default:
160 BUG();
161 }
162}
Christoffer Dall5b3e5e52013-01-20 18:28:09 -0500163
164/**
165 * kvm_handle_wfi - handle a wait-for-interrupts instruction executed by a guest
166 * @vcpu: the vcpu pointer
167 * @run: the kvm_run structure pointer
168 *
169 * Simply sets the wait_for_interrupts flag on the vcpu structure, which will
170 * halt execution of world-switches and schedule other host processes until
171 * there is an incoming IRQ or FIQ to the VM.
172 */
173int kvm_handle_wfi(struct kvm_vcpu *vcpu, struct kvm_run *run)
174{
175 trace_kvm_wfi(*vcpu_pc(vcpu));
176 kvm_vcpu_block(vcpu);
177 return 1;
178}
179
Marc Zyngierc5997562012-12-08 18:13:18 +0000180/*
181 * A conditional instruction is allowed to trap, even though it
182 * wouldn't be executed. So let's re-implement the hardware, in
183 * software!
184 */
185bool kvm_condition_valid(struct kvm_vcpu *vcpu)
186{
187 unsigned long cpsr, cond, insn;
188
189 /*
190 * Exception Code 0 can only happen if we set HCR.TGE to 1, to
191 * catch undefined instructions, and then we won't get past
192 * the arm_exit_handlers test anyway.
193 */
194 BUG_ON(!kvm_vcpu_trap_get_class(vcpu));
195
196 /* Top two bits non-zero? Unconditional. */
197 if (kvm_vcpu_get_hsr(vcpu) >> 30)
198 return true;
199
200 cpsr = *vcpu_cpsr(vcpu);
201
202 /* Is condition field valid? */
203 if ((kvm_vcpu_get_hsr(vcpu) & HSR_CV) >> HSR_CV_SHIFT)
204 cond = (kvm_vcpu_get_hsr(vcpu) & HSR_COND) >> HSR_COND_SHIFT;
205 else {
206 /* This can happen in Thumb mode: examine IT state. */
207 unsigned long it;
208
209 it = ((cpsr >> 8) & 0xFC) | ((cpsr >> 25) & 0x3);
210
211 /* it == 0 => unconditional. */
212 if (it == 0)
213 return true;
214
215 /* The cond for this insn works out as the top 4 bits. */
216 cond = (it >> 4);
217 }
218
219 /* Shift makes it look like an ARM-mode instruction */
220 insn = cond << 28;
221 return arm_check_condition(insn, cpsr) != ARM_OPCODE_CONDTEST_FAIL;
222}
223
Christoffer Dall5b3e5e52013-01-20 18:28:09 -0500224/**
225 * adjust_itstate - adjust ITSTATE when emulating instructions in IT-block
226 * @vcpu: The VCPU pointer
227 *
228 * When exceptions occur while instructions are executed in Thumb IF-THEN
229 * blocks, the ITSTATE field of the CPSR is not advanved (updated), so we have
230 * to do this little bit of work manually. The fields map like this:
231 *
232 * IT[7:0] -> CPSR[26:25],CPSR[15:10]
233 */
234static void kvm_adjust_itstate(struct kvm_vcpu *vcpu)
235{
236 unsigned long itbits, cond;
237 unsigned long cpsr = *vcpu_cpsr(vcpu);
238 bool is_arm = !(cpsr & PSR_T_BIT);
239
240 BUG_ON(is_arm && (cpsr & PSR_IT_MASK));
241
242 if (!(cpsr & PSR_IT_MASK))
243 return;
244
245 cond = (cpsr & 0xe000) >> 13;
246 itbits = (cpsr & 0x1c00) >> (10 - 2);
247 itbits |= (cpsr & (0x3 << 25)) >> 25;
248
249 /* Perform ITAdvance (see page A-52 in ARM DDI 0406C) */
250 if ((itbits & 0x7) == 0)
251 itbits = cond = 0;
252 else
253 itbits = (itbits << 1) & 0x1f;
254
255 cpsr &= ~PSR_IT_MASK;
256 cpsr |= cond << 13;
257 cpsr |= (itbits & 0x1c) << (10 - 2);
258 cpsr |= (itbits & 0x3) << 25;
259 *vcpu_cpsr(vcpu) = cpsr;
260}
261
262/**
263 * kvm_skip_instr - skip a trapped instruction and proceed to the next
264 * @vcpu: The vcpu pointer
265 */
266void kvm_skip_instr(struct kvm_vcpu *vcpu, bool is_wide_instr)
267{
268 bool is_thumb;
269
270 is_thumb = !!(*vcpu_cpsr(vcpu) & PSR_T_BIT);
271 if (is_thumb && !is_wide_instr)
272 *vcpu_pc(vcpu) += 2;
273 else
274 *vcpu_pc(vcpu) += 4;
275 kvm_adjust_itstate(vcpu);
276}
277
278
279/******************************************************************************
280 * Inject exceptions into the guest
281 */
282
283static u32 exc_vector_base(struct kvm_vcpu *vcpu)
284{
285 u32 sctlr = vcpu->arch.cp15[c1_SCTLR];
286 u32 vbar = vcpu->arch.cp15[c12_VBAR];
287
288 if (sctlr & SCTLR_V)
289 return 0xffff0000;
290 else /* always have security exceptions */
291 return vbar;
292}
293
294/**
295 * kvm_inject_undefined - inject an undefined exception into the guest
296 * @vcpu: The VCPU to receive the undefined exception
297 *
298 * It is assumed that this code is called from the VCPU thread and that the
299 * VCPU therefore is not currently executing guest code.
300 *
301 * Modelled after TakeUndefInstrException() pseudocode.
302 */
303void kvm_inject_undefined(struct kvm_vcpu *vcpu)
304{
Marc Zyngierdb730d82012-10-03 11:17:02 +0100305 unsigned long new_lr_value;
306 unsigned long new_spsr_value;
307 unsigned long cpsr = *vcpu_cpsr(vcpu);
Christoffer Dall5b3e5e52013-01-20 18:28:09 -0500308 u32 sctlr = vcpu->arch.cp15[c1_SCTLR];
309 bool is_thumb = (cpsr & PSR_T_BIT);
310 u32 vect_offset = 4;
311 u32 return_offset = (is_thumb) ? 2 : 4;
312
313 new_spsr_value = cpsr;
314 new_lr_value = *vcpu_pc(vcpu) - return_offset;
315
316 *vcpu_cpsr(vcpu) = (cpsr & ~MODE_MASK) | UND_MODE;
317 *vcpu_cpsr(vcpu) |= PSR_I_BIT;
318 *vcpu_cpsr(vcpu) &= ~(PSR_IT_MASK | PSR_J_BIT | PSR_E_BIT | PSR_T_BIT);
319
320 if (sctlr & SCTLR_TE)
321 *vcpu_cpsr(vcpu) |= PSR_T_BIT;
322 if (sctlr & SCTLR_EE)
323 *vcpu_cpsr(vcpu) |= PSR_E_BIT;
324
325 /* Note: These now point to UND banked copies */
326 *vcpu_spsr(vcpu) = cpsr;
327 *vcpu_reg(vcpu, 14) = new_lr_value;
328
329 /* Branch to exception vector */
330 *vcpu_pc(vcpu) = exc_vector_base(vcpu) + vect_offset;
331}
332
333/*
334 * Modelled after TakeDataAbortException() and TakePrefetchAbortException
335 * pseudocode.
336 */
337static void inject_abt(struct kvm_vcpu *vcpu, bool is_pabt, unsigned long addr)
338{
Marc Zyngierdb730d82012-10-03 11:17:02 +0100339 unsigned long new_lr_value;
340 unsigned long new_spsr_value;
341 unsigned long cpsr = *vcpu_cpsr(vcpu);
Christoffer Dall5b3e5e52013-01-20 18:28:09 -0500342 u32 sctlr = vcpu->arch.cp15[c1_SCTLR];
343 bool is_thumb = (cpsr & PSR_T_BIT);
344 u32 vect_offset;
345 u32 return_offset = (is_thumb) ? 4 : 0;
346 bool is_lpae;
347
348 new_spsr_value = cpsr;
349 new_lr_value = *vcpu_pc(vcpu) + return_offset;
350
351 *vcpu_cpsr(vcpu) = (cpsr & ~MODE_MASK) | ABT_MODE;
352 *vcpu_cpsr(vcpu) |= PSR_I_BIT | PSR_A_BIT;
353 *vcpu_cpsr(vcpu) &= ~(PSR_IT_MASK | PSR_J_BIT | PSR_E_BIT | PSR_T_BIT);
354
355 if (sctlr & SCTLR_TE)
356 *vcpu_cpsr(vcpu) |= PSR_T_BIT;
357 if (sctlr & SCTLR_EE)
358 *vcpu_cpsr(vcpu) |= PSR_E_BIT;
359
360 /* Note: These now point to ABT banked copies */
361 *vcpu_spsr(vcpu) = cpsr;
362 *vcpu_reg(vcpu, 14) = new_lr_value;
363
364 if (is_pabt)
365 vect_offset = 12;
366 else
367 vect_offset = 16;
368
369 /* Branch to exception vector */
370 *vcpu_pc(vcpu) = exc_vector_base(vcpu) + vect_offset;
371
372 if (is_pabt) {
373 /* Set DFAR and DFSR */
374 vcpu->arch.cp15[c6_IFAR] = addr;
375 is_lpae = (vcpu->arch.cp15[c2_TTBCR] >> 31);
376 /* Always give debug fault for now - should give guest a clue */
377 if (is_lpae)
378 vcpu->arch.cp15[c5_IFSR] = 1 << 9 | 0x22;
379 else
380 vcpu->arch.cp15[c5_IFSR] = 2;
381 } else { /* !iabt */
382 /* Set DFAR and DFSR */
383 vcpu->arch.cp15[c6_DFAR] = addr;
384 is_lpae = (vcpu->arch.cp15[c2_TTBCR] >> 31);
385 /* Always give debug fault for now - should give guest a clue */
386 if (is_lpae)
387 vcpu->arch.cp15[c5_DFSR] = 1 << 9 | 0x22;
388 else
389 vcpu->arch.cp15[c5_DFSR] = 2;
390 }
391
392}
393
394/**
395 * kvm_inject_dabt - inject a data abort into the guest
396 * @vcpu: The VCPU to receive the undefined exception
397 * @addr: The address to report in the DFAR
398 *
399 * It is assumed that this code is called from the VCPU thread and that the
400 * VCPU therefore is not currently executing guest code.
401 */
402void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr)
403{
404 inject_abt(vcpu, false, addr);
405}
406
407/**
408 * kvm_inject_pabt - inject a prefetch abort into the guest
409 * @vcpu: The VCPU to receive the undefined exception
410 * @addr: The address to report in the DFAR
411 *
412 * It is assumed that this code is called from the VCPU thread and that the
413 * VCPU therefore is not currently executing guest code.
414 */
415void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr)
416{
417 inject_abt(vcpu, true, addr);
418}