KVM: VMX: handle IO when emulation is due to #GP in real mode.
[deliverable/linux.git] / arch / s390 / kvm / intercept.c
CommitLineData
8f2abe6a 1/*
a53c8fab 2 * in-kernel handling for sie intercepts
8f2abe6a 3 *
a53c8fab 4 * Copyright IBM Corp. 2008, 2009
8f2abe6a
CB
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License (version 2 only)
8 * as published by the Free Software Foundation.
9 *
10 * Author(s): Carsten Otte <cotte@de.ibm.com>
11 * Christian Borntraeger <borntraeger@de.ibm.com>
12 */
13
14#include <linux/kvm_host.h>
15#include <linux/errno.h>
16#include <linux/pagemap.h>
17
18#include <asm/kvm_host.h>
19
20#include "kvm-s390.h"
ba5c1e9b 21#include "gaccess.h"
5786fffa 22#include "trace.h"
ade38c31 23#include "trace-s390.h"
ba5c1e9b 24
f5e10b09 25static int handle_lctlg(struct kvm_vcpu *vcpu)
ba5c1e9b
CO
26{
27 int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
28 int reg3 = vcpu->arch.sie_block->ipa & 0x000f;
29 int base2 = vcpu->arch.sie_block->ipb >> 28;
30 int disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16) +
31 ((vcpu->arch.sie_block->ipb & 0xff00) << 4);
32 u64 useraddr;
33 int reg, rc;
34
f5e10b09 35 vcpu->stat.instruction_lctlg++;
ba5c1e9b 36 if ((vcpu->arch.sie_block->ipb & 0xff) != 0x2f)
b8e660b8 37 return -EOPNOTSUPP;
ba5c1e9b
CO
38
39 useraddr = disp2;
40 if (base2)
5a32c1af 41 useraddr += vcpu->run->s.regs.gprs[base2];
ba5c1e9b 42
5a00a5e7
CB
43 if (useraddr & 7)
44 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION);
45
ba5c1e9b
CO
46 reg = reg1;
47
f5e10b09 48 VCPU_EVENT(vcpu, 5, "lctlg r1:%x, r3:%x,b2:%x,d2:%x", reg1, reg3, base2,
ba5c1e9b 49 disp2);
5786fffa 50 trace_kvm_s390_handle_lctl(vcpu, 1, reg1, reg3, useraddr);
ba5c1e9b
CO
51
52 do {
53 rc = get_guest_u64(vcpu, useraddr,
54 &vcpu->arch.sie_block->gcr[reg]);
55 if (rc == -EFAULT) {
56 kvm_s390_inject_program_int(vcpu, PGM_ADDRESSING);
57 break;
58 }
59 useraddr += 8;
60 if (reg == reg3)
61 break;
62 reg = (reg + 1) % 16;
63 } while (1);
64 return 0;
65}
66
67static int handle_lctl(struct kvm_vcpu *vcpu)
68{
69 int reg1 = (vcpu->arch.sie_block->ipa & 0x00f0) >> 4;
70 int reg3 = vcpu->arch.sie_block->ipa & 0x000f;
71 int base2 = vcpu->arch.sie_block->ipb >> 28;
72 int disp2 = ((vcpu->arch.sie_block->ipb & 0x0fff0000) >> 16);
73 u64 useraddr;
74 u32 val = 0;
75 int reg, rc;
76
77 vcpu->stat.instruction_lctl++;
78
79 useraddr = disp2;
80 if (base2)
5a32c1af 81 useraddr += vcpu->run->s.regs.gprs[base2];
ba5c1e9b 82
5a00a5e7
CB
83 if (useraddr & 3)
84 return kvm_s390_inject_program_int(vcpu, PGM_SPECIFICATION);
85
ba5c1e9b
CO
86 VCPU_EVENT(vcpu, 5, "lctl r1:%x, r3:%x,b2:%x,d2:%x", reg1, reg3, base2,
87 disp2);
5786fffa 88 trace_kvm_s390_handle_lctl(vcpu, 0, reg1, reg3, useraddr);
ba5c1e9b
CO
89
90 reg = reg1;
91 do {
92 rc = get_guest_u32(vcpu, useraddr, &val);
93 if (rc == -EFAULT) {
94 kvm_s390_inject_program_int(vcpu, PGM_ADDRESSING);
95 break;
96 }
97 vcpu->arch.sie_block->gcr[reg] &= 0xffffffff00000000ul;
98 vcpu->arch.sie_block->gcr[reg] |= val;
99 useraddr += 4;
100 if (reg == reg3)
101 break;
102 reg = (reg + 1) % 16;
103 } while (1);
104 return 0;
105}
106
107static intercept_handler_t instruction_handlers[256] = {
8c3f61e2 108 [0x01] = kvm_s390_handle_01,
e28acfea 109 [0x83] = kvm_s390_handle_diag,
5288fbf0 110 [0xae] = kvm_s390_handle_sigp,
70455a36 111 [0xb2] = kvm_s390_handle_b2,
ba5c1e9b 112 [0xb7] = handle_lctl,
bb25b9ba 113 [0xe5] = kvm_s390_handle_e5,
f5e10b09 114 [0xeb] = handle_lctlg,
ba5c1e9b 115};
8f2abe6a
CB
116
117static int handle_noop(struct kvm_vcpu *vcpu)
118{
119 switch (vcpu->arch.sie_block->icptcode) {
0eaeafa1
CB
120 case 0x0:
121 vcpu->stat.exit_null++;
122 break;
8f2abe6a
CB
123 case 0x10:
124 vcpu->stat.exit_external_request++;
125 break;
126 case 0x14:
127 vcpu->stat.exit_external_interrupt++;
128 break;
129 default:
130 break; /* nothing */
131 }
132 return 0;
133}
134
135static int handle_stop(struct kvm_vcpu *vcpu)
136{
9ace903d 137 int rc = 0;
5288fbf0 138
8f2abe6a 139 vcpu->stat.exit_stop_request++;
5288fbf0 140 spin_lock_bh(&vcpu->arch.local_int.lock);
5288fbf0 141
ade38c31
CH
142 trace_kvm_s390_stop_request(vcpu->arch.local_int.action_bits);
143
9ace903d
CE
144 if (vcpu->arch.local_int.action_bits & ACTION_RELOADVCPU_ON_STOP) {
145 vcpu->arch.local_int.action_bits &= ~ACTION_RELOADVCPU_ON_STOP;
146 rc = SIE_INTERCEPT_RERUNVCPU;
147 vcpu->run->exit_reason = KVM_EXIT_INTR;
148 }
149
5288fbf0 150 if (vcpu->arch.local_int.action_bits & ACTION_STOP_ON_STOP) {
9e6dabef
CH
151 atomic_set_mask(CPUSTAT_STOPPED,
152 &vcpu->arch.sie_block->cpuflags);
5288fbf0
CB
153 vcpu->arch.local_int.action_bits &= ~ACTION_STOP_ON_STOP;
154 VCPU_EVENT(vcpu, 3, "%s", "cpu stopped");
b8e660b8 155 rc = -EOPNOTSUPP;
9ace903d
CE
156 }
157
9e0d5473
JF
158 if (vcpu->arch.local_int.action_bits & ACTION_STORE_ON_STOP) {
159 vcpu->arch.local_int.action_bits &= ~ACTION_STORE_ON_STOP;
160 /* store status must be called unlocked. Since local_int.lock
161 * only protects local_int.* and not guest memory we can give
162 * up the lock here */
163 spin_unlock_bh(&vcpu->arch.local_int.lock);
164 rc = kvm_s390_vcpu_store_status(vcpu,
165 KVM_S390_STORE_STATUS_NOADDR);
166 if (rc >= 0)
167 rc = -EOPNOTSUPP;
168 } else
169 spin_unlock_bh(&vcpu->arch.local_int.lock);
5288fbf0 170 return rc;
8f2abe6a
CB
171}
172
173static int handle_validity(struct kvm_vcpu *vcpu)
174{
598841ca 175 unsigned long vmaddr;
8f2abe6a 176 int viwhy = vcpu->arch.sie_block->ipb >> 16;
3edbcff9
CO
177 int rc;
178
8f2abe6a 179 vcpu->stat.exit_validity++;
5786fffa 180 trace_kvm_s390_intercept_validity(vcpu, viwhy);
092670cd
CO
181 if (viwhy == 0x37) {
182 vmaddr = gmap_fault(vcpu->arch.sie_block->prefix,
183 vcpu->arch.gmap);
184 if (IS_ERR_VALUE(vmaddr)) {
185 rc = -EOPNOTSUPP;
186 goto out;
187 }
188 rc = fault_in_pages_writeable((char __user *) vmaddr,
189 PAGE_SIZE);
598841ca 190 if (rc) {
3edbcff9 191 /* user will receive sigsegv, exit to user */
b8e660b8 192 rc = -EOPNOTSUPP;
598841ca
CO
193 goto out;
194 }
092670cd 195 vmaddr = gmap_fault(vcpu->arch.sie_block->prefix + PAGE_SIZE,
598841ca
CO
196 vcpu->arch.gmap);
197 if (IS_ERR_VALUE(vmaddr)) {
198 rc = -EOPNOTSUPP;
199 goto out;
200 }
092670cd
CO
201 rc = fault_in_pages_writeable((char __user *) vmaddr,
202 PAGE_SIZE);
203 if (rc) {
204 /* user will receive sigsegv, exit to user */
598841ca
CO
205 rc = -EOPNOTSUPP;
206 goto out;
207 }
3edbcff9 208 } else
b8e660b8 209 rc = -EOPNOTSUPP;
3edbcff9 210
598841ca 211out:
3edbcff9
CO
212 if (rc)
213 VCPU_EVENT(vcpu, 2, "unhandled validity intercept code %d",
214 viwhy);
215 return rc;
8f2abe6a
CB
216}
217
ba5c1e9b
CO
218static int handle_instruction(struct kvm_vcpu *vcpu)
219{
220 intercept_handler_t handler;
221
222 vcpu->stat.exit_instruction++;
5786fffa
CH
223 trace_kvm_s390_intercept_instruction(vcpu,
224 vcpu->arch.sie_block->ipa,
225 vcpu->arch.sie_block->ipb);
ba5c1e9b
CO
226 handler = instruction_handlers[vcpu->arch.sie_block->ipa >> 8];
227 if (handler)
228 return handler(vcpu);
b8e660b8 229 return -EOPNOTSUPP;
ba5c1e9b
CO
230}
231
232static int handle_prog(struct kvm_vcpu *vcpu)
233{
234 vcpu->stat.exit_program_interruption++;
5786fffa 235 trace_kvm_s390_intercept_prog(vcpu, vcpu->arch.sie_block->iprcc);
ba5c1e9b
CO
236 return kvm_s390_inject_program_int(vcpu, vcpu->arch.sie_block->iprcc);
237}
238
239static int handle_instruction_and_prog(struct kvm_vcpu *vcpu)
240{
241 int rc, rc2;
242
243 vcpu->stat.exit_instr_and_program++;
244 rc = handle_instruction(vcpu);
245 rc2 = handle_prog(vcpu);
246
b8e660b8 247 if (rc == -EOPNOTSUPP)
ba5c1e9b
CO
248 vcpu->arch.sie_block->icptcode = 0x04;
249 if (rc)
250 return rc;
251 return rc2;
252}
253
062d5e9b 254static const intercept_handler_t intercept_funcs[] = {
8f2abe6a 255 [0x00 >> 2] = handle_noop,
ba5c1e9b
CO
256 [0x04 >> 2] = handle_instruction,
257 [0x08 >> 2] = handle_prog,
258 [0x0C >> 2] = handle_instruction_and_prog,
8f2abe6a
CB
259 [0x10 >> 2] = handle_noop,
260 [0x14 >> 2] = handle_noop,
ba5c1e9b 261 [0x1C >> 2] = kvm_s390_handle_wait,
8f2abe6a
CB
262 [0x20 >> 2] = handle_validity,
263 [0x28 >> 2] = handle_stop,
264};
265
266int kvm_handle_sie_intercept(struct kvm_vcpu *vcpu)
267{
268 intercept_handler_t func;
269 u8 code = vcpu->arch.sie_block->icptcode;
270
062d5e9b 271 if (code & 3 || (code >> 2) >= ARRAY_SIZE(intercept_funcs))
b8e660b8 272 return -EOPNOTSUPP;
8f2abe6a
CB
273 func = intercept_funcs[code >> 2];
274 if (func)
275 return func(vcpu);
b8e660b8 276 return -EOPNOTSUPP;
8f2abe6a 277}
This page took 0.384712 seconds and 5 git commands to generate.