1 /*
2  * mmio.c: MMIO emulation components.
3  * Copyright (c) 2004, Intel Corporation.
4  *  Yaozu Dong (Eddie Dong) (Eddie.dong@intel.com)
5  *  Kun Tian (Kevin Tian) (Kevin.tian@intel.com)
6  *
7  * Copyright (c) 2007 Intel Corporation  KVM support.
8  * Xuefei Xu (Anthony Xu) (anthony.xu@intel.com)
9  * Xiantao Zhang  (xiantao.zhang@intel.com)
10  *
11  * This program is free software; you can redistribute it and/or modify it
12  * under the terms and conditions of the GNU General Public License,
13  * version 2, as published by the Free Software Foundation.
14  *
15  * This program is distributed in the hope it will be useful, but WITHOUT
16  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
17  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
18  * more details.
19  *
20  * You should have received a copy of the GNU General Public License along with
21  * this program; if not, write to the Free Software Foundation, Inc., 59 Temple
22  * Place - Suite 330, Boston, MA 02111-1307 USA.
23  *
24  */
25 
26 #include <linux/kvm_host.h>
27 
28 #include "vcpu.h"
29 
vlsapic_write_xtp(struct kvm_vcpu * v,uint8_t val)30 static void vlsapic_write_xtp(struct kvm_vcpu *v, uint8_t val)
31 {
32 	VLSAPIC_XTP(v) = val;
33 }
34 
35 /*
36  * LSAPIC OFFSET
37  */
38 #define PIB_LOW_HALF(ofst)     !(ofst & (1 << 20))
39 #define PIB_OFST_INTA          0x1E0000
40 #define PIB_OFST_XTP           0x1E0008
41 
42 /*
43  * execute write IPI op.
44  */
vlsapic_write_ipi(struct kvm_vcpu * vcpu,uint64_t addr,uint64_t data)45 static void vlsapic_write_ipi(struct kvm_vcpu *vcpu,
46 					uint64_t addr, uint64_t data)
47 {
48 	struct exit_ctl_data *p = &current_vcpu->arch.exit_data;
49 	unsigned long psr;
50 
51 	local_irq_save(psr);
52 
53 	p->exit_reason = EXIT_REASON_IPI;
54 	p->u.ipi_data.addr.val = addr;
55 	p->u.ipi_data.data.val = data;
56 	vmm_transition(current_vcpu);
57 
58 	local_irq_restore(psr);
59 
60 }
61 
lsapic_write(struct kvm_vcpu * v,unsigned long addr,unsigned long length,unsigned long val)62 void lsapic_write(struct kvm_vcpu *v, unsigned long addr,
63 			unsigned long length, unsigned long val)
64 {
65 	addr &= (PIB_SIZE - 1);
66 
67 	switch (addr) {
68 	case PIB_OFST_INTA:
69 		panic_vm(v, "Undefined write on PIB INTA\n");
70 		break;
71 	case PIB_OFST_XTP:
72 		if (length == 1) {
73 			vlsapic_write_xtp(v, val);
74 		} else {
75 			panic_vm(v, "Undefined write on PIB XTP\n");
76 		}
77 		break;
78 	default:
79 		if (PIB_LOW_HALF(addr)) {
80 			/*Lower half */
81 			if (length != 8)
82 				panic_vm(v, "Can't LHF write with size %ld!\n",
83 						length);
84 			else
85 				vlsapic_write_ipi(v, addr, val);
86 		} else {   /*Upper half */
87 			panic_vm(v, "IPI-UHF write %lx\n", addr);
88 		}
89 		break;
90 	}
91 }
92 
lsapic_read(struct kvm_vcpu * v,unsigned long addr,unsigned long length)93 unsigned long lsapic_read(struct kvm_vcpu *v, unsigned long addr,
94 		unsigned long length)
95 {
96 	uint64_t result = 0;
97 
98 	addr &= (PIB_SIZE - 1);
99 
100 	switch (addr) {
101 	case PIB_OFST_INTA:
102 		if (length == 1) /* 1 byte load */
103 			; /* There is no i8259, there is no INTA access*/
104 		else
105 			panic_vm(v, "Undefined read on PIB INTA\n");
106 
107 		break;
108 	case PIB_OFST_XTP:
109 		if (length == 1) {
110 			result = VLSAPIC_XTP(v);
111 		} else {
112 			panic_vm(v, "Undefined read on PIB XTP\n");
113 		}
114 		break;
115 	default:
116 		panic_vm(v, "Undefined addr access for lsapic!\n");
117 		break;
118 	}
119 	return result;
120 }
121 
mmio_access(struct kvm_vcpu * vcpu,u64 src_pa,u64 * dest,u16 s,int ma,int dir)122 static void mmio_access(struct kvm_vcpu *vcpu, u64 src_pa, u64 *dest,
123 					u16 s, int ma, int dir)
124 {
125 	unsigned long iot;
126 	struct exit_ctl_data *p = &vcpu->arch.exit_data;
127 	unsigned long psr;
128 
129 	iot = __gpfn_is_io(src_pa >> PAGE_SHIFT);
130 
131 	local_irq_save(psr);
132 
133 	/*Intercept the access for PIB range*/
134 	if (iot == GPFN_PIB) {
135 		if (!dir)
136 			lsapic_write(vcpu, src_pa, s, *dest);
137 		else
138 			*dest = lsapic_read(vcpu, src_pa, s);
139 		goto out;
140 	}
141 	p->exit_reason = EXIT_REASON_MMIO_INSTRUCTION;
142 	p->u.ioreq.addr = src_pa;
143 	p->u.ioreq.size = s;
144 	p->u.ioreq.dir = dir;
145 	if (dir == IOREQ_WRITE)
146 		p->u.ioreq.data = *dest;
147 	p->u.ioreq.state = STATE_IOREQ_READY;
148 	vmm_transition(vcpu);
149 
150 	if (p->u.ioreq.state == STATE_IORESP_READY) {
151 		if (dir == IOREQ_READ)
152 			/* it's necessary to ensure zero extending */
153 			*dest = p->u.ioreq.data & (~0UL >> (64-(s*8)));
154 	} else
155 		panic_vm(vcpu, "Unhandled mmio access returned!\n");
156 out:
157 	local_irq_restore(psr);
158 	return ;
159 }
160 
161 /*
162    dir 1: read 0:write
163    inst_type 0:integer 1:floating point
164  */
165 #define SL_INTEGER	0	/* store/load interger*/
166 #define SL_FLOATING	1     	/* store/load floating*/
167 
emulate_io_inst(struct kvm_vcpu * vcpu,u64 padr,u64 ma)168 void emulate_io_inst(struct kvm_vcpu *vcpu, u64 padr, u64 ma)
169 {
170 	struct kvm_pt_regs *regs;
171 	IA64_BUNDLE bundle;
172 	int slot, dir = 0;
173 	int inst_type = -1;
174 	u16 size = 0;
175 	u64 data, slot1a, slot1b, temp, update_reg;
176 	s32 imm;
177 	INST64 inst;
178 
179 	regs = vcpu_regs(vcpu);
180 
181 	if (fetch_code(vcpu, regs->cr_iip, &bundle)) {
182 		/* if fetch code fail, return and try again */
183 		return;
184 	}
185 	slot = ((struct ia64_psr *)&(regs->cr_ipsr))->ri;
186 	if (!slot)
187 		inst.inst = bundle.slot0;
188 	else if (slot == 1) {
189 		slot1a = bundle.slot1a;
190 		slot1b = bundle.slot1b;
191 		inst.inst = slot1a + (slot1b << 18);
192 	} else if (slot == 2)
193 		inst.inst = bundle.slot2;
194 
195 	/* Integer Load/Store */
196 	if (inst.M1.major == 4 && inst.M1.m == 0 && inst.M1.x == 0) {
197 		inst_type = SL_INTEGER;
198 		size = (inst.M1.x6 & 0x3);
199 		if ((inst.M1.x6 >> 2) > 0xb) {
200 			/*write*/
201 			dir = IOREQ_WRITE;
202 			data = vcpu_get_gr(vcpu, inst.M4.r2);
203 		} else if ((inst.M1.x6 >> 2) < 0xb) {
204 			/*read*/
205 			dir = IOREQ_READ;
206 		}
207 	} else if (inst.M2.major == 4 && inst.M2.m == 1 && inst.M2.x == 0) {
208 		/* Integer Load + Reg update */
209 		inst_type = SL_INTEGER;
210 		dir = IOREQ_READ;
211 		size = (inst.M2.x6 & 0x3);
212 		temp = vcpu_get_gr(vcpu, inst.M2.r3);
213 		update_reg = vcpu_get_gr(vcpu, inst.M2.r2);
214 		temp += update_reg;
215 		vcpu_set_gr(vcpu, inst.M2.r3, temp, 0);
216 	} else if (inst.M3.major == 5) {
217 		/*Integer Load/Store + Imm update*/
218 		inst_type = SL_INTEGER;
219 		size = (inst.M3.x6&0x3);
220 		if ((inst.M5.x6 >> 2) > 0xb) {
221 			/*write*/
222 			dir = IOREQ_WRITE;
223 			data = vcpu_get_gr(vcpu, inst.M5.r2);
224 			temp = vcpu_get_gr(vcpu, inst.M5.r3);
225 			imm = (inst.M5.s << 31) | (inst.M5.i << 30) |
226 				(inst.M5.imm7 << 23);
227 			temp += imm >> 23;
228 			vcpu_set_gr(vcpu, inst.M5.r3, temp, 0);
229 
230 		} else if ((inst.M3.x6 >> 2) < 0xb) {
231 			/*read*/
232 			dir = IOREQ_READ;
233 			temp = vcpu_get_gr(vcpu, inst.M3.r3);
234 			imm = (inst.M3.s << 31) | (inst.M3.i << 30) |
235 				(inst.M3.imm7 << 23);
236 			temp += imm >> 23;
237 			vcpu_set_gr(vcpu, inst.M3.r3, temp, 0);
238 
239 		}
240 	} else if (inst.M9.major == 6 && inst.M9.x6 == 0x3B
241 				&& inst.M9.m == 0 && inst.M9.x == 0) {
242 		/* Floating-point spill*/
243 		struct ia64_fpreg v;
244 
245 		inst_type = SL_FLOATING;
246 		dir = IOREQ_WRITE;
247 		vcpu_get_fpreg(vcpu, inst.M9.f2, &v);
248 		/* Write high word. FIXME: this is a kludge!  */
249 		v.u.bits[1] &= 0x3ffff;
250 		mmio_access(vcpu, padr + 8, (u64 *)&v.u.bits[1], 8,
251 			    ma, IOREQ_WRITE);
252 		data = v.u.bits[0];
253 		size = 3;
254 	} else if (inst.M10.major == 7 && inst.M10.x6 == 0x3B) {
255 		/* Floating-point spill + Imm update */
256 		struct ia64_fpreg v;
257 
258 		inst_type = SL_FLOATING;
259 		dir = IOREQ_WRITE;
260 		vcpu_get_fpreg(vcpu, inst.M10.f2, &v);
261 		temp = vcpu_get_gr(vcpu, inst.M10.r3);
262 		imm = (inst.M10.s << 31) | (inst.M10.i << 30) |
263 			(inst.M10.imm7 << 23);
264 		temp += imm >> 23;
265 		vcpu_set_gr(vcpu, inst.M10.r3, temp, 0);
266 
267 		/* Write high word.FIXME: this is a kludge!  */
268 		v.u.bits[1] &= 0x3ffff;
269 		mmio_access(vcpu, padr + 8, (u64 *)&v.u.bits[1],
270 			    8, ma, IOREQ_WRITE);
271 		data = v.u.bits[0];
272 		size = 3;
273 	} else if (inst.M10.major == 7 && inst.M10.x6 == 0x31) {
274 		/* Floating-point stf8 + Imm update */
275 		struct ia64_fpreg v;
276 		inst_type = SL_FLOATING;
277 		dir = IOREQ_WRITE;
278 		size = 3;
279 		vcpu_get_fpreg(vcpu, inst.M10.f2, &v);
280 		data = v.u.bits[0]; /* Significand.  */
281 		temp = vcpu_get_gr(vcpu, inst.M10.r3);
282 		imm = (inst.M10.s << 31) | (inst.M10.i << 30) |
283 			(inst.M10.imm7 << 23);
284 		temp += imm >> 23;
285 		vcpu_set_gr(vcpu, inst.M10.r3, temp, 0);
286 	} else if (inst.M15.major == 7 && inst.M15.x6 >= 0x2c
287 			&& inst.M15.x6 <= 0x2f) {
288 		temp = vcpu_get_gr(vcpu, inst.M15.r3);
289 		imm = (inst.M15.s << 31) | (inst.M15.i << 30) |
290 			(inst.M15.imm7 << 23);
291 		temp += imm >> 23;
292 		vcpu_set_gr(vcpu, inst.M15.r3, temp, 0);
293 
294 		vcpu_increment_iip(vcpu);
295 		return;
296 	} else if (inst.M12.major == 6 && inst.M12.m == 1
297 			&& inst.M12.x == 1 && inst.M12.x6 == 1) {
298 		/* Floating-point Load Pair + Imm ldfp8 M12*/
299 		struct ia64_fpreg v;
300 
301 		inst_type = SL_FLOATING;
302 		dir = IOREQ_READ;
303 		size = 8;     /*ldfd*/
304 		mmio_access(vcpu, padr, &data, size, ma, dir);
305 		v.u.bits[0] = data;
306 		v.u.bits[1] = 0x1003E;
307 		vcpu_set_fpreg(vcpu, inst.M12.f1, &v);
308 		padr += 8;
309 		mmio_access(vcpu, padr, &data, size, ma, dir);
310 		v.u.bits[0] = data;
311 		v.u.bits[1] = 0x1003E;
312 		vcpu_set_fpreg(vcpu, inst.M12.f2, &v);
313 		padr += 8;
314 		vcpu_set_gr(vcpu, inst.M12.r3, padr, 0);
315 		vcpu_increment_iip(vcpu);
316 		return;
317 	} else {
318 		inst_type = -1;
319 		panic_vm(vcpu, "Unsupported MMIO access instruction! "
320 				"Bunld[0]=0x%lx, Bundle[1]=0x%lx\n",
321 				bundle.i64[0], bundle.i64[1]);
322 	}
323 
324 	size = 1 << size;
325 	if (dir == IOREQ_WRITE) {
326 		mmio_access(vcpu, padr, &data, size, ma, dir);
327 	} else {
328 		mmio_access(vcpu, padr, &data, size, ma, dir);
329 		if (inst_type == SL_INTEGER)
330 			vcpu_set_gr(vcpu, inst.M1.r1, data, 0);
331 		else
332 			panic_vm(vcpu, "Unsupported instruction type!\n");
333 
334 	}
335 	vcpu_increment_iip(vcpu);
336 }
337