1/*
2 * This program is free software; you can redistribute it and/or modify
3 * it under the terms of the GNU General Public License, version 2, as
4 * published by the Free Software Foundation.
5 *
6 * This program is distributed in the hope that it will be useful,
7 * but WITHOUT ANY WARRANTY; without even the implied warranty of
8 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
9 * GNU General Public License for more details.
10 *
11 * You should have received a copy of the GNU General Public License
12 * along with this program; if not, write to the Free Software
13 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
14 *
15 * Copyright SUSE Linux Products GmbH 2009
16 *
17 * Authors: Alexander Graf <agraf@suse.de>
18 */
19
20#define SHADOW_SLB_ESID(num)	(SLBSHADOW_SAVEAREA + (num * 0x10))
21#define SHADOW_SLB_VSID(num)	(SLBSHADOW_SAVEAREA + (num * 0x10) + 0x8)
22#define UNBOLT_SLB_ENTRY(num) \
23	ld	r9, SHADOW_SLB_ESID(num)(r12); \
24	/* Invalid? Skip. */; \
25	rldicl. r0, r9, 37, 63; \
26	beq	slb_entry_skip_ ## num; \
27	xoris	r9, r9, SLB_ESID_V@h; \
28	std	r9, SHADOW_SLB_ESID(num)(r12); \
29  slb_entry_skip_ ## num:
30
31#define REBOLT_SLB_ENTRY(num) \
32	ld	r10, SHADOW_SLB_ESID(num)(r11); \
33	cmpdi	r10, 0; \
34	beq	slb_exit_skip_ ## num; \
35	oris	r10, r10, SLB_ESID_V@h; \
36	ld	r9, SHADOW_SLB_VSID(num)(r11); \
37	slbmte	r9, r10; \
38	std	r10, SHADOW_SLB_ESID(num)(r11); \
39slb_exit_skip_ ## num:
40
41/******************************************************************************
42 *                                                                            *
43 *                               Entry code                                   *
44 *                                                                            *
45 *****************************************************************************/
46
47.macro LOAD_GUEST_SEGMENTS
48
49	/* Required state:
50	 *
51	 * MSR = ~IR|DR
52	 * R13 = PACA
53	 * R1 = host R1
54	 * R2 = host R2
55	 * R3 = shadow vcpu
56	 * all other volatile GPRS = free except R4, R6
57	 * SVCPU[CR]  = guest CR
58	 * SVCPU[XER] = guest XER
59	 * SVCPU[CTR] = guest CTR
60	 * SVCPU[LR]  = guest LR
61	 */
62
63	/* Remove LPAR shadow entries */
64
65#if SLB_NUM_BOLTED == 3
66
67	ld	r12, PACA_SLBSHADOWPTR(r13)
68
69	/* Save off the first entry so we can slbie it later */
70	ld	r10, SHADOW_SLB_ESID(0)(r12)
71	ld	r11, SHADOW_SLB_VSID(0)(r12)
72
73	/* Remove bolted entries */
74	UNBOLT_SLB_ENTRY(0)
75	UNBOLT_SLB_ENTRY(1)
76	UNBOLT_SLB_ENTRY(2)
77
78#else
79#error unknown number of bolted entries
80#endif
81
82	/* Flush SLB */
83
84	slbia
85
86	/* r0 = esid & ESID_MASK */
87	rldicr  r10, r10, 0, 35
88	/* r0 |= CLASS_BIT(VSID) */
89	rldic   r12, r11, 56 - 36, 36
90	or      r10, r10, r12
91	slbie	r10
92
93	isync
94
95	/* Fill SLB with our shadow */
96
97	lbz	r12, SVCPU_SLB_MAX(r3)
98	mulli	r12, r12, 16
99	addi	r12, r12, SVCPU_SLB
100	add	r12, r12, r3
101
102	/* for (r11 = kvm_slb; r11 < kvm_slb + kvm_slb_size; r11+=slb_entry) */
103	li	r11, SVCPU_SLB
104	add	r11, r11, r3
105
106slb_loop_enter:
107
108	ld	r10, 0(r11)
109
110	rldicl. r0, r10, 37, 63
111	beq	slb_loop_enter_skip
112
113	ld	r9, 8(r11)
114	slbmte	r9, r10
115
116slb_loop_enter_skip:
117	addi	r11, r11, 16
118	cmpd	cr0, r11, r12
119	blt	slb_loop_enter
120
121slb_do_enter:
122
123.endm
124
125/******************************************************************************
126 *                                                                            *
127 *                               Exit code                                    *
128 *                                                                            *
129 *****************************************************************************/
130
131.macro LOAD_HOST_SEGMENTS
132
133	/* Register usage at this point:
134	 *
135	 * R1         = host R1
136	 * R2         = host R2
137	 * R12        = exit handler id
138	 * R13        = shadow vcpu - SHADOW_VCPU_OFF [=PACA on PPC64]
139	 * SVCPU.*    = guest *
140	 * SVCPU[CR]  = guest CR
141	 * SVCPU[XER] = guest XER
142	 * SVCPU[CTR] = guest CTR
143	 * SVCPU[LR]  = guest LR
144	 *
145	 */
146
147	/* Restore bolted entries from the shadow and fix it along the way */
148
149	/* We don't store anything in entry 0, so we don't need to take care of it */
150	slbia
151	isync
152
153#if SLB_NUM_BOLTED == 3
154
155	ld	r11, PACA_SLBSHADOWPTR(r13)
156
157	REBOLT_SLB_ENTRY(0)
158	REBOLT_SLB_ENTRY(1)
159	REBOLT_SLB_ENTRY(2)
160
161#else
162#error unknown number of bolted entries
163#endif
164
165slb_do_exit:
166
167.endm
168