Merge branch 'chandan/prep-subpage-blocksize' into for-chris-4.6
[deliverable/linux.git] / arch / powerpc / kvm / e500.h
1 /*
2 * Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
3 *
4 * Author: Yu Liu <yu.liu@freescale.com>
5 * Scott Wood <scottwood@freescale.com>
6 * Ashish Kalra <ashish.kalra@freescale.com>
7 * Varun Sethi <varun.sethi@freescale.com>
8 *
9 * Description:
10 * This file is based on arch/powerpc/kvm/44x_tlb.h and
11 * arch/powerpc/include/asm/kvm_44x.h by Hollis Blanchard <hollisb@us.ibm.com>,
12 * Copyright IBM Corp. 2007-2008
13 *
14 * This program is free software; you can redistribute it and/or modify
15 * it under the terms of the GNU General Public License, version 2, as
16 * published by the Free Software Foundation.
17 */
18
19 #ifndef KVM_E500_H
20 #define KVM_E500_H
21
22 #include <linux/kvm_host.h>
23 #include <asm/mmu-book3e.h>
24 #include <asm/tlb.h>
25 #include <asm/cputhreads.h>
26
27 enum vcpu_ftr {
28 VCPU_FTR_MMU_V2
29 };
30
31 #define E500_PID_NUM 3
32 #define E500_TLB_NUM 2
33
34 /* entry is mapped somewhere in host TLB */
35 #define E500_TLB_VALID (1 << 31)
36 /* TLB1 entry is mapped by host TLB1, tracked by bitmaps */
37 #define E500_TLB_BITMAP (1 << 30)
38 /* TLB1 entry is mapped by host TLB0 */
39 #define E500_TLB_TLB0 (1 << 29)
40 /* bits [6-5] MAS2_X1 and MAS2_X0 and [4-0] bits for WIMGE */
41 #define E500_TLB_MAS2_ATTR (0x7f)
42
43 struct tlbe_ref {
44 kvm_pfn_t pfn; /* valid only for TLB0, except briefly */
45 unsigned int flags; /* E500_TLB_* */
46 };
47
48 struct tlbe_priv {
49 struct tlbe_ref ref;
50 };
51
52 #ifdef CONFIG_KVM_E500V2
53 struct vcpu_id_table;
54 #endif
55
56 struct kvmppc_e500_tlb_params {
57 int entries, ways, sets;
58 };
59
60 struct kvmppc_vcpu_e500 {
61 struct kvm_vcpu vcpu;
62
63 /* Unmodified copy of the guest's TLB -- shared with host userspace. */
64 struct kvm_book3e_206_tlb_entry *gtlb_arch;
65
66 /* Starting entry number in gtlb_arch[] */
67 int gtlb_offset[E500_TLB_NUM];
68
69 /* KVM internal information associated with each guest TLB entry */
70 struct tlbe_priv *gtlb_priv[E500_TLB_NUM];
71
72 struct kvmppc_e500_tlb_params gtlb_params[E500_TLB_NUM];
73
74 unsigned int gtlb_nv[E500_TLB_NUM];
75
76 unsigned int host_tlb1_nv;
77
78 u32 svr;
79 u32 l1csr0;
80 u32 l1csr1;
81 u32 hid0;
82 u32 hid1;
83 u64 mcar;
84
85 struct page **shared_tlb_pages;
86 int num_shared_tlb_pages;
87
88 u64 *g2h_tlb1_map;
89 unsigned int *h2g_tlb1_rmap;
90
91 /* Minimum and maximum address mapped my TLB1 */
92 unsigned long tlb1_min_eaddr;
93 unsigned long tlb1_max_eaddr;
94
95 #ifdef CONFIG_KVM_E500V2
96 u32 pid[E500_PID_NUM];
97
98 /* vcpu id table */
99 struct vcpu_id_table *idt;
100 #endif
101 };
102
103 static inline struct kvmppc_vcpu_e500 *to_e500(struct kvm_vcpu *vcpu)
104 {
105 return container_of(vcpu, struct kvmppc_vcpu_e500, vcpu);
106 }
107
108
109 /* This geometry is the legacy default -- can be overridden by userspace */
110 #define KVM_E500_TLB0_WAY_SIZE 128
111 #define KVM_E500_TLB0_WAY_NUM 2
112
113 #define KVM_E500_TLB0_SIZE (KVM_E500_TLB0_WAY_SIZE * KVM_E500_TLB0_WAY_NUM)
114 #define KVM_E500_TLB1_SIZE 16
115
116 #define index_of(tlbsel, esel) (((tlbsel) << 16) | ((esel) & 0xFFFF))
117 #define tlbsel_of(index) ((index) >> 16)
118 #define esel_of(index) ((index) & 0xFFFF)
119
120 #define E500_TLB_USER_PERM_MASK (MAS3_UX|MAS3_UR|MAS3_UW)
121 #define E500_TLB_SUPER_PERM_MASK (MAS3_SX|MAS3_SR|MAS3_SW)
122 #define MAS2_ATTRIB_MASK \
123 (MAS2_X0 | MAS2_X1 | MAS2_E | MAS2_G)
124 #define MAS3_ATTRIB_MASK \
125 (MAS3_U0 | MAS3_U1 | MAS3_U2 | MAS3_U3 \
126 | E500_TLB_USER_PERM_MASK | E500_TLB_SUPER_PERM_MASK)
127
128 int kvmppc_e500_emul_mt_mmucsr0(struct kvmppc_vcpu_e500 *vcpu_e500,
129 ulong value);
130 int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu);
131 int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu);
132 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea);
133 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea);
134 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea);
135 int kvmppc_e500_tlb_init(struct kvmppc_vcpu_e500 *vcpu_e500);
136 void kvmppc_e500_tlb_uninit(struct kvmppc_vcpu_e500 *vcpu_e500);
137
138 void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
139 int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
140
141 int kvmppc_get_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
142 union kvmppc_one_reg *val);
143 int kvmppc_set_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
144 union kvmppc_one_reg *val);
145
146 #ifdef CONFIG_KVM_E500V2
147 unsigned int kvmppc_e500_get_sid(struct kvmppc_vcpu_e500 *vcpu_e500,
148 unsigned int as, unsigned int gid,
149 unsigned int pr, int avoid_recursion);
150 #endif
151
152 /* TLB helper functions */
153 static inline unsigned int
154 get_tlb_size(const struct kvm_book3e_206_tlb_entry *tlbe)
155 {
156 return (tlbe->mas1 >> 7) & 0x1f;
157 }
158
159 static inline gva_t get_tlb_eaddr(const struct kvm_book3e_206_tlb_entry *tlbe)
160 {
161 return tlbe->mas2 & MAS2_EPN;
162 }
163
164 static inline u64 get_tlb_bytes(const struct kvm_book3e_206_tlb_entry *tlbe)
165 {
166 unsigned int pgsize = get_tlb_size(tlbe);
167 return 1ULL << 10 << pgsize;
168 }
169
170 static inline gva_t get_tlb_end(const struct kvm_book3e_206_tlb_entry *tlbe)
171 {
172 u64 bytes = get_tlb_bytes(tlbe);
173 return get_tlb_eaddr(tlbe) + bytes - 1;
174 }
175
176 static inline u64 get_tlb_raddr(const struct kvm_book3e_206_tlb_entry *tlbe)
177 {
178 return tlbe->mas7_3 & ~0xfffULL;
179 }
180
181 static inline unsigned int
182 get_tlb_tid(const struct kvm_book3e_206_tlb_entry *tlbe)
183 {
184 return (tlbe->mas1 >> 16) & 0xff;
185 }
186
187 static inline unsigned int
188 get_tlb_ts(const struct kvm_book3e_206_tlb_entry *tlbe)
189 {
190 return (tlbe->mas1 >> 12) & 0x1;
191 }
192
193 static inline unsigned int
194 get_tlb_v(const struct kvm_book3e_206_tlb_entry *tlbe)
195 {
196 return (tlbe->mas1 >> 31) & 0x1;
197 }
198
199 static inline unsigned int
200 get_tlb_iprot(const struct kvm_book3e_206_tlb_entry *tlbe)
201 {
202 return (tlbe->mas1 >> 30) & 0x1;
203 }
204
205 static inline unsigned int
206 get_tlb_tsize(const struct kvm_book3e_206_tlb_entry *tlbe)
207 {
208 return (tlbe->mas1 & MAS1_TSIZE_MASK) >> MAS1_TSIZE_SHIFT;
209 }
210
211 static inline unsigned int get_cur_pid(struct kvm_vcpu *vcpu)
212 {
213 return vcpu->arch.pid & 0xff;
214 }
215
216 static inline unsigned int get_cur_as(struct kvm_vcpu *vcpu)
217 {
218 return !!(vcpu->arch.shared->msr & (MSR_IS | MSR_DS));
219 }
220
221 static inline unsigned int get_cur_pr(struct kvm_vcpu *vcpu)
222 {
223 return !!(vcpu->arch.shared->msr & MSR_PR);
224 }
225
226 static inline unsigned int get_cur_spid(const struct kvm_vcpu *vcpu)
227 {
228 return (vcpu->arch.shared->mas6 >> 16) & 0xff;
229 }
230
231 static inline unsigned int get_cur_sas(const struct kvm_vcpu *vcpu)
232 {
233 return vcpu->arch.shared->mas6 & 0x1;
234 }
235
236 static inline unsigned int get_tlb_tlbsel(const struct kvm_vcpu *vcpu)
237 {
238 /*
239 * Manual says that tlbsel has 2 bits wide.
240 * Since we only have two TLBs, only lower bit is used.
241 */
242 return (vcpu->arch.shared->mas0 >> 28) & 0x1;
243 }
244
245 static inline unsigned int get_tlb_nv_bit(const struct kvm_vcpu *vcpu)
246 {
247 return vcpu->arch.shared->mas0 & 0xfff;
248 }
249
250 static inline unsigned int get_tlb_esel_bit(const struct kvm_vcpu *vcpu)
251 {
252 return (vcpu->arch.shared->mas0 >> 16) & 0xfff;
253 }
254
255 static inline int tlbe_is_host_safe(const struct kvm_vcpu *vcpu,
256 const struct kvm_book3e_206_tlb_entry *tlbe)
257 {
258 gpa_t gpa;
259
260 if (!get_tlb_v(tlbe))
261 return 0;
262
263 #ifndef CONFIG_KVM_BOOKE_HV
264 /* Does it match current guest AS? */
265 /* XXX what about IS != DS? */
266 if (get_tlb_ts(tlbe) != !!(vcpu->arch.shared->msr & MSR_IS))
267 return 0;
268 #endif
269
270 gpa = get_tlb_raddr(tlbe);
271 if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT))
272 /* Mapping is not for RAM. */
273 return 0;
274
275 return 1;
276 }
277
278 static inline struct kvm_book3e_206_tlb_entry *get_entry(
279 struct kvmppc_vcpu_e500 *vcpu_e500, int tlbsel, int entry)
280 {
281 int offset = vcpu_e500->gtlb_offset[tlbsel];
282 return &vcpu_e500->gtlb_arch[offset + entry];
283 }
284
285 void kvmppc_e500_tlbil_one(struct kvmppc_vcpu_e500 *vcpu_e500,
286 struct kvm_book3e_206_tlb_entry *gtlbe);
287 void kvmppc_e500_tlbil_all(struct kvmppc_vcpu_e500 *vcpu_e500);
288
289 #ifdef CONFIG_KVM_BOOKE_HV
290 #define kvmppc_e500_get_tlb_stid(vcpu, gtlbe) get_tlb_tid(gtlbe)
291 #define get_tlbmiss_tid(vcpu) get_cur_pid(vcpu)
292 #define get_tlb_sts(gtlbe) (gtlbe->mas1 & MAS1_TS)
293
294 /*
295 * These functions should be called with preemption disabled
296 * and the returned value is valid only in that context
297 */
298 static inline int get_thread_specific_lpid(int vm_lpid)
299 {
300 int vcpu_lpid = vm_lpid;
301
302 if (threads_per_core == 2)
303 vcpu_lpid |= smp_processor_id() & 1;
304
305 return vcpu_lpid;
306 }
307
308 static inline int get_lpid(struct kvm_vcpu *vcpu)
309 {
310 return get_thread_specific_lpid(vcpu->kvm->arch.lpid);
311 }
312 #else
313 unsigned int kvmppc_e500_get_tlb_stid(struct kvm_vcpu *vcpu,
314 struct kvm_book3e_206_tlb_entry *gtlbe);
315
316 static inline unsigned int get_tlbmiss_tid(struct kvm_vcpu *vcpu)
317 {
318 struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
319 unsigned int tidseld = (vcpu->arch.shared->mas4 >> 16) & 0xf;
320
321 return vcpu_e500->pid[tidseld];
322 }
323
324 /* Force TS=1 for all guest mappings. */
325 #define get_tlb_sts(gtlbe) (MAS1_TS)
326 #endif /* !BOOKE_HV */
327
328 static inline bool has_feature(const struct kvm_vcpu *vcpu,
329 enum vcpu_ftr ftr)
330 {
331 bool has_ftr;
332 switch (ftr) {
333 case VCPU_FTR_MMU_V2:
334 has_ftr = ((vcpu->arch.mmucfg & MMUCFG_MAVN) == MMUCFG_MAVN_V2);
335 break;
336 default:
337 return false;
338 }
339 return has_ftr;
340 }
341
342 #endif /* KVM_E500_H */
This page took 0.052545 seconds and 6 git commands to generate.