Lines Matching refs:svcpu

273 	struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu);  in kvmppc_mmu_next_segment()  local
280 for (i = 0; i < svcpu->slb_max; i++) { in kvmppc_mmu_next_segment()
281 if (!(svcpu->slb[i].esid & SLB_ESID_V)) in kvmppc_mmu_next_segment()
283 else if ((svcpu->slb[i].esid & ESID_MASK) == esid) { in kvmppc_mmu_next_segment()
301 if ((svcpu->slb_max) == max_slb_size) in kvmppc_mmu_next_segment()
304 r = svcpu->slb_max; in kvmppc_mmu_next_segment()
305 svcpu->slb_max++; in kvmppc_mmu_next_segment()
308 svcpu_put(svcpu); in kvmppc_mmu_next_segment()
314 struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu); in kvmppc_mmu_map_segment() local
327 svcpu->slb[slb_index].esid = 0; in kvmppc_mmu_map_segment()
348 svcpu->slb[slb_index].esid = slb_esid; in kvmppc_mmu_map_segment()
349 svcpu->slb[slb_index].vsid = slb_vsid; in kvmppc_mmu_map_segment()
354 svcpu_put(svcpu); in kvmppc_mmu_map_segment()
360 struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu); in kvmppc_mmu_flush_segment() local
364 for (i = 0; i < svcpu->slb_max; i++) { in kvmppc_mmu_flush_segment()
365 if ((svcpu->slb[i].esid & SLB_ESID_V) && in kvmppc_mmu_flush_segment()
366 (svcpu->slb[i].esid & seg_mask) == ea) { in kvmppc_mmu_flush_segment()
368 svcpu->slb[i].esid = 0; in kvmppc_mmu_flush_segment()
372 svcpu_put(svcpu); in kvmppc_mmu_flush_segment()
377 struct kvmppc_book3s_shadow_vcpu *svcpu = svcpu_get(vcpu); in kvmppc_mmu_flush_segments() local
378 svcpu->slb_max = 0; in kvmppc_mmu_flush_segments()
379 svcpu->slb[0].esid = 0; in kvmppc_mmu_flush_segments()
380 svcpu_put(svcpu); in kvmppc_mmu_flush_segments()