Kyle Swenson | 8d8f654 | 2021-03-15 11:02:55 -0600 | [diff] [blame^] | 1 | /* |
| 2 | * Copyright (C) 2010,2012 Freescale Semiconductor, Inc. All rights reserved. |
| 3 | * |
| 4 | * Author: Varun Sethi, <varun.sethi@freescale.com> |
| 5 | * |
| 6 | * Description: |
| 7 | * This file is derived from arch/powerpc/kvm/e500.c, |
| 8 | * by Yu Liu <yu.liu@freescale.com>. |
| 9 | * |
| 10 | * This program is free software; you can redistribute it and/or modify |
| 11 | * it under the terms of the GNU General Public License, version 2, as |
| 12 | * published by the Free Software Foundation. |
| 13 | */ |
| 14 | |
| 15 | #include <linux/kvm_host.h> |
| 16 | #include <linux/slab.h> |
| 17 | #include <linux/err.h> |
| 18 | #include <linux/export.h> |
| 19 | #include <linux/miscdevice.h> |
| 20 | #include <linux/module.h> |
| 21 | |
| 22 | #include <asm/reg.h> |
| 23 | #include <asm/cputable.h> |
| 24 | #include <asm/tlbflush.h> |
| 25 | #include <asm/kvm_ppc.h> |
| 26 | #include <asm/dbell.h> |
| 27 | |
| 28 | #include "booke.h" |
| 29 | #include "e500.h" |
| 30 | |
| 31 | void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type) |
| 32 | { |
| 33 | enum ppc_dbell dbell_type; |
| 34 | unsigned long tag; |
| 35 | |
| 36 | switch (type) { |
| 37 | case INT_CLASS_NONCRIT: |
| 38 | dbell_type = PPC_G_DBELL; |
| 39 | break; |
| 40 | case INT_CLASS_CRIT: |
| 41 | dbell_type = PPC_G_DBELL_CRIT; |
| 42 | break; |
| 43 | case INT_CLASS_MC: |
| 44 | dbell_type = PPC_G_DBELL_MC; |
| 45 | break; |
| 46 | default: |
| 47 | WARN_ONCE(1, "%s: unknown int type %d\n", __func__, type); |
| 48 | return; |
| 49 | } |
| 50 | |
| 51 | preempt_disable(); |
| 52 | tag = PPC_DBELL_LPID(get_lpid(vcpu)) | vcpu->vcpu_id; |
| 53 | mb(); |
| 54 | ppc_msgsnd(dbell_type, 0, tag); |
| 55 | preempt_enable(); |
| 56 | } |
| 57 | |
| 58 | /* gtlbe must not be mapped by more than one host tlb entry */ |
| 59 | void kvmppc_e500_tlbil_one(struct kvmppc_vcpu_e500 *vcpu_e500, |
| 60 | struct kvm_book3e_206_tlb_entry *gtlbe) |
| 61 | { |
| 62 | unsigned int tid, ts; |
| 63 | gva_t eaddr; |
| 64 | u32 val; |
| 65 | unsigned long flags; |
| 66 | |
| 67 | ts = get_tlb_ts(gtlbe); |
| 68 | tid = get_tlb_tid(gtlbe); |
| 69 | |
| 70 | /* We search the host TLB to invalidate its shadow TLB entry */ |
| 71 | val = (tid << 16) | ts; |
| 72 | eaddr = get_tlb_eaddr(gtlbe); |
| 73 | |
| 74 | local_irq_save(flags); |
| 75 | |
| 76 | mtspr(SPRN_MAS6, val); |
| 77 | mtspr(SPRN_MAS5, MAS5_SGS | get_lpid(&vcpu_e500->vcpu)); |
| 78 | |
| 79 | asm volatile("tlbsx 0, %[eaddr]\n" : : [eaddr] "r" (eaddr)); |
| 80 | val = mfspr(SPRN_MAS1); |
| 81 | if (val & MAS1_VALID) { |
| 82 | mtspr(SPRN_MAS1, val & ~MAS1_VALID); |
| 83 | asm volatile("tlbwe"); |
| 84 | } |
| 85 | mtspr(SPRN_MAS5, 0); |
| 86 | /* NOTE: tlbsx also updates mas8, so clear it for host tlbwe */ |
| 87 | mtspr(SPRN_MAS8, 0); |
| 88 | isync(); |
| 89 | |
| 90 | local_irq_restore(flags); |
| 91 | } |
| 92 | |
| 93 | void kvmppc_e500_tlbil_all(struct kvmppc_vcpu_e500 *vcpu_e500) |
| 94 | { |
| 95 | unsigned long flags; |
| 96 | |
| 97 | local_irq_save(flags); |
| 98 | mtspr(SPRN_MAS5, MAS5_SGS | get_lpid(&vcpu_e500->vcpu)); |
| 99 | asm volatile("tlbilxlpid"); |
| 100 | mtspr(SPRN_MAS5, 0); |
| 101 | local_irq_restore(flags); |
| 102 | } |
| 103 | |
| 104 | void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid) |
| 105 | { |
| 106 | vcpu->arch.pid = pid; |
| 107 | } |
| 108 | |
| 109 | void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr) |
| 110 | { |
| 111 | } |
| 112 | |
| 113 | /* We use two lpids per VM */ |
| 114 | static DEFINE_PER_CPU(struct kvm_vcpu *[KVMPPC_NR_LPIDS], last_vcpu_of_lpid); |
| 115 | |
| 116 | static void kvmppc_core_vcpu_load_e500mc(struct kvm_vcpu *vcpu, int cpu) |
| 117 | { |
| 118 | struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu); |
| 119 | |
| 120 | kvmppc_booke_vcpu_load(vcpu, cpu); |
| 121 | |
| 122 | mtspr(SPRN_LPID, get_lpid(vcpu)); |
| 123 | mtspr(SPRN_EPCR, vcpu->arch.shadow_epcr); |
| 124 | mtspr(SPRN_GPIR, vcpu->vcpu_id); |
| 125 | mtspr(SPRN_MSRP, vcpu->arch.shadow_msrp); |
| 126 | vcpu->arch.eplc = EPC_EGS | (get_lpid(vcpu) << EPC_ELPID_SHIFT); |
| 127 | vcpu->arch.epsc = vcpu->arch.eplc; |
| 128 | mtspr(SPRN_EPLC, vcpu->arch.eplc); |
| 129 | mtspr(SPRN_EPSC, vcpu->arch.epsc); |
| 130 | |
| 131 | mtspr(SPRN_GIVPR, vcpu->arch.ivpr); |
| 132 | mtspr(SPRN_GIVOR2, vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE]); |
| 133 | mtspr(SPRN_GIVOR8, vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL]); |
| 134 | mtspr(SPRN_GSPRG0, (unsigned long)vcpu->arch.shared->sprg0); |
| 135 | mtspr(SPRN_GSPRG1, (unsigned long)vcpu->arch.shared->sprg1); |
| 136 | mtspr(SPRN_GSPRG2, (unsigned long)vcpu->arch.shared->sprg2); |
| 137 | mtspr(SPRN_GSPRG3, (unsigned long)vcpu->arch.shared->sprg3); |
| 138 | |
| 139 | mtspr(SPRN_GSRR0, vcpu->arch.shared->srr0); |
| 140 | mtspr(SPRN_GSRR1, vcpu->arch.shared->srr1); |
| 141 | |
| 142 | mtspr(SPRN_GEPR, vcpu->arch.epr); |
| 143 | mtspr(SPRN_GDEAR, vcpu->arch.shared->dar); |
| 144 | mtspr(SPRN_GESR, vcpu->arch.shared->esr); |
| 145 | |
| 146 | if (vcpu->arch.oldpir != mfspr(SPRN_PIR) || |
| 147 | __this_cpu_read(last_vcpu_of_lpid[get_lpid(vcpu)]) != vcpu) { |
| 148 | kvmppc_e500_tlbil_all(vcpu_e500); |
| 149 | __this_cpu_write(last_vcpu_of_lpid[get_lpid(vcpu)], vcpu); |
| 150 | } |
| 151 | } |
| 152 | |
| 153 | static void kvmppc_core_vcpu_put_e500mc(struct kvm_vcpu *vcpu) |
| 154 | { |
| 155 | vcpu->arch.eplc = mfspr(SPRN_EPLC); |
| 156 | vcpu->arch.epsc = mfspr(SPRN_EPSC); |
| 157 | |
| 158 | vcpu->arch.shared->sprg0 = mfspr(SPRN_GSPRG0); |
| 159 | vcpu->arch.shared->sprg1 = mfspr(SPRN_GSPRG1); |
| 160 | vcpu->arch.shared->sprg2 = mfspr(SPRN_GSPRG2); |
| 161 | vcpu->arch.shared->sprg3 = mfspr(SPRN_GSPRG3); |
| 162 | |
| 163 | vcpu->arch.shared->srr0 = mfspr(SPRN_GSRR0); |
| 164 | vcpu->arch.shared->srr1 = mfspr(SPRN_GSRR1); |
| 165 | |
| 166 | vcpu->arch.epr = mfspr(SPRN_GEPR); |
| 167 | vcpu->arch.shared->dar = mfspr(SPRN_GDEAR); |
| 168 | vcpu->arch.shared->esr = mfspr(SPRN_GESR); |
| 169 | |
| 170 | vcpu->arch.oldpir = mfspr(SPRN_PIR); |
| 171 | |
| 172 | kvmppc_booke_vcpu_put(vcpu); |
| 173 | } |
| 174 | |
| 175 | int kvmppc_core_check_processor_compat(void) |
| 176 | { |
| 177 | int r; |
| 178 | |
| 179 | if (strcmp(cur_cpu_spec->cpu_name, "e500mc") == 0) |
| 180 | r = 0; |
| 181 | else if (strcmp(cur_cpu_spec->cpu_name, "e5500") == 0) |
| 182 | r = 0; |
| 183 | #ifdef CONFIG_ALTIVEC |
| 184 | /* |
| 185 | * Since guests have the priviledge to enable AltiVec, we need AltiVec |
| 186 | * support in the host to save/restore their context. |
| 187 | * Don't use CPU_FTR_ALTIVEC to identify cores with AltiVec unit |
| 188 | * because it's cleared in the absence of CONFIG_ALTIVEC! |
| 189 | */ |
| 190 | else if (strcmp(cur_cpu_spec->cpu_name, "e6500") == 0) |
| 191 | r = 0; |
| 192 | #endif |
| 193 | else |
| 194 | r = -ENOTSUPP; |
| 195 | |
| 196 | return r; |
| 197 | } |
| 198 | |
| 199 | int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu) |
| 200 | { |
| 201 | struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu); |
| 202 | |
| 203 | vcpu->arch.shadow_epcr = SPRN_EPCR_DSIGS | SPRN_EPCR_DGTMI | \ |
| 204 | SPRN_EPCR_DUVD; |
| 205 | #ifdef CONFIG_64BIT |
| 206 | vcpu->arch.shadow_epcr |= SPRN_EPCR_ICM; |
| 207 | #endif |
| 208 | vcpu->arch.shadow_msrp = MSRP_UCLEP | MSRP_PMMP; |
| 209 | |
| 210 | vcpu->arch.pvr = mfspr(SPRN_PVR); |
| 211 | vcpu_e500->svr = mfspr(SPRN_SVR); |
| 212 | |
| 213 | vcpu->arch.cpu_type = KVM_CPU_E500MC; |
| 214 | |
| 215 | return 0; |
| 216 | } |
| 217 | |
| 218 | static int kvmppc_core_get_sregs_e500mc(struct kvm_vcpu *vcpu, |
| 219 | struct kvm_sregs *sregs) |
| 220 | { |
| 221 | struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu); |
| 222 | |
| 223 | sregs->u.e.features |= KVM_SREGS_E_ARCH206_MMU | KVM_SREGS_E_PM | |
| 224 | KVM_SREGS_E_PC; |
| 225 | sregs->u.e.impl_id = KVM_SREGS_E_IMPL_FSL; |
| 226 | |
| 227 | sregs->u.e.impl.fsl.features = 0; |
| 228 | sregs->u.e.impl.fsl.svr = vcpu_e500->svr; |
| 229 | sregs->u.e.impl.fsl.hid0 = vcpu_e500->hid0; |
| 230 | sregs->u.e.impl.fsl.mcar = vcpu_e500->mcar; |
| 231 | |
| 232 | kvmppc_get_sregs_e500_tlb(vcpu, sregs); |
| 233 | |
| 234 | sregs->u.e.ivor_high[3] = |
| 235 | vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR]; |
| 236 | sregs->u.e.ivor_high[4] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL]; |
| 237 | sregs->u.e.ivor_high[5] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT]; |
| 238 | |
| 239 | return kvmppc_get_sregs_ivor(vcpu, sregs); |
| 240 | } |
| 241 | |
| 242 | static int kvmppc_core_set_sregs_e500mc(struct kvm_vcpu *vcpu, |
| 243 | struct kvm_sregs *sregs) |
| 244 | { |
| 245 | struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu); |
| 246 | int ret; |
| 247 | |
| 248 | if (sregs->u.e.impl_id == KVM_SREGS_E_IMPL_FSL) { |
| 249 | vcpu_e500->svr = sregs->u.e.impl.fsl.svr; |
| 250 | vcpu_e500->hid0 = sregs->u.e.impl.fsl.hid0; |
| 251 | vcpu_e500->mcar = sregs->u.e.impl.fsl.mcar; |
| 252 | } |
| 253 | |
| 254 | ret = kvmppc_set_sregs_e500_tlb(vcpu, sregs); |
| 255 | if (ret < 0) |
| 256 | return ret; |
| 257 | |
| 258 | if (!(sregs->u.e.features & KVM_SREGS_E_IVOR)) |
| 259 | return 0; |
| 260 | |
| 261 | if (sregs->u.e.features & KVM_SREGS_E_PM) { |
| 262 | vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = |
| 263 | sregs->u.e.ivor_high[3]; |
| 264 | } |
| 265 | |
| 266 | if (sregs->u.e.features & KVM_SREGS_E_PC) { |
| 267 | vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = |
| 268 | sregs->u.e.ivor_high[4]; |
| 269 | vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = |
| 270 | sregs->u.e.ivor_high[5]; |
| 271 | } |
| 272 | |
| 273 | return kvmppc_set_sregs_ivor(vcpu, sregs); |
| 274 | } |
| 275 | |
| 276 | static int kvmppc_get_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id, |
| 277 | union kvmppc_one_reg *val) |
| 278 | { |
| 279 | int r = 0; |
| 280 | |
| 281 | switch (id) { |
| 282 | case KVM_REG_PPC_SPRG9: |
| 283 | *val = get_reg_val(id, vcpu->arch.sprg9); |
| 284 | break; |
| 285 | default: |
| 286 | r = kvmppc_get_one_reg_e500_tlb(vcpu, id, val); |
| 287 | } |
| 288 | |
| 289 | return r; |
| 290 | } |
| 291 | |
| 292 | static int kvmppc_set_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id, |
| 293 | union kvmppc_one_reg *val) |
| 294 | { |
| 295 | int r = 0; |
| 296 | |
| 297 | switch (id) { |
| 298 | case KVM_REG_PPC_SPRG9: |
| 299 | vcpu->arch.sprg9 = set_reg_val(id, *val); |
| 300 | break; |
| 301 | default: |
| 302 | r = kvmppc_set_one_reg_e500_tlb(vcpu, id, val); |
| 303 | } |
| 304 | |
| 305 | return r; |
| 306 | } |
| 307 | |
| 308 | static struct kvm_vcpu *kvmppc_core_vcpu_create_e500mc(struct kvm *kvm, |
| 309 | unsigned int id) |
| 310 | { |
| 311 | struct kvmppc_vcpu_e500 *vcpu_e500; |
| 312 | struct kvm_vcpu *vcpu; |
| 313 | int err; |
| 314 | |
| 315 | vcpu_e500 = kmem_cache_zalloc(kvm_vcpu_cache, GFP_KERNEL); |
| 316 | if (!vcpu_e500) { |
| 317 | err = -ENOMEM; |
| 318 | goto out; |
| 319 | } |
| 320 | vcpu = &vcpu_e500->vcpu; |
| 321 | |
| 322 | /* Invalid PIR value -- this LPID dosn't have valid state on any cpu */ |
| 323 | vcpu->arch.oldpir = 0xffffffff; |
| 324 | |
| 325 | err = kvm_vcpu_init(vcpu, kvm, id); |
| 326 | if (err) |
| 327 | goto free_vcpu; |
| 328 | |
| 329 | err = kvmppc_e500_tlb_init(vcpu_e500); |
| 330 | if (err) |
| 331 | goto uninit_vcpu; |
| 332 | |
| 333 | vcpu->arch.shared = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO); |
| 334 | if (!vcpu->arch.shared) |
| 335 | goto uninit_tlb; |
| 336 | |
| 337 | return vcpu; |
| 338 | |
| 339 | uninit_tlb: |
| 340 | kvmppc_e500_tlb_uninit(vcpu_e500); |
| 341 | uninit_vcpu: |
| 342 | kvm_vcpu_uninit(vcpu); |
| 343 | |
| 344 | free_vcpu: |
| 345 | kmem_cache_free(kvm_vcpu_cache, vcpu_e500); |
| 346 | out: |
| 347 | return ERR_PTR(err); |
| 348 | } |
| 349 | |
| 350 | static void kvmppc_core_vcpu_free_e500mc(struct kvm_vcpu *vcpu) |
| 351 | { |
| 352 | struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu); |
| 353 | |
| 354 | free_page((unsigned long)vcpu->arch.shared); |
| 355 | kvmppc_e500_tlb_uninit(vcpu_e500); |
| 356 | kvm_vcpu_uninit(vcpu); |
| 357 | kmem_cache_free(kvm_vcpu_cache, vcpu_e500); |
| 358 | } |
| 359 | |
| 360 | static int kvmppc_core_init_vm_e500mc(struct kvm *kvm) |
| 361 | { |
| 362 | int lpid; |
| 363 | |
| 364 | lpid = kvmppc_alloc_lpid(); |
| 365 | if (lpid < 0) |
| 366 | return lpid; |
| 367 | |
| 368 | /* |
| 369 | * Use two lpids per VM on cores with two threads like e6500. Use |
| 370 | * even numbers to speedup vcpu lpid computation with consecutive lpids |
| 371 | * per VM. vm1 will use lpids 2 and 3, vm2 lpids 4 and 5, and so on. |
| 372 | */ |
| 373 | if (threads_per_core == 2) |
| 374 | lpid <<= 1; |
| 375 | |
| 376 | kvm->arch.lpid = lpid; |
| 377 | return 0; |
| 378 | } |
| 379 | |
| 380 | static void kvmppc_core_destroy_vm_e500mc(struct kvm *kvm) |
| 381 | { |
| 382 | int lpid = kvm->arch.lpid; |
| 383 | |
| 384 | if (threads_per_core == 2) |
| 385 | lpid >>= 1; |
| 386 | |
| 387 | kvmppc_free_lpid(lpid); |
| 388 | } |
| 389 | |
| 390 | static struct kvmppc_ops kvm_ops_e500mc = { |
| 391 | .get_sregs = kvmppc_core_get_sregs_e500mc, |
| 392 | .set_sregs = kvmppc_core_set_sregs_e500mc, |
| 393 | .get_one_reg = kvmppc_get_one_reg_e500mc, |
| 394 | .set_one_reg = kvmppc_set_one_reg_e500mc, |
| 395 | .vcpu_load = kvmppc_core_vcpu_load_e500mc, |
| 396 | .vcpu_put = kvmppc_core_vcpu_put_e500mc, |
| 397 | .vcpu_create = kvmppc_core_vcpu_create_e500mc, |
| 398 | .vcpu_free = kvmppc_core_vcpu_free_e500mc, |
| 399 | .mmu_destroy = kvmppc_mmu_destroy_e500, |
| 400 | .init_vm = kvmppc_core_init_vm_e500mc, |
| 401 | .destroy_vm = kvmppc_core_destroy_vm_e500mc, |
| 402 | .emulate_op = kvmppc_core_emulate_op_e500, |
| 403 | .emulate_mtspr = kvmppc_core_emulate_mtspr_e500, |
| 404 | .emulate_mfspr = kvmppc_core_emulate_mfspr_e500, |
| 405 | }; |
| 406 | |
| 407 | static int __init kvmppc_e500mc_init(void) |
| 408 | { |
| 409 | int r; |
| 410 | |
| 411 | r = kvmppc_booke_init(); |
| 412 | if (r) |
| 413 | goto err_out; |
| 414 | |
| 415 | /* |
| 416 | * Use two lpids per VM on dual threaded processors like e6500 |
| 417 | * to workarround the lack of tlb write conditional instruction. |
| 418 | * Expose half the number of available hardware lpids to the lpid |
| 419 | * allocator. |
| 420 | */ |
| 421 | kvmppc_init_lpid(KVMPPC_NR_LPIDS/threads_per_core); |
| 422 | kvmppc_claim_lpid(0); /* host */ |
| 423 | |
| 424 | r = kvm_init(NULL, sizeof(struct kvmppc_vcpu_e500), 0, THIS_MODULE); |
| 425 | if (r) |
| 426 | goto err_out; |
| 427 | kvm_ops_e500mc.owner = THIS_MODULE; |
| 428 | kvmppc_pr_ops = &kvm_ops_e500mc; |
| 429 | |
| 430 | err_out: |
| 431 | return r; |
| 432 | } |
| 433 | |
| 434 | static void __exit kvmppc_e500mc_exit(void) |
| 435 | { |
| 436 | kvmppc_pr_ops = NULL; |
| 437 | kvmppc_booke_exit(); |
| 438 | } |
| 439 | |
| 440 | module_init(kvmppc_e500mc_init); |
| 441 | module_exit(kvmppc_e500mc_exit); |
| 442 | MODULE_ALIAS_MISCDEV(KVM_MINOR); |
| 443 | MODULE_ALIAS("devname:kvm"); |