mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-25 03:55:09 +08:00
46808a4cb8
Everywhere we use kvm_for_each_vpcu(), we use an int as the vcpu index. Unfortunately, we're about to move rework the iterator, which requires this to be upgrade to an unsigned long. Let's bite the bullet and repaint all of it in one go. Signed-off-by: Marc Zyngier <maz@kernel.org> Message-Id: <20211116160403.4074052-7-maz@kernel.org> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
453 lines
9.6 KiB
C
453 lines
9.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
|
|
*
|
|
* Author: Yu Liu, <yu.liu@freescale.com>
|
|
*
|
|
* Description:
|
|
* This file is derived from arch/powerpc/kvm/44x_emulate.c,
|
|
* by Hollis Blanchard <hollisb@us.ibm.com>.
|
|
*/
|
|
|
|
#include <asm/kvm_ppc.h>
|
|
#include <asm/disassemble.h>
|
|
#include <asm/dbell.h>
|
|
#include <asm/reg_booke.h>
|
|
|
|
#include "booke.h"
|
|
#include "e500.h"
|
|
|
|
#define XOP_DCBTLS 166
|
|
#define XOP_MSGSND 206
|
|
#define XOP_MSGCLR 238
|
|
#define XOP_MFTMR 366
|
|
#define XOP_TLBIVAX 786
|
|
#define XOP_TLBSX 914
|
|
#define XOP_TLBRE 946
|
|
#define XOP_TLBWE 978
|
|
#define XOP_TLBILX 18
|
|
#define XOP_EHPRIV 270
|
|
|
|
#ifdef CONFIG_KVM_E500MC
|
|
static int dbell2prio(ulong param)
|
|
{
|
|
int msg = param & PPC_DBELL_TYPE_MASK;
|
|
int prio = -1;
|
|
|
|
switch (msg) {
|
|
case PPC_DBELL_TYPE(PPC_DBELL):
|
|
prio = BOOKE_IRQPRIO_DBELL;
|
|
break;
|
|
case PPC_DBELL_TYPE(PPC_DBELL_CRIT):
|
|
prio = BOOKE_IRQPRIO_DBELL_CRIT;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return prio;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb)
|
|
{
|
|
ulong param = vcpu->arch.regs.gpr[rb];
|
|
int prio = dbell2prio(param);
|
|
|
|
if (prio < 0)
|
|
return EMULATE_FAIL;
|
|
|
|
clear_bit(prio, &vcpu->arch.pending_exceptions);
|
|
return EMULATE_DONE;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb)
|
|
{
|
|
ulong param = vcpu->arch.regs.gpr[rb];
|
|
int prio = dbell2prio(rb);
|
|
int pir = param & PPC_DBELL_PIR_MASK;
|
|
unsigned long i;
|
|
struct kvm_vcpu *cvcpu;
|
|
|
|
if (prio < 0)
|
|
return EMULATE_FAIL;
|
|
|
|
kvm_for_each_vcpu(i, cvcpu, vcpu->kvm) {
|
|
int cpir = cvcpu->arch.shared->pir;
|
|
if ((param & PPC_DBELL_MSG_BRDCAST) || (cpir == pir)) {
|
|
set_bit(prio, &cvcpu->arch.pending_exceptions);
|
|
kvm_vcpu_kick(cvcpu);
|
|
}
|
|
}
|
|
|
|
return EMULATE_DONE;
|
|
}
|
|
#endif
|
|
|
|
static int kvmppc_e500_emul_ehpriv(struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance)
|
|
{
|
|
int emulated = EMULATE_DONE;
|
|
|
|
switch (get_oc(inst)) {
|
|
case EHPRIV_OC_DEBUG:
|
|
vcpu->run->exit_reason = KVM_EXIT_DEBUG;
|
|
vcpu->run->debug.arch.address = vcpu->arch.regs.nip;
|
|
vcpu->run->debug.arch.status = 0;
|
|
kvmppc_account_exit(vcpu, DEBUG_EXITS);
|
|
emulated = EMULATE_EXIT_USER;
|
|
*advance = 0;
|
|
break;
|
|
default:
|
|
emulated = EMULATE_FAIL;
|
|
}
|
|
return emulated;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
|
|
|
|
/* Always fail to lock the cache */
|
|
vcpu_e500->l1csr0 |= L1CSR0_CUL;
|
|
return EMULATE_DONE;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst,
|
|
int rt)
|
|
{
|
|
/* Expose one thread per vcpu */
|
|
if (get_tmrn(inst) == TMRN_TMCFG0) {
|
|
kvmppc_set_gpr(vcpu, rt,
|
|
1 | (1 << TMRN_TMCFG0_NATHRD_SHIFT));
|
|
return EMULATE_DONE;
|
|
}
|
|
|
|
return EMULATE_FAIL;
|
|
}
|
|
|
|
int kvmppc_core_emulate_op_e500(struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance)
|
|
{
|
|
int emulated = EMULATE_DONE;
|
|
int ra = get_ra(inst);
|
|
int rb = get_rb(inst);
|
|
int rt = get_rt(inst);
|
|
gva_t ea;
|
|
|
|
switch (get_op(inst)) {
|
|
case 31:
|
|
switch (get_xop(inst)) {
|
|
|
|
case XOP_DCBTLS:
|
|
emulated = kvmppc_e500_emul_dcbtls(vcpu);
|
|
break;
|
|
|
|
#ifdef CONFIG_KVM_E500MC
|
|
case XOP_MSGSND:
|
|
emulated = kvmppc_e500_emul_msgsnd(vcpu, rb);
|
|
break;
|
|
|
|
case XOP_MSGCLR:
|
|
emulated = kvmppc_e500_emul_msgclr(vcpu, rb);
|
|
break;
|
|
#endif
|
|
|
|
case XOP_TLBRE:
|
|
emulated = kvmppc_e500_emul_tlbre(vcpu);
|
|
break;
|
|
|
|
case XOP_TLBWE:
|
|
emulated = kvmppc_e500_emul_tlbwe(vcpu);
|
|
break;
|
|
|
|
case XOP_TLBSX:
|
|
ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
|
|
emulated = kvmppc_e500_emul_tlbsx(vcpu, ea);
|
|
break;
|
|
|
|
case XOP_TLBILX: {
|
|
int type = rt & 0x3;
|
|
ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
|
|
emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea);
|
|
break;
|
|
}
|
|
|
|
case XOP_TLBIVAX:
|
|
ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
|
|
emulated = kvmppc_e500_emul_tlbivax(vcpu, ea);
|
|
break;
|
|
|
|
case XOP_MFTMR:
|
|
emulated = kvmppc_e500_emul_mftmr(vcpu, inst, rt);
|
|
break;
|
|
|
|
case XOP_EHPRIV:
|
|
emulated = kvmppc_e500_emul_ehpriv(vcpu, inst, advance);
|
|
break;
|
|
|
|
default:
|
|
emulated = EMULATE_FAIL;
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
emulated = EMULATE_FAIL;
|
|
}
|
|
|
|
if (emulated == EMULATE_FAIL)
|
|
emulated = kvmppc_booke_emulate_op(vcpu, inst, advance);
|
|
|
|
return emulated;
|
|
}
|
|
|
|
int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
|
|
{
|
|
struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
|
|
int emulated = EMULATE_DONE;
|
|
|
|
switch (sprn) {
|
|
#ifndef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_PID:
|
|
kvmppc_set_pid(vcpu, spr_val);
|
|
break;
|
|
case SPRN_PID1:
|
|
if (spr_val != 0)
|
|
return EMULATE_FAIL;
|
|
vcpu_e500->pid[1] = spr_val;
|
|
break;
|
|
case SPRN_PID2:
|
|
if (spr_val != 0)
|
|
return EMULATE_FAIL;
|
|
vcpu_e500->pid[2] = spr_val;
|
|
break;
|
|
case SPRN_MAS0:
|
|
vcpu->arch.shared->mas0 = spr_val;
|
|
break;
|
|
case SPRN_MAS1:
|
|
vcpu->arch.shared->mas1 = spr_val;
|
|
break;
|
|
case SPRN_MAS2:
|
|
vcpu->arch.shared->mas2 = spr_val;
|
|
break;
|
|
case SPRN_MAS3:
|
|
vcpu->arch.shared->mas7_3 &= ~(u64)0xffffffff;
|
|
vcpu->arch.shared->mas7_3 |= spr_val;
|
|
break;
|
|
case SPRN_MAS4:
|
|
vcpu->arch.shared->mas4 = spr_val;
|
|
break;
|
|
case SPRN_MAS6:
|
|
vcpu->arch.shared->mas6 = spr_val;
|
|
break;
|
|
case SPRN_MAS7:
|
|
vcpu->arch.shared->mas7_3 &= (u64)0xffffffff;
|
|
vcpu->arch.shared->mas7_3 |= (u64)spr_val << 32;
|
|
break;
|
|
#endif
|
|
case SPRN_L1CSR0:
|
|
vcpu_e500->l1csr0 = spr_val;
|
|
vcpu_e500->l1csr0 &= ~(L1CSR0_DCFI | L1CSR0_CLFC);
|
|
break;
|
|
case SPRN_L1CSR1:
|
|
vcpu_e500->l1csr1 = spr_val;
|
|
vcpu_e500->l1csr1 &= ~(L1CSR1_ICFI | L1CSR1_ICLFR);
|
|
break;
|
|
case SPRN_HID0:
|
|
vcpu_e500->hid0 = spr_val;
|
|
break;
|
|
case SPRN_HID1:
|
|
vcpu_e500->hid1 = spr_val;
|
|
break;
|
|
|
|
case SPRN_MMUCSR0:
|
|
emulated = kvmppc_e500_emul_mt_mmucsr0(vcpu_e500,
|
|
spr_val);
|
|
break;
|
|
|
|
case SPRN_PWRMGTCR0:
|
|
/*
|
|
* Guest relies on host power management configurations
|
|
* Treat the request as a general store
|
|
*/
|
|
vcpu->arch.pwrmgtcr0 = spr_val;
|
|
break;
|
|
|
|
case SPRN_BUCSR:
|
|
/*
|
|
* If we are here, it means that we have already flushed the
|
|
* branch predictor, so just return to guest.
|
|
*/
|
|
break;
|
|
|
|
/* extra exceptions */
|
|
#ifdef CONFIG_SPE_POSSIBLE
|
|
case SPRN_IVOR32:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] = spr_val;
|
|
break;
|
|
case SPRN_IVOR33:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] = spr_val;
|
|
break;
|
|
case SPRN_IVOR34:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] = spr_val;
|
|
break;
|
|
#endif
|
|
#ifdef CONFIG_ALTIVEC
|
|
case SPRN_IVOR32:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL] = spr_val;
|
|
break;
|
|
case SPRN_IVOR33:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST] = spr_val;
|
|
break;
|
|
#endif
|
|
case SPRN_IVOR35:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = spr_val;
|
|
break;
|
|
#ifdef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_IVOR36:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = spr_val;
|
|
break;
|
|
case SPRN_IVOR37:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = spr_val;
|
|
break;
|
|
#endif
|
|
default:
|
|
emulated = kvmppc_booke_emulate_mtspr(vcpu, sprn, spr_val);
|
|
}
|
|
|
|
return emulated;
|
|
}
|
|
|
|
int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
|
|
{
|
|
struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
|
|
int emulated = EMULATE_DONE;
|
|
|
|
switch (sprn) {
|
|
#ifndef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_PID:
|
|
*spr_val = vcpu_e500->pid[0];
|
|
break;
|
|
case SPRN_PID1:
|
|
*spr_val = vcpu_e500->pid[1];
|
|
break;
|
|
case SPRN_PID2:
|
|
*spr_val = vcpu_e500->pid[2];
|
|
break;
|
|
case SPRN_MAS0:
|
|
*spr_val = vcpu->arch.shared->mas0;
|
|
break;
|
|
case SPRN_MAS1:
|
|
*spr_val = vcpu->arch.shared->mas1;
|
|
break;
|
|
case SPRN_MAS2:
|
|
*spr_val = vcpu->arch.shared->mas2;
|
|
break;
|
|
case SPRN_MAS3:
|
|
*spr_val = (u32)vcpu->arch.shared->mas7_3;
|
|
break;
|
|
case SPRN_MAS4:
|
|
*spr_val = vcpu->arch.shared->mas4;
|
|
break;
|
|
case SPRN_MAS6:
|
|
*spr_val = vcpu->arch.shared->mas6;
|
|
break;
|
|
case SPRN_MAS7:
|
|
*spr_val = vcpu->arch.shared->mas7_3 >> 32;
|
|
break;
|
|
#endif
|
|
case SPRN_DECAR:
|
|
*spr_val = vcpu->arch.decar;
|
|
break;
|
|
case SPRN_TLB0CFG:
|
|
*spr_val = vcpu->arch.tlbcfg[0];
|
|
break;
|
|
case SPRN_TLB1CFG:
|
|
*spr_val = vcpu->arch.tlbcfg[1];
|
|
break;
|
|
case SPRN_TLB0PS:
|
|
if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
|
|
return EMULATE_FAIL;
|
|
*spr_val = vcpu->arch.tlbps[0];
|
|
break;
|
|
case SPRN_TLB1PS:
|
|
if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
|
|
return EMULATE_FAIL;
|
|
*spr_val = vcpu->arch.tlbps[1];
|
|
break;
|
|
case SPRN_L1CSR0:
|
|
*spr_val = vcpu_e500->l1csr0;
|
|
break;
|
|
case SPRN_L1CSR1:
|
|
*spr_val = vcpu_e500->l1csr1;
|
|
break;
|
|
case SPRN_HID0:
|
|
*spr_val = vcpu_e500->hid0;
|
|
break;
|
|
case SPRN_HID1:
|
|
*spr_val = vcpu_e500->hid1;
|
|
break;
|
|
case SPRN_SVR:
|
|
*spr_val = vcpu_e500->svr;
|
|
break;
|
|
|
|
case SPRN_MMUCSR0:
|
|
*spr_val = 0;
|
|
break;
|
|
|
|
case SPRN_MMUCFG:
|
|
*spr_val = vcpu->arch.mmucfg;
|
|
break;
|
|
case SPRN_EPTCFG:
|
|
if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
|
|
return EMULATE_FAIL;
|
|
/*
|
|
* Legacy Linux guests access EPTCFG register even if the E.PT
|
|
* category is disabled in the VM. Give them a chance to live.
|
|
*/
|
|
*spr_val = vcpu->arch.eptcfg;
|
|
break;
|
|
|
|
case SPRN_PWRMGTCR0:
|
|
*spr_val = vcpu->arch.pwrmgtcr0;
|
|
break;
|
|
|
|
/* extra exceptions */
|
|
#ifdef CONFIG_SPE_POSSIBLE
|
|
case SPRN_IVOR32:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
|
|
break;
|
|
case SPRN_IVOR33:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
|
|
break;
|
|
case SPRN_IVOR34:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
|
|
break;
|
|
#endif
|
|
#ifdef CONFIG_ALTIVEC
|
|
case SPRN_IVOR32:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL];
|
|
break;
|
|
case SPRN_IVOR33:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST];
|
|
break;
|
|
#endif
|
|
case SPRN_IVOR35:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
|
|
break;
|
|
#ifdef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_IVOR36:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
|
|
break;
|
|
case SPRN_IVOR37:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
|
|
break;
|
|
#endif
|
|
default:
|
|
emulated = kvmppc_booke_emulate_mfspr(vcpu, sprn, spr_val);
|
|
}
|
|
|
|
return emulated;
|
|
}
|
|
|