forked from Qortal/Brooklyn
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
219 lines
6.1 KiB
219 lines
6.1 KiB
// SPDX-License-Identifier: GPL-2.0-only |
|
|
|
#include <asm/kvm_ppc.h> |
|
#include <asm/pmc.h> |
|
|
|
#include "book3s_hv.h" |
|
|
|
static void freeze_pmu(unsigned long mmcr0, unsigned long mmcra) |
|
{ |
|
if (!(mmcr0 & MMCR0_FC)) |
|
goto do_freeze; |
|
if (mmcra & MMCRA_SAMPLE_ENABLE) |
|
goto do_freeze; |
|
if (cpu_has_feature(CPU_FTR_ARCH_31)) { |
|
if (!(mmcr0 & MMCR0_PMCCEXT)) |
|
goto do_freeze; |
|
if (!(mmcra & MMCRA_BHRB_DISABLE)) |
|
goto do_freeze; |
|
} |
|
return; |
|
|
|
do_freeze: |
|
mmcr0 = MMCR0_FC; |
|
mmcra = 0; |
|
if (cpu_has_feature(CPU_FTR_ARCH_31)) { |
|
mmcr0 |= MMCR0_PMCCEXT; |
|
mmcra = MMCRA_BHRB_DISABLE; |
|
} |
|
|
|
mtspr(SPRN_MMCR0, mmcr0); |
|
mtspr(SPRN_MMCRA, mmcra); |
|
isync(); |
|
} |
|
|
|
void switch_pmu_to_guest(struct kvm_vcpu *vcpu, |
|
struct p9_host_os_sprs *host_os_sprs) |
|
{ |
|
struct lppaca *lp; |
|
int load_pmu = 1; |
|
|
|
lp = vcpu->arch.vpa.pinned_addr; |
|
if (lp) |
|
load_pmu = lp->pmcregs_in_use; |
|
|
|
/* Save host */ |
|
if (ppc_get_pmu_inuse()) { |
|
/* POWER9, POWER10 do not implement HPMC or SPMC */ |
|
|
|
host_os_sprs->mmcr0 = mfspr(SPRN_MMCR0); |
|
host_os_sprs->mmcra = mfspr(SPRN_MMCRA); |
|
|
|
freeze_pmu(host_os_sprs->mmcr0, host_os_sprs->mmcra); |
|
|
|
host_os_sprs->pmc1 = mfspr(SPRN_PMC1); |
|
host_os_sprs->pmc2 = mfspr(SPRN_PMC2); |
|
host_os_sprs->pmc3 = mfspr(SPRN_PMC3); |
|
host_os_sprs->pmc4 = mfspr(SPRN_PMC4); |
|
host_os_sprs->pmc5 = mfspr(SPRN_PMC5); |
|
host_os_sprs->pmc6 = mfspr(SPRN_PMC6); |
|
host_os_sprs->mmcr1 = mfspr(SPRN_MMCR1); |
|
host_os_sprs->mmcr2 = mfspr(SPRN_MMCR2); |
|
host_os_sprs->sdar = mfspr(SPRN_SDAR); |
|
host_os_sprs->siar = mfspr(SPRN_SIAR); |
|
host_os_sprs->sier1 = mfspr(SPRN_SIER); |
|
|
|
if (cpu_has_feature(CPU_FTR_ARCH_31)) { |
|
host_os_sprs->mmcr3 = mfspr(SPRN_MMCR3); |
|
host_os_sprs->sier2 = mfspr(SPRN_SIER2); |
|
host_os_sprs->sier3 = mfspr(SPRN_SIER3); |
|
} |
|
} |
|
|
|
#ifdef CONFIG_PPC_PSERIES |
|
/* After saving PMU, before loading guest PMU, flip pmcregs_in_use */ |
|
if (kvmhv_on_pseries()) { |
|
barrier(); |
|
get_lppaca()->pmcregs_in_use = load_pmu; |
|
barrier(); |
|
} |
|
#endif |
|
|
|
/* |
|
* Load guest. If the VPA said the PMCs are not in use but the guest |
|
* tried to access them anyway, HFSCR[PM] will be set by the HFAC |
|
* fault so we can make forward progress. |
|
*/ |
|
if (load_pmu || (vcpu->arch.hfscr & HFSCR_PM)) { |
|
mtspr(SPRN_PMC1, vcpu->arch.pmc[0]); |
|
mtspr(SPRN_PMC2, vcpu->arch.pmc[1]); |
|
mtspr(SPRN_PMC3, vcpu->arch.pmc[2]); |
|
mtspr(SPRN_PMC4, vcpu->arch.pmc[3]); |
|
mtspr(SPRN_PMC5, vcpu->arch.pmc[4]); |
|
mtspr(SPRN_PMC6, vcpu->arch.pmc[5]); |
|
mtspr(SPRN_MMCR1, vcpu->arch.mmcr[1]); |
|
mtspr(SPRN_MMCR2, vcpu->arch.mmcr[2]); |
|
mtspr(SPRN_SDAR, vcpu->arch.sdar); |
|
mtspr(SPRN_SIAR, vcpu->arch.siar); |
|
mtspr(SPRN_SIER, vcpu->arch.sier[0]); |
|
|
|
if (cpu_has_feature(CPU_FTR_ARCH_31)) { |
|
mtspr(SPRN_MMCR3, vcpu->arch.mmcr[3]); |
|
mtspr(SPRN_SIER2, vcpu->arch.sier[1]); |
|
mtspr(SPRN_SIER3, vcpu->arch.sier[2]); |
|
} |
|
|
|
/* Set MMCRA then MMCR0 last */ |
|
mtspr(SPRN_MMCRA, vcpu->arch.mmcra); |
|
mtspr(SPRN_MMCR0, vcpu->arch.mmcr[0]); |
|
/* No isync necessary because we're starting counters */ |
|
|
|
if (!vcpu->arch.nested && |
|
(vcpu->arch.hfscr_permitted & HFSCR_PM)) |
|
vcpu->arch.hfscr |= HFSCR_PM; |
|
} |
|
} |
|
EXPORT_SYMBOL_GPL(switch_pmu_to_guest); |
|
|
|
void switch_pmu_to_host(struct kvm_vcpu *vcpu, |
|
struct p9_host_os_sprs *host_os_sprs) |
|
{ |
|
struct lppaca *lp; |
|
int save_pmu = 1; |
|
|
|
lp = vcpu->arch.vpa.pinned_addr; |
|
if (lp) |
|
save_pmu = lp->pmcregs_in_use; |
|
if (IS_ENABLED(CONFIG_KVM_BOOK3S_HV_NESTED_PMU_WORKAROUND)) { |
|
/* |
|
* Save pmu if this guest is capable of running nested guests. |
|
* This is option is for old L1s that do not set their |
|
* lppaca->pmcregs_in_use properly when entering their L2. |
|
*/ |
|
save_pmu |= nesting_enabled(vcpu->kvm); |
|
} |
|
|
|
if (save_pmu) { |
|
vcpu->arch.mmcr[0] = mfspr(SPRN_MMCR0); |
|
vcpu->arch.mmcra = mfspr(SPRN_MMCRA); |
|
|
|
freeze_pmu(vcpu->arch.mmcr[0], vcpu->arch.mmcra); |
|
|
|
vcpu->arch.pmc[0] = mfspr(SPRN_PMC1); |
|
vcpu->arch.pmc[1] = mfspr(SPRN_PMC2); |
|
vcpu->arch.pmc[2] = mfspr(SPRN_PMC3); |
|
vcpu->arch.pmc[3] = mfspr(SPRN_PMC4); |
|
vcpu->arch.pmc[4] = mfspr(SPRN_PMC5); |
|
vcpu->arch.pmc[5] = mfspr(SPRN_PMC6); |
|
vcpu->arch.mmcr[1] = mfspr(SPRN_MMCR1); |
|
vcpu->arch.mmcr[2] = mfspr(SPRN_MMCR2); |
|
vcpu->arch.sdar = mfspr(SPRN_SDAR); |
|
vcpu->arch.siar = mfspr(SPRN_SIAR); |
|
vcpu->arch.sier[0] = mfspr(SPRN_SIER); |
|
|
|
if (cpu_has_feature(CPU_FTR_ARCH_31)) { |
|
vcpu->arch.mmcr[3] = mfspr(SPRN_MMCR3); |
|
vcpu->arch.sier[1] = mfspr(SPRN_SIER2); |
|
vcpu->arch.sier[2] = mfspr(SPRN_SIER3); |
|
} |
|
|
|
} else if (vcpu->arch.hfscr & HFSCR_PM) { |
|
/* |
|
* The guest accessed PMC SPRs without specifying they should |
|
* be preserved, or it cleared pmcregs_in_use after the last |
|
* access. Just ensure they are frozen. |
|
*/ |
|
freeze_pmu(mfspr(SPRN_MMCR0), mfspr(SPRN_MMCRA)); |
|
|
|
/* |
|
* Demand-fault PMU register access in the guest. |
|
* |
|
* This is used to grab the guest's VPA pmcregs_in_use value |
|
* and reflect it into the host's VPA in the case of a nested |
|
* hypervisor. |
|
* |
|
* It also avoids having to zero-out SPRs after each guest |
|
* exit to avoid side-channels when. |
|
* |
|
* This is cleared here when we exit the guest, so later HFSCR |
|
* interrupt handling can add it back to run the guest with |
|
* PM enabled next time. |
|
*/ |
|
if (!vcpu->arch.nested) |
|
vcpu->arch.hfscr &= ~HFSCR_PM; |
|
} /* otherwise the PMU should still be frozen */ |
|
|
|
#ifdef CONFIG_PPC_PSERIES |
|
if (kvmhv_on_pseries()) { |
|
barrier(); |
|
get_lppaca()->pmcregs_in_use = ppc_get_pmu_inuse(); |
|
barrier(); |
|
} |
|
#endif |
|
|
|
if (ppc_get_pmu_inuse()) { |
|
mtspr(SPRN_PMC1, host_os_sprs->pmc1); |
|
mtspr(SPRN_PMC2, host_os_sprs->pmc2); |
|
mtspr(SPRN_PMC3, host_os_sprs->pmc3); |
|
mtspr(SPRN_PMC4, host_os_sprs->pmc4); |
|
mtspr(SPRN_PMC5, host_os_sprs->pmc5); |
|
mtspr(SPRN_PMC6, host_os_sprs->pmc6); |
|
mtspr(SPRN_MMCR1, host_os_sprs->mmcr1); |
|
mtspr(SPRN_MMCR2, host_os_sprs->mmcr2); |
|
mtspr(SPRN_SDAR, host_os_sprs->sdar); |
|
mtspr(SPRN_SIAR, host_os_sprs->siar); |
|
mtspr(SPRN_SIER, host_os_sprs->sier1); |
|
|
|
if (cpu_has_feature(CPU_FTR_ARCH_31)) { |
|
mtspr(SPRN_MMCR3, host_os_sprs->mmcr3); |
|
mtspr(SPRN_SIER2, host_os_sprs->sier2); |
|
mtspr(SPRN_SIER3, host_os_sprs->sier3); |
|
} |
|
|
|
/* Set MMCRA then MMCR0 last */ |
|
mtspr(SPRN_MMCRA, host_os_sprs->mmcra); |
|
mtspr(SPRN_MMCR0, host_os_sprs->mmcr0); |
|
isync(); |
|
} |
|
} |
|
EXPORT_SYMBOL_GPL(switch_pmu_to_host);
|
|
|