mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-10-02 23:27:06 +00:00
arm64: KVM: Free perf event of PMU when destroying vcpu
When KVM frees VCPU, it needs to free the perf_event of PMU. Signed-off-by: Shannon Zhao <shannon.zhao@linaro.org> Reviewed-by: Marc Zyngier <marc.zyngier@arm.com> Reviewed-by: Andrew Jones <drjones@redhat.com> Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
This commit is contained in:
parent
2aa36e9840
commit
5f0a714a2b
3 changed files with 24 additions and 0 deletions
|
@ -266,6 +266,7 @@ void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu)
|
||||||
kvm_mmu_free_memory_caches(vcpu);
|
kvm_mmu_free_memory_caches(vcpu);
|
||||||
kvm_timer_vcpu_terminate(vcpu);
|
kvm_timer_vcpu_terminate(vcpu);
|
||||||
kvm_vgic_vcpu_destroy(vcpu);
|
kvm_vgic_vcpu_destroy(vcpu);
|
||||||
|
kvm_pmu_vcpu_destroy(vcpu);
|
||||||
kmem_cache_free(kvm_vcpu_cache, vcpu);
|
kmem_cache_free(kvm_vcpu_cache, vcpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -43,6 +43,7 @@ u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx);
|
||||||
void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
|
void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
|
||||||
u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu);
|
u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu);
|
||||||
void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu);
|
void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu);
|
||||||
|
void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu);
|
||||||
void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 val);
|
void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 val);
|
||||||
void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 val);
|
void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 val);
|
||||||
void kvm_pmu_overflow_set(struct kvm_vcpu *vcpu, u64 val);
|
void kvm_pmu_overflow_set(struct kvm_vcpu *vcpu, u64 val);
|
||||||
|
@ -69,6 +70,7 @@ static inline u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
static inline void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) {}
|
static inline void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu) {}
|
||||||
|
static inline void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) {}
|
||||||
static inline void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 val) {}
|
static inline void kvm_pmu_disable_counter(struct kvm_vcpu *vcpu, u64 val) {}
|
||||||
static inline void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 val) {}
|
static inline void kvm_pmu_enable_counter(struct kvm_vcpu *vcpu, u64 val) {}
|
||||||
static inline void kvm_pmu_overflow_set(struct kvm_vcpu *vcpu, u64 val) {}
|
static inline void kvm_pmu_overflow_set(struct kvm_vcpu *vcpu, u64 val) {}
|
||||||
|
|
|
@ -101,6 +101,27 @@ void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* kvm_pmu_vcpu_destroy - free perf event of PMU for cpu
|
||||||
|
* @vcpu: The vcpu pointer
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
struct kvm_pmu *pmu = &vcpu->arch.pmu;
|
||||||
|
|
||||||
|
for (i = 0; i < ARMV8_PMU_MAX_COUNTERS; i++) {
|
||||||
|
struct kvm_pmc *pmc = &pmu->pmc[i];
|
||||||
|
|
||||||
|
if (pmc->perf_event) {
|
||||||
|
perf_event_disable(pmc->perf_event);
|
||||||
|
perf_event_release_kernel(pmc->perf_event);
|
||||||
|
pmc->perf_event = NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu)
|
u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
u64 val = vcpu_sys_reg(vcpu, PMCR_EL0) >> ARMV8_PMU_PMCR_N_SHIFT;
|
u64 val = vcpu_sys_reg(vcpu, PMCR_EL0) >> ARMV8_PMU_PMCR_N_SHIFT;
|
||||||
|
|
Loading…
Reference in a new issue