73b0195416
Coverity complained about the possible out-of-bounds access with counter_virt/counter_virt_prev because these two arrays are accessed with privilege mode. However, these two arrays are accessed only when virt is enabled. Thus, the privilege mode can't be M mode. Add the asserts anyways to detect any wrong usage of these arrays in the future. Suggested-by: Peter Maydell <peter.maydell@linaro.org> Signed-off-by: Atish Patra <atishp@rivosinc.com> Fixes: Coverity CID 1558459 Fixes: Coverity CID 1558462 Reviewed-by: Alistair Francis <alistair.francis@wdc.com> Message-ID: <20240724-fixes-v1-1-4a64596b0d64@rivosinc.com> Signed-off-by: Alistair Francis <alistair.francis@wdc.com>
603 lines
18 KiB
C
603 lines
18 KiB
C
/*
|
|
* RISC-V PMU file.
|
|
*
|
|
* Copyright (c) 2021 Western Digital Corporation or its affiliates.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2 or later, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/log.h"
|
|
#include "qemu/error-report.h"
|
|
#include "qemu/timer.h"
|
|
#include "cpu.h"
|
|
#include "pmu.h"
|
|
#include "sysemu/cpu-timers.h"
|
|
#include "sysemu/device_tree.h"
|
|
|
|
#define RISCV_TIMEBASE_FREQ 1000000000 /* 1Ghz */
|
|
|
|
/*
|
|
* To keep it simple, any event can be mapped to any programmable counters in
|
|
* QEMU. The generic cycle & instruction count events can also be monitored
|
|
* using programmable counters. In that case, mcycle & minstret must continue
|
|
* to provide the correct value as well. Heterogeneous PMU per hart is not
|
|
* supported yet. Thus, number of counters are same across all harts.
|
|
*/
|
|
void riscv_pmu_generate_fdt_node(void *fdt, uint32_t cmask, char *pmu_name)
|
|
{
|
|
uint32_t fdt_event_ctr_map[15] = {};
|
|
|
|
/*
|
|
* The event encoding is specified in the SBI specification
|
|
* Event idx is a 20bits wide number encoded as follows:
|
|
* event_idx[19:16] = type
|
|
* event_idx[15:0] = code
|
|
* The code field in cache events are encoded as follows:
|
|
* event_idx.code[15:3] = cache_id
|
|
* event_idx.code[2:1] = op_id
|
|
* event_idx.code[0:0] = result_id
|
|
*/
|
|
|
|
/* SBI_PMU_HW_CPU_CYCLES: 0x01 : type(0x00) */
|
|
fdt_event_ctr_map[0] = cpu_to_be32(0x00000001);
|
|
fdt_event_ctr_map[1] = cpu_to_be32(0x00000001);
|
|
fdt_event_ctr_map[2] = cpu_to_be32(cmask | 1 << 0);
|
|
|
|
/* SBI_PMU_HW_INSTRUCTIONS: 0x02 : type(0x00) */
|
|
fdt_event_ctr_map[3] = cpu_to_be32(0x00000002);
|
|
fdt_event_ctr_map[4] = cpu_to_be32(0x00000002);
|
|
fdt_event_ctr_map[5] = cpu_to_be32(cmask | 1 << 2);
|
|
|
|
/* SBI_PMU_HW_CACHE_DTLB : 0x03 READ : 0x00 MISS : 0x00 type(0x01) */
|
|
fdt_event_ctr_map[6] = cpu_to_be32(0x00010019);
|
|
fdt_event_ctr_map[7] = cpu_to_be32(0x00010019);
|
|
fdt_event_ctr_map[8] = cpu_to_be32(cmask);
|
|
|
|
/* SBI_PMU_HW_CACHE_DTLB : 0x03 WRITE : 0x01 MISS : 0x00 type(0x01) */
|
|
fdt_event_ctr_map[9] = cpu_to_be32(0x0001001B);
|
|
fdt_event_ctr_map[10] = cpu_to_be32(0x0001001B);
|
|
fdt_event_ctr_map[11] = cpu_to_be32(cmask);
|
|
|
|
/* SBI_PMU_HW_CACHE_ITLB : 0x04 READ : 0x00 MISS : 0x00 type(0x01) */
|
|
fdt_event_ctr_map[12] = cpu_to_be32(0x00010021);
|
|
fdt_event_ctr_map[13] = cpu_to_be32(0x00010021);
|
|
fdt_event_ctr_map[14] = cpu_to_be32(cmask);
|
|
|
|
/* This a OpenSBI specific DT property documented in OpenSBI docs */
|
|
qemu_fdt_setprop(fdt, pmu_name, "riscv,event-to-mhpmcounters",
|
|
fdt_event_ctr_map, sizeof(fdt_event_ctr_map));
|
|
}
|
|
|
|
static bool riscv_pmu_counter_valid(RISCVCPU *cpu, uint32_t ctr_idx)
|
|
{
|
|
if (ctr_idx < 3 || ctr_idx >= RV_MAX_MHPMCOUNTERS ||
|
|
!(cpu->pmu_avail_ctrs & BIT(ctr_idx))) {
|
|
return false;
|
|
} else {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
static bool riscv_pmu_counter_enabled(RISCVCPU *cpu, uint32_t ctr_idx)
|
|
{
|
|
CPURISCVState *env = &cpu->env;
|
|
|
|
if (riscv_pmu_counter_valid(cpu, ctr_idx) &&
|
|
!get_field(env->mcountinhibit, BIT(ctr_idx))) {
|
|
return true;
|
|
} else {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static int riscv_pmu_incr_ctr_rv32(RISCVCPU *cpu, uint32_t ctr_idx)
|
|
{
|
|
CPURISCVState *env = &cpu->env;
|
|
target_ulong max_val = UINT32_MAX;
|
|
PMUCTRState *counter = &env->pmu_ctrs[ctr_idx];
|
|
bool virt_on = env->virt_enabled;
|
|
|
|
/* Privilege mode filtering */
|
|
if ((env->priv == PRV_M &&
|
|
(env->mhpmeventh_val[ctr_idx] & MHPMEVENTH_BIT_MINH)) ||
|
|
(env->priv == PRV_S && virt_on &&
|
|
(env->mhpmeventh_val[ctr_idx] & MHPMEVENTH_BIT_VSINH)) ||
|
|
(env->priv == PRV_U && virt_on &&
|
|
(env->mhpmeventh_val[ctr_idx] & MHPMEVENTH_BIT_VUINH)) ||
|
|
(env->priv == PRV_S && !virt_on &&
|
|
(env->mhpmeventh_val[ctr_idx] & MHPMEVENTH_BIT_SINH)) ||
|
|
(env->priv == PRV_U && !virt_on &&
|
|
(env->mhpmeventh_val[ctr_idx] & MHPMEVENTH_BIT_UINH))) {
|
|
return 0;
|
|
}
|
|
|
|
/* Handle the overflow scenario */
|
|
if (counter->mhpmcounter_val == max_val) {
|
|
if (counter->mhpmcounterh_val == max_val) {
|
|
counter->mhpmcounter_val = 0;
|
|
counter->mhpmcounterh_val = 0;
|
|
/* Generate interrupt only if OF bit is clear */
|
|
if (!(env->mhpmeventh_val[ctr_idx] & MHPMEVENTH_BIT_OF)) {
|
|
env->mhpmeventh_val[ctr_idx] |= MHPMEVENTH_BIT_OF;
|
|
riscv_cpu_update_mip(env, MIP_LCOFIP, BOOL_TO_MASK(1));
|
|
}
|
|
} else {
|
|
counter->mhpmcounterh_val++;
|
|
}
|
|
} else {
|
|
counter->mhpmcounter_val++;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int riscv_pmu_incr_ctr_rv64(RISCVCPU *cpu, uint32_t ctr_idx)
|
|
{
|
|
CPURISCVState *env = &cpu->env;
|
|
PMUCTRState *counter = &env->pmu_ctrs[ctr_idx];
|
|
uint64_t max_val = UINT64_MAX;
|
|
bool virt_on = env->virt_enabled;
|
|
|
|
/* Privilege mode filtering */
|
|
if ((env->priv == PRV_M &&
|
|
(env->mhpmevent_val[ctr_idx] & MHPMEVENT_BIT_MINH)) ||
|
|
(env->priv == PRV_S && virt_on &&
|
|
(env->mhpmevent_val[ctr_idx] & MHPMEVENT_BIT_VSINH)) ||
|
|
(env->priv == PRV_U && virt_on &&
|
|
(env->mhpmevent_val[ctr_idx] & MHPMEVENT_BIT_VUINH)) ||
|
|
(env->priv == PRV_S && !virt_on &&
|
|
(env->mhpmevent_val[ctr_idx] & MHPMEVENT_BIT_SINH)) ||
|
|
(env->priv == PRV_U && !virt_on &&
|
|
(env->mhpmevent_val[ctr_idx] & MHPMEVENT_BIT_UINH))) {
|
|
return 0;
|
|
}
|
|
|
|
/* Handle the overflow scenario */
|
|
if (counter->mhpmcounter_val == max_val) {
|
|
counter->mhpmcounter_val = 0;
|
|
/* Generate interrupt only if OF bit is clear */
|
|
if (!(env->mhpmevent_val[ctr_idx] & MHPMEVENT_BIT_OF)) {
|
|
env->mhpmevent_val[ctr_idx] |= MHPMEVENT_BIT_OF;
|
|
riscv_cpu_update_mip(env, MIP_LCOFIP, BOOL_TO_MASK(1));
|
|
}
|
|
} else {
|
|
counter->mhpmcounter_val++;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Information needed to update counters:
|
|
* new_priv, new_virt: To correctly save starting snapshot for the newly
|
|
* started mode. Look at array being indexed with newprv.
|
|
* old_priv, old_virt: To correctly select previous snapshot for old priv
|
|
* and compute delta. Also to select correct counter
|
|
* to inc. Look at arrays being indexed with env->priv.
|
|
*
|
|
* To avoid the complexity of calling this function, we assume that
|
|
* env->priv and env->virt_enabled contain old priv and old virt and
|
|
* new priv and new virt values are passed in as arguments.
|
|
*/
|
|
static void riscv_pmu_icount_update_priv(CPURISCVState *env,
|
|
target_ulong newpriv, bool new_virt)
|
|
{
|
|
uint64_t *snapshot_prev, *snapshot_new;
|
|
uint64_t current_icount;
|
|
uint64_t *counter_arr;
|
|
uint64_t delta;
|
|
|
|
if (icount_enabled()) {
|
|
current_icount = icount_get_raw();
|
|
} else {
|
|
current_icount = cpu_get_host_ticks();
|
|
}
|
|
|
|
if (env->virt_enabled) {
|
|
g_assert(env->priv <= PRV_S);
|
|
counter_arr = env->pmu_fixed_ctrs[1].counter_virt;
|
|
snapshot_prev = env->pmu_fixed_ctrs[1].counter_virt_prev;
|
|
} else {
|
|
counter_arr = env->pmu_fixed_ctrs[1].counter;
|
|
snapshot_prev = env->pmu_fixed_ctrs[1].counter_prev;
|
|
}
|
|
|
|
if (new_virt) {
|
|
g_assert(newpriv <= PRV_S);
|
|
snapshot_new = env->pmu_fixed_ctrs[1].counter_virt_prev;
|
|
} else {
|
|
snapshot_new = env->pmu_fixed_ctrs[1].counter_prev;
|
|
}
|
|
|
|
/*
|
|
* new_priv can be same as env->priv. So we need to calculate
|
|
* delta first before updating snapshot_new[new_priv].
|
|
*/
|
|
delta = current_icount - snapshot_prev[env->priv];
|
|
snapshot_new[newpriv] = current_icount;
|
|
|
|
counter_arr[env->priv] += delta;
|
|
}
|
|
|
|
static void riscv_pmu_cycle_update_priv(CPURISCVState *env,
|
|
target_ulong newpriv, bool new_virt)
|
|
{
|
|
uint64_t *snapshot_prev, *snapshot_new;
|
|
uint64_t current_ticks;
|
|
uint64_t *counter_arr;
|
|
uint64_t delta;
|
|
|
|
if (icount_enabled()) {
|
|
current_ticks = icount_get();
|
|
} else {
|
|
current_ticks = cpu_get_host_ticks();
|
|
}
|
|
|
|
if (env->virt_enabled) {
|
|
g_assert(env->priv <= PRV_S);
|
|
counter_arr = env->pmu_fixed_ctrs[0].counter_virt;
|
|
snapshot_prev = env->pmu_fixed_ctrs[0].counter_virt_prev;
|
|
} else {
|
|
counter_arr = env->pmu_fixed_ctrs[0].counter;
|
|
snapshot_prev = env->pmu_fixed_ctrs[0].counter_prev;
|
|
}
|
|
|
|
if (new_virt) {
|
|
g_assert(newpriv <= PRV_S);
|
|
snapshot_new = env->pmu_fixed_ctrs[0].counter_virt_prev;
|
|
} else {
|
|
snapshot_new = env->pmu_fixed_ctrs[0].counter_prev;
|
|
}
|
|
|
|
delta = current_ticks - snapshot_prev[env->priv];
|
|
snapshot_new[newpriv] = current_ticks;
|
|
|
|
counter_arr[env->priv] += delta;
|
|
}
|
|
|
|
void riscv_pmu_update_fixed_ctrs(CPURISCVState *env, target_ulong newpriv,
|
|
bool new_virt)
|
|
{
|
|
riscv_pmu_cycle_update_priv(env, newpriv, new_virt);
|
|
riscv_pmu_icount_update_priv(env, newpriv, new_virt);
|
|
}
|
|
|
|
int riscv_pmu_incr_ctr(RISCVCPU *cpu, enum riscv_pmu_event_idx event_idx)
|
|
{
|
|
uint32_t ctr_idx;
|
|
int ret;
|
|
CPURISCVState *env = &cpu->env;
|
|
gpointer value;
|
|
|
|
if (!cpu->cfg.pmu_mask) {
|
|
return 0;
|
|
}
|
|
value = g_hash_table_lookup(cpu->pmu_event_ctr_map,
|
|
GUINT_TO_POINTER(event_idx));
|
|
if (!value) {
|
|
return -1;
|
|
}
|
|
|
|
ctr_idx = GPOINTER_TO_UINT(value);
|
|
if (!riscv_pmu_counter_enabled(cpu, ctr_idx)) {
|
|
return -1;
|
|
}
|
|
|
|
if (riscv_cpu_mxl(env) == MXL_RV32) {
|
|
ret = riscv_pmu_incr_ctr_rv32(cpu, ctr_idx);
|
|
} else {
|
|
ret = riscv_pmu_incr_ctr_rv64(cpu, ctr_idx);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
bool riscv_pmu_ctr_monitor_instructions(CPURISCVState *env,
|
|
uint32_t target_ctr)
|
|
{
|
|
RISCVCPU *cpu;
|
|
uint32_t event_idx;
|
|
uint32_t ctr_idx;
|
|
|
|
/* Fixed instret counter */
|
|
if (target_ctr == 2) {
|
|
return true;
|
|
}
|
|
|
|
cpu = env_archcpu(env);
|
|
if (!cpu->pmu_event_ctr_map) {
|
|
return false;
|
|
}
|
|
|
|
event_idx = RISCV_PMU_EVENT_HW_INSTRUCTIONS;
|
|
ctr_idx = GPOINTER_TO_UINT(g_hash_table_lookup(cpu->pmu_event_ctr_map,
|
|
GUINT_TO_POINTER(event_idx)));
|
|
if (!ctr_idx) {
|
|
return false;
|
|
}
|
|
|
|
return target_ctr == ctr_idx ? true : false;
|
|
}
|
|
|
|
bool riscv_pmu_ctr_monitor_cycles(CPURISCVState *env, uint32_t target_ctr)
|
|
{
|
|
RISCVCPU *cpu;
|
|
uint32_t event_idx;
|
|
uint32_t ctr_idx;
|
|
|
|
/* Fixed mcycle counter */
|
|
if (target_ctr == 0) {
|
|
return true;
|
|
}
|
|
|
|
cpu = env_archcpu(env);
|
|
if (!cpu->pmu_event_ctr_map) {
|
|
return false;
|
|
}
|
|
|
|
event_idx = RISCV_PMU_EVENT_HW_CPU_CYCLES;
|
|
ctr_idx = GPOINTER_TO_UINT(g_hash_table_lookup(cpu->pmu_event_ctr_map,
|
|
GUINT_TO_POINTER(event_idx)));
|
|
|
|
/* Counter zero is not used for event_ctr_map */
|
|
if (!ctr_idx) {
|
|
return false;
|
|
}
|
|
|
|
return (target_ctr == ctr_idx) ? true : false;
|
|
}
|
|
|
|
static gboolean pmu_remove_event_map(gpointer key, gpointer value,
|
|
gpointer udata)
|
|
{
|
|
return (GPOINTER_TO_UINT(value) == GPOINTER_TO_UINT(udata)) ? true : false;
|
|
}
|
|
|
|
static int64_t pmu_icount_ticks_to_ns(int64_t value)
|
|
{
|
|
int64_t ret = 0;
|
|
|
|
if (icount_enabled()) {
|
|
ret = icount_to_ns(value);
|
|
} else {
|
|
ret = (NANOSECONDS_PER_SECOND / RISCV_TIMEBASE_FREQ) * value;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int riscv_pmu_update_event_map(CPURISCVState *env, uint64_t value,
|
|
uint32_t ctr_idx)
|
|
{
|
|
uint32_t event_idx;
|
|
RISCVCPU *cpu = env_archcpu(env);
|
|
|
|
if (!riscv_pmu_counter_valid(cpu, ctr_idx) || !cpu->pmu_event_ctr_map) {
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* Expected mhpmevent value is zero for reset case. Remove the current
|
|
* mapping.
|
|
*/
|
|
if (!value) {
|
|
g_hash_table_foreach_remove(cpu->pmu_event_ctr_map,
|
|
pmu_remove_event_map,
|
|
GUINT_TO_POINTER(ctr_idx));
|
|
return 0;
|
|
}
|
|
|
|
event_idx = value & MHPMEVENT_IDX_MASK;
|
|
if (g_hash_table_lookup(cpu->pmu_event_ctr_map,
|
|
GUINT_TO_POINTER(event_idx))) {
|
|
return 0;
|
|
}
|
|
|
|
switch (event_idx) {
|
|
case RISCV_PMU_EVENT_HW_CPU_CYCLES:
|
|
case RISCV_PMU_EVENT_HW_INSTRUCTIONS:
|
|
case RISCV_PMU_EVENT_CACHE_DTLB_READ_MISS:
|
|
case RISCV_PMU_EVENT_CACHE_DTLB_WRITE_MISS:
|
|
case RISCV_PMU_EVENT_CACHE_ITLB_PREFETCH_MISS:
|
|
break;
|
|
default:
|
|
/* We don't support any raw events right now */
|
|
return -1;
|
|
}
|
|
g_hash_table_insert(cpu->pmu_event_ctr_map, GUINT_TO_POINTER(event_idx),
|
|
GUINT_TO_POINTER(ctr_idx));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool pmu_hpmevent_is_of_set(CPURISCVState *env, uint32_t ctr_idx)
|
|
{
|
|
target_ulong mhpmevent_val;
|
|
uint64_t of_bit_mask;
|
|
|
|
if (riscv_cpu_mxl(env) == MXL_RV32) {
|
|
mhpmevent_val = env->mhpmeventh_val[ctr_idx];
|
|
of_bit_mask = MHPMEVENTH_BIT_OF;
|
|
} else {
|
|
mhpmevent_val = env->mhpmevent_val[ctr_idx];
|
|
of_bit_mask = MHPMEVENT_BIT_OF;
|
|
}
|
|
|
|
return get_field(mhpmevent_val, of_bit_mask);
|
|
}
|
|
|
|
static bool pmu_hpmevent_set_of_if_clear(CPURISCVState *env, uint32_t ctr_idx)
|
|
{
|
|
target_ulong *mhpmevent_val;
|
|
uint64_t of_bit_mask;
|
|
|
|
if (riscv_cpu_mxl(env) == MXL_RV32) {
|
|
mhpmevent_val = &env->mhpmeventh_val[ctr_idx];
|
|
of_bit_mask = MHPMEVENTH_BIT_OF;
|
|
} else {
|
|
mhpmevent_val = &env->mhpmevent_val[ctr_idx];
|
|
of_bit_mask = MHPMEVENT_BIT_OF;
|
|
}
|
|
|
|
if (!get_field(*mhpmevent_val, of_bit_mask)) {
|
|
*mhpmevent_val |= of_bit_mask;
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
static void pmu_timer_trigger_irq(RISCVCPU *cpu,
|
|
enum riscv_pmu_event_idx evt_idx)
|
|
{
|
|
uint32_t ctr_idx;
|
|
CPURISCVState *env = &cpu->env;
|
|
PMUCTRState *counter;
|
|
int64_t irq_trigger_at;
|
|
uint64_t curr_ctr_val, curr_ctrh_val;
|
|
uint64_t ctr_val;
|
|
|
|
if (evt_idx != RISCV_PMU_EVENT_HW_CPU_CYCLES &&
|
|
evt_idx != RISCV_PMU_EVENT_HW_INSTRUCTIONS) {
|
|
return;
|
|
}
|
|
|
|
ctr_idx = GPOINTER_TO_UINT(g_hash_table_lookup(cpu->pmu_event_ctr_map,
|
|
GUINT_TO_POINTER(evt_idx)));
|
|
if (!riscv_pmu_counter_enabled(cpu, ctr_idx)) {
|
|
return;
|
|
}
|
|
|
|
/* Generate interrupt only if OF bit is clear */
|
|
if (pmu_hpmevent_is_of_set(env, ctr_idx)) {
|
|
return;
|
|
}
|
|
|
|
counter = &env->pmu_ctrs[ctr_idx];
|
|
if (counter->irq_overflow_left > 0) {
|
|
irq_trigger_at = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) +
|
|
counter->irq_overflow_left;
|
|
timer_mod_anticipate_ns(cpu->pmu_timer, irq_trigger_at);
|
|
counter->irq_overflow_left = 0;
|
|
return;
|
|
}
|
|
|
|
riscv_pmu_read_ctr(env, (target_ulong *)&curr_ctr_val, false, ctr_idx);
|
|
ctr_val = counter->mhpmcounter_val;
|
|
if (riscv_cpu_mxl(env) == MXL_RV32) {
|
|
riscv_pmu_read_ctr(env, (target_ulong *)&curr_ctrh_val, true, ctr_idx);
|
|
curr_ctr_val = curr_ctr_val | (curr_ctrh_val << 32);
|
|
ctr_val = ctr_val |
|
|
((uint64_t)counter->mhpmcounterh_val << 32);
|
|
}
|
|
|
|
/*
|
|
* We can not accommodate for inhibited modes when setting up timer. Check
|
|
* if the counter has actually overflowed or not by comparing current
|
|
* counter value (accommodated for inhibited modes) with software written
|
|
* counter value.
|
|
*/
|
|
if (curr_ctr_val >= ctr_val) {
|
|
riscv_pmu_setup_timer(env, curr_ctr_val, ctr_idx);
|
|
return;
|
|
}
|
|
|
|
if (cpu->pmu_avail_ctrs & BIT(ctr_idx)) {
|
|
if (pmu_hpmevent_set_of_if_clear(env, ctr_idx)) {
|
|
riscv_cpu_update_mip(env, MIP_LCOFIP, BOOL_TO_MASK(1));
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Timer callback for instret and cycle counter overflow */
|
|
void riscv_pmu_timer_cb(void *priv)
|
|
{
|
|
RISCVCPU *cpu = priv;
|
|
|
|
/* Timer event was triggered only for these events */
|
|
pmu_timer_trigger_irq(cpu, RISCV_PMU_EVENT_HW_CPU_CYCLES);
|
|
pmu_timer_trigger_irq(cpu, RISCV_PMU_EVENT_HW_INSTRUCTIONS);
|
|
}
|
|
|
|
int riscv_pmu_setup_timer(CPURISCVState *env, uint64_t value, uint32_t ctr_idx)
|
|
{
|
|
uint64_t overflow_delta, overflow_at, curr_ns;
|
|
int64_t overflow_ns, overflow_left = 0;
|
|
RISCVCPU *cpu = env_archcpu(env);
|
|
PMUCTRState *counter = &env->pmu_ctrs[ctr_idx];
|
|
|
|
/* No need to setup a timer if LCOFI is disabled when OF is set */
|
|
if (!riscv_pmu_counter_valid(cpu, ctr_idx) || !cpu->cfg.ext_sscofpmf ||
|
|
pmu_hpmevent_is_of_set(env, ctr_idx)) {
|
|
return -1;
|
|
}
|
|
|
|
if (value) {
|
|
overflow_delta = UINT64_MAX - value + 1;
|
|
} else {
|
|
overflow_delta = UINT64_MAX;
|
|
}
|
|
|
|
/*
|
|
* QEMU supports only int64_t timers while RISC-V counters are uint64_t.
|
|
* Compute the leftover and save it so that it can be reprogrammed again
|
|
* when timer expires.
|
|
*/
|
|
if (overflow_delta > INT64_MAX) {
|
|
overflow_left = overflow_delta - INT64_MAX;
|
|
}
|
|
|
|
if (riscv_pmu_ctr_monitor_cycles(env, ctr_idx) ||
|
|
riscv_pmu_ctr_monitor_instructions(env, ctr_idx)) {
|
|
overflow_ns = pmu_icount_ticks_to_ns((int64_t)overflow_delta);
|
|
overflow_left = pmu_icount_ticks_to_ns(overflow_left) ;
|
|
} else {
|
|
return -1;
|
|
}
|
|
curr_ns = (uint64_t)qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL);
|
|
overflow_at = curr_ns + overflow_ns;
|
|
if (overflow_at <= curr_ns)
|
|
overflow_at = UINT64_MAX;
|
|
|
|
if (overflow_at > INT64_MAX) {
|
|
overflow_left += overflow_at - INT64_MAX;
|
|
counter->irq_overflow_left = overflow_left;
|
|
overflow_at = INT64_MAX;
|
|
}
|
|
timer_mod_anticipate_ns(cpu->pmu_timer, overflow_at);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
void riscv_pmu_init(RISCVCPU *cpu, Error **errp)
|
|
{
|
|
if (cpu->cfg.pmu_mask & (COUNTEREN_CY | COUNTEREN_TM | COUNTEREN_IR)) {
|
|
error_setg(errp, "\"pmu-mask\" contains invalid bits (0-2) set");
|
|
return;
|
|
}
|
|
|
|
if (ctpop32(cpu->cfg.pmu_mask) > (RV_MAX_MHPMCOUNTERS - 3)) {
|
|
error_setg(errp, "Number of counters exceeds maximum available");
|
|
return;
|
|
}
|
|
|
|
cpu->pmu_event_ctr_map = g_hash_table_new(g_direct_hash, g_direct_equal);
|
|
if (!cpu->pmu_event_ctr_map) {
|
|
error_setg(errp, "Unable to allocate PMU event hash table");
|
|
return;
|
|
}
|
|
|
|
cpu->pmu_avail_ctrs = cpu->cfg.pmu_mask;
|
|
}
|