acf949411f
Make the BASEPRI register banked if v8M security extensions are enabled. Note that we do not yet implement the functionality of the new AIRCR.PRIS bit (which allows the effect of the NS copy of BASEPRI to be restricted). Signed-off-by: Peter Maydell <peter.maydell@linaro.org> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Message-id: 1503414539-28762-7-git-send-email-peter.maydell@linaro.org
1229 lines
39 KiB
C
1229 lines
39 KiB
C
/*
|
|
* ARM Nested Vectored Interrupt Controller
|
|
*
|
|
* Copyright (c) 2006-2007 CodeSourcery.
|
|
* Written by Paul Brook
|
|
*
|
|
* This code is licensed under the GPL.
|
|
*
|
|
* The ARMv7M System controller is fairly tightly tied in with the
|
|
* NVIC. Much of that is also implemented here.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qapi/error.h"
|
|
#include "qemu-common.h"
|
|
#include "cpu.h"
|
|
#include "hw/sysbus.h"
|
|
#include "qemu/timer.h"
|
|
#include "hw/arm/arm.h"
|
|
#include "hw/intc/armv7m_nvic.h"
|
|
#include "target/arm/cpu.h"
|
|
#include "exec/exec-all.h"
|
|
#include "qemu/log.h"
|
|
#include "trace.h"
|
|
|
|
/* IRQ number counting:
|
|
*
|
|
* the num-irq property counts the number of external IRQ lines
|
|
*
|
|
* NVICState::num_irq counts the total number of exceptions
|
|
* (external IRQs, the 15 internal exceptions including reset,
|
|
* and one for the unused exception number 0).
|
|
*
|
|
* NVIC_MAX_IRQ is the highest permitted number of external IRQ lines.
|
|
*
|
|
* NVIC_MAX_VECTORS is the highest permitted number of exceptions.
|
|
*
|
|
* Iterating through all exceptions should typically be done with
|
|
* for (i = 1; i < s->num_irq; i++) to avoid the unused slot 0.
|
|
*
|
|
* The external qemu_irq lines are the NVIC's external IRQ lines,
|
|
* so line 0 is exception 16.
|
|
*
|
|
* In the terminology of the architecture manual, "interrupts" are
|
|
* a subcategory of exception referring to the external interrupts
|
|
* (which are exception numbers NVIC_FIRST_IRQ and upward).
|
|
* For historical reasons QEMU tends to use "interrupt" and
|
|
* "exception" more or less interchangeably.
|
|
*/
|
|
#define NVIC_FIRST_IRQ 16
|
|
#define NVIC_MAX_IRQ (NVIC_MAX_VECTORS - NVIC_FIRST_IRQ)
|
|
|
|
/* Effective running priority of the CPU when no exception is active
|
|
* (higher than the highest possible priority value)
|
|
*/
|
|
#define NVIC_NOEXC_PRIO 0x100
|
|
|
|
static const uint8_t nvic_id[] = {
|
|
0x00, 0xb0, 0x1b, 0x00, 0x0d, 0xe0, 0x05, 0xb1
|
|
};
|
|
|
|
static int nvic_pending_prio(NVICState *s)
|
|
{
|
|
/* return the priority of the current pending interrupt,
|
|
* or NVIC_NOEXC_PRIO if no interrupt is pending
|
|
*/
|
|
return s->vectpending ? s->vectors[s->vectpending].prio : NVIC_NOEXC_PRIO;
|
|
}
|
|
|
|
/* Return the value of the ISCR RETTOBASE bit:
|
|
* 1 if there is exactly one active exception
|
|
* 0 if there is more than one active exception
|
|
* UNKNOWN if there are no active exceptions (we choose 1,
|
|
* which matches the choice Cortex-M3 is documented as making).
|
|
*
|
|
* NB: some versions of the documentation talk about this
|
|
* counting "active exceptions other than the one shown by IPSR";
|
|
* this is only different in the obscure corner case where guest
|
|
* code has manually deactivated an exception and is about
|
|
* to fail an exception-return integrity check. The definition
|
|
* above is the one from the v8M ARM ARM and is also in line
|
|
* with the behaviour documented for the Cortex-M3.
|
|
*/
|
|
static bool nvic_rettobase(NVICState *s)
|
|
{
|
|
int irq, nhand = 0;
|
|
|
|
for (irq = ARMV7M_EXCP_RESET; irq < s->num_irq; irq++) {
|
|
if (s->vectors[irq].active) {
|
|
nhand++;
|
|
if (nhand == 2) {
|
|
return 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/* Return the value of the ISCR ISRPENDING bit:
|
|
* 1 if an external interrupt is pending
|
|
* 0 if no external interrupt is pending
|
|
*/
|
|
static bool nvic_isrpending(NVICState *s)
|
|
{
|
|
int irq;
|
|
|
|
/* We can shortcut if the highest priority pending interrupt
|
|
* happens to be external or if there is nothing pending.
|
|
*/
|
|
if (s->vectpending > NVIC_FIRST_IRQ) {
|
|
return true;
|
|
}
|
|
if (s->vectpending == 0) {
|
|
return false;
|
|
}
|
|
|
|
for (irq = NVIC_FIRST_IRQ; irq < s->num_irq; irq++) {
|
|
if (s->vectors[irq].pending) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/* Return a mask word which clears the subpriority bits from
|
|
* a priority value for an M-profile exception, leaving only
|
|
* the group priority.
|
|
*/
|
|
static inline uint32_t nvic_gprio_mask(NVICState *s)
|
|
{
|
|
return ~0U << (s->prigroup + 1);
|
|
}
|
|
|
|
/* Recompute vectpending and exception_prio */
|
|
static void nvic_recompute_state(NVICState *s)
|
|
{
|
|
int i;
|
|
int pend_prio = NVIC_NOEXC_PRIO;
|
|
int active_prio = NVIC_NOEXC_PRIO;
|
|
int pend_irq = 0;
|
|
|
|
for (i = 1; i < s->num_irq; i++) {
|
|
VecInfo *vec = &s->vectors[i];
|
|
|
|
if (vec->enabled && vec->pending && vec->prio < pend_prio) {
|
|
pend_prio = vec->prio;
|
|
pend_irq = i;
|
|
}
|
|
if (vec->active && vec->prio < active_prio) {
|
|
active_prio = vec->prio;
|
|
}
|
|
}
|
|
|
|
s->vectpending = pend_irq;
|
|
s->exception_prio = active_prio & nvic_gprio_mask(s);
|
|
|
|
trace_nvic_recompute_state(s->vectpending, s->exception_prio);
|
|
}
|
|
|
|
/* Return the current execution priority of the CPU
|
|
* (equivalent to the pseudocode ExecutionPriority function).
|
|
* This is a value between -2 (NMI priority) and NVIC_NOEXC_PRIO.
|
|
*/
|
|
static inline int nvic_exec_prio(NVICState *s)
|
|
{
|
|
CPUARMState *env = &s->cpu->env;
|
|
int running;
|
|
|
|
if (env->v7m.faultmask) {
|
|
running = -1;
|
|
} else if (env->v7m.primask) {
|
|
running = 0;
|
|
} else if (env->v7m.basepri[env->v7m.secure] > 0) {
|
|
running = env->v7m.basepri[env->v7m.secure] & nvic_gprio_mask(s);
|
|
} else {
|
|
running = NVIC_NOEXC_PRIO; /* lower than any possible priority */
|
|
}
|
|
/* consider priority of active handler */
|
|
return MIN(running, s->exception_prio);
|
|
}
|
|
|
|
bool armv7m_nvic_can_take_pending_exception(void *opaque)
|
|
{
|
|
NVICState *s = opaque;
|
|
|
|
return nvic_exec_prio(s) > nvic_pending_prio(s);
|
|
}
|
|
|
|
/* caller must call nvic_irq_update() after this */
|
|
static void set_prio(NVICState *s, unsigned irq, uint8_t prio)
|
|
{
|
|
assert(irq > ARMV7M_EXCP_NMI); /* only use for configurable prios */
|
|
assert(irq < s->num_irq);
|
|
|
|
s->vectors[irq].prio = prio;
|
|
|
|
trace_nvic_set_prio(irq, prio);
|
|
}
|
|
|
|
/* Recompute state and assert irq line accordingly.
|
|
* Must be called after changes to:
|
|
* vec->active, vec->enabled, vec->pending or vec->prio for any vector
|
|
* prigroup
|
|
*/
|
|
static void nvic_irq_update(NVICState *s)
|
|
{
|
|
int lvl;
|
|
int pend_prio;
|
|
|
|
nvic_recompute_state(s);
|
|
pend_prio = nvic_pending_prio(s);
|
|
|
|
/* Raise NVIC output if this IRQ would be taken, except that we
|
|
* ignore the effects of the BASEPRI, FAULTMASK and PRIMASK (which
|
|
* will be checked for in arm_v7m_cpu_exec_interrupt()); changes
|
|
* to those CPU registers don't cause us to recalculate the NVIC
|
|
* pending info.
|
|
*/
|
|
lvl = (pend_prio < s->exception_prio);
|
|
trace_nvic_irq_update(s->vectpending, pend_prio, s->exception_prio, lvl);
|
|
qemu_set_irq(s->excpout, lvl);
|
|
}
|
|
|
|
static void armv7m_nvic_clear_pending(void *opaque, int irq)
|
|
{
|
|
NVICState *s = (NVICState *)opaque;
|
|
VecInfo *vec;
|
|
|
|
assert(irq > ARMV7M_EXCP_RESET && irq < s->num_irq);
|
|
|
|
vec = &s->vectors[irq];
|
|
trace_nvic_clear_pending(irq, vec->enabled, vec->prio);
|
|
if (vec->pending) {
|
|
vec->pending = 0;
|
|
nvic_irq_update(s);
|
|
}
|
|
}
|
|
|
|
void armv7m_nvic_set_pending(void *opaque, int irq)
|
|
{
|
|
NVICState *s = (NVICState *)opaque;
|
|
VecInfo *vec;
|
|
|
|
assert(irq > ARMV7M_EXCP_RESET && irq < s->num_irq);
|
|
|
|
vec = &s->vectors[irq];
|
|
trace_nvic_set_pending(irq, vec->enabled, vec->prio);
|
|
|
|
|
|
if (irq >= ARMV7M_EXCP_HARD && irq < ARMV7M_EXCP_PENDSV) {
|
|
/* If a synchronous exception is pending then it may be
|
|
* escalated to HardFault if:
|
|
* * it is equal or lower priority to current execution
|
|
* * it is disabled
|
|
* (ie we need to take it immediately but we can't do so).
|
|
* Asynchronous exceptions (and interrupts) simply remain pending.
|
|
*
|
|
* For QEMU, we don't have any imprecise (asynchronous) faults,
|
|
* so we can assume that PREFETCH_ABORT and DATA_ABORT are always
|
|
* synchronous.
|
|
* Debug exceptions are awkward because only Debug exceptions
|
|
* resulting from the BKPT instruction should be escalated,
|
|
* but we don't currently implement any Debug exceptions other
|
|
* than those that result from BKPT, so we treat all debug exceptions
|
|
* as needing escalation.
|
|
*
|
|
* This all means we can identify whether to escalate based only on
|
|
* the exception number and don't (yet) need the caller to explicitly
|
|
* tell us whether this exception is synchronous or not.
|
|
*/
|
|
int running = nvic_exec_prio(s);
|
|
bool escalate = false;
|
|
|
|
if (vec->prio >= running) {
|
|
trace_nvic_escalate_prio(irq, vec->prio, running);
|
|
escalate = true;
|
|
} else if (!vec->enabled) {
|
|
trace_nvic_escalate_disabled(irq);
|
|
escalate = true;
|
|
}
|
|
|
|
if (escalate) {
|
|
if (running < 0) {
|
|
/* We want to escalate to HardFault but we can't take a
|
|
* synchronous HardFault at this point either. This is a
|
|
* Lockup condition due to a guest bug. We don't model
|
|
* Lockup, so report via cpu_abort() instead.
|
|
*/
|
|
cpu_abort(&s->cpu->parent_obj,
|
|
"Lockup: can't escalate %d to HardFault "
|
|
"(current priority %d)\n", irq, running);
|
|
}
|
|
|
|
/* We can do the escalation, so we take HardFault instead */
|
|
irq = ARMV7M_EXCP_HARD;
|
|
vec = &s->vectors[irq];
|
|
s->cpu->env.v7m.hfsr |= R_V7M_HFSR_FORCED_MASK;
|
|
}
|
|
}
|
|
|
|
if (!vec->pending) {
|
|
vec->pending = 1;
|
|
nvic_irq_update(s);
|
|
}
|
|
}
|
|
|
|
/* Make pending IRQ active. */
|
|
void armv7m_nvic_acknowledge_irq(void *opaque)
|
|
{
|
|
NVICState *s = (NVICState *)opaque;
|
|
CPUARMState *env = &s->cpu->env;
|
|
const int pending = s->vectpending;
|
|
const int running = nvic_exec_prio(s);
|
|
int pendgroupprio;
|
|
VecInfo *vec;
|
|
|
|
assert(pending > ARMV7M_EXCP_RESET && pending < s->num_irq);
|
|
|
|
vec = &s->vectors[pending];
|
|
|
|
assert(vec->enabled);
|
|
assert(vec->pending);
|
|
|
|
pendgroupprio = vec->prio & nvic_gprio_mask(s);
|
|
assert(pendgroupprio < running);
|
|
|
|
trace_nvic_acknowledge_irq(pending, vec->prio);
|
|
|
|
vec->active = 1;
|
|
vec->pending = 0;
|
|
|
|
env->v7m.exception = s->vectpending;
|
|
|
|
nvic_irq_update(s);
|
|
}
|
|
|
|
int armv7m_nvic_complete_irq(void *opaque, int irq)
|
|
{
|
|
NVICState *s = (NVICState *)opaque;
|
|
VecInfo *vec;
|
|
int ret;
|
|
|
|
assert(irq > ARMV7M_EXCP_RESET && irq < s->num_irq);
|
|
|
|
vec = &s->vectors[irq];
|
|
|
|
trace_nvic_complete_irq(irq);
|
|
|
|
if (!vec->active) {
|
|
/* Tell the caller this was an illegal exception return */
|
|
return -1;
|
|
}
|
|
|
|
ret = nvic_rettobase(s);
|
|
|
|
vec->active = 0;
|
|
if (vec->level) {
|
|
/* Re-pend the exception if it's still held high; only
|
|
* happens for extenal IRQs
|
|
*/
|
|
assert(irq >= NVIC_FIRST_IRQ);
|
|
vec->pending = 1;
|
|
}
|
|
|
|
nvic_irq_update(s);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* callback when external interrupt line is changed */
|
|
static void set_irq_level(void *opaque, int n, int level)
|
|
{
|
|
NVICState *s = opaque;
|
|
VecInfo *vec;
|
|
|
|
n += NVIC_FIRST_IRQ;
|
|
|
|
assert(n >= NVIC_FIRST_IRQ && n < s->num_irq);
|
|
|
|
trace_nvic_set_irq_level(n, level);
|
|
|
|
/* The pending status of an external interrupt is
|
|
* latched on rising edge and exception handler return.
|
|
*
|
|
* Pulsing the IRQ will always run the handler
|
|
* once, and the handler will re-run until the
|
|
* level is low when the handler completes.
|
|
*/
|
|
vec = &s->vectors[n];
|
|
if (level != vec->level) {
|
|
vec->level = level;
|
|
if (level) {
|
|
armv7m_nvic_set_pending(s, n);
|
|
}
|
|
}
|
|
}
|
|
|
|
static uint32_t nvic_readl(NVICState *s, uint32_t offset)
|
|
{
|
|
ARMCPU *cpu = s->cpu;
|
|
uint32_t val;
|
|
|
|
switch (offset) {
|
|
case 4: /* Interrupt Control Type. */
|
|
return ((s->num_irq - NVIC_FIRST_IRQ) / 32) - 1;
|
|
case 0xd00: /* CPUID Base. */
|
|
return cpu->midr;
|
|
case 0xd04: /* Interrupt Control State. */
|
|
/* VECTACTIVE */
|
|
val = cpu->env.v7m.exception;
|
|
/* VECTPENDING */
|
|
val |= (s->vectpending & 0xff) << 12;
|
|
/* ISRPENDING - set if any external IRQ is pending */
|
|
if (nvic_isrpending(s)) {
|
|
val |= (1 << 22);
|
|
}
|
|
/* RETTOBASE - set if only one handler is active */
|
|
if (nvic_rettobase(s)) {
|
|
val |= (1 << 11);
|
|
}
|
|
/* PENDSTSET */
|
|
if (s->vectors[ARMV7M_EXCP_SYSTICK].pending) {
|
|
val |= (1 << 26);
|
|
}
|
|
/* PENDSVSET */
|
|
if (s->vectors[ARMV7M_EXCP_PENDSV].pending) {
|
|
val |= (1 << 28);
|
|
}
|
|
/* NMIPENDSET */
|
|
if (s->vectors[ARMV7M_EXCP_NMI].pending) {
|
|
val |= (1 << 31);
|
|
}
|
|
/* ISRPREEMPT not implemented */
|
|
return val;
|
|
case 0xd08: /* Vector Table Offset. */
|
|
return cpu->env.v7m.vecbase;
|
|
case 0xd0c: /* Application Interrupt/Reset Control. */
|
|
return 0xfa050000 | (s->prigroup << 8);
|
|
case 0xd10: /* System Control. */
|
|
/* TODO: Implement SLEEPONEXIT. */
|
|
return 0;
|
|
case 0xd14: /* Configuration Control. */
|
|
return cpu->env.v7m.ccr;
|
|
case 0xd24: /* System Handler Status. */
|
|
val = 0;
|
|
if (s->vectors[ARMV7M_EXCP_MEM].active) {
|
|
val |= (1 << 0);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_BUS].active) {
|
|
val |= (1 << 1);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_USAGE].active) {
|
|
val |= (1 << 3);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_SVC].active) {
|
|
val |= (1 << 7);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_DEBUG].active) {
|
|
val |= (1 << 8);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_PENDSV].active) {
|
|
val |= (1 << 10);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_SYSTICK].active) {
|
|
val |= (1 << 11);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_USAGE].pending) {
|
|
val |= (1 << 12);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_MEM].pending) {
|
|
val |= (1 << 13);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_BUS].pending) {
|
|
val |= (1 << 14);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_SVC].pending) {
|
|
val |= (1 << 15);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_MEM].enabled) {
|
|
val |= (1 << 16);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_BUS].enabled) {
|
|
val |= (1 << 17);
|
|
}
|
|
if (s->vectors[ARMV7M_EXCP_USAGE].enabled) {
|
|
val |= (1 << 18);
|
|
}
|
|
return val;
|
|
case 0xd28: /* Configurable Fault Status. */
|
|
return cpu->env.v7m.cfsr;
|
|
case 0xd2c: /* Hard Fault Status. */
|
|
return cpu->env.v7m.hfsr;
|
|
case 0xd30: /* Debug Fault Status. */
|
|
return cpu->env.v7m.dfsr;
|
|
case 0xd34: /* MMFAR MemManage Fault Address */
|
|
return cpu->env.v7m.mmfar;
|
|
case 0xd38: /* Bus Fault Address. */
|
|
return cpu->env.v7m.bfar;
|
|
case 0xd3c: /* Aux Fault Status. */
|
|
/* TODO: Implement fault status registers. */
|
|
qemu_log_mask(LOG_UNIMP,
|
|
"Aux Fault status registers unimplemented\n");
|
|
return 0;
|
|
case 0xd40: /* PFR0. */
|
|
return 0x00000030;
|
|
case 0xd44: /* PRF1. */
|
|
return 0x00000200;
|
|
case 0xd48: /* DFR0. */
|
|
return 0x00100000;
|
|
case 0xd4c: /* AFR0. */
|
|
return 0x00000000;
|
|
case 0xd50: /* MMFR0. */
|
|
return 0x00000030;
|
|
case 0xd54: /* MMFR1. */
|
|
return 0x00000000;
|
|
case 0xd58: /* MMFR2. */
|
|
return 0x00000000;
|
|
case 0xd5c: /* MMFR3. */
|
|
return 0x00000000;
|
|
case 0xd60: /* ISAR0. */
|
|
return 0x01141110;
|
|
case 0xd64: /* ISAR1. */
|
|
return 0x02111000;
|
|
case 0xd68: /* ISAR2. */
|
|
return 0x21112231;
|
|
case 0xd6c: /* ISAR3. */
|
|
return 0x01111110;
|
|
case 0xd70: /* ISAR4. */
|
|
return 0x01310102;
|
|
/* TODO: Implement debug registers. */
|
|
case 0xd90: /* MPU_TYPE */
|
|
/* Unified MPU; if the MPU is not present this value is zero */
|
|
return cpu->pmsav7_dregion << 8;
|
|
break;
|
|
case 0xd94: /* MPU_CTRL */
|
|
return cpu->env.v7m.mpu_ctrl;
|
|
case 0xd98: /* MPU_RNR */
|
|
return cpu->env.pmsav7.rnr;
|
|
case 0xd9c: /* MPU_RBAR */
|
|
case 0xda4: /* MPU_RBAR_A1 */
|
|
case 0xdac: /* MPU_RBAR_A2 */
|
|
case 0xdb4: /* MPU_RBAR_A3 */
|
|
{
|
|
int region = cpu->env.pmsav7.rnr;
|
|
|
|
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
/* PMSAv8M handling of the aliases is different from v7M:
|
|
* aliases A1, A2, A3 override the low two bits of the region
|
|
* number in MPU_RNR, and there is no 'region' field in the
|
|
* RBAR register.
|
|
*/
|
|
int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
|
|
if (aliasno) {
|
|
region = deposit32(region, 0, 2, aliasno);
|
|
}
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return 0;
|
|
}
|
|
return cpu->env.pmsav8.rbar[region];
|
|
}
|
|
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return 0;
|
|
}
|
|
return (cpu->env.pmsav7.drbar[region] & 0x1f) | (region & 0xf);
|
|
}
|
|
case 0xda0: /* MPU_RASR (v7M), MPU_RLAR (v8M) */
|
|
case 0xda8: /* MPU_RASR_A1 (v7M), MPU_RLAR_A1 (v8M) */
|
|
case 0xdb0: /* MPU_RASR_A2 (v7M), MPU_RLAR_A2 (v8M) */
|
|
case 0xdb8: /* MPU_RASR_A3 (v7M), MPU_RLAR_A3 (v8M) */
|
|
{
|
|
int region = cpu->env.pmsav7.rnr;
|
|
|
|
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
/* PMSAv8M handling of the aliases is different from v7M:
|
|
* aliases A1, A2, A3 override the low two bits of the region
|
|
* number in MPU_RNR.
|
|
*/
|
|
int aliasno = (offset - 0xda0) / 8; /* 0..3 */
|
|
if (aliasno) {
|
|
region = deposit32(region, 0, 2, aliasno);
|
|
}
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return 0;
|
|
}
|
|
return cpu->env.pmsav8.rlar[region];
|
|
}
|
|
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return 0;
|
|
}
|
|
return ((cpu->env.pmsav7.dracr[region] & 0xffff) << 16) |
|
|
(cpu->env.pmsav7.drsr[region] & 0xffff);
|
|
}
|
|
case 0xdc0: /* MPU_MAIR0 */
|
|
if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
goto bad_offset;
|
|
}
|
|
return cpu->env.pmsav8.mair0;
|
|
case 0xdc4: /* MPU_MAIR1 */
|
|
if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
goto bad_offset;
|
|
}
|
|
return cpu->env.pmsav8.mair1;
|
|
default:
|
|
bad_offset:
|
|
qemu_log_mask(LOG_GUEST_ERROR, "NVIC: Bad read offset 0x%x\n", offset);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static void nvic_writel(NVICState *s, uint32_t offset, uint32_t value)
|
|
{
|
|
ARMCPU *cpu = s->cpu;
|
|
|
|
switch (offset) {
|
|
case 0xd04: /* Interrupt Control State. */
|
|
if (value & (1 << 31)) {
|
|
armv7m_nvic_set_pending(s, ARMV7M_EXCP_NMI);
|
|
}
|
|
if (value & (1 << 28)) {
|
|
armv7m_nvic_set_pending(s, ARMV7M_EXCP_PENDSV);
|
|
} else if (value & (1 << 27)) {
|
|
armv7m_nvic_clear_pending(s, ARMV7M_EXCP_PENDSV);
|
|
}
|
|
if (value & (1 << 26)) {
|
|
armv7m_nvic_set_pending(s, ARMV7M_EXCP_SYSTICK);
|
|
} else if (value & (1 << 25)) {
|
|
armv7m_nvic_clear_pending(s, ARMV7M_EXCP_SYSTICK);
|
|
}
|
|
break;
|
|
case 0xd08: /* Vector Table Offset. */
|
|
cpu->env.v7m.vecbase = value & 0xffffff80;
|
|
break;
|
|
case 0xd0c: /* Application Interrupt/Reset Control. */
|
|
if ((value >> 16) == 0x05fa) {
|
|
if (value & 4) {
|
|
qemu_irq_pulse(s->sysresetreq);
|
|
}
|
|
if (value & 2) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"Setting VECTCLRACTIVE when not in DEBUG mode "
|
|
"is UNPREDICTABLE\n");
|
|
}
|
|
if (value & 1) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"Setting VECTRESET when not in DEBUG mode "
|
|
"is UNPREDICTABLE\n");
|
|
}
|
|
s->prigroup = extract32(value, 8, 3);
|
|
nvic_irq_update(s);
|
|
}
|
|
break;
|
|
case 0xd10: /* System Control. */
|
|
/* TODO: Implement control registers. */
|
|
qemu_log_mask(LOG_UNIMP, "NVIC: SCR unimplemented\n");
|
|
break;
|
|
case 0xd14: /* Configuration Control. */
|
|
/* Enforce RAZ/WI on reserved and must-RAZ/WI bits */
|
|
value &= (R_V7M_CCR_STKALIGN_MASK |
|
|
R_V7M_CCR_BFHFNMIGN_MASK |
|
|
R_V7M_CCR_DIV_0_TRP_MASK |
|
|
R_V7M_CCR_UNALIGN_TRP_MASK |
|
|
R_V7M_CCR_USERSETMPEND_MASK |
|
|
R_V7M_CCR_NONBASETHRDENA_MASK);
|
|
|
|
cpu->env.v7m.ccr = value;
|
|
break;
|
|
case 0xd24: /* System Handler Control. */
|
|
s->vectors[ARMV7M_EXCP_MEM].active = (value & (1 << 0)) != 0;
|
|
s->vectors[ARMV7M_EXCP_BUS].active = (value & (1 << 1)) != 0;
|
|
s->vectors[ARMV7M_EXCP_USAGE].active = (value & (1 << 3)) != 0;
|
|
s->vectors[ARMV7M_EXCP_SVC].active = (value & (1 << 7)) != 0;
|
|
s->vectors[ARMV7M_EXCP_DEBUG].active = (value & (1 << 8)) != 0;
|
|
s->vectors[ARMV7M_EXCP_PENDSV].active = (value & (1 << 10)) != 0;
|
|
s->vectors[ARMV7M_EXCP_SYSTICK].active = (value & (1 << 11)) != 0;
|
|
s->vectors[ARMV7M_EXCP_USAGE].pending = (value & (1 << 12)) != 0;
|
|
s->vectors[ARMV7M_EXCP_MEM].pending = (value & (1 << 13)) != 0;
|
|
s->vectors[ARMV7M_EXCP_BUS].pending = (value & (1 << 14)) != 0;
|
|
s->vectors[ARMV7M_EXCP_SVC].pending = (value & (1 << 15)) != 0;
|
|
s->vectors[ARMV7M_EXCP_MEM].enabled = (value & (1 << 16)) != 0;
|
|
s->vectors[ARMV7M_EXCP_BUS].enabled = (value & (1 << 17)) != 0;
|
|
s->vectors[ARMV7M_EXCP_USAGE].enabled = (value & (1 << 18)) != 0;
|
|
nvic_irq_update(s);
|
|
break;
|
|
case 0xd28: /* Configurable Fault Status. */
|
|
cpu->env.v7m.cfsr &= ~value; /* W1C */
|
|
break;
|
|
case 0xd2c: /* Hard Fault Status. */
|
|
cpu->env.v7m.hfsr &= ~value; /* W1C */
|
|
break;
|
|
case 0xd30: /* Debug Fault Status. */
|
|
cpu->env.v7m.dfsr &= ~value; /* W1C */
|
|
break;
|
|
case 0xd34: /* Mem Manage Address. */
|
|
cpu->env.v7m.mmfar = value;
|
|
return;
|
|
case 0xd38: /* Bus Fault Address. */
|
|
cpu->env.v7m.bfar = value;
|
|
return;
|
|
case 0xd3c: /* Aux Fault Status. */
|
|
qemu_log_mask(LOG_UNIMP,
|
|
"NVIC: Aux fault status registers unimplemented\n");
|
|
break;
|
|
case 0xd90: /* MPU_TYPE */
|
|
return; /* RO */
|
|
case 0xd94: /* MPU_CTRL */
|
|
if ((value &
|
|
(R_V7M_MPU_CTRL_HFNMIENA_MASK | R_V7M_MPU_CTRL_ENABLE_MASK))
|
|
== R_V7M_MPU_CTRL_HFNMIENA_MASK) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "MPU_CTRL: HFNMIENA and !ENABLE is "
|
|
"UNPREDICTABLE\n");
|
|
}
|
|
cpu->env.v7m.mpu_ctrl = value & (R_V7M_MPU_CTRL_ENABLE_MASK |
|
|
R_V7M_MPU_CTRL_HFNMIENA_MASK |
|
|
R_V7M_MPU_CTRL_PRIVDEFENA_MASK);
|
|
tlb_flush(CPU(cpu));
|
|
break;
|
|
case 0xd98: /* MPU_RNR */
|
|
if (value >= cpu->pmsav7_dregion) {
|
|
qemu_log_mask(LOG_GUEST_ERROR, "MPU region out of range %"
|
|
PRIu32 "/%" PRIu32 "\n",
|
|
value, cpu->pmsav7_dregion);
|
|
} else {
|
|
cpu->env.pmsav7.rnr = value;
|
|
}
|
|
break;
|
|
case 0xd9c: /* MPU_RBAR */
|
|
case 0xda4: /* MPU_RBAR_A1 */
|
|
case 0xdac: /* MPU_RBAR_A2 */
|
|
case 0xdb4: /* MPU_RBAR_A3 */
|
|
{
|
|
int region;
|
|
|
|
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
/* PMSAv8M handling of the aliases is different from v7M:
|
|
* aliases A1, A2, A3 override the low two bits of the region
|
|
* number in MPU_RNR, and there is no 'region' field in the
|
|
* RBAR register.
|
|
*/
|
|
int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
|
|
|
|
region = cpu->env.pmsav7.rnr;
|
|
if (aliasno) {
|
|
region = deposit32(region, 0, 2, aliasno);
|
|
}
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return;
|
|
}
|
|
cpu->env.pmsav8.rbar[region] = value;
|
|
tlb_flush(CPU(cpu));
|
|
return;
|
|
}
|
|
|
|
if (value & (1 << 4)) {
|
|
/* VALID bit means use the region number specified in this
|
|
* value and also update MPU_RNR.REGION with that value.
|
|
*/
|
|
region = extract32(value, 0, 4);
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"MPU region out of range %u/%" PRIu32 "\n",
|
|
region, cpu->pmsav7_dregion);
|
|
return;
|
|
}
|
|
cpu->env.pmsav7.rnr = region;
|
|
} else {
|
|
region = cpu->env.pmsav7.rnr;
|
|
}
|
|
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return;
|
|
}
|
|
|
|
cpu->env.pmsav7.drbar[region] = value & ~0x1f;
|
|
tlb_flush(CPU(cpu));
|
|
break;
|
|
}
|
|
case 0xda0: /* MPU_RASR (v7M), MPU_RLAR (v8M) */
|
|
case 0xda8: /* MPU_RASR_A1 (v7M), MPU_RLAR_A1 (v8M) */
|
|
case 0xdb0: /* MPU_RASR_A2 (v7M), MPU_RLAR_A2 (v8M) */
|
|
case 0xdb8: /* MPU_RASR_A3 (v7M), MPU_RLAR_A3 (v8M) */
|
|
{
|
|
int region = cpu->env.pmsav7.rnr;
|
|
|
|
if (arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
/* PMSAv8M handling of the aliases is different from v7M:
|
|
* aliases A1, A2, A3 override the low two bits of the region
|
|
* number in MPU_RNR.
|
|
*/
|
|
int aliasno = (offset - 0xd9c) / 8; /* 0..3 */
|
|
|
|
region = cpu->env.pmsav7.rnr;
|
|
if (aliasno) {
|
|
region = deposit32(region, 0, 2, aliasno);
|
|
}
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return;
|
|
}
|
|
cpu->env.pmsav8.rlar[region] = value;
|
|
tlb_flush(CPU(cpu));
|
|
return;
|
|
}
|
|
|
|
if (region >= cpu->pmsav7_dregion) {
|
|
return;
|
|
}
|
|
|
|
cpu->env.pmsav7.drsr[region] = value & 0xff3f;
|
|
cpu->env.pmsav7.dracr[region] = (value >> 16) & 0x173f;
|
|
tlb_flush(CPU(cpu));
|
|
break;
|
|
}
|
|
case 0xdc0: /* MPU_MAIR0 */
|
|
if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
goto bad_offset;
|
|
}
|
|
if (cpu->pmsav7_dregion) {
|
|
/* Register is RES0 if no MPU regions are implemented */
|
|
cpu->env.pmsav8.mair0 = value;
|
|
}
|
|
/* We don't need to do anything else because memory attributes
|
|
* only affect cacheability, and we don't implement caching.
|
|
*/
|
|
break;
|
|
case 0xdc4: /* MPU_MAIR1 */
|
|
if (!arm_feature(&cpu->env, ARM_FEATURE_V8)) {
|
|
goto bad_offset;
|
|
}
|
|
if (cpu->pmsav7_dregion) {
|
|
/* Register is RES0 if no MPU regions are implemented */
|
|
cpu->env.pmsav8.mair1 = value;
|
|
}
|
|
/* We don't need to do anything else because memory attributes
|
|
* only affect cacheability, and we don't implement caching.
|
|
*/
|
|
break;
|
|
case 0xf00: /* Software Triggered Interrupt Register */
|
|
{
|
|
int excnum = (value & 0x1ff) + NVIC_FIRST_IRQ;
|
|
if (excnum < s->num_irq) {
|
|
armv7m_nvic_set_pending(s, excnum);
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
bad_offset:
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"NVIC: Bad write offset 0x%x\n", offset);
|
|
}
|
|
}
|
|
|
|
static bool nvic_user_access_ok(NVICState *s, hwaddr offset)
|
|
{
|
|
/* Return true if unprivileged access to this register is permitted. */
|
|
switch (offset) {
|
|
case 0xf00: /* STIR: accessible only if CCR.USERSETMPEND permits */
|
|
return s->cpu->env.v7m.ccr & R_V7M_CCR_USERSETMPEND_MASK;
|
|
default:
|
|
/* All other user accesses cause a BusFault unconditionally */
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static MemTxResult nvic_sysreg_read(void *opaque, hwaddr addr,
|
|
uint64_t *data, unsigned size,
|
|
MemTxAttrs attrs)
|
|
{
|
|
NVICState *s = (NVICState *)opaque;
|
|
uint32_t offset = addr;
|
|
unsigned i, startvec, end;
|
|
uint32_t val;
|
|
|
|
if (attrs.user && !nvic_user_access_ok(s, addr)) {
|
|
/* Generate BusFault for unprivileged accesses */
|
|
return MEMTX_ERROR;
|
|
}
|
|
|
|
switch (offset) {
|
|
/* reads of set and clear both return the status */
|
|
case 0x100 ... 0x13f: /* NVIC Set enable */
|
|
offset += 0x80;
|
|
/* fall through */
|
|
case 0x180 ... 0x1bf: /* NVIC Clear enable */
|
|
val = 0;
|
|
startvec = offset - 0x180 + NVIC_FIRST_IRQ; /* vector # */
|
|
|
|
for (i = 0, end = size * 8; i < end && startvec + i < s->num_irq; i++) {
|
|
if (s->vectors[startvec + i].enabled) {
|
|
val |= (1 << i);
|
|
}
|
|
}
|
|
break;
|
|
case 0x200 ... 0x23f: /* NVIC Set pend */
|
|
offset += 0x80;
|
|
/* fall through */
|
|
case 0x280 ... 0x2bf: /* NVIC Clear pend */
|
|
val = 0;
|
|
startvec = offset - 0x280 + NVIC_FIRST_IRQ; /* vector # */
|
|
for (i = 0, end = size * 8; i < end && startvec + i < s->num_irq; i++) {
|
|
if (s->vectors[startvec + i].pending) {
|
|
val |= (1 << i);
|
|
}
|
|
}
|
|
break;
|
|
case 0x300 ... 0x33f: /* NVIC Active */
|
|
val = 0;
|
|
startvec = offset - 0x300 + NVIC_FIRST_IRQ; /* vector # */
|
|
|
|
for (i = 0, end = size * 8; i < end && startvec + i < s->num_irq; i++) {
|
|
if (s->vectors[startvec + i].active) {
|
|
val |= (1 << i);
|
|
}
|
|
}
|
|
break;
|
|
case 0x400 ... 0x5ef: /* NVIC Priority */
|
|
val = 0;
|
|
startvec = offset - 0x400 + NVIC_FIRST_IRQ; /* vector # */
|
|
|
|
for (i = 0; i < size && startvec + i < s->num_irq; i++) {
|
|
val |= s->vectors[startvec + i].prio << (8 * i);
|
|
}
|
|
break;
|
|
case 0xd18 ... 0xd23: /* System Handler Priority. */
|
|
val = 0;
|
|
for (i = 0; i < size; i++) {
|
|
val |= s->vectors[(offset - 0xd14) + i].prio << (i * 8);
|
|
}
|
|
break;
|
|
case 0xfe0 ... 0xfff: /* ID. */
|
|
if (offset & 3) {
|
|
val = 0;
|
|
} else {
|
|
val = nvic_id[(offset - 0xfe0) >> 2];
|
|
}
|
|
break;
|
|
default:
|
|
if (size == 4) {
|
|
val = nvic_readl(s, offset);
|
|
} else {
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"NVIC: Bad read of size %d at offset 0x%x\n",
|
|
size, offset);
|
|
val = 0;
|
|
}
|
|
}
|
|
|
|
trace_nvic_sysreg_read(addr, val, size);
|
|
*data = val;
|
|
return MEMTX_OK;
|
|
}
|
|
|
|
static MemTxResult nvic_sysreg_write(void *opaque, hwaddr addr,
|
|
uint64_t value, unsigned size,
|
|
MemTxAttrs attrs)
|
|
{
|
|
NVICState *s = (NVICState *)opaque;
|
|
uint32_t offset = addr;
|
|
unsigned i, startvec, end;
|
|
unsigned setval = 0;
|
|
|
|
trace_nvic_sysreg_write(addr, value, size);
|
|
|
|
if (attrs.user && !nvic_user_access_ok(s, addr)) {
|
|
/* Generate BusFault for unprivileged accesses */
|
|
return MEMTX_ERROR;
|
|
}
|
|
|
|
switch (offset) {
|
|
case 0x100 ... 0x13f: /* NVIC Set enable */
|
|
offset += 0x80;
|
|
setval = 1;
|
|
/* fall through */
|
|
case 0x180 ... 0x1bf: /* NVIC Clear enable */
|
|
startvec = 8 * (offset - 0x180) + NVIC_FIRST_IRQ;
|
|
|
|
for (i = 0, end = size * 8; i < end && startvec + i < s->num_irq; i++) {
|
|
if (value & (1 << i)) {
|
|
s->vectors[startvec + i].enabled = setval;
|
|
}
|
|
}
|
|
nvic_irq_update(s);
|
|
return MEMTX_OK;
|
|
case 0x200 ... 0x23f: /* NVIC Set pend */
|
|
/* the special logic in armv7m_nvic_set_pending()
|
|
* is not needed since IRQs are never escalated
|
|
*/
|
|
offset += 0x80;
|
|
setval = 1;
|
|
/* fall through */
|
|
case 0x280 ... 0x2bf: /* NVIC Clear pend */
|
|
startvec = 8 * (offset - 0x280) + NVIC_FIRST_IRQ; /* vector # */
|
|
|
|
for (i = 0, end = size * 8; i < end && startvec + i < s->num_irq; i++) {
|
|
if (value & (1 << i)) {
|
|
s->vectors[startvec + i].pending = setval;
|
|
}
|
|
}
|
|
nvic_irq_update(s);
|
|
return MEMTX_OK;
|
|
case 0x300 ... 0x33f: /* NVIC Active */
|
|
return MEMTX_OK; /* R/O */
|
|
case 0x400 ... 0x5ef: /* NVIC Priority */
|
|
startvec = 8 * (offset - 0x400) + NVIC_FIRST_IRQ; /* vector # */
|
|
|
|
for (i = 0; i < size && startvec + i < s->num_irq; i++) {
|
|
set_prio(s, startvec + i, (value >> (i * 8)) & 0xff);
|
|
}
|
|
nvic_irq_update(s);
|
|
return MEMTX_OK;
|
|
case 0xd18 ... 0xd23: /* System Handler Priority. */
|
|
for (i = 0; i < size; i++) {
|
|
unsigned hdlidx = (offset - 0xd14) + i;
|
|
set_prio(s, hdlidx, (value >> (i * 8)) & 0xff);
|
|
}
|
|
nvic_irq_update(s);
|
|
return MEMTX_OK;
|
|
}
|
|
if (size == 4) {
|
|
nvic_writel(s, offset, value);
|
|
return MEMTX_OK;
|
|
}
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"NVIC: Bad write of size %d at offset 0x%x\n", size, offset);
|
|
/* This is UNPREDICTABLE; treat as RAZ/WI */
|
|
return MEMTX_OK;
|
|
}
|
|
|
|
static const MemoryRegionOps nvic_sysreg_ops = {
|
|
.read_with_attrs = nvic_sysreg_read,
|
|
.write_with_attrs = nvic_sysreg_write,
|
|
.endianness = DEVICE_NATIVE_ENDIAN,
|
|
};
|
|
|
|
static int nvic_post_load(void *opaque, int version_id)
|
|
{
|
|
NVICState *s = opaque;
|
|
unsigned i;
|
|
|
|
/* Check for out of range priority settings */
|
|
if (s->vectors[ARMV7M_EXCP_RESET].prio != -3 ||
|
|
s->vectors[ARMV7M_EXCP_NMI].prio != -2 ||
|
|
s->vectors[ARMV7M_EXCP_HARD].prio != -1) {
|
|
return 1;
|
|
}
|
|
for (i = ARMV7M_EXCP_MEM; i < s->num_irq; i++) {
|
|
if (s->vectors[i].prio & ~0xff) {
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
nvic_recompute_state(s);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const VMStateDescription vmstate_VecInfo = {
|
|
.name = "armv7m_nvic_info",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_INT16(prio, VecInfo),
|
|
VMSTATE_UINT8(enabled, VecInfo),
|
|
VMSTATE_UINT8(pending, VecInfo),
|
|
VMSTATE_UINT8(active, VecInfo),
|
|
VMSTATE_UINT8(level, VecInfo),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
static const VMStateDescription vmstate_nvic = {
|
|
.name = "armv7m_nvic",
|
|
.version_id = 4,
|
|
.minimum_version_id = 4,
|
|
.post_load = &nvic_post_load,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_STRUCT_ARRAY(vectors, NVICState, NVIC_MAX_VECTORS, 1,
|
|
vmstate_VecInfo, VecInfo),
|
|
VMSTATE_UINT32(prigroup, NVICState),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
static Property props_nvic[] = {
|
|
/* Number of external IRQ lines (so excluding the 16 internal exceptions) */
|
|
DEFINE_PROP_UINT32("num-irq", NVICState, num_irq, 64),
|
|
DEFINE_PROP_END_OF_LIST()
|
|
};
|
|
|
|
static void armv7m_nvic_reset(DeviceState *dev)
|
|
{
|
|
NVICState *s = NVIC(dev);
|
|
|
|
s->vectors[ARMV7M_EXCP_NMI].enabled = 1;
|
|
s->vectors[ARMV7M_EXCP_HARD].enabled = 1;
|
|
/* MEM, BUS, and USAGE are enabled through
|
|
* the System Handler Control register
|
|
*/
|
|
s->vectors[ARMV7M_EXCP_SVC].enabled = 1;
|
|
s->vectors[ARMV7M_EXCP_DEBUG].enabled = 1;
|
|
s->vectors[ARMV7M_EXCP_PENDSV].enabled = 1;
|
|
s->vectors[ARMV7M_EXCP_SYSTICK].enabled = 1;
|
|
|
|
s->vectors[ARMV7M_EXCP_RESET].prio = -3;
|
|
s->vectors[ARMV7M_EXCP_NMI].prio = -2;
|
|
s->vectors[ARMV7M_EXCP_HARD].prio = -1;
|
|
|
|
/* Strictly speaking the reset handler should be enabled.
|
|
* However, we don't simulate soft resets through the NVIC,
|
|
* and the reset vector should never be pended.
|
|
* So we leave it disabled to catch logic errors.
|
|
*/
|
|
|
|
s->exception_prio = NVIC_NOEXC_PRIO;
|
|
s->vectpending = 0;
|
|
}
|
|
|
|
static void nvic_systick_trigger(void *opaque, int n, int level)
|
|
{
|
|
NVICState *s = opaque;
|
|
|
|
if (level) {
|
|
/* SysTick just asked us to pend its exception.
|
|
* (This is different from an external interrupt line's
|
|
* behaviour.)
|
|
*/
|
|
armv7m_nvic_set_pending(s, ARMV7M_EXCP_SYSTICK);
|
|
}
|
|
}
|
|
|
|
static void armv7m_nvic_realize(DeviceState *dev, Error **errp)
|
|
{
|
|
NVICState *s = NVIC(dev);
|
|
SysBusDevice *systick_sbd;
|
|
Error *err = NULL;
|
|
|
|
s->cpu = ARM_CPU(qemu_get_cpu(0));
|
|
assert(s->cpu);
|
|
|
|
if (s->num_irq > NVIC_MAX_IRQ) {
|
|
error_setg(errp, "num-irq %d exceeds NVIC maximum", s->num_irq);
|
|
return;
|
|
}
|
|
|
|
qdev_init_gpio_in(dev, set_irq_level, s->num_irq);
|
|
|
|
/* include space for internal exception vectors */
|
|
s->num_irq += NVIC_FIRST_IRQ;
|
|
|
|
object_property_set_bool(OBJECT(&s->systick), true, "realized", &err);
|
|
if (err != NULL) {
|
|
error_propagate(errp, err);
|
|
return;
|
|
}
|
|
systick_sbd = SYS_BUS_DEVICE(&s->systick);
|
|
sysbus_connect_irq(systick_sbd, 0,
|
|
qdev_get_gpio_in_named(dev, "systick-trigger", 0));
|
|
|
|
/* The NVIC and System Control Space (SCS) starts at 0xe000e000
|
|
* and looks like this:
|
|
* 0x004 - ICTR
|
|
* 0x010 - 0xff - systick
|
|
* 0x100..0x7ec - NVIC
|
|
* 0x7f0..0xcff - Reserved
|
|
* 0xd00..0xd3c - SCS registers
|
|
* 0xd40..0xeff - Reserved or Not implemented
|
|
* 0xf00 - STIR
|
|
*/
|
|
memory_region_init(&s->container, OBJECT(s), "nvic", 0x1000);
|
|
/* The system register region goes at the bottom of the priority
|
|
* stack as it covers the whole page.
|
|
*/
|
|
memory_region_init_io(&s->sysregmem, OBJECT(s), &nvic_sysreg_ops, s,
|
|
"nvic_sysregs", 0x1000);
|
|
memory_region_add_subregion(&s->container, 0, &s->sysregmem);
|
|
memory_region_add_subregion_overlap(&s->container, 0x10,
|
|
sysbus_mmio_get_region(systick_sbd, 0),
|
|
1);
|
|
|
|
sysbus_init_mmio(SYS_BUS_DEVICE(dev), &s->container);
|
|
}
|
|
|
|
static void armv7m_nvic_instance_init(Object *obj)
|
|
{
|
|
/* We have a different default value for the num-irq property
|
|
* than our superclass. This function runs after qdev init
|
|
* has set the defaults from the Property array and before
|
|
* any user-specified property setting, so just modify the
|
|
* value in the GICState struct.
|
|
*/
|
|
DeviceState *dev = DEVICE(obj);
|
|
NVICState *nvic = NVIC(obj);
|
|
SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
|
|
|
|
object_initialize(&nvic->systick, sizeof(nvic->systick), TYPE_SYSTICK);
|
|
qdev_set_parent_bus(DEVICE(&nvic->systick), sysbus_get_default());
|
|
|
|
sysbus_init_irq(sbd, &nvic->excpout);
|
|
qdev_init_gpio_out_named(dev, &nvic->sysresetreq, "SYSRESETREQ", 1);
|
|
qdev_init_gpio_in_named(dev, nvic_systick_trigger, "systick-trigger", 1);
|
|
}
|
|
|
|
static void armv7m_nvic_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
|
|
dc->vmsd = &vmstate_nvic;
|
|
dc->props = props_nvic;
|
|
dc->reset = armv7m_nvic_reset;
|
|
dc->realize = armv7m_nvic_realize;
|
|
}
|
|
|
|
static const TypeInfo armv7m_nvic_info = {
|
|
.name = TYPE_NVIC,
|
|
.parent = TYPE_SYS_BUS_DEVICE,
|
|
.instance_init = armv7m_nvic_instance_init,
|
|
.instance_size = sizeof(NVICState),
|
|
.class_init = armv7m_nvic_class_init,
|
|
.class_size = sizeof(SysBusDeviceClass),
|
|
};
|
|
|
|
static void armv7m_nvic_register_types(void)
|
|
{
|
|
type_register_static(&armv7m_nvic_info);
|
|
}
|
|
|
|
type_init(armv7m_nvic_register_types)
|