arm: spelling fixes
Signed-off-by: Michael Tokarev <mjt@tls.msk.ru> Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
parent
cced0d6539
commit
673d821541
@ -1,6 +1,6 @@
|
|||||||
/*
|
/*
|
||||||
* SPDX-License-Identifier: GPL-2.0-or-later
|
* SPDX-License-Identifier: GPL-2.0-or-later
|
||||||
* Host specific cpu indentification for AArch64.
|
* Host specific cpu identification for AArch64.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#ifndef HOST_CPUINFO_H
|
#ifndef HOST_CPUINFO_H
|
||||||
|
@ -1565,7 +1565,7 @@ static void ast1030_evb_i2c_init(AspeedMachineState *bmc)
|
|||||||
{
|
{
|
||||||
AspeedSoCState *soc = &bmc->soc;
|
AspeedSoCState *soc = &bmc->soc;
|
||||||
|
|
||||||
/* U10 24C08 connects to SDA/SCL Groupt 1 by default */
|
/* U10 24C08 connects to SDA/SCL Group 1 by default */
|
||||||
uint8_t *eeprom_buf = g_malloc0(32 * 1024);
|
uint8_t *eeprom_buf = g_malloc0(32 * 1024);
|
||||||
smbus_eeprom_init_one(aspeed_i2c_get_bus(&soc->i2c, 0), 0x50, eeprom_buf);
|
smbus_eeprom_init_one(aspeed_i2c_get_bus(&soc->i2c, 0), 0x50, eeprom_buf);
|
||||||
|
|
||||||
|
@ -1205,7 +1205,7 @@ static void mps2_tz_idau_check(IDAUInterface *ii, uint32_t address,
|
|||||||
{
|
{
|
||||||
/*
|
/*
|
||||||
* The MPS2 TZ FPGA images have IDAUs in them which are connected to
|
* The MPS2 TZ FPGA images have IDAUs in them which are connected to
|
||||||
* the Master Security Controllers. Thes have the same logic as
|
* the Master Security Controllers. These have the same logic as
|
||||||
* is used by the IoTKit for the IDAU connected to the CPU, except
|
* is used by the IoTKit for the IDAU connected to the CPU, except
|
||||||
* that MSCs don't care about the NSC attribute.
|
* that MSCs don't care about the NSC attribute.
|
||||||
*/
|
*/
|
||||||
|
@ -239,7 +239,7 @@ static inline bool gic_lr_entry_is_free(uint32_t entry)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Return true if this LR should trigger an EOI maintenance interrupt, i.e. the
|
/* Return true if this LR should trigger an EOI maintenance interrupt, i.e. the
|
||||||
* corrsponding bit in EISR is set.
|
* corresponding bit in EISR is set.
|
||||||
*/
|
*/
|
||||||
static inline bool gic_lr_entry_is_eoi(uint32_t entry)
|
static inline bool gic_lr_entry_is_eoi(uint32_t entry)
|
||||||
{
|
{
|
||||||
@ -1333,7 +1333,7 @@ static void gic_dist_writeb(void *opaque, hwaddr offset,
|
|||||||
|
|
||||||
/* ??? This currently clears the pending bit for all CPUs, even
|
/* ??? This currently clears the pending bit for all CPUs, even
|
||||||
for per-CPU interrupts. It's unclear whether this is the
|
for per-CPU interrupts. It's unclear whether this is the
|
||||||
corect behavior. */
|
correct behavior. */
|
||||||
if (value & (1 << i)) {
|
if (value & (1 << i)) {
|
||||||
GIC_DIST_CLEAR_PENDING(irq + i, ALL_CPU_MASK);
|
GIC_DIST_CLEAR_PENDING(irq + i, ALL_CPU_MASK);
|
||||||
}
|
}
|
||||||
|
@ -494,7 +494,7 @@ static MemTxResult gicr_writel(GICv3CPUState *cs, hwaddr offset,
|
|||||||
/* Only the ProcessorSleep bit is writable. When the guest sets
|
/* Only the ProcessorSleep bit is writable. When the guest sets
|
||||||
* it, it requests that we transition the channel between the
|
* it, it requests that we transition the channel between the
|
||||||
* redistributor and the cpu interface to quiescent, and that
|
* redistributor and the cpu interface to quiescent, and that
|
||||||
* we set the ChildrenAsleep bit once the inteface has reached the
|
* we set the ChildrenAsleep bit once the interface has reached the
|
||||||
* quiescent state.
|
* quiescent state.
|
||||||
* Setting the ProcessorSleep to 0 reverses the quiescing, and
|
* Setting the ProcessorSleep to 0 reverses the quiescing, and
|
||||||
* ChildrenAsleep is cleared once the transition is complete.
|
* ChildrenAsleep is cleared once the transition is complete.
|
||||||
|
@ -894,7 +894,7 @@ int armv7m_nvic_complete_irq(NVICState *s, int irq, bool secure)
|
|||||||
vec->active = 0;
|
vec->active = 0;
|
||||||
if (vec->level) {
|
if (vec->level) {
|
||||||
/* Re-pend the exception if it's still held high; only
|
/* Re-pend the exception if it's still held high; only
|
||||||
* happens for extenal IRQs
|
* happens for external IRQs
|
||||||
*/
|
*/
|
||||||
assert(irq >= NVIC_FIRST_IRQ);
|
assert(irq >= NVIC_FIRST_IRQ);
|
||||||
vec->pending = 1;
|
vec->pending = 1;
|
||||||
|
@ -368,7 +368,7 @@ static const MemoryRegionOps allwinner_r40_detect_ops = {
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* mctl_r40_detect_rank_count in u-boot will write the high 1G of DDR
|
* mctl_r40_detect_rank_count in u-boot will write the high 1G of DDR
|
||||||
* to detect wether the board support dual_rank or not. Create a virtual memory
|
* to detect whether the board support dual_rank or not. Create a virtual memory
|
||||||
* if the board's ram_size less or equal than 1G, and set read time out flag of
|
* if the board's ram_size less or equal than 1G, and set read time out flag of
|
||||||
* REG_DRAMCTL_PGSR when the user touch this high dram.
|
* REG_DRAMCTL_PGSR when the user touch this high dram.
|
||||||
*/
|
*/
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
/*
|
/*
|
||||||
* Exynos4210 Pseudo Random Nubmer Generator Emulation
|
* Exynos4210 Pseudo Random Number Generator Emulation
|
||||||
*
|
*
|
||||||
* Copyright (c) 2017 Krzysztof Kozlowski <krzk@kernel.org>
|
* Copyright (c) 2017 Krzysztof Kozlowski <krzk@kernel.org>
|
||||||
*
|
*
|
||||||
|
@ -165,7 +165,7 @@ enum FslIMX7MemoryMap {
|
|||||||
* Some versions of the reference manual claim that UART2 is @
|
* Some versions of the reference manual claim that UART2 is @
|
||||||
* 0x30870000, but experiments with HW + DT files in upstream
|
* 0x30870000, but experiments with HW + DT files in upstream
|
||||||
* Linux kernel show that not to be true and that block is
|
* Linux kernel show that not to be true and that block is
|
||||||
* acutally located @ 0x30890000
|
* actually located @ 0x30890000
|
||||||
*/
|
*/
|
||||||
FSL_IMX7_UART2_ADDR = 0x30890000,
|
FSL_IMX7_UART2_ADDR = 0x30890000,
|
||||||
FSL_IMX7_UART3_ADDR = 0x30880000,
|
FSL_IMX7_UART3_ADDR = 0x30880000,
|
||||||
|
@ -74,7 +74,7 @@ struct NVICState {
|
|||||||
*/
|
*/
|
||||||
bool vectpending_is_s_banked;
|
bool vectpending_is_s_banked;
|
||||||
int exception_prio; /* group prio of the highest prio active exception */
|
int exception_prio; /* group prio of the highest prio active exception */
|
||||||
int vectpending_prio; /* group prio of the exeception in vectpending */
|
int vectpending_prio; /* group prio of the exception in vectpending */
|
||||||
|
|
||||||
MemoryRegion sysregmem;
|
MemoryRegion sysregmem;
|
||||||
|
|
||||||
|
@ -677,7 +677,7 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx,
|
|||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The PSTATE bits only mask the interrupt if we have not overriden the
|
* The PSTATE bits only mask the interrupt if we have not overridden the
|
||||||
* ability above.
|
* ability above.
|
||||||
*/
|
*/
|
||||||
return unmasked || pstate_unmasked;
|
return unmasked || pstate_unmasked;
|
||||||
|
@ -2592,7 +2592,7 @@ static inline bool arm_el_is_aa64(CPUARMState *env, int el)
|
|||||||
return aa64;
|
return aa64;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Function for determing whether guest cp register reads and writes should
|
/* Function for determining whether guest cp register reads and writes should
|
||||||
* access the secure or non-secure bank of a cp register. When EL3 is
|
* access the secure or non-secure bank of a cp register. When EL3 is
|
||||||
* operating in AArch32 state, the NS-bit determines whether the secure
|
* operating in AArch32 state, the NS-bit determines whether the secure
|
||||||
* instance of a cp register should be used. When EL3 is AArch64 (or if
|
* instance of a cp register should be used. When EL3 is AArch64 (or if
|
||||||
|
@ -95,7 +95,7 @@ void arm_cpu_sve_finalize(ARMCPU *cpu, Error **errp)
|
|||||||
|
|
||||||
if (kvm_enabled()) {
|
if (kvm_enabled()) {
|
||||||
/*
|
/*
|
||||||
* For KVM we have to automatically enable all supported unitialized
|
* For KVM we have to automatically enable all supported uninitialized
|
||||||
* lengths, even when the smaller lengths are not all powers-of-two.
|
* lengths, even when the smaller lengths are not all powers-of-two.
|
||||||
*/
|
*/
|
||||||
vq_map |= vq_supported & ~vq_init & vq_mask;
|
vq_map |= vq_supported & ~vq_init & vq_mask;
|
||||||
|
@ -1674,7 +1674,7 @@ static void pmevtyper_rawwrite(CPUARMState *env, const ARMCPRegInfo *ri,
|
|||||||
* pmevtyper_rawwrite is called between a pair of pmu_op_start and
|
* pmevtyper_rawwrite is called between a pair of pmu_op_start and
|
||||||
* pmu_op_finish calls when loading saved state for a migration. Because
|
* pmu_op_finish calls when loading saved state for a migration. Because
|
||||||
* we're potentially updating the type of event here, the value written to
|
* we're potentially updating the type of event here, the value written to
|
||||||
* c14_pmevcntr_delta by the preceeding pmu_op_start call may be for a
|
* c14_pmevcntr_delta by the preceding pmu_op_start call may be for a
|
||||||
* different counter type. Therefore, we need to set this value to the
|
* different counter type. Therefore, we need to set this value to the
|
||||||
* current count for the counter type we're writing so that pmu_op_finish
|
* current count for the counter type we're writing so that pmu_op_finish
|
||||||
* has the correct count for its calculation.
|
* has the correct count for its calculation.
|
||||||
@ -7009,7 +7009,7 @@ static const ARMCPRegInfo rme_reginfo[] = {
|
|||||||
/*
|
/*
|
||||||
* QEMU does not have a way to invalidate by physical address, thus
|
* QEMU does not have a way to invalidate by physical address, thus
|
||||||
* invalidating a range of physical addresses is accomplished by
|
* invalidating a range of physical addresses is accomplished by
|
||||||
* flushing all tlb entries in the outer sharable domain,
|
* flushing all tlb entries in the outer shareable domain,
|
||||||
* just like PAALLOS.
|
* just like PAALLOS.
|
||||||
*/
|
*/
|
||||||
{ .name = "TLBI_RPALOS", .state = ARM_CP_STATE_AA64,
|
{ .name = "TLBI_RPALOS", .state = ARM_CP_STATE_AA64,
|
||||||
|
@ -148,7 +148,7 @@ uint32_t HELPER(v7m_tt)(CPUARMState *env, uint32_t addr, uint32_t op)
|
|||||||
* R: 0 because unpriv and A flag not set
|
* R: 0 because unpriv and A flag not set
|
||||||
* SRVALID: 0 because NS
|
* SRVALID: 0 because NS
|
||||||
* MRVALID: 0 because unpriv and A flag not set
|
* MRVALID: 0 because unpriv and A flag not set
|
||||||
* SREGION: 0 becaus SRVALID is 0
|
* SREGION: 0 because SRVALID is 0
|
||||||
* MREGION: 0 because MRVALID is 0
|
* MREGION: 0 because MRVALID is 0
|
||||||
*/
|
*/
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -182,7 +182,7 @@ void gen_a64_update_pc(DisasContext *s, target_long diff)
|
|||||||
* + for EL2 and EL3 there is only one TBI bit, and if it is set
|
* + for EL2 and EL3 there is only one TBI bit, and if it is set
|
||||||
* then the address is zero-extended, clearing bits [63:56]
|
* then the address is zero-extended, clearing bits [63:56]
|
||||||
* + for EL0 and EL1, TBI0 controls addresses with bit 55 == 0
|
* + for EL0 and EL1, TBI0 controls addresses with bit 55 == 0
|
||||||
* and TBI1 controls addressses with bit 55 == 1.
|
* and TBI1 controls addresses with bit 55 == 1.
|
||||||
* If the appropriate TBI bit is set for the address then
|
* If the appropriate TBI bit is set for the address then
|
||||||
* the address is sign-extended from bit 55 into bits [63:56]
|
* the address is sign-extended from bit 55 into bits [63:56]
|
||||||
*
|
*
|
||||||
@ -2313,7 +2313,7 @@ static void handle_sys(DisasContext *s, bool isread,
|
|||||||
|
|
||||||
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
|
if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
|
||||||
/*
|
/*
|
||||||
* A write to any coprocessor regiser that ends a TB
|
* A write to any coprocessor register that ends a TB
|
||||||
* must rebuild the hflags for the next TB.
|
* must rebuild the hflags for the next TB.
|
||||||
*/
|
*/
|
||||||
gen_rebuild_hflags(s);
|
gen_rebuild_hflags(s);
|
||||||
|
@ -2182,7 +2182,7 @@ static bool trans_VMOV_to_2gp(DisasContext *s, arg_VMOV_to_2gp *a)
|
|||||||
* execution if it is not in an IT block. For us this means
|
* execution if it is not in an IT block. For us this means
|
||||||
* only that if PSR.ECI says we should not be executing the beat
|
* only that if PSR.ECI says we should not be executing the beat
|
||||||
* corresponding to the lane of the vector register being accessed
|
* corresponding to the lane of the vector register being accessed
|
||||||
* then we should skip perfoming the move, and that we need to do
|
* then we should skip performing the move, and that we need to do
|
||||||
* the usual check for bad ECI state and advance of ECI state.
|
* the usual check for bad ECI state and advance of ECI state.
|
||||||
* (If PSR.ECI is non-zero then we cannot be in an IT block.)
|
* (If PSR.ECI is non-zero then we cannot be in an IT block.)
|
||||||
*/
|
*/
|
||||||
@ -2225,7 +2225,7 @@ static bool trans_VMOV_from_2gp(DisasContext *s, arg_VMOV_to_2gp *a)
|
|||||||
* execution if it is not in an IT block. For us this means
|
* execution if it is not in an IT block. For us this means
|
||||||
* only that if PSR.ECI says we should not be executing the beat
|
* only that if PSR.ECI says we should not be executing the beat
|
||||||
* corresponding to the lane of the vector register being accessed
|
* corresponding to the lane of the vector register being accessed
|
||||||
* then we should skip perfoming the move, and that we need to do
|
* then we should skip performing the move, and that we need to do
|
||||||
* the usual check for bad ECI state and advance of ECI state.
|
* the usual check for bad ECI state and advance of ECI state.
|
||||||
* (If PSR.ECI is non-zero then we cannot be in an IT block.)
|
* (If PSR.ECI is non-zero then we cannot be in an IT block.)
|
||||||
*/
|
*/
|
||||||
|
@ -1841,7 +1841,7 @@ TRANS_FEAT(PNEXT, aa64_sve, do_pfirst_pnext, a, gen_helper_sve_pnext)
|
|||||||
|
|
||||||
/* Perform an inline saturating addition of a 32-bit value within
|
/* Perform an inline saturating addition of a 32-bit value within
|
||||||
* a 64-bit register. The second operand is known to be positive,
|
* a 64-bit register. The second operand is known to be positive,
|
||||||
* which halves the comparisions we must perform to bound the result.
|
* which halves the comparisons we must perform to bound the result.
|
||||||
*/
|
*/
|
||||||
static void do_sat_addsub_32(TCGv_i64 reg, TCGv_i64 val, bool u, bool d)
|
static void do_sat_addsub_32(TCGv_i64 reg, TCGv_i64 val, bool u, bool d)
|
||||||
{
|
{
|
||||||
|
@ -144,7 +144,7 @@ static void gen_preserve_fp_state(DisasContext *s, bool skip_context_update)
|
|||||||
* Generate code for M-profile FP context handling: update the
|
* Generate code for M-profile FP context handling: update the
|
||||||
* ownership of the FP context, and create a new context if
|
* ownership of the FP context, and create a new context if
|
||||||
* necessary. This corresponds to the parts of the pseudocode
|
* necessary. This corresponds to the parts of the pseudocode
|
||||||
* ExecuteFPCheck() after the inital PreserveFPState() call.
|
* ExecuteFPCheck() after the initial PreserveFPState() call.
|
||||||
*/
|
*/
|
||||||
static void gen_update_fp_context(DisasContext *s)
|
static void gen_update_fp_context(DisasContext *s)
|
||||||
{
|
{
|
||||||
|
@ -2626,7 +2626,7 @@ void HELPER(gvec_bfmmla)(void *vd, void *vn, void *vm, void *va, uint32_t desc)
|
|||||||
* Process the entire segment at once, writing back the
|
* Process the entire segment at once, writing back the
|
||||||
* results only after we've consumed all of the inputs.
|
* results only after we've consumed all of the inputs.
|
||||||
*
|
*
|
||||||
* Key to indicies by column:
|
* Key to indices by column:
|
||||||
* i j i k j k
|
* i j i k j k
|
||||||
*/
|
*/
|
||||||
sum00 = a[s + H4(0 + 0)];
|
sum00 = a[s + H4(0 + 0)];
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
from __future__ import print_function
|
from __future__ import print_function
|
||||||
#
|
#
|
||||||
# Test the SVE registers are visable and changeable via gdbstub
|
# Test the SVE registers are visible and changeable via gdbstub
|
||||||
#
|
#
|
||||||
# This is launched via tests/guest-debug/run-test.py
|
# This is launched via tests/guest-debug/run-test.py
|
||||||
#
|
#
|
||||||
|
@ -28,7 +28,7 @@ asm(
|
|||||||
" fmopa za1.s, p0/m, p0/m, z0.s, z0.s\n"
|
" fmopa za1.s, p0/m, p0/m, z0.s, z0.s\n"
|
||||||
/*
|
/*
|
||||||
* Read the first 4x4 sub-matrix of elements from tile 1:
|
* Read the first 4x4 sub-matrix of elements from tile 1:
|
||||||
* Note that za1h should be interchangable here.
|
* Note that za1h should be interchangeable here.
|
||||||
*/
|
*/
|
||||||
" mov w12, #0\n"
|
" mov w12, #0\n"
|
||||||
" mova z0.s, p0/m, za1v.s[w12, #0]\n"
|
" mova z0.s, p0/m, za1v.s[w12, #0]\n"
|
||||||
|
@ -9,7 +9,7 @@
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* Semihosting interface on ARM AArch64
|
* Semihosting interface on ARM AArch64
|
||||||
* See "Semihosting for AArch32 and AArch64 Relase 2.0" by ARM
|
* See "Semihosting for AArch32 and AArch64 Release 2.0" by ARM
|
||||||
* w0 - semihosting call number
|
* w0 - semihosting call number
|
||||||
* x1 - semihosting parameter
|
* x1 - semihosting parameter
|
||||||
*/
|
*/
|
||||||
@ -147,7 +147,7 @@ __start:
|
|||||||
* T0SZ[5:0] = 2^(64 - 25)
|
* T0SZ[5:0] = 2^(64 - 25)
|
||||||
*
|
*
|
||||||
* The size of T0SZ controls what the initial lookup level. It
|
* The size of T0SZ controls what the initial lookup level. It
|
||||||
* would be nice to start at level 2 but unfortunatly for a
|
* would be nice to start at level 2 but unfortunately for a
|
||||||
* flat-mapping on the virt machine we need to handle IA's
|
* flat-mapping on the virt machine we need to handle IA's
|
||||||
* with at least 1gb range to see RAM. So we start with a
|
* with at least 1gb range to see RAM. So we start with a
|
||||||
* level 1 lookup.
|
* level 1 lookup.
|
||||||
@ -189,7 +189,7 @@ __start:
|
|||||||
msr cpacr_el1, x0
|
msr cpacr_el1, x0
|
||||||
|
|
||||||
/* Setup some stack space and enter the test code.
|
/* Setup some stack space and enter the test code.
|
||||||
* Assume everthing except the return value is garbage when we
|
* Assume everything except the return value is garbage when we
|
||||||
* return, we won't need it.
|
* return, we won't need it.
|
||||||
*/
|
*/
|
||||||
adrp x0, stack_end
|
adrp x0, stack_end
|
||||||
|
@ -86,7 +86,7 @@ int main(int argc, char *argv[argc])
|
|||||||
}
|
}
|
||||||
ptr_to_heap++;
|
ptr_to_heap++;
|
||||||
}
|
}
|
||||||
ml_printf("r/w to heap upto %p\n", ptr_to_heap);
|
ml_printf("r/w to heap up to %p\n", ptr_to_heap);
|
||||||
|
|
||||||
ml_printf("Passed HeapInfo checks\n");
|
ml_printf("Passed HeapInfo checks\n");
|
||||||
return 0;
|
return 0;
|
||||||
|
Loading…
Reference in New Issue
Block a user