b709da5d29
Different code paths handle memory accesses: - tcg generated code - load/store helpers - atomic helpers This value is saved in cpu->neg.plugin_mem_value_{high,low}. Values are written only for accessed word size (upper bits are not set). Atomic operations are doing read/write at the same time, so we generate two memory callbacks instead of one, to allow plugins to access distinct values. For now, we can have access only up to 128 bits, thus split this in two 64 bits words. When QEMU will support wider operations, we'll be able to reconsider this. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Alex Bennée <alex.bennee@linaro.org> Signed-off-by: Pierrick Bouvier <pierrick.bouvier@linaro.org> Message-Id: <20240724194708.1843704-2-pierrick.bouvier@linaro.org> Signed-off-by: Alex Bennée <alex.bennee@linaro.org> Message-Id: <20240916085400.1046925-5-alex.bennee@linaro.org>
334 lines
12 KiB
C
334 lines
12 KiB
C
/*
|
|
* Atomic helper templates
|
|
* Included from tcg-runtime.c and cputlb.c.
|
|
*
|
|
* Copyright (c) 2016 Red Hat, Inc
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "qemu/plugin.h"
|
|
|
|
#if DATA_SIZE == 16
|
|
# define SUFFIX o
|
|
# define DATA_TYPE Int128
|
|
# define BSWAP bswap128
|
|
# define SHIFT 4
|
|
#elif DATA_SIZE == 8
|
|
# define SUFFIX q
|
|
# define DATA_TYPE aligned_uint64_t
|
|
# define SDATA_TYPE aligned_int64_t
|
|
# define BSWAP bswap64
|
|
# define SHIFT 3
|
|
#elif DATA_SIZE == 4
|
|
# define SUFFIX l
|
|
# define DATA_TYPE uint32_t
|
|
# define SDATA_TYPE int32_t
|
|
# define BSWAP bswap32
|
|
# define SHIFT 2
|
|
#elif DATA_SIZE == 2
|
|
# define SUFFIX w
|
|
# define DATA_TYPE uint16_t
|
|
# define SDATA_TYPE int16_t
|
|
# define BSWAP bswap16
|
|
# define SHIFT 1
|
|
#elif DATA_SIZE == 1
|
|
# define SUFFIX b
|
|
# define DATA_TYPE uint8_t
|
|
# define SDATA_TYPE int8_t
|
|
# define BSWAP
|
|
# define SHIFT 0
|
|
#else
|
|
# error unsupported data size
|
|
#endif
|
|
|
|
#if DATA_SIZE == 16
|
|
# define VALUE_LOW(val) int128_getlo(val)
|
|
# define VALUE_HIGH(val) int128_gethi(val)
|
|
#else
|
|
# define VALUE_LOW(val) val
|
|
# define VALUE_HIGH(val) 0
|
|
#endif
|
|
|
|
#if DATA_SIZE >= 4
|
|
# define ABI_TYPE DATA_TYPE
|
|
#else
|
|
# define ABI_TYPE uint32_t
|
|
#endif
|
|
|
|
/* Define host-endian atomic operations. Note that END is used within
|
|
the ATOMIC_NAME macro, and redefined below. */
|
|
#if DATA_SIZE == 1
|
|
# define END
|
|
#elif HOST_BIG_ENDIAN
|
|
# define END _be
|
|
#else
|
|
# define END _le
|
|
#endif
|
|
|
|
ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, abi_ptr addr,
|
|
ABI_TYPE cmpv, ABI_TYPE newv,
|
|
MemOpIdx oi, uintptr_t retaddr)
|
|
{
|
|
DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
|
|
DATA_SIZE, retaddr);
|
|
DATA_TYPE ret;
|
|
|
|
#if DATA_SIZE == 16
|
|
ret = atomic16_cmpxchg(haddr, cmpv, newv);
|
|
#else
|
|
ret = qatomic_cmpxchg__nocheck(haddr, cmpv, newv);
|
|
#endif
|
|
ATOMIC_MMU_CLEANUP;
|
|
atomic_trace_rmw_post(env, addr,
|
|
VALUE_LOW(ret),
|
|
VALUE_HIGH(ret),
|
|
VALUE_LOW(newv),
|
|
VALUE_HIGH(newv),
|
|
oi);
|
|
return ret;
|
|
}
|
|
|
|
#if DATA_SIZE < 16
|
|
ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, abi_ptr addr, ABI_TYPE val,
|
|
MemOpIdx oi, uintptr_t retaddr)
|
|
{
|
|
DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
|
|
DATA_SIZE, retaddr);
|
|
DATA_TYPE ret;
|
|
|
|
ret = qatomic_xchg__nocheck(haddr, val);
|
|
ATOMIC_MMU_CLEANUP;
|
|
atomic_trace_rmw_post(env, addr,
|
|
VALUE_LOW(ret),
|
|
VALUE_HIGH(ret),
|
|
VALUE_LOW(val),
|
|
VALUE_HIGH(val),
|
|
oi);
|
|
return ret;
|
|
}
|
|
|
|
#define GEN_ATOMIC_HELPER(X) \
|
|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr, \
|
|
ABI_TYPE val, MemOpIdx oi, uintptr_t retaddr) \
|
|
{ \
|
|
DATA_TYPE *haddr, ret; \
|
|
haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr); \
|
|
ret = qatomic_##X(haddr, val); \
|
|
ATOMIC_MMU_CLEANUP; \
|
|
atomic_trace_rmw_post(env, addr, \
|
|
VALUE_LOW(ret), \
|
|
VALUE_HIGH(ret), \
|
|
VALUE_LOW(val), \
|
|
VALUE_HIGH(val), \
|
|
oi); \
|
|
return ret; \
|
|
}
|
|
|
|
GEN_ATOMIC_HELPER(fetch_add)
|
|
GEN_ATOMIC_HELPER(fetch_and)
|
|
GEN_ATOMIC_HELPER(fetch_or)
|
|
GEN_ATOMIC_HELPER(fetch_xor)
|
|
GEN_ATOMIC_HELPER(add_fetch)
|
|
GEN_ATOMIC_HELPER(and_fetch)
|
|
GEN_ATOMIC_HELPER(or_fetch)
|
|
GEN_ATOMIC_HELPER(xor_fetch)
|
|
|
|
#undef GEN_ATOMIC_HELPER
|
|
|
|
/*
|
|
* These helpers are, as a whole, full barriers. Within the helper,
|
|
* the leading barrier is explicit and the trailing barrier is within
|
|
* cmpxchg primitive.
|
|
*
|
|
* Trace this load + RMW loop as a single RMW op. This way, regardless
|
|
* of CF_PARALLEL's value, we'll trace just a read and a write.
|
|
*/
|
|
#define GEN_ATOMIC_HELPER_FN(X, FN, XDATA_TYPE, RET) \
|
|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr, \
|
|
ABI_TYPE xval, MemOpIdx oi, uintptr_t retaddr) \
|
|
{ \
|
|
XDATA_TYPE *haddr, cmp, old, new, val = xval; \
|
|
haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr); \
|
|
smp_mb(); \
|
|
cmp = qatomic_read__nocheck(haddr); \
|
|
do { \
|
|
old = cmp; new = FN(old, val); \
|
|
cmp = qatomic_cmpxchg__nocheck(haddr, old, new); \
|
|
} while (cmp != old); \
|
|
ATOMIC_MMU_CLEANUP; \
|
|
atomic_trace_rmw_post(env, addr, \
|
|
VALUE_LOW(old), \
|
|
VALUE_HIGH(old), \
|
|
VALUE_LOW(xval), \
|
|
VALUE_HIGH(xval), \
|
|
oi); \
|
|
return RET; \
|
|
}
|
|
|
|
GEN_ATOMIC_HELPER_FN(fetch_smin, MIN, SDATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(fetch_umin, MIN, DATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(fetch_smax, MAX, SDATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(fetch_umax, MAX, DATA_TYPE, old)
|
|
|
|
GEN_ATOMIC_HELPER_FN(smin_fetch, MIN, SDATA_TYPE, new)
|
|
GEN_ATOMIC_HELPER_FN(umin_fetch, MIN, DATA_TYPE, new)
|
|
GEN_ATOMIC_HELPER_FN(smax_fetch, MAX, SDATA_TYPE, new)
|
|
GEN_ATOMIC_HELPER_FN(umax_fetch, MAX, DATA_TYPE, new)
|
|
|
|
#undef GEN_ATOMIC_HELPER_FN
|
|
#endif /* DATA SIZE < 16 */
|
|
|
|
#undef END
|
|
|
|
#if DATA_SIZE > 1
|
|
|
|
/* Define reverse-host-endian atomic operations. Note that END is used
|
|
within the ATOMIC_NAME macro. */
|
|
#if HOST_BIG_ENDIAN
|
|
# define END _le
|
|
#else
|
|
# define END _be
|
|
#endif
|
|
|
|
ABI_TYPE ATOMIC_NAME(cmpxchg)(CPUArchState *env, abi_ptr addr,
|
|
ABI_TYPE cmpv, ABI_TYPE newv,
|
|
MemOpIdx oi, uintptr_t retaddr)
|
|
{
|
|
DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
|
|
DATA_SIZE, retaddr);
|
|
DATA_TYPE ret;
|
|
|
|
#if DATA_SIZE == 16
|
|
ret = atomic16_cmpxchg(haddr, BSWAP(cmpv), BSWAP(newv));
|
|
#else
|
|
ret = qatomic_cmpxchg__nocheck(haddr, BSWAP(cmpv), BSWAP(newv));
|
|
#endif
|
|
ATOMIC_MMU_CLEANUP;
|
|
atomic_trace_rmw_post(env, addr,
|
|
VALUE_LOW(ret),
|
|
VALUE_HIGH(ret),
|
|
VALUE_LOW(newv),
|
|
VALUE_HIGH(newv),
|
|
oi);
|
|
return BSWAP(ret);
|
|
}
|
|
|
|
#if DATA_SIZE < 16
|
|
ABI_TYPE ATOMIC_NAME(xchg)(CPUArchState *env, abi_ptr addr, ABI_TYPE val,
|
|
MemOpIdx oi, uintptr_t retaddr)
|
|
{
|
|
DATA_TYPE *haddr = atomic_mmu_lookup(env_cpu(env), addr, oi,
|
|
DATA_SIZE, retaddr);
|
|
ABI_TYPE ret;
|
|
|
|
ret = qatomic_xchg__nocheck(haddr, BSWAP(val));
|
|
ATOMIC_MMU_CLEANUP;
|
|
atomic_trace_rmw_post(env, addr,
|
|
VALUE_LOW(ret),
|
|
VALUE_HIGH(ret),
|
|
VALUE_LOW(val),
|
|
VALUE_HIGH(val),
|
|
oi);
|
|
return BSWAP(ret);
|
|
}
|
|
|
|
#define GEN_ATOMIC_HELPER(X) \
|
|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr, \
|
|
ABI_TYPE val, MemOpIdx oi, uintptr_t retaddr) \
|
|
{ \
|
|
DATA_TYPE *haddr, ret; \
|
|
haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr); \
|
|
ret = qatomic_##X(haddr, BSWAP(val)); \
|
|
ATOMIC_MMU_CLEANUP; \
|
|
atomic_trace_rmw_post(env, addr, \
|
|
VALUE_LOW(ret), \
|
|
VALUE_HIGH(ret), \
|
|
VALUE_LOW(val), \
|
|
VALUE_HIGH(val), \
|
|
oi); \
|
|
return BSWAP(ret); \
|
|
}
|
|
|
|
GEN_ATOMIC_HELPER(fetch_and)
|
|
GEN_ATOMIC_HELPER(fetch_or)
|
|
GEN_ATOMIC_HELPER(fetch_xor)
|
|
GEN_ATOMIC_HELPER(and_fetch)
|
|
GEN_ATOMIC_HELPER(or_fetch)
|
|
GEN_ATOMIC_HELPER(xor_fetch)
|
|
|
|
#undef GEN_ATOMIC_HELPER
|
|
|
|
/* These helpers are, as a whole, full barriers. Within the helper,
|
|
* the leading barrier is explicit and the trailing barrier is within
|
|
* cmpxchg primitive.
|
|
*
|
|
* Trace this load + RMW loop as a single RMW op. This way, regardless
|
|
* of CF_PARALLEL's value, we'll trace just a read and a write.
|
|
*/
|
|
#define GEN_ATOMIC_HELPER_FN(X, FN, XDATA_TYPE, RET) \
|
|
ABI_TYPE ATOMIC_NAME(X)(CPUArchState *env, abi_ptr addr, \
|
|
ABI_TYPE xval, MemOpIdx oi, uintptr_t retaddr) \
|
|
{ \
|
|
XDATA_TYPE *haddr, ldo, ldn, old, new, val = xval; \
|
|
haddr = atomic_mmu_lookup(env_cpu(env), addr, oi, DATA_SIZE, retaddr); \
|
|
smp_mb(); \
|
|
ldn = qatomic_read__nocheck(haddr); \
|
|
do { \
|
|
ldo = ldn; old = BSWAP(ldo); new = FN(old, val); \
|
|
ldn = qatomic_cmpxchg__nocheck(haddr, ldo, BSWAP(new)); \
|
|
} while (ldo != ldn); \
|
|
ATOMIC_MMU_CLEANUP; \
|
|
atomic_trace_rmw_post(env, addr, \
|
|
VALUE_LOW(old), \
|
|
VALUE_HIGH(old), \
|
|
VALUE_LOW(xval), \
|
|
VALUE_HIGH(xval), \
|
|
oi); \
|
|
return RET; \
|
|
}
|
|
|
|
GEN_ATOMIC_HELPER_FN(fetch_smin, MIN, SDATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(fetch_umin, MIN, DATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(fetch_smax, MAX, SDATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(fetch_umax, MAX, DATA_TYPE, old)
|
|
|
|
GEN_ATOMIC_HELPER_FN(smin_fetch, MIN, SDATA_TYPE, new)
|
|
GEN_ATOMIC_HELPER_FN(umin_fetch, MIN, DATA_TYPE, new)
|
|
GEN_ATOMIC_HELPER_FN(smax_fetch, MAX, SDATA_TYPE, new)
|
|
GEN_ATOMIC_HELPER_FN(umax_fetch, MAX, DATA_TYPE, new)
|
|
|
|
/* Note that for addition, we need to use a separate cmpxchg loop instead
|
|
of bswaps for the reverse-host-endian helpers. */
|
|
#define ADD(X, Y) (X + Y)
|
|
GEN_ATOMIC_HELPER_FN(fetch_add, ADD, DATA_TYPE, old)
|
|
GEN_ATOMIC_HELPER_FN(add_fetch, ADD, DATA_TYPE, new)
|
|
#undef ADD
|
|
|
|
#undef GEN_ATOMIC_HELPER_FN
|
|
#endif /* DATA_SIZE < 16 */
|
|
|
|
#undef END
|
|
#endif /* DATA_SIZE > 1 */
|
|
|
|
#undef BSWAP
|
|
#undef ABI_TYPE
|
|
#undef DATA_TYPE
|
|
#undef SDATA_TYPE
|
|
#undef SUFFIX
|
|
#undef DATA_SIZE
|
|
#undef SHIFT
|
|
#undef VALUE_LOW
|
|
#undef VALUE_HIGH
|