qemu/tests/tcg/xtensa/test_dfp0_arith.S

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

163 lines
6.1 KiB
ArmAsm
Raw Normal View History

#include "macros.inc"
#include "fpu.h"
test_suite fp0_arith
#if XCHAL_HAVE_DFP
.macro movfp fr, v
movi a2, ((\v) >> 32) & 0xffffffff
movi a3, ((\v) & 0xffffffff)
wfrd \fr, a2, a3
.endm
.macro check_res fr, r, sr
rfrd a2, \fr
dump a2
movi a3, ((\r) >> 32) & 0xffffffff
assert eq, a2, a3
rfr a2, \fr
dump a2
movi a3, ((\r) & 0xffffffff)
assert eq, a2, a3
rur a2, fsr
movi a3, \sr
assert eq, a2, a3
.endm
test add_d
movi a2, 1
wsr a2, cpenable
/* MAX_FLOAT + MAX_FLOAT = +inf/MAX_FLOAT */
test_op2 add.d, f6, f7, f8, F64_MAX, F64_MAX, \
F64_PINF, F64_MAX, F64_PINF, F64_MAX, \
FSR_OI, FSR_OI, FSR_OI, FSR_OI
test_end
test add_d_inf
/* 1 + +inf = +inf */
test_op2 add.d, f6, f7, f8, F64_1, F64_PINF, \
F64_PINF, F64_PINF, F64_PINF, F64_PINF, \
FSR__, FSR__, FSR__, FSR__
/* +inf + -inf = default NaN */
test_op2 add.d, f0, f1, f2, F64_PINF, F64_NINF, \
F64_DNAN, F64_DNAN, F64_DNAN, F64_DNAN, \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
test add_d_nan_dfpu
/* 1 + QNaN = QNaN */
test_op2 add.d, f9, f10, f11, F64_1, F64_QNAN(1), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
/* 1 + SNaN = QNaN */
test_op2 add.d, f12, f13, f14, F64_1, F64_SNAN(1), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* SNaN1 + SNaN2 = QNaN2 */
test_op2 add.d, f15, f0, f1, F64_SNAN(1), F64_SNAN(2), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* QNaN1 + SNaN2 = QNaN2 */
test_op2 add.d, f5, f6, f7, F64_QNAN(1), F64_SNAN(2), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* SNaN1 + QNaN2 = QNaN2 */
test_op2 add.d, f8, f9, f10, F64_SNAN(1), F64_QNAN(2), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
test sub_d
/* norm - norm = denorm */
test_op2 sub.d, f6, f7, f8, F64_MIN_NORM | 1, F64_MIN_NORM, \
0x00000001, 0x00000001, 0x00000001, 0x00000001, \
FSR__, FSR__, FSR__, FSR__
test_end
test mul_d
test_op2 mul.d, f0, f1, f2, F64_1 | 1, F64_1 | 1, \
F64_1 | 2, F64_1 | 2, F64_1 | 3, F64_1 | 2, \
FSR_I, FSR_I, FSR_I, FSR_I
/* MAX_FLOAT/2 * MAX_FLOAT/2 = +inf/MAX_FLOAT */
test_op2 mul.d, f6, f7, f8, F64_MAX_2, F64_MAX_2, \
F64_PINF, F64_MAX, F64_PINF, F64_MAX, \
FSR_OI, FSR_OI, FSR_OI, FSR_OI
/* min norm * min norm = 0/denorm */
test_op2 mul.d, f6, f7, f8, F64_MIN_NORM, F64_MIN_NORM, \
F64_0, F64_0, 0x00000001, F64_0, \
FSR_UI, FSR_UI, FSR_UI, FSR_UI
/* inf * 0 = default NaN */
test_op2 mul.d, f6, f7, f8, F64_PINF, F64_0, \
F64_DNAN, F64_DNAN, F64_DNAN, F64_DNAN, \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
test madd_d
test_op3 madd.d, f0, f1, f2, f0, F64_0, F64_1 | 1, F64_1 | 1, \
F64_1 | 2, F64_1 | 2, F64_1 | 3, F64_1 | 2, \
FSR_I, FSR_I, FSR_I, FSR_I
test_end
test madd_d_precision
test_op3 madd.d, f0, f1, f2, f0, \
F64_MINUS | F64_1 | 2, F64_1 | 1, F64_1 | 1, \
0x3970000000000000, 0x3970000000000000, 0x3970000000000000, 0x3970000000000000, \
FSR__, FSR__, FSR__, FSR__
test_end
test madd_d_nan_dfpu
/* DFPU madd/msub NaN1, NaN2, NaN3 priority: NaN1, NaN3, NaN2 */
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_1, F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_QNAN(2), F64_1, \
F64_QNAN(2), F64_QNAN(2), F64_QNAN(2), F64_QNAN(2), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_1, F64_QNAN(3), \
F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_QNAN(2), F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_1, F64_QNAN(3), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_QNAN(2), F64_QNAN(3), \
F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), F64_QNAN(3), \
FSR__, FSR__, FSR__, FSR__
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_QNAN(2), F64_QNAN(3), \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR__, FSR__, FSR__, FSR__
/* inf * 0 = default NaN */
test_op3 madd.d, f0, f1, f2, f0, F64_1, F64_PINF, F64_0, \
F64_DNAN, F64_DNAN, F64_DNAN, F64_DNAN, \
FSR_V, FSR_V, FSR_V, FSR_V
/* inf * 0 + SNaN1 = QNaN1 */
test_op3 madd.d, f0, f1, f2, f0, F64_SNAN(1), F64_PINF, F64_0, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* inf * 0 + QNaN1 = QNaN1 */
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_PINF, F64_0, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
/* madd/msub SNaN turns to QNaN and sets Invalid flag */
test_op3 madd.d, f0, f1, f2, f0, F64_SNAN(1), F64_1, F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
test_op3 madd.d, f0, f1, f2, f0, F64_QNAN(1), F64_SNAN(2), F64_1, \
F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), F64_QNAN(1), \
FSR_V, FSR_V, FSR_V, FSR_V
test_end
#endif
test_suite_end