qemu/tests/tcg/hexagon/misc.c
Mukilan Thiyagarajan eaee3b6faf tests/tcg/hexagon: fix underspecifed asm constraints
There are two test cases where the inline asm doesn't
have the correct constraints causing them to fail.

In misc.c, the 'result' output needs the early clobber
modifier since the rest of the inputs are read after
assignment to the output register.

In mem_noshuf.c, the register r7 is written to but
not specified in the clobber list.

Signed-off-by: Mukilan Thiyagarajan <quic_mthiyaga@quicinc.com>
Signed-off-by: Taylor Simpson <tsimpson@quicinc.com>
Reviewed-by: Taylor Simpson <tsimpson@quicinc.com>
Message-Id: <20221229081836.12130-1-quic_mthiyaga@quicinc.com>
2023-01-05 09:19:02 -08:00

474 lines
12 KiB
C

/*
* Copyright(c) 2019-2021 Qualcomm Innovation Center, Inc. All Rights Reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, see <http://www.gnu.org/licenses/>.
*/
#include <stdio.h>
#include <string.h>
typedef unsigned char uint8_t;
typedef unsigned short uint16_t;
typedef unsigned int uint32_t;
static inline void S4_storerhnew_rr(void *p, int index, uint16_t v)
{
asm volatile("{\n\t"
" r0 = %0\n\n"
" memh(%1+%2<<#2) = r0.new\n\t"
"}\n"
:: "r"(v), "r"(p), "r"(index)
: "r0", "memory");
}
static uint32_t data;
static inline void *S4_storerbnew_ap(uint8_t v)
{
void *ret;
asm volatile("{\n\t"
" r0 = %1\n\n"
" memb(%0 = ##data) = r0.new\n\t"
"}\n"
: "=r"(ret)
: "r"(v)
: "r0", "memory");
return ret;
}
static inline void *S4_storerhnew_ap(uint16_t v)
{
void *ret;
asm volatile("{\n\t"
" r0 = %1\n\n"
" memh(%0 = ##data) = r0.new\n\t"
"}\n"
: "=r"(ret)
: "r"(v)
: "r0", "memory");
return ret;
}
static inline void *S4_storerinew_ap(uint32_t v)
{
void *ret;
asm volatile("{\n\t"
" r0 = %1\n\n"
" memw(%0 = ##data) = r0.new\n\t"
"}\n"
: "=r"(ret)
: "r"(v)
: "r0", "memory");
return ret;
}
static inline void S4_storeirbt_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (p0) memb(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirbf_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (!p0) memb(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirbtnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (p0.new) memb(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirbfnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (!p0.new) memb(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirht_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (p0) memh(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirhf_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (!p0) memh(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirhtnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (p0.new) memh(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirhfnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (!p0.new) memh(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirit_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (p0) memw(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirif_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (!p0) memw(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeiritnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (p0.new) memw(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirifnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (!p0.new) memw(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static int L2_ploadrifnew_pi(void *p, int pred)
{
int result;
asm volatile("%0 = #31\n\t"
"{\n\t"
" p0 = cmp.eq(%2, #1)\n\t"
" if (!p0.new) %0 = memw(%1++#4)\n\t"
"}\n\t"
: "=&r"(result), "+r"(p) : "r"(pred)
: "p0");
return result;
}
/*
* Test that compound-compare-jump is executed in 2 parts
* First we have to do all the compares in the packet and
* account for auto-anding. Then, we can do the predicated
* jump.
*/
static inline int cmpnd_cmp_jump(void)
{
int retval;
asm ("r5 = #7\n\t"
"r6 = #9\n\t"
"{\n\t"
" p0 = cmp.eq(r5, #7)\n\t"
" if (p0.new) jump:nt 1f\n\t"
" p0 = cmp.eq(r6, #7)\n\t"
"}\n\t"
"%0 = #12\n\t"
"jump 2f\n\t"
"1:\n\t"
"%0 = #13\n\t"
"2:\n\t"
: "=r"(retval) :: "r5", "r6", "p0");
return retval;
}
static inline int test_clrtnew(int arg1, int old_val)
{
int ret;
asm volatile("r5 = %2\n\t"
"{\n\t"
"p0 = cmp.eq(%1, #1)\n\t"
"if (p0.new) r5=#0\n\t"
"}\n\t"
"%0 = r5\n\t"
: "=r"(ret)
: "r"(arg1), "r"(old_val)
: "p0", "r5");
return ret;
}
int err;
static void check(int val, int expect)
{
if (val != expect) {
printf("ERROR: 0x%04x != 0x%04x\n", val, expect);
err++;
}
}
static void check64(long long val, long long expect)
{
if (val != expect) {
printf("ERROR: 0x%016llx != 0x%016llx\n", val, expect);
err++;
}
}
uint32_t init[10] = { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9 };
uint32_t array[10];
uint32_t early_exit;
/*
* Write this as a function because we can't guarantee the compiler will
* allocate a frame with just the SL2_return_tnew packet.
*/
static void SL2_return_tnew(int x);
asm ("SL2_return_tnew:\n\t"
" allocframe(#0)\n\t"
" r1 = #1\n\t"
" memw(##early_exit) = r1\n\t"
" {\n\t"
" p0 = cmp.eq(r0, #1)\n\t"
" if (p0.new) dealloc_return:nt\n\t" /* SL2_return_tnew */
" }\n\t"
" r1 = #0\n\t"
" memw(##early_exit) = r1\n\t"
" dealloc_return\n\t"
);
static long long creg_pair(int x, int y)
{
long long retval;
asm ("m0 = %1\n\t"
"m1 = %2\n\t"
"%0 = c7:6\n\t"
: "=r"(retval) : "r"(x), "r"(y) : "m0", "m1");
return retval;
}
static long long decbin(long long x, long long y, int *pred)
{
long long retval;
asm ("%0 = decbin(%2, %3)\n\t"
"%1 = p0\n\t"
: "=r"(retval), "=r"(*pred)
: "r"(x), "r"(y));
return retval;
}
/* Check that predicates are auto-and'ed in a packet */
static int auto_and(void)
{
int retval;
asm ("r5 = #1\n\t"
"{\n\t"
" p0 = cmp.eq(r1, #1)\n\t"
" p0 = cmp.eq(r1, #2)\n\t"
"}\n\t"
"%0 = p0\n\t"
: "=r"(retval)
:
: "r5", "p0");
return retval;
}
void test_lsbnew(void)
{
int result;
asm("r0 = #2\n\t"
"r1 = #5\n\t"
"{\n\t"
" p0 = r0\n\t"
" if (p0.new) r1 = #3\n\t"
"}\n\t"
"%0 = r1\n\t"
: "=r"(result) :: "r0", "r1", "p0");
check(result, 5);
}
void test_l2fetch(void)
{
/* These don't do anything in qemu, just make sure they don't assert */
asm volatile ("l2fetch(r0, r1)\n\t"
"l2fetch(r0, r3:2)\n\t");
}
int main()
{
int res;
long long res64;
int pred;
memcpy(array, init, sizeof(array));
S4_storerhnew_rr(array, 4, 0xffff);
check(array[4], 0xffff);
data = ~0;
check((uint32_t)S4_storerbnew_ap(0x12), (uint32_t)&data);
check(data, 0xffffff12);
data = ~0;
check((uint32_t)S4_storerhnew_ap(0x1234), (uint32_t)&data);
check(data, 0xffff1234);
data = ~0;
check((uint32_t)S4_storerinew_ap(0x12345678), (uint32_t)&data);
check(data, 0x12345678);
/* Byte */
memcpy(array, init, sizeof(array));
S4_storeirbt_io(&array[1], 1);
check(array[2], 27);
S4_storeirbt_io(&array[2], 0);
check(array[3], 3);
memcpy(array, init, sizeof(array));
S4_storeirbf_io(&array[3], 0);
check(array[4], 27);
S4_storeirbf_io(&array[4], 1);
check(array[5], 5);
memcpy(array, init, sizeof(array));
S4_storeirbtnew_io(&array[5], 1);
check(array[6], 27);
S4_storeirbtnew_io(&array[6], 0);
check(array[7], 7);
memcpy(array, init, sizeof(array));
S4_storeirbfnew_io(&array[7], 0);
check(array[8], 27);
S4_storeirbfnew_io(&array[8], 1);
check(array[9], 9);
/* Half word */
memcpy(array, init, sizeof(array));
S4_storeirht_io(&array[1], 1);
check(array[2], 27);
S4_storeirht_io(&array[2], 0);
check(array[3], 3);
memcpy(array, init, sizeof(array));
S4_storeirhf_io(&array[3], 0);
check(array[4], 27);
S4_storeirhf_io(&array[4], 1);
check(array[5], 5);
memcpy(array, init, sizeof(array));
S4_storeirhtnew_io(&array[5], 1);
check(array[6], 27);
S4_storeirhtnew_io(&array[6], 0);
check(array[7], 7);
memcpy(array, init, sizeof(array));
S4_storeirhfnew_io(&array[7], 0);
check(array[8], 27);
S4_storeirhfnew_io(&array[8], 1);
check(array[9], 9);
/* Word */
memcpy(array, init, sizeof(array));
S4_storeirit_io(&array[1], 1);
check(array[2], 27);
S4_storeirit_io(&array[2], 0);
check(array[3], 3);
memcpy(array, init, sizeof(array));
S4_storeirif_io(&array[3], 0);
check(array[4], 27);
S4_storeirif_io(&array[4], 1);
check(array[5], 5);
memcpy(array, init, sizeof(array));
S4_storeiritnew_io(&array[5], 1);
check(array[6], 27);
S4_storeiritnew_io(&array[6], 0);
check(array[7], 7);
memcpy(array, init, sizeof(array));
S4_storeirifnew_io(&array[7], 0);
check(array[8], 27);
S4_storeirifnew_io(&array[8], 1);
check(array[9], 9);
memcpy(array, init, sizeof(array));
res = L2_ploadrifnew_pi(&array[6], 0);
check(res, 6);
res = L2_ploadrifnew_pi(&array[7], 1);
check(res, 31);
int x = cmpnd_cmp_jump();
check(x, 12);
SL2_return_tnew(0);
check(early_exit, 0);
SL2_return_tnew(1);
check(early_exit, 1);
long long pair = creg_pair(5, 7);
check((int)pair, 5);
check((int)(pair >> 32), 7);
res = test_clrtnew(1, 7);
check(res, 0);
res = test_clrtnew(2, 7);
check(res, 7);
res64 = decbin(0xf0f1f2f3f4f5f6f7LL, 0x7f6f5f4f3f2f1f0fLL, &pred);
check64(res64, 0x357980003700010cLL);
check(pred, 0);
res64 = decbin(0xfLL, 0x1bLL, &pred);
check64(res64, 0x78000100LL);
check(pred, 1);
res = auto_and();
check(res, 0);
test_lsbnew();
test_l2fetch();
puts(err ? "FAIL" : "PASS");
return err;
}