Viewing file: st2_s64.c (4.74 KB) -rw-r--r-- Select action/file-type: (+) | (+) | (+) | Code (+) | Session (+) | (+) | SDB (+) | (+) | (+) | (+) | (+) | (+) |
/* { dg-final { check-function-bodies "**" "" "-DCHECK_ASM" { target { ! ilp32 } } } } */
#include "test_sve_acle.h"
/* ** st2_s64_base: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_s64_base, svint64x2_t, int64_t, svst2_s64 (p0, x0, z0), svst2 (p0, x0, z0))
/* ** st2_s64_index: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, x1, lsl 3\] ** ret */ TEST_STORE (st2_s64_index, svint64x2_t, int64_t, svst2_s64 (p0, x0 + x1, z0), svst2 (p0, x0 + x1, z0))
/* Moving the constant into a register would also be OK. */ /* ** st2_s64_1: ** incb x0 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_s64_1, svint64x2_t, int64_t, svst2_s64 (p0, x0 + svcntd (), z0), svst2 (p0, x0 + svcntd (), z0))
/* ** st2_s64_2: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #2, mul vl\] ** ret */ TEST_STORE (st2_s64_2, svint64x2_t, int64_t, svst2_s64 (p0, x0 + svcntd () * 2, z0), svst2 (p0, x0 + svcntd () * 2, z0))
/* ** st2_s64_14: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #14, mul vl\] ** ret */ TEST_STORE (st2_s64_14, svint64x2_t, int64_t, svst2_s64 (p0, x0 + svcntd () * 14, z0), svst2 (p0, x0 + svcntd () * 14, z0))
/* Moving the constant into a register would also be OK. */ /* ** st2_s64_16: ** incb x0, all, mul #16 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_s64_16, svint64x2_t, int64_t, svst2_s64 (p0, x0 + svcntd () * 16, z0), svst2 (p0, x0 + svcntd () * 16, z0))
/* Moving the constant into a register would also be OK. */ /* ** st2_s64_m1: ** decb x0 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_s64_m1, svint64x2_t, int64_t, svst2_s64 (p0, x0 - svcntd (), z0), svst2 (p0, x0 - svcntd (), z0))
/* ** st2_s64_m2: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #-2, mul vl\] ** ret */ TEST_STORE (st2_s64_m2, svint64x2_t, int64_t, svst2_s64 (p0, x0 - svcntd () * 2, z0), svst2 (p0, x0 - svcntd () * 2, z0))
/* ** st2_s64_m16: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #-16, mul vl\] ** ret */ TEST_STORE (st2_s64_m16, svint64x2_t, int64_t, svst2_s64 (p0, x0 - svcntd () * 16, z0), svst2 (p0, x0 - svcntd () * 16, z0))
/* ** st2_s64_m18: ** addvl (x[0-9]+), x0, #-18 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[\1\] ** ret */ TEST_STORE (st2_s64_m18, svint64x2_t, int64_t, svst2_s64 (p0, x0 - svcntd () * 18, z0), svst2 (p0, x0 - svcntd () * 18, z0))
/* ** st2_vnum_s64_0: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_vnum_s64_0, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, 0, z0), svst2_vnum (p0, x0, 0, z0))
/* Moving the constant into a register would also be OK. */ /* ** st2_vnum_s64_1: ** incb x0 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_vnum_s64_1, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, 1, z0), svst2_vnum (p0, x0, 1, z0))
/* ** st2_vnum_s64_2: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #2, mul vl\] ** ret */ TEST_STORE (st2_vnum_s64_2, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, 2, z0), svst2_vnum (p0, x0, 2, z0))
/* ** st2_vnum_s64_14: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #14, mul vl\] ** ret */ TEST_STORE (st2_vnum_s64_14, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, 14, z0), svst2_vnum (p0, x0, 14, z0))
/* Moving the constant into a register would also be OK. */ /* ** st2_vnum_s64_16: ** incb x0, all, mul #16 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_vnum_s64_16, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, 16, z0), svst2_vnum (p0, x0, 16, z0))
/* Moving the constant into a register would also be OK. */ /* ** st2_vnum_s64_m1: ** decb x0 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0\] ** ret */ TEST_STORE (st2_vnum_s64_m1, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, -1, z0), svst2_vnum (p0, x0, -1, z0))
/* ** st2_vnum_s64_m2: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #-2, mul vl\] ** ret */ TEST_STORE (st2_vnum_s64_m2, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, -2, z0), svst2_vnum (p0, x0, -2, z0))
/* ** st2_vnum_s64_m16: ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[x0, #-16, mul vl\] ** ret */ TEST_STORE (st2_vnum_s64_m16, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, -16, z0), svst2_vnum (p0, x0, -16, z0))
/* ** st2_vnum_s64_m18: ** addvl (x[0-9]+), x0, #-18 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[\1\] ** ret */ TEST_STORE (st2_vnum_s64_m18, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, -18, z0), svst2_vnum (p0, x0, -18, z0))
/* Using MUL to calculate an index would also be OK. */ /* ** st2_vnum_s64_x1: ** cntb (x[0-9]+) ** madd (x[0-9]+), (x1, \1|\1, x1), x0 ** st2d {z0\.d(?: - |, )z1\.d}, p0, \[\2\] ** ret */ TEST_STORE (st2_vnum_s64_x1, svint64x2_t, int64_t, svst2_vnum_s64 (p0, x0, x1, z0), svst2_vnum (p0, x0, x1, z0))
|