mirror of https://github.com/xemu-project/xemu.git
target/loongarch: Implement xvsrlrn xvsrarn
This patch includes: - XVSRLRN.{B.H/H.W/W.D}; - XVSRARN.{B.H/H.W/W.D}; - XVSRLRNI.{B.H/H.W/W.D/D.Q}; - XVSRARNI.{B.H/H.W/W.D/D.Q}. Signed-off-by: Song Gao <gaosong@loongson.cn> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Message-Id: <20230914022645.1151356-40-gaosong@loongson.cn>
This commit is contained in:
parent
40c7674e9e
commit
c50ce38a47
|
@ -2120,6 +2120,22 @@ INSN_LASX(xvsrani_h_w, vv_i)
|
|||
INSN_LASX(xvsrani_w_d, vv_i)
|
||||
INSN_LASX(xvsrani_d_q, vv_i)
|
||||
|
||||
INSN_LASX(xvsrlrn_b_h, vvv)
|
||||
INSN_LASX(xvsrlrn_h_w, vvv)
|
||||
INSN_LASX(xvsrlrn_w_d, vvv)
|
||||
INSN_LASX(xvsrarn_b_h, vvv)
|
||||
INSN_LASX(xvsrarn_h_w, vvv)
|
||||
INSN_LASX(xvsrarn_w_d, vvv)
|
||||
|
||||
INSN_LASX(xvsrlrni_b_h, vv_i)
|
||||
INSN_LASX(xvsrlrni_h_w, vv_i)
|
||||
INSN_LASX(xvsrlrni_w_d, vv_i)
|
||||
INSN_LASX(xvsrlrni_d_q, vv_i)
|
||||
INSN_LASX(xvsrarni_b_h, vv_i)
|
||||
INSN_LASX(xvsrarni_h_w, vv_i)
|
||||
INSN_LASX(xvsrarni_w_d, vv_i)
|
||||
INSN_LASX(xvsrarni_d_q, vv_i)
|
||||
|
||||
INSN_LASX(xvreplgr2vr_b, vr)
|
||||
INSN_LASX(xvreplgr2vr_h, vr)
|
||||
INSN_LASX(xvreplgr2vr_w, vr)
|
||||
|
|
|
@ -3801,6 +3801,12 @@ TRANS(vsrlrn_w_d, LSX, gen_vvv, gen_helper_vsrlrn_w_d)
|
|||
TRANS(vsrarn_b_h, LSX, gen_vvv, gen_helper_vsrarn_b_h)
|
||||
TRANS(vsrarn_h_w, LSX, gen_vvv, gen_helper_vsrarn_h_w)
|
||||
TRANS(vsrarn_w_d, LSX, gen_vvv, gen_helper_vsrarn_w_d)
|
||||
TRANS(xvsrlrn_b_h, LASX, gen_xxx, gen_helper_vsrlrn_b_h)
|
||||
TRANS(xvsrlrn_h_w, LASX, gen_xxx, gen_helper_vsrlrn_h_w)
|
||||
TRANS(xvsrlrn_w_d, LASX, gen_xxx, gen_helper_vsrlrn_w_d)
|
||||
TRANS(xvsrarn_b_h, LASX, gen_xxx, gen_helper_vsrarn_b_h)
|
||||
TRANS(xvsrarn_h_w, LASX, gen_xxx, gen_helper_vsrarn_h_w)
|
||||
TRANS(xvsrarn_w_d, LASX, gen_xxx, gen_helper_vsrarn_w_d)
|
||||
|
||||
TRANS(vsrlrni_b_h, LSX, gen_vv_i, gen_helper_vsrlrni_b_h)
|
||||
TRANS(vsrlrni_h_w, LSX, gen_vv_i, gen_helper_vsrlrni_h_w)
|
||||
|
@ -3810,6 +3816,14 @@ TRANS(vsrarni_b_h, LSX, gen_vv_i, gen_helper_vsrarni_b_h)
|
|||
TRANS(vsrarni_h_w, LSX, gen_vv_i, gen_helper_vsrarni_h_w)
|
||||
TRANS(vsrarni_w_d, LSX, gen_vv_i, gen_helper_vsrarni_w_d)
|
||||
TRANS(vsrarni_d_q, LSX, gen_vv_i, gen_helper_vsrarni_d_q)
|
||||
TRANS(xvsrlrni_b_h, LASX, gen_xx_i, gen_helper_vsrlrni_b_h)
|
||||
TRANS(xvsrlrni_h_w, LASX, gen_xx_i, gen_helper_vsrlrni_h_w)
|
||||
TRANS(xvsrlrni_w_d, LASX, gen_xx_i, gen_helper_vsrlrni_w_d)
|
||||
TRANS(xvsrlrni_d_q, LASX, gen_xx_i, gen_helper_vsrlrni_d_q)
|
||||
TRANS(xvsrarni_b_h, LASX, gen_xx_i, gen_helper_vsrarni_b_h)
|
||||
TRANS(xvsrarni_h_w, LASX, gen_xx_i, gen_helper_vsrarni_h_w)
|
||||
TRANS(xvsrarni_w_d, LASX, gen_xx_i, gen_helper_vsrarni_w_d)
|
||||
TRANS(xvsrarni_d_q, LASX, gen_xx_i, gen_helper_vsrarni_d_q)
|
||||
|
||||
TRANS(vssrln_b_h, LSX, gen_vvv, gen_helper_vssrln_b_h)
|
||||
TRANS(vssrln_h_w, LSX, gen_vvv, gen_helper_vssrln_h_w)
|
||||
|
|
|
@ -1694,6 +1694,22 @@ xvsrani_h_w 0111 01110101 10001 ..... ..... ..... @vv_ui5
|
|||
xvsrani_w_d 0111 01110101 1001 ...... ..... ..... @vv_ui6
|
||||
xvsrani_d_q 0111 01110101 101 ....... ..... ..... @vv_ui7
|
||||
|
||||
xvsrlrn_b_h 0111 01001111 10001 ..... ..... ..... @vvv
|
||||
xvsrlrn_h_w 0111 01001111 10010 ..... ..... ..... @vvv
|
||||
xvsrlrn_w_d 0111 01001111 10011 ..... ..... ..... @vvv
|
||||
xvsrarn_b_h 0111 01001111 10101 ..... ..... ..... @vvv
|
||||
xvsrarn_h_w 0111 01001111 10110 ..... ..... ..... @vvv
|
||||
xvsrarn_w_d 0111 01001111 10111 ..... ..... ..... @vvv
|
||||
|
||||
xvsrlrni_b_h 0111 01110100 01000 1 .... ..... ..... @vv_ui4
|
||||
xvsrlrni_h_w 0111 01110100 01001 ..... ..... ..... @vv_ui5
|
||||
xvsrlrni_w_d 0111 01110100 0101 ...... ..... ..... @vv_ui6
|
||||
xvsrlrni_d_q 0111 01110100 011 ....... ..... ..... @vv_ui7
|
||||
xvsrarni_b_h 0111 01110101 11000 1 .... ..... ..... @vv_ui4
|
||||
xvsrarni_h_w 0111 01110101 11001 ..... ..... ..... @vv_ui5
|
||||
xvsrarni_w_d 0111 01110101 1101 ...... ..... ..... @vv_ui6
|
||||
xvsrarni_d_q 0111 01110101 111 ....... ..... ..... @vv_ui7
|
||||
|
||||
xvreplgr2vr_b 0111 01101001 11110 00000 ..... ..... @vr
|
||||
xvreplgr2vr_h 0111 01101001 11110 00001 ..... ..... @vr
|
||||
xvreplgr2vr_w 0111 01101001 11110 00010 ..... ..... @vr
|
||||
|
|
|
@ -1231,76 +1231,95 @@ VSRANI(vsrani_b_h, 16, B, H)
|
|||
VSRANI(vsrani_h_w, 32, H, W)
|
||||
VSRANI(vsrani_w_d, 64, W, D)
|
||||
|
||||
#define VSRLRN(NAME, BIT, T, E1, E2) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t desc) \
|
||||
{ \
|
||||
int i; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
\
|
||||
for (i = 0; i < LSX_LEN/BIT; i++) { \
|
||||
Vd->E1(i) = do_vsrlr_ ## E2(Vj->E2(i), ((T)Vk->E2(i))%BIT); \
|
||||
} \
|
||||
Vd->D(1) = 0; \
|
||||
#define VSRLRN(NAME, BIT, E1, E2, E3) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t desc) \
|
||||
{ \
|
||||
int i, j, ofs; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
int oprsz = simd_oprsz(desc); \
|
||||
\
|
||||
ofs = LSX_LEN / BIT; \
|
||||
for (i = 0; i < oprsz / 16; i++) { \
|
||||
for (j = 0; j < ofs; j++) { \
|
||||
Vd->E1(j + ofs * 2 * i) = do_vsrlr_ ##E2(Vj->E2(j + ofs * i), \
|
||||
Vk->E3(j + ofs * i) % BIT); \
|
||||
} \
|
||||
Vd->D(2 * i + 1) = 0; \
|
||||
} \
|
||||
}
|
||||
|
||||
VSRLRN(vsrlrn_b_h, 16, uint16_t, B, H)
|
||||
VSRLRN(vsrlrn_h_w, 32, uint32_t, H, W)
|
||||
VSRLRN(vsrlrn_w_d, 64, uint64_t, W, D)
|
||||
VSRLRN(vsrlrn_b_h, 16, B, H, UH)
|
||||
VSRLRN(vsrlrn_h_w, 32, H, W, UW)
|
||||
VSRLRN(vsrlrn_w_d, 64, W, D, UD)
|
||||
|
||||
#define VSRARN(NAME, BIT, T, E1, E2) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t desc) \
|
||||
{ \
|
||||
int i; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
\
|
||||
for (i = 0; i < LSX_LEN/BIT; i++) { \
|
||||
Vd->E1(i) = do_vsrar_ ## E2(Vj->E2(i), ((T)Vk->E2(i))%BIT); \
|
||||
} \
|
||||
Vd->D(1) = 0; \
|
||||
#define VSRARN(NAME, BIT, E1, E2, E3) \
|
||||
void HELPER(NAME)(void *vd, void *vj, void *vk, uint32_t desc) \
|
||||
{ \
|
||||
int i, j, ofs; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
VReg *Vk = (VReg *)vk; \
|
||||
int oprsz = simd_oprsz(desc); \
|
||||
\
|
||||
ofs = LSX_LEN / BIT; \
|
||||
for (i = 0; i < oprsz / 16; i++) { \
|
||||
for (j = 0; j < ofs; j++) { \
|
||||
Vd->E1(j + ofs * 2 * i) = do_vsrar_ ## E2(Vj->E2(j + ofs * i), \
|
||||
Vk->E3(j + ofs * i) % BIT); \
|
||||
} \
|
||||
Vd->D(2 * i + 1) = 0; \
|
||||
} \
|
||||
}
|
||||
|
||||
VSRARN(vsrarn_b_h, 16, uint8_t, B, H)
|
||||
VSRARN(vsrarn_h_w, 32, uint16_t, H, W)
|
||||
VSRARN(vsrarn_w_d, 64, uint32_t, W, D)
|
||||
VSRARN(vsrarn_b_h, 16, B, H, UH)
|
||||
VSRARN(vsrarn_h_w, 32, H, W, UW)
|
||||
VSRARN(vsrarn_w_d, 64, W, D, UD)
|
||||
|
||||
#define VSRLRNI(NAME, BIT, E1, E2) \
|
||||
void HELPER(NAME)(void *vd, void *vj, uint64_t imm, uint32_t desc) \
|
||||
{ \
|
||||
int i, max; \
|
||||
VReg temp; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
\
|
||||
temp.D(0) = 0; \
|
||||
temp.D(1) = 0; \
|
||||
max = LSX_LEN/BIT; \
|
||||
for (i = 0; i < max; i++) { \
|
||||
temp.E1(i) = do_vsrlr_ ## E2(Vj->E2(i), imm); \
|
||||
temp.E1(i + max) = do_vsrlr_ ## E2(Vd->E2(i), imm); \
|
||||
} \
|
||||
*Vd = temp; \
|
||||
#define VSRLRNI(NAME, BIT, E1, E2) \
|
||||
void HELPER(NAME)(void *vd, void *vj, uint64_t imm, uint32_t desc) \
|
||||
{ \
|
||||
int i, j, ofs; \
|
||||
VReg temp = {}; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
int oprsz = simd_oprsz(desc); \
|
||||
\
|
||||
ofs = LSX_LEN / BIT; \
|
||||
for (i = 0; i < oprsz / 16; i++) { \
|
||||
for (j = 0; j < ofs; j++) { \
|
||||
temp.E1(j + ofs * 2 * i) = do_vsrlr_ ## E2(Vj->E2(j + ofs * i), imm); \
|
||||
temp.E1(j + ofs * (2 * i + 1)) = do_vsrlr_ ## E2(Vd->E2(j + ofs * i), \
|
||||
imm); \
|
||||
} \
|
||||
} \
|
||||
*Vd = temp; \
|
||||
}
|
||||
|
||||
void HELPER(vsrlrni_d_q)(void *vd, void *vj, uint64_t imm, uint32_t desc)
|
||||
{
|
||||
VReg temp;
|
||||
int i;
|
||||
VReg temp = {};
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
Int128 r1, r2;
|
||||
Int128 r[4];
|
||||
int oprsz = simd_oprsz(desc);
|
||||
|
||||
if (imm == 0) {
|
||||
temp.D(0) = int128_getlo(Vj->Q(0));
|
||||
temp.D(1) = int128_getlo(Vd->Q(0));
|
||||
} else {
|
||||
r1 = int128_and(int128_urshift(Vj->Q(0), (imm -1)), int128_one());
|
||||
r2 = int128_and(int128_urshift(Vd->Q(0), (imm -1)), int128_one());
|
||||
|
||||
temp.D(0) = int128_getlo(int128_add(int128_urshift(Vj->Q(0), imm), r1));
|
||||
temp.D(1) = int128_getlo(int128_add(int128_urshift(Vd->Q(0), imm), r2));
|
||||
for (i = 0; i < oprsz / 16; i++) {
|
||||
if (imm == 0) {
|
||||
temp.D(2 * i) = int128_getlo(Vj->Q(i));
|
||||
temp.D(2 * i + 1) = int128_getlo(Vd->Q(i));
|
||||
} else {
|
||||
r[2 * i] = int128_and(int128_urshift(Vj->Q(i), (imm - 1)),
|
||||
int128_one());
|
||||
r[2 * i + 1] = int128_and(int128_urshift(Vd->Q(i), (imm - 1)),
|
||||
int128_one());
|
||||
temp.D(2 * i) = int128_getlo(int128_add(int128_urshift(Vj->Q(i),
|
||||
imm), r[2 * i]));
|
||||
temp.D(2 * i + 1) = int128_getlo(int128_add(int128_urshift(Vd->Q(i),
|
||||
imm), r[ 2 * i + 1]));
|
||||
}
|
||||
}
|
||||
*Vd = temp;
|
||||
}
|
||||
|
@ -1309,40 +1328,49 @@ VSRLRNI(vsrlrni_b_h, 16, B, H)
|
|||
VSRLRNI(vsrlrni_h_w, 32, H, W)
|
||||
VSRLRNI(vsrlrni_w_d, 64, W, D)
|
||||
|
||||
#define VSRARNI(NAME, BIT, E1, E2) \
|
||||
void HELPER(NAME)(void *vd, void *vj, uint64_t imm, uint32_t desc) \
|
||||
{ \
|
||||
int i, max; \
|
||||
VReg temp; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
\
|
||||
temp.D(0) = 0; \
|
||||
temp.D(1) = 0; \
|
||||
max = LSX_LEN/BIT; \
|
||||
for (i = 0; i < max; i++) { \
|
||||
temp.E1(i) = do_vsrar_ ## E2(Vj->E2(i), imm); \
|
||||
temp.E1(i + max) = do_vsrar_ ## E2(Vd->E2(i), imm); \
|
||||
} \
|
||||
*Vd = temp; \
|
||||
#define VSRARNI(NAME, BIT, E1, E2) \
|
||||
void HELPER(NAME)(void *vd, void *vj, uint64_t imm, uint32_t desc) \
|
||||
{ \
|
||||
int i, j, ofs; \
|
||||
VReg temp = {}; \
|
||||
VReg *Vd = (VReg *)vd; \
|
||||
VReg *Vj = (VReg *)vj; \
|
||||
int oprsz = simd_oprsz(desc); \
|
||||
\
|
||||
ofs = LSX_LEN / BIT; \
|
||||
for (i = 0; i < oprsz / 16; i++) { \
|
||||
for (j = 0; j < ofs; j++) { \
|
||||
temp.E1(j + ofs * 2 * i) = do_vsrar_ ## E2(Vj->E2(j + ofs * i), imm); \
|
||||
temp.E1(j + ofs * (2 * i + 1)) = do_vsrar_ ## E2(Vd->E2(j + ofs * i), \
|
||||
imm); \
|
||||
} \
|
||||
} \
|
||||
*Vd = temp; \
|
||||
}
|
||||
|
||||
void HELPER(vsrarni_d_q)(void *vd, void *vj, uint64_t imm, uint32_t desc)
|
||||
{
|
||||
VReg temp;
|
||||
int i;
|
||||
VReg temp = {};
|
||||
VReg *Vd = (VReg *)vd;
|
||||
VReg *Vj = (VReg *)vj;
|
||||
Int128 r1, r2;
|
||||
Int128 r[4];
|
||||
int oprsz = simd_oprsz(desc);
|
||||
|
||||
if (imm == 0) {
|
||||
temp.D(0) = int128_getlo(Vj->Q(0));
|
||||
temp.D(1) = int128_getlo(Vd->Q(0));
|
||||
} else {
|
||||
r1 = int128_and(int128_rshift(Vj->Q(0), (imm -1)), int128_one());
|
||||
r2 = int128_and(int128_rshift(Vd->Q(0), (imm -1)), int128_one());
|
||||
|
||||
temp.D(0) = int128_getlo(int128_add(int128_rshift(Vj->Q(0), imm), r1));
|
||||
temp.D(1) = int128_getlo(int128_add(int128_rshift(Vd->Q(0), imm), r2));
|
||||
for (i = 0; i < oprsz / 16; i++) {
|
||||
if (imm == 0) {
|
||||
temp.D(2 * i) = int128_getlo(Vj->Q(i));
|
||||
temp.D(2 * i + 1) = int128_getlo(Vd->Q(i));
|
||||
} else {
|
||||
r[2 * i] = int128_and(int128_rshift(Vj->Q(i), (imm - 1)),
|
||||
int128_one());
|
||||
r[2 * i + 1] = int128_and(int128_rshift(Vd->Q(i), (imm - 1)),
|
||||
int128_one());
|
||||
temp.D(2 * i) = int128_getlo(int128_add(int128_rshift(Vj->Q(i),
|
||||
imm), r[2 * i]));
|
||||
temp.D(2 * i + 1) = int128_getlo(int128_add(int128_rshift(Vd->Q(i),
|
||||
imm), r[2 * i + 1]));
|
||||
}
|
||||
}
|
||||
*Vd = temp;
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue