A handy shorthand for specifying the shufflevector(insertelement(poison,
foo, 0), poison, zeroinitializer) splat pattern was introduced in
#74620.
Some of the RISC-V tests were converted over to use this new form in
dbb65dd330, this patch handles the rest
which didn't have any codegen diffs.
This not only converts some constant expressions to the new form, but
also instruction sequences that weren't previously constant expressions
to constant expressions as well. In some cases this affects codegen, but
these have been omitted here and will be handled in a separate PR.
1835 lines
76 KiB
LLVM
1835 lines
76 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=riscv32 -mattr=+m,+d,+zfh,+zvfh,+v -target-abi=ilp32d \
|
|
; RUN: -verify-machineinstrs < %s | FileCheck %s --check-prefixes=CHECK,RV32
|
|
; RUN: llc -mtriple=riscv64 -mattr=+m,+d,+zfh,+zvfh,+v -target-abi=lp64d \
|
|
; RUN: -verify-machineinstrs < %s | FileCheck %s --check-prefixes=CHECK,RV64
|
|
|
|
declare void @llvm.masked.scatter.nxv1i8.nxv1p0(<vscale x 1 x i8>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1i8(<vscale x 1 x i8> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e8, mf8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e8, mf8, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1i8.nxv1p0(<vscale x 1 x i8> %val, <vscale x 1 x ptr> %ptrs, i32 1, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2i8.nxv2p0(<vscale x 2 x i8>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2i8(<vscale x 2 x i8> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2i8.nxv2p0(<vscale x 2 x i8> %val, <vscale x 2 x ptr> %ptrs, i32 1, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_nxv2i16_truncstore_nxv2i8(<vscale x 2 x i16> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i16_truncstore_nxv2i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i16_truncstore_nxv2i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
%tval = trunc <vscale x 2 x i16> %val to <vscale x 2 x i8>
|
|
call void @llvm.masked.scatter.nxv2i8.nxv2p0(<vscale x 2 x i8> %tval, <vscale x 2 x ptr> %ptrs, i32 1, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_nxv2i32_truncstore_nxv2i8(<vscale x 2 x i32> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i32_truncstore_nxv2i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV32-NEXT: vsetvli zero, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i32_truncstore_nxv2i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV64-NEXT: vsetvli zero, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
%tval = trunc <vscale x 2 x i32> %val to <vscale x 2 x i8>
|
|
call void @llvm.masked.scatter.nxv2i8.nxv2p0(<vscale x 2 x i8> %tval, <vscale x 2 x ptr> %ptrs, i32 1, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_nxv2i64_truncstore_nxv2i8(<vscale x 2 x i64> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i64_truncstore_nxv2i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v11, v8, 0
|
|
; RV32-NEXT: vsetvli zero, zero, e16, mf2, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v11, 0
|
|
; RV32-NEXT: vsetvli zero, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i64_truncstore_nxv2i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v12, v8, 0
|
|
; RV64-NEXT: vsetvli zero, zero, e16, mf2, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v12, 0
|
|
; RV64-NEXT: vsetvli zero, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
%tval = trunc <vscale x 2 x i64> %val to <vscale x 2 x i8>
|
|
call void @llvm.masked.scatter.nxv2i8.nxv2p0(<vscale x 2 x i8> %tval, <vscale x 2 x ptr> %ptrs, i32 1, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4i8.nxv4p0(<vscale x 4 x i8>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4i8(<vscale x 4 x i8> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e8, mf2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e8, mf2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i8.nxv4p0(<vscale x 4 x i8> %val, <vscale x 4 x ptr> %ptrs, i32 1, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4i8(<vscale x 4 x i8> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e8, mf2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e8, mf2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i8.nxv4p0(<vscale x 4 x i8> %val, <vscale x 4 x ptr> %ptrs, i32 1, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4i8(<vscale x 4 x i8> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4i8:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i8.nxv4p0(<vscale x 4 x i8> %val, <vscale x 4 x ptr> %ptrs, i32 1, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8i8.nxv8p0(<vscale x 8 x i8>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8i8(<vscale x 8 x i8> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e8, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e8, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8i8.nxv8p0(<vscale x 8 x i8> %val, <vscale x 8 x ptr> %ptrs, i32 1, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8(<vscale x 8 x i8> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v12, v9
|
|
; RV32-NEXT: vsetvli zero, zero, e8, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v9
|
|
; RV64-NEXT: vsetvli zero, zero, e8, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i8, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8i8.nxv8p0(<vscale x 8 x i8> %val, <vscale x 8 x ptr> %ptrs, i32 1, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv1i16.nxv1p0(<vscale x 1 x i16>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1i16(<vscale x 1 x i16> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, mf4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, mf4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1i16.nxv1p0(<vscale x 1 x i16> %val, <vscale x 1 x ptr> %ptrs, i32 2, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2i16.nxv2p0(<vscale x 2 x i16>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2i16(<vscale x 2 x i16> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2i16.nxv2p0(<vscale x 2 x i16> %val, <vscale x 2 x ptr> %ptrs, i32 2, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_nxv2i32_truncstore_nxv2i16(<vscale x 2 x i32> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i32_truncstore_nxv2i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i32_truncstore_nxv2i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v8, 0
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
%tval = trunc <vscale x 2 x i32> %val to <vscale x 2 x i16>
|
|
call void @llvm.masked.scatter.nxv2i16.nxv2p0(<vscale x 2 x i16> %tval, <vscale x 2 x ptr> %ptrs, i32 2, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_nxv2i64_truncstore_nxv2i16(<vscale x 2 x i64> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i64_truncstore_nxv2i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v11, v8, 0
|
|
; RV32-NEXT: vsetvli zero, zero, e16, mf2, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v8, v11, 0
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i64_truncstore_nxv2i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v12, v8, 0
|
|
; RV64-NEXT: vsetvli zero, zero, e16, mf2, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v8, v12, 0
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
%tval = trunc <vscale x 2 x i64> %val to <vscale x 2 x i16>
|
|
call void @llvm.masked.scatter.nxv2i16.nxv2p0(<vscale x 2 x i16> %tval, <vscale x 2 x ptr> %ptrs, i32 2, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4i16.nxv4p0(<vscale x 4 x i16>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4i16(<vscale x 4 x i16> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i16.nxv4p0(<vscale x 4 x i16> %val, <vscale x 4 x ptr> %ptrs, i32 2, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4i16(<vscale x 4 x i16> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i16.nxv4p0(<vscale x 4 x i16> %val, <vscale x 4 x ptr> %ptrs, i32 2, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4i16(<vscale x 4 x i16> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4i16:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i16.nxv4p0(<vscale x 4 x i16> %val, <vscale x 4 x ptr> %ptrs, i32 2, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8i16.nxv8p0(<vscale x 8 x i16>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8i16(<vscale x 8 x i16> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8i16.nxv8p0(<vscale x 8 x i16> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8_nxv8i16(<vscale x 8 x i16> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8_nxv8i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v12, v10
|
|
; RV32-NEXT: vadd.vv v12, v12, v12
|
|
; RV32-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8_nxv8i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v10
|
|
; RV64-NEXT: vadd.vv v16, v16, v16
|
|
; RV64-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i16, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8i16.nxv8p0(<vscale x 8 x i16> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i8_nxv8i16(<vscale x 8 x i16> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v12, v10
|
|
; RV32-NEXT: vadd.vv v12, v12, v12
|
|
; RV32-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v10
|
|
; RV64-NEXT: vadd.vv v16, v16, v16
|
|
; RV64-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i8> %idxs to <vscale x 8 x i16>
|
|
%ptrs = getelementptr inbounds i16, ptr %base, <vscale x 8 x i16> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i16.nxv8p0(<vscale x 8 x i16> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i8_nxv8i16(<vscale x 8 x i16> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i8_nxv8i16:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e8, m1, ta, ma
|
|
; CHECK-NEXT: vwaddu.vv v12, v10, v10
|
|
; CHECK-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; CHECK-NEXT: vsoxei16.v v8, (a0), v12, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i8> %idxs to <vscale x 8 x i16>
|
|
%ptrs = getelementptr inbounds i16, ptr %base, <vscale x 8 x i16> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i16.nxv8p0(<vscale x 8 x i16> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i16(<vscale x 8 x i16> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vwadd.vv v12, v10, v10
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v16, v10
|
|
; RV64-NEXT: vadd.vv v16, v16, v16
|
|
; RV64-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i16, ptr %base, <vscale x 8 x i16> %idxs
|
|
call void @llvm.masked.scatter.nxv8i16.nxv8p0(<vscale x 8 x i16> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv1i32.nxv1p0(<vscale x 1 x i32>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1i32(<vscale x 1 x i32> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, mf2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, mf2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1i32.nxv1p0(<vscale x 1 x i32> %val, <vscale x 1 x ptr> %ptrs, i32 4, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2i32.nxv2p0(<vscale x 2 x i32>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2i32(<vscale x 2 x i32> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2i32.nxv2p0(<vscale x 2 x i32> %val, <vscale x 2 x ptr> %ptrs, i32 4, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_nxv2i64_truncstore_nxv2i32(<vscale x 2 x i64> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i64_truncstore_nxv2i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v11, v8, 0
|
|
; RV32-NEXT: vsoxei32.v v11, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i64_truncstore_nxv2i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV64-NEXT: vnsrl.wi v12, v8, 0
|
|
; RV64-NEXT: vsoxei64.v v12, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
%tval = trunc <vscale x 2 x i64> %val to <vscale x 2 x i32>
|
|
call void @llvm.masked.scatter.nxv2i32.nxv2p0(<vscale x 2 x i32> %tval, <vscale x 2 x ptr> %ptrs, i32 4, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4i32.nxv4p0(<vscale x 4 x i32>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4i32(<vscale x 4 x i32> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i32.nxv4p0(<vscale x 4 x i32> %val, <vscale x 4 x ptr> %ptrs, i32 4, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4i32(<vscale x 4 x i32> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i32.nxv4p0(<vscale x 4 x i32> %val, <vscale x 4 x ptr> %ptrs, i32 4, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4i32(<vscale x 4 x i32> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4i32:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i32.nxv4p0(<vscale x 4 x i32> %val, <vscale x 4 x ptr> %ptrs, i32 4, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8i32(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8_nxv8i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8_nxv8i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i8_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i8> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i8_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i8_nxv8i32:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e16, m2, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v14, v12
|
|
; CHECK-NEXT: vsll.vi v12, v14, 2
|
|
; CHECK-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; CHECK-NEXT: vsoxei16.v v8, (a0), v12, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i8> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i16_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i16_nxv8i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i16_nxv8i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i16> %idxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i16_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i16> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i16_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i16_nxv8i32:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v16, v12
|
|
; CHECK-NEXT: vsll.vi v12, v16, 2
|
|
; CHECK-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i16> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i32(<vscale x 8 x i32> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v12, v12, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf2 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i32, ptr %base, <vscale x 8 x i32> %idxs
|
|
call void @llvm.masked.scatter.nxv8i32.nxv8p0(<vscale x 8 x i32> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv1i64.nxv1p0(<vscale x 1 x i64>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1i64(<vscale x 1 x i64> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1i64.nxv1p0(<vscale x 1 x i64> %val, <vscale x 1 x ptr> %ptrs, i32 8, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2i64.nxv2p0(<vscale x 2 x i64>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2i64(<vscale x 2 x i64> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2i64.nxv2p0(<vscale x 2 x i64> %val, <vscale x 2 x ptr> %ptrs, i32 8, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4i64.nxv4p0(<vscale x 4 x i64>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4i64(<vscale x 4 x i64> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i64.nxv4p0(<vscale x 4 x i64> %val, <vscale x 4 x ptr> %ptrs, i32 8, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4i64(<vscale x 4 x i64> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i64.nxv4p0(<vscale x 4 x i64> %val, <vscale x 4 x ptr> %ptrs, i32 8, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4i64(<vscale x 4 x i64> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4i64:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4i64.nxv4p0(<vscale x 4 x i64> %val, <vscale x 4 x ptr> %ptrs, i32 8, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8i64(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i8_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i8> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i8_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i8_nxv8i64:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e16, m2, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v18, v16
|
|
; CHECK-NEXT: vsll.vi v16, v18, 3
|
|
; CHECK-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; CHECK-NEXT: vsoxei16.v v8, (a0), v16, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i8> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i16_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i16_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i16_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i16> %idxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i16_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i16> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i16_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i16_nxv8i64:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v20, v16
|
|
; CHECK-NEXT: vsll.vi v16, v20, 3
|
|
; CHECK-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; CHECK-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i16> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i32_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i32_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v16, v16, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i32_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf2 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i32> %idxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i32_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i32_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v16, v16, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i32_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf2 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i32> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i32_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_zext_nxv8i32_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v16, v16, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_zext_nxv8i32_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vzext.vf2 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i32> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i64(<vscale x 8 x i64> %val, ptr %base, <vscale x 8 x i64> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v24, v16, 0
|
|
; RV32-NEXT: vsll.vi v16, v24, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsll.vi v16, v16, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds i64, ptr %base, <vscale x 8 x i64> %idxs
|
|
call void @llvm.masked.scatter.nxv8i64.nxv8p0(<vscale x 8 x i64> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv1f16.nxv1p0(<vscale x 1 x half>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1f16(<vscale x 1 x half> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, mf4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, mf4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1f16.nxv1p0(<vscale x 1 x half> %val, <vscale x 1 x ptr> %ptrs, i32 2, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2f16.nxv2p0(<vscale x 2 x half>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2f16(<vscale x 2 x half> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, mf2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2f16.nxv2p0(<vscale x 2 x half> %val, <vscale x 2 x ptr> %ptrs, i32 2, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4f16.nxv4p0(<vscale x 4 x half>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4f16(<vscale x 4 x half> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f16.nxv4p0(<vscale x 4 x half> %val, <vscale x 4 x ptr> %ptrs, i32 2, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4f16(<vscale x 4 x half> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f16.nxv4p0(<vscale x 4 x half> %val, <vscale x 4 x ptr> %ptrs, i32 2, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4f16(<vscale x 4 x half> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4f16:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f16.nxv4p0(<vscale x 4 x half> %val, <vscale x 4 x ptr> %ptrs, i32 2, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8f16.nxv8p0(<vscale x 8 x half>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8f16(<vscale x 8 x half> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8f16.nxv8p0(<vscale x 8 x half> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8_nxv8f16(<vscale x 8 x half> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8_nxv8f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v12, v10
|
|
; RV32-NEXT: vadd.vv v12, v12, v12
|
|
; RV32-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8_nxv8f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v10
|
|
; RV64-NEXT: vadd.vv v16, v16, v16
|
|
; RV64-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds half, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8f16.nxv8p0(<vscale x 8 x half> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i8_nxv8f16(<vscale x 8 x half> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v12, v10
|
|
; RV32-NEXT: vadd.vv v12, v12, v12
|
|
; RV32-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v10
|
|
; RV64-NEXT: vadd.vv v16, v16, v16
|
|
; RV64-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i8> %idxs to <vscale x 8 x i16>
|
|
%ptrs = getelementptr inbounds half, ptr %base, <vscale x 8 x i16> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f16.nxv8p0(<vscale x 8 x half> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i8_nxv8f16(<vscale x 8 x half> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i8_nxv8f16:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e8, m1, ta, ma
|
|
; CHECK-NEXT: vwaddu.vv v12, v10, v10
|
|
; CHECK-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; CHECK-NEXT: vsoxei16.v v8, (a0), v12, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i8> %idxs to <vscale x 8 x i16>
|
|
%ptrs = getelementptr inbounds half, ptr %base, <vscale x 8 x i16> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f16.nxv8p0(<vscale x 8 x half> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8f16(<vscale x 8 x half> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8f16:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e16, m2, ta, ma
|
|
; RV32-NEXT: vwadd.vv v12, v10, v10
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8f16:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v16, v10
|
|
; RV64-NEXT: vadd.vv v16, v16, v16
|
|
; RV64-NEXT: vsetvli zero, zero, e16, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds half, ptr %base, <vscale x 8 x i16> %idxs
|
|
call void @llvm.masked.scatter.nxv8f16.nxv8p0(<vscale x 8 x half> %val, <vscale x 8 x ptr> %ptrs, i32 2, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv1f32.nxv1p0(<vscale x 1 x float>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1f32(<vscale x 1 x float> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, mf2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, mf2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1f32.nxv1p0(<vscale x 1 x float> %val, <vscale x 1 x ptr> %ptrs, i32 4, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2f32.nxv2p0(<vscale x 2 x float>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2f32(<vscale x 2 x float> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2f32.nxv2p0(<vscale x 2 x float> %val, <vscale x 2 x ptr> %ptrs, i32 4, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4f32.nxv4p0(<vscale x 4 x float>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4f32(<vscale x 4 x float> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f32.nxv4p0(<vscale x 4 x float> %val, <vscale x 4 x ptr> %ptrs, i32 4, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4f32(<vscale x 4 x float> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f32.nxv4p0(<vscale x 4 x float> %val, <vscale x 4 x ptr> %ptrs, i32 4, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4f32(<vscale x 4 x float> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4f32:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f32.nxv4p0(<vscale x 4 x float> %val, <vscale x 4 x ptr> %ptrs, i32 4, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8f32(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8_nxv8f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8_nxv8f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i8_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i8> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i8_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i8_nxv8f32:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e16, m2, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v14, v12
|
|
; CHECK-NEXT: vsll.vi v12, v14, 2
|
|
; CHECK-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; CHECK-NEXT: vsoxei16.v v8, (a0), v12, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i8> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i16_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i16_nxv8f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i16_nxv8f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i16> %idxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i16_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v16, v12
|
|
; RV32-NEXT: vsll.vi v12, v16, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i16> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i16_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i16_nxv8f32:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v16, v12
|
|
; CHECK-NEXT: vsll.vi v12, v16, 2
|
|
; CHECK-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i16> %idxs to <vscale x 8 x i32>
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i32> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8f32(<vscale x 8 x float> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8f32:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v12, v12, 2
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8f32:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf2 v16, v12
|
|
; RV64-NEXT: vsll.vi v16, v16, 2
|
|
; RV64-NEXT: vsetvli zero, zero, e32, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds float, ptr %base, <vscale x 8 x i32> %idxs
|
|
call void @llvm.masked.scatter.nxv8f32.nxv8p0(<vscale x 8 x float> %val, <vscale x 8 x ptr> %ptrs, i32 4, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv1f64.nxv1p0(<vscale x 1 x double>, <vscale x 1 x ptr>, i32, <vscale x 1 x i1>)
|
|
|
|
define void @mscatter_nxv1f64(<vscale x 1 x double> %val, <vscale x 1 x ptr> %ptrs, <vscale x 1 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv1f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m1, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v9, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv1f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m1, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v9, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv1f64.nxv1p0(<vscale x 1 x double> %val, <vscale x 1 x ptr> %ptrs, i32 8, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv2f64.nxv2p0(<vscale x 2 x double>, <vscale x 2 x ptr>, i32, <vscale x 2 x i1>)
|
|
|
|
define void @mscatter_nxv2f64(<vscale x 2 x double> %val, <vscale x 2 x ptr> %ptrs, <vscale x 2 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv2f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m2, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v10, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv2f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m2, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v10, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv2f64.nxv2p0(<vscale x 2 x double> %val, <vscale x 2 x ptr> %ptrs, i32 8, <vscale x 2 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv4f64.nxv4p0(<vscale x 4 x double>, <vscale x 4 x ptr>, i32, <vscale x 4 x i1>)
|
|
|
|
define void @mscatter_nxv4f64(<vscale x 4 x double> %val, <vscale x 4 x ptr> %ptrs, <vscale x 4 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv4f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv4f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f64.nxv4p0(<vscale x 4 x double> %val, <vscale x 4 x ptr> %ptrs, i32 8, <vscale x 4 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_truemask_nxv4f64(<vscale x 4 x double> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; RV32-LABEL: mscatter_truemask_nxv4f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v12
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_truemask_nxv4f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m4, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v12
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f64.nxv4p0(<vscale x 4 x double> %val, <vscale x 4 x ptr> %ptrs, i32 8, <vscale x 4 x i1> splat (i1 1))
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_falsemask_nxv4f64(<vscale x 4 x double> %val, <vscale x 4 x ptr> %ptrs) {
|
|
; CHECK-LABEL: mscatter_falsemask_nxv4f64:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv4f64.nxv4p0(<vscale x 4 x double> %val, <vscale x 4 x ptr> %ptrs, i32 8, <vscale x 4 x i1> zeroinitializer)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double>, <vscale x 8 x ptr>, i32, <vscale x 8 x i1>)
|
|
|
|
define void @mscatter_nxv8f64(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i8_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i8_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i8_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i8> %idxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i8_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i8_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i8> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i8_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i8> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i8_nxv8f64:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e16, m2, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v18, v16
|
|
; CHECK-NEXT: vsll.vi v16, v18, 3
|
|
; CHECK-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; CHECK-NEXT: vsoxei16.v v8, (a0), v16, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i8> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i16_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i16_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i16_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i16> %idxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i16_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v20, v16
|
|
; RV32-NEXT: vsll.vi v16, v20, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i16_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i16> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i16_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i16> %idxs, <vscale x 8 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv8i16_nxv8f64:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; CHECK-NEXT: vzext.vf2 v20, v16
|
|
; CHECK-NEXT: vsll.vi v16, v20, 3
|
|
; CHECK-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; CHECK-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i16> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8i32_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8i32_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v16, v16, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8i32_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf2 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i32> %idxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_sext_nxv8i32_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_sext_nxv8i32_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v16, v16, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_sext_nxv8i32_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf2 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = sext <vscale x 8 x i32> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv8i32_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i32> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_zext_nxv8i32_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vsll.vi v16, v16, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_zext_nxv8i32_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vzext.vf2 v24, v16
|
|
; RV64-NEXT: vsll.vi v16, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%eidxs = zext <vscale x 8 x i32> %idxs to <vscale x 8 x i64>
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %eidxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv8f64(<vscale x 8 x double> %val, ptr %base, <vscale x 8 x i64> %idxs, <vscale x 8 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv8f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m4, ta, ma
|
|
; RV32-NEXT: vnsrl.wi v24, v16, 0
|
|
; RV32-NEXT: vsll.vi v16, v24, 3
|
|
; RV32-NEXT: vsetvli zero, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v16, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv8f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsll.vi v16, v16, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v16, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 8 x i64> %idxs
|
|
call void @llvm.masked.scatter.nxv8f64.nxv8p0(<vscale x 8 x double> %val, <vscale x 8 x ptr> %ptrs, i32 8, <vscale x 8 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
declare void @llvm.masked.scatter.nxv16f64.nxv16p0(<vscale x 16 x double>, <vscale x 16 x ptr>, i32, <vscale x 16 x i1>)
|
|
|
|
declare <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double>, <vscale x 8 x double>, i64)
|
|
declare <vscale x 16 x ptr> @llvm.vector.insert.nxv8p0.nxv16p0(<vscale x 16 x ptr>, <vscale x 8 x ptr>, i64)
|
|
|
|
define void @mscatter_nxv16f64(<vscale x 8 x double> %val0, <vscale x 8 x double> %val1, <vscale x 8 x ptr> %ptrs0, <vscale x 8 x ptr> %ptrs1, <vscale x 16 x i1> %m) {
|
|
; RV32-LABEL: mscatter_nxv16f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vl4re32.v v24, (a0)
|
|
; RV32-NEXT: vl4re32.v v28, (a1)
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (zero), v24, v0.t
|
|
; RV32-NEXT: csrr a0, vlenb
|
|
; RV32-NEXT: srli a0, a0, 3
|
|
; RV32-NEXT: vsetvli a1, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vslidedown.vx v0, v0, a0
|
|
; RV32-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v16, (zero), v28, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_nxv16f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: addi sp, sp, -16
|
|
; RV64-NEXT: .cfi_def_cfa_offset 16
|
|
; RV64-NEXT: csrr a2, vlenb
|
|
; RV64-NEXT: slli a2, a2, 3
|
|
; RV64-NEXT: sub sp, sp, a2
|
|
; RV64-NEXT: .cfi_escape 0x0f, 0x0d, 0x72, 0x00, 0x11, 0x10, 0x22, 0x11, 0x08, 0x92, 0xa2, 0x38, 0x00, 0x1e, 0x22 # sp + 16 + 8 * vlenb
|
|
; RV64-NEXT: vl8re64.v v24, (a0)
|
|
; RV64-NEXT: addi a0, sp, 16
|
|
; RV64-NEXT: vs8r.v v16, (a0) # Unknown-size Folded Spill
|
|
; RV64-NEXT: vl8re64.v v16, (a1)
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v24, v0.t
|
|
; RV64-NEXT: csrr a0, vlenb
|
|
; RV64-NEXT: srli a0, a0, 3
|
|
; RV64-NEXT: vsetvli a1, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vslidedown.vx v0, v0, a0
|
|
; RV64-NEXT: vsetvli a0, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: addi a0, sp, 16
|
|
; RV64-NEXT: vl8r.v v8, (a0) # Unknown-size Folded Reload
|
|
; RV64-NEXT: vsoxei64.v v8, (zero), v16, v0.t
|
|
; RV64-NEXT: csrr a0, vlenb
|
|
; RV64-NEXT: slli a0, a0, 3
|
|
; RV64-NEXT: add sp, sp, a0
|
|
; RV64-NEXT: addi sp, sp, 16
|
|
; RV64-NEXT: ret
|
|
%p0 = call <vscale x 16 x ptr> @llvm.vector.insert.nxv8p0.nxv16p0(<vscale x 16 x ptr> undef, <vscale x 8 x ptr> %ptrs0, i64 0)
|
|
%p1 = call <vscale x 16 x ptr> @llvm.vector.insert.nxv8p0.nxv16p0(<vscale x 16 x ptr> %p0, <vscale x 8 x ptr> %ptrs1, i64 8)
|
|
%v0 = call <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double> undef, <vscale x 8 x double> %val0, i64 0)
|
|
%v1 = call <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double> %v0, <vscale x 8 x double> %val1, i64 8)
|
|
call void @llvm.masked.scatter.nxv16f64.nxv16p0(<vscale x 16 x double> %v1, <vscale x 16 x ptr> %p1, i32 8, <vscale x 16 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv16i8_nxv16f64(<vscale x 8 x double> %val0, <vscale x 8 x double> %val1, ptr %base, <vscale x 16 x i8> %idxs, <vscale x 16 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv16i8_nxv16f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vl2r.v v6, (a1)
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m8, ta, ma
|
|
; RV32-NEXT: vsext.vf4 v24, v6
|
|
; RV32-NEXT: vsll.vi v24, v24, 3
|
|
; RV32-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v24, v0.t
|
|
; RV32-NEXT: csrr a1, vlenb
|
|
; RV32-NEXT: srli a1, a1, 3
|
|
; RV32-NEXT: vsetvli a2, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vslidedown.vx v0, v0, a1
|
|
; RV32-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v16, (a0), v28, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv16i8_nxv16f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vl2r.v v6, (a1)
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf8 v24, v6
|
|
; RV64-NEXT: vsll.vi v24, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v24, v0.t
|
|
; RV64-NEXT: vsext.vf8 v8, v7
|
|
; RV64-NEXT: vsll.vi v8, v8, 3
|
|
; RV64-NEXT: csrr a1, vlenb
|
|
; RV64-NEXT: srli a1, a1, 3
|
|
; RV64-NEXT: vsetvli a2, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vslidedown.vx v0, v0, a1
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v16, (a0), v8, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 16 x i8> %idxs
|
|
%v0 = call <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double> undef, <vscale x 8 x double> %val0, i64 0)
|
|
%v1 = call <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double> %v0, <vscale x 8 x double> %val1, i64 8)
|
|
call void @llvm.masked.scatter.nxv16f64.nxv16p0(<vscale x 16 x double> %v1, <vscale x 16 x ptr> %ptrs, i32 8, <vscale x 16 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_nxv16i16_nxv16f64(<vscale x 8 x double> %val0, <vscale x 8 x double> %val1, ptr %base, <vscale x 16 x i16> %idxs, <vscale x 16 x i1> %m) {
|
|
; RV32-LABEL: mscatter_baseidx_nxv16i16_nxv16f64:
|
|
; RV32: # %bb.0:
|
|
; RV32-NEXT: vl4re16.v v4, (a1)
|
|
; RV32-NEXT: vsetvli a1, zero, e32, m8, ta, ma
|
|
; RV32-NEXT: vsext.vf2 v24, v4
|
|
; RV32-NEXT: vsll.vi v24, v24, 3
|
|
; RV32-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v8, (a0), v24, v0.t
|
|
; RV32-NEXT: csrr a1, vlenb
|
|
; RV32-NEXT: srli a1, a1, 3
|
|
; RV32-NEXT: vsetvli a2, zero, e8, mf4, ta, ma
|
|
; RV32-NEXT: vslidedown.vx v0, v0, a1
|
|
; RV32-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV32-NEXT: vsoxei32.v v16, (a0), v28, v0.t
|
|
; RV32-NEXT: ret
|
|
;
|
|
; RV64-LABEL: mscatter_baseidx_nxv16i16_nxv16f64:
|
|
; RV64: # %bb.0:
|
|
; RV64-NEXT: vl4re16.v v4, (a1)
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsext.vf4 v24, v4
|
|
; RV64-NEXT: vsll.vi v24, v24, 3
|
|
; RV64-NEXT: vsoxei64.v v8, (a0), v24, v0.t
|
|
; RV64-NEXT: vsext.vf4 v8, v6
|
|
; RV64-NEXT: vsll.vi v8, v8, 3
|
|
; RV64-NEXT: csrr a1, vlenb
|
|
; RV64-NEXT: srli a1, a1, 3
|
|
; RV64-NEXT: vsetvli a2, zero, e8, mf4, ta, ma
|
|
; RV64-NEXT: vslidedown.vx v0, v0, a1
|
|
; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, ma
|
|
; RV64-NEXT: vsoxei64.v v16, (a0), v8, v0.t
|
|
; RV64-NEXT: ret
|
|
%ptrs = getelementptr inbounds double, ptr %base, <vscale x 16 x i16> %idxs
|
|
%v0 = call <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double> undef, <vscale x 8 x double> %val0, i64 0)
|
|
%v1 = call <vscale x 16 x double> @llvm.vector.insert.nxv8f64.nxv16f64(<vscale x 16 x double> %v0, <vscale x 8 x double> %val1, i64 8)
|
|
call void @llvm.masked.scatter.nxv16f64.nxv16p0(<vscale x 16 x double> %v1, <vscale x 16 x ptr> %ptrs, i32 8, <vscale x 16 x i1> %m)
|
|
ret void
|
|
}
|
|
|
|
define void @mscatter_baseidx_zext_nxv1i1_nxv1i8(<vscale x 1 x i8> %val, ptr %base, <vscale x 1 x i1> %idxs, <vscale x 1 x i1> %m) {
|
|
; CHECK-LABEL: mscatter_baseidx_zext_nxv1i1_nxv1i8:
|
|
; CHECK: # %bb.0:
|
|
; CHECK-NEXT: vsetvli a1, zero, e8, mf8, ta, ma
|
|
; CHECK-NEXT: vmv.v.i v10, 0
|
|
; CHECK-NEXT: vmerge.vim v10, v10, 1, v0
|
|
; CHECK-NEXT: vmv1r.v v0, v9
|
|
; CHECK-NEXT: vsoxei8.v v8, (a0), v10, v0.t
|
|
; CHECK-NEXT: ret
|
|
%eidxs = zext <vscale x 1 x i1> %idxs to <vscale x 1 x i8>
|
|
%ptrs = getelementptr inbounds i8, ptr %base, <vscale x 1 x i8> %eidxs
|
|
call void @llvm.masked.scatter.nxv1i8.nxv1p0(<vscale x 1 x i8> %val, <vscale x 1 x ptr> %ptrs, i32 1, <vscale x 1 x i1> %m)
|
|
ret void
|
|
}
|