Currently SLP vectorizer walks through the instructions and selects 3 main classes of values: 1) reduction operations - instructions with same reduction opcode (add, mul, min/max, etc.), which build the reduction, 2) reduced values - instructions with the same opcodes, but different from the reduction opcode, 3) extra arguments - all other values, instructions from the different basic block rather than the root node, instructions with to many/less uses. This scheme is not very efficient. It excludes some instructions and all non-instruction values from the reductions (constants, proficient gathers), to many possibly reduced values are marked as extra arguments. Patch improves this process by introducing a bit extended analysis stage. During this stage, we still try to select 3 classes of the values: 1) reduction operations - same as before, 2) possibly reduced values - all instructions from the current block/non-instructions, which may build a vectorization tree, 3) extra arguments - instructions from the different basic blocks. Additionally, an extra sorting of the possibly reduced values occurs to build the scalar sequences which highly likely will bed vectorized, e.g. loads are grouped by the distance between them, constants are grouped together, cmp instructions are sorted by their compare types and predicates, extractelement instructions are sorted by the vector operand, etc. Also, these groups are reordered by their length so the longest group is the first in the list of the possibly reduced values. The vectorization process tries to emit the reductions for all these groups. These reductions, remaining non-vectorized possible reduced values and extra arguments are then combined into the final expression just like it was before. Differential Revision: https://reviews.llvm.org/D114171
341 lines
13 KiB
LLVM
341 lines
13 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
|
; RUN: opt -O2 -S -mattr=avx < %s | FileCheck %s
|
|
; RUN: opt -passes="default<O2>" -S -mattr=avx < %s | FileCheck %s
|
|
|
|
target triple = "x86_64--"
|
|
target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128"
|
|
|
|
define i32 @ext_ext_or_reduction_v4i32(<4 x i32> %x, <4 x i32> %y) {
|
|
; CHECK-LABEL: @ext_ext_or_reduction_v4i32(
|
|
; CHECK-NEXT: [[Z:%.*]] = and <4 x i32> [[Y:%.*]], [[X:%.*]]
|
|
; CHECK-NEXT: [[TMP1:%.*]] = call i32 @llvm.vector.reduce.or.v4i32(<4 x i32> [[Z]])
|
|
; CHECK-NEXT: ret i32 [[TMP1]]
|
|
;
|
|
%z = and <4 x i32> %x, %y
|
|
%z0 = extractelement <4 x i32> %z, i32 0
|
|
%z1 = extractelement <4 x i32> %z, i32 1
|
|
%z01 = or i32 %z0, %z1
|
|
%z2 = extractelement <4 x i32> %z, i32 2
|
|
%z012 = or i32 %z01, %z2
|
|
%z3 = extractelement <4 x i32> %z, i32 3
|
|
%z0123 = or i32 %z3, %z012
|
|
ret i32 %z0123
|
|
}
|
|
|
|
define i32 @ext_ext_partial_add_reduction_v4i32(<4 x i32> %x) {
|
|
; CHECK-LABEL: @ext_ext_partial_add_reduction_v4i32(
|
|
; CHECK-NEXT: [[SHIFT:%.*]] = shufflevector <4 x i32> [[X:%.*]], <4 x i32> poison, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef>
|
|
; CHECK-NEXT: [[TMP1:%.*]] = add <4 x i32> [[SHIFT]], [[X]]
|
|
; CHECK-NEXT: [[SHIFT1:%.*]] = shufflevector <4 x i32> [[X]], <4 x i32> poison, <4 x i32> <i32 2, i32 undef, i32 undef, i32 undef>
|
|
; CHECK-NEXT: [[TMP2:%.*]] = add <4 x i32> [[TMP1]], [[SHIFT1]]
|
|
; CHECK-NEXT: [[X210:%.*]] = extractelement <4 x i32> [[TMP2]], i64 0
|
|
; CHECK-NEXT: ret i32 [[X210]]
|
|
;
|
|
%x0 = extractelement <4 x i32> %x, i32 0
|
|
%x1 = extractelement <4 x i32> %x, i32 1
|
|
%x10 = add i32 %x1, %x0
|
|
%x2 = extractelement <4 x i32> %x, i32 2
|
|
%x210 = add i32 %x2, %x10
|
|
ret i32 %x210
|
|
}
|
|
|
|
define i32 @ext_ext_partial_add_reduction_and_extra_add_v4i32(<4 x i32> %x, <4 x i32> %y) {
|
|
; CHECK-LABEL: @ext_ext_partial_add_reduction_and_extra_add_v4i32(
|
|
; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <4 x i32> [[Y:%.*]], <4 x i32> [[X:%.*]], <4 x i32> <i32 0, i32 1, i32 2, i32 6>
|
|
; CHECK-NEXT: [[TMP2:%.*]] = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> [[TMP1]])
|
|
; CHECK-NEXT: ret i32 [[TMP2]]
|
|
;
|
|
%y0 = extractelement <4 x i32> %y, i32 0
|
|
%y1 = extractelement <4 x i32> %y, i32 1
|
|
%y10 = add i32 %y1, %y0
|
|
%y2 = extractelement <4 x i32> %y, i32 2
|
|
%y210 = add i32 %y2, %y10
|
|
%x2 = extractelement <4 x i32> %x, i32 2
|
|
%x2y210 = add i32 %x2, %y210
|
|
ret i32 %x2y210
|
|
}
|
|
|
|
; PR43953 - https://bugs.llvm.org/show_bug.cgi?id=43953
|
|
; We want to end up with a single reduction on the next 4 tests.
|
|
|
|
define i32 @TestVectorsEqual(i32* noalias %Vec0, i32* noalias %Vec1, i32 %Tolerance) {
|
|
; CHECK-LABEL: @TestVectorsEqual(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast i32* [[VEC0:%.*]] to <4 x i32>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, <4 x i32>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = bitcast i32* [[VEC1:%.*]] to <4 x i32>*
|
|
; CHECK-NEXT: [[TMP3:%.*]] = load <4 x i32>, <4 x i32>* [[TMP2]], align 4
|
|
; CHECK-NEXT: [[TMP4:%.*]] = sub nsw <4 x i32> [[TMP1]], [[TMP3]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = call <4 x i32> @llvm.abs.v4i32(<4 x i32> [[TMP4]], i1 true)
|
|
; CHECK-NEXT: [[TMP6:%.*]] = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> [[TMP5]])
|
|
; CHECK-NEXT: [[CMP5_NOT:%.*]] = icmp sle i32 [[TMP6]], [[TOLERANCE:%.*]]
|
|
; CHECK-NEXT: [[COND6:%.*]] = zext i1 [[CMP5_NOT]] to i32
|
|
; CHECK-NEXT: ret i32 [[COND6]]
|
|
;
|
|
entry:
|
|
br label %for.cond
|
|
|
|
for.cond:
|
|
%sum.0 = phi i32 [ 0, %entry ], [ %add, %for.inc ]
|
|
%Component.0 = phi i32 [ 0, %entry ], [ %inc, %for.inc ]
|
|
%cmp = icmp slt i32 %Component.0, 4
|
|
br i1 %cmp, label %for.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
br label %for.end
|
|
|
|
for.body:
|
|
%idxprom = sext i32 %Component.0 to i64
|
|
%arrayidx = getelementptr inbounds i32, i32* %Vec0, i64 %idxprom
|
|
%0 = load i32, i32* %arrayidx, align 4
|
|
%idxprom1 = sext i32 %Component.0 to i64
|
|
%arrayidx2 = getelementptr inbounds i32, i32* %Vec1, i64 %idxprom1
|
|
%1 = load i32, i32* %arrayidx2, align 4
|
|
%sub = sub nsw i32 %0, %1
|
|
%cmp3 = icmp sge i32 %sub, 0
|
|
br i1 %cmp3, label %cond.true, label %cond.false
|
|
|
|
cond.true:
|
|
br label %cond.end
|
|
|
|
cond.false:
|
|
%sub4 = sub nsw i32 0, %sub
|
|
br label %cond.end
|
|
|
|
cond.end:
|
|
%cond = phi i32 [ %sub, %cond.true ], [ %sub4, %cond.false ]
|
|
%add = add nsw i32 %sum.0, %cond
|
|
br label %for.inc
|
|
|
|
for.inc:
|
|
%inc = add nsw i32 %Component.0, 1
|
|
br label %for.cond
|
|
|
|
for.end:
|
|
%cmp5 = icmp sle i32 %sum.0, %Tolerance
|
|
%2 = zext i1 %cmp5 to i64
|
|
%cond6 = select i1 %cmp5, i32 1, i32 0
|
|
ret i32 %cond6
|
|
}
|
|
|
|
define i32 @TestVectorsEqual_alt(i32* noalias %Vec0, i32* noalias %Vec1, i32 %Tolerance) {
|
|
; CHECK-LABEL: @TestVectorsEqual_alt(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast i32* [[VEC0:%.*]] to <4 x i32>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, <4 x i32>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = bitcast i32* [[VEC1:%.*]] to <4 x i32>*
|
|
; CHECK-NEXT: [[TMP3:%.*]] = load <4 x i32>, <4 x i32>* [[TMP2]], align 4
|
|
; CHECK-NEXT: [[TMP4:%.*]] = sub <4 x i32> [[TMP1]], [[TMP3]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = call i32 @llvm.vector.reduce.add.v4i32(<4 x i32> [[TMP4]])
|
|
; CHECK-NEXT: [[CMP3_NOT:%.*]] = icmp ule i32 [[TMP5]], [[TOLERANCE:%.*]]
|
|
; CHECK-NEXT: [[COND:%.*]] = zext i1 [[CMP3_NOT]] to i32
|
|
; CHECK-NEXT: ret i32 [[COND]]
|
|
;
|
|
entry:
|
|
br label %for.cond
|
|
|
|
for.cond:
|
|
%sum.0 = phi i32 [ 0, %entry ], [ %add, %for.inc ]
|
|
%Component.0 = phi i32 [ 0, %entry ], [ %inc, %for.inc ]
|
|
%cmp = icmp slt i32 %Component.0, 4
|
|
br i1 %cmp, label %for.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
br label %for.end
|
|
|
|
for.body:
|
|
%idxprom = sext i32 %Component.0 to i64
|
|
%arrayidx = getelementptr inbounds i32, i32* %Vec0, i64 %idxprom
|
|
%0 = load i32, i32* %arrayidx, align 4
|
|
%idxprom1 = sext i32 %Component.0 to i64
|
|
%arrayidx2 = getelementptr inbounds i32, i32* %Vec1, i64 %idxprom1
|
|
%1 = load i32, i32* %arrayidx2, align 4
|
|
%sub = sub i32 %0, %1
|
|
%add = add i32 %sum.0, %sub
|
|
br label %for.inc
|
|
|
|
for.inc:
|
|
%inc = add nsw i32 %Component.0, 1
|
|
br label %for.cond
|
|
|
|
for.end:
|
|
%cmp3 = icmp ule i32 %sum.0, %Tolerance
|
|
%2 = zext i1 %cmp3 to i64
|
|
%cond = select i1 %cmp3, i32 1, i32 0
|
|
ret i32 %cond
|
|
}
|
|
|
|
define i32 @TestVectorsEqualFP(float* noalias %Vec0, float* noalias %Vec1, float %Tolerance) {
|
|
; CHECK-LABEL: @TestVectorsEqualFP(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast float* [[VEC0:%.*]] to <4 x float>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x float>, <4 x float>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = bitcast float* [[VEC1:%.*]] to <4 x float>*
|
|
; CHECK-NEXT: [[TMP3:%.*]] = load <4 x float>, <4 x float>* [[TMP2]], align 4
|
|
; CHECK-NEXT: [[TMP4:%.*]] = fsub fast <4 x float> [[TMP1]], [[TMP3]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = call fast <4 x float> @llvm.fabs.v4f32(<4 x float> [[TMP4]])
|
|
; CHECK-NEXT: [[TMP6:%.*]] = call fast float @llvm.vector.reduce.fadd.v4f32(float -0.000000e+00, <4 x float> [[TMP5]])
|
|
; CHECK-NEXT: [[CMP4:%.*]] = fcmp fast ole float [[TMP6]], [[TOLERANCE:%.*]]
|
|
; CHECK-NEXT: [[COND5:%.*]] = zext i1 [[CMP4]] to i32
|
|
; CHECK-NEXT: ret i32 [[COND5]]
|
|
;
|
|
entry:
|
|
br label %for.cond
|
|
|
|
for.cond:
|
|
%sum.0 = phi float [ 0.000000e+00, %entry ], [ %add, %for.inc ]
|
|
%Component.0 = phi i32 [ 0, %entry ], [ %inc, %for.inc ]
|
|
%cmp = icmp slt i32 %Component.0, 4
|
|
br i1 %cmp, label %for.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
br label %for.end
|
|
|
|
for.body:
|
|
%idxprom = sext i32 %Component.0 to i64
|
|
%arrayidx = getelementptr inbounds float, float* %Vec0, i64 %idxprom
|
|
%0 = load float, float* %arrayidx, align 4
|
|
%idxprom1 = sext i32 %Component.0 to i64
|
|
%arrayidx2 = getelementptr inbounds float, float* %Vec1, i64 %idxprom1
|
|
%1 = load float, float* %arrayidx2, align 4
|
|
%sub = fsub fast float %0, %1
|
|
%cmp3 = fcmp fast oge float %sub, 0.000000e+00
|
|
br i1 %cmp3, label %cond.true, label %cond.false
|
|
|
|
cond.true:
|
|
br label %cond.end
|
|
|
|
cond.false:
|
|
%fneg = fneg fast float %sub
|
|
br label %cond.end
|
|
|
|
cond.end:
|
|
%cond = phi fast float [ %sub, %cond.true ], [ %fneg, %cond.false ]
|
|
%add = fadd fast float %sum.0, %cond
|
|
br label %for.inc
|
|
|
|
for.inc:
|
|
%inc = add nsw i32 %Component.0, 1
|
|
br label %for.cond
|
|
|
|
for.end:
|
|
%cmp4 = fcmp fast ole float %sum.0, %Tolerance
|
|
%2 = zext i1 %cmp4 to i64
|
|
%cond5 = select i1 %cmp4, i32 1, i32 0
|
|
ret i32 %cond5
|
|
}
|
|
|
|
define i32 @TestVectorsEqualFP_alt(float* noalias %Vec0, float* noalias %Vec1, float %Tolerance) {
|
|
; CHECK-LABEL: @TestVectorsEqualFP_alt(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast float* [[VEC0:%.*]] to <4 x float>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x float>, <4 x float>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = bitcast float* [[VEC1:%.*]] to <4 x float>*
|
|
; CHECK-NEXT: [[TMP3:%.*]] = load <4 x float>, <4 x float>* [[TMP2]], align 4
|
|
; CHECK-NEXT: [[TMP4:%.*]] = fsub fast <4 x float> [[TMP1]], [[TMP3]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = call fast float @llvm.vector.reduce.fadd.v4f32(float -0.000000e+00, <4 x float> [[TMP4]])
|
|
; CHECK-NEXT: [[CMP3:%.*]] = fcmp fast ole float [[TMP5]], [[TOLERANCE:%.*]]
|
|
; CHECK-NEXT: [[COND:%.*]] = zext i1 [[CMP3]] to i32
|
|
; CHECK-NEXT: ret i32 [[COND]]
|
|
;
|
|
entry:
|
|
br label %for.cond
|
|
|
|
for.cond:
|
|
%sum.0 = phi float [ 0.000000e+00, %entry ], [ %add, %for.inc ]
|
|
%Component.0 = phi i32 [ 0, %entry ], [ %inc, %for.inc ]
|
|
%cmp = icmp slt i32 %Component.0, 4
|
|
br i1 %cmp, label %for.body, label %for.cond.cleanup
|
|
|
|
for.cond.cleanup:
|
|
br label %for.end
|
|
|
|
for.body:
|
|
%idxprom = sext i32 %Component.0 to i64
|
|
%arrayidx = getelementptr inbounds float, float* %Vec0, i64 %idxprom
|
|
%0 = load float, float* %arrayidx, align 4
|
|
%idxprom1 = sext i32 %Component.0 to i64
|
|
%arrayidx2 = getelementptr inbounds float, float* %Vec1, i64 %idxprom1
|
|
%1 = load float, float* %arrayidx2, align 4
|
|
%sub = fsub fast float %0, %1
|
|
%add = fadd fast float %sum.0, %sub
|
|
br label %for.inc
|
|
|
|
for.inc:
|
|
%inc = add nsw i32 %Component.0, 1
|
|
br label %for.cond
|
|
|
|
for.end:
|
|
%cmp3 = fcmp fast ole float %sum.0, %Tolerance
|
|
%2 = zext i1 %cmp3 to i64
|
|
%cond = select i1 %cmp3, i32 1, i32 0
|
|
ret i32 %cond
|
|
}
|
|
|
|
; PR43745 - https://bugs.llvm.org/show_bug.cgi?id=43745
|
|
|
|
; FIXME: this should be vectorized
|
|
define i1 @cmp_lt_gt(double %a, double %b, double %c) {
|
|
; CHECK-LABEL: @cmp_lt_gt(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: [[FNEG:%.*]] = fneg double [[B:%.*]]
|
|
; CHECK-NEXT: [[MUL:%.*]] = fmul double [[A:%.*]], 2.000000e+00
|
|
; CHECK-NEXT: [[TMP0:%.*]] = insertelement <2 x double> poison, double [[C:%.*]], i64 1
|
|
; CHECK-NEXT: [[TMP1:%.*]] = insertelement <2 x double> [[TMP0]], double [[FNEG]], i64 0
|
|
; CHECK-NEXT: [[TMP2:%.*]] = insertelement <2 x double> poison, double [[C]], i64 0
|
|
; CHECK-NEXT: [[TMP3:%.*]] = insertelement <2 x double> [[TMP2]], double [[B]], i64 1
|
|
; CHECK-NEXT: [[TMP4:%.*]] = fsub <2 x double> [[TMP1]], [[TMP3]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = insertelement <2 x double> poison, double [[MUL]], i64 0
|
|
; CHECK-NEXT: [[TMP6:%.*]] = shufflevector <2 x double> [[TMP5]], <2 x double> poison, <2 x i32> zeroinitializer
|
|
; CHECK-NEXT: [[TMP7:%.*]] = fdiv <2 x double> [[TMP4]], [[TMP6]]
|
|
; CHECK-NEXT: [[TMP8:%.*]] = extractelement <2 x double> [[TMP7]], i64 1
|
|
; CHECK-NEXT: [[CMP:%.*]] = fcmp olt double [[TMP8]], 0x3EB0C6F7A0B5ED8D
|
|
; CHECK-NEXT: [[TMP9:%.*]] = extractelement <2 x double> [[TMP7]], i64 0
|
|
; CHECK-NEXT: [[CMP4:%.*]] = fcmp olt double [[TMP9]], 0x3EB0C6F7A0B5ED8D
|
|
; CHECK-NEXT: [[OR_COND:%.*]] = select i1 [[CMP]], i1 [[CMP4]], i1 false
|
|
; CHECK-NEXT: br i1 [[OR_COND]], label [[CLEANUP:%.*]], label [[LOR_LHS_FALSE:%.*]]
|
|
; CHECK: lor.lhs.false:
|
|
; CHECK-NEXT: [[TMP10:%.*]] = fcmp ule <2 x double> [[TMP7]], <double 1.000000e+00, double 1.000000e+00>
|
|
; CHECK-NEXT: [[TMP11:%.*]] = extractelement <2 x i1> [[TMP10]], i64 0
|
|
; CHECK-NEXT: [[TMP12:%.*]] = extractelement <2 x i1> [[TMP10]], i64 1
|
|
; CHECK-NEXT: [[OR_COND1:%.*]] = select i1 [[TMP12]], i1 true, i1 [[TMP11]]
|
|
; CHECK-NEXT: br label [[CLEANUP]]
|
|
; CHECK: cleanup:
|
|
; CHECK-NEXT: [[RETVAL_0:%.*]] = phi i1 [ false, [[ENTRY:%.*]] ], [ [[OR_COND1]], [[LOR_LHS_FALSE]] ]
|
|
; CHECK-NEXT: ret i1 [[RETVAL_0]]
|
|
;
|
|
entry:
|
|
%fneg = fneg double %b
|
|
%add = fadd double %fneg, %c
|
|
%mul = fmul double 2.0, %a
|
|
%div = fdiv double %add, %mul
|
|
%fneg1 = fneg double %b
|
|
%sub = fsub double %fneg1, %c
|
|
%mul2 = fmul double 2.0, %a
|
|
%div3 = fdiv double %sub, %mul2
|
|
%cmp = fcmp olt double %div, 0x3EB0C6F7A0B5ED8D
|
|
br i1 %cmp, label %land.lhs.true, label %lor.lhs.false
|
|
|
|
land.lhs.true:
|
|
%cmp4 = fcmp olt double %div3, 0x3EB0C6F7A0B5ED8D
|
|
br i1 %cmp4, label %if.then, label %lor.lhs.false
|
|
|
|
lor.lhs.false:
|
|
%cmp5 = fcmp ogt double %div, 1.0
|
|
br i1 %cmp5, label %land.lhs.true6, label %if.end
|
|
|
|
land.lhs.true6:
|
|
%cmp7 = fcmp ogt double %div3, 1.0
|
|
br i1 %cmp7, label %if.then, label %if.end
|
|
|
|
if.then:
|
|
br label %cleanup
|
|
|
|
if.end:
|
|
br label %cleanup
|
|
|
|
cleanup:
|
|
%retval.0 = phi i1 [ false, %if.then ], [ true, %if.end ]
|
|
ret i1 %retval.0
|
|
}
|