Currently SLP vectorizer walks through the instructions and selects 3 main classes of values: 1) reduction operations - instructions with same reduction opcode (add, mul, min/max, etc.), which build the reduction, 2) reduced values - instructions with the same opcodes, but different from the reduction opcode, 3) extra arguments - all other values, instructions from the different basic block rather than the root node, instructions with to many/less uses. This scheme is not very efficient. It excludes some instructions and all non-instruction values from the reductions (constants, proficient gathers), to many possibly reduced values are marked as extra arguments. Patch improves this process by introducing a bit extended analysis stage. During this stage, we still try to select 3 classes of the values: 1) reduction operations - same as before, 2) possibly reduced values - all instructions from the current block/non-instructions, which may build a vectorization tree, 3) extra arguments - instructions from the different basic blocks. Additionally, an extra sorting of the possibly reduced values occurs to build the scalar sequences which highly likely will bed vectorized, e.g. loads are grouped by the distance between them, constants are grouped together, cmp instructions are sorted by their compare types and predicates, extractelement instructions are sorted by the vector operand, etc. Also, these groups are reordered by their length so the longest group is the first in the list of the possibly reduced values. The vectorization process tries to emit the reductions for all these groups. These reductions, remaining non-vectorized possible reduced values and extra arguments are then combined into the final expression just like it was before. Differential Revision: https://reviews.llvm.org/D114171
261 lines
9.8 KiB
LLVM
261 lines
9.8 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
|
|
; RUN: opt < %s -slp-vectorizer -S -mtriple=x86_64-apple-macosx10.10.0 -mattr=+sse4.2 | FileCheck %s
|
|
|
|
; PR28474
|
|
|
|
;void foo();
|
|
;
|
|
;int test1(unsigned int *p) {
|
|
; int sum = 0;
|
|
; #pragma nounroll
|
|
; for (int y = 0; y < 2; y++) {
|
|
; // Inner loop gets unrolled
|
|
; for (int x = 0; x < 8; x++) {
|
|
; sum += p[x] * 42;
|
|
; }
|
|
; // Dummy call to keep outer loop alive
|
|
; foo();
|
|
; }
|
|
; return sum;
|
|
;}
|
|
|
|
define i32 @test(i32* nocapture readonly %p) {
|
|
; CHECK-LABEL: @test(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: br label [[FOR_BODY:%.*]]
|
|
; CHECK: for.body:
|
|
; CHECK-NEXT: [[SUM:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[OP_RDX:%.*]], [[FOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast i32* [[P:%.*]] to <8 x i32>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i32>, <8 x i32>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = mul <8 x i32> [[TMP1]], <i32 42, i32 42, i32 42, i32 42, i32 42, i32 42, i32 42, i32 42>
|
|
; CHECK-NEXT: [[TMP3:%.*]] = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> [[TMP2]])
|
|
; CHECK-NEXT: [[OP_RDX]] = add i32 [[TMP3]], [[SUM]]
|
|
; CHECK-NEXT: br i1 true, label [[FOR_END:%.*]], label [[FOR_BODY]]
|
|
; CHECK: for.end:
|
|
; CHECK-NEXT: ret i32 [[OP_RDX]]
|
|
;
|
|
entry:
|
|
%arrayidx.1 = getelementptr inbounds i32, i32* %p, i64 1
|
|
%arrayidx.2 = getelementptr inbounds i32, i32* %p, i64 2
|
|
%arrayidx.3 = getelementptr inbounds i32, i32* %p, i64 3
|
|
%arrayidx.4 = getelementptr inbounds i32, i32* %p, i64 4
|
|
%arrayidx.5 = getelementptr inbounds i32, i32* %p, i64 5
|
|
%arrayidx.6 = getelementptr inbounds i32, i32* %p, i64 6
|
|
%arrayidx.7 = getelementptr inbounds i32, i32* %p, i64 7
|
|
br label %for.body
|
|
|
|
for.body:
|
|
%sum = phi i32 [ 0, %entry ], [ %add.7, %for.body ]
|
|
%tmp = load i32, i32* %p, align 4
|
|
%mul = mul i32 %tmp, 42
|
|
%add = add i32 %mul, %sum
|
|
%tmp5 = load i32, i32* %arrayidx.1, align 4
|
|
%mul.1 = mul i32 %tmp5, 42
|
|
%add.1 = add i32 %mul.1, %add
|
|
%tmp6 = load i32, i32* %arrayidx.2, align 4
|
|
%mul.2 = mul i32 %tmp6, 42
|
|
%add.2 = add i32 %mul.2, %add.1
|
|
%tmp7 = load i32, i32* %arrayidx.3, align 4
|
|
%mul.3 = mul i32 %tmp7, 42
|
|
%add.3 = add i32 %mul.3, %add.2
|
|
%tmp8 = load i32, i32* %arrayidx.4, align 4
|
|
%mul.4 = mul i32 %tmp8, 42
|
|
%add.4 = add i32 %mul.4, %add.3
|
|
%tmp9 = load i32, i32* %arrayidx.5, align 4
|
|
%mul.5 = mul i32 %tmp9, 42
|
|
%add.5 = add i32 %mul.5, %add.4
|
|
%tmp10 = load i32, i32* %arrayidx.6, align 4
|
|
%mul.6 = mul i32 %tmp10, 42
|
|
%add.6 = add i32 %mul.6, %add.5
|
|
%tmp11 = load i32, i32* %arrayidx.7, align 4
|
|
%mul.7 = mul i32 %tmp11, 42
|
|
%add.7 = add i32 %mul.7, %add.6
|
|
br i1 true, label %for.end, label %for.body
|
|
|
|
for.end:
|
|
ret i32 %add.7
|
|
}
|
|
|
|
;void foo();
|
|
;
|
|
;int test2(unsigned int *p, unsigned int *q) {
|
|
; int sum = 0;
|
|
; #pragma nounroll
|
|
; for (int y = 0; y < 2; y++) {
|
|
; // Inner loop gets unrolled
|
|
; for (int x = 0; x < 8; x++) {
|
|
; sum += p[x] * q[x];
|
|
; }
|
|
; // Dummy call to keep outer loop alive
|
|
; foo();
|
|
; }
|
|
; return sum;
|
|
;}
|
|
|
|
define i32 @test2(i32* nocapture readonly %p, i32* nocapture readonly %q) {
|
|
; CHECK-LABEL: @test2(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: br label [[FOR_BODY:%.*]]
|
|
; CHECK: for.body:
|
|
; CHECK-NEXT: [[SUM:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[OP_RDX:%.*]], [[FOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast i32* [[P:%.*]] to <8 x i32>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i32>, <8 x i32>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = bitcast i32* [[Q:%.*]] to <8 x i32>*
|
|
; CHECK-NEXT: [[TMP3:%.*]] = load <8 x i32>, <8 x i32>* [[TMP2]], align 4
|
|
; CHECK-NEXT: [[TMP4:%.*]] = mul <8 x i32> [[TMP1]], [[TMP3]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> [[TMP4]])
|
|
; CHECK-NEXT: [[OP_RDX]] = add i32 [[TMP5]], [[SUM]]
|
|
; CHECK-NEXT: br i1 true, label [[FOR_END:%.*]], label [[FOR_BODY]]
|
|
; CHECK: for.end:
|
|
; CHECK-NEXT: ret i32 [[OP_RDX]]
|
|
;
|
|
entry:
|
|
%arrayidx.p.1 = getelementptr inbounds i32, i32* %p, i64 1
|
|
%arrayidx.p.2 = getelementptr inbounds i32, i32* %p, i64 2
|
|
%arrayidx.p.3 = getelementptr inbounds i32, i32* %p, i64 3
|
|
%arrayidx.p.4 = getelementptr inbounds i32, i32* %p, i64 4
|
|
%arrayidx.p.5 = getelementptr inbounds i32, i32* %p, i64 5
|
|
%arrayidx.p.6 = getelementptr inbounds i32, i32* %p, i64 6
|
|
%arrayidx.p.7 = getelementptr inbounds i32, i32* %p, i64 7
|
|
|
|
%arrayidx.q.1 = getelementptr inbounds i32, i32* %q, i64 1
|
|
%arrayidx.q.2 = getelementptr inbounds i32, i32* %q, i64 2
|
|
%arrayidx.q.3 = getelementptr inbounds i32, i32* %q, i64 3
|
|
%arrayidx.q.4 = getelementptr inbounds i32, i32* %q, i64 4
|
|
%arrayidx.q.5 = getelementptr inbounds i32, i32* %q, i64 5
|
|
%arrayidx.q.6 = getelementptr inbounds i32, i32* %q, i64 6
|
|
%arrayidx.q.7 = getelementptr inbounds i32, i32* %q, i64 7
|
|
br label %for.body
|
|
|
|
for.body:
|
|
%sum = phi i32 [ 0, %entry ], [ %add.7, %for.body ]
|
|
%tmpp = load i32, i32* %p, align 4
|
|
%tmpq = load i32, i32* %q, align 4
|
|
%mul = mul i32 %tmpp, %tmpq
|
|
%add = add i32 %mul, %sum
|
|
%tmp5p = load i32, i32* %arrayidx.p.1, align 4
|
|
%tmp5q = load i32, i32* %arrayidx.q.1, align 4
|
|
%mul.1 = mul i32 %tmp5p, %tmp5q
|
|
%add.1 = add i32 %mul.1, %add
|
|
%tmp6p = load i32, i32* %arrayidx.p.2, align 4
|
|
%tmp6q = load i32, i32* %arrayidx.q.2, align 4
|
|
%mul.2 = mul i32 %tmp6p, %tmp6q
|
|
%add.2 = add i32 %mul.2, %add.1
|
|
%tmp7p = load i32, i32* %arrayidx.p.3, align 4
|
|
%tmp7q = load i32, i32* %arrayidx.q.3, align 4
|
|
%mul.3 = mul i32 %tmp7p, %tmp7q
|
|
%add.3 = add i32 %mul.3, %add.2
|
|
%tmp8p = load i32, i32* %arrayidx.p.4, align 4
|
|
%tmp8q = load i32, i32* %arrayidx.q.4, align 4
|
|
%mul.4 = mul i32 %tmp8p, %tmp8q
|
|
%add.4 = add i32 %mul.4, %add.3
|
|
%tmp9p = load i32, i32* %arrayidx.p.5, align 4
|
|
%tmp9q = load i32, i32* %arrayidx.q.5, align 4
|
|
%mul.5 = mul i32 %tmp9p, %tmp9q
|
|
%add.5 = add i32 %mul.5, %add.4
|
|
%tmp10p = load i32, i32* %arrayidx.p.6, align 4
|
|
%tmp10q = load i32, i32* %arrayidx.q.6, align 4
|
|
%mul.6 = mul i32 %tmp10p, %tmp10q
|
|
%add.6 = add i32 %mul.6, %add.5
|
|
%tmp11p = load i32, i32* %arrayidx.p.7, align 4
|
|
%tmp11q = load i32, i32* %arrayidx.q.7, align 4
|
|
%mul.7 = mul i32 %tmp11p, %tmp11q
|
|
%add.7 = add i32 %mul.7, %add.6
|
|
br i1 true, label %for.end, label %for.body
|
|
|
|
for.end:
|
|
ret i32 %add.7
|
|
}
|
|
|
|
;void foo();
|
|
;
|
|
;int test3(unsigned int *p, unsigned int *q) {
|
|
; int sum = 0;
|
|
; #pragma nounroll
|
|
; for (int y = 0; y < 2; y++) {
|
|
; // Inner loop gets unrolled
|
|
; for (int x = 0; x < 8; x++) {
|
|
; sum += p[x] * q[7-x];
|
|
; }
|
|
; // Dummy call to keep outer loop alive
|
|
; foo();
|
|
; }
|
|
; return sum;
|
|
;}
|
|
|
|
define i32 @test3(i32* nocapture readonly %p, i32* nocapture readonly %q) {
|
|
; CHECK-LABEL: @test3(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: br label [[FOR_BODY:%.*]]
|
|
; CHECK: for.body:
|
|
; CHECK-NEXT: [[SUM:%.*]] = phi i32 [ 0, [[ENTRY:%.*]] ], [ [[OP_RDX:%.*]], [[FOR_BODY]] ]
|
|
; CHECK-NEXT: [[TMP0:%.*]] = bitcast i32* [[P:%.*]] to <8 x i32>*
|
|
; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i32>, <8 x i32>* [[TMP0]], align 4
|
|
; CHECK-NEXT: [[TMP2:%.*]] = bitcast i32* [[Q:%.*]] to <8 x i32>*
|
|
; CHECK-NEXT: [[TMP3:%.*]] = load <8 x i32>, <8 x i32>* [[TMP2]], align 4
|
|
; CHECK-NEXT: [[SHUFFLE:%.*]] = shufflevector <8 x i32> [[TMP3]], <8 x i32> poison, <8 x i32> <i32 7, i32 6, i32 5, i32 4, i32 3, i32 2, i32 1, i32 0>
|
|
; CHECK-NEXT: [[TMP4:%.*]] = mul <8 x i32> [[TMP1]], [[SHUFFLE]]
|
|
; CHECK-NEXT: [[TMP5:%.*]] = call i32 @llvm.vector.reduce.add.v8i32(<8 x i32> [[TMP4]])
|
|
; CHECK-NEXT: [[OP_RDX]] = add i32 [[TMP5]], [[SUM]]
|
|
; CHECK-NEXT: br i1 true, label [[FOR_END:%.*]], label [[FOR_BODY]]
|
|
; CHECK: for.end:
|
|
; CHECK-NEXT: ret i32 [[OP_RDX]]
|
|
;
|
|
entry:
|
|
%arrayidx.p.1 = getelementptr inbounds i32, i32* %p, i64 1
|
|
%arrayidx.p.2 = getelementptr inbounds i32, i32* %p, i64 2
|
|
%arrayidx.p.3 = getelementptr inbounds i32, i32* %p, i64 3
|
|
%arrayidx.p.4 = getelementptr inbounds i32, i32* %p, i64 4
|
|
%arrayidx.p.5 = getelementptr inbounds i32, i32* %p, i64 5
|
|
%arrayidx.p.6 = getelementptr inbounds i32, i32* %p, i64 6
|
|
%arrayidx.p.7 = getelementptr inbounds i32, i32* %p, i64 7
|
|
|
|
%arrayidx.q.1 = getelementptr inbounds i32, i32* %q, i64 1
|
|
%arrayidx.q.2 = getelementptr inbounds i32, i32* %q, i64 2
|
|
%arrayidx.q.3 = getelementptr inbounds i32, i32* %q, i64 3
|
|
%arrayidx.q.4 = getelementptr inbounds i32, i32* %q, i64 4
|
|
%arrayidx.q.5 = getelementptr inbounds i32, i32* %q, i64 5
|
|
%arrayidx.q.6 = getelementptr inbounds i32, i32* %q, i64 6
|
|
%arrayidx.q.7 = getelementptr inbounds i32, i32* %q, i64 7
|
|
br label %for.body
|
|
|
|
for.body:
|
|
%sum = phi i32 [ 0, %entry ], [ %add.7, %for.body ]
|
|
%tmpp = load i32, i32* %p, align 4
|
|
%tmpq = load i32, i32* %arrayidx.q.7, align 4
|
|
%mul = mul i32 %tmpp, %tmpq
|
|
%add = add i32 %mul, %sum
|
|
%tmp5p = load i32, i32* %arrayidx.p.1, align 4
|
|
%tmp5q = load i32, i32* %arrayidx.q.6, align 4
|
|
%mul.1 = mul i32 %tmp5p, %tmp5q
|
|
%add.1 = add i32 %mul.1, %add
|
|
%tmp6p = load i32, i32* %arrayidx.p.2, align 4
|
|
%tmp6q = load i32, i32* %arrayidx.q.5, align 4
|
|
%mul.2 = mul i32 %tmp6p, %tmp6q
|
|
%add.2 = add i32 %mul.2, %add.1
|
|
%tmp7p = load i32, i32* %arrayidx.p.3, align 4
|
|
%tmp7q = load i32, i32* %arrayidx.q.4, align 4
|
|
%mul.3 = mul i32 %tmp7p, %tmp7q
|
|
%add.3 = add i32 %mul.3, %add.2
|
|
%tmp8p = load i32, i32* %arrayidx.p.4, align 4
|
|
%tmp8q = load i32, i32* %arrayidx.q.3, align 4
|
|
%mul.4 = mul i32 %tmp8p, %tmp8q
|
|
%add.4 = add i32 %mul.4, %add.3
|
|
%tmp9p = load i32, i32* %arrayidx.p.5, align 4
|
|
%tmp9q = load i32, i32* %arrayidx.q.2, align 4
|
|
%mul.5 = mul i32 %tmp9p, %tmp9q
|
|
%add.5 = add i32 %mul.5, %add.4
|
|
%tmp10p = load i32, i32* %arrayidx.p.6, align 4
|
|
%tmp10q = load i32, i32* %arrayidx.q.1, align 4
|
|
%mul.6 = mul i32 %tmp10p, %tmp10q
|
|
%add.6 = add i32 %mul.6, %add.5
|
|
%tmp11p = load i32, i32* %arrayidx.p.7, align 4
|
|
%tmp11q = load i32, i32* %q, align 4
|
|
%mul.7 = mul i32 %tmp11p, %tmp11q
|
|
%add.7 = add i32 %mul.7, %add.6
|
|
br i1 true, label %for.end, label %for.body
|
|
|
|
for.end:
|
|
ret i32 %add.7
|
|
}
|