This reverts commit 77a0da926c as we've
received multiple reports of this significantly impacting performance,
in ways that don't seem to just be target specific cost models going
wrong. I would offer some reproducers, but the test changes here seem to
be full of them!
Reverting for now and hopefully we can remove the "hack" more carefully
as we go.
385 lines
13 KiB
LLVM
385 lines
13 KiB
LLVM
; This test verifies that the loop vectorizer will NOT produce a tail
|
|
; loop with the optimize for size or the minimize size attributes.
|
|
; REQUIRES: asserts
|
|
; RUN: opt < %s -enable-new-pm=0 -loop-vectorize -S | FileCheck %s
|
|
; RUN: opt < %s -enable-new-pm=0 -loop-vectorize -pgso -S | FileCheck %s -check-prefix=PGSO
|
|
; RUN: opt < %s -enable-new-pm=0 -loop-vectorize -pgso=false -S | FileCheck %s -check-prefix=NPGSO
|
|
; RUN: opt < %s -passes='require<profile-summary>,loop-vectorize' -S | FileCheck %s
|
|
; RUN: opt < %s -passes='require<profile-summary>,loop-vectorize' -pgso -S | FileCheck %s -check-prefix=PGSO
|
|
; RUN: opt < %s -passes='require<profile-summary>,loop-vectorize' -pgso=false -S | FileCheck %s -check-prefix=NPGSO
|
|
|
|
target datalayout = "E-m:e-p:32:32-i64:32-f64:32:64-a:0:32-n32-S128"
|
|
|
|
@tab = common global [32 x i8] zeroinitializer, align 1
|
|
|
|
define i32 @foo_optsize() #0 {
|
|
; CHECK-LABEL: @foo_optsize(
|
|
; CHECK-NOT: <2 x i8>
|
|
; CHECK-NOT: <4 x i8>
|
|
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body: ; preds = %for.body, %entry
|
|
%i.08 = phi i32 [ 0, %entry ], [ %inc, %for.body ]
|
|
%arrayidx = getelementptr inbounds [32 x i8], [32 x i8]* @tab, i32 0, i32 %i.08
|
|
%0 = load i8, i8* %arrayidx, align 1
|
|
%cmp1 = icmp eq i8 %0, 0
|
|
%. = select i1 %cmp1, i8 2, i8 1
|
|
store i8 %., i8* %arrayidx, align 1
|
|
%inc = add nsw i32 %i.08, 1
|
|
%exitcond = icmp eq i32 %i.08, 202
|
|
br i1 %exitcond, label %for.end, label %for.body
|
|
|
|
for.end: ; preds = %for.body
|
|
ret i32 0
|
|
}
|
|
|
|
attributes #0 = { optsize }
|
|
|
|
define i32 @foo_minsize() #1 {
|
|
; CHECK-LABEL: @foo_minsize(
|
|
; CHECK-NOT: <2 x i8>
|
|
; CHECK-NOT: <4 x i8>
|
|
; CHECK-LABEL: @foo_pgso(
|
|
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body: ; preds = %for.body, %entry
|
|
%i.08 = phi i32 [ 0, %entry ], [ %inc, %for.body ]
|
|
%arrayidx = getelementptr inbounds [32 x i8], [32 x i8]* @tab, i32 0, i32 %i.08
|
|
%0 = load i8, i8* %arrayidx, align 1
|
|
%cmp1 = icmp eq i8 %0, 0
|
|
%. = select i1 %cmp1, i8 2, i8 1
|
|
store i8 %., i8* %arrayidx, align 1
|
|
%inc = add nsw i32 %i.08, 1
|
|
%exitcond = icmp eq i32 %i.08, 202
|
|
br i1 %exitcond, label %for.end, label %for.body
|
|
|
|
for.end: ; preds = %for.body
|
|
ret i32 0
|
|
}
|
|
|
|
attributes #1 = { minsize }
|
|
|
|
define i32 @foo_pgso() !prof !14 {
|
|
; PGSO-LABEL: @foo_pgso(
|
|
; PGSO-NOT: <{{[0-9]+}} x i8>
|
|
; NPGSO-LABEL: @foo_pgso(
|
|
; NPGSO: <{{[0-9]+}} x i8>
|
|
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body: ; preds = %for.body, %entry
|
|
%i.08 = phi i32 [ 0, %entry ], [ %inc, %for.body ]
|
|
%arrayidx = getelementptr inbounds [32 x i8], [32 x i8]* @tab, i32 0, i32 %i.08
|
|
%0 = load i8, i8* %arrayidx, align 1
|
|
%cmp1 = icmp eq i8 %0, 0
|
|
%. = select i1 %cmp1, i8 2, i8 1
|
|
store i8 %., i8* %arrayidx, align 1
|
|
%inc = add nsw i32 %i.08, 1
|
|
%exitcond = icmp eq i32 %i.08, 202
|
|
br i1 %exitcond, label %for.end, label %for.body
|
|
|
|
for.end: ; preds = %for.body
|
|
ret i32 0
|
|
}
|
|
|
|
; PR43371: don't run into an assert due to emitting SCEV runtime checks
|
|
; with OptForSize.
|
|
;
|
|
@cm_array = external global [2592 x i16], align 1
|
|
|
|
define void @pr43371() optsize {
|
|
;
|
|
; CHECK-LABEL: @pr43371
|
|
; CHECK-NOT: vector.scevcheck
|
|
;
|
|
; We do not want to generate SCEV predicates when optimising for size, because
|
|
; that will lead to extra code generation such as the SCEV overflow runtime
|
|
; checks. Not generating SCEV predicates can still result in vectorisation as
|
|
; the non-consecutive loads/stores can be scalarized:
|
|
;
|
|
; CHECK: vector.body:
|
|
; CHECK: store i16 0, i16* %{{.*}}, align 1
|
|
; CHECK: store i16 0, i16* %{{.*}}, align 1
|
|
; CHECK: br i1 {{.*}}, label %vector.body
|
|
;
|
|
entry:
|
|
br label %for.body29
|
|
|
|
for.cond.cleanup28:
|
|
unreachable
|
|
|
|
for.body29:
|
|
%i24.0170 = phi i16 [ 0, %entry], [ %inc37, %for.body29]
|
|
%add33 = add i16 undef, %i24.0170
|
|
%idxprom34 = zext i16 %add33 to i32
|
|
%arrayidx35 = getelementptr [2592 x i16], [2592 x i16] * @cm_array, i32 0, i32 %idxprom34
|
|
store i16 0, i16 * %arrayidx35, align 1
|
|
%inc37 = add i16 %i24.0170, 1
|
|
%cmp26 = icmp ult i16 %inc37, 756
|
|
br i1 %cmp26, label %for.body29, label %for.cond.cleanup28
|
|
}
|
|
|
|
define void @pr43371_pgso() !prof !14 {
|
|
;
|
|
; CHECK-LABEL: @pr43371_pgso
|
|
; CHECK-NOT: vector.scevcheck
|
|
;
|
|
; We do not want to generate SCEV predicates when optimising for size, because
|
|
; that will lead to extra code generation such as the SCEV overflow runtime
|
|
; checks. Not generating SCEV predicates can still result in vectorisation as
|
|
; the non-consecutive loads/stores can be scalarized:
|
|
;
|
|
; CHECK: vector.body:
|
|
; CHECK: store i16 0, i16* %{{.*}}, align 1
|
|
; CHECK: store i16 0, i16* %{{.*}}, align 1
|
|
; CHECK: br i1 {{.*}}, label %vector.body
|
|
;
|
|
entry:
|
|
br label %for.body29
|
|
|
|
for.cond.cleanup28:
|
|
unreachable
|
|
|
|
for.body29:
|
|
%i24.0170 = phi i16 [ 0, %entry], [ %inc37, %for.body29]
|
|
%add33 = add i16 undef, %i24.0170
|
|
%idxprom34 = zext i16 %add33 to i32
|
|
%arrayidx35 = getelementptr [2592 x i16], [2592 x i16] * @cm_array, i32 0, i32 %idxprom34
|
|
store i16 0, i16 * %arrayidx35, align 1
|
|
%inc37 = add i16 %i24.0170, 1
|
|
%cmp26 = icmp ult i16 %inc37, 756
|
|
br i1 %cmp26, label %for.body29, label %for.cond.cleanup28
|
|
}
|
|
|
|
; PR45526: don't vectorize with fold-tail if first-order-recurrence is live-out.
|
|
;
|
|
define i32 @pr45526() optsize {
|
|
;
|
|
; CHECK-LABEL: @pr45526
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: br label %loop
|
|
; CHECK-EMPTY:
|
|
; CHECK-NEXT: loop:
|
|
; CHECK-NEXT: %piv = phi i32 [ 0, %entry ], [ %pivPlus1, %loop ]
|
|
; CHECK-NEXT: %for = phi i32 [ 5, %entry ], [ %pivPlus1, %loop ]
|
|
; CHECK-NEXT: %pivPlus1 = add nuw nsw i32 %piv, 1
|
|
; CHECK-NEXT: %cond = icmp ult i32 %piv, 510
|
|
; CHECK-NEXT: br i1 %cond, label %loop, label %exit
|
|
; CHECK-EMPTY:
|
|
; CHECK-NEXT: exit:
|
|
; CHECK-NEXT: %for.lcssa = phi i32 [ %for, %loop ]
|
|
; CHECK-NEXT: ret i32 %for.lcssa
|
|
;
|
|
entry:
|
|
br label %loop
|
|
|
|
loop:
|
|
%piv = phi i32 [ 0, %entry ], [ %pivPlus1, %loop ]
|
|
%for = phi i32 [ 5, %entry ], [ %pivPlus1, %loop ]
|
|
%pivPlus1 = add nuw nsw i32 %piv, 1
|
|
%cond = icmp ult i32 %piv, 510
|
|
br i1 %cond, label %loop, label %exit
|
|
|
|
exit:
|
|
ret i32 %for
|
|
}
|
|
|
|
define i32 @pr45526_pgso() !prof !14 {
|
|
;
|
|
; CHECK-LABEL: @pr45526_pgso
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: br label %loop
|
|
; CHECK-EMPTY:
|
|
; CHECK-NEXT: loop:
|
|
; CHECK-NEXT: %piv = phi i32 [ 0, %entry ], [ %pivPlus1, %loop ]
|
|
; CHECK-NEXT: %for = phi i32 [ 5, %entry ], [ %pivPlus1, %loop ]
|
|
; CHECK-NEXT: %pivPlus1 = add nuw nsw i32 %piv, 1
|
|
; CHECK-NEXT: %cond = icmp ult i32 %piv, 510
|
|
; CHECK-NEXT: br i1 %cond, label %loop, label %exit
|
|
; CHECK-EMPTY:
|
|
; CHECK-NEXT: exit:
|
|
; CHECK-NEXT: %for.lcssa = phi i32 [ %for, %loop ]
|
|
; CHECK-NEXT: ret i32 %for.lcssa
|
|
;
|
|
entry:
|
|
br label %loop
|
|
|
|
loop:
|
|
%piv = phi i32 [ 0, %entry ], [ %pivPlus1, %loop ]
|
|
%for = phi i32 [ 5, %entry ], [ %pivPlus1, %loop ]
|
|
%pivPlus1 = add nuw nsw i32 %piv, 1
|
|
%cond = icmp ult i32 %piv, 510
|
|
br i1 %cond, label %loop, label %exit
|
|
|
|
exit:
|
|
ret i32 %for
|
|
}
|
|
|
|
; PR46228: Vectorize w/o versioning for unit stride under optsize and enabled
|
|
; vectorization.
|
|
|
|
; NOTE: Some assertions have been autogenerated by utils/update_test_checks.py
|
|
define void @stride1(i16* noalias %B, i32 %BStride) optsize {
|
|
; CHECK-LABEL: @stride1(
|
|
; CHECK-NEXT: entry:
|
|
; CHECK-NEXT: br i1 false, label [[SCALAR_PH:%.*]], label [[VECTOR_PH:%.*]]
|
|
; CHECK: vector.ph:
|
|
; CHECK-NEXT: [[BROADCAST_SPLATINSERT:%.*]] = insertelement <2 x i32> poison, i32 [[BSTRIDE:%.*]], i32 0
|
|
; CHECK-NEXT: [[BROADCAST_SPLAT:%.*]] = shufflevector <2 x i32> [[BROADCAST_SPLATINSERT]], <2 x i32> poison, <2 x i32> zeroinitializer
|
|
; CHECK-NEXT: br label [[VECTOR_BODY:%.*]]
|
|
; CHECK: vector.body:
|
|
; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, [[VECTOR_PH]] ], [ [[INDEX_NEXT:%.*]], [[PRED_STORE_CONTINUE2:%.*]] ]
|
|
; CHECK-NEXT: [[VEC_IND:%.*]] = phi <2 x i32> [ <i32 0, i32 1>, [[VECTOR_PH]] ], [ [[VEC_IND_NEXT:%.*]], [[PRED_STORE_CONTINUE2]] ]
|
|
; CHECK-NEXT: [[TMP1:%.*]] = icmp ule <2 x i32> [[VEC_IND]], <i32 1024, i32 1024>
|
|
; CHECK-NEXT: [[TMP0:%.*]] = mul nsw <2 x i32> [[VEC_IND]], [[BROADCAST_SPLAT]]
|
|
; CHECK-NEXT: [[TMP2:%.*]] = extractelement <2 x i1> [[TMP1]], i32 0
|
|
; CHECK-NEXT: br i1 [[TMP2]], label [[PRED_STORE_IF:%.*]], label [[PRED_STORE_CONTINUE:%.*]]
|
|
; CHECK: pred.store.if:
|
|
; CHECK-NEXT: [[TMP3:%.*]] = extractelement <2 x i32> [[TMP0]], i32 0
|
|
; CHECK-NEXT: [[TMP4:%.*]] = getelementptr inbounds i16, i16* [[B:%.*]], i32 [[TMP3]]
|
|
; CHECK-NEXT: store i16 42, i16* [[TMP4]], align 4
|
|
; CHECK-NEXT: br label [[PRED_STORE_CONTINUE]]
|
|
; CHECK: pred.store.continue:
|
|
; CHECK-NEXT: [[TMP5:%.*]] = extractelement <2 x i1> [[TMP1]], i32 1
|
|
; CHECK-NEXT: br i1 [[TMP5]], label [[PRED_STORE_IF1:%.*]], label [[PRED_STORE_CONTINUE2]]
|
|
; CHECK: pred.store.if1:
|
|
; CHECK-NEXT: [[TMP6:%.*]] = extractelement <2 x i32> [[TMP0]], i32 1
|
|
; CHECK-NEXT: [[TMP7:%.*]] = getelementptr inbounds i16, i16* [[B]], i32 [[TMP6]]
|
|
; CHECK-NEXT: store i16 42, i16* [[TMP7]], align 4
|
|
; CHECK-NEXT: br label [[PRED_STORE_CONTINUE2]]
|
|
; CHECK: pred.store.continue2:
|
|
; CHECK-NEXT: [[INDEX_NEXT]] = add i32 [[INDEX]], 2
|
|
; CHECK-NEXT: [[VEC_IND_NEXT]] = add <2 x i32> [[VEC_IND]], <i32 2, i32 2>
|
|
; CHECK-NEXT: [[TMP8:%.*]] = icmp eq i32 [[INDEX_NEXT]], 1026
|
|
; CHECK-NEXT: br i1 [[TMP8]], label [[MIDDLE_BLOCK:%.*]], label [[VECTOR_BODY]], !llvm.loop !21
|
|
; CHECK: middle.block:
|
|
; CHECK-NEXT: br i1 true, label [[FOR_END:%.*]], label [[SCALAR_PH]]
|
|
; CHECK: scalar.ph:
|
|
; CHECK: for.end:
|
|
; CHECK-NEXT: ret void
|
|
;
|
|
; PGSO-LABEL: @stride1(
|
|
; PGSO-NEXT: entry:
|
|
; PGSO-NEXT: br i1 false, label %scalar.ph, label %vector.ph
|
|
;
|
|
; NPGSO-LABEL: @stride1(
|
|
; NPGSO-NEXT: entry:
|
|
; NPGSO-NEXT: br i1 false, label %scalar.ph, label %vector.ph
|
|
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body:
|
|
%iv = phi i32 [ %iv.next, %for.body ], [ 0, %entry ]
|
|
%mulB = mul nsw i32 %iv, %BStride
|
|
%gepOfB = getelementptr inbounds i16, i16* %B, i32 %mulB
|
|
store i16 42, i16* %gepOfB, align 4
|
|
%iv.next = add nuw nsw i32 %iv, 1
|
|
%exitcond = icmp eq i32 %iv.next, 1025
|
|
br i1 %exitcond, label %for.end, label %for.body, !llvm.loop !15
|
|
|
|
for.end:
|
|
ret void
|
|
}
|
|
|
|
; Vectorize with versioning for unit stride for PGSO and enabled vectorization.
|
|
;
|
|
define void @stride1_pgso(i16* noalias %B, i32 %BStride) !prof !14 {
|
|
; CHECK-LABEL: @stride1_pgso(
|
|
; CHECK: vector.body
|
|
;
|
|
; PGSO-LABEL: @stride1_pgso(
|
|
; PGSO: vector.body
|
|
;
|
|
; NPGSO-LABEL: @stride1_pgso(
|
|
; NPGSO: vector.body
|
|
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body:
|
|
%iv = phi i32 [ %iv.next, %for.body ], [ 0, %entry ]
|
|
%mulB = mul nsw i32 %iv, %BStride
|
|
%gepOfB = getelementptr inbounds i16, i16* %B, i32 %mulB
|
|
store i16 42, i16* %gepOfB, align 4
|
|
%iv.next = add nuw nsw i32 %iv, 1
|
|
%exitcond = icmp eq i32 %iv.next, 1025
|
|
br i1 %exitcond, label %for.end, label %for.body, !llvm.loop !15
|
|
|
|
for.end:
|
|
ret void
|
|
}
|
|
|
|
; PR46652: Check that the need for stride==1 check prevents vectorizing a loop
|
|
; having tiny trip count, when compiling w/o -Os/-Oz.
|
|
; CHECK-LABEL: @pr46652
|
|
; CHECK-NOT: vector.scevcheck
|
|
; CHECK-NOT: vector.body
|
|
; CHECK-LABEL: for.body
|
|
|
|
@g = external global [1 x i16], align 1
|
|
|
|
define void @pr46652(i16 %stride) {
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body: ; preds = %for.body, %entry
|
|
%l1.02 = phi i16 [ 1, %entry ], [ %inc9, %for.body ]
|
|
%mul = mul nsw i16 %l1.02, %stride
|
|
%arrayidx6 = getelementptr inbounds [1 x i16], [1 x i16]* @g, i16 0, i16 %mul
|
|
%0 = load i16, i16* %arrayidx6, align 1
|
|
%inc9 = add nuw nsw i16 %l1.02, 1
|
|
%exitcond.not = icmp eq i16 %inc9, 16
|
|
br i1 %exitcond.not, label %for.end, label %for.body
|
|
|
|
for.end: ; preds = %for.body
|
|
ret void
|
|
}
|
|
|
|
; Make sure we do not crash while building the VPlan for the loop with the
|
|
; select below.
|
|
define i32 @PR48142(i32* %ptr.start, i32* %ptr.end) optsize {
|
|
; CHECK-LABEL: PR48142
|
|
; CHECK-NOT: vector.body
|
|
entry:
|
|
br label %for.body
|
|
|
|
for.body:
|
|
%i.014 = phi i32 [ 20, %entry ], [ %cond, %for.body ]
|
|
%ptr.iv = phi i32* [ %ptr.start, %entry ], [ %ptr.next, %for.body ]
|
|
%cmp4 = icmp slt i32 %i.014, 99
|
|
%cond = select i1 %cmp4, i32 99, i32 %i.014
|
|
store i32 0, i32* %ptr.iv
|
|
%ptr.next = getelementptr inbounds i32, i32* %ptr.iv, i64 1
|
|
%cmp.not = icmp eq i32* %ptr.next, %ptr.end
|
|
br i1 %cmp.not, label %exit, label %for.body
|
|
|
|
exit:
|
|
%res = phi i32 [ %cond, %for.body ]
|
|
ret i32 %res
|
|
}
|
|
|
|
!llvm.module.flags = !{!0}
|
|
!0 = !{i32 1, !"ProfileSummary", !1}
|
|
!1 = !{!2, !3, !4, !5, !6, !7, !8, !9}
|
|
!2 = !{!"ProfileFormat", !"InstrProf"}
|
|
!3 = !{!"TotalCount", i64 10000}
|
|
!4 = !{!"MaxCount", i64 10}
|
|
!5 = !{!"MaxInternalCount", i64 1}
|
|
!6 = !{!"MaxFunctionCount", i64 1000}
|
|
!7 = !{!"NumCounts", i64 3}
|
|
!8 = !{!"NumFunctions", i64 3}
|
|
!9 = !{!"DetailedSummary", !10}
|
|
!10 = !{!11, !12, !13}
|
|
!11 = !{i32 10000, i64 100, i32 1}
|
|
!12 = !{i32 999000, i64 100, i32 1}
|
|
!13 = !{i32 999999, i64 1, i32 2}
|
|
!14 = !{!"function_entry_count", i64 0}
|
|
!15 = distinct !{!15, !16}
|
|
!16 = !{!"llvm.loop.vectorize.enable", i1 true}
|