We added a new post-isel CTRLoop pass in D122125. That pass will expand the hardware loop related intrinsic to CTR loop or normal loop based on the loop context. So we don't need to conservatively check the CTR clobber now on the IR level. Reviewed By: lkail Differential Revision: https://reviews.llvm.org/D135847
107 lines
3.2 KiB
LLVM
107 lines
3.2 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=powerpcspe -verify-machineinstrs < %s | FileCheck %s
|
|
|
|
define void @baz() #0 {
|
|
; CHECK-LABEL: baz:
|
|
; CHECK: # %bb.0: # %bb
|
|
; CHECK-NEXT: mflr 0
|
|
; CHECK-NEXT: stwu 1, -16(1)
|
|
; CHECK-NEXT: stw 0, 20(1)
|
|
; CHECK-NEXT: mtctr 3
|
|
; CHECK-NEXT: .p2align 4
|
|
; CHECK-NEXT: .LBB0_1: # %bb1
|
|
; CHECK-NEXT: #
|
|
; CHECK-NEXT: bdnz .LBB0_1
|
|
; CHECK-NEXT: # %bb.2: # %bb8
|
|
; CHECK-NEXT: bl wibble
|
|
; CHECK-NEXT: lwz 0, 20(1)
|
|
; CHECK-NEXT: addi 1, 1, 16
|
|
; CHECK-NEXT: mtlr 0
|
|
; CHECK-NEXT: blr
|
|
bb:
|
|
br label %bb1
|
|
|
|
bb1:
|
|
%tmp = phi i32 [ %tmp6, %bb1 ], [ undef, %bb ]
|
|
%tmp2 = phi i32 [ %tmp3, %bb1 ], [ undef, %bb ]
|
|
%tmp3 = add nsw i32 %tmp2, 1
|
|
%tmp4 = sitofp i32 %tmp to double
|
|
%tmp5 = tail call double @llvm.fmuladd.f64(double 0.000000e+00, double -0.000000e+00, double %tmp4)
|
|
%tmp6 = fptosi double %tmp5 to i32
|
|
%tmp7 = icmp eq i32 %tmp2, 0
|
|
br i1 %tmp7, label %bb8, label %bb1
|
|
|
|
bb8:
|
|
call void @wibble(i32 %tmp6)
|
|
ret void
|
|
}
|
|
|
|
define void @wombat() #0 {
|
|
; CHECK-LABEL: wombat:
|
|
; CHECK: # %bb.0: # %bb
|
|
; CHECK-NEXT: mflr 0
|
|
; CHECK-NEXT: stwu 1, -48(1)
|
|
; CHECK-NEXT: li 3, .LCPI1_0@l
|
|
; CHECK-NEXT: li 5, .LCPI1_1@l
|
|
; CHECK-NEXT: lis 4, .LCPI1_0@ha
|
|
; CHECK-NEXT: lis 6, .LCPI1_1@ha
|
|
; CHECK-NEXT: stw 0, 52(1)
|
|
; CHECK-NEXT: evstdd 29, 24(1) # 8-byte Folded Spill
|
|
; CHECK-NEXT: evstdd 30, 32(1) # 8-byte Folded Spill
|
|
; CHECK-NEXT: evlddx 30, 4, 3
|
|
; CHECK-NEXT: # implicit-def: $r3
|
|
; CHECK-NEXT: evlddx 29, 6, 5
|
|
; CHECK-NEXT: evstdd 28, 16(1) # 8-byte Folded Spill
|
|
; CHECK-NEXT: # implicit-def: $r28
|
|
; CHECK-NEXT: .p2align 4
|
|
; CHECK-NEXT: .LBB1_1: # %bb1
|
|
; CHECK-NEXT: #
|
|
; CHECK-NEXT: efdcfsi 8, 3
|
|
; CHECK-NEXT: mr 4, 30
|
|
; CHECK-NEXT: mr 6, 29
|
|
; CHECK-NEXT: evmergehi 3, 30, 30
|
|
; CHECK-NEXT: evmergehi 5, 29, 29
|
|
; CHECK-NEXT: # kill: def $r3 killed $r3 killed $s3
|
|
; CHECK-NEXT: # kill: def $r5 killed $r5 killed $s5
|
|
; CHECK-NEXT: evmergehi 7, 8, 8
|
|
; CHECK-NEXT: # kill: def $r8 killed $r8 killed $s8
|
|
; CHECK-NEXT: # kill: def $r7 killed $r7 killed $s7
|
|
; CHECK-NEXT: bl fma
|
|
; CHECK-NEXT: evmergelo 3, 3, 4
|
|
; CHECK-NEXT: addi 28, 28, -1
|
|
; CHECK-NEXT: cmplwi 28, 0
|
|
; CHECK-NEXT: efdctsiz 3, 3
|
|
; CHECK-NEXT: bc 12, 1, .LBB1_1
|
|
; CHECK-NEXT: # %bb.2: # %bb8
|
|
; CHECK-NEXT: bl wibble
|
|
; CHECK-NEXT: evldd 30, 32(1) # 8-byte Folded Reload
|
|
; CHECK-NEXT: evldd 29, 24(1) # 8-byte Folded Reload
|
|
; CHECK-NEXT: evldd 28, 16(1) # 8-byte Folded Reload
|
|
; CHECK-NEXT: lwz 0, 52(1)
|
|
; CHECK-NEXT: addi 1, 1, 48
|
|
; CHECK-NEXT: mtlr 0
|
|
; CHECK-NEXT: blr
|
|
bb:
|
|
br label %bb1
|
|
|
|
bb1:
|
|
%tmp = phi i32 [ %tmp6, %bb1 ], [ undef, %bb ]
|
|
%tmp2 = phi i32 [ %tmp3, %bb1 ], [ undef, %bb ]
|
|
%tmp3 = add nsw i32 %tmp2, 1
|
|
%tmp4 = sitofp i32 %tmp to double
|
|
%tmp5 = tail call double @llvm.fma.f64(double 0.000000e+00, double -0.000000e+00, double %tmp4)
|
|
%tmp6 = fptosi double %tmp5 to i32
|
|
%tmp7 = icmp eq i32 %tmp2, 0
|
|
br i1 %tmp7, label %bb8, label %bb1
|
|
|
|
bb8:
|
|
call void @wibble(i32 %tmp6)
|
|
ret void
|
|
}
|
|
|
|
declare void @wibble(i32)
|
|
declare double @llvm.fmuladd.f64(double, double, double)
|
|
declare double @llvm.fma.f64(double, double, double)
|
|
|
|
attributes #0 = { nounwind }
|