Files
clang-p2996/llvm/test/CodeGen/PowerPC/pr55463.ll
Chen Zheng b5e1fc19da [PowerPC] don't check CTR clobber in hardware loop insertion pass
We added a new post-isel CTRLoop pass in D122125. That pass will expand
the hardware loop related intrinsic to CTR loop or normal loop based
on the loop context. So we don't need to conservatively check the CTR
clobber now on the IR level.

Reviewed By: lkail

Differential Revision: https://reviews.llvm.org/D135847
2022-12-04 20:53:49 -05:00

107 lines
3.2 KiB
LLVM

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple=powerpcspe -verify-machineinstrs < %s | FileCheck %s
define void @baz() #0 {
; CHECK-LABEL: baz:
; CHECK: # %bb.0: # %bb
; CHECK-NEXT: mflr 0
; CHECK-NEXT: stwu 1, -16(1)
; CHECK-NEXT: stw 0, 20(1)
; CHECK-NEXT: mtctr 3
; CHECK-NEXT: .p2align 4
; CHECK-NEXT: .LBB0_1: # %bb1
; CHECK-NEXT: #
; CHECK-NEXT: bdnz .LBB0_1
; CHECK-NEXT: # %bb.2: # %bb8
; CHECK-NEXT: bl wibble
; CHECK-NEXT: lwz 0, 20(1)
; CHECK-NEXT: addi 1, 1, 16
; CHECK-NEXT: mtlr 0
; CHECK-NEXT: blr
bb:
br label %bb1
bb1:
%tmp = phi i32 [ %tmp6, %bb1 ], [ undef, %bb ]
%tmp2 = phi i32 [ %tmp3, %bb1 ], [ undef, %bb ]
%tmp3 = add nsw i32 %tmp2, 1
%tmp4 = sitofp i32 %tmp to double
%tmp5 = tail call double @llvm.fmuladd.f64(double 0.000000e+00, double -0.000000e+00, double %tmp4)
%tmp6 = fptosi double %tmp5 to i32
%tmp7 = icmp eq i32 %tmp2, 0
br i1 %tmp7, label %bb8, label %bb1
bb8:
call void @wibble(i32 %tmp6)
ret void
}
define void @wombat() #0 {
; CHECK-LABEL: wombat:
; CHECK: # %bb.0: # %bb
; CHECK-NEXT: mflr 0
; CHECK-NEXT: stwu 1, -48(1)
; CHECK-NEXT: li 3, .LCPI1_0@l
; CHECK-NEXT: li 5, .LCPI1_1@l
; CHECK-NEXT: lis 4, .LCPI1_0@ha
; CHECK-NEXT: lis 6, .LCPI1_1@ha
; CHECK-NEXT: stw 0, 52(1)
; CHECK-NEXT: evstdd 29, 24(1) # 8-byte Folded Spill
; CHECK-NEXT: evstdd 30, 32(1) # 8-byte Folded Spill
; CHECK-NEXT: evlddx 30, 4, 3
; CHECK-NEXT: # implicit-def: $r3
; CHECK-NEXT: evlddx 29, 6, 5
; CHECK-NEXT: evstdd 28, 16(1) # 8-byte Folded Spill
; CHECK-NEXT: # implicit-def: $r28
; CHECK-NEXT: .p2align 4
; CHECK-NEXT: .LBB1_1: # %bb1
; CHECK-NEXT: #
; CHECK-NEXT: efdcfsi 8, 3
; CHECK-NEXT: mr 4, 30
; CHECK-NEXT: mr 6, 29
; CHECK-NEXT: evmergehi 3, 30, 30
; CHECK-NEXT: evmergehi 5, 29, 29
; CHECK-NEXT: # kill: def $r3 killed $r3 killed $s3
; CHECK-NEXT: # kill: def $r5 killed $r5 killed $s5
; CHECK-NEXT: evmergehi 7, 8, 8
; CHECK-NEXT: # kill: def $r8 killed $r8 killed $s8
; CHECK-NEXT: # kill: def $r7 killed $r7 killed $s7
; CHECK-NEXT: bl fma
; CHECK-NEXT: evmergelo 3, 3, 4
; CHECK-NEXT: addi 28, 28, -1
; CHECK-NEXT: cmplwi 28, 0
; CHECK-NEXT: efdctsiz 3, 3
; CHECK-NEXT: bc 12, 1, .LBB1_1
; CHECK-NEXT: # %bb.2: # %bb8
; CHECK-NEXT: bl wibble
; CHECK-NEXT: evldd 30, 32(1) # 8-byte Folded Reload
; CHECK-NEXT: evldd 29, 24(1) # 8-byte Folded Reload
; CHECK-NEXT: evldd 28, 16(1) # 8-byte Folded Reload
; CHECK-NEXT: lwz 0, 52(1)
; CHECK-NEXT: addi 1, 1, 48
; CHECK-NEXT: mtlr 0
; CHECK-NEXT: blr
bb:
br label %bb1
bb1:
%tmp = phi i32 [ %tmp6, %bb1 ], [ undef, %bb ]
%tmp2 = phi i32 [ %tmp3, %bb1 ], [ undef, %bb ]
%tmp3 = add nsw i32 %tmp2, 1
%tmp4 = sitofp i32 %tmp to double
%tmp5 = tail call double @llvm.fma.f64(double 0.000000e+00, double -0.000000e+00, double %tmp4)
%tmp6 = fptosi double %tmp5 to i32
%tmp7 = icmp eq i32 %tmp2, 0
br i1 %tmp7, label %bb8, label %bb1
bb8:
call void @wibble(i32 %tmp6)
ret void
}
declare void @wibble(i32)
declare double @llvm.fmuladd.f64(double, double, double)
declare double @llvm.fma.f64(double, double, double)
attributes #0 = { nounwind }