82 lines
3.8 KiB
LLVM
82 lines
3.8 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -mtriple=x86_64-pc-linux < %s | FileCheck %s --check-prefix=X64
|
|
; RUN: llc -mtriple=i686-pc-linux < %s | FileCheck %s --check-prefix=X86
|
|
|
|
define double @foo(double %0) #0 {
|
|
; X64-LABEL: foo:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: subq $24, %rsp
|
|
; X64-NEXT: movsd %xmm0, (%rsp) # 8-byte Spill
|
|
; X64-NEXT: movl $1024, %edi # imm = 0x400
|
|
; X64-NEXT: callq fesetround@PLT
|
|
; X64-NEXT: movsd {{.*#+}} xmm1 = [1.0E+0,0.0E+0]
|
|
; X64-NEXT: divsd (%rsp), %xmm1 # 8-byte Folded Reload
|
|
; X64-NEXT: movsd %xmm1, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill
|
|
; X64-NEXT: movl $1024, %edi # imm = 0x400
|
|
; X64-NEXT: callq fesetround@PLT
|
|
; X64-NEXT: movsd {{.*#+}} xmm0 = [1.0E+0,0.0E+0]
|
|
; X64-NEXT: divsd (%rsp), %xmm0 # 8-byte Folded Reload
|
|
; X64-NEXT: movsd %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill
|
|
; X64-NEXT: movl $1024, %edi # imm = 0x400
|
|
; X64-NEXT: callq fesetround@PLT
|
|
; X64-NEXT: movsd {{.*#+}} xmm2 = [1.0E+0,0.0E+0]
|
|
; X64-NEXT: divsd (%rsp), %xmm2 # 8-byte Folded Reload
|
|
; X64-NEXT: movsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 8-byte Reload
|
|
; X64-NEXT: # xmm0 = mem[0],zero
|
|
; X64-NEXT: movsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm1 # 8-byte Reload
|
|
; X64-NEXT: # xmm1 = mem[0],zero
|
|
; X64-NEXT: callq fma@PLT
|
|
; X64-NEXT: addq $24, %rsp
|
|
; X64-NEXT: retq
|
|
;
|
|
; X86-LABEL: foo:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: subl $60, %esp
|
|
; X86-NEXT: fldl {{[0-9]+}}(%esp)
|
|
; X86-NEXT: fstpl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Spill
|
|
; X86-NEXT: wait
|
|
; X86-NEXT: movl $1024, (%esp) # imm = 0x400
|
|
; X86-NEXT: calll fesetround@PLT
|
|
; X86-NEXT: fld1
|
|
; X86-NEXT: fstl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Spill
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fdivrp %st, %st(1)
|
|
; X86-NEXT: fstpl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Spill
|
|
; X86-NEXT: wait
|
|
; X86-NEXT: movl $1024, (%esp) # imm = 0x400
|
|
; X86-NEXT: calll fesetround@PLT
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fdivp %st, %st(1)
|
|
; X86-NEXT: fstpl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Spill
|
|
; X86-NEXT: wait
|
|
; X86-NEXT: movl $1024, (%esp) # imm = 0x400
|
|
; X86-NEXT: calll fesetround@PLT
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fdivp %st, %st(1)
|
|
; X86-NEXT: fstpl {{[0-9]+}}(%esp)
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fstpl {{[0-9]+}}(%esp)
|
|
; X86-NEXT: fldl {{[-0-9]+}}(%e{{[sb]}}p) # 8-byte Folded Reload
|
|
; X86-NEXT: fstpl (%esp)
|
|
; X86-NEXT: wait
|
|
; X86-NEXT: calll fma
|
|
; X86-NEXT: addl $60, %esp
|
|
; X86-NEXT: retl
|
|
%2 = call i32 @fesetround(i32 noundef 1024)
|
|
%3 = call double @llvm.experimental.constrained.fdiv.f64(double 1.000000e+00, double %0, metadata !"round.dynamic", metadata !"fpexcept.ignore") #0
|
|
%4 = call i32 @fesetround(i32 noundef 1024)
|
|
%5 = call double @llvm.experimental.constrained.fdiv.f64(double 1.000000e+00, double %0, metadata !"round.dynamic", metadata !"fpexcept.ignore") #0
|
|
%6 = call i32 @fesetround(i32 noundef 1024)
|
|
%7 = call double @llvm.experimental.constrained.fdiv.f64(double 1.000000e+00, double %0, metadata !"round.dynamic", metadata !"fpexcept.ignore") #0
|
|
%8 = call double @llvm.experimental.constrained.fma.f64(double %3, double %5, double %7, metadata !"round.dynamic", metadata !"fpexcept.ignore") #0
|
|
ret double %8
|
|
}
|
|
|
|
declare i32 @fesetround(i32) #0
|
|
declare double @llvm.experimental.constrained.fdiv.f64(double, double, metadata, metadata) #0
|
|
declare double @llvm.experimental.constrained.fma.f64(double, double, double, metadata, metadata) #0
|
|
|
|
attributes #0 = { nounwind strictfp }
|