Files
clang-p2996/llvm/test/CodeGen/SystemZ/fp-strict-mul-04.ll
Kai Nacke a1710eb3cd [SystemZ][NFC] Opaque pointer migration.
The LIT test cases were migrated with the script provided by
Nikita Popov.

No manual changes were made. Committed without review since
no functional changes, after consultation with uweigand.
2022-10-11 21:09:43 +00:00

436 lines
20 KiB
LLVM

; Test strict multiplication of two f64s, producing an f128 result.
;
; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
declare fp128 @llvm.experimental.constrained.fmul.f128(fp128, fp128, metadata, metadata)
declare double @llvm.experimental.constrained.fadd.f64(double, double, metadata, metadata)
declare double @llvm.experimental.constrained.fptrunc.f64.f128(fp128, metadata, metadata)
declare fp128 @llvm.experimental.constrained.fpext.f128.f64(double, metadata)
declare double @foo()
; Check register multiplication. "mxdbr %f0, %f2" is not valid from LLVM's
; point of view, because %f2 is the low register of the FP128 %f0. Pass the
; multiplier in %f4 instead.
define void @f1(double %f1, double %dummy, double %f2, ptr %dst) #0 {
; CHECK-LABEL: f1:
; CHECK: mxdbr %f0, %f4
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f1,
metadata !"fpexcept.strict") #0
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f2,
metadata !"fpexcept.strict") #0
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %f1x, fp128 %f2x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store fp128 %res, ptr %dst
ret void
}
; Check the low end of the MXDB range.
define void @f2(double %f1, ptr %ptr, ptr %dst) #0 {
; CHECK-LABEL: f2:
; CHECK: mxdb %f0, 0(%r2)
; CHECK: std %f0, 0(%r3)
; CHECK: std %f2, 8(%r3)
; CHECK: br %r14
%f2 = load double, ptr %ptr
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f1,
metadata !"fpexcept.strict") #0
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f2,
metadata !"fpexcept.strict") #0
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %f1x, fp128 %f2x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store fp128 %res, ptr %dst
ret void
}
; Check the high end of the aligned MXDB range.
define void @f3(double %f1, ptr %base, ptr %dst) #0 {
; CHECK-LABEL: f3:
; CHECK: mxdb %f0, 4088(%r2)
; CHECK: std %f0, 0(%r3)
; CHECK: std %f2, 8(%r3)
; CHECK: br %r14
%ptr = getelementptr double, ptr %base, i64 511
%f2 = load double, ptr %ptr
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f1,
metadata !"fpexcept.strict") #0
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f2,
metadata !"fpexcept.strict") #0
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %f1x, fp128 %f2x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store fp128 %res, ptr %dst
ret void
}
; Check the next doubleword up, which needs separate address logic.
; Other sequences besides this one would be OK.
define void @f4(double %f1, ptr %base, ptr %dst) #0 {
; CHECK-LABEL: f4:
; CHECK: aghi %r2, 4096
; CHECK: mxdb %f0, 0(%r2)
; CHECK: std %f0, 0(%r3)
; CHECK: std %f2, 8(%r3)
; CHECK: br %r14
%ptr = getelementptr double, ptr %base, i64 512
%f2 = load double, ptr %ptr
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f1,
metadata !"fpexcept.strict") #0
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f2,
metadata !"fpexcept.strict") #0
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %f1x, fp128 %f2x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store fp128 %res, ptr %dst
ret void
}
; Check negative displacements, which also need separate address logic.
define void @f5(double %f1, ptr %base, ptr %dst) #0 {
; CHECK-LABEL: f5:
; CHECK: aghi %r2, -8
; CHECK: mxdb %f0, 0(%r2)
; CHECK: std %f0, 0(%r3)
; CHECK: std %f2, 8(%r3)
; CHECK: br %r14
%ptr = getelementptr double, ptr %base, i64 -1
%f2 = load double, ptr %ptr
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f1,
metadata !"fpexcept.strict") #0
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f2,
metadata !"fpexcept.strict") #0
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %f1x, fp128 %f2x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store fp128 %res, ptr %dst
ret void
}
; Check that MXDB allows indices.
define void @f6(double %f1, ptr %base, i64 %index, ptr %dst) #0 {
; CHECK-LABEL: f6:
; CHECK: sllg %r1, %r3, 3
; CHECK: mxdb %f0, 800(%r1,%r2)
; CHECK: std %f0, 0(%r4)
; CHECK: std %f2, 8(%r4)
; CHECK: br %r14
%ptr1 = getelementptr double, ptr %base, i64 %index
%ptr2 = getelementptr double, ptr %ptr1, i64 100
%f2 = load double, ptr %ptr2
%f1x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f1,
metadata !"fpexcept.strict") #0
%f2x = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %f2,
metadata !"fpexcept.strict") #0
%res = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %f1x, fp128 %f2x,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store fp128 %res, ptr %dst
ret void
}
; Check that multiplications of spilled values can use MXDB rather than MXDBR.
define double @f7(ptr %ptr0) #0 {
; CHECK-LABEL: f7:
; CHECK: brasl %r14, foo@PLT
; CHECK: mxdb %f0, 160(%r15)
; CHECK: br %r14
%ptr1 = getelementptr double, ptr %ptr0, i64 2
%ptr2 = getelementptr double, ptr %ptr0, i64 4
%ptr3 = getelementptr double, ptr %ptr0, i64 6
%ptr4 = getelementptr double, ptr %ptr0, i64 8
%ptr5 = getelementptr double, ptr %ptr0, i64 10
%ptr6 = getelementptr double, ptr %ptr0, i64 12
%ptr7 = getelementptr double, ptr %ptr0, i64 14
%ptr8 = getelementptr double, ptr %ptr0, i64 16
%ptr9 = getelementptr double, ptr %ptr0, i64 18
%ptr10 = getelementptr double, ptr %ptr0, i64 20
%val0 = load double, ptr %ptr0
%val1 = load double, ptr %ptr1
%val2 = load double, ptr %ptr2
%val3 = load double, ptr %ptr3
%val4 = load double, ptr %ptr4
%val5 = load double, ptr %ptr5
%val6 = load double, ptr %ptr6
%val7 = load double, ptr %ptr7
%val8 = load double, ptr %ptr8
%val9 = load double, ptr %ptr9
%val10 = load double, ptr %ptr10
%frob0 = call double @llvm.experimental.constrained.fadd.f64(
double %val0, double %val0,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob1 = call double @llvm.experimental.constrained.fadd.f64(
double %val1, double %val1,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob2 = call double @llvm.experimental.constrained.fadd.f64(
double %val2, double %val2,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob3 = call double @llvm.experimental.constrained.fadd.f64(
double %val3, double %val3,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob4 = call double @llvm.experimental.constrained.fadd.f64(
double %val4, double %val4,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob5 = call double @llvm.experimental.constrained.fadd.f64(
double %val5, double %val5,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob6 = call double @llvm.experimental.constrained.fadd.f64(
double %val6, double %val6,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob7 = call double @llvm.experimental.constrained.fadd.f64(
double %val7, double %val7,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob8 = call double @llvm.experimental.constrained.fadd.f64(
double %val8, double %val8,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob9 = call double @llvm.experimental.constrained.fadd.f64(
double %val9, double %val9,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%frob10 = call double @llvm.experimental.constrained.fadd.f64(
double %val10, double %val10,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
store double %frob0, ptr %ptr0
store double %frob1, ptr %ptr1
store double %frob2, ptr %ptr2
store double %frob3, ptr %ptr3
store double %frob4, ptr %ptr4
store double %frob5, ptr %ptr5
store double %frob6, ptr %ptr6
store double %frob7, ptr %ptr7
store double %frob8, ptr %ptr8
store double %frob9, ptr %ptr9
store double %frob10, ptr %ptr10
%ret = call double @foo() #0
%accext0 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %ret,
metadata !"fpexcept.strict") #0
%ext0 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob0,
metadata !"fpexcept.strict") #0
%mul0 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext0, fp128 %ext0,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra0 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul0, fp128 0xL00000000000000003fff000001000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc0 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra0,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext1 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc0,
metadata !"fpexcept.strict") #0
%ext1 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob1,
metadata !"fpexcept.strict") #0
%mul1 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext1, fp128 %ext1,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra1 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul1, fp128 0xL00000000000000003fff000002000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc1 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra1,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext2 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc1,
metadata !"fpexcept.strict") #0
%ext2 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob2,
metadata !"fpexcept.strict") #0
%mul2 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext2, fp128 %ext2,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra2 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul2, fp128 0xL00000000000000003fff000003000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc2 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra2,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext3 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc2,
metadata !"fpexcept.strict") #0
%ext3 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob3,
metadata !"fpexcept.strict") #0
%mul3 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext3, fp128 %ext3,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra3 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul3, fp128 0xL00000000000000003fff000004000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc3 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra3,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext4 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc3,
metadata !"fpexcept.strict") #0
%ext4 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob4,
metadata !"fpexcept.strict") #0
%mul4 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext4, fp128 %ext4,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra4 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul4, fp128 0xL00000000000000003fff000005000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc4 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra4,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext5 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc4,
metadata !"fpexcept.strict") #0
%ext5 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob5,
metadata !"fpexcept.strict") #0
%mul5 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext5, fp128 %ext5,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra5 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul5, fp128 0xL00000000000000003fff000006000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc5 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra5,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext6 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc5,
metadata !"fpexcept.strict") #0
%ext6 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob6,
metadata !"fpexcept.strict") #0
%mul6 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext6, fp128 %ext6,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra6 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul6, fp128 0xL00000000000000003fff000007000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc6 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra6,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext7 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc6,
metadata !"fpexcept.strict") #0
%ext7 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob7,
metadata !"fpexcept.strict") #0
%mul7 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext7, fp128 %ext7,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra7 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul7, fp128 0xL00000000000000003fff000008000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc7 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra7,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext8 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc7,
metadata !"fpexcept.strict") #0
%ext8 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob8,
metadata !"fpexcept.strict") #0
%mul8 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext8, fp128 %ext8,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra8 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul8, fp128 0xL00000000000000003fff000009000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc8 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra8,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%accext9 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %trunc8,
metadata !"fpexcept.strict") #0
%ext9 = call fp128 @llvm.experimental.constrained.fpext.f128.f64(
double %frob9,
metadata !"fpexcept.strict") #0
%mul9 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %accext9, fp128 %ext9,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%extra9 = call fp128 @llvm.experimental.constrained.fmul.f128(
fp128 %mul9, fp128 0xL00000000000000003fff00000a000000,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
%trunc9 = call double @llvm.experimental.constrained.fptrunc.f64.f128(
fp128 %extra9,
metadata !"round.dynamic",
metadata !"fpexcept.strict") #0
ret double %trunc9
}
attributes #0 = { strictfp }