There are some patterns in td files without MVT/class set for some operands in target pattern that are from the source pattern. This prevents GlobalISelEmitter from adding them as a valid rule, because the target child operand is an unsupported kind operand. For now, for a leaf child, only IntInit and DefInit are handled in GlobalISelEmitter. This issue can be workaround by adding MVT/class to the patterns in the td files, like the workarounds for patterns anyext and setcc in PPCInstrInfo.td in D140878. To avoid adding the same workarounds for other patterns in td files, this patch tries to handle the UnsetInit case in GlobalISelEmitter. Adding the new handling allows us to remove the workarounds in the td files and also generates many selection rules for PPC target. Reviewed By: arsenm Differential Revision: https://reviews.llvm.org/D141247
1324 lines
51 KiB
LLVM
1324 lines
51 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -global-isel -march=amdgcn -mcpu=tahiti < %s | FileCheck -check-prefix=GFX6 %s
|
|
; RUN: llc -global-isel -march=amdgcn -mcpu=hawaii < %s | FileCheck -check-prefix=GFX7 %s
|
|
; RUN: llc -global-isel -march=amdgcn -mcpu=fiji < %s | FileCheck -check-prefix=GFX8 %s
|
|
; RUN: llc -global-isel -march=amdgcn -mcpu=gfx900 < %s | FileCheck -check-prefix=GFX9 %s
|
|
; RUN: llc -global-isel -march=amdgcn -mcpu=gfx1010 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX10 %s
|
|
; RUN: llc -global-isel -march=amdgcn -mcpu=gfx1100 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX11 %s
|
|
; RUN: llc -global-isel=0 -march=amdgcn -mcpu=tahiti < %s | FileCheck -check-prefix=SDAG_GFX6 %s
|
|
; RUN: llc -global-isel=0 -march=amdgcn -mcpu=hawaii < %s | FileCheck -check-prefix=SDAG_GFX7 %s
|
|
; RUN: llc -global-isel=0 -march=amdgcn -mcpu=fiji < %s | FileCheck -check-prefix=SDAG_GFX8 %s
|
|
; RUN: llc -global-isel=0 -march=amdgcn -mcpu=gfx900 < %s | FileCheck -check-prefix=SDAG_GFX9 %s
|
|
; RUN: llc -global-isel=0 -march=amdgcn -mcpu=gfx1010 < %s | FileCheck -check-prefixes=SDAG_GFX10PLUS,SDAG_GFX10 %s
|
|
; RUN: llc -global-isel=0 -march=amdgcn -mcpu=gfx1100 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=SDAG_GFX10PLUS,SDAG_GFX11 %s
|
|
|
|
define float @v_roundeven_f32(float %x) {
|
|
; GFX6-LABEL: v_roundeven_f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call float @llvm.roundeven.f32(float %x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define <2 x float> @v_roundeven_v2f32(<2 x float> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v2f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <2 x float> @llvm.roundeven.v2f32(<2 x float> %x)
|
|
ret <2 x float> %roundeven
|
|
}
|
|
|
|
define <3 x float> @v_roundeven_v3f32(<3 x float> %x) {
|
|
; GFX6-LABEL: v_roundeven_v3f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v3f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v3f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v3f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v3f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <3 x float> @llvm.roundeven.v3f32(<3 x float> %x)
|
|
ret <3 x float> %roundeven
|
|
}
|
|
|
|
define <4 x float> @v_roundeven_v4f32(<4 x float> %x) {
|
|
; GFX6-LABEL: v_roundeven_v4f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v4f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v4f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX8-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v4f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX9-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v4f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <4 x float> @llvm.roundeven.v4f32(<4 x float> %x)
|
|
ret <4 x float> %roundeven
|
|
}
|
|
|
|
define half @v_roundeven_f16(half %x) {
|
|
; GFX6-LABEL: v_roundeven_f16:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f16:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f16:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f16:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f16:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call half @llvm.roundeven.f16(half %x)
|
|
ret half %roundeven
|
|
}
|
|
|
|
define <2 x half> @v_roundeven_v2f16(<2 x half> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f16:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f16:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f16:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f16:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_v2f16:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-LABEL: v_roundeven_v2f16:
|
|
; GFX11: ; %bb.0:
|
|
; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX11-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; GFX11-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-NEXT: v_rndne_f16_e32 v1, v1
|
|
; GFX11-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; GFX11-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v1, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v1, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v1, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX11: ; %bb.0:
|
|
; SDAG_GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX11-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX11-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX11-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <2 x half> @llvm.roundeven.v2f16(<2 x half> %x)
|
|
ret <2 x half> %roundeven
|
|
}
|
|
|
|
define <2 x half> @v_roundeven_v2f16_fneg(<2 x half> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; GFX6-NEXT: v_and_b32_e32 v0, 0xffff, v0
|
|
; GFX6-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX6-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v0
|
|
; GFX6-NEXT: v_lshrrev_b32_e32 v0, 16, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v2, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v1
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v2
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; GFX7-NEXT: v_and_b32_e32 v0, 0xffff, v0
|
|
; GFX7-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX7-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v0
|
|
; GFX7-NEXT: v_lshrrev_b32_e32 v0, 16, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v2, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v1
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v2
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX8-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX9-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX10-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX11: ; %bb.0:
|
|
; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX11-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX11-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; GFX11-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-NEXT: v_rndne_f16_e32 v1, v1
|
|
; GFX11-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; GFX11-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; SDAG_GFX6-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX6-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; SDAG_GFX6-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; SDAG_GFX7-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX7-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; SDAG_GFX7-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v1, -v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v1, -v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v1, -v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX11: ; %bb.0:
|
|
; SDAG_GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX11-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e64 v1, -v1
|
|
; SDAG_GFX11-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX11-NEXT: s_setpc_b64 s[30:31]
|
|
%x.fneg = fneg <2 x half> %x
|
|
%roundeven = call <2 x half> @llvm.roundeven.v2f16(<2 x half> %x.fneg)
|
|
ret <2 x half> %roundeven
|
|
}
|
|
|
|
define <4 x half> @v_roundeven_v4f16(<4 x half> %x) {
|
|
; GFX6-LABEL: v_roundeven_v4f16:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v4f16:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v4f16:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f16_e32 v2, v0
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_rndne_f16_e32 v3, v1
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v1, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_or_b32_e32 v0, v2, v0
|
|
; GFX8-NEXT: v_or_b32_e32 v1, v3, v1
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v4f16:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f16_e32 v2, v0
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_rndne_f16_e32 v3, v1
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v1, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_pack_b32_f16 v0, v2, v0
|
|
; GFX9-NEXT: v_pack_b32_f16 v1, v3, v1
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_v4f16:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10-NEXT: v_rndne_f16_e32 v2, v0
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_rndne_f16_e32 v3, v1
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v1, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_pack_b32_f16 v0, v2, v0
|
|
; GFX10-NEXT: v_pack_b32_f16 v1, v3, v1
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-LABEL: v_roundeven_v4f16:
|
|
; GFX11: ; %bb.0:
|
|
; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX11-NEXT: v_lshrrev_b32_e32 v2, 16, v0
|
|
; GFX11-NEXT: v_lshrrev_b32_e32 v3, 16, v1
|
|
; GFX11-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-NEXT: v_rndne_f16_e32 v1, v1
|
|
; GFX11-NEXT: v_rndne_f16_e32 v2, v2
|
|
; GFX11-NEXT: v_rndne_f16_e32 v3, v3
|
|
; GFX11-NEXT: v_pack_b32_f16 v0, v0, v2
|
|
; GFX11-NEXT: v_pack_b32_f16 v1, v1, v3
|
|
; GFX11-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v2, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v3, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v0, v0, v3
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v1, v1, v2
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v2, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v3, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v0, v0, v3
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v1, v1, v2
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v2, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v3, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v0, v0, v3
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v1, v1, v2
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX11: ; %bb.0:
|
|
; SDAG_GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX11-NEXT: v_lshrrev_b32_e32 v2, 16, v0
|
|
; SDAG_GFX11-NEXT: v_lshrrev_b32_e32 v3, 16, v1
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e32 v2, v2
|
|
; SDAG_GFX11-NEXT: v_rndne_f16_e32 v3, v3
|
|
; SDAG_GFX11-NEXT: v_pack_b32_f16 v0, v0, v2
|
|
; SDAG_GFX11-NEXT: v_pack_b32_f16 v1, v1, v3
|
|
; SDAG_GFX11-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <4 x half> @llvm.roundeven.v4f16(<4 x half> %x)
|
|
ret <4 x half> %roundeven
|
|
}
|
|
|
|
|
|
define float @v_roundeven_f32_fabs(float %x) {
|
|
; GFX6-LABEL: v_roundeven_f32_fabs:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f32_fabs:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f32_fabs:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f32_fabs:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f32_fabs:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%fabs.x = call float @llvm.fabs.f32(float %x)
|
|
%roundeven = call float @llvm.roundeven.f32(float %fabs.x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define amdgpu_ps float @s_roundeven_f32(float inreg %x) {
|
|
; GFX6-LABEL: s_roundeven_f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX6-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX7-LABEL: s_roundeven_f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX7-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX8-LABEL: s_roundeven_f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX8-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX9-LABEL: s_roundeven_f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX9-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX10PLUS-LABEL: s_roundeven_f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX10PLUS-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX6-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX6-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX7-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX7-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX8-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX8-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX9-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX9-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX10PLUS-NEXT: ; return to shader part epilog
|
|
%roundeven = call float @llvm.roundeven.f32(float %x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define float @v_roundeven_f32_fneg(float %x) {
|
|
; GFX6-LABEL: v_roundeven_f32_fneg:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f32_fneg:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f32_fneg:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f32_fneg:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f32_fneg:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%neg.x = fneg float %x
|
|
%roundeven = call float @llvm.roundeven.f32(float %neg.x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define double @v_roundeven_f64(double %x) {
|
|
; GFX6-LABEL: v_roundeven_f64:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_and_b32_e32 v3, 0x80000000, v1
|
|
; GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; GFX6-NEXT: v_or_b32_e32 v3, 0x43300000, v3
|
|
; GFX6-NEXT: v_add_f64 v[4:5], v[0:1], v[2:3]
|
|
; GFX6-NEXT: s_mov_b32 s4, -1
|
|
; GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v1, vcc
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f64:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f64:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f64:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f64:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: s_brev_b32 s4, -2
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0x43300000
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v3, s4, v2, v1
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[4:5], v[0:1], v[2:3]
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s4, -1
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v1, vcc
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call double @llvm.roundeven.f64(double %x)
|
|
ret double %roundeven
|
|
}
|
|
|
|
define double @v_roundeven_f64_fneg(double %x) {
|
|
; GFX6-LABEL: v_roundeven_f64_fneg:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_xor_b32_e32 v6, 0x80000000, v1
|
|
; GFX6-NEXT: v_and_b32_e32 v3, 0x80000000, v6
|
|
; GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; GFX6-NEXT: v_or_b32_e32 v3, 0x43300000, v3
|
|
; GFX6-NEXT: v_add_f64 v[4:5], -v[0:1], v[2:3]
|
|
; GFX6-NEXT: s_mov_b32 s4, -1
|
|
; GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v6, vcc
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f64_fneg:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f64_fneg:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f64_fneg:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f64_fneg:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_xor_b32_e32 v6, 0x80000000, v1
|
|
; SDAG_GFX6-NEXT: s_brev_b32 s4, -2
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0x43300000
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v3, s4, v2, v6
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[4:5], -v[0:1], v[2:3]
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s4, -1
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v6, vcc
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%neg.x = fneg double %x
|
|
%roundeven = call double @llvm.roundeven.f64(double %neg.x)
|
|
ret double %roundeven
|
|
}
|
|
|
|
define <2 x double> @v_roundeven_v2f64(<2 x double> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f64:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_and_b32_e32 v5, 0x80000000, v1
|
|
; GFX6-NEXT: v_mov_b32_e32 v4, 0
|
|
; GFX6-NEXT: v_or_b32_e32 v5, 0x43300000, v5
|
|
; GFX6-NEXT: v_add_f64 v[6:7], v[0:1], v[4:5]
|
|
; GFX6-NEXT: s_mov_b32 s4, -1
|
|
; GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; GFX6-NEXT: v_add_f64 v[5:6], v[6:7], -v[4:5]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v0, v5, v0, vcc
|
|
; GFX6-NEXT: v_and_b32_e32 v5, 0x80000000, v3
|
|
; GFX6-NEXT: v_or_b32_e32 v5, 0x43300000, v5
|
|
; GFX6-NEXT: v_add_f64 v[7:8], v[2:3], v[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v1, v6, v1, vcc
|
|
; GFX6-NEXT: v_add_f64 v[4:5], v[7:8], -v[4:5]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[2:3]|, s[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v2, v4, v2, vcc
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v3, v5, v3, vcc
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f64:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX7-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f64:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX8-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f64:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX9-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v2f64:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: s_brev_b32 s6, -2
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v8, 0x43300000
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v5, s6, v8, v1
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v4, 0
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[6:7], v[0:1], v[4:5]
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s4, -1
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[5:6], v[6:7], -v[4:5]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v0, v5, v0, vcc
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v5, s6, v8, v3
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[7:8], v[2:3], v[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v1, v6, v1, vcc
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[4:5], v[7:8], -v[4:5]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[2:3]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v2, v4, v2, vcc
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v3, v5, v3, vcc
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <2 x double> @llvm.roundeven.v2f64(<2 x double> %x)
|
|
ret <2 x double> %roundeven
|
|
}
|
|
|
|
declare half @llvm.roundeven.f16(half) #0
|
|
declare <2 x half> @llvm.roundeven.v2f16(<2 x half>) #0
|
|
declare <4 x half> @llvm.roundeven.v4f16(<4 x half>) #0
|
|
|
|
declare float @llvm.roundeven.f32(float) #0
|
|
declare <2 x float> @llvm.roundeven.v2f32(<2 x float>) #0
|
|
declare <3 x float> @llvm.roundeven.v3f32(<3 x float>) #0
|
|
declare <4 x float> @llvm.roundeven.v4f32(<4 x float>) #0
|
|
|
|
declare double @llvm.roundeven.f64(double) #0
|
|
declare <2 x double> @llvm.roundeven.v2f64(<2 x double>) #0
|
|
|
|
declare half @llvm.fabs.f16(half) #0
|
|
declare float @llvm.fabs.f32(float) #0
|
|
|
|
attributes #0 = { nounwind readnone speculatable willreturn }
|