This is a NFC patch. This patch run a bulk update on CodeGen tests that are impacted by the true16 features. This patch applies: 1. duplicate GFX11plus runlines and apply them with "+mattr=+real-true16" and "+mattr=-real-true16" 2. update the test with the update script For some GISEL runlines, the current CodeGen do not fully support the true16 version. Still update the runlines, but comment out the failing one, and added a "FIXME-TRUE16" comment to that test for easier tracking. These test will be fixed in the following patches. This is in a transition state that we support both "+real-true16/-real-true16" in our code base. We plan to move to "+real-true16" as default, and finally remove "-real-true16" mode and test lines.
1376 lines
54 KiB
LLVM
1376 lines
54 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=tahiti < %s | FileCheck -check-prefix=GFX6 %s
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=hawaii < %s | FileCheck -check-prefix=GFX7 %s
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=fiji < %s | FileCheck -check-prefix=GFX8 %s
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=gfx900 < %s | FileCheck -check-prefix=GFX9 %s
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=gfx1010 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX10 %s
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=gfx1100 -mattr=+real-true16 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX11,GFX11-TRUE16 %s
|
|
; RUN: llc -global-isel -mtriple=amdgcn -mcpu=gfx1100 -mattr=-real-true16 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX11,GFX11-FAKE16 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=tahiti < %s | FileCheck -check-prefix=SDAG_GFX6 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=hawaii < %s | FileCheck -check-prefix=SDAG_GFX7 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=fiji < %s | FileCheck -check-prefix=SDAG_GFX8 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=gfx900 < %s | FileCheck -check-prefix=SDAG_GFX9 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=gfx1010 < %s | FileCheck -check-prefixes=SDAG_GFX10PLUS,SDAG_GFX10 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=gfx1100 -mattr=+real-true16 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=SDAG_GFX10PLUS,SDAG_GFX11,SDAG_GFX11-TRUE16 %s
|
|
; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=gfx1100 -mattr=-real-true16 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=SDAG_GFX10PLUS,SDAG_GFX11,SDAG_GFX11-FAKE16 %s
|
|
|
|
define float @v_roundeven_f32(float %x) {
|
|
; GFX6-LABEL: v_roundeven_f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call float @llvm.roundeven.f32(float %x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define <2 x float> @v_roundeven_v2f32(<2 x float> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v2f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v2f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <2 x float> @llvm.roundeven.v2f32(<2 x float> %x)
|
|
ret <2 x float> %roundeven
|
|
}
|
|
|
|
define <3 x float> @v_roundeven_v3f32(<3 x float> %x) {
|
|
; GFX6-LABEL: v_roundeven_v3f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v3f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v3f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v3f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v3f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v3f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <3 x float> @llvm.roundeven.v3f32(<3 x float> %x)
|
|
ret <3 x float> %roundeven
|
|
}
|
|
|
|
define <4 x float> @v_roundeven_v4f32(<4 x float> %x) {
|
|
; GFX6-LABEL: v_roundeven_v4f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v4f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v4f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX8-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v4f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX9-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v4f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v4f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <4 x float> @llvm.roundeven.v4f32(<4 x float> %x)
|
|
ret <4 x float> %roundeven
|
|
}
|
|
|
|
define half @v_roundeven_f16(half %x) {
|
|
; GFX6-LABEL: v_roundeven_f16:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f16:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f16:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f16:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_f16:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-TRUE16-LABEL: v_roundeven_f16:
|
|
; GFX11-TRUE16: ; %bb.0:
|
|
; GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-FAKE16-LABEL: v_roundeven_f16:
|
|
; GFX11-FAKE16: ; %bb.0:
|
|
; GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-TRUE16-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX11-TRUE16: ; %bb.0:
|
|
; SDAG_GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; SDAG_GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-FAKE16-LABEL: v_roundeven_f16:
|
|
; SDAG_GFX11-FAKE16: ; %bb.0:
|
|
; SDAG_GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call half @llvm.roundeven.f16(half %x)
|
|
ret half %roundeven
|
|
}
|
|
|
|
define <2 x half> @v_roundeven_v2f16(<2 x half> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f16:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f16:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f16:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f16:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_v2f16:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-TRUE16-LABEL: v_roundeven_v2f16:
|
|
; GFX11-TRUE16: ; %bb.0:
|
|
; GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.h, v0.h
|
|
; GFX11-TRUE16-NEXT: v_pack_b32_f16 v0, v0.l, v0.h
|
|
; GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-FAKE16-LABEL: v_roundeven_v2f16:
|
|
; GFX11-FAKE16: ; %bb.0:
|
|
; GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v1, v1
|
|
; GFX11-FAKE16-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v1, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v1, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v1, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-TRUE16-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX11-TRUE16: ; %bb.0:
|
|
; SDAG_GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.h, v0.h
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; SDAG_GFX11-TRUE16-NEXT: v_pack_b32_f16 v0, v0.l, v0.h
|
|
; SDAG_GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-FAKE16-LABEL: v_roundeven_v2f16:
|
|
; SDAG_GFX11-FAKE16: ; %bb.0:
|
|
; SDAG_GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX11-FAKE16-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <2 x half> @llvm.roundeven.v2f16(<2 x half> %x)
|
|
ret <2 x half> %roundeven
|
|
}
|
|
|
|
define <2 x half> @v_roundeven_v2f16_fneg(<2 x half> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; GFX6-NEXT: v_and_b32_e32 v0, 0xffff, v0
|
|
; GFX6-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX6-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v0
|
|
; GFX6-NEXT: v_lshrrev_b32_e32 v0, 16, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v2, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v1
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v2
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; GFX7-NEXT: v_and_b32_e32 v0, 0xffff, v0
|
|
; GFX7-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX7-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v0
|
|
; GFX7-NEXT: v_lshrrev_b32_e32 v0, 16, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v2, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v1
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v2
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX8-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX9-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX10-NEXT: v_rndne_f16_e32 v1, v0
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_pack_b32_f16 v0, v1, v0
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-TRUE16-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX11-TRUE16: ; %bb.0:
|
|
; GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-TRUE16-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.h, v0.h
|
|
; GFX11-TRUE16-NEXT: v_pack_b32_f16 v0, v0.l, v0.h
|
|
; GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-FAKE16-LABEL: v_roundeven_v2f16_fneg:
|
|
; GFX11-FAKE16: ; %bb.0:
|
|
; GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-FAKE16-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v1, v1
|
|
; GFX11-FAKE16-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; SDAG_GFX6-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX6-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; SDAG_GFX6-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_lshlrev_b32_e32 v1, 16, v1
|
|
; SDAG_GFX7-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX7-NEXT: v_xor_b32_e32 v0, 0x80008000, v0
|
|
; SDAG_GFX7-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v1, -v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v0, v0, v1
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v1, -v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v1, -v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-TRUE16-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX11-TRUE16: ; %bb.0:
|
|
; SDAG_GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e64 v0.h, -v0.h
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e64 v0.l, -v0.l
|
|
; SDAG_GFX11-TRUE16-NEXT: v_pack_b32_f16 v0, v0.l, v0.h
|
|
; SDAG_GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-FAKE16-LABEL: v_roundeven_v2f16_fneg:
|
|
; SDAG_GFX11-FAKE16: ; %bb.0:
|
|
; SDAG_GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v1, 16, v0
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e64 v0, -v0
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e64 v1, -v1
|
|
; SDAG_GFX11-FAKE16-NEXT: v_pack_b32_f16 v0, v0, v1
|
|
; SDAG_GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
%x.fneg = fneg <2 x half> %x
|
|
%roundeven = call <2 x half> @llvm.roundeven.v2f16(<2 x half> %x.fneg)
|
|
ret <2 x half> %roundeven
|
|
}
|
|
|
|
define <4 x half> @v_roundeven_v4f16(<4 x half> %x) {
|
|
; GFX6-LABEL: v_roundeven_v4f16:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; GFX6-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; GFX6-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v4f16:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; GFX7-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; GFX7-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v4f16:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f16_e32 v2, v0
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_rndne_f16_e32 v3, v1
|
|
; GFX8-NEXT: v_rndne_f16_sdwa v1, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX8-NEXT: v_or_b32_e32 v0, v2, v0
|
|
; GFX8-NEXT: v_or_b32_e32 v1, v3, v1
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v4f16:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f16_e32 v2, v0
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_rndne_f16_e32 v3, v1
|
|
; GFX9-NEXT: v_rndne_f16_sdwa v1, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX9-NEXT: v_pack_b32_f16 v0, v2, v0
|
|
; GFX9-NEXT: v_pack_b32_f16 v1, v3, v1
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10-LABEL: v_roundeven_v4f16:
|
|
; GFX10: ; %bb.0:
|
|
; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10-NEXT: v_rndne_f16_e32 v2, v0
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v0, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_rndne_f16_e32 v3, v1
|
|
; GFX10-NEXT: v_rndne_f16_sdwa v1, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; GFX10-NEXT: v_pack_b32_f16 v0, v2, v0
|
|
; GFX10-NEXT: v_pack_b32_f16 v1, v3, v1
|
|
; GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-TRUE16-LABEL: v_roundeven_v4f16:
|
|
; GFX11-TRUE16: ; %bb.0:
|
|
; GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.h, v0.h
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v1.l, v1.l
|
|
; GFX11-TRUE16-NEXT: v_rndne_f16_e32 v1.h, v1.h
|
|
; GFX11-TRUE16-NEXT: v_pack_b32_f16 v0, v0.l, v0.h
|
|
; GFX11-TRUE16-NEXT: v_pack_b32_f16 v1, v1.l, v1.h
|
|
; GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX11-FAKE16-LABEL: v_roundeven_v4f16:
|
|
; GFX11-FAKE16: ; %bb.0:
|
|
; GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v2, 16, v0
|
|
; GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v3, 16, v1
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v1, v1
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v2, v2
|
|
; GFX11-FAKE16-NEXT: v_rndne_f16_e32 v3, v3
|
|
; GFX11-FAKE16-NEXT: v_pack_b32_f16 v0, v0, v2
|
|
; GFX11-FAKE16-NEXT: v_pack_b32_f16 v1, v1, v3
|
|
; GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_cvt_f16_f32_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_cvt_f32_f16_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, v0
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v1, v1
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v2, v2
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v3, v3
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v2, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_sdwa v3, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX8-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v0, v0, v3
|
|
; SDAG_GFX8-NEXT: v_or_b32_e32 v1, v1, v2
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v2, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_sdwa v3, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX9-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v0, v0, v3
|
|
; SDAG_GFX9-NEXT: v_pack_b32_f16 v1, v1, v2
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX10: ; %bb.0:
|
|
; SDAG_GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v2, v1 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_sdwa v3, v0 dst_sel:DWORD dst_unused:UNUSED_PAD src0_sel:WORD_1
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX10-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v0, v0, v3
|
|
; SDAG_GFX10-NEXT: v_pack_b32_f16 v1, v1, v2
|
|
; SDAG_GFX10-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-TRUE16-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX11-TRUE16: ; %bb.0:
|
|
; SDAG_GFX11-TRUE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v1.h, v1.h
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.h, v0.h
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v0.l, v0.l
|
|
; SDAG_GFX11-TRUE16-NEXT: v_rndne_f16_e32 v1.l, v1.l
|
|
; SDAG_GFX11-TRUE16-NEXT: v_pack_b32_f16 v0, v0.l, v0.h
|
|
; SDAG_GFX11-TRUE16-NEXT: v_pack_b32_f16 v1, v1.l, v1.h
|
|
; SDAG_GFX11-TRUE16-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX11-FAKE16-LABEL: v_roundeven_v4f16:
|
|
; SDAG_GFX11-FAKE16: ; %bb.0:
|
|
; SDAG_GFX11-FAKE16-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v2, 16, v0
|
|
; SDAG_GFX11-FAKE16-NEXT: v_lshrrev_b32_e32 v3, 16, v1
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v1, v1
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v0, v0
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v2, v2
|
|
; SDAG_GFX11-FAKE16-NEXT: v_rndne_f16_e32 v3, v3
|
|
; SDAG_GFX11-FAKE16-NEXT: v_pack_b32_f16 v0, v0, v2
|
|
; SDAG_GFX11-FAKE16-NEXT: v_pack_b32_f16 v1, v1, v3
|
|
; SDAG_GFX11-FAKE16-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <4 x half> @llvm.roundeven.v4f16(<4 x half> %x)
|
|
ret <4 x half> %roundeven
|
|
}
|
|
|
|
|
|
define float @v_roundeven_f32_fabs(float %x) {
|
|
; GFX6-LABEL: v_roundeven_f32_fabs:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f32_fabs:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f32_fabs:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f32_fabs:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f32_fabs:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f32_fabs:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e64 v0, |v0|
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%fabs.x = call float @llvm.fabs.f32(float %x)
|
|
%roundeven = call float @llvm.roundeven.f32(float %fabs.x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define amdgpu_ps float @s_roundeven_f32(float inreg %x) {
|
|
; GFX6-LABEL: s_roundeven_f32:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX6-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX7-LABEL: s_roundeven_f32:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX7-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX8-LABEL: s_roundeven_f32:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX8-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX9-LABEL: s_roundeven_f32:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX9-NEXT: ; return to shader part epilog
|
|
;
|
|
; GFX10PLUS-LABEL: s_roundeven_f32:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e32 v0, s0
|
|
; GFX10PLUS-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX6-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX6-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX7-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX7-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX8-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX8-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX9-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX9-NEXT: ; return to shader part epilog
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: s_roundeven_f32:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e32 v0, s0
|
|
; SDAG_GFX10PLUS-NEXT: ; return to shader part epilog
|
|
%roundeven = call float @llvm.roundeven.f32(float %x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define float @v_roundeven_f32_fneg(float %x) {
|
|
; GFX6-LABEL: v_roundeven_f32_fneg:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f32_fneg:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f32_fneg:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f32_fneg:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f32_fneg:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f32_fneg:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f32_e64 v0, -v0
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%neg.x = fneg float %x
|
|
%roundeven = call float @llvm.roundeven.f32(float %neg.x)
|
|
ret float %roundeven
|
|
}
|
|
|
|
define double @v_roundeven_f64(double %x) {
|
|
; GFX6-LABEL: v_roundeven_f64:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_and_b32_e32 v3, 0x80000000, v1
|
|
; GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; GFX6-NEXT: v_or_b32_e32 v3, 0x43300000, v3
|
|
; GFX6-NEXT: v_add_f64 v[4:5], v[0:1], v[2:3]
|
|
; GFX6-NEXT: v_mov_b32_e32 v6, -1
|
|
; GFX6-NEXT: v_mov_b32_e32 v7, 0x432fffff
|
|
; GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, v[6:7]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v1, vcc
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f64:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f64:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f64:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f64:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: s_brev_b32 s6, -2
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0x43300000
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v3, s6, v2, v1
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s4, -1
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[4:5], v[0:1], v[2:3]
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v1, vcc
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f64:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call double @llvm.roundeven.f64(double %x)
|
|
ret double %roundeven
|
|
}
|
|
|
|
define double @v_roundeven_f64_fneg(double %x) {
|
|
; GFX6-LABEL: v_roundeven_f64_fneg:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_xor_b32_e32 v8, 0x80000000, v1
|
|
; GFX6-NEXT: v_and_b32_e32 v3, 0x80000000, v8
|
|
; GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; GFX6-NEXT: v_or_b32_e32 v3, 0x43300000, v3
|
|
; GFX6-NEXT: v_add_f64 v[4:5], -v[0:1], v[2:3]
|
|
; GFX6-NEXT: v_mov_b32_e32 v6, -1
|
|
; GFX6-NEXT: v_mov_b32_e32 v7, 0x432fffff
|
|
; GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, v[6:7]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v8, vcc
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_f64_fneg:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_f64_fneg:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_f64_fneg:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_f64_fneg:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: v_xor_b32_e32 v6, 0x80000000, v1
|
|
; SDAG_GFX6-NEXT: s_brev_b32 s4, -2
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0x43300000
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v3, s4, v2, v6
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v2, 0
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[4:5], -v[0:1], v[2:3]
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s4, -1
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[2:3], v[4:5], -v[2:3]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v0, v2, v0, vcc
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v1, v3, v6, vcc
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_f64_fneg:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e64 v[0:1], -v[0:1]
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%neg.x = fneg double %x
|
|
%roundeven = call double @llvm.roundeven.f64(double %neg.x)
|
|
ret double %roundeven
|
|
}
|
|
|
|
define <2 x double> @v_roundeven_v2f64(<2 x double> %x) {
|
|
; GFX6-LABEL: v_roundeven_v2f64:
|
|
; GFX6: ; %bb.0:
|
|
; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX6-NEXT: v_and_b32_e32 v5, 0x80000000, v1
|
|
; GFX6-NEXT: v_mov_b32_e32 v4, 0
|
|
; GFX6-NEXT: v_or_b32_e32 v5, 0x43300000, v5
|
|
; GFX6-NEXT: v_add_f64 v[6:7], v[0:1], v[4:5]
|
|
; GFX6-NEXT: s_mov_b32 s4, -1
|
|
; GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; GFX6-NEXT: v_add_f64 v[5:6], v[6:7], -v[4:5]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v0, v5, v0, vcc
|
|
; GFX6-NEXT: v_and_b32_e32 v5, 0x80000000, v3
|
|
; GFX6-NEXT: v_or_b32_e32 v5, 0x43300000, v5
|
|
; GFX6-NEXT: v_add_f64 v[7:8], v[2:3], v[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v1, v6, v1, vcc
|
|
; GFX6-NEXT: v_add_f64 v[4:5], v[7:8], -v[4:5]
|
|
; GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[2:3]|, s[4:5]
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v2, v4, v2, vcc
|
|
; GFX6-NEXT: v_cndmask_b32_e32 v3, v5, v3, vcc
|
|
; GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX7-LABEL: v_roundeven_v2f64:
|
|
; GFX7: ; %bb.0:
|
|
; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX7-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX8-LABEL: v_roundeven_v2f64:
|
|
; GFX8: ; %bb.0:
|
|
; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX8-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX9-LABEL: v_roundeven_v2f64:
|
|
; GFX9: ; %bb.0:
|
|
; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX9-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; GFX10PLUS-LABEL: v_roundeven_v2f64:
|
|
; GFX10PLUS: ; %bb.0:
|
|
; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; GFX10PLUS-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX6-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX6: ; %bb.0:
|
|
; SDAG_GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX6-NEXT: s_brev_b32 s6, -2
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v8, 0x43300000
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v5, s6, v8, v1
|
|
; SDAG_GFX6-NEXT: v_mov_b32_e32 v4, 0
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[6:7], v[0:1], v[4:5]
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s4, -1
|
|
; SDAG_GFX6-NEXT: s_mov_b32 s5, 0x432fffff
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[5:6], v[6:7], -v[4:5]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[0:1]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v0, v5, v0, vcc
|
|
; SDAG_GFX6-NEXT: v_bfi_b32 v5, s6, v8, v3
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[7:8], v[2:3], v[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v1, v6, v1, vcc
|
|
; SDAG_GFX6-NEXT: v_add_f64 v[4:5], v[7:8], -v[4:5]
|
|
; SDAG_GFX6-NEXT: v_cmp_gt_f64_e64 vcc, |v[2:3]|, s[4:5]
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v2, v4, v2, vcc
|
|
; SDAG_GFX6-NEXT: v_cndmask_b32_e32 v3, v5, v3, vcc
|
|
; SDAG_GFX6-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX7-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX7: ; %bb.0:
|
|
; SDAG_GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX7-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX7-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX8-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX8: ; %bb.0:
|
|
; SDAG_GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX8-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX8-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX9-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX9: ; %bb.0:
|
|
; SDAG_GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX9-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX9-NEXT: s_setpc_b64 s[30:31]
|
|
;
|
|
; SDAG_GFX10PLUS-LABEL: v_roundeven_v2f64:
|
|
; SDAG_GFX10PLUS: ; %bb.0:
|
|
; SDAG_GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0)
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e32 v[0:1], v[0:1]
|
|
; SDAG_GFX10PLUS-NEXT: v_rndne_f64_e32 v[2:3], v[2:3]
|
|
; SDAG_GFX10PLUS-NEXT: s_setpc_b64 s[30:31]
|
|
%roundeven = call <2 x double> @llvm.roundeven.v2f64(<2 x double> %x)
|
|
ret <2 x double> %roundeven
|
|
}
|
|
|
|
declare half @llvm.roundeven.f16(half) #0
|
|
declare <2 x half> @llvm.roundeven.v2f16(<2 x half>) #0
|
|
declare <4 x half> @llvm.roundeven.v4f16(<4 x half>) #0
|
|
|
|
declare float @llvm.roundeven.f32(float) #0
|
|
declare <2 x float> @llvm.roundeven.v2f32(<2 x float>) #0
|
|
declare <3 x float> @llvm.roundeven.v3f32(<3 x float>) #0
|
|
declare <4 x float> @llvm.roundeven.v4f32(<4 x float>) #0
|
|
|
|
declare double @llvm.roundeven.f64(double) #0
|
|
declare <2 x double> @llvm.roundeven.v2f64(<2 x double>) #0
|
|
|
|
declare half @llvm.fabs.f16(half) #0
|
|
declare float @llvm.fabs.f32(float) #0
|
|
|
|
attributes #0 = { nounwind readnone speculatable willreturn }
|
|
;; NOTE: These prefixes are unused and the list is autogenerated. Do not add tests below this line:
|
|
; GFX11: {{.*}}
|
|
; SDAG_GFX11: {{.*}}
|