In order to enable the LLVM frontend to better analyze buffer operations (and to potentially enable more precise analyses on the backend), define versions of the raw and structured buffer intrinsics that use `ptr addrspace(8)` instead of `<4 x i32>` to represent their rsrc arguments. The new intrinsics are named by replacing `buffer.` with `buffer.ptr`. One advantage to these intrinsic definitions is that, instead of specifying that a buffer load/store will read/write some memory, we can indicate that the memory read or written will be based on the pointer argument. This means that, for example, a read from a `noalias` buffer can be pulled out of a loop that is modifying a distinct buffer. In the future, we will define custom PseudoSourceValues that will allow us to package up the (buffer, index, offset) triples that buffer intrinsics contain and allow for more precise backend analysis. This work also enables creating address space 7, which represents manipulation of raw buffers using native LLVM load and store instructions. Where tests simply used a buffer intrinsic while testing some other code path (such as the tests for VGPR spills), they have been updated to use the new intrinsic form. Tests that are "about" buffer intrinsics (for instance, those that ensure that they codegen as expected) have been duplicated, either within existing files or into new ones. Depends on D145441 Reviewed By: arsenm, #amdgpu Differential Revision: https://reviews.llvm.org/D147547
135 lines
4.8 KiB
LLVM
135 lines
4.8 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -march=amdgcn -mcpu=verde -verify-machineinstrs | FileCheck --check-prefix=SI %s
|
|
; RUN: llc < %s -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs | FileCheck --check-prefix=FLAT %s
|
|
|
|
define amdgpu_ps float @uniform_kill(float %a, i32 %b, float %c) {
|
|
; SI-LABEL: uniform_kill:
|
|
; SI: ; %bb.0: ; %entry
|
|
; SI-NEXT: v_cvt_i32_f32_e32 v0, v0
|
|
; SI-NEXT: s_mov_b64 s[0:1], exec
|
|
; SI-NEXT: s_mov_b64 s[2:3], -1
|
|
; SI-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; SI-NEXT: v_and_b32_e32 v0, 1, v0
|
|
; SI-NEXT: v_cmp_eq_u32_e32 vcc, 1, v0
|
|
; SI-NEXT: s_and_saveexec_b64 s[4:5], vcc
|
|
; SI-NEXT: ; %bb.1: ; %if1
|
|
; SI-NEXT: s_xor_b64 s[2:3], exec, -1
|
|
; SI-NEXT: ; %bb.2: ; %endif1
|
|
; SI-NEXT: s_or_b64 exec, exec, s[4:5]
|
|
; SI-NEXT: s_wqm_b64 s[4:5], s[2:3]
|
|
; SI-NEXT: s_xor_b64 s[4:5], s[4:5], exec
|
|
; SI-NEXT: s_andn2_b64 s[0:1], s[0:1], s[4:5]
|
|
; SI-NEXT: s_cbranch_scc0 .LBB0_6
|
|
; SI-NEXT: ; %bb.3: ; %endif1
|
|
; SI-NEXT: s_and_b64 exec, exec, s[0:1]
|
|
; SI-NEXT: v_mov_b32_e32 v0, 0
|
|
; SI-NEXT: s_and_saveexec_b64 s[0:1], s[2:3]
|
|
; SI-NEXT: s_cbranch_execz .LBB0_5
|
|
; SI-NEXT: ; %bb.4: ; %if2
|
|
; SI-NEXT: s_mov_b32 s3, 0
|
|
; SI-NEXT: s_load_dwordx2 s[4:5], s[2:3], 0x0
|
|
; SI-NEXT: v_add_f32_e32 v0, 1.0, v2
|
|
; SI-NEXT: v_cvt_i32_f32_e32 v0, v0
|
|
; SI-NEXT: s_waitcnt lgkmcnt(0)
|
|
; SI-NEXT: s_mov_b32 s6, s4
|
|
; SI-NEXT: s_mov_b32 s7, s5
|
|
; SI-NEXT: buffer_atomic_swap v0, off, s[4:7], 0 offset:4 glc
|
|
; SI-NEXT: s_waitcnt vmcnt(0) expcnt(0)
|
|
; SI-NEXT: v_cvt_f32_i32_e32 v0, v0
|
|
; SI-NEXT: .LBB0_5: ; %endif2
|
|
; SI-NEXT: s_or_b64 exec, exec, s[0:1]
|
|
; SI-NEXT: s_branch .LBB0_7
|
|
; SI-NEXT: .LBB0_6:
|
|
; SI-NEXT: s_mov_b64 exec, 0
|
|
; SI-NEXT: exp null off, off, off, off done vm
|
|
; SI-NEXT: s_endpgm
|
|
; SI-NEXT: .LBB0_7:
|
|
;
|
|
; FLAT-LABEL: uniform_kill:
|
|
; FLAT: ; %bb.0: ; %entry
|
|
; FLAT-NEXT: v_cvt_i32_f32_e32 v0, v0
|
|
; FLAT-NEXT: s_mov_b64 s[0:1], exec
|
|
; FLAT-NEXT: s_mov_b64 s[2:3], -1
|
|
; FLAT-NEXT: v_or_b32_e32 v0, v1, v0
|
|
; FLAT-NEXT: v_and_b32_e32 v0, 1, v0
|
|
; FLAT-NEXT: v_cmp_eq_u32_e32 vcc, 1, v0
|
|
; FLAT-NEXT: s_and_saveexec_b64 s[4:5], vcc
|
|
; FLAT-NEXT: ; %bb.1: ; %if1
|
|
; FLAT-NEXT: s_xor_b64 s[2:3], exec, -1
|
|
; FLAT-NEXT: ; %bb.2: ; %endif1
|
|
; FLAT-NEXT: s_or_b64 exec, exec, s[4:5]
|
|
; FLAT-NEXT: s_wqm_b64 s[4:5], s[2:3]
|
|
; FLAT-NEXT: s_xor_b64 s[4:5], s[4:5], exec
|
|
; FLAT-NEXT: s_andn2_b64 s[0:1], s[0:1], s[4:5]
|
|
; FLAT-NEXT: s_cbranch_scc0 .LBB0_6
|
|
; FLAT-NEXT: ; %bb.3: ; %endif1
|
|
; FLAT-NEXT: s_and_b64 exec, exec, s[0:1]
|
|
; FLAT-NEXT: v_mov_b32_e32 v0, 0
|
|
; FLAT-NEXT: s_and_saveexec_b64 s[0:1], s[2:3]
|
|
; FLAT-NEXT: s_cbranch_execz .LBB0_5
|
|
; FLAT-NEXT: ; %bb.4: ; %if2
|
|
; FLAT-NEXT: s_mov_b32 s3, 0
|
|
; FLAT-NEXT: s_load_dwordx2 s[4:5], s[2:3], 0x0
|
|
; FLAT-NEXT: v_add_f32_e32 v0, 1.0, v2
|
|
; FLAT-NEXT: v_cvt_i32_f32_e32 v0, v0
|
|
; FLAT-NEXT: s_waitcnt lgkmcnt(0)
|
|
; FLAT-NEXT: s_mov_b32 s6, s4
|
|
; FLAT-NEXT: s_mov_b32 s7, s5
|
|
; FLAT-NEXT: buffer_atomic_swap v0, off, s[4:7], 0 offset:4 glc
|
|
; FLAT-NEXT: s_waitcnt vmcnt(0)
|
|
; FLAT-NEXT: v_cvt_f32_i32_e32 v0, v0
|
|
; FLAT-NEXT: .LBB0_5: ; %endif2
|
|
; FLAT-NEXT: s_or_b64 exec, exec, s[0:1]
|
|
; FLAT-NEXT: s_branch .LBB0_7
|
|
; FLAT-NEXT: .LBB0_6:
|
|
; FLAT-NEXT: s_mov_b64 exec, 0
|
|
; FLAT-NEXT: exp null off, off, off, off done vm
|
|
; FLAT-NEXT: s_endpgm
|
|
; FLAT-NEXT: .LBB0_7:
|
|
entry:
|
|
%.1 = fptosi float %a to i32
|
|
%.2 = or i32 %b, %.1
|
|
%.3 = and i32 %.2, 1
|
|
%.not = icmp eq i32 %.3, 0
|
|
br i1 %.not, label %endif1, label %if1
|
|
|
|
if1:
|
|
br i1 false, label %if3, label %endif1
|
|
|
|
if3:
|
|
br label %endif1
|
|
|
|
endif1:
|
|
%.0 = phi i1 [ false, %if3 ], [ false, %if1 ], [ true, %entry ]
|
|
%.4 = call i1 @llvm.amdgcn.wqm.vote(i1 %.0)
|
|
; This kill must be uniformly executed
|
|
call void @llvm.amdgcn.kill(i1 %.4)
|
|
%.test0 = fadd nsz arcp float %c, 1.0
|
|
%.test1 = fptosi float %.test0 to i32
|
|
br i1 %.0, label %if2, label %endif2
|
|
|
|
if2:
|
|
%.5 = getelementptr inbounds ptr addrspace(8), ptr addrspace(6) undef, i32 31, !amdgpu.uniform !0
|
|
%.6 = load ptr addrspace(8), ptr addrspace(6) %.5, align 16, !invariant.load !0
|
|
%.7 = call i32 @llvm.amdgcn.raw.ptr.buffer.atomic.swap.i32(i32 %.test1, ptr addrspace(8) %.6, i32 4, i32 0, i32 0)
|
|
%.8 = sitofp i32 %.7 to float
|
|
br label %endif2
|
|
|
|
endif2:
|
|
%.9 = phi float [ %.8, %if2 ], [ 0.0, %endif1 ]
|
|
ret float %.9
|
|
}
|
|
|
|
|
|
declare i32 @llvm.amdgcn.raw.ptr.buffer.atomic.swap.i32(i32, ptr addrspace(8), i32, i32, i32 immarg) #2
|
|
declare i1 @llvm.amdgcn.wqm.vote(i1) #3
|
|
declare void @llvm.amdgcn.kill(i1) #4
|
|
declare float @llvm.amdgcn.wqm.f32(float) #1
|
|
|
|
attributes #1 = { nounwind readnone speculatable willreturn }
|
|
attributes #2 = { nounwind willreturn memory(argmem: readwrite) }
|
|
attributes #3 = { convergent nounwind readnone willreturn }
|
|
attributes #4 = { nounwind }
|
|
|
|
!0 = !{}
|