Summary: The documentation at https://llvm.org/docs/AMDGPUUsage.html#memory-scopes states that these 'one-as' modifiers are more specific versions of the scopes that only apply to a specific address space. This doesn't make sense for fences which have no associated address space to use, and it's a more restrictive version the normal scope. This should not tbe the default behavior, but it is currently emitted in all cases except for sequentially consistent.
243 lines
15 KiB
Plaintext
243 lines
15 KiB
Plaintext
// RUN: %clang_cc1 -x hip %s -emit-llvm -o - -triple=amdgcn-amd-amdhsa \
|
|
// RUN: -fcuda-is-device -target-cpu gfx906 -fnative-half-type \
|
|
// RUN: -fnative-half-arguments-and-returns | FileCheck -check-prefixes=CHECK,SAFEIR %s
|
|
|
|
// RUN: %clang_cc1 -x hip %s -emit-llvm -o - -triple=amdgcn-amd-amdhsa \
|
|
// RUN: -fcuda-is-device -target-cpu gfx906 -fnative-half-type \
|
|
// RUN: -fnative-half-arguments-and-returns -munsafe-fp-atomics | FileCheck -check-prefixes=CHECK,UNSAFEIR %s
|
|
|
|
// RUN: %clang_cc1 -x hip %s -O3 -S -o - -triple=amdgcn-amd-amdhsa \
|
|
// RUN: -fcuda-is-device -target-cpu gfx1100 -fnative-half-type \
|
|
// RUN: -fnative-half-arguments-and-returns | FileCheck -check-prefix=SAFE %s
|
|
|
|
// RUN: %clang_cc1 -x hip %s -O3 -S -o - -triple=amdgcn-amd-amdhsa \
|
|
// RUN: -fcuda-is-device -target-cpu gfx940 -fnative-half-type \
|
|
// RUN: -fnative-half-arguments-and-returns -munsafe-fp-atomics \
|
|
// RUN: | FileCheck -check-prefix=UNSAFE %s
|
|
|
|
// REQUIRES: amdgpu-registered-target
|
|
|
|
#include "Inputs/cuda.h"
|
|
#include <stdatomic.h>
|
|
|
|
__global__ void ffp1(float *p) {
|
|
// CHECK-LABEL: @_Z4ffp1Pf
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 4{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 4{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 4{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 4{{$}}
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} syncscope("agent") monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE:[0-9]+]]{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} syncscope("workgroup") monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 4, !amdgpu.no.fine.grained.memory !{{[0-9]+}}, !amdgpu.ignore.denormal.mode !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 4, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 4, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 4, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE:[0-9]+]], !amdgpu.no.fine.grained.memory !{{[0-9]+}}, !amdgpu.ignore.denormal.mode !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
|
|
// SAFE: _Z4ffp1Pf
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
|
|
// UNSAFE: _Z4ffp1Pf
|
|
// UNSAFE: global_atomic_add_f32
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
|
|
__atomic_fetch_add(p, 1.0f, memory_order_relaxed);
|
|
__atomic_fetch_sub(p, 1.0f, memory_order_relaxed);
|
|
__atomic_fetch_max(p, 1.0f, memory_order_relaxed);
|
|
__atomic_fetch_min(p, 1.0f, memory_order_relaxed);
|
|
|
|
__hip_atomic_fetch_add(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_sub(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
__hip_atomic_fetch_max(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_min(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
}
|
|
|
|
__global__ void ffp2(double *p) {
|
|
// CHECK-LABEL: @_Z4ffp2Pd
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
|
|
// SAFE-LABEL: @_Z4ffp2Pd
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
|
|
// UNSAFE-LABEL: @_Z4ffp2Pd
|
|
// UNSAFE: global_atomic_add_f64
|
|
// UNSAFE: global_atomic_cmpswap_x2
|
|
// UNSAFE: global_atomic_max_f64
|
|
// UNSAFE: global_atomic_min_f64
|
|
// UNSAFE: global_atomic_max_f64
|
|
// UNSAFE: global_atomic_min_f64
|
|
__atomic_fetch_add(p, 1.0, memory_order_relaxed);
|
|
__atomic_fetch_sub(p, 1.0, memory_order_relaxed);
|
|
__atomic_fetch_max(p, 1.0, memory_order_relaxed);
|
|
__atomic_fetch_min(p, 1.0, memory_order_relaxed);
|
|
__hip_atomic_fetch_add(p, 1.0, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_sub(p, 1.0, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
__hip_atomic_fetch_max(p, 1.0, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_min(p, 1.0, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
}
|
|
|
|
// long double is the same as double for amdgcn.
|
|
__global__ void ffp3(long double *p) {
|
|
// CHECK-LABEL: @_Z4ffp3Pe
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 8{{$}}
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
|
|
// SAFE-LABEL: @_Z4ffp3Pe
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// SAFE: global_atomic_cmpswap_b64
|
|
// UNSAFE-LABEL: @_Z4ffp3Pe
|
|
// UNSAFE: global_atomic_cmpswap_x2
|
|
// UNSAFE: global_atomic_max_f64
|
|
// UNSAFE: global_atomic_min_f64
|
|
// UNSAFE: global_atomic_max_f64
|
|
// UNSAFE: global_atomic_min_f64
|
|
__atomic_fetch_add(p, 1.0L, memory_order_relaxed);
|
|
__atomic_fetch_sub(p, 1.0L, memory_order_relaxed);
|
|
__atomic_fetch_max(p, 1.0L, memory_order_relaxed);
|
|
__atomic_fetch_min(p, 1.0L, memory_order_relaxed);
|
|
__hip_atomic_fetch_add(p, 1.0L, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_sub(p, 1.0L, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
__hip_atomic_fetch_max(p, 1.0L, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_min(p, 1.0L, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
}
|
|
|
|
__device__ double ffp4(double *p, float f) {
|
|
// CHECK-LABEL: @_Z4ffp4Pdf
|
|
// CHECK: fpext contract float {{.*}} to double
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8{{$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
__atomic_fetch_sub(p, f, memory_order_relaxed);
|
|
return __hip_atomic_fetch_sub(p, f, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
}
|
|
|
|
__device__ double ffp5(double *p, int i) {
|
|
// CHECK-LABEL: @_Z4ffp5Pdi
|
|
// CHECK: sitofp i32 {{.*}} to double
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8{{$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
__atomic_fetch_sub(p, i, memory_order_relaxed);
|
|
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 8, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
return __hip_atomic_fetch_sub(p, i, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
}
|
|
|
|
__global__ void ffp6(_Float16 *p) {
|
|
// CHECK-LABEL: @_Z4ffp6PDF16
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 2{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 2{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 2{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 2{{$}}
|
|
// SAFEIR: atomicrmw fadd ptr {{.*}} syncscope("agent") monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fsub ptr {{.*}} syncscope("workgroup") monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// SAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 2, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 2, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} monotonic, align 2, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} monotonic, align 2, !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fadd ptr {{.*}} monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fsub ptr {{.*}} monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmax ptr {{.*}} syncscope("agent") monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
// UNSAFEIR: atomicrmw fmin ptr {{.*}} syncscope("workgroup") monotonic, align 2, !noalias.addrspace ![[$NO_PRIVATE]], !amdgpu.no.fine.grained.memory !{{[0-9]+$}}
|
|
|
|
// SAFE: _Z4ffp6PDF16
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
// SAFE: global_atomic_cmpswap
|
|
|
|
// UNSAFE: _Z4ffp6PDF16
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
// UNSAFE: global_atomic_cmpswap
|
|
__atomic_fetch_add(p, 1.0, memory_order_relaxed);
|
|
__atomic_fetch_sub(p, 1.0, memory_order_relaxed);
|
|
__atomic_fetch_max(p, 1.0, memory_order_relaxed);
|
|
__atomic_fetch_min(p, 1.0, memory_order_relaxed);
|
|
|
|
__hip_atomic_fetch_add(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_sub(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
__hip_atomic_fetch_max(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_AGENT);
|
|
__hip_atomic_fetch_min(p, 1.0f, memory_order_relaxed, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
}
|
|
|
|
// CHECK-LABEL: @_Z12test_cmpxchgPiii
|
|
// CHECK: cmpxchg ptr %{{.+}}, i32 %{{.+}}, i32 %{{.+}} acquire acquire, align 4{{$}}
|
|
// CHECK: cmpxchg weak ptr %{{.+}}, i32 %{{.+}}, i32 %{{.+}} acquire acquire, align 4{{$}}
|
|
// CHECK: cmpxchg ptr %{{.+}}, i32 %{{.+}}, i32 %{{.+}} syncscope("workgroup") monotonic monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
// CHECK: cmpxchg weak ptr %{{.+}}, i32 %{{.+}}, i32 %{{.+}} syncscope("workgroup") monotonic monotonic, align 4, !noalias.addrspace ![[$NO_PRIVATE]]{{$}}
|
|
__device__ int test_cmpxchg(int *ptr, int cmp, int desired) {
|
|
bool flag = __atomic_compare_exchange(ptr, &cmp, &desired, 0, memory_order_acquire, memory_order_acquire);
|
|
flag = __atomic_compare_exchange_n(ptr, &cmp, desired, 1, memory_order_acquire, memory_order_acquire);
|
|
flag = __hip_atomic_compare_exchange_strong(ptr, &cmp, desired, __ATOMIC_RELAXED, __ATOMIC_RELAXED, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
flag = __hip_atomic_compare_exchange_weak(ptr, &cmp, desired, __ATOMIC_RELAXED, __ATOMIC_RELAXED, __HIP_MEMORY_SCOPE_WORKGROUP);
|
|
return flag;
|
|
}
|
|
|
|
// SAFEIR: ![[$NO_PRIVATE]] = !{i32 5, i32 6}
|
|
// UNSAFEIR: ![[$NO_PRIVATE]] = !{i32 5, i32 6}
|