Summary: We did a great job getting this wrong: - We messed up which LLVM IR types to use for arguments and return values. The optimized libcalls use integer types for values. Clang attempted to use the IR type which corresponds to the value passed in instead of using an appropriately sized integer type. This would result in violations of the ABI for, as an example, floating point types. - We didn't bother recording the result of the atomic libcall in the destination memory. Instead, call the functions with arguments matching the type of the libcall prototype's parameters. This fixes PR20780. Differential Revision: http://reviews.llvm.org/D5098 llvm-svn: 216714
524 lines
18 KiB
C
524 lines
18 KiB
C
// RUN: %clang_cc1 %s -emit-llvm -o - -triple=i686-apple-darwin9 | FileCheck %s
|
|
|
|
// Also test serialization of atomic operations here, to avoid duplicating the
|
|
// test.
|
|
// RUN: %clang_cc1 %s -emit-pch -o %t -triple=i686-apple-darwin9
|
|
// RUN: %clang_cc1 %s -include-pch %t -triple=i686-apple-darwin9 -emit-llvm -o - | FileCheck %s
|
|
#ifndef ALREADY_INCLUDED
|
|
#define ALREADY_INCLUDED
|
|
|
|
// Basic IRGen tests for __c11_atomic_* and GNU __atomic_*
|
|
|
|
typedef enum memory_order {
|
|
memory_order_relaxed, memory_order_consume, memory_order_acquire,
|
|
memory_order_release, memory_order_acq_rel, memory_order_seq_cst
|
|
} memory_order;
|
|
|
|
int fi1(_Atomic(int) *i) {
|
|
// CHECK-LABEL: @fi1
|
|
// CHECK: load atomic i32* {{.*}} seq_cst
|
|
return __c11_atomic_load(i, memory_order_seq_cst);
|
|
}
|
|
|
|
int fi1a(int *i) {
|
|
// CHECK-LABEL: @fi1a
|
|
// CHECK: load atomic i32* {{.*}} seq_cst
|
|
int v;
|
|
__atomic_load(i, &v, memory_order_seq_cst);
|
|
return v;
|
|
}
|
|
|
|
int fi1b(int *i) {
|
|
// CHECK-LABEL: @fi1b
|
|
// CHECK: load atomic i32* {{.*}} seq_cst
|
|
return __atomic_load_n(i, memory_order_seq_cst);
|
|
}
|
|
|
|
void fi2(_Atomic(int) *i) {
|
|
// CHECK-LABEL: @fi2
|
|
// CHECK: store atomic i32 {{.*}} seq_cst
|
|
__c11_atomic_store(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
void fi2a(int *i) {
|
|
// CHECK-LABEL: @fi2a
|
|
// CHECK: store atomic i32 {{.*}} seq_cst
|
|
int v = 1;
|
|
__atomic_store(i, &v, memory_order_seq_cst);
|
|
}
|
|
|
|
void fi2b(int *i) {
|
|
// CHECK-LABEL: @fi2b
|
|
// CHECK: store atomic i32 {{.*}} seq_cst
|
|
__atomic_store_n(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
int fi3(_Atomic(int) *i) {
|
|
// CHECK-LABEL: @fi3
|
|
// CHECK: atomicrmw and
|
|
// CHECK-NOT: and
|
|
return __c11_atomic_fetch_and(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
int fi3a(int *i) {
|
|
// CHECK-LABEL: @fi3a
|
|
// CHECK: atomicrmw xor
|
|
// CHECK-NOT: xor
|
|
return __atomic_fetch_xor(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
int fi3b(int *i) {
|
|
// CHECK-LABEL: @fi3b
|
|
// CHECK: atomicrmw add
|
|
// CHECK: add
|
|
return __atomic_add_fetch(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
int fi3c(int *i) {
|
|
// CHECK-LABEL: @fi3c
|
|
// CHECK: atomicrmw nand
|
|
// CHECK-NOT: and
|
|
return __atomic_fetch_nand(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
int fi3d(int *i) {
|
|
// CHECK-LABEL: @fi3d
|
|
// CHECK: atomicrmw nand
|
|
// CHECK: and
|
|
// CHECK: xor
|
|
return __atomic_nand_fetch(i, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
_Bool fi4(_Atomic(int) *i) {
|
|
// CHECK-LABEL: @fi4
|
|
// CHECK: [[PAIR:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]]
|
|
// CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 0
|
|
// CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 1
|
|
// CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]]
|
|
// CHECK: store i32 [[OLD]]
|
|
int cmp = 0;
|
|
return __c11_atomic_compare_exchange_strong(i, &cmp, 1, memory_order_acquire, memory_order_acquire);
|
|
}
|
|
|
|
_Bool fi4a(int *i) {
|
|
// CHECK-LABEL: @fi4
|
|
// CHECK: [[PAIR:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]]
|
|
// CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 0
|
|
// CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 1
|
|
// CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]]
|
|
// CHECK: store i32 [[OLD]]
|
|
int cmp = 0;
|
|
int desired = 1;
|
|
return __atomic_compare_exchange(i, &cmp, &desired, 0, memory_order_acquire, memory_order_acquire);
|
|
}
|
|
|
|
_Bool fi4b(int *i) {
|
|
// CHECK-LABEL: @fi4
|
|
// CHECK: [[PAIR:%[.0-9A-Z_a-z]+]] = cmpxchg weak i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]]
|
|
// CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 0
|
|
// CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 1
|
|
// CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]]
|
|
// CHECK: store i32 [[OLD]]
|
|
int cmp = 0;
|
|
return __atomic_compare_exchange_n(i, &cmp, 1, 1, memory_order_acquire, memory_order_acquire);
|
|
}
|
|
|
|
float ff1(_Atomic(float) *d) {
|
|
// CHECK-LABEL: @ff1
|
|
// CHECK: load atomic i32* {{.*}} monotonic
|
|
return __c11_atomic_load(d, memory_order_relaxed);
|
|
}
|
|
|
|
void ff2(_Atomic(float) *d) {
|
|
// CHECK-LABEL: @ff2
|
|
// CHECK: store atomic i32 {{.*}} release
|
|
__c11_atomic_store(d, 1, memory_order_release);
|
|
}
|
|
|
|
float ff3(_Atomic(float) *d) {
|
|
return __c11_atomic_exchange(d, 2, memory_order_seq_cst);
|
|
}
|
|
|
|
struct S {
|
|
double x;
|
|
};
|
|
|
|
struct S fd1(struct S *a) {
|
|
// CHECK-LABEL: @fd1
|
|
// CHECK: [[RETVAL:%.*]] = alloca %struct.S, align 4
|
|
// CHECK: [[RET:%.*]] = alloca %struct.S, align 4
|
|
// CHECK: [[CALL:%.*]] = call i64 @__atomic_load_8(
|
|
// CHECK: [[CAST:%.*]] = bitcast %struct.S* [[RET]] to i64*
|
|
// CHECK: store i64 [[CALL]], i64* [[CAST]], align 4
|
|
struct S ret;
|
|
__atomic_load(a, &ret, memory_order_seq_cst);
|
|
return ret;
|
|
}
|
|
|
|
void fd2(struct S *a, struct S *b) {
|
|
// CHECK-LABEL: @fd2
|
|
// CHECK: [[A_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: [[B_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: store %struct.S* %a, %struct.S** [[A_ADDR]], align 4
|
|
// CHECK-NEXT: store %struct.S* %b, %struct.S** [[B_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_A_PTR:%.*]] = load %struct.S** [[A_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_B_PTR:%.*]] = load %struct.S** [[B_ADDR]], align 4
|
|
// CHECK-NEXT: [[COERCED_A:%.*]] = bitcast %struct.S* [[LOAD_A_PTR]] to i8*
|
|
// CHECK-NEXT: [[COERCED_B:%.*]] = bitcast %struct.S* [[LOAD_B_PTR]] to i64*
|
|
// CHECK-NEXT: [[LOAD_B:%.*]] = load i64* [[COERCED_B]], align 4
|
|
// CHECK-NEXT: call void @__atomic_store_8(i8* [[COERCED_A]], i64 [[LOAD_B]],
|
|
// CHECK-NEXT: ret void
|
|
__atomic_store(a, b, memory_order_seq_cst);
|
|
}
|
|
|
|
void fd3(struct S *a, struct S *b, struct S *c) {
|
|
// CHECK-LABEL: @fd3
|
|
// CHECK: [[A_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: [[B_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: [[C_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: store %struct.S* %a, %struct.S** [[A_ADDR]], align 4
|
|
// CHECK-NEXT: store %struct.S* %b, %struct.S** [[B_ADDR]], align 4
|
|
// CHECK-NEXT: store %struct.S* %c, %struct.S** [[C_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_A_PTR:%.*]] = load %struct.S** [[A_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_B_PTR:%.*]] = load %struct.S** [[B_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_C_PTR:%.*]] = load %struct.S** [[C_ADDR]], align 4
|
|
// CHECK-NEXT: [[COERCED_A:%.*]] = bitcast %struct.S* [[LOAD_A_PTR]] to i8*
|
|
// CHECK-NEXT: [[COERCED_B:%.*]] = bitcast %struct.S* [[LOAD_B_PTR]] to i64*
|
|
// CHECK-NEXT: [[LOAD_B:%.*]] = load i64* [[COERCED_B]], align 4
|
|
// CHECK-NEXT: [[CALL:%.*]] = call i64 @__atomic_exchange_8(i8* [[COERCED_A]], i64 [[LOAD_B]],
|
|
// CHECK-NEXT: [[COERCED_C:%.*]] = bitcast %struct.S* [[LOAD_C_PTR]] to i64*
|
|
// CHECK-NEXT: store i64 [[CALL]], i64* [[COERCED_C]], align 4
|
|
|
|
__atomic_exchange(a, b, c, memory_order_seq_cst);
|
|
}
|
|
|
|
_Bool fd4(struct S *a, struct S *b, struct S *c) {
|
|
// CHECK-LABEL: @fd4
|
|
// CHECK: [[A_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: [[B_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK-NEXT: [[C_ADDR:%.*]] = alloca %struct.S*, align 4
|
|
// CHECK: store %struct.S* %a, %struct.S** [[A_ADDR]], align 4
|
|
// CHECK-NEXT: store %struct.S* %b, %struct.S** [[B_ADDR]], align 4
|
|
// CHECK-NEXT: store %struct.S* %c, %struct.S** [[C_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_A_PTR:%.*]] = load %struct.S** [[A_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_B_PTR:%.*]] = load %struct.S** [[B_ADDR]], align 4
|
|
// CHECK-NEXT: [[LOAD_C_PTR:%.*]] = load %struct.S** [[C_ADDR]], align 4
|
|
// CHECK-NEXT: [[COERCED_A:%.*]] = bitcast %struct.S* [[LOAD_A_PTR]] to i8*
|
|
// CHECK-NEXT: [[COERCED_B:%.*]] = bitcast %struct.S* [[LOAD_B_PTR]] to i8*
|
|
// CHECK-NEXT: [[COERCED_C:%.*]] = bitcast %struct.S* [[LOAD_C_PTR]] to i64*
|
|
// CHECK-NEXT: [[LOAD_C:%.*]] = load i64* [[COERCED_C]], align 4
|
|
// CHECK-NEXT: [[CALL:%.*]] = call zeroext i1 @__atomic_compare_exchange_8(i8* [[COERCED_A]], i8* [[COERCED_B]], i64 [[LOAD_C]]
|
|
// CHECK-NEXT: ret i1 [[CALL]]
|
|
return __atomic_compare_exchange(a, b, c, 1, 5, 5);
|
|
}
|
|
|
|
int* fp1(_Atomic(int*) *p) {
|
|
// CHECK-LABEL: @fp1
|
|
// CHECK: load atomic i32* {{.*}} seq_cst
|
|
return __c11_atomic_load(p, memory_order_seq_cst);
|
|
}
|
|
|
|
int* fp2(_Atomic(int*) *p) {
|
|
// CHECK-LABEL: @fp2
|
|
// CHECK: store i32 4
|
|
// CHECK: atomicrmw add {{.*}} monotonic
|
|
return __c11_atomic_fetch_add(p, 1, memory_order_relaxed);
|
|
}
|
|
|
|
int *fp2a(int **p) {
|
|
// CHECK-LABEL: @fp2a
|
|
// CHECK: store i32 4
|
|
// CHECK: atomicrmw sub {{.*}} monotonic
|
|
// Note, the GNU builtins do not multiply by sizeof(T)!
|
|
return __atomic_fetch_sub(p, 4, memory_order_relaxed);
|
|
}
|
|
|
|
_Complex float fc(_Atomic(_Complex float) *c) {
|
|
// CHECK-LABEL: @fc
|
|
// CHECK: atomicrmw xchg i64*
|
|
return __c11_atomic_exchange(c, 2, memory_order_seq_cst);
|
|
}
|
|
|
|
typedef struct X { int x; } X;
|
|
X fs(_Atomic(X) *c) {
|
|
// CHECK-LABEL: @fs
|
|
// CHECK: atomicrmw xchg i32*
|
|
return __c11_atomic_exchange(c, (X){2}, memory_order_seq_cst);
|
|
}
|
|
|
|
X fsa(X *c, X *d) {
|
|
// CHECK-LABEL: @fsa
|
|
// CHECK: atomicrmw xchg i32*
|
|
X ret;
|
|
__atomic_exchange(c, d, &ret, memory_order_seq_cst);
|
|
return ret;
|
|
}
|
|
|
|
_Bool fsb(_Bool *c) {
|
|
// CHECK-LABEL: @fsb
|
|
// CHECK: atomicrmw xchg i8*
|
|
return __atomic_exchange_n(c, 1, memory_order_seq_cst);
|
|
}
|
|
|
|
char flag1;
|
|
volatile char flag2;
|
|
void test_and_set() {
|
|
// CHECK: atomicrmw xchg i8* @flag1, i8 1 seq_cst
|
|
__atomic_test_and_set(&flag1, memory_order_seq_cst);
|
|
// CHECK: atomicrmw volatile xchg i8* @flag2, i8 1 acquire
|
|
__atomic_test_and_set(&flag2, memory_order_acquire);
|
|
// CHECK: store atomic volatile i8 0, i8* @flag2 release
|
|
__atomic_clear(&flag2, memory_order_release);
|
|
// CHECK: store atomic i8 0, i8* @flag1 seq_cst
|
|
__atomic_clear(&flag1, memory_order_seq_cst);
|
|
}
|
|
|
|
struct Sixteen {
|
|
char c[16];
|
|
} sixteen;
|
|
struct Seventeen {
|
|
char c[17];
|
|
} seventeen;
|
|
|
|
int lock_free(struct Incomplete *incomplete) {
|
|
// CHECK-LABEL: @lock_free
|
|
|
|
// CHECK: call i32 @__atomic_is_lock_free(i32 3, i8* null)
|
|
__c11_atomic_is_lock_free(3);
|
|
|
|
// CHECK: call i32 @__atomic_is_lock_free(i32 16, i8* {{.*}}@sixteen{{.*}})
|
|
__atomic_is_lock_free(16, &sixteen);
|
|
|
|
// CHECK: call i32 @__atomic_is_lock_free(i32 17, i8* {{.*}}@seventeen{{.*}})
|
|
__atomic_is_lock_free(17, &seventeen);
|
|
|
|
// CHECK: call i32 @__atomic_is_lock_free(i32 4, {{.*}})
|
|
__atomic_is_lock_free(4, incomplete);
|
|
|
|
char cs[20];
|
|
// CHECK: call i32 @__atomic_is_lock_free(i32 4, {{.*}})
|
|
__atomic_is_lock_free(4, cs+1);
|
|
|
|
// CHECK-NOT: call
|
|
__atomic_always_lock_free(3, 0);
|
|
__atomic_always_lock_free(16, 0);
|
|
__atomic_always_lock_free(17, 0);
|
|
__atomic_always_lock_free(16, &sixteen);
|
|
__atomic_always_lock_free(17, &seventeen);
|
|
|
|
int n;
|
|
__atomic_is_lock_free(4, &n);
|
|
|
|
// CHECK: ret i32 1
|
|
return __c11_atomic_is_lock_free(sizeof(_Atomic(int)));
|
|
}
|
|
|
|
// Tests for atomic operations on big values. These should call the functions
|
|
// defined here:
|
|
// http://gcc.gnu.org/wiki/Atomic/GCCMM/LIbrary#The_Library_interface
|
|
|
|
struct foo {
|
|
int big[128];
|
|
};
|
|
struct bar {
|
|
char c[3];
|
|
};
|
|
|
|
struct bar smallThing, thing1, thing2;
|
|
struct foo bigThing;
|
|
_Atomic(struct foo) bigAtomic;
|
|
|
|
void structAtomicStore() {
|
|
// CHECK-LABEL: @structAtomicStore
|
|
struct foo f = {0};
|
|
struct bar b = {0};
|
|
__atomic_store(&smallThing, &b, 5);
|
|
// CHECK: call void @__atomic_store(i32 3, i8* {{.*}} @smallThing
|
|
|
|
__atomic_store(&bigThing, &f, 5);
|
|
// CHECK: call void @__atomic_store(i32 512, i8* {{.*}} @bigThing
|
|
}
|
|
void structAtomicLoad() {
|
|
// CHECK-LABEL: @structAtomicLoad
|
|
struct bar b;
|
|
__atomic_load(&smallThing, &b, 5);
|
|
// CHECK: call void @__atomic_load(i32 3, i8* {{.*}} @smallThing
|
|
|
|
struct foo f = {0};
|
|
__atomic_load(&bigThing, &f, 5);
|
|
// CHECK: call void @__atomic_load(i32 512, i8* {{.*}} @bigThing
|
|
}
|
|
struct foo structAtomicExchange() {
|
|
// CHECK-LABEL: @structAtomicExchange
|
|
struct foo f = {0};
|
|
struct foo old;
|
|
__atomic_exchange(&f, &bigThing, &old, 5);
|
|
// CHECK: call void @__atomic_exchange(i32 512, {{.*}}, i8* bitcast ({{.*}} @bigThing to i8*),
|
|
|
|
return __c11_atomic_exchange(&bigAtomic, f, 5);
|
|
// CHECK: call void @__atomic_exchange(i32 512, i8* bitcast ({{.*}} @bigAtomic to i8*),
|
|
}
|
|
int structAtomicCmpExchange() {
|
|
// CHECK-LABEL: @structAtomicCmpExchange
|
|
_Bool x = __atomic_compare_exchange(&smallThing, &thing1, &thing2, 1, 5, 5);
|
|
// CHECK: call zeroext i1 @__atomic_compare_exchange(i32 3, {{.*}} @smallThing{{.*}} @thing1{{.*}} @thing2
|
|
|
|
struct foo f = {0};
|
|
struct foo g = {0};
|
|
g.big[12] = 12;
|
|
return x & __c11_atomic_compare_exchange_strong(&bigAtomic, &f, g, 5, 5);
|
|
// CHECK: call zeroext i1 @__atomic_compare_exchange(i32 512, i8* bitcast ({{.*}} @bigAtomic to i8*),
|
|
}
|
|
|
|
// Check that no atomic operations are used in any initialisation of _Atomic
|
|
// types.
|
|
_Atomic(int) atomic_init_i = 42;
|
|
|
|
// CHECK-LABEL: @atomic_init_foo
|
|
void atomic_init_foo()
|
|
{
|
|
// CHECK-NOT: }
|
|
// CHECK-NOT: atomic
|
|
// CHECK: store
|
|
_Atomic(int) j = 12;
|
|
|
|
// CHECK-NOT: }
|
|
// CHECK-NOT: atomic
|
|
// CHECK: store
|
|
__c11_atomic_init(&j, 42);
|
|
|
|
// CHECK-NOT: atomic
|
|
// CHECK: }
|
|
}
|
|
|
|
// CHECK-LABEL: @failureOrder
|
|
void failureOrder(_Atomic(int) *ptr, int *ptr2) {
|
|
__c11_atomic_compare_exchange_strong(ptr, ptr2, 43, memory_order_acquire, memory_order_relaxed);
|
|
// CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} acquire monotonic
|
|
|
|
__c11_atomic_compare_exchange_weak(ptr, ptr2, 43, memory_order_seq_cst, memory_order_acquire);
|
|
// CHECK: cmpxchg weak i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} seq_cst acquire
|
|
|
|
// Unknown ordering: conservatively pick strongest valid option (for now!).
|
|
__atomic_compare_exchange(ptr2, ptr2, ptr2, 0, memory_order_acq_rel, *ptr2);
|
|
// CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} acq_rel acquire
|
|
|
|
// Undefined behaviour: don't really care what that last ordering is so leave
|
|
// it out:
|
|
__atomic_compare_exchange_n(ptr2, ptr2, 43, 1, memory_order_seq_cst, 42);
|
|
// CHECK: cmpxchg weak i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} seq_cst
|
|
}
|
|
|
|
// CHECK-LABEL: @generalFailureOrder
|
|
void generalFailureOrder(_Atomic(int) *ptr, int *ptr2, int success, int fail) {
|
|
__c11_atomic_compare_exchange_strong(ptr, ptr2, 42, success, fail);
|
|
// CHECK: switch i32 {{.*}}, label %[[MONOTONIC:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: i32 1, label %[[ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 2, label %[[ACQUIRE]]
|
|
// CHECK-NEXT: i32 3, label %[[RELEASE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 4, label %[[ACQREL:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 5, label %[[SEQCST:[0-9a-zA-Z._]+]]
|
|
|
|
// CHECK: [[MONOTONIC]]
|
|
// CHECK: switch {{.*}}, label %[[MONOTONIC_MONOTONIC:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: ]
|
|
|
|
// CHECK: [[ACQUIRE]]
|
|
// CHECK: switch {{.*}}, label %[[ACQUIRE_MONOTONIC:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: i32 1, label %[[ACQUIRE_ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 2, label %[[ACQUIRE_ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: ]
|
|
|
|
// CHECK: [[RELEASE]]
|
|
// CHECK: switch {{.*}}, label %[[RELEASE_MONOTONIC:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: ]
|
|
|
|
// CHECK: [[ACQREL]]
|
|
// CHECK: switch {{.*}}, label %[[ACQREL_MONOTONIC:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: i32 1, label %[[ACQREL_ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 2, label %[[ACQREL_ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: ]
|
|
|
|
// CHECK: [[SEQCST]]
|
|
// CHECK: switch {{.*}}, label %[[SEQCST_MONOTONIC:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: i32 1, label %[[SEQCST_ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 2, label %[[SEQCST_ACQUIRE:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: i32 5, label %[[SEQCST_SEQCST:[0-9a-zA-Z._]+]]
|
|
// CHECK-NEXT: ]
|
|
|
|
// CHECK: [[MONOTONIC_MONOTONIC]]
|
|
// CHECK: cmpxchg {{.*}} monotonic monotonic
|
|
// CHECK: br
|
|
|
|
// CHECK: [[ACQUIRE_MONOTONIC]]
|
|
// CHECK: cmpxchg {{.*}} acquire monotonic
|
|
// CHECK: br
|
|
|
|
// CHECK: [[ACQUIRE_ACQUIRE]]
|
|
// CHECK: cmpxchg {{.*}} acquire acquire
|
|
// CHECK: br
|
|
|
|
// CHECK: [[ACQREL_MONOTONIC]]
|
|
// CHECK: cmpxchg {{.*}} acq_rel monotonic
|
|
// CHECK: br
|
|
|
|
// CHECK: [[ACQREL_ACQUIRE]]
|
|
// CHECK: cmpxchg {{.*}} acq_rel acquire
|
|
// CHECK: br
|
|
|
|
// CHECK: [[SEQCST_MONOTONIC]]
|
|
// CHECK: cmpxchg {{.*}} seq_cst monotonic
|
|
// CHECK: br
|
|
|
|
// CHECK: [[SEQCST_ACQUIRE]]
|
|
// CHECK: cmpxchg {{.*}} seq_cst acquire
|
|
// CHECK: br
|
|
|
|
// CHECK: [[SEQCST_SEQCST]]
|
|
// CHECK: cmpxchg {{.*}} seq_cst seq_cst
|
|
// CHECK: br
|
|
}
|
|
|
|
void generalWeakness(int *ptr, int *ptr2, _Bool weak) {
|
|
__atomic_compare_exchange_n(ptr, ptr2, 42, weak, memory_order_seq_cst, memory_order_seq_cst);
|
|
// CHECK: switch i1 {{.*}}, label %[[WEAK:[0-9a-zA-Z._]+]] [
|
|
// CHECK-NEXT: i1 false, label %[[STRONG:[0-9a-zA-Z._]+]]
|
|
|
|
// CHECK: [[STRONG]]
|
|
// CHECK-NOT: br
|
|
// CHECK: cmpxchg {{.*}} seq_cst seq_cst
|
|
// CHECK: br
|
|
|
|
// CHECK: [[WEAK]]
|
|
// CHECK-NOT: br
|
|
// CHECK: cmpxchg weak {{.*}} seq_cst seq_cst
|
|
// CHECK: br
|
|
}
|
|
|
|
// Having checked the flow in the previous two cases, we'll trust clang to
|
|
// combine them sanely.
|
|
void EMIT_ALL_THE_THINGS(int *ptr, int *ptr2, int new, _Bool weak, int success, int fail) {
|
|
__atomic_compare_exchange(ptr, ptr2, &new, weak, success, fail);
|
|
|
|
// CHECK: = cmpxchg {{.*}} monotonic monotonic
|
|
// CHECK: = cmpxchg weak {{.*}} monotonic monotonic
|
|
// CHECK: = cmpxchg {{.*}} acquire monotonic
|
|
// CHECK: = cmpxchg {{.*}} acquire acquire
|
|
// CHECK: = cmpxchg weak {{.*}} acquire monotonic
|
|
// CHECK: = cmpxchg weak {{.*}} acquire acquire
|
|
// CHECK: = cmpxchg {{.*}} release monotonic
|
|
// CHECK: = cmpxchg weak {{.*}} release monotonic
|
|
// CHECK: = cmpxchg {{.*}} acq_rel monotonic
|
|
// CHECK: = cmpxchg {{.*}} acq_rel acquire
|
|
// CHECK: = cmpxchg weak {{.*}} acq_rel monotonic
|
|
// CHECK: = cmpxchg weak {{.*}} acq_rel acquire
|
|
// CHECK: = cmpxchg {{.*}} seq_cst monotonic
|
|
// CHECK: = cmpxchg {{.*}} seq_cst acquire
|
|
// CHECK: = cmpxchg {{.*}} seq_cst seq_cst
|
|
// CHECK: = cmpxchg weak {{.*}} seq_cst monotonic
|
|
// CHECK: = cmpxchg weak {{.*}} seq_cst acquire
|
|
// CHECK: = cmpxchg weak {{.*}} seq_cst seq_cst
|
|
}
|
|
|
|
#endif
|