| 1 | // RUN: %clang_cc1 < %s -triple armv5e-none-linux-gnueabi -emit-llvm -O1 | FileCheck %s |
| 2 | |
| 3 | // FIXME: This file should not be checking -O1 output. |
| 4 | // Ie, it is testing many IR optimizer passes as part of front-end verification. |
| 5 | |
| 6 | enum memory_order { |
| 7 | memory_order_relaxed, memory_order_consume, memory_order_acquire, |
| 8 | memory_order_release, memory_order_acq_rel, memory_order_seq_cst |
| 9 | }; |
| 10 | |
| 11 | int *test_c11_atomic_fetch_add_int_ptr(_Atomic(int *) *p) { |
| 12 | // CHECK: test_c11_atomic_fetch_add_int_ptr |
| 13 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 12, i32 5) |
| 14 | return __c11_atomic_fetch_add(p, 3, memory_order_seq_cst); |
| 15 | } |
| 16 | |
| 17 | int *test_c11_atomic_fetch_sub_int_ptr(_Atomic(int *) *p) { |
| 18 | // CHECK: test_c11_atomic_fetch_sub_int_ptr |
| 19 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 20, i32 5) |
| 20 | return __c11_atomic_fetch_sub(p, 5, memory_order_seq_cst); |
| 21 | } |
| 22 | |
| 23 | int test_c11_atomic_fetch_add_int(_Atomic(int) *p) { |
| 24 | // CHECK: test_c11_atomic_fetch_add_int |
| 25 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 3, i32 5) |
| 26 | return __c11_atomic_fetch_add(p, 3, memory_order_seq_cst); |
| 27 | } |
| 28 | |
| 29 | int test_c11_atomic_fetch_sub_int(_Atomic(int) *p) { |
| 30 | // CHECK: test_c11_atomic_fetch_sub_int |
| 31 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 5, i32 5) |
| 32 | return __c11_atomic_fetch_sub(p, 5, memory_order_seq_cst); |
| 33 | } |
| 34 | |
| 35 | int *fp2a(int **p) { |
| 36 | // CHECK: @fp2a |
| 37 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 4, i32 0) |
| 38 | // Note, the GNU builtins do not multiply by sizeof(T)! |
| 39 | return __atomic_fetch_sub(p, 4, memory_order_relaxed); |
| 40 | } |
| 41 | |
| 42 | int test_atomic_fetch_add(int *p) { |
| 43 | // CHECK: test_atomic_fetch_add |
| 44 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 45 | return __atomic_fetch_add(p, 55, memory_order_seq_cst); |
| 46 | } |
| 47 | |
| 48 | int test_atomic_fetch_sub(int *p) { |
| 49 | // CHECK: test_atomic_fetch_sub |
| 50 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 51 | return __atomic_fetch_sub(p, 55, memory_order_seq_cst); |
| 52 | } |
| 53 | |
| 54 | int test_atomic_fetch_and(int *p) { |
| 55 | // CHECK: test_atomic_fetch_and |
| 56 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_and_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 57 | return __atomic_fetch_and(p, 55, memory_order_seq_cst); |
| 58 | } |
| 59 | |
| 60 | int test_atomic_fetch_or(int *p) { |
| 61 | // CHECK: test_atomic_fetch_or |
| 62 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_or_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 63 | return __atomic_fetch_or(p, 55, memory_order_seq_cst); |
| 64 | } |
| 65 | |
| 66 | int test_atomic_fetch_xor(int *p) { |
| 67 | // CHECK: test_atomic_fetch_xor |
| 68 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_xor_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 69 | return __atomic_fetch_xor(p, 55, memory_order_seq_cst); |
| 70 | } |
| 71 | |
| 72 | int test_atomic_fetch_nand(int *p) { |
| 73 | // CHECK: test_atomic_fetch_nand |
| 74 | // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_nand_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 75 | return __atomic_fetch_nand(p, 55, memory_order_seq_cst); |
| 76 | } |
| 77 | |
| 78 | int test_atomic_add_fetch(int *p) { |
| 79 | // CHECK: test_atomic_add_fetch |
| 80 | // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 81 | // CHECK: {{%[^ ]*}} = add i32 [[CALL]], 55 |
| 82 | return __atomic_add_fetch(p, 55, memory_order_seq_cst); |
| 83 | } |
| 84 | |
| 85 | int test_atomic_sub_fetch(int *p) { |
| 86 | // CHECK: test_atomic_sub_fetch |
| 87 | // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 88 | // CHECK: {{%[^ ]*}} = add i32 [[CALL]], -55 |
| 89 | return __atomic_sub_fetch(p, 55, memory_order_seq_cst); |
| 90 | } |
| 91 | |
| 92 | int test_atomic_and_fetch(int *p) { |
| 93 | // CHECK: test_atomic_and_fetch |
| 94 | // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_and_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 95 | // CHECK: {{%[^ ]*}} = and i32 [[CALL]], 55 |
| 96 | return __atomic_and_fetch(p, 55, memory_order_seq_cst); |
| 97 | } |
| 98 | |
| 99 | int test_atomic_or_fetch(int *p) { |
| 100 | // CHECK: test_atomic_or_fetch |
| 101 | // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_or_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 102 | // CHECK: {{%[^ ]*}} = or i32 [[CALL]], 55 |
| 103 | return __atomic_or_fetch(p, 55, memory_order_seq_cst); |
| 104 | } |
| 105 | |
| 106 | int test_atomic_xor_fetch(int *p) { |
| 107 | // CHECK: test_atomic_xor_fetch |
| 108 | // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_xor_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 109 | // CHECK: {{%[^ ]*}} = xor i32 [[CALL]], 55 |
| 110 | return __atomic_xor_fetch(p, 55, memory_order_seq_cst); |
| 111 | } |
| 112 | |
| 113 | int test_atomic_nand_fetch(int *p) { |
| 114 | // CHECK: test_atomic_nand_fetch |
| 115 | // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_nand_4(i8* {{%[0-9]+}}, i32 55, i32 5) |
| 116 | // FIXME: We should not be checking optimized IR. It changes independently of clang. |
| 117 | // FIXME-CHECK: [[AND:%[^ ]*]] = and i32 [[CALL]], 55 |
| 118 | // FIXME-CHECK: {{%[^ ]*}} = xor i32 [[AND]], -1 |
| 119 | return __atomic_nand_fetch(p, 55, memory_order_seq_cst); |
| 120 | } |
| 121 | |