1// RUN: %clang_cc1 -triple arm-linux-gnueabi -emit-llvm %s -o - | FileCheck %s -check-prefix=ARM 2// RUN: %clang_cc1 -triple powerpc-linux-gnu -emit-llvm %s -o - | FileCheck %s -check-prefix=PPC32 3// RUN: %clang_cc1 -triple powerpc64-linux-gnu -emit-llvm %s -o - | FileCheck %s -check-prefix=PPC64 4// RUN: %clang_cc1 -triple mipsel-linux-gnu -emit-llvm %s -o - | FileCheck %s -check-prefix=MIPS32 5// RUN: %clang_cc1 -triple mips64el-linux-gnu -emit-llvm %s -o - | FileCheck %s -check-prefix=MIPS64 6// RUN: %clang_cc1 -triple sparc-unknown-eabi -emit-llvm %s -o - | FileCheck %s -check-prefix=SPARC 7 8unsigned char c1, c2; 9unsigned short s1, s2; 10unsigned int i1, i2; 11unsigned long long ll1, ll2; 12unsigned char a1[100], a2[100]; 13 14enum memory_order { 15 memory_order_relaxed, 16 memory_order_consume, 17 memory_order_acquire, 18 memory_order_release, 19 memory_order_acq_rel, 20 memory_order_seq_cst 21}; 22 23void test1(void) { 24 (void)__atomic_load(&c1, &c2, memory_order_seq_cst); 25 (void)__atomic_store(&c1, &c2, memory_order_seq_cst); 26 (void)__atomic_load(&s1, &s2, memory_order_seq_cst); 27 (void)__atomic_store(&s1, &s2, memory_order_seq_cst); 28 (void)__atomic_load(&i1, &i2, memory_order_seq_cst); 29 (void)__atomic_store(&i1, &i2, memory_order_seq_cst); 30 (void)__atomic_load(&ll1, &ll2, memory_order_seq_cst); 31 (void)__atomic_store(&ll1, &ll2, memory_order_seq_cst); 32 (void)__atomic_load(&a1, &a2, memory_order_seq_cst); 33 (void)__atomic_store(&a1, &a2, memory_order_seq_cst); 34 35// ARM-LABEL: define{{.*}} void @test1 36// ARM: = call{{.*}} zeroext i8 @__atomic_load_1(i8* @c1 37// ARM: call{{.*}} void @__atomic_store_1(i8* @c1, i8 zeroext 38// ARM: = call{{.*}} zeroext i16 @__atomic_load_2(i8* bitcast (i16* @s1 to i8*) 39// ARM: call{{.*}} void @__atomic_store_2(i8* bitcast (i16* @s1 to i8*), i16 zeroext 40// ARM: = call{{.*}} i32 @__atomic_load_4(i8* bitcast (i32* @i1 to i8*) 41// ARM: call{{.*}} void @__atomic_store_4(i8* bitcast (i32* @i1 to i8*), i32 42// ARM: = call{{.*}} i64 @__atomic_load_8(i8* bitcast (i64* @ll1 to i8*) 43// ARM: call{{.*}} void @__atomic_store_8(i8* bitcast (i64* @ll1 to i8*), i64 44// ARM: call{{.*}} void @__atomic_load(i32 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 45// ARM: call{{.*}} void @__atomic_store(i32 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 46 47// PPC32-LABEL: define void @test1 48// PPC32: = load atomic i8, i8* @c1 seq_cst 49// PPC32: store atomic i8 {{.*}}, i8* @c1 seq_cst 50// PPC32: = load atomic i16, i16* @s1 seq_cst 51// PPC32: store atomic i16 {{.*}}, i16* @s1 seq_cst 52// PPC32: = load atomic i32, i32* @i1 seq_cst 53// PPC32: store atomic i32 {{.*}}, i32* @i1 seq_cst 54// PPC32: = call i64 @__atomic_load_8(i8* bitcast (i64* @ll1 to i8*) 55// PPC32: call void @__atomic_store_8(i8* bitcast (i64* @ll1 to i8*), i64 56// PPC32: call void @__atomic_load(i32 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 57// PPC32: call void @__atomic_store(i32 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 58 59// PPC64-LABEL: define void @test1 60// PPC64: = load atomic i8, i8* @c1 seq_cst 61// PPC64: store atomic i8 {{.*}}, i8* @c1 seq_cst 62// PPC64: = load atomic i16, i16* @s1 seq_cst 63// PPC64: store atomic i16 {{.*}}, i16* @s1 seq_cst 64// PPC64: = load atomic i32, i32* @i1 seq_cst 65// PPC64: store atomic i32 {{.*}}, i32* @i1 seq_cst 66// PPC64: = load atomic i64, i64* @ll1 seq_cst 67// PPC64: store atomic i64 {{.*}}, i64* @ll1 seq_cst 68// PPC64: call void @__atomic_load(i64 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 69// PPC64: call void @__atomic_store(i64 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 70 71// MIPS32-LABEL: define void @test1 72// MIPS32: = load atomic i8, i8* @c1 seq_cst 73// MIPS32: store atomic i8 {{.*}}, i8* @c1 seq_cst 74// MIPS32: = load atomic i16, i16* @s1 seq_cst 75// MIPS32: store atomic i16 {{.*}}, i16* @s1 seq_cst 76// MIPS32: = load atomic i32, i32* @i1 seq_cst 77// MIPS32: store atomic i32 {{.*}}, i32* @i1 seq_cst 78// MIPS32: call i64 @__atomic_load_8(i8* bitcast (i64* @ll1 to i8*) 79// MIPS32: call void @__atomic_store_8(i8* bitcast (i64* @ll1 to i8*), i64 80// MIPS32: call void @__atomic_load(i32 signext 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 81// MIPS32: call void @__atomic_store(i32 signext 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 82 83// MIPS64-LABEL: define void @test1 84// MIPS64: = load atomic i8, i8* @c1 seq_cst 85// MIPS64: store atomic i8 {{.*}}, i8* @c1 seq_cst 86// MIPS64: = load atomic i16, i16* @s1 seq_cst 87// MIPS64: store atomic i16 {{.*}}, i16* @s1 seq_cst 88// MIPS64: = load atomic i32, i32* @i1 seq_cst 89// MIPS64: store atomic i32 {{.*}}, i32* @i1 seq_cst 90// MIPS64: = load atomic i64, i64* @ll1 seq_cst 91// MIPS64: store atomic i64 {{.*}}, i64* @ll1 seq_cst 92// MIPS64: call void @__atomic_load(i64 zeroext 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0) 93// MIPS64: call void @__atomic_store(i64 zeroext 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 94 95// SPARC-LABEL: define void @test1 96// SPARC: = load atomic i8, i8* @c1 seq_cst 97// SPARC: store atomic i8 {{.*}}, i8* @c1 seq_cst 98// SPARC: = load atomic i16, i16* @s1 seq_cst 99// SPARC: store atomic i16 {{.*}}, i16* @s1 seq_cst 100// SPARC: = load atomic i32, i32* @i1 seq_cst 101// SPARC: store atomic i32 {{.*}}, i32* @i1 seq_cst 102// SPARC: = load atomic i64, i64* @ll1 seq_cst 103// SPARC: store atomic i64 {{.*}}, i64* @ll1 seq_cst 104// SPARC: call void @__atomic_load(i32 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 105// SPARC: call void @__atomic_store(i32 100, i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a1, i32 0, i32 0), i8* getelementptr inbounds ([100 x i8], [100 x i8]* @a2, i32 0, i32 0) 106} 107