1; RUN: opt -basicaa -gvn -S < %s | FileCheck %s
2
3target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64"
4target triple = "x86_64-apple-macosx10.7.0"
5
6@x = common global i32 0, align 4
7@y = common global i32 0, align 4
8
9; GVN across unordered store (allowed)
10define i32 @test1() nounwind uwtable ssp {
11; CHECK: test1
12; CHECK: add i32 %x, %x
13entry:
14  %x = load i32* @y
15  store atomic i32 %x, i32* @x unordered, align 4
16  %y = load i32* @y
17  %z = add i32 %x, %y
18  ret i32 %z
19}
20
21; GVN across seq_cst store (allowed in theory; not implemented ATM)
22define i32 @test2() nounwind uwtable ssp {
23; CHECK: test2
24; CHECK: add i32 %x, %y
25entry:
26  %x = load i32* @y
27  store atomic i32 %x, i32* @x seq_cst, align 4
28  %y = load i32* @y
29  %z = add i32 %x, %y
30  ret i32 %z
31}
32
33; GVN across unordered load (allowed)
34define i32 @test3() nounwind uwtable ssp {
35; CHECK: test3
36; CHECK: add i32 %x, %x
37entry:
38  %x = load i32* @y
39  %y = load atomic i32* @x unordered, align 4
40  %z = load i32* @y
41  %a = add i32 %x, %z
42  %b = add i32 %y, %a
43  ret i32 %b
44}
45
46; GVN across acquire load (load after atomic load must not be removed)
47define i32 @test4() nounwind uwtable ssp {
48; CHECK: test4
49; CHECK: load atomic i32* @x
50; CHECK: load i32* @y
51entry:
52  %x = load i32* @y
53  %y = load atomic i32* @x seq_cst, align 4
54  %x2 = load i32* @y
55  %x3 = add i32 %x, %x2
56  %y2 = add i32 %y, %x3
57  ret i32 %y2
58}
59
60; GVN load to unordered load (allowed)
61define i32 @test5() nounwind uwtable ssp {
62; CHECK: test5
63; CHECK: add i32 %x, %x
64entry:
65  %x = load atomic i32* @x unordered, align 4
66  %y = load i32* @x
67  %z = add i32 %x, %y
68  ret i32 %z
69}
70
71; GVN unordered load to load (unordered load must not be removed)
72define i32 @test6() nounwind uwtable ssp {
73; CHECK: test6
74; CHECK: load atomic i32* @x unordered
75entry:
76  %x = load i32* @x
77  %x2 = load atomic i32* @x unordered, align 4
78  %x3 = add i32 %x, %x2
79  ret i32 %x3
80}
81