atom-call-reg-indirect-foldedreload64.ll
2.99 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
; RUN: llc < %s -mtriple=x86_64-linux-gnu -mcpu=atom | \
; RUN: FileCheck --check-prefix=ATOM %s
; RUN: llc < %s -mtriple=x86_64-linux-gnu -mcpu=core2 | \
; RUN: FileCheck --check-prefix=CORE2 %s
; ATOM: callq *{{%[a-z]+[0-9]*}}
; CORE2: callq *funcp
;
; Original source code built with clang -S -emit-llvm -m64 test64.c:
; int a, b, c, d, e, f, g, h, i, j, k, l, m, n;
; extern int (*funcp)(int, int, int, int, int, int,
; int, int, int, int, int, int,
; int, int);
; extern int sum;
;
; void func()
; {
; sum = 0;
; for( i = a; i < b; ++i )
; {
; sum += (*funcp)(a, i, i*2, i/b, c, d, e, f, g, h, j, k, l, n);
; }
; }
;
@sum = external global i32
@a = common global i32 0, align 4
@i = common global i32 0, align 4
@b = common global i32 0, align 4
@funcp = external global i32 (i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32)*
@c = common global i32 0, align 4
@d = common global i32 0, align 4
@e = common global i32 0, align 4
@f = common global i32 0, align 4
@g = common global i32 0, align 4
@h = common global i32 0, align 4
@j = common global i32 0, align 4
@k = common global i32 0, align 4
@l = common global i32 0, align 4
@n = common global i32 0, align 4
@m = common global i32 0, align 4
define void @func() #0 {
entry:
store i32 0, i32* @sum, align 4
%0 = load i32, i32* @a, align 4
store i32 %0, i32* @i, align 4
br label %for.cond
for.cond: ; preds = %for.inc, %entry
%1 = load i32, i32* @i, align 4
%2 = load i32, i32* @b, align 4
%cmp = icmp slt i32 %1, %2
br i1 %cmp, label %for.body, label %for.end
for.body: ; preds = %for.cond
%3 = load i32 (i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32)*, i32 (i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32)** @funcp, align 8
%4 = load i32, i32* @a, align 4
%5 = load i32, i32* @i, align 4
%6 = load i32, i32* @i, align 4
%mul = mul nsw i32 %6, 2
%7 = load i32, i32* @i, align 4
%8 = load i32, i32* @b, align 4
%div = sdiv i32 %7, %8
%9 = load i32, i32* @c, align 4
%10 = load i32, i32* @d, align 4
%11 = load i32, i32* @e, align 4
%12 = load i32, i32* @f, align 4
%13 = load i32, i32* @g, align 4
%14 = load i32, i32* @h, align 4
%15 = load i32, i32* @j, align 4
%16 = load i32, i32* @k, align 4
%17 = load i32, i32* @l, align 4
%18 = load i32, i32* @n, align 4
%call = call i32 %3(i32 %4, i32 %5, i32 %mul, i32 %div, i32 %9, i32 %10, i32 %11, i32 %12, i32 %13, i32 %14, i32 %15, i32 %16, i32 %17, i32 %18)
%19 = load i32, i32* @sum, align 4
%add = add nsw i32 %19, %call
store i32 %add, i32* @sum, align 4
br label %for.inc
for.inc: ; preds = %for.body
%20 = load i32, i32* @i, align 4
%inc = add nsw i32 %20, 1
store i32 %inc, i32* @i, align 4
br label %for.cond
for.end: ; preds = %for.cond
ret void
}