1*67e74705SXin Li // RUN: %clang_cc1 -triple mips-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=O32
2*67e74705SXin Li // RUN: %clang_cc1 -triple mipsel-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=O32
3*67e74705SXin Li // RUN: %clang_cc1 -triple mips64-unknown-linux -o - -emit-llvm -target-abi n32 %s | FileCheck %s -check-prefix=ALL -check-prefix=N32 -check-prefix=NEW
4*67e74705SXin Li // RUN: %clang_cc1 -triple mips64-unknown-linux -o - -emit-llvm -target-abi n32 %s | FileCheck %s -check-prefix=ALL -check-prefix=N32 -check-prefix=NEW
5*67e74705SXin Li // RUN: %clang_cc1 -triple mips64-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=N64 -check-prefix=NEW
6*67e74705SXin Li // RUN: %clang_cc1 -triple mips64el-unknown-linux -o - -emit-llvm %s | FileCheck %s -check-prefix=ALL -check-prefix=N64 -check-prefix=NEW
7*67e74705SXin Li
8*67e74705SXin Li #include <stdarg.h>
9*67e74705SXin Li
10*67e74705SXin Li typedef int v4i32 __attribute__ ((__vector_size__ (16)));
11*67e74705SXin Li
test_i32(char * fmt,...)12*67e74705SXin Li int test_i32(char *fmt, ...) {
13*67e74705SXin Li va_list va;
14*67e74705SXin Li
15*67e74705SXin Li va_start(va, fmt);
16*67e74705SXin Li int v = va_arg(va, int);
17*67e74705SXin Li va_end(va);
18*67e74705SXin Li
19*67e74705SXin Li return v;
20*67e74705SXin Li }
21*67e74705SXin Li
22*67e74705SXin Li // ALL-LABEL: define i32 @test_i32(i8*{{.*}} %fmt, ...)
23*67e74705SXin Li //
24*67e74705SXin Li // O32: %va = alloca i8*, align [[PTRALIGN:4]]
25*67e74705SXin Li // N32: %va = alloca i8*, align [[PTRALIGN:4]]
26*67e74705SXin Li // N64: %va = alloca i8*, align [[PTRALIGN:8]]
27*67e74705SXin Li // ALL: [[V:%.*]] = alloca i32, align 4
28*67e74705SXin Li // NEW: [[PROMOTION_TEMP:%.*]] = alloca i32, align 4
29*67e74705SXin Li //
30*67e74705SXin Li // ALL: [[VA:%.+]] = bitcast i8** %va to i8*
31*67e74705SXin Li // ALL: call void @llvm.va_start(i8* [[VA]])
32*67e74705SXin Li // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]]
33*67e74705SXin Li // O32: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T:i32]] [[CHUNKSIZE:4]]
34*67e74705SXin Li // NEW: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T:i32|i64]] [[CHUNKSIZE:8]]
35*67e74705SXin Li //
36*67e74705SXin Li // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]]
37*67e74705SXin Li //
38*67e74705SXin Li // O32: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to [[CHUNK_T:i32]]*
39*67e74705SXin Li // O32: [[ARG:%.+]] = load i32, i32* [[AP_CAST]], align [[CHUNKALIGN:4]]
40*67e74705SXin Li //
41*67e74705SXin Li // N32: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to [[CHUNK_T:i64]]*
42*67e74705SXin Li // N32: [[TMP:%.+]] = load i64, i64* [[AP_CAST]], align [[CHUNKALIGN:8]]
43*67e74705SXin Li // N64: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to [[CHUNK_T:i64]]*
44*67e74705SXin Li // N64: [[TMP:%.+]] = load i64, i64* [[AP_CAST]], align [[CHUNKALIGN:8]]
45*67e74705SXin Li // NEW: [[TMP2:%.+]] = trunc i64 [[TMP]] to i32
46*67e74705SXin Li // NEW: store i32 [[TMP2]], i32* [[PROMOTION_TEMP]], align 4
47*67e74705SXin Li // NEW: [[ARG:%.+]] = load i32, i32* [[PROMOTION_TEMP]], align 4
48*67e74705SXin Li // ALL: store i32 [[ARG]], i32* [[V]], align 4
49*67e74705SXin Li //
50*67e74705SXin Li // ALL: [[VA1:%.+]] = bitcast i8** %va to i8*
51*67e74705SXin Li // ALL: call void @llvm.va_end(i8* [[VA1]])
52*67e74705SXin Li // ALL: }
53*67e74705SXin Li
test_i64(char * fmt,...)54*67e74705SXin Li long long test_i64(char *fmt, ...) {
55*67e74705SXin Li va_list va;
56*67e74705SXin Li
57*67e74705SXin Li va_start(va, fmt);
58*67e74705SXin Li long long v = va_arg(va, long long);
59*67e74705SXin Li va_end(va);
60*67e74705SXin Li
61*67e74705SXin Li return v;
62*67e74705SXin Li }
63*67e74705SXin Li
64*67e74705SXin Li // ALL-LABEL: define i64 @test_i64(i8*{{.*}} %fmt, ...)
65*67e74705SXin Li //
66*67e74705SXin Li // ALL: %va = alloca i8*, align [[PTRALIGN]]
67*67e74705SXin Li // ALL: [[VA:%.+]] = bitcast i8** %va to i8*
68*67e74705SXin Li // ALL: call void @llvm.va_start(i8* [[VA]])
69*67e74705SXin Li // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]]
70*67e74705SXin Li //
71*67e74705SXin Li // i64 is 8-byte aligned, while this is within O32's stack alignment there's no
72*67e74705SXin Li // guarantee that the offset is still 8-byte aligned after earlier reads.
73*67e74705SXin Li // O32: [[TMP1:%.+]] = ptrtoint i8* [[AP_CUR]] to i32
74*67e74705SXin Li // O32: [[TMP2:%.+]] = add i32 [[TMP1]], 7
75*67e74705SXin Li // O32: [[TMP3:%.+]] = and i32 [[TMP2]], -8
76*67e74705SXin Li // O32: [[AP_CUR:%.+]] = inttoptr i32 [[TMP3]] to i8*
77*67e74705SXin Li //
78*67e74705SXin Li // ALL: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T]] 8
79*67e74705SXin Li // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]]
80*67e74705SXin Li //
81*67e74705SXin Li // ALL: [[AP_CAST:%.*]] = bitcast i8* [[AP_CUR]] to i64*
82*67e74705SXin Li // ALL: [[ARG:%.+]] = load i64, i64* [[AP_CAST]], align 8
83*67e74705SXin Li //
84*67e74705SXin Li // ALL: [[VA1:%.+]] = bitcast i8** %va to i8*
85*67e74705SXin Li // ALL: call void @llvm.va_end(i8* [[VA1]])
86*67e74705SXin Li // ALL: }
87*67e74705SXin Li
test_ptr(char * fmt,...)88*67e74705SXin Li char *test_ptr(char *fmt, ...) {
89*67e74705SXin Li va_list va;
90*67e74705SXin Li
91*67e74705SXin Li va_start(va, fmt);
92*67e74705SXin Li char *v = va_arg(va, char *);
93*67e74705SXin Li va_end(va);
94*67e74705SXin Li
95*67e74705SXin Li return v;
96*67e74705SXin Li }
97*67e74705SXin Li
98*67e74705SXin Li // ALL-LABEL: define i8* @test_ptr(i8*{{.*}} %fmt, ...)
99*67e74705SXin Li //
100*67e74705SXin Li // ALL: %va = alloca i8*, align [[PTRALIGN]]
101*67e74705SXin Li // ALL: [[V:%.*]] = alloca i8*, align [[PTRALIGN]]
102*67e74705SXin Li // N32: [[AP_CAST:%.+]] = alloca i8*, align 4
103*67e74705SXin Li // ALL: [[VA:%.+]] = bitcast i8** %va to i8*
104*67e74705SXin Li // ALL: call void @llvm.va_start(i8* [[VA]])
105*67e74705SXin Li // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]]
106*67e74705SXin Li // ALL: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T]] [[CHUNKSIZE]]
107*67e74705SXin Li // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]]
108*67e74705SXin Li //
109*67e74705SXin Li // When the chunk size matches the pointer size, this is easy.
110*67e74705SXin Li // O32: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to i8**
111*67e74705SXin Li // N64: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to i8**
112*67e74705SXin Li // Otherwise we need a promotion temporary.
113*67e74705SXin Li // N32: [[TMP1:%.+]] = bitcast i8* [[AP_CUR]] to i64*
114*67e74705SXin Li // N32: [[TMP2:%.+]] = load i64, i64* [[TMP1]], align 8
115*67e74705SXin Li // N32: [[TMP3:%.+]] = trunc i64 [[TMP2]] to i32
116*67e74705SXin Li // N32: [[PTR:%.+]] = inttoptr i32 [[TMP3]] to i8*
117*67e74705SXin Li // N32: store i8* [[PTR]], i8** [[AP_CAST]], align 4
118*67e74705SXin Li //
119*67e74705SXin Li // ALL: [[ARG:%.+]] = load i8*, i8** [[AP_CAST]], align [[PTRALIGN]]
120*67e74705SXin Li // ALL: store i8* [[ARG]], i8** [[V]], align [[PTRALIGN]]
121*67e74705SXin Li //
122*67e74705SXin Li // ALL: [[VA1:%.+]] = bitcast i8** %va to i8*
123*67e74705SXin Li // ALL: call void @llvm.va_end(i8* [[VA1]])
124*67e74705SXin Li // ALL: }
125*67e74705SXin Li
test_v4i32(char * fmt,...)126*67e74705SXin Li int test_v4i32(char *fmt, ...) {
127*67e74705SXin Li va_list va;
128*67e74705SXin Li
129*67e74705SXin Li va_start(va, fmt);
130*67e74705SXin Li v4i32 v = va_arg(va, v4i32);
131*67e74705SXin Li va_end(va);
132*67e74705SXin Li
133*67e74705SXin Li return v[0];
134*67e74705SXin Li }
135*67e74705SXin Li
136*67e74705SXin Li // ALL-LABEL: define i32 @test_v4i32(i8*{{.*}} %fmt, ...)
137*67e74705SXin Li //
138*67e74705SXin Li // ALL: %va = alloca i8*, align [[PTRALIGN]]
139*67e74705SXin Li // ALL: [[V]] = alloca <4 x i32>, align 16
140*67e74705SXin Li // ALL: [[VA1:%.+]] = bitcast i8** %va to i8*
141*67e74705SXin Li // ALL: call void @llvm.va_start(i8* [[VA1]])
142*67e74705SXin Li // ALL: [[AP_CUR:%.+]] = load i8*, i8** %va, align [[PTRALIGN]]
143*67e74705SXin Li //
144*67e74705SXin Li // Vectors are 16-byte aligned, however the O32 ABI has a maximum alignment of
145*67e74705SXin Li // 8-bytes since the base of the stack is 8-byte aligned.
146*67e74705SXin Li // O32: [[TMP1:%.+]] = ptrtoint i8* [[AP_CUR]] to i32
147*67e74705SXin Li // O32: [[TMP2:%.+]] = add i32 [[TMP1]], 7
148*67e74705SXin Li // O32: [[TMP3:%.+]] = and i32 [[TMP2]], -8
149*67e74705SXin Li // O32: [[AP_CUR:%.+]] = inttoptr i32 [[TMP3]] to i8*
150*67e74705SXin Li //
151*67e74705SXin Li // NEW: [[TMP1:%.+]] = ptrtoint i8* [[AP_CUR]] to [[INTPTR_T]]
152*67e74705SXin Li // NEW: [[TMP2:%.+]] = add [[INTPTR_T]] [[TMP1]], 15
153*67e74705SXin Li // NEW: [[TMP3:%.+]] = and [[INTPTR_T]] [[TMP2]], -16
154*67e74705SXin Li // NEW: [[AP_CUR:%.+]] = inttoptr [[INTPTR_T]] [[TMP3]] to i8*
155*67e74705SXin Li //
156*67e74705SXin Li // ALL: [[AP_NEXT:%.+]] = getelementptr inbounds i8, i8* [[AP_CUR]], [[INTPTR_T]] 16
157*67e74705SXin Li // ALL: store i8* [[AP_NEXT]], i8** %va, align [[PTRALIGN]]
158*67e74705SXin Li //
159*67e74705SXin Li // ALL: [[AP_CAST:%.+]] = bitcast i8* [[AP_CUR]] to <4 x i32>*
160*67e74705SXin Li // O32: [[ARG:%.+]] = load <4 x i32>, <4 x i32>* [[AP_CAST]], align 8
161*67e74705SXin Li // N64: [[ARG:%.+]] = load <4 x i32>, <4 x i32>* [[AP_CAST]], align 16
162*67e74705SXin Li // ALL: store <4 x i32> [[ARG]], <4 x i32>* [[V]], align 16
163*67e74705SXin Li //
164*67e74705SXin Li // ALL: [[VA1:%.+]] = bitcast i8** %va to i8*
165*67e74705SXin Li // ALL: call void @llvm.va_end(i8* [[VA1]])
166*67e74705SXin Li // ALL: [[VECEXT:%.+]] = extractelement <4 x i32> {{.*}}, i32 0
167*67e74705SXin Li // ALL: ret i32 [[VECEXT]]
168*67e74705SXin Li // ALL: }
169