1 // RUN: %clang_cc1 -verify -fopenmp -x c++ -triple x86_64-apple-darwin10 -emit-llvm %s -o - | FileCheck %s 2 // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple x86_64-apple-darwin10 -emit-pch -o %t %s 3 // RUN: %clang_cc1 -fopenmp -x c++ -triple x86_64-apple-darwin10 -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck %s 4 // RUN: %clang_cc1 -verify -fopenmp -x c++ -std=c++11 -DLAMBDA -triple x86_64-apple-darwin10 -emit-llvm %s -o - | FileCheck -check-prefix=LAMBDA %s 5 // RUN: %clang_cc1 -verify -fopenmp -x c++ -fblocks -DBLOCKS -triple x86_64-apple-darwin10 -emit-llvm %s -o - | FileCheck -check-prefix=BLOCKS %s 6 // expected-no-diagnostics 7 #ifndef HEADER 8 #define HEADER 9 10 struct SS { 11 int a; 12 int b : 4; 13 int &c; 14 SS(int &d) : a(0), b(0), c(d) { 15 #pragma omp parallel 16 #pragma omp for lastprivate(a, b, c) 17 for (int i = 0; i < 2; ++i) 18 #ifdef LAMBDA 19 [&]() { 20 ++this->a, --b, (this)->c /= 1; 21 #pragma omp parallel 22 #pragma omp for lastprivate(a, b, c) 23 for (int i = 0; i < 2; ++i) 24 ++(this)->a, --b, this->c /= 1; 25 }(); 26 #elif defined(BLOCKS) 27 ^{ 28 ++a; 29 --this->b; 30 (this)->c /= 1; 31 #pragma omp parallel 32 #pragma omp for lastprivate(a, b, c) 33 for (int i = 0; i < 2; ++i) 34 ++(this)->a, --b, this->c /= 1; 35 }(); 36 #else 37 ++this->a, --b, c /= 1; 38 #endif 39 #pragma omp for 40 for (a = 0; a < 2; ++a) 41 #ifdef LAMBDA 42 [&]() { 43 --this->a, ++b, (this)->c *= 2; 44 #pragma omp parallel 45 #pragma omp for lastprivate(b) 46 for (b = 0; b < 2; ++b) 47 ++(this)->a, --b, this->c /= 1; 48 }(); 49 #elif defined(BLOCKS) 50 ^{ 51 ++a; 52 --this->b; 53 (this)->c /= 1; 54 #pragma omp parallel 55 #pragma omp for 56 for (c = 0; c < 2; ++c) 57 ++(this)->a, --b, this->c /= 1; 58 }(); 59 #else 60 ++this->a, --b, c /= 1; 61 #endif 62 } 63 }; 64 65 template <typename T> 66 struct SST { 67 T a; 68 SST() : a(T()) { 69 #pragma omp parallel 70 #pragma omp for lastprivate(a) 71 for (int i = 0; i < 2; ++i) 72 #ifdef LAMBDA 73 [&]() { 74 [&]() { 75 ++this->a; 76 #pragma omp parallel 77 #pragma omp for lastprivate(a) 78 for (int i = 0; i < 2; ++i) 79 ++(this)->a; 80 }(); 81 }(); 82 #elif defined(BLOCKS) 83 ^{ 84 ^{ 85 ++a; 86 #pragma omp parallel 87 #pragma omp for lastprivate(a) 88 for (int i = 0; i < 2; ++i) 89 ++(this)->a; 90 }(); 91 }(); 92 #else 93 ++(this)->a; 94 #endif 95 #pragma omp for 96 for (a = 0; a < 2; ++a) 97 #ifdef LAMBDA 98 [&]() { 99 ++this->a; 100 #pragma omp parallel 101 #pragma omp for 102 for (a = 0; a < 2; ++(this)->a) 103 ++(this)->a; 104 }(); 105 #elif defined(BLOCKS) 106 ^{ 107 ++a; 108 #pragma omp parallel 109 #pragma omp for 110 for (this->a = 0; a < 2; ++a) 111 ++(this)->a; 112 }(); 113 #else 114 ++(this)->a; 115 #endif 116 } 117 }; 118 119 template <class T> 120 struct S { 121 T f; 122 S(T a) : f(a) {} 123 S() : f() {} 124 S<T> &operator=(const S<T> &); 125 operator T() { return T(); } 126 ~S() {} 127 }; 128 129 volatile int g __attribute__((aligned(128)))= 1212; 130 volatile int &g1 = g; 131 float f; 132 char cnt; 133 134 // CHECK: [[SS_TY:%.+]] = type { i{{[0-9]+}}, i8 135 // LAMBDA: [[SS_TY:%.+]] = type { i{{[0-9]+}}, i8 136 // BLOCKS: [[SS_TY:%.+]] = type { i{{[0-9]+}}, i8 137 // CHECK: [[S_FLOAT_TY:%.+]] = type { float } 138 // CHECK: [[S_INT_TY:%.+]] = type { i32 } 139 // CHECK-DAG: [[IMPLICIT_BARRIER_LOC:@.+]] = private unnamed_addr constant %{{.+}} { i32 0, i32 66, i32 0, i32 0, i8* 140 // CHECK-DAG: [[X:@.+]] = global double 0.0 141 // CHECK-DAG: [[F:@.+]] = global float 0.0 142 // CHECK-DAG: [[CNT:@.+]] = global i8 0 143 template <typename T> 144 T tmain() { 145 S<T> test; 146 SST<T> sst; 147 T t_var __attribute__((aligned(128))) = T(); 148 T vec[] __attribute__((aligned(128))) = {1, 2}; 149 S<T> s_arr[] __attribute__((aligned(128))) = {1, 2}; 150 S<T> &var __attribute__((aligned(128))) = test; 151 #pragma omp parallel 152 #pragma omp for lastprivate(t_var, vec, s_arr, var) 153 for (int i = 0; i < 2; ++i) { 154 vec[i] = t_var; 155 s_arr[i] = var; 156 } 157 return T(); 158 } 159 160 namespace A { 161 double x; 162 } 163 namespace B { 164 using A::x; 165 } 166 167 int main() { 168 static int sivar; 169 SS ss(sivar); 170 #ifdef LAMBDA 171 // LAMBDA: [[G:@.+]] = global i{{[0-9]+}} 1212, 172 // LAMBDA: [[SIVAR:@.+]] = internal global i{{[0-9]+}} 0, 173 // LAMBDA-LABEL: @main 174 // LAMBDA: alloca [[SS_TY]], 175 // LAMBDA: alloca [[CAP_TY:%.+]], 176 // LAMBDA: call void [[OUTER_LAMBDA:@.+]]([[CAP_TY]]* 177 [&]() { 178 // LAMBDA: define{{.*}} internal{{.*}} void [[OUTER_LAMBDA]]( 179 // LAMBDA: call void {{.+}} @__kmpc_fork_call({{.+}}, i32 1, {{.+}}* [[OMP_REGION:@.+]] to {{.+}}, i32* %{{.+}}) 180 #pragma omp parallel 181 #pragma omp for lastprivate(g, g1, sivar) 182 for (int i = 0; i < 2; ++i) { 183 // LAMBDA: define {{.+}} @{{.+}}([[SS_TY]]* 184 // LAMBDA: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 0 185 // LAMBDA: store i{{[0-9]+}} 0, i{{[0-9]+}}* % 186 // LAMBDA: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 1 187 // LAMBDA: store i8 188 // LAMBDA: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 2 189 // LAMBDA: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 1, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*, [[SS_TY]]*)* [[SS_MICROTASK:@.+]] to void 190 // LAMBDA: call void @__kmpc_for_static_init_4( 191 // LAMBDA-NOT: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 0 192 // LAMBDA: call{{.*}} void [[SS_LAMBDA1:@[^ ]+]] 193 // LAMBDA: call void @__kmpc_for_static_fini(% 194 // LAMBDA: ret 195 196 // LAMBDA: define internal void [[SS_MICROTASK]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, [[SS_TY]]* %{{.+}}) 197 // LAMBDA: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 0 198 // LAMBDA-NOT: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 1 199 // LAMBDA: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 2 200 // LAMBDA: call void @__kmpc_for_static_init_4( 201 // LAMBDA-NOT: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* 202 // LAMBDA: call{{.*}} void [[SS_LAMBDA:@[^ ]+]] 203 // LAMBDA: call void @__kmpc_for_static_fini( 204 // LAMBDA: br i1 205 // LAMBDA: [[B_REF:%.+]] = getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 1 206 // LAMBDA: store i8 %{{.+}}, i8* [[B_REF]], 207 // LAMBDA: br label 208 // LAMBDA: ret void 209 210 // LAMBDA: define internal void @{{.+}}(i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, [[SS_TY]]* %{{.+}}, i32* {{.+}}, i32* {{.+}}, i32* {{.+}}) 211 // LAMBDA: alloca i{{[0-9]+}}, 212 // LAMBDA: alloca i{{[0-9]+}}, 213 // LAMBDA: alloca i{{[0-9]+}}, 214 // LAMBDA: alloca i{{[0-9]+}}, 215 // LAMBDA: alloca i{{[0-9]+}}, 216 // LAMBDA: [[A_PRIV:%.+]] = alloca i{{[0-9]+}}, 217 // LAMBDA: [[B_PRIV:%.+]] = alloca i{{[0-9]+}}, 218 // LAMBDA: [[C_PRIV:%.+]] = alloca i{{[0-9]+}}, 219 // LAMBDA: store i{{[0-9]+}}* [[A_PRIV]], i{{[0-9]+}}** [[REFA:%.+]], 220 // LAMBDA: store i{{[0-9]+}}* [[C_PRIV]], i{{[0-9]+}}** [[REFC:%.+]], 221 // LAMBDA: call void @__kmpc_for_static_init_4( 222 // LAMBDA: [[A_PRIV:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[REFA]], 223 // LAMBDA-NEXT: [[A_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[A_PRIV]], 224 // LAMBDA-NEXT: [[INC:%.+]] = add nsw i{{[0-9]+}} [[A_VAL]], 1 225 // LAMBDA-NEXT: store i{{[0-9]+}} [[INC]], i{{[0-9]+}}* [[A_PRIV]], 226 // LAMBDA-NEXT: [[B_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[B_PRIV]], 227 // LAMBDA-NEXT: [[DEC:%.+]] = add nsw i{{[0-9]+}} [[B_VAL]], -1 228 // LAMBDA-NEXT: store i{{[0-9]+}} [[DEC]], i{{[0-9]+}}* [[B_PRIV]], 229 // LAMBDA-NEXT: [[C_PRIV:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[REFC]], 230 // LAMBDA-NEXT: [[C_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[C_PRIV]], 231 // LAMBDA-NEXT: [[DIV:%.+]] = sdiv i{{[0-9]+}} [[C_VAL]], 1 232 // LAMBDA-NEXT: store i{{[0-9]+}} [[DIV]], i{{[0-9]+}}* [[C_PRIV]], 233 // LAMBDA: call void @__kmpc_for_static_fini( 234 // LAMBDA: br i1 235 // LAMBDA: br label 236 // LAMBDA: ret void 237 238 // LAMBDA: define internal void @{{.+}}(i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, [[SS_TY]]* %{{.+}}) 239 // LAMBDA: ret void 240 241 // LAMBDA: define{{.*}} internal{{.*}} void [[OMP_REGION]](i32* noalias %{{.+}}, i32* noalias %{{.+}}, i32* dereferenceable(4) [[SIVAR:%.+]]) 242 // LAMBDA: alloca i{{[0-9]+}}, 243 // LAMBDA: alloca i{{[0-9]+}}, 244 // LAMBDA: alloca i{{[0-9]+}}, 245 // LAMBDA: alloca i{{[0-9]+}}, 246 // LAMBDA: alloca i{{[0-9]+}}, 247 // LAMBDA: [[G_PRIVATE_ADDR:%.+]] = alloca i{{[0-9]+}}, align 128 248 // LAMBDA: [[G1_PRIVATE_ADDR:%.+]] = alloca i{{[0-9]+}}, 249 // LAMBDA: [[SIVAR_PRIVATE_ADDR:%.+]] = alloca i{{[0-9]+}}, 250 // LAMBDA: [[SIVAR_PRIVATE_ADDR_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** %{{.+}}, 251 252 // LAMBDA: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** %{{.+}} 253 // LAMBDA: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 254 255 // LAMBDA: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 [[GTID]], i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 256 // LAMBDA: store i{{[0-9]+}} 1, i{{[0-9]+}}* [[G_PRIVATE_ADDR]], 257 // LAMBDA: store i{{[0-9]+}} 2, i{{[0-9]+}}* [[SIVAR_PRIVATE_ADDR]], 258 // LAMBDA: [[G_PRIVATE_ADDR_REF:%.+]] = getelementptr inbounds %{{.+}}, %{{.+}}* [[ARG:%.+]], i{{[0-9]+}} 0, i{{[0-9]+}} 0 259 // LAMBDA: store i{{[0-9]+}}* [[G_PRIVATE_ADDR]], i{{[0-9]+}}** [[G_PRIVATE_ADDR_REF]] 260 // LAMBDA: [[SIVAR_PRIVATE_ADDR_REF:%.+]] = getelementptr inbounds %{{.+}}, %{{.+}}* [[ARG:%.+]], i{{[0-9]+}} 0, i{{[0-9]+}} 1 261 // LAMBDA: store i{{[0-9]+}}* [[SIVAR_PRIVATE_ADDR]], i{{[0-9]+}}** [[SIVAR_PRIVATE_ADDR_REF]] 262 // LAMBDA: call void [[INNER_LAMBDA:@.+]](%{{.+}}* [[ARG]]) 263 // LAMBDA: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 [[GTID]]) 264 g = 1; 265 g1 = 1; 266 sivar = 2; 267 // Check for final copying of private values back to original vars. 268 // LAMBDA: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 269 // LAMBDA: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 270 // LAMBDA: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 271 // LAMBDA: [[LAST_THEN]] 272 // Actual copying. 273 274 // original g=private_g; 275 // LAMBDA: [[G_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[G_PRIVATE_ADDR]], 276 // LAMBDA: store volatile i{{[0-9]+}} [[G_VAL]], i{{[0-9]+}}* [[G]], 277 278 // original sivar=private_sivar; 279 // LAMBDA: [[SIVAR_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[SIVAR_PRIVATE_ADDR]], 280 // LAMBDA: store i{{[0-9]+}} [[SIVAR_VAL]], i{{[0-9]+}}* %{{.+}}, 281 // LAMBDA: br label %[[LAST_DONE]] 282 // LAMBDA: [[LAST_DONE]] 283 // LAMBDA: call void @__kmpc_barrier(%{{.+}}* @{{.+}}, i{{[0-9]+}} [[GTID]]) 284 [&]() { 285 // LAMBDA: define {{.+}} void [[INNER_LAMBDA]](%{{.+}}* [[ARG_PTR:%.+]]) 286 // LAMBDA: store %{{.+}}* [[ARG_PTR]], %{{.+}}** [[ARG_PTR_REF:%.+]], 287 g = 2; 288 g1 = 2; 289 sivar = 4; 290 // LAMBDA: [[ARG_PTR:%.+]] = load %{{.+}}*, %{{.+}}** [[ARG_PTR_REF]] 291 // LAMBDA: [[G_PTR_REF:%.+]] = getelementptr inbounds %{{.+}}, %{{.+}}* [[ARG_PTR]], i{{[0-9]+}} 0, i{{[0-9]+}} 0 292 // LAMBDA: [[G_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[G_PTR_REF]] 293 // LAMBDA: store i{{[0-9]+}} 2, i{{[0-9]+}}* [[G_REF]] 294 // LAMBDA: [[SIVAR_PTR_REF:%.+]] = getelementptr inbounds %{{.+}}, %{{.+}}* [[ARG_PTR]], i{{[0-9]+}} 0, i{{[0-9]+}} 1 295 // LAMBDA: [[SIVAR_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[SIVAR_PTR_REF]] 296 // LAMBDA: store i{{[0-9]+}} 4, i{{[0-9]+}}* [[SIVAR_REF]] 297 }(); 298 } 299 }(); 300 return 0; 301 #elif defined(BLOCKS) 302 // BLOCKS: [[G:@.+]] = global i{{[0-9]+}} 1212, 303 // BLOCKS-LABEL: @main 304 // BLOCKS: call 305 // BLOCKS: call void {{%.+}}(i8 306 ^{ 307 // BLOCKS: define{{.*}} internal{{.*}} void {{.+}}(i8* 308 // BLOCKS: call void {{.+}} @__kmpc_fork_call({{.+}}, i32 1, {{.+}}* [[OMP_REGION:@.+]] to {{.+}}) 309 #pragma omp parallel 310 #pragma omp for lastprivate(g, g1, sivar) 311 for (int i = 0; i < 2; ++i) { 312 // BLOCKS: define{{.*}} internal{{.*}} void [[OMP_REGION]](i32* noalias %{{.+}}, i32* noalias %{{.+}}, i32* dereferenceable(4) [[SIVAR:%.+]]) 313 // BLOCKS: alloca i{{[0-9]+}}, 314 // BLOCKS: alloca i{{[0-9]+}}, 315 // BLOCKS: alloca i{{[0-9]+}}, 316 // BLOCKS: alloca i{{[0-9]+}}, 317 // BLOCKS: alloca i{{[0-9]+}}, 318 // BLOCKS: [[G_PRIVATE_ADDR:%.+]] = alloca i{{[0-9]+}}, align 128 319 // BLOCKS: [[G1_PRIVATE_ADDR:%.+]] = alloca i{{[0-9]+}}, align 4 320 // BLOCKS: [[SIVAR_PRIVATE_ADDR:%.+]] = alloca i{{[0-9]+}}, 321 // BLOCKS: store i{{[0-9]+}}* [[SIVAR]], i{{[0-9]+}}** [[SIVAR_ADDR:%.+]], 322 // BLOCKS: {{.+}} = load i{{[0-9]+}}*, i{{[0-9]+}}** [[SIVAR_ADDR]] 323 // BLOCKS: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** %{{.+}} 324 // BLOCKS: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 325 // BLOCKS: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 [[GTID]], i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 326 // BLOCKS: store i{{[0-9]+}} 1, i{{[0-9]+}}* [[G_PRIVATE_ADDR]], 327 // BLOCKS-NOT: [[G]]{{[[^:word:]]}} 328 // BLOCKS: i{{[0-9]+}}* [[G_PRIVATE_ADDR]] 329 // BLOCKS-NOT: [[G]]{{[[^:word:]]}} 330 // BLOCKS: call void {{%.+}}(i8 331 // BLOCKS: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 [[GTID]]) 332 g = 1; 333 g1 = 1; 334 sivar = 2; 335 // Check for final copying of private values back to original vars. 336 // BLOCKS: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 337 // BLOCKS: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 338 // BLOCKS: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 339 // BLOCKS: [[LAST_THEN]] 340 // Actual copying. 341 342 // original g=private_g; 343 // BLOCKS: [[G_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[G_PRIVATE_ADDR]], 344 // BLOCKS: store volatile i{{[0-9]+}} [[G_VAL]], i{{[0-9]+}}* [[G]], 345 // BLOCKS: [[SIVAR_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[SIVAR_PRIVATE_ADDR]], 346 // BLOCKS: store i{{[0-9]+}} [[SIVAR_VAL]], i{{[0-9]+}}* %{{.+}}, 347 // BLOCKS: br label %[[LAST_DONE]] 348 // BLOCKS: [[LAST_DONE]] 349 // BLOCKS: call void @__kmpc_barrier(%{{.+}}* @{{.+}}, i{{[0-9]+}} [[GTID]]) 350 g = 1; 351 g1 = 1; 352 ^{ 353 // BLOCKS: define {{.+}} void {{@.+}}(i8* 354 g = 2; 355 g1 = 1; 356 sivar = 4; 357 // BLOCKS-NOT: [[G]]{{[[^:word:]]}} 358 // BLOCKS: store i{{[0-9]+}} 2, i{{[0-9]+}}* 359 // BLOCKS-NOT: [[G]]{{[[^:word:]]}} 360 // BLOCKS-NOT: [[SIVAR]]{{[[^:word:]]}} 361 // BLOCKS: store i{{[0-9]+}} 4, i{{[0-9]+}}* 362 // BLOCKS-NOT: [[SIVAR]]{{[[^:word:]]}} 363 // BLOCKS: ret 364 }(); 365 } 366 }(); 367 return 0; 368 // BLOCKS: define {{.+}} @{{.+}}([[SS_TY]]* 369 // BLOCKS: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 0 370 // BLOCKS: store i{{[0-9]+}} 0, i{{[0-9]+}}* % 371 // BLOCKS: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 1 372 // BLOCKS: store i8 373 // BLOCKS: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 2 374 // BLOCKS: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 1, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*, [[SS_TY]]*)* [[SS_MICROTASK:@.+]] to void 375 // BLOCKS: call void @__kmpc_for_static_init_4( 376 // BLOCKS-NOT: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 0 377 // BLOCKS: call void 378 // BLOCKS: call void @__kmpc_for_static_fini(% 379 // BLOCKS: ret 380 381 // BLOCKS: define internal void [[SS_MICROTASK]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, [[SS_TY]]* %{{.+}}) 382 // BLOCKS: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 0 383 // BLOCKS-NOT: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 1 384 // BLOCKS: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 2 385 // BLOCKS: call void @__kmpc_for_static_init_4( 386 // BLOCKS-NOT: getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* 387 // BLOCKS: call{{.*}} void 388 // BLOCKS: call void @__kmpc_for_static_fini( 389 // BLOCKS: br i1 390 // BLOCKS: [[B_REF:%.+]] = getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 1 391 // BLOCKS: store i8 %{{.+}}, i8* [[B_REF]], 392 // BLOCKS: br label 393 // BLOCKS: ret void 394 395 // BLOCKS: define internal void @{{.+}}(i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, [[SS_TY]]* %{{.+}}, i32* {{.+}}, i32* {{.+}}, i32* {{.+}}) 396 // BLOCKS: alloca i{{[0-9]+}}, 397 // BLOCKS: alloca i{{[0-9]+}}, 398 // BLOCKS: alloca i{{[0-9]+}}, 399 // BLOCKS: alloca i{{[0-9]+}}, 400 // BLOCKS: alloca i{{[0-9]+}}, 401 // BLOCKS: [[A_PRIV:%.+]] = alloca i{{[0-9]+}}, 402 // BLOCKS: [[B_PRIV:%.+]] = alloca i{{[0-9]+}}, 403 // BLOCKS: [[C_PRIV:%.+]] = alloca i{{[0-9]+}}, 404 // BLOCKS: store i{{[0-9]+}}* [[A_PRIV]], i{{[0-9]+}}** [[REFA:%.+]], 405 // BLOCKS: store i{{[0-9]+}}* [[C_PRIV]], i{{[0-9]+}}** [[REFC:%.+]], 406 // BLOCKS: call void @__kmpc_for_static_init_4( 407 // BLOCKS: [[A_PRIV:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[REFA]], 408 // BLOCKS-NEXT: [[A_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[A_PRIV]], 409 // BLOCKS-NEXT: [[INC:%.+]] = add nsw i{{[0-9]+}} [[A_VAL]], 1 410 // BLOCKS-NEXT: store i{{[0-9]+}} [[INC]], i{{[0-9]+}}* [[A_PRIV]], 411 // BLOCKS-NEXT: [[B_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[B_PRIV]], 412 // BLOCKS-NEXT: [[DEC:%.+]] = add nsw i{{[0-9]+}} [[B_VAL]], -1 413 // BLOCKS-NEXT: store i{{[0-9]+}} [[DEC]], i{{[0-9]+}}* [[B_PRIV]], 414 // BLOCKS-NEXT: [[C_PRIV:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[REFC]], 415 // BLOCKS-NEXT: [[C_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[C_PRIV]], 416 // BLOCKS-NEXT: [[DIV:%.+]] = sdiv i{{[0-9]+}} [[C_VAL]], 1 417 // BLOCKS-NEXT: store i{{[0-9]+}} [[DIV]], i{{[0-9]+}}* [[C_PRIV]], 418 // BLOCKS: call void @__kmpc_for_static_fini( 419 // BLOCKS: br i1 420 // BLOCKS: br label 421 // BLOCKS: ret void 422 #else 423 S<float> test; 424 int t_var = 0; 425 int vec[] = {1, 2}; 426 S<float> s_arr[] = {1, 2}; 427 S<float> var(3); 428 #pragma omp parallel 429 #pragma omp for lastprivate(t_var, vec, s_arr, var, sivar) 430 for (int i = 0; i < 2; ++i) { 431 vec[i] = t_var; 432 s_arr[i] = var; 433 sivar += i; 434 } 435 #pragma omp parallel 436 #pragma omp for lastprivate(A::x, B::x) firstprivate(f) lastprivate(f) 437 for (int i = 0; i < 2; ++i) { 438 A::x++; 439 } 440 #pragma omp parallel 441 #pragma omp for firstprivate(f) lastprivate(f) 442 for (int i = 0; i < 2; ++i) { 443 A::x++; 444 } 445 #pragma omp parallel 446 #pragma omp for lastprivate(cnt) 447 for (cnt = 0; cnt < 2; ++cnt) { 448 A::x++; 449 } 450 return tmain<int>(); 451 #endif 452 } 453 454 // CHECK: define i{{[0-9]+}} @main() 455 // CHECK: [[TEST:%.+]] = alloca [[S_FLOAT_TY]], 456 // CHECK: call {{.*}} [[S_FLOAT_TY_DEF_CONSTR:@.+]]([[S_FLOAT_TY]]* [[TEST]]) 457 // CHECK: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 5, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*, i32*, [2 x i32]*, [2 x [[S_FLOAT_TY]]]*, [[S_FLOAT_TY]]*, i32*)* [[MAIN_MICROTASK:@.+]] to void 458 // CHECK: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 0, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*)* [[MAIN_MICROTASK1:@.+]] to void 459 // CHECK: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 0, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*)* [[MAIN_MICROTASK2:@.+]] to void 460 // CHECK: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 0, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*)* [[MAIN_MICROTASK3:@.+]] to void 461 // CHECK: = call {{.+}} [[TMAIN_INT:@.+]]() 462 // CHECK: call void [[S_FLOAT_TY_DESTR:@.+]]([[S_FLOAT_TY]]* 463 // CHECK: ret 464 465 // CHECK: define internal void [[MAIN_MICROTASK]](i32* noalias [[GTID_ADDR:%.+]], i32* noalias %{{.+}}, i32* dereferenceable(4) %{{.+}}, [2 x i32]* dereferenceable(8) %{{.+}}, [2 x [[S_FLOAT_TY]]]* dereferenceable(8) %{{.+}}, [[S_FLOAT_TY]]* dereferenceable(4) %{{.+}}) 466 // CHECK: alloca i{{[0-9]+}}, 467 // CHECK: alloca i{{[0-9]+}}, 468 // CHECK: alloca i{{[0-9]+}}, 469 // CHECK: alloca i{{[0-9]+}}, 470 // CHECK: alloca i{{[0-9]+}}, 471 // CHECK: [[T_VAR_PRIV:%.+]] = alloca i{{[0-9]+}}, 472 // CHECK: [[VEC_PRIV:%.+]] = alloca [2 x i{{[0-9]+}}], 473 // CHECK: [[S_ARR_PRIV:%.+]] = alloca [2 x [[S_FLOAT_TY]]], 474 // CHECK: [[VAR_PRIV:%.+]] = alloca [[S_FLOAT_TY]], 475 // CHECK: [[SIVAR_PRIV:%.+]] = alloca i{{[0-9]+}}, 476 // CHECK: store i{{[0-9]+}}* [[GTID_ADDR]], i{{[0-9]+}}** [[GTID_ADDR_REF:%.+]] 477 478 // CHECK: [[T_VAR_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** % 479 // CHECK: [[VEC_REF:%.+]] = load [2 x i32]*, [2 x i32]** % 480 // CHECK: [[S_ARR_REF:%.+]] = load [2 x [[S_FLOAT_TY]]]*, [2 x [[S_FLOAT_TY]]]** % 481 // CHECK: [[VAR_REF:%.+]] = load [[S_FLOAT_TY]]*, [[S_FLOAT_TY]]** % 482 483 // Check for default initialization. 484 // CHECK-NOT: [[T_VAR_PRIV]] 485 // CHECK-NOT: [[VEC_PRIV]] 486 // CHECK: [[S_ARR_PRIV_ITEM:%.+]] = phi [[S_FLOAT_TY]]* 487 // CHECK: call {{.*}} [[S_FLOAT_TY_DEF_CONSTR]]([[S_FLOAT_TY]]* [[S_ARR_PRIV_ITEM]]) 488 // CHECK: call {{.*}} [[S_FLOAT_TY_DEF_CONSTR]]([[S_FLOAT_TY]]* [[VAR_PRIV]]) 489 // CHECK: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 %{{.+}}, i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 490 // <Skip loop body> 491 // CHECK: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 %{{.+}}) 492 493 // Check for final copying of private values back to original vars. 494 // CHECK: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 495 // CHECK: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 496 // CHECK: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 497 // CHECK: [[LAST_THEN]] 498 // Actual copying. 499 500 // original t_var=private_t_var; 501 // CHECK: [[T_VAR_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[T_VAR_PRIV]], 502 // CHECK: store i{{[0-9]+}} [[T_VAR_VAL]], i{{[0-9]+}}* [[T_VAR_REF]], 503 504 // original vec[]=private_vec[]; 505 // CHECK: [[VEC_DEST:%.+]] = bitcast [2 x i{{[0-9]+}}]* [[VEC_REF]] to i8* 506 // CHECK: [[VEC_SRC:%.+]] = bitcast [2 x i{{[0-9]+}}]* [[VEC_PRIV]] to i8* 507 // CHECK: call void @llvm.memcpy.{{.+}}(i8* [[VEC_DEST]], i8* [[VEC_SRC]], 508 509 // original s_arr[]=private_s_arr[]; 510 // CHECK: [[S_ARR_BEGIN:%.+]] = getelementptr inbounds [2 x [[S_FLOAT_TY]]], [2 x [[S_FLOAT_TY]]]* [[S_ARR_REF]], i{{[0-9]+}} 0, i{{[0-9]+}} 0 511 // CHECK: [[S_ARR_PRIV_BEGIN:%.+]] = bitcast [2 x [[S_FLOAT_TY]]]* [[S_ARR_PRIV]] to [[S_FLOAT_TY]]* 512 // CHECK: [[S_ARR_END:%.+]] = getelementptr [[S_FLOAT_TY]], [[S_FLOAT_TY]]* [[S_ARR_BEGIN]], i{{[0-9]+}} 2 513 // CHECK: [[IS_EMPTY:%.+]] = icmp eq [[S_FLOAT_TY]]* [[S_ARR_BEGIN]], [[S_ARR_END]] 514 // CHECK: br i1 [[IS_EMPTY]], label %[[S_ARR_BODY_DONE:.+]], label %[[S_ARR_BODY:.+]] 515 // CHECK: [[S_ARR_BODY]] 516 // CHECK: call {{.*}} [[S_FLOAT_TY_COPY_ASSIGN:@.+]]([[S_FLOAT_TY]]* {{.+}}, [[S_FLOAT_TY]]* {{.+}}) 517 // CHECK: br i1 {{.+}}, label %[[S_ARR_BODY_DONE]], label %[[S_ARR_BODY]] 518 // CHECK: [[S_ARR_BODY_DONE]] 519 520 // original var=private_var; 521 // CHECK: call {{.*}} [[S_FLOAT_TY_COPY_ASSIGN:@.+]]([[S_FLOAT_TY]]* [[VAR_REF]], [[S_FLOAT_TY]]* {{.*}} [[VAR_PRIV]]) 522 // CHECK: [[SIVAR_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[SIVAR_PRIV]], 523 // CHECK: br label %[[LAST_DONE]] 524 // CHECK: [[LAST_DONE]] 525 // CHECK-DAG: call void [[S_FLOAT_TY_DESTR]]([[S_FLOAT_TY]]* [[VAR_PRIV]]) 526 // CHECK-DAG: call void [[S_FLOAT_TY_DESTR]]([[S_FLOAT_TY]]* 527 // CHECK: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[GTID_ADDR_REF]] 528 // CHECK: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 529 // CHECK: call void @__kmpc_barrier(%{{.+}}* [[IMPLICIT_BARRIER_LOC]], i{{[0-9]+}} [[GTID]]) 530 // CHECK: ret void 531 532 // 533 // CHECK: define internal void [[MAIN_MICROTASK1]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}) 534 // CHECK: [[F_PRIV:%.+]] = alloca float, 535 // CHECK-NOT: alloca float 536 // CHECK: [[X_PRIV:%.+]] = alloca double, 537 // CHECK-NOT: alloca float 538 // CHECK-NOT: alloca double 539 540 // Check for default initialization. 541 // CHECK-NOT: [[X_PRIV]] 542 // CHECK: [[F_VAL:%.+]] = load float, float* [[F]], 543 // CHECK: store float [[F_VAL]], float* [[F_PRIV]], 544 // CHECK-NOT: [[X_PRIV]] 545 546 // CHECK: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[GTID_ADDR_REF]] 547 // CHECK: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 548 // CHECK: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 [[GTID]], i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 549 // <Skip loop body> 550 // CHECK: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 [[GTID]]) 551 552 // Check for final copying of private values back to original vars. 553 // CHECK: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 554 // CHECK: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 555 // CHECK: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 556 // CHECK: [[LAST_THEN]] 557 // Actual copying. 558 559 // original x=private_x; 560 // CHECK: [[X_VAL:%.+]] = load double, double* [[X_PRIV]], 561 // CHECK: store double [[X_VAL]], double* [[X]], 562 563 // original f=private_f; 564 // CHECK: [[F_VAL:%.+]] = load float, float* [[F_PRIV]], 565 // CHECK: store float [[F_VAL]], float* [[F]], 566 567 // CHECK-NEXT: br label %[[LAST_DONE]] 568 // CHECK: [[LAST_DONE]] 569 570 // CHECK: call void @__kmpc_barrier(%{{.+}}* [[IMPLICIT_BARRIER_LOC]], i{{[0-9]+}} [[GTID]]) 571 // CHECK: ret void 572 573 // CHECK: define internal void [[MAIN_MICROTASK2]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}) 574 // CHECK: [[F_PRIV:%.+]] = alloca float, 575 // CHECK-NOT: alloca float 576 577 // Check for default initialization. 578 // CHECK: [[F_VAL:%.+]] = load float, float* [[F]], 579 // CHECK: store float [[F_VAL]], float* [[F_PRIV]], 580 581 // CHECK: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[GTID_ADDR_REF]] 582 // CHECK: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 583 // CHECK: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 [[GTID]], i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 584 // <Skip loop body> 585 // CHECK: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 [[GTID]]) 586 587 // Check for final copying of private values back to original vars. 588 // CHECK: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 589 // CHECK: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 590 // CHECK: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 591 // CHECK: [[LAST_THEN]] 592 // Actual copying. 593 594 // original f=private_f; 595 // CHECK: [[F_VAL:%.+]] = load float, float* [[F_PRIV]], 596 // CHECK: store float [[F_VAL]], float* [[F]], 597 598 // CHECK-NEXT: br label %[[LAST_DONE]] 599 // CHECK: [[LAST_DONE]] 600 601 // CHECK: call void @__kmpc_barrier(%{{.+}}* [[IMPLICIT_BARRIER_LOC]], i{{[0-9]+}} [[GTID]]) 602 // CHECK: ret void 603 604 // CHECK: define internal void [[MAIN_MICROTASK3]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}) 605 // CHECK: [[CNT_PRIV:%.+]] = alloca i8, 606 607 // CHECK: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[GTID_ADDR_REF]] 608 // CHECK: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 609 // CHECK: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 [[GTID]], i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* [[OMP_LB:%[^,]+]], i32* [[OMP_UB:%[^,]+]], i32* [[OMP_ST:%[^,]+]], i32 1, i32 1) 610 // UB = min(UB, GlobalUB) 611 // CHECK-NEXT: [[UB:%.+]] = load i32, i32* [[OMP_UB]] 612 // CHECK-NEXT: [[UBCMP:%.+]] = icmp sgt i32 [[UB]], 1 613 // CHECK-NEXT: br i1 [[UBCMP]], label [[UB_TRUE:%[^,]+]], label [[UB_FALSE:%[^,]+]] 614 // CHECK: [[UBRESULT:%.+]] = phi i32 [ 1, [[UB_TRUE]] ], [ [[UBVAL:%[^,]+]], [[UB_FALSE]] ] 615 // CHECK-NEXT: store i32 [[UBRESULT]], i32* [[OMP_UB]] 616 // CHECK-NEXT: [[LB:%.+]] = load i32, i32* [[OMP_LB]] 617 // CHECK-NEXT: store i32 [[LB]], i32* [[OMP_IV:[^,]+]] 618 // <Skip loop body> 619 // CHECK: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 [[GTID]]) 620 621 // Check for final copying of private values back to original vars. 622 // CHECK: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 623 // CHECK: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 624 // CHECK: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 625 // CHECK: [[LAST_THEN]] 626 627 // Calculate private cnt value. 628 // CHECK: store i8 2, i8* [[CNT_PRIV]] 629 // original cnt=private_cnt; 630 // CHECK: [[CNT_VAL:%.+]] = load i8, i8* [[CNT_PRIV]], 631 // CHECK: store i8 [[CNT_VAL]], i8* [[CNT]], 632 633 // CHECK-NEXT: br label %[[LAST_DONE]] 634 // CHECK: [[LAST_DONE]] 635 636 // CHECK: call void @__kmpc_barrier(%{{.+}}* [[IMPLICIT_BARRIER_LOC]], i{{[0-9]+}} [[GTID]]) 637 // CHECK: ret void 638 639 // CHECK: define {{.*}} i{{[0-9]+}} [[TMAIN_INT]]() 640 // CHECK: [[TEST:%.+]] = alloca [[S_INT_TY]], 641 // CHECK: call {{.*}} [[S_INT_TY_DEF_CONSTR:@.+]]([[S_INT_TY]]* [[TEST]]) 642 // CHECK: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 4, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*, i32*, [2 x i32]*, [2 x [[S_INT_TY]]]*, [[S_INT_TY]]*)* [[TMAIN_MICROTASK:@.+]] to void 643 // CHECK: call void [[S_INT_TY_DESTR:@.+]]([[S_INT_TY]]* 644 // CHECK: ret 645 646 // CHECK: define {{.+}} @{{.+}}([[SS_TY]]* 647 // CHECK: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 0 648 // CHECK: store i{{[0-9]+}} 0, i{{[0-9]+}}* % 649 // CHECK: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 1 650 // CHECK: store i8 651 // CHECK: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 2 652 // CHECK: call void (%{{.+}}*, i{{[0-9]+}}, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)*, ...) @__kmpc_fork_call(%{{.+}}* @{{.+}}, i{{[0-9]+}} 1, void (i{{[0-9]+}}*, i{{[0-9]+}}*, ...)* bitcast (void (i{{[0-9]+}}*, i{{[0-9]+}}*, [[SS_TY]]*)* [[SS_MICROTASK:@.+]] to void 653 // CHECK: call void @__kmpc_for_static_init_4( 654 // CHECK-NOT: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 0 655 // CHECK: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 1 656 // CHECK: getelementptr inbounds [[SS_TY]], [[SS_TY]]* %{{.+}}, i32 0, i32 2 657 // CHECK: call void @__kmpc_for_static_fini(% 658 // CHECK: ret 659 660 // CHECK: define internal void [[SS_MICROTASK]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, [[SS_TY]]* %{{.+}}) 661 // CHECK: alloca i{{[0-9]+}}, 662 // CHECK: alloca i{{[0-9]+}}, 663 // CHECK: alloca i{{[0-9]+}}, 664 // CHECK: alloca i{{[0-9]+}}, 665 // CHECK: alloca i{{[0-9]+}}, 666 // CHECK: alloca i{{[0-9]+}}, 667 // CHECK: [[A_PRIV:%.+]] = alloca i{{[0-9]+}}, 668 // CHECK: [[B_PRIV:%.+]] = alloca i{{[0-9]+}}, 669 // CHECK: [[C_PRIV:%.+]] = alloca i{{[0-9]+}}, 670 // CHECK: store i{{[0-9]+}}* [[A_PRIV]], i{{[0-9]+}}** [[REFA:%.+]], 671 // CHECK: store i{{[0-9]+}}* [[C_PRIV]], i{{[0-9]+}}** [[REFC:%.+]], 672 // CHECK: call void @__kmpc_for_static_init_4( 673 // CHECK: [[A_PRIV:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[REFA]], 674 // CHECK-NEXT: [[A_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[A_PRIV]], 675 // CHECK-NEXT: [[INC:%.+]] = add nsw i{{[0-9]+}} [[A_VAL]], 1 676 // CHECK-NEXT: store i{{[0-9]+}} [[INC]], i{{[0-9]+}}* [[A_PRIV]], 677 // CHECK-NEXT: [[B_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[B_PRIV]], 678 // CHECK-NEXT: [[DEC:%.+]] = add nsw i{{[0-9]+}} [[B_VAL]], -1 679 // CHECK-NEXT: store i{{[0-9]+}} [[DEC]], i{{[0-9]+}}* [[B_PRIV]], 680 // CHECK-NEXT: [[C_PRIV:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[REFC]], 681 // CHECK-NEXT: [[C_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[C_PRIV]], 682 // CHECK-NEXT: [[DIV:%.+]] = sdiv i{{[0-9]+}} [[C_VAL]], 1 683 // CHECK-NEXT: store i{{[0-9]+}} [[DIV]], i{{[0-9]+}}* [[C_PRIV]], 684 // CHECK: call void @__kmpc_for_static_fini( 685 // CHECK: br i1 686 // CHECK: [[B_REF:%.+]] = getelementptr {{.*}}[[SS_TY]], [[SS_TY]]* %{{.*}}, i32 0, i32 1 687 // CHECK: store i8 %{{.+}}, i8* [[B_REF]], 688 // CHECK: br label 689 // CHECK: ret void 690 691 // CHECK: define internal void [[TMAIN_MICROTASK]](i{{[0-9]+}}* noalias [[GTID_ADDR:%.+]], i{{[0-9]+}}* noalias %{{.+}}, i32* dereferenceable(4) %{{.+}}, [2 x i32]* dereferenceable(8) %{{.+}}, [2 x [[S_INT_TY]]]* dereferenceable(8) %{{.+}}, [[S_INT_TY]]* dereferenceable(4) %{{.+}}) 692 // CHECK: alloca i{{[0-9]+}}, 693 // CHECK: alloca i{{[0-9]+}}, 694 // CHECK: alloca i{{[0-9]+}}, 695 // CHECK: alloca i{{[0-9]+}}, 696 // CHECK: alloca i{{[0-9]+}}, 697 // CHECK: [[T_VAR_PRIV:%.+]] = alloca i{{[0-9]+}}, align 128 698 // CHECK: [[VEC_PRIV:%.+]] = alloca [2 x i{{[0-9]+}}], align 128 699 // CHECK: [[S_ARR_PRIV:%.+]] = alloca [2 x [[S_INT_TY]]], align 128 700 // CHECK: [[VAR_PRIV:%.+]] = alloca [[S_INT_TY]], align 128 701 // CHECK: [[VAR_PRIV_REF:%.+]] = alloca [[S_INT_TY]]*, 702 // CHECK: store i{{[0-9]+}}* [[GTID_ADDR]], i{{[0-9]+}}** [[GTID_ADDR_REF:%.+]] 703 704 // CHECK: [[T_VAR_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** % 705 // CHECK: [[VEC_REF:%.+]] = load [2 x i{{[0-9]+}}]*, [2 x i{{[0-9]+}}]** % 706 // CHECK: [[S_ARR_REF:%.+]] = load [2 x [[S_INT_TY]]]*, [2 x [[S_INT_TY]]]** % 707 708 // Check for default initialization. 709 // CHECK-NOT: [[T_VAR_PRIV]] 710 // CHECK-NOT: [[VEC_PRIV]] 711 // CHECK: [[S_ARR_PRIV_ITEM:%.+]] = phi [[S_INT_TY]]* 712 // CHECK: call {{.*}} [[S_INT_TY_DEF_CONSTR]]([[S_INT_TY]]* [[S_ARR_PRIV_ITEM]]) 713 // CHECK: [[VAR_REF:%.+]] = load [[S_INT_TY]]*, [[S_INT_TY]]** % 714 // CHECK: call {{.*}} [[S_INT_TY_DEF_CONSTR]]([[S_INT_TY]]* [[VAR_PRIV]]) 715 // CHECK: store [[S_INT_TY]]* [[VAR_PRIV]], [[S_INT_TY]]** [[VAR_PRIV_REF]] 716 // CHECK: call {{.+}} @__kmpc_for_static_init_4(%{{.+}}* @{{.+}}, i32 %{{.+}}, i32 34, i32* [[IS_LAST_ADDR:%.+]], i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 717 // <Skip loop body> 718 // CHECK: call void @__kmpc_for_static_fini(%{{.+}}* @{{.+}}, i32 %{{.+}}) 719 720 // Check for final copying of private values back to original vars. 721 // CHECK: [[IS_LAST_VAL:%.+]] = load i32, i32* [[IS_LAST_ADDR]], 722 // CHECK: [[IS_LAST_ITER:%.+]] = icmp ne i32 [[IS_LAST_VAL]], 0 723 // CHECK: br i1 [[IS_LAST_ITER:%.+]], label %[[LAST_THEN:.+]], label %[[LAST_DONE:.+]] 724 // CHECK: [[LAST_THEN]] 725 // Actual copying. 726 727 // original t_var=private_t_var; 728 // CHECK: [[T_VAR_VAL:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[T_VAR_PRIV]], 729 // CHECK: store i{{[0-9]+}} [[T_VAR_VAL]], i{{[0-9]+}}* [[T_VAR_REF]], 730 731 // original vec[]=private_vec[]; 732 // CHECK: [[VEC_DEST:%.+]] = bitcast [2 x i{{[0-9]+}}]* [[VEC_REF]] to i8* 733 // CHECK: [[VEC_SRC:%.+]] = bitcast [2 x i{{[0-9]+}}]* [[VEC_PRIV]] to i8* 734 // CHECK: call void @llvm.memcpy.{{.+}}(i8* [[VEC_DEST]], i8* [[VEC_SRC]], 735 736 // original s_arr[]=private_s_arr[]; 737 // CHECK: [[S_ARR_BEGIN:%.+]] = getelementptr inbounds [2 x [[S_INT_TY]]], [2 x [[S_INT_TY]]]* [[S_ARR_REF]], i{{[0-9]+}} 0, i{{[0-9]+}} 0 738 // CHECK: [[S_ARR_PRIV_BEGIN:%.+]] = bitcast [2 x [[S_INT_TY]]]* [[S_ARR_PRIV]] to [[S_INT_TY]]* 739 // CHECK: [[S_ARR_END:%.+]] = getelementptr [[S_INT_TY]], [[S_INT_TY]]* [[S_ARR_BEGIN]], i{{[0-9]+}} 2 740 // CHECK: [[IS_EMPTY:%.+]] = icmp eq [[S_INT_TY]]* [[S_ARR_BEGIN]], [[S_ARR_END]] 741 // CHECK: br i1 [[IS_EMPTY]], label %[[S_ARR_BODY_DONE:.+]], label %[[S_ARR_BODY:.+]] 742 // CHECK: [[S_ARR_BODY]] 743 // CHECK: call {{.*}} [[S_INT_TY_COPY_ASSIGN:@.+]]([[S_INT_TY]]* {{.+}}, [[S_INT_TY]]* {{.+}}) 744 // CHECK: br i1 {{.+}}, label %[[S_ARR_BODY_DONE]], label %[[S_ARR_BODY]] 745 // CHECK: [[S_ARR_BODY_DONE]] 746 747 // original var=private_var; 748 // CHECK: [[VAR_PRIV1:%.+]] = load [[S_INT_TY]]*, [[S_INT_TY]]** [[VAR_PRIV_REF]], 749 // CHECK: call {{.*}} [[S_INT_TY_COPY_ASSIGN:@.+]]([[S_INT_TY]]* [[VAR_REF]], [[S_INT_TY]]* {{.*}} [[VAR_PRIV1]]) 750 // CHECK: br label %[[LAST_DONE]] 751 // CHECK: [[LAST_DONE]] 752 // CHECK-DAG: call void [[S_INT_TY_DESTR]]([[S_INT_TY]]* [[VAR_PRIV]]) 753 // CHECK-DAG: call void [[S_INT_TY_DESTR]]([[S_INT_TY]]* 754 // CHECK: [[GTID_REF:%.+]] = load i{{[0-9]+}}*, i{{[0-9]+}}** [[GTID_ADDR_REF]] 755 // CHECK: [[GTID:%.+]] = load i{{[0-9]+}}, i{{[0-9]+}}* [[GTID_REF]] 756 // CHECK: call void @__kmpc_barrier(%{{.+}}* [[IMPLICIT_BARRIER_LOC]], i{{[0-9]+}} [[GTID]]) 757 // CHECK: ret void 758 #endif 759 760