1 // RUN: %clang_cc1 -verify -fopenmp -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -fexceptions -fcxx-exceptions -o - | FileCheck %s 2 // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -emit-pch -o %t %s 3 // RUN: %clang_cc1 -fopenmp -x c++ -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck %s 4 5 // RUN: %clang_cc1 -verify -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -fexceptions -fcxx-exceptions -o - | FileCheck --check-prefix SIMD-ONLY0 %s 6 // RUN: %clang_cc1 -fopenmp-simd -x c++ -std=c++11 -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -emit-pch -o %t %s 7 // RUN: %clang_cc1 -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck --check-prefix SIMD-ONLY0 %s 8 // SIMD-ONLY0-NOT: {{__kmpc|__tgt}} 9 // expected-no-diagnostics 10 11 #ifndef HEADER 12 #define HEADER 13 14 // CHECK: [[KMP_DIM:%.+]] = type { i64, i64, i64 } 15 extern int n; 16 int a[10], b[10], c[10], d[10]; 17 void foo(); 18 19 // CHECK-LABEL:bar 20 void bar() { 21 int i,j; 22 // CHECK: call void @__kmpc_doacross_init( 23 // CHECK: call void @__kmpc_doacross_fini( 24 #pragma omp parallel for ordered(2) 25 for (i = 0; i < n; ++i) 26 for (j = 0; j < n; ++j) 27 a[i] = b[i] + 1; 28 // CHECK: call void @__kmpc_doacross_init( 29 // CHECK: call void @__kmpc_doacross_fini( 30 #pragma omp for collapse(2) ordered(2) 31 for (int i = 0; i < n; i++) 32 for (int j = 0; j < n; j++) 33 ; 34 } 35 36 // CHECK-LABEL: @main() 37 int main() { 38 int i; 39 // CHECK: [[DIMS:%.+]] = alloca [1 x [[KMP_DIM]]], 40 // CHECK: [[GTID:%.+]] = call i32 @__kmpc_global_thread_num([[IDENT:%.+]]) 41 // CHECK: icmp 42 // CHECK-NEXT: br i1 % 43 // CHECK: [[CAST:%.+]] = bitcast [1 x [[KMP_DIM]]]* [[DIMS]] to i8* 44 // CHECK: call void @llvm.memset.p0i8.i64(i8* align 8 [[CAST]], i8 0, i64 24, i1 false) 45 // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0 46 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1 47 // CHECK: store i64 %{{.+}}, i64* % 48 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2 49 // CHECK: store i64 1, i64* % 50 // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0 51 // CHECK: [[CAST:%.+]] = bitcast [[KMP_DIM]]* [[DIM]] to i8* 52 // CHECK: call void @__kmpc_doacross_init([[IDENT]], i32 [[GTID]], i32 1, i8* [[CAST]]) 53 // CHECK: call void @__kmpc_for_static_init_4(%struct.ident_t* @{{.+}}, i32 [[GTID]], i32 33, i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 54 #pragma omp for ordered(1) 55 for (int i = 0; i < n; ++i) { 56 a[i] = b[i] + 1; 57 foo(); 58 // CHECK: invoke void [[FOO:.+]]( 59 // CHECK: load i32, i32* [[I:%.+]], 60 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 61 // CHECK-NEXT: sdiv i32 %{{.+}}, 1 62 // CHECK-NEXT: sext i32 %{{.+}} to i64 63 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0 64 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 65 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0 66 // CHECK-NEXT: call void @__kmpc_doacross_post([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 67 #pragma omp ordered depend(source) 68 c[i] = c[i] + 1; 69 foo(); 70 // CHECK: invoke void [[FOO]] 71 // CHECK: load i32, i32* [[I]], 72 // CHECK-NEXT: sub nsw i32 %{{.+}}, 2 73 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 74 // CHECK-NEXT: sdiv i32 %{{.+}}, 1 75 // CHECK-NEXT: sext i32 %{{.+}} to i64 76 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0 77 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 78 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0 79 // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 80 #pragma omp ordered depend(sink : i - 2) 81 d[i] = a[i - 2]; 82 } 83 // CHECK: landingpad 84 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]]) 85 // CHECK: br label % 86 87 // CHECK: call void @__kmpc_for_static_fini( 88 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]]) 89 // CHECK: ret i32 0 90 return 0; 91 } 92 93 // CHECK-LABEL: main1 94 int main1() { 95 // CHECK: [[DIMS:%.+]] = alloca [1 x [[KMP_DIM]]], 96 // CHECK: [[GTID:%.+]] = call i32 @__kmpc_global_thread_num([[IDENT:%.+]]) 97 // CHECK: icmp 98 // CHECK-NEXT: br i1 % 99 // CHECK: [[CAST:%.+]] = bitcast [1 x [[KMP_DIM]]]* [[DIMS]] to i8* 100 // CHECK: call void @llvm.memset.p0i8.i64(i8* align 8 [[CAST]], i8 0, i64 24, i1 false) 101 // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0 102 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1 103 // CHECK: store i64 %{{.+}}, i64* % 104 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2 105 // CHECK: store i64 1, i64* % 106 // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0 107 // CHECK: [[CAST:%.+]] = bitcast [[KMP_DIM]]* [[DIM]] to i8* 108 // CHECK: call void @__kmpc_doacross_init([[IDENT]], i32 [[GTID]], i32 1, i8* [[CAST]]) 109 // CHECK: call void @__kmpc_for_static_init_4(%struct.ident_t* @{{.+}}, i32 [[GTID]], i32 33, i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 110 #pragma omp for ordered(1) 111 for (int i = n; i > 0; --i) { 112 a[i] = b[i] + 1; 113 foo(); 114 // CHECK: invoke void [[FOO:.+]]( 115 // CHECK: [[UB_VAL:%.+]] = load i32, i32* [[UB:%.+]], 116 // CHECK-NEXT: [[I_VAL:%.+]] = load i32, i32* [[I:%.+]], 117 // CHECK-NEXT: sub i32 [[UB_VAL]], [[I_VAL]] 118 // CHECK-NEXT: udiv i32 %{{.+}}, 1 119 // CHECK-NEXT: zext i32 %{{.+}} to i64 120 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0 121 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 122 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0 123 // CHECK-NEXT: call void @__kmpc_doacross_post([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 124 #pragma omp ordered depend(source) 125 c[i] = c[i] + 1; 126 foo(); 127 // CHECK: invoke void [[FOO]] 128 // CHECK: [[UB_VAL:%.+]] = load i32, i32* [[UB]], 129 // CHECK-NEXT: [[I_VAL:%.+]] = load i32, i32* [[I]], 130 // CHECK-NEXT: [[SUB:%.+]] = sub nsw i32 [[I_VAL]], 2 131 // CHECK-NEXT: sub i32 [[UB_VAL]], [[SUB]] 132 // CHECK-NEXT: udiv i32 %{{.+}}, 1 133 // CHECK-NEXT: zext i32 %{{.+}} to i64 134 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0 135 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 136 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0 137 // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 138 #pragma omp ordered depend(sink : i - 2) 139 d[i] = a[i - 2]; 140 } 141 // CHECK: landingpad 142 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]]) 143 // CHECK: br label % 144 145 // CHECK: call void @__kmpc_for_static_fini( 146 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]]) 147 // CHECK: ret i32 0 148 return 0; 149 } 150 151 // CHECK: define {{.+}}TestStruct 152 template <typename T> 153 struct TestStruct { 154 static const int M = 10; 155 static const int N = 20; 156 T i; 157 T a[N][M]; 158 T b[N][M]; 159 T foo(T, T); 160 T bar(T, T, T); 161 void baz(T, T); 162 TestStruct() { 163 // CHECK: [[DIMS:%.+]] = alloca [2 x [[KMP_DIM]]], 164 // CHECK: [[GTID:%.+]] = call i32 @__kmpc_global_thread_num([[IDENT:%.+]]) 165 // CHECK: [[CAST:%.+]] = bitcast [2 x [[KMP_DIM]]]* [[DIMS]] to i8* 166 // CHECK: call void @llvm.memset.p0i8.i64(i8* align 8 [[CAST]], i8 0, i64 48, i1 false) 167 // CHECK: [[DIM:%.+]] = getelementptr inbounds [2 x [[KMP_DIM]]], [2 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0 168 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1 169 // CHECK: store i64 10, i64* % 170 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2 171 // CHECK: store i64 1, i64* % 172 // CHECK: [[DIM:%.+]] = getelementptr inbounds [2 x [[KMP_DIM]]], [2 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 1 173 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1 174 // CHECK: store i64 %{{.+}}, i64* % 175 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2 176 // CHECK: store i64 1, i64* % 177 // CHECK: [[DIM:%.+]] = getelementptr inbounds [2 x [[KMP_DIM]]], [2 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0 178 // CHECK: [[CAST:%.+]] = bitcast [[KMP_DIM]]* [[DIM]] to i8* 179 // CHECK: call void @__kmpc_doacross_init([[IDENT]], i32 [[GTID]], i32 2, i8* [[CAST]]) 180 // CHECK: call void @__kmpc_for_static_init_4(%struct.ident_t* @{{.+}}, i32 [[GTID]], i32 33, i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32* %{{.+}}, i32 1, i32 1) 181 #pragma omp for ordered(2) 182 for (T j = 0; j < M; j++) 183 for (i = 0; i < n; i += 2) { 184 a[i][j] = foo(i, j); 185 // CHECK: invoke {{.+TestStruct.+foo}} 186 // CHECK: load i32*, i32** % 187 // CHECK: load i32, i32* % 188 // CHECK: load i32, i32* % 189 // CHECK: load i32, i32* [[J:%.+]], 190 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 191 // CHECK-NEXT: sdiv i32 %{{.+}}, 1 192 // CHECK-NEXT: sext i32 %{{.+}} to i64 193 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT:%.+]], i64 0, i64 0 194 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP:%.+]], 195 // CHECK-NEXT: [[I:%.+]] = load i32*, i32** [[I_REF:%.+]], 196 // CHECK-NEXT: load i32, i32* [[I]], 197 // CHECK-NEXT: sub nsw i32 %{{.+}}, 2 198 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 199 // CHECK-NEXT: sdiv i32 %{{.+}}, 2 200 // CHECK-NEXT: sext i32 %{{.+}} to i64 201 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 1 202 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 203 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 0 204 // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 205 // CHECK-NEXT: load i32, i32* [[J:%.+]], 206 // CHECK-NEXT: sub nsw i32 %{{.+}}, 1 207 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 208 // CHECK-NEXT: sdiv i32 %{{.+}}, 1 209 // CHECK-NEXT: sext i32 %{{.+}} to i64 210 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT:%.+]], i64 0, i64 0 211 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP:%.+]], 212 // CHECK-NEXT: [[I:%.+]] = load i32*, i32** [[I_REF]], 213 // CHECK-NEXT: load i32, i32* [[I]], 214 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 215 // CHECK-NEXT: sdiv i32 %{{.+}}, 2 216 // CHECK-NEXT: sext i32 %{{.+}} to i64 217 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 1 218 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 219 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 0 220 // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 221 #pragma omp ordered depend(sink : j, i - 2) depend(sink : j - 1, i) 222 b[i][j] = bar(a[i][j], b[i - 1][j], b[i][j - 1]); 223 // CHECK: invoke {{.+TestStruct.+bar}} 224 // CHECK: load i32*, i32** % 225 // CHECK: load i32, i32* % 226 // CHECK: load i32, i32* % 227 // CHECK: load i32, i32* [[J]], 228 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 229 // CHECK-NEXT: sdiv i32 %{{.+}}, 1 230 // CHECK-NEXT: sext i32 %{{.+}} to i64 231 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT:%.+]], i64 0, i64 0 232 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP:%.+]], 233 // CHECK-NEXT: [[I:%.+]] = load i32*, i32** [[I_REF]], 234 // CHECK-NEXT: load i32, i32* [[I]], 235 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0 236 // CHECK-NEXT: sdiv i32 %{{.+}}, 2 237 // CHECK-NEXT: sext i32 %{{.+}} to i64 238 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 1 239 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]], 240 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 0 241 // CHECK-NEXT: call void @__kmpc_doacross_post([[IDENT]], i32 [[GTID]], i64* [[TMP]]) 242 #pragma omp ordered depend(source) 243 baz(a[i][j], b[i][j]); 244 } 245 } 246 // CHECK: landingpad 247 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]]) 248 // CHECK: br label % 249 250 // CHECK: call void @__kmpc_for_static_fini( 251 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]]) 252 // CHECK: ret 253 }; 254 255 TestStruct<int> s; 256 #endif // HEADER 257