1 // RUN: %clang_cc1 -verify -fopenmp -triple x86_64-unknown-unknown -emit-llvm %s -o - | FileCheck %s
2 // RUN: %clang_cc1 -fopenmp -triple x86_64-unknown-unknown -emit-pch -o %t %s
3 // RUN: %clang_cc1 -fopenmp -triple x86_64-unknown-unknown -include-pch %t -verify %s -emit-llvm -o - | FileCheck %s
5 // RUN: %clang_cc1 -verify -fopenmp-simd -triple x86_64-unknown-unknown -emit-llvm %s -o - | FileCheck --check-prefix SIMD-ONLY0 %s
6 // RUN: %clang_cc1 -fopenmp-simd -triple x86_64-unknown-unknown -emit-pch -o %t %s
7 // RUN: %clang_cc1 -fopenmp-simd -triple x86_64-unknown-unknown -include-pch %t -verify %s -emit-llvm -o - | FileCheck --check-prefix SIMD-ONLY0 %s
8 // SIMD-ONLY0-NOT: {{__kmpc|__tgt}}
9 // expected-no-diagnostics
14 // CHECK: [[KMP_DIM:%.+]] = type { i64, i64, i64 }
16 int a[10], b[10], c[10], d[10];
19 // CHECK-LABEL: @main()
22 // CHECK: [[DIMS:%.+]] = alloca [1 x [[KMP_DIM]]],
23 // CHECK: [[GTID:%.+]] = call i32 @__kmpc_global_thread_num([[IDENT:%.+]])
25 // CHECK-NEXT: br i1 %
26 // CHECK: [[CAST:%.+]] = bitcast [1 x [[KMP_DIM]]]* [[DIMS]] to i8*
27 // CHECK: call void @llvm.memset.p0i8.i64(i8* align 8 [[CAST]], i8 0, i64 24, i1 false)
28 // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0
29 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1
30 // CHECK: store i64 %{{.+}}, i64* %
31 // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2
32 // CHECK: store i64 1, i64* %
33 // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0
34 // CHECK: [[CAST:%.+]] = bitcast [[KMP_DIM]]* [[DIM]] to i8*
35 // CHECK: call void @__kmpc_doacross_init([[IDENT]], i32 [[GTID]], i32 1, i8* [[CAST]])
36 // CHECK: call void @__kmpc_for_static_init_4(
37 #pragma omp for ordered(1)
38 for (i = 0; i < n; ++i) {
41 // CHECK: call void [[FOO:.+]](
42 // CHECK: load i32, i32* [[I:%.+]],
43 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
44 // CHECK-NEXT: sdiv i32 %{{.+}}, 1
45 // CHECK-NEXT: sext i32 %{{.+}} to i64
46 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0
47 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
48 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0
49 // CHECK-NEXT: call void @__kmpc_doacross_post([[IDENT]], i32 [[GTID]], i64* [[TMP]])
50 #pragma omp ordered depend(source)
53 // CHECK: call void [[FOO]]
54 // CHECK: load i32, i32* [[I]],
55 // CHECK-NEXT: sub nsw i32 %{{.+}}, 2
56 // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
57 // CHECK-NEXT: sdiv i32 %{{.+}}, 1
58 // CHECK-NEXT: sext i32 %{{.+}} to i64
59 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0
60 // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
61 // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0
62 // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]])
63 #pragma omp ordered depend(sink : i - 2)
66 // CHECK: call void @__kmpc_for_static_fini(
67 // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]])