ordered_doacross_codegen.cpp 8.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192
  1. // RUN: %clang_cc1 -verify -fopenmp -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -fexceptions -fcxx-exceptions -o - | FileCheck %s
  2. // RUN: %clang_cc1 -fopenmp -x c++ -std=c++11 -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -emit-pch -o %t %s
  3. // RUN: %clang_cc1 -fopenmp -x c++ -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck %s
  4. // RUN: %clang_cc1 -verify -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -emit-llvm %s -fexceptions -fcxx-exceptions -o - | FileCheck --check-prefix SIMD-ONLY0 %s
  5. // RUN: %clang_cc1 -fopenmp-simd -x c++ -std=c++11 -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -emit-pch -o %t %s
  6. // RUN: %clang_cc1 -fopenmp-simd -x c++ -triple x86_64-unknown-unknown -fexceptions -fcxx-exceptions -std=c++11 -include-pch %t -verify %s -emit-llvm -o - | FileCheck --check-prefix SIMD-ONLY0 %s
  7. // SIMD-ONLY0-NOT: {{__kmpc|__tgt}}
  8. // expected-no-diagnostics
  9. #ifndef HEADER
  10. #define HEADER
  11. // CHECK: [[KMP_DIM:%.+]] = type { i64, i64, i64 }
  12. extern int n;
  13. int a[10], b[10], c[10], d[10];
  14. void foo();
  15. // CHECK-LABEL:bar
  16. void bar() {
  17. int i,j;
  18. // CHECK: call void @__kmpc_doacross_init(
  19. // CHECK: call void @__kmpc_doacross_fini(
  20. #pragma omp parallel for ordered(2)
  21. for (i = 0; i < n; ++i)
  22. for (j = 0; j < n; ++j)
  23. a[i] = b[i] + 1;
  24. }
  25. // CHECK-LABEL: @main()
  26. int main() {
  27. int i;
  28. // CHECK: [[DIMS:%.+]] = alloca [1 x [[KMP_DIM]]],
  29. // CHECK: [[GTID:%.+]] = call i32 @__kmpc_global_thread_num([[IDENT:%.+]])
  30. // CHECK: icmp
  31. // CHECK-NEXT: br i1 %
  32. // CHECK: [[CAST:%.+]] = bitcast [1 x [[KMP_DIM]]]* [[DIMS]] to i8*
  33. // CHECK: call void @llvm.memset.p0i8.i64(i8* align 8 [[CAST]], i8 0, i64 24, i1 false)
  34. // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0
  35. // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1
  36. // CHECK: store i64 %{{.+}}, i64* %
  37. // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2
  38. // CHECK: store i64 1, i64* %
  39. // CHECK: [[DIM:%.+]] = getelementptr inbounds [1 x [[KMP_DIM]]], [1 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0
  40. // CHECK: [[CAST:%.+]] = bitcast [[KMP_DIM]]* [[DIM]] to i8*
  41. // CHECK: call void @__kmpc_doacross_init([[IDENT]], i32 [[GTID]], i32 1, i8* [[CAST]])
  42. // CHECK: call void @__kmpc_for_static_init_4(
  43. #pragma omp for ordered(1)
  44. for (int i = 0; i < n; ++i) {
  45. a[i] = b[i] + 1;
  46. foo();
  47. // CHECK: invoke void [[FOO:.+]](
  48. // CHECK: load i32, i32* [[I:%.+]],
  49. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  50. // CHECK-NEXT: sdiv i32 %{{.+}}, 1
  51. // CHECK-NEXT: sext i32 %{{.+}} to i64
  52. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0
  53. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
  54. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0
  55. // CHECK-NEXT: call void @__kmpc_doacross_post([[IDENT]], i32 [[GTID]], i64* [[TMP]])
  56. #pragma omp ordered depend(source)
  57. c[i] = c[i] + 1;
  58. foo();
  59. // CHECK: invoke void [[FOO]]
  60. // CHECK: load i32, i32* [[I]],
  61. // CHECK-NEXT: sub nsw i32 %{{.+}}, 2
  62. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  63. // CHECK-NEXT: sdiv i32 %{{.+}}, 1
  64. // CHECK-NEXT: sext i32 %{{.+}} to i64
  65. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT:%.+]], i64 0, i64 0
  66. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
  67. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [1 x i64], [1 x i64]* [[CNT]], i64 0, i64 0
  68. // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]])
  69. #pragma omp ordered depend(sink : i - 2)
  70. d[i] = a[i - 2];
  71. }
  72. // CHECK: landingpad
  73. // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]])
  74. // CHECK: br label %
  75. // CHECK: call void @__kmpc_for_static_fini(
  76. // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]])
  77. // CHECK: ret i32 0
  78. return 0;
  79. }
  80. // CHECK: define {{.+}}TestStruct
  81. template <typename T>
  82. struct TestStruct {
  83. static const int M = 10;
  84. static const int N = 20;
  85. T i;
  86. T a[N][M];
  87. T b[N][M];
  88. T foo(T, T);
  89. T bar(T, T, T);
  90. void baz(T, T);
  91. TestStruct() {
  92. // CHECK: [[DIMS:%.+]] = alloca [2 x [[KMP_DIM]]],
  93. // CHECK: [[GTID:%.+]] = call i32 @__kmpc_global_thread_num([[IDENT:%.+]])
  94. // CHECK: [[CAST:%.+]] = bitcast [2 x [[KMP_DIM]]]* [[DIMS]] to i8*
  95. // CHECK: call void @llvm.memset.p0i8.i64(i8* align 8 [[CAST]], i8 0, i64 48, i1 false)
  96. // CHECK: [[DIM:%.+]] = getelementptr inbounds [2 x [[KMP_DIM]]], [2 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0
  97. // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1
  98. // CHECK: store i64 10, i64* %
  99. // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2
  100. // CHECK: store i64 1, i64* %
  101. // CHECK: [[DIM:%.+]] = getelementptr inbounds [2 x [[KMP_DIM]]], [2 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 1
  102. // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 1
  103. // CHECK: store i64 %{{.+}}, i64* %
  104. // CHECK: getelementptr inbounds [[KMP_DIM]], [[KMP_DIM]]* [[DIM]], i32 0, i32 2
  105. // CHECK: store i64 1, i64* %
  106. // CHECK: [[DIM:%.+]] = getelementptr inbounds [2 x [[KMP_DIM]]], [2 x [[KMP_DIM]]]* [[DIMS]], i64 0, i64 0
  107. // CHECK: [[CAST:%.+]] = bitcast [[KMP_DIM]]* [[DIM]] to i8*
  108. // CHECK: call void @__kmpc_doacross_init([[IDENT]], i32 [[GTID]], i32 2, i8* [[CAST]])
  109. // CHECK: call void @__kmpc_for_static_init_4(
  110. #pragma omp for ordered(2)
  111. for (T j = 0; j < M; j++)
  112. for (i = 0; i < n; i += 2) {
  113. a[i][j] = foo(i, j);
  114. // CHECK: invoke {{.+TestStruct.+foo}}
  115. // CHECK: load i32*, i32** %
  116. // CHECK: load i32, i32* %
  117. // CHECK: load i32, i32* %
  118. // CHECK: load i32, i32* [[J:%.+]],
  119. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  120. // CHECK-NEXT: sdiv i32 %{{.+}}, 1
  121. // CHECK-NEXT: sext i32 %{{.+}} to i64
  122. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT:%.+]], i64 0, i64 0
  123. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP:%.+]],
  124. // CHECK-NEXT: [[I:%.+]] = load i32*, i32** [[I_REF:%.+]],
  125. // CHECK-NEXT: load i32, i32* [[I]],
  126. // CHECK-NEXT: sub nsw i32 %{{.+}}, 2
  127. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  128. // CHECK-NEXT: sdiv i32 %{{.+}}, 2
  129. // CHECK-NEXT: sext i32 %{{.+}} to i64
  130. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 1
  131. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
  132. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 0
  133. // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]])
  134. // CHECK-NEXT: load i32, i32* [[J:%.+]],
  135. // CHECK-NEXT: sub nsw i32 %{{.+}}, 1
  136. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  137. // CHECK-NEXT: sdiv i32 %{{.+}}, 1
  138. // CHECK-NEXT: sext i32 %{{.+}} to i64
  139. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT:%.+]], i64 0, i64 0
  140. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP:%.+]],
  141. // CHECK-NEXT: [[I:%.+]] = load i32*, i32** [[I_REF]],
  142. // CHECK-NEXT: load i32, i32* [[I]],
  143. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  144. // CHECK-NEXT: sdiv i32 %{{.+}}, 2
  145. // CHECK-NEXT: sext i32 %{{.+}} to i64
  146. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 1
  147. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
  148. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 0
  149. // CHECK-NEXT: call void @__kmpc_doacross_wait([[IDENT]], i32 [[GTID]], i64* [[TMP]])
  150. #pragma omp ordered depend(sink : j, i - 2) depend(sink : j - 1, i)
  151. b[i][j] = bar(a[i][j], b[i - 1][j], b[i][j - 1]);
  152. // CHECK: invoke {{.+TestStruct.+bar}}
  153. // CHECK: load i32*, i32** %
  154. // CHECK: load i32, i32* %
  155. // CHECK: load i32, i32* %
  156. // CHECK: load i32, i32* [[J]],
  157. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  158. // CHECK-NEXT: sdiv i32 %{{.+}}, 1
  159. // CHECK-NEXT: sext i32 %{{.+}} to i64
  160. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT:%.+]], i64 0, i64 0
  161. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP:%.+]],
  162. // CHECK-NEXT: [[I:%.+]] = load i32*, i32** [[I_REF]],
  163. // CHECK-NEXT: load i32, i32* [[I]],
  164. // CHECK-NEXT: sub nsw i32 %{{.+}}, 0
  165. // CHECK-NEXT: sdiv i32 %{{.+}}, 2
  166. // CHECK-NEXT: sext i32 %{{.+}} to i64
  167. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 1
  168. // CHECK-NEXT: store i64 %{{.+}}, i64* [[TMP]],
  169. // CHECK-NEXT: [[TMP:%.+]] = getelementptr inbounds [2 x i64], [2 x i64]* [[CNT]], i64 0, i64 0
  170. // CHECK-NEXT: call void @__kmpc_doacross_post([[IDENT]], i32 [[GTID]], i64* [[TMP]])
  171. #pragma omp ordered depend(source)
  172. baz(a[i][j], b[i][j]);
  173. }
  174. }
  175. // CHECK: landingpad
  176. // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]])
  177. // CHECK: br label %
  178. // CHECK: call void @__kmpc_for_static_fini(
  179. // CHECK: call void @__kmpc_doacross_fini([[IDENT]], i32 [[GTID]])
  180. // CHECK: ret
  181. };
  182. TestStruct<int> s;
  183. #endif // HEADER