1 // NOTE: Assertions have been autogenerated by utils/update_cc_test_checks.py UTC_ARGS: --function-signature --include-generated-funcs --replace-value-regex "__omp_offloading_[0-9a-z]+_[0-9a-z]+" 2 // RUN: %clang_cc1 -no-opaque-pointers -verify -fopenmp -x c++ -triple powerpc64le-unknown-unknown -fopenmp-targets=nvptx64-nvidia-cuda -fopenmp-offload-mandatory -emit-llvm %s -o - | FileCheck %s --check-prefix=MANDATORY 3 // expected-no-diagnostics 4 5 void foo() {} 6 #pragma omp declare target(foo) 7 8 void bar() {} 9 #pragma omp declare target device_type(nohost) to(bar) 10 11 void host() { 12 #pragma omp target 13 { bar(); } 14 } 15 16 void host_if(bool cond) { 17 #pragma omp target if(cond) 18 { bar(); } 19 } 20 21 void host_dev(int device) { 22 #pragma omp target device(device) 23 { bar(); } 24 } 25 // MANDATORY-LABEL: define {{[^@]+}}@_Z3foov 26 // MANDATORY-SAME: () #[[ATTR0:[0-9]+]] { 27 // MANDATORY-NEXT: entry: 28 // MANDATORY-NEXT: ret void 29 // 30 // 31 // MANDATORY-LABEL: define {{[^@]+}}@_Z4hostv 32 // MANDATORY-SAME: () #[[ATTR0]] { 33 // MANDATORY-NEXT: entry: 34 // MANDATORY-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8 35 // MANDATORY-NEXT: [[TMP0:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 0 36 // MANDATORY-NEXT: store i32 1, i32* [[TMP0]], align 4 37 // MANDATORY-NEXT: [[TMP1:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 1 38 // MANDATORY-NEXT: store i32 0, i32* [[TMP1]], align 4 39 // MANDATORY-NEXT: [[TMP2:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 2 40 // MANDATORY-NEXT: store i8** null, i8*** [[TMP2]], align 8 41 // MANDATORY-NEXT: [[TMP3:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 3 42 // MANDATORY-NEXT: store i8** null, i8*** [[TMP3]], align 8 43 // MANDATORY-NEXT: [[TMP4:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 4 44 // MANDATORY-NEXT: store i64* null, i64** [[TMP4]], align 8 45 // MANDATORY-NEXT: [[TMP5:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 5 46 // MANDATORY-NEXT: store i64* null, i64** [[TMP5]], align 8 47 // MANDATORY-NEXT: [[TMP6:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 6 48 // MANDATORY-NEXT: store i8** null, i8*** [[TMP6]], align 8 49 // MANDATORY-NEXT: [[TMP7:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 7 50 // MANDATORY-NEXT: store i8** null, i8*** [[TMP7]], align 8 51 // MANDATORY-NEXT: [[TMP8:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 8 52 // MANDATORY-NEXT: store i64 0, i64* [[TMP8]], align 8 53 // MANDATORY-NEXT: [[TMP9:%.*]] = call i32 @__tgt_target_kernel(%struct.ident_t* @[[GLOB1:[0-9]+]], i64 -1, i32 -1, i32 0, i8* @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4hostv_l12.region_id, %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]]) 54 // MANDATORY-NEXT: [[TMP10:%.*]] = icmp ne i32 [[TMP9]], 0 55 // MANDATORY-NEXT: br i1 [[TMP10]], label [[OMP_OFFLOAD_FAILED:%.*]], label [[OMP_OFFLOAD_CONT:%.*]] 56 // MANDATORY: omp_offload.failed: 57 // MANDATORY-NEXT: unreachable 58 // MANDATORY: omp_offload.cont: 59 // MANDATORY-NEXT: ret void 60 // 61 // 62 // MANDATORY-LABEL: define {{[^@]+}}@_Z7host_ifb 63 // MANDATORY-SAME: (i1 noundef zeroext [[COND:%.*]]) #[[ATTR0]] { 64 // MANDATORY-NEXT: entry: 65 // MANDATORY-NEXT: [[COND_ADDR:%.*]] = alloca i8, align 1 66 // MANDATORY-NEXT: [[FROMBOOL:%.*]] = zext i1 [[COND]] to i8 67 // MANDATORY-NEXT: store i8 [[FROMBOOL]], i8* [[COND_ADDR]], align 1 68 // MANDATORY-NEXT: [[TMP0:%.*]] = load i8, i8* [[COND_ADDR]], align 1 69 // MANDATORY-NEXT: [[TOBOOL:%.*]] = trunc i8 [[TMP0]] to i1 70 // MANDATORY-NEXT: br i1 [[TOBOOL]], label [[OMP_IF_THEN:%.*]], label [[OMP_IF_ELSE:%.*]] 71 // MANDATORY: omp_if.then: 72 // MANDATORY-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8 73 // MANDATORY-NEXT: [[TMP1:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 0 74 // MANDATORY-NEXT: store i32 1, i32* [[TMP1]], align 4 75 // MANDATORY-NEXT: [[TMP2:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 1 76 // MANDATORY-NEXT: store i32 0, i32* [[TMP2]], align 4 77 // MANDATORY-NEXT: [[TMP3:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 2 78 // MANDATORY-NEXT: store i8** null, i8*** [[TMP3]], align 8 79 // MANDATORY-NEXT: [[TMP4:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 3 80 // MANDATORY-NEXT: store i8** null, i8*** [[TMP4]], align 8 81 // MANDATORY-NEXT: [[TMP5:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 4 82 // MANDATORY-NEXT: store i64* null, i64** [[TMP5]], align 8 83 // MANDATORY-NEXT: [[TMP6:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 5 84 // MANDATORY-NEXT: store i64* null, i64** [[TMP6]], align 8 85 // MANDATORY-NEXT: [[TMP7:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 6 86 // MANDATORY-NEXT: store i8** null, i8*** [[TMP7]], align 8 87 // MANDATORY-NEXT: [[TMP8:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 7 88 // MANDATORY-NEXT: store i8** null, i8*** [[TMP8]], align 8 89 // MANDATORY-NEXT: [[TMP9:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 8 90 // MANDATORY-NEXT: store i64 0, i64* [[TMP9]], align 8 91 // MANDATORY-NEXT: [[TMP10:%.*]] = call i32 @__tgt_target_kernel(%struct.ident_t* @[[GLOB1]], i64 -1, i32 -1, i32 0, i8* @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z7host_ifb_l17.region_id, %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]]) 92 // MANDATORY-NEXT: [[TMP11:%.*]] = icmp ne i32 [[TMP10]], 0 93 // MANDATORY-NEXT: br i1 [[TMP11]], label [[OMP_OFFLOAD_FAILED:%.*]], label [[OMP_OFFLOAD_CONT:%.*]] 94 // MANDATORY: omp_offload.failed: 95 // MANDATORY-NEXT: unreachable 96 // MANDATORY: omp_offload.cont: 97 // MANDATORY-NEXT: br label [[OMP_IF_END:%.*]] 98 // MANDATORY: omp_if.else: 99 // MANDATORY-NEXT: unreachable 100 // MANDATORY: omp_if.end: 101 // MANDATORY-NEXT: ret void 102 // 103 // 104 // MANDATORY-LABEL: define {{[^@]+}}@_Z8host_devi 105 // MANDATORY-SAME: (i32 noundef signext [[DEVICE:%.*]]) #[[ATTR0]] { 106 // MANDATORY-NEXT: entry: 107 // MANDATORY-NEXT: [[DEVICE_ADDR:%.*]] = alloca i32, align 4 108 // MANDATORY-NEXT: [[DOTCAPTURE_EXPR_:%.*]] = alloca i32, align 4 109 // MANDATORY-NEXT: store i32 [[DEVICE]], i32* [[DEVICE_ADDR]], align 4 110 // MANDATORY-NEXT: [[TMP0:%.*]] = load i32, i32* [[DEVICE_ADDR]], align 4 111 // MANDATORY-NEXT: store i32 [[TMP0]], i32* [[DOTCAPTURE_EXPR_]], align 4 112 // MANDATORY-NEXT: [[TMP1:%.*]] = load i32, i32* [[DOTCAPTURE_EXPR_]], align 4 113 // MANDATORY-NEXT: [[TMP2:%.*]] = sext i32 [[TMP1]] to i64 114 // MANDATORY-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8 115 // MANDATORY-NEXT: [[TMP3:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 0 116 // MANDATORY-NEXT: store i32 1, i32* [[TMP3]], align 4 117 // MANDATORY-NEXT: [[TMP4:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 1 118 // MANDATORY-NEXT: store i32 0, i32* [[TMP4]], align 4 119 // MANDATORY-NEXT: [[TMP5:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 2 120 // MANDATORY-NEXT: store i8** null, i8*** [[TMP5]], align 8 121 // MANDATORY-NEXT: [[TMP6:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 3 122 // MANDATORY-NEXT: store i8** null, i8*** [[TMP6]], align 8 123 // MANDATORY-NEXT: [[TMP7:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 4 124 // MANDATORY-NEXT: store i64* null, i64** [[TMP7]], align 8 125 // MANDATORY-NEXT: [[TMP8:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 5 126 // MANDATORY-NEXT: store i64* null, i64** [[TMP8]], align 8 127 // MANDATORY-NEXT: [[TMP9:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 6 128 // MANDATORY-NEXT: store i8** null, i8*** [[TMP9]], align 8 129 // MANDATORY-NEXT: [[TMP10:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 7 130 // MANDATORY-NEXT: store i8** null, i8*** [[TMP10]], align 8 131 // MANDATORY-NEXT: [[TMP11:%.*]] = getelementptr inbounds [[STRUCT___TGT_KERNEL_ARGUMENTS]], %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]], i32 0, i32 8 132 // MANDATORY-NEXT: store i64 0, i64* [[TMP11]], align 8 133 // MANDATORY-NEXT: [[TMP12:%.*]] = call i32 @__tgt_target_kernel(%struct.ident_t* @[[GLOB1]], i64 [[TMP2]], i32 -1, i32 0, i8* @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z8host_devi_l22.region_id, %struct.__tgt_kernel_arguments* [[KERNEL_ARGS]]) 134 // MANDATORY-NEXT: [[TMP13:%.*]] = icmp ne i32 [[TMP12]], 0 135 // MANDATORY-NEXT: br i1 [[TMP13]], label [[OMP_OFFLOAD_FAILED:%.*]], label [[OMP_OFFLOAD_CONT:%.*]] 136 // MANDATORY: omp_offload.failed: 137 // MANDATORY-NEXT: unreachable 138 // MANDATORY: omp_offload.cont: 139 // MANDATORY-NEXT: ret void 140 // 141 // 142 // MANDATORY-LABEL: define {{[^@]+}}@.omp_offloading.requires_reg 143 // MANDATORY-SAME: () #[[ATTR3:[0-9]+]] { 144 // MANDATORY-NEXT: entry: 145 // MANDATORY-NEXT: call void @__tgt_register_requires(i64 1) 146 // MANDATORY-NEXT: ret void 147 // 148