clang 23.0.0git
SemaAMDGPU.cpp
Go to the documentation of this file.
1//===------ SemaAMDGPU.cpp ------- AMDGPU target-specific routines --------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements semantic analysis functions specific to AMDGPU.
10//
11//===----------------------------------------------------------------------===//
12
14#include "clang/AST/Decl.h"
16#include "clang/AST/Expr.h"
22#include "clang/Sema/Scope.h"
23#include "clang/Sema/Sema.h"
24#include "llvm/ADT/SmallVector.h"
25#include "llvm/ADT/StringExtras.h"
26#include "llvm/ADT/StringMap.h"
27#include "llvm/Support/AMDGPUAddrSpace.h"
28#include "llvm/Support/AtomicOrdering.h"
29#include "llvm/TargetParser/TargetParser.h"
30#include <cstdint>
31#include <utility>
32
33namespace clang {
34
36
38 CallExpr *TheCall) {
39 // position of memory order and scope arguments in the builtin
40 unsigned OrderIndex, ScopeIndex;
41
42 const auto *FD = SemaRef.getCurFunctionDecl(/*AllowLambda=*/true);
43 assert(FD && "AMDGPU builtins should not be used outside of a function");
44 llvm::StringMap<bool> CallerFeatureMap;
45 getASTContext().getFunctionFeatureMap(CallerFeatureMap, FD);
46 bool HasGFX950Insts =
47 Builtin::evaluateRequiredTargetFeatures("gfx950-insts", CallerFeatureMap);
48
49 switch (BuiltinID) {
50 case AMDGPU::BI__builtin_amdgcn_raw_ptr_buffer_load_lds:
51 case AMDGPU::BI__builtin_amdgcn_raw_ptr_buffer_load_async_lds:
52 case AMDGPU::BI__builtin_amdgcn_struct_ptr_buffer_load_lds:
53 case AMDGPU::BI__builtin_amdgcn_struct_ptr_buffer_load_async_lds:
54 case AMDGPU::BI__builtin_amdgcn_load_to_lds:
55 case AMDGPU::BI__builtin_amdgcn_load_async_to_lds:
56 case AMDGPU::BI__builtin_amdgcn_global_load_lds:
57 case AMDGPU::BI__builtin_amdgcn_global_load_async_lds: {
58 constexpr const int SizeIdx = 2;
59 llvm::APSInt Size;
60 Expr *ArgExpr = TheCall->getArg(SizeIdx);
61 // Check for instantiation-dependent expressions (e.g., involving template
62 // parameters). These will be checked again during template instantiation.
63 if (ArgExpr->isInstantiationDependent())
64 return false;
65 [[maybe_unused]] ExprResult R =
66 SemaRef.VerifyIntegerConstantExpression(ArgExpr, &Size);
67 assert(!R.isInvalid());
68 switch (Size.getSExtValue()) {
69 case 1:
70 case 2:
71 case 4:
72 return false;
73 case 12:
74 case 16: {
75 if (HasGFX950Insts)
76 return false;
77 [[fallthrough]];
78 }
79 default:
80 SemaRef.targetDiag(ArgExpr->getExprLoc(),
81 diag::err_amdgcn_load_lds_size_invalid_value)
82 << ArgExpr->getSourceRange();
83 SemaRef.targetDiag(ArgExpr->getExprLoc(),
84 diag::note_amdgcn_load_lds_size_valid_value)
85 << HasGFX950Insts << ArgExpr->getSourceRange();
86 return true;
87 }
88 }
89 case AMDGPU::BI__builtin_amdgcn_get_fpenv:
90 case AMDGPU::BI__builtin_amdgcn_set_fpenv:
91 return false;
92 case AMDGPU::BI__builtin_amdgcn_atomic_inc32:
93 case AMDGPU::BI__builtin_amdgcn_atomic_inc64:
94 case AMDGPU::BI__builtin_amdgcn_atomic_dec32:
95 case AMDGPU::BI__builtin_amdgcn_atomic_dec64:
96 OrderIndex = 2;
97 ScopeIndex = 3;
98 break;
99 case AMDGPU::BI__builtin_amdgcn_fence:
100 OrderIndex = 0;
101 ScopeIndex = 1;
102 break;
103 case AMDGPU::BI__builtin_amdgcn_s_setreg:
104 return SemaRef.BuiltinConstantArgRange(TheCall, /*ArgNum=*/0, /*Low=*/0,
105 /*High=*/UINT16_MAX);
106 case AMDGPU::BI__builtin_amdgcn_s_wait_event: {
107 llvm::APSInt Result;
108 if (SemaRef.BuiltinConstantArg(TheCall, 0, Result))
109 return true;
110
112 "gfx12-insts", CallerFeatureMap);
113
114 // gfx11 -> gfx12 changed the interpretation of the bitmask. gfx12 inverted
115 // the intepretation for export_ready, but shifted the used bit by 1. Thus
116 // waiting for the export_ready event can use a value of 2 universally.
117 if (((IsGFX12Plus && !Result[1]) || (!IsGFX12Plus && Result[0])) ||
118 Result.getZExtValue() > 2) {
119 Expr *ArgExpr = TheCall->getArg(0);
120 SemaRef.targetDiag(ArgExpr->getExprLoc(),
121 diag::warn_amdgpu_s_wait_event_mask_no_effect_target)
122 << ArgExpr->getSourceRange();
123 SemaRef.targetDiag(ArgExpr->getExprLoc(),
124 diag::note_amdgpu_s_wait_event_suggested_value)
125 << ArgExpr->getSourceRange();
126 }
127
128 return false;
129 }
130 case AMDGPU::BI__builtin_amdgcn_mov_dpp:
131 return checkMovDPPFunctionCall(TheCall, 5, 1);
132 case AMDGPU::BI__builtin_amdgcn_mov_dpp8:
133 return checkMovDPPFunctionCall(TheCall, 2, 1);
134 case AMDGPU::BI__builtin_amdgcn_update_dpp:
135 return checkMovDPPFunctionCall(TheCall, 6, 2);
136 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f16_fp8:
137 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_bf16_fp8:
138 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f16_bf8:
139 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_bf16_bf8:
140 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f16_fp4:
141 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_bf16_fp4:
142 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f32_fp8:
143 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f32_bf8:
144 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f32_fp4:
145 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f16_fp6:
146 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_bf16_fp6:
147 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f16_bf6:
148 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_bf16_bf6:
149 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f32_fp6:
150 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f32_bf6:
151 return SemaRef.BuiltinConstantArgRange(TheCall, 2, 0, 15);
152 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_load_32x4B:
153 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_load_16x8B:
154 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_load_8x16B:
155 return checkCoopAtomicFunctionCall(TheCall, /*IsStore=*/false);
156 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_store_32x4B:
157 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_store_16x8B:
158 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_store_8x16B:
159 return checkCoopAtomicFunctionCall(TheCall, /*IsStore=*/true);
160 case AMDGPU::BI__builtin_amdgcn_flat_load_monitor_b32:
161 case AMDGPU::BI__builtin_amdgcn_flat_load_monitor_b64:
162 case AMDGPU::BI__builtin_amdgcn_flat_load_monitor_b128:
163 case AMDGPU::BI__builtin_amdgcn_global_load_monitor_b32:
164 case AMDGPU::BI__builtin_amdgcn_global_load_monitor_b64:
165 case AMDGPU::BI__builtin_amdgcn_global_load_monitor_b128:
166 return checkAtomicMonitorLoad(TheCall);
167 case AMDGPU::BI__builtin_amdgcn_image_load_1d_v4f32_i32:
168 case AMDGPU::BI__builtin_amdgcn_image_load_1darray_v4f32_i32:
169 case AMDGPU::BI__builtin_amdgcn_image_load_1d_v4f16_i32:
170 case AMDGPU::BI__builtin_amdgcn_image_load_1darray_v4f16_i32:
171 case AMDGPU::BI__builtin_amdgcn_image_load_2d_f32_i32:
172 case AMDGPU::BI__builtin_amdgcn_image_load_2d_v4f32_i32:
173 case AMDGPU::BI__builtin_amdgcn_image_load_2d_v4f16_i32:
174 case AMDGPU::BI__builtin_amdgcn_image_load_2darray_f32_i32:
175 case AMDGPU::BI__builtin_amdgcn_image_load_2darray_v4f32_i32:
176 case AMDGPU::BI__builtin_amdgcn_image_load_2darray_v4f16_i32:
177 case AMDGPU::BI__builtin_amdgcn_image_load_3d_v4f32_i32:
178 case AMDGPU::BI__builtin_amdgcn_image_load_3d_v4f16_i32:
179 case AMDGPU::BI__builtin_amdgcn_image_load_cube_v4f32_i32:
180 case AMDGPU::BI__builtin_amdgcn_image_load_cube_v4f16_i32:
181 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1d_v4f32_i32:
182 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1d_v4f16_i32:
183 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1darray_v4f32_i32:
184 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1darray_v4f16_i32:
185 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2d_f32_i32:
186 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2d_v4f32_i32:
187 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2d_v4f16_i32:
188 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2darray_f32_i32:
189 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2darray_v4f32_i32:
190 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2darray_v4f16_i32:
191 case AMDGPU::BI__builtin_amdgcn_image_load_mip_3d_v4f32_i32:
192 case AMDGPU::BI__builtin_amdgcn_image_load_mip_3d_v4f16_i32:
193 case AMDGPU::BI__builtin_amdgcn_image_load_mip_cube_v4f32_i32:
194 case AMDGPU::BI__builtin_amdgcn_image_load_mip_cube_v4f16_i32:
195 case AMDGPU::BI__builtin_amdgcn_image_sample_1d_v4f32_f32:
196 case AMDGPU::BI__builtin_amdgcn_image_sample_1darray_v4f32_f32:
197 case AMDGPU::BI__builtin_amdgcn_image_sample_1d_v4f16_f32:
198 case AMDGPU::BI__builtin_amdgcn_image_sample_1darray_v4f16_f32:
199 case AMDGPU::BI__builtin_amdgcn_image_sample_2d_f32_f32:
200 case AMDGPU::BI__builtin_amdgcn_image_sample_2d_v4f32_f32:
201 case AMDGPU::BI__builtin_amdgcn_image_sample_2d_v4f16_f32:
202 case AMDGPU::BI__builtin_amdgcn_image_sample_2darray_f32_f32:
203 case AMDGPU::BI__builtin_amdgcn_image_sample_2darray_v4f32_f32:
204 case AMDGPU::BI__builtin_amdgcn_image_sample_2darray_v4f16_f32:
205 case AMDGPU::BI__builtin_amdgcn_image_sample_3d_v4f32_f32:
206 case AMDGPU::BI__builtin_amdgcn_image_sample_3d_v4f16_f32:
207 case AMDGPU::BI__builtin_amdgcn_image_sample_cube_v4f32_f32:
208 case AMDGPU::BI__builtin_amdgcn_image_sample_cube_v4f16_f32:
209 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1d_v4f32_f32:
210 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1d_v4f16_f32:
211 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1darray_v4f32_f32:
212 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1darray_v4f16_f32:
213 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2d_f32_f32:
214 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2d_v4f32_f32:
215 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2d_v4f16_f32:
216 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2darray_f32_f32:
217 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2darray_v4f32_f32:
218 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2darray_v4f16_f32:
219 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_3d_v4f32_f32:
220 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_3d_v4f16_f32:
221 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_cube_v4f32_f32:
222 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_cube_v4f16_f32:
223 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1d_v4f32_f32:
224 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1d_v4f16_f32:
225 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1darray_v4f32_f32:
226 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1darray_v4f16_f32:
227 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2d_f32_f32:
228 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2d_v4f16_f32:
229 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2d_v4f32_f32:
230 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2darray_f32_f32:
231 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2darray_v4f32_f32:
232 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2darray_v4f16_f32:
233 case AMDGPU::BI__builtin_amdgcn_image_sample_l_3d_v4f32_f32:
234 case AMDGPU::BI__builtin_amdgcn_image_sample_l_3d_v4f16_f32:
235 case AMDGPU::BI__builtin_amdgcn_image_sample_l_cube_v4f32_f32:
236 case AMDGPU::BI__builtin_amdgcn_image_sample_l_cube_v4f16_f32:
237 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1d_v4f32_f32:
238 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1d_v4f16_f32:
239 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1darray_v4f32_f32:
240 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1darray_v4f16_f32:
241 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2d_f32_f32:
242 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2d_v4f32_f32:
243 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2d_v4f16_f32:
244 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2darray_f32_f32:
245 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2darray_v4f32_f32:
246 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2darray_v4f16_f32:
247 case AMDGPU::BI__builtin_amdgcn_image_sample_d_3d_v4f32_f32:
248 case AMDGPU::BI__builtin_amdgcn_image_sample_d_3d_v4f16_f32:
249 case AMDGPU::BI__builtin_amdgcn_image_gather4_lz_2d_v4f32_f32: {
250 StringRef FeatureList(
251 getASTContext().BuiltinInfo.getRequiredFeatures(BuiltinID));
253 CallerFeatureMap)) {
254 Diag(TheCall->getBeginLoc(), diag::err_builtin_needs_feature)
255 << FD->getDeclName() << FeatureList;
256 return false;
257 }
258
259 unsigned ArgCount = TheCall->getNumArgs() - 1;
260 llvm::APSInt Result;
261
262 // Compilain about dmask values which are too huge to fully fit into 4 bits
263 // (which is the actual size of the dmask in corresponding HW instructions).
264 constexpr unsigned DMaskArgNo = 0;
265 constexpr int Low = 0;
266 constexpr int High = 15;
267 if (SemaRef.BuiltinConstantArg(TheCall, DMaskArgNo, Result) ||
268 SemaRef.BuiltinConstantArgRange(TheCall, DMaskArgNo, Low, High,
269 /* RangeIsError = */ true))
270 return true;
271
272 // Dmask indicates which elements should be returned and it is not possible
273 // to return more values than there are elements in return type.
274 int NumElementsInRetTy = 1;
275 const Type *RetTy = TheCall->getType().getTypePtr();
276 if (auto *VTy = dyn_cast<VectorType>(RetTy))
277 NumElementsInRetTy = VTy->getNumElements();
278 int NumActiveBitsInDMask =
279 llvm::popcount(static_cast<uint8_t>(Result.getExtValue()));
280 if (NumActiveBitsInDMask > NumElementsInRetTy) {
281 Diag(TheCall->getBeginLoc(),
282 diag::err_amdgcn_dmask_has_too_many_bits_set);
283 return true;
284 }
285
286 // For gather, only one bit can be set indicating which exact component to
287 // return.
288 bool ExtraGatherChecks =
289 BuiltinID == AMDGPU::BI__builtin_amdgcn_image_gather4_lz_2d_v4f32_f32 &&
290 SemaRef.BuiltinConstantArgPower2(TheCall, 0);
291
292 return ExtraGatherChecks ||
293 (SemaRef.BuiltinConstantArg(TheCall, ArgCount, Result)) ||
294 (SemaRef.BuiltinConstantArg(TheCall, (ArgCount - 1), Result));
295 }
296 case AMDGPU::BI__builtin_amdgcn_image_store_1d_v4f32_i32:
297 case AMDGPU::BI__builtin_amdgcn_image_store_1darray_v4f32_i32:
298 case AMDGPU::BI__builtin_amdgcn_image_store_1d_v4f16_i32:
299 case AMDGPU::BI__builtin_amdgcn_image_store_1darray_v4f16_i32:
300 case AMDGPU::BI__builtin_amdgcn_image_store_2d_f32_i32:
301 case AMDGPU::BI__builtin_amdgcn_image_store_2d_v4f32_i32:
302 case AMDGPU::BI__builtin_amdgcn_image_store_2d_v4f16_i32:
303 case AMDGPU::BI__builtin_amdgcn_image_store_2darray_f32_i32:
304 case AMDGPU::BI__builtin_amdgcn_image_store_2darray_v4f32_i32:
305 case AMDGPU::BI__builtin_amdgcn_image_store_2darray_v4f16_i32:
306 case AMDGPU::BI__builtin_amdgcn_image_store_3d_v4f32_i32:
307 case AMDGPU::BI__builtin_amdgcn_image_store_3d_v4f16_i32:
308 case AMDGPU::BI__builtin_amdgcn_image_store_cube_v4f32_i32:
309 case AMDGPU::BI__builtin_amdgcn_image_store_cube_v4f16_i32:
310 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1d_v4f32_i32:
311 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1d_v4f16_i32:
312 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1darray_v4f32_i32:
313 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1darray_v4f16_i32:
314 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2d_f32_i32:
315 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2d_v4f32_i32:
316 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2d_v4f16_i32:
317 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2darray_f32_i32:
318 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2darray_v4f32_i32:
319 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2darray_v4f16_i32:
320 case AMDGPU::BI__builtin_amdgcn_image_store_mip_3d_v4f32_i32:
321 case AMDGPU::BI__builtin_amdgcn_image_store_mip_3d_v4f16_i32:
322 case AMDGPU::BI__builtin_amdgcn_image_store_mip_cube_v4f32_i32:
323 case AMDGPU::BI__builtin_amdgcn_image_store_mip_cube_v4f16_i32: {
324 StringRef FeatureList(
325 getASTContext().BuiltinInfo.getRequiredFeatures(BuiltinID));
327 CallerFeatureMap)) {
328 Diag(TheCall->getBeginLoc(), diag::err_builtin_needs_feature)
329 << FD->getDeclName() << FeatureList;
330 return false;
331 }
332
333 unsigned ArgCount = TheCall->getNumArgs() - 1;
334 llvm::APSInt Result;
335
336 // Complain about dmask values which are too huge to fully fit into 4 bits
337 // (which is the actual size of the dmask in corresponding HW instructions).
338 constexpr unsigned DMaskArgNo = 1;
339 return SemaRef.BuiltinConstantArgRange(TheCall, DMaskArgNo, /*Low=*/0,
340 /*High=*/15,
341 /*RangeIsError=*/true) ||
342 SemaRef.BuiltinConstantArg(TheCall, ArgCount, Result) ||
343 SemaRef.BuiltinConstantArg(TheCall, (ArgCount - 1), Result);
344 }
345 case AMDGPU::BI__builtin_amdgcn_wmma_i32_16x16x64_iu8:
346 case AMDGPU::BI__builtin_amdgcn_swmmac_i32_16x16x128_iu8: {
347 if (BuiltinID == AMDGPU::BI__builtin_amdgcn_wmma_i32_16x16x64_iu8) {
348 if (SemaRef.checkArgCountRange(TheCall, 7, 8))
349 return true;
350 if (TheCall->getNumArgs() == 7)
351 return false;
352 } else if (BuiltinID ==
353 AMDGPU::BI__builtin_amdgcn_swmmac_i32_16x16x128_iu8) {
354 if (SemaRef.checkArgCountRange(TheCall, 8, 9))
355 return true;
356 if (TheCall->getNumArgs() == 8)
357 return false;
358 }
359 // Check if the last argument (clamp operand) is a constant and is
360 // convertible to bool.
361 Expr *ClampArg = TheCall->getArg(TheCall->getNumArgs() - 1);
362 // 1) Ensure clamp argument is a constant expression
363 llvm::APSInt ClampValue;
364 if (!SemaRef.VerifyIntegerConstantExpression(ClampArg, &ClampValue)
365 .isUsable())
366 return true;
367 // 2) Check if the argument can be converted to bool type
368 if (!SemaRef.Context.hasSameType(ClampArg->getType(),
369 SemaRef.Context.BoolTy)) {
370 // Try to convert to bool
371 QualType BoolTy = SemaRef.Context.BoolTy;
372 ExprResult ClampExpr(ClampArg);
373 SemaRef.CheckSingleAssignmentConstraints(BoolTy, ClampExpr);
374 if (ClampExpr.isInvalid())
375 return true;
376 }
377 return false;
378 }
379 case AMDGPU::BI__builtin_amdgcn_wmma_f32_16x16x32_bf16:
380 case AMDGPU::BI__builtin_amdgcn_wmma_f32_16x16x4_f32:
381 case AMDGPU::BI__builtin_amdgcn_wmma_f32_16x16x32_f16:
382 case AMDGPU::BI__builtin_amdgcn_wmma_f16_16x16x32_f16:
383 case AMDGPU::BI__builtin_amdgcn_wmma_bf16_16x16x32_bf16:
384 case AMDGPU::BI__builtin_amdgcn_wmma_bf16f32_16x16x32_bf16:
385 return SemaRef.BuiltinConstantArgRange(TheCall, /*ArgNum=*/0, /*Low=*/0,
386 /*High=*/0) ||
387 SemaRef.BuiltinConstantArgRange(TheCall, /*ArgNum=*/2, /*Low=*/0,
388 /*High=*/0);
389 default:
390 return false;
391 }
392
393 ExprResult Arg = TheCall->getArg(OrderIndex);
394 auto ArgExpr = Arg.get();
395 Expr::EvalResult ArgResult;
396
397 if (!ArgExpr->EvaluateAsInt(ArgResult, getASTContext()))
398 return Diag(ArgExpr->getExprLoc(), diag::err_typecheck_expect_int)
399 << ArgExpr->getType();
400 auto Ord = ArgResult.Val.getInt().getZExtValue();
401
402 // Check validity of memory ordering as per C11 / C++11's memory model.
403 // Only fence needs check. Atomic dec/inc allow all memory orders.
404 if (!llvm::isValidAtomicOrderingCABI(Ord))
405 return Diag(ArgExpr->getBeginLoc(),
406 diag::warn_atomic_op_has_invalid_memory_order)
407 << 0 << ArgExpr->getSourceRange();
408 switch (static_cast<llvm::AtomicOrderingCABI>(Ord)) {
409 case llvm::AtomicOrderingCABI::relaxed:
410 case llvm::AtomicOrderingCABI::consume:
411 if (BuiltinID == AMDGPU::BI__builtin_amdgcn_fence)
412 return Diag(ArgExpr->getBeginLoc(),
413 diag::warn_atomic_op_has_invalid_memory_order)
414 << 0 << ArgExpr->getSourceRange();
415 break;
416 case llvm::AtomicOrderingCABI::acquire:
417 case llvm::AtomicOrderingCABI::release:
418 case llvm::AtomicOrderingCABI::acq_rel:
419 case llvm::AtomicOrderingCABI::seq_cst:
420 break;
421 }
422
423 Arg = TheCall->getArg(ScopeIndex);
424 ArgExpr = Arg.get();
425 Expr::EvalResult ArgResult1;
426 // Check that sync scope is a constant literal
427 if (!ArgExpr->EvaluateAsConstantExpr(ArgResult1, getASTContext()))
428 return Diag(ArgExpr->getExprLoc(), diag::err_expr_not_string_literal)
429 << ArgExpr->getType();
430
431 return false;
432}
433
435 bool MayStore) {
436 Expr::EvalResult AtomicOrdArgRes;
437 if (!E->EvaluateAsInt(AtomicOrdArgRes, getASTContext()))
438 llvm_unreachable("Intrinsic requires imm for atomic ordering argument!");
439 auto Ord =
440 llvm::AtomicOrderingCABI(AtomicOrdArgRes.Val.getInt().getZExtValue());
441
442 // Atomic ordering cannot be acq_rel in any case, acquire for stores or
443 // release for loads.
444 if (!llvm::isValidAtomicOrderingCABI((unsigned)Ord) ||
445 (!(MayLoad && MayStore) && (Ord == llvm::AtomicOrderingCABI::acq_rel)) ||
446 (!MayLoad && Ord == llvm::AtomicOrderingCABI::acquire) ||
447 (!MayStore && Ord == llvm::AtomicOrderingCABI::release)) {
448 return Diag(E->getBeginLoc(), diag::warn_atomic_op_has_invalid_memory_order)
449 << 0 << E->getSourceRange();
450 }
451
452 return false;
453}
454
456 bool Fail = false;
457
458 // First argument is a global or generic pointer.
459 Expr *PtrArg = TheCall->getArg(0);
460 QualType PtrTy = PtrArg->getType()->getPointeeType();
461 unsigned AS = getASTContext().getTargetAddressSpace(PtrTy.getAddressSpace());
462 if (AS != llvm::AMDGPUAS::FLAT_ADDRESS &&
463 AS != llvm::AMDGPUAS::GLOBAL_ADDRESS) {
464 Fail = true;
465 Diag(TheCall->getBeginLoc(), diag::err_amdgcn_coop_atomic_invalid_as)
466 << PtrArg->getSourceRange();
467 }
468
469 Expr *AO = TheCall->getArg(IsStore ? 2 : 1);
470 Expr *Scope = TheCall->getArg(TheCall->getNumArgs() - 1);
471
472 if (AO->isValueDependent() || Scope->isValueDependent())
473 return false;
474
475 // Check atomic ordering
476 Fail |=
477 checkAtomicOrderingCABIArg(TheCall->getArg(IsStore ? 2 : 1),
478 /*MayLoad=*/!IsStore, /*MayStore=*/IsStore);
479
480 // Last argument is the syncscope as a string literal.
481 if (!isa<StringLiteral>(Scope->IgnoreParenImpCasts())) {
482 Diag(TheCall->getBeginLoc(), diag::err_expr_not_string_literal)
483 << Scope->getSourceRange();
484 Fail = true;
485 }
486
487 return Fail;
488}
489
491 bool Fail = false;
492
493 Expr *AO = TheCall->getArg(1);
494 Expr *Scope = TheCall->getArg(TheCall->getNumArgs() - 1);
495
496 if (AO->isValueDependent() || Scope->isValueDependent())
497 return false;
498
499 Fail |= checkAtomicOrderingCABIArg(TheCall->getArg(1), /*MayLoad=*/true,
500 /*MayStore=*/false);
501
503 if (std::optional<llvm::APSInt> Result =
504 Scope->getIntegerConstantExpr(SemaRef.Context)) {
505 if (!ScopeModel->isValid(Result->getZExtValue())) {
506 Diag(Scope->getBeginLoc(), diag::err_atomic_op_has_invalid_sync_scope)
507 << Scope->getSourceRange();
508 Fail = true;
509 }
510 }
511
512 return Fail;
513}
514
515bool SemaAMDGPU::checkMovDPPFunctionCall(CallExpr *TheCall, unsigned NumArgs,
516 unsigned NumDataArgs) {
517 assert(NumDataArgs <= 2);
518 if (SemaRef.checkArgCountRange(TheCall, NumArgs, NumArgs))
519 return true;
520 Expr *Args[2];
521 QualType ArgTys[2];
522 for (unsigned I = 0; I != NumDataArgs; ++I) {
523 Args[I] = TheCall->getArg(I);
524 ArgTys[I] = Args[I]->getType();
525 // TODO: Vectors can also be supported.
526 if (!ArgTys[I]->isArithmeticType() || ArgTys[I]->isAnyComplexType()) {
527 SemaRef.Diag(Args[I]->getBeginLoc(),
528 diag::err_typecheck_cond_expect_int_float)
529 << ArgTys[I] << Args[I]->getSourceRange();
530 return true;
531 }
532 }
533 if (NumDataArgs < 2)
534 return false;
535
536 if (getASTContext().hasSameUnqualifiedType(ArgTys[0], ArgTys[1]))
537 return false;
538
539 if (((ArgTys[0]->isUnsignedIntegerType() &&
540 ArgTys[1]->isSignedIntegerType()) ||
541 (ArgTys[0]->isSignedIntegerType() &&
542 ArgTys[1]->isUnsignedIntegerType())) &&
543 getASTContext().getTypeSize(ArgTys[0]) ==
544 getASTContext().getTypeSize(ArgTys[1]))
545 return false;
546
547 SemaRef.Diag(Args[1]->getBeginLoc(),
548 diag::err_typecheck_call_different_arg_types)
549 << ArgTys[0] << ArgTys[1];
550 return true;
551}
552
553static bool
555 const AMDGPUFlatWorkGroupSizeAttr &Attr) {
556 // Accept template arguments for now as they depend on something else.
557 // We'll get to check them when they eventually get instantiated.
558 if (MinExpr->isValueDependent() || MaxExpr->isValueDependent())
559 return false;
560
561 uint32_t Min = 0;
562 if (!S.checkUInt32Argument(Attr, MinExpr, Min, 0))
563 return true;
564
565 uint32_t Max = 0;
566 if (!S.checkUInt32Argument(Attr, MaxExpr, Max, 1))
567 return true;
568
569 if (Min == 0 && Max != 0) {
570 S.Diag(Attr.getLocation(), diag::err_attribute_argument_invalid)
571 << &Attr << 0;
572 return true;
573 }
574 if (Min > Max) {
575 S.Diag(Attr.getLocation(), diag::err_attribute_argument_invalid)
576 << &Attr << 1;
577 return true;
578 }
579
580 return false;
581}
582
583AMDGPUFlatWorkGroupSizeAttr *
585 Expr *MinExpr, Expr *MaxExpr) {
586 ASTContext &Context = getASTContext();
587 AMDGPUFlatWorkGroupSizeAttr TmpAttr(Context, CI, MinExpr, MaxExpr);
588
589 if (checkAMDGPUFlatWorkGroupSizeArguments(SemaRef, MinExpr, MaxExpr, TmpAttr))
590 return nullptr;
591 return ::new (Context)
592 AMDGPUFlatWorkGroupSizeAttr(Context, CI, MinExpr, MaxExpr);
593}
594
596 const AttributeCommonInfo &CI,
597 Expr *MinExpr, Expr *MaxExpr) {
598 if (auto *Attr = CreateAMDGPUFlatWorkGroupSizeAttr(CI, MinExpr, MaxExpr))
599 D->addAttr(Attr);
600}
601
603 const ParsedAttr &AL) {
604 Expr *MinExpr = AL.getArgAsExpr(0);
605 Expr *MaxExpr = AL.getArgAsExpr(1);
606
607 addAMDGPUFlatWorkGroupSizeAttr(D, AL, MinExpr, MaxExpr);
608}
609
610static bool checkAMDGPUWavesPerEUArguments(Sema &S, Expr *MinExpr,
611 Expr *MaxExpr,
612 const AMDGPUWavesPerEUAttr &Attr) {
613 if (S.DiagnoseUnexpandedParameterPack(MinExpr) ||
614 (MaxExpr && S.DiagnoseUnexpandedParameterPack(MaxExpr)))
615 return true;
616
617 // Accept template arguments for now as they depend on something else.
618 // We'll get to check them when they eventually get instantiated.
619 if (MinExpr->isValueDependent() || (MaxExpr && MaxExpr->isValueDependent()))
620 return false;
621
622 uint32_t Min = 0;
623 if (!S.checkUInt32Argument(Attr, MinExpr, Min, 0))
624 return true;
625
626 uint32_t Max = 0;
627 if (MaxExpr && !S.checkUInt32Argument(Attr, MaxExpr, Max, 1))
628 return true;
629
630 if (Min == 0 && Max != 0) {
631 S.Diag(Attr.getLocation(), diag::err_attribute_argument_invalid)
632 << &Attr << 0;
633 return true;
634 }
635 if (Max != 0 && Min > Max) {
636 S.Diag(Attr.getLocation(), diag::err_attribute_argument_invalid)
637 << &Attr << 1;
638 return true;
639 }
640
641 return false;
642}
643
644AMDGPUWavesPerEUAttr *
646 Expr *MinExpr, Expr *MaxExpr) {
647 ASTContext &Context = getASTContext();
648 AMDGPUWavesPerEUAttr TmpAttr(Context, CI, MinExpr, MaxExpr);
649
650 if (checkAMDGPUWavesPerEUArguments(SemaRef, MinExpr, MaxExpr, TmpAttr))
651 return nullptr;
652
653 return ::new (Context) AMDGPUWavesPerEUAttr(Context, CI, MinExpr, MaxExpr);
654}
655
657 Expr *MinExpr, Expr *MaxExpr) {
658 if (auto *Attr = CreateAMDGPUWavesPerEUAttr(CI, MinExpr, MaxExpr))
659 D->addAttr(Attr);
660}
661
664 return;
665
666 Expr *MinExpr = AL.getArgAsExpr(0);
667 Expr *MaxExpr = (AL.getNumArgs() > 1) ? AL.getArgAsExpr(1) : nullptr;
668
669 addAMDGPUWavesPerEUAttr(D, AL, MinExpr, MaxExpr);
670}
671
673 Diag(AL.getLoc(), diag::warn_amdgpu_num_reg_attr_deprecated) << AL;
674
675 uint32_t NumSGPR = 0;
676 Expr *NumSGPRExpr = AL.getArgAsExpr(0);
677 if (!SemaRef.checkUInt32Argument(AL, NumSGPRExpr, NumSGPR))
678 return;
679
680 D->addAttr(::new (getASTContext())
681 AMDGPUNumSGPRAttr(getASTContext(), AL, NumSGPR));
682}
683
685 Diag(AL.getLoc(), diag::warn_amdgpu_num_reg_attr_deprecated) << AL;
686
687 uint32_t NumVGPR = 0;
688 Expr *NumVGPRExpr = AL.getArgAsExpr(0);
689 if (!SemaRef.checkUInt32Argument(AL, NumVGPRExpr, NumVGPR))
690 return;
691
692 D->addAttr(::new (getASTContext())
693 AMDGPUNumVGPRAttr(getASTContext(), AL, NumVGPR));
694}
695
696static bool
698 Expr *ZExpr,
699 const AMDGPUMaxNumWorkGroupsAttr &Attr) {
700 if (S.DiagnoseUnexpandedParameterPack(XExpr) ||
701 (YExpr && S.DiagnoseUnexpandedParameterPack(YExpr)) ||
702 (ZExpr && S.DiagnoseUnexpandedParameterPack(ZExpr)))
703 return true;
704
705 // Accept template arguments for now as they depend on something else.
706 // We'll get to check them when they eventually get instantiated.
707 if (XExpr->isValueDependent() || (YExpr && YExpr->isValueDependent()) ||
708 (ZExpr && ZExpr->isValueDependent()))
709 return false;
710
711 uint32_t NumWG = 0;
712 Expr *Exprs[3] = {XExpr, YExpr, ZExpr};
713 for (int i = 0; i < 3; i++) {
714 if (Exprs[i]) {
715 if (!S.checkUInt32Argument(Attr, Exprs[i], NumWG, i,
716 /*StrictlyUnsigned=*/true))
717 return true;
718 if (NumWG == 0) {
719 S.Diag(Attr.getLoc(), diag::err_attribute_argument_is_zero)
720 << &Attr << Exprs[i]->getSourceRange();
721 return true;
722 }
723 }
724 }
725
726 return false;
727}
728
730 const AttributeCommonInfo &CI, Expr *XExpr, Expr *YExpr, Expr *ZExpr) {
731 ASTContext &Context = getASTContext();
732 AMDGPUMaxNumWorkGroupsAttr TmpAttr(Context, CI, XExpr, YExpr, ZExpr);
733 assert(!SemaRef.isSFINAEContext() &&
734 "Can't produce SFINAE diagnostic pointing to temporary attribute");
735
736 if (checkAMDGPUMaxNumWorkGroupsArguments(SemaRef, XExpr, YExpr, ZExpr,
737 TmpAttr))
738 return nullptr;
739
740 return ::new (Context)
741 AMDGPUMaxNumWorkGroupsAttr(Context, CI, XExpr, YExpr, ZExpr);
742}
743
745 const AttributeCommonInfo &CI,
746 Expr *XExpr, Expr *YExpr,
747 Expr *ZExpr) {
748 if (auto *Attr = CreateAMDGPUMaxNumWorkGroupsAttr(CI, XExpr, YExpr, ZExpr))
749 D->addAttr(Attr);
750}
751
753 const ParsedAttr &AL) {
754 Expr *YExpr = (AL.getNumArgs() > 1) ? AL.getArgAsExpr(1) : nullptr;
755 Expr *ZExpr = (AL.getNumArgs() > 2) ? AL.getArgAsExpr(2) : nullptr;
756 addAMDGPUMaxNumWorkGroupsAttr(D, AL, AL.getArgAsExpr(0), YExpr, ZExpr);
757}
758
761 ASTContext &Ctx = getASTContext();
762 QualType BoolTy = Ctx.getLogicalOperationType();
763 llvm::APInt False = llvm::APInt::getZero(Ctx.getIntWidth(BoolTy));
764 llvm::APInt True = llvm::APInt::getAllOnes(Ctx.getIntWidth(BoolTy));
765 SourceLocation Loc = CE->getExprLoc();
766
767 if (!CE->getBuiltinCallee())
768 return *ExpandedPredicates
769 .insert(IntegerLiteral::Create(Ctx, False, BoolTy, Loc))
770 .first;
771
772 bool P = false;
773 unsigned BI = CE->getBuiltinCallee();
774 if (Ctx.BuiltinInfo.isAuxBuiltinID(BI))
775 BI = Ctx.BuiltinInfo.getAuxBuiltinID(BI);
776
777 if (BI == AMDGPU::BI__builtin_amdgcn_processor_is) {
778 auto *GFX = dyn_cast<StringLiteral>(CE->getArg(0)->IgnoreParenCasts());
779 if (!GFX) {
780 Diag(Loc, diag::err_amdgcn_processor_is_arg_not_literal);
781 return nullptr;
782 }
783
784 StringRef N = GFX->getString();
785 const TargetInfo &TI = Ctx.getTargetInfo();
786 const TargetInfo *AuxTI = Ctx.getAuxTargetInfo();
787 if (!TI.isValidCPUName(N) && (!AuxTI || !AuxTI->isValidCPUName(N))) {
788 Diag(Loc, diag::err_amdgcn_processor_is_arg_invalid_value) << N;
790 if (TI.getTriple().getVendor() == llvm::Triple::VendorType::AMD)
791 TI.fillValidCPUList(ValidList);
792 else if (AuxTI) // Since the BI is present it must be an AMDGPU triple.
793 AuxTI->fillValidCPUList(ValidList);
794 if (!ValidList.empty())
795 Diag(Loc, diag::note_amdgcn_processor_is_valid_options)
796 << llvm::join(ValidList, ", ");
797 return nullptr;
798 }
799 if (Ctx.getTargetInfo().getTriple().isSPIRV()) {
800 CE->setType(BoolTy);
801 return *ExpandedPredicates.insert(CE).first;
802 }
803
804 if (auto TID = Ctx.getTargetInfo().getTargetID())
805 P = TID->find(N) == 0;
806 } else {
807 Expr *Arg = CE->getArg(0);
808 if (!Arg || Arg->getType() != Ctx.BuiltinFnTy) {
809 Diag(Loc, diag::err_amdgcn_is_invocable_arg_invalid_value) << Arg;
810 return nullptr;
811 }
812
813 if (Ctx.getTargetInfo().getTriple().isSPIRV()) {
814 CE->setType(BoolTy);
815 return *ExpandedPredicates.insert(CE).first;
816 }
817
819
820 StringRef RF = Ctx.BuiltinInfo.getRequiredFeatures(FD->getBuiltinID());
821 llvm::StringMap<bool> CF;
822 Ctx.getFunctionFeatureMap(CF, FD);
823
825 }
826
827 return *ExpandedPredicates
828 .insert(
829 IntegerLiteral::Create(Ctx, P ? True : False, BoolTy, Loc))
830 .first;
831}
832
834 return ExpandedPredicates.contains(E);
835}
836
838 PotentiallyUnguardedBuiltinUsers.insert(FD);
839}
840
842 return PotentiallyUnguardedBuiltinUsers.contains(FD);
843}
844
845namespace {
846/// This class implements -Wamdgpu-unguarded-builtin-usage.
847///
848/// This is done with a traversal of the AST of a function that includes a
849/// call to a target specific builtin. Whenever we encounter an \c if of the
850/// form: \c if(__builtin_amdgcn_is_invocable), we consider the then statement
851/// guarded.
852class DiagnoseUnguardedBuiltins : public DynamicRecursiveASTVisitor {
853 // TODO: this could eventually be extended to consider what happens when there
854 // are multiple target architectures specified via target("arch=gfxXXX")
855 // target("arch=gfxyyy") etc., as well as feature disabling via "-XXX".
856 Sema &SemaRef;
857
858 SmallVector<StringRef> TargetFeatures;
860 SmallVector<unsigned> GuardedBuiltins;
861
862 static Expr *FindPredicate(Expr *Cond) {
863 if (auto *CE = dyn_cast<CallExpr>(Cond)) {
864 if (CE->getBuiltinCallee() == AMDGPU::BI__builtin_amdgcn_is_invocable ||
865 CE->getBuiltinCallee() == AMDGPU::BI__builtin_amdgcn_processor_is)
866 return Cond;
867 } else if (auto *UO = dyn_cast<UnaryOperator>(Cond)) {
868 return FindPredicate(UO->getSubExpr());
869 } else if (auto *BO = dyn_cast<BinaryOperator>(Cond)) {
870 if ((Cond = FindPredicate(BO->getLHS())))
871 return Cond;
872 return FindPredicate(BO->getRHS());
873 }
874 return nullptr;
875 }
876
877 bool EnterPredicateGuardedContext(CallExpr *P);
878 void ExitPredicateGuardedContext(bool WasProcessorCheck);
879 bool TraverseGuardedStmt(Stmt *S, CallExpr *P);
880
881public:
882 DiagnoseUnguardedBuiltins(Sema &SemaRef) : SemaRef(SemaRef) {
883 if (auto *TAT = SemaRef.getCurFunctionDecl(true)->getAttr<TargetAttr>()) {
884 // We use the somewhat misnamed x86 accessors because they provide exactly
885 // what we require.
886 TAT->getX86AddedFeatures(TargetFeatures);
887 if (auto GFXIP = TAT->getX86Architecture())
888 CurrentGFXIP.emplace_back(TAT->getLocation(), *GFXIP);
889 }
890 }
891
892 bool TraverseLambdaExpr(LambdaExpr *LE) override {
893 if (SemaRef.AMDGPU().HasPotentiallyUnguardedBuiltinUsage(
894 LE->getCallOperator()))
895 return true; // We have already handled this.
896 return DynamicRecursiveASTVisitor::TraverseLambdaExpr(LE);
897 }
898
899 bool TraverseStmt(Stmt *S) override {
900 if (!S)
901 return true;
903 }
904
905 void IssueDiagnostics(Stmt *S) { TraverseStmt(S); }
906
907 bool TraverseIfStmt(IfStmt *If) override {
908 if (auto *CE = dyn_cast_or_null<CallExpr>(FindPredicate(If->getCond())))
909 return TraverseGuardedStmt(If, CE);
910 return DynamicRecursiveASTVisitor::TraverseIfStmt(If);
911 }
912
913 bool TraverseCaseStmt(CaseStmt *CS) override {
914 return TraverseStmt(CS->getSubStmt());
915 }
916
917 bool TraverseConditionalOperator(ConditionalOperator *CO) override {
918 if (auto *CE = dyn_cast_or_null<CallExpr>(FindPredicate(CO->getCond())))
919 return TraverseGuardedStmt(CO, CE);
920 return DynamicRecursiveASTVisitor::TraverseConditionalOperator(CO);
921 }
922
923 bool VisitAsmStmt(AsmStmt *ASM) override;
924 bool VisitCallExpr(CallExpr *CE) override;
925};
926
927bool DiagnoseUnguardedBuiltins::EnterPredicateGuardedContext(CallExpr *P) {
928 bool IsProcessorCheck =
929 P->getBuiltinCallee() == AMDGPU::BI__builtin_amdgcn_processor_is;
930
931 if (IsProcessorCheck) {
932 StringRef G = cast<clang::StringLiteral>(P->getArg(0))->getString();
933 // TODO: handle generic ISAs.
934 if (!CurrentGFXIP.empty() && G != CurrentGFXIP.back().second) {
935 SemaRef.Diag(P->getExprLoc(),
936 diag::err_amdgcn_conflicting_is_processor_options)
937 << P;
938 SemaRef.Diag(CurrentGFXIP.back().first,
939 diag::note_amdgcn_previous_is_processor_guard);
940 }
941 CurrentGFXIP.emplace_back(P->getExprLoc(), G);
942 } else {
943 auto *FD = cast<FunctionDecl>(
944 cast<DeclRefExpr>(P->getArg(0))->getReferencedDeclOfCallee());
945 GuardedBuiltins.push_back(FD->getBuiltinID());
946 }
947
948 return IsProcessorCheck;
949}
950
951void DiagnoseUnguardedBuiltins::ExitPredicateGuardedContext(bool WasProcCheck) {
952 if (WasProcCheck)
953 CurrentGFXIP.pop_back();
954 else
955 GuardedBuiltins.pop_back();
956}
957
958inline std::pair<Stmt *, Stmt *> GetTraversalOrder(Stmt *S) {
959 std::pair<Stmt *, Stmt *> Ordered;
960 Expr *Condition = nullptr;
961
962 if (auto *CO = dyn_cast<ConditionalOperator>(S)) {
963 Condition = CO->getCond();
964 Ordered = {CO->getTrueExpr(), CO->getFalseExpr()};
965 } else if (auto *If = dyn_cast<IfStmt>(S)) {
966 Condition = If->getCond();
967 Ordered = {If->getThen(), If->getElse()};
968 }
969
970 if (auto *UO = dyn_cast<UnaryOperator>(Condition))
971 if (UO->getOpcode() == UnaryOperatorKind::UO_LNot)
972 std::swap(Ordered.first, Ordered.second);
973
974 return Ordered;
975}
976
977bool DiagnoseUnguardedBuiltins::TraverseGuardedStmt(Stmt *S, CallExpr *P) {
978 assert(S && "Unexpected missing Statement!");
979 assert(P && "Unexpected missing Predicate!");
980
981 auto [Guarded, Unguarded] = GetTraversalOrder(S);
982
983 bool WasProcessorCheck = EnterPredicateGuardedContext(P);
984
985 bool Continue = TraverseStmt(Guarded);
986
987 ExitPredicateGuardedContext(WasProcessorCheck);
988
989 return Continue && TraverseStmt(Unguarded);
990}
991
992bool DiagnoseUnguardedBuiltins::VisitAsmStmt(AsmStmt *ASM) {
993 // TODO: should we check if the ASM is valid for the target? Can we?
994 if (!CurrentGFXIP.empty())
995 return true;
996
997 std::string S = ASM->generateAsmString(SemaRef.getASTContext());
998 SemaRef.Diag(ASM->getAsmLoc(), diag::warn_amdgcn_unguarded_asm_stmt) << S;
999 SemaRef.Diag(ASM->getAsmLoc(), diag::note_amdgcn_unguarded_asm_silence) << S;
1000
1001 return true;
1002}
1003
1004bool DiagnoseUnguardedBuiltins::VisitCallExpr(CallExpr *CE) {
1005 unsigned ID = CE->getBuiltinCallee();
1006 Builtin::Context &BInfo = SemaRef.getASTContext().BuiltinInfo;
1007
1008 if (!ID)
1009 return true;
1010 if (!BInfo.isTSBuiltin(ID))
1011 return true;
1012 if (ID == AMDGPU::BI__builtin_amdgcn_processor_is ||
1013 ID == AMDGPU::BI__builtin_amdgcn_is_invocable)
1014 return true;
1015 if (llvm::find(GuardedBuiltins, ID) != GuardedBuiltins.end())
1016 return true;
1017
1018 StringRef FL(BInfo.getRequiredFeatures(ID));
1019 llvm::StringMap<bool> FeatureMap;
1020 if (CurrentGFXIP.empty()) {
1021 for (auto &&F : TargetFeatures)
1022 FeatureMap[F] = true;
1023 for (auto &&GID : GuardedBuiltins)
1024 for (auto &&F : llvm::split(BInfo.getRequiredFeatures(GID), ','))
1025 FeatureMap[F] = true;
1026 } else {
1027 static const llvm::Triple AMDGCN(llvm::Triple::amdgcn,
1028 llvm::Triple::NoSubArch, llvm::Triple::AMD,
1029 llvm::Triple::AMDHSA);
1030 llvm::AMDGPU::fillAMDGPUFeatureMap(CurrentGFXIP.back().second, AMDGCN,
1031 FeatureMap);
1032 }
1033
1034 FunctionDecl *BI = CE->getDirectCallee();
1035 SourceLocation BICallLoc = CE->getExprLoc();
1036 if (Builtin::evaluateRequiredTargetFeatures(FL, FeatureMap)) {
1037 SemaRef.Diag(BICallLoc, diag::warn_amdgcn_unguarded_builtin) << BI;
1038 SemaRef.Diag(BICallLoc, diag::note_amdgcn_unguarded_builtin_silence) << BI;
1039 } else {
1040 StringRef GFXIP = CurrentGFXIP.empty() ? "" : CurrentGFXIP.back().second;
1041 SemaRef.Diag(BICallLoc, diag::err_amdgcn_incompatible_builtin)
1042 << BI << FL << !CurrentGFXIP.empty() << GFXIP;
1043 if (!CurrentGFXIP.empty())
1044 SemaRef.Diag(CurrentGFXIP.back().first,
1045 diag::note_amdgcn_previous_is_processor_guard);
1046 }
1047
1048 return true;
1049}
1050} // Unnamed namespace
1051
1053 DiagnoseUnguardedBuiltins(SemaRef).IssueDiagnostics(FD->getBody());
1054}
1055} // namespace clang
#define GFX(gpu)
This file declares semantic analysis functions specific to AMDGPU.
Enumerates target-specific builtins in their own namespaces within namespace clang.
APSInt & getInt()
Definition APValue.h:508
Holds long-lived AST nodes (such as types and decls) that can be referred to throughout the semantic ...
Definition ASTContext.h:227
unsigned getIntWidth(QualType T) const
Builtin::Context & BuiltinInfo
Definition ASTContext.h:807
CanQualType getLogicalOperationType() const
The result type of logical operations, '<', '>', '!=', etc.
const TargetInfo * getAuxTargetInfo() const
Definition ASTContext.h:925
CanQualType BuiltinFnTy
const TargetInfo & getTargetInfo() const
Definition ASTContext.h:924
void getFunctionFeatureMap(llvm::StringMap< bool > &FeatureMap, const FunctionDecl *) const
unsigned getTargetAddressSpace(LangAS AS) const
PtrTy get() const
Definition Ownership.h:171
bool isInvalid() const
Definition Ownership.h:167
static std::unique_ptr< AtomicScopeModel > create(AtomicScopeModelKind K)
Create an atomic scope model by AtomicScopeModelKind.
Definition SyncScope.h:298
Attr - This represents one attribute.
Definition Attr.h:46
SourceLocation getLocation() const
Definition Attr.h:99
SourceLocation getLoc() const
bool isAuxBuiltinID(unsigned ID) const
Return true if the builtin ID belongs exclusively to the AuxTarget, and false if it belongs to both p...
Definition Builtins.h:442
unsigned getAuxBuiltinID(unsigned ID) const
Return real builtin ID (i.e.
Definition Builtins.h:448
const char * getRequiredFeatures(unsigned ID) const
Definition Builtins.cpp:102
CallExpr - Represents a function call (C99 6.5.2.2, C++ [expr.call]).
Definition Expr.h:2946
Expr * getArg(unsigned Arg)
getArg - Return the specified argument.
Definition Expr.h:3150
SourceLocation getBeginLoc() const
Definition Expr.h:3280
unsigned getBuiltinCallee() const
getBuiltinCallee - If this is a call to a builtin, return the builtin ID of the callee.
Definition Expr.cpp:1592
unsigned getNumArgs() const
getNumArgs - Return the number of actual arguments to this call.
Definition Expr.h:3137
Decl - This represents one declaration (or definition), e.g.
Definition DeclBase.h:86
void addAttr(Attr *A)
virtual bool TraverseStmt(MaybeConst< Stmt > *S)
This represents one expression.
Definition Expr.h:112
bool EvaluateAsInt(EvalResult &Result, const ASTContext &Ctx, SideEffectsKind AllowSideEffects=SE_NoSideEffects, bool InConstantContext=false) const
EvaluateAsInt - Return true if this is a constant which we can fold and convert to an integer,...
Expr * IgnoreParenCasts() LLVM_READONLY
Skip past any parentheses and casts which might surround this expression until reaching a fixed point...
Definition Expr.cpp:3095
void setType(QualType t)
Definition Expr.h:145
bool isValueDependent() const
Determines whether the value of this expression depends on.
Definition Expr.h:177
Expr * IgnoreParens() LLVM_READONLY
Skip past any parentheses which might surround this expression until reaching a fixed point.
Definition Expr.cpp:3086
Decl * getReferencedDeclOfCallee()
Definition Expr.cpp:1546
bool isInstantiationDependent() const
Whether this expression is instantiation-dependent, meaning that it depends in some way on.
Definition Expr.h:223
SourceLocation getExprLoc() const LLVM_READONLY
getExprLoc - Return the preferred location for the arrow when diagnosing a problem with a generic exp...
Definition Expr.cpp:277
QualType getType() const
Definition Expr.h:144
Represents a function declaration or definition.
Definition Decl.h:2015
Stmt * getBody(const FunctionDecl *&Definition) const
Retrieve the body (definition) of the function.
Definition Decl.cpp:3245
static IntegerLiteral * Create(const ASTContext &C, const llvm::APInt &V, QualType type, SourceLocation l)
Returns a new integer literal with value 'V' and type 'type'.
Definition Expr.cpp:975
ParsedAttr - Represents a syntactic attribute.
Definition ParsedAttr.h:119
unsigned getNumArgs() const
getNumArgs - Return the number of actual arguments to this attribute.
Definition ParsedAttr.h:371
Expr * getArgAsExpr(unsigned Arg) const
Definition ParsedAttr.h:383
bool checkAtLeastNumArgs(class Sema &S, unsigned Num) const
Check if the attribute has at least as many args as Num.
bool checkAtMostNumArgs(class Sema &S, unsigned Num) const
Check if the attribute has at most as many args as Num.
A (possibly-)qualified type.
Definition TypeBase.h:937
const Type * getTypePtr() const
Retrieves a pointer to the underlying (unqualified) type.
Definition TypeBase.h:8436
Scope - A scope is a transient data structure that is used while parsing the program.
Definition Scope.h:41
void handleAMDGPUMaxNumWorkGroupsAttr(Decl *D, const ParsedAttr &AL)
void addAMDGPUFlatWorkGroupSizeAttr(Decl *D, const AttributeCommonInfo &CI, Expr *Min, Expr *Max)
addAMDGPUFlatWorkGroupSizeAttr - Adds an amdgpu_flat_work_group_size attribute to a particular declar...
bool checkCoopAtomicFunctionCall(CallExpr *TheCall, bool IsStore)
bool HasPotentiallyUnguardedBuiltinUsage(FunctionDecl *FD) const
void handleAMDGPUFlatWorkGroupSizeAttr(Decl *D, const ParsedAttr &AL)
bool checkAtomicMonitorLoad(CallExpr *TheCall)
bool checkAtomicOrderingCABIArg(Expr *E, bool MayLoad, bool MayStore)
Emits a diagnostic if the E is not an atomic ordering encoded in the C ABI format,...
void handleAMDGPUNumSGPRAttr(Decl *D, const ParsedAttr &AL)
AMDGPUMaxNumWorkGroupsAttr * CreateAMDGPUMaxNumWorkGroupsAttr(const AttributeCommonInfo &CI, Expr *XExpr, Expr *YExpr, Expr *ZExpr)
Create an AMDGPUMaxNumWorkGroupsAttr attribute.
Expr * ExpandAMDGPUPredicateBuiltIn(Expr *CE)
Expand a valid use of the feature identification builtins into its corresponding sequence of instruct...
AMDGPUWavesPerEUAttr * CreateAMDGPUWavesPerEUAttr(const AttributeCommonInfo &CI, Expr *Min, Expr *Max)
Create an AMDGPUWavesPerEUAttr attribute.
void DiagnoseUnguardedBuiltinUsage(FunctionDecl *FD)
void handleAMDGPUNumVGPRAttr(Decl *D, const ParsedAttr &AL)
AMDGPUFlatWorkGroupSizeAttr * CreateAMDGPUFlatWorkGroupSizeAttr(const AttributeCommonInfo &CI, Expr *Min, Expr *Max)
Create an AMDGPUWavesPerEUAttr attribute.
void AddPotentiallyUnguardedBuiltinUser(FunctionDecl *FD)
Diagnose unguarded usages of AMDGPU builtins and recommend guarding with __builtin_amdgcn_is_invocabl...
bool checkMovDPPFunctionCall(CallExpr *TheCall, unsigned NumArgs, unsigned NumDataArgs)
void handleAMDGPUWavesPerEUAttr(Decl *D, const ParsedAttr &AL)
bool IsPredicate(Expr *E) const
bool CheckAMDGCNBuiltinFunctionCall(unsigned BuiltinID, CallExpr *TheCall)
void addAMDGPUWavesPerEUAttr(Decl *D, const AttributeCommonInfo &CI, Expr *Min, Expr *Max)
addAMDGPUWavePersEUAttr - Adds an amdgpu_waves_per_eu attribute to a particular declaration.
void addAMDGPUMaxNumWorkGroupsAttr(Decl *D, const AttributeCommonInfo &CI, Expr *XExpr, Expr *YExpr, Expr *ZExpr)
addAMDGPUMaxNumWorkGroupsAttr - Adds an amdgpu_max_num_work_groups attribute to a particular declarat...
SemaBase(Sema &S)
Definition SemaBase.cpp:7
ASTContext & getASTContext() const
Definition SemaBase.cpp:9
Sema & SemaRef
Definition SemaBase.h:40
SemaDiagnosticBuilder Diag(SourceLocation Loc, unsigned DiagID)
Emit a diagnostic.
Definition SemaBase.cpp:61
Sema - This implements semantic analysis and AST building for C.
Definition Sema.h:868
bool DiagnoseUnexpandedParameterPack(SourceLocation Loc, TypeSourceInfo *T, UnexpandedParameterPackContext UPPC)
If the given type contains an unexpanded parameter pack, diagnose the error.
bool checkUInt32Argument(const AttrInfo &AI, const Expr *Expr, uint32_t &Val, unsigned Idx=UINT_MAX, bool StrictlyUnsigned=false)
If Expr is a valid integer constant, get the value of the integer expression and return success or fa...
Definition Sema.h:4906
Encodes a location in the source.
SourceRange getSourceRange() const LLVM_READONLY
SourceLocation tokens are not useful in isolation - they are low level value objects created/interpre...
Definition Stmt.cpp:343
SourceLocation getBeginLoc() const LLVM_READONLY
Definition Stmt.cpp:355
Exposes information about the current target.
Definition TargetInfo.h:227
virtual std::optional< std::string > getTargetID() const
Returns the target ID if supported.
const llvm::Triple & getTriple() const
Returns the target triple of the primary target.
virtual void fillValidCPUList(SmallVectorImpl< StringRef > &Values) const
Fill a SmallVectorImpl with the valid values to setCPU.
virtual bool isValidCPUName(StringRef Name) const
Determine whether this TargetInfo supports the given CPU name.
The base class of the type hierarchy.
Definition TypeBase.h:1871
QualType getPointeeType() const
If this is a pointer, ObjC object pointer, or block pointer, this returns the respective pointee.
Definition Type.cpp:789
Defines the clang::TargetInfo interface.
bool evaluateRequiredTargetFeatures(llvm::StringRef RequiredFatures, const llvm::StringMap< bool > &TargetFetureMap)
Returns true if the required target features of a builtin function are enabled.
bool LE(InterpState &S, CodePtr OpPC)
Definition Interp.h:1438
The JSON file list parser is used to communicate input to InstallAPI.
bool isa(CodeGen::Address addr)
Definition Address.h:330
@ If
'if' clause, allowed on all the Compute Constructs, Data Constructs, Executable Constructs,...
Expr * Cond
};
@ Result
The result type of a method or function.
Definition TypeBase.h:905
static bool checkAMDGPUMaxNumWorkGroupsArguments(Sema &S, Expr *XExpr, Expr *YExpr, Expr *ZExpr, const AMDGPUMaxNumWorkGroupsAttr &Attr)
DynamicRecursiveASTVisitorBase< false > DynamicRecursiveASTVisitor
U cast(CodeGen::Address addr)
Definition Address.h:327
static bool checkAMDGPUFlatWorkGroupSizeArguments(Sema &S, Expr *MinExpr, Expr *MaxExpr, const AMDGPUFlatWorkGroupSizeAttr &Attr)
ActionResult< Expr * > ExprResult
Definition Ownership.h:249
static bool checkAMDGPUWavesPerEUArguments(Sema &S, Expr *MinExpr, Expr *MaxExpr, const AMDGPUWavesPerEUAttr &Attr)
EvalResult is a struct with detailed info about an evaluated expression.
Definition Expr.h:648
APValue Val
Val - This is the value the expression can be folded to.
Definition Expr.h:650