40 unsigned OrderIndex, ScopeIndex;
42 const auto *FD =
SemaRef.getCurFunctionDecl(
true);
43 assert(FD &&
"AMDGPU builtins should not be used outside of a function");
44 llvm::StringMap<bool> CallerFeatureMap;
50 case AMDGPU::BI__builtin_amdgcn_raw_ptr_buffer_load_lds:
51 case AMDGPU::BI__builtin_amdgcn_raw_ptr_buffer_load_async_lds:
52 case AMDGPU::BI__builtin_amdgcn_struct_ptr_buffer_load_lds:
53 case AMDGPU::BI__builtin_amdgcn_struct_ptr_buffer_load_async_lds:
54 case AMDGPU::BI__builtin_amdgcn_load_to_lds:
55 case AMDGPU::BI__builtin_amdgcn_load_async_to_lds:
56 case AMDGPU::BI__builtin_amdgcn_global_load_lds:
57 case AMDGPU::BI__builtin_amdgcn_global_load_async_lds: {
58 constexpr const int SizeIdx = 2;
66 SemaRef.VerifyIntegerConstantExpression(ArgExpr, &Size);
67 assert(!R.isInvalid());
68 switch (Size.getSExtValue()) {
81 diag::err_amdgcn_load_lds_size_invalid_value)
84 diag::note_amdgcn_load_lds_size_valid_value)
89 case AMDGPU::BI__builtin_amdgcn_get_fpenv:
90 case AMDGPU::BI__builtin_amdgcn_set_fpenv:
92 case AMDGPU::BI__builtin_amdgcn_atomic_inc32:
93 case AMDGPU::BI__builtin_amdgcn_atomic_inc64:
94 case AMDGPU::BI__builtin_amdgcn_atomic_dec32:
95 case AMDGPU::BI__builtin_amdgcn_atomic_dec64:
99 case AMDGPU::BI__builtin_amdgcn_fence:
103 case AMDGPU::BI__builtin_amdgcn_s_setreg:
104 return SemaRef.BuiltinConstantArgRange(TheCall, 0, 0,
106 case AMDGPU::BI__builtin_amdgcn_s_wait_event: {
112 "gfx12-insts", CallerFeatureMap);
117 if (((IsGFX12Plus && !
Result[1]) || (!IsGFX12Plus &&
Result[0])) ||
118 Result.getZExtValue() > 2) {
121 diag::warn_amdgpu_s_wait_event_mask_no_effect_target)
124 diag::note_amdgpu_s_wait_event_suggested_value)
130 case AMDGPU::BI__builtin_amdgcn_mov_dpp:
132 case AMDGPU::BI__builtin_amdgcn_mov_dpp8:
134 case AMDGPU::BI__builtin_amdgcn_update_dpp:
136 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f16_fp8:
137 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_bf16_fp8:
138 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f16_bf8:
139 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_bf16_bf8:
140 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f16_fp4:
141 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_bf16_fp4:
142 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f32_fp8:
143 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f32_bf8:
144 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk8_f32_fp4:
145 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f16_fp6:
146 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_bf16_fp6:
147 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f16_bf6:
148 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_bf16_bf6:
149 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f32_fp6:
150 case AMDGPU::BI__builtin_amdgcn_cvt_scale_pk16_f32_bf6:
151 return SemaRef.BuiltinConstantArgRange(TheCall, 2, 0, 15);
152 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_load_32x4B:
153 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_load_16x8B:
154 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_load_8x16B:
156 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_store_32x4B:
157 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_store_16x8B:
158 case AMDGPU::BI__builtin_amdgcn_cooperative_atomic_store_8x16B:
160 case AMDGPU::BI__builtin_amdgcn_flat_load_monitor_b32:
161 case AMDGPU::BI__builtin_amdgcn_flat_load_monitor_b64:
162 case AMDGPU::BI__builtin_amdgcn_flat_load_monitor_b128:
163 case AMDGPU::BI__builtin_amdgcn_global_load_monitor_b32:
164 case AMDGPU::BI__builtin_amdgcn_global_load_monitor_b64:
165 case AMDGPU::BI__builtin_amdgcn_global_load_monitor_b128:
167 case AMDGPU::BI__builtin_amdgcn_image_load_1d_v4f32_i32:
168 case AMDGPU::BI__builtin_amdgcn_image_load_1darray_v4f32_i32:
169 case AMDGPU::BI__builtin_amdgcn_image_load_1d_v4f16_i32:
170 case AMDGPU::BI__builtin_amdgcn_image_load_1darray_v4f16_i32:
171 case AMDGPU::BI__builtin_amdgcn_image_load_2d_f32_i32:
172 case AMDGPU::BI__builtin_amdgcn_image_load_2d_v4f32_i32:
173 case AMDGPU::BI__builtin_amdgcn_image_load_2d_v4f16_i32:
174 case AMDGPU::BI__builtin_amdgcn_image_load_2darray_f32_i32:
175 case AMDGPU::BI__builtin_amdgcn_image_load_2darray_v4f32_i32:
176 case AMDGPU::BI__builtin_amdgcn_image_load_2darray_v4f16_i32:
177 case AMDGPU::BI__builtin_amdgcn_image_load_3d_v4f32_i32:
178 case AMDGPU::BI__builtin_amdgcn_image_load_3d_v4f16_i32:
179 case AMDGPU::BI__builtin_amdgcn_image_load_cube_v4f32_i32:
180 case AMDGPU::BI__builtin_amdgcn_image_load_cube_v4f16_i32:
181 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1d_v4f32_i32:
182 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1d_v4f16_i32:
183 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1darray_v4f32_i32:
184 case AMDGPU::BI__builtin_amdgcn_image_load_mip_1darray_v4f16_i32:
185 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2d_f32_i32:
186 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2d_v4f32_i32:
187 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2d_v4f16_i32:
188 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2darray_f32_i32:
189 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2darray_v4f32_i32:
190 case AMDGPU::BI__builtin_amdgcn_image_load_mip_2darray_v4f16_i32:
191 case AMDGPU::BI__builtin_amdgcn_image_load_mip_3d_v4f32_i32:
192 case AMDGPU::BI__builtin_amdgcn_image_load_mip_3d_v4f16_i32:
193 case AMDGPU::BI__builtin_amdgcn_image_load_mip_cube_v4f32_i32:
194 case AMDGPU::BI__builtin_amdgcn_image_load_mip_cube_v4f16_i32:
195 case AMDGPU::BI__builtin_amdgcn_image_sample_1d_v4f32_f32:
196 case AMDGPU::BI__builtin_amdgcn_image_sample_1darray_v4f32_f32:
197 case AMDGPU::BI__builtin_amdgcn_image_sample_1d_v4f16_f32:
198 case AMDGPU::BI__builtin_amdgcn_image_sample_1darray_v4f16_f32:
199 case AMDGPU::BI__builtin_amdgcn_image_sample_2d_f32_f32:
200 case AMDGPU::BI__builtin_amdgcn_image_sample_2d_v4f32_f32:
201 case AMDGPU::BI__builtin_amdgcn_image_sample_2d_v4f16_f32:
202 case AMDGPU::BI__builtin_amdgcn_image_sample_2darray_f32_f32:
203 case AMDGPU::BI__builtin_amdgcn_image_sample_2darray_v4f32_f32:
204 case AMDGPU::BI__builtin_amdgcn_image_sample_2darray_v4f16_f32:
205 case AMDGPU::BI__builtin_amdgcn_image_sample_3d_v4f32_f32:
206 case AMDGPU::BI__builtin_amdgcn_image_sample_3d_v4f16_f32:
207 case AMDGPU::BI__builtin_amdgcn_image_sample_cube_v4f32_f32:
208 case AMDGPU::BI__builtin_amdgcn_image_sample_cube_v4f16_f32:
209 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1d_v4f32_f32:
210 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1d_v4f16_f32:
211 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1darray_v4f32_f32:
212 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_1darray_v4f16_f32:
213 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2d_f32_f32:
214 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2d_v4f32_f32:
215 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2d_v4f16_f32:
216 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2darray_f32_f32:
217 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2darray_v4f32_f32:
218 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_2darray_v4f16_f32:
219 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_3d_v4f32_f32:
220 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_3d_v4f16_f32:
221 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_cube_v4f32_f32:
222 case AMDGPU::BI__builtin_amdgcn_image_sample_lz_cube_v4f16_f32:
223 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1d_v4f32_f32:
224 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1d_v4f16_f32:
225 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1darray_v4f32_f32:
226 case AMDGPU::BI__builtin_amdgcn_image_sample_l_1darray_v4f16_f32:
227 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2d_f32_f32:
228 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2d_v4f16_f32:
229 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2d_v4f32_f32:
230 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2darray_f32_f32:
231 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2darray_v4f32_f32:
232 case AMDGPU::BI__builtin_amdgcn_image_sample_l_2darray_v4f16_f32:
233 case AMDGPU::BI__builtin_amdgcn_image_sample_l_3d_v4f32_f32:
234 case AMDGPU::BI__builtin_amdgcn_image_sample_l_3d_v4f16_f32:
235 case AMDGPU::BI__builtin_amdgcn_image_sample_l_cube_v4f32_f32:
236 case AMDGPU::BI__builtin_amdgcn_image_sample_l_cube_v4f16_f32:
237 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1d_v4f32_f32:
238 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1d_v4f16_f32:
239 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1darray_v4f32_f32:
240 case AMDGPU::BI__builtin_amdgcn_image_sample_d_1darray_v4f16_f32:
241 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2d_f32_f32:
242 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2d_v4f32_f32:
243 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2d_v4f16_f32:
244 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2darray_f32_f32:
245 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2darray_v4f32_f32:
246 case AMDGPU::BI__builtin_amdgcn_image_sample_d_2darray_v4f16_f32:
247 case AMDGPU::BI__builtin_amdgcn_image_sample_d_3d_v4f32_f32:
248 case AMDGPU::BI__builtin_amdgcn_image_sample_d_3d_v4f16_f32:
249 case AMDGPU::BI__builtin_amdgcn_image_gather4_lz_2d_v4f32_f32: {
250 StringRef FeatureList(
255 << FD->getDeclName() << FeatureList;
259 unsigned ArgCount = TheCall->
getNumArgs() - 1;
264 constexpr unsigned DMaskArgNo = 0;
265 constexpr int Low = 0;
266 constexpr int High = 15;
267 if (
SemaRef.BuiltinConstantArg(TheCall, DMaskArgNo,
Result) ||
268 SemaRef.BuiltinConstantArgRange(TheCall, DMaskArgNo, Low, High,
274 int NumElementsInRetTy = 1;
276 if (
auto *VTy = dyn_cast<VectorType>(RetTy))
277 NumElementsInRetTy = VTy->getNumElements();
278 int NumActiveBitsInDMask =
279 llvm::popcount(
static_cast<uint8_t
>(
Result.getExtValue()));
280 if (NumActiveBitsInDMask > NumElementsInRetTy) {
282 diag::err_amdgcn_dmask_has_too_many_bits_set);
288 bool ExtraGatherChecks =
289 BuiltinID == AMDGPU::BI__builtin_amdgcn_image_gather4_lz_2d_v4f32_f32 &&
290 SemaRef.BuiltinConstantArgPower2(TheCall, 0);
292 return ExtraGatherChecks ||
294 (
SemaRef.BuiltinConstantArg(TheCall, (ArgCount - 1),
Result));
296 case AMDGPU::BI__builtin_amdgcn_image_store_1d_v4f32_i32:
297 case AMDGPU::BI__builtin_amdgcn_image_store_1darray_v4f32_i32:
298 case AMDGPU::BI__builtin_amdgcn_image_store_1d_v4f16_i32:
299 case AMDGPU::BI__builtin_amdgcn_image_store_1darray_v4f16_i32:
300 case AMDGPU::BI__builtin_amdgcn_image_store_2d_f32_i32:
301 case AMDGPU::BI__builtin_amdgcn_image_store_2d_v4f32_i32:
302 case AMDGPU::BI__builtin_amdgcn_image_store_2d_v4f16_i32:
303 case AMDGPU::BI__builtin_amdgcn_image_store_2darray_f32_i32:
304 case AMDGPU::BI__builtin_amdgcn_image_store_2darray_v4f32_i32:
305 case AMDGPU::BI__builtin_amdgcn_image_store_2darray_v4f16_i32:
306 case AMDGPU::BI__builtin_amdgcn_image_store_3d_v4f32_i32:
307 case AMDGPU::BI__builtin_amdgcn_image_store_3d_v4f16_i32:
308 case AMDGPU::BI__builtin_amdgcn_image_store_cube_v4f32_i32:
309 case AMDGPU::BI__builtin_amdgcn_image_store_cube_v4f16_i32:
310 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1d_v4f32_i32:
311 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1d_v4f16_i32:
312 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1darray_v4f32_i32:
313 case AMDGPU::BI__builtin_amdgcn_image_store_mip_1darray_v4f16_i32:
314 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2d_f32_i32:
315 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2d_v4f32_i32:
316 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2d_v4f16_i32:
317 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2darray_f32_i32:
318 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2darray_v4f32_i32:
319 case AMDGPU::BI__builtin_amdgcn_image_store_mip_2darray_v4f16_i32:
320 case AMDGPU::BI__builtin_amdgcn_image_store_mip_3d_v4f32_i32:
321 case AMDGPU::BI__builtin_amdgcn_image_store_mip_3d_v4f16_i32:
322 case AMDGPU::BI__builtin_amdgcn_image_store_mip_cube_v4f32_i32:
323 case AMDGPU::BI__builtin_amdgcn_image_store_mip_cube_v4f16_i32: {
324 StringRef FeatureList(
329 << FD->getDeclName() << FeatureList;
333 unsigned ArgCount = TheCall->
getNumArgs() - 1;
338 constexpr unsigned DMaskArgNo = 1;
339 return SemaRef.BuiltinConstantArgRange(TheCall, DMaskArgNo, 0,
345 case AMDGPU::BI__builtin_amdgcn_wmma_i32_16x16x64_iu8:
346 case AMDGPU::BI__builtin_amdgcn_swmmac_i32_16x16x128_iu8: {
347 if (BuiltinID == AMDGPU::BI__builtin_amdgcn_wmma_i32_16x16x64_iu8) {
348 if (
SemaRef.checkArgCountRange(TheCall, 7, 8))
352 }
else if (BuiltinID ==
353 AMDGPU::BI__builtin_amdgcn_swmmac_i32_16x16x128_iu8) {
354 if (
SemaRef.checkArgCountRange(TheCall, 8, 9))
363 llvm::APSInt ClampValue;
364 if (!
SemaRef.VerifyIntegerConstantExpression(ClampArg, &ClampValue)
373 SemaRef.CheckSingleAssignmentConstraints(BoolTy, ClampExpr);
379 case AMDGPU::BI__builtin_amdgcn_wmma_f32_16x16x32_bf16:
380 case AMDGPU::BI__builtin_amdgcn_wmma_f32_16x16x4_f32:
381 case AMDGPU::BI__builtin_amdgcn_wmma_f32_16x16x32_f16:
382 case AMDGPU::BI__builtin_amdgcn_wmma_f16_16x16x32_f16:
383 case AMDGPU::BI__builtin_amdgcn_wmma_bf16_16x16x32_bf16:
384 case AMDGPU::BI__builtin_amdgcn_wmma_bf16f32_16x16x32_bf16:
385 return SemaRef.BuiltinConstantArgRange(TheCall, 0, 0,
387 SemaRef.BuiltinConstantArgRange(TheCall, 2, 0,
394 auto ArgExpr = Arg.
get();
398 return Diag(ArgExpr->getExprLoc(), diag::err_typecheck_expect_int)
399 << ArgExpr->getType();
400 auto Ord = ArgResult.
Val.
getInt().getZExtValue();
404 if (!llvm::isValidAtomicOrderingCABI(Ord))
405 return Diag(ArgExpr->getBeginLoc(),
406 diag::warn_atomic_op_has_invalid_memory_order)
407 << 0 << ArgExpr->getSourceRange();
408 switch (
static_cast<llvm::AtomicOrderingCABI
>(Ord)) {
409 case llvm::AtomicOrderingCABI::relaxed:
410 case llvm::AtomicOrderingCABI::consume:
411 if (BuiltinID == AMDGPU::BI__builtin_amdgcn_fence)
412 return Diag(ArgExpr->getBeginLoc(),
413 diag::warn_atomic_op_has_invalid_memory_order)
414 << 0 << ArgExpr->getSourceRange();
416 case llvm::AtomicOrderingCABI::acquire:
417 case llvm::AtomicOrderingCABI::release:
418 case llvm::AtomicOrderingCABI::acq_rel:
419 case llvm::AtomicOrderingCABI::seq_cst:
423 Arg = TheCall->
getArg(ScopeIndex);
427 if (!ArgExpr->EvaluateAsConstantExpr(ArgResult1,
getASTContext()))
428 return Diag(ArgExpr->getExprLoc(), diag::err_expr_not_string_literal)
429 << ArgExpr->getType();