541 NEONMAP1(__a32_vcvt_bf16_f32, arm_neon_vcvtfp2bf, 0),
549 NEONMAP1(vabsq_v, arm_neon_vabs, 0),
553 NEONMAP1(vaesdq_u8, arm_neon_aesd, 0),
554 NEONMAP1(vaeseq_u8, arm_neon_aese, 0),
555 NEONMAP1(vaesimcq_u8, arm_neon_aesimc, 0),
556 NEONMAP1(vaesmcq_u8, arm_neon_aesmc, 0),
557 NEONMAP1(vbfdot_f32, arm_neon_bfdot, 0),
558 NEONMAP1(vbfdotq_f32, arm_neon_bfdot, 0),
559 NEONMAP1(vbfmlalbq_f32, arm_neon_bfmlalb, 0),
560 NEONMAP1(vbfmlaltq_f32, arm_neon_bfmlalt, 0),
561 NEONMAP1(vbfmmlaq_f32, arm_neon_bfmmla, 0),
574 NEONMAP1(vcage_v, arm_neon_vacge, 0),
575 NEONMAP1(vcageq_v, arm_neon_vacge, 0),
576 NEONMAP1(vcagt_v, arm_neon_vacgt, 0),
577 NEONMAP1(vcagtq_v, arm_neon_vacgt, 0),
578 NEONMAP1(vcale_v, arm_neon_vacge, 0),
579 NEONMAP1(vcaleq_v, arm_neon_vacge, 0),
580 NEONMAP1(vcalt_v, arm_neon_vacgt, 0),
581 NEONMAP1(vcaltq_v, arm_neon_vacgt, 0),
598 NEONMAP1(vcvt_f16_f32, arm_neon_vcvtfp2hf, 0),
601 NEONMAP1(vcvt_f32_f16, arm_neon_vcvthf2fp, 0),
603 NEONMAP1(vcvt_n_f16_s16, arm_neon_vcvtfxs2fp, 0),
604 NEONMAP1(vcvt_n_f16_u16, arm_neon_vcvtfxu2fp, 0),
605 NEONMAP2(vcvt_n_f32_v, arm_neon_vcvtfxu2fp, arm_neon_vcvtfxs2fp, 0),
606 NEONMAP1(vcvt_n_s16_f16, arm_neon_vcvtfp2fxs, 0),
607 NEONMAP1(vcvt_n_s32_v, arm_neon_vcvtfp2fxs, 0),
608 NEONMAP1(vcvt_n_s64_v, arm_neon_vcvtfp2fxs, 0),
609 NEONMAP1(vcvt_n_u16_f16, arm_neon_vcvtfp2fxu, 0),
610 NEONMAP1(vcvt_n_u32_v, arm_neon_vcvtfp2fxu, 0),
611 NEONMAP1(vcvt_n_u64_v, arm_neon_vcvtfp2fxu, 0),
618 NEONMAP1(vcvta_s16_f16, arm_neon_vcvtas, 0),
619 NEONMAP1(vcvta_s32_v, arm_neon_vcvtas, 0),
620 NEONMAP1(vcvta_s64_v, arm_neon_vcvtas, 0),
621 NEONMAP1(vcvta_u16_f16, arm_neon_vcvtau, 0),
622 NEONMAP1(vcvta_u32_v, arm_neon_vcvtau, 0),
623 NEONMAP1(vcvta_u64_v, arm_neon_vcvtau, 0),
624 NEONMAP1(vcvtaq_s16_f16, arm_neon_vcvtas, 0),
625 NEONMAP1(vcvtaq_s32_v, arm_neon_vcvtas, 0),
626 NEONMAP1(vcvtaq_s64_v, arm_neon_vcvtas, 0),
627 NEONMAP1(vcvtaq_u16_f16, arm_neon_vcvtau, 0),
628 NEONMAP1(vcvtaq_u32_v, arm_neon_vcvtau, 0),
629 NEONMAP1(vcvtaq_u64_v, arm_neon_vcvtau, 0),
630 NEONMAP1(vcvth_bf16_f32, arm_neon_vcvtbfp2bf, 0),
631 NEONMAP1(vcvtm_s16_f16, arm_neon_vcvtms, 0),
632 NEONMAP1(vcvtm_s32_v, arm_neon_vcvtms, 0),
633 NEONMAP1(vcvtm_s64_v, arm_neon_vcvtms, 0),
634 NEONMAP1(vcvtm_u16_f16, arm_neon_vcvtmu, 0),
635 NEONMAP1(vcvtm_u32_v, arm_neon_vcvtmu, 0),
636 NEONMAP1(vcvtm_u64_v, arm_neon_vcvtmu, 0),
637 NEONMAP1(vcvtmq_s16_f16, arm_neon_vcvtms, 0),
638 NEONMAP1(vcvtmq_s32_v, arm_neon_vcvtms, 0),
639 NEONMAP1(vcvtmq_s64_v, arm_neon_vcvtms, 0),
640 NEONMAP1(vcvtmq_u16_f16, arm_neon_vcvtmu, 0),
641 NEONMAP1(vcvtmq_u32_v, arm_neon_vcvtmu, 0),
642 NEONMAP1(vcvtmq_u64_v, arm_neon_vcvtmu, 0),
643 NEONMAP1(vcvtn_s16_f16, arm_neon_vcvtns, 0),
644 NEONMAP1(vcvtn_s32_v, arm_neon_vcvtns, 0),
645 NEONMAP1(vcvtn_s64_v, arm_neon_vcvtns, 0),
646 NEONMAP1(vcvtn_u16_f16, arm_neon_vcvtnu, 0),
647 NEONMAP1(vcvtn_u32_v, arm_neon_vcvtnu, 0),
648 NEONMAP1(vcvtn_u64_v, arm_neon_vcvtnu, 0),
649 NEONMAP1(vcvtnq_s16_f16, arm_neon_vcvtns, 0),
650 NEONMAP1(vcvtnq_s32_v, arm_neon_vcvtns, 0),
651 NEONMAP1(vcvtnq_s64_v, arm_neon_vcvtns, 0),
652 NEONMAP1(vcvtnq_u16_f16, arm_neon_vcvtnu, 0),
653 NEONMAP1(vcvtnq_u32_v, arm_neon_vcvtnu, 0),
654 NEONMAP1(vcvtnq_u64_v, arm_neon_vcvtnu, 0),
655 NEONMAP1(vcvtp_s16_f16, arm_neon_vcvtps, 0),
656 NEONMAP1(vcvtp_s32_v, arm_neon_vcvtps, 0),
657 NEONMAP1(vcvtp_s64_v, arm_neon_vcvtps, 0),
658 NEONMAP1(vcvtp_u16_f16, arm_neon_vcvtpu, 0),
659 NEONMAP1(vcvtp_u32_v, arm_neon_vcvtpu, 0),
660 NEONMAP1(vcvtp_u64_v, arm_neon_vcvtpu, 0),
661 NEONMAP1(vcvtpq_s16_f16, arm_neon_vcvtps, 0),
662 NEONMAP1(vcvtpq_s32_v, arm_neon_vcvtps, 0),
663 NEONMAP1(vcvtpq_s64_v, arm_neon_vcvtps, 0),
664 NEONMAP1(vcvtpq_u16_f16, arm_neon_vcvtpu, 0),
665 NEONMAP1(vcvtpq_u32_v, arm_neon_vcvtpu, 0),
666 NEONMAP1(vcvtpq_u64_v, arm_neon_vcvtpu, 0),
670 NEONMAP1(vcvtq_n_f16_s16, arm_neon_vcvtfxs2fp, 0),
671 NEONMAP1(vcvtq_n_f16_u16, arm_neon_vcvtfxu2fp, 0),
672 NEONMAP2(vcvtq_n_f32_v, arm_neon_vcvtfxu2fp, arm_neon_vcvtfxs2fp, 0),
673 NEONMAP1(vcvtq_n_s16_f16, arm_neon_vcvtfp2fxs, 0),
674 NEONMAP1(vcvtq_n_s32_v, arm_neon_vcvtfp2fxs, 0),
675 NEONMAP1(vcvtq_n_s64_v, arm_neon_vcvtfp2fxs, 0),
676 NEONMAP1(vcvtq_n_u16_f16, arm_neon_vcvtfp2fxu, 0),
677 NEONMAP1(vcvtq_n_u32_v, arm_neon_vcvtfp2fxu, 0),
678 NEONMAP1(vcvtq_n_u64_v, arm_neon_vcvtfp2fxu, 0),
685 NEONMAP1(vdot_s32, arm_neon_sdot, 0),
686 NEONMAP1(vdot_u32, arm_neon_udot, 0),
687 NEONMAP1(vdotq_s32, arm_neon_sdot, 0),
688 NEONMAP1(vdotq_u32, arm_neon_udot, 0),
699 NEONMAP1(vld1_x2_v, arm_neon_vld1x2, 0),
700 NEONMAP1(vld1_x3_v, arm_neon_vld1x3, 0),
701 NEONMAP1(vld1_x4_v, arm_neon_vld1x4, 0),
703 NEONMAP1(vld1q_v, arm_neon_vld1, 0),
704 NEONMAP1(vld1q_x2_v, arm_neon_vld1x2, 0),
705 NEONMAP1(vld1q_x3_v, arm_neon_vld1x3, 0),
706 NEONMAP1(vld1q_x4_v, arm_neon_vld1x4, 0),
707 NEONMAP1(vld2_dup_v, arm_neon_vld2dup, 0),
708 NEONMAP1(vld2_lane_v, arm_neon_vld2lane, 0),
710 NEONMAP1(vld2q_dup_v, arm_neon_vld2dup, 0),
711 NEONMAP1(vld2q_lane_v, arm_neon_vld2lane, 0),
712 NEONMAP1(vld2q_v, arm_neon_vld2, 0),
713 NEONMAP1(vld3_dup_v, arm_neon_vld3dup, 0),
714 NEONMAP1(vld3_lane_v, arm_neon_vld3lane, 0),
716 NEONMAP1(vld3q_dup_v, arm_neon_vld3dup, 0),
717 NEONMAP1(vld3q_lane_v, arm_neon_vld3lane, 0),
718 NEONMAP1(vld3q_v, arm_neon_vld3, 0),
719 NEONMAP1(vld4_dup_v, arm_neon_vld4dup, 0),
720 NEONMAP1(vld4_lane_v, arm_neon_vld4lane, 0),
722 NEONMAP1(vld4q_dup_v, arm_neon_vld4dup, 0),
723 NEONMAP1(vld4q_lane_v, arm_neon_vld4lane, 0),
724 NEONMAP1(vld4q_v, arm_neon_vld4, 0),
733 NEONMAP1(vmmlaq_s32, arm_neon_smmla, 0),
734 NEONMAP1(vmmlaq_u32, arm_neon_ummla, 0),
752 NEONMAP2(vqdmlal_v, arm_neon_vqdmull, sadd_sat, 0),
753 NEONMAP2(vqdmlsl_v, arm_neon_vqdmull, ssub_sat, 0),
777 NEONMAP1(vqshlu_n_v, arm_neon_vqshiftsu, 0),
778 NEONMAP1(vqshluq_n_v, arm_neon_vqshiftsu, 0),
782 NEONMAP2(vrecpe_v, arm_neon_vrecpe, arm_neon_vrecpe, 0),
783 NEONMAP2(vrecpeq_v, arm_neon_vrecpe, arm_neon_vrecpe, 0),
806 NEONMAP2(vrsqrte_v, arm_neon_vrsqrte, arm_neon_vrsqrte, 0),
807 NEONMAP2(vrsqrteq_v, arm_neon_vrsqrte, arm_neon_vrsqrte, 0),
811 NEONMAP1(vsha1su0q_u32, arm_neon_sha1su0, 0),
812 NEONMAP1(vsha1su1q_u32, arm_neon_sha1su1, 0),
813 NEONMAP1(vsha256h2q_u32, arm_neon_sha256h2, 0),
814 NEONMAP1(vsha256hq_u32, arm_neon_sha256h, 0),
815 NEONMAP1(vsha256su0q_u32, arm_neon_sha256su0, 0),
816 NEONMAP1(vsha256su1q_u32, arm_neon_sha256su1, 0),
826 NEONMAP1(vst1_x2_v, arm_neon_vst1x2, 0),
827 NEONMAP1(vst1_x3_v, arm_neon_vst1x3, 0),
828 NEONMAP1(vst1_x4_v, arm_neon_vst1x4, 0),
829 NEONMAP1(vst1q_v, arm_neon_vst1, 0),
830 NEONMAP1(vst1q_x2_v, arm_neon_vst1x2, 0),
831 NEONMAP1(vst1q_x3_v, arm_neon_vst1x3, 0),
832 NEONMAP1(vst1q_x4_v, arm_neon_vst1x4, 0),
833 NEONMAP1(vst2_lane_v, arm_neon_vst2lane, 0),
835 NEONMAP1(vst2q_lane_v, arm_neon_vst2lane, 0),
836 NEONMAP1(vst2q_v, arm_neon_vst2, 0),
837 NEONMAP1(vst3_lane_v, arm_neon_vst3lane, 0),
839 NEONMAP1(vst3q_lane_v, arm_neon_vst3lane, 0),
840 NEONMAP1(vst3q_v, arm_neon_vst3, 0),
841 NEONMAP1(vst4_lane_v, arm_neon_vst4lane, 0),
843 NEONMAP1(vst4q_lane_v, arm_neon_vst4lane, 0),
844 NEONMAP1(vst4q_v, arm_neon_vst4, 0),
850 NEONMAP1(vusdot_s32, arm_neon_usdot, 0),
851 NEONMAP1(vusdotq_s32, arm_neon_usdot, 0),
852 NEONMAP1(vusmmlaq_s32, arm_neon_usmmla, 0),
1185 unsigned BuiltinID,
unsigned LLVMIntrinsic,
unsigned AltLLVMIntrinsic,
1186 const char *NameHint,
unsigned Modifier,
const CallExpr *E,
1188 llvm::Triple::ArchType
Arch) {
1194 std::optional<llvm::APSInt> NeonTypeConst =
1201 const bool Usgn =
Type.isUnsigned();
1202 const bool Quad =
Type.isQuad();
1203 const bool Floating =
Type.isFloatingPoint();
1205 const bool AllowBFloatArgsAndRet =
1208 llvm::FixedVectorType *VTy =
1209 GetNeonType(
this,
Type, HasFastHalfType,
false, AllowBFloatArgsAndRet);
1210 llvm::Type *Ty = VTy;
1214 auto getAlignmentValue32 = [&](
Address addr) ->
Value* {
1215 return Builder.getInt32(addr.getAlignment().getQuantity());
1218 unsigned Int = LLVMIntrinsic;
1220 Int = AltLLVMIntrinsic;
1222 switch (BuiltinID) {
1224 case NEON::BI__builtin_neon_splat_lane_v:
1225 case NEON::BI__builtin_neon_splat_laneq_v:
1226 case NEON::BI__builtin_neon_splatq_lane_v:
1227 case NEON::BI__builtin_neon_splatq_laneq_v: {
1228 auto NumElements = VTy->getElementCount();
1229 if (BuiltinID == NEON::BI__builtin_neon_splatq_lane_v)
1230 NumElements = NumElements * 2;
1231 if (BuiltinID == NEON::BI__builtin_neon_splat_laneq_v)
1232 NumElements = NumElements.divideCoefficientBy(2);
1234 Ops[0] =
Builder.CreateBitCast(Ops[0], VTy);
1237 case NEON::BI__builtin_neon_vpadd_v:
1238 case NEON::BI__builtin_neon_vpaddq_v:
1240 if (VTy->getElementType()->isFloatingPointTy() &&
1241 Int == Intrinsic::aarch64_neon_addp)
1242 Int = Intrinsic::aarch64_neon_faddp;
1244 case NEON::BI__builtin_neon_vabs_v:
1245 case NEON::BI__builtin_neon_vabsq_v:
1246 if (VTy->getElementType()->isFloatingPointTy())
1247 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::fabs, Ty), Ops,
"vabs");
1248 return EmitNeonCall(
CGM.getIntrinsic(LLVMIntrinsic, Ty), Ops,
"vabs");
1249 case NEON::BI__builtin_neon_vadd_v:
1250 case NEON::BI__builtin_neon_vaddq_v: {
1251 llvm::Type *VTy = llvm::FixedVectorType::get(
Int8Ty, Quad ? 16 : 8);
1252 Ops[0] =
Builder.CreateBitCast(Ops[0], VTy);
1253 Ops[1] =
Builder.CreateBitCast(Ops[1], VTy);
1254 Ops[0] =
Builder.CreateXor(Ops[0], Ops[1]);
1255 return Builder.CreateBitCast(Ops[0], Ty);
1257 case NEON::BI__builtin_neon_vaddhn_v: {
1258 llvm::FixedVectorType *SrcTy =
1259 llvm::FixedVectorType::getExtendedElementVectorType(VTy);
1262 Ops[0] =
Builder.CreateBitCast(Ops[0], SrcTy);
1263 Ops[1] =
Builder.CreateBitCast(Ops[1], SrcTy);
1264 Ops[0] =
Builder.CreateAdd(Ops[0], Ops[1],
"vaddhn");
1267 Constant *ShiftAmt =
1268 ConstantInt::get(SrcTy, SrcTy->getScalarSizeInBits() / 2);
1269 Ops[0] =
Builder.CreateLShr(Ops[0], ShiftAmt,
"vaddhn");
1272 return Builder.CreateTrunc(Ops[0], VTy,
"vaddhn");
1274 case NEON::BI__builtin_neon_vcale_v:
1275 case NEON::BI__builtin_neon_vcaleq_v:
1276 case NEON::BI__builtin_neon_vcalt_v:
1277 case NEON::BI__builtin_neon_vcaltq_v:
1278 std::swap(Ops[0], Ops[1]);
1280 case NEON::BI__builtin_neon_vcage_v:
1281 case NEON::BI__builtin_neon_vcageq_v:
1282 case NEON::BI__builtin_neon_vcagt_v:
1283 case NEON::BI__builtin_neon_vcagtq_v: {
1285 switch (VTy->getScalarSizeInBits()) {
1286 default: llvm_unreachable(
"unexpected type");
1297 auto *VecFlt = llvm::FixedVectorType::get(Ty, VTy->getNumElements());
1298 llvm::Type *Tys[] = { VTy, VecFlt };
1299 Function *F =
CGM.getIntrinsic(LLVMIntrinsic, Tys);
1302 case NEON::BI__builtin_neon_vceqz_v:
1303 case NEON::BI__builtin_neon_vceqzq_v:
1305 Ops[0], Ty, Floating ? ICmpInst::FCMP_OEQ : ICmpInst::ICMP_EQ,
"vceqz");
1306 case NEON::BI__builtin_neon_vcgez_v:
1307 case NEON::BI__builtin_neon_vcgezq_v:
1309 Ops[0], Ty, Floating ? ICmpInst::FCMP_OGE : ICmpInst::ICMP_SGE,
1311 case NEON::BI__builtin_neon_vclez_v:
1312 case NEON::BI__builtin_neon_vclezq_v:
1314 Ops[0], Ty, Floating ? ICmpInst::FCMP_OLE : ICmpInst::ICMP_SLE,
1316 case NEON::BI__builtin_neon_vcgtz_v:
1317 case NEON::BI__builtin_neon_vcgtzq_v:
1319 Ops[0], Ty, Floating ? ICmpInst::FCMP_OGT : ICmpInst::ICMP_SGT,
1321 case NEON::BI__builtin_neon_vcltz_v:
1322 case NEON::BI__builtin_neon_vcltzq_v:
1324 Ops[0], Ty, Floating ? ICmpInst::FCMP_OLT : ICmpInst::ICMP_SLT,
1326 case NEON::BI__builtin_neon_vclz_v:
1327 case NEON::BI__builtin_neon_vclzq_v:
1332 case NEON::BI__builtin_neon_vcvt_f32_v:
1333 case NEON::BI__builtin_neon_vcvtq_f32_v:
1334 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
1337 return Usgn ?
Builder.CreateUIToFP(Ops[0], Ty,
"vcvt")
1338 :
Builder.CreateSIToFP(Ops[0], Ty,
"vcvt");
1339 case NEON::BI__builtin_neon_vcvt_f16_s16:
1340 case NEON::BI__builtin_neon_vcvt_f16_u16:
1341 case NEON::BI__builtin_neon_vcvtq_f16_s16:
1342 case NEON::BI__builtin_neon_vcvtq_f16_u16:
1343 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
1346 return Usgn ?
Builder.CreateUIToFP(Ops[0], Ty,
"vcvt")
1347 :
Builder.CreateSIToFP(Ops[0], Ty,
"vcvt");
1348 case NEON::BI__builtin_neon_vcvt_n_f16_s16:
1349 case NEON::BI__builtin_neon_vcvt_n_f16_u16:
1350 case NEON::BI__builtin_neon_vcvtq_n_f16_s16:
1351 case NEON::BI__builtin_neon_vcvtq_n_f16_u16: {
1356 case NEON::BI__builtin_neon_vcvt_n_f32_v:
1357 case NEON::BI__builtin_neon_vcvt_n_f64_v:
1358 case NEON::BI__builtin_neon_vcvtq_n_f32_v:
1359 case NEON::BI__builtin_neon_vcvtq_n_f64_v: {
1361 Int = Usgn ? LLVMIntrinsic : AltLLVMIntrinsic;
1365 case NEON::BI__builtin_neon_vcvt_n_s16_f16:
1366 case NEON::BI__builtin_neon_vcvt_n_s32_v:
1367 case NEON::BI__builtin_neon_vcvt_n_u16_f16:
1368 case NEON::BI__builtin_neon_vcvt_n_u32_v:
1369 case NEON::BI__builtin_neon_vcvt_n_s64_v:
1370 case NEON::BI__builtin_neon_vcvt_n_u64_v:
1371 case NEON::BI__builtin_neon_vcvtq_n_s16_f16:
1372 case NEON::BI__builtin_neon_vcvtq_n_s32_v:
1373 case NEON::BI__builtin_neon_vcvtq_n_u16_f16:
1374 case NEON::BI__builtin_neon_vcvtq_n_u32_v:
1375 case NEON::BI__builtin_neon_vcvtq_n_s64_v:
1376 case NEON::BI__builtin_neon_vcvtq_n_u64_v: {
1378 Function *F =
CGM.getIntrinsic(LLVMIntrinsic, Tys);
1381 case NEON::BI__builtin_neon_vcvt_s32_v:
1382 case NEON::BI__builtin_neon_vcvt_u32_v:
1383 case NEON::BI__builtin_neon_vcvt_s64_v:
1384 case NEON::BI__builtin_neon_vcvt_u64_v:
1385 case NEON::BI__builtin_neon_vcvt_s16_f16:
1386 case NEON::BI__builtin_neon_vcvt_u16_f16:
1387 case NEON::BI__builtin_neon_vcvtq_s32_v:
1388 case NEON::BI__builtin_neon_vcvtq_u32_v:
1389 case NEON::BI__builtin_neon_vcvtq_s64_v:
1390 case NEON::BI__builtin_neon_vcvtq_u64_v:
1391 case NEON::BI__builtin_neon_vcvtq_s16_f16:
1392 case NEON::BI__builtin_neon_vcvtq_u16_f16: {
1396 if (!
Builder.getIsFPConstrained())
1397 Int = Usgn ? Intrinsic::fptoui_sat : Intrinsic::fptosi_sat;
1398 llvm::Type *Tys[2] = {Ty, Ops[0]->getType()};
1403 return Usgn ?
Builder.CreateFPToUI(Ops[0], Ty,
"vcvt")
1404 :
Builder.CreateFPToSI(Ops[0], Ty,
"vcvt");
1406 case NEON::BI__builtin_neon_vcvta_s16_f16:
1407 case NEON::BI__builtin_neon_vcvta_s32_v:
1408 case NEON::BI__builtin_neon_vcvta_s64_v:
1409 case NEON::BI__builtin_neon_vcvta_u16_f16:
1410 case NEON::BI__builtin_neon_vcvta_u32_v:
1411 case NEON::BI__builtin_neon_vcvta_u64_v:
1412 case NEON::BI__builtin_neon_vcvtaq_s16_f16:
1413 case NEON::BI__builtin_neon_vcvtaq_s32_v:
1414 case NEON::BI__builtin_neon_vcvtaq_s64_v:
1415 case NEON::BI__builtin_neon_vcvtaq_u16_f16:
1416 case NEON::BI__builtin_neon_vcvtaq_u32_v:
1417 case NEON::BI__builtin_neon_vcvtaq_u64_v:
1418 case NEON::BI__builtin_neon_vcvtn_s16_f16:
1419 case NEON::BI__builtin_neon_vcvtn_s32_v:
1420 case NEON::BI__builtin_neon_vcvtn_s64_v:
1421 case NEON::BI__builtin_neon_vcvtn_u16_f16:
1422 case NEON::BI__builtin_neon_vcvtn_u32_v:
1423 case NEON::BI__builtin_neon_vcvtn_u64_v:
1424 case NEON::BI__builtin_neon_vcvtnq_s16_f16:
1425 case NEON::BI__builtin_neon_vcvtnq_s32_v:
1426 case NEON::BI__builtin_neon_vcvtnq_s64_v:
1427 case NEON::BI__builtin_neon_vcvtnq_u16_f16:
1428 case NEON::BI__builtin_neon_vcvtnq_u32_v:
1429 case NEON::BI__builtin_neon_vcvtnq_u64_v:
1430 case NEON::BI__builtin_neon_vcvtp_s16_f16:
1431 case NEON::BI__builtin_neon_vcvtp_s32_v:
1432 case NEON::BI__builtin_neon_vcvtp_s64_v:
1433 case NEON::BI__builtin_neon_vcvtp_u16_f16:
1434 case NEON::BI__builtin_neon_vcvtp_u32_v:
1435 case NEON::BI__builtin_neon_vcvtp_u64_v:
1436 case NEON::BI__builtin_neon_vcvtpq_s16_f16:
1437 case NEON::BI__builtin_neon_vcvtpq_s32_v:
1438 case NEON::BI__builtin_neon_vcvtpq_s64_v:
1439 case NEON::BI__builtin_neon_vcvtpq_u16_f16:
1440 case NEON::BI__builtin_neon_vcvtpq_u32_v:
1441 case NEON::BI__builtin_neon_vcvtpq_u64_v:
1442 case NEON::BI__builtin_neon_vcvtm_s16_f16:
1443 case NEON::BI__builtin_neon_vcvtm_s32_v:
1444 case NEON::BI__builtin_neon_vcvtm_s64_v:
1445 case NEON::BI__builtin_neon_vcvtm_u16_f16:
1446 case NEON::BI__builtin_neon_vcvtm_u32_v:
1447 case NEON::BI__builtin_neon_vcvtm_u64_v:
1448 case NEON::BI__builtin_neon_vcvtmq_s16_f16:
1449 case NEON::BI__builtin_neon_vcvtmq_s32_v:
1450 case NEON::BI__builtin_neon_vcvtmq_s64_v:
1451 case NEON::BI__builtin_neon_vcvtmq_u16_f16:
1452 case NEON::BI__builtin_neon_vcvtmq_u32_v:
1453 case NEON::BI__builtin_neon_vcvtmq_u64_v: {
1455 return EmitNeonCall(
CGM.getIntrinsic(LLVMIntrinsic, Tys), Ops, NameHint);
1457 case NEON::BI__builtin_neon_vcvtx_f32_v: {
1458 llvm::Type *Tys[2] = { VTy->getTruncatedElementVectorType(VTy), Ty};
1459 return EmitNeonCall(
CGM.getIntrinsic(LLVMIntrinsic, Tys), Ops, NameHint);
1462 case NEON::BI__builtin_neon_vext_v:
1463 case NEON::BI__builtin_neon_vextq_v: {
1466 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; ++i)
1467 Indices.push_back(i+CV);
1469 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
1470 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
1471 return Builder.CreateShuffleVector(Ops[0], Ops[1], Indices,
"vext");
1473 case NEON::BI__builtin_neon_vfma_v:
1474 case NEON::BI__builtin_neon_vfmaq_v: {
1475 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
1476 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
1477 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
1481 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma, Ty,
1482 {Ops[1], Ops[2], Ops[0]});
1484 case NEON::BI__builtin_neon_vld1_v:
1485 case NEON::BI__builtin_neon_vld1q_v: {
1487 Ops.push_back(getAlignmentValue32(PtrOp0));
1488 return EmitNeonCall(
CGM.getIntrinsic(LLVMIntrinsic, Tys), Ops,
"vld1");
1490 case NEON::BI__builtin_neon_vld1_x2_v:
1491 case NEON::BI__builtin_neon_vld1q_x2_v:
1492 case NEON::BI__builtin_neon_vld1_x3_v:
1493 case NEON::BI__builtin_neon_vld1q_x3_v:
1494 case NEON::BI__builtin_neon_vld1_x4_v:
1495 case NEON::BI__builtin_neon_vld1q_x4_v: {
1497 Function *F =
CGM.getIntrinsic(LLVMIntrinsic, Tys);
1498 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld1xN");
1499 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
1501 case NEON::BI__builtin_neon_vld2_v:
1502 case NEON::BI__builtin_neon_vld2q_v:
1503 case NEON::BI__builtin_neon_vld3_v:
1504 case NEON::BI__builtin_neon_vld3q_v:
1505 case NEON::BI__builtin_neon_vld4_v:
1506 case NEON::BI__builtin_neon_vld4q_v:
1507 case NEON::BI__builtin_neon_vld2_dup_v:
1508 case NEON::BI__builtin_neon_vld2q_dup_v:
1509 case NEON::BI__builtin_neon_vld3_dup_v:
1510 case NEON::BI__builtin_neon_vld3q_dup_v:
1511 case NEON::BI__builtin_neon_vld4_dup_v:
1512 case NEON::BI__builtin_neon_vld4q_dup_v: {
1514 Function *F =
CGM.getIntrinsic(LLVMIntrinsic, Tys);
1515 Value *Align = getAlignmentValue32(PtrOp1);
1516 Ops[1] =
Builder.CreateCall(F, {Ops[1], Align}, NameHint);
1517 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
1519 case NEON::BI__builtin_neon_vld1_dup_v:
1520 case NEON::BI__builtin_neon_vld1q_dup_v: {
1521 Value *
V = PoisonValue::get(Ty);
1523 LoadInst *Ld =
Builder.CreateLoad(PtrOp0);
1524 llvm::Constant *CI = ConstantInt::get(
SizeTy, 0);
1525 Ops[0] =
Builder.CreateInsertElement(
V, Ld, CI);
1528 case NEON::BI__builtin_neon_vld2_lane_v:
1529 case NEON::BI__builtin_neon_vld2q_lane_v:
1530 case NEON::BI__builtin_neon_vld3_lane_v:
1531 case NEON::BI__builtin_neon_vld3q_lane_v:
1532 case NEON::BI__builtin_neon_vld4_lane_v:
1533 case NEON::BI__builtin_neon_vld4q_lane_v: {
1535 Function *F =
CGM.getIntrinsic(LLVMIntrinsic, Tys);
1536 for (
unsigned I = 2; I < Ops.size() - 1; ++I)
1537 Ops[I] =
Builder.CreateBitCast(Ops[I], Ty);
1538 Ops.push_back(getAlignmentValue32(PtrOp1));
1540 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
1542 case NEON::BI__builtin_neon_vmovl_v: {
1543 llvm::FixedVectorType *DTy =
1544 llvm::FixedVectorType::getTruncatedElementVectorType(VTy);
1545 Ops[0] =
Builder.CreateBitCast(Ops[0], DTy);
1547 return Builder.CreateZExt(Ops[0], Ty,
"vmovl");
1548 return Builder.CreateSExt(Ops[0], Ty,
"vmovl");
1550 case NEON::BI__builtin_neon_vmovn_v: {
1551 llvm::FixedVectorType *QTy =
1552 llvm::FixedVectorType::getExtendedElementVectorType(VTy);
1553 Ops[0] =
Builder.CreateBitCast(Ops[0], QTy);
1554 return Builder.CreateTrunc(Ops[0], Ty,
"vmovn");
1556 case NEON::BI__builtin_neon_vmull_v:
1562 Int = Usgn ? Intrinsic::arm_neon_vmullu : Intrinsic::arm_neon_vmulls;
1563 Int =
Type.isPoly() ? (
unsigned)Intrinsic::arm_neon_vmullp : Int;
1565 case NEON::BI__builtin_neon_vpadal_v:
1566 case NEON::BI__builtin_neon_vpadalq_v: {
1568 unsigned EltBits = VTy->getElementType()->getPrimitiveSizeInBits();
1572 llvm::FixedVectorType::get(EltTy, VTy->getNumElements() * 2);
1573 llvm::Type *Tys[2] = { Ty, NarrowTy };
1576 case NEON::BI__builtin_neon_vpaddl_v:
1577 case NEON::BI__builtin_neon_vpaddlq_v: {
1579 unsigned EltBits = VTy->getElementType()->getPrimitiveSizeInBits();
1580 llvm::Type *EltTy = llvm::IntegerType::get(
getLLVMContext(), EltBits / 2);
1582 llvm::FixedVectorType::get(EltTy, VTy->getNumElements() * 2);
1583 llvm::Type *Tys[2] = { Ty, NarrowTy };
1586 case NEON::BI__builtin_neon_vqdmlal_v:
1587 case NEON::BI__builtin_neon_vqdmlsl_v: {
1592 return EmitNeonCall(
CGM.getIntrinsic(AltLLVMIntrinsic, Ty), Ops, NameHint);
1594 case NEON::BI__builtin_neon_vqdmulhq_lane_v:
1595 case NEON::BI__builtin_neon_vqdmulh_lane_v:
1596 case NEON::BI__builtin_neon_vqrdmulhq_lane_v:
1597 case NEON::BI__builtin_neon_vqrdmulh_lane_v: {
1599 if (BuiltinID == NEON::BI__builtin_neon_vqdmulhq_lane_v ||
1600 BuiltinID == NEON::BI__builtin_neon_vqrdmulhq_lane_v)
1601 RTy = llvm::FixedVectorType::get(RTy->getElementType(),
1602 RTy->getNumElements() * 2);
1603 llvm::Type *Tys[2] = {
1608 case NEON::BI__builtin_neon_vqdmulhq_laneq_v:
1609 case NEON::BI__builtin_neon_vqdmulh_laneq_v:
1610 case NEON::BI__builtin_neon_vqrdmulhq_laneq_v:
1611 case NEON::BI__builtin_neon_vqrdmulh_laneq_v: {
1612 llvm::Type *Tys[2] = {
1617 case NEON::BI__builtin_neon_vqshl_n_v:
1618 case NEON::BI__builtin_neon_vqshlq_n_v:
1621 case NEON::BI__builtin_neon_vqshlu_n_v:
1622 case NEON::BI__builtin_neon_vqshluq_n_v:
1625 case NEON::BI__builtin_neon_vrecpe_v:
1626 case NEON::BI__builtin_neon_vrecpeq_v:
1627 case NEON::BI__builtin_neon_vrsqrte_v:
1628 case NEON::BI__builtin_neon_vrsqrteq_v:
1629 Int = Ty->isFPOrFPVectorTy() ? LLVMIntrinsic : AltLLVMIntrinsic;
1631 case NEON::BI__builtin_neon_vrndi_v:
1632 case NEON::BI__builtin_neon_vrndiq_v:
1633 Int =
Builder.getIsFPConstrained()
1634 ? Intrinsic::experimental_constrained_nearbyint
1635 : Intrinsic::nearbyint;
1637 case NEON::BI__builtin_neon_vrshr_n_v:
1638 case NEON::BI__builtin_neon_vrshrq_n_v:
1641 case NEON::BI__builtin_neon_vsha512hq_u64:
1642 case NEON::BI__builtin_neon_vsha512h2q_u64:
1643 case NEON::BI__builtin_neon_vsha512su0q_u64:
1644 case NEON::BI__builtin_neon_vsha512su1q_u64: {
1648 case NEON::BI__builtin_neon_vshl_n_v:
1649 case NEON::BI__builtin_neon_vshlq_n_v:
1651 return Builder.CreateShl(
Builder.CreateBitCast(Ops[0],Ty), Ops[1],
1653 case NEON::BI__builtin_neon_vshll_n_v: {
1654 llvm::FixedVectorType *SrcTy =
1655 llvm::FixedVectorType::getTruncatedElementVectorType(VTy);
1656 Ops[0] =
Builder.CreateBitCast(Ops[0], SrcTy);
1658 Ops[0] =
Builder.CreateZExt(Ops[0], VTy);
1660 Ops[0] =
Builder.CreateSExt(Ops[0], VTy);
1662 return Builder.CreateShl(Ops[0], Ops[1],
"vshll_n");
1664 case NEON::BI__builtin_neon_vshrn_n_v: {
1665 llvm::FixedVectorType *SrcTy =
1666 llvm::FixedVectorType::getExtendedElementVectorType(VTy);
1667 Ops[0] =
Builder.CreateBitCast(Ops[0], SrcTy);
1670 Ops[0] =
Builder.CreateLShr(Ops[0], Ops[1]);
1672 Ops[0] =
Builder.CreateAShr(Ops[0], Ops[1]);
1673 return Builder.CreateTrunc(Ops[0], Ty,
"vshrn_n");
1675 case NEON::BI__builtin_neon_vshr_n_v:
1676 case NEON::BI__builtin_neon_vshrq_n_v:
1678 case NEON::BI__builtin_neon_vst1_v:
1679 case NEON::BI__builtin_neon_vst1q_v:
1680 case NEON::BI__builtin_neon_vst2_v:
1681 case NEON::BI__builtin_neon_vst2q_v:
1682 case NEON::BI__builtin_neon_vst3_v:
1683 case NEON::BI__builtin_neon_vst3q_v:
1684 case NEON::BI__builtin_neon_vst4_v:
1685 case NEON::BI__builtin_neon_vst4q_v:
1686 case NEON::BI__builtin_neon_vst2_lane_v:
1687 case NEON::BI__builtin_neon_vst2q_lane_v:
1688 case NEON::BI__builtin_neon_vst3_lane_v:
1689 case NEON::BI__builtin_neon_vst3q_lane_v:
1690 case NEON::BI__builtin_neon_vst4_lane_v:
1691 case NEON::BI__builtin_neon_vst4q_lane_v: {
1693 Ops.push_back(getAlignmentValue32(PtrOp0));
1696 case NEON::BI__builtin_neon_vsm3partw1q_u32:
1697 case NEON::BI__builtin_neon_vsm3partw2q_u32:
1698 case NEON::BI__builtin_neon_vsm3ss1q_u32:
1699 case NEON::BI__builtin_neon_vsm4ekeyq_u32:
1700 case NEON::BI__builtin_neon_vsm4eq_u32: {
1704 case NEON::BI__builtin_neon_vsm3tt1aq_u32:
1705 case NEON::BI__builtin_neon_vsm3tt1bq_u32:
1706 case NEON::BI__builtin_neon_vsm3tt2aq_u32:
1707 case NEON::BI__builtin_neon_vsm3tt2bq_u32: {
1712 case NEON::BI__builtin_neon_vst1_x2_v:
1713 case NEON::BI__builtin_neon_vst1q_x2_v:
1714 case NEON::BI__builtin_neon_vst1_x3_v:
1715 case NEON::BI__builtin_neon_vst1q_x3_v:
1716 case NEON::BI__builtin_neon_vst1_x4_v:
1717 case NEON::BI__builtin_neon_vst1q_x4_v: {
1720 if (
Arch == llvm::Triple::aarch64 ||
Arch == llvm::Triple::aarch64_be ||
1721 Arch == llvm::Triple::aarch64_32) {
1723 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
1729 case NEON::BI__builtin_neon_vsubhn_v: {
1730 llvm::FixedVectorType *SrcTy =
1731 llvm::FixedVectorType::getExtendedElementVectorType(VTy);
1734 Ops[0] =
Builder.CreateBitCast(Ops[0], SrcTy);
1735 Ops[1] =
Builder.CreateBitCast(Ops[1], SrcTy);
1736 Ops[0] =
Builder.CreateSub(Ops[0], Ops[1],
"vsubhn");
1739 Constant *ShiftAmt =
1740 ConstantInt::get(SrcTy, SrcTy->getScalarSizeInBits() / 2);
1741 Ops[0] =
Builder.CreateLShr(Ops[0], ShiftAmt,
"vsubhn");
1744 return Builder.CreateTrunc(Ops[0], VTy,
"vsubhn");
1746 case NEON::BI__builtin_neon_vtrn_v:
1747 case NEON::BI__builtin_neon_vtrnq_v: {
1748 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
1749 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
1750 Value *SV =
nullptr;
1752 for (
unsigned vi = 0; vi != 2; ++vi) {
1754 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; i += 2) {
1755 Indices.push_back(i+vi);
1756 Indices.push_back(i+e+vi);
1759 SV =
Builder.CreateShuffleVector(Ops[1], Ops[2], Indices,
"vtrn");
1764 case NEON::BI__builtin_neon_vtst_v:
1765 case NEON::BI__builtin_neon_vtstq_v: {
1766 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
1767 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
1768 Ops[0] =
Builder.CreateAnd(Ops[0], Ops[1]);
1769 Ops[0] =
Builder.CreateICmp(ICmpInst::ICMP_NE, Ops[0],
1770 ConstantAggregateZero::get(Ty));
1771 return Builder.CreateSExt(Ops[0], Ty,
"vtst");
1773 case NEON::BI__builtin_neon_vuzp_v:
1774 case NEON::BI__builtin_neon_vuzpq_v: {
1775 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
1776 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
1777 Value *SV =
nullptr;
1779 for (
unsigned vi = 0; vi != 2; ++vi) {
1781 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; ++i)
1782 Indices.push_back(2*i+vi);
1785 SV =
Builder.CreateShuffleVector(Ops[1], Ops[2], Indices,
"vuzp");
1790 case NEON::BI__builtin_neon_vxarq_u64: {
1795 case NEON::BI__builtin_neon_vzip_v:
1796 case NEON::BI__builtin_neon_vzipq_v: {
1797 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
1798 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
1799 Value *SV =
nullptr;
1801 for (
unsigned vi = 0; vi != 2; ++vi) {
1803 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; i += 2) {
1804 Indices.push_back((i + vi*e) >> 1);
1805 Indices.push_back(((i + vi*e) >> 1)+e);
1808 SV =
Builder.CreateShuffleVector(Ops[1], Ops[2], Indices,
"vzip");
1813 case NEON::BI__builtin_neon_vdot_s32:
1814 case NEON::BI__builtin_neon_vdot_u32:
1815 case NEON::BI__builtin_neon_vdotq_s32:
1816 case NEON::BI__builtin_neon_vdotq_u32: {
1818 llvm::FixedVectorType::get(
Int8Ty, Ty->getPrimitiveSizeInBits() / 8);
1819 llvm::Type *Tys[2] = { Ty, InputTy };
1822 case NEON::BI__builtin_neon_vfmlal_low_f16:
1823 case NEON::BI__builtin_neon_vfmlalq_low_f16: {
1825 llvm::FixedVectorType::get(
HalfTy, Ty->getPrimitiveSizeInBits() / 16);
1826 llvm::Type *Tys[2] = { Ty, InputTy };
1829 case NEON::BI__builtin_neon_vfmlsl_low_f16:
1830 case NEON::BI__builtin_neon_vfmlslq_low_f16: {
1832 llvm::FixedVectorType::get(
HalfTy, Ty->getPrimitiveSizeInBits() / 16);
1833 llvm::Type *Tys[2] = { Ty, InputTy };
1836 case NEON::BI__builtin_neon_vfmlal_high_f16:
1837 case NEON::BI__builtin_neon_vfmlalq_high_f16: {
1839 llvm::FixedVectorType::get(
HalfTy, Ty->getPrimitiveSizeInBits() / 16);
1840 llvm::Type *Tys[2] = { Ty, InputTy };
1843 case NEON::BI__builtin_neon_vfmlsl_high_f16:
1844 case NEON::BI__builtin_neon_vfmlslq_high_f16: {
1846 llvm::FixedVectorType::get(
HalfTy, Ty->getPrimitiveSizeInBits() / 16);
1847 llvm::Type *Tys[2] = { Ty, InputTy };
1850 case NEON::BI__builtin_neon_vmmlaq_s32:
1851 case NEON::BI__builtin_neon_vmmlaq_u32: {
1853 llvm::FixedVectorType::get(
Int8Ty, Ty->getPrimitiveSizeInBits() / 8);
1854 llvm::Type *Tys[2] = { Ty, InputTy };
1855 return EmitNeonCall(
CGM.getIntrinsic(LLVMIntrinsic, Tys), Ops,
"vmmla");
1857 case NEON::BI__builtin_neon_vusmmlaq_s32: {
1859 llvm::FixedVectorType::get(
Int8Ty, Ty->getPrimitiveSizeInBits() / 8);
1860 llvm::Type *Tys[2] = { Ty, InputTy };
1863 case NEON::BI__builtin_neon_vusdot_s32:
1864 case NEON::BI__builtin_neon_vusdotq_s32: {
1866 llvm::FixedVectorType::get(
Int8Ty, Ty->getPrimitiveSizeInBits() / 8);
1867 llvm::Type *Tys[2] = { Ty, InputTy };
1870 case NEON::BI__builtin_neon_vbfdot_f32:
1871 case NEON::BI__builtin_neon_vbfdotq_f32: {
1872 llvm::Type *InputTy =
1873 llvm::FixedVectorType::get(
BFloatTy, Ty->getPrimitiveSizeInBits() / 16);
1874 llvm::Type *Tys[2] = { Ty, InputTy };
1877 case NEON::BI__builtin_neon___a32_vcvt_bf16_f32: {
1878 llvm::Type *Tys[1] = { Ty };
1885 assert(Int &&
"Expected valid intrinsic number");
4490 llvm::Triple::ArchType
Arch) {
4499 if (BuiltinID == Builtin::BI__builtin_cpu_supports)
4500 return EmitAArch64CpuSupports(E);
4502 unsigned HintID =
static_cast<unsigned>(-1);
4503 switch (BuiltinID) {
4505 case clang::AArch64::BI__builtin_arm_nop:
4508 case clang::AArch64::BI__builtin_arm_yield:
4509 case clang::AArch64::BI__yield:
4512 case clang::AArch64::BI__builtin_arm_wfe:
4513 case clang::AArch64::BI__wfe:
4516 case clang::AArch64::BI__builtin_arm_wfi:
4517 case clang::AArch64::BI__wfi:
4520 case clang::AArch64::BI__builtin_arm_sev:
4521 case clang::AArch64::BI__sev:
4524 case clang::AArch64::BI__builtin_arm_sevl:
4525 case clang::AArch64::BI__sevl:
4530 if (HintID !=
static_cast<unsigned>(-1)) {
4531 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_hint);
4532 return Builder.CreateCall(F, llvm::ConstantInt::get(
Int32Ty, HintID));
4535 if (BuiltinID == clang::AArch64::BI__builtin_arm_trap) {
4536 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_break);
4541 if (BuiltinID == clang::AArch64::BI__builtin_arm_get_sme_state) {
4544 llvm::FunctionType::get(StructType::get(
CGM.Int64Ty,
CGM.Int64Ty), {},
4546 "__arm_sme_state"));
4548 "aarch64_pstate_sm_compatible");
4549 CI->setAttributes(Attrs);
4552 AArch64_SME_ABI_Support_Routines_PreserveMost_From_X2);
4559 if (BuiltinID == clang::AArch64::BI__builtin_arm_rbit) {
4561 "rbit of unusual size!");
4564 CGM.getIntrinsic(Intrinsic::bitreverse, Arg->getType()), Arg,
"rbit");
4566 if (BuiltinID == clang::AArch64::BI__builtin_arm_rbit64) {
4568 "rbit of unusual size!");
4571 CGM.getIntrinsic(Intrinsic::bitreverse, Arg->getType()), Arg,
"rbit");
4574 if (BuiltinID == clang::AArch64::BI__builtin_arm_clz ||
4575 BuiltinID == clang::AArch64::BI__builtin_arm_clz64) {
4577 Function *F =
CGM.getIntrinsic(Intrinsic::ctlz, Arg->getType());
4579 if (BuiltinID == clang::AArch64::BI__builtin_arm_clz64)
4584 if (BuiltinID == clang::AArch64::BI__builtin_arm_cls) {
4586 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_cls), Arg,
4589 if (BuiltinID == clang::AArch64::BI__builtin_arm_cls64) {
4591 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_cls64), Arg,
4595 if (BuiltinID == clang::AArch64::BI__builtin_arm_rint32zf ||
4596 BuiltinID == clang::AArch64::BI__builtin_arm_rint32z) {
4598 llvm::Type *Ty = Arg->getType();
4599 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_frint32z, Ty),
4603 if (BuiltinID == clang::AArch64::BI__builtin_arm_rint64zf ||
4604 BuiltinID == clang::AArch64::BI__builtin_arm_rint64z) {
4606 llvm::Type *Ty = Arg->getType();
4607 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_frint64z, Ty),
4611 if (BuiltinID == clang::AArch64::BI__builtin_arm_rint32xf ||
4612 BuiltinID == clang::AArch64::BI__builtin_arm_rint32x) {
4614 llvm::Type *Ty = Arg->getType();
4615 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_frint32x, Ty),
4619 if (BuiltinID == clang::AArch64::BI__builtin_arm_rint64xf ||
4620 BuiltinID == clang::AArch64::BI__builtin_arm_rint64x) {
4622 llvm::Type *Ty = Arg->getType();
4623 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_frint64x, Ty),
4627 if (BuiltinID == clang::AArch64::BI__builtin_arm_jcvt) {
4629 "__jcvt of unusual size!");
4632 CGM.getIntrinsic(Intrinsic::aarch64_fjcvtzs), Arg);
4635 if (BuiltinID == clang::AArch64::BI__builtin_arm_ld64b ||
4636 BuiltinID == clang::AArch64::BI__builtin_arm_st64b ||
4637 BuiltinID == clang::AArch64::BI__builtin_arm_st64bv ||
4638 BuiltinID == clang::AArch64::BI__builtin_arm_st64bv0) {
4642 if (BuiltinID == clang::AArch64::BI__builtin_arm_ld64b) {
4645 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_ld64b);
4646 llvm::Value *Val =
Builder.CreateCall(F, MemAddr);
4648 for (
size_t i = 0; i < 8; i++) {
4649 llvm::Value *ValOffsetPtr =
4661 Args.push_back(MemAddr);
4662 for (
size_t i = 0; i < 8; i++) {
4663 llvm::Value *ValOffsetPtr =
4669 auto Intr = (BuiltinID == clang::AArch64::BI__builtin_arm_st64b
4670 ? Intrinsic::aarch64_st64b
4671 : BuiltinID == clang::AArch64::BI__builtin_arm_st64bv
4672 ? Intrinsic::aarch64_st64bv
4673 : Intrinsic::aarch64_st64bv0);
4675 return Builder.CreateCall(F, Args);
4678 if (BuiltinID == clang::AArch64::BI__builtin_arm_rndr ||
4679 BuiltinID == clang::AArch64::BI__builtin_arm_rndrrs) {
4681 auto Intr = (BuiltinID == clang::AArch64::BI__builtin_arm_rndr
4682 ? Intrinsic::aarch64_rndr
4683 : Intrinsic::aarch64_rndrrs);
4685 llvm::Value *Val =
Builder.CreateCall(F);
4686 Value *RandomValue =
Builder.CreateExtractValue(Val, 0);
4690 Builder.CreateStore(RandomValue, MemAddress);
4695 if (BuiltinID == clang::AArch64::BI__clear_cache) {
4696 assert(E->
getNumArgs() == 2 &&
"__clear_cache takes 2 arguments");
4699 for (
unsigned i = 0; i < 2; i++)
4701 llvm::Type *Ty =
CGM.getTypes().ConvertType(FD->
getType());
4703 StringRef Name = FD->
getName();
4707 if ((BuiltinID == clang::AArch64::BI__builtin_arm_ldrex ||
4708 BuiltinID == clang::AArch64::BI__builtin_arm_ldaex) &&
4711 CGM.getIntrinsic(BuiltinID == clang::AArch64::BI__builtin_arm_ldaex
4712 ? Intrinsic::aarch64_ldaxp
4713 : Intrinsic::aarch64_ldxp);
4720 llvm::Type *Int128Ty = llvm::IntegerType::get(
getLLVMContext(), 128);
4721 Val0 =
Builder.CreateZExt(Val0, Int128Ty);
4722 Val1 =
Builder.CreateZExt(Val1, Int128Ty);
4724 Value *ShiftCst = llvm::ConstantInt::get(Int128Ty, 64);
4725 Val =
Builder.CreateShl(Val0, ShiftCst,
"shl",
true );
4726 Val =
Builder.CreateOr(Val, Val1);
4728 }
else if (BuiltinID == clang::AArch64::BI__builtin_arm_ldrex ||
4729 BuiltinID == clang::AArch64::BI__builtin_arm_ldaex) {
4738 CGM.getIntrinsic(BuiltinID == clang::AArch64::BI__builtin_arm_ldaex
4739 ? Intrinsic::aarch64_ldaxr
4740 : Intrinsic::aarch64_ldxr,
4742 CallInst *Val =
Builder.CreateCall(F, LoadAddr,
"ldxr");
4746 if (RealResTy->isPointerTy())
4747 return Builder.CreateIntToPtr(Val, RealResTy);
4749 llvm::Type *IntResTy = llvm::IntegerType::get(
4751 return Builder.CreateBitCast(
Builder.CreateTruncOrBitCast(Val, IntResTy),
4755 if ((BuiltinID == clang::AArch64::BI__builtin_arm_strex ||
4756 BuiltinID == clang::AArch64::BI__builtin_arm_stlex) &&
4759 CGM.getIntrinsic(BuiltinID == clang::AArch64::BI__builtin_arm_stlex
4760 ? Intrinsic::aarch64_stlxp
4761 : Intrinsic::aarch64_stxp);
4768 llvm::Value *Val =
Builder.CreateLoad(Tmp);
4773 return Builder.CreateCall(F, {Arg0, Arg1, StPtr},
"stxp");
4776 if (BuiltinID == clang::AArch64::BI__builtin_arm_strex ||
4777 BuiltinID == clang::AArch64::BI__builtin_arm_stlex) {
4782 llvm::Type *StoreTy =
4785 if (StoreVal->
getType()->isPointerTy())
4788 llvm::Type *
IntTy = llvm::IntegerType::get(
4790 CGM.getDataLayout().getTypeSizeInBits(StoreVal->
getType()));
4796 CGM.getIntrinsic(BuiltinID == clang::AArch64::BI__builtin_arm_stlex
4797 ? Intrinsic::aarch64_stlxr
4798 : Intrinsic::aarch64_stxr,
4800 CallInst *CI =
Builder.CreateCall(F, {StoreVal, StoreAddr},
"stxr");
4802 1, Attribute::get(
getLLVMContext(), Attribute::ElementType, StoreTy));
4806 if (BuiltinID == clang::AArch64::BI__getReg) {
4809 llvm_unreachable(
"Sema will ensure that the parameter is constant");
4812 LLVMContext &Context =
CGM.getLLVMContext();
4815 llvm::Metadata *Ops[] = {llvm::MDString::get(Context, Reg)};
4816 llvm::MDNode *RegName = llvm::MDNode::get(Context, Ops);
4817 llvm::Value *Metadata = llvm::MetadataAsValue::get(Context, RegName);
4820 CGM.getIntrinsic(Intrinsic::read_register, {
Int64Ty});
4821 return Builder.CreateCall(F, Metadata);
4824 if (BuiltinID == clang::AArch64::BI__break) {
4827 llvm_unreachable(
"Sema will ensure that the parameter is constant");
4829 llvm::Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_break);
4833 if (BuiltinID == clang::AArch64::BI__builtin_arm_clrex) {
4834 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_clrex);
4838 if (BuiltinID == clang::AArch64::BI_ReadWriteBarrier)
4839 return Builder.CreateFence(llvm::AtomicOrdering::SequentiallyConsistent,
4840 llvm::SyncScope::SingleThread);
4843 Intrinsic::ID CRCIntrinsicID = Intrinsic::not_intrinsic;
4844 switch (BuiltinID) {
4845 case clang::AArch64::BI__builtin_arm_crc32b:
4846 CRCIntrinsicID = Intrinsic::aarch64_crc32b;
break;
4847 case clang::AArch64::BI__builtin_arm_crc32cb:
4848 CRCIntrinsicID = Intrinsic::aarch64_crc32cb;
break;
4849 case clang::AArch64::BI__builtin_arm_crc32h:
4850 CRCIntrinsicID = Intrinsic::aarch64_crc32h;
break;
4851 case clang::AArch64::BI__builtin_arm_crc32ch:
4852 CRCIntrinsicID = Intrinsic::aarch64_crc32ch;
break;
4853 case clang::AArch64::BI__builtin_arm_crc32w:
4854 CRCIntrinsicID = Intrinsic::aarch64_crc32w;
break;
4855 case clang::AArch64::BI__builtin_arm_crc32cw:
4856 CRCIntrinsicID = Intrinsic::aarch64_crc32cw;
break;
4857 case clang::AArch64::BI__builtin_arm_crc32d:
4858 CRCIntrinsicID = Intrinsic::aarch64_crc32x;
break;
4859 case clang::AArch64::BI__builtin_arm_crc32cd:
4860 CRCIntrinsicID = Intrinsic::aarch64_crc32cx;
break;
4863 if (CRCIntrinsicID != Intrinsic::not_intrinsic) {
4868 llvm::Type *DataTy = F->getFunctionType()->getParamType(1);
4869 Arg1 =
Builder.CreateZExtOrBitCast(Arg1, DataTy);
4871 return Builder.CreateCall(F, {Arg0, Arg1});
4875 if (BuiltinID == AArch64::BI__builtin_arm_mops_memset_tag) {
4882 CGM.getIntrinsic(Intrinsic::aarch64_mops_memset_tag), {Dst, Val, Size});
4885 if (BuiltinID == AArch64::BI__builtin_arm_range_prefetch ||
4886 BuiltinID == AArch64::BI__builtin_arm_range_prefetch_x)
4890 Intrinsic::ID MTEIntrinsicID = Intrinsic::not_intrinsic;
4891 switch (BuiltinID) {
4892 case clang::AArch64::BI__builtin_arm_irg:
4893 MTEIntrinsicID = Intrinsic::aarch64_irg;
break;
4894 case clang::AArch64::BI__builtin_arm_addg:
4895 MTEIntrinsicID = Intrinsic::aarch64_addg;
break;
4896 case clang::AArch64::BI__builtin_arm_gmi:
4897 MTEIntrinsicID = Intrinsic::aarch64_gmi;
break;
4898 case clang::AArch64::BI__builtin_arm_ldg:
4899 MTEIntrinsicID = Intrinsic::aarch64_ldg;
break;
4900 case clang::AArch64::BI__builtin_arm_stg:
4901 MTEIntrinsicID = Intrinsic::aarch64_stg;
break;
4902 case clang::AArch64::BI__builtin_arm_subp:
4903 MTEIntrinsicID = Intrinsic::aarch64_subp;
break;
4906 if (MTEIntrinsicID != Intrinsic::not_intrinsic) {
4907 if (MTEIntrinsicID == Intrinsic::aarch64_irg) {
4912 return Builder.CreateCall(
CGM.getIntrinsic(MTEIntrinsicID),
4915 if (MTEIntrinsicID == Intrinsic::aarch64_addg) {
4920 return Builder.CreateCall(
CGM.getIntrinsic(MTEIntrinsicID),
4921 {Pointer, TagOffset});
4923 if (MTEIntrinsicID == Intrinsic::aarch64_gmi) {
4929 CGM.getIntrinsic(MTEIntrinsicID), {Pointer, ExcludedMask});
4934 if (MTEIntrinsicID == Intrinsic::aarch64_ldg) {
4936 return Builder.CreateCall(
CGM.getIntrinsic(MTEIntrinsicID),
4937 {TagAddress, TagAddress});
4942 if (MTEIntrinsicID == Intrinsic::aarch64_stg) {
4944 return Builder.CreateCall(
CGM.getIntrinsic(MTEIntrinsicID),
4945 {TagAddress, TagAddress});
4947 if (MTEIntrinsicID == Intrinsic::aarch64_subp) {
4951 CGM.getIntrinsic(MTEIntrinsicID), {PointerA, PointerB});
4955 if (BuiltinID == clang::AArch64::BI__builtin_arm_rsr ||
4956 BuiltinID == clang::AArch64::BI__builtin_arm_rsr64 ||
4957 BuiltinID == clang::AArch64::BI__builtin_arm_rsr128 ||
4958 BuiltinID == clang::AArch64::BI__builtin_arm_rsrp ||
4959 BuiltinID == clang::AArch64::BI__builtin_arm_wsr ||
4960 BuiltinID == clang::AArch64::BI__builtin_arm_wsr64 ||
4961 BuiltinID == clang::AArch64::BI__builtin_arm_wsr128 ||
4962 BuiltinID == clang::AArch64::BI__builtin_arm_wsrp) {
4965 if (BuiltinID == clang::AArch64::BI__builtin_arm_rsr ||
4966 BuiltinID == clang::AArch64::BI__builtin_arm_rsr64 ||
4967 BuiltinID == clang::AArch64::BI__builtin_arm_rsr128 ||
4968 BuiltinID == clang::AArch64::BI__builtin_arm_rsrp)
4971 bool IsPointerBuiltin = BuiltinID == clang::AArch64::BI__builtin_arm_rsrp ||
4972 BuiltinID == clang::AArch64::BI__builtin_arm_wsrp;
4974 bool Is32Bit = BuiltinID == clang::AArch64::BI__builtin_arm_rsr ||
4975 BuiltinID == clang::AArch64::BI__builtin_arm_wsr;
4977 bool Is128Bit = BuiltinID == clang::AArch64::BI__builtin_arm_rsr128 ||
4978 BuiltinID == clang::AArch64::BI__builtin_arm_wsr128;
4980 llvm::Type *ValueType;
4984 }
else if (Is128Bit) {
4985 llvm::Type *Int128Ty =
4986 llvm::IntegerType::getInt128Ty(
CGM.getLLVMContext());
4987 ValueType = Int128Ty;
4989 }
else if (IsPointerBuiltin) {
4999 if (BuiltinID == clang::AArch64::BI_ReadStatusReg ||
5000 BuiltinID == clang::AArch64::BI_WriteStatusReg) {
5001 LLVMContext &Context =
CGM.getLLVMContext();
5006 std::string SysRegStr;
5007 llvm::raw_string_ostream(SysRegStr)
5008 << (0b10 | SysReg >> 14) <<
":" << ((SysReg >> 11) & 7) <<
":"
5009 << ((SysReg >> 7) & 15) <<
":" << ((SysReg >> 3) & 15) <<
":"
5012 llvm::Metadata *Ops[] = { llvm::MDString::get(Context, SysRegStr) };
5013 llvm::MDNode *RegName = llvm::MDNode::get(Context, Ops);
5014 llvm::Value *Metadata = llvm::MetadataAsValue::get(Context, RegName);
5019 if (BuiltinID == clang::AArch64::BI_ReadStatusReg) {
5020 llvm::Function *F =
CGM.getIntrinsic(Intrinsic::read_register, Types);
5022 return Builder.CreateCall(F, Metadata);
5025 llvm::Function *F =
CGM.getIntrinsic(Intrinsic::write_register, Types);
5027 llvm::Value *
Result =
Builder.CreateCall(F, {Metadata, ArgValue});
5032 if (BuiltinID == clang::AArch64::BI__sys) {
5035 const unsigned Op1 = SysReg >> 11;
5036 const unsigned CRn = (SysReg >> 7) & 0xf;
5037 const unsigned CRm = (SysReg >> 3) & 0xf;
5038 const unsigned Op2 = SysReg & 0x7;
5040 Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::aarch64_sys),
5041 {Builder.getInt32(Op1), Builder.getInt32(CRn),
5042 Builder.getInt32(CRm), Builder.getInt32(Op2),
5043 EmitScalarExpr(E->getArg(1))});
5047 return ConstantInt::get(
Builder.getInt32Ty(), 0);
5050 if (BuiltinID == clang::AArch64::BI_AddressOfReturnAddress) {
5056 if (BuiltinID == clang::AArch64::BI__builtin_sponentry) {
5061 if (BuiltinID == clang::AArch64::BI__mulh ||
5062 BuiltinID == clang::AArch64::BI__umulh) {
5064 llvm::Type *Int128Ty = llvm::IntegerType::get(
getLLVMContext(), 128);
5066 bool IsSigned = BuiltinID == clang::AArch64::BI__mulh;
5072 Value *MulResult, *HigherBits;
5074 MulResult =
Builder.CreateNSWMul(LHS, RHS);
5075 HigherBits =
Builder.CreateAShr(MulResult, 64);
5077 MulResult =
Builder.CreateNUWMul(LHS, RHS);
5078 HigherBits =
Builder.CreateLShr(MulResult, 64);
5080 HigherBits =
Builder.CreateIntCast(HigherBits, ResType, IsSigned);
5085 if (BuiltinID == AArch64::BI__writex18byte ||
5086 BuiltinID == AArch64::BI__writex18word ||
5087 BuiltinID == AArch64::BI__writex18dword ||
5088 BuiltinID == AArch64::BI__writex18qword) {
5104 if (BuiltinID == AArch64::BI__readx18byte ||
5105 BuiltinID == AArch64::BI__readx18word ||
5106 BuiltinID == AArch64::BI__readx18dword ||
5107 BuiltinID == AArch64::BI__readx18qword) {
5122 if (BuiltinID == AArch64::BI__addx18byte ||
5123 BuiltinID == AArch64::BI__addx18word ||
5124 BuiltinID == AArch64::BI__addx18dword ||
5125 BuiltinID == AArch64::BI__addx18qword ||
5126 BuiltinID == AArch64::BI__incx18byte ||
5127 BuiltinID == AArch64::BI__incx18word ||
5128 BuiltinID == AArch64::BI__incx18dword ||
5129 BuiltinID == AArch64::BI__incx18qword) {
5132 switch (BuiltinID) {
5133 case AArch64::BI__incx18byte:
5137 case AArch64::BI__incx18word:
5141 case AArch64::BI__incx18dword:
5145 case AArch64::BI__incx18qword:
5151 isIncrement =
false;
5176 if (BuiltinID == AArch64::BI_CopyDoubleFromInt64 ||
5177 BuiltinID == AArch64::BI_CopyFloatFromInt32 ||
5178 BuiltinID == AArch64::BI_CopyInt32FromFloat ||
5179 BuiltinID == AArch64::BI_CopyInt64FromDouble) {
5182 return Builder.CreateBitCast(Arg, RetTy);
5185 if (BuiltinID == AArch64::BI_CountLeadingOnes ||
5186 BuiltinID == AArch64::BI_CountLeadingOnes64 ||
5187 BuiltinID == AArch64::BI_CountLeadingZeros ||
5188 BuiltinID == AArch64::BI_CountLeadingZeros64) {
5192 if (BuiltinID == AArch64::BI_CountLeadingOnes ||
5193 BuiltinID == AArch64::BI_CountLeadingOnes64)
5194 Arg =
Builder.CreateXor(Arg, Constant::getAllOnesValue(
ArgType));
5199 if (BuiltinID == AArch64::BI_CountLeadingOnes64 ||
5200 BuiltinID == AArch64::BI_CountLeadingZeros64)
5205 if (BuiltinID == AArch64::BI_CountLeadingSigns ||
5206 BuiltinID == AArch64::BI_CountLeadingSigns64) {
5209 Function *F = (BuiltinID == AArch64::BI_CountLeadingSigns)
5210 ?
CGM.getIntrinsic(Intrinsic::aarch64_cls)
5211 :
CGM.getIntrinsic(Intrinsic::aarch64_cls64);
5214 if (BuiltinID == AArch64::BI_CountLeadingSigns64)
5219 if (BuiltinID == AArch64::BI_CountOneBits ||
5220 BuiltinID == AArch64::BI_CountOneBits64) {
5226 if (BuiltinID == AArch64::BI_CountOneBits64)
5231 if (BuiltinID == AArch64::BI__prefetch) {
5240 if (BuiltinID == AArch64::BI__hlt) {
5241 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_hlt);
5246 return ConstantInt::get(
Builder.getInt32Ty(), 0);
5249 if (BuiltinID == NEON::BI__builtin_neon_vcvth_bf16_f32)
5257 if (std::optional<MSVCIntrin> MsvcIntId =
5263 return P.first == BuiltinID;
5266 BuiltinID = It->second;
5272 bool IsSISD = (
Builtin !=
nullptr);
5276 unsigned ICEArguments = 0;
5287 unsigned NumArgs = E->
getNumArgs() - (HasExtraArg ? 1 : 0);
5288 for (
unsigned i = 0, e = NumArgs; i != e; i++) {
5290 switch (BuiltinID) {
5291 case NEON::BI__builtin_neon_vld1_v:
5292 case NEON::BI__builtin_neon_vld1q_v:
5293 case NEON::BI__builtin_neon_vld1_dup_v:
5294 case NEON::BI__builtin_neon_vld1q_dup_v:
5295 case NEON::BI__builtin_neon_vld1_lane_v:
5296 case NEON::BI__builtin_neon_vld1q_lane_v:
5297 case NEON::BI__builtin_neon_vst1_v:
5298 case NEON::BI__builtin_neon_vst1q_v:
5299 case NEON::BI__builtin_neon_vst1_lane_v:
5300 case NEON::BI__builtin_neon_vst1q_lane_v:
5301 case NEON::BI__builtin_neon_vldap1_lane_s64:
5302 case NEON::BI__builtin_neon_vldap1q_lane_s64:
5303 case NEON::BI__builtin_neon_vstl1_lane_s64:
5304 case NEON::BI__builtin_neon_vstl1q_lane_s64:
5317 assert(
Result &&
"SISD intrinsic should have been handled");
5323 if (std::optional<llvm::APSInt>
Result =
5328 bool usgn =
Type.isUnsigned();
5329 bool quad =
Type.isQuad();
5347 switch (BuiltinID) {
5349 case NEON::BI__builtin_neon_vabsh_f16:
5351 case NEON::BI__builtin_neon_vaddq_p128: {
5353 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
5354 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
5355 Ops[0] =
Builder.CreateXor(Ops[0], Ops[1]);
5356 llvm::Type *Int128Ty = llvm::Type::getIntNTy(
getLLVMContext(), 128);
5357 return Builder.CreateBitCast(Ops[0], Int128Ty);
5359 case NEON::BI__builtin_neon_vldrq_p128: {
5360 llvm::Type *Int128Ty = llvm::Type::getIntNTy(
getLLVMContext(), 128);
5361 return Builder.CreateAlignedLoad(Int128Ty, Ops[0],
5364 case NEON::BI__builtin_neon_vstrq_p128: {
5365 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
5367 case NEON::BI__builtin_neon_vcvts_f32_u32:
5368 case NEON::BI__builtin_neon_vcvtd_f64_u64:
5371 case NEON::BI__builtin_neon_vcvts_f32_s32:
5372 case NEON::BI__builtin_neon_vcvtd_f64_s64: {
5373 bool Is64 = Ops[0]->getType()->getPrimitiveSizeInBits() == 64;
5376 Ops[0] =
Builder.CreateBitCast(Ops[0], InTy);
5378 return Builder.CreateUIToFP(Ops[0], FTy);
5379 return Builder.CreateSIToFP(Ops[0], FTy);
5381 case NEON::BI__builtin_neon_vcvth_f16_u16:
5382 case NEON::BI__builtin_neon_vcvth_f16_u32:
5383 case NEON::BI__builtin_neon_vcvth_f16_u64:
5386 case NEON::BI__builtin_neon_vcvth_f16_s16:
5387 case NEON::BI__builtin_neon_vcvth_f16_s32:
5388 case NEON::BI__builtin_neon_vcvth_f16_s64: {
5389 llvm::Type *FTy =
HalfTy;
5391 if (Ops[0]->
getType()->getPrimitiveSizeInBits() == 64)
5393 else if (Ops[0]->
getType()->getPrimitiveSizeInBits() == 32)
5397 Ops[0] =
Builder.CreateBitCast(Ops[0], InTy);
5399 return Builder.CreateUIToFP(Ops[0], FTy);
5400 return Builder.CreateSIToFP(Ops[0], FTy);
5402 case NEON::BI__builtin_neon_vcvtah_u16_f16:
5403 case NEON::BI__builtin_neon_vcvtmh_u16_f16:
5404 case NEON::BI__builtin_neon_vcvtnh_u16_f16:
5405 case NEON::BI__builtin_neon_vcvtph_u16_f16:
5406 case NEON::BI__builtin_neon_vcvtah_s16_f16:
5407 case NEON::BI__builtin_neon_vcvtmh_s16_f16:
5408 case NEON::BI__builtin_neon_vcvtnh_s16_f16:
5409 case NEON::BI__builtin_neon_vcvtph_s16_f16: {
5411 llvm::Type* FTy =
HalfTy;
5412 llvm::Type *Tys[2] = {InTy, FTy};
5413 switch (BuiltinID) {
5414 default: llvm_unreachable(
"missing builtin ID in switch!");
5415 case NEON::BI__builtin_neon_vcvtah_u16_f16:
5416 Int = Intrinsic::aarch64_neon_fcvtau;
break;
5417 case NEON::BI__builtin_neon_vcvtmh_u16_f16:
5418 Int = Intrinsic::aarch64_neon_fcvtmu;
break;
5419 case NEON::BI__builtin_neon_vcvtnh_u16_f16:
5420 Int = Intrinsic::aarch64_neon_fcvtnu;
break;
5421 case NEON::BI__builtin_neon_vcvtph_u16_f16:
5422 Int = Intrinsic::aarch64_neon_fcvtpu;
break;
5423 case NEON::BI__builtin_neon_vcvtah_s16_f16:
5424 Int = Intrinsic::aarch64_neon_fcvtas;
break;
5425 case NEON::BI__builtin_neon_vcvtmh_s16_f16:
5426 Int = Intrinsic::aarch64_neon_fcvtms;
break;
5427 case NEON::BI__builtin_neon_vcvtnh_s16_f16:
5428 Int = Intrinsic::aarch64_neon_fcvtns;
break;
5429 case NEON::BI__builtin_neon_vcvtph_s16_f16:
5430 Int = Intrinsic::aarch64_neon_fcvtps;
break;
5434 case NEON::BI__builtin_neon_vcaleh_f16:
5435 case NEON::BI__builtin_neon_vcalth_f16:
5436 case NEON::BI__builtin_neon_vcageh_f16:
5437 case NEON::BI__builtin_neon_vcagth_f16: {
5439 llvm::Type* FTy =
HalfTy;
5440 llvm::Type *Tys[2] = {InTy, FTy};
5441 switch (BuiltinID) {
5442 default: llvm_unreachable(
"missing builtin ID in switch!");
5443 case NEON::BI__builtin_neon_vcageh_f16:
5444 Int = Intrinsic::aarch64_neon_facge;
break;
5445 case NEON::BI__builtin_neon_vcagth_f16:
5446 Int = Intrinsic::aarch64_neon_facgt;
break;
5447 case NEON::BI__builtin_neon_vcaleh_f16:
5448 Int = Intrinsic::aarch64_neon_facge; std::swap(Ops[0], Ops[1]);
break;
5449 case NEON::BI__builtin_neon_vcalth_f16:
5450 Int = Intrinsic::aarch64_neon_facgt; std::swap(Ops[0], Ops[1]);
break;
5455 case NEON::BI__builtin_neon_vcvth_n_s16_f16:
5456 case NEON::BI__builtin_neon_vcvth_n_u16_f16: {
5458 llvm::Type* FTy =
HalfTy;
5459 llvm::Type *Tys[2] = {InTy, FTy};
5460 switch (BuiltinID) {
5461 default: llvm_unreachable(
"missing builtin ID in switch!");
5462 case NEON::BI__builtin_neon_vcvth_n_s16_f16:
5463 Int = Intrinsic::aarch64_neon_vcvtfp2fxs;
break;
5464 case NEON::BI__builtin_neon_vcvth_n_u16_f16:
5465 Int = Intrinsic::aarch64_neon_vcvtfp2fxu;
break;
5470 case NEON::BI__builtin_neon_vcvth_n_f16_s16:
5471 case NEON::BI__builtin_neon_vcvth_n_f16_u16: {
5472 llvm::Type* FTy =
HalfTy;
5474 llvm::Type *Tys[2] = {FTy, InTy};
5475 switch (BuiltinID) {
5476 default: llvm_unreachable(
"missing builtin ID in switch!");
5477 case NEON::BI__builtin_neon_vcvth_n_f16_s16:
5478 Int = Intrinsic::aarch64_neon_vcvtfxs2fp;
5479 Ops[0] =
Builder.CreateSExt(Ops[0], InTy,
"sext");
5481 case NEON::BI__builtin_neon_vcvth_n_f16_u16:
5482 Int = Intrinsic::aarch64_neon_vcvtfxu2fp;
5483 Ops[0] =
Builder.CreateZExt(Ops[0], InTy);
5488 case NEON::BI__builtin_neon_vpaddd_s64: {
5491 auto *Ty = llvm::FixedVectorType::get(
Int64Ty, 2);
5493 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty,
"v2i64");
5494 llvm::Value *Idx0 = llvm::ConstantInt::get(
SizeTy, 0);
5495 llvm::Value *Idx1 = llvm::ConstantInt::get(
SizeTy, 1);
5496 Value *Op0 =
Builder.CreateExtractElement(Ops[0], Idx0,
"lane0");
5497 Value *Op1 =
Builder.CreateExtractElement(Ops[0], Idx1,
"lane1");
5499 return Builder.CreateAdd(Op0, Op1,
"vpaddd");
5501 case NEON::BI__builtin_neon_vpaddd_f64: {
5502 auto *Ty = llvm::FixedVectorType::get(
DoubleTy, 2);
5504 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty,
"v2f64");
5505 llvm::Value *Idx0 = llvm::ConstantInt::get(
SizeTy, 0);
5506 llvm::Value *Idx1 = llvm::ConstantInt::get(
SizeTy, 1);
5507 Value *Op0 =
Builder.CreateExtractElement(Ops[0], Idx0,
"lane0");
5508 Value *Op1 =
Builder.CreateExtractElement(Ops[0], Idx1,
"lane1");
5510 return Builder.CreateFAdd(Op0, Op1,
"vpaddd");
5512 case NEON::BI__builtin_neon_vpadds_f32: {
5513 auto *Ty = llvm::FixedVectorType::get(
FloatTy, 2);
5515 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty,
"v2f32");
5516 llvm::Value *Idx0 = llvm::ConstantInt::get(
SizeTy, 0);
5517 llvm::Value *Idx1 = llvm::ConstantInt::get(
SizeTy, 1);
5518 Value *Op0 =
Builder.CreateExtractElement(Ops[0], Idx0,
"lane0");
5519 Value *Op1 =
Builder.CreateExtractElement(Ops[0], Idx1,
"lane1");
5521 return Builder.CreateFAdd(Op0, Op1,
"vpaddd");
5523 case NEON::BI__builtin_neon_vceqzd_s64:
5526 ICmpInst::ICMP_EQ,
"vceqz");
5527 case NEON::BI__builtin_neon_vceqzd_f64:
5528 case NEON::BI__builtin_neon_vceqzs_f32:
5529 case NEON::BI__builtin_neon_vceqzh_f16:
5532 ICmpInst::FCMP_OEQ,
"vceqz");
5533 case NEON::BI__builtin_neon_vcgezd_s64:
5536 ICmpInst::ICMP_SGE,
"vcgez");
5537 case NEON::BI__builtin_neon_vcgezd_f64:
5538 case NEON::BI__builtin_neon_vcgezs_f32:
5539 case NEON::BI__builtin_neon_vcgezh_f16:
5542 ICmpInst::FCMP_OGE,
"vcgez");
5543 case NEON::BI__builtin_neon_vclezd_s64:
5546 ICmpInst::ICMP_SLE,
"vclez");
5547 case NEON::BI__builtin_neon_vclezd_f64:
5548 case NEON::BI__builtin_neon_vclezs_f32:
5549 case NEON::BI__builtin_neon_vclezh_f16:
5552 ICmpInst::FCMP_OLE,
"vclez");
5553 case NEON::BI__builtin_neon_vcgtzd_s64:
5556 ICmpInst::ICMP_SGT,
"vcgtz");
5557 case NEON::BI__builtin_neon_vcgtzd_f64:
5558 case NEON::BI__builtin_neon_vcgtzs_f32:
5559 case NEON::BI__builtin_neon_vcgtzh_f16:
5562 ICmpInst::FCMP_OGT,
"vcgtz");
5563 case NEON::BI__builtin_neon_vcltzd_s64:
5566 ICmpInst::ICMP_SLT,
"vcltz");
5568 case NEON::BI__builtin_neon_vcltzd_f64:
5569 case NEON::BI__builtin_neon_vcltzs_f32:
5570 case NEON::BI__builtin_neon_vcltzh_f16:
5573 ICmpInst::FCMP_OLT,
"vcltz");
5575 case NEON::BI__builtin_neon_vceqzd_u64: {
5578 ICmpInst::ICMP_EQ,
"vceqzd");
5580 case NEON::BI__builtin_neon_vceqd_f64:
5581 case NEON::BI__builtin_neon_vcled_f64:
5582 case NEON::BI__builtin_neon_vcltd_f64:
5583 case NEON::BI__builtin_neon_vcged_f64:
5584 case NEON::BI__builtin_neon_vcgtd_f64: {
5585 llvm::CmpInst::Predicate P;
5586 switch (BuiltinID) {
5587 default: llvm_unreachable(
"missing builtin ID in switch!");
5588 case NEON::BI__builtin_neon_vceqd_f64: P = llvm::FCmpInst::FCMP_OEQ;
break;
5589 case NEON::BI__builtin_neon_vcled_f64: P = llvm::FCmpInst::FCMP_OLE;
break;
5590 case NEON::BI__builtin_neon_vcltd_f64: P = llvm::FCmpInst::FCMP_OLT;
break;
5591 case NEON::BI__builtin_neon_vcged_f64: P = llvm::FCmpInst::FCMP_OGE;
break;
5592 case NEON::BI__builtin_neon_vcgtd_f64: P = llvm::FCmpInst::FCMP_OGT;
break;
5596 if (P == llvm::FCmpInst::FCMP_OEQ)
5597 Ops[0] =
Builder.CreateFCmp(P, Ops[0], Ops[1]);
5599 Ops[0] =
Builder.CreateFCmpS(P, Ops[0], Ops[1]);
5602 case NEON::BI__builtin_neon_vceqs_f32:
5603 case NEON::BI__builtin_neon_vcles_f32:
5604 case NEON::BI__builtin_neon_vclts_f32:
5605 case NEON::BI__builtin_neon_vcges_f32:
5606 case NEON::BI__builtin_neon_vcgts_f32: {
5607 llvm::CmpInst::Predicate P;
5608 switch (BuiltinID) {
5609 default: llvm_unreachable(
"missing builtin ID in switch!");
5610 case NEON::BI__builtin_neon_vceqs_f32: P = llvm::FCmpInst::FCMP_OEQ;
break;
5611 case NEON::BI__builtin_neon_vcles_f32: P = llvm::FCmpInst::FCMP_OLE;
break;
5612 case NEON::BI__builtin_neon_vclts_f32: P = llvm::FCmpInst::FCMP_OLT;
break;
5613 case NEON::BI__builtin_neon_vcges_f32: P = llvm::FCmpInst::FCMP_OGE;
break;
5614 case NEON::BI__builtin_neon_vcgts_f32: P = llvm::FCmpInst::FCMP_OGT;
break;
5618 if (P == llvm::FCmpInst::FCMP_OEQ)
5619 Ops[0] =
Builder.CreateFCmp(P, Ops[0], Ops[1]);
5621 Ops[0] =
Builder.CreateFCmpS(P, Ops[0], Ops[1]);
5624 case NEON::BI__builtin_neon_vceqh_f16:
5625 case NEON::BI__builtin_neon_vcleh_f16:
5626 case NEON::BI__builtin_neon_vclth_f16:
5627 case NEON::BI__builtin_neon_vcgeh_f16:
5628 case NEON::BI__builtin_neon_vcgth_f16: {
5629 llvm::CmpInst::Predicate P;
5630 switch (BuiltinID) {
5631 default: llvm_unreachable(
"missing builtin ID in switch!");
5632 case NEON::BI__builtin_neon_vceqh_f16: P = llvm::FCmpInst::FCMP_OEQ;
break;
5633 case NEON::BI__builtin_neon_vcleh_f16: P = llvm::FCmpInst::FCMP_OLE;
break;
5634 case NEON::BI__builtin_neon_vclth_f16: P = llvm::FCmpInst::FCMP_OLT;
break;
5635 case NEON::BI__builtin_neon_vcgeh_f16: P = llvm::FCmpInst::FCMP_OGE;
break;
5636 case NEON::BI__builtin_neon_vcgth_f16: P = llvm::FCmpInst::FCMP_OGT;
break;
5640 if (P == llvm::FCmpInst::FCMP_OEQ)
5641 Ops[0] =
Builder.CreateFCmp(P, Ops[0], Ops[1]);
5643 Ops[0] =
Builder.CreateFCmpS(P, Ops[0], Ops[1]);
5646 case NEON::BI__builtin_neon_vceqd_s64:
5647 case NEON::BI__builtin_neon_vceqd_u64:
5648 case NEON::BI__builtin_neon_vcgtd_s64:
5649 case NEON::BI__builtin_neon_vcgtd_u64:
5650 case NEON::BI__builtin_neon_vcltd_s64:
5651 case NEON::BI__builtin_neon_vcltd_u64:
5652 case NEON::BI__builtin_neon_vcged_u64:
5653 case NEON::BI__builtin_neon_vcged_s64:
5654 case NEON::BI__builtin_neon_vcled_u64:
5655 case NEON::BI__builtin_neon_vcled_s64: {
5656 llvm::CmpInst::Predicate P;
5657 switch (BuiltinID) {
5658 default: llvm_unreachable(
"missing builtin ID in switch!");
5659 case NEON::BI__builtin_neon_vceqd_s64:
5660 case NEON::BI__builtin_neon_vceqd_u64:P = llvm::ICmpInst::ICMP_EQ;
break;
5661 case NEON::BI__builtin_neon_vcgtd_s64:P = llvm::ICmpInst::ICMP_SGT;
break;
5662 case NEON::BI__builtin_neon_vcgtd_u64:P = llvm::ICmpInst::ICMP_UGT;
break;
5663 case NEON::BI__builtin_neon_vcltd_s64:P = llvm::ICmpInst::ICMP_SLT;
break;
5664 case NEON::BI__builtin_neon_vcltd_u64:P = llvm::ICmpInst::ICMP_ULT;
break;
5665 case NEON::BI__builtin_neon_vcged_u64:P = llvm::ICmpInst::ICMP_UGE;
break;
5666 case NEON::BI__builtin_neon_vcged_s64:P = llvm::ICmpInst::ICMP_SGE;
break;
5667 case NEON::BI__builtin_neon_vcled_u64:P = llvm::ICmpInst::ICMP_ULE;
break;
5668 case NEON::BI__builtin_neon_vcled_s64:P = llvm::ICmpInst::ICMP_SLE;
break;
5672 Ops[0] =
Builder.CreateICmp(P, Ops[0], Ops[1]);
5675 case NEON::BI__builtin_neon_vnegd_s64:
5676 return Builder.CreateNeg(Ops[0],
"vnegd");
5677 case NEON::BI__builtin_neon_vnegh_f16:
5678 return Builder.CreateFNeg(Ops[0],
"vnegh");
5679 case NEON::BI__builtin_neon_vtstd_s64:
5680 case NEON::BI__builtin_neon_vtstd_u64: {
5683 Ops[0] =
Builder.CreateAnd(Ops[0], Ops[1]);
5684 Ops[0] =
Builder.CreateICmp(ICmpInst::ICMP_NE, Ops[0],
5685 llvm::Constant::getNullValue(
Int64Ty));
5688 case NEON::BI__builtin_neon_vset_lane_i8:
5689 case NEON::BI__builtin_neon_vset_lane_i16:
5690 case NEON::BI__builtin_neon_vset_lane_i32:
5691 case NEON::BI__builtin_neon_vset_lane_i64:
5692 case NEON::BI__builtin_neon_vset_lane_bf16:
5693 case NEON::BI__builtin_neon_vset_lane_f32:
5694 case NEON::BI__builtin_neon_vsetq_lane_i8:
5695 case NEON::BI__builtin_neon_vsetq_lane_i16:
5696 case NEON::BI__builtin_neon_vsetq_lane_i32:
5697 case NEON::BI__builtin_neon_vsetq_lane_i64:
5698 case NEON::BI__builtin_neon_vsetq_lane_bf16:
5699 case NEON::BI__builtin_neon_vsetq_lane_f32:
5700 return Builder.CreateInsertElement(Ops[1], Ops[0], Ops[2],
"vset_lane");
5701 case NEON::BI__builtin_neon_vset_lane_f64:
5704 Builder.CreateBitCast(Ops[1], llvm::FixedVectorType::get(
DoubleTy, 1));
5705 return Builder.CreateInsertElement(Ops[1], Ops[0], Ops[2],
"vset_lane");
5706 case NEON::BI__builtin_neon_vset_lane_mf8:
5707 case NEON::BI__builtin_neon_vsetq_lane_mf8:
5711 return Builder.CreateInsertElement(Ops[1], Ops[0], Ops[2],
"vset_lane");
5712 case NEON::BI__builtin_neon_vsetq_lane_f64:
5715 Builder.CreateBitCast(Ops[1], llvm::FixedVectorType::get(
DoubleTy, 2));
5716 return Builder.CreateInsertElement(Ops[1], Ops[0], Ops[2],
"vset_lane");
5718 case NEON::BI__builtin_neon_vget_lane_i8:
5719 case NEON::BI__builtin_neon_vdupb_lane_i8:
5721 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int8Ty, 8));
5722 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5723 case NEON::BI__builtin_neon_vgetq_lane_i8:
5724 case NEON::BI__builtin_neon_vdupb_laneq_i8:
5726 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int8Ty, 16));
5727 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5728 case NEON::BI__builtin_neon_vget_lane_mf8:
5729 case NEON::BI__builtin_neon_vdupb_lane_mf8:
5730 case NEON::BI__builtin_neon_vgetq_lane_mf8:
5731 case NEON::BI__builtin_neon_vdupb_laneq_mf8:
5732 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5733 case NEON::BI__builtin_neon_vget_lane_i16:
5734 case NEON::BI__builtin_neon_vduph_lane_i16:
5736 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int16Ty, 4));
5737 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5738 case NEON::BI__builtin_neon_vgetq_lane_i16:
5739 case NEON::BI__builtin_neon_vduph_laneq_i16:
5741 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int16Ty, 8));
5742 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5743 case NEON::BI__builtin_neon_vget_lane_i32:
5744 case NEON::BI__builtin_neon_vdups_lane_i32:
5746 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int32Ty, 2));
5747 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5748 case NEON::BI__builtin_neon_vdups_lane_f32:
5750 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
FloatTy, 2));
5751 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vdups_lane");
5752 case NEON::BI__builtin_neon_vgetq_lane_i32:
5753 case NEON::BI__builtin_neon_vdups_laneq_i32:
5755 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int32Ty, 4));
5756 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5757 case NEON::BI__builtin_neon_vget_lane_i64:
5758 case NEON::BI__builtin_neon_vdupd_lane_i64:
5760 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int64Ty, 1));
5761 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5762 case NEON::BI__builtin_neon_vdupd_lane_f64:
5764 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
DoubleTy, 1));
5765 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vdupd_lane");
5766 case NEON::BI__builtin_neon_vgetq_lane_i64:
5767 case NEON::BI__builtin_neon_vdupd_laneq_i64:
5769 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
Int64Ty, 2));
5770 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5771 case NEON::BI__builtin_neon_vget_lane_f32:
5773 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
FloatTy, 2));
5774 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5775 case NEON::BI__builtin_neon_vget_lane_f64:
5777 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
DoubleTy, 1));
5778 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5779 case NEON::BI__builtin_neon_vgetq_lane_f32:
5780 case NEON::BI__builtin_neon_vdups_laneq_f32:
5782 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
FloatTy, 4));
5783 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5784 case NEON::BI__builtin_neon_vgetq_lane_f64:
5785 case NEON::BI__builtin_neon_vdupd_laneq_f64:
5787 Builder.CreateBitCast(Ops[0], llvm::FixedVectorType::get(
DoubleTy, 2));
5788 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5789 case NEON::BI__builtin_neon_vaddh_f16:
5790 return Builder.CreateFAdd(Ops[0], Ops[1],
"vaddh");
5791 case NEON::BI__builtin_neon_vsubh_f16:
5792 return Builder.CreateFSub(Ops[0], Ops[1],
"vsubh");
5793 case NEON::BI__builtin_neon_vmulh_f16:
5794 return Builder.CreateFMul(Ops[0], Ops[1],
"vmulh");
5795 case NEON::BI__builtin_neon_vdivh_f16:
5796 return Builder.CreateFDiv(Ops[0], Ops[1],
"vdivh");
5797 case NEON::BI__builtin_neon_vfmah_f16:
5800 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma,
HalfTy,
5801 {Ops[1], Ops[2], Ops[0]});
5802 case NEON::BI__builtin_neon_vfmsh_f16: {
5807 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma,
HalfTy,
5808 {Neg, Ops[2], Ops[0]});
5810 case NEON::BI__builtin_neon_vaddd_s64:
5811 case NEON::BI__builtin_neon_vaddd_u64:
5812 return Builder.CreateAdd(Ops[0], Ops[1],
"vaddd");
5813 case NEON::BI__builtin_neon_vsubd_s64:
5814 case NEON::BI__builtin_neon_vsubd_u64:
5815 return Builder.CreateSub(Ops[0], Ops[1],
"vsubd");
5816 case NEON::BI__builtin_neon_vqdmlalh_s16:
5817 case NEON::BI__builtin_neon_vqdmlslh_s16: {
5821 auto *VTy = llvm::FixedVectorType::get(
Int32Ty, 4);
5822 Ops[1] =
EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_sqdmull, VTy),
5823 ProductOps,
"vqdmlXl");
5824 Constant *CI = ConstantInt::get(
SizeTy, 0);
5825 Ops[1] =
Builder.CreateExtractElement(Ops[1], CI,
"lane0");
5827 unsigned AccumInt = BuiltinID == NEON::BI__builtin_neon_vqdmlalh_s16
5828 ? Intrinsic::aarch64_neon_sqadd
5829 : Intrinsic::aarch64_neon_sqsub;
5834 case NEON::BI__builtin_neon_vqshlud_n_s64: {
5839 case NEON::BI__builtin_neon_vqshld_n_u64:
5840 case NEON::BI__builtin_neon_vqshld_n_s64: {
5841 Int = BuiltinID == NEON::BI__builtin_neon_vqshld_n_u64
5842 ? Intrinsic::aarch64_neon_uqshl
5843 : Intrinsic::aarch64_neon_sqshl;
5847 case NEON::BI__builtin_neon_vrshrd_n_u64:
5848 case NEON::BI__builtin_neon_vrshrd_n_s64: {
5849 Int = BuiltinID == NEON::BI__builtin_neon_vrshrd_n_u64
5850 ? Intrinsic::aarch64_neon_urshl
5851 : Intrinsic::aarch64_neon_srshl;
5853 Ops[1] = ConstantInt::get(
Int64Ty, -SV);
5856 case NEON::BI__builtin_neon_vrsrad_n_u64:
5857 case NEON::BI__builtin_neon_vrsrad_n_s64: {
5858 Int = BuiltinID == NEON::BI__builtin_neon_vrsrad_n_u64
5859 ? Intrinsic::aarch64_neon_urshl
5860 : Intrinsic::aarch64_neon_srshl;
5862 Ops[2] =
Builder.CreateNeg(Ops[2]);
5864 {Ops[1], Builder.CreateSExt(Ops[2], Int64Ty)});
5867 case NEON::BI__builtin_neon_vshld_n_s64:
5868 case NEON::BI__builtin_neon_vshld_n_u64: {
5871 Ops[0], ConstantInt::get(
Int64Ty, Amt->getZExtValue()),
"shld_n");
5873 case NEON::BI__builtin_neon_vshrd_n_s64: {
5876 Ops[0], ConstantInt::get(
Int64Ty, std::min(
static_cast<uint64_t
>(63),
5877 Amt->getZExtValue())),
5880 case NEON::BI__builtin_neon_vshrd_n_u64: {
5882 uint64_t ShiftAmt = Amt->getZExtValue();
5885 return ConstantInt::get(
Int64Ty, 0);
5886 return Builder.CreateLShr(Ops[0], ConstantInt::get(
Int64Ty, ShiftAmt),
5889 case NEON::BI__builtin_neon_vsrad_n_s64: {
5892 Ops[1], ConstantInt::get(
Int64Ty, std::min(
static_cast<uint64_t
>(63),
5893 Amt->getZExtValue())),
5895 return Builder.CreateAdd(Ops[0], Ops[1]);
5897 case NEON::BI__builtin_neon_vsrad_n_u64: {
5899 uint64_t ShiftAmt = Amt->getZExtValue();
5904 Ops[1] =
Builder.CreateLShr(Ops[1], ConstantInt::get(
Int64Ty, ShiftAmt),
5906 return Builder.CreateAdd(Ops[0], Ops[1]);
5908 case NEON::BI__builtin_neon_vqdmlalh_lane_s16:
5909 case NEON::BI__builtin_neon_vqdmlalh_laneq_s16:
5910 case NEON::BI__builtin_neon_vqdmlslh_lane_s16:
5911 case NEON::BI__builtin_neon_vqdmlslh_laneq_s16: {
5912 Ops[2] =
Builder.CreateExtractElement(Ops[2], Ops[3],
"lane");
5916 auto *VTy = llvm::FixedVectorType::get(
Int32Ty, 4);
5917 Ops[1] =
EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_sqdmull, VTy),
5918 ProductOps,
"vqdmlXl");
5919 Constant *CI = ConstantInt::get(
SizeTy, 0);
5920 Ops[1] =
Builder.CreateExtractElement(Ops[1], CI,
"lane0");
5925 unsigned AccInt = (BuiltinID == NEON::BI__builtin_neon_vqdmlalh_lane_s16 ||
5926 BuiltinID == NEON::BI__builtin_neon_vqdmlalh_laneq_s16)
5927 ? Intrinsic::aarch64_neon_sqadd
5928 : Intrinsic::aarch64_neon_sqsub;
5931 case NEON::BI__builtin_neon_vqdmlals_s32:
5932 case NEON::BI__builtin_neon_vqdmlsls_s32: {
5934 ProductOps.push_back(Ops[1]);
5935 ProductOps.push_back(Ops[2]);
5937 EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_sqdmulls_scalar),
5938 ProductOps,
"vqdmlXl");
5940 unsigned AccumInt = BuiltinID == NEON::BI__builtin_neon_vqdmlals_s32
5941 ? Intrinsic::aarch64_neon_sqadd
5942 : Intrinsic::aarch64_neon_sqsub;
5947 case NEON::BI__builtin_neon_vqdmlals_lane_s32:
5948 case NEON::BI__builtin_neon_vqdmlals_laneq_s32:
5949 case NEON::BI__builtin_neon_vqdmlsls_lane_s32:
5950 case NEON::BI__builtin_neon_vqdmlsls_laneq_s32: {
5951 Ops[2] =
Builder.CreateExtractElement(Ops[2], Ops[3],
"lane");
5953 ProductOps.push_back(Ops[1]);
5954 ProductOps.push_back(Ops[2]);
5956 EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_sqdmulls_scalar),
5957 ProductOps,
"vqdmlXl");
5962 unsigned AccInt = (BuiltinID == NEON::BI__builtin_neon_vqdmlals_lane_s32 ||
5963 BuiltinID == NEON::BI__builtin_neon_vqdmlals_laneq_s32)
5964 ? Intrinsic::aarch64_neon_sqadd
5965 : Intrinsic::aarch64_neon_sqsub;
5968 case NEON::BI__builtin_neon_vget_lane_bf16:
5969 case NEON::BI__builtin_neon_vduph_lane_bf16:
5970 case NEON::BI__builtin_neon_vduph_lane_f16: {
5971 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vget_lane");
5973 case NEON::BI__builtin_neon_vgetq_lane_bf16:
5974 case NEON::BI__builtin_neon_vduph_laneq_bf16:
5975 case NEON::BI__builtin_neon_vduph_laneq_f16: {
5976 return Builder.CreateExtractElement(Ops[0], Ops[1],
"vgetq_lane");
5978 case NEON::BI__builtin_neon_vcvt_bf16_f32: {
5979 llvm::Type *V4F32 = FixedVectorType::get(
Builder.getFloatTy(), 4);
5980 llvm::Type *V4BF16 = FixedVectorType::get(
Builder.getBFloatTy(), 4);
5981 return Builder.CreateFPTrunc(
Builder.CreateBitCast(Ops[0], V4F32), V4BF16);
5983 case NEON::BI__builtin_neon_vcvtq_low_bf16_f32: {
5985 std::iota(ConcatMask.begin(), ConcatMask.end(), 0);
5986 llvm::Type *V4F32 = FixedVectorType::get(
Builder.getFloatTy(), 4);
5987 llvm::Type *V4BF16 = FixedVectorType::get(
Builder.getBFloatTy(), 4);
5988 llvm::Value *Trunc =
5989 Builder.CreateFPTrunc(
Builder.CreateBitCast(Ops[0], V4F32), V4BF16);
5990 return Builder.CreateShuffleVector(
5991 Trunc, ConstantAggregateZero::get(V4BF16), ConcatMask);
5993 case NEON::BI__builtin_neon_vcvtq_high_bf16_f32: {
5995 std::iota(ConcatMask.begin(), ConcatMask.end(), 0);
5997 std::iota(LoMask.begin(), LoMask.end(), 0);
5998 llvm::Type *V4F32 = FixedVectorType::get(
Builder.getFloatTy(), 4);
5999 llvm::Type *V4BF16 = FixedVectorType::get(
Builder.getBFloatTy(), 4);
6000 llvm::Type *V8BF16 = FixedVectorType::get(
Builder.getBFloatTy(), 8);
6001 llvm::Value *Inactive =
Builder.CreateShuffleVector(
6002 Builder.CreateBitCast(Ops[0], V8BF16), LoMask);
6003 llvm::Value *Trunc =
6004 Builder.CreateFPTrunc(
Builder.CreateBitCast(Ops[1], V4F32), V4BF16);
6005 return Builder.CreateShuffleVector(Inactive, Trunc, ConcatMask);
6008 case clang::AArch64::BI_InterlockedAdd:
6009 case clang::AArch64::BI_InterlockedAdd_acq:
6010 case clang::AArch64::BI_InterlockedAdd_rel:
6011 case clang::AArch64::BI_InterlockedAdd_nf:
6012 case clang::AArch64::BI_InterlockedAdd64:
6013 case clang::AArch64::BI_InterlockedAdd64_acq:
6014 case clang::AArch64::BI_InterlockedAdd64_rel:
6015 case clang::AArch64::BI_InterlockedAdd64_nf: {
6017 Value *Val = Ops[1];
6018 llvm::AtomicOrdering Ordering;
6019 switch (BuiltinID) {
6020 case clang::AArch64::BI_InterlockedAdd:
6021 case clang::AArch64::BI_InterlockedAdd64:
6022 Ordering = llvm::AtomicOrdering::SequentiallyConsistent;
6024 case clang::AArch64::BI_InterlockedAdd_acq:
6025 case clang::AArch64::BI_InterlockedAdd64_acq:
6026 Ordering = llvm::AtomicOrdering::Acquire;
6028 case clang::AArch64::BI_InterlockedAdd_rel:
6029 case clang::AArch64::BI_InterlockedAdd64_rel:
6030 Ordering = llvm::AtomicOrdering::Release;
6032 case clang::AArch64::BI_InterlockedAdd_nf:
6033 case clang::AArch64::BI_InterlockedAdd64_nf:
6034 Ordering = llvm::AtomicOrdering::Monotonic;
6037 llvm_unreachable(
"missing builtin ID in switch!");
6039 AtomicRMWInst *RMWI =
6040 Builder.CreateAtomicRMW(AtomicRMWInst::Add, DestAddr, Val, Ordering);
6041 return Builder.CreateAdd(RMWI, Val);
6046 llvm::Type *Ty = VTy;
6050 bool ExtractLow =
false;
6051 bool ExtendLaneArg =
false;
6052 switch (BuiltinID) {
6053 default:
return nullptr;
6054 case NEON::BI__builtin_neon_vbsl_v:
6055 case NEON::BI__builtin_neon_vbslq_v: {
6056 llvm::Type *BitTy = llvm::VectorType::getInteger(VTy);
6057 Ops[0] =
Builder.CreateBitCast(Ops[0], BitTy,
"vbsl");
6058 Ops[1] =
Builder.CreateBitCast(Ops[1], BitTy,
"vbsl");
6059 Ops[2] =
Builder.CreateBitCast(Ops[2], BitTy,
"vbsl");
6061 Ops[1] =
Builder.CreateAnd(Ops[0], Ops[1],
"vbsl");
6062 Ops[2] =
Builder.CreateAnd(
Builder.CreateNot(Ops[0]), Ops[2],
"vbsl");
6063 Ops[0] =
Builder.CreateOr(Ops[1], Ops[2],
"vbsl");
6064 return Builder.CreateBitCast(Ops[0], Ty);
6066 case NEON::BI__builtin_neon_vfma_lane_v:
6067 case NEON::BI__builtin_neon_vfmaq_lane_v: {
6070 Value *Addend = Ops[0];
6071 Value *Multiplicand = Ops[1];
6072 Value *LaneSource = Ops[2];
6073 Ops[0] = Multiplicand;
6074 Ops[1] = LaneSource;
6078 auto *SourceTy = BuiltinID == NEON::BI__builtin_neon_vfmaq_lane_v
6079 ? llvm::FixedVectorType::get(VTy->getElementType(),
6080 VTy->getNumElements() / 2)
6083 Value *SV = llvm::ConstantVector::getSplat(VTy->getElementCount(), cst);
6084 Ops[1] =
Builder.CreateBitCast(Ops[1], SourceTy);
6085 Ops[1] =
Builder.CreateShuffleVector(Ops[1], Ops[1], SV,
"lane");
6088 Int =
Builder.getIsFPConstrained() ? Intrinsic::experimental_constrained_fma
6092 case NEON::BI__builtin_neon_vfma_laneq_v: {
6095 if (VTy && VTy->getElementType() ==
DoubleTy) {
6098 llvm::FixedVectorType *VTy =
6100 Ops[2] =
Builder.CreateBitCast(Ops[2], VTy);
6101 Ops[2] =
Builder.CreateExtractElement(Ops[2], Ops[3],
"extract");
6104 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma,
6105 DoubleTy, {Ops[1], Ops[2], Ops[0]});
6108 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
6109 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6111 auto *STy = llvm::FixedVectorType::get(VTy->getElementType(),
6112 VTy->getNumElements() * 2);
6113 Ops[2] =
Builder.CreateBitCast(Ops[2], STy);
6114 Value *SV = llvm::ConstantVector::getSplat(VTy->getElementCount(),
6116 Ops[2] =
Builder.CreateShuffleVector(Ops[2], Ops[2], SV,
"lane");
6119 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma, Ty,
6120 {Ops[2], Ops[1], Ops[0]});
6122 case NEON::BI__builtin_neon_vfmaq_laneq_v: {
6123 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
6124 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6126 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6129 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma, Ty,
6130 {Ops[2], Ops[1], Ops[0]});
6132 case NEON::BI__builtin_neon_vfmah_lane_f16:
6133 case NEON::BI__builtin_neon_vfmas_lane_f32:
6134 case NEON::BI__builtin_neon_vfmah_laneq_f16:
6135 case NEON::BI__builtin_neon_vfmas_laneq_f32:
6136 case NEON::BI__builtin_neon_vfmad_lane_f64:
6137 case NEON::BI__builtin_neon_vfmad_laneq_f64: {
6139 Ops[2] =
Builder.CreateExtractElement(Ops[2], Ops[3],
"extract");
6141 *
this, Intrinsic::fma, Intrinsic::experimental_constrained_fma, Ty,
6142 {Ops[1], Ops[2], Ops[0]});
6144 case NEON::BI__builtin_neon_vmull_v:
6146 Int = usgn ? Intrinsic::aarch64_neon_umull : Intrinsic::aarch64_neon_smull;
6147 if (
Type.isPoly()) Int = Intrinsic::aarch64_neon_pmull;
6149 case NEON::BI__builtin_neon_vmax_v:
6150 case NEON::BI__builtin_neon_vmaxq_v:
6152 Int = usgn ? Intrinsic::aarch64_neon_umax : Intrinsic::aarch64_neon_smax;
6153 if (Ty->isFPOrFPVectorTy()) Int = Intrinsic::aarch64_neon_fmax;
6155 case NEON::BI__builtin_neon_vmaxh_f16: {
6156 Int = Intrinsic::aarch64_neon_fmax;
6159 case NEON::BI__builtin_neon_vmin_v:
6160 case NEON::BI__builtin_neon_vminq_v:
6162 Int = usgn ? Intrinsic::aarch64_neon_umin : Intrinsic::aarch64_neon_smin;
6163 if (Ty->isFPOrFPVectorTy()) Int = Intrinsic::aarch64_neon_fmin;
6165 case NEON::BI__builtin_neon_vminh_f16: {
6166 Int = Intrinsic::aarch64_neon_fmin;
6169 case NEON::BI__builtin_neon_vabd_v:
6170 case NEON::BI__builtin_neon_vabdq_v:
6172 Int = usgn ? Intrinsic::aarch64_neon_uabd : Intrinsic::aarch64_neon_sabd;
6173 if (Ty->isFPOrFPVectorTy()) Int = Intrinsic::aarch64_neon_fabd;
6175 case NEON::BI__builtin_neon_vpadal_v:
6176 case NEON::BI__builtin_neon_vpadalq_v: {
6177 unsigned ArgElts = VTy->getNumElements();
6179 unsigned BitWidth = EltTy->getBitWidth();
6180 auto *ArgTy = llvm::FixedVectorType::get(
6181 llvm::IntegerType::get(
getLLVMContext(), BitWidth / 2), 2 * ArgElts);
6182 llvm::Type* Tys[2] = { VTy, ArgTy };
6183 Int = usgn ? Intrinsic::aarch64_neon_uaddlp : Intrinsic::aarch64_neon_saddlp;
6185 TmpOps.push_back(Ops[1]);
6188 llvm::Value *addend =
Builder.CreateBitCast(Ops[0], tmp->getType());
6189 return Builder.CreateAdd(tmp, addend);
6191 case NEON::BI__builtin_neon_vpmin_v:
6192 case NEON::BI__builtin_neon_vpminq_v:
6194 Int = usgn ? Intrinsic::aarch64_neon_uminp : Intrinsic::aarch64_neon_sminp;
6195 if (Ty->isFPOrFPVectorTy()) Int = Intrinsic::aarch64_neon_fminp;
6197 case NEON::BI__builtin_neon_vpmax_v:
6198 case NEON::BI__builtin_neon_vpmaxq_v:
6200 Int = usgn ? Intrinsic::aarch64_neon_umaxp : Intrinsic::aarch64_neon_smaxp;
6201 if (Ty->isFPOrFPVectorTy()) Int = Intrinsic::aarch64_neon_fmaxp;
6203 case NEON::BI__builtin_neon_vminnm_v:
6204 case NEON::BI__builtin_neon_vminnmq_v:
6205 Int = Intrinsic::aarch64_neon_fminnm;
6207 case NEON::BI__builtin_neon_vminnmh_f16:
6208 Int = Intrinsic::aarch64_neon_fminnm;
6210 case NEON::BI__builtin_neon_vmaxnm_v:
6211 case NEON::BI__builtin_neon_vmaxnmq_v:
6212 Int = Intrinsic::aarch64_neon_fmaxnm;
6214 case NEON::BI__builtin_neon_vmaxnmh_f16:
6215 Int = Intrinsic::aarch64_neon_fmaxnm;
6217 case NEON::BI__builtin_neon_vrecpss_f32: {
6221 case NEON::BI__builtin_neon_vrecpsd_f64:
6224 case NEON::BI__builtin_neon_vrecpsh_f16:
6227 case NEON::BI__builtin_neon_vqshrun_n_v:
6228 Int = Intrinsic::aarch64_neon_sqshrun;
6230 case NEON::BI__builtin_neon_vqrshrun_n_v:
6231 Int = Intrinsic::aarch64_neon_sqrshrun;
6233 case NEON::BI__builtin_neon_vqshrn_n_v:
6234 Int = usgn ? Intrinsic::aarch64_neon_uqshrn : Intrinsic::aarch64_neon_sqshrn;
6236 case NEON::BI__builtin_neon_vrshrn_n_v:
6237 Int = Intrinsic::aarch64_neon_rshrn;
6239 case NEON::BI__builtin_neon_vqrshrn_n_v:
6240 Int = usgn ? Intrinsic::aarch64_neon_uqrshrn : Intrinsic::aarch64_neon_sqrshrn;
6242 case NEON::BI__builtin_neon_vrndah_f16: {
6243 Int =
Builder.getIsFPConstrained()
6244 ? Intrinsic::experimental_constrained_round
6248 case NEON::BI__builtin_neon_vrnda_v:
6249 case NEON::BI__builtin_neon_vrndaq_v: {
6250 Int =
Builder.getIsFPConstrained()
6251 ? Intrinsic::experimental_constrained_round
6255 case NEON::BI__builtin_neon_vrndih_f16: {
6256 Int =
Builder.getIsFPConstrained()
6257 ? Intrinsic::experimental_constrained_nearbyint
6258 : Intrinsic::nearbyint;
6261 case NEON::BI__builtin_neon_vrndmh_f16: {
6262 Int =
Builder.getIsFPConstrained()
6263 ? Intrinsic::experimental_constrained_floor
6267 case NEON::BI__builtin_neon_vrndm_v:
6268 case NEON::BI__builtin_neon_vrndmq_v: {
6269 Int =
Builder.getIsFPConstrained()
6270 ? Intrinsic::experimental_constrained_floor
6274 case NEON::BI__builtin_neon_vrndnh_f16: {
6275 Int =
Builder.getIsFPConstrained()
6276 ? Intrinsic::experimental_constrained_roundeven
6277 : Intrinsic::roundeven;
6280 case NEON::BI__builtin_neon_vrndn_v:
6281 case NEON::BI__builtin_neon_vrndnq_v: {
6282 Int =
Builder.getIsFPConstrained()
6283 ? Intrinsic::experimental_constrained_roundeven
6284 : Intrinsic::roundeven;
6287 case NEON::BI__builtin_neon_vrndns_f32: {
6288 Int =
Builder.getIsFPConstrained()
6289 ? Intrinsic::experimental_constrained_roundeven
6290 : Intrinsic::roundeven;
6293 case NEON::BI__builtin_neon_vrndph_f16: {
6294 Int =
Builder.getIsFPConstrained()
6295 ? Intrinsic::experimental_constrained_ceil
6299 case NEON::BI__builtin_neon_vrndp_v:
6300 case NEON::BI__builtin_neon_vrndpq_v: {
6301 Int =
Builder.getIsFPConstrained()
6302 ? Intrinsic::experimental_constrained_ceil
6306 case NEON::BI__builtin_neon_vrndxh_f16: {
6307 Int =
Builder.getIsFPConstrained()
6308 ? Intrinsic::experimental_constrained_rint
6312 case NEON::BI__builtin_neon_vrndx_v:
6313 case NEON::BI__builtin_neon_vrndxq_v: {
6314 Int =
Builder.getIsFPConstrained()
6315 ? Intrinsic::experimental_constrained_rint
6319 case NEON::BI__builtin_neon_vrndh_f16: {
6320 Int =
Builder.getIsFPConstrained()
6321 ? Intrinsic::experimental_constrained_trunc
6325 case NEON::BI__builtin_neon_vrnd32x_f32:
6326 case NEON::BI__builtin_neon_vrnd32xq_f32:
6327 case NEON::BI__builtin_neon_vrnd32x_f64:
6328 case NEON::BI__builtin_neon_vrnd32xq_f64: {
6329 Int = Intrinsic::aarch64_neon_frint32x;
6332 case NEON::BI__builtin_neon_vrnd32z_f32:
6333 case NEON::BI__builtin_neon_vrnd32zq_f32:
6334 case NEON::BI__builtin_neon_vrnd32z_f64:
6335 case NEON::BI__builtin_neon_vrnd32zq_f64: {
6336 Int = Intrinsic::aarch64_neon_frint32z;
6339 case NEON::BI__builtin_neon_vrnd64x_f32:
6340 case NEON::BI__builtin_neon_vrnd64xq_f32:
6341 case NEON::BI__builtin_neon_vrnd64x_f64:
6342 case NEON::BI__builtin_neon_vrnd64xq_f64: {
6343 Int = Intrinsic::aarch64_neon_frint64x;
6346 case NEON::BI__builtin_neon_vrnd64z_f32:
6347 case NEON::BI__builtin_neon_vrnd64zq_f32:
6348 case NEON::BI__builtin_neon_vrnd64z_f64:
6349 case NEON::BI__builtin_neon_vrnd64zq_f64: {
6350 Int = Intrinsic::aarch64_neon_frint64z;
6353 case NEON::BI__builtin_neon_vrnd_v:
6354 case NEON::BI__builtin_neon_vrndq_v: {
6355 Int =
Builder.getIsFPConstrained()
6356 ? Intrinsic::experimental_constrained_trunc
6360 case NEON::BI__builtin_neon_vcvt_f64_v:
6361 case NEON::BI__builtin_neon_vcvtq_f64_v:
6362 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
6364 return usgn ?
Builder.CreateUIToFP(Ops[0], Ty,
"vcvt")
6365 :
Builder.CreateSIToFP(Ops[0], Ty,
"vcvt");
6366 case NEON::BI__builtin_neon_vcvt_f64_f32: {
6368 "unexpected vcvt_f64_f32 builtin");
6372 return Builder.CreateFPExt(Ops[0], Ty,
"vcvt");
6374 case NEON::BI__builtin_neon_vcvt_f32_f64: {
6376 "unexpected vcvt_f32_f64 builtin");
6380 return Builder.CreateFPTrunc(Ops[0], Ty,
"vcvt");
6382 case NEON::BI__builtin_neon_vcvta_s16_f16:
6383 case NEON::BI__builtin_neon_vcvta_u16_f16:
6384 case NEON::BI__builtin_neon_vcvta_s32_v:
6385 case NEON::BI__builtin_neon_vcvtaq_s16_f16:
6386 case NEON::BI__builtin_neon_vcvtaq_s32_v:
6387 case NEON::BI__builtin_neon_vcvta_u32_v:
6388 case NEON::BI__builtin_neon_vcvtaq_u16_f16:
6389 case NEON::BI__builtin_neon_vcvtaq_u32_v:
6390 case NEON::BI__builtin_neon_vcvta_s64_v:
6391 case NEON::BI__builtin_neon_vcvtaq_s64_v:
6392 case NEON::BI__builtin_neon_vcvta_u64_v:
6393 case NEON::BI__builtin_neon_vcvtaq_u64_v: {
6394 Int = usgn ? Intrinsic::aarch64_neon_fcvtau : Intrinsic::aarch64_neon_fcvtas;
6398 case NEON::BI__builtin_neon_vcvtm_s16_f16:
6399 case NEON::BI__builtin_neon_vcvtm_s32_v:
6400 case NEON::BI__builtin_neon_vcvtmq_s16_f16:
6401 case NEON::BI__builtin_neon_vcvtmq_s32_v:
6402 case NEON::BI__builtin_neon_vcvtm_u16_f16:
6403 case NEON::BI__builtin_neon_vcvtm_u32_v:
6404 case NEON::BI__builtin_neon_vcvtmq_u16_f16:
6405 case NEON::BI__builtin_neon_vcvtmq_u32_v:
6406 case NEON::BI__builtin_neon_vcvtm_s64_v:
6407 case NEON::BI__builtin_neon_vcvtmq_s64_v:
6408 case NEON::BI__builtin_neon_vcvtm_u64_v:
6409 case NEON::BI__builtin_neon_vcvtmq_u64_v: {
6410 Int = usgn ? Intrinsic::aarch64_neon_fcvtmu : Intrinsic::aarch64_neon_fcvtms;
6414 case NEON::BI__builtin_neon_vcvtn_s16_f16:
6415 case NEON::BI__builtin_neon_vcvtn_s32_v:
6416 case NEON::BI__builtin_neon_vcvtnq_s16_f16:
6417 case NEON::BI__builtin_neon_vcvtnq_s32_v:
6418 case NEON::BI__builtin_neon_vcvtn_u16_f16:
6419 case NEON::BI__builtin_neon_vcvtn_u32_v:
6420 case NEON::BI__builtin_neon_vcvtnq_u16_f16:
6421 case NEON::BI__builtin_neon_vcvtnq_u32_v:
6422 case NEON::BI__builtin_neon_vcvtn_s64_v:
6423 case NEON::BI__builtin_neon_vcvtnq_s64_v:
6424 case NEON::BI__builtin_neon_vcvtn_u64_v:
6425 case NEON::BI__builtin_neon_vcvtnq_u64_v: {
6426 Int = usgn ? Intrinsic::aarch64_neon_fcvtnu : Intrinsic::aarch64_neon_fcvtns;
6430 case NEON::BI__builtin_neon_vcvtp_s16_f16:
6431 case NEON::BI__builtin_neon_vcvtp_s32_v:
6432 case NEON::BI__builtin_neon_vcvtpq_s16_f16:
6433 case NEON::BI__builtin_neon_vcvtpq_s32_v:
6434 case NEON::BI__builtin_neon_vcvtp_u16_f16:
6435 case NEON::BI__builtin_neon_vcvtp_u32_v:
6436 case NEON::BI__builtin_neon_vcvtpq_u16_f16:
6437 case NEON::BI__builtin_neon_vcvtpq_u32_v:
6438 case NEON::BI__builtin_neon_vcvtp_s64_v:
6439 case NEON::BI__builtin_neon_vcvtpq_s64_v:
6440 case NEON::BI__builtin_neon_vcvtp_u64_v:
6441 case NEON::BI__builtin_neon_vcvtpq_u64_v: {
6442 Int = usgn ? Intrinsic::aarch64_neon_fcvtpu : Intrinsic::aarch64_neon_fcvtps;
6446 case NEON::BI__builtin_neon_vmulx_v:
6447 case NEON::BI__builtin_neon_vmulxq_v: {
6448 Int = Intrinsic::aarch64_neon_fmulx;
6451 case NEON::BI__builtin_neon_vmulxh_lane_f16:
6452 case NEON::BI__builtin_neon_vmulxh_laneq_f16: {
6455 Ops[1] =
Builder.CreateExtractElement(Ops[1], Ops[2],
"extract");
6457 Int = Intrinsic::aarch64_neon_fmulx;
6460 case NEON::BI__builtin_neon_vmul_lane_v:
6461 case NEON::BI__builtin_neon_vmul_laneq_v: {
6464 if (BuiltinID == NEON::BI__builtin_neon_vmul_laneq_v)
6467 llvm::FixedVectorType *VTy =
6469 Ops[1] =
Builder.CreateBitCast(Ops[1], VTy);
6470 Ops[1] =
Builder.CreateExtractElement(Ops[1], Ops[2],
"extract");
6474 case NEON::BI__builtin_neon_vpmaxnm_v:
6475 case NEON::BI__builtin_neon_vpmaxnmq_v: {
6476 Int = Intrinsic::aarch64_neon_fmaxnmp;
6479 case NEON::BI__builtin_neon_vpminnm_v:
6480 case NEON::BI__builtin_neon_vpminnmq_v: {
6481 Int = Intrinsic::aarch64_neon_fminnmp;
6484 case NEON::BI__builtin_neon_vsqrth_f16: {
6485 Int =
Builder.getIsFPConstrained()
6486 ? Intrinsic::experimental_constrained_sqrt
6490 case NEON::BI__builtin_neon_vsqrt_v:
6491 case NEON::BI__builtin_neon_vsqrtq_v: {
6492 Int =
Builder.getIsFPConstrained()
6493 ? Intrinsic::experimental_constrained_sqrt
6495 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
6498 case NEON::BI__builtin_neon_vrbit_v:
6499 case NEON::BI__builtin_neon_vrbitq_v: {
6500 Int = Intrinsic::bitreverse;
6503 case NEON::BI__builtin_neon_vmaxv_f16: {
6504 Int = Intrinsic::aarch64_neon_fmaxv;
6506 VTy = llvm::FixedVectorType::get(
HalfTy, 4);
6507 llvm::Type *Tys[2] = {Ty, VTy};
6511 case NEON::BI__builtin_neon_vmaxvq_f16: {
6512 Int = Intrinsic::aarch64_neon_fmaxv;
6514 VTy = llvm::FixedVectorType::get(
HalfTy, 8);
6515 llvm::Type *Tys[2] = {Ty, VTy};
6519 case NEON::BI__builtin_neon_vminv_f16: {
6520 Int = Intrinsic::aarch64_neon_fminv;
6522 VTy = llvm::FixedVectorType::get(
HalfTy, 4);
6523 llvm::Type *Tys[2] = {Ty, VTy};
6527 case NEON::BI__builtin_neon_vminvq_f16: {
6528 Int = Intrinsic::aarch64_neon_fminv;
6530 VTy = llvm::FixedVectorType::get(
HalfTy, 8);
6531 llvm::Type *Tys[2] = {Ty, VTy};
6535 case NEON::BI__builtin_neon_vmaxnmv_f16: {
6536 Int = Intrinsic::aarch64_neon_fmaxnmv;
6538 VTy = llvm::FixedVectorType::get(
HalfTy, 4);
6539 llvm::Type *Tys[2] = {Ty, VTy};
6543 case NEON::BI__builtin_neon_vmaxnmvq_f16: {
6544 Int = Intrinsic::aarch64_neon_fmaxnmv;
6546 VTy = llvm::FixedVectorType::get(
HalfTy, 8);
6547 llvm::Type *Tys[2] = {Ty, VTy};
6551 case NEON::BI__builtin_neon_vminnmv_f16: {
6552 Int = Intrinsic::aarch64_neon_fminnmv;
6554 VTy = llvm::FixedVectorType::get(
HalfTy, 4);
6555 llvm::Type *Tys[2] = {Ty, VTy};
6559 case NEON::BI__builtin_neon_vminnmvq_f16: {
6560 Int = Intrinsic::aarch64_neon_fminnmv;
6562 VTy = llvm::FixedVectorType::get(
HalfTy, 8);
6563 llvm::Type *Tys[2] = {Ty, VTy};
6567 case NEON::BI__builtin_neon_vmul_n_f64: {
6570 return Builder.CreateFMul(Ops[0], RHS);
6572 case NEON::BI__builtin_neon_vaddlv_u8: {
6573 Int = Intrinsic::aarch64_neon_uaddlv;
6575 VTy = llvm::FixedVectorType::get(
Int8Ty, 8);
6576 llvm::Type *Tys[2] = {Ty, VTy};
6580 case NEON::BI__builtin_neon_vaddlv_u16: {
6581 Int = Intrinsic::aarch64_neon_uaddlv;
6583 VTy = llvm::FixedVectorType::get(
Int16Ty, 4);
6584 llvm::Type *Tys[2] = {Ty, VTy};
6587 case NEON::BI__builtin_neon_vaddlvq_u8: {
6588 Int = Intrinsic::aarch64_neon_uaddlv;
6590 VTy = llvm::FixedVectorType::get(
Int8Ty, 16);
6591 llvm::Type *Tys[2] = {Ty, VTy};
6595 case NEON::BI__builtin_neon_vaddlvq_u16: {
6596 Int = Intrinsic::aarch64_neon_uaddlv;
6598 VTy = llvm::FixedVectorType::get(
Int16Ty, 8);
6599 llvm::Type *Tys[2] = {Ty, VTy};
6602 case NEON::BI__builtin_neon_vaddlv_s8: {
6603 Int = Intrinsic::aarch64_neon_saddlv;
6605 VTy = llvm::FixedVectorType::get(
Int8Ty, 8);
6606 llvm::Type *Tys[2] = {Ty, VTy};
6610 case NEON::BI__builtin_neon_vaddlv_s16: {
6611 Int = Intrinsic::aarch64_neon_saddlv;
6613 VTy = llvm::FixedVectorType::get(
Int16Ty, 4);
6614 llvm::Type *Tys[2] = {Ty, VTy};
6617 case NEON::BI__builtin_neon_vaddlvq_s8: {
6618 Int = Intrinsic::aarch64_neon_saddlv;
6620 VTy = llvm::FixedVectorType::get(
Int8Ty, 16);
6621 llvm::Type *Tys[2] = {Ty, VTy};
6625 case NEON::BI__builtin_neon_vaddlvq_s16: {
6626 Int = Intrinsic::aarch64_neon_saddlv;
6628 VTy = llvm::FixedVectorType::get(
Int16Ty, 8);
6629 llvm::Type *Tys[2] = {Ty, VTy};
6632 case NEON::BI__builtin_neon_vsri_n_v:
6633 case NEON::BI__builtin_neon_vsriq_n_v: {
6634 Int = Intrinsic::aarch64_neon_vsri;
6635 llvm::Function *Intrin =
CGM.getIntrinsic(Int, Ty);
6638 case NEON::BI__builtin_neon_vsli_n_v:
6639 case NEON::BI__builtin_neon_vsliq_n_v: {
6640 Int = Intrinsic::aarch64_neon_vsli;
6641 llvm::Function *Intrin =
CGM.getIntrinsic(Int, Ty);
6644 case NEON::BI__builtin_neon_vsra_n_v:
6645 case NEON::BI__builtin_neon_vsraq_n_v:
6646 Ops[0] =
Builder.CreateBitCast(Ops[0], Ty);
6648 return Builder.CreateAdd(Ops[0], Ops[1]);
6649 case NEON::BI__builtin_neon_vrsra_n_v:
6650 case NEON::BI__builtin_neon_vrsraq_n_v: {
6651 Int = usgn ? Intrinsic::aarch64_neon_urshl : Intrinsic::aarch64_neon_srshl;
6653 TmpOps.push_back(Ops[1]);
6654 TmpOps.push_back(Ops[2]);
6656 llvm::Value *tmp =
EmitNeonCall(F, TmpOps,
"vrshr_n", 1,
true);
6657 Ops[0] =
Builder.CreateBitCast(Ops[0], VTy);
6658 return Builder.CreateAdd(Ops[0], tmp);
6660 case NEON::BI__builtin_neon_vld1_v:
6661 case NEON::BI__builtin_neon_vld1q_v: {
6664 case NEON::BI__builtin_neon_vst1_v:
6665 case NEON::BI__builtin_neon_vst1q_v:
6666 Ops[1] =
Builder.CreateBitCast(Ops[1], VTy);
6668 case NEON::BI__builtin_neon_vld1_lane_v:
6669 case NEON::BI__builtin_neon_vld1q_lane_v: {
6670 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6671 Ops[0] =
Builder.CreateAlignedLoad(VTy->getElementType(), Ops[0],
6673 return Builder.CreateInsertElement(Ops[1], Ops[0], Ops[2],
"vld1_lane");
6675 case NEON::BI__builtin_neon_vldap1_lane_s64:
6676 case NEON::BI__builtin_neon_vldap1q_lane_s64: {
6677 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6678 llvm::LoadInst *LI =
Builder.CreateAlignedLoad(
6680 LI->setAtomic(llvm::AtomicOrdering::Acquire);
6682 return Builder.CreateInsertElement(Ops[1], Ops[0], Ops[2],
"vldap1_lane");
6684 case NEON::BI__builtin_neon_vld1_dup_v:
6685 case NEON::BI__builtin_neon_vld1q_dup_v: {
6686 Value *
V = PoisonValue::get(Ty);
6687 Ops[0] =
Builder.CreateAlignedLoad(VTy->getElementType(), Ops[0],
6689 llvm::Constant *CI = ConstantInt::get(
Int32Ty, 0);
6690 Ops[0] =
Builder.CreateInsertElement(
V, Ops[0], CI);
6693 case NEON::BI__builtin_neon_vst1_lane_v:
6694 case NEON::BI__builtin_neon_vst1q_lane_v:
6695 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6696 Ops[1] =
Builder.CreateExtractElement(Ops[1], Ops[2]);
6698 case NEON::BI__builtin_neon_vstl1_lane_s64:
6699 case NEON::BI__builtin_neon_vstl1q_lane_s64: {
6700 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6701 Ops[1] =
Builder.CreateExtractElement(Ops[1], Ops[2]);
6702 llvm::StoreInst *SI =
6704 SI->setAtomic(llvm::AtomicOrdering::Release);
6707 case NEON::BI__builtin_neon_vld2_v:
6708 case NEON::BI__builtin_neon_vld2q_v: {
6710 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld2, Tys);
6711 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld2");
6712 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6714 case NEON::BI__builtin_neon_vld3_v:
6715 case NEON::BI__builtin_neon_vld3q_v: {
6717 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld3, Tys);
6718 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld3");
6719 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6721 case NEON::BI__builtin_neon_vld4_v:
6722 case NEON::BI__builtin_neon_vld4q_v: {
6724 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld4, Tys);
6725 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld4");
6726 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6728 case NEON::BI__builtin_neon_vld2_dup_v:
6729 case NEON::BI__builtin_neon_vld2q_dup_v: {
6731 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld2r, Tys);
6732 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld2");
6733 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6735 case NEON::BI__builtin_neon_vld3_dup_v:
6736 case NEON::BI__builtin_neon_vld3q_dup_v: {
6738 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld3r, Tys);
6739 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld3");
6740 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6742 case NEON::BI__builtin_neon_vld4_dup_v:
6743 case NEON::BI__builtin_neon_vld4q_dup_v: {
6745 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld4r, Tys);
6746 Ops[1] =
Builder.CreateCall(F, Ops[1],
"vld4");
6747 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6749 case NEON::BI__builtin_neon_vld2_lane_v:
6750 case NEON::BI__builtin_neon_vld2q_lane_v: {
6751 llvm::Type *Tys[2] = { VTy, Ops[1]->getType() };
6752 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld2lane, Tys);
6753 std::rotate(Ops.begin() + 1, Ops.begin() + 2, Ops.end());
6754 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6755 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6758 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6760 case NEON::BI__builtin_neon_vld3_lane_v:
6761 case NEON::BI__builtin_neon_vld3q_lane_v: {
6762 llvm::Type *Tys[2] = { VTy, Ops[1]->getType() };
6763 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld3lane, Tys);
6764 std::rotate(Ops.begin() + 1, Ops.begin() + 2, Ops.end());
6765 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6766 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6767 Ops[3] =
Builder.CreateBitCast(Ops[3], Ty);
6770 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6772 case NEON::BI__builtin_neon_vld4_lane_v:
6773 case NEON::BI__builtin_neon_vld4q_lane_v: {
6774 llvm::Type *Tys[2] = { VTy, Ops[1]->getType() };
6775 Function *F =
CGM.getIntrinsic(Intrinsic::aarch64_neon_ld4lane, Tys);
6776 std::rotate(Ops.begin() + 1, Ops.begin() + 2, Ops.end());
6777 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6778 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6779 Ops[3] =
Builder.CreateBitCast(Ops[3], Ty);
6780 Ops[4] =
Builder.CreateBitCast(Ops[4], Ty);
6783 return Builder.CreateDefaultAlignedStore(Ops[1], Ops[0]);
6785 case NEON::BI__builtin_neon_vst2_v:
6786 case NEON::BI__builtin_neon_vst2q_v: {
6787 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
6788 llvm::Type *Tys[2] = { VTy, Ops[2]->getType() };
6789 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_st2, Tys),
6792 case NEON::BI__builtin_neon_vst2_lane_v:
6793 case NEON::BI__builtin_neon_vst2q_lane_v: {
6794 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
6796 llvm::Type *Tys[2] = { VTy, Ops[3]->getType() };
6797 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_st2lane, Tys),
6800 case NEON::BI__builtin_neon_vst3_v:
6801 case NEON::BI__builtin_neon_vst3q_v: {
6802 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
6803 llvm::Type *Tys[2] = { VTy, Ops[3]->getType() };
6804 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_st3, Tys),
6807 case NEON::BI__builtin_neon_vst3_lane_v:
6808 case NEON::BI__builtin_neon_vst3q_lane_v: {
6809 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
6811 llvm::Type *Tys[2] = { VTy, Ops[4]->getType() };
6812 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_st3lane, Tys),
6815 case NEON::BI__builtin_neon_vst4_v:
6816 case NEON::BI__builtin_neon_vst4q_v: {
6817 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
6818 llvm::Type *Tys[2] = { VTy, Ops[4]->getType() };
6819 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_st4, Tys),
6822 case NEON::BI__builtin_neon_vst4_lane_v:
6823 case NEON::BI__builtin_neon_vst4q_lane_v: {
6824 std::rotate(Ops.begin(), Ops.begin() + 1, Ops.end());
6826 llvm::Type *Tys[2] = { VTy, Ops[5]->getType() };
6827 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_st4lane, Tys),
6830 case NEON::BI__builtin_neon_vtrn_v:
6831 case NEON::BI__builtin_neon_vtrnq_v: {
6832 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6833 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6834 Value *SV =
nullptr;
6836 for (
unsigned vi = 0; vi != 2; ++vi) {
6838 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; i += 2) {
6839 Indices.push_back(i+vi);
6840 Indices.push_back(i+e+vi);
6843 SV =
Builder.CreateShuffleVector(Ops[1], Ops[2], Indices,
"vtrn");
6848 case NEON::BI__builtin_neon_vuzp_v:
6849 case NEON::BI__builtin_neon_vuzpq_v: {
6850 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6851 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6852 Value *SV =
nullptr;
6854 for (
unsigned vi = 0; vi != 2; ++vi) {
6856 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; ++i)
6857 Indices.push_back(2*i+vi);
6860 SV =
Builder.CreateShuffleVector(Ops[1], Ops[2], Indices,
"vuzp");
6865 case NEON::BI__builtin_neon_vzip_v:
6866 case NEON::BI__builtin_neon_vzipq_v: {
6867 Ops[1] =
Builder.CreateBitCast(Ops[1], Ty);
6868 Ops[2] =
Builder.CreateBitCast(Ops[2], Ty);
6869 Value *SV =
nullptr;
6871 for (
unsigned vi = 0; vi != 2; ++vi) {
6873 for (
unsigned i = 0, e = VTy->getNumElements(); i != e; i += 2) {
6874 Indices.push_back((i + vi*e) >> 1);
6875 Indices.push_back(((i + vi*e) >> 1)+e);
6878 SV =
Builder.CreateShuffleVector(Ops[1], Ops[2], Indices,
"vzip");
6883 case NEON::BI__builtin_neon_vqtbl1q_v: {
6884 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbl1, Ty),
6887 case NEON::BI__builtin_neon_vqtbl2q_v: {
6888 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbl2, Ty),
6891 case NEON::BI__builtin_neon_vqtbl3q_v: {
6892 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbl3, Ty),
6895 case NEON::BI__builtin_neon_vqtbl4q_v: {
6896 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbl4, Ty),
6899 case NEON::BI__builtin_neon_vqtbx1q_v: {
6900 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbx1, Ty),
6903 case NEON::BI__builtin_neon_vqtbx2q_v: {
6904 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbx2, Ty),
6907 case NEON::BI__builtin_neon_vqtbx3q_v: {
6908 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbx3, Ty),
6911 case NEON::BI__builtin_neon_vqtbx4q_v: {
6912 return EmitNeonCall(
CGM.getIntrinsic(Intrinsic::aarch64_neon_tbx4, Ty),
6915 case NEON::BI__builtin_neon_vsqadd_v:
6916 case NEON::BI__builtin_neon_vsqaddq_v: {
6917 Int = Intrinsic::aarch64_neon_usqadd;
6920 case NEON::BI__builtin_neon_vuqadd_v:
6921 case NEON::BI__builtin_neon_vuqaddq_v: {
6922 Int = Intrinsic::aarch64_neon_suqadd;
6926 case NEON::BI__builtin_neon_vluti2_laneq_mf8:
6927 case NEON::BI__builtin_neon_vluti2_laneq_bf16:
6928 case NEON::BI__builtin_neon_vluti2_laneq_f16:
6929 case NEON::BI__builtin_neon_vluti2_laneq_p16:
6930 case NEON::BI__builtin_neon_vluti2_laneq_p8:
6931 case NEON::BI__builtin_neon_vluti2_laneq_s16:
6932 case NEON::BI__builtin_neon_vluti2_laneq_s8:
6933 case NEON::BI__builtin_neon_vluti2_laneq_u16:
6934 case NEON::BI__builtin_neon_vluti2_laneq_u8: {
6935 Int = Intrinsic::aarch64_neon_vluti2_laneq;
6942 case NEON::BI__builtin_neon_vluti2q_laneq_mf8:
6943 case NEON::BI__builtin_neon_vluti2q_laneq_bf16:
6944 case NEON::BI__builtin_neon_vluti2q_laneq_f16:
6945 case NEON::BI__builtin_neon_vluti2q_laneq_p16:
6946 case NEON::BI__builtin_neon_vluti2q_laneq_p8:
6947 case NEON::BI__builtin_neon_vluti2q_laneq_s16:
6948 case NEON::BI__builtin_neon_vluti2q_laneq_s8:
6949 case NEON::BI__builtin_neon_vluti2q_laneq_u16:
6950 case NEON::BI__builtin_neon_vluti2q_laneq_u8: {
6951 Int = Intrinsic::aarch64_neon_vluti2_laneq;
6958 case NEON::BI__builtin_neon_vluti2_lane_mf8:
6959 case NEON::BI__builtin_neon_vluti2_lane_bf16:
6960 case NEON::BI__builtin_neon_vluti2_lane_f16:
6961 case NEON::BI__builtin_neon_vluti2_lane_p16:
6962 case NEON::BI__builtin_neon_vluti2_lane_p8:
6963 case NEON::BI__builtin_neon_vluti2_lane_s16:
6964 case NEON::BI__builtin_neon_vluti2_lane_s8:
6965 case NEON::BI__builtin_neon_vluti2_lane_u16:
6966 case NEON::BI__builtin_neon_vluti2_lane_u8: {
6967 Int = Intrinsic::aarch64_neon_vluti2_lane;
6974 case NEON::BI__builtin_neon_vluti2q_lane_mf8:
6975 case NEON::BI__builtin_neon_vluti2q_lane_bf16:
6976 case NEON::BI__builtin_neon_vluti2q_lane_f16:
6977 case NEON::BI__builtin_neon_vluti2q_lane_p16:
6978 case NEON::BI__builtin_neon_vluti2q_lane_p8:
6979 case NEON::BI__builtin_neon_vluti2q_lane_s16:
6980 case NEON::BI__builtin_neon_vluti2q_lane_s8:
6981 case NEON::BI__builtin_neon_vluti2q_lane_u16:
6982 case NEON::BI__builtin_neon_vluti2q_lane_u8: {
6983 Int = Intrinsic::aarch64_neon_vluti2_lane;
6990 case NEON::BI__builtin_neon_vluti4q_lane_mf8:
6991 case NEON::BI__builtin_neon_vluti4q_lane_p8:
6992 case NEON::BI__builtin_neon_vluti4q_lane_s8:
6993 case NEON::BI__builtin_neon_vluti4q_lane_u8: {
6994 Int = Intrinsic::aarch64_neon_vluti4q_lane;
6997 case NEON::BI__builtin_neon_vluti4q_laneq_mf8:
6998 case NEON::BI__builtin_neon_vluti4q_laneq_p8:
6999 case NEON::BI__builtin_neon_vluti4q_laneq_s8:
7000 case NEON::BI__builtin_neon_vluti4q_laneq_u8: {
7001 Int = Intrinsic::aarch64_neon_vluti4q_laneq;
7004 case NEON::BI__builtin_neon_vluti4q_lane_bf16_x2:
7005 case NEON::BI__builtin_neon_vluti4q_lane_f16_x2:
7006 case NEON::BI__builtin_neon_vluti4q_lane_p16_x2:
7007 case NEON::BI__builtin_neon_vluti4q_lane_s16_x2:
7008 case NEON::BI__builtin_neon_vluti4q_lane_u16_x2: {
7009 Int = Intrinsic::aarch64_neon_vluti4q_lane_x2;
7010 return EmitNeonCall(
CGM.getIntrinsic(Int, Ty), Ops,
"vluti4q_lane_x2");
7012 case NEON::BI__builtin_neon_vluti4q_laneq_bf16_x2:
7013 case NEON::BI__builtin_neon_vluti4q_laneq_f16_x2:
7014 case NEON::BI__builtin_neon_vluti4q_laneq_p16_x2:
7015 case NEON::BI__builtin_neon_vluti4q_laneq_s16_x2:
7016 case NEON::BI__builtin_neon_vluti4q_laneq_u16_x2: {
7017 Int = Intrinsic::aarch64_neon_vluti4q_laneq_x2;
7018 return EmitNeonCall(
CGM.getIntrinsic(Int, Ty), Ops,
"vluti4q_laneq_x2");
7020 case NEON::BI__builtin_neon_vmmlaq_f16_mf8_fpm:
7022 {llvm::FixedVectorType::get(
HalfTy, 8),
7023 llvm::FixedVectorType::get(
Int8Ty, 16)},
7025 case NEON::BI__builtin_neon_vmmlaq_f32_mf8_fpm:
7027 {llvm::FixedVectorType::get(
FloatTy, 4),
7028 llvm::FixedVectorType::get(
Int8Ty, 16)},
7030 case NEON::BI__builtin_neon_vcvt1_low_bf16_mf8_fpm:
7033 case NEON::BI__builtin_neon_vcvt1_bf16_mf8_fpm:
7034 case NEON::BI__builtin_neon_vcvt1_high_bf16_mf8_fpm:
7036 llvm::FixedVectorType::get(
BFloatTy, 8),
7037 Ops[0]->
getType(), ExtractLow, Ops, E,
"vbfcvt1");
7038 case NEON::BI__builtin_neon_vcvt2_low_bf16_mf8_fpm:
7041 case NEON::BI__builtin_neon_vcvt2_bf16_mf8_fpm:
7042 case NEON::BI__builtin_neon_vcvt2_high_bf16_mf8_fpm:
7044 llvm::FixedVectorType::get(
BFloatTy, 8),
7045 Ops[0]->
getType(), ExtractLow, Ops, E,
"vbfcvt2");
7046 case NEON::BI__builtin_neon_vcvt1_low_f16_mf8_fpm:
7049 case NEON::BI__builtin_neon_vcvt1_f16_mf8_fpm:
7050 case NEON::BI__builtin_neon_vcvt1_high_f16_mf8_fpm:
7052 llvm::FixedVectorType::get(
HalfTy, 8),
7053 Ops[0]->
getType(), ExtractLow, Ops, E,
"vbfcvt1");
7054 case NEON::BI__builtin_neon_vcvt2_low_f16_mf8_fpm:
7057 case NEON::BI__builtin_neon_vcvt2_f16_mf8_fpm:
7058 case NEON::BI__builtin_neon_vcvt2_high_f16_mf8_fpm:
7060 llvm::FixedVectorType::get(
HalfTy, 8),
7061 Ops[0]->
getType(), ExtractLow, Ops, E,
"vbfcvt2");
7062 case NEON::BI__builtin_neon_vcvt_mf8_f32_fpm:
7064 llvm::FixedVectorType::get(
Int8Ty, 8),
7065 Ops[0]->
getType(),
false, Ops, E,
"vfcvtn");
7066 case NEON::BI__builtin_neon_vcvt_mf8_f16_fpm:
7068 llvm::FixedVectorType::get(
Int8Ty, 8),
7069 llvm::FixedVectorType::get(
HalfTy, 4),
false, Ops,
7071 case NEON::BI__builtin_neon_vcvtq_mf8_f16_fpm:
7073 llvm::FixedVectorType::get(
Int8Ty, 16),
7074 llvm::FixedVectorType::get(
HalfTy, 8),
false, Ops,
7076 case NEON::BI__builtin_neon_vcvt_high_mf8_f32_fpm: {
7077 llvm::Type *Ty = llvm::FixedVectorType::get(
Int8Ty, 16);
7078 Ops[0] =
Builder.CreateInsertVector(Ty, PoisonValue::get(Ty), Ops[0],
7081 Ops[1]->
getType(),
false, Ops, E,
"vfcvtn2");
7084 case NEON::BI__builtin_neon_vdot_f16_mf8_fpm:
7085 case NEON::BI__builtin_neon_vdotq_f16_mf8_fpm:
7088 case NEON::BI__builtin_neon_vdot_lane_f16_mf8_fpm:
7089 case NEON::BI__builtin_neon_vdotq_lane_f16_mf8_fpm:
7090 ExtendLaneArg =
true;
7092 case NEON::BI__builtin_neon_vdot_laneq_f16_mf8_fpm:
7093 case NEON::BI__builtin_neon_vdotq_laneq_f16_mf8_fpm:
7095 ExtendLaneArg,
HalfTy, Ops, E,
"fdot2_lane");
7096 case NEON::BI__builtin_neon_vdot_f32_mf8_fpm:
7097 case NEON::BI__builtin_neon_vdotq_f32_mf8_fpm:
7100 case NEON::BI__builtin_neon_vdot_lane_f32_mf8_fpm:
7101 case NEON::BI__builtin_neon_vdotq_lane_f32_mf8_fpm:
7102 ExtendLaneArg =
true;
7104 case NEON::BI__builtin_neon_vdot_laneq_f32_mf8_fpm:
7105 case NEON::BI__builtin_neon_vdotq_laneq_f32_mf8_fpm:
7107 ExtendLaneArg,
FloatTy, Ops, E,
"fdot4_lane");
7109 case NEON::BI__builtin_neon_vmlalbq_f16_mf8_fpm:
7111 {llvm::FixedVectorType::get(
HalfTy, 8)}, Ops, E,
7113 case NEON::BI__builtin_neon_vmlaltq_f16_mf8_fpm:
7115 {llvm::FixedVectorType::get(
HalfTy, 8)}, Ops, E,
7117 case NEON::BI__builtin_neon_vmlallbbq_f32_mf8_fpm:
7119 {llvm::FixedVectorType::get(
FloatTy, 4)}, Ops, E,
7121 case NEON::BI__builtin_neon_vmlallbtq_f32_mf8_fpm:
7123 {llvm::FixedVectorType::get(
FloatTy, 4)}, Ops, E,
7125 case NEON::BI__builtin_neon_vmlalltbq_f32_mf8_fpm:
7127 {llvm::FixedVectorType::get(
FloatTy, 4)}, Ops, E,
7129 case NEON::BI__builtin_neon_vmlallttq_f32_mf8_fpm:
7131 {llvm::FixedVectorType::get(
FloatTy, 4)}, Ops, E,
7133 case NEON::BI__builtin_neon_vmlalbq_lane_f16_mf8_fpm:
7134 ExtendLaneArg =
true;
7136 case NEON::BI__builtin_neon_vmlalbq_laneq_f16_mf8_fpm:
7138 ExtendLaneArg,
HalfTy, Ops, E,
"vmlal_lane");
7139 case NEON::BI__builtin_neon_vmlaltq_lane_f16_mf8_fpm:
7140 ExtendLaneArg =
true;
7142 case NEON::BI__builtin_neon_vmlaltq_laneq_f16_mf8_fpm:
7144 ExtendLaneArg,
HalfTy, Ops, E,
"vmlal_lane");
7145 case NEON::BI__builtin_neon_vmlallbbq_lane_f32_mf8_fpm:
7146 ExtendLaneArg =
true;
7148 case NEON::BI__builtin_neon_vmlallbbq_laneq_f32_mf8_fpm:
7150 ExtendLaneArg,
FloatTy, Ops, E,
"vmlall_lane");
7151 case NEON::BI__builtin_neon_vmlallbtq_lane_f32_mf8_fpm:
7152 ExtendLaneArg =
true;
7154 case NEON::BI__builtin_neon_vmlallbtq_laneq_f32_mf8_fpm:
7156 ExtendLaneArg,
FloatTy, Ops, E,
"vmlall_lane");
7157 case NEON::BI__builtin_neon_vmlalltbq_lane_f32_mf8_fpm:
7158 ExtendLaneArg =
true;
7160 case NEON::BI__builtin_neon_vmlalltbq_laneq_f32_mf8_fpm:
7162 ExtendLaneArg,
FloatTy, Ops, E,
"vmlall_lane");
7163 case NEON::BI__builtin_neon_vmlallttq_lane_f32_mf8_fpm:
7164 ExtendLaneArg =
true;
7166 case NEON::BI__builtin_neon_vmlallttq_laneq_f32_mf8_fpm:
7168 ExtendLaneArg,
FloatTy, Ops, E,
"vmlall_lane");
7169 case NEON::BI__builtin_neon_vamin_f16:
7170 case NEON::BI__builtin_neon_vaminq_f16:
7171 case NEON::BI__builtin_neon_vamin_f32:
7172 case NEON::BI__builtin_neon_vaminq_f32:
7173 case NEON::BI__builtin_neon_vaminq_f64: {
7174 Int = Intrinsic::aarch64_neon_famin;
7177 case NEON::BI__builtin_neon_vamax_f16:
7178 case NEON::BI__builtin_neon_vamaxq_f16:
7179 case NEON::BI__builtin_neon_vamax_f32:
7180 case NEON::BI__builtin_neon_vamaxq_f32:
7181 case NEON::BI__builtin_neon_vamaxq_f64: {
7182 Int = Intrinsic::aarch64_neon_famax;
7185 case NEON::BI__builtin_neon_vscale_f16:
7186 case NEON::BI__builtin_neon_vscaleq_f16:
7187 case NEON::BI__builtin_neon_vscale_f32:
7188 case NEON::BI__builtin_neon_vscaleq_f32:
7189 case NEON::BI__builtin_neon_vscaleq_f64: {
7190 Int = Intrinsic::aarch64_neon_fp8_fscale;