795 if (BuiltinID == Builtin::BI__builtin_cpu_is)
796 return EmitX86CpuIs(E);
797 if (BuiltinID == Builtin::BI__builtin_cpu_supports)
798 return EmitX86CpuSupports(E);
799 if (BuiltinID == Builtin::BI__builtin_cpu_init)
800 return EmitX86CpuInit();
808 bool IsMaskFCmp =
false;
809 bool IsConjFMA =
false;
812 unsigned ICEArguments = 0;
817 for (
unsigned i = 0, e = E->
getNumArgs(); i != e; i++) {
827 auto getCmpIntrinsicCall = [
this, &Ops](Intrinsic::ID ID,
unsigned Imm) {
828 Ops.push_back(llvm::ConstantInt::get(
Int8Ty, Imm));
829 llvm::Function *F =
CGM.getIntrinsic(ID);
830 return Builder.CreateCall(F, Ops);
838 auto getVectorFCmpIR = [
this, &Ops, E](CmpInst::Predicate Pred,
843 Cmp =
Builder.CreateFCmpS(Pred, Ops[0], Ops[1]);
845 Cmp =
Builder.CreateFCmp(Pred, Ops[0], Ops[1]);
847 llvm::VectorType *IntVecTy = llvm::VectorType::getInteger(FPVecTy);
849 return Builder.CreateBitCast(Sext, FPVecTy);
853 default:
return nullptr;
854 case X86::BI_mm_prefetch: {
857 Value *RW = ConstantInt::get(
Int32Ty, (
C->getZExtValue() >> 2) & 0x1);
858 Value *Locality = ConstantInt::get(
Int32Ty,
C->getZExtValue() & 0x3);
863 case X86::BI_m_prefetch:
864 case X86::BI_m_prefetchw: {
868 ConstantInt::get(
Int32Ty, BuiltinID == X86::BI_m_prefetchw ? 1 : 0);
874 case X86::BI_mm_clflush: {
875 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse2_clflush),
878 case X86::BI_mm_lfence: {
879 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse2_lfence));
881 case X86::BI_mm_mfence: {
882 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse2_mfence));
884 case X86::BI_mm_sfence: {
885 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse_sfence));
887 case X86::BI_mm_pause: {
888 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse2_pause));
890 case X86::BI__rdtsc: {
891 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_rdtsc));
893 case X86::BI__builtin_ia32_rdtscp: {
899 case X86::BI__builtin_ia32_roundps:
900 case X86::BI__builtin_ia32_roundpd:
901 case X86::BI__builtin_ia32_roundps256:
902 case X86::BI__builtin_ia32_roundpd256: {
904 unsigned MXCSRMask = 0b100;
905 unsigned FRoundNoExcMask = 0b1000;
906 unsigned UseMXCSR = MXCSRMask & M;
907 unsigned FRoundNoExc = FRoundNoExcMask & M;
909 if (UseMXCSR || !FRoundNoExc) {
911 Intrinsic::ID ID = Intrinsic::not_intrinsic;
914 case X86::BI__builtin_ia32_roundps:
915 ID = Intrinsic::x86_sse41_round_ps;
917 case X86::BI__builtin_ia32_roundps256:
918 ID = Intrinsic::x86_avx_round_ps_256;
920 case X86::BI__builtin_ia32_roundpd:
921 ID = Intrinsic::x86_sse41_round_pd;
923 case X86::BI__builtin_ia32_roundpd256:
924 ID = Intrinsic::x86_avx_round_pd_256;
927 llvm_unreachable(
"must return from switch");
931 return Builder.CreateCall(F, Ops);
936 case X86::BI__builtin_ia32_roundss:
937 case X86::BI__builtin_ia32_roundsd: {
939 unsigned MXCSRMask = 0b100;
940 unsigned FRoundNoExcMask = 0b1000;
941 unsigned UseMXCSR = MXCSRMask & M;
942 unsigned FRoundNoExc = FRoundNoExcMask & M;
944 if (UseMXCSR || !FRoundNoExc) {
946 Intrinsic::ID ID = Intrinsic::not_intrinsic;
949 case X86::BI__builtin_ia32_roundss:
950 ID = Intrinsic::x86_sse41_round_ss;
952 case X86::BI__builtin_ia32_roundsd:
953 ID = Intrinsic::x86_sse41_round_sd;
956 llvm_unreachable(
"must return from switch");
960 return Builder.CreateCall(F, Ops);
964 Value *ValAt0 =
Builder.CreateExtractElement(Ops[1], Idx);
967 return Builder.CreateInsertElement(Ops[0], RoundedAt0, Idx);
969 case X86::BI__builtin_ia32_lzcnt_u16:
970 case X86::BI__builtin_ia32_lzcnt_u32:
971 case X86::BI__builtin_ia32_lzcnt_u64: {
975 case X86::BI__builtin_ia32_tzcnt_u16:
976 case X86::BI__builtin_ia32_tzcnt_u32:
977 case X86::BI__builtin_ia32_tzcnt_u64: {
981 case X86::BI__builtin_ia32_undef128:
982 case X86::BI__builtin_ia32_undef256:
983 case X86::BI__builtin_ia32_undef512:
990 case X86::BI__builtin_ia32_vec_ext_v4hi:
991 case X86::BI__builtin_ia32_vec_ext_v16qi:
992 case X86::BI__builtin_ia32_vec_ext_v8hi:
993 case X86::BI__builtin_ia32_vec_ext_v4si:
994 case X86::BI__builtin_ia32_vec_ext_v4sf:
995 case X86::BI__builtin_ia32_vec_ext_v2di:
996 case X86::BI__builtin_ia32_vec_ext_v32qi:
997 case X86::BI__builtin_ia32_vec_ext_v16hi:
998 case X86::BI__builtin_ia32_vec_ext_v8si:
999 case X86::BI__builtin_ia32_vec_ext_v4di: {
1003 Index &= NumElts - 1;
1006 return Builder.CreateExtractElement(Ops[0], Index);
1008 case X86::BI__builtin_ia32_vec_set_v4hi:
1009 case X86::BI__builtin_ia32_vec_set_v16qi:
1010 case X86::BI__builtin_ia32_vec_set_v8hi:
1011 case X86::BI__builtin_ia32_vec_set_v4si:
1012 case X86::BI__builtin_ia32_vec_set_v2di:
1013 case X86::BI__builtin_ia32_vec_set_v32qi:
1014 case X86::BI__builtin_ia32_vec_set_v16hi:
1015 case X86::BI__builtin_ia32_vec_set_v8si:
1016 case X86::BI__builtin_ia32_vec_set_v4di: {
1020 Index &= NumElts - 1;
1023 return Builder.CreateInsertElement(Ops[0], Ops[1], Index);
1025 case X86::BI_mm_setcsr:
1026 case X86::BI__builtin_ia32_ldmxcsr: {
1028 Builder.CreateStore(Ops[0], Tmp);
1029 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse_ldmxcsr),
1032 case X86::BI_mm_getcsr:
1033 case X86::BI__builtin_ia32_stmxcsr: {
1035 Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_sse_stmxcsr),
1037 return Builder.CreateLoad(Tmp,
"stmxcsr");
1039 case X86::BI__builtin_ia32_xsave:
1040 case X86::BI__builtin_ia32_xsave64:
1041 case X86::BI__builtin_ia32_xrstor:
1042 case X86::BI__builtin_ia32_xrstor64:
1043 case X86::BI__builtin_ia32_xsaveopt:
1044 case X86::BI__builtin_ia32_xsaveopt64:
1045 case X86::BI__builtin_ia32_xrstors:
1046 case X86::BI__builtin_ia32_xrstors64:
1047 case X86::BI__builtin_ia32_xsavec:
1048 case X86::BI__builtin_ia32_xsavec64:
1049 case X86::BI__builtin_ia32_xsaves:
1050 case X86::BI__builtin_ia32_xsaves64:
1051 case X86::BI__builtin_ia32_xsetbv:
1052 case X86::BI_xsetbv: {
1054#define INTRINSIC_X86_XSAVE_ID(NAME) \
1055 case X86::BI__builtin_ia32_##NAME: \
1056 ID = Intrinsic::x86_##NAME; \
1058 switch (BuiltinID) {
1059 default: llvm_unreachable(
"Unsupported intrinsic!");
1073 case X86::BI_xsetbv:
1074 ID = Intrinsic::x86_xsetbv;
1077#undef INTRINSIC_X86_XSAVE_ID
1083 return Builder.CreateCall(
CGM.getIntrinsic(ID), Ops);
1085 case X86::BI__builtin_ia32_xgetbv:
1086 case X86::BI_xgetbv:
1087 return Builder.CreateCall(
CGM.getIntrinsic(Intrinsic::x86_xgetbv), Ops);
1088 case X86::BI__builtin_ia32_storedqudi128_mask:
1089 case X86::BI__builtin_ia32_storedqusi128_mask:
1090 case X86::BI__builtin_ia32_storedquhi128_mask:
1091 case X86::BI__builtin_ia32_storedquqi128_mask:
1092 case X86::BI__builtin_ia32_storeupd128_mask:
1093 case X86::BI__builtin_ia32_storeups128_mask:
1094 case X86::BI__builtin_ia32_storedqudi256_mask:
1095 case X86::BI__builtin_ia32_storedqusi256_mask:
1096 case X86::BI__builtin_ia32_storedquhi256_mask:
1097 case X86::BI__builtin_ia32_storedquqi256_mask:
1098 case X86::BI__builtin_ia32_storeupd256_mask:
1099 case X86::BI__builtin_ia32_storeups256_mask:
1100 case X86::BI__builtin_ia32_storedqudi512_mask:
1101 case X86::BI__builtin_ia32_storedqusi512_mask:
1102 case X86::BI__builtin_ia32_storedquhi512_mask:
1103 case X86::BI__builtin_ia32_storedquqi512_mask:
1104 case X86::BI__builtin_ia32_storeupd512_mask:
1105 case X86::BI__builtin_ia32_storeups512_mask:
1108 case X86::BI__builtin_ia32_storesbf16128_mask:
1109 case X86::BI__builtin_ia32_storesh128_mask:
1110 case X86::BI__builtin_ia32_storess128_mask:
1111 case X86::BI__builtin_ia32_storesd128_mask:
1114 case X86::BI__builtin_ia32_cvtmask2b128:
1115 case X86::BI__builtin_ia32_cvtmask2b256:
1116 case X86::BI__builtin_ia32_cvtmask2b512:
1117 case X86::BI__builtin_ia32_cvtmask2w128:
1118 case X86::BI__builtin_ia32_cvtmask2w256:
1119 case X86::BI__builtin_ia32_cvtmask2w512:
1120 case X86::BI__builtin_ia32_cvtmask2d128:
1121 case X86::BI__builtin_ia32_cvtmask2d256:
1122 case X86::BI__builtin_ia32_cvtmask2d512:
1123 case X86::BI__builtin_ia32_cvtmask2q128:
1124 case X86::BI__builtin_ia32_cvtmask2q256:
1125 case X86::BI__builtin_ia32_cvtmask2q512:
1128 case X86::BI__builtin_ia32_cvtb2mask128:
1129 case X86::BI__builtin_ia32_cvtb2mask256:
1130 case X86::BI__builtin_ia32_cvtb2mask512:
1131 case X86::BI__builtin_ia32_cvtw2mask128:
1132 case X86::BI__builtin_ia32_cvtw2mask256:
1133 case X86::BI__builtin_ia32_cvtw2mask512:
1134 case X86::BI__builtin_ia32_cvtd2mask128:
1135 case X86::BI__builtin_ia32_cvtd2mask256:
1136 case X86::BI__builtin_ia32_cvtd2mask512:
1137 case X86::BI__builtin_ia32_cvtq2mask128:
1138 case X86::BI__builtin_ia32_cvtq2mask256:
1139 case X86::BI__builtin_ia32_cvtq2mask512:
1142 case X86::BI__builtin_ia32_cvtdq2ps512_mask:
1143 case X86::BI__builtin_ia32_cvtqq2ps512_mask:
1144 case X86::BI__builtin_ia32_cvtqq2pd512_mask:
1145 case X86::BI__builtin_ia32_vcvtw2ph512_mask:
1146 case X86::BI__builtin_ia32_vcvtdq2ph512_mask:
1147 case X86::BI__builtin_ia32_vcvtqq2ph512_mask:
1149 case X86::BI__builtin_ia32_cvtudq2ps512_mask:
1150 case X86::BI__builtin_ia32_cvtuqq2ps512_mask:
1151 case X86::BI__builtin_ia32_cvtuqq2pd512_mask:
1152 case X86::BI__builtin_ia32_vcvtuw2ph512_mask:
1153 case X86::BI__builtin_ia32_vcvtudq2ph512_mask:
1154 case X86::BI__builtin_ia32_vcvtuqq2ph512_mask:
1157 case X86::BI__builtin_ia32_vfmaddsh3_mask:
1158 case X86::BI__builtin_ia32_vfmaddss3_mask:
1159 case X86::BI__builtin_ia32_vfmaddsd3_mask:
1161 case X86::BI__builtin_ia32_vfmaddsh3_maskz:
1162 case X86::BI__builtin_ia32_vfmaddss3_maskz:
1163 case X86::BI__builtin_ia32_vfmaddsd3_maskz:
1165 case X86::BI__builtin_ia32_vfmaddsh3_mask3:
1166 case X86::BI__builtin_ia32_vfmaddss3_mask3:
1167 case X86::BI__builtin_ia32_vfmaddsd3_mask3:
1169 case X86::BI__builtin_ia32_vfmsubsh3_mask3:
1170 case X86::BI__builtin_ia32_vfmsubss3_mask3:
1171 case X86::BI__builtin_ia32_vfmsubsd3_mask3:
1174 case X86::BI__builtin_ia32_vfmaddph512_mask:
1175 case X86::BI__builtin_ia32_vfmaddph512_maskz:
1176 case X86::BI__builtin_ia32_vfmaddph512_mask3:
1177 case X86::BI__builtin_ia32_vfmaddps512_mask:
1178 case X86::BI__builtin_ia32_vfmaddps512_maskz:
1179 case X86::BI__builtin_ia32_vfmaddps512_mask3:
1180 case X86::BI__builtin_ia32_vfmsubps512_mask3:
1181 case X86::BI__builtin_ia32_vfmaddpd512_mask:
1182 case X86::BI__builtin_ia32_vfmaddpd512_maskz:
1183 case X86::BI__builtin_ia32_vfmaddpd512_mask3:
1184 case X86::BI__builtin_ia32_vfmsubpd512_mask3:
1185 case X86::BI__builtin_ia32_vfmsubph512_mask3:
1187 case X86::BI__builtin_ia32_vfmaddsubph512_mask:
1188 case X86::BI__builtin_ia32_vfmaddsubph512_maskz:
1189 case X86::BI__builtin_ia32_vfmaddsubph512_mask3:
1190 case X86::BI__builtin_ia32_vfmsubaddph512_mask3:
1191 case X86::BI__builtin_ia32_vfmaddsubps512_mask:
1192 case X86::BI__builtin_ia32_vfmaddsubps512_maskz:
1193 case X86::BI__builtin_ia32_vfmaddsubps512_mask3:
1194 case X86::BI__builtin_ia32_vfmsubaddps512_mask3:
1195 case X86::BI__builtin_ia32_vfmaddsubpd512_mask:
1196 case X86::BI__builtin_ia32_vfmaddsubpd512_maskz:
1197 case X86::BI__builtin_ia32_vfmaddsubpd512_mask3:
1198 case X86::BI__builtin_ia32_vfmsubaddpd512_mask3:
1201 case X86::BI__builtin_ia32_movdqa32store128_mask:
1202 case X86::BI__builtin_ia32_movdqa64store128_mask:
1203 case X86::BI__builtin_ia32_storeaps128_mask:
1204 case X86::BI__builtin_ia32_storeapd128_mask:
1205 case X86::BI__builtin_ia32_movdqa32store256_mask:
1206 case X86::BI__builtin_ia32_movdqa64store256_mask:
1207 case X86::BI__builtin_ia32_storeaps256_mask:
1208 case X86::BI__builtin_ia32_storeapd256_mask:
1209 case X86::BI__builtin_ia32_movdqa32store512_mask:
1210 case X86::BI__builtin_ia32_movdqa64store512_mask:
1211 case X86::BI__builtin_ia32_storeaps512_mask:
1212 case X86::BI__builtin_ia32_storeapd512_mask:
1217 case X86::BI__builtin_ia32_loadups128_mask:
1218 case X86::BI__builtin_ia32_loadups256_mask:
1219 case X86::BI__builtin_ia32_loadups512_mask:
1220 case X86::BI__builtin_ia32_loadupd128_mask:
1221 case X86::BI__builtin_ia32_loadupd256_mask:
1222 case X86::BI__builtin_ia32_loadupd512_mask:
1223 case X86::BI__builtin_ia32_loaddquqi128_mask:
1224 case X86::BI__builtin_ia32_loaddquqi256_mask:
1225 case X86::BI__builtin_ia32_loaddquqi512_mask:
1226 case X86::BI__builtin_ia32_loaddquhi128_mask:
1227 case X86::BI__builtin_ia32_loaddquhi256_mask:
1228 case X86::BI__builtin_ia32_loaddquhi512_mask:
1229 case X86::BI__builtin_ia32_loaddqusi128_mask:
1230 case X86::BI__builtin_ia32_loaddqusi256_mask:
1231 case X86::BI__builtin_ia32_loaddqusi512_mask:
1232 case X86::BI__builtin_ia32_loaddqudi128_mask:
1233 case X86::BI__builtin_ia32_loaddqudi256_mask:
1234 case X86::BI__builtin_ia32_loaddqudi512_mask:
1237 case X86::BI__builtin_ia32_loadsbf16128_mask:
1238 case X86::BI__builtin_ia32_loadsh128_mask:
1239 case X86::BI__builtin_ia32_loadss128_mask:
1240 case X86::BI__builtin_ia32_loadsd128_mask:
1243 case X86::BI__builtin_ia32_loadaps128_mask:
1244 case X86::BI__builtin_ia32_loadaps256_mask:
1245 case X86::BI__builtin_ia32_loadaps512_mask:
1246 case X86::BI__builtin_ia32_loadapd128_mask:
1247 case X86::BI__builtin_ia32_loadapd256_mask:
1248 case X86::BI__builtin_ia32_loadapd512_mask:
1249 case X86::BI__builtin_ia32_movdqa32load128_mask:
1250 case X86::BI__builtin_ia32_movdqa32load256_mask:
1251 case X86::BI__builtin_ia32_movdqa32load512_mask:
1252 case X86::BI__builtin_ia32_movdqa64load128_mask:
1253 case X86::BI__builtin_ia32_movdqa64load256_mask:
1254 case X86::BI__builtin_ia32_movdqa64load512_mask:
1259 case X86::BI__builtin_ia32_expandloaddf128_mask:
1260 case X86::BI__builtin_ia32_expandloaddf256_mask:
1261 case X86::BI__builtin_ia32_expandloaddf512_mask:
1262 case X86::BI__builtin_ia32_expandloadsf128_mask:
1263 case X86::BI__builtin_ia32_expandloadsf256_mask:
1264 case X86::BI__builtin_ia32_expandloadsf512_mask:
1265 case X86::BI__builtin_ia32_expandloaddi128_mask:
1266 case X86::BI__builtin_ia32_expandloaddi256_mask:
1267 case X86::BI__builtin_ia32_expandloaddi512_mask:
1268 case X86::BI__builtin_ia32_expandloadsi128_mask:
1269 case X86::BI__builtin_ia32_expandloadsi256_mask:
1270 case X86::BI__builtin_ia32_expandloadsi512_mask:
1271 case X86::BI__builtin_ia32_expandloadhi128_mask:
1272 case X86::BI__builtin_ia32_expandloadhi256_mask:
1273 case X86::BI__builtin_ia32_expandloadhi512_mask:
1274 case X86::BI__builtin_ia32_expandloadqi128_mask:
1275 case X86::BI__builtin_ia32_expandloadqi256_mask:
1276 case X86::BI__builtin_ia32_expandloadqi512_mask:
1279 case X86::BI__builtin_ia32_compressstoredf128_mask:
1280 case X86::BI__builtin_ia32_compressstoredf256_mask:
1281 case X86::BI__builtin_ia32_compressstoredf512_mask:
1282 case X86::BI__builtin_ia32_compressstoresf128_mask:
1283 case X86::BI__builtin_ia32_compressstoresf256_mask:
1284 case X86::BI__builtin_ia32_compressstoresf512_mask:
1285 case X86::BI__builtin_ia32_compressstoredi128_mask:
1286 case X86::BI__builtin_ia32_compressstoredi256_mask:
1287 case X86::BI__builtin_ia32_compressstoredi512_mask:
1288 case X86::BI__builtin_ia32_compressstoresi128_mask:
1289 case X86::BI__builtin_ia32_compressstoresi256_mask:
1290 case X86::BI__builtin_ia32_compressstoresi512_mask:
1291 case X86::BI__builtin_ia32_compressstorehi128_mask:
1292 case X86::BI__builtin_ia32_compressstorehi256_mask:
1293 case X86::BI__builtin_ia32_compressstorehi512_mask:
1294 case X86::BI__builtin_ia32_compressstoreqi128_mask:
1295 case X86::BI__builtin_ia32_compressstoreqi256_mask:
1296 case X86::BI__builtin_ia32_compressstoreqi512_mask:
1299 case X86::BI__builtin_ia32_expanddf128_mask:
1300 case X86::BI__builtin_ia32_expanddf256_mask:
1301 case X86::BI__builtin_ia32_expanddf512_mask:
1302 case X86::BI__builtin_ia32_expandsf128_mask:
1303 case X86::BI__builtin_ia32_expandsf256_mask:
1304 case X86::BI__builtin_ia32_expandsf512_mask:
1305 case X86::BI__builtin_ia32_expanddi128_mask:
1306 case X86::BI__builtin_ia32_expanddi256_mask:
1307 case X86::BI__builtin_ia32_expanddi512_mask:
1308 case X86::BI__builtin_ia32_expandsi128_mask:
1309 case X86::BI__builtin_ia32_expandsi256_mask:
1310 case X86::BI__builtin_ia32_expandsi512_mask:
1311 case X86::BI__builtin_ia32_expandhi128_mask:
1312 case X86::BI__builtin_ia32_expandhi256_mask:
1313 case X86::BI__builtin_ia32_expandhi512_mask:
1314 case X86::BI__builtin_ia32_expandqi128_mask:
1315 case X86::BI__builtin_ia32_expandqi256_mask:
1316 case X86::BI__builtin_ia32_expandqi512_mask:
1319 case X86::BI__builtin_ia32_compressdf128_mask:
1320 case X86::BI__builtin_ia32_compressdf256_mask:
1321 case X86::BI__builtin_ia32_compressdf512_mask:
1322 case X86::BI__builtin_ia32_compresssf128_mask:
1323 case X86::BI__builtin_ia32_compresssf256_mask:
1324 case X86::BI__builtin_ia32_compresssf512_mask:
1325 case X86::BI__builtin_ia32_compressdi128_mask:
1326 case X86::BI__builtin_ia32_compressdi256_mask:
1327 case X86::BI__builtin_ia32_compressdi512_mask:
1328 case X86::BI__builtin_ia32_compresssi128_mask:
1329 case X86::BI__builtin_ia32_compresssi256_mask:
1330 case X86::BI__builtin_ia32_compresssi512_mask:
1331 case X86::BI__builtin_ia32_compresshi128_mask:
1332 case X86::BI__builtin_ia32_compresshi256_mask:
1333 case X86::BI__builtin_ia32_compresshi512_mask:
1334 case X86::BI__builtin_ia32_compressqi128_mask:
1335 case X86::BI__builtin_ia32_compressqi256_mask:
1336 case X86::BI__builtin_ia32_compressqi512_mask:
1339 case X86::BI__builtin_ia32_gather3div2df:
1340 case X86::BI__builtin_ia32_gather3div2di:
1341 case X86::BI__builtin_ia32_gather3div4df:
1342 case X86::BI__builtin_ia32_gather3div4di:
1343 case X86::BI__builtin_ia32_gather3div4sf:
1344 case X86::BI__builtin_ia32_gather3div4si:
1345 case X86::BI__builtin_ia32_gather3div8sf:
1346 case X86::BI__builtin_ia32_gather3div8si:
1347 case X86::BI__builtin_ia32_gather3siv2df:
1348 case X86::BI__builtin_ia32_gather3siv2di:
1349 case X86::BI__builtin_ia32_gather3siv4df:
1350 case X86::BI__builtin_ia32_gather3siv4di:
1351 case X86::BI__builtin_ia32_gather3siv4sf:
1352 case X86::BI__builtin_ia32_gather3siv4si:
1353 case X86::BI__builtin_ia32_gather3siv8sf:
1354 case X86::BI__builtin_ia32_gather3siv8si:
1355 case X86::BI__builtin_ia32_gathersiv8df:
1356 case X86::BI__builtin_ia32_gathersiv16sf:
1357 case X86::BI__builtin_ia32_gatherdiv8df:
1358 case X86::BI__builtin_ia32_gatherdiv16sf:
1359 case X86::BI__builtin_ia32_gathersiv8di:
1360 case X86::BI__builtin_ia32_gathersiv16si:
1361 case X86::BI__builtin_ia32_gatherdiv8di:
1362 case X86::BI__builtin_ia32_gatherdiv16si: {
1364 switch (BuiltinID) {
1365 default: llvm_unreachable(
"Unexpected builtin");
1366 case X86::BI__builtin_ia32_gather3div2df:
1367 IID = Intrinsic::x86_avx512_mask_gather3div2_df;
1369 case X86::BI__builtin_ia32_gather3div2di:
1370 IID = Intrinsic::x86_avx512_mask_gather3div2_di;
1372 case X86::BI__builtin_ia32_gather3div4df:
1373 IID = Intrinsic::x86_avx512_mask_gather3div4_df;
1375 case X86::BI__builtin_ia32_gather3div4di:
1376 IID = Intrinsic::x86_avx512_mask_gather3div4_di;
1378 case X86::BI__builtin_ia32_gather3div4sf:
1379 IID = Intrinsic::x86_avx512_mask_gather3div4_sf;
1381 case X86::BI__builtin_ia32_gather3div4si:
1382 IID = Intrinsic::x86_avx512_mask_gather3div4_si;
1384 case X86::BI__builtin_ia32_gather3div8sf:
1385 IID = Intrinsic::x86_avx512_mask_gather3div8_sf;
1387 case X86::BI__builtin_ia32_gather3div8si:
1388 IID = Intrinsic::x86_avx512_mask_gather3div8_si;
1390 case X86::BI__builtin_ia32_gather3siv2df:
1391 IID = Intrinsic::x86_avx512_mask_gather3siv2_df;
1393 case X86::BI__builtin_ia32_gather3siv2di:
1394 IID = Intrinsic::x86_avx512_mask_gather3siv2_di;
1396 case X86::BI__builtin_ia32_gather3siv4df:
1397 IID = Intrinsic::x86_avx512_mask_gather3siv4_df;
1399 case X86::BI__builtin_ia32_gather3siv4di:
1400 IID = Intrinsic::x86_avx512_mask_gather3siv4_di;
1402 case X86::BI__builtin_ia32_gather3siv4sf:
1403 IID = Intrinsic::x86_avx512_mask_gather3siv4_sf;
1405 case X86::BI__builtin_ia32_gather3siv4si:
1406 IID = Intrinsic::x86_avx512_mask_gather3siv4_si;
1408 case X86::BI__builtin_ia32_gather3siv8sf:
1409 IID = Intrinsic::x86_avx512_mask_gather3siv8_sf;
1411 case X86::BI__builtin_ia32_gather3siv8si:
1412 IID = Intrinsic::x86_avx512_mask_gather3siv8_si;
1414 case X86::BI__builtin_ia32_gathersiv8df:
1415 IID = Intrinsic::x86_avx512_mask_gather_dpd_512;
1417 case X86::BI__builtin_ia32_gathersiv16sf:
1418 IID = Intrinsic::x86_avx512_mask_gather_dps_512;
1420 case X86::BI__builtin_ia32_gatherdiv8df:
1421 IID = Intrinsic::x86_avx512_mask_gather_qpd_512;
1423 case X86::BI__builtin_ia32_gatherdiv16sf:
1424 IID = Intrinsic::x86_avx512_mask_gather_qps_512;
1426 case X86::BI__builtin_ia32_gathersiv8di:
1427 IID = Intrinsic::x86_avx512_mask_gather_dpq_512;
1429 case X86::BI__builtin_ia32_gathersiv16si:
1430 IID = Intrinsic::x86_avx512_mask_gather_dpi_512;
1432 case X86::BI__builtin_ia32_gatherdiv8di:
1433 IID = Intrinsic::x86_avx512_mask_gather_qpq_512;
1435 case X86::BI__builtin_ia32_gatherdiv16si:
1436 IID = Intrinsic::x86_avx512_mask_gather_qpi_512;
1440 unsigned MinElts = std::min(
1445 return Builder.CreateCall(Intr, Ops);
1448 case X86::BI__builtin_ia32_scattersiv8df:
1449 case X86::BI__builtin_ia32_scattersiv16sf:
1450 case X86::BI__builtin_ia32_scatterdiv8df:
1451 case X86::BI__builtin_ia32_scatterdiv16sf:
1452 case X86::BI__builtin_ia32_scattersiv8di:
1453 case X86::BI__builtin_ia32_scattersiv16si:
1454 case X86::BI__builtin_ia32_scatterdiv8di:
1455 case X86::BI__builtin_ia32_scatterdiv16si:
1456 case X86::BI__builtin_ia32_scatterdiv2df:
1457 case X86::BI__builtin_ia32_scatterdiv2di:
1458 case X86::BI__builtin_ia32_scatterdiv4df:
1459 case X86::BI__builtin_ia32_scatterdiv4di:
1460 case X86::BI__builtin_ia32_scatterdiv4sf:
1461 case X86::BI__builtin_ia32_scatterdiv4si:
1462 case X86::BI__builtin_ia32_scatterdiv8sf:
1463 case X86::BI__builtin_ia32_scatterdiv8si:
1464 case X86::BI__builtin_ia32_scattersiv2df:
1465 case X86::BI__builtin_ia32_scattersiv2di:
1466 case X86::BI__builtin_ia32_scattersiv4df:
1467 case X86::BI__builtin_ia32_scattersiv4di:
1468 case X86::BI__builtin_ia32_scattersiv4sf:
1469 case X86::BI__builtin_ia32_scattersiv4si:
1470 case X86::BI__builtin_ia32_scattersiv8sf:
1471 case X86::BI__builtin_ia32_scattersiv8si: {
1473 switch (BuiltinID) {
1474 default: llvm_unreachable(
"Unexpected builtin");
1475 case X86::BI__builtin_ia32_scattersiv8df:
1476 IID = Intrinsic::x86_avx512_mask_scatter_dpd_512;
1478 case X86::BI__builtin_ia32_scattersiv16sf:
1479 IID = Intrinsic::x86_avx512_mask_scatter_dps_512;
1481 case X86::BI__builtin_ia32_scatterdiv8df:
1482 IID = Intrinsic::x86_avx512_mask_scatter_qpd_512;
1484 case X86::BI__builtin_ia32_scatterdiv16sf:
1485 IID = Intrinsic::x86_avx512_mask_scatter_qps_512;
1487 case X86::BI__builtin_ia32_scattersiv8di:
1488 IID = Intrinsic::x86_avx512_mask_scatter_dpq_512;
1490 case X86::BI__builtin_ia32_scattersiv16si:
1491 IID = Intrinsic::x86_avx512_mask_scatter_dpi_512;
1493 case X86::BI__builtin_ia32_scatterdiv8di:
1494 IID = Intrinsic::x86_avx512_mask_scatter_qpq_512;
1496 case X86::BI__builtin_ia32_scatterdiv16si:
1497 IID = Intrinsic::x86_avx512_mask_scatter_qpi_512;
1499 case X86::BI__builtin_ia32_scatterdiv2df:
1500 IID = Intrinsic::x86_avx512_mask_scatterdiv2_df;
1502 case X86::BI__builtin_ia32_scatterdiv2di:
1503 IID = Intrinsic::x86_avx512_mask_scatterdiv2_di;
1505 case X86::BI__builtin_ia32_scatterdiv4df:
1506 IID = Intrinsic::x86_avx512_mask_scatterdiv4_df;
1508 case X86::BI__builtin_ia32_scatterdiv4di:
1509 IID = Intrinsic::x86_avx512_mask_scatterdiv4_di;
1511 case X86::BI__builtin_ia32_scatterdiv4sf:
1512 IID = Intrinsic::x86_avx512_mask_scatterdiv4_sf;
1514 case X86::BI__builtin_ia32_scatterdiv4si:
1515 IID = Intrinsic::x86_avx512_mask_scatterdiv4_si;
1517 case X86::BI__builtin_ia32_scatterdiv8sf:
1518 IID = Intrinsic::x86_avx512_mask_scatterdiv8_sf;
1520 case X86::BI__builtin_ia32_scatterdiv8si:
1521 IID = Intrinsic::x86_avx512_mask_scatterdiv8_si;
1523 case X86::BI__builtin_ia32_scattersiv2df:
1524 IID = Intrinsic::x86_avx512_mask_scattersiv2_df;
1526 case X86::BI__builtin_ia32_scattersiv2di:
1527 IID = Intrinsic::x86_avx512_mask_scattersiv2_di;
1529 case X86::BI__builtin_ia32_scattersiv4df:
1530 IID = Intrinsic::x86_avx512_mask_scattersiv4_df;
1532 case X86::BI__builtin_ia32_scattersiv4di:
1533 IID = Intrinsic::x86_avx512_mask_scattersiv4_di;
1535 case X86::BI__builtin_ia32_scattersiv4sf:
1536 IID = Intrinsic::x86_avx512_mask_scattersiv4_sf;
1538 case X86::BI__builtin_ia32_scattersiv4si:
1539 IID = Intrinsic::x86_avx512_mask_scattersiv4_si;
1541 case X86::BI__builtin_ia32_scattersiv8sf:
1542 IID = Intrinsic::x86_avx512_mask_scattersiv8_sf;
1544 case X86::BI__builtin_ia32_scattersiv8si:
1545 IID = Intrinsic::x86_avx512_mask_scattersiv8_si;
1549 unsigned MinElts = std::min(
1554 return Builder.CreateCall(Intr, Ops);
1557 case X86::BI__builtin_ia32_vextractf128_pd256:
1558 case X86::BI__builtin_ia32_vextractf128_ps256:
1559 case X86::BI__builtin_ia32_vextractf128_si256:
1560 case X86::BI__builtin_ia32_extract128i256:
1561 case X86::BI__builtin_ia32_extractf64x4_mask:
1562 case X86::BI__builtin_ia32_extractf32x4_mask:
1563 case X86::BI__builtin_ia32_extracti64x4_mask:
1564 case X86::BI__builtin_ia32_extracti32x4_mask:
1565 case X86::BI__builtin_ia32_extractf32x8_mask:
1566 case X86::BI__builtin_ia32_extracti32x8_mask:
1567 case X86::BI__builtin_ia32_extractf32x4_256_mask:
1568 case X86::BI__builtin_ia32_extracti32x4_256_mask:
1569 case X86::BI__builtin_ia32_extractf64x2_256_mask:
1570 case X86::BI__builtin_ia32_extracti64x2_256_mask:
1571 case X86::BI__builtin_ia32_extractf64x2_512_mask:
1572 case X86::BI__builtin_ia32_extracti64x2_512_mask: {
1574 unsigned NumElts = DstTy->getNumElements();
1575 unsigned SrcNumElts =
1577 unsigned SubVectors = SrcNumElts / NumElts;
1579 assert(llvm::isPowerOf2_32(SubVectors) &&
"Expected power of 2 subvectors");
1580 Index &= SubVectors - 1;
1584 for (
unsigned i = 0; i != NumElts; ++i)
1585 Indices[i] = i + Index;
1590 if (Ops.size() == 4)
1595 case X86::BI__builtin_ia32_vinsertf128_pd256:
1596 case X86::BI__builtin_ia32_vinsertf128_ps256:
1597 case X86::BI__builtin_ia32_vinsertf128_si256:
1598 case X86::BI__builtin_ia32_insert128i256:
1599 case X86::BI__builtin_ia32_insertf64x4:
1600 case X86::BI__builtin_ia32_insertf32x4:
1601 case X86::BI__builtin_ia32_inserti64x4:
1602 case X86::BI__builtin_ia32_inserti32x4:
1603 case X86::BI__builtin_ia32_insertf32x8:
1604 case X86::BI__builtin_ia32_inserti32x8:
1605 case X86::BI__builtin_ia32_insertf32x4_256:
1606 case X86::BI__builtin_ia32_inserti32x4_256:
1607 case X86::BI__builtin_ia32_insertf64x2_256:
1608 case X86::BI__builtin_ia32_inserti64x2_256:
1609 case X86::BI__builtin_ia32_insertf64x2_512:
1610 case X86::BI__builtin_ia32_inserti64x2_512: {
1611 unsigned DstNumElts =
1613 unsigned SrcNumElts =
1615 unsigned SubVectors = DstNumElts / SrcNumElts;
1617 assert(llvm::isPowerOf2_32(SubVectors) &&
"Expected power of 2 subvectors");
1618 Index &= SubVectors - 1;
1619 Index *= SrcNumElts;
1622 for (
unsigned i = 0; i != DstNumElts; ++i)
1623 Indices[i] = (i >= SrcNumElts) ? SrcNumElts + (i % SrcNumElts) : i;
1626 Ops[1],
ArrayRef(Indices, DstNumElts),
"widen");
1628 for (
unsigned i = 0; i != DstNumElts; ++i) {
1629 if (i >= Index && i < (Index + SrcNumElts))
1630 Indices[i] = (i - Index) + DstNumElts;
1635 return Builder.CreateShuffleVector(Ops[0], Op1,
1636 ArrayRef(Indices, DstNumElts),
"insert");
1638 case X86::BI__builtin_ia32_pmovqd512_mask:
1639 case X86::BI__builtin_ia32_pmovwb512_mask: {
1643 case X86::BI__builtin_ia32_pmovdb512_mask:
1644 case X86::BI__builtin_ia32_pmovdw512_mask:
1645 case X86::BI__builtin_ia32_pmovqw512_mask: {
1646 if (
const auto *
C = dyn_cast<Constant>(Ops[2]))
1647 if (
C->isAllOnesValue())
1651 switch (BuiltinID) {
1652 default: llvm_unreachable(
"Unsupported intrinsic!");
1653 case X86::BI__builtin_ia32_pmovdb512_mask:
1654 IID = Intrinsic::x86_avx512_mask_pmov_db_512;
1656 case X86::BI__builtin_ia32_pmovdw512_mask:
1657 IID = Intrinsic::x86_avx512_mask_pmov_dw_512;
1659 case X86::BI__builtin_ia32_pmovqw512_mask:
1660 IID = Intrinsic::x86_avx512_mask_pmov_qw_512;
1665 return Builder.CreateCall(Intr, Ops);
1667 case X86::BI__builtin_ia32_pblendw128:
1668 case X86::BI__builtin_ia32_blendpd:
1669 case X86::BI__builtin_ia32_blendps:
1670 case X86::BI__builtin_ia32_blendpd256:
1671 case X86::BI__builtin_ia32_blendps256:
1672 case X86::BI__builtin_ia32_pblendw256:
1673 case X86::BI__builtin_ia32_pblendd128:
1674 case X86::BI__builtin_ia32_pblendd256: {
1682 for (
unsigned i = 0; i != NumElts; ++i)
1683 Indices[i] = ((Imm >> (i % 8)) & 0x1) ? NumElts + i : i;
1685 return Builder.CreateShuffleVector(Ops[0], Ops[1],
1686 ArrayRef(Indices, NumElts),
"blend");
1688 case X86::BI__builtin_ia32_pshuflw:
1689 case X86::BI__builtin_ia32_pshuflw256:
1690 case X86::BI__builtin_ia32_pshuflw512: {
1693 unsigned NumElts = Ty->getNumElements();
1696 Imm = (Imm & 0xff) * 0x01010101;
1699 for (
unsigned l = 0; l != NumElts; l += 8) {
1700 for (
unsigned i = 0; i != 4; ++i) {
1701 Indices[l + i] = l + (Imm & 3);
1704 for (
unsigned i = 4; i != 8; ++i)
1705 Indices[l + i] = l + i;
1708 return Builder.CreateShuffleVector(Ops[0],
ArrayRef(Indices, NumElts),
1711 case X86::BI__builtin_ia32_pshufhw:
1712 case X86::BI__builtin_ia32_pshufhw256:
1713 case X86::BI__builtin_ia32_pshufhw512: {
1716 unsigned NumElts = Ty->getNumElements();
1719 Imm = (Imm & 0xff) * 0x01010101;
1722 for (
unsigned l = 0; l != NumElts; l += 8) {
1723 for (
unsigned i = 0; i != 4; ++i)
1724 Indices[l + i] = l + i;
1725 for (
unsigned i = 4; i != 8; ++i) {
1726 Indices[l + i] = l + 4 + (Imm & 3);
1731 return Builder.CreateShuffleVector(Ops[0],
ArrayRef(Indices, NumElts),
1734 case X86::BI__builtin_ia32_pshufd:
1735 case X86::BI__builtin_ia32_pshufd256:
1736 case X86::BI__builtin_ia32_pshufd512:
1737 case X86::BI__builtin_ia32_vpermilpd:
1738 case X86::BI__builtin_ia32_vpermilps:
1739 case X86::BI__builtin_ia32_vpermilpd256:
1740 case X86::BI__builtin_ia32_vpermilps256:
1741 case X86::BI__builtin_ia32_vpermilpd512:
1742 case X86::BI__builtin_ia32_vpermilps512: {
1745 unsigned NumElts = Ty->getNumElements();
1746 unsigned NumLanes = Ty->getPrimitiveSizeInBits() / 128;
1747 unsigned NumLaneElts = NumElts / NumLanes;
1750 Imm = (Imm & 0xff) * 0x01010101;
1753 for (
unsigned l = 0; l != NumElts; l += NumLaneElts) {
1754 for (
unsigned i = 0; i != NumLaneElts; ++i) {
1755 Indices[i + l] = (Imm % NumLaneElts) + l;
1760 return Builder.CreateShuffleVector(Ops[0],
ArrayRef(Indices, NumElts),
1763 case X86::BI__builtin_ia32_shufpd:
1764 case X86::BI__builtin_ia32_shufpd256:
1765 case X86::BI__builtin_ia32_shufpd512:
1766 case X86::BI__builtin_ia32_shufps:
1767 case X86::BI__builtin_ia32_shufps256:
1768 case X86::BI__builtin_ia32_shufps512: {
1771 unsigned NumElts = Ty->getNumElements();
1772 unsigned NumLanes = Ty->getPrimitiveSizeInBits() / 128;
1773 unsigned NumLaneElts = NumElts / NumLanes;
1776 Imm = (Imm & 0xff) * 0x01010101;
1779 for (
unsigned l = 0; l != NumElts; l += NumLaneElts) {
1780 for (
unsigned i = 0; i != NumLaneElts; ++i) {
1781 unsigned Index = Imm % NumLaneElts;
1783 if (i >= (NumLaneElts / 2))
1785 Indices[l + i] = l + Index;
1789 return Builder.CreateShuffleVector(Ops[0], Ops[1],
1790 ArrayRef(Indices, NumElts),
"shufp");
1792 case X86::BI__builtin_ia32_permdi256:
1793 case X86::BI__builtin_ia32_permdf256:
1794 case X86::BI__builtin_ia32_permdi512:
1795 case X86::BI__builtin_ia32_permdf512: {
1798 unsigned NumElts = Ty->getNumElements();
1802 for (
unsigned l = 0; l != NumElts; l += 4)
1803 for (
unsigned i = 0; i != 4; ++i)
1804 Indices[l + i] = l + ((Imm >> (2 * i)) & 0x3);
1806 return Builder.CreateShuffleVector(Ops[0],
ArrayRef(Indices, NumElts),
1809 case X86::BI__builtin_ia32_palignr128:
1810 case X86::BI__builtin_ia32_palignr256:
1811 case X86::BI__builtin_ia32_palignr512: {
1816 assert(NumElts % 16 == 0);
1825 if (ShiftVal > 16) {
1828 Ops[0] = llvm::Constant::getNullValue(Ops[0]->
getType());
1833 for (
unsigned l = 0; l != NumElts; l += 16) {
1834 for (
unsigned i = 0; i != 16; ++i) {
1835 unsigned Idx = ShiftVal + i;
1837 Idx += NumElts - 16;
1838 Indices[l + i] = Idx + l;
1842 return Builder.CreateShuffleVector(Ops[1], Ops[0],
1843 ArrayRef(Indices, NumElts),
"palignr");
1845 case X86::BI__builtin_ia32_alignd128:
1846 case X86::BI__builtin_ia32_alignd256:
1847 case X86::BI__builtin_ia32_alignd512:
1848 case X86::BI__builtin_ia32_alignq128:
1849 case X86::BI__builtin_ia32_alignq256:
1850 case X86::BI__builtin_ia32_alignq512: {
1856 ShiftVal &= NumElts - 1;
1859 for (
unsigned i = 0; i != NumElts; ++i)
1860 Indices[i] = i + ShiftVal;
1862 return Builder.CreateShuffleVector(Ops[1], Ops[0],
1863 ArrayRef(Indices, NumElts),
"valign");
1865 case X86::BI__builtin_ia32_shuf_f32x4_256:
1866 case X86::BI__builtin_ia32_shuf_f64x2_256:
1867 case X86::BI__builtin_ia32_shuf_i32x4_256:
1868 case X86::BI__builtin_ia32_shuf_i64x2_256:
1869 case X86::BI__builtin_ia32_shuf_f32x4:
1870 case X86::BI__builtin_ia32_shuf_f64x2:
1871 case X86::BI__builtin_ia32_shuf_i32x4:
1872 case X86::BI__builtin_ia32_shuf_i64x2: {
1875 unsigned NumElts = Ty->getNumElements();
1876 unsigned NumLanes = Ty->getPrimitiveSizeInBits() == 512 ? 4 : 2;
1877 unsigned NumLaneElts = NumElts / NumLanes;
1880 for (
unsigned l = 0; l != NumElts; l += NumLaneElts) {
1881 unsigned Index = (Imm % NumLanes) * NumLaneElts;
1883 if (l >= (NumElts / 2))
1885 for (
unsigned i = 0; i != NumLaneElts; ++i) {
1886 Indices[l + i] = Index + i;
1890 return Builder.CreateShuffleVector(Ops[0], Ops[1],
1891 ArrayRef(Indices, NumElts),
"shuf");
1894 case X86::BI__builtin_ia32_vperm2f128_pd256:
1895 case X86::BI__builtin_ia32_vperm2f128_ps256:
1896 case X86::BI__builtin_ia32_vperm2f128_si256:
1897 case X86::BI__builtin_ia32_permti256: {
1909 for (
unsigned l = 0; l != 2; ++l) {
1911 if (Imm & (1 << ((l * 4) + 3)))
1912 OutOps[l] = llvm::ConstantAggregateZero::get(Ops[0]->
getType());
1913 else if (Imm & (1 << ((l * 4) + 1)))
1918 for (
unsigned i = 0; i != NumElts/2; ++i) {
1920 unsigned Idx = (l * NumElts) + i;
1923 if (Imm & (1 << (l * 4)))
1925 Indices[(l * (NumElts/2)) + i] = Idx;
1929 return Builder.CreateShuffleVector(OutOps[0], OutOps[1],
1930 ArrayRef(Indices, NumElts),
"vperm");
1933 case X86::BI__builtin_ia32_pslldqi128_byteshift:
1934 case X86::BI__builtin_ia32_pslldqi256_byteshift:
1935 case X86::BI__builtin_ia32_pslldqi512_byteshift: {
1939 unsigned NumElts = VecTy->getNumElements();
1940 Value *
Zero = llvm::Constant::getNullValue(VecTy);
1948 for (
unsigned l = 0; l != NumElts; l += 16) {
1949 for (
unsigned i = 0; i != 16; ++i) {
1950 unsigned Idx = NumElts + i - ShiftVal;
1952 Idx -= NumElts - 16;
1953 Indices[l + i] = Idx + l;
1959 case X86::BI__builtin_ia32_psrldqi128_byteshift:
1960 case X86::BI__builtin_ia32_psrldqi256_byteshift:
1961 case X86::BI__builtin_ia32_psrldqi512_byteshift: {
1965 unsigned NumElts = VecTy->getNumElements();
1966 Value *
Zero = llvm::Constant::getNullValue(VecTy);
1974 for (
unsigned l = 0; l != NumElts; l += 16) {
1975 for (
unsigned i = 0; i != 16; ++i) {
1976 unsigned Idx = i + ShiftVal;
1978 Idx += NumElts - 16;
1979 Indices[l + i] = Idx + l;
1985 case X86::BI__builtin_ia32_kshiftliqi:
1986 case X86::BI__builtin_ia32_kshiftlihi:
1987 case X86::BI__builtin_ia32_kshiftlisi:
1988 case X86::BI__builtin_ia32_kshiftlidi: {
1990 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
1992 if (ShiftVal >= NumElts)
1993 return llvm::Constant::getNullValue(Ops[0]->
getType());
1998 for (
unsigned i = 0; i != NumElts; ++i)
1999 Indices[i] = NumElts + i - ShiftVal;
2001 Value *
Zero = llvm::Constant::getNullValue(In->getType());
2006 case X86::BI__builtin_ia32_kshiftriqi:
2007 case X86::BI__builtin_ia32_kshiftrihi:
2008 case X86::BI__builtin_ia32_kshiftrisi:
2009 case X86::BI__builtin_ia32_kshiftridi: {
2011 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
2013 if (ShiftVal >= NumElts)
2014 return llvm::Constant::getNullValue(Ops[0]->
getType());
2019 for (
unsigned i = 0; i != NumElts; ++i)
2020 Indices[i] = i + ShiftVal;
2022 Value *
Zero = llvm::Constant::getNullValue(In->getType());
2027 case X86::BI__builtin_ia32_movnti:
2028 case X86::BI__builtin_ia32_movnti64:
2029 case X86::BI__builtin_ia32_movntsd:
2030 case X86::BI__builtin_ia32_movntss: {
2031 llvm::MDNode *Node = llvm::MDNode::get(
2034 Value *Ptr = Ops[0];
2035 Value *Src = Ops[1];
2038 if (BuiltinID == X86::BI__builtin_ia32_movntsd ||
2039 BuiltinID == X86::BI__builtin_ia32_movntss)
2040 Src =
Builder.CreateExtractElement(Src, (uint64_t)0,
"extract");
2043 StoreInst *SI =
Builder.CreateDefaultAlignedStore(Src, Ptr);
2044 SI->setMetadata(llvm::LLVMContext::MD_nontemporal, Node);
2045 SI->setAlignment(llvm::Align(1));
2049 case X86::BI__builtin_ia32_vprotbi:
2050 case X86::BI__builtin_ia32_vprotwi:
2051 case X86::BI__builtin_ia32_vprotdi:
2052 case X86::BI__builtin_ia32_vprotqi:
2053 case X86::BI__builtin_ia32_prold128:
2054 case X86::BI__builtin_ia32_prold256:
2055 case X86::BI__builtin_ia32_prold512:
2056 case X86::BI__builtin_ia32_prolq128:
2057 case X86::BI__builtin_ia32_prolq256:
2058 case X86::BI__builtin_ia32_prolq512:
2060 case X86::BI__builtin_ia32_prord128:
2061 case X86::BI__builtin_ia32_prord256:
2062 case X86::BI__builtin_ia32_prord512:
2063 case X86::BI__builtin_ia32_prorq128:
2064 case X86::BI__builtin_ia32_prorq256:
2065 case X86::BI__builtin_ia32_prorq512:
2067 case X86::BI__builtin_ia32_selectb_128:
2068 case X86::BI__builtin_ia32_selectb_256:
2069 case X86::BI__builtin_ia32_selectb_512:
2070 case X86::BI__builtin_ia32_selectw_128:
2071 case X86::BI__builtin_ia32_selectw_256:
2072 case X86::BI__builtin_ia32_selectw_512:
2073 case X86::BI__builtin_ia32_selectd_128:
2074 case X86::BI__builtin_ia32_selectd_256:
2075 case X86::BI__builtin_ia32_selectd_512:
2076 case X86::BI__builtin_ia32_selectq_128:
2077 case X86::BI__builtin_ia32_selectq_256:
2078 case X86::BI__builtin_ia32_selectq_512:
2079 case X86::BI__builtin_ia32_selectph_128:
2080 case X86::BI__builtin_ia32_selectph_256:
2081 case X86::BI__builtin_ia32_selectph_512:
2082 case X86::BI__builtin_ia32_selectpbf_128:
2083 case X86::BI__builtin_ia32_selectpbf_256:
2084 case X86::BI__builtin_ia32_selectpbf_512:
2085 case X86::BI__builtin_ia32_selectps_128:
2086 case X86::BI__builtin_ia32_selectps_256:
2087 case X86::BI__builtin_ia32_selectps_512:
2088 case X86::BI__builtin_ia32_selectpd_128:
2089 case X86::BI__builtin_ia32_selectpd_256:
2090 case X86::BI__builtin_ia32_selectpd_512:
2092 case X86::BI__builtin_ia32_selectsh_128:
2093 case X86::BI__builtin_ia32_selectsbf_128:
2094 case X86::BI__builtin_ia32_selectss_128:
2095 case X86::BI__builtin_ia32_selectsd_128: {
2096 Value *A =
Builder.CreateExtractElement(Ops[1], (uint64_t)0);
2097 Value *B =
Builder.CreateExtractElement(Ops[2], (uint64_t)0);
2099 return Builder.CreateInsertElement(Ops[1], A, (uint64_t)0);
2101 case X86::BI__builtin_ia32_cmpb128_mask:
2102 case X86::BI__builtin_ia32_cmpb256_mask:
2103 case X86::BI__builtin_ia32_cmpb512_mask:
2104 case X86::BI__builtin_ia32_cmpw128_mask:
2105 case X86::BI__builtin_ia32_cmpw256_mask:
2106 case X86::BI__builtin_ia32_cmpw512_mask:
2107 case X86::BI__builtin_ia32_cmpd128_mask:
2108 case X86::BI__builtin_ia32_cmpd256_mask:
2109 case X86::BI__builtin_ia32_cmpd512_mask:
2110 case X86::BI__builtin_ia32_cmpq128_mask:
2111 case X86::BI__builtin_ia32_cmpq256_mask:
2112 case X86::BI__builtin_ia32_cmpq512_mask: {
2116 case X86::BI__builtin_ia32_ucmpb128_mask:
2117 case X86::BI__builtin_ia32_ucmpb256_mask:
2118 case X86::BI__builtin_ia32_ucmpb512_mask:
2119 case X86::BI__builtin_ia32_ucmpw128_mask:
2120 case X86::BI__builtin_ia32_ucmpw256_mask:
2121 case X86::BI__builtin_ia32_ucmpw512_mask:
2122 case X86::BI__builtin_ia32_ucmpd128_mask:
2123 case X86::BI__builtin_ia32_ucmpd256_mask:
2124 case X86::BI__builtin_ia32_ucmpd512_mask:
2125 case X86::BI__builtin_ia32_ucmpq128_mask:
2126 case X86::BI__builtin_ia32_ucmpq256_mask:
2127 case X86::BI__builtin_ia32_ucmpq512_mask: {
2131 case X86::BI__builtin_ia32_vpcomb:
2132 case X86::BI__builtin_ia32_vpcomw:
2133 case X86::BI__builtin_ia32_vpcomd:
2134 case X86::BI__builtin_ia32_vpcomq:
2136 case X86::BI__builtin_ia32_vpcomub:
2137 case X86::BI__builtin_ia32_vpcomuw:
2138 case X86::BI__builtin_ia32_vpcomud:
2139 case X86::BI__builtin_ia32_vpcomuq:
2142 case X86::BI__builtin_ia32_kortestcqi:
2143 case X86::BI__builtin_ia32_kortestchi:
2144 case X86::BI__builtin_ia32_kortestcsi:
2145 case X86::BI__builtin_ia32_kortestcdi: {
2147 Value *
C = llvm::Constant::getAllOnesValue(Ops[0]->
getType());
2151 case X86::BI__builtin_ia32_kortestzqi:
2152 case X86::BI__builtin_ia32_kortestzhi:
2153 case X86::BI__builtin_ia32_kortestzsi:
2154 case X86::BI__builtin_ia32_kortestzdi: {
2161 case X86::BI__builtin_ia32_ktestcqi:
2162 case X86::BI__builtin_ia32_ktestzqi:
2163 case X86::BI__builtin_ia32_ktestchi:
2164 case X86::BI__builtin_ia32_ktestzhi:
2165 case X86::BI__builtin_ia32_ktestcsi:
2166 case X86::BI__builtin_ia32_ktestzsi:
2167 case X86::BI__builtin_ia32_ktestcdi:
2168 case X86::BI__builtin_ia32_ktestzdi: {
2170 switch (BuiltinID) {
2171 default: llvm_unreachable(
"Unsupported intrinsic!");
2172 case X86::BI__builtin_ia32_ktestcqi:
2173 IID = Intrinsic::x86_avx512_ktestc_b;
2175 case X86::BI__builtin_ia32_ktestzqi:
2176 IID = Intrinsic::x86_avx512_ktestz_b;
2178 case X86::BI__builtin_ia32_ktestchi:
2179 IID = Intrinsic::x86_avx512_ktestc_w;
2181 case X86::BI__builtin_ia32_ktestzhi:
2182 IID = Intrinsic::x86_avx512_ktestz_w;
2184 case X86::BI__builtin_ia32_ktestcsi:
2185 IID = Intrinsic::x86_avx512_ktestc_d;
2187 case X86::BI__builtin_ia32_ktestzsi:
2188 IID = Intrinsic::x86_avx512_ktestz_d;
2190 case X86::BI__builtin_ia32_ktestcdi:
2191 IID = Intrinsic::x86_avx512_ktestc_q;
2193 case X86::BI__builtin_ia32_ktestzdi:
2194 IID = Intrinsic::x86_avx512_ktestz_q;
2198 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
2202 return Builder.CreateCall(Intr, {LHS, RHS});
2205 case X86::BI__builtin_ia32_kaddqi:
2206 case X86::BI__builtin_ia32_kaddhi:
2207 case X86::BI__builtin_ia32_kaddsi:
2208 case X86::BI__builtin_ia32_kadddi: {
2210 switch (BuiltinID) {
2211 default: llvm_unreachable(
"Unsupported intrinsic!");
2212 case X86::BI__builtin_ia32_kaddqi:
2213 IID = Intrinsic::x86_avx512_kadd_b;
2215 case X86::BI__builtin_ia32_kaddhi:
2216 IID = Intrinsic::x86_avx512_kadd_w;
2218 case X86::BI__builtin_ia32_kaddsi:
2219 IID = Intrinsic::x86_avx512_kadd_d;
2221 case X86::BI__builtin_ia32_kadddi:
2222 IID = Intrinsic::x86_avx512_kadd_q;
2226 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
2233 case X86::BI__builtin_ia32_kandqi:
2234 case X86::BI__builtin_ia32_kandhi:
2235 case X86::BI__builtin_ia32_kandsi:
2236 case X86::BI__builtin_ia32_kanddi:
2238 case X86::BI__builtin_ia32_kandnqi:
2239 case X86::BI__builtin_ia32_kandnhi:
2240 case X86::BI__builtin_ia32_kandnsi:
2241 case X86::BI__builtin_ia32_kandndi:
2243 case X86::BI__builtin_ia32_korqi:
2244 case X86::BI__builtin_ia32_korhi:
2245 case X86::BI__builtin_ia32_korsi:
2246 case X86::BI__builtin_ia32_kordi:
2248 case X86::BI__builtin_ia32_kxnorqi:
2249 case X86::BI__builtin_ia32_kxnorhi:
2250 case X86::BI__builtin_ia32_kxnorsi:
2251 case X86::BI__builtin_ia32_kxnordi:
2253 case X86::BI__builtin_ia32_kxorqi:
2254 case X86::BI__builtin_ia32_kxorhi:
2255 case X86::BI__builtin_ia32_kxorsi:
2256 case X86::BI__builtin_ia32_kxordi:
2258 case X86::BI__builtin_ia32_knotqi:
2259 case X86::BI__builtin_ia32_knothi:
2260 case X86::BI__builtin_ia32_knotsi:
2261 case X86::BI__builtin_ia32_knotdi: {
2262 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
2267 case X86::BI__builtin_ia32_kmovb:
2268 case X86::BI__builtin_ia32_kmovw:
2269 case X86::BI__builtin_ia32_kmovd:
2270 case X86::BI__builtin_ia32_kmovq: {
2274 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
2279 case X86::BI__builtin_ia32_kunpckdi:
2280 case X86::BI__builtin_ia32_kunpcksi:
2281 case X86::BI__builtin_ia32_kunpckhi: {
2282 unsigned NumElts = Ops[0]->getType()->getIntegerBitWidth();
2286 for (
unsigned i = 0; i != NumElts; ++i)
2291 LHS =
Builder.CreateShuffleVector(LHS, LHS,
ArrayRef(Indices, NumElts / 2));
2292 RHS =
Builder.CreateShuffleVector(RHS, RHS,
ArrayRef(Indices, NumElts / 2));
2300 case X86::BI__builtin_ia32_sqrtsh_round_mask:
2301 case X86::BI__builtin_ia32_sqrtsd_round_mask:
2302 case X86::BI__builtin_ia32_sqrtss_round_mask: {
2309 switch (BuiltinID) {
2311 llvm_unreachable(
"Unsupported intrinsic!");
2312 case X86::BI__builtin_ia32_sqrtsh_round_mask:
2313 IID = Intrinsic::x86_avx512fp16_mask_sqrt_sh;
2315 case X86::BI__builtin_ia32_sqrtsd_round_mask:
2316 IID = Intrinsic::x86_avx512_mask_sqrt_sd;
2318 case X86::BI__builtin_ia32_sqrtss_round_mask:
2319 IID = Intrinsic::x86_avx512_mask_sqrt_ss;
2322 return Builder.CreateCall(
CGM.getIntrinsic(IID), Ops);
2324 Value *A =
Builder.CreateExtractElement(Ops[1], (uint64_t)0);
2326 if (
Builder.getIsFPConstrained()) {
2328 F =
CGM.getIntrinsic(Intrinsic::experimental_constrained_sqrt,
2330 A =
Builder.CreateConstrainedFPCall(F, A);
2332 F =
CGM.getIntrinsic(Intrinsic::sqrt, A->
getType());
2335 Value *Src =
Builder.CreateExtractElement(Ops[2], (uint64_t)0);
2337 return Builder.CreateInsertElement(Ops[0], A, (uint64_t)0);
2339 case X86::BI__builtin_ia32_sqrtph512:
2340 case X86::BI__builtin_ia32_sqrtps512:
2341 case X86::BI__builtin_ia32_sqrtpd512: {
2348 switch (BuiltinID) {
2350 llvm_unreachable(
"Unsupported intrinsic!");
2351 case X86::BI__builtin_ia32_sqrtph512:
2352 IID = Intrinsic::x86_avx512fp16_sqrt_ph_512;
2354 case X86::BI__builtin_ia32_sqrtps512:
2355 IID = Intrinsic::x86_avx512_sqrt_ps_512;
2357 case X86::BI__builtin_ia32_sqrtpd512:
2358 IID = Intrinsic::x86_avx512_sqrt_pd_512;
2361 return Builder.CreateCall(
CGM.getIntrinsic(IID), Ops);
2363 if (
Builder.getIsFPConstrained()) {
2365 Function *F =
CGM.getIntrinsic(Intrinsic::experimental_constrained_sqrt,
2367 return Builder.CreateConstrainedFPCall(F, Ops[0]);
2370 return Builder.CreateCall(F, Ops[0]);
2374 case X86::BI__builtin_ia32_pmuludq128:
2375 case X86::BI__builtin_ia32_pmuludq256:
2376 case X86::BI__builtin_ia32_pmuludq512:
2379 case X86::BI__builtin_ia32_pmuldq128:
2380 case X86::BI__builtin_ia32_pmuldq256:
2381 case X86::BI__builtin_ia32_pmuldq512:
2384 case X86::BI__builtin_ia32_pternlogd512_mask:
2385 case X86::BI__builtin_ia32_pternlogq512_mask:
2386 case X86::BI__builtin_ia32_pternlogd128_mask:
2387 case X86::BI__builtin_ia32_pternlogd256_mask:
2388 case X86::BI__builtin_ia32_pternlogq128_mask:
2389 case X86::BI__builtin_ia32_pternlogq256_mask:
2392 case X86::BI__builtin_ia32_pternlogd512_maskz:
2393 case X86::BI__builtin_ia32_pternlogq512_maskz:
2394 case X86::BI__builtin_ia32_pternlogd128_maskz:
2395 case X86::BI__builtin_ia32_pternlogd256_maskz:
2396 case X86::BI__builtin_ia32_pternlogq128_maskz:
2397 case X86::BI__builtin_ia32_pternlogq256_maskz:
2400 case X86::BI__builtin_ia32_vpshldd128:
2401 case X86::BI__builtin_ia32_vpshldd256:
2402 case X86::BI__builtin_ia32_vpshldd512:
2403 case X86::BI__builtin_ia32_vpshldq128:
2404 case X86::BI__builtin_ia32_vpshldq256:
2405 case X86::BI__builtin_ia32_vpshldq512:
2406 case X86::BI__builtin_ia32_vpshldw128:
2407 case X86::BI__builtin_ia32_vpshldw256:
2408 case X86::BI__builtin_ia32_vpshldw512:
2411 case X86::BI__builtin_ia32_vpshrdd128:
2412 case X86::BI__builtin_ia32_vpshrdd256:
2413 case X86::BI__builtin_ia32_vpshrdd512:
2414 case X86::BI__builtin_ia32_vpshrdq128:
2415 case X86::BI__builtin_ia32_vpshrdq256:
2416 case X86::BI__builtin_ia32_vpshrdq512:
2417 case X86::BI__builtin_ia32_vpshrdw128:
2418 case X86::BI__builtin_ia32_vpshrdw256:
2419 case X86::BI__builtin_ia32_vpshrdw512:
2424 case X86::BI__builtin_ia32_reduce_fadd_pd512:
2425 case X86::BI__builtin_ia32_reduce_fadd_ps512:
2426 case X86::BI__builtin_ia32_reduce_fadd_ph512:
2427 case X86::BI__builtin_ia32_reduce_fadd_ph256:
2428 case X86::BI__builtin_ia32_reduce_fadd_ph128: {
2430 CGM.getIntrinsic(Intrinsic::vector_reduce_fadd, Ops[1]->
getType());
2431 IRBuilder<>::FastMathFlagGuard FMFGuard(
Builder);
2432 Builder.getFastMathFlags().setAllowReassoc();
2433 return Builder.CreateCall(F, {Ops[0], Ops[1]});
2435 case X86::BI__builtin_ia32_reduce_fmul_pd512:
2436 case X86::BI__builtin_ia32_reduce_fmul_ps512:
2437 case X86::BI__builtin_ia32_reduce_fmul_ph512:
2438 case X86::BI__builtin_ia32_reduce_fmul_ph256:
2439 case X86::BI__builtin_ia32_reduce_fmul_ph128: {
2441 CGM.getIntrinsic(Intrinsic::vector_reduce_fmul, Ops[1]->
getType());
2442 IRBuilder<>::FastMathFlagGuard FMFGuard(
Builder);
2443 Builder.getFastMathFlags().setAllowReassoc();
2444 return Builder.CreateCall(F, {Ops[0], Ops[1]});
2446 case X86::BI__builtin_ia32_reduce_fmax_pd512:
2447 case X86::BI__builtin_ia32_reduce_fmax_ps512:
2448 case X86::BI__builtin_ia32_reduce_fmax_ph512:
2449 case X86::BI__builtin_ia32_reduce_fmax_ph256:
2450 case X86::BI__builtin_ia32_reduce_fmax_ph128: {
2452 CGM.getIntrinsic(Intrinsic::vector_reduce_fmax, Ops[0]->
getType());
2453 IRBuilder<>::FastMathFlagGuard FMFGuard(
Builder);
2454 Builder.getFastMathFlags().setNoNaNs();
2455 return Builder.CreateCall(F, {Ops[0]});
2457 case X86::BI__builtin_ia32_reduce_fmin_pd512:
2458 case X86::BI__builtin_ia32_reduce_fmin_ps512:
2459 case X86::BI__builtin_ia32_reduce_fmin_ph512:
2460 case X86::BI__builtin_ia32_reduce_fmin_ph256:
2461 case X86::BI__builtin_ia32_reduce_fmin_ph128: {
2463 CGM.getIntrinsic(Intrinsic::vector_reduce_fmin, Ops[0]->
getType());
2464 IRBuilder<>::FastMathFlagGuard FMFGuard(
Builder);
2465 Builder.getFastMathFlags().setNoNaNs();
2466 return Builder.CreateCall(F, {Ops[0]});
2469 case X86::BI__builtin_ia32_rdrand16_step:
2470 case X86::BI__builtin_ia32_rdrand32_step:
2471 case X86::BI__builtin_ia32_rdrand64_step:
2472 case X86::BI__builtin_ia32_rdseed16_step:
2473 case X86::BI__builtin_ia32_rdseed32_step:
2474 case X86::BI__builtin_ia32_rdseed64_step: {
2476 switch (BuiltinID) {
2477 default: llvm_unreachable(
"Unsupported intrinsic!");
2478 case X86::BI__builtin_ia32_rdrand16_step:
2479 ID = Intrinsic::x86_rdrand_16;
2481 case X86::BI__builtin_ia32_rdrand32_step:
2482 ID = Intrinsic::x86_rdrand_32;
2484 case X86::BI__builtin_ia32_rdrand64_step:
2485 ID = Intrinsic::x86_rdrand_64;
2487 case X86::BI__builtin_ia32_rdseed16_step:
2488 ID = Intrinsic::x86_rdseed_16;
2490 case X86::BI__builtin_ia32_rdseed32_step:
2491 ID = Intrinsic::x86_rdseed_32;
2493 case X86::BI__builtin_ia32_rdseed64_step:
2494 ID = Intrinsic::x86_rdseed_64;
2503 case X86::BI__builtin_ia32_addcarryx_u32:
2504 case X86::BI__builtin_ia32_addcarryx_u64:
2505 case X86::BI__builtin_ia32_subborrow_u32:
2506 case X86::BI__builtin_ia32_subborrow_u64: {
2508 switch (BuiltinID) {
2509 default: llvm_unreachable(
"Unsupported intrinsic!");
2510 case X86::BI__builtin_ia32_addcarryx_u32:
2511 IID = Intrinsic::x86_addcarry_32;
2513 case X86::BI__builtin_ia32_addcarryx_u64:
2514 IID = Intrinsic::x86_addcarry_64;
2516 case X86::BI__builtin_ia32_subborrow_u32:
2517 IID = Intrinsic::x86_subborrow_32;
2519 case X86::BI__builtin_ia32_subborrow_u64:
2520 IID = Intrinsic::x86_subborrow_64;
2525 { Ops[0], Ops[1], Ops[2] });
2531 case X86::BI__builtin_ia32_fpclassps128_mask:
2532 case X86::BI__builtin_ia32_fpclassps256_mask:
2533 case X86::BI__builtin_ia32_fpclassps512_mask:
2534 case X86::BI__builtin_ia32_vfpclassbf16128_mask:
2535 case X86::BI__builtin_ia32_vfpclassbf16256_mask:
2536 case X86::BI__builtin_ia32_vfpclassbf16512_mask:
2537 case X86::BI__builtin_ia32_fpclassph128_mask:
2538 case X86::BI__builtin_ia32_fpclassph256_mask:
2539 case X86::BI__builtin_ia32_fpclassph512_mask:
2540 case X86::BI__builtin_ia32_fpclasspd128_mask:
2541 case X86::BI__builtin_ia32_fpclasspd256_mask:
2542 case X86::BI__builtin_ia32_fpclasspd512_mask: {
2545 Value *MaskIn = Ops[2];
2549 switch (BuiltinID) {
2550 default: llvm_unreachable(
"Unsupported intrinsic!");
2551 case X86::BI__builtin_ia32_vfpclassbf16128_mask:
2552 ID = Intrinsic::x86_avx10_fpclass_bf16_128;
2554 case X86::BI__builtin_ia32_vfpclassbf16256_mask:
2555 ID = Intrinsic::x86_avx10_fpclass_bf16_256;
2557 case X86::BI__builtin_ia32_vfpclassbf16512_mask:
2558 ID = Intrinsic::x86_avx10_fpclass_bf16_512;
2560 case X86::BI__builtin_ia32_fpclassph128_mask:
2561 ID = Intrinsic::x86_avx512fp16_fpclass_ph_128;
2563 case X86::BI__builtin_ia32_fpclassph256_mask:
2564 ID = Intrinsic::x86_avx512fp16_fpclass_ph_256;
2566 case X86::BI__builtin_ia32_fpclassph512_mask:
2567 ID = Intrinsic::x86_avx512fp16_fpclass_ph_512;
2569 case X86::BI__builtin_ia32_fpclassps128_mask:
2570 ID = Intrinsic::x86_avx512_fpclass_ps_128;
2572 case X86::BI__builtin_ia32_fpclassps256_mask:
2573 ID = Intrinsic::x86_avx512_fpclass_ps_256;
2575 case X86::BI__builtin_ia32_fpclassps512_mask:
2576 ID = Intrinsic::x86_avx512_fpclass_ps_512;
2578 case X86::BI__builtin_ia32_fpclasspd128_mask:
2579 ID = Intrinsic::x86_avx512_fpclass_pd_128;
2581 case X86::BI__builtin_ia32_fpclasspd256_mask:
2582 ID = Intrinsic::x86_avx512_fpclass_pd_256;
2584 case X86::BI__builtin_ia32_fpclasspd512_mask:
2585 ID = Intrinsic::x86_avx512_fpclass_pd_512;
2593 case X86::BI__builtin_ia32_vp2intersect_q_512:
2594 case X86::BI__builtin_ia32_vp2intersect_q_256:
2595 case X86::BI__builtin_ia32_vp2intersect_q_128:
2596 case X86::BI__builtin_ia32_vp2intersect_d_512:
2597 case X86::BI__builtin_ia32_vp2intersect_d_256:
2598 case X86::BI__builtin_ia32_vp2intersect_d_128: {
2603 switch (BuiltinID) {
2604 default: llvm_unreachable(
"Unsupported intrinsic!");
2605 case X86::BI__builtin_ia32_vp2intersect_q_512:
2606 ID = Intrinsic::x86_avx512_vp2intersect_q_512;
2608 case X86::BI__builtin_ia32_vp2intersect_q_256:
2609 ID = Intrinsic::x86_avx512_vp2intersect_q_256;
2611 case X86::BI__builtin_ia32_vp2intersect_q_128:
2612 ID = Intrinsic::x86_avx512_vp2intersect_q_128;
2614 case X86::BI__builtin_ia32_vp2intersect_d_512:
2615 ID = Intrinsic::x86_avx512_vp2intersect_d_512;
2617 case X86::BI__builtin_ia32_vp2intersect_d_256:
2618 ID = Intrinsic::x86_avx512_vp2intersect_d_256;
2620 case X86::BI__builtin_ia32_vp2intersect_d_128:
2621 ID = Intrinsic::x86_avx512_vp2intersect_d_128;
2635 case X86::BI__builtin_ia32_vpmultishiftqb128:
2636 case X86::BI__builtin_ia32_vpmultishiftqb256:
2637 case X86::BI__builtin_ia32_vpmultishiftqb512: {
2639 switch (BuiltinID) {
2640 default: llvm_unreachable(
"Unsupported intrinsic!");
2641 case X86::BI__builtin_ia32_vpmultishiftqb128:
2642 ID = Intrinsic::x86_avx512_pmultishift_qb_128;
2644 case X86::BI__builtin_ia32_vpmultishiftqb256:
2645 ID = Intrinsic::x86_avx512_pmultishift_qb_256;
2647 case X86::BI__builtin_ia32_vpmultishiftqb512:
2648 ID = Intrinsic::x86_avx512_pmultishift_qb_512;
2652 return Builder.CreateCall(
CGM.getIntrinsic(ID), Ops);
2655 case X86::BI__builtin_ia32_vpshufbitqmb128_mask:
2656 case X86::BI__builtin_ia32_vpshufbitqmb256_mask:
2657 case X86::BI__builtin_ia32_vpshufbitqmb512_mask: {
2660 Value *MaskIn = Ops[2];
2664 switch (BuiltinID) {
2665 default: llvm_unreachable(
"Unsupported intrinsic!");
2666 case X86::BI__builtin_ia32_vpshufbitqmb128_mask:
2667 ID = Intrinsic::x86_avx512_vpshufbitqmb_128;
2669 case X86::BI__builtin_ia32_vpshufbitqmb256_mask:
2670 ID = Intrinsic::x86_avx512_vpshufbitqmb_256;
2672 case X86::BI__builtin_ia32_vpshufbitqmb512_mask:
2673 ID = Intrinsic::x86_avx512_vpshufbitqmb_512;
2682 case X86::BI__builtin_ia32_cmpeqps:
2683 case X86::BI__builtin_ia32_cmpeqpd:
2684 return getVectorFCmpIR(CmpInst::FCMP_OEQ,
false);
2685 case X86::BI__builtin_ia32_cmpltps:
2686 case X86::BI__builtin_ia32_cmpltpd:
2687 return getVectorFCmpIR(CmpInst::FCMP_OLT,
true);
2688 case X86::BI__builtin_ia32_cmpleps:
2689 case X86::BI__builtin_ia32_cmplepd:
2690 return getVectorFCmpIR(CmpInst::FCMP_OLE,
true);
2691 case X86::BI__builtin_ia32_cmpunordps:
2692 case X86::BI__builtin_ia32_cmpunordpd:
2693 return getVectorFCmpIR(CmpInst::FCMP_UNO,
false);
2694 case X86::BI__builtin_ia32_cmpneqps:
2695 case X86::BI__builtin_ia32_cmpneqpd:
2696 return getVectorFCmpIR(CmpInst::FCMP_UNE,
false);
2697 case X86::BI__builtin_ia32_cmpnltps:
2698 case X86::BI__builtin_ia32_cmpnltpd:
2699 return getVectorFCmpIR(CmpInst::FCMP_UGE,
true);
2700 case X86::BI__builtin_ia32_cmpnleps:
2701 case X86::BI__builtin_ia32_cmpnlepd:
2702 return getVectorFCmpIR(CmpInst::FCMP_UGT,
true);
2703 case X86::BI__builtin_ia32_cmpordps:
2704 case X86::BI__builtin_ia32_cmpordpd:
2705 return getVectorFCmpIR(CmpInst::FCMP_ORD,
false);
2706 case X86::BI__builtin_ia32_cmpph128_mask:
2707 case X86::BI__builtin_ia32_cmpph256_mask:
2708 case X86::BI__builtin_ia32_cmpph512_mask:
2709 case X86::BI__builtin_ia32_cmpps128_mask:
2710 case X86::BI__builtin_ia32_cmpps256_mask:
2711 case X86::BI__builtin_ia32_cmpps512_mask:
2712 case X86::BI__builtin_ia32_cmppd128_mask:
2713 case X86::BI__builtin_ia32_cmppd256_mask:
2714 case X86::BI__builtin_ia32_cmppd512_mask:
2715 case X86::BI__builtin_ia32_vcmpbf16512_mask:
2716 case X86::BI__builtin_ia32_vcmpbf16256_mask:
2717 case X86::BI__builtin_ia32_vcmpbf16128_mask:
2720 case X86::BI__builtin_ia32_cmpps:
2721 case X86::BI__builtin_ia32_cmpps256:
2722 case X86::BI__builtin_ia32_cmppd:
2723 case X86::BI__builtin_ia32_cmppd256: {
2736 FCmpInst::Predicate Pred;
2741 case 0x00: Pred = FCmpInst::FCMP_OEQ; IsSignaling =
false;
break;
2742 case 0x01: Pred = FCmpInst::FCMP_OLT; IsSignaling =
true;
break;
2743 case 0x02: Pred = FCmpInst::FCMP_OLE; IsSignaling =
true;
break;
2744 case 0x03: Pred = FCmpInst::FCMP_UNO; IsSignaling =
false;
break;
2745 case 0x04: Pred = FCmpInst::FCMP_UNE; IsSignaling =
false;
break;
2746 case 0x05: Pred = FCmpInst::FCMP_UGE; IsSignaling =
true;
break;
2747 case 0x06: Pred = FCmpInst::FCMP_UGT; IsSignaling =
true;
break;
2748 case 0x07: Pred = FCmpInst::FCMP_ORD; IsSignaling =
false;
break;
2749 case 0x08: Pred = FCmpInst::FCMP_UEQ; IsSignaling =
false;
break;
2750 case 0x09: Pred = FCmpInst::FCMP_ULT; IsSignaling =
true;
break;
2751 case 0x0a: Pred = FCmpInst::FCMP_ULE; IsSignaling =
true;
break;
2752 case 0x0b: Pred = FCmpInst::FCMP_FALSE; IsSignaling =
false;
break;
2753 case 0x0c: Pred = FCmpInst::FCMP_ONE; IsSignaling =
false;
break;
2754 case 0x0d: Pred = FCmpInst::FCMP_OGE; IsSignaling =
true;
break;
2755 case 0x0e: Pred = FCmpInst::FCMP_OGT; IsSignaling =
true;
break;
2756 case 0x0f: Pred = FCmpInst::FCMP_TRUE; IsSignaling =
false;
break;
2757 default: llvm_unreachable(
"Unhandled CC");
2762 IsSignaling = !IsSignaling;
2769 if (
Builder.getIsFPConstrained() &&
2770 (Pred == FCmpInst::FCMP_TRUE || Pred == FCmpInst::FCMP_FALSE ||
2774 switch (BuiltinID) {
2775 default: llvm_unreachable(
"Unexpected builtin");
2776 case X86::BI__builtin_ia32_cmpps:
2777 IID = Intrinsic::x86_sse_cmp_ps;
2779 case X86::BI__builtin_ia32_cmpps256:
2780 IID = Intrinsic::x86_avx_cmp_ps_256;
2782 case X86::BI__builtin_ia32_cmppd:
2783 IID = Intrinsic::x86_sse2_cmp_pd;
2785 case X86::BI__builtin_ia32_cmppd256:
2786 IID = Intrinsic::x86_avx_cmp_pd_256;
2788 case X86::BI__builtin_ia32_cmpph128_mask:
2789 IID = Intrinsic::x86_avx512fp16_mask_cmp_ph_128;
2791 case X86::BI__builtin_ia32_cmpph256_mask:
2792 IID = Intrinsic::x86_avx512fp16_mask_cmp_ph_256;
2794 case X86::BI__builtin_ia32_cmpph512_mask:
2795 IID = Intrinsic::x86_avx512fp16_mask_cmp_ph_512;
2797 case X86::BI__builtin_ia32_cmpps512_mask:
2798 IID = Intrinsic::x86_avx512_mask_cmp_ps_512;
2800 case X86::BI__builtin_ia32_cmppd512_mask:
2801 IID = Intrinsic::x86_avx512_mask_cmp_pd_512;
2803 case X86::BI__builtin_ia32_cmpps128_mask:
2804 IID = Intrinsic::x86_avx512_mask_cmp_ps_128;
2806 case X86::BI__builtin_ia32_cmpps256_mask:
2807 IID = Intrinsic::x86_avx512_mask_cmp_ps_256;
2809 case X86::BI__builtin_ia32_cmppd128_mask:
2810 IID = Intrinsic::x86_avx512_mask_cmp_pd_128;
2812 case X86::BI__builtin_ia32_cmppd256_mask:
2813 IID = Intrinsic::x86_avx512_mask_cmp_pd_256;
2826 return Builder.CreateCall(Intr, Ops);
2840 Cmp =
Builder.CreateFCmpS(Pred, Ops[0], Ops[1]);
2842 Cmp =
Builder.CreateFCmp(Pred, Ops[0], Ops[1]);
2846 return getVectorFCmpIR(Pred, IsSignaling);
2850 case X86::BI__builtin_ia32_cmpeqss:
2851 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 0);
2852 case X86::BI__builtin_ia32_cmpltss:
2853 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 1);
2854 case X86::BI__builtin_ia32_cmpless:
2855 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 2);
2856 case X86::BI__builtin_ia32_cmpunordss:
2857 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 3);
2858 case X86::BI__builtin_ia32_cmpneqss:
2859 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 4);
2860 case X86::BI__builtin_ia32_cmpnltss:
2861 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 5);
2862 case X86::BI__builtin_ia32_cmpnless:
2863 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 6);
2864 case X86::BI__builtin_ia32_cmpordss:
2865 return getCmpIntrinsicCall(Intrinsic::x86_sse_cmp_ss, 7);
2866 case X86::BI__builtin_ia32_cmpeqsd:
2867 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 0);
2868 case X86::BI__builtin_ia32_cmpltsd:
2869 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 1);
2870 case X86::BI__builtin_ia32_cmplesd:
2871 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 2);
2872 case X86::BI__builtin_ia32_cmpunordsd:
2873 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 3);
2874 case X86::BI__builtin_ia32_cmpneqsd:
2875 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 4);
2876 case X86::BI__builtin_ia32_cmpnltsd:
2877 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 5);
2878 case X86::BI__builtin_ia32_cmpnlesd:
2879 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 6);
2880 case X86::BI__builtin_ia32_cmpordsd:
2881 return getCmpIntrinsicCall(Intrinsic::x86_sse2_cmp_sd, 7);
2884 case X86::BI__builtin_ia32_vcvtph2ps_mask:
2885 case X86::BI__builtin_ia32_vcvtph2ps256_mask:
2886 case X86::BI__builtin_ia32_vcvtph2ps512_mask: {
2892 case X86::BI__builtin_ia32_cvtneps2bf16_128_mask: {
2896 Intrinsic::ID IID = Intrinsic::x86_avx512bf16_mask_cvtneps2bf16_128;
2897 return Builder.CreateCall(
CGM.getIntrinsic(IID), Ops);
2900 case X86::BI__builtin_ia32_cvtneps2bf16_256_mask:
2901 case X86::BI__builtin_ia32_cvtneps2bf16_512_mask: {
2903 switch (BuiltinID) {
2904 default: llvm_unreachable(
"Unsupported intrinsic!");
2905 case X86::BI__builtin_ia32_cvtneps2bf16_256_mask:
2906 IID = Intrinsic::x86_avx512bf16_cvtneps2bf16_256;
2908 case X86::BI__builtin_ia32_cvtneps2bf16_512_mask:
2909 IID = Intrinsic::x86_avx512bf16_cvtneps2bf16_512;
2916 case X86::BI__cpuid:
2917 case X86::BI__cpuidex: {
2919 Value *SubFuncId = BuiltinID == X86::BI__cpuidex
2921 : llvm::ConstantInt::get(
Int32Ty, 0);
2923 llvm::StructType *CpuidRetTy =
2925 llvm::FunctionType *FTy =
2928 StringRef
Asm, Constraints;
2929 if (
getTarget().getTriple().getArch() == llvm::Triple::x86) {
2931 Constraints =
"={ax},={bx},={cx},={dx},{ax},{cx}";
2934 Asm =
"xchgq %rbx, ${1:q}\n"
2936 "xchgq %rbx, ${1:q}";
2937 Constraints =
"={ax},=r,={cx},={dx},0,2";
2940 llvm::InlineAsm *IA = llvm::InlineAsm::get(FTy,
Asm, Constraints,
2942 Value *IACall =
Builder.CreateCall(IA, {FuncId, SubFuncId});
2944 Value *Store =
nullptr;
2945 for (
unsigned i = 0; i < 4; i++) {
2946 Value *Extracted =
Builder.CreateExtractValue(IACall, i);
2957 case X86::BI__emulu: {
2959 bool isSigned = (BuiltinID == X86::BI__emul);
2962 return Builder.CreateMul(LHS, RHS,
"", !isSigned, isSigned);
2965 case X86::BI__umulh:
2966 case X86::BI_mul128:
2967 case X86::BI_umul128: {
2969 llvm::Type *Int128Ty = llvm::IntegerType::get(
getLLVMContext(), 128);
2971 bool IsSigned = (BuiltinID == X86::BI__mulh || BuiltinID == X86::BI_mul128);
2972 Value *LHS =
Builder.CreateIntCast(Ops[0], Int128Ty, IsSigned);
2973 Value *RHS =
Builder.CreateIntCast(Ops[1], Int128Ty, IsSigned);
2975 Value *MulResult, *HigherBits;
2977 MulResult =
Builder.CreateNSWMul(LHS, RHS);
2978 HigherBits =
Builder.CreateAShr(MulResult, 64);
2980 MulResult =
Builder.CreateNUWMul(LHS, RHS);
2981 HigherBits =
Builder.CreateLShr(MulResult, 64);
2983 HigherBits =
Builder.CreateIntCast(HigherBits, ResType, IsSigned);
2985 if (BuiltinID == X86::BI__mulh || BuiltinID == X86::BI__umulh)
2989 Builder.CreateStore(HigherBits, HighBitsAddress);
2990 return Builder.CreateIntCast(MulResult, ResType, IsSigned);
2993 case X86::BI__faststorefence: {
2994 return Builder.CreateFence(llvm::AtomicOrdering::SequentiallyConsistent,
2995 llvm::SyncScope::System);
2997 case X86::BI__shiftleft128:
2998 case X86::BI__shiftright128: {
2999 llvm::Function *F =
CGM.getIntrinsic(
3000 BuiltinID == X86::BI__shiftleft128 ? Intrinsic::fshl : Intrinsic::fshr,
3005 std::swap(Ops[0], Ops[1]);
3007 return Builder.CreateCall(F, Ops);
3009 case X86::BI_ReadWriteBarrier:
3010 case X86::BI_ReadBarrier:
3011 case X86::BI_WriteBarrier: {
3012 return Builder.CreateFence(llvm::AtomicOrdering::SequentiallyConsistent,
3013 llvm::SyncScope::SingleThread);
3016 case X86::BI_AddressOfReturnAddress: {
3021 case X86::BI__stosb: {
3024 return Builder.CreateMemSet(Ops[0], Ops[1], Ops[2], Align(1),
true);
3029 case X86::BI__int2c: {
3031 llvm::FunctionType *FTy = llvm::FunctionType::get(
VoidTy,
false);
3032 llvm::InlineAsm *IA =
3033 llvm::InlineAsm::get(FTy,
"int $$0x2c",
"",
true);
3034 llvm::AttributeList NoReturnAttr = llvm::AttributeList::get(
3036 llvm::Attribute::NoReturn);
3037 llvm::CallInst *CI =
Builder.CreateCall(IA);
3038 CI->setAttributes(NoReturnAttr);
3041 case X86::BI__readfsbyte:
3042 case X86::BI__readfsword:
3043 case X86::BI__readfsdword:
3044 case X86::BI__readfsqword: {
3048 LoadInst *Load =
Builder.CreateAlignedLoad(
3050 Load->setVolatile(
true);
3053 case X86::BI__readgsbyte:
3054 case X86::BI__readgsword:
3055 case X86::BI__readgsdword:
3056 case X86::BI__readgsqword: {
3060 LoadInst *Load =
Builder.CreateAlignedLoad(
3062 Load->setVolatile(
true);
3065 case X86::BI__builtin_ia32_encodekey128_u32: {
3066 Intrinsic::ID IID = Intrinsic::x86_encodekey128;
3070 for (
int i = 0; i < 3; ++i) {
3073 Builder.CreateAlignedStore(Extract, Ptr, Align(1));
3078 case X86::BI__builtin_ia32_encodekey256_u32: {
3079 Intrinsic::ID IID = Intrinsic::x86_encodekey256;
3082 Builder.CreateCall(
CGM.getIntrinsic(IID), {Ops[0], Ops[1], Ops[2]});
3084 for (
int i = 0; i < 4; ++i) {
3087 Builder.CreateAlignedStore(Extract, Ptr, Align(1));
3092 case X86::BI__builtin_ia32_aesenc128kl_u8:
3093 case X86::BI__builtin_ia32_aesdec128kl_u8:
3094 case X86::BI__builtin_ia32_aesenc256kl_u8:
3095 case X86::BI__builtin_ia32_aesdec256kl_u8: {
3097 StringRef BlockName;
3098 switch (BuiltinID) {
3100 llvm_unreachable(
"Unexpected builtin");
3101 case X86::BI__builtin_ia32_aesenc128kl_u8:
3102 IID = Intrinsic::x86_aesenc128kl;
3103 BlockName =
"aesenc128kl";
3105 case X86::BI__builtin_ia32_aesdec128kl_u8:
3106 IID = Intrinsic::x86_aesdec128kl;
3107 BlockName =
"aesdec128kl";
3109 case X86::BI__builtin_ia32_aesenc256kl_u8:
3110 IID = Intrinsic::x86_aesenc256kl;
3111 BlockName =
"aesenc256kl";
3113 case X86::BI__builtin_ia32_aesdec256kl_u8:
3114 IID = Intrinsic::x86_aesdec256kl;
3115 BlockName =
"aesdec256kl";
3121 BasicBlock *NoError =
3131 Builder.SetInsertPoint(NoError);
3132 Builder.CreateDefaultAlignedStore(Out, Ops[0]);
3136 Constant *
Zero = llvm::Constant::getNullValue(Out->getType());
3143 case X86::BI__builtin_ia32_aesencwide128kl_u8:
3144 case X86::BI__builtin_ia32_aesdecwide128kl_u8:
3145 case X86::BI__builtin_ia32_aesencwide256kl_u8:
3146 case X86::BI__builtin_ia32_aesdecwide256kl_u8: {
3148 StringRef BlockName;
3149 switch (BuiltinID) {
3150 case X86::BI__builtin_ia32_aesencwide128kl_u8:
3151 IID = Intrinsic::x86_aesencwide128kl;
3152 BlockName =
"aesencwide128kl";
3154 case X86::BI__builtin_ia32_aesdecwide128kl_u8:
3155 IID = Intrinsic::x86_aesdecwide128kl;
3156 BlockName =
"aesdecwide128kl";
3158 case X86::BI__builtin_ia32_aesencwide256kl_u8:
3159 IID = Intrinsic::x86_aesencwide256kl;
3160 BlockName =
"aesencwide256kl";
3162 case X86::BI__builtin_ia32_aesdecwide256kl_u8:
3163 IID = Intrinsic::x86_aesdecwide256kl;
3164 BlockName =
"aesdecwide256kl";
3168 llvm::Type *Ty = FixedVectorType::get(
Builder.getInt64Ty(), 2);
3171 for (
int i = 0; i != 8; ++i) {
3172 Value *Ptr =
Builder.CreateConstGEP1_32(Ty, Ops[1], i);
3173 InOps[i + 1] =
Builder.CreateAlignedLoad(Ty, Ptr, Align(16));
3178 BasicBlock *NoError =
3187 Builder.SetInsertPoint(NoError);
3188 for (
int i = 0; i != 8; ++i) {
3191 Builder.CreateAlignedStore(Extract, Ptr, Align(16));
3196 for (
int i = 0; i != 8; ++i) {
3198 Constant *
Zero = llvm::Constant::getNullValue(Out->getType());
3199 Value *Ptr =
Builder.CreateConstGEP1_32(Out->getType(), Ops[0], i);
3207 case X86::BI__builtin_ia32_vfcmaddcph512_mask:
3210 case X86::BI__builtin_ia32_vfmaddcph512_mask: {
3211 Intrinsic::ID IID = IsConjFMA
3212 ? Intrinsic::x86_avx512fp16_mask_vfcmadd_cph_512
3213 : Intrinsic::x86_avx512fp16_mask_vfmadd_cph_512;
3217 case X86::BI__builtin_ia32_vfcmaddcsh_round_mask:
3220 case X86::BI__builtin_ia32_vfmaddcsh_round_mask: {
3221 Intrinsic::ID IID = IsConjFMA ? Intrinsic::x86_avx512fp16_mask_vfcmadd_csh
3222 : Intrinsic::x86_avx512fp16_mask_vfmadd_csh;
3227 case X86::BI__builtin_ia32_vfcmaddcsh_round_mask3:
3230 case X86::BI__builtin_ia32_vfmaddcsh_round_mask3: {
3231 Intrinsic::ID IID = IsConjFMA ? Intrinsic::x86_avx512fp16_mask_vfcmadd_csh
3232 : Intrinsic::x86_avx512fp16_mask_vfmadd_csh;
3234 static constexpr int Mask[] = {0, 5, 6, 7};
3235 return Builder.CreateShuffleVector(
Call, Ops[2], Mask);
3237 case X86::BI__builtin_ia32_prefetchi:
3239 CGM.getIntrinsic(Intrinsic::prefetch, Ops[0]->getType()),
3240 {Ops[0], llvm::ConstantInt::get(Int32Ty, 0), Ops[1],
3241 llvm::ConstantInt::get(Int32Ty, 0)});