36 bool Is64Bit = Subtarget.is64Bit();
37 bool HasCMOV = Subtarget.canUseCMOV();
38 bool HasSSE1 = Subtarget.hasSSE1();
39 bool HasSSE2 = Subtarget.hasSSE2();
40 bool HasSSE41 = Subtarget.hasSSE41();
41 bool HasAVX = Subtarget.hasAVX();
42 bool HasAVX2 = Subtarget.hasAVX2();
43 bool HasAVX512 = Subtarget.hasAVX512();
44 bool HasVLX = Subtarget.hasVLX();
45 bool HasDQI = Subtarget.hasAVX512() && Subtarget.hasDQI();
46 bool HasBWI = Subtarget.hasAVX512() && Subtarget.hasBWI();
47 bool UseX87 = !Subtarget.useSoftFloat() && Subtarget.hasX87();
48 bool HasPOPCNT = Subtarget.hasPOPCNT();
49 bool HasLZCNT = Subtarget.hasLZCNT();
50 bool HasBMI = Subtarget.hasBMI();
60 const LLT sMaxScalar = Subtarget.is64Bit() ? s64 : s32;
81 const LLT s8MaxVector = HasAVX512 ? v64s8 : HasAVX ? v32s8 : v16s8;
82 const LLT s16MaxVector = HasAVX512 ? v32s16 : HasAVX ? v16s16 : v8s16;
83 const LLT s32MaxVector = HasAVX512 ? v16s32 : HasAVX ? v8s32 : v4s32;
84 const LLT s64MaxVector = HasAVX512 ? v8s64 : HasAVX ? v4s64 : v2s64;
93 {G_IMPLICIT_DEF, G_PHI, G_FREEZE, G_CONSTANT_FOLD_BARRIER})
94 .legalFor({p0, s1, s8, s16, s32, s64})
95 .legalFor(UseX87, {s80})
96 .legalFor(Is64Bit, {s128})
97 .legalFor(HasSSE2, {v16s8, v8s16, v4s32, v2s64})
98 .legalFor(HasAVX, {v32s8, v16s16, v8s32, v4s64})
99 .legalFor(HasAVX512, {v64s8, v32s16, v16s32, v8s64})
100 .widenScalarOrEltToNextPow2(0, 8)
101 .clampScalarOrElt(0, s8, sMaxScalar)
102 .moreElementsToNextPow2(0)
103 .clampNumElements(0, v16s8, s8MaxVector)
104 .clampNumElements(0, v8s16, s16MaxVector)
105 .clampNumElements(0, v4s32, s32MaxVector)
106 .clampNumElements(0, v2s64, s64MaxVector)
107 .clampMaxNumElements(0, p0,
108 Is64Bit ? s64MaxVector.getNumElements()
109 : s32MaxVector.getNumElements())
114 .legalFor(Is64Bit, {s64})
115 .widenScalarToNextPow2(0, 8)
119 .widenScalarIf(
typeIs(1, s16),
121 return std::pair<unsigned, LLT>(1, s32);
126 {G_FCOS, G_FCOSH, G_FACOS, G_FSIN, G_FSINH, G_FASIN, G_FTAN,
127 G_FTANH, G_FATAN, G_FATAN2, G_FPOW, G_FEXP, G_FEXP2, G_FEXP10,
128 G_FLOG, G_FLOG2, G_FLOG10, G_FPOWI, G_FSINCOS, G_FCEIL, G_FFLOOR})
132 .
legalFor(UseX87 && !HasSSE1, {s32})
133 .legalFor(UseX87 && !HasSSE2, {s64})
134 .legalFor(UseX87, {s80})
139 .legalFor(HasSSE2 || UseX87, {s64})
140 .legalFor(UseX87, {s80});
146 for (
unsigned Op : {G_MERGE_VALUES, G_UNMERGE_VALUES}) {
147 unsigned BigTyIdx =
Op == G_MERGE_VALUES ? 0 : 1;
148 unsigned LitTyIdx =
Op == G_MERGE_VALUES ? 1 : 0;
155 switch (Q.
Types[BigTyIdx].getSizeInBits()) {
166 switch (Q.
Types[LitTyIdx].getSizeInBits()) {
181 .widenScalarToNextPow2(0, 32)
186 .legalFor({s8, s16, s32})
187 .legalFor(Is64Bit, {s64})
188 .legalFor(HasSSE2, {v16s8, v8s16, v4s32, v2s64})
189 .legalFor(HasAVX2, {v32s8, v16s16, v8s32, v4s64})
190 .legalFor(HasAVX512, {v16s32, v8s64})
191 .legalFor(HasBWI, {v64s8, v32s16})
192 .clampMinNumElements(0, s8, 16)
205 .legalFor({{s8, s8}, {s16, s8}, {s32, s8}})
206 .legalFor(Is64Bit, {{s64, s8}})
207 .widenScalarToNextPow2(0, 32)
215 .legalFor(Is64Bit, {s64})
216 .legalFor(HasSSE2, {v8s16})
217 .legalFor(HasSSE41, {v4s32})
218 .legalFor(HasAVX2, {v16s16, v8s32})
219 .legalFor(HasAVX512, {v16s32})
220 .legalFor(HasDQI, {v8s64})
221 .legalFor(HasDQI && HasVLX, {v2s64, v4s64})
222 .legalFor(HasBWI, {v32s16})
223 .clampMinNumElements(0, s16, 8)
234 .legalFor({s8, s16, s32})
235 .legalFor(Is64Bit, {s64})
236 .widenScalarToNextPow2(0, 32)
242 .legalFor({s8, s16, s32})
243 .legalFor(Is64Bit, {s64})
245 .clampScalar(0, s8, sMaxScalar);
249 .legalFor({{s8, s8}, {s16, s8}, {s32, s8}})
250 .legalFor(Is64Bit, {{s64, s8}})
251 .clampScalar(0, s8, sMaxScalar)
256 .legalFor({s8, s16, s32})
257 .legalFor(Is64Bit, {s64})
258 .legalFor(HasSSE2, {v16s8, v8s16, v4s32, v2s64})
259 .legalFor(HasAVX, {v32s8, v16s16, v8s32, v4s64})
260 .legalFor(HasAVX512, {v64s8, v32s16, v16s32, v8s64})
261 .clampNumElements(0, v16s8, s8MaxVector)
270 const std::initializer_list<LLT> IntTypes32 = {s8, s16, s32, p0};
271 const std::initializer_list<LLT> IntTypes64 = {s8, s16, s32, s64, p0};
275 .clampScalar(0, s8, s8)
282 .legalFor(Is64Bit, {s64})
283 .widenScalarToNextPow2(0, 32)
288 .
legalFor(HasPOPCNT, {{s16, s16}, {s32, s32}})
289 .legalFor(HasPOPCNT && Is64Bit, {{s64, s64}})
290 .widenScalarToNextPow2(1, 16)
296 .
legalFor(HasLZCNT, {{s16, s16}, {s32, s32}})
297 .legalFor(HasLZCNT && Is64Bit, {{s64, s64}})
298 .widenScalarToNextPow2(1, 16)
305 .legalFor(Is64Bit, {{s64, s64}})
306 .widenScalarToNextPow2(1, 16)
311 .
legalFor(HasBMI, {{s16, s16}, {s32, s32}})
312 .legalFor(HasBMI && Is64Bit, {{s64, s64}})
313 .widenScalarToNextPow2(1, 16)
320 const std::initializer_list<LLT> PtrTypes32 = {s1, s8, s16, s32};
321 const std::initializer_list<LLT> PtrTypes64 = {s1, s8, s16, s32, s64};
325 .maxScalar(0, sMaxScalar)
334 .legalFor(Is64Bit, {{p0, s64}})
335 .widenScalarToNextPow2(1, 32)
343 for (
unsigned Op : {G_LOAD, G_STORE}) {
345 Action.legalForTypesWithMemDesc({{s8, p0, s8, 1},
350 {v4s8, p0, v4s8, 1}});
352 Action.legalForTypesWithMemDesc(
353 {{s64, p0, s64, 1}, {v2s32, p0, v2s32, 1}});
356 Action.legalForTypesWithMemDesc({{v4s32, p0, v4s32, 1}});
358 Action.legalForTypesWithMemDesc({{v16s8, p0, v16s8, 1},
359 {v8s16, p0, v8s16, 1},
360 {v2s64, p0, v2s64, 1},
361 {v2p0, p0, v2p0, 1}});
363 Action.legalForTypesWithMemDesc({{v32s8, p0, v32s8, 1},
364 {v16s16, p0, v16s16, 1},
365 {v8s32, p0, v8s32, 1},
366 {v4s64, p0, v4s64, 1},
367 {v4p0, p0, v4p0, 1}});
369 Action.legalForTypesWithMemDesc({{v64s8, p0, v64s8, 1},
370 {v32s16, p0, v32s16, 1},
371 {v16s32, p0, v16s32, 1},
372 {v8s64, p0, v8s64, 1}});
376 Action.legalForTypesWithMemDesc({{s8, p0, s1, 1},
381 Action.legalForTypesWithMemDesc(
382 {{s64, p0, s8, 1}, {s64, p0, s16, 1}, {s64, p0, s32, 1}});
388 Action.widenScalarToNextPow2(0, 8)
389 .clampScalar(0, s8, sMaxScalar)
393 for (
unsigned Op : {G_SEXTLOAD, G_ZEXTLOAD}) {
395 Action.legalForTypesWithMemDesc(
396 {{s16, p0, s8, 1}, {s32, p0, s8, 1}, {s32, p0, s16, 1}});
398 Action.legalForTypesWithMemDesc(
399 {{s64, p0, s8, 1}, {s64, p0, s16, 1}, {s64, p0, s32, 1}});
403 for (
unsigned Op : {G_FPEXTLOAD, G_FPTRUNCSTORE}) {
405 Action.legalForTypesWithMemDesc(
406 UseX87, {{s80, p0, s32, 1}, {s80, p0, s64, 1}, {s64, p0, s32, 1}});
412 .legalFor(Is64Bit, {s64})
413 .widenScalarToNextPow2(0, 8)
420 .legalFor({s8, s16, s32})
421 .legalFor(Is64Bit, {s64})
422 .widenScalarToNextPow2(0, 8)
433 .legalFor(UseX87, {s80});
437 .legalFor({s32, s64})
438 .legalFor(HasSSE1, {v4s32})
439 .legalFor(HasSSE2, {v2s64})
440 .legalFor(HasAVX, {v8s32, v4s64})
441 .legalFor(HasAVX512, {v16s32, v8s64})
442 .legalFor(UseX87, {s80});
446 .legalFor(UseX87 && !Is64Bit, {s64})
451 .
legalFor(HasSSE1 || UseX87, {s8, s32})
452 .legalFor(HasSSE2 || UseX87, {s8, s64})
453 .legalFor(UseX87, {s8, s80})
454 .clampScalar(0, s8, s8)
461 .legalFor(HasAVX, {{v4s64, v4s32}})
462 .legalFor(HasAVX512, {{v8s64, v8s32}})
463 .lowerFor(UseX87, {{s64, s32}, {s80, s32}, {s80, s64}})
468 .legalFor(HasAVX, {{v4s32, v4s64}})
469 .legalFor(HasAVX512, {{v8s32, v8s64}})
470 .lowerFor(UseX87, {{s32, s64}, {s32, s80}, {s64, s80}});
474 .legalFor(HasSSE1 && Is64Bit, {{s32, s64}})
475 .legalFor(HasSSE2, {{s64, s32}})
476 .legalFor(HasSSE2 && Is64Bit, {{s64, s64}})
477 .clampScalar(1, (UseX87 && !HasSSE1) ? s16 : s32, sMaxScalar)
480 .clampScalar(0, s32, HasSSE2 ? s64 : s32)
485 .legalFor(HasSSE1 && Is64Bit, {{s64, s32}})
486 .legalFor(HasSSE2, {{s32, s64}})
487 .legalFor(HasSSE2 && Is64Bit, {{s64, s64}})
488 .clampScalar(0, (UseX87 && !HasSSE1) ? s16 : s32, sMaxScalar)
491 .clampScalar(1, s32, HasSSE2 ? s64 : s32)
502 .
legalFor(HasAVX512, {{s32, s32}, {s32, s64}, {s64, s32}, {s64, s64}})
505 ((HasSSE1 &&
typeIs(0, s32)(Query)) ||
506 (HasSSE2 &&
typeIs(0, s64)(Query))) &&
512 ((HasSSE1 &&
typeIs(0, s32)(Query)) ||
513 (HasSSE2 &&
typeIs(0, s64)(Query))) &&
514 (Is64Bit &&
typeIs(1, s64)(Query));
516 .clampScalar(0, s32, HasSSE2 ? s64 : s32)
522 .
legalFor(HasAVX512, {{s32, s32}, {s32, s64}, {s64, s32}, {s64, s64}})
525 ((HasSSE1 &&
typeIs(1, s32)(Query)) ||
526 (HasSSE2 &&
typeIs(1, s64)(Query))) &&
535 ((HasSSE1 &&
typeIs(1, s32)(Query)) ||
536 (HasSSE2 &&
typeIs(1, s64)(Query))) &&
537 (Is64Bit &&
typeIs(0, s64)(Query));
539 .clampScalar(0, s32, sMaxScalar)
547 return (HasSSE1 &&
typeInSet(0, {v4s32})(Query)) ||
548 (HasSSE2 &&
typeInSet(0, {v2s64, v8s16, v16s8})(Query)) ||
549 (HasAVX &&
typeInSet(0, {v4s64, v8s32, v16s16, v32s8})(Query)) ||
550 (HasAVX512 &&
typeInSet(0, {v8s64, v16s32, v32s16, v64s8}));
552 .clampNumElements(0, v16s8, s8MaxVector)
560 unsigned SubIdx = Query.
Opcode == G_EXTRACT ? 0 : 1;
561 unsigned FullIdx = Query.
Opcode == G_EXTRACT ? 1 : 0;
566 {v2s64, v4s64}})(Query)) ||
575 {v4s64, v8s64}})(Query));
582 {{v32s8, v16s8}, {v16s16, v8s16}, {v8s32, v4s32}, {v4s64, v2s64}})
583 .legalFor(HasAVX, {{v64s8, v16s8},
594 .
legalFor({{s16, s32}, {s32, s32}, {p0, s32}})
595 .legalFor(!HasCMOV, {{s8, s32}})
596 .legalFor(Is64Bit, {{s64, s32}})
597 .legalFor(UseX87, {{s80, s32}})
598 .clampScalar(1, s32, s32)
611 .
lowerFor(Is64Bit, {{s1, s32}, {s1, s64}, {s1, s80}});