1 ;; Predicate definitions for IA-32 and x86-64.
2 ;; Copyright (C) 2004, 2005 Free Software Foundation, Inc.
4 ;; This file is part of GCC.
6 ;; GCC is free software; you can redistribute it and/or modify
7 ;; it under the terms of the GNU General Public License as published by
8 ;; the Free Software Foundation; either version 2, or (at your option)
11 ;; GCC is distributed in the hope that it will be useful,
12 ;; but WITHOUT ANY WARRANTY; without even the implied warranty of
13 ;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 ;; GNU General Public License for more details.
16 ;; You should have received a copy of the GNU General Public License
17 ;; along with GCC; see the file COPYING. If not, write to
18 ;; the Free Software Foundation, 51 Franklin Street, Fifth Floor,
19 ;; Boston, MA 02110-1301, USA.
21 ;; Return nonzero if OP is either a i387 or SSE fp register.
22 (define_predicate "any_fp_register_operand"
23 (and (match_code "reg")
24 (match_test "ANY_FP_REGNO_P (REGNO (op))")))
26 ;; Return nonzero if OP is an i387 fp register.
27 (define_predicate "fp_register_operand"
28 (and (match_code "reg")
29 (match_test "FP_REGNO_P (REGNO (op))")))
31 ;; Return nonzero if OP is a non-fp register_operand.
32 (define_predicate "register_and_not_any_fp_reg_operand"
33 (and (match_code "reg")
34 (not (match_test "ANY_FP_REGNO_P (REGNO (op))"))))
36 ;; Return nonzero if OP is a register operand other than an i387 fp register.
37 (define_predicate "register_and_not_fp_reg_operand"
38 (and (match_code "reg")
39 (not (match_test "FP_REGNO_P (REGNO (op))"))))
41 ;; True if the operand is an MMX register.
42 (define_predicate "mmx_reg_operand"
43 (and (match_code "reg")
44 (match_test "MMX_REGNO_P (REGNO (op))")))
46 ;; True if the operand is a Q_REGS class register.
47 (define_predicate "q_regs_operand"
48 (match_operand 0 "register_operand")
50 if (GET_CODE (op) == SUBREG)
52 return ANY_QI_REG_P (op);
55 ;; Return true if op is a NON_Q_REGS class register.
56 (define_predicate "non_q_regs_operand"
57 (match_operand 0 "register_operand")
59 if (GET_CODE (op) == SUBREG)
61 return NON_QI_REG_P (op);
64 ;; Match an SI or HImode register for a zero_extract.
65 (define_special_predicate "ext_register_operand"
66 (match_operand 0 "register_operand")
68 if ((!TARGET_64BIT || GET_MODE (op) != DImode)
69 && GET_MODE (op) != SImode && GET_MODE (op) != HImode)
71 if (GET_CODE (op) == SUBREG)
74 /* Be careful to accept only registers having upper parts. */
75 return REGNO (op) > LAST_VIRTUAL_REGISTER || REGNO (op) < 4;
78 ;; Return true if op is the flags register.
79 (define_predicate "flags_reg_operand"
80 (and (match_code "reg")
81 (match_test "REGNO (op) == FLAGS_REG")))
83 ;; Return 1 if VALUE can be stored in a sign extended immediate field.
84 (define_predicate "x86_64_immediate_operand"
85 (match_code "const_int,symbol_ref,label_ref,const")
88 return immediate_operand (op, mode);
90 switch (GET_CODE (op))
93 /* CONST_DOUBLES never match, since HOST_BITS_PER_WIDE_INT is known
94 to be at least 32 and this all acceptable constants are
95 represented as CONST_INT. */
96 if (HOST_BITS_PER_WIDE_INT == 32)
100 HOST_WIDE_INT val = trunc_int_for_mode (INTVAL (op), DImode);
101 return trunc_int_for_mode (val, SImode) == val;
106 /* For certain code models, the symbolic references are known to fit.
107 in CM_SMALL_PIC model we know it fits if it is local to the shared
108 library. Don't count TLS SYMBOL_REFs here, since they should fit
109 only if inside of UNSPEC handled below. */
110 /* TLS symbols are not constant. */
111 if (SYMBOL_REF_TLS_MODEL (op))
113 return (ix86_cmodel == CM_SMALL || ix86_cmodel == CM_KERNEL
114 || (ix86_cmodel == CM_MEDIUM && !SYMBOL_REF_FAR_ADDR_P (op)));
117 /* For certain code models, the code is near as well. */
118 return (ix86_cmodel == CM_SMALL || ix86_cmodel == CM_MEDIUM
119 || ix86_cmodel == CM_KERNEL);
122 /* We also may accept the offsetted memory references in certain
124 if (GET_CODE (XEXP (op, 0)) == UNSPEC)
125 switch (XINT (XEXP (op, 0), 1))
127 case UNSPEC_GOTPCREL:
129 case UNSPEC_GOTNTPOFF:
136 if (GET_CODE (XEXP (op, 0)) == PLUS)
138 rtx op1 = XEXP (XEXP (op, 0), 0);
139 rtx op2 = XEXP (XEXP (op, 0), 1);
140 HOST_WIDE_INT offset;
142 if (ix86_cmodel == CM_LARGE)
144 if (GET_CODE (op2) != CONST_INT)
146 offset = trunc_int_for_mode (INTVAL (op2), DImode);
147 switch (GET_CODE (op1))
150 /* TLS symbols are not constant. */
151 if (SYMBOL_REF_TLS_MODEL (op1))
153 /* For CM_SMALL assume that latest object is 16MB before
154 end of 31bits boundary. We may also accept pretty
155 large negative constants knowing that all objects are
156 in the positive half of address space. */
157 if ((ix86_cmodel == CM_SMALL
158 || (ix86_cmodel == CM_MEDIUM
159 && !SYMBOL_REF_FAR_ADDR_P (op1)))
160 && offset < 16*1024*1024
161 && trunc_int_for_mode (offset, SImode) == offset)
163 /* For CM_KERNEL we know that all object resist in the
164 negative half of 32bits address space. We may not
165 accept negative offsets, since they may be just off
166 and we may accept pretty large positive ones. */
167 if (ix86_cmodel == CM_KERNEL
169 && trunc_int_for_mode (offset, SImode) == offset)
174 /* These conditions are similar to SYMBOL_REF ones, just the
175 constraints for code models differ. */
176 if ((ix86_cmodel == CM_SMALL || ix86_cmodel == CM_MEDIUM)
177 && offset < 16*1024*1024
178 && trunc_int_for_mode (offset, SImode) == offset)
180 if (ix86_cmodel == CM_KERNEL
182 && trunc_int_for_mode (offset, SImode) == offset)
187 switch (XINT (op1, 1))
192 && trunc_int_for_mode (offset, SImode) == offset)
210 ;; Return 1 if VALUE can be stored in the zero extended immediate field.
211 (define_predicate "x86_64_zext_immediate_operand"
212 (match_code "const_double,const_int,symbol_ref,label_ref,const")
214 switch (GET_CODE (op))
217 if (HOST_BITS_PER_WIDE_INT == 32)
218 return (GET_MODE (op) == VOIDmode && !CONST_DOUBLE_HIGH (op));
223 if (HOST_BITS_PER_WIDE_INT == 32)
224 return INTVAL (op) >= 0;
226 return !(INTVAL (op) & ~(HOST_WIDE_INT) 0xffffffff);
229 /* For certain code models, the symbolic references are known to fit. */
230 /* TLS symbols are not constant. */
231 if (SYMBOL_REF_TLS_MODEL (op))
233 return (ix86_cmodel == CM_SMALL
234 || (ix86_cmodel == CM_MEDIUM
235 && !SYMBOL_REF_FAR_ADDR_P (op)));
238 /* For certain code models, the code is near as well. */
239 return ix86_cmodel == CM_SMALL || ix86_cmodel == CM_MEDIUM;
242 /* We also may accept the offsetted memory references in certain
244 if (GET_CODE (XEXP (op, 0)) == PLUS)
246 rtx op1 = XEXP (XEXP (op, 0), 0);
247 rtx op2 = XEXP (XEXP (op, 0), 1);
249 if (ix86_cmodel == CM_LARGE)
251 switch (GET_CODE (op1))
254 /* TLS symbols are not constant. */
255 if (SYMBOL_REF_TLS_MODEL (op1))
257 /* For small code model we may accept pretty large positive
258 offsets, since one bit is available for free. Negative
259 offsets are limited by the size of NULL pointer area
260 specified by the ABI. */
261 if ((ix86_cmodel == CM_SMALL
262 || (ix86_cmodel == CM_MEDIUM
263 && !SYMBOL_REF_FAR_ADDR_P (op1)))
264 && GET_CODE (op2) == CONST_INT
265 && trunc_int_for_mode (INTVAL (op2), DImode) > -0x10000
266 && trunc_int_for_mode (INTVAL (op2), SImode) == INTVAL (op2))
268 /* ??? For the kernel, we may accept adjustment of
269 -0x10000000, since we know that it will just convert
270 negative address space to positive, but perhaps this
271 is not worthwhile. */
275 /* These conditions are similar to SYMBOL_REF ones, just the
276 constraints for code models differ. */
277 if ((ix86_cmodel == CM_SMALL || ix86_cmodel == CM_MEDIUM)
278 && GET_CODE (op2) == CONST_INT
279 && trunc_int_for_mode (INTVAL (op2), DImode) > -0x10000
280 && trunc_int_for_mode (INTVAL (op2), SImode) == INTVAL (op2))
296 ;; Return nonzero if OP is general operand representable on x86_64.
297 (define_predicate "x86_64_general_operand"
298 (if_then_else (match_test "TARGET_64BIT")
299 (ior (match_operand 0 "nonimmediate_operand")
300 (match_operand 0 "x86_64_immediate_operand"))
301 (match_operand 0 "general_operand")))
303 ;; Return nonzero if OP is general operand representable on x86_64
304 ;; as either sign extended or zero extended constant.
305 (define_predicate "x86_64_szext_general_operand"
306 (if_then_else (match_test "TARGET_64BIT")
307 (ior (match_operand 0 "nonimmediate_operand")
308 (ior (match_operand 0 "x86_64_immediate_operand")
309 (match_operand 0 "x86_64_zext_immediate_operand")))
310 (match_operand 0 "general_operand")))
312 ;; Return nonzero if OP is nonmemory operand representable on x86_64.
313 (define_predicate "x86_64_nonmemory_operand"
314 (if_then_else (match_test "TARGET_64BIT")
315 (ior (match_operand 0 "register_operand")
316 (match_operand 0 "x86_64_immediate_operand"))
317 (match_operand 0 "nonmemory_operand")))
319 ;; Return nonzero if OP is nonmemory operand representable on x86_64.
320 (define_predicate "x86_64_szext_nonmemory_operand"
321 (if_then_else (match_test "TARGET_64BIT")
322 (ior (match_operand 0 "register_operand")
323 (ior (match_operand 0 "x86_64_immediate_operand")
324 (match_operand 0 "x86_64_zext_immediate_operand")))
325 (match_operand 0 "nonmemory_operand")))
327 ;; Return true when operand is PIC expression that can be computed by lea
329 (define_predicate "pic_32bit_operand"
330 (match_code "const,symbol_ref,label_ref")
334 /* Rule out relocations that translate into 64bit constants. */
335 if (TARGET_64BIT && GET_CODE (op) == CONST)
338 if (GET_CODE (op) == PLUS && GET_CODE (XEXP (op, 1)) == CONST_INT)
340 if (GET_CODE (op) == UNSPEC
341 && (XINT (op, 1) == UNSPEC_GOTOFF
342 || XINT (op, 1) == UNSPEC_GOT))
345 return symbolic_operand (op, mode);
349 ;; Return nonzero if OP is nonmemory operand acceptable by movabs patterns.
350 (define_predicate "x86_64_movabs_operand"
351 (if_then_else (match_test "!TARGET_64BIT || !flag_pic")
352 (match_operand 0 "nonmemory_operand")
353 (ior (match_operand 0 "register_operand")
354 (and (match_operand 0 "const_double_operand")
355 (match_test "GET_MODE_SIZE (mode) <= 8")))))
357 ;; Returns nonzero if OP is either a symbol reference or a sum of a symbol
358 ;; reference and a constant.
359 (define_predicate "symbolic_operand"
360 (match_code "symbol_ref,label_ref,const")
362 switch (GET_CODE (op))
370 if (GET_CODE (op) == SYMBOL_REF
371 || GET_CODE (op) == LABEL_REF
372 || (GET_CODE (op) == UNSPEC
373 && (XINT (op, 1) == UNSPEC_GOT
374 || XINT (op, 1) == UNSPEC_GOTOFF
375 || XINT (op, 1) == UNSPEC_GOTPCREL)))
377 if (GET_CODE (op) != PLUS
378 || GET_CODE (XEXP (op, 1)) != CONST_INT)
382 if (GET_CODE (op) == SYMBOL_REF
383 || GET_CODE (op) == LABEL_REF)
385 /* Only @GOTOFF gets offsets. */
386 if (GET_CODE (op) != UNSPEC
387 || XINT (op, 1) != UNSPEC_GOTOFF)
390 op = XVECEXP (op, 0, 0);
391 if (GET_CODE (op) == SYMBOL_REF
392 || GET_CODE (op) == LABEL_REF)
401 ;; Return true if the operand contains a @GOT or @GOTOFF reference.
402 (define_predicate "pic_symbolic_operand"
408 if (GET_CODE (op) == UNSPEC
409 && XINT (op, 1) == UNSPEC_GOTPCREL)
411 if (GET_CODE (op) == PLUS
412 && GET_CODE (XEXP (op, 0)) == UNSPEC
413 && XINT (XEXP (op, 0), 1) == UNSPEC_GOTPCREL)
418 if (GET_CODE (op) == UNSPEC)
420 if (GET_CODE (op) != PLUS
421 || GET_CODE (XEXP (op, 1)) != CONST_INT)
424 if (GET_CODE (op) == UNSPEC)
430 ;; Return true if OP is a symbolic operand that resolves locally.
431 (define_predicate "local_symbolic_operand"
432 (match_code "const,label_ref,symbol_ref")
434 if (GET_CODE (op) == CONST
435 && GET_CODE (XEXP (op, 0)) == PLUS
436 && GET_CODE (XEXP (XEXP (op, 0), 1)) == CONST_INT)
437 op = XEXP (XEXP (op, 0), 0);
439 if (GET_CODE (op) == LABEL_REF)
442 if (GET_CODE (op) != SYMBOL_REF)
445 if (SYMBOL_REF_TLS_MODEL (op) != 0)
448 if (SYMBOL_REF_LOCAL_P (op))
451 /* There is, however, a not insubstantial body of code in the rest of
452 the compiler that assumes it can just stick the results of
453 ASM_GENERATE_INTERNAL_LABEL in a symbol_ref and have done. */
454 /* ??? This is a hack. Should update the body of the compiler to
455 always create a DECL an invoke targetm.encode_section_info. */
456 if (strncmp (XSTR (op, 0), internal_label_prefix,
457 internal_label_prefix_len) == 0)
463 ;; Test for various thread-local symbols.
464 (define_predicate "tls_symbolic_operand"
465 (and (match_code "symbol_ref")
466 (match_test "SYMBOL_REF_TLS_MODEL (op) != 0")))
468 ;; Test for a pc-relative call operand
469 (define_predicate "constant_call_address_operand"
470 (ior (match_code "symbol_ref")
471 (match_operand 0 "local_symbolic_operand")))
473 ;; True for any non-virtual or eliminable register. Used in places where
474 ;; instantiation of such a register may cause the pattern to not be recognized.
475 (define_predicate "register_no_elim_operand"
476 (match_operand 0 "register_operand")
478 if (GET_CODE (op) == SUBREG)
479 op = SUBREG_REG (op);
480 return !(op == arg_pointer_rtx
481 || op == frame_pointer_rtx
482 || (REGNO (op) >= FIRST_PSEUDO_REGISTER
483 && REGNO (op) <= LAST_VIRTUAL_REGISTER));
486 ;; Similarly, but include the stack pointer. This is used to prevent esp
487 ;; from being used as an index reg.
488 (define_predicate "index_register_operand"
489 (match_operand 0 "register_operand")
491 if (GET_CODE (op) == SUBREG)
492 op = SUBREG_REG (op);
493 if (reload_in_progress || reload_completed)
494 return REG_OK_FOR_INDEX_STRICT_P (op);
496 return REG_OK_FOR_INDEX_NONSTRICT_P (op);
499 ;; Return false if this is any eliminable register. Otherwise general_operand.
500 (define_predicate "general_no_elim_operand"
501 (if_then_else (match_code "reg,subreg")
502 (match_operand 0 "register_no_elim_operand")
503 (match_operand 0 "general_operand")))
505 ;; Return false if this is any eliminable register. Otherwise
506 ;; register_operand or a constant.
507 (define_predicate "nonmemory_no_elim_operand"
508 (ior (match_operand 0 "register_no_elim_operand")
509 (match_operand 0 "immediate_operand")))
511 ;; Test for a valid operand for a call instruction.
512 (define_predicate "call_insn_operand"
513 (ior (match_operand 0 "constant_call_address_operand")
514 (ior (match_operand 0 "register_no_elim_operand")
515 (match_operand 0 "memory_operand"))))
517 ;; Similarly, but for tail calls, in which we cannot allow memory references.
518 (define_predicate "sibcall_insn_operand"
519 (ior (match_operand 0 "constant_call_address_operand")
520 (match_operand 0 "register_no_elim_operand")))
522 ;; Match exactly zero.
523 (define_predicate "const0_operand"
524 (match_code "const_int,const_double,const_vector")
526 if (mode == VOIDmode)
527 mode = GET_MODE (op);
528 return op == CONST0_RTX (mode);
531 ;; Match exactly one.
532 (define_predicate "const1_operand"
533 (and (match_code "const_int")
534 (match_test "op == const1_rtx")))
536 ;; Match exactly eight.
537 (define_predicate "const8_operand"
538 (and (match_code "const_int")
539 (match_test "INTVAL (op) == 8")))
541 ;; Match 2, 4, or 8. Used for leal multiplicands.
542 (define_predicate "const248_operand"
543 (match_code "const_int")
545 HOST_WIDE_INT i = INTVAL (op);
546 return i == 2 || i == 4 || i == 8;
550 (define_predicate "const_0_to_1_operand"
551 (and (match_code "const_int")
552 (match_test "op == const0_rtx || op == const1_rtx")))
555 (define_predicate "const_0_to_3_operand"
556 (and (match_code "const_int")
557 (match_test "INTVAL (op) >= 0 && INTVAL (op) <= 3")))
560 (define_predicate "const_0_to_7_operand"
561 (and (match_code "const_int")
562 (match_test "INTVAL (op) >= 0 && INTVAL (op) <= 7")))
565 (define_predicate "const_0_to_15_operand"
566 (and (match_code "const_int")
567 (match_test "INTVAL (op) >= 0 && INTVAL (op) <= 15")))
570 (define_predicate "const_0_to_63_operand"
571 (and (match_code "const_int")
572 (match_test "INTVAL (op) >= 0 && INTVAL (op) <= 63")))
575 (define_predicate "const_0_to_255_operand"
576 (and (match_code "const_int")
577 (match_test "INTVAL (op) >= 0 && INTVAL (op) <= 255")))
579 ;; Match (0 to 255) * 8
580 (define_predicate "const_0_to_255_mul_8_operand"
581 (match_code "const_int")
583 unsigned HOST_WIDE_INT val = INTVAL (op);
584 return val <= 255*8 && val % 8 == 0;
587 ;; Return nonzero if OP is CONST_INT >= 1 and <= 31 (a valid operand
588 ;; for shift & compare patterns, as shifting by 0 does not change flags).
589 (define_predicate "const_1_to_31_operand"
590 (and (match_code "const_int")
591 (match_test "INTVAL (op) >= 1 && INTVAL (op) <= 31")))
594 (define_predicate "const_2_to_3_operand"
595 (and (match_code "const_int")
596 (match_test "INTVAL (op) == 2 || INTVAL (op) == 3")))
599 (define_predicate "const_4_to_7_operand"
600 (and (match_code "const_int")
601 (match_test "INTVAL (op) >= 4 && INTVAL (op) <= 7")))
603 ;; Match exactly one bit in 4-bit mask.
604 (define_predicate "const_pow2_1_to_8_operand"
605 (match_code "const_int")
607 unsigned int log = exact_log2 (INTVAL (op));
611 ;; Match exactly one bit in 8-bit mask.
612 (define_predicate "const_pow2_1_to_128_operand"
613 (match_code "const_int")
615 unsigned int log = exact_log2 (INTVAL (op));
619 ;; True if this is a constant appropriate for an increment or decrement.
620 (define_predicate "incdec_operand"
621 (match_code "const_int")
623 /* On Pentium4, the inc and dec operations causes extra dependency on flag
624 registers, since carry flag is not set. */
625 if ((TARGET_PENTIUM4 || TARGET_NOCONA) && !optimize_size)
627 return op == const1_rtx || op == constm1_rtx;
630 ;; True for registers, or 1 or -1. Used to optimize double-word shifts.
631 (define_predicate "reg_or_pm1_operand"
632 (ior (match_operand 0 "register_operand")
633 (and (match_code "const_int")
634 (match_test "op == const1_rtx || op == constm1_rtx"))))
636 ;; True if OP is acceptable as operand of DImode shift expander.
637 (define_predicate "shiftdi_operand"
638 (if_then_else (match_test "TARGET_64BIT")
639 (match_operand 0 "nonimmediate_operand")
640 (match_operand 0 "register_operand")))
642 (define_predicate "ashldi_input_operand"
643 (if_then_else (match_test "TARGET_64BIT")
644 (match_operand 0 "nonimmediate_operand")
645 (match_operand 0 "reg_or_pm1_operand")))
647 ;; Return true if OP is a vector load from the constant pool with just
648 ;; the first element nonzero.
649 (define_predicate "zero_extended_scalar_load_operand"
653 op = maybe_get_pool_constant (op);
656 if (GET_CODE (op) != CONST_VECTOR)
659 (GET_MODE_SIZE (GET_MODE (op)) /
660 GET_MODE_SIZE (GET_MODE_INNER (GET_MODE (op))));
661 for (n_elts--; n_elts > 0; n_elts--)
663 rtx elt = CONST_VECTOR_ELT (op, n_elts);
664 if (elt != CONST0_RTX (GET_MODE_INNER (GET_MODE (op))))
670 ;; Return 1 when OP is operand acceptable for standard SSE move.
671 (define_predicate "vector_move_operand"
672 (ior (match_operand 0 "nonimmediate_operand")
673 (match_operand 0 "const0_operand")))
675 ;; Return true if OP is a register or a zero.
676 (define_predicate "reg_or_0_operand"
677 (ior (match_operand 0 "register_operand")
678 (match_operand 0 "const0_operand")))
680 ;; Return true if op if a valid address, and does not contain
681 ;; a segment override.
682 (define_special_predicate "no_seg_address_operand"
683 (match_operand 0 "address_operand")
685 struct ix86_address parts;
688 ok = ix86_decompose_address (op, &parts);
690 return parts.seg == SEG_DEFAULT;
693 ;; Return nonzero if the rtx is known to be at least 32 bits aligned.
694 (define_predicate "aligned_operand"
695 (match_operand 0 "general_operand")
697 struct ix86_address parts;
700 /* Registers and immediate operands are always "aligned". */
701 if (GET_CODE (op) != MEM)
704 /* Don't even try to do any aligned optimizations with volatiles. */
705 if (MEM_VOLATILE_P (op))
708 if (MEM_ALIGN (op) >= 32)
713 /* Pushes and pops are only valid on the stack pointer. */
714 if (GET_CODE (op) == PRE_DEC
715 || GET_CODE (op) == POST_INC)
718 /* Decode the address. */
719 ok = ix86_decompose_address (op, &parts);
722 /* Look for some component that isn't known to be aligned. */
725 if (REGNO_POINTER_ALIGN (REGNO (parts.index)) * parts.scale < 32)
730 if (REGNO_POINTER_ALIGN (REGNO (parts.base)) < 32)
735 if (GET_CODE (parts.disp) != CONST_INT
736 || (INTVAL (parts.disp) & 3) != 0)
740 /* Didn't find one -- this must be an aligned address. */
744 ;; Returns 1 if OP is memory operand with a displacement.
745 (define_predicate "memory_displacement_operand"
746 (match_operand 0 "memory_operand")
748 struct ix86_address parts;
751 ok = ix86_decompose_address (XEXP (op, 0), &parts);
753 return parts.disp != NULL_RTX;
756 ;; Returns 1 if OP is memory operand that cannot be represented
757 ;; by the modRM array.
758 (define_predicate "long_memory_operand"
759 (and (match_operand 0 "memory_operand")
760 (match_test "memory_address_length (op) != 0")))
762 ;; Return 1 if OP is a comparison operator that can be issued by fcmov.
763 (define_predicate "fcmov_comparison_operator"
764 (match_operand 0 "comparison_operator")
766 enum machine_mode inmode = GET_MODE (XEXP (op, 0));
767 enum rtx_code code = GET_CODE (op);
769 if (inmode == CCFPmode || inmode == CCFPUmode)
771 enum rtx_code second_code, bypass_code;
772 ix86_fp_comparison_codes (code, &bypass_code, &code, &second_code);
773 if (bypass_code != UNKNOWN || second_code != UNKNOWN)
775 code = ix86_fp_compare_code_to_integer (code);
777 /* i387 supports just limited amount of conditional codes. */
780 case LTU: case GTU: case LEU: case GEU:
781 if (inmode == CCmode || inmode == CCFPmode || inmode == CCFPUmode)
784 case ORDERED: case UNORDERED:
792 ;; Return 1 if OP is a comparison that can be used in the CMPSS/CMPPS insns.
793 ;; The first set are supported directly; the second set can't be done with
794 ;; full IEEE support, i.e. NaNs.
796 ;; ??? It would seem that we have a lot of uses of this predicate that pass
797 ;; it the wrong mode. We got away with this because the old function didn't
798 ;; check the mode at all. Mirror that for now by calling this a special
801 (define_special_predicate "sse_comparison_operator"
802 (match_code "eq,lt,le,unordered,ne,unge,ungt,ordered"))
804 ;; Return 1 if OP is a valid comparison operator in valid mode.
805 (define_predicate "ix86_comparison_operator"
806 (match_operand 0 "comparison_operator")
808 enum machine_mode inmode = GET_MODE (XEXP (op, 0));
809 enum rtx_code code = GET_CODE (op);
811 if (inmode == CCFPmode || inmode == CCFPUmode)
813 enum rtx_code second_code, bypass_code;
814 ix86_fp_comparison_codes (code, &bypass_code, &code, &second_code);
815 return (bypass_code == UNKNOWN && second_code == UNKNOWN);
822 if (inmode == CCmode || inmode == CCGCmode
823 || inmode == CCGOCmode || inmode == CCNOmode)
826 case LTU: case GTU: case LEU: case ORDERED: case UNORDERED: case GEU:
827 if (inmode == CCmode)
831 if (inmode == CCmode || inmode == CCGCmode || inmode == CCNOmode)
839 ;; Return 1 if OP is a valid comparison operator testing carry flag to be set.
840 (define_predicate "ix86_carry_flag_operator"
841 (match_code "ltu,lt,unlt,gt,ungt,le,unle,ge,unge,ltgt,uneq")
843 enum machine_mode inmode = GET_MODE (XEXP (op, 0));
844 enum rtx_code code = GET_CODE (op);
846 if (GET_CODE (XEXP (op, 0)) != REG
847 || REGNO (XEXP (op, 0)) != FLAGS_REG
848 || XEXP (op, 1) != const0_rtx)
851 if (inmode == CCFPmode || inmode == CCFPUmode)
853 enum rtx_code second_code, bypass_code;
854 ix86_fp_comparison_codes (code, &bypass_code, &code, &second_code);
855 if (bypass_code != UNKNOWN || second_code != UNKNOWN)
857 code = ix86_fp_compare_code_to_integer (code);
859 else if (inmode != CCmode)
865 ;; Nearly general operand, but accept any const_double, since we wish
866 ;; to be able to drop them into memory rather than have them get pulled
868 (define_predicate "cmp_fp_expander_operand"
869 (ior (match_code "const_double")
870 (match_operand 0 "general_operand")))
872 ;; Return true if this is a valid binary floating-point operation.
873 (define_predicate "binary_fp_operator"
874 (match_code "plus,minus,mult,div"))
876 ;; Return true if this is a multiply operation.
877 (define_predicate "mult_operator"
880 ;; Return true if this is a division operation.
881 (define_predicate "div_operator"
884 ;; Return true if this is a float extend operation.
885 (define_predicate "float_operator"
886 (match_code "float"))
888 ;; Return true for ARITHMETIC_P.
889 (define_predicate "arith_or_logical_operator"
890 (match_code "plus,mult,and,ior,xor,smin,smax,umin,umax,compare,minus,div,
891 mod,udiv,umod,ashift,rotate,ashiftrt,lshiftrt,rotatert"))
893 ;; Return 1 if OP is a binary operator that can be promoted to wider mode.
894 ;; Modern CPUs have same latency for HImode and SImode multiply,
895 ;; but 386 and 486 do HImode multiply faster. */
896 (define_predicate "promotable_binary_operator"
897 (ior (match_code "plus,and,ior,xor,ashift")
898 (and (match_code "mult")
899 (match_test "ix86_tune > PROCESSOR_I486"))))
901 ;; To avoid problems when jump re-emits comparisons like testqi_ext_ccno_0,
902 ;; re-recognize the operand to avoid a copy_to_mode_reg that will fail.
904 ;; ??? It seems likely that this will only work because cmpsi is an
905 ;; expander, and no actual insns use this.
907 (define_predicate "cmpsi_operand_1"
910 return (GET_MODE (op) == SImode
911 && GET_CODE (XEXP (op, 0)) == ZERO_EXTRACT
912 && GET_CODE (XEXP (XEXP (op, 0), 1)) == CONST_INT
913 && GET_CODE (XEXP (XEXP (op, 0), 2)) == CONST_INT
914 && INTVAL (XEXP (XEXP (op, 0), 1)) == 8
915 && INTVAL (XEXP (XEXP (op, 0), 2)) == 8
916 && GET_CODE (XEXP (op, 1)) == CONST_INT);
919 (define_predicate "cmpsi_operand"
920 (ior (match_operand 0 "nonimmediate_operand")
921 (match_operand 0 "cmpsi_operand_1")))
923 (define_predicate "compare_operator"
924 (match_code "compare"))
926 (define_predicate "absneg_operator"
927 (match_code "abs,neg"))