alias.c (get_alias_set, [...]): Use MEM_P.
[gcc.git] / gcc / ifcvt.c
1 /* If-conversion support.
2 Copyright (C) 2000, 2001, 2002, 2003, 2004 Free Software Foundation, Inc.
3
4 This file is part of GCC.
5
6 GCC is free software; you can redistribute it and/or modify it
7 under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 2, or (at your option)
9 any later version.
10
11 GCC is distributed in the hope that it will be useful, but WITHOUT
12 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
13 or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
14 License for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING. If not, write to the Free
18 Software Foundation, 59 Temple Place - Suite 330, Boston, MA
19 02111-1307, USA. */
20
21 #include "config.h"
22 #include "system.h"
23 #include "coretypes.h"
24 #include "tm.h"
25
26 #include "rtl.h"
27 #include "regs.h"
28 #include "function.h"
29 #include "flags.h"
30 #include "insn-config.h"
31 #include "recog.h"
32 #include "except.h"
33 #include "hard-reg-set.h"
34 #include "basic-block.h"
35 #include "expr.h"
36 #include "real.h"
37 #include "output.h"
38 #include "optabs.h"
39 #include "toplev.h"
40 #include "tm_p.h"
41 #include "cfgloop.h"
42 #include "target.h"
43
44
45 #ifndef HAVE_conditional_execution
46 #define HAVE_conditional_execution 0
47 #endif
48 #ifndef HAVE_conditional_move
49 #define HAVE_conditional_move 0
50 #endif
51 #ifndef HAVE_incscc
52 #define HAVE_incscc 0
53 #endif
54 #ifndef HAVE_decscc
55 #define HAVE_decscc 0
56 #endif
57 #ifndef HAVE_trap
58 #define HAVE_trap 0
59 #endif
60 #ifndef HAVE_conditional_trap
61 #define HAVE_conditional_trap 0
62 #endif
63
64 #ifndef MAX_CONDITIONAL_EXECUTE
65 #define MAX_CONDITIONAL_EXECUTE (BRANCH_COST + 1)
66 #endif
67
68 #define NULL_EDGE ((struct edge_def *)NULL)
69 #define NULL_BLOCK ((struct basic_block_def *)NULL)
70
71 /* # of IF-THEN or IF-THEN-ELSE blocks we looked at */
72 static int num_possible_if_blocks;
73
74 /* # of IF-THEN or IF-THEN-ELSE blocks were converted to conditional
75 execution. */
76 static int num_updated_if_blocks;
77
78 /* # of changes made which require life information to be updated. */
79 static int num_true_changes;
80
81 /* Whether conditional execution changes were made. */
82 static int cond_exec_changed_p;
83
84 /* True if life data ok at present. */
85 static bool life_data_ok;
86
87 /* Forward references. */
88 static int count_bb_insns (basic_block);
89 static rtx first_active_insn (basic_block);
90 static rtx last_active_insn (basic_block, int);
91 static basic_block block_fallthru (basic_block);
92 static int cond_exec_process_insns (ce_if_block_t *, rtx, rtx, rtx, rtx, int);
93 static rtx cond_exec_get_condition (rtx);
94 static int cond_exec_process_if_block (ce_if_block_t *, int);
95 static rtx noce_get_condition (rtx, rtx *);
96 static int noce_operand_ok (rtx);
97 static int noce_process_if_block (ce_if_block_t *);
98 static int process_if_block (ce_if_block_t *);
99 static void merge_if_block (ce_if_block_t *);
100 static int find_cond_trap (basic_block, edge, edge);
101 static basic_block find_if_header (basic_block, int);
102 static int block_jumps_and_fallthru_p (basic_block, basic_block);
103 static int find_if_block (ce_if_block_t *);
104 static int find_if_case_1 (basic_block, edge, edge);
105 static int find_if_case_2 (basic_block, edge, edge);
106 static int find_memory (rtx *, void *);
107 static int dead_or_predicable (basic_block, basic_block, basic_block,
108 basic_block, int);
109 static void noce_emit_move_insn (rtx, rtx);
110 static rtx block_has_only_trap (basic_block);
111 static void mark_loop_exit_edges (void);
112 \f
113 /* Sets EDGE_LOOP_EXIT flag for all loop exits. */
114 static void
115 mark_loop_exit_edges (void)
116 {
117 struct loops loops;
118 basic_block bb;
119 edge e;
120
121 flow_loops_find (&loops, LOOP_TREE);
122 free_dominance_info (CDI_DOMINATORS);
123
124 if (loops.num > 1)
125 {
126 FOR_EACH_BB (bb)
127 {
128 for (e = bb->succ; e; e = e->succ_next)
129 {
130 if (find_common_loop (bb->loop_father, e->dest->loop_father)
131 != bb->loop_father)
132 e->flags |= EDGE_LOOP_EXIT;
133 else
134 e->flags &= ~EDGE_LOOP_EXIT;
135 }
136 }
137 }
138
139 flow_loops_free (&loops);
140 }
141
142 /* Count the number of non-jump active insns in BB. */
143
144 static int
145 count_bb_insns (basic_block bb)
146 {
147 int count = 0;
148 rtx insn = BB_HEAD (bb);
149
150 while (1)
151 {
152 if (GET_CODE (insn) == CALL_INSN || GET_CODE (insn) == INSN)
153 count++;
154
155 if (insn == BB_END (bb))
156 break;
157 insn = NEXT_INSN (insn);
158 }
159
160 return count;
161 }
162
163 /* Return the first non-jump active insn in the basic block. */
164
165 static rtx
166 first_active_insn (basic_block bb)
167 {
168 rtx insn = BB_HEAD (bb);
169
170 if (GET_CODE (insn) == CODE_LABEL)
171 {
172 if (insn == BB_END (bb))
173 return NULL_RTX;
174 insn = NEXT_INSN (insn);
175 }
176
177 while (GET_CODE (insn) == NOTE)
178 {
179 if (insn == BB_END (bb))
180 return NULL_RTX;
181 insn = NEXT_INSN (insn);
182 }
183
184 if (GET_CODE (insn) == JUMP_INSN)
185 return NULL_RTX;
186
187 return insn;
188 }
189
190 /* Return the last non-jump active (non-jump) insn in the basic block. */
191
192 static rtx
193 last_active_insn (basic_block bb, int skip_use_p)
194 {
195 rtx insn = BB_END (bb);
196 rtx head = BB_HEAD (bb);
197
198 while (GET_CODE (insn) == NOTE
199 || GET_CODE (insn) == JUMP_INSN
200 || (skip_use_p
201 && GET_CODE (insn) == INSN
202 && GET_CODE (PATTERN (insn)) == USE))
203 {
204 if (insn == head)
205 return NULL_RTX;
206 insn = PREV_INSN (insn);
207 }
208
209 if (GET_CODE (insn) == CODE_LABEL)
210 return NULL_RTX;
211
212 return insn;
213 }
214
215 /* Return the basic block reached by falling though the basic block BB. */
216
217 static basic_block
218 block_fallthru (basic_block bb)
219 {
220 edge e;
221
222 for (e = bb->succ;
223 e != NULL_EDGE && (e->flags & EDGE_FALLTHRU) == 0;
224 e = e->succ_next)
225 ;
226
227 return (e) ? e->dest : NULL_BLOCK;
228 }
229 \f
230 /* Go through a bunch of insns, converting them to conditional
231 execution format if possible. Return TRUE if all of the non-note
232 insns were processed. */
233
234 static int
235 cond_exec_process_insns (ce_if_block_t *ce_info ATTRIBUTE_UNUSED,
236 /* if block information */rtx start,
237 /* first insn to look at */rtx end,
238 /* last insn to look at */rtx test,
239 /* conditional execution test */rtx prob_val,
240 /* probability of branch taken. */int mod_ok)
241 {
242 int must_be_last = FALSE;
243 rtx insn;
244 rtx xtest;
245 rtx pattern;
246
247 if (!start || !end)
248 return FALSE;
249
250 for (insn = start; ; insn = NEXT_INSN (insn))
251 {
252 if (GET_CODE (insn) == NOTE)
253 goto insn_done;
254
255 if (GET_CODE (insn) != INSN && GET_CODE (insn) != CALL_INSN)
256 abort ();
257
258 /* Remove USE insns that get in the way. */
259 if (reload_completed && GET_CODE (PATTERN (insn)) == USE)
260 {
261 /* ??? Ug. Actually unlinking the thing is problematic,
262 given what we'd have to coordinate with our callers. */
263 SET_INSN_DELETED (insn);
264 goto insn_done;
265 }
266
267 /* Last insn wasn't last? */
268 if (must_be_last)
269 return FALSE;
270
271 if (modified_in_p (test, insn))
272 {
273 if (!mod_ok)
274 return FALSE;
275 must_be_last = TRUE;
276 }
277
278 /* Now build the conditional form of the instruction. */
279 pattern = PATTERN (insn);
280 xtest = copy_rtx (test);
281
282 /* If this is already a COND_EXEC, rewrite the test to be an AND of the
283 two conditions. */
284 if (GET_CODE (pattern) == COND_EXEC)
285 {
286 if (GET_MODE (xtest) != GET_MODE (COND_EXEC_TEST (pattern)))
287 return FALSE;
288
289 xtest = gen_rtx_AND (GET_MODE (xtest), xtest,
290 COND_EXEC_TEST (pattern));
291 pattern = COND_EXEC_CODE (pattern);
292 }
293
294 pattern = gen_rtx_COND_EXEC (VOIDmode, xtest, pattern);
295
296 /* If the machine needs to modify the insn being conditionally executed,
297 say for example to force a constant integer operand into a temp
298 register, do so here. */
299 #ifdef IFCVT_MODIFY_INSN
300 IFCVT_MODIFY_INSN (ce_info, pattern, insn);
301 if (! pattern)
302 return FALSE;
303 #endif
304
305 validate_change (insn, &PATTERN (insn), pattern, 1);
306
307 if (GET_CODE (insn) == CALL_INSN && prob_val)
308 validate_change (insn, &REG_NOTES (insn),
309 alloc_EXPR_LIST (REG_BR_PROB, prob_val,
310 REG_NOTES (insn)), 1);
311
312 insn_done:
313 if (insn == end)
314 break;
315 }
316
317 return TRUE;
318 }
319
320 /* Return the condition for a jump. Do not do any special processing. */
321
322 static rtx
323 cond_exec_get_condition (rtx jump)
324 {
325 rtx test_if, cond;
326
327 if (any_condjump_p (jump))
328 test_if = SET_SRC (pc_set (jump));
329 else
330 return NULL_RTX;
331 cond = XEXP (test_if, 0);
332
333 /* If this branches to JUMP_LABEL when the condition is false,
334 reverse the condition. */
335 if (GET_CODE (XEXP (test_if, 2)) == LABEL_REF
336 && XEXP (XEXP (test_if, 2), 0) == JUMP_LABEL (jump))
337 {
338 enum rtx_code rev = reversed_comparison_code (cond, jump);
339 if (rev == UNKNOWN)
340 return NULL_RTX;
341
342 cond = gen_rtx_fmt_ee (rev, GET_MODE (cond), XEXP (cond, 0),
343 XEXP (cond, 1));
344 }
345
346 return cond;
347 }
348
349 /* Given a simple IF-THEN or IF-THEN-ELSE block, attempt to convert it
350 to conditional execution. Return TRUE if we were successful at
351 converting the block. */
352
353 static int
354 cond_exec_process_if_block (ce_if_block_t * ce_info,
355 /* if block information */int do_multiple_p)
356 {
357 basic_block test_bb = ce_info->test_bb; /* last test block */
358 basic_block then_bb = ce_info->then_bb; /* THEN */
359 basic_block else_bb = ce_info->else_bb; /* ELSE or NULL */
360 rtx test_expr; /* expression in IF_THEN_ELSE that is tested */
361 rtx then_start; /* first insn in THEN block */
362 rtx then_end; /* last insn + 1 in THEN block */
363 rtx else_start = NULL_RTX; /* first insn in ELSE block or NULL */
364 rtx else_end = NULL_RTX; /* last insn + 1 in ELSE block */
365 int max; /* max # of insns to convert. */
366 int then_mod_ok; /* whether conditional mods are ok in THEN */
367 rtx true_expr; /* test for else block insns */
368 rtx false_expr; /* test for then block insns */
369 rtx true_prob_val; /* probability of else block */
370 rtx false_prob_val; /* probability of then block */
371 int n_insns;
372 enum rtx_code false_code;
373
374 /* If test is comprised of && or || elements, and we've failed at handling
375 all of them together, just use the last test if it is the special case of
376 && elements without an ELSE block. */
377 if (!do_multiple_p && ce_info->num_multiple_test_blocks)
378 {
379 if (else_bb || ! ce_info->and_and_p)
380 return FALSE;
381
382 ce_info->test_bb = test_bb = ce_info->last_test_bb;
383 ce_info->num_multiple_test_blocks = 0;
384 ce_info->num_and_and_blocks = 0;
385 ce_info->num_or_or_blocks = 0;
386 }
387
388 /* Find the conditional jump to the ELSE or JOIN part, and isolate
389 the test. */
390 test_expr = cond_exec_get_condition (BB_END (test_bb));
391 if (! test_expr)
392 return FALSE;
393
394 /* If the conditional jump is more than just a conditional jump,
395 then we can not do conditional execution conversion on this block. */
396 if (! onlyjump_p (BB_END (test_bb)))
397 return FALSE;
398
399 /* Collect the bounds of where we're to search, skipping any labels, jumps
400 and notes at the beginning and end of the block. Then count the total
401 number of insns and see if it is small enough to convert. */
402 then_start = first_active_insn (then_bb);
403 then_end = last_active_insn (then_bb, TRUE);
404 n_insns = ce_info->num_then_insns = count_bb_insns (then_bb);
405 max = MAX_CONDITIONAL_EXECUTE;
406
407 if (else_bb)
408 {
409 max *= 2;
410 else_start = first_active_insn (else_bb);
411 else_end = last_active_insn (else_bb, TRUE);
412 n_insns += ce_info->num_else_insns = count_bb_insns (else_bb);
413 }
414
415 if (n_insns > max)
416 return FALSE;
417
418 /* Map test_expr/test_jump into the appropriate MD tests to use on
419 the conditionally executed code. */
420
421 true_expr = test_expr;
422
423 false_code = reversed_comparison_code (true_expr, BB_END (test_bb));
424 if (false_code != UNKNOWN)
425 false_expr = gen_rtx_fmt_ee (false_code, GET_MODE (true_expr),
426 XEXP (true_expr, 0), XEXP (true_expr, 1));
427 else
428 false_expr = NULL_RTX;
429
430 #ifdef IFCVT_MODIFY_TESTS
431 /* If the machine description needs to modify the tests, such as setting a
432 conditional execution register from a comparison, it can do so here. */
433 IFCVT_MODIFY_TESTS (ce_info, true_expr, false_expr);
434
435 /* See if the conversion failed. */
436 if (!true_expr || !false_expr)
437 goto fail;
438 #endif
439
440 true_prob_val = find_reg_note (BB_END (test_bb), REG_BR_PROB, NULL_RTX);
441 if (true_prob_val)
442 {
443 true_prob_val = XEXP (true_prob_val, 0);
444 false_prob_val = GEN_INT (REG_BR_PROB_BASE - INTVAL (true_prob_val));
445 }
446 else
447 false_prob_val = NULL_RTX;
448
449 /* If we have && or || tests, do them here. These tests are in the adjacent
450 blocks after the first block containing the test. */
451 if (ce_info->num_multiple_test_blocks > 0)
452 {
453 basic_block bb = test_bb;
454 basic_block last_test_bb = ce_info->last_test_bb;
455
456 if (! false_expr)
457 goto fail;
458
459 do
460 {
461 rtx start, end;
462 rtx t, f;
463
464 bb = block_fallthru (bb);
465 start = first_active_insn (bb);
466 end = last_active_insn (bb, TRUE);
467 if (start
468 && ! cond_exec_process_insns (ce_info, start, end, false_expr,
469 false_prob_val, FALSE))
470 goto fail;
471
472 /* If the conditional jump is more than just a conditional jump, then
473 we can not do conditional execution conversion on this block. */
474 if (! onlyjump_p (BB_END (bb)))
475 goto fail;
476
477 /* Find the conditional jump and isolate the test. */
478 t = cond_exec_get_condition (BB_END (bb));
479 if (! t)
480 goto fail;
481
482 f = gen_rtx_fmt_ee (reverse_condition (GET_CODE (t)),
483 GET_MODE (t),
484 XEXP (t, 0),
485 XEXP (t, 1));
486
487 if (ce_info->and_and_p)
488 {
489 t = gen_rtx_AND (GET_MODE (t), true_expr, t);
490 f = gen_rtx_IOR (GET_MODE (t), false_expr, f);
491 }
492 else
493 {
494 t = gen_rtx_IOR (GET_MODE (t), true_expr, t);
495 f = gen_rtx_AND (GET_MODE (t), false_expr, f);
496 }
497
498 /* If the machine description needs to modify the tests, such as
499 setting a conditional execution register from a comparison, it can
500 do so here. */
501 #ifdef IFCVT_MODIFY_MULTIPLE_TESTS
502 IFCVT_MODIFY_MULTIPLE_TESTS (ce_info, bb, t, f);
503
504 /* See if the conversion failed. */
505 if (!t || !f)
506 goto fail;
507 #endif
508
509 true_expr = t;
510 false_expr = f;
511 }
512 while (bb != last_test_bb);
513 }
514
515 /* For IF-THEN-ELSE blocks, we don't allow modifications of the test
516 on then THEN block. */
517 then_mod_ok = (else_bb == NULL_BLOCK);
518
519 /* Go through the THEN and ELSE blocks converting the insns if possible
520 to conditional execution. */
521
522 if (then_end
523 && (! false_expr
524 || ! cond_exec_process_insns (ce_info, then_start, then_end,
525 false_expr, false_prob_val,
526 then_mod_ok)))
527 goto fail;
528
529 if (else_bb && else_end
530 && ! cond_exec_process_insns (ce_info, else_start, else_end,
531 true_expr, true_prob_val, TRUE))
532 goto fail;
533
534 /* If we cannot apply the changes, fail. Do not go through the normal fail
535 processing, since apply_change_group will call cancel_changes. */
536 if (! apply_change_group ())
537 {
538 #ifdef IFCVT_MODIFY_CANCEL
539 /* Cancel any machine dependent changes. */
540 IFCVT_MODIFY_CANCEL (ce_info);
541 #endif
542 return FALSE;
543 }
544
545 #ifdef IFCVT_MODIFY_FINAL
546 /* Do any machine dependent final modifications. */
547 IFCVT_MODIFY_FINAL (ce_info);
548 #endif
549
550 /* Conversion succeeded. */
551 if (dump_file)
552 fprintf (dump_file, "%d insn%s converted to conditional execution.\n",
553 n_insns, (n_insns == 1) ? " was" : "s were");
554
555 /* Merge the blocks! */
556 merge_if_block (ce_info);
557 cond_exec_changed_p = TRUE;
558 return TRUE;
559
560 fail:
561 #ifdef IFCVT_MODIFY_CANCEL
562 /* Cancel any machine dependent changes. */
563 IFCVT_MODIFY_CANCEL (ce_info);
564 #endif
565
566 cancel_changes (0);
567 return FALSE;
568 }
569 \f
570 /* Used by noce_process_if_block to communicate with its subroutines.
571
572 The subroutines know that A and B may be evaluated freely. They
573 know that X is a register. They should insert new instructions
574 before cond_earliest. */
575
576 struct noce_if_info
577 {
578 basic_block test_bb;
579 rtx insn_a, insn_b;
580 rtx x, a, b;
581 rtx jump, cond, cond_earliest;
582 };
583
584 static rtx noce_emit_store_flag (struct noce_if_info *, rtx, int, int);
585 static int noce_try_move (struct noce_if_info *);
586 static int noce_try_store_flag (struct noce_if_info *);
587 static int noce_try_addcc (struct noce_if_info *);
588 static int noce_try_store_flag_constants (struct noce_if_info *);
589 static int noce_try_store_flag_mask (struct noce_if_info *);
590 static rtx noce_emit_cmove (struct noce_if_info *, rtx, enum rtx_code, rtx,
591 rtx, rtx, rtx);
592 static int noce_try_cmove (struct noce_if_info *);
593 static int noce_try_cmove_arith (struct noce_if_info *);
594 static rtx noce_get_alt_condition (struct noce_if_info *, rtx, rtx *);
595 static int noce_try_minmax (struct noce_if_info *);
596 static int noce_try_abs (struct noce_if_info *);
597 static int noce_try_sign_mask (struct noce_if_info *);
598
599 /* Helper function for noce_try_store_flag*. */
600
601 static rtx
602 noce_emit_store_flag (struct noce_if_info *if_info, rtx x, int reversep,
603 int normalize)
604 {
605 rtx cond = if_info->cond;
606 int cond_complex;
607 enum rtx_code code;
608
609 cond_complex = (! general_operand (XEXP (cond, 0), VOIDmode)
610 || ! general_operand (XEXP (cond, 1), VOIDmode));
611
612 /* If earliest == jump, or when the condition is complex, try to
613 build the store_flag insn directly. */
614
615 if (cond_complex)
616 cond = XEXP (SET_SRC (pc_set (if_info->jump)), 0);
617
618 if (reversep)
619 code = reversed_comparison_code (cond, if_info->jump);
620 else
621 code = GET_CODE (cond);
622
623 if ((if_info->cond_earliest == if_info->jump || cond_complex)
624 && (normalize == 0 || STORE_FLAG_VALUE == normalize))
625 {
626 rtx tmp;
627
628 tmp = gen_rtx_fmt_ee (code, GET_MODE (x), XEXP (cond, 0),
629 XEXP (cond, 1));
630 tmp = gen_rtx_SET (VOIDmode, x, tmp);
631
632 start_sequence ();
633 tmp = emit_insn (tmp);
634
635 if (recog_memoized (tmp) >= 0)
636 {
637 tmp = get_insns ();
638 end_sequence ();
639 emit_insn (tmp);
640
641 if_info->cond_earliest = if_info->jump;
642
643 return x;
644 }
645
646 end_sequence ();
647 }
648
649 /* Don't even try if the comparison operands or the mode of X are weird. */
650 if (cond_complex || !SCALAR_INT_MODE_P (GET_MODE (x)))
651 return NULL_RTX;
652
653 return emit_store_flag (x, code, XEXP (cond, 0),
654 XEXP (cond, 1), VOIDmode,
655 (code == LTU || code == LEU
656 || code == GEU || code == GTU), normalize);
657 }
658
659 /* Emit instruction to move an rtx, possibly into STRICT_LOW_PART.
660 X is the destination/target and Y is the value to copy. */
661
662 static void
663 noce_emit_move_insn (rtx x, rtx y)
664 {
665 enum machine_mode outmode, inmode;
666 rtx outer, inner;
667 int bitpos;
668
669 if (GET_CODE (x) != STRICT_LOW_PART)
670 {
671 emit_move_insn (x, y);
672 return;
673 }
674
675 outer = XEXP (x, 0);
676 inner = XEXP (outer, 0);
677 outmode = GET_MODE (outer);
678 inmode = GET_MODE (inner);
679 bitpos = SUBREG_BYTE (outer) * BITS_PER_UNIT;
680 store_bit_field (inner, GET_MODE_BITSIZE (outmode), bitpos, outmode, y,
681 GET_MODE_BITSIZE (inmode));
682 }
683
684 /* Return sequence of instructions generated by if conversion. This
685 function calls end_sequence() to end the current stream, ensures
686 that are instructions are unshared, recognizable non-jump insns.
687 On failure, this function returns a NULL_RTX. */
688
689 static rtx
690 end_ifcvt_sequence (struct noce_if_info *if_info)
691 {
692 rtx insn;
693 rtx seq = get_insns ();
694
695 set_used_flags (if_info->x);
696 set_used_flags (if_info->cond);
697 unshare_all_rtl_in_chain (seq);
698 end_sequence ();
699
700 /* Make sure that all of the instructions emitted are recognizable,
701 and that we haven't introduced a new jump instruction.
702 As an excersise for the reader, build a general mechanism that
703 allows proper placement of required clobbers. */
704 for (insn = seq; insn; insn = NEXT_INSN (insn))
705 if (GET_CODE (insn) == JUMP_INSN
706 || recog_memoized (insn) == -1)
707 return NULL_RTX;
708
709 return seq;
710 }
711
712 /* Convert "if (a != b) x = a; else x = b" into "x = a" and
713 "if (a == b) x = a; else x = b" into "x = b". */
714
715 static int
716 noce_try_move (struct noce_if_info *if_info)
717 {
718 rtx cond = if_info->cond;
719 enum rtx_code code = GET_CODE (cond);
720 rtx y, seq;
721
722 if (code != NE && code != EQ)
723 return FALSE;
724
725 /* This optimization isn't valid if either A or B could be a NaN
726 or a signed zero. */
727 if (HONOR_NANS (GET_MODE (if_info->x))
728 || HONOR_SIGNED_ZEROS (GET_MODE (if_info->x)))
729 return FALSE;
730
731 /* Check whether the operands of the comparison are A and in
732 either order. */
733 if ((rtx_equal_p (if_info->a, XEXP (cond, 0))
734 && rtx_equal_p (if_info->b, XEXP (cond, 1)))
735 || (rtx_equal_p (if_info->a, XEXP (cond, 1))
736 && rtx_equal_p (if_info->b, XEXP (cond, 0))))
737 {
738 y = (code == EQ) ? if_info->a : if_info->b;
739
740 /* Avoid generating the move if the source is the destination. */
741 if (! rtx_equal_p (if_info->x, y))
742 {
743 start_sequence ();
744 noce_emit_move_insn (if_info->x, y);
745 seq = end_ifcvt_sequence (if_info);
746 if (!seq)
747 return FALSE;
748
749 emit_insn_before_setloc (seq, if_info->jump,
750 INSN_LOCATOR (if_info->insn_a));
751 }
752 return TRUE;
753 }
754 return FALSE;
755 }
756
757 /* Convert "if (test) x = 1; else x = 0".
758
759 Only try 0 and STORE_FLAG_VALUE here. Other combinations will be
760 tried in noce_try_store_flag_constants after noce_try_cmove has had
761 a go at the conversion. */
762
763 static int
764 noce_try_store_flag (struct noce_if_info *if_info)
765 {
766 int reversep;
767 rtx target, seq;
768
769 if (GET_CODE (if_info->b) == CONST_INT
770 && INTVAL (if_info->b) == STORE_FLAG_VALUE
771 && if_info->a == const0_rtx)
772 reversep = 0;
773 else if (if_info->b == const0_rtx
774 && GET_CODE (if_info->a) == CONST_INT
775 && INTVAL (if_info->a) == STORE_FLAG_VALUE
776 && (reversed_comparison_code (if_info->cond, if_info->jump)
777 != UNKNOWN))
778 reversep = 1;
779 else
780 return FALSE;
781
782 start_sequence ();
783
784 target = noce_emit_store_flag (if_info, if_info->x, reversep, 0);
785 if (target)
786 {
787 if (target != if_info->x)
788 noce_emit_move_insn (if_info->x, target);
789
790 seq = end_ifcvt_sequence (if_info);
791 if (! seq)
792 return FALSE;
793
794 emit_insn_before_setloc (seq, if_info->jump,
795 INSN_LOCATOR (if_info->insn_a));
796 return TRUE;
797 }
798 else
799 {
800 end_sequence ();
801 return FALSE;
802 }
803 }
804
805 /* Convert "if (test) x = a; else x = b", for A and B constant. */
806
807 static int
808 noce_try_store_flag_constants (struct noce_if_info *if_info)
809 {
810 rtx target, seq;
811 int reversep;
812 HOST_WIDE_INT itrue, ifalse, diff, tmp;
813 int normalize, can_reverse;
814 enum machine_mode mode;
815
816 if (! no_new_pseudos
817 && GET_CODE (if_info->a) == CONST_INT
818 && GET_CODE (if_info->b) == CONST_INT)
819 {
820 mode = GET_MODE (if_info->x);
821 ifalse = INTVAL (if_info->a);
822 itrue = INTVAL (if_info->b);
823
824 /* Make sure we can represent the difference between the two values. */
825 if ((itrue - ifalse > 0)
826 != ((ifalse < 0) != (itrue < 0) ? ifalse < 0 : ifalse < itrue))
827 return FALSE;
828
829 diff = trunc_int_for_mode (itrue - ifalse, mode);
830
831 can_reverse = (reversed_comparison_code (if_info->cond, if_info->jump)
832 != UNKNOWN);
833
834 reversep = 0;
835 if (diff == STORE_FLAG_VALUE || diff == -STORE_FLAG_VALUE)
836 normalize = 0;
837 else if (ifalse == 0 && exact_log2 (itrue) >= 0
838 && (STORE_FLAG_VALUE == 1
839 || BRANCH_COST >= 2))
840 normalize = 1;
841 else if (itrue == 0 && exact_log2 (ifalse) >= 0 && can_reverse
842 && (STORE_FLAG_VALUE == 1 || BRANCH_COST >= 2))
843 normalize = 1, reversep = 1;
844 else if (itrue == -1
845 && (STORE_FLAG_VALUE == -1
846 || BRANCH_COST >= 2))
847 normalize = -1;
848 else if (ifalse == -1 && can_reverse
849 && (STORE_FLAG_VALUE == -1 || BRANCH_COST >= 2))
850 normalize = -1, reversep = 1;
851 else if ((BRANCH_COST >= 2 && STORE_FLAG_VALUE == -1)
852 || BRANCH_COST >= 3)
853 normalize = -1;
854 else
855 return FALSE;
856
857 if (reversep)
858 {
859 tmp = itrue; itrue = ifalse; ifalse = tmp;
860 diff = trunc_int_for_mode (-diff, mode);
861 }
862
863 start_sequence ();
864 target = noce_emit_store_flag (if_info, if_info->x, reversep, normalize);
865 if (! target)
866 {
867 end_sequence ();
868 return FALSE;
869 }
870
871 /* if (test) x = 3; else x = 4;
872 => x = 3 + (test == 0); */
873 if (diff == STORE_FLAG_VALUE || diff == -STORE_FLAG_VALUE)
874 {
875 target = expand_simple_binop (mode,
876 (diff == STORE_FLAG_VALUE
877 ? PLUS : MINUS),
878 GEN_INT (ifalse), target, if_info->x, 0,
879 OPTAB_WIDEN);
880 }
881
882 /* if (test) x = 8; else x = 0;
883 => x = (test != 0) << 3; */
884 else if (ifalse == 0 && (tmp = exact_log2 (itrue)) >= 0)
885 {
886 target = expand_simple_binop (mode, ASHIFT,
887 target, GEN_INT (tmp), if_info->x, 0,
888 OPTAB_WIDEN);
889 }
890
891 /* if (test) x = -1; else x = b;
892 => x = -(test != 0) | b; */
893 else if (itrue == -1)
894 {
895 target = expand_simple_binop (mode, IOR,
896 target, GEN_INT (ifalse), if_info->x, 0,
897 OPTAB_WIDEN);
898 }
899
900 /* if (test) x = a; else x = b;
901 => x = (-(test != 0) & (b - a)) + a; */
902 else
903 {
904 target = expand_simple_binop (mode, AND,
905 target, GEN_INT (diff), if_info->x, 0,
906 OPTAB_WIDEN);
907 if (target)
908 target = expand_simple_binop (mode, PLUS,
909 target, GEN_INT (ifalse),
910 if_info->x, 0, OPTAB_WIDEN);
911 }
912
913 if (! target)
914 {
915 end_sequence ();
916 return FALSE;
917 }
918
919 if (target != if_info->x)
920 noce_emit_move_insn (if_info->x, target);
921
922 seq = end_ifcvt_sequence (if_info);
923 if (!seq)
924 return FALSE;
925
926 emit_insn_before_setloc (seq, if_info->jump,
927 INSN_LOCATOR (if_info->insn_a));
928 return TRUE;
929 }
930
931 return FALSE;
932 }
933
934 /* Convert "if (test) foo++" into "foo += (test != 0)", and
935 similarly for "foo--". */
936
937 static int
938 noce_try_addcc (struct noce_if_info *if_info)
939 {
940 rtx target, seq;
941 int subtract, normalize;
942
943 if (! no_new_pseudos
944 && GET_CODE (if_info->a) == PLUS
945 && rtx_equal_p (XEXP (if_info->a, 0), if_info->b)
946 && (reversed_comparison_code (if_info->cond, if_info->jump)
947 != UNKNOWN))
948 {
949 rtx cond = if_info->cond;
950 enum rtx_code code = reversed_comparison_code (cond, if_info->jump);
951
952 /* First try to use addcc pattern. */
953 if (general_operand (XEXP (cond, 0), VOIDmode)
954 && general_operand (XEXP (cond, 1), VOIDmode))
955 {
956 start_sequence ();
957 target = emit_conditional_add (if_info->x, code,
958 XEXP (cond, 0),
959 XEXP (cond, 1),
960 VOIDmode,
961 if_info->b,
962 XEXP (if_info->a, 1),
963 GET_MODE (if_info->x),
964 (code == LTU || code == GEU
965 || code == LEU || code == GTU));
966 if (target)
967 {
968 if (target != if_info->x)
969 noce_emit_move_insn (if_info->x, target);
970
971 seq = end_ifcvt_sequence (if_info);
972 if (!seq)
973 return FALSE;
974
975 emit_insn_before_setloc (seq, if_info->jump,
976 INSN_LOCATOR (if_info->insn_a));
977 return TRUE;
978 }
979 end_sequence ();
980 }
981
982 /* If that fails, construct conditional increment or decrement using
983 setcc. */
984 if (BRANCH_COST >= 2
985 && (XEXP (if_info->a, 1) == const1_rtx
986 || XEXP (if_info->a, 1) == constm1_rtx))
987 {
988 start_sequence ();
989 if (STORE_FLAG_VALUE == INTVAL (XEXP (if_info->a, 1)))
990 subtract = 0, normalize = 0;
991 else if (-STORE_FLAG_VALUE == INTVAL (XEXP (if_info->a, 1)))
992 subtract = 1, normalize = 0;
993 else
994 subtract = 0, normalize = INTVAL (XEXP (if_info->a, 1));
995
996
997 target = noce_emit_store_flag (if_info,
998 gen_reg_rtx (GET_MODE (if_info->x)),
999 1, normalize);
1000
1001 if (target)
1002 target = expand_simple_binop (GET_MODE (if_info->x),
1003 subtract ? MINUS : PLUS,
1004 if_info->b, target, if_info->x,
1005 0, OPTAB_WIDEN);
1006 if (target)
1007 {
1008 if (target != if_info->x)
1009 noce_emit_move_insn (if_info->x, target);
1010
1011 seq = end_ifcvt_sequence (if_info);
1012 if (!seq)
1013 return FALSE;
1014
1015 emit_insn_before_setloc (seq, if_info->jump,
1016 INSN_LOCATOR (if_info->insn_a));
1017 return TRUE;
1018 }
1019 end_sequence ();
1020 }
1021 }
1022
1023 return FALSE;
1024 }
1025
1026 /* Convert "if (test) x = 0;" to "x &= -(test == 0);" */
1027
1028 static int
1029 noce_try_store_flag_mask (struct noce_if_info *if_info)
1030 {
1031 rtx target, seq;
1032 int reversep;
1033
1034 reversep = 0;
1035 if (! no_new_pseudos
1036 && (BRANCH_COST >= 2
1037 || STORE_FLAG_VALUE == -1)
1038 && ((if_info->a == const0_rtx
1039 && rtx_equal_p (if_info->b, if_info->x))
1040 || ((reversep = (reversed_comparison_code (if_info->cond,
1041 if_info->jump)
1042 != UNKNOWN))
1043 && if_info->b == const0_rtx
1044 && rtx_equal_p (if_info->a, if_info->x))))
1045 {
1046 start_sequence ();
1047 target = noce_emit_store_flag (if_info,
1048 gen_reg_rtx (GET_MODE (if_info->x)),
1049 reversep, -1);
1050 if (target)
1051 target = expand_simple_binop (GET_MODE (if_info->x), AND,
1052 if_info->x,
1053 target, if_info->x, 0,
1054 OPTAB_WIDEN);
1055
1056 if (target)
1057 {
1058 if (target != if_info->x)
1059 noce_emit_move_insn (if_info->x, target);
1060
1061 seq = end_ifcvt_sequence (if_info);
1062 if (!seq)
1063 return FALSE;
1064
1065 emit_insn_before_setloc (seq, if_info->jump,
1066 INSN_LOCATOR (if_info->insn_a));
1067 return TRUE;
1068 }
1069
1070 end_sequence ();
1071 }
1072
1073 return FALSE;
1074 }
1075
1076 /* Helper function for noce_try_cmove and noce_try_cmove_arith. */
1077
1078 static rtx
1079 noce_emit_cmove (struct noce_if_info *if_info, rtx x, enum rtx_code code,
1080 rtx cmp_a, rtx cmp_b, rtx vfalse, rtx vtrue)
1081 {
1082 /* If earliest == jump, try to build the cmove insn directly.
1083 This is helpful when combine has created some complex condition
1084 (like for alpha's cmovlbs) that we can't hope to regenerate
1085 through the normal interface. */
1086
1087 if (if_info->cond_earliest == if_info->jump)
1088 {
1089 rtx tmp;
1090
1091 tmp = gen_rtx_fmt_ee (code, GET_MODE (if_info->cond), cmp_a, cmp_b);
1092 tmp = gen_rtx_IF_THEN_ELSE (GET_MODE (x), tmp, vtrue, vfalse);
1093 tmp = gen_rtx_SET (VOIDmode, x, tmp);
1094
1095 start_sequence ();
1096 tmp = emit_insn (tmp);
1097
1098 if (recog_memoized (tmp) >= 0)
1099 {
1100 tmp = get_insns ();
1101 end_sequence ();
1102 emit_insn (tmp);
1103
1104 return x;
1105 }
1106
1107 end_sequence ();
1108 }
1109
1110 /* Don't even try if the comparison operands are weird. */
1111 if (! general_operand (cmp_a, GET_MODE (cmp_a))
1112 || ! general_operand (cmp_b, GET_MODE (cmp_b)))
1113 return NULL_RTX;
1114
1115 #if HAVE_conditional_move
1116 return emit_conditional_move (x, code, cmp_a, cmp_b, VOIDmode,
1117 vtrue, vfalse, GET_MODE (x),
1118 (code == LTU || code == GEU
1119 || code == LEU || code == GTU));
1120 #else
1121 /* We'll never get here, as noce_process_if_block doesn't call the
1122 functions involved. Ifdef code, however, should be discouraged
1123 because it leads to typos in the code not selected. However,
1124 emit_conditional_move won't exist either. */
1125 return NULL_RTX;
1126 #endif
1127 }
1128
1129 /* Try only simple constants and registers here. More complex cases
1130 are handled in noce_try_cmove_arith after noce_try_store_flag_arith
1131 has had a go at it. */
1132
1133 static int
1134 noce_try_cmove (struct noce_if_info *if_info)
1135 {
1136 enum rtx_code code;
1137 rtx target, seq;
1138
1139 if ((CONSTANT_P (if_info->a) || register_operand (if_info->a, VOIDmode))
1140 && (CONSTANT_P (if_info->b) || register_operand (if_info->b, VOIDmode)))
1141 {
1142 start_sequence ();
1143
1144 code = GET_CODE (if_info->cond);
1145 target = noce_emit_cmove (if_info, if_info->x, code,
1146 XEXP (if_info->cond, 0),
1147 XEXP (if_info->cond, 1),
1148 if_info->a, if_info->b);
1149
1150 if (target)
1151 {
1152 if (target != if_info->x)
1153 noce_emit_move_insn (if_info->x, target);
1154
1155 seq = end_ifcvt_sequence (if_info);
1156 if (!seq)
1157 return FALSE;
1158
1159 emit_insn_before_setloc (seq, if_info->jump,
1160 INSN_LOCATOR (if_info->insn_a));
1161 return TRUE;
1162 }
1163 else
1164 {
1165 end_sequence ();
1166 return FALSE;
1167 }
1168 }
1169
1170 return FALSE;
1171 }
1172
1173 /* Try more complex cases involving conditional_move. */
1174
1175 static int
1176 noce_try_cmove_arith (struct noce_if_info *if_info)
1177 {
1178 rtx a = if_info->a;
1179 rtx b = if_info->b;
1180 rtx x = if_info->x;
1181 rtx insn_a, insn_b;
1182 rtx tmp, target;
1183 int is_mem = 0;
1184 enum rtx_code code;
1185
1186 /* A conditional move from two memory sources is equivalent to a
1187 conditional on their addresses followed by a load. Don't do this
1188 early because it'll screw alias analysis. Note that we've
1189 already checked for no side effects. */
1190 if (! no_new_pseudos && cse_not_expected
1191 && MEM_P (a) && MEM_P (b)
1192 && BRANCH_COST >= 5)
1193 {
1194 a = XEXP (a, 0);
1195 b = XEXP (b, 0);
1196 x = gen_reg_rtx (Pmode);
1197 is_mem = 1;
1198 }
1199
1200 /* ??? We could handle this if we knew that a load from A or B could
1201 not fault. This is also true if we've already loaded
1202 from the address along the path from ENTRY. */
1203 else if (may_trap_p (a) || may_trap_p (b))
1204 return FALSE;
1205
1206 /* if (test) x = a + b; else x = c - d;
1207 => y = a + b;
1208 x = c - d;
1209 if (test)
1210 x = y;
1211 */
1212
1213 code = GET_CODE (if_info->cond);
1214 insn_a = if_info->insn_a;
1215 insn_b = if_info->insn_b;
1216
1217 /* Possibly rearrange operands to make things come out more natural. */
1218 if (reversed_comparison_code (if_info->cond, if_info->jump) != UNKNOWN)
1219 {
1220 int reversep = 0;
1221 if (rtx_equal_p (b, x))
1222 reversep = 1;
1223 else if (general_operand (b, GET_MODE (b)))
1224 reversep = 1;
1225
1226 if (reversep)
1227 {
1228 code = reversed_comparison_code (if_info->cond, if_info->jump);
1229 tmp = a, a = b, b = tmp;
1230 tmp = insn_a, insn_a = insn_b, insn_b = tmp;
1231 }
1232 }
1233
1234 start_sequence ();
1235
1236 /* If either operand is complex, load it into a register first.
1237 The best way to do this is to copy the original insn. In this
1238 way we preserve any clobbers etc that the insn may have had.
1239 This is of course not possible in the IS_MEM case. */
1240 if (! general_operand (a, GET_MODE (a)))
1241 {
1242 rtx set;
1243
1244 if (no_new_pseudos)
1245 goto end_seq_and_fail;
1246
1247 if (is_mem)
1248 {
1249 tmp = gen_reg_rtx (GET_MODE (a));
1250 tmp = emit_insn (gen_rtx_SET (VOIDmode, tmp, a));
1251 }
1252 else if (! insn_a)
1253 goto end_seq_and_fail;
1254 else
1255 {
1256 a = gen_reg_rtx (GET_MODE (a));
1257 tmp = copy_rtx (insn_a);
1258 set = single_set (tmp);
1259 SET_DEST (set) = a;
1260 tmp = emit_insn (PATTERN (tmp));
1261 }
1262 if (recog_memoized (tmp) < 0)
1263 goto end_seq_and_fail;
1264 }
1265 if (! general_operand (b, GET_MODE (b)))
1266 {
1267 rtx set;
1268
1269 if (no_new_pseudos)
1270 goto end_seq_and_fail;
1271
1272 if (is_mem)
1273 {
1274 tmp = gen_reg_rtx (GET_MODE (b));
1275 tmp = emit_insn (gen_rtx_SET (VOIDmode,
1276 tmp,
1277 b));
1278 }
1279 else if (! insn_b)
1280 goto end_seq_and_fail;
1281 else
1282 {
1283 b = gen_reg_rtx (GET_MODE (b));
1284 tmp = copy_rtx (insn_b);
1285 set = single_set (tmp);
1286 SET_DEST (set) = b;
1287 tmp = emit_insn (PATTERN (tmp));
1288 }
1289 if (recog_memoized (tmp) < 0)
1290 goto end_seq_and_fail;
1291 }
1292
1293 target = noce_emit_cmove (if_info, x, code, XEXP (if_info->cond, 0),
1294 XEXP (if_info->cond, 1), a, b);
1295
1296 if (! target)
1297 goto end_seq_and_fail;
1298
1299 /* If we're handling a memory for above, emit the load now. */
1300 if (is_mem)
1301 {
1302 tmp = gen_rtx_MEM (GET_MODE (if_info->x), target);
1303
1304 /* Copy over flags as appropriate. */
1305 if (MEM_VOLATILE_P (if_info->a) || MEM_VOLATILE_P (if_info->b))
1306 MEM_VOLATILE_P (tmp) = 1;
1307 if (MEM_IN_STRUCT_P (if_info->a) && MEM_IN_STRUCT_P (if_info->b))
1308 MEM_IN_STRUCT_P (tmp) = 1;
1309 if (MEM_SCALAR_P (if_info->a) && MEM_SCALAR_P (if_info->b))
1310 MEM_SCALAR_P (tmp) = 1;
1311 if (MEM_ALIAS_SET (if_info->a) == MEM_ALIAS_SET (if_info->b))
1312 set_mem_alias_set (tmp, MEM_ALIAS_SET (if_info->a));
1313 set_mem_align (tmp,
1314 MIN (MEM_ALIGN (if_info->a), MEM_ALIGN (if_info->b)));
1315
1316 noce_emit_move_insn (if_info->x, tmp);
1317 }
1318 else if (target != x)
1319 noce_emit_move_insn (x, target);
1320
1321 tmp = end_ifcvt_sequence (if_info);
1322 if (!tmp)
1323 return FALSE;
1324
1325 emit_insn_before_setloc (tmp, if_info->jump, INSN_LOCATOR (if_info->insn_a));
1326 return TRUE;
1327
1328 end_seq_and_fail:
1329 end_sequence ();
1330 return FALSE;
1331 }
1332
1333 /* For most cases, the simplified condition we found is the best
1334 choice, but this is not the case for the min/max/abs transforms.
1335 For these we wish to know that it is A or B in the condition. */
1336
1337 static rtx
1338 noce_get_alt_condition (struct noce_if_info *if_info, rtx target,
1339 rtx *earliest)
1340 {
1341 rtx cond, set, insn;
1342 int reverse;
1343
1344 /* If target is already mentioned in the known condition, return it. */
1345 if (reg_mentioned_p (target, if_info->cond))
1346 {
1347 *earliest = if_info->cond_earliest;
1348 return if_info->cond;
1349 }
1350
1351 set = pc_set (if_info->jump);
1352 cond = XEXP (SET_SRC (set), 0);
1353 reverse
1354 = GET_CODE (XEXP (SET_SRC (set), 2)) == LABEL_REF
1355 && XEXP (XEXP (SET_SRC (set), 2), 0) == JUMP_LABEL (if_info->jump);
1356
1357 /* If we're looking for a constant, try to make the conditional
1358 have that constant in it. There are two reasons why it may
1359 not have the constant we want:
1360
1361 1. GCC may have needed to put the constant in a register, because
1362 the target can't compare directly against that constant. For
1363 this case, we look for a SET immediately before the comparison
1364 that puts a constant in that register.
1365
1366 2. GCC may have canonicalized the conditional, for example
1367 replacing "if x < 4" with "if x <= 3". We can undo that (or
1368 make equivalent types of changes) to get the constants we need
1369 if they're off by one in the right direction. */
1370
1371 if (GET_CODE (target) == CONST_INT)
1372 {
1373 enum rtx_code code = GET_CODE (if_info->cond);
1374 rtx op_a = XEXP (if_info->cond, 0);
1375 rtx op_b = XEXP (if_info->cond, 1);
1376 rtx prev_insn;
1377
1378 /* First, look to see if we put a constant in a register. */
1379 prev_insn = PREV_INSN (if_info->cond_earliest);
1380 if (prev_insn
1381 && INSN_P (prev_insn)
1382 && GET_CODE (PATTERN (prev_insn)) == SET)
1383 {
1384 rtx src = find_reg_equal_equiv_note (prev_insn);
1385 if (!src)
1386 src = SET_SRC (PATTERN (prev_insn));
1387 if (GET_CODE (src) == CONST_INT)
1388 {
1389 if (rtx_equal_p (op_a, SET_DEST (PATTERN (prev_insn))))
1390 op_a = src;
1391 else if (rtx_equal_p (op_b, SET_DEST (PATTERN (prev_insn))))
1392 op_b = src;
1393
1394 if (GET_CODE (op_a) == CONST_INT)
1395 {
1396 rtx tmp = op_a;
1397 op_a = op_b;
1398 op_b = tmp;
1399 code = swap_condition (code);
1400 }
1401 }
1402 }
1403
1404 /* Now, look to see if we can get the right constant by
1405 adjusting the conditional. */
1406 if (GET_CODE (op_b) == CONST_INT)
1407 {
1408 HOST_WIDE_INT desired_val = INTVAL (target);
1409 HOST_WIDE_INT actual_val = INTVAL (op_b);
1410
1411 switch (code)
1412 {
1413 case LT:
1414 if (actual_val == desired_val + 1)
1415 {
1416 code = LE;
1417 op_b = GEN_INT (desired_val);
1418 }
1419 break;
1420 case LE:
1421 if (actual_val == desired_val - 1)
1422 {
1423 code = LT;
1424 op_b = GEN_INT (desired_val);
1425 }
1426 break;
1427 case GT:
1428 if (actual_val == desired_val - 1)
1429 {
1430 code = GE;
1431 op_b = GEN_INT (desired_val);
1432 }
1433 break;
1434 case GE:
1435 if (actual_val == desired_val + 1)
1436 {
1437 code = GT;
1438 op_b = GEN_INT (desired_val);
1439 }
1440 break;
1441 default:
1442 break;
1443 }
1444 }
1445
1446 /* If we made any changes, generate a new conditional that is
1447 equivalent to what we started with, but has the right
1448 constants in it. */
1449 if (code != GET_CODE (if_info->cond)
1450 || op_a != XEXP (if_info->cond, 0)
1451 || op_b != XEXP (if_info->cond, 1))
1452 {
1453 cond = gen_rtx_fmt_ee (code, GET_MODE (cond), op_a, op_b);
1454 *earliest = if_info->cond_earliest;
1455 return cond;
1456 }
1457 }
1458
1459 cond = canonicalize_condition (if_info->jump, cond, reverse,
1460 earliest, target, false);
1461 if (! cond || ! reg_mentioned_p (target, cond))
1462 return NULL;
1463
1464 /* We almost certainly searched back to a different place.
1465 Need to re-verify correct lifetimes. */
1466
1467 /* X may not be mentioned in the range (cond_earliest, jump]. */
1468 for (insn = if_info->jump; insn != *earliest; insn = PREV_INSN (insn))
1469 if (INSN_P (insn) && reg_overlap_mentioned_p (if_info->x, PATTERN (insn)))
1470 return NULL;
1471
1472 /* A and B may not be modified in the range [cond_earliest, jump). */
1473 for (insn = *earliest; insn != if_info->jump; insn = NEXT_INSN (insn))
1474 if (INSN_P (insn)
1475 && (modified_in_p (if_info->a, insn)
1476 || modified_in_p (if_info->b, insn)))
1477 return NULL;
1478
1479 return cond;
1480 }
1481
1482 /* Convert "if (a < b) x = a; else x = b;" to "x = min(a, b);", etc. */
1483
1484 static int
1485 noce_try_minmax (struct noce_if_info *if_info)
1486 {
1487 rtx cond, earliest, target, seq;
1488 enum rtx_code code, op;
1489 int unsignedp;
1490
1491 /* ??? Can't guarantee that expand_binop won't create pseudos. */
1492 if (no_new_pseudos)
1493 return FALSE;
1494
1495 /* ??? Reject modes with NaNs or signed zeros since we don't know how
1496 they will be resolved with an SMIN/SMAX. It wouldn't be too hard
1497 to get the target to tell us... */
1498 if (HONOR_SIGNED_ZEROS (GET_MODE (if_info->x))
1499 || HONOR_NANS (GET_MODE (if_info->x)))
1500 return FALSE;
1501
1502 cond = noce_get_alt_condition (if_info, if_info->a, &earliest);
1503 if (!cond)
1504 return FALSE;
1505
1506 /* Verify the condition is of the form we expect, and canonicalize
1507 the comparison code. */
1508 code = GET_CODE (cond);
1509 if (rtx_equal_p (XEXP (cond, 0), if_info->a))
1510 {
1511 if (! rtx_equal_p (XEXP (cond, 1), if_info->b))
1512 return FALSE;
1513 }
1514 else if (rtx_equal_p (XEXP (cond, 1), if_info->a))
1515 {
1516 if (! rtx_equal_p (XEXP (cond, 0), if_info->b))
1517 return FALSE;
1518 code = swap_condition (code);
1519 }
1520 else
1521 return FALSE;
1522
1523 /* Determine what sort of operation this is. Note that the code is for
1524 a taken branch, so the code->operation mapping appears backwards. */
1525 switch (code)
1526 {
1527 case LT:
1528 case LE:
1529 case UNLT:
1530 case UNLE:
1531 op = SMAX;
1532 unsignedp = 0;
1533 break;
1534 case GT:
1535 case GE:
1536 case UNGT:
1537 case UNGE:
1538 op = SMIN;
1539 unsignedp = 0;
1540 break;
1541 case LTU:
1542 case LEU:
1543 op = UMAX;
1544 unsignedp = 1;
1545 break;
1546 case GTU:
1547 case GEU:
1548 op = UMIN;
1549 unsignedp = 1;
1550 break;
1551 default:
1552 return FALSE;
1553 }
1554
1555 start_sequence ();
1556
1557 target = expand_simple_binop (GET_MODE (if_info->x), op,
1558 if_info->a, if_info->b,
1559 if_info->x, unsignedp, OPTAB_WIDEN);
1560 if (! target)
1561 {
1562 end_sequence ();
1563 return FALSE;
1564 }
1565 if (target != if_info->x)
1566 noce_emit_move_insn (if_info->x, target);
1567
1568 seq = end_ifcvt_sequence (if_info);
1569 if (!seq)
1570 return FALSE;
1571
1572 emit_insn_before_setloc (seq, if_info->jump, INSN_LOCATOR (if_info->insn_a));
1573 if_info->cond = cond;
1574 if_info->cond_earliest = earliest;
1575
1576 return TRUE;
1577 }
1578
1579 /* Convert "if (a < 0) x = -a; else x = a;" to "x = abs(a);", etc. */
1580
1581 static int
1582 noce_try_abs (struct noce_if_info *if_info)
1583 {
1584 rtx cond, earliest, target, seq, a, b, c;
1585 int negate;
1586
1587 /* ??? Can't guarantee that expand_binop won't create pseudos. */
1588 if (no_new_pseudos)
1589 return FALSE;
1590
1591 /* Recognize A and B as constituting an ABS or NABS. */
1592 a = if_info->a;
1593 b = if_info->b;
1594 if (GET_CODE (a) == NEG && rtx_equal_p (XEXP (a, 0), b))
1595 negate = 0;
1596 else if (GET_CODE (b) == NEG && rtx_equal_p (XEXP (b, 0), a))
1597 {
1598 c = a; a = b; b = c;
1599 negate = 1;
1600 }
1601 else
1602 return FALSE;
1603
1604 cond = noce_get_alt_condition (if_info, b, &earliest);
1605 if (!cond)
1606 return FALSE;
1607
1608 /* Verify the condition is of the form we expect. */
1609 if (rtx_equal_p (XEXP (cond, 0), b))
1610 c = XEXP (cond, 1);
1611 else if (rtx_equal_p (XEXP (cond, 1), b))
1612 c = XEXP (cond, 0);
1613 else
1614 return FALSE;
1615
1616 /* Verify that C is zero. Search backward through the block for
1617 a REG_EQUAL note if necessary. */
1618 if (REG_P (c))
1619 {
1620 rtx insn, note = NULL;
1621 for (insn = earliest;
1622 insn != BB_HEAD (if_info->test_bb);
1623 insn = PREV_INSN (insn))
1624 if (INSN_P (insn)
1625 && ((note = find_reg_note (insn, REG_EQUAL, c))
1626 || (note = find_reg_note (insn, REG_EQUIV, c))))
1627 break;
1628 if (! note)
1629 return FALSE;
1630 c = XEXP (note, 0);
1631 }
1632 if (MEM_P (c)
1633 && GET_CODE (XEXP (c, 0)) == SYMBOL_REF
1634 && CONSTANT_POOL_ADDRESS_P (XEXP (c, 0)))
1635 c = get_pool_constant (XEXP (c, 0));
1636
1637 /* Work around funny ideas get_condition has wrt canonicalization.
1638 Note that these rtx constants are known to be CONST_INT, and
1639 therefore imply integer comparisons. */
1640 if (c == constm1_rtx && GET_CODE (cond) == GT)
1641 ;
1642 else if (c == const1_rtx && GET_CODE (cond) == LT)
1643 ;
1644 else if (c != CONST0_RTX (GET_MODE (b)))
1645 return FALSE;
1646
1647 /* Determine what sort of operation this is. */
1648 switch (GET_CODE (cond))
1649 {
1650 case LT:
1651 case LE:
1652 case UNLT:
1653 case UNLE:
1654 negate = !negate;
1655 break;
1656 case GT:
1657 case GE:
1658 case UNGT:
1659 case UNGE:
1660 break;
1661 default:
1662 return FALSE;
1663 }
1664
1665 start_sequence ();
1666
1667 target = expand_abs_nojump (GET_MODE (if_info->x), b, if_info->x, 1);
1668
1669 /* ??? It's a quandary whether cmove would be better here, especially
1670 for integers. Perhaps combine will clean things up. */
1671 if (target && negate)
1672 target = expand_simple_unop (GET_MODE (target), NEG, target, if_info->x, 0);
1673
1674 if (! target)
1675 {
1676 end_sequence ();
1677 return FALSE;
1678 }
1679
1680 if (target != if_info->x)
1681 noce_emit_move_insn (if_info->x, target);
1682
1683 seq = end_ifcvt_sequence (if_info);
1684 if (!seq)
1685 return FALSE;
1686
1687 emit_insn_before_setloc (seq, if_info->jump, INSN_LOCATOR (if_info->insn_a));
1688 if_info->cond = cond;
1689 if_info->cond_earliest = earliest;
1690
1691 return TRUE;
1692 }
1693
1694 /* Convert "if (m < 0) x = b; else x = 0;" to "x = (m >> C) & b;". */
1695
1696 static int
1697 noce_try_sign_mask (struct noce_if_info *if_info)
1698 {
1699 rtx cond, t, m, c, seq;
1700 enum machine_mode mode;
1701 enum rtx_code code;
1702
1703 if (no_new_pseudos)
1704 return FALSE;
1705
1706 cond = if_info->cond;
1707 code = GET_CODE (cond);
1708 m = XEXP (cond, 0);
1709 c = XEXP (cond, 1);
1710
1711 t = NULL_RTX;
1712 if (if_info->a == const0_rtx)
1713 {
1714 if ((code == LT && c == const0_rtx)
1715 || (code == LE && c == constm1_rtx))
1716 t = if_info->b;
1717 }
1718 else if (if_info->b == const0_rtx)
1719 {
1720 if ((code == GE && c == const0_rtx)
1721 || (code == GT && c == constm1_rtx))
1722 t = if_info->a;
1723 }
1724
1725 if (! t || side_effects_p (t))
1726 return FALSE;
1727
1728 /* We currently don't handle different modes. */
1729 mode = GET_MODE (t);
1730 if (GET_MODE (m) != mode)
1731 return FALSE;
1732
1733 /* This is only profitable if T is cheap. */
1734 if (rtx_cost (t, SET) >= COSTS_N_INSNS (2))
1735 return FALSE;
1736
1737 start_sequence ();
1738 c = gen_int_mode (GET_MODE_BITSIZE (mode) - 1, mode);
1739 m = expand_binop (mode, ashr_optab, m, c, NULL_RTX, 0, OPTAB_DIRECT);
1740 t = m ? expand_binop (mode, and_optab, m, t, NULL_RTX, 0, OPTAB_DIRECT)
1741 : NULL_RTX;
1742
1743 if (!t)
1744 {
1745 end_sequence ();
1746 return FALSE;
1747 }
1748
1749 noce_emit_move_insn (if_info->x, t);
1750
1751 seq = end_ifcvt_sequence (if_info);
1752 if (!seq)
1753 return FALSE;
1754
1755 emit_insn_before_setloc (seq, if_info->jump, INSN_LOCATOR (if_info->insn_a));
1756 return TRUE;
1757 }
1758
1759
1760 /* Similar to get_condition, only the resulting condition must be
1761 valid at JUMP, instead of at EARLIEST. */
1762
1763 static rtx
1764 noce_get_condition (rtx jump, rtx *earliest)
1765 {
1766 rtx cond, set, tmp, insn;
1767 bool reverse;
1768
1769 if (! any_condjump_p (jump))
1770 return NULL_RTX;
1771
1772 set = pc_set (jump);
1773
1774 /* If this branches to JUMP_LABEL when the condition is false,
1775 reverse the condition. */
1776 reverse = (GET_CODE (XEXP (SET_SRC (set), 2)) == LABEL_REF
1777 && XEXP (XEXP (SET_SRC (set), 2), 0) == JUMP_LABEL (jump));
1778
1779 /* If the condition variable is a register and is MODE_INT, accept it. */
1780
1781 cond = XEXP (SET_SRC (set), 0);
1782 tmp = XEXP (cond, 0);
1783 if (REG_P (tmp) && GET_MODE_CLASS (GET_MODE (tmp)) == MODE_INT)
1784 {
1785 *earliest = jump;
1786
1787 if (reverse)
1788 cond = gen_rtx_fmt_ee (reverse_condition (GET_CODE (cond)),
1789 GET_MODE (cond), tmp, XEXP (cond, 1));
1790 return cond;
1791 }
1792
1793 /* Otherwise, fall back on canonicalize_condition to do the dirty
1794 work of manipulating MODE_CC values and COMPARE rtx codes. */
1795
1796 tmp = canonicalize_condition (jump, cond, reverse, earliest, NULL_RTX,
1797 false);
1798 if (!tmp)
1799 return NULL_RTX;
1800
1801 /* We are going to insert code before JUMP, not before EARLIEST.
1802 We must therefore be certain that the given condition is valid
1803 at JUMP by virtue of not having been modified since. */
1804 for (insn = *earliest; insn != jump; insn = NEXT_INSN (insn))
1805 if (INSN_P (insn) && modified_in_p (tmp, insn))
1806 break;
1807 if (insn == jump)
1808 return tmp;
1809
1810 /* The condition was modified. See if we can get a partial result
1811 that doesn't follow all the reversals. Perhaps combine can fold
1812 them together later. */
1813 tmp = XEXP (tmp, 0);
1814 if (!REG_P (tmp) || GET_MODE_CLASS (GET_MODE (tmp)) != MODE_INT)
1815 return NULL_RTX;
1816 tmp = canonicalize_condition (jump, cond, reverse, earliest, tmp,
1817 false);
1818 if (!tmp)
1819 return NULL_RTX;
1820
1821 /* For sanity's sake, re-validate the new result. */
1822 for (insn = *earliest; insn != jump; insn = NEXT_INSN (insn))
1823 if (INSN_P (insn) && modified_in_p (tmp, insn))
1824 return NULL_RTX;
1825
1826 return tmp;
1827 }
1828
1829 /* Return true if OP is ok for if-then-else processing. */
1830
1831 static int
1832 noce_operand_ok (rtx op)
1833 {
1834 /* We special-case memories, so handle any of them with
1835 no address side effects. */
1836 if (MEM_P (op))
1837 return ! side_effects_p (XEXP (op, 0));
1838
1839 if (side_effects_p (op))
1840 return FALSE;
1841
1842 return ! may_trap_p (op);
1843 }
1844
1845 /* Given a simple IF-THEN or IF-THEN-ELSE block, attempt to convert it
1846 without using conditional execution. Return TRUE if we were
1847 successful at converting the block. */
1848
1849 static int
1850 noce_process_if_block (struct ce_if_block * ce_info)
1851 {
1852 basic_block test_bb = ce_info->test_bb; /* test block */
1853 basic_block then_bb = ce_info->then_bb; /* THEN */
1854 basic_block else_bb = ce_info->else_bb; /* ELSE or NULL */
1855 struct noce_if_info if_info;
1856 rtx insn_a, insn_b;
1857 rtx set_a, set_b;
1858 rtx orig_x, x, a, b;
1859 rtx jump, cond;
1860
1861 /* We're looking for patterns of the form
1862
1863 (1) if (...) x = a; else x = b;
1864 (2) x = b; if (...) x = a;
1865 (3) if (...) x = a; // as if with an initial x = x.
1866
1867 The later patterns require jumps to be more expensive.
1868
1869 ??? For future expansion, look for multiple X in such patterns. */
1870
1871 /* If test is comprised of && or || elements, don't handle it unless it is
1872 the special case of && elements without an ELSE block. */
1873 if (ce_info->num_multiple_test_blocks)
1874 {
1875 if (else_bb || ! ce_info->and_and_p)
1876 return FALSE;
1877
1878 ce_info->test_bb = test_bb = ce_info->last_test_bb;
1879 ce_info->num_multiple_test_blocks = 0;
1880 ce_info->num_and_and_blocks = 0;
1881 ce_info->num_or_or_blocks = 0;
1882 }
1883
1884 /* If this is not a standard conditional jump, we can't parse it. */
1885 jump = BB_END (test_bb);
1886 cond = noce_get_condition (jump, &if_info.cond_earliest);
1887 if (! cond)
1888 return FALSE;
1889
1890 /* If the conditional jump is more than just a conditional
1891 jump, then we can not do if-conversion on this block. */
1892 if (! onlyjump_p (jump))
1893 return FALSE;
1894
1895 /* We must be comparing objects whose modes imply the size. */
1896 if (GET_MODE (XEXP (cond, 0)) == BLKmode)
1897 return FALSE;
1898
1899 /* Look for one of the potential sets. */
1900 insn_a = first_active_insn (then_bb);
1901 if (! insn_a
1902 || insn_a != last_active_insn (then_bb, FALSE)
1903 || (set_a = single_set (insn_a)) == NULL_RTX)
1904 return FALSE;
1905
1906 x = SET_DEST (set_a);
1907 a = SET_SRC (set_a);
1908
1909 /* Look for the other potential set. Make sure we've got equivalent
1910 destinations. */
1911 /* ??? This is overconservative. Storing to two different mems is
1912 as easy as conditionally computing the address. Storing to a
1913 single mem merely requires a scratch memory to use as one of the
1914 destination addresses; often the memory immediately below the
1915 stack pointer is available for this. */
1916 set_b = NULL_RTX;
1917 if (else_bb)
1918 {
1919 insn_b = first_active_insn (else_bb);
1920 if (! insn_b
1921 || insn_b != last_active_insn (else_bb, FALSE)
1922 || (set_b = single_set (insn_b)) == NULL_RTX
1923 || ! rtx_equal_p (x, SET_DEST (set_b)))
1924 return FALSE;
1925 }
1926 else
1927 {
1928 insn_b = prev_nonnote_insn (if_info.cond_earliest);
1929 /* We're going to be moving the evaluation of B down from above
1930 COND_EARLIEST to JUMP. Make sure the relevant data is still
1931 intact. */
1932 if (! insn_b
1933 || GET_CODE (insn_b) != INSN
1934 || (set_b = single_set (insn_b)) == NULL_RTX
1935 || ! rtx_equal_p (x, SET_DEST (set_b))
1936 || reg_overlap_mentioned_p (x, SET_SRC (set_b))
1937 || modified_between_p (SET_SRC (set_b),
1938 PREV_INSN (if_info.cond_earliest), jump)
1939 /* Likewise with X. In particular this can happen when
1940 noce_get_condition looks farther back in the instruction
1941 stream than one might expect. */
1942 || reg_overlap_mentioned_p (x, cond)
1943 || reg_overlap_mentioned_p (x, a)
1944 || modified_between_p (x, PREV_INSN (if_info.cond_earliest), jump))
1945 insn_b = set_b = NULL_RTX;
1946 }
1947
1948 /* If x has side effects then only the if-then-else form is safe to
1949 convert. But even in that case we would need to restore any notes
1950 (such as REG_INC) at then end. That can be tricky if
1951 noce_emit_move_insn expands to more than one insn, so disable the
1952 optimization entirely for now if there are side effects. */
1953 if (side_effects_p (x))
1954 return FALSE;
1955
1956 b = (set_b ? SET_SRC (set_b) : x);
1957
1958 /* Only operate on register destinations, and even then avoid extending
1959 the lifetime of hard registers on small register class machines. */
1960 orig_x = x;
1961 if (!REG_P (x)
1962 || (SMALL_REGISTER_CLASSES
1963 && REGNO (x) < FIRST_PSEUDO_REGISTER))
1964 {
1965 if (no_new_pseudos || GET_MODE (x) == BLKmode)
1966 return FALSE;
1967 x = gen_reg_rtx (GET_MODE (GET_CODE (x) == STRICT_LOW_PART
1968 ? XEXP (x, 0) : x));
1969 }
1970
1971 /* Don't operate on sources that may trap or are volatile. */
1972 if (! noce_operand_ok (a) || ! noce_operand_ok (b))
1973 return FALSE;
1974
1975 /* Set up the info block for our subroutines. */
1976 if_info.test_bb = test_bb;
1977 if_info.cond = cond;
1978 if_info.jump = jump;
1979 if_info.insn_a = insn_a;
1980 if_info.insn_b = insn_b;
1981 if_info.x = x;
1982 if_info.a = a;
1983 if_info.b = b;
1984
1985 /* Try optimizations in some approximation of a useful order. */
1986 /* ??? Should first look to see if X is live incoming at all. If it
1987 isn't, we don't need anything but an unconditional set. */
1988
1989 /* Look and see if A and B are really the same. Avoid creating silly
1990 cmove constructs that no one will fix up later. */
1991 if (rtx_equal_p (a, b))
1992 {
1993 /* If we have an INSN_B, we don't have to create any new rtl. Just
1994 move the instruction that we already have. If we don't have an
1995 INSN_B, that means that A == X, and we've got a noop move. In
1996 that case don't do anything and let the code below delete INSN_A. */
1997 if (insn_b && else_bb)
1998 {
1999 rtx note;
2000
2001 if (else_bb && insn_b == BB_END (else_bb))
2002 BB_END (else_bb) = PREV_INSN (insn_b);
2003 reorder_insns (insn_b, insn_b, PREV_INSN (jump));
2004
2005 /* If there was a REG_EQUAL note, delete it since it may have been
2006 true due to this insn being after a jump. */
2007 if ((note = find_reg_note (insn_b, REG_EQUAL, NULL_RTX)) != 0)
2008 remove_note (insn_b, note);
2009
2010 insn_b = NULL_RTX;
2011 }
2012 /* If we have "x = b; if (...) x = a;", and x has side-effects, then
2013 x must be executed twice. */
2014 else if (insn_b && side_effects_p (orig_x))
2015 return FALSE;
2016
2017 x = orig_x;
2018 goto success;
2019 }
2020
2021 /* Disallow the "if (...) x = a;" form (with an implicit "else x = x;")
2022 for most optimizations if writing to x may trap, i.e. it's a memory
2023 other than a static var or a stack slot. */
2024 if (! set_b
2025 && MEM_P (orig_x)
2026 && ! MEM_NOTRAP_P (orig_x)
2027 && rtx_addr_can_trap_p (XEXP (orig_x, 0)))
2028 {
2029 if (HAVE_conditional_move)
2030 {
2031 if (noce_try_cmove (&if_info))
2032 goto success;
2033 if (! HAVE_conditional_execution
2034 && noce_try_cmove_arith (&if_info))
2035 goto success;
2036 }
2037 return FALSE;
2038 }
2039
2040 if (noce_try_move (&if_info))
2041 goto success;
2042 if (noce_try_store_flag (&if_info))
2043 goto success;
2044 if (noce_try_minmax (&if_info))
2045 goto success;
2046 if (noce_try_abs (&if_info))
2047 goto success;
2048 if (HAVE_conditional_move
2049 && noce_try_cmove (&if_info))
2050 goto success;
2051 if (! HAVE_conditional_execution)
2052 {
2053 if (noce_try_store_flag_constants (&if_info))
2054 goto success;
2055 if (noce_try_addcc (&if_info))
2056 goto success;
2057 if (noce_try_store_flag_mask (&if_info))
2058 goto success;
2059 if (HAVE_conditional_move
2060 && noce_try_cmove_arith (&if_info))
2061 goto success;
2062 if (noce_try_sign_mask (&if_info))
2063 goto success;
2064 }
2065
2066 return FALSE;
2067
2068 success:
2069 /* The original sets may now be killed. */
2070 delete_insn (insn_a);
2071
2072 /* Several special cases here: First, we may have reused insn_b above,
2073 in which case insn_b is now NULL. Second, we want to delete insn_b
2074 if it came from the ELSE block, because follows the now correct
2075 write that appears in the TEST block. However, if we got insn_b from
2076 the TEST block, it may in fact be loading data needed for the comparison.
2077 We'll let life_analysis remove the insn if it's really dead. */
2078 if (insn_b && else_bb)
2079 delete_insn (insn_b);
2080
2081 /* The new insns will have been inserted immediately before the jump. We
2082 should be able to remove the jump with impunity, but the condition itself
2083 may have been modified by gcse to be shared across basic blocks. */
2084 delete_insn (jump);
2085
2086 /* If we used a temporary, fix it up now. */
2087 if (orig_x != x)
2088 {
2089 start_sequence ();
2090 noce_emit_move_insn (orig_x, x);
2091 insn_b = get_insns ();
2092 set_used_flags (orig_x);
2093 unshare_all_rtl_in_chain (insn_b);
2094 end_sequence ();
2095
2096 emit_insn_after_setloc (insn_b, BB_END (test_bb), INSN_LOCATOR (insn_a));
2097 }
2098
2099 /* Merge the blocks! */
2100 merge_if_block (ce_info);
2101
2102 return TRUE;
2103 }
2104 \f
2105 /* Attempt to convert an IF-THEN or IF-THEN-ELSE block into
2106 straight line code. Return true if successful. */
2107
2108 static int
2109 process_if_block (struct ce_if_block * ce_info)
2110 {
2111 if (! reload_completed
2112 && noce_process_if_block (ce_info))
2113 return TRUE;
2114
2115 if (HAVE_conditional_execution && reload_completed)
2116 {
2117 /* If we have && and || tests, try to first handle combining the && and
2118 || tests into the conditional code, and if that fails, go back and
2119 handle it without the && and ||, which at present handles the && case
2120 if there was no ELSE block. */
2121 if (cond_exec_process_if_block (ce_info, TRUE))
2122 return TRUE;
2123
2124 if (ce_info->num_multiple_test_blocks)
2125 {
2126 cancel_changes (0);
2127
2128 if (cond_exec_process_if_block (ce_info, FALSE))
2129 return TRUE;
2130 }
2131 }
2132
2133 return FALSE;
2134 }
2135
2136 /* Merge the blocks and mark for local life update. */
2137
2138 static void
2139 merge_if_block (struct ce_if_block * ce_info)
2140 {
2141 basic_block test_bb = ce_info->test_bb; /* last test block */
2142 basic_block then_bb = ce_info->then_bb; /* THEN */
2143 basic_block else_bb = ce_info->else_bb; /* ELSE or NULL */
2144 basic_block join_bb = ce_info->join_bb; /* join block */
2145 basic_block combo_bb;
2146
2147 /* All block merging is done into the lower block numbers. */
2148
2149 combo_bb = test_bb;
2150
2151 /* Merge any basic blocks to handle && and || subtests. Each of
2152 the blocks are on the fallthru path from the predecessor block. */
2153 if (ce_info->num_multiple_test_blocks > 0)
2154 {
2155 basic_block bb = test_bb;
2156 basic_block last_test_bb = ce_info->last_test_bb;
2157 basic_block fallthru = block_fallthru (bb);
2158
2159 do
2160 {
2161 bb = fallthru;
2162 fallthru = block_fallthru (bb);
2163 merge_blocks (combo_bb, bb);
2164 num_true_changes++;
2165 }
2166 while (bb != last_test_bb);
2167 }
2168
2169 /* Merge TEST block into THEN block. Normally the THEN block won't have a
2170 label, but it might if there were || tests. That label's count should be
2171 zero, and it normally should be removed. */
2172
2173 if (then_bb)
2174 {
2175 if (combo_bb->global_live_at_end)
2176 COPY_REG_SET (combo_bb->global_live_at_end,
2177 then_bb->global_live_at_end);
2178 merge_blocks (combo_bb, then_bb);
2179 num_true_changes++;
2180 }
2181
2182 /* The ELSE block, if it existed, had a label. That label count
2183 will almost always be zero, but odd things can happen when labels
2184 get their addresses taken. */
2185 if (else_bb)
2186 {
2187 merge_blocks (combo_bb, else_bb);
2188 num_true_changes++;
2189 }
2190
2191 /* If there was no join block reported, that means it was not adjacent
2192 to the others, and so we cannot merge them. */
2193
2194 if (! join_bb)
2195 {
2196 rtx last = BB_END (combo_bb);
2197
2198 /* The outgoing edge for the current COMBO block should already
2199 be correct. Verify this. */
2200 if (combo_bb->succ == NULL_EDGE)
2201 {
2202 if (find_reg_note (last, REG_NORETURN, NULL))
2203 ;
2204 else if (GET_CODE (last) == INSN
2205 && GET_CODE (PATTERN (last)) == TRAP_IF
2206 && TRAP_CONDITION (PATTERN (last)) == const_true_rtx)
2207 ;
2208 else
2209 abort ();
2210 }
2211
2212 /* There should still be something at the end of the THEN or ELSE
2213 blocks taking us to our final destination. */
2214 else if (GET_CODE (last) == JUMP_INSN)
2215 ;
2216 else if (combo_bb->succ->dest == EXIT_BLOCK_PTR
2217 && GET_CODE (last) == CALL_INSN
2218 && SIBLING_CALL_P (last))
2219 ;
2220 else if ((combo_bb->succ->flags & EDGE_EH)
2221 && can_throw_internal (last))
2222 ;
2223 else
2224 abort ();
2225 }
2226
2227 /* The JOIN block may have had quite a number of other predecessors too.
2228 Since we've already merged the TEST, THEN and ELSE blocks, we should
2229 have only one remaining edge from our if-then-else diamond. If there
2230 is more than one remaining edge, it must come from elsewhere. There
2231 may be zero incoming edges if the THEN block didn't actually join
2232 back up (as with a call to abort). */
2233 else if ((join_bb->pred == NULL
2234 || join_bb->pred->pred_next == NULL)
2235 && join_bb != EXIT_BLOCK_PTR)
2236 {
2237 /* We can merge the JOIN. */
2238 if (combo_bb->global_live_at_end)
2239 COPY_REG_SET (combo_bb->global_live_at_end,
2240 join_bb->global_live_at_end);
2241
2242 merge_blocks (combo_bb, join_bb);
2243 num_true_changes++;
2244 }
2245 else
2246 {
2247 /* We cannot merge the JOIN. */
2248
2249 /* The outgoing edge for the current COMBO block should already
2250 be correct. Verify this. */
2251 if (combo_bb->succ->succ_next != NULL_EDGE
2252 || combo_bb->succ->dest != join_bb)
2253 abort ();
2254
2255 /* Remove the jump and cruft from the end of the COMBO block. */
2256 if (join_bb != EXIT_BLOCK_PTR)
2257 tidy_fallthru_edge (combo_bb->succ);
2258 }
2259
2260 num_updated_if_blocks++;
2261 }
2262 \f
2263 /* Find a block ending in a simple IF condition and try to transform it
2264 in some way. When converting a multi-block condition, put the new code
2265 in the first such block and delete the rest. Return a pointer to this
2266 first block if some transformation was done. Return NULL otherwise. */
2267
2268 static basic_block
2269 find_if_header (basic_block test_bb, int pass)
2270 {
2271 ce_if_block_t ce_info;
2272 edge then_edge;
2273 edge else_edge;
2274
2275 /* The kind of block we're looking for has exactly two successors. */
2276 if ((then_edge = test_bb->succ) == NULL_EDGE
2277 || (else_edge = then_edge->succ_next) == NULL_EDGE
2278 || else_edge->succ_next != NULL_EDGE)
2279 return NULL;
2280
2281 /* Neither edge should be abnormal. */
2282 if ((then_edge->flags & EDGE_COMPLEX)
2283 || (else_edge->flags & EDGE_COMPLEX))
2284 return NULL;
2285
2286 /* Nor exit the loop. */
2287 if ((then_edge->flags & EDGE_LOOP_EXIT)
2288 || (else_edge->flags & EDGE_LOOP_EXIT))
2289 return NULL;
2290
2291 /* The THEN edge is canonically the one that falls through. */
2292 if (then_edge->flags & EDGE_FALLTHRU)
2293 ;
2294 else if (else_edge->flags & EDGE_FALLTHRU)
2295 {
2296 edge e = else_edge;
2297 else_edge = then_edge;
2298 then_edge = e;
2299 }
2300 else
2301 /* Otherwise this must be a multiway branch of some sort. */
2302 return NULL;
2303
2304 memset (&ce_info, '\0', sizeof (ce_info));
2305 ce_info.test_bb = test_bb;
2306 ce_info.then_bb = then_edge->dest;
2307 ce_info.else_bb = else_edge->dest;
2308 ce_info.pass = pass;
2309
2310 #ifdef IFCVT_INIT_EXTRA_FIELDS
2311 IFCVT_INIT_EXTRA_FIELDS (&ce_info);
2312 #endif
2313
2314 if (find_if_block (&ce_info))
2315 goto success;
2316
2317 if (HAVE_trap && HAVE_conditional_trap
2318 && find_cond_trap (test_bb, then_edge, else_edge))
2319 goto success;
2320
2321 if (dom_computed[CDI_POST_DOMINATORS] >= DOM_NO_FAST_QUERY
2322 && (! HAVE_conditional_execution || reload_completed))
2323 {
2324 if (find_if_case_1 (test_bb, then_edge, else_edge))
2325 goto success;
2326 if (find_if_case_2 (test_bb, then_edge, else_edge))
2327 goto success;
2328 }
2329
2330 return NULL;
2331
2332 success:
2333 if (dump_file)
2334 fprintf (dump_file, "Conversion succeeded on pass %d.\n", pass);
2335 return ce_info.test_bb;
2336 }
2337
2338 /* Return true if a block has two edges, one of which falls through to the next
2339 block, and the other jumps to a specific block, so that we can tell if the
2340 block is part of an && test or an || test. Returns either -1 or the number
2341 of non-note, non-jump, non-USE/CLOBBER insns in the block. */
2342
2343 static int
2344 block_jumps_and_fallthru_p (basic_block cur_bb, basic_block target_bb)
2345 {
2346 edge cur_edge;
2347 int fallthru_p = FALSE;
2348 int jump_p = FALSE;
2349 rtx insn;
2350 rtx end;
2351 int n_insns = 0;
2352
2353 if (!cur_bb || !target_bb)
2354 return -1;
2355
2356 /* If no edges, obviously it doesn't jump or fallthru. */
2357 if (cur_bb->succ == NULL_EDGE)
2358 return FALSE;
2359
2360 for (cur_edge = cur_bb->succ;
2361 cur_edge != NULL_EDGE;
2362 cur_edge = cur_edge->succ_next)
2363 {
2364 if (cur_edge->flags & EDGE_COMPLEX)
2365 /* Anything complex isn't what we want. */
2366 return -1;
2367
2368 else if (cur_edge->flags & EDGE_FALLTHRU)
2369 fallthru_p = TRUE;
2370
2371 else if (cur_edge->dest == target_bb)
2372 jump_p = TRUE;
2373
2374 else
2375 return -1;
2376 }
2377
2378 if ((jump_p & fallthru_p) == 0)
2379 return -1;
2380
2381 /* Don't allow calls in the block, since this is used to group && and ||
2382 together for conditional execution support. ??? we should support
2383 conditional execution support across calls for IA-64 some day, but
2384 for now it makes the code simpler. */
2385 end = BB_END (cur_bb);
2386 insn = BB_HEAD (cur_bb);
2387
2388 while (insn != NULL_RTX)
2389 {
2390 if (GET_CODE (insn) == CALL_INSN)
2391 return -1;
2392
2393 if (INSN_P (insn)
2394 && GET_CODE (insn) != JUMP_INSN
2395 && GET_CODE (PATTERN (insn)) != USE
2396 && GET_CODE (PATTERN (insn)) != CLOBBER)
2397 n_insns++;
2398
2399 if (insn == end)
2400 break;
2401
2402 insn = NEXT_INSN (insn);
2403 }
2404
2405 return n_insns;
2406 }
2407
2408 /* Determine if a given basic block heads a simple IF-THEN or IF-THEN-ELSE
2409 block. If so, we'll try to convert the insns to not require the branch.
2410 Return TRUE if we were successful at converting the block. */
2411
2412 static int
2413 find_if_block (struct ce_if_block * ce_info)
2414 {
2415 basic_block test_bb = ce_info->test_bb;
2416 basic_block then_bb = ce_info->then_bb;
2417 basic_block else_bb = ce_info->else_bb;
2418 basic_block join_bb = NULL_BLOCK;
2419 edge then_succ = then_bb->succ;
2420 edge else_succ = else_bb->succ;
2421 int then_predecessors;
2422 int else_predecessors;
2423 edge cur_edge;
2424 basic_block next;
2425
2426 ce_info->last_test_bb = test_bb;
2427
2428 /* Discover if any fall through predecessors of the current test basic block
2429 were && tests (which jump to the else block) or || tests (which jump to
2430 the then block). */
2431 if (HAVE_conditional_execution && reload_completed
2432 && test_bb->pred != NULL_EDGE
2433 && test_bb->pred->pred_next == NULL_EDGE
2434 && test_bb->pred->flags == EDGE_FALLTHRU)
2435 {
2436 basic_block bb = test_bb->pred->src;
2437 basic_block target_bb;
2438 int max_insns = MAX_CONDITIONAL_EXECUTE;
2439 int n_insns;
2440
2441 /* Determine if the preceding block is an && or || block. */
2442 if ((n_insns = block_jumps_and_fallthru_p (bb, else_bb)) >= 0)
2443 {
2444 ce_info->and_and_p = TRUE;
2445 target_bb = else_bb;
2446 }
2447 else if ((n_insns = block_jumps_and_fallthru_p (bb, then_bb)) >= 0)
2448 {
2449 ce_info->and_and_p = FALSE;
2450 target_bb = then_bb;
2451 }
2452 else
2453 target_bb = NULL_BLOCK;
2454
2455 if (target_bb && n_insns <= max_insns)
2456 {
2457 int total_insns = 0;
2458 int blocks = 0;
2459
2460 ce_info->last_test_bb = test_bb;
2461
2462 /* Found at least one && or || block, look for more. */
2463 do
2464 {
2465 ce_info->test_bb = test_bb = bb;
2466 total_insns += n_insns;
2467 blocks++;
2468
2469 if (bb->pred == NULL_EDGE || bb->pred->pred_next != NULL_EDGE)
2470 break;
2471
2472 bb = bb->pred->src;
2473 n_insns = block_jumps_and_fallthru_p (bb, target_bb);
2474 }
2475 while (n_insns >= 0 && (total_insns + n_insns) <= max_insns);
2476
2477 ce_info->num_multiple_test_blocks = blocks;
2478 ce_info->num_multiple_test_insns = total_insns;
2479
2480 if (ce_info->and_and_p)
2481 ce_info->num_and_and_blocks = blocks;
2482 else
2483 ce_info->num_or_or_blocks = blocks;
2484 }
2485 }
2486
2487 /* Count the number of edges the THEN and ELSE blocks have. */
2488 then_predecessors = 0;
2489 for (cur_edge = then_bb->pred;
2490 cur_edge != NULL_EDGE;
2491 cur_edge = cur_edge->pred_next)
2492 {
2493 then_predecessors++;
2494 if (cur_edge->flags & EDGE_COMPLEX)
2495 return FALSE;
2496 }
2497
2498 else_predecessors = 0;
2499 for (cur_edge = else_bb->pred;
2500 cur_edge != NULL_EDGE;
2501 cur_edge = cur_edge->pred_next)
2502 {
2503 else_predecessors++;
2504 if (cur_edge->flags & EDGE_COMPLEX)
2505 return FALSE;
2506 }
2507
2508 /* The THEN block of an IF-THEN combo must have exactly one predecessor,
2509 other than any || blocks which jump to the THEN block. */
2510 if ((then_predecessors - ce_info->num_or_or_blocks) != 1)
2511 return FALSE;
2512
2513 /* The THEN block of an IF-THEN combo must have zero or one successors. */
2514 if (then_succ != NULL_EDGE
2515 && (then_succ->succ_next != NULL_EDGE
2516 || (then_succ->flags & EDGE_COMPLEX)
2517 || (flow2_completed && tablejump_p (BB_END (then_bb), NULL, NULL))))
2518 return FALSE;
2519
2520 /* If the THEN block has no successors, conditional execution can still
2521 make a conditional call. Don't do this unless the ELSE block has
2522 only one incoming edge -- the CFG manipulation is too ugly otherwise.
2523 Check for the last insn of the THEN block being an indirect jump, which
2524 is listed as not having any successors, but confuses the rest of the CE
2525 code processing. ??? we should fix this in the future. */
2526 if (then_succ == NULL)
2527 {
2528 if (else_bb->pred->pred_next == NULL_EDGE)
2529 {
2530 rtx last_insn = BB_END (then_bb);
2531
2532 while (last_insn
2533 && GET_CODE (last_insn) == NOTE
2534 && last_insn != BB_HEAD (then_bb))
2535 last_insn = PREV_INSN (last_insn);
2536
2537 if (last_insn
2538 && GET_CODE (last_insn) == JUMP_INSN
2539 && ! simplejump_p (last_insn))
2540 return FALSE;
2541
2542 join_bb = else_bb;
2543 else_bb = NULL_BLOCK;
2544 }
2545 else
2546 return FALSE;
2547 }
2548
2549 /* If the THEN block's successor is the other edge out of the TEST block,
2550 then we have an IF-THEN combo without an ELSE. */
2551 else if (then_succ->dest == else_bb)
2552 {
2553 join_bb = else_bb;
2554 else_bb = NULL_BLOCK;
2555 }
2556
2557 /* If the THEN and ELSE block meet in a subsequent block, and the ELSE
2558 has exactly one predecessor and one successor, and the outgoing edge
2559 is not complex, then we have an IF-THEN-ELSE combo. */
2560 else if (else_succ != NULL_EDGE
2561 && then_succ->dest == else_succ->dest
2562 && else_bb->pred->pred_next == NULL_EDGE
2563 && else_succ->succ_next == NULL_EDGE
2564 && ! (else_succ->flags & EDGE_COMPLEX)
2565 && ! (flow2_completed && tablejump_p (BB_END (else_bb), NULL, NULL)))
2566 join_bb = else_succ->dest;
2567
2568 /* Otherwise it is not an IF-THEN or IF-THEN-ELSE combination. */
2569 else
2570 return FALSE;
2571
2572 num_possible_if_blocks++;
2573
2574 if (dump_file)
2575 {
2576 fprintf (dump_file,
2577 "\nIF-THEN%s block found, pass %d, start block %d "
2578 "[insn %d], then %d [%d]",
2579 (else_bb) ? "-ELSE" : "",
2580 ce_info->pass,
2581 test_bb->index,
2582 BB_HEAD (test_bb) ? (int)INSN_UID (BB_HEAD (test_bb)) : -1,
2583 then_bb->index,
2584 BB_HEAD (then_bb) ? (int)INSN_UID (BB_HEAD (then_bb)) : -1);
2585
2586 if (else_bb)
2587 fprintf (dump_file, ", else %d [%d]",
2588 else_bb->index,
2589 BB_HEAD (else_bb) ? (int)INSN_UID (BB_HEAD (else_bb)) : -1);
2590
2591 fprintf (dump_file, ", join %d [%d]",
2592 join_bb->index,
2593 BB_HEAD (join_bb) ? (int)INSN_UID (BB_HEAD (join_bb)) : -1);
2594
2595 if (ce_info->num_multiple_test_blocks > 0)
2596 fprintf (dump_file, ", %d %s block%s last test %d [%d]",
2597 ce_info->num_multiple_test_blocks,
2598 (ce_info->and_and_p) ? "&&" : "||",
2599 (ce_info->num_multiple_test_blocks == 1) ? "" : "s",
2600 ce_info->last_test_bb->index,
2601 ((BB_HEAD (ce_info->last_test_bb))
2602 ? (int)INSN_UID (BB_HEAD (ce_info->last_test_bb))
2603 : -1));
2604
2605 fputc ('\n', dump_file);
2606 }
2607
2608 /* Make sure IF, THEN, and ELSE, blocks are adjacent. Actually, we get the
2609 first condition for free, since we've already asserted that there's a
2610 fallthru edge from IF to THEN. Likewise for the && and || blocks, since
2611 we checked the FALLTHRU flag, those are already adjacent to the last IF
2612 block. */
2613 /* ??? As an enhancement, move the ELSE block. Have to deal with
2614 BLOCK notes, if by no other means than aborting the merge if they
2615 exist. Sticky enough I don't want to think about it now. */
2616 next = then_bb;
2617 if (else_bb && (next = next->next_bb) != else_bb)
2618 return FALSE;
2619 if ((next = next->next_bb) != join_bb && join_bb != EXIT_BLOCK_PTR)
2620 {
2621 if (else_bb)
2622 join_bb = NULL;
2623 else
2624 return FALSE;
2625 }
2626
2627 /* Do the real work. */
2628 ce_info->else_bb = else_bb;
2629 ce_info->join_bb = join_bb;
2630
2631 return process_if_block (ce_info);
2632 }
2633
2634 /* Convert a branch over a trap, or a branch
2635 to a trap, into a conditional trap. */
2636
2637 static int
2638 find_cond_trap (basic_block test_bb, edge then_edge, edge else_edge)
2639 {
2640 basic_block then_bb = then_edge->dest;
2641 basic_block else_bb = else_edge->dest;
2642 basic_block other_bb, trap_bb;
2643 rtx trap, jump, cond, cond_earliest, seq;
2644 enum rtx_code code;
2645
2646 /* Locate the block with the trap instruction. */
2647 /* ??? While we look for no successors, we really ought to allow
2648 EH successors. Need to fix merge_if_block for that to work. */
2649 if ((trap = block_has_only_trap (then_bb)) != NULL)
2650 trap_bb = then_bb, other_bb = else_bb;
2651 else if ((trap = block_has_only_trap (else_bb)) != NULL)
2652 trap_bb = else_bb, other_bb = then_bb;
2653 else
2654 return FALSE;
2655
2656 if (dump_file)
2657 {
2658 fprintf (dump_file, "\nTRAP-IF block found, start %d, trap %d\n",
2659 test_bb->index, trap_bb->index);
2660 }
2661
2662 /* If this is not a standard conditional jump, we can't parse it. */
2663 jump = BB_END (test_bb);
2664 cond = noce_get_condition (jump, &cond_earliest);
2665 if (! cond)
2666 return FALSE;
2667
2668 /* If the conditional jump is more than just a conditional jump, then
2669 we can not do if-conversion on this block. */
2670 if (! onlyjump_p (jump))
2671 return FALSE;
2672
2673 /* We must be comparing objects whose modes imply the size. */
2674 if (GET_MODE (XEXP (cond, 0)) == BLKmode)
2675 return FALSE;
2676
2677 /* Reverse the comparison code, if necessary. */
2678 code = GET_CODE (cond);
2679 if (then_bb == trap_bb)
2680 {
2681 code = reversed_comparison_code (cond, jump);
2682 if (code == UNKNOWN)
2683 return FALSE;
2684 }
2685
2686 /* Attempt to generate the conditional trap. */
2687 seq = gen_cond_trap (code, XEXP (cond, 0),
2688 XEXP (cond, 1),
2689 TRAP_CODE (PATTERN (trap)));
2690 if (seq == NULL)
2691 return FALSE;
2692
2693 num_true_changes++;
2694
2695 /* Emit the new insns before cond_earliest. */
2696 emit_insn_before_setloc (seq, cond_earliest, INSN_LOCATOR (trap));
2697
2698 /* Delete the trap block if possible. */
2699 remove_edge (trap_bb == then_bb ? then_edge : else_edge);
2700 if (trap_bb->pred == NULL)
2701 delete_basic_block (trap_bb);
2702
2703 /* If the non-trap block and the test are now adjacent, merge them.
2704 Otherwise we must insert a direct branch. */
2705 if (test_bb->next_bb == other_bb)
2706 {
2707 struct ce_if_block new_ce_info;
2708 delete_insn (jump);
2709 memset (&new_ce_info, '\0', sizeof (new_ce_info));
2710 new_ce_info.test_bb = test_bb;
2711 new_ce_info.then_bb = NULL;
2712 new_ce_info.else_bb = NULL;
2713 new_ce_info.join_bb = other_bb;
2714 merge_if_block (&new_ce_info);
2715 }
2716 else
2717 {
2718 rtx lab, newjump;
2719
2720 lab = JUMP_LABEL (jump);
2721 newjump = emit_jump_insn_after (gen_jump (lab), jump);
2722 LABEL_NUSES (lab) += 1;
2723 JUMP_LABEL (newjump) = lab;
2724 emit_barrier_after (newjump);
2725
2726 delete_insn (jump);
2727 }
2728
2729 return TRUE;
2730 }
2731
2732 /* Subroutine of find_cond_trap: if BB contains only a trap insn,
2733 return it. */
2734
2735 static rtx
2736 block_has_only_trap (basic_block bb)
2737 {
2738 rtx trap;
2739
2740 /* We're not the exit block. */
2741 if (bb == EXIT_BLOCK_PTR)
2742 return NULL_RTX;
2743
2744 /* The block must have no successors. */
2745 if (bb->succ)
2746 return NULL_RTX;
2747
2748 /* The only instruction in the THEN block must be the trap. */
2749 trap = first_active_insn (bb);
2750 if (! (trap == BB_END (bb)
2751 && GET_CODE (PATTERN (trap)) == TRAP_IF
2752 && TRAP_CONDITION (PATTERN (trap)) == const_true_rtx))
2753 return NULL_RTX;
2754
2755 return trap;
2756 }
2757
2758 /* Look for IF-THEN-ELSE cases in which one of THEN or ELSE is
2759 transformable, but not necessarily the other. There need be no
2760 JOIN block.
2761
2762 Return TRUE if we were successful at converting the block.
2763
2764 Cases we'd like to look at:
2765
2766 (1)
2767 if (test) goto over; // x not live
2768 x = a;
2769 goto label;
2770 over:
2771
2772 becomes
2773
2774 x = a;
2775 if (! test) goto label;
2776
2777 (2)
2778 if (test) goto E; // x not live
2779 x = big();
2780 goto L;
2781 E:
2782 x = b;
2783 goto M;
2784
2785 becomes
2786
2787 x = b;
2788 if (test) goto M;
2789 x = big();
2790 goto L;
2791
2792 (3) // This one's really only interesting for targets that can do
2793 // multiway branching, e.g. IA-64 BBB bundles. For other targets
2794 // it results in multiple branches on a cache line, which often
2795 // does not sit well with predictors.
2796
2797 if (test1) goto E; // predicted not taken
2798 x = a;
2799 if (test2) goto F;
2800 ...
2801 E:
2802 x = b;
2803 J:
2804
2805 becomes
2806
2807 x = a;
2808 if (test1) goto E;
2809 if (test2) goto F;
2810
2811 Notes:
2812
2813 (A) Don't do (2) if the branch is predicted against the block we're
2814 eliminating. Do it anyway if we can eliminate a branch; this requires
2815 that the sole successor of the eliminated block postdominate the other
2816 side of the if.
2817
2818 (B) With CE, on (3) we can steal from both sides of the if, creating
2819
2820 if (test1) x = a;
2821 if (!test1) x = b;
2822 if (test1) goto J;
2823 if (test2) goto F;
2824 ...
2825 J:
2826
2827 Again, this is most useful if J postdominates.
2828
2829 (C) CE substitutes for helpful life information.
2830
2831 (D) These heuristics need a lot of work. */
2832
2833 /* Tests for case 1 above. */
2834
2835 static int
2836 find_if_case_1 (basic_block test_bb, edge then_edge, edge else_edge)
2837 {
2838 basic_block then_bb = then_edge->dest;
2839 basic_block else_bb = else_edge->dest, new_bb;
2840 edge then_succ = then_bb->succ;
2841 int then_bb_index;
2842
2843 /* If we are partitioning hot/cold basic blocks, we don't want to
2844 mess up unconditional or indirect jumps that cross between hot
2845 and cold sections. */
2846
2847 if (flag_reorder_blocks_and_partition
2848 && ((BB_END (then_bb)
2849 && find_reg_note (BB_END (then_bb), REG_CROSSING_JUMP, NULL_RTX))
2850 || (BB_END (else_bb)
2851 && find_reg_note (BB_END (else_bb), REG_CROSSING_JUMP,
2852 NULL_RTX))))
2853 return FALSE;
2854
2855 /* THEN has one successor. */
2856 if (!then_succ || then_succ->succ_next != NULL)
2857 return FALSE;
2858
2859 /* THEN does not fall through, but is not strange either. */
2860 if (then_succ->flags & (EDGE_COMPLEX | EDGE_FALLTHRU))
2861 return FALSE;
2862
2863 /* THEN has one predecessor. */
2864 if (then_bb->pred->pred_next != NULL)
2865 return FALSE;
2866
2867 /* THEN must do something. */
2868 if (forwarder_block_p (then_bb))
2869 return FALSE;
2870
2871 num_possible_if_blocks++;
2872 if (dump_file)
2873 fprintf (dump_file,
2874 "\nIF-CASE-1 found, start %d, then %d\n",
2875 test_bb->index, then_bb->index);
2876
2877 /* THEN is small. */
2878 if (count_bb_insns (then_bb) > BRANCH_COST)
2879 return FALSE;
2880
2881 /* Registers set are dead, or are predicable. */
2882 if (! dead_or_predicable (test_bb, then_bb, else_bb,
2883 then_bb->succ->dest, 1))
2884 return FALSE;
2885
2886 /* Conversion went ok, including moving the insns and fixing up the
2887 jump. Adjust the CFG to match. */
2888
2889 bitmap_operation (test_bb->global_live_at_end,
2890 else_bb->global_live_at_start,
2891 then_bb->global_live_at_end, BITMAP_IOR);
2892
2893 new_bb = redirect_edge_and_branch_force (FALLTHRU_EDGE (test_bb), else_bb);
2894 then_bb_index = then_bb->index;
2895 delete_basic_block (then_bb);
2896
2897 /* Make rest of code believe that the newly created block is the THEN_BB
2898 block we removed. */
2899 if (new_bb)
2900 {
2901 new_bb->index = then_bb_index;
2902 BASIC_BLOCK (then_bb_index) = new_bb;
2903 }
2904 /* We've possibly created jump to next insn, cleanup_cfg will solve that
2905 later. */
2906
2907 num_true_changes++;
2908 num_updated_if_blocks++;
2909
2910 return TRUE;
2911 }
2912
2913 /* Test for case 2 above. */
2914
2915 static int
2916 find_if_case_2 (basic_block test_bb, edge then_edge, edge else_edge)
2917 {
2918 basic_block then_bb = then_edge->dest;
2919 basic_block else_bb = else_edge->dest;
2920 edge else_succ = else_bb->succ;
2921 rtx note;
2922
2923 /* If we are partitioning hot/cold basic blocks, we don't want to
2924 mess up unconditional or indirect jumps that cross between hot
2925 and cold sections. */
2926
2927 if (flag_reorder_blocks_and_partition
2928 && ((BB_END (then_bb)
2929 && find_reg_note (BB_END (then_bb), REG_CROSSING_JUMP, NULL_RTX))
2930 || (BB_END (else_bb)
2931 && find_reg_note (BB_END (else_bb), REG_CROSSING_JUMP,
2932 NULL_RTX))))
2933 return FALSE;
2934
2935 /* ELSE has one successor. */
2936 if (!else_succ || else_succ->succ_next != NULL)
2937 return FALSE;
2938
2939 /* ELSE outgoing edge is not complex. */
2940 if (else_succ->flags & EDGE_COMPLEX)
2941 return FALSE;
2942
2943 /* ELSE has one predecessor. */
2944 if (else_bb->pred->pred_next != NULL)
2945 return FALSE;
2946
2947 /* THEN is not EXIT. */
2948 if (then_bb->index < 0)
2949 return FALSE;
2950
2951 /* ELSE is predicted or SUCC(ELSE) postdominates THEN. */
2952 note = find_reg_note (BB_END (test_bb), REG_BR_PROB, NULL_RTX);
2953 if (note && INTVAL (XEXP (note, 0)) >= REG_BR_PROB_BASE / 2)
2954 ;
2955 else if (else_succ->dest->index < 0
2956 || dominated_by_p (CDI_POST_DOMINATORS, then_bb,
2957 else_succ->dest))
2958 ;
2959 else
2960 return FALSE;
2961
2962 num_possible_if_blocks++;
2963 if (dump_file)
2964 fprintf (dump_file,
2965 "\nIF-CASE-2 found, start %d, else %d\n",
2966 test_bb->index, else_bb->index);
2967
2968 /* ELSE is small. */
2969 if (count_bb_insns (else_bb) > BRANCH_COST)
2970 return FALSE;
2971
2972 /* Registers set are dead, or are predicable. */
2973 if (! dead_or_predicable (test_bb, else_bb, then_bb, else_succ->dest, 0))
2974 return FALSE;
2975
2976 /* Conversion went ok, including moving the insns and fixing up the
2977 jump. Adjust the CFG to match. */
2978
2979 bitmap_operation (test_bb->global_live_at_end,
2980 then_bb->global_live_at_start,
2981 else_bb->global_live_at_end, BITMAP_IOR);
2982
2983 delete_basic_block (else_bb);
2984
2985 num_true_changes++;
2986 num_updated_if_blocks++;
2987
2988 /* ??? We may now fallthru from one of THEN's successors into a join
2989 block. Rerun cleanup_cfg? Examine things manually? Wait? */
2990
2991 return TRUE;
2992 }
2993
2994 /* A subroutine of dead_or_predicable called through for_each_rtx.
2995 Return 1 if a memory is found. */
2996
2997 static int
2998 find_memory (rtx *px, void *data ATTRIBUTE_UNUSED)
2999 {
3000 return MEM_P (*px);
3001 }
3002
3003 /* Used by the code above to perform the actual rtl transformations.
3004 Return TRUE if successful.
3005
3006 TEST_BB is the block containing the conditional branch. MERGE_BB
3007 is the block containing the code to manipulate. NEW_DEST is the
3008 label TEST_BB should be branching to after the conversion.
3009 REVERSEP is true if the sense of the branch should be reversed. */
3010
3011 static int
3012 dead_or_predicable (basic_block test_bb, basic_block merge_bb,
3013 basic_block other_bb, basic_block new_dest, int reversep)
3014 {
3015 rtx head, end, jump, earliest = NULL_RTX, old_dest, new_label = NULL_RTX;
3016
3017 jump = BB_END (test_bb);
3018
3019 /* Find the extent of the real code in the merge block. */
3020 head = BB_HEAD (merge_bb);
3021 end = BB_END (merge_bb);
3022
3023 if (GET_CODE (head) == CODE_LABEL)
3024 head = NEXT_INSN (head);
3025 if (GET_CODE (head) == NOTE)
3026 {
3027 if (head == end)
3028 {
3029 head = end = NULL_RTX;
3030 goto no_body;
3031 }
3032 head = NEXT_INSN (head);
3033 }
3034
3035 if (GET_CODE (end) == JUMP_INSN)
3036 {
3037 if (head == end)
3038 {
3039 head = end = NULL_RTX;
3040 goto no_body;
3041 }
3042 end = PREV_INSN (end);
3043 }
3044
3045 /* Disable handling dead code by conditional execution if the machine needs
3046 to do anything funny with the tests, etc. */
3047 #ifndef IFCVT_MODIFY_TESTS
3048 if (HAVE_conditional_execution)
3049 {
3050 /* In the conditional execution case, we have things easy. We know
3051 the condition is reversible. We don't have to check life info
3052 because we're going to conditionally execute the code anyway.
3053 All that's left is making sure the insns involved can actually
3054 be predicated. */
3055
3056 rtx cond, prob_val;
3057
3058 cond = cond_exec_get_condition (jump);
3059 if (! cond)
3060 return FALSE;
3061
3062 prob_val = find_reg_note (jump, REG_BR_PROB, NULL_RTX);
3063 if (prob_val)
3064 prob_val = XEXP (prob_val, 0);
3065
3066 if (reversep)
3067 {
3068 enum rtx_code rev = reversed_comparison_code (cond, jump);
3069 if (rev == UNKNOWN)
3070 return FALSE;
3071 cond = gen_rtx_fmt_ee (rev, GET_MODE (cond), XEXP (cond, 0),
3072 XEXP (cond, 1));
3073 if (prob_val)
3074 prob_val = GEN_INT (REG_BR_PROB_BASE - INTVAL (prob_val));
3075 }
3076
3077 if (! cond_exec_process_insns ((ce_if_block_t *)0, head, end, cond,
3078 prob_val, 0))
3079 goto cancel;
3080
3081 earliest = jump;
3082 }
3083 else
3084 #endif
3085 {
3086 /* In the non-conditional execution case, we have to verify that there
3087 are no trapping operations, no calls, no references to memory, and
3088 that any registers modified are dead at the branch site. */
3089
3090 rtx insn, cond, prev;
3091 regset_head merge_set_head, tmp_head, test_live_head, test_set_head;
3092 regset merge_set, tmp, test_live, test_set;
3093 struct propagate_block_info *pbi;
3094 int i, fail = 0;
3095
3096 /* Check for no calls or trapping operations. */
3097 for (insn = head; ; insn = NEXT_INSN (insn))
3098 {
3099 if (GET_CODE (insn) == CALL_INSN)
3100 return FALSE;
3101 if (INSN_P (insn))
3102 {
3103 if (may_trap_p (PATTERN (insn)))
3104 return FALSE;
3105
3106 /* ??? Even non-trapping memories such as stack frame
3107 references must be avoided. For stores, we collect
3108 no lifetime info; for reads, we'd have to assert
3109 true_dependence false against every store in the
3110 TEST range. */
3111 if (for_each_rtx (&PATTERN (insn), find_memory, NULL))
3112 return FALSE;
3113 }
3114 if (insn == end)
3115 break;
3116 }
3117
3118 if (! any_condjump_p (jump))
3119 return FALSE;
3120
3121 /* Find the extent of the conditional. */
3122 cond = noce_get_condition (jump, &earliest);
3123 if (! cond)
3124 return FALSE;
3125
3126 /* Collect:
3127 MERGE_SET = set of registers set in MERGE_BB
3128 TEST_LIVE = set of registers live at EARLIEST
3129 TEST_SET = set of registers set between EARLIEST and the
3130 end of the block. */
3131
3132 tmp = INITIALIZE_REG_SET (tmp_head);
3133 merge_set = INITIALIZE_REG_SET (merge_set_head);
3134 test_live = INITIALIZE_REG_SET (test_live_head);
3135 test_set = INITIALIZE_REG_SET (test_set_head);
3136
3137 /* ??? bb->local_set is only valid during calculate_global_regs_live,
3138 so we must recompute usage for MERGE_BB. Not so bad, I suppose,
3139 since we've already asserted that MERGE_BB is small. */
3140 propagate_block (merge_bb, tmp, merge_set, merge_set, 0);
3141
3142 /* For small register class machines, don't lengthen lifetimes of
3143 hard registers before reload. */
3144 if (SMALL_REGISTER_CLASSES && ! reload_completed)
3145 {
3146 EXECUTE_IF_SET_IN_BITMAP
3147 (merge_set, 0, i,
3148 {
3149 if (i < FIRST_PSEUDO_REGISTER
3150 && ! fixed_regs[i]
3151 && ! global_regs[i])
3152 fail = 1;
3153 });
3154 }
3155
3156 /* For TEST, we're interested in a range of insns, not a whole block.
3157 Moreover, we're interested in the insns live from OTHER_BB. */
3158
3159 COPY_REG_SET (test_live, other_bb->global_live_at_start);
3160 pbi = init_propagate_block_info (test_bb, test_live, test_set, test_set,
3161 0);
3162
3163 for (insn = jump; ; insn = prev)
3164 {
3165 prev = propagate_one_insn (pbi, insn);
3166 if (insn == earliest)
3167 break;
3168 }
3169
3170 free_propagate_block_info (pbi);
3171
3172 /* We can perform the transformation if
3173 MERGE_SET & (TEST_SET | TEST_LIVE)
3174 and
3175 TEST_SET & merge_bb->global_live_at_start
3176 are empty. */
3177
3178 bitmap_operation (tmp, test_set, test_live, BITMAP_IOR);
3179 bitmap_operation (tmp, tmp, merge_set, BITMAP_AND);
3180 EXECUTE_IF_SET_IN_BITMAP(tmp, 0, i, fail = 1);
3181
3182 bitmap_operation (tmp, test_set, merge_bb->global_live_at_start,
3183 BITMAP_AND);
3184 EXECUTE_IF_SET_IN_BITMAP(tmp, 0, i, fail = 1);
3185
3186 FREE_REG_SET (tmp);
3187 FREE_REG_SET (merge_set);
3188 FREE_REG_SET (test_live);
3189 FREE_REG_SET (test_set);
3190
3191 if (fail)
3192 return FALSE;
3193 }
3194
3195 no_body:
3196 /* We don't want to use normal invert_jump or redirect_jump because
3197 we don't want to delete_insn called. Also, we want to do our own
3198 change group management. */
3199
3200 old_dest = JUMP_LABEL (jump);
3201 if (other_bb != new_dest)
3202 {
3203 new_label = block_label (new_dest);
3204 if (reversep
3205 ? ! invert_jump_1 (jump, new_label)
3206 : ! redirect_jump_1 (jump, new_label))
3207 goto cancel;
3208 }
3209
3210 if (! apply_change_group ())
3211 return FALSE;
3212
3213 if (other_bb != new_dest)
3214 {
3215 if (old_dest)
3216 LABEL_NUSES (old_dest) -= 1;
3217 if (new_label)
3218 LABEL_NUSES (new_label) += 1;
3219 JUMP_LABEL (jump) = new_label;
3220 if (reversep)
3221 invert_br_probabilities (jump);
3222
3223 redirect_edge_succ (BRANCH_EDGE (test_bb), new_dest);
3224 if (reversep)
3225 {
3226 gcov_type count, probability;
3227 count = BRANCH_EDGE (test_bb)->count;
3228 BRANCH_EDGE (test_bb)->count = FALLTHRU_EDGE (test_bb)->count;
3229 FALLTHRU_EDGE (test_bb)->count = count;
3230 probability = BRANCH_EDGE (test_bb)->probability;
3231 BRANCH_EDGE (test_bb)->probability
3232 = FALLTHRU_EDGE (test_bb)->probability;
3233 FALLTHRU_EDGE (test_bb)->probability = probability;
3234 update_br_prob_note (test_bb);
3235 }
3236 }
3237
3238 /* Move the insns out of MERGE_BB to before the branch. */
3239 if (head != NULL)
3240 {
3241 if (end == BB_END (merge_bb))
3242 BB_END (merge_bb) = PREV_INSN (head);
3243
3244 if (squeeze_notes (&head, &end))
3245 return TRUE;
3246
3247 reorder_insns (head, end, PREV_INSN (earliest));
3248 }
3249
3250 /* Remove the jump and edge if we can. */
3251 if (other_bb == new_dest)
3252 {
3253 delete_insn (jump);
3254 remove_edge (BRANCH_EDGE (test_bb));
3255 /* ??? Can't merge blocks here, as then_bb is still in use.
3256 At minimum, the merge will get done just before bb-reorder. */
3257 }
3258
3259 return TRUE;
3260
3261 cancel:
3262 cancel_changes (0);
3263 return FALSE;
3264 }
3265 \f
3266 /* Main entry point for all if-conversion. */
3267
3268 void
3269 if_convert (int x_life_data_ok)
3270 {
3271 basic_block bb;
3272 int pass;
3273
3274 num_possible_if_blocks = 0;
3275 num_updated_if_blocks = 0;
3276 num_true_changes = 0;
3277 life_data_ok = (x_life_data_ok != 0);
3278
3279 if ((! targetm.cannot_modify_jumps_p ())
3280 && (!flag_reorder_blocks_and_partition || !no_new_pseudos))
3281 mark_loop_exit_edges ();
3282
3283 /* Compute postdominators if we think we'll use them. */
3284 if (HAVE_conditional_execution || life_data_ok)
3285 calculate_dominance_info (CDI_POST_DOMINATORS);
3286
3287 if (life_data_ok)
3288 clear_bb_flags ();
3289
3290 /* Go through each of the basic blocks looking for things to convert. If we
3291 have conditional execution, we make multiple passes to allow us to handle
3292 IF-THEN{-ELSE} blocks within other IF-THEN{-ELSE} blocks. */
3293 pass = 0;
3294 do
3295 {
3296 cond_exec_changed_p = FALSE;
3297 pass++;
3298
3299 #ifdef IFCVT_MULTIPLE_DUMPS
3300 if (dump_file && pass > 1)
3301 fprintf (dump_file, "\n\n========== Pass %d ==========\n", pass);
3302 #endif
3303
3304 FOR_EACH_BB (bb)
3305 {
3306 basic_block new_bb;
3307 while ((new_bb = find_if_header (bb, pass)))
3308 bb = new_bb;
3309 }
3310
3311 #ifdef IFCVT_MULTIPLE_DUMPS
3312 if (dump_file && cond_exec_changed_p)
3313 print_rtl_with_bb (dump_file, get_insns ());
3314 #endif
3315 }
3316 while (cond_exec_changed_p);
3317
3318 #ifdef IFCVT_MULTIPLE_DUMPS
3319 if (dump_file)
3320 fprintf (dump_file, "\n\n========== no more changes\n");
3321 #endif
3322
3323 free_dominance_info (CDI_POST_DOMINATORS);
3324
3325 if (dump_file)
3326 fflush (dump_file);
3327
3328 clear_aux_for_blocks ();
3329
3330 /* Rebuild life info for basic blocks that require it. */
3331 if (num_true_changes && life_data_ok)
3332 {
3333 /* If we allocated new pseudos, we must resize the array for sched1. */
3334 if (max_regno < max_reg_num ())
3335 {
3336 max_regno = max_reg_num ();
3337 allocate_reg_info (max_regno, FALSE, FALSE);
3338 }
3339 update_life_info_in_dirty_blocks (UPDATE_LIFE_GLOBAL_RM_NOTES,
3340 PROP_DEATH_NOTES | PROP_SCAN_DEAD_CODE
3341 | PROP_KILL_DEAD_CODE);
3342 }
3343
3344 /* Write the final stats. */
3345 if (dump_file && num_possible_if_blocks > 0)
3346 {
3347 fprintf (dump_file,
3348 "\n%d possible IF blocks searched.\n",
3349 num_possible_if_blocks);
3350 fprintf (dump_file,
3351 "%d IF blocks converted.\n",
3352 num_updated_if_blocks);
3353 fprintf (dump_file,
3354 "%d true changes made.\n\n\n",
3355 num_true_changes);
3356 }
3357
3358 #ifdef ENABLE_CHECKING
3359 verify_flow_info ();
3360 #endif
3361 }