re PR tree-optimization/82129 (ICE in compute_antic, at tree-ssa-pre.c:2447)
[gcc.git] / gcc / tree-ssa-pre.c
1 /* Full and partial redundancy elimination and code hoisting on SSA GIMPLE.
2 Copyright (C) 2001-2017 Free Software Foundation, Inc.
3 Contributed by Daniel Berlin <dan@dberlin.org> and Steven Bosscher
4 <stevenb@suse.de>
5
6 This file is part of GCC.
7
8 GCC is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 3, or (at your option)
11 any later version.
12
13 GCC is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
17
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
21
22 #include "config.h"
23 #include "system.h"
24 #include "coretypes.h"
25 #include "backend.h"
26 #include "rtl.h"
27 #include "tree.h"
28 #include "gimple.h"
29 #include "predict.h"
30 #include "alloc-pool.h"
31 #include "tree-pass.h"
32 #include "ssa.h"
33 #include "cgraph.h"
34 #include "gimple-pretty-print.h"
35 #include "fold-const.h"
36 #include "cfganal.h"
37 #include "gimple-fold.h"
38 #include "tree-eh.h"
39 #include "gimplify.h"
40 #include "gimple-iterator.h"
41 #include "tree-cfg.h"
42 #include "tree-ssa-loop.h"
43 #include "tree-into-ssa.h"
44 #include "tree-dfa.h"
45 #include "tree-ssa.h"
46 #include "cfgloop.h"
47 #include "tree-ssa-sccvn.h"
48 #include "tree-scalar-evolution.h"
49 #include "params.h"
50 #include "dbgcnt.h"
51 #include "domwalk.h"
52 #include "tree-ssa-propagate.h"
53 #include "ipa-utils.h"
54 #include "tree-cfgcleanup.h"
55 #include "langhooks.h"
56 #include "alias.h"
57
58 /* Even though this file is called tree-ssa-pre.c, we actually
59 implement a bit more than just PRE here. All of them piggy-back
60 on GVN which is implemented in tree-ssa-sccvn.c.
61
62 1. Full Redundancy Elimination (FRE)
63 This is the elimination phase of GVN.
64
65 2. Partial Redundancy Elimination (PRE)
66 This is adds computation of AVAIL_OUT and ANTIC_IN and
67 doing expression insertion to form GVN-PRE.
68
69 3. Code hoisting
70 This optimization uses the ANTIC_IN sets computed for PRE
71 to move expressions further up than PRE would do, to make
72 multiple computations of the same value fully redundant.
73 This pass is explained below (after the explanation of the
74 basic algorithm for PRE).
75 */
76
77 /* TODO:
78
79 1. Avail sets can be shared by making an avail_find_leader that
80 walks up the dominator tree and looks in those avail sets.
81 This might affect code optimality, it's unclear right now.
82 Currently the AVAIL_OUT sets are the remaining quadraticness in
83 memory of GVN-PRE.
84 2. Strength reduction can be performed by anticipating expressions
85 we can repair later on.
86 3. We can do back-substitution or smarter value numbering to catch
87 commutative expressions split up over multiple statements.
88 */
89
90 /* For ease of terminology, "expression node" in the below refers to
91 every expression node but GIMPLE_ASSIGN, because GIMPLE_ASSIGNs
92 represent the actual statement containing the expressions we care about,
93 and we cache the value number by putting it in the expression. */
94
95 /* Basic algorithm for Partial Redundancy Elimination:
96
97 First we walk the statements to generate the AVAIL sets, the
98 EXP_GEN sets, and the tmp_gen sets. EXP_GEN sets represent the
99 generation of values/expressions by a given block. We use them
100 when computing the ANTIC sets. The AVAIL sets consist of
101 SSA_NAME's that represent values, so we know what values are
102 available in what blocks. AVAIL is a forward dataflow problem. In
103 SSA, values are never killed, so we don't need a kill set, or a
104 fixpoint iteration, in order to calculate the AVAIL sets. In
105 traditional parlance, AVAIL sets tell us the downsafety of the
106 expressions/values.
107
108 Next, we generate the ANTIC sets. These sets represent the
109 anticipatable expressions. ANTIC is a backwards dataflow
110 problem. An expression is anticipatable in a given block if it could
111 be generated in that block. This means that if we had to perform
112 an insertion in that block, of the value of that expression, we
113 could. Calculating the ANTIC sets requires phi translation of
114 expressions, because the flow goes backwards through phis. We must
115 iterate to a fixpoint of the ANTIC sets, because we have a kill
116 set. Even in SSA form, values are not live over the entire
117 function, only from their definition point onwards. So we have to
118 remove values from the ANTIC set once we go past the definition
119 point of the leaders that make them up.
120 compute_antic/compute_antic_aux performs this computation.
121
122 Third, we perform insertions to make partially redundant
123 expressions fully redundant.
124
125 An expression is partially redundant (excluding partial
126 anticipation) if:
127
128 1. It is AVAIL in some, but not all, of the predecessors of a
129 given block.
130 2. It is ANTIC in all the predecessors.
131
132 In order to make it fully redundant, we insert the expression into
133 the predecessors where it is not available, but is ANTIC.
134
135 When optimizing for size, we only eliminate the partial redundancy
136 if we need to insert in only one predecessor. This avoids almost
137 completely the code size increase that PRE usually causes.
138
139 For the partial anticipation case, we only perform insertion if it
140 is partially anticipated in some block, and fully available in all
141 of the predecessors.
142
143 do_pre_regular_insertion/do_pre_partial_partial_insertion
144 performs these steps, driven by insert/insert_aux.
145
146 Fourth, we eliminate fully redundant expressions.
147 This is a simple statement walk that replaces redundant
148 calculations with the now available values. */
149
150 /* Basic algorithm for Code Hoisting:
151
152 Code hoisting is: Moving value computations up in the control flow
153 graph to make multiple copies redundant. Typically this is a size
154 optimization, but there are cases where it also is helpful for speed.
155
156 A simple code hoisting algorithm is implemented that piggy-backs on
157 the PRE infrastructure. For code hoisting, we have to know ANTIC_OUT
158 which is effectively ANTIC_IN - AVAIL_OUT. The latter two have to be
159 computed for PRE, and we can use them to perform a limited version of
160 code hoisting, too.
161
162 For the purpose of this implementation, a value is hoistable to a basic
163 block B if the following properties are met:
164
165 1. The value is in ANTIC_IN(B) -- the value will be computed on all
166 paths from B to function exit and it can be computed in B);
167
168 2. The value is not in AVAIL_OUT(B) -- there would be no need to
169 compute the value again and make it available twice;
170
171 3. All successors of B are dominated by B -- makes sure that inserting
172 a computation of the value in B will make the remaining
173 computations fully redundant;
174
175 4. At least one successor has the value in AVAIL_OUT -- to avoid
176 hoisting values up too far;
177
178 5. There are at least two successors of B -- hoisting in straight
179 line code is pointless.
180
181 The third condition is not strictly necessary, but it would complicate
182 the hoisting pass a lot. In fact, I don't know of any code hoisting
183 algorithm that does not have this requirement. Fortunately, experiments
184 have show that most candidate hoistable values are in regions that meet
185 this condition (e.g. diamond-shape regions).
186
187 The forth condition is necessary to avoid hoisting things up too far
188 away from the uses of the value. Nothing else limits the algorithm
189 from hoisting everything up as far as ANTIC_IN allows. Experiments
190 with SPEC and CSiBE have shown that hoisting up too far results in more
191 spilling, less benefits for code size, and worse benchmark scores.
192 Fortunately, in practice most of the interesting hoisting opportunities
193 are caught despite this limitation.
194
195 For hoistable values that meet all conditions, expressions are inserted
196 to make the calculation of the hoistable value fully redundant. We
197 perform code hoisting insertions after each round of PRE insertions,
198 because code hoisting never exposes new PRE opportunities, but PRE can
199 create new code hoisting opportunities.
200
201 The code hoisting algorithm is implemented in do_hoist_insert, driven
202 by insert/insert_aux. */
203
204 /* Representations of value numbers:
205
206 Value numbers are represented by a representative SSA_NAME. We
207 will create fake SSA_NAME's in situations where we need a
208 representative but do not have one (because it is a complex
209 expression). In order to facilitate storing the value numbers in
210 bitmaps, and keep the number of wasted SSA_NAME's down, we also
211 associate a value_id with each value number, and create full blown
212 ssa_name's only where we actually need them (IE in operands of
213 existing expressions).
214
215 Theoretically you could replace all the value_id's with
216 SSA_NAME_VERSION, but this would allocate a large number of
217 SSA_NAME's (which are each > 30 bytes) just to get a 4 byte number.
218 It would also require an additional indirection at each point we
219 use the value id. */
220
221 /* Representation of expressions on value numbers:
222
223 Expressions consisting of value numbers are represented the same
224 way as our VN internally represents them, with an additional
225 "pre_expr" wrapping around them in order to facilitate storing all
226 of the expressions in the same sets. */
227
228 /* Representation of sets:
229
230 The dataflow sets do not need to be sorted in any particular order
231 for the majority of their lifetime, are simply represented as two
232 bitmaps, one that keeps track of values present in the set, and one
233 that keeps track of expressions present in the set.
234
235 When we need them in topological order, we produce it on demand by
236 transforming the bitmap into an array and sorting it into topo
237 order. */
238
239 /* Type of expression, used to know which member of the PRE_EXPR union
240 is valid. */
241
242 enum pre_expr_kind
243 {
244 NAME,
245 NARY,
246 REFERENCE,
247 CONSTANT
248 };
249
250 union pre_expr_union
251 {
252 tree name;
253 tree constant;
254 vn_nary_op_t nary;
255 vn_reference_t reference;
256 };
257
258 typedef struct pre_expr_d : nofree_ptr_hash <pre_expr_d>
259 {
260 enum pre_expr_kind kind;
261 unsigned int id;
262 pre_expr_union u;
263
264 /* hash_table support. */
265 static inline hashval_t hash (const pre_expr_d *);
266 static inline int equal (const pre_expr_d *, const pre_expr_d *);
267 } *pre_expr;
268
269 #define PRE_EXPR_NAME(e) (e)->u.name
270 #define PRE_EXPR_NARY(e) (e)->u.nary
271 #define PRE_EXPR_REFERENCE(e) (e)->u.reference
272 #define PRE_EXPR_CONSTANT(e) (e)->u.constant
273
274 /* Compare E1 and E1 for equality. */
275
276 inline int
277 pre_expr_d::equal (const pre_expr_d *e1, const pre_expr_d *e2)
278 {
279 if (e1->kind != e2->kind)
280 return false;
281
282 switch (e1->kind)
283 {
284 case CONSTANT:
285 return vn_constant_eq_with_type (PRE_EXPR_CONSTANT (e1),
286 PRE_EXPR_CONSTANT (e2));
287 case NAME:
288 return PRE_EXPR_NAME (e1) == PRE_EXPR_NAME (e2);
289 case NARY:
290 return vn_nary_op_eq (PRE_EXPR_NARY (e1), PRE_EXPR_NARY (e2));
291 case REFERENCE:
292 return vn_reference_eq (PRE_EXPR_REFERENCE (e1),
293 PRE_EXPR_REFERENCE (e2));
294 default:
295 gcc_unreachable ();
296 }
297 }
298
299 /* Hash E. */
300
301 inline hashval_t
302 pre_expr_d::hash (const pre_expr_d *e)
303 {
304 switch (e->kind)
305 {
306 case CONSTANT:
307 return vn_hash_constant_with_type (PRE_EXPR_CONSTANT (e));
308 case NAME:
309 return SSA_NAME_VERSION (PRE_EXPR_NAME (e));
310 case NARY:
311 return PRE_EXPR_NARY (e)->hashcode;
312 case REFERENCE:
313 return PRE_EXPR_REFERENCE (e)->hashcode;
314 default:
315 gcc_unreachable ();
316 }
317 }
318
319 /* Next global expression id number. */
320 static unsigned int next_expression_id;
321
322 /* Mapping from expression to id number we can use in bitmap sets. */
323 static vec<pre_expr> expressions;
324 static hash_table<pre_expr_d> *expression_to_id;
325 static vec<unsigned> name_to_id;
326
327 /* Allocate an expression id for EXPR. */
328
329 static inline unsigned int
330 alloc_expression_id (pre_expr expr)
331 {
332 struct pre_expr_d **slot;
333 /* Make sure we won't overflow. */
334 gcc_assert (next_expression_id + 1 > next_expression_id);
335 expr->id = next_expression_id++;
336 expressions.safe_push (expr);
337 if (expr->kind == NAME)
338 {
339 unsigned version = SSA_NAME_VERSION (PRE_EXPR_NAME (expr));
340 /* vec::safe_grow_cleared allocates no headroom. Avoid frequent
341 re-allocations by using vec::reserve upfront. */
342 unsigned old_len = name_to_id.length ();
343 name_to_id.reserve (num_ssa_names - old_len);
344 name_to_id.quick_grow_cleared (num_ssa_names);
345 gcc_assert (name_to_id[version] == 0);
346 name_to_id[version] = expr->id;
347 }
348 else
349 {
350 slot = expression_to_id->find_slot (expr, INSERT);
351 gcc_assert (!*slot);
352 *slot = expr;
353 }
354 return next_expression_id - 1;
355 }
356
357 /* Return the expression id for tree EXPR. */
358
359 static inline unsigned int
360 get_expression_id (const pre_expr expr)
361 {
362 return expr->id;
363 }
364
365 static inline unsigned int
366 lookup_expression_id (const pre_expr expr)
367 {
368 struct pre_expr_d **slot;
369
370 if (expr->kind == NAME)
371 {
372 unsigned version = SSA_NAME_VERSION (PRE_EXPR_NAME (expr));
373 if (name_to_id.length () <= version)
374 return 0;
375 return name_to_id[version];
376 }
377 else
378 {
379 slot = expression_to_id->find_slot (expr, NO_INSERT);
380 if (!slot)
381 return 0;
382 return ((pre_expr)*slot)->id;
383 }
384 }
385
386 /* Return the existing expression id for EXPR, or create one if one
387 does not exist yet. */
388
389 static inline unsigned int
390 get_or_alloc_expression_id (pre_expr expr)
391 {
392 unsigned int id = lookup_expression_id (expr);
393 if (id == 0)
394 return alloc_expression_id (expr);
395 return expr->id = id;
396 }
397
398 /* Return the expression that has expression id ID */
399
400 static inline pre_expr
401 expression_for_id (unsigned int id)
402 {
403 return expressions[id];
404 }
405
406 /* Free the expression id field in all of our expressions,
407 and then destroy the expressions array. */
408
409 static void
410 clear_expression_ids (void)
411 {
412 expressions.release ();
413 }
414
415 static object_allocator<pre_expr_d> pre_expr_pool ("pre_expr nodes");
416
417 /* Given an SSA_NAME NAME, get or create a pre_expr to represent it. */
418
419 static pre_expr
420 get_or_alloc_expr_for_name (tree name)
421 {
422 struct pre_expr_d expr;
423 pre_expr result;
424 unsigned int result_id;
425
426 expr.kind = NAME;
427 expr.id = 0;
428 PRE_EXPR_NAME (&expr) = name;
429 result_id = lookup_expression_id (&expr);
430 if (result_id != 0)
431 return expression_for_id (result_id);
432
433 result = pre_expr_pool.allocate ();
434 result->kind = NAME;
435 PRE_EXPR_NAME (result) = name;
436 alloc_expression_id (result);
437 return result;
438 }
439
440 /* An unordered bitmap set. One bitmap tracks values, the other,
441 expressions. */
442 typedef struct bitmap_set
443 {
444 bitmap_head expressions;
445 bitmap_head values;
446 } *bitmap_set_t;
447
448 #define FOR_EACH_EXPR_ID_IN_SET(set, id, bi) \
449 EXECUTE_IF_SET_IN_BITMAP (&(set)->expressions, 0, (id), (bi))
450
451 #define FOR_EACH_VALUE_ID_IN_SET(set, id, bi) \
452 EXECUTE_IF_SET_IN_BITMAP (&(set)->values, 0, (id), (bi))
453
454 /* Mapping from value id to expressions with that value_id. */
455 static vec<bitmap> value_expressions;
456
457 /* Sets that we need to keep track of. */
458 typedef struct bb_bitmap_sets
459 {
460 /* The EXP_GEN set, which represents expressions/values generated in
461 a basic block. */
462 bitmap_set_t exp_gen;
463
464 /* The PHI_GEN set, which represents PHI results generated in a
465 basic block. */
466 bitmap_set_t phi_gen;
467
468 /* The TMP_GEN set, which represents results/temporaries generated
469 in a basic block. IE the LHS of an expression. */
470 bitmap_set_t tmp_gen;
471
472 /* The AVAIL_OUT set, which represents which values are available in
473 a given basic block. */
474 bitmap_set_t avail_out;
475
476 /* The ANTIC_IN set, which represents which values are anticipatable
477 in a given basic block. */
478 bitmap_set_t antic_in;
479
480 /* The PA_IN set, which represents which values are
481 partially anticipatable in a given basic block. */
482 bitmap_set_t pa_in;
483
484 /* The NEW_SETS set, which is used during insertion to augment the
485 AVAIL_OUT set of blocks with the new insertions performed during
486 the current iteration. */
487 bitmap_set_t new_sets;
488
489 /* A cache for value_dies_in_block_x. */
490 bitmap expr_dies;
491
492 /* The live virtual operand on successor edges. */
493 tree vop_on_exit;
494
495 /* True if we have visited this block during ANTIC calculation. */
496 unsigned int visited : 1;
497
498 /* True when the block contains a call that might not return. */
499 unsigned int contains_may_not_return_call : 1;
500 } *bb_value_sets_t;
501
502 #define EXP_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->exp_gen
503 #define PHI_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->phi_gen
504 #define TMP_GEN(BB) ((bb_value_sets_t) ((BB)->aux))->tmp_gen
505 #define AVAIL_OUT(BB) ((bb_value_sets_t) ((BB)->aux))->avail_out
506 #define ANTIC_IN(BB) ((bb_value_sets_t) ((BB)->aux))->antic_in
507 #define PA_IN(BB) ((bb_value_sets_t) ((BB)->aux))->pa_in
508 #define NEW_SETS(BB) ((bb_value_sets_t) ((BB)->aux))->new_sets
509 #define EXPR_DIES(BB) ((bb_value_sets_t) ((BB)->aux))->expr_dies
510 #define BB_VISITED(BB) ((bb_value_sets_t) ((BB)->aux))->visited
511 #define BB_MAY_NOTRETURN(BB) ((bb_value_sets_t) ((BB)->aux))->contains_may_not_return_call
512 #define BB_LIVE_VOP_ON_EXIT(BB) ((bb_value_sets_t) ((BB)->aux))->vop_on_exit
513
514
515 /* This structure is used to keep track of statistics on what
516 optimization PRE was able to perform. */
517 static struct
518 {
519 /* The number of RHS computations eliminated by PRE. */
520 int eliminations;
521
522 /* The number of new expressions/temporaries generated by PRE. */
523 int insertions;
524
525 /* The number of inserts found due to partial anticipation */
526 int pa_insert;
527
528 /* The number of inserts made for code hoisting. */
529 int hoist_insert;
530
531 /* The number of new PHI nodes added by PRE. */
532 int phis;
533 } pre_stats;
534
535 static bool do_partial_partial;
536 static pre_expr bitmap_find_leader (bitmap_set_t, unsigned int);
537 static void bitmap_value_insert_into_set (bitmap_set_t, pre_expr);
538 static void bitmap_value_replace_in_set (bitmap_set_t, pre_expr);
539 static void bitmap_set_copy (bitmap_set_t, bitmap_set_t);
540 static bool bitmap_set_contains_value (bitmap_set_t, unsigned int);
541 static void bitmap_insert_into_set (bitmap_set_t, pre_expr);
542 static bitmap_set_t bitmap_set_new (void);
543 static tree create_expression_by_pieces (basic_block, pre_expr, gimple_seq *,
544 tree);
545 static tree find_or_generate_expression (basic_block, tree, gimple_seq *);
546 static unsigned int get_expr_value_id (pre_expr);
547
548 /* We can add and remove elements and entries to and from sets
549 and hash tables, so we use alloc pools for them. */
550
551 static object_allocator<bitmap_set> bitmap_set_pool ("Bitmap sets");
552 static bitmap_obstack grand_bitmap_obstack;
553
554 /* Set of blocks with statements that have had their EH properties changed. */
555 static bitmap need_eh_cleanup;
556
557 /* Set of blocks with statements that have had their AB properties changed. */
558 static bitmap need_ab_cleanup;
559
560 /* A three tuple {e, pred, v} used to cache phi translations in the
561 phi_translate_table. */
562
563 typedef struct expr_pred_trans_d : free_ptr_hash<expr_pred_trans_d>
564 {
565 /* The expression. */
566 pre_expr e;
567
568 /* The predecessor block along which we translated the expression. */
569 basic_block pred;
570
571 /* The value that resulted from the translation. */
572 pre_expr v;
573
574 /* The hashcode for the expression, pred pair. This is cached for
575 speed reasons. */
576 hashval_t hashcode;
577
578 /* hash_table support. */
579 static inline hashval_t hash (const expr_pred_trans_d *);
580 static inline int equal (const expr_pred_trans_d *, const expr_pred_trans_d *);
581 } *expr_pred_trans_t;
582 typedef const struct expr_pred_trans_d *const_expr_pred_trans_t;
583
584 inline hashval_t
585 expr_pred_trans_d::hash (const expr_pred_trans_d *e)
586 {
587 return e->hashcode;
588 }
589
590 inline int
591 expr_pred_trans_d::equal (const expr_pred_trans_d *ve1,
592 const expr_pred_trans_d *ve2)
593 {
594 basic_block b1 = ve1->pred;
595 basic_block b2 = ve2->pred;
596
597 /* If they are not translations for the same basic block, they can't
598 be equal. */
599 if (b1 != b2)
600 return false;
601 return pre_expr_d::equal (ve1->e, ve2->e);
602 }
603
604 /* The phi_translate_table caches phi translations for a given
605 expression and predecessor. */
606 static hash_table<expr_pred_trans_d> *phi_translate_table;
607
608 /* Add the tuple mapping from {expression E, basic block PRED} to
609 the phi translation table and return whether it pre-existed. */
610
611 static inline bool
612 phi_trans_add (expr_pred_trans_t *entry, pre_expr e, basic_block pred)
613 {
614 expr_pred_trans_t *slot;
615 expr_pred_trans_d tem;
616 hashval_t hash = iterative_hash_hashval_t (pre_expr_d::hash (e),
617 pred->index);
618 tem.e = e;
619 tem.pred = pred;
620 tem.hashcode = hash;
621 slot = phi_translate_table->find_slot_with_hash (&tem, hash, INSERT);
622 if (*slot)
623 {
624 *entry = *slot;
625 return true;
626 }
627
628 *entry = *slot = XNEW (struct expr_pred_trans_d);
629 (*entry)->e = e;
630 (*entry)->pred = pred;
631 (*entry)->hashcode = hash;
632 return false;
633 }
634
635
636 /* Add expression E to the expression set of value id V. */
637
638 static void
639 add_to_value (unsigned int v, pre_expr e)
640 {
641 bitmap set;
642
643 gcc_checking_assert (get_expr_value_id (e) == v);
644
645 if (v >= value_expressions.length ())
646 {
647 value_expressions.safe_grow_cleared (v + 1);
648 }
649
650 set = value_expressions[v];
651 if (!set)
652 {
653 set = BITMAP_ALLOC (&grand_bitmap_obstack);
654 value_expressions[v] = set;
655 }
656
657 bitmap_set_bit (set, get_or_alloc_expression_id (e));
658 }
659
660 /* Create a new bitmap set and return it. */
661
662 static bitmap_set_t
663 bitmap_set_new (void)
664 {
665 bitmap_set_t ret = bitmap_set_pool.allocate ();
666 bitmap_initialize (&ret->expressions, &grand_bitmap_obstack);
667 bitmap_initialize (&ret->values, &grand_bitmap_obstack);
668 return ret;
669 }
670
671 /* Return the value id for a PRE expression EXPR. */
672
673 static unsigned int
674 get_expr_value_id (pre_expr expr)
675 {
676 unsigned int id;
677 switch (expr->kind)
678 {
679 case CONSTANT:
680 id = get_constant_value_id (PRE_EXPR_CONSTANT (expr));
681 break;
682 case NAME:
683 id = VN_INFO (PRE_EXPR_NAME (expr))->value_id;
684 break;
685 case NARY:
686 id = PRE_EXPR_NARY (expr)->value_id;
687 break;
688 case REFERENCE:
689 id = PRE_EXPR_REFERENCE (expr)->value_id;
690 break;
691 default:
692 gcc_unreachable ();
693 }
694 /* ??? We cannot assert that expr has a value-id (it can be 0), because
695 we assign value-ids only to expressions that have a result
696 in set_hashtable_value_ids. */
697 return id;
698 }
699
700 /* Return a SCCVN valnum (SSA name or constant) for the PRE value-id VAL. */
701
702 static tree
703 sccvn_valnum_from_value_id (unsigned int val)
704 {
705 bitmap_iterator bi;
706 unsigned int i;
707 bitmap exprset = value_expressions[val];
708 EXECUTE_IF_SET_IN_BITMAP (exprset, 0, i, bi)
709 {
710 pre_expr vexpr = expression_for_id (i);
711 if (vexpr->kind == NAME)
712 return VN_INFO (PRE_EXPR_NAME (vexpr))->valnum;
713 else if (vexpr->kind == CONSTANT)
714 return PRE_EXPR_CONSTANT (vexpr);
715 }
716 return NULL_TREE;
717 }
718
719 /* Remove an expression EXPR from a bitmapped set. */
720
721 static void
722 bitmap_remove_from_set (bitmap_set_t set, pre_expr expr)
723 {
724 unsigned int val = get_expr_value_id (expr);
725 if (!value_id_constant_p (val))
726 {
727 bitmap_clear_bit (&set->values, val);
728 bitmap_clear_bit (&set->expressions, get_expression_id (expr));
729 }
730 }
731
732 /* Insert an expression EXPR into a bitmapped set. */
733
734 static void
735 bitmap_insert_into_set (bitmap_set_t set, pre_expr expr)
736 {
737 unsigned int val = get_expr_value_id (expr);
738 if (! value_id_constant_p (val))
739 {
740 /* Note this is the only function causing multiple expressions
741 for the same value to appear in a set. This is needed for
742 TMP_GEN, PHI_GEN and NEW_SETs. */
743 bitmap_set_bit (&set->values, val);
744 bitmap_set_bit (&set->expressions, get_or_alloc_expression_id (expr));
745 }
746 }
747
748 /* Copy a bitmapped set ORIG, into bitmapped set DEST. */
749
750 static void
751 bitmap_set_copy (bitmap_set_t dest, bitmap_set_t orig)
752 {
753 bitmap_copy (&dest->expressions, &orig->expressions);
754 bitmap_copy (&dest->values, &orig->values);
755 }
756
757
758 /* Free memory used up by SET. */
759 static void
760 bitmap_set_free (bitmap_set_t set)
761 {
762 bitmap_clear (&set->expressions);
763 bitmap_clear (&set->values);
764 }
765
766
767 /* Generate an topological-ordered array of bitmap set SET. */
768
769 static vec<pre_expr>
770 sorted_array_from_bitmap_set (bitmap_set_t set)
771 {
772 unsigned int i, j;
773 bitmap_iterator bi, bj;
774 vec<pre_expr> result;
775
776 /* Pre-allocate enough space for the array. */
777 result.create (bitmap_count_bits (&set->expressions));
778
779 FOR_EACH_VALUE_ID_IN_SET (set, i, bi)
780 {
781 /* The number of expressions having a given value is usually
782 relatively small. Thus, rather than making a vector of all
783 the expressions and sorting it by value-id, we walk the values
784 and check in the reverse mapping that tells us what expressions
785 have a given value, to filter those in our set. As a result,
786 the expressions are inserted in value-id order, which means
787 topological order.
788
789 If this is somehow a significant lose for some cases, we can
790 choose which set to walk based on the set size. */
791 bitmap exprset = value_expressions[i];
792 EXECUTE_IF_SET_IN_BITMAP (exprset, 0, j, bj)
793 {
794 if (bitmap_bit_p (&set->expressions, j))
795 result.quick_push (expression_for_id (j));
796 }
797 }
798
799 return result;
800 }
801
802 /* Subtract all expressions contained in ORIG from DEST. */
803
804 static bitmap_set_t
805 bitmap_set_subtract (bitmap_set_t dest, bitmap_set_t orig)
806 {
807 bitmap_set_t result = bitmap_set_new ();
808 bitmap_iterator bi;
809 unsigned int i;
810
811 bitmap_and_compl (&result->expressions, &dest->expressions,
812 &orig->expressions);
813
814 FOR_EACH_EXPR_ID_IN_SET (result, i, bi)
815 {
816 pre_expr expr = expression_for_id (i);
817 unsigned int value_id = get_expr_value_id (expr);
818 bitmap_set_bit (&result->values, value_id);
819 }
820
821 return result;
822 }
823
824 /* Subtract all values in bitmap set B from bitmap set A. */
825
826 static void
827 bitmap_set_subtract_values (bitmap_set_t a, bitmap_set_t b)
828 {
829 unsigned int i;
830 bitmap_iterator bi;
831 pre_expr to_remove = NULL;
832 FOR_EACH_EXPR_ID_IN_SET (a, i, bi)
833 {
834 if (to_remove)
835 {
836 bitmap_remove_from_set (a, to_remove);
837 to_remove = NULL;
838 }
839 pre_expr expr = expression_for_id (i);
840 if (bitmap_set_contains_value (b, get_expr_value_id (expr)))
841 to_remove = expr;
842 }
843 if (to_remove)
844 bitmap_remove_from_set (a, to_remove);
845 }
846
847
848 /* Return true if bitmapped set SET contains the value VALUE_ID. */
849
850 static bool
851 bitmap_set_contains_value (bitmap_set_t set, unsigned int value_id)
852 {
853 if (value_id_constant_p (value_id))
854 return true;
855
856 if (!set || bitmap_empty_p (&set->expressions))
857 return false;
858
859 return bitmap_bit_p (&set->values, value_id);
860 }
861
862 static inline bool
863 bitmap_set_contains_expr (bitmap_set_t set, const pre_expr expr)
864 {
865 return bitmap_bit_p (&set->expressions, get_expression_id (expr));
866 }
867
868 /* Replace an instance of value LOOKFOR with expression EXPR in SET. */
869
870 static void
871 bitmap_set_replace_value (bitmap_set_t set, unsigned int lookfor,
872 const pre_expr expr)
873 {
874 bitmap exprset;
875 unsigned int i;
876 bitmap_iterator bi;
877
878 if (value_id_constant_p (lookfor))
879 return;
880
881 if (!bitmap_set_contains_value (set, lookfor))
882 return;
883
884 /* The number of expressions having a given value is usually
885 significantly less than the total number of expressions in SET.
886 Thus, rather than check, for each expression in SET, whether it
887 has the value LOOKFOR, we walk the reverse mapping that tells us
888 what expressions have a given value, and see if any of those
889 expressions are in our set. For large testcases, this is about
890 5-10x faster than walking the bitmap. If this is somehow a
891 significant lose for some cases, we can choose which set to walk
892 based on the set size. */
893 exprset = value_expressions[lookfor];
894 EXECUTE_IF_SET_IN_BITMAP (exprset, 0, i, bi)
895 {
896 if (bitmap_clear_bit (&set->expressions, i))
897 {
898 bitmap_set_bit (&set->expressions, get_expression_id (expr));
899 return;
900 }
901 }
902
903 gcc_unreachable ();
904 }
905
906 /* Return true if two bitmap sets are equal. */
907
908 static bool
909 bitmap_set_equal (bitmap_set_t a, bitmap_set_t b)
910 {
911 return bitmap_equal_p (&a->values, &b->values);
912 }
913
914 /* Replace an instance of EXPR's VALUE with EXPR in SET if it exists,
915 and add it otherwise. */
916
917 static void
918 bitmap_value_replace_in_set (bitmap_set_t set, pre_expr expr)
919 {
920 unsigned int val = get_expr_value_id (expr);
921
922 if (bitmap_set_contains_value (set, val))
923 bitmap_set_replace_value (set, val, expr);
924 else
925 bitmap_insert_into_set (set, expr);
926 }
927
928 /* Insert EXPR into SET if EXPR's value is not already present in
929 SET. */
930
931 static void
932 bitmap_value_insert_into_set (bitmap_set_t set, pre_expr expr)
933 {
934 unsigned int val = get_expr_value_id (expr);
935
936 gcc_checking_assert (expr->id == get_or_alloc_expression_id (expr));
937
938 /* Constant values are always considered to be part of the set. */
939 if (value_id_constant_p (val))
940 return;
941
942 /* If the value membership changed, add the expression. */
943 if (bitmap_set_bit (&set->values, val))
944 bitmap_set_bit (&set->expressions, expr->id);
945 }
946
947 /* Print out EXPR to outfile. */
948
949 static void
950 print_pre_expr (FILE *outfile, const pre_expr expr)
951 {
952 if (! expr)
953 {
954 fprintf (outfile, "NULL");
955 return;
956 }
957 switch (expr->kind)
958 {
959 case CONSTANT:
960 print_generic_expr (outfile, PRE_EXPR_CONSTANT (expr));
961 break;
962 case NAME:
963 print_generic_expr (outfile, PRE_EXPR_NAME (expr));
964 break;
965 case NARY:
966 {
967 unsigned int i;
968 vn_nary_op_t nary = PRE_EXPR_NARY (expr);
969 fprintf (outfile, "{%s,", get_tree_code_name (nary->opcode));
970 for (i = 0; i < nary->length; i++)
971 {
972 print_generic_expr (outfile, nary->op[i]);
973 if (i != (unsigned) nary->length - 1)
974 fprintf (outfile, ",");
975 }
976 fprintf (outfile, "}");
977 }
978 break;
979
980 case REFERENCE:
981 {
982 vn_reference_op_t vro;
983 unsigned int i;
984 vn_reference_t ref = PRE_EXPR_REFERENCE (expr);
985 fprintf (outfile, "{");
986 for (i = 0;
987 ref->operands.iterate (i, &vro);
988 i++)
989 {
990 bool closebrace = false;
991 if (vro->opcode != SSA_NAME
992 && TREE_CODE_CLASS (vro->opcode) != tcc_declaration)
993 {
994 fprintf (outfile, "%s", get_tree_code_name (vro->opcode));
995 if (vro->op0)
996 {
997 fprintf (outfile, "<");
998 closebrace = true;
999 }
1000 }
1001 if (vro->op0)
1002 {
1003 print_generic_expr (outfile, vro->op0);
1004 if (vro->op1)
1005 {
1006 fprintf (outfile, ",");
1007 print_generic_expr (outfile, vro->op1);
1008 }
1009 if (vro->op2)
1010 {
1011 fprintf (outfile, ",");
1012 print_generic_expr (outfile, vro->op2);
1013 }
1014 }
1015 if (closebrace)
1016 fprintf (outfile, ">");
1017 if (i != ref->operands.length () - 1)
1018 fprintf (outfile, ",");
1019 }
1020 fprintf (outfile, "}");
1021 if (ref->vuse)
1022 {
1023 fprintf (outfile, "@");
1024 print_generic_expr (outfile, ref->vuse);
1025 }
1026 }
1027 break;
1028 }
1029 }
1030 void debug_pre_expr (pre_expr);
1031
1032 /* Like print_pre_expr but always prints to stderr. */
1033 DEBUG_FUNCTION void
1034 debug_pre_expr (pre_expr e)
1035 {
1036 print_pre_expr (stderr, e);
1037 fprintf (stderr, "\n");
1038 }
1039
1040 /* Print out SET to OUTFILE. */
1041
1042 static void
1043 print_bitmap_set (FILE *outfile, bitmap_set_t set,
1044 const char *setname, int blockindex)
1045 {
1046 fprintf (outfile, "%s[%d] := { ", setname, blockindex);
1047 if (set)
1048 {
1049 bool first = true;
1050 unsigned i;
1051 bitmap_iterator bi;
1052
1053 FOR_EACH_EXPR_ID_IN_SET (set, i, bi)
1054 {
1055 const pre_expr expr = expression_for_id (i);
1056
1057 if (!first)
1058 fprintf (outfile, ", ");
1059 first = false;
1060 print_pre_expr (outfile, expr);
1061
1062 fprintf (outfile, " (%04d)", get_expr_value_id (expr));
1063 }
1064 }
1065 fprintf (outfile, " }\n");
1066 }
1067
1068 void debug_bitmap_set (bitmap_set_t);
1069
1070 DEBUG_FUNCTION void
1071 debug_bitmap_set (bitmap_set_t set)
1072 {
1073 print_bitmap_set (stderr, set, "debug", 0);
1074 }
1075
1076 void debug_bitmap_sets_for (basic_block);
1077
1078 DEBUG_FUNCTION void
1079 debug_bitmap_sets_for (basic_block bb)
1080 {
1081 print_bitmap_set (stderr, AVAIL_OUT (bb), "avail_out", bb->index);
1082 print_bitmap_set (stderr, EXP_GEN (bb), "exp_gen", bb->index);
1083 print_bitmap_set (stderr, PHI_GEN (bb), "phi_gen", bb->index);
1084 print_bitmap_set (stderr, TMP_GEN (bb), "tmp_gen", bb->index);
1085 print_bitmap_set (stderr, ANTIC_IN (bb), "antic_in", bb->index);
1086 if (do_partial_partial)
1087 print_bitmap_set (stderr, PA_IN (bb), "pa_in", bb->index);
1088 print_bitmap_set (stderr, NEW_SETS (bb), "new_sets", bb->index);
1089 }
1090
1091 /* Print out the expressions that have VAL to OUTFILE. */
1092
1093 static void
1094 print_value_expressions (FILE *outfile, unsigned int val)
1095 {
1096 bitmap set = value_expressions[val];
1097 if (set)
1098 {
1099 bitmap_set x;
1100 char s[10];
1101 sprintf (s, "%04d", val);
1102 x.expressions = *set;
1103 print_bitmap_set (outfile, &x, s, 0);
1104 }
1105 }
1106
1107
1108 DEBUG_FUNCTION void
1109 debug_value_expressions (unsigned int val)
1110 {
1111 print_value_expressions (stderr, val);
1112 }
1113
1114 /* Given a CONSTANT, allocate a new CONSTANT type PRE_EXPR to
1115 represent it. */
1116
1117 static pre_expr
1118 get_or_alloc_expr_for_constant (tree constant)
1119 {
1120 unsigned int result_id;
1121 unsigned int value_id;
1122 struct pre_expr_d expr;
1123 pre_expr newexpr;
1124
1125 expr.kind = CONSTANT;
1126 PRE_EXPR_CONSTANT (&expr) = constant;
1127 result_id = lookup_expression_id (&expr);
1128 if (result_id != 0)
1129 return expression_for_id (result_id);
1130
1131 newexpr = pre_expr_pool.allocate ();
1132 newexpr->kind = CONSTANT;
1133 PRE_EXPR_CONSTANT (newexpr) = constant;
1134 alloc_expression_id (newexpr);
1135 value_id = get_or_alloc_constant_value_id (constant);
1136 add_to_value (value_id, newexpr);
1137 return newexpr;
1138 }
1139
1140 /* Get or allocate a pre_expr for a piece of GIMPLE, and return it.
1141 Currently only supports constants and SSA_NAMES. */
1142 static pre_expr
1143 get_or_alloc_expr_for (tree t)
1144 {
1145 if (TREE_CODE (t) == SSA_NAME)
1146 return get_or_alloc_expr_for_name (t);
1147 else if (is_gimple_min_invariant (t))
1148 return get_or_alloc_expr_for_constant (t);
1149 gcc_unreachable ();
1150 }
1151
1152 /* Return the folded version of T if T, when folded, is a gimple
1153 min_invariant or an SSA name. Otherwise, return T. */
1154
1155 static pre_expr
1156 fully_constant_expression (pre_expr e)
1157 {
1158 switch (e->kind)
1159 {
1160 case CONSTANT:
1161 return e;
1162 case NARY:
1163 {
1164 vn_nary_op_t nary = PRE_EXPR_NARY (e);
1165 tree res = vn_nary_simplify (nary);
1166 if (!res)
1167 return e;
1168 if (is_gimple_min_invariant (res))
1169 return get_or_alloc_expr_for_constant (res);
1170 if (TREE_CODE (res) == SSA_NAME)
1171 return get_or_alloc_expr_for_name (res);
1172 return e;
1173 }
1174 case REFERENCE:
1175 {
1176 vn_reference_t ref = PRE_EXPR_REFERENCE (e);
1177 tree folded;
1178 if ((folded = fully_constant_vn_reference_p (ref)))
1179 return get_or_alloc_expr_for_constant (folded);
1180 return e;
1181 }
1182 default:
1183 return e;
1184 }
1185 return e;
1186 }
1187
1188 /* Translate the VUSE backwards through phi nodes in PHIBLOCK, so that
1189 it has the value it would have in BLOCK. Set *SAME_VALID to true
1190 in case the new vuse doesn't change the value id of the OPERANDS. */
1191
1192 static tree
1193 translate_vuse_through_block (vec<vn_reference_op_s> operands,
1194 alias_set_type set, tree type, tree vuse,
1195 basic_block phiblock,
1196 basic_block block, bool *same_valid)
1197 {
1198 gimple *phi = SSA_NAME_DEF_STMT (vuse);
1199 ao_ref ref;
1200 edge e = NULL;
1201 bool use_oracle;
1202
1203 *same_valid = true;
1204
1205 if (gimple_bb (phi) != phiblock)
1206 return vuse;
1207
1208 use_oracle = ao_ref_init_from_vn_reference (&ref, set, type, operands);
1209
1210 /* Use the alias-oracle to find either the PHI node in this block,
1211 the first VUSE used in this block that is equivalent to vuse or
1212 the first VUSE which definition in this block kills the value. */
1213 if (gimple_code (phi) == GIMPLE_PHI)
1214 e = find_edge (block, phiblock);
1215 else if (use_oracle)
1216 while (!stmt_may_clobber_ref_p_1 (phi, &ref))
1217 {
1218 vuse = gimple_vuse (phi);
1219 phi = SSA_NAME_DEF_STMT (vuse);
1220 if (gimple_bb (phi) != phiblock)
1221 return vuse;
1222 if (gimple_code (phi) == GIMPLE_PHI)
1223 {
1224 e = find_edge (block, phiblock);
1225 break;
1226 }
1227 }
1228 else
1229 return NULL_TREE;
1230
1231 if (e)
1232 {
1233 if (use_oracle)
1234 {
1235 bitmap visited = NULL;
1236 unsigned int cnt;
1237 /* Try to find a vuse that dominates this phi node by skipping
1238 non-clobbering statements. */
1239 vuse = get_continuation_for_phi (phi, &ref, &cnt, &visited, false,
1240 NULL, NULL);
1241 if (visited)
1242 BITMAP_FREE (visited);
1243 }
1244 else
1245 vuse = NULL_TREE;
1246 if (!vuse)
1247 {
1248 /* If we didn't find any, the value ID can't stay the same,
1249 but return the translated vuse. */
1250 *same_valid = false;
1251 vuse = PHI_ARG_DEF (phi, e->dest_idx);
1252 }
1253 /* ??? We would like to return vuse here as this is the canonical
1254 upmost vdef that this reference is associated with. But during
1255 insertion of the references into the hash tables we only ever
1256 directly insert with their direct gimple_vuse, hence returning
1257 something else would make us not find the other expression. */
1258 return PHI_ARG_DEF (phi, e->dest_idx);
1259 }
1260
1261 return NULL_TREE;
1262 }
1263
1264 /* Like bitmap_find_leader, but checks for the value existing in SET1 *or*
1265 SET2 *or* SET3. This is used to avoid making a set consisting of the union
1266 of PA_IN and ANTIC_IN during insert and phi-translation. */
1267
1268 static inline pre_expr
1269 find_leader_in_sets (unsigned int val, bitmap_set_t set1, bitmap_set_t set2,
1270 bitmap_set_t set3 = NULL)
1271 {
1272 pre_expr result;
1273
1274 result = bitmap_find_leader (set1, val);
1275 if (!result && set2)
1276 result = bitmap_find_leader (set2, val);
1277 if (!result && set3)
1278 result = bitmap_find_leader (set3, val);
1279 return result;
1280 }
1281
1282 /* Get the tree type for our PRE expression e. */
1283
1284 static tree
1285 get_expr_type (const pre_expr e)
1286 {
1287 switch (e->kind)
1288 {
1289 case NAME:
1290 return TREE_TYPE (PRE_EXPR_NAME (e));
1291 case CONSTANT:
1292 return TREE_TYPE (PRE_EXPR_CONSTANT (e));
1293 case REFERENCE:
1294 return PRE_EXPR_REFERENCE (e)->type;
1295 case NARY:
1296 return PRE_EXPR_NARY (e)->type;
1297 }
1298 gcc_unreachable ();
1299 }
1300
1301 /* Get a representative SSA_NAME for a given expression.
1302 Since all of our sub-expressions are treated as values, we require
1303 them to be SSA_NAME's for simplicity.
1304 Prior versions of GVNPRE used to use "value handles" here, so that
1305 an expression would be VH.11 + VH.10 instead of d_3 + e_6. In
1306 either case, the operands are really values (IE we do not expect
1307 them to be usable without finding leaders). */
1308
1309 static tree
1310 get_representative_for (const pre_expr e)
1311 {
1312 tree name;
1313 unsigned int value_id = get_expr_value_id (e);
1314
1315 switch (e->kind)
1316 {
1317 case NAME:
1318 return VN_INFO (PRE_EXPR_NAME (e))->valnum;
1319 case CONSTANT:
1320 return PRE_EXPR_CONSTANT (e);
1321 case NARY:
1322 case REFERENCE:
1323 {
1324 /* Go through all of the expressions representing this value
1325 and pick out an SSA_NAME. */
1326 unsigned int i;
1327 bitmap_iterator bi;
1328 bitmap exprs = value_expressions[value_id];
1329 EXECUTE_IF_SET_IN_BITMAP (exprs, 0, i, bi)
1330 {
1331 pre_expr rep = expression_for_id (i);
1332 if (rep->kind == NAME)
1333 return VN_INFO (PRE_EXPR_NAME (rep))->valnum;
1334 else if (rep->kind == CONSTANT)
1335 return PRE_EXPR_CONSTANT (rep);
1336 }
1337 }
1338 break;
1339 }
1340
1341 /* If we reached here we couldn't find an SSA_NAME. This can
1342 happen when we've discovered a value that has never appeared in
1343 the program as set to an SSA_NAME, as the result of phi translation.
1344 Create one here.
1345 ??? We should be able to re-use this when we insert the statement
1346 to compute it. */
1347 name = make_temp_ssa_name (get_expr_type (e), gimple_build_nop (), "pretmp");
1348 VN_INFO_GET (name)->value_id = value_id;
1349 VN_INFO (name)->valnum = name;
1350 /* ??? For now mark this SSA name for release by SCCVN. */
1351 VN_INFO (name)->needs_insertion = true;
1352 add_to_value (value_id, get_or_alloc_expr_for_name (name));
1353 if (dump_file && (dump_flags & TDF_DETAILS))
1354 {
1355 fprintf (dump_file, "Created SSA_NAME representative ");
1356 print_generic_expr (dump_file, name);
1357 fprintf (dump_file, " for expression:");
1358 print_pre_expr (dump_file, e);
1359 fprintf (dump_file, " (%04d)\n", value_id);
1360 }
1361
1362 return name;
1363 }
1364
1365
1366
1367 static pre_expr
1368 phi_translate (pre_expr expr, bitmap_set_t set1, bitmap_set_t set2,
1369 basic_block pred, basic_block phiblock);
1370
1371 /* Translate EXPR using phis in PHIBLOCK, so that it has the values of
1372 the phis in PRED. Return NULL if we can't find a leader for each part
1373 of the translated expression. */
1374
1375 static pre_expr
1376 phi_translate_1 (pre_expr expr, bitmap_set_t set1, bitmap_set_t set2,
1377 basic_block pred, basic_block phiblock)
1378 {
1379 switch (expr->kind)
1380 {
1381 case NARY:
1382 {
1383 unsigned int i;
1384 bool changed = false;
1385 vn_nary_op_t nary = PRE_EXPR_NARY (expr);
1386 vn_nary_op_t newnary = XALLOCAVAR (struct vn_nary_op_s,
1387 sizeof_vn_nary_op (nary->length));
1388 memcpy (newnary, nary, sizeof_vn_nary_op (nary->length));
1389
1390 for (i = 0; i < newnary->length; i++)
1391 {
1392 if (TREE_CODE (newnary->op[i]) != SSA_NAME)
1393 continue;
1394 else
1395 {
1396 pre_expr leader, result;
1397 unsigned int op_val_id = VN_INFO (newnary->op[i])->value_id;
1398 leader = find_leader_in_sets (op_val_id, set1, set2);
1399 result = phi_translate (leader, set1, set2, pred, phiblock);
1400 if (result && result != leader)
1401 newnary->op[i] = get_representative_for (result);
1402 else if (!result)
1403 return NULL;
1404
1405 changed |= newnary->op[i] != nary->op[i];
1406 }
1407 }
1408 if (changed)
1409 {
1410 pre_expr constant;
1411 unsigned int new_val_id;
1412
1413 PRE_EXPR_NARY (expr) = newnary;
1414 constant = fully_constant_expression (expr);
1415 PRE_EXPR_NARY (expr) = nary;
1416 if (constant != expr)
1417 {
1418 /* For non-CONSTANTs we have to make sure we can eventually
1419 insert the expression. Which means we need to have a
1420 leader for it. */
1421 if (constant->kind != CONSTANT)
1422 {
1423 /* Do not allow simplifications to non-constants over
1424 backedges as this will likely result in a loop PHI node
1425 to be inserted and increased register pressure.
1426 See PR77498 - this avoids doing predcoms work in
1427 a less efficient way. */
1428 if (find_edge (pred, phiblock)->flags & EDGE_DFS_BACK)
1429 ;
1430 else
1431 {
1432 unsigned value_id = get_expr_value_id (constant);
1433 constant = find_leader_in_sets (value_id, set1, set2,
1434 AVAIL_OUT (pred));
1435 if (constant)
1436 return constant;
1437 }
1438 }
1439 else
1440 return constant;
1441 }
1442
1443 tree result = vn_nary_op_lookup_pieces (newnary->length,
1444 newnary->opcode,
1445 newnary->type,
1446 &newnary->op[0],
1447 &nary);
1448 if (result && is_gimple_min_invariant (result))
1449 return get_or_alloc_expr_for_constant (result);
1450
1451 expr = pre_expr_pool.allocate ();
1452 expr->kind = NARY;
1453 expr->id = 0;
1454 if (nary)
1455 {
1456 PRE_EXPR_NARY (expr) = nary;
1457 new_val_id = nary->value_id;
1458 get_or_alloc_expression_id (expr);
1459 /* When we end up re-using a value number make sure that
1460 doesn't have unrelated (which we can't check here)
1461 range or points-to info on it. */
1462 if (result
1463 && INTEGRAL_TYPE_P (TREE_TYPE (result))
1464 && SSA_NAME_RANGE_INFO (result)
1465 && ! SSA_NAME_IS_DEFAULT_DEF (result))
1466 {
1467 if (! VN_INFO (result)->info.range_info)
1468 {
1469 VN_INFO (result)->info.range_info
1470 = SSA_NAME_RANGE_INFO (result);
1471 VN_INFO (result)->range_info_anti_range_p
1472 = SSA_NAME_ANTI_RANGE_P (result);
1473 }
1474 if (dump_file && (dump_flags & TDF_DETAILS))
1475 {
1476 fprintf (dump_file, "clearing range info of ");
1477 print_generic_expr (dump_file, result);
1478 fprintf (dump_file, "\n");
1479 }
1480 SSA_NAME_RANGE_INFO (result) = NULL;
1481 }
1482 else if (result
1483 && POINTER_TYPE_P (TREE_TYPE (result))
1484 && SSA_NAME_PTR_INFO (result)
1485 && ! SSA_NAME_IS_DEFAULT_DEF (result))
1486 {
1487 if (! VN_INFO (result)->info.ptr_info)
1488 VN_INFO (result)->info.ptr_info
1489 = SSA_NAME_PTR_INFO (result);
1490 if (dump_file && (dump_flags & TDF_DETAILS))
1491 {
1492 fprintf (dump_file, "clearing points-to info of ");
1493 print_generic_expr (dump_file, result);
1494 fprintf (dump_file, "\n");
1495 }
1496 SSA_NAME_PTR_INFO (result) = NULL;
1497 }
1498 }
1499 else
1500 {
1501 new_val_id = get_next_value_id ();
1502 value_expressions.safe_grow_cleared (get_max_value_id () + 1);
1503 nary = vn_nary_op_insert_pieces (newnary->length,
1504 newnary->opcode,
1505 newnary->type,
1506 &newnary->op[0],
1507 result, new_val_id);
1508 PRE_EXPR_NARY (expr) = nary;
1509 get_or_alloc_expression_id (expr);
1510 }
1511 add_to_value (new_val_id, expr);
1512 }
1513 return expr;
1514 }
1515 break;
1516
1517 case REFERENCE:
1518 {
1519 vn_reference_t ref = PRE_EXPR_REFERENCE (expr);
1520 vec<vn_reference_op_s> operands = ref->operands;
1521 tree vuse = ref->vuse;
1522 tree newvuse = vuse;
1523 vec<vn_reference_op_s> newoperands = vNULL;
1524 bool changed = false, same_valid = true;
1525 unsigned int i, n;
1526 vn_reference_op_t operand;
1527 vn_reference_t newref;
1528
1529 for (i = 0; operands.iterate (i, &operand); i++)
1530 {
1531 pre_expr opresult;
1532 pre_expr leader;
1533 tree op[3];
1534 tree type = operand->type;
1535 vn_reference_op_s newop = *operand;
1536 op[0] = operand->op0;
1537 op[1] = operand->op1;
1538 op[2] = operand->op2;
1539 for (n = 0; n < 3; ++n)
1540 {
1541 unsigned int op_val_id;
1542 if (!op[n])
1543 continue;
1544 if (TREE_CODE (op[n]) != SSA_NAME)
1545 {
1546 /* We can't possibly insert these. */
1547 if (n != 0
1548 && !is_gimple_min_invariant (op[n]))
1549 break;
1550 continue;
1551 }
1552 op_val_id = VN_INFO (op[n])->value_id;
1553 leader = find_leader_in_sets (op_val_id, set1, set2);
1554 opresult = phi_translate (leader, set1, set2, pred, phiblock);
1555 if (opresult && opresult != leader)
1556 {
1557 tree name = get_representative_for (opresult);
1558 changed |= name != op[n];
1559 op[n] = name;
1560 }
1561 else if (!opresult)
1562 break;
1563 }
1564 if (n != 3)
1565 {
1566 newoperands.release ();
1567 return NULL;
1568 }
1569 if (!changed)
1570 continue;
1571 if (!newoperands.exists ())
1572 newoperands = operands.copy ();
1573 /* We may have changed from an SSA_NAME to a constant */
1574 if (newop.opcode == SSA_NAME && TREE_CODE (op[0]) != SSA_NAME)
1575 newop.opcode = TREE_CODE (op[0]);
1576 newop.type = type;
1577 newop.op0 = op[0];
1578 newop.op1 = op[1];
1579 newop.op2 = op[2];
1580 newoperands[i] = newop;
1581 }
1582 gcc_checking_assert (i == operands.length ());
1583
1584 if (vuse)
1585 {
1586 newvuse = translate_vuse_through_block (newoperands.exists ()
1587 ? newoperands : operands,
1588 ref->set, ref->type,
1589 vuse, phiblock, pred,
1590 &same_valid);
1591 if (newvuse == NULL_TREE)
1592 {
1593 newoperands.release ();
1594 return NULL;
1595 }
1596 }
1597
1598 if (changed || newvuse != vuse)
1599 {
1600 unsigned int new_val_id;
1601 pre_expr constant;
1602
1603 tree result = vn_reference_lookup_pieces (newvuse, ref->set,
1604 ref->type,
1605 newoperands.exists ()
1606 ? newoperands : operands,
1607 &newref, VN_WALK);
1608 if (result)
1609 newoperands.release ();
1610
1611 /* We can always insert constants, so if we have a partial
1612 redundant constant load of another type try to translate it
1613 to a constant of appropriate type. */
1614 if (result && is_gimple_min_invariant (result))
1615 {
1616 tree tem = result;
1617 if (!useless_type_conversion_p (ref->type, TREE_TYPE (result)))
1618 {
1619 tem = fold_unary (VIEW_CONVERT_EXPR, ref->type, result);
1620 if (tem && !is_gimple_min_invariant (tem))
1621 tem = NULL_TREE;
1622 }
1623 if (tem)
1624 return get_or_alloc_expr_for_constant (tem);
1625 }
1626
1627 /* If we'd have to convert things we would need to validate
1628 if we can insert the translated expression. So fail
1629 here for now - we cannot insert an alias with a different
1630 type in the VN tables either, as that would assert. */
1631 if (result
1632 && !useless_type_conversion_p (ref->type, TREE_TYPE (result)))
1633 return NULL;
1634 else if (!result && newref
1635 && !useless_type_conversion_p (ref->type, newref->type))
1636 {
1637 newoperands.release ();
1638 return NULL;
1639 }
1640
1641 expr = pre_expr_pool.allocate ();
1642 expr->kind = REFERENCE;
1643 expr->id = 0;
1644
1645 if (newref)
1646 {
1647 PRE_EXPR_REFERENCE (expr) = newref;
1648 constant = fully_constant_expression (expr);
1649 if (constant != expr)
1650 return constant;
1651
1652 new_val_id = newref->value_id;
1653 get_or_alloc_expression_id (expr);
1654 }
1655 else
1656 {
1657 if (changed || !same_valid)
1658 {
1659 new_val_id = get_next_value_id ();
1660 value_expressions.safe_grow_cleared
1661 (get_max_value_id () + 1);
1662 }
1663 else
1664 new_val_id = ref->value_id;
1665 if (!newoperands.exists ())
1666 newoperands = operands.copy ();
1667 newref = vn_reference_insert_pieces (newvuse, ref->set,
1668 ref->type,
1669 newoperands,
1670 result, new_val_id);
1671 newoperands = vNULL;
1672 PRE_EXPR_REFERENCE (expr) = newref;
1673 constant = fully_constant_expression (expr);
1674 if (constant != expr)
1675 return constant;
1676 get_or_alloc_expression_id (expr);
1677 }
1678 add_to_value (new_val_id, expr);
1679 }
1680 newoperands.release ();
1681 return expr;
1682 }
1683 break;
1684
1685 case NAME:
1686 {
1687 tree name = PRE_EXPR_NAME (expr);
1688 gimple *def_stmt = SSA_NAME_DEF_STMT (name);
1689 /* If the SSA name is defined by a PHI node in this block,
1690 translate it. */
1691 if (gimple_code (def_stmt) == GIMPLE_PHI
1692 && gimple_bb (def_stmt) == phiblock)
1693 {
1694 edge e = find_edge (pred, gimple_bb (def_stmt));
1695 tree def = PHI_ARG_DEF (def_stmt, e->dest_idx);
1696
1697 /* Handle constant. */
1698 if (is_gimple_min_invariant (def))
1699 return get_or_alloc_expr_for_constant (def);
1700
1701 return get_or_alloc_expr_for_name (def);
1702 }
1703 /* Otherwise return it unchanged - it will get removed if its
1704 value is not available in PREDs AVAIL_OUT set of expressions
1705 by the subtraction of TMP_GEN. */
1706 return expr;
1707 }
1708
1709 default:
1710 gcc_unreachable ();
1711 }
1712 }
1713
1714 /* Wrapper around phi_translate_1 providing caching functionality. */
1715
1716 static pre_expr
1717 phi_translate (pre_expr expr, bitmap_set_t set1, bitmap_set_t set2,
1718 basic_block pred, basic_block phiblock)
1719 {
1720 expr_pred_trans_t slot = NULL;
1721 pre_expr phitrans;
1722
1723 if (!expr)
1724 return NULL;
1725
1726 /* Constants contain no values that need translation. */
1727 if (expr->kind == CONSTANT)
1728 return expr;
1729
1730 if (value_id_constant_p (get_expr_value_id (expr)))
1731 return expr;
1732
1733 /* Don't add translations of NAMEs as those are cheap to translate. */
1734 if (expr->kind != NAME)
1735 {
1736 if (phi_trans_add (&slot, expr, pred))
1737 return slot->v;
1738 /* Store NULL for the value we want to return in the case of
1739 recursing. */
1740 slot->v = NULL;
1741 }
1742
1743 /* Translate. */
1744 phitrans = phi_translate_1 (expr, set1, set2, pred, phiblock);
1745
1746 if (slot)
1747 {
1748 if (phitrans)
1749 slot->v = phitrans;
1750 else
1751 /* Remove failed translations again, they cause insert
1752 iteration to not pick up new opportunities reliably. */
1753 phi_translate_table->remove_elt_with_hash (slot, slot->hashcode);
1754 }
1755
1756 return phitrans;
1757 }
1758
1759
1760 /* For each expression in SET, translate the values through phi nodes
1761 in PHIBLOCK using edge PHIBLOCK->PRED, and store the resulting
1762 expressions in DEST. */
1763
1764 static void
1765 phi_translate_set (bitmap_set_t dest, bitmap_set_t set, basic_block pred,
1766 basic_block phiblock)
1767 {
1768 vec<pre_expr> exprs;
1769 pre_expr expr;
1770 int i;
1771
1772 if (gimple_seq_empty_p (phi_nodes (phiblock)))
1773 {
1774 bitmap_set_copy (dest, set);
1775 return;
1776 }
1777
1778 exprs = sorted_array_from_bitmap_set (set);
1779 FOR_EACH_VEC_ELT (exprs, i, expr)
1780 {
1781 pre_expr translated;
1782 translated = phi_translate (expr, set, NULL, pred, phiblock);
1783 if (!translated)
1784 continue;
1785
1786 /* We might end up with multiple expressions from SET being
1787 translated to the same value. In this case we do not want
1788 to retain the NARY or REFERENCE expression but prefer a NAME
1789 which would be the leader. */
1790 if (translated->kind == NAME)
1791 bitmap_value_replace_in_set (dest, translated);
1792 else
1793 bitmap_value_insert_into_set (dest, translated);
1794 }
1795 exprs.release ();
1796 }
1797
1798 /* Find the leader for a value (i.e., the name representing that
1799 value) in a given set, and return it. Return NULL if no leader
1800 is found. */
1801
1802 static pre_expr
1803 bitmap_find_leader (bitmap_set_t set, unsigned int val)
1804 {
1805 if (value_id_constant_p (val))
1806 {
1807 unsigned int i;
1808 bitmap_iterator bi;
1809 bitmap exprset = value_expressions[val];
1810
1811 EXECUTE_IF_SET_IN_BITMAP (exprset, 0, i, bi)
1812 {
1813 pre_expr expr = expression_for_id (i);
1814 if (expr->kind == CONSTANT)
1815 return expr;
1816 }
1817 }
1818 if (bitmap_set_contains_value (set, val))
1819 {
1820 /* Rather than walk the entire bitmap of expressions, and see
1821 whether any of them has the value we are looking for, we look
1822 at the reverse mapping, which tells us the set of expressions
1823 that have a given value (IE value->expressions with that
1824 value) and see if any of those expressions are in our set.
1825 The number of expressions per value is usually significantly
1826 less than the number of expressions in the set. In fact, for
1827 large testcases, doing it this way is roughly 5-10x faster
1828 than walking the bitmap.
1829 If this is somehow a significant lose for some cases, we can
1830 choose which set to walk based on which set is smaller. */
1831 unsigned int i;
1832 bitmap_iterator bi;
1833 bitmap exprset = value_expressions[val];
1834
1835 EXECUTE_IF_AND_IN_BITMAP (exprset, &set->expressions, 0, i, bi)
1836 return expression_for_id (i);
1837 }
1838 return NULL;
1839 }
1840
1841 /* Determine if EXPR, a memory expression, is ANTIC_IN at the top of
1842 BLOCK by seeing if it is not killed in the block. Note that we are
1843 only determining whether there is a store that kills it. Because
1844 of the order in which clean iterates over values, we are guaranteed
1845 that altered operands will have caused us to be eliminated from the
1846 ANTIC_IN set already. */
1847
1848 static bool
1849 value_dies_in_block_x (pre_expr expr, basic_block block)
1850 {
1851 tree vuse = PRE_EXPR_REFERENCE (expr)->vuse;
1852 vn_reference_t refx = PRE_EXPR_REFERENCE (expr);
1853 gimple *def;
1854 gimple_stmt_iterator gsi;
1855 unsigned id = get_expression_id (expr);
1856 bool res = false;
1857 ao_ref ref;
1858
1859 if (!vuse)
1860 return false;
1861
1862 /* Lookup a previously calculated result. */
1863 if (EXPR_DIES (block)
1864 && bitmap_bit_p (EXPR_DIES (block), id * 2))
1865 return bitmap_bit_p (EXPR_DIES (block), id * 2 + 1);
1866
1867 /* A memory expression {e, VUSE} dies in the block if there is a
1868 statement that may clobber e. If, starting statement walk from the
1869 top of the basic block, a statement uses VUSE there can be no kill
1870 inbetween that use and the original statement that loaded {e, VUSE},
1871 so we can stop walking. */
1872 ref.base = NULL_TREE;
1873 for (gsi = gsi_start_bb (block); !gsi_end_p (gsi); gsi_next (&gsi))
1874 {
1875 tree def_vuse, def_vdef;
1876 def = gsi_stmt (gsi);
1877 def_vuse = gimple_vuse (def);
1878 def_vdef = gimple_vdef (def);
1879
1880 /* Not a memory statement. */
1881 if (!def_vuse)
1882 continue;
1883
1884 /* Not a may-def. */
1885 if (!def_vdef)
1886 {
1887 /* A load with the same VUSE, we're done. */
1888 if (def_vuse == vuse)
1889 break;
1890
1891 continue;
1892 }
1893
1894 /* Init ref only if we really need it. */
1895 if (ref.base == NULL_TREE
1896 && !ao_ref_init_from_vn_reference (&ref, refx->set, refx->type,
1897 refx->operands))
1898 {
1899 res = true;
1900 break;
1901 }
1902 /* If the statement may clobber expr, it dies. */
1903 if (stmt_may_clobber_ref_p_1 (def, &ref))
1904 {
1905 res = true;
1906 break;
1907 }
1908 }
1909
1910 /* Remember the result. */
1911 if (!EXPR_DIES (block))
1912 EXPR_DIES (block) = BITMAP_ALLOC (&grand_bitmap_obstack);
1913 bitmap_set_bit (EXPR_DIES (block), id * 2);
1914 if (res)
1915 bitmap_set_bit (EXPR_DIES (block), id * 2 + 1);
1916
1917 return res;
1918 }
1919
1920
1921 /* Determine if OP is valid in SET1 U SET2, which it is when the union
1922 contains its value-id. */
1923
1924 static bool
1925 op_valid_in_sets (bitmap_set_t set1, bitmap_set_t set2, tree op)
1926 {
1927 if (op && TREE_CODE (op) == SSA_NAME)
1928 {
1929 unsigned int value_id = VN_INFO (op)->value_id;
1930 if (!(bitmap_set_contains_value (set1, value_id)
1931 || (set2 && bitmap_set_contains_value (set2, value_id))))
1932 return false;
1933 }
1934 return true;
1935 }
1936
1937 /* Determine if the expression EXPR is valid in SET1 U SET2.
1938 ONLY SET2 CAN BE NULL.
1939 This means that we have a leader for each part of the expression
1940 (if it consists of values), or the expression is an SSA_NAME.
1941 For loads/calls, we also see if the vuse is killed in this block. */
1942
1943 static bool
1944 valid_in_sets (bitmap_set_t set1, bitmap_set_t set2, pre_expr expr)
1945 {
1946 switch (expr->kind)
1947 {
1948 case NAME:
1949 /* By construction all NAMEs are available. Non-available
1950 NAMEs are removed by subtracting TMP_GEN from the sets. */
1951 return true;
1952 case NARY:
1953 {
1954 unsigned int i;
1955 vn_nary_op_t nary = PRE_EXPR_NARY (expr);
1956 for (i = 0; i < nary->length; i++)
1957 if (!op_valid_in_sets (set1, set2, nary->op[i]))
1958 return false;
1959 return true;
1960 }
1961 break;
1962 case REFERENCE:
1963 {
1964 vn_reference_t ref = PRE_EXPR_REFERENCE (expr);
1965 vn_reference_op_t vro;
1966 unsigned int i;
1967
1968 FOR_EACH_VEC_ELT (ref->operands, i, vro)
1969 {
1970 if (!op_valid_in_sets (set1, set2, vro->op0)
1971 || !op_valid_in_sets (set1, set2, vro->op1)
1972 || !op_valid_in_sets (set1, set2, vro->op2))
1973 return false;
1974 }
1975 return true;
1976 }
1977 default:
1978 gcc_unreachable ();
1979 }
1980 }
1981
1982 /* Clean the set of expressions that are no longer valid in SET1 or
1983 SET2. This means expressions that are made up of values we have no
1984 leaders for in SET1 or SET2. This version is used for partial
1985 anticipation, which means it is not valid in either ANTIC_IN or
1986 PA_IN. */
1987
1988 static void
1989 dependent_clean (bitmap_set_t set1, bitmap_set_t set2)
1990 {
1991 vec<pre_expr> exprs = sorted_array_from_bitmap_set (set1);
1992 pre_expr expr;
1993 int i;
1994
1995 FOR_EACH_VEC_ELT (exprs, i, expr)
1996 {
1997 if (!valid_in_sets (set1, set2, expr))
1998 bitmap_remove_from_set (set1, expr);
1999 }
2000 exprs.release ();
2001 }
2002
2003 /* Clean the set of expressions that are no longer valid in SET. This
2004 means expressions that are made up of values we have no leaders for
2005 in SET. */
2006
2007 static void
2008 clean (bitmap_set_t set)
2009 {
2010 vec<pre_expr> exprs = sorted_array_from_bitmap_set (set);
2011 pre_expr expr;
2012 int i;
2013
2014 FOR_EACH_VEC_ELT (exprs, i, expr)
2015 {
2016 if (!valid_in_sets (set, NULL, expr))
2017 bitmap_remove_from_set (set, expr);
2018 }
2019 exprs.release ();
2020 }
2021
2022 /* Clean the set of expressions that are no longer valid in SET because
2023 they are clobbered in BLOCK or because they trap and may not be executed. */
2024
2025 static void
2026 prune_clobbered_mems (bitmap_set_t set, basic_block block)
2027 {
2028 bitmap_iterator bi;
2029 unsigned i;
2030 pre_expr to_remove = NULL;
2031
2032 FOR_EACH_EXPR_ID_IN_SET (set, i, bi)
2033 {
2034 /* Remove queued expr. */
2035 if (to_remove)
2036 {
2037 bitmap_remove_from_set (set, to_remove);
2038 to_remove = NULL;
2039 }
2040
2041 pre_expr expr = expression_for_id (i);
2042 if (expr->kind == REFERENCE)
2043 {
2044 vn_reference_t ref = PRE_EXPR_REFERENCE (expr);
2045 if (ref->vuse)
2046 {
2047 gimple *def_stmt = SSA_NAME_DEF_STMT (ref->vuse);
2048 if (!gimple_nop_p (def_stmt)
2049 && ((gimple_bb (def_stmt) != block
2050 && !dominated_by_p (CDI_DOMINATORS,
2051 block, gimple_bb (def_stmt)))
2052 || (gimple_bb (def_stmt) == block
2053 && value_dies_in_block_x (expr, block))))
2054 to_remove = expr;
2055 }
2056 }
2057 else if (expr->kind == NARY)
2058 {
2059 vn_nary_op_t nary = PRE_EXPR_NARY (expr);
2060 /* If the NARY may trap make sure the block does not contain
2061 a possible exit point.
2062 ??? This is overly conservative if we translate AVAIL_OUT
2063 as the available expression might be after the exit point. */
2064 if (BB_MAY_NOTRETURN (block)
2065 && vn_nary_may_trap (nary))
2066 to_remove = expr;
2067 }
2068 }
2069
2070 /* Remove queued expr. */
2071 if (to_remove)
2072 bitmap_remove_from_set (set, to_remove);
2073 }
2074
2075 static sbitmap has_abnormal_preds;
2076
2077 /* Compute the ANTIC set for BLOCK.
2078
2079 If succs(BLOCK) > 1 then
2080 ANTIC_OUT[BLOCK] = intersection of ANTIC_IN[b] for all succ(BLOCK)
2081 else if succs(BLOCK) == 1 then
2082 ANTIC_OUT[BLOCK] = phi_translate (ANTIC_IN[succ(BLOCK)])
2083
2084 ANTIC_IN[BLOCK] = clean(ANTIC_OUT[BLOCK] U EXP_GEN[BLOCK] - TMP_GEN[BLOCK])
2085
2086 Note that clean() is deferred until after the iteration. */
2087
2088 static bool
2089 compute_antic_aux (basic_block block, bool block_has_abnormal_pred_edge)
2090 {
2091 bitmap_set_t S, old, ANTIC_OUT;
2092 bitmap_iterator bi;
2093 unsigned int bii;
2094 edge e;
2095 edge_iterator ei;
2096
2097 bool changed = ! BB_VISITED (block);
2098 BB_VISITED (block) = 1;
2099 old = ANTIC_OUT = S = NULL;
2100
2101 /* If any edges from predecessors are abnormal, antic_in is empty,
2102 so do nothing. */
2103 if (block_has_abnormal_pred_edge)
2104 goto maybe_dump_sets;
2105
2106 old = ANTIC_IN (block);
2107 ANTIC_OUT = bitmap_set_new ();
2108
2109 /* If the block has no successors, ANTIC_OUT is empty. */
2110 if (EDGE_COUNT (block->succs) == 0)
2111 ;
2112 /* If we have one successor, we could have some phi nodes to
2113 translate through. */
2114 else if (single_succ_p (block))
2115 {
2116 basic_block succ_bb = single_succ (block);
2117 gcc_assert (BB_VISITED (succ_bb));
2118 phi_translate_set (ANTIC_OUT, ANTIC_IN (succ_bb), block, succ_bb);
2119 }
2120 /* If we have multiple successors, we take the intersection of all of
2121 them. Note that in the case of loop exit phi nodes, we may have
2122 phis to translate through. */
2123 else
2124 {
2125 size_t i;
2126 basic_block bprime, first = NULL;
2127
2128 auto_vec<basic_block> worklist (EDGE_COUNT (block->succs));
2129 FOR_EACH_EDGE (e, ei, block->succs)
2130 {
2131 if (!first
2132 && BB_VISITED (e->dest))
2133 first = e->dest;
2134 else if (BB_VISITED (e->dest))
2135 worklist.quick_push (e->dest);
2136 else
2137 {
2138 /* Unvisited successors get their ANTIC_IN replaced by the
2139 maximal set to arrive at a maximum ANTIC_IN solution.
2140 We can ignore them in the intersection operation and thus
2141 need not explicitely represent that maximum solution. */
2142 if (dump_file && (dump_flags & TDF_DETAILS))
2143 fprintf (dump_file, "ANTIC_IN is MAX on %d->%d\n",
2144 e->src->index, e->dest->index);
2145 }
2146 }
2147
2148 /* Of multiple successors we have to have visited one already
2149 which is guaranteed by iteration order. */
2150 gcc_assert (first != NULL);
2151
2152 phi_translate_set (ANTIC_OUT, ANTIC_IN (first), block, first);
2153
2154 /* If we have multiple successors we need to intersect the ANTIC_OUT
2155 sets. For values that's a simple intersection but for
2156 expressions it is a union. Given we want to have a single
2157 expression per value in our sets we have to canonicalize.
2158 Avoid randomness and running into cycles like for PR82129 and
2159 canonicalize the expression we choose to the one with the
2160 lowest id. This requires we actually compute the union first. */
2161 FOR_EACH_VEC_ELT (worklist, i, bprime)
2162 {
2163 if (!gimple_seq_empty_p (phi_nodes (bprime)))
2164 {
2165 bitmap_set_t tmp = bitmap_set_new ();
2166 phi_translate_set (tmp, ANTIC_IN (bprime), block, bprime);
2167 bitmap_and_into (&ANTIC_OUT->values, &tmp->values);
2168 bitmap_ior_into (&ANTIC_OUT->expressions, &tmp->expressions);
2169 bitmap_set_free (tmp);
2170 }
2171 else
2172 {
2173 bitmap_and_into (&ANTIC_OUT->values, &ANTIC_IN (bprime)->values);
2174 bitmap_ior_into (&ANTIC_OUT->expressions,
2175 &ANTIC_IN (bprime)->expressions);
2176 }
2177 }
2178 if (! worklist.is_empty ())
2179 {
2180 /* Prune expressions not in the value set, canonicalizing to
2181 expression with lowest ID. */
2182 bitmap_iterator bi;
2183 unsigned int i;
2184 unsigned int to_clear = -1U;
2185 bitmap seen_value = BITMAP_ALLOC (NULL);
2186 FOR_EACH_EXPR_ID_IN_SET (ANTIC_OUT, i, bi)
2187 {
2188 if (to_clear != -1U)
2189 {
2190 bitmap_clear_bit (&ANTIC_OUT->expressions, to_clear);
2191 to_clear = -1U;
2192 }
2193 pre_expr expr = expression_for_id (i);
2194 unsigned int value_id = get_expr_value_id (expr);
2195 if (!bitmap_bit_p (&ANTIC_OUT->values, value_id)
2196 || !bitmap_set_bit (seen_value, value_id))
2197 to_clear = i;
2198 }
2199 if (to_clear != -1U)
2200 bitmap_clear_bit (&ANTIC_OUT->expressions, to_clear);
2201 BITMAP_FREE (seen_value);
2202 }
2203 }
2204
2205 /* Prune expressions that are clobbered in block and thus become
2206 invalid if translated from ANTIC_OUT to ANTIC_IN. */
2207 prune_clobbered_mems (ANTIC_OUT, block);
2208
2209 /* Generate ANTIC_OUT - TMP_GEN. */
2210 S = bitmap_set_subtract (ANTIC_OUT, TMP_GEN (block));
2211
2212 /* Start ANTIC_IN with EXP_GEN - TMP_GEN. */
2213 ANTIC_IN (block) = bitmap_set_subtract (EXP_GEN (block),
2214 TMP_GEN (block));
2215
2216 /* Then union in the ANTIC_OUT - TMP_GEN values,
2217 to get ANTIC_OUT U EXP_GEN - TMP_GEN */
2218 FOR_EACH_EXPR_ID_IN_SET (S, bii, bi)
2219 bitmap_value_insert_into_set (ANTIC_IN (block),
2220 expression_for_id (bii));
2221
2222 /* clean (ANTIC_IN (block)) is defered to after the iteration converged
2223 because it can cause non-convergence, see for example PR81181. */
2224
2225 if (!bitmap_set_equal (old, ANTIC_IN (block)))
2226 changed = true;
2227
2228 maybe_dump_sets:
2229 if (dump_file && (dump_flags & TDF_DETAILS))
2230 {
2231 if (ANTIC_OUT)
2232 print_bitmap_set (dump_file, ANTIC_OUT, "ANTIC_OUT", block->index);
2233
2234 if (changed)
2235 fprintf (dump_file, "[changed] ");
2236 print_bitmap_set (dump_file, ANTIC_IN (block), "ANTIC_IN",
2237 block->index);
2238
2239 if (S)
2240 print_bitmap_set (dump_file, S, "S", block->index);
2241 }
2242 if (old)
2243 bitmap_set_free (old);
2244 if (S)
2245 bitmap_set_free (S);
2246 if (ANTIC_OUT)
2247 bitmap_set_free (ANTIC_OUT);
2248 return changed;
2249 }
2250
2251 /* Compute PARTIAL_ANTIC for BLOCK.
2252
2253 If succs(BLOCK) > 1 then
2254 PA_OUT[BLOCK] = value wise union of PA_IN[b] + all ANTIC_IN not
2255 in ANTIC_OUT for all succ(BLOCK)
2256 else if succs(BLOCK) == 1 then
2257 PA_OUT[BLOCK] = phi_translate (PA_IN[succ(BLOCK)])
2258
2259 PA_IN[BLOCK] = dependent_clean(PA_OUT[BLOCK] - TMP_GEN[BLOCK]
2260 - ANTIC_IN[BLOCK])
2261
2262 */
2263 static void
2264 compute_partial_antic_aux (basic_block block,
2265 bool block_has_abnormal_pred_edge)
2266 {
2267 bitmap_set_t old_PA_IN;
2268 bitmap_set_t PA_OUT;
2269 edge e;
2270 edge_iterator ei;
2271 unsigned long max_pa = PARAM_VALUE (PARAM_MAX_PARTIAL_ANTIC_LENGTH);
2272
2273 old_PA_IN = PA_OUT = NULL;
2274
2275 /* If any edges from predecessors are abnormal, antic_in is empty,
2276 so do nothing. */
2277 if (block_has_abnormal_pred_edge)
2278 goto maybe_dump_sets;
2279
2280 /* If there are too many partially anticipatable values in the
2281 block, phi_translate_set can take an exponential time: stop
2282 before the translation starts. */
2283 if (max_pa
2284 && single_succ_p (block)
2285 && bitmap_count_bits (&PA_IN (single_succ (block))->values) > max_pa)
2286 goto maybe_dump_sets;
2287
2288 old_PA_IN = PA_IN (block);
2289 PA_OUT = bitmap_set_new ();
2290
2291 /* If the block has no successors, ANTIC_OUT is empty. */
2292 if (EDGE_COUNT (block->succs) == 0)
2293 ;
2294 /* If we have one successor, we could have some phi nodes to
2295 translate through. Note that we can't phi translate across DFS
2296 back edges in partial antic, because it uses a union operation on
2297 the successors. For recurrences like IV's, we will end up
2298 generating a new value in the set on each go around (i + 3 (VH.1)
2299 VH.1 + 1 (VH.2), VH.2 + 1 (VH.3), etc), forever. */
2300 else if (single_succ_p (block))
2301 {
2302 basic_block succ = single_succ (block);
2303 if (!(single_succ_edge (block)->flags & EDGE_DFS_BACK))
2304 phi_translate_set (PA_OUT, PA_IN (succ), block, succ);
2305 }
2306 /* If we have multiple successors, we take the union of all of
2307 them. */
2308 else
2309 {
2310 size_t i;
2311 basic_block bprime;
2312
2313 auto_vec<basic_block> worklist (EDGE_COUNT (block->succs));
2314 FOR_EACH_EDGE (e, ei, block->succs)
2315 {
2316 if (e->flags & EDGE_DFS_BACK)
2317 continue;
2318 worklist.quick_push (e->dest);
2319 }
2320 if (worklist.length () > 0)
2321 {
2322 FOR_EACH_VEC_ELT (worklist, i, bprime)
2323 {
2324 unsigned int i;
2325 bitmap_iterator bi;
2326
2327 FOR_EACH_EXPR_ID_IN_SET (ANTIC_IN (bprime), i, bi)
2328 bitmap_value_insert_into_set (PA_OUT,
2329 expression_for_id (i));
2330 if (!gimple_seq_empty_p (phi_nodes (bprime)))
2331 {
2332 bitmap_set_t pa_in = bitmap_set_new ();
2333 phi_translate_set (pa_in, PA_IN (bprime), block, bprime);
2334 FOR_EACH_EXPR_ID_IN_SET (pa_in, i, bi)
2335 bitmap_value_insert_into_set (PA_OUT,
2336 expression_for_id (i));
2337 bitmap_set_free (pa_in);
2338 }
2339 else
2340 FOR_EACH_EXPR_ID_IN_SET (PA_IN (bprime), i, bi)
2341 bitmap_value_insert_into_set (PA_OUT,
2342 expression_for_id (i));
2343 }
2344 }
2345 }
2346
2347 /* Prune expressions that are clobbered in block and thus become
2348 invalid if translated from PA_OUT to PA_IN. */
2349 prune_clobbered_mems (PA_OUT, block);
2350
2351 /* PA_IN starts with PA_OUT - TMP_GEN.
2352 Then we subtract things from ANTIC_IN. */
2353 PA_IN (block) = bitmap_set_subtract (PA_OUT, TMP_GEN (block));
2354
2355 /* For partial antic, we want to put back in the phi results, since
2356 we will properly avoid making them partially antic over backedges. */
2357 bitmap_ior_into (&PA_IN (block)->values, &PHI_GEN (block)->values);
2358 bitmap_ior_into (&PA_IN (block)->expressions, &PHI_GEN (block)->expressions);
2359
2360 /* PA_IN[block] = PA_IN[block] - ANTIC_IN[block] */
2361 bitmap_set_subtract_values (PA_IN (block), ANTIC_IN (block));
2362
2363 dependent_clean (PA_IN (block), ANTIC_IN (block));
2364
2365 maybe_dump_sets:
2366 if (dump_file && (dump_flags & TDF_DETAILS))
2367 {
2368 if (PA_OUT)
2369 print_bitmap_set (dump_file, PA_OUT, "PA_OUT", block->index);
2370
2371 print_bitmap_set (dump_file, PA_IN (block), "PA_IN", block->index);
2372 }
2373 if (old_PA_IN)
2374 bitmap_set_free (old_PA_IN);
2375 if (PA_OUT)
2376 bitmap_set_free (PA_OUT);
2377 }
2378
2379 /* Compute ANTIC and partial ANTIC sets. */
2380
2381 static void
2382 compute_antic (void)
2383 {
2384 bool changed = true;
2385 int num_iterations = 0;
2386 basic_block block;
2387 int i;
2388 edge_iterator ei;
2389 edge e;
2390
2391 /* If any predecessor edges are abnormal, we punt, so antic_in is empty.
2392 We pre-build the map of blocks with incoming abnormal edges here. */
2393 has_abnormal_preds = sbitmap_alloc (last_basic_block_for_fn (cfun));
2394 bitmap_clear (has_abnormal_preds);
2395
2396 FOR_ALL_BB_FN (block, cfun)
2397 {
2398 BB_VISITED (block) = 0;
2399
2400 FOR_EACH_EDGE (e, ei, block->preds)
2401 if (e->flags & EDGE_ABNORMAL)
2402 {
2403 bitmap_set_bit (has_abnormal_preds, block->index);
2404 break;
2405 }
2406
2407 /* While we are here, give empty ANTIC_IN sets to each block. */
2408 ANTIC_IN (block) = bitmap_set_new ();
2409 if (do_partial_partial)
2410 PA_IN (block) = bitmap_set_new ();
2411 }
2412
2413 /* At the exit block we anticipate nothing. */
2414 BB_VISITED (EXIT_BLOCK_PTR_FOR_FN (cfun)) = 1;
2415
2416 /* For ANTIC computation we need a postorder that also guarantees that
2417 a block with a single successor is visited after its successor.
2418 RPO on the inverted CFG has this property. */
2419 auto_vec<int, 20> postorder;
2420 inverted_post_order_compute (&postorder);
2421
2422 auto_sbitmap worklist (last_basic_block_for_fn (cfun) + 1);
2423 bitmap_clear (worklist);
2424 FOR_EACH_EDGE (e, ei, EXIT_BLOCK_PTR_FOR_FN (cfun)->preds)
2425 bitmap_set_bit (worklist, e->src->index);
2426 while (changed)
2427 {
2428 if (dump_file && (dump_flags & TDF_DETAILS))
2429 fprintf (dump_file, "Starting iteration %d\n", num_iterations);
2430 /* ??? We need to clear our PHI translation cache here as the
2431 ANTIC sets shrink and we restrict valid translations to
2432 those having operands with leaders in ANTIC. Same below
2433 for PA ANTIC computation. */
2434 num_iterations++;
2435 changed = false;
2436 for (i = postorder.length () - 1; i >= 0; i--)
2437 {
2438 if (bitmap_bit_p (worklist, postorder[i]))
2439 {
2440 basic_block block = BASIC_BLOCK_FOR_FN (cfun, postorder[i]);
2441 bitmap_clear_bit (worklist, block->index);
2442 if (compute_antic_aux (block,
2443 bitmap_bit_p (has_abnormal_preds,
2444 block->index)))
2445 {
2446 FOR_EACH_EDGE (e, ei, block->preds)
2447 bitmap_set_bit (worklist, e->src->index);
2448 changed = true;
2449 }
2450 }
2451 }
2452 /* Theoretically possible, but *highly* unlikely. */
2453 gcc_checking_assert (num_iterations < 500);
2454 }
2455
2456 /* We have to clean after the dataflow problem converged as cleaning
2457 can cause non-convergence because it is based on expressions
2458 rather than values. */
2459 FOR_EACH_BB_FN (block, cfun)
2460 clean (ANTIC_IN (block));
2461
2462 statistics_histogram_event (cfun, "compute_antic iterations",
2463 num_iterations);
2464
2465 if (do_partial_partial)
2466 {
2467 /* For partial antic we ignore backedges and thus we do not need
2468 to perform any iteration when we process blocks in postorder. */
2469 int postorder_num
2470 = pre_and_rev_post_order_compute (NULL, postorder.address (), false);
2471 for (i = postorder_num - 1 ; i >= 0; i--)
2472 {
2473 basic_block block = BASIC_BLOCK_FOR_FN (cfun, postorder[i]);
2474 compute_partial_antic_aux (block,
2475 bitmap_bit_p (has_abnormal_preds,
2476 block->index));
2477 }
2478 }
2479
2480 sbitmap_free (has_abnormal_preds);
2481 }
2482
2483
2484 /* Inserted expressions are placed onto this worklist, which is used
2485 for performing quick dead code elimination of insertions we made
2486 that didn't turn out to be necessary. */
2487 static bitmap inserted_exprs;
2488
2489 /* The actual worker for create_component_ref_by_pieces. */
2490
2491 static tree
2492 create_component_ref_by_pieces_1 (basic_block block, vn_reference_t ref,
2493 unsigned int *operand, gimple_seq *stmts)
2494 {
2495 vn_reference_op_t currop = &ref->operands[*operand];
2496 tree genop;
2497 ++*operand;
2498 switch (currop->opcode)
2499 {
2500 case CALL_EXPR:
2501 gcc_unreachable ();
2502
2503 case MEM_REF:
2504 {
2505 tree baseop = create_component_ref_by_pieces_1 (block, ref, operand,
2506 stmts);
2507 if (!baseop)
2508 return NULL_TREE;
2509 tree offset = currop->op0;
2510 if (TREE_CODE (baseop) == ADDR_EXPR
2511 && handled_component_p (TREE_OPERAND (baseop, 0)))
2512 {
2513 HOST_WIDE_INT off;
2514 tree base;
2515 base = get_addr_base_and_unit_offset (TREE_OPERAND (baseop, 0),
2516 &off);
2517 gcc_assert (base);
2518 offset = int_const_binop (PLUS_EXPR, offset,
2519 build_int_cst (TREE_TYPE (offset),
2520 off));
2521 baseop = build_fold_addr_expr (base);
2522 }
2523 genop = build2 (MEM_REF, currop->type, baseop, offset);
2524 MR_DEPENDENCE_CLIQUE (genop) = currop->clique;
2525 MR_DEPENDENCE_BASE (genop) = currop->base;
2526 REF_REVERSE_STORAGE_ORDER (genop) = currop->reverse;
2527 return genop;
2528 }
2529
2530 case TARGET_MEM_REF:
2531 {
2532 tree genop0 = NULL_TREE, genop1 = NULL_TREE;
2533 vn_reference_op_t nextop = &ref->operands[++*operand];
2534 tree baseop = create_component_ref_by_pieces_1 (block, ref, operand,
2535 stmts);
2536 if (!baseop)
2537 return NULL_TREE;
2538 if (currop->op0)
2539 {
2540 genop0 = find_or_generate_expression (block, currop->op0, stmts);
2541 if (!genop0)
2542 return NULL_TREE;
2543 }
2544 if (nextop->op0)
2545 {
2546 genop1 = find_or_generate_expression (block, nextop->op0, stmts);
2547 if (!genop1)
2548 return NULL_TREE;
2549 }
2550 genop = build5 (TARGET_MEM_REF, currop->type,
2551 baseop, currop->op2, genop0, currop->op1, genop1);
2552
2553 MR_DEPENDENCE_CLIQUE (genop) = currop->clique;
2554 MR_DEPENDENCE_BASE (genop) = currop->base;
2555 return genop;
2556 }
2557
2558 case ADDR_EXPR:
2559 if (currop->op0)
2560 {
2561 gcc_assert (is_gimple_min_invariant (currop->op0));
2562 return currop->op0;
2563 }
2564 /* Fallthrough. */
2565 case REALPART_EXPR:
2566 case IMAGPART_EXPR:
2567 case VIEW_CONVERT_EXPR:
2568 {
2569 tree genop0 = create_component_ref_by_pieces_1 (block, ref, operand,
2570 stmts);
2571 if (!genop0)
2572 return NULL_TREE;
2573 return fold_build1 (currop->opcode, currop->type, genop0);
2574 }
2575
2576 case WITH_SIZE_EXPR:
2577 {
2578 tree genop0 = create_component_ref_by_pieces_1 (block, ref, operand,
2579 stmts);
2580 if (!genop0)
2581 return NULL_TREE;
2582 tree genop1 = find_or_generate_expression (block, currop->op0, stmts);
2583 if (!genop1)
2584 return NULL_TREE;
2585 return fold_build2 (currop->opcode, currop->type, genop0, genop1);
2586 }
2587
2588 case BIT_FIELD_REF:
2589 {
2590 tree genop0 = create_component_ref_by_pieces_1 (block, ref, operand,
2591 stmts);
2592 if (!genop0)
2593 return NULL_TREE;
2594 tree op1 = currop->op0;
2595 tree op2 = currop->op1;
2596 tree t = build3 (BIT_FIELD_REF, currop->type, genop0, op1, op2);
2597 REF_REVERSE_STORAGE_ORDER (t) = currop->reverse;
2598 return fold (t);
2599 }
2600
2601 /* For array ref vn_reference_op's, operand 1 of the array ref
2602 is op0 of the reference op and operand 3 of the array ref is
2603 op1. */
2604 case ARRAY_RANGE_REF:
2605 case ARRAY_REF:
2606 {
2607 tree genop0;
2608 tree genop1 = currop->op0;
2609 tree genop2 = currop->op1;
2610 tree genop3 = currop->op2;
2611 genop0 = create_component_ref_by_pieces_1 (block, ref, operand,
2612 stmts);
2613 if (!genop0)
2614 return NULL_TREE;
2615 genop1 = find_or_generate_expression (block, genop1, stmts);
2616 if (!genop1)
2617 return NULL_TREE;
2618 if (genop2)
2619 {
2620 tree domain_type = TYPE_DOMAIN (TREE_TYPE (genop0));
2621 /* Drop zero minimum index if redundant. */
2622 if (integer_zerop (genop2)
2623 && (!domain_type
2624 || integer_zerop (TYPE_MIN_VALUE (domain_type))))
2625 genop2 = NULL_TREE;
2626 else
2627 {
2628 genop2 = find_or_generate_expression (block, genop2, stmts);
2629 if (!genop2)
2630 return NULL_TREE;
2631 }
2632 }
2633 if (genop3)
2634 {
2635 tree elmt_type = TREE_TYPE (TREE_TYPE (genop0));
2636 /* We can't always put a size in units of the element alignment
2637 here as the element alignment may be not visible. See
2638 PR43783. Simply drop the element size for constant
2639 sizes. */
2640 if (TREE_CODE (genop3) == INTEGER_CST
2641 && TREE_CODE (TYPE_SIZE_UNIT (elmt_type)) == INTEGER_CST
2642 && wi::eq_p (wi::to_offset (TYPE_SIZE_UNIT (elmt_type)),
2643 (wi::to_offset (genop3)
2644 * vn_ref_op_align_unit (currop))))
2645 genop3 = NULL_TREE;
2646 else
2647 {
2648 genop3 = find_or_generate_expression (block, genop3, stmts);
2649 if (!genop3)
2650 return NULL_TREE;
2651 }
2652 }
2653 return build4 (currop->opcode, currop->type, genop0, genop1,
2654 genop2, genop3);
2655 }
2656 case COMPONENT_REF:
2657 {
2658 tree op0;
2659 tree op1;
2660 tree genop2 = currop->op1;
2661 op0 = create_component_ref_by_pieces_1 (block, ref, operand, stmts);
2662 if (!op0)
2663 return NULL_TREE;
2664 /* op1 should be a FIELD_DECL, which are represented by themselves. */
2665 op1 = currop->op0;
2666 if (genop2)
2667 {
2668 genop2 = find_or_generate_expression (block, genop2, stmts);
2669 if (!genop2)
2670 return NULL_TREE;
2671 }
2672 return fold_build3 (COMPONENT_REF, TREE_TYPE (op1), op0, op1, genop2);
2673 }
2674
2675 case SSA_NAME:
2676 {
2677 genop = find_or_generate_expression (block, currop->op0, stmts);
2678 return genop;
2679 }
2680 case STRING_CST:
2681 case INTEGER_CST:
2682 case COMPLEX_CST:
2683 case VECTOR_CST:
2684 case REAL_CST:
2685 case CONSTRUCTOR:
2686 case VAR_DECL:
2687 case PARM_DECL:
2688 case CONST_DECL:
2689 case RESULT_DECL:
2690 case FUNCTION_DECL:
2691 return currop->op0;
2692
2693 default:
2694 gcc_unreachable ();
2695 }
2696 }
2697
2698 /* For COMPONENT_REF's and ARRAY_REF's, we can't have any intermediates for the
2699 COMPONENT_REF or MEM_REF or ARRAY_REF portion, because we'd end up with
2700 trying to rename aggregates into ssa form directly, which is a no no.
2701
2702 Thus, this routine doesn't create temporaries, it just builds a
2703 single access expression for the array, calling
2704 find_or_generate_expression to build the innermost pieces.
2705
2706 This function is a subroutine of create_expression_by_pieces, and
2707 should not be called on it's own unless you really know what you
2708 are doing. */
2709
2710 static tree
2711 create_component_ref_by_pieces (basic_block block, vn_reference_t ref,
2712 gimple_seq *stmts)
2713 {
2714 unsigned int op = 0;
2715 return create_component_ref_by_pieces_1 (block, ref, &op, stmts);
2716 }
2717
2718 /* Find a simple leader for an expression, or generate one using
2719 create_expression_by_pieces from a NARY expression for the value.
2720 BLOCK is the basic_block we are looking for leaders in.
2721 OP is the tree expression to find a leader for or generate.
2722 Returns the leader or NULL_TREE on failure. */
2723
2724 static tree
2725 find_or_generate_expression (basic_block block, tree op, gimple_seq *stmts)
2726 {
2727 pre_expr expr = get_or_alloc_expr_for (op);
2728 unsigned int lookfor = get_expr_value_id (expr);
2729 pre_expr leader = bitmap_find_leader (AVAIL_OUT (block), lookfor);
2730 if (leader)
2731 {
2732 if (leader->kind == NAME)
2733 return PRE_EXPR_NAME (leader);
2734 else if (leader->kind == CONSTANT)
2735 return PRE_EXPR_CONSTANT (leader);
2736
2737 /* Defer. */
2738 return NULL_TREE;
2739 }
2740
2741 /* It must be a complex expression, so generate it recursively. Note
2742 that this is only necessary to handle gcc.dg/tree-ssa/ssa-pre28.c
2743 where the insert algorithm fails to insert a required expression. */
2744 bitmap exprset = value_expressions[lookfor];
2745 bitmap_iterator bi;
2746 unsigned int i;
2747 EXECUTE_IF_SET_IN_BITMAP (exprset, 0, i, bi)
2748 {
2749 pre_expr temp = expression_for_id (i);
2750 /* We cannot insert random REFERENCE expressions at arbitrary
2751 places. We can insert NARYs which eventually re-materializes
2752 its operand values. */
2753 if (temp->kind == NARY)
2754 return create_expression_by_pieces (block, temp, stmts,
2755 get_expr_type (expr));
2756 }
2757
2758 /* Defer. */
2759 return NULL_TREE;
2760 }
2761
2762 /* Create an expression in pieces, so that we can handle very complex
2763 expressions that may be ANTIC, but not necessary GIMPLE.
2764 BLOCK is the basic block the expression will be inserted into,
2765 EXPR is the expression to insert (in value form)
2766 STMTS is a statement list to append the necessary insertions into.
2767
2768 This function will die if we hit some value that shouldn't be
2769 ANTIC but is (IE there is no leader for it, or its components).
2770 The function returns NULL_TREE in case a different antic expression
2771 has to be inserted first.
2772 This function may also generate expressions that are themselves
2773 partially or fully redundant. Those that are will be either made
2774 fully redundant during the next iteration of insert (for partially
2775 redundant ones), or eliminated by eliminate (for fully redundant
2776 ones). */
2777
2778 static tree
2779 create_expression_by_pieces (basic_block block, pre_expr expr,
2780 gimple_seq *stmts, tree type)
2781 {
2782 tree name;
2783 tree folded;
2784 gimple_seq forced_stmts = NULL;
2785 unsigned int value_id;
2786 gimple_stmt_iterator gsi;
2787 tree exprtype = type ? type : get_expr_type (expr);
2788 pre_expr nameexpr;
2789 gassign *newstmt;
2790
2791 switch (expr->kind)
2792 {
2793 /* We may hit the NAME/CONSTANT case if we have to convert types
2794 that value numbering saw through. */
2795 case NAME:
2796 folded = PRE_EXPR_NAME (expr);
2797 if (useless_type_conversion_p (exprtype, TREE_TYPE (folded)))
2798 return folded;
2799 break;
2800 case CONSTANT:
2801 {
2802 folded = PRE_EXPR_CONSTANT (expr);
2803 tree tem = fold_convert (exprtype, folded);
2804 if (is_gimple_min_invariant (tem))
2805 return tem;
2806 break;
2807 }
2808 case REFERENCE:
2809 if (PRE_EXPR_REFERENCE (expr)->operands[0].opcode == CALL_EXPR)
2810 {
2811 vn_reference_t ref = PRE_EXPR_REFERENCE (expr);
2812 unsigned int operand = 1;
2813 vn_reference_op_t currop = &ref->operands[0];
2814 tree sc = NULL_TREE;
2815 tree fn;
2816 if (TREE_CODE (currop->op0) == FUNCTION_DECL)
2817 fn = currop->op0;
2818 else
2819 fn = find_or_generate_expression (block, currop->op0, stmts);
2820 if (!fn)
2821 return NULL_TREE;
2822 if (currop->op1)
2823 {
2824 sc = find_or_generate_expression (block, currop->op1, stmts);
2825 if (!sc)
2826 return NULL_TREE;
2827 }
2828 auto_vec<tree> args (ref->operands.length () - 1);
2829 while (operand < ref->operands.length ())
2830 {
2831 tree arg = create_component_ref_by_pieces_1 (block, ref,
2832 &operand, stmts);
2833 if (!arg)
2834 return NULL_TREE;
2835 args.quick_push (arg);
2836 }
2837 gcall *call
2838 = gimple_build_call_vec ((TREE_CODE (fn) == FUNCTION_DECL
2839 ? build_fold_addr_expr (fn) : fn), args);
2840 gimple_call_set_with_bounds (call, currop->with_bounds);
2841 if (sc)
2842 gimple_call_set_chain (call, sc);
2843 tree forcedname = make_ssa_name (currop->type);
2844 gimple_call_set_lhs (call, forcedname);
2845 gimple_set_vuse (call, BB_LIVE_VOP_ON_EXIT (block));
2846 gimple_seq_add_stmt_without_update (&forced_stmts, call);
2847 folded = forcedname;
2848 }
2849 else
2850 {
2851 folded = create_component_ref_by_pieces (block,
2852 PRE_EXPR_REFERENCE (expr),
2853 stmts);
2854 if (!folded)
2855 return NULL_TREE;
2856 name = make_temp_ssa_name (exprtype, NULL, "pretmp");
2857 newstmt = gimple_build_assign (name, folded);
2858 gimple_seq_add_stmt_without_update (&forced_stmts, newstmt);
2859 gimple_set_vuse (newstmt, BB_LIVE_VOP_ON_EXIT (block));
2860 folded = name;
2861 }
2862 break;
2863 case NARY:
2864 {
2865 vn_nary_op_t nary = PRE_EXPR_NARY (expr);
2866 tree *genop = XALLOCAVEC (tree, nary->length);
2867 unsigned i;
2868 for (i = 0; i < nary->length; ++i)
2869 {
2870 genop[i] = find_or_generate_expression (block, nary->op[i], stmts);
2871 if (!genop[i])
2872 return NULL_TREE;
2873 /* Ensure genop[] is properly typed for POINTER_PLUS_EXPR. It
2874 may have conversions stripped. */
2875 if (nary->opcode == POINTER_PLUS_EXPR)
2876 {
2877 if (i == 0)
2878 genop[i] = gimple_convert (&forced_stmts,
2879 nary->type, genop[i]);
2880 else if (i == 1)
2881 genop[i] = gimple_convert (&forced_stmts,
2882 sizetype, genop[i]);
2883 }
2884 else
2885 genop[i] = gimple_convert (&forced_stmts,
2886 TREE_TYPE (nary->op[i]), genop[i]);
2887 }
2888 if (nary->opcode == CONSTRUCTOR)
2889 {
2890 vec<constructor_elt, va_gc> *elts = NULL;
2891 for (i = 0; i < nary->length; ++i)
2892 CONSTRUCTOR_APPEND_ELT (elts, NULL_TREE, genop[i]);
2893 folded = build_constructor (nary->type, elts);
2894 name = make_temp_ssa_name (exprtype, NULL, "pretmp");
2895 newstmt = gimple_build_assign (name, folded);
2896 gimple_seq_add_stmt_without_update (&forced_stmts, newstmt);
2897 folded = name;
2898 }
2899 else
2900 {
2901 switch (nary->length)
2902 {
2903 case 1:
2904 folded = gimple_build (&forced_stmts, nary->opcode, nary->type,
2905 genop[0]);
2906 break;
2907 case 2:
2908 folded = gimple_build (&forced_stmts, nary->opcode, nary->type,
2909 genop[0], genop[1]);
2910 break;
2911 case 3:
2912 folded = gimple_build (&forced_stmts, nary->opcode, nary->type,
2913 genop[0], genop[1], genop[2]);
2914 break;
2915 default:
2916 gcc_unreachable ();
2917 }
2918 }
2919 }
2920 break;
2921 default:
2922 gcc_unreachable ();
2923 }
2924
2925 folded = gimple_convert (&forced_stmts, exprtype, folded);
2926
2927 /* If there is nothing to insert, return the simplified result. */
2928 if (gimple_seq_empty_p (forced_stmts))
2929 return folded;
2930 /* If we simplified to a constant return it and discard eventually
2931 built stmts. */
2932 if (is_gimple_min_invariant (folded))
2933 {
2934 gimple_seq_discard (forced_stmts);
2935 return folded;
2936 }
2937 /* Likewise if we simplified to sth not queued for insertion. */
2938 bool found = false;
2939 gsi = gsi_last (forced_stmts);
2940 for (; !gsi_end_p (gsi); gsi_prev (&gsi))
2941 {
2942 gimple *stmt = gsi_stmt (gsi);
2943 tree forcedname = gimple_get_lhs (stmt);
2944 if (forcedname == folded)
2945 {
2946 found = true;
2947 break;
2948 }
2949 }
2950 if (! found)
2951 {
2952 gimple_seq_discard (forced_stmts);
2953 return folded;
2954 }
2955 gcc_assert (TREE_CODE (folded) == SSA_NAME);
2956
2957 /* If we have any intermediate expressions to the value sets, add them
2958 to the value sets and chain them in the instruction stream. */
2959 if (forced_stmts)
2960 {
2961 gsi = gsi_start (forced_stmts);
2962 for (; !gsi_end_p (gsi); gsi_next (&gsi))
2963 {
2964 gimple *stmt = gsi_stmt (gsi);
2965 tree forcedname = gimple_get_lhs (stmt);
2966 pre_expr nameexpr;
2967
2968 if (forcedname != folded)
2969 {
2970 VN_INFO_GET (forcedname)->valnum = forcedname;
2971 VN_INFO (forcedname)->value_id = get_next_value_id ();
2972 nameexpr = get_or_alloc_expr_for_name (forcedname);
2973 add_to_value (VN_INFO (forcedname)->value_id, nameexpr);
2974 bitmap_value_replace_in_set (NEW_SETS (block), nameexpr);
2975 bitmap_value_replace_in_set (AVAIL_OUT (block), nameexpr);
2976 }
2977
2978 bitmap_set_bit (inserted_exprs, SSA_NAME_VERSION (forcedname));
2979 }
2980 gimple_seq_add_seq (stmts, forced_stmts);
2981 }
2982
2983 name = folded;
2984
2985 /* Fold the last statement. */
2986 gsi = gsi_last (*stmts);
2987 if (fold_stmt_inplace (&gsi))
2988 update_stmt (gsi_stmt (gsi));
2989
2990 /* Add a value number to the temporary.
2991 The value may already exist in either NEW_SETS, or AVAIL_OUT, because
2992 we are creating the expression by pieces, and this particular piece of
2993 the expression may have been represented. There is no harm in replacing
2994 here. */
2995 value_id = get_expr_value_id (expr);
2996 VN_INFO_GET (name)->value_id = value_id;
2997 VN_INFO (name)->valnum = sccvn_valnum_from_value_id (value_id);
2998 if (VN_INFO (name)->valnum == NULL_TREE)
2999 VN_INFO (name)->valnum = name;
3000 gcc_assert (VN_INFO (name)->valnum != NULL_TREE);
3001 nameexpr = get_or_alloc_expr_for_name (name);
3002 add_to_value (value_id, nameexpr);
3003 if (NEW_SETS (block))
3004 bitmap_value_replace_in_set (NEW_SETS (block), nameexpr);
3005 bitmap_value_replace_in_set (AVAIL_OUT (block), nameexpr);
3006
3007 pre_stats.insertions++;
3008 if (dump_file && (dump_flags & TDF_DETAILS))
3009 {
3010 fprintf (dump_file, "Inserted ");
3011 print_gimple_stmt (dump_file, gsi_stmt (gsi_last (*stmts)), 0);
3012 fprintf (dump_file, " in predecessor %d (%04d)\n",
3013 block->index, value_id);
3014 }
3015
3016 return name;
3017 }
3018
3019
3020 /* Insert the to-be-made-available values of expression EXPRNUM for each
3021 predecessor, stored in AVAIL, into the predecessors of BLOCK, and
3022 merge the result with a phi node, given the same value number as
3023 NODE. Return true if we have inserted new stuff. */
3024
3025 static bool
3026 insert_into_preds_of_block (basic_block block, unsigned int exprnum,
3027 vec<pre_expr> avail)
3028 {
3029 pre_expr expr = expression_for_id (exprnum);
3030 pre_expr newphi;
3031 unsigned int val = get_expr_value_id (expr);
3032 edge pred;
3033 bool insertions = false;
3034 bool nophi = false;
3035 basic_block bprime;
3036 pre_expr eprime;
3037 edge_iterator ei;
3038 tree type = get_expr_type (expr);
3039 tree temp;
3040 gphi *phi;
3041
3042 /* Make sure we aren't creating an induction variable. */
3043 if (bb_loop_depth (block) > 0 && EDGE_COUNT (block->preds) == 2)
3044 {
3045 bool firstinsideloop = false;
3046 bool secondinsideloop = false;
3047 firstinsideloop = flow_bb_inside_loop_p (block->loop_father,
3048 EDGE_PRED (block, 0)->src);
3049 secondinsideloop = flow_bb_inside_loop_p (block->loop_father,
3050 EDGE_PRED (block, 1)->src);
3051 /* Induction variables only have one edge inside the loop. */
3052 if ((firstinsideloop ^ secondinsideloop)
3053 && expr->kind != REFERENCE)
3054 {
3055 if (dump_file && (dump_flags & TDF_DETAILS))
3056 fprintf (dump_file, "Skipping insertion of phi for partial redundancy: Looks like an induction variable\n");
3057 nophi = true;
3058 }
3059 }
3060
3061 /* Make the necessary insertions. */
3062 FOR_EACH_EDGE (pred, ei, block->preds)
3063 {
3064 gimple_seq stmts = NULL;
3065 tree builtexpr;
3066 bprime = pred->src;
3067 eprime = avail[pred->dest_idx];
3068 builtexpr = create_expression_by_pieces (bprime, eprime,
3069 &stmts, type);
3070 gcc_assert (!(pred->flags & EDGE_ABNORMAL));
3071 if (!gimple_seq_empty_p (stmts))
3072 {
3073 gsi_insert_seq_on_edge (pred, stmts);
3074 insertions = true;
3075 }
3076 if (!builtexpr)
3077 {
3078 /* We cannot insert a PHI node if we failed to insert
3079 on one edge. */
3080 nophi = true;
3081 continue;
3082 }
3083 if (is_gimple_min_invariant (builtexpr))
3084 avail[pred->dest_idx] = get_or_alloc_expr_for_constant (builtexpr);
3085 else
3086 avail[pred->dest_idx] = get_or_alloc_expr_for_name (builtexpr);
3087 }
3088 /* If we didn't want a phi node, and we made insertions, we still have
3089 inserted new stuff, and thus return true. If we didn't want a phi node,
3090 and didn't make insertions, we haven't added anything new, so return
3091 false. */
3092 if (nophi && insertions)
3093 return true;
3094 else if (nophi && !insertions)
3095 return false;
3096
3097 /* Now build a phi for the new variable. */
3098 temp = make_temp_ssa_name (type, NULL, "prephitmp");
3099 phi = create_phi_node (temp, block);
3100
3101 VN_INFO_GET (temp)->value_id = val;
3102 VN_INFO (temp)->valnum = sccvn_valnum_from_value_id (val);
3103 if (VN_INFO (temp)->valnum == NULL_TREE)
3104 VN_INFO (temp)->valnum = temp;
3105 bitmap_set_bit (inserted_exprs, SSA_NAME_VERSION (temp));
3106 FOR_EACH_EDGE (pred, ei, block->preds)
3107 {
3108 pre_expr ae = avail[pred->dest_idx];
3109 gcc_assert (get_expr_type (ae) == type
3110 || useless_type_conversion_p (type, get_expr_type (ae)));
3111 if (ae->kind == CONSTANT)
3112 add_phi_arg (phi, unshare_expr (PRE_EXPR_CONSTANT (ae)),
3113 pred, UNKNOWN_LOCATION);
3114 else
3115 add_phi_arg (phi, PRE_EXPR_NAME (ae), pred, UNKNOWN_LOCATION);
3116 }
3117
3118 newphi = get_or_alloc_expr_for_name (temp);
3119 add_to_value (val, newphi);
3120
3121 /* The value should *not* exist in PHI_GEN, or else we wouldn't be doing
3122 this insertion, since we test for the existence of this value in PHI_GEN
3123 before proceeding with the partial redundancy checks in insert_aux.
3124
3125 The value may exist in AVAIL_OUT, in particular, it could be represented
3126 by the expression we are trying to eliminate, in which case we want the
3127 replacement to occur. If it's not existing in AVAIL_OUT, we want it
3128 inserted there.
3129
3130 Similarly, to the PHI_GEN case, the value should not exist in NEW_SETS of
3131 this block, because if it did, it would have existed in our dominator's
3132 AVAIL_OUT, and would have been skipped due to the full redundancy check.
3133 */
3134
3135 bitmap_insert_into_set (PHI_GEN (block), newphi);
3136 bitmap_value_replace_in_set (AVAIL_OUT (block),
3137 newphi);
3138 bitmap_insert_into_set (NEW_SETS (block),
3139 newphi);
3140
3141 /* If we insert a PHI node for a conversion of another PHI node
3142 in the same basic-block try to preserve range information.
3143 This is important so that followup loop passes receive optimal
3144 number of iteration analysis results. See PR61743. */
3145 if (expr->kind == NARY
3146 && CONVERT_EXPR_CODE_P (expr->u.nary->opcode)
3147 && TREE_CODE (expr->u.nary->op[0]) == SSA_NAME
3148 && gimple_bb (SSA_NAME_DEF_STMT (expr->u.nary->op[0])) == block
3149 && INTEGRAL_TYPE_P (type)
3150 && INTEGRAL_TYPE_P (TREE_TYPE (expr->u.nary->op[0]))
3151 && (TYPE_PRECISION (type)
3152 >= TYPE_PRECISION (TREE_TYPE (expr->u.nary->op[0])))
3153 && SSA_NAME_RANGE_INFO (expr->u.nary->op[0]))
3154 {
3155 wide_int min, max;
3156 if (get_range_info (expr->u.nary->op[0], &min, &max) == VR_RANGE
3157 && !wi::neg_p (min, SIGNED)
3158 && !wi::neg_p (max, SIGNED))
3159 /* Just handle extension and sign-changes of all-positive ranges. */
3160 set_range_info (temp,
3161 SSA_NAME_RANGE_TYPE (expr->u.nary->op[0]),
3162 wide_int_storage::from (min, TYPE_PRECISION (type),
3163 TYPE_SIGN (type)),
3164 wide_int_storage::from (max, TYPE_PRECISION (type),
3165 TYPE_SIGN (type)));
3166 }
3167
3168 if (dump_file && (dump_flags & TDF_DETAILS))
3169 {
3170 fprintf (dump_file, "Created phi ");
3171 print_gimple_stmt (dump_file, phi, 0);
3172 fprintf (dump_file, " in block %d (%04d)\n", block->index, val);
3173 }
3174 pre_stats.phis++;
3175 return true;
3176 }
3177
3178
3179
3180 /* Perform insertion of partially redundant or hoistable values.
3181 For BLOCK, do the following:
3182 1. Propagate the NEW_SETS of the dominator into the current block.
3183 If the block has multiple predecessors,
3184 2a. Iterate over the ANTIC expressions for the block to see if
3185 any of them are partially redundant.
3186 2b. If so, insert them into the necessary predecessors to make
3187 the expression fully redundant.
3188 2c. Insert a new PHI merging the values of the predecessors.
3189 2d. Insert the new PHI, and the new expressions, into the
3190 NEW_SETS set.
3191 If the block has multiple successors,
3192 3a. Iterate over the ANTIC values for the block to see if
3193 any of them are good candidates for hoisting.
3194 3b. If so, insert expressions computing the values in BLOCK,
3195 and add the new expressions into the NEW_SETS set.
3196 4. Recursively call ourselves on the dominator children of BLOCK.
3197
3198 Steps 1, 2a, and 4 are done by insert_aux. 2b, 2c and 2d are done by
3199 do_pre_regular_insertion and do_partial_insertion. 3a and 3b are
3200 done in do_hoist_insertion.
3201 */
3202
3203 static bool
3204 do_pre_regular_insertion (basic_block block, basic_block dom)
3205 {
3206 bool new_stuff = false;
3207 vec<pre_expr> exprs;
3208 pre_expr expr;
3209 auto_vec<pre_expr> avail;
3210 int i;
3211
3212 exprs = sorted_array_from_bitmap_set (ANTIC_IN (block));
3213 avail.safe_grow (EDGE_COUNT (block->preds));
3214
3215 FOR_EACH_VEC_ELT (exprs, i, expr)
3216 {
3217 if (expr->kind == NARY
3218 || expr->kind == REFERENCE)
3219 {
3220 unsigned int val;
3221 bool by_some = false;
3222 bool cant_insert = false;
3223 bool all_same = true;
3224 pre_expr first_s = NULL;
3225 edge pred;
3226 basic_block bprime;
3227 pre_expr eprime = NULL;
3228 edge_iterator ei;
3229 pre_expr edoubleprime = NULL;
3230 bool do_insertion = false;
3231
3232 val = get_expr_value_id (expr);
3233 if (bitmap_set_contains_value (PHI_GEN (block), val))
3234 continue;
3235 if (bitmap_set_contains_value (AVAIL_OUT (dom), val))
3236 {
3237 if (dump_file && (dump_flags & TDF_DETAILS))
3238 {
3239 fprintf (dump_file, "Found fully redundant value: ");
3240 print_pre_expr (dump_file, expr);
3241 fprintf (dump_file, "\n");
3242 }
3243 continue;
3244 }
3245
3246 FOR_EACH_EDGE (pred, ei, block->preds)
3247 {
3248 unsigned int vprime;
3249
3250 /* We should never run insertion for the exit block
3251 and so not come across fake pred edges. */
3252 gcc_assert (!(pred->flags & EDGE_FAKE));
3253 bprime = pred->src;
3254 /* We are looking at ANTIC_OUT of bprime. */
3255 eprime = phi_translate (expr, ANTIC_IN (block), NULL,
3256 bprime, block);
3257
3258 /* eprime will generally only be NULL if the
3259 value of the expression, translated
3260 through the PHI for this predecessor, is
3261 undefined. If that is the case, we can't
3262 make the expression fully redundant,
3263 because its value is undefined along a
3264 predecessor path. We can thus break out
3265 early because it doesn't matter what the
3266 rest of the results are. */
3267 if (eprime == NULL)
3268 {
3269 avail[pred->dest_idx] = NULL;
3270 cant_insert = true;
3271 break;
3272 }
3273
3274 vprime = get_expr_value_id (eprime);
3275 edoubleprime = bitmap_find_leader (AVAIL_OUT (bprime),
3276 vprime);
3277 if (edoubleprime == NULL)
3278 {
3279 avail[pred->dest_idx] = eprime;
3280 all_same = false;
3281 }
3282 else
3283 {
3284 avail[pred->dest_idx] = edoubleprime;
3285 by_some = true;
3286 /* We want to perform insertions to remove a redundancy on
3287 a path in the CFG we want to optimize for speed. */
3288 if (optimize_edge_for_speed_p (pred))
3289 do_insertion = true;
3290 if (first_s == NULL)
3291 first_s = edoubleprime;
3292 else if (!pre_expr_d::equal (first_s, edoubleprime))
3293 all_same = false;
3294 }
3295 }
3296 /* If we can insert it, it's not the same value
3297 already existing along every predecessor, and
3298 it's defined by some predecessor, it is
3299 partially redundant. */
3300 if (!cant_insert && !all_same && by_some)
3301 {
3302 if (!do_insertion)
3303 {
3304 if (dump_file && (dump_flags & TDF_DETAILS))
3305 {
3306 fprintf (dump_file, "Skipping partial redundancy for "
3307 "expression ");
3308 print_pre_expr (dump_file, expr);
3309 fprintf (dump_file, " (%04d), no redundancy on to be "
3310 "optimized for speed edge\n", val);
3311 }
3312 }
3313 else if (dbg_cnt (treepre_insert))
3314 {
3315 if (dump_file && (dump_flags & TDF_DETAILS))
3316 {
3317 fprintf (dump_file, "Found partial redundancy for "
3318 "expression ");
3319 print_pre_expr (dump_file, expr);
3320 fprintf (dump_file, " (%04d)\n",
3321 get_expr_value_id (expr));
3322 }
3323 if (insert_into_preds_of_block (block,
3324 get_expression_id (expr),
3325 avail))
3326 new_stuff = true;
3327 }
3328 }
3329 /* If all edges produce the same value and that value is
3330 an invariant, then the PHI has the same value on all
3331 edges. Note this. */
3332 else if (!cant_insert && all_same)
3333 {
3334 gcc_assert (edoubleprime->kind == CONSTANT
3335 || edoubleprime->kind == NAME);
3336
3337 tree temp = make_temp_ssa_name (get_expr_type (expr),
3338 NULL, "pretmp");
3339 gassign *assign
3340 = gimple_build_assign (temp,
3341 edoubleprime->kind == CONSTANT ?
3342 PRE_EXPR_CONSTANT (edoubleprime) :
3343 PRE_EXPR_NAME (edoubleprime));
3344 gimple_stmt_iterator gsi = gsi_after_labels (block);
3345 gsi_insert_before (&gsi, assign, GSI_NEW_STMT);
3346
3347 VN_INFO_GET (temp)->value_id = val;
3348 VN_INFO (temp)->valnum = sccvn_valnum_from_value_id (val);
3349 if (VN_INFO (temp)->valnum == NULL_TREE)
3350 VN_INFO (temp)->valnum = temp;
3351 bitmap_set_bit (inserted_exprs, SSA_NAME_VERSION (temp));
3352 pre_expr newe = get_or_alloc_expr_for_name (temp);
3353 add_to_value (val, newe);
3354 bitmap_value_replace_in_set (AVAIL_OUT (block), newe);
3355 bitmap_insert_into_set (NEW_SETS (block), newe);
3356 }
3357 }
3358 }
3359
3360 exprs.release ();
3361 return new_stuff;
3362 }
3363
3364
3365 /* Perform insertion for partially anticipatable expressions. There
3366 is only one case we will perform insertion for these. This case is
3367 if the expression is partially anticipatable, and fully available.
3368 In this case, we know that putting it earlier will enable us to
3369 remove the later computation. */
3370
3371 static bool
3372 do_pre_partial_partial_insertion (basic_block block, basic_block dom)
3373 {
3374 bool new_stuff = false;
3375 vec<pre_expr> exprs;
3376 pre_expr expr;
3377 auto_vec<pre_expr> avail;
3378 int i;
3379
3380 exprs = sorted_array_from_bitmap_set (PA_IN (block));
3381 avail.safe_grow (EDGE_COUNT (block->preds));
3382
3383 FOR_EACH_VEC_ELT (exprs, i, expr)
3384 {
3385 if (expr->kind == NARY
3386 || expr->kind == REFERENCE)
3387 {
3388 unsigned int val;
3389 bool by_all = true;
3390 bool cant_insert = false;
3391 edge pred;
3392 basic_block bprime;
3393 pre_expr eprime = NULL;
3394 edge_iterator ei;
3395
3396 val = get_expr_value_id (expr);
3397 if (bitmap_set_contains_value (PHI_GEN (block), val))
3398 continue;
3399 if (bitmap_set_contains_value (AVAIL_OUT (dom), val))
3400 continue;
3401
3402 FOR_EACH_EDGE (pred, ei, block->preds)
3403 {
3404 unsigned int vprime;
3405 pre_expr edoubleprime;
3406
3407 /* We should never run insertion for the exit block
3408 and so not come across fake pred edges. */
3409 gcc_assert (!(pred->flags & EDGE_FAKE));
3410 bprime = pred->src;
3411 eprime = phi_translate (expr, ANTIC_IN (block),
3412 PA_IN (block),
3413 bprime, block);
3414
3415 /* eprime will generally only be NULL if the
3416 value of the expression, translated
3417 through the PHI for this predecessor, is
3418 undefined. If that is the case, we can't
3419 make the expression fully redundant,
3420 because its value is undefined along a
3421 predecessor path. We can thus break out
3422 early because it doesn't matter what the
3423 rest of the results are. */
3424 if (eprime == NULL)
3425 {
3426 avail[pred->dest_idx] = NULL;
3427 cant_insert = true;
3428 break;
3429 }
3430
3431 vprime = get_expr_value_id (eprime);
3432 edoubleprime = bitmap_find_leader (AVAIL_OUT (bprime), vprime);
3433 avail[pred->dest_idx] = edoubleprime;
3434 if (edoubleprime == NULL)
3435 {
3436 by_all = false;
3437 break;
3438 }
3439 }
3440
3441 /* If we can insert it, it's not the same value
3442 already existing along every predecessor, and
3443 it's defined by some predecessor, it is
3444 partially redundant. */
3445 if (!cant_insert && by_all)
3446 {
3447 edge succ;
3448 bool do_insertion = false;
3449
3450 /* Insert only if we can remove a later expression on a path
3451 that we want to optimize for speed.
3452 The phi node that we will be inserting in BLOCK is not free,
3453 and inserting it for the sake of !optimize_for_speed successor
3454 may cause regressions on the speed path. */
3455 FOR_EACH_EDGE (succ, ei, block->succs)
3456 {
3457 if (bitmap_set_contains_value (PA_IN (succ->dest), val)
3458 || bitmap_set_contains_value (ANTIC_IN (succ->dest), val))
3459 {
3460 if (optimize_edge_for_speed_p (succ))
3461 do_insertion = true;
3462 }
3463 }
3464
3465 if (!do_insertion)
3466 {
3467 if (dump_file && (dump_flags & TDF_DETAILS))
3468 {
3469 fprintf (dump_file, "Skipping partial partial redundancy "
3470 "for expression ");
3471 print_pre_expr (dump_file, expr);
3472 fprintf (dump_file, " (%04d), not (partially) anticipated "
3473 "on any to be optimized for speed edges\n", val);
3474 }
3475 }
3476 else if (dbg_cnt (treepre_insert))
3477 {
3478 pre_stats.pa_insert++;
3479 if (dump_file && (dump_flags & TDF_DETAILS))
3480 {
3481 fprintf (dump_file, "Found partial partial redundancy "
3482 "for expression ");
3483 print_pre_expr (dump_file, expr);
3484 fprintf (dump_file, " (%04d)\n",
3485 get_expr_value_id (expr));
3486 }
3487 if (insert_into_preds_of_block (block,
3488 get_expression_id (expr),
3489 avail))
3490 new_stuff = true;
3491 }
3492 }
3493 }
3494 }
3495
3496 exprs.release ();
3497 return new_stuff;
3498 }
3499
3500 /* Insert expressions in BLOCK to compute hoistable values up.
3501 Return TRUE if something was inserted, otherwise return FALSE.
3502 The caller has to make sure that BLOCK has at least two successors. */
3503
3504 static bool
3505 do_hoist_insertion (basic_block block)
3506 {
3507 edge e;
3508 edge_iterator ei;
3509 bool new_stuff = false;
3510 unsigned i;
3511 gimple_stmt_iterator last;
3512
3513 /* At least two successors, or else... */
3514 gcc_assert (EDGE_COUNT (block->succs) >= 2);
3515
3516 /* Check that all successors of BLOCK are dominated by block.
3517 We could use dominated_by_p() for this, but actually there is a much
3518 quicker check: any successor that is dominated by BLOCK can't have
3519 more than one predecessor edge. */
3520 FOR_EACH_EDGE (e, ei, block->succs)
3521 if (! single_pred_p (e->dest))
3522 return false;
3523
3524 /* Determine the insertion point. If we cannot safely insert before
3525 the last stmt if we'd have to, bail out. */
3526 last = gsi_last_bb (block);
3527 if (!gsi_end_p (last)
3528 && !is_ctrl_stmt (gsi_stmt (last))
3529 && stmt_ends_bb_p (gsi_stmt (last)))
3530 return false;
3531
3532 /* Compute the set of hoistable expressions from ANTIC_IN. First compute
3533 hoistable values. */
3534 bitmap_set hoistable_set;
3535
3536 /* A hoistable value must be in ANTIC_IN(block)
3537 but not in AVAIL_OUT(BLOCK). */
3538 bitmap_initialize (&hoistable_set.values, &grand_bitmap_obstack);
3539 bitmap_and_compl (&hoistable_set.values,
3540 &ANTIC_IN (block)->values, &AVAIL_OUT (block)->values);
3541
3542 /* Short-cut for a common case: hoistable_set is empty. */
3543 if (bitmap_empty_p (&hoistable_set.values))
3544 return false;
3545
3546 /* Compute which of the hoistable values is in AVAIL_OUT of
3547 at least one of the successors of BLOCK. */
3548 bitmap_head availout_in_some;
3549 bitmap_initialize (&availout_in_some, &grand_bitmap_obstack);
3550 FOR_EACH_EDGE (e, ei, block->succs)
3551 /* Do not consider expressions solely because their availability
3552 on loop exits. They'd be ANTIC-IN throughout the whole loop
3553 and thus effectively hoisted across loops by combination of
3554 PRE and hoisting. */
3555 if (! loop_exit_edge_p (block->loop_father, e))
3556 bitmap_ior_and_into (&availout_in_some, &hoistable_set.values,
3557 &AVAIL_OUT (e->dest)->values);
3558 bitmap_clear (&hoistable_set.values);
3559
3560 /* Short-cut for a common case: availout_in_some is empty. */
3561 if (bitmap_empty_p (&availout_in_some))
3562 return false;
3563
3564 /* Hack hoitable_set in-place so we can use sorted_array_from_bitmap_set. */
3565 hoistable_set.values = availout_in_some;
3566 hoistable_set.expressions = ANTIC_IN (block)->expressions;
3567
3568 /* Now finally construct the topological-ordered expression set. */
3569 vec<pre_expr> exprs = sorted_array_from_bitmap_set (&hoistable_set);
3570
3571 bitmap_clear (&hoistable_set.values);
3572
3573 /* If there are candidate values for hoisting, insert expressions
3574 strategically to make the hoistable expressions fully redundant. */
3575 pre_expr expr;
3576 FOR_EACH_VEC_ELT (exprs, i, expr)
3577 {
3578 /* While we try to sort expressions topologically above the
3579 sorting doesn't work out perfectly. Catch expressions we
3580 already inserted. */
3581 unsigned int value_id = get_expr_value_id (expr);
3582 if (bitmap_set_contains_value (AVAIL_OUT (block), value_id))
3583 {
3584 if (dump_file && (dump_flags & TDF_DETAILS))
3585 {
3586 fprintf (dump_file,
3587 "Already inserted expression for ");
3588 print_pre_expr (dump_file, expr);
3589 fprintf (dump_file, " (%04d)\n", value_id);
3590 }
3591 continue;
3592 }
3593
3594 /* OK, we should hoist this value. Perform the transformation. */
3595 pre_stats.hoist_insert++;
3596 if (dump_file && (dump_flags & TDF_DETAILS))
3597 {
3598 fprintf (dump_file,
3599 "Inserting expression in block %d for code hoisting: ",
3600 block->index);
3601 print_pre_expr (dump_file, expr);
3602 fprintf (dump_file, " (%04d)\n", value_id);
3603 }
3604
3605 gimple_seq stmts = NULL;
3606 tree res = create_expression_by_pieces (block, expr, &stmts,
3607 get_expr_type (expr));
3608
3609 /* Do not return true if expression creation ultimately
3610 did not insert any statements. */
3611 if (gimple_seq_empty_p (stmts))
3612 res = NULL_TREE;
3613 else
3614 {
3615 if (gsi_end_p (last) || is_ctrl_stmt (gsi_stmt (last)))
3616 gsi_insert_seq_before (&last, stmts, GSI_SAME_STMT);
3617 else
3618 gsi_insert_seq_after (&last, stmts, GSI_NEW_STMT);
3619 }
3620
3621 /* Make sure to not return true if expression creation ultimately
3622 failed but also make sure to insert any stmts produced as they
3623 are tracked in inserted_exprs. */
3624 if (! res)
3625 continue;
3626
3627 new_stuff = true;
3628 }
3629
3630 exprs.release ();
3631
3632 return new_stuff;
3633 }
3634
3635 /* Do a dominator walk on the control flow graph, and insert computations
3636 of values as necessary for PRE and hoisting. */
3637
3638 static bool
3639 insert_aux (basic_block block, bool do_pre, bool do_hoist)
3640 {
3641 basic_block son;
3642 bool new_stuff = false;
3643
3644 if (block)
3645 {
3646 basic_block dom;
3647 dom = get_immediate_dominator (CDI_DOMINATORS, block);
3648 if (dom)
3649 {
3650 unsigned i;
3651 bitmap_iterator bi;
3652 bitmap_set_t newset;
3653
3654 /* First, update the AVAIL_OUT set with anything we may have
3655 inserted higher up in the dominator tree. */
3656 newset = NEW_SETS (dom);
3657 if (newset)
3658 {
3659 /* Note that we need to value_replace both NEW_SETS, and
3660 AVAIL_OUT. For both the case of NEW_SETS, the value may be
3661 represented by some non-simple expression here that we want
3662 to replace it with. */
3663 FOR_EACH_EXPR_ID_IN_SET (newset, i, bi)
3664 {
3665 pre_expr expr = expression_for_id (i);
3666 bitmap_value_replace_in_set (NEW_SETS (block), expr);
3667 bitmap_value_replace_in_set (AVAIL_OUT (block), expr);
3668 }
3669 }
3670
3671 /* Insert expressions for partial redundancies. */
3672 if (do_pre && !single_pred_p (block))
3673 {
3674 new_stuff |= do_pre_regular_insertion (block, dom);
3675 if (do_partial_partial)
3676 new_stuff |= do_pre_partial_partial_insertion (block, dom);
3677 }
3678
3679 /* Insert expressions for hoisting. */
3680 if (do_hoist && EDGE_COUNT (block->succs) >= 2)
3681 new_stuff |= do_hoist_insertion (block);
3682 }
3683 }
3684 for (son = first_dom_son (CDI_DOMINATORS, block);
3685 son;
3686 son = next_dom_son (CDI_DOMINATORS, son))
3687 {
3688 new_stuff |= insert_aux (son, do_pre, do_hoist);
3689 }
3690
3691 return new_stuff;
3692 }
3693
3694 /* Perform insertion of partially redundant and hoistable values. */
3695
3696 static void
3697 insert (void)
3698 {
3699 bool new_stuff = true;
3700 basic_block bb;
3701 int num_iterations = 0;
3702
3703 FOR_ALL_BB_FN (bb, cfun)
3704 NEW_SETS (bb) = bitmap_set_new ();
3705
3706 while (new_stuff)
3707 {
3708 num_iterations++;
3709 if (dump_file && dump_flags & TDF_DETAILS)
3710 fprintf (dump_file, "Starting insert iteration %d\n", num_iterations);
3711 new_stuff = insert_aux (ENTRY_BLOCK_PTR_FOR_FN (cfun), flag_tree_pre,
3712 flag_code_hoisting);
3713
3714 /* Clear the NEW sets before the next iteration. We have already
3715 fully propagated its contents. */
3716 if (new_stuff)
3717 FOR_ALL_BB_FN (bb, cfun)
3718 bitmap_set_free (NEW_SETS (bb));
3719 }
3720 statistics_histogram_event (cfun, "insert iterations", num_iterations);
3721 }
3722
3723
3724 /* Compute the AVAIL set for all basic blocks.
3725
3726 This function performs value numbering of the statements in each basic
3727 block. The AVAIL sets are built from information we glean while doing
3728 this value numbering, since the AVAIL sets contain only one entry per
3729 value.
3730
3731 AVAIL_IN[BLOCK] = AVAIL_OUT[dom(BLOCK)].
3732 AVAIL_OUT[BLOCK] = AVAIL_IN[BLOCK] U PHI_GEN[BLOCK] U TMP_GEN[BLOCK]. */
3733
3734 static void
3735 compute_avail (void)
3736 {
3737
3738 basic_block block, son;
3739 basic_block *worklist;
3740 size_t sp = 0;
3741 unsigned i;
3742 tree name;
3743
3744 /* We pretend that default definitions are defined in the entry block.
3745 This includes function arguments and the static chain decl. */
3746 FOR_EACH_SSA_NAME (i, name, cfun)
3747 {
3748 pre_expr e;
3749 if (!SSA_NAME_IS_DEFAULT_DEF (name)
3750 || has_zero_uses (name)
3751 || virtual_operand_p (name))
3752 continue;
3753
3754 e = get_or_alloc_expr_for_name (name);
3755 add_to_value (get_expr_value_id (e), e);
3756 bitmap_insert_into_set (TMP_GEN (ENTRY_BLOCK_PTR_FOR_FN (cfun)), e);
3757 bitmap_value_insert_into_set (AVAIL_OUT (ENTRY_BLOCK_PTR_FOR_FN (cfun)),
3758 e);
3759 }
3760
3761 if (dump_file && (dump_flags & TDF_DETAILS))
3762 {
3763 print_bitmap_set (dump_file, TMP_GEN (ENTRY_BLOCK_PTR_FOR_FN (cfun)),
3764 "tmp_gen", ENTRY_BLOCK);
3765 print_bitmap_set (dump_file, AVAIL_OUT (ENTRY_BLOCK_PTR_FOR_FN (cfun)),
3766 "avail_out", ENTRY_BLOCK);
3767 }
3768
3769 /* Allocate the worklist. */
3770 worklist = XNEWVEC (basic_block, n_basic_blocks_for_fn (cfun));
3771
3772 /* Seed the algorithm by putting the dominator children of the entry
3773 block on the worklist. */
3774 for (son = first_dom_son (CDI_DOMINATORS, ENTRY_BLOCK_PTR_FOR_FN (cfun));
3775 son;
3776 son = next_dom_son (CDI_DOMINATORS, son))
3777 worklist[sp++] = son;
3778
3779 BB_LIVE_VOP_ON_EXIT (ENTRY_BLOCK_PTR_FOR_FN (cfun))
3780 = ssa_default_def (cfun, gimple_vop (cfun));
3781
3782 /* Loop until the worklist is empty. */
3783 while (sp)
3784 {
3785 gimple *stmt;
3786 basic_block dom;
3787
3788 /* Pick a block from the worklist. */
3789 block = worklist[--sp];
3790
3791 /* Initially, the set of available values in BLOCK is that of
3792 its immediate dominator. */
3793 dom = get_immediate_dominator (CDI_DOMINATORS, block);
3794 if (dom)
3795 {
3796 bitmap_set_copy (AVAIL_OUT (block), AVAIL_OUT (dom));
3797 BB_LIVE_VOP_ON_EXIT (block) = BB_LIVE_VOP_ON_EXIT (dom);
3798 }
3799
3800 /* Generate values for PHI nodes. */
3801 for (gphi_iterator gsi = gsi_start_phis (block); !gsi_end_p (gsi);
3802 gsi_next (&gsi))
3803 {
3804 tree result = gimple_phi_result (gsi.phi ());
3805
3806 /* We have no need for virtual phis, as they don't represent
3807 actual computations. */
3808 if (virtual_operand_p (result))
3809 {
3810 BB_LIVE_VOP_ON_EXIT (block) = result;
3811 continue;
3812 }
3813
3814 pre_expr e = get_or_alloc_expr_for_name (result);
3815 add_to_value (get_expr_value_id (e), e);
3816 bitmap_value_insert_into_set (AVAIL_OUT (block), e);
3817 bitmap_insert_into_set (PHI_GEN (block), e);
3818 }
3819
3820 BB_MAY_NOTRETURN (block) = 0;
3821
3822 /* Now compute value numbers and populate value sets with all
3823 the expressions computed in BLOCK. */
3824 for (gimple_stmt_iterator gsi = gsi_start_bb (block); !gsi_end_p (gsi);
3825 gsi_next (&gsi))
3826 {
3827 ssa_op_iter iter;
3828 tree op;
3829
3830 stmt = gsi_stmt (gsi);
3831
3832 /* Cache whether the basic-block has any non-visible side-effect
3833 or control flow.
3834 If this isn't a call or it is the last stmt in the
3835 basic-block then the CFG represents things correctly. */
3836 if (is_gimple_call (stmt) && !stmt_ends_bb_p (stmt))
3837 {
3838 /* Non-looping const functions always return normally.
3839 Otherwise the call might not return or have side-effects
3840 that forbids hoisting possibly trapping expressions
3841 before it. */
3842 int flags = gimple_call_flags (stmt);
3843 if (!(flags & ECF_CONST)
3844 || (flags & ECF_LOOPING_CONST_OR_PURE))
3845 BB_MAY_NOTRETURN (block) = 1;
3846 }
3847
3848 FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_DEF)
3849 {
3850 pre_expr e = get_or_alloc_expr_for_name (op);
3851
3852 add_to_value (get_expr_value_id (e), e);
3853 bitmap_insert_into_set (TMP_GEN (block), e);
3854 bitmap_value_insert_into_set (AVAIL_OUT (block), e);
3855 }
3856
3857 if (gimple_vdef (stmt))
3858 BB_LIVE_VOP_ON_EXIT (block) = gimple_vdef (stmt);
3859
3860 if (gimple_has_side_effects (stmt)
3861 || stmt_could_throw_p (stmt)
3862 || is_gimple_debug (stmt))
3863 continue;
3864
3865 FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_USE)
3866 {
3867 if (ssa_undefined_value_p (op))
3868 continue;
3869 pre_expr e = get_or_alloc_expr_for_name (op);
3870 bitmap_value_insert_into_set (EXP_GEN (block), e);
3871 }
3872
3873 switch (gimple_code (stmt))
3874 {
3875 case GIMPLE_RETURN:
3876 continue;
3877
3878 case GIMPLE_CALL:
3879 {
3880 vn_reference_t ref;
3881 vn_reference_s ref1;
3882 pre_expr result = NULL;
3883
3884 /* We can value number only calls to real functions. */
3885 if (gimple_call_internal_p (stmt))
3886 continue;
3887
3888 vn_reference_lookup_call (as_a <gcall *> (stmt), &ref, &ref1);
3889 if (!ref)
3890 continue;
3891
3892 /* If the value of the call is not invalidated in
3893 this block until it is computed, add the expression
3894 to EXP_GEN. */
3895 if (!gimple_vuse (stmt)
3896 || gimple_code
3897 (SSA_NAME_DEF_STMT (gimple_vuse (stmt))) == GIMPLE_PHI
3898 || gimple_bb (SSA_NAME_DEF_STMT
3899 (gimple_vuse (stmt))) != block)
3900 {
3901 result = pre_expr_pool.allocate ();
3902 result->kind = REFERENCE;
3903 result->id = 0;
3904 PRE_EXPR_REFERENCE (result) = ref;
3905
3906 get_or_alloc_expression_id (result);
3907 add_to_value (get_expr_value_id (result), result);
3908 bitmap_value_insert_into_set (EXP_GEN (block), result);
3909 }
3910 continue;
3911 }
3912
3913 case GIMPLE_ASSIGN:
3914 {
3915 pre_expr result = NULL;
3916 switch (vn_get_stmt_kind (stmt))
3917 {
3918 case VN_NARY:
3919 {
3920 enum tree_code code = gimple_assign_rhs_code (stmt);
3921 vn_nary_op_t nary;
3922
3923 /* COND_EXPR and VEC_COND_EXPR are awkward in
3924 that they contain an embedded complex expression.
3925 Don't even try to shove those through PRE. */
3926 if (code == COND_EXPR
3927 || code == VEC_COND_EXPR)
3928 continue;
3929
3930 vn_nary_op_lookup_stmt (stmt, &nary);
3931 if (!nary)
3932 continue;
3933
3934 /* If the NARY traps and there was a preceding
3935 point in the block that might not return avoid
3936 adding the nary to EXP_GEN. */
3937 if (BB_MAY_NOTRETURN (block)
3938 && vn_nary_may_trap (nary))
3939 continue;
3940
3941 result = pre_expr_pool.allocate ();
3942 result->kind = NARY;
3943 result->id = 0;
3944 PRE_EXPR_NARY (result) = nary;
3945 break;
3946 }
3947
3948 case VN_REFERENCE:
3949 {
3950 tree rhs1 = gimple_assign_rhs1 (stmt);
3951 alias_set_type set = get_alias_set (rhs1);
3952 vec<vn_reference_op_s> operands
3953 = vn_reference_operands_for_lookup (rhs1);
3954 vn_reference_t ref;
3955 vn_reference_lookup_pieces (gimple_vuse (stmt), set,
3956 TREE_TYPE (rhs1),
3957 operands, &ref, VN_WALK);
3958 if (!ref)
3959 {
3960 operands.release ();
3961 continue;
3962 }
3963
3964 /* If the value of the reference is not invalidated in
3965 this block until it is computed, add the expression
3966 to EXP_GEN. */
3967 if (gimple_vuse (stmt))
3968 {
3969 gimple *def_stmt;
3970 bool ok = true;
3971 def_stmt = SSA_NAME_DEF_STMT (gimple_vuse (stmt));
3972 while (!gimple_nop_p (def_stmt)
3973 && gimple_code (def_stmt) != GIMPLE_PHI
3974 && gimple_bb (def_stmt) == block)
3975 {
3976 if (stmt_may_clobber_ref_p
3977 (def_stmt, gimple_assign_rhs1 (stmt)))
3978 {
3979 ok = false;
3980 break;
3981 }
3982 def_stmt
3983 = SSA_NAME_DEF_STMT (gimple_vuse (def_stmt));
3984 }
3985 if (!ok)
3986 {
3987 operands.release ();
3988 continue;
3989 }
3990 }
3991
3992 /* If the load was value-numbered to another
3993 load make sure we do not use its expression
3994 for insertion if it wouldn't be a valid
3995 replacement. */
3996 /* At the momemt we have a testcase
3997 for hoist insertion of aligned vs. misaligned
3998 variants in gcc.dg/torture/pr65270-1.c thus
3999 with just alignment to be considered we can
4000 simply replace the expression in the hashtable
4001 with the most conservative one. */
4002 vn_reference_op_t ref1 = &ref->operands.last ();
4003 while (ref1->opcode != TARGET_MEM_REF
4004 && ref1->opcode != MEM_REF
4005 && ref1 != &ref->operands[0])
4006 --ref1;
4007 vn_reference_op_t ref2 = &operands.last ();
4008 while (ref2->opcode != TARGET_MEM_REF
4009 && ref2->opcode != MEM_REF
4010 && ref2 != &operands[0])
4011 --ref2;
4012 if ((ref1->opcode == TARGET_MEM_REF
4013 || ref1->opcode == MEM_REF)
4014 && (TYPE_ALIGN (ref1->type)
4015 > TYPE_ALIGN (ref2->type)))
4016 ref1->type
4017 = build_aligned_type (ref1->type,
4018 TYPE_ALIGN (ref2->type));
4019 /* TBAA behavior is an obvious part so make sure
4020 that the hashtable one covers this as well
4021 by adjusting the ref alias set and its base. */
4022 if (ref->set == set
4023 || alias_set_subset_of (set, ref->set))
4024 ;
4025 else if (alias_set_subset_of (ref->set, set))
4026 {
4027 ref->set = set;
4028 if (ref1->opcode == MEM_REF)
4029 ref1->op0
4030 = wide_int_to_tree (TREE_TYPE (ref2->op0),
4031 wi::to_wide (ref1->op0));
4032 else
4033 ref1->op2
4034 = wide_int_to_tree (TREE_TYPE (ref2->op2),
4035 wi::to_wide (ref1->op2));
4036 }
4037 else
4038 {
4039 ref->set = 0;
4040 if (ref1->opcode == MEM_REF)
4041 ref1->op0
4042 = wide_int_to_tree (ptr_type_node,
4043 wi::to_wide (ref1->op0));
4044 else
4045 ref1->op2
4046 = wide_int_to_tree (ptr_type_node,
4047 wi::to_wide (ref1->op2));
4048 }
4049 operands.release ();
4050
4051 result = pre_expr_pool.allocate ();
4052 result->kind = REFERENCE;
4053 result->id = 0;
4054 PRE_EXPR_REFERENCE (result) = ref;
4055 break;
4056 }
4057
4058 default:
4059 continue;
4060 }
4061
4062 get_or_alloc_expression_id (result);
4063 add_to_value (get_expr_value_id (result), result);
4064 bitmap_value_insert_into_set (EXP_GEN (block), result);
4065 continue;
4066 }
4067 default:
4068 break;
4069 }
4070 }
4071
4072 if (dump_file && (dump_flags & TDF_DETAILS))
4073 {
4074 print_bitmap_set (dump_file, EXP_GEN (block),
4075 "exp_gen", block->index);
4076 print_bitmap_set (dump_file, PHI_GEN (block),
4077 "phi_gen", block->index);
4078 print_bitmap_set (dump_file, TMP_GEN (block),
4079 "tmp_gen", block->index);
4080 print_bitmap_set (dump_file, AVAIL_OUT (block),
4081 "avail_out", block->index);
4082 }
4083
4084 /* Put the dominator children of BLOCK on the worklist of blocks
4085 to compute available sets for. */
4086 for (son = first_dom_son (CDI_DOMINATORS, block);
4087 son;
4088 son = next_dom_son (CDI_DOMINATORS, son))
4089 worklist[sp++] = son;
4090 }
4091
4092 free (worklist);
4093 }
4094
4095
4096 /* Local state for the eliminate domwalk. */
4097 static vec<gimple *> el_to_remove;
4098 static vec<gimple *> el_to_fixup;
4099 static unsigned int el_todo;
4100 static vec<tree> el_avail;
4101 static vec<tree> el_avail_stack;
4102
4103 /* Return a leader for OP that is available at the current point of the
4104 eliminate domwalk. */
4105
4106 static tree
4107 eliminate_avail (tree op)
4108 {
4109 tree valnum = VN_INFO (op)->valnum;
4110 if (TREE_CODE (valnum) == SSA_NAME)
4111 {
4112 if (SSA_NAME_IS_DEFAULT_DEF (valnum))
4113 return valnum;
4114 if (el_avail.length () > SSA_NAME_VERSION (valnum))
4115 return el_avail[SSA_NAME_VERSION (valnum)];
4116 }
4117 else if (is_gimple_min_invariant (valnum))
4118 return valnum;
4119 return NULL_TREE;
4120 }
4121
4122 /* At the current point of the eliminate domwalk make OP available. */
4123
4124 static void
4125 eliminate_push_avail (tree op)
4126 {
4127 tree valnum = VN_INFO (op)->valnum;
4128 if (TREE_CODE (valnum) == SSA_NAME)
4129 {
4130 if (el_avail.length () <= SSA_NAME_VERSION (valnum))
4131 el_avail.safe_grow_cleared (SSA_NAME_VERSION (valnum) + 1);
4132 tree pushop = op;
4133 if (el_avail[SSA_NAME_VERSION (valnum)])
4134 pushop = el_avail[SSA_NAME_VERSION (valnum)];
4135 el_avail_stack.safe_push (pushop);
4136 el_avail[SSA_NAME_VERSION (valnum)] = op;
4137 }
4138 }
4139
4140 /* Insert the expression recorded by SCCVN for VAL at *GSI. Returns
4141 the leader for the expression if insertion was successful. */
4142
4143 static tree
4144 eliminate_insert (gimple_stmt_iterator *gsi, tree val)
4145 {
4146 /* We can insert a sequence with a single assignment only. */
4147 gimple_seq stmts = VN_INFO (val)->expr;
4148 if (!gimple_seq_singleton_p (stmts))
4149 return NULL_TREE;
4150 gassign *stmt = dyn_cast <gassign *> (gimple_seq_first_stmt (stmts));
4151 if (!stmt
4152 || (!CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (stmt))
4153 && gimple_assign_rhs_code (stmt) != VIEW_CONVERT_EXPR
4154 && gimple_assign_rhs_code (stmt) != BIT_FIELD_REF
4155 && (gimple_assign_rhs_code (stmt) != BIT_AND_EXPR
4156 || TREE_CODE (gimple_assign_rhs2 (stmt)) != INTEGER_CST)))
4157 return NULL_TREE;
4158
4159 tree op = gimple_assign_rhs1 (stmt);
4160 if (gimple_assign_rhs_code (stmt) == VIEW_CONVERT_EXPR
4161 || gimple_assign_rhs_code (stmt) == BIT_FIELD_REF)
4162 op = TREE_OPERAND (op, 0);
4163 tree leader = TREE_CODE (op) == SSA_NAME ? eliminate_avail (op) : op;
4164 if (!leader)
4165 return NULL_TREE;
4166
4167 tree res;
4168 stmts = NULL;
4169 if (gimple_assign_rhs_code (stmt) == BIT_FIELD_REF)
4170 res = gimple_build (&stmts, BIT_FIELD_REF,
4171 TREE_TYPE (val), leader,
4172 TREE_OPERAND (gimple_assign_rhs1 (stmt), 1),
4173 TREE_OPERAND (gimple_assign_rhs1 (stmt), 2));
4174 else if (gimple_assign_rhs_code (stmt) == BIT_AND_EXPR)
4175 res = gimple_build (&stmts, BIT_AND_EXPR,
4176 TREE_TYPE (val), leader, gimple_assign_rhs2 (stmt));
4177 else
4178 res = gimple_build (&stmts, gimple_assign_rhs_code (stmt),
4179 TREE_TYPE (val), leader);
4180 if (TREE_CODE (res) != SSA_NAME
4181 || SSA_NAME_IS_DEFAULT_DEF (res)
4182 || gimple_bb (SSA_NAME_DEF_STMT (res)))
4183 {
4184 gimple_seq_discard (stmts);
4185
4186 /* During propagation we have to treat SSA info conservatively
4187 and thus we can end up simplifying the inserted expression
4188 at elimination time to sth not defined in stmts. */
4189 /* But then this is a redundancy we failed to detect. Which means
4190 res now has two values. That doesn't play well with how
4191 we track availability here, so give up. */
4192 if (dump_file && (dump_flags & TDF_DETAILS))
4193 {
4194 if (TREE_CODE (res) == SSA_NAME)
4195 res = eliminate_avail (res);
4196 if (res)
4197 {
4198 fprintf (dump_file, "Failed to insert expression for value ");
4199 print_generic_expr (dump_file, val);
4200 fprintf (dump_file, " which is really fully redundant to ");
4201 print_generic_expr (dump_file, res);
4202 fprintf (dump_file, "\n");
4203 }
4204 }
4205
4206 return NULL_TREE;
4207 }
4208 else
4209 {
4210 gsi_insert_seq_before (gsi, stmts, GSI_SAME_STMT);
4211 VN_INFO_GET (res)->valnum = val;
4212 }
4213
4214 pre_stats.insertions++;
4215 if (dump_file && (dump_flags & TDF_DETAILS))
4216 {
4217 fprintf (dump_file, "Inserted ");
4218 print_gimple_stmt (dump_file, SSA_NAME_DEF_STMT (res), 0);
4219 }
4220
4221 return res;
4222 }
4223
4224 class eliminate_dom_walker : public dom_walker
4225 {
4226 public:
4227 eliminate_dom_walker (cdi_direction direction, bool do_pre_)
4228 : dom_walker (direction), do_pre (do_pre_) {}
4229
4230 virtual edge before_dom_children (basic_block);
4231 virtual void after_dom_children (basic_block);
4232
4233 bool do_pre;
4234 };
4235
4236 /* Perform elimination for the basic-block B during the domwalk. */
4237
4238 edge
4239 eliminate_dom_walker::before_dom_children (basic_block b)
4240 {
4241 /* Mark new bb. */
4242 el_avail_stack.safe_push (NULL_TREE);
4243
4244 /* Skip unreachable blocks marked unreachable during the SCCVN domwalk. */
4245 edge_iterator ei;
4246 edge e;
4247 FOR_EACH_EDGE (e, ei, b->preds)
4248 if (e->flags & EDGE_EXECUTABLE)
4249 break;
4250 if (! e)
4251 return NULL;
4252
4253 for (gphi_iterator gsi = gsi_start_phis (b); !gsi_end_p (gsi);)
4254 {
4255 gphi *phi = gsi.phi ();
4256 tree res = PHI_RESULT (phi);
4257
4258 if (virtual_operand_p (res))
4259 {
4260 gsi_next (&gsi);
4261 continue;
4262 }
4263
4264 tree sprime = eliminate_avail (res);
4265 if (sprime
4266 && sprime != res)
4267 {
4268 if (dump_file && (dump_flags & TDF_DETAILS))
4269 {
4270 fprintf (dump_file, "Replaced redundant PHI node defining ");
4271 print_generic_expr (dump_file, res);
4272 fprintf (dump_file, " with ");
4273 print_generic_expr (dump_file, sprime);
4274 fprintf (dump_file, "\n");
4275 }
4276
4277 /* If we inserted this PHI node ourself, it's not an elimination. */
4278 if (inserted_exprs
4279 && bitmap_bit_p (inserted_exprs, SSA_NAME_VERSION (res)))
4280 pre_stats.phis--;
4281 else
4282 pre_stats.eliminations++;
4283
4284 /* If we will propagate into all uses don't bother to do
4285 anything. */
4286 if (may_propagate_copy (res, sprime))
4287 {
4288 /* Mark the PHI for removal. */
4289 el_to_remove.safe_push (phi);
4290 gsi_next (&gsi);
4291 continue;
4292 }
4293
4294 remove_phi_node (&gsi, false);
4295
4296 if (!useless_type_conversion_p (TREE_TYPE (res), TREE_TYPE (sprime)))
4297 sprime = fold_convert (TREE_TYPE (res), sprime);
4298 gimple *stmt = gimple_build_assign (res, sprime);
4299 gimple_stmt_iterator gsi2 = gsi_after_labels (b);
4300 gsi_insert_before (&gsi2, stmt, GSI_NEW_STMT);
4301 continue;
4302 }
4303
4304 eliminate_push_avail (res);
4305 gsi_next (&gsi);
4306 }
4307
4308 for (gimple_stmt_iterator gsi = gsi_start_bb (b);
4309 !gsi_end_p (gsi);
4310 gsi_next (&gsi))
4311 {
4312 tree sprime = NULL_TREE;
4313 gimple *stmt = gsi_stmt (gsi);
4314 tree lhs = gimple_get_lhs (stmt);
4315 if (lhs && TREE_CODE (lhs) == SSA_NAME
4316 && !gimple_has_volatile_ops (stmt)
4317 /* See PR43491. Do not replace a global register variable when
4318 it is a the RHS of an assignment. Do replace local register
4319 variables since gcc does not guarantee a local variable will
4320 be allocated in register.
4321 ??? The fix isn't effective here. This should instead
4322 be ensured by not value-numbering them the same but treating
4323 them like volatiles? */
4324 && !(gimple_assign_single_p (stmt)
4325 && (TREE_CODE (gimple_assign_rhs1 (stmt)) == VAR_DECL
4326 && DECL_HARD_REGISTER (gimple_assign_rhs1 (stmt))
4327 && is_global_var (gimple_assign_rhs1 (stmt)))))
4328 {
4329 sprime = eliminate_avail (lhs);
4330 if (!sprime)
4331 {
4332 /* If there is no existing usable leader but SCCVN thinks
4333 it has an expression it wants to use as replacement,
4334 insert that. */
4335 tree val = VN_INFO (lhs)->valnum;
4336 if (val != VN_TOP
4337 && TREE_CODE (val) == SSA_NAME
4338 && VN_INFO (val)->needs_insertion
4339 && VN_INFO (val)->expr != NULL
4340 && (sprime = eliminate_insert (&gsi, val)) != NULL_TREE)
4341 eliminate_push_avail (sprime);
4342 }
4343
4344 /* If this now constitutes a copy duplicate points-to
4345 and range info appropriately. This is especially
4346 important for inserted code. See tree-ssa-copy.c
4347 for similar code. */
4348 if (sprime
4349 && TREE_CODE (sprime) == SSA_NAME)
4350 {
4351 basic_block sprime_b = gimple_bb (SSA_NAME_DEF_STMT (sprime));
4352 if (POINTER_TYPE_P (TREE_TYPE (lhs))
4353 && VN_INFO_PTR_INFO (lhs)
4354 && ! VN_INFO_PTR_INFO (sprime))
4355 {
4356 duplicate_ssa_name_ptr_info (sprime,
4357 VN_INFO_PTR_INFO (lhs));
4358 if (b != sprime_b)
4359 mark_ptr_info_alignment_unknown
4360 (SSA_NAME_PTR_INFO (sprime));
4361 }
4362 else if (INTEGRAL_TYPE_P (TREE_TYPE (lhs))
4363 && VN_INFO_RANGE_INFO (lhs)
4364 && ! VN_INFO_RANGE_INFO (sprime)
4365 && b == sprime_b)
4366 duplicate_ssa_name_range_info (sprime,
4367 VN_INFO_RANGE_TYPE (lhs),
4368 VN_INFO_RANGE_INFO (lhs));
4369 }
4370
4371 /* Inhibit the use of an inserted PHI on a loop header when
4372 the address of the memory reference is a simple induction
4373 variable. In other cases the vectorizer won't do anything
4374 anyway (either it's loop invariant or a complicated
4375 expression). */
4376 if (sprime
4377 && TREE_CODE (sprime) == SSA_NAME
4378 && do_pre
4379 && (flag_tree_loop_vectorize || flag_tree_parallelize_loops > 1)
4380 && loop_outer (b->loop_father)
4381 && has_zero_uses (sprime)
4382 && bitmap_bit_p (inserted_exprs, SSA_NAME_VERSION (sprime))
4383 && gimple_assign_load_p (stmt))
4384 {
4385 gimple *def_stmt = SSA_NAME_DEF_STMT (sprime);
4386 basic_block def_bb = gimple_bb (def_stmt);
4387 if (gimple_code (def_stmt) == GIMPLE_PHI
4388 && def_bb->loop_father->header == def_bb)
4389 {
4390 loop_p loop = def_bb->loop_father;
4391 ssa_op_iter iter;
4392 tree op;
4393 bool found = false;
4394 FOR_EACH_SSA_TREE_OPERAND (op, stmt, iter, SSA_OP_USE)
4395 {
4396 affine_iv iv;
4397 def_bb = gimple_bb (SSA_NAME_DEF_STMT (op));
4398 if (def_bb
4399 && flow_bb_inside_loop_p (loop, def_bb)
4400 && simple_iv (loop, loop, op, &iv, true))
4401 {
4402 found = true;
4403 break;
4404 }
4405 }
4406 if (found)
4407 {
4408 if (dump_file && (dump_flags & TDF_DETAILS))
4409 {
4410 fprintf (dump_file, "Not replacing ");
4411 print_gimple_expr (dump_file, stmt, 0);
4412 fprintf (dump_file, " with ");
4413 print_generic_expr (dump_file, sprime);
4414 fprintf (dump_file, " which would add a loop"
4415 " carried dependence to loop %d\n",
4416 loop->num);
4417 }
4418 /* Don't keep sprime available. */
4419 sprime = NULL_TREE;
4420 }
4421 }
4422 }
4423
4424 if (sprime)
4425 {
4426 /* If we can propagate the value computed for LHS into
4427 all uses don't bother doing anything with this stmt. */
4428 if (may_propagate_copy (lhs, sprime))
4429 {
4430 /* Mark it for removal. */
4431 el_to_remove.safe_push (stmt);
4432
4433 /* ??? Don't count copy/constant propagations. */
4434 if (gimple_assign_single_p (stmt)
4435 && (TREE_CODE (gimple_assign_rhs1 (stmt)) == SSA_NAME
4436 || gimple_assign_rhs1 (stmt) == sprime))
4437 continue;
4438
4439 if (dump_file && (dump_flags & TDF_DETAILS))
4440 {
4441 fprintf (dump_file, "Replaced ");
4442 print_gimple_expr (dump_file, stmt, 0);
4443 fprintf (dump_file, " with ");
4444 print_generic_expr (dump_file, sprime);
4445 fprintf (dump_file, " in all uses of ");
4446 print_gimple_stmt (dump_file, stmt, 0);
4447 }
4448
4449 pre_stats.eliminations++;
4450 continue;
4451 }
4452
4453 /* If this is an assignment from our leader (which
4454 happens in the case the value-number is a constant)
4455 then there is nothing to do. */
4456 if (gimple_assign_single_p (stmt)
4457 && sprime == gimple_assign_rhs1 (stmt))
4458 continue;
4459
4460 /* Else replace its RHS. */
4461 bool can_make_abnormal_goto
4462 = is_gimple_call (stmt)
4463 && stmt_can_make_abnormal_goto (stmt);
4464
4465 if (dump_file && (dump_flags & TDF_DETAILS))
4466 {
4467 fprintf (dump_file, "Replaced ");
4468 print_gimple_expr (dump_file, stmt, 0);
4469 fprintf (dump_file, " with ");
4470 print_generic_expr (dump_file, sprime);
4471 fprintf (dump_file, " in ");
4472 print_gimple_stmt (dump_file, stmt, 0);
4473 }
4474
4475 pre_stats.eliminations++;
4476 gimple *orig_stmt = stmt;
4477 if (!useless_type_conversion_p (TREE_TYPE (lhs),
4478 TREE_TYPE (sprime)))
4479 sprime = fold_convert (TREE_TYPE (lhs), sprime);
4480 tree vdef = gimple_vdef (stmt);
4481 tree vuse = gimple_vuse (stmt);
4482 propagate_tree_value_into_stmt (&gsi, sprime);
4483 stmt = gsi_stmt (gsi);
4484 update_stmt (stmt);
4485 if (vdef != gimple_vdef (stmt))
4486 VN_INFO (vdef)->valnum = vuse;
4487
4488 /* If we removed EH side-effects from the statement, clean
4489 its EH information. */
4490 if (maybe_clean_or_replace_eh_stmt (orig_stmt, stmt))
4491 {
4492 bitmap_set_bit (need_eh_cleanup,
4493 gimple_bb (stmt)->index);
4494 if (dump_file && (dump_flags & TDF_DETAILS))
4495 fprintf (dump_file, " Removed EH side-effects.\n");
4496 }
4497
4498 /* Likewise for AB side-effects. */
4499 if (can_make_abnormal_goto
4500 && !stmt_can_make_abnormal_goto (stmt))
4501 {
4502 bitmap_set_bit (need_ab_cleanup,
4503 gimple_bb (stmt)->index);
4504 if (dump_file && (dump_flags & TDF_DETAILS))
4505 fprintf (dump_file, " Removed AB side-effects.\n");
4506 }
4507
4508 continue;
4509 }
4510 }
4511
4512 /* If the statement is a scalar store, see if the expression
4513 has the same value number as its rhs. If so, the store is
4514 dead. */
4515 if (gimple_assign_single_p (stmt)
4516 && !gimple_has_volatile_ops (stmt)
4517 && !is_gimple_reg (gimple_assign_lhs (stmt))
4518 && (TREE_CODE (gimple_assign_rhs1 (stmt)) == SSA_NAME
4519 || is_gimple_min_invariant (gimple_assign_rhs1 (stmt))))
4520 {
4521 tree val;
4522 tree rhs = gimple_assign_rhs1 (stmt);
4523 vn_reference_t vnresult;
4524 val = vn_reference_lookup (lhs, gimple_vuse (stmt), VN_WALKREWRITE,
4525 &vnresult, false);
4526 if (TREE_CODE (rhs) == SSA_NAME)
4527 rhs = VN_INFO (rhs)->valnum;
4528 if (val
4529 && operand_equal_p (val, rhs, 0))
4530 {
4531 /* We can only remove the later store if the former aliases
4532 at least all accesses the later one does or if the store
4533 was to readonly memory storing the same value. */
4534 alias_set_type set = get_alias_set (lhs);
4535 if (! vnresult
4536 || vnresult->set == set
4537 || alias_set_subset_of (set, vnresult->set))
4538 {
4539 if (dump_file && (dump_flags & TDF_DETAILS))
4540 {
4541 fprintf (dump_file, "Deleted redundant store ");
4542 print_gimple_stmt (dump_file, stmt, 0);
4543 }
4544
4545 /* Queue stmt for removal. */
4546 el_to_remove.safe_push (stmt);
4547 continue;
4548 }
4549 }
4550 }
4551
4552 /* If this is a control statement value numbering left edges
4553 unexecuted on force the condition in a way consistent with
4554 that. */
4555 if (gcond *cond = dyn_cast <gcond *> (stmt))
4556 {
4557 if ((EDGE_SUCC (b, 0)->flags & EDGE_EXECUTABLE)
4558 ^ (EDGE_SUCC (b, 1)->flags & EDGE_EXECUTABLE))
4559 {
4560 if (dump_file && (dump_flags & TDF_DETAILS))
4561 {
4562 fprintf (dump_file, "Removing unexecutable edge from ");
4563 print_gimple_stmt (dump_file, stmt, 0);
4564 }
4565 if (((EDGE_SUCC (b, 0)->flags & EDGE_TRUE_VALUE) != 0)
4566 == ((EDGE_SUCC (b, 0)->flags & EDGE_EXECUTABLE) != 0))
4567 gimple_cond_make_true (cond);
4568 else
4569 gimple_cond_make_false (cond);
4570 update_stmt (cond);
4571 el_todo |= TODO_cleanup_cfg;
4572 continue;
4573 }
4574 }
4575
4576 bool can_make_abnormal_goto = stmt_can_make_abnormal_goto (stmt);
4577 bool was_noreturn = (is_gimple_call (stmt)
4578 && gimple_call_noreturn_p (stmt));
4579 tree vdef = gimple_vdef (stmt);
4580 tree vuse = gimple_vuse (stmt);
4581
4582 /* If we didn't replace the whole stmt (or propagate the result
4583 into all uses), replace all uses on this stmt with their
4584 leaders. */
4585 bool modified = false;
4586 use_operand_p use_p;
4587 ssa_op_iter iter;
4588 FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE)
4589 {
4590 tree use = USE_FROM_PTR (use_p);
4591 /* ??? The call code above leaves stmt operands un-updated. */
4592 if (TREE_CODE (use) != SSA_NAME)
4593 continue;
4594 tree sprime = eliminate_avail (use);
4595 if (sprime && sprime != use
4596 && may_propagate_copy (use, sprime)
4597 /* We substitute into debug stmts to avoid excessive
4598 debug temporaries created by removed stmts, but we need
4599 to avoid doing so for inserted sprimes as we never want
4600 to create debug temporaries for them. */
4601 && (!inserted_exprs
4602 || TREE_CODE (sprime) != SSA_NAME
4603 || !is_gimple_debug (stmt)
4604 || !bitmap_bit_p (inserted_exprs, SSA_NAME_VERSION (sprime))))
4605 {
4606 propagate_value (use_p, sprime);
4607 modified = true;
4608 }
4609 }
4610
4611 /* Fold the stmt if modified, this canonicalizes MEM_REFs we propagated
4612 into which is a requirement for the IPA devirt machinery. */
4613 gimple *old_stmt = stmt;
4614 if (modified)
4615 {
4616 /* If a formerly non-invariant ADDR_EXPR is turned into an
4617 invariant one it was on a separate stmt. */
4618 if (gimple_assign_single_p (stmt)
4619 && TREE_CODE (gimple_assign_rhs1 (stmt)) == ADDR_EXPR)
4620 recompute_tree_invariant_for_addr_expr (gimple_assign_rhs1 (stmt));
4621 gimple_stmt_iterator prev = gsi;
4622 gsi_prev (&prev);
4623 if (fold_stmt (&gsi))
4624 {
4625 /* fold_stmt may have created new stmts inbetween
4626 the previous stmt and the folded stmt. Mark
4627 all defs created there as varying to not confuse
4628 the SCCVN machinery as we're using that even during
4629 elimination. */
4630 if (gsi_end_p (prev))
4631 prev = gsi_start_bb (b);
4632 else
4633 gsi_next (&prev);
4634 if (gsi_stmt (prev) != gsi_stmt (gsi))
4635 do
4636 {
4637 tree def;
4638 ssa_op_iter dit;
4639 FOR_EACH_SSA_TREE_OPERAND (def, gsi_stmt (prev),
4640 dit, SSA_OP_ALL_DEFS)
4641 /* As existing DEFs may move between stmts
4642 we have to guard VN_INFO_GET. */
4643 if (! has_VN_INFO (def))
4644 VN_INFO_GET (def)->valnum = def;
4645 if (gsi_stmt (prev) == gsi_stmt (gsi))
4646 break;
4647 gsi_next (&prev);
4648 }
4649 while (1);
4650 }
4651 stmt = gsi_stmt (gsi);
4652 /* In case we folded the stmt away schedule the NOP for removal. */
4653 if (gimple_nop_p (stmt))
4654 el_to_remove.safe_push (stmt);
4655 }
4656
4657 /* Visit indirect calls and turn them into direct calls if
4658 possible using the devirtualization machinery. Do this before
4659 checking for required EH/abnormal/noreturn cleanup as devird
4660 may expose more of those. */
4661 if (gcall *call_stmt = dyn_cast <gcall *> (stmt))
4662 {
4663 tree fn = gimple_call_fn (call_stmt);
4664 if (fn
4665 && flag_devirtualize
4666 && virtual_method_call_p (fn))
4667 {
4668 tree otr_type = obj_type_ref_class (fn);
4669 unsigned HOST_WIDE_INT otr_tok
4670 = tree_to_uhwi (OBJ_TYPE_REF_TOKEN (fn));
4671 tree instance;
4672 ipa_polymorphic_call_context context (current_function_decl,
4673 fn, stmt, &instance);
4674 context.get_dynamic_type (instance, OBJ_TYPE_REF_OBJECT (fn),
4675 otr_type, stmt);
4676 bool final;
4677 vec <cgraph_node *> targets
4678 = possible_polymorphic_call_targets (obj_type_ref_class (fn),
4679 otr_tok, context, &final);
4680 if (dump_file)
4681 dump_possible_polymorphic_call_targets (dump_file,
4682 obj_type_ref_class (fn),
4683 otr_tok, context);
4684 if (final && targets.length () <= 1 && dbg_cnt (devirt))
4685 {
4686 tree fn;
4687 if (targets.length () == 1)
4688 fn = targets[0]->decl;
4689 else
4690 fn = builtin_decl_implicit (BUILT_IN_UNREACHABLE);
4691 if (dump_enabled_p ())
4692 {
4693 location_t loc = gimple_location (stmt);
4694 dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, loc,
4695 "converting indirect call to "
4696 "function %s\n",
4697 lang_hooks.decl_printable_name (fn, 2));
4698 }
4699 gimple_call_set_fndecl (call_stmt, fn);
4700 /* If changing the call to __builtin_unreachable
4701 or similar noreturn function, adjust gimple_call_fntype
4702 too. */
4703 if (gimple_call_noreturn_p (call_stmt)
4704 && VOID_TYPE_P (TREE_TYPE (TREE_TYPE (fn)))
4705 && TYPE_ARG_TYPES (TREE_TYPE (fn))
4706 && (TREE_VALUE (TYPE_ARG_TYPES (TREE_TYPE (fn)))
4707 == void_type_node))
4708 gimple_call_set_fntype (call_stmt, TREE_TYPE (fn));
4709 maybe_remove_unused_call_args (cfun, call_stmt);
4710 modified = true;
4711 }
4712 }
4713 }
4714
4715 if (modified)
4716 {
4717 /* When changing a call into a noreturn call, cfg cleanup
4718 is needed to fix up the noreturn call. */
4719 if (!was_noreturn
4720 && is_gimple_call (stmt) && gimple_call_noreturn_p (stmt))
4721 el_to_fixup.safe_push (stmt);
4722 /* When changing a condition or switch into one we know what
4723 edge will be executed, schedule a cfg cleanup. */
4724 if ((gimple_code (stmt) == GIMPLE_COND
4725 && (gimple_cond_true_p (as_a <gcond *> (stmt))
4726 || gimple_cond_false_p (as_a <gcond *> (stmt))))
4727 || (gimple_code (stmt) == GIMPLE_SWITCH
4728 && TREE_CODE (gimple_switch_index
4729 (as_a <gswitch *> (stmt))) == INTEGER_CST))
4730 el_todo |= TODO_cleanup_cfg;
4731 /* If we removed EH side-effects from the statement, clean
4732 its EH information. */
4733 if (maybe_clean_or_replace_eh_stmt (old_stmt, stmt))
4734 {
4735 bitmap_set_bit (need_eh_cleanup,
4736 gimple_bb (stmt)->index);
4737 if (dump_file && (dump_flags & TDF_DETAILS))
4738 fprintf (dump_file, " Removed EH side-effects.\n");
4739 }
4740 /* Likewise for AB side-effects. */
4741 if (can_make_abnormal_goto
4742 && !stmt_can_make_abnormal_goto (stmt))
4743 {
4744 bitmap_set_bit (need_ab_cleanup,
4745 gimple_bb (stmt)->index);
4746 if (dump_file && (dump_flags & TDF_DETAILS))
4747 fprintf (dump_file, " Removed AB side-effects.\n");
4748 }
4749 update_stmt (stmt);
4750 if (vdef != gimple_vdef (stmt))
4751 VN_INFO (vdef)->valnum = vuse;
4752 }
4753
4754 /* Make new values available - for fully redundant LHS we
4755 continue with the next stmt above and skip this. */
4756 def_operand_p defp;
4757 FOR_EACH_SSA_DEF_OPERAND (defp, stmt, iter, SSA_OP_DEF)
4758 eliminate_push_avail (DEF_FROM_PTR (defp));
4759 }
4760
4761 /* Replace destination PHI arguments. */
4762 FOR_EACH_EDGE (e, ei, b->succs)
4763 if (e->flags & EDGE_EXECUTABLE)
4764 for (gphi_iterator gsi = gsi_start_phis (e->dest);
4765 !gsi_end_p (gsi);
4766 gsi_next (&gsi))
4767 {
4768 gphi *phi = gsi.phi ();
4769 use_operand_p use_p = PHI_ARG_DEF_PTR_FROM_EDGE (phi, e);
4770 tree arg = USE_FROM_PTR (use_p);
4771 if (TREE_CODE (arg) != SSA_NAME
4772 || virtual_operand_p (arg))
4773 continue;
4774 tree sprime = eliminate_avail (arg);
4775 if (sprime && may_propagate_copy (arg, sprime))
4776 propagate_value (use_p, sprime);
4777 }
4778 return NULL;
4779 }
4780
4781 /* Make no longer available leaders no longer available. */
4782
4783 void
4784 eliminate_dom_walker::after_dom_children (basic_block)
4785 {
4786 tree entry;
4787 while ((entry = el_avail_stack.pop ()) != NULL_TREE)
4788 {
4789 tree valnum = VN_INFO (entry)->valnum;
4790 tree old = el_avail[SSA_NAME_VERSION (valnum)];
4791 if (old == entry)
4792 el_avail[SSA_NAME_VERSION (valnum)] = NULL_TREE;
4793 else
4794 el_avail[SSA_NAME_VERSION (valnum)] = entry;
4795 }
4796 }
4797
4798 /* Eliminate fully redundant computations. */
4799
4800 static unsigned int
4801 eliminate (bool do_pre)
4802 {
4803 need_eh_cleanup = BITMAP_ALLOC (NULL);
4804 need_ab_cleanup = BITMAP_ALLOC (NULL);
4805
4806 el_to_remove.create (0);
4807 el_to_fixup.create (0);
4808 el_todo = 0;
4809 el_avail.create (num_ssa_names);
4810 el_avail_stack.create (0);
4811
4812 eliminate_dom_walker (CDI_DOMINATORS,
4813 do_pre).walk (cfun->cfg->x_entry_block_ptr);
4814
4815 el_avail.release ();
4816 el_avail_stack.release ();
4817
4818 return el_todo;
4819 }
4820
4821 /* Perform CFG cleanups made necessary by elimination. */
4822
4823 static unsigned
4824 fini_eliminate (void)
4825 {
4826 gimple_stmt_iterator gsi;
4827 gimple *stmt;
4828 unsigned todo = 0;
4829
4830 /* We cannot remove stmts during BB walk, especially not release SSA
4831 names there as this confuses the VN machinery. The stmts ending
4832 up in el_to_remove are either stores or simple copies.
4833 Remove stmts in reverse order to make debug stmt creation possible. */
4834 while (!el_to_remove.is_empty ())
4835 {
4836 stmt = el_to_remove.pop ();
4837
4838 if (dump_file && (dump_flags & TDF_DETAILS))
4839 {
4840 fprintf (dump_file, "Removing dead stmt ");
4841 print_gimple_stmt (dump_file, stmt, 0, 0);
4842 }
4843
4844 gsi = gsi_for_stmt (stmt);
4845 if (gimple_code (stmt) == GIMPLE_PHI)
4846 remove_phi_node (&gsi, true);
4847 else
4848 {
4849 basic_block bb = gimple_bb (stmt);
4850 unlink_stmt_vdef (stmt);
4851 if (gsi_remove (&gsi, true))
4852 bitmap_set_bit (need_eh_cleanup, bb->index);
4853 if (is_gimple_call (stmt) && stmt_can_make_abnormal_goto (stmt))
4854 bitmap_set_bit (need_ab_cleanup, bb->index);
4855 release_defs (stmt);
4856 }
4857
4858 /* Removing a stmt may expose a forwarder block. */
4859 todo |= TODO_cleanup_cfg;
4860 }
4861 el_to_remove.release ();
4862
4863 /* Fixup stmts that became noreturn calls. This may require splitting
4864 blocks and thus isn't possible during the dominator walk. Do this
4865 in reverse order so we don't inadvertedly remove a stmt we want to
4866 fixup by visiting a dominating now noreturn call first. */
4867 while (!el_to_fixup.is_empty ())
4868 {
4869 stmt = el_to_fixup.pop ();
4870
4871 if (dump_file && (dump_flags & TDF_DETAILS))
4872 {
4873 fprintf (dump_file, "Fixing up noreturn call ");
4874 print_gimple_stmt (dump_file, stmt, 0);
4875 }
4876
4877 if (fixup_noreturn_call (stmt))
4878 todo |= TODO_cleanup_cfg;
4879 }
4880 el_to_fixup.release ();
4881
4882 bool do_eh_cleanup = !bitmap_empty_p (need_eh_cleanup);
4883 bool do_ab_cleanup = !bitmap_empty_p (need_ab_cleanup);
4884
4885 if (do_eh_cleanup)
4886 gimple_purge_all_dead_eh_edges (need_eh_cleanup);
4887
4888 if (do_ab_cleanup)
4889 gimple_purge_all_dead_abnormal_call_edges (need_ab_cleanup);
4890
4891 BITMAP_FREE (need_eh_cleanup);
4892 BITMAP_FREE (need_ab_cleanup);
4893
4894 if (do_eh_cleanup || do_ab_cleanup)
4895 todo |= TODO_cleanup_cfg;
4896 return todo;
4897 }
4898
4899 /* Cheap DCE of a known set of possibly dead stmts.
4900
4901 Because we don't follow exactly the standard PRE algorithm, and decide not
4902 to insert PHI nodes sometimes, and because value numbering of casts isn't
4903 perfect, we sometimes end up inserting dead code. This simple DCE-like
4904 pass removes any insertions we made that weren't actually used. */
4905
4906 static void
4907 remove_dead_inserted_code (void)
4908 {
4909 /* ??? Re-use inserted_exprs as worklist not only as initial set.
4910 This may end up removing non-inserted code as well. If we
4911 keep inserted_exprs unchanged we could restrict new worklist
4912 elements to members of inserted_exprs. */
4913 bitmap worklist = inserted_exprs;
4914 while (! bitmap_empty_p (worklist))
4915 {
4916 /* Pop item. */
4917 unsigned i = bitmap_first_set_bit (worklist);
4918 bitmap_clear_bit (worklist, i);
4919
4920 tree def = ssa_name (i);
4921 /* Removed by somebody else or still in use. */
4922 if (! def || ! has_zero_uses (def))
4923 continue;
4924
4925 gimple *t = SSA_NAME_DEF_STMT (def);
4926 if (gimple_has_side_effects (t))
4927 continue;
4928
4929 /* Add uses to the worklist. */
4930 ssa_op_iter iter;
4931 use_operand_p use_p;
4932 FOR_EACH_PHI_OR_STMT_USE (use_p, t, iter, SSA_OP_USE)
4933 {
4934 tree use = USE_FROM_PTR (use_p);
4935 if (TREE_CODE (use) == SSA_NAME
4936 && ! SSA_NAME_IS_DEFAULT_DEF (use))
4937 bitmap_set_bit (worklist, SSA_NAME_VERSION (use));
4938 }
4939
4940 /* Remove stmt. */
4941 if (dump_file && (dump_flags & TDF_DETAILS))
4942 {
4943 fprintf (dump_file, "Removing unnecessary insertion:");
4944 print_gimple_stmt (dump_file, t, 0);
4945 }
4946 gimple_stmt_iterator gsi = gsi_for_stmt (t);
4947 if (gimple_code (t) == GIMPLE_PHI)
4948 remove_phi_node (&gsi, true);
4949 else
4950 {
4951 gsi_remove (&gsi, true);
4952 release_defs (t);
4953 }
4954 }
4955 }
4956
4957
4958 /* Initialize data structures used by PRE. */
4959
4960 static void
4961 init_pre (void)
4962 {
4963 basic_block bb;
4964
4965 next_expression_id = 1;
4966 expressions.create (0);
4967 expressions.safe_push (NULL);
4968 value_expressions.create (get_max_value_id () + 1);
4969 value_expressions.safe_grow_cleared (get_max_value_id () + 1);
4970 name_to_id.create (0);
4971
4972 inserted_exprs = BITMAP_ALLOC (NULL);
4973
4974 connect_infinite_loops_to_exit ();
4975 memset (&pre_stats, 0, sizeof (pre_stats));
4976
4977 alloc_aux_for_blocks (sizeof (struct bb_bitmap_sets));
4978
4979 calculate_dominance_info (CDI_DOMINATORS);
4980
4981 bitmap_obstack_initialize (&grand_bitmap_obstack);
4982 phi_translate_table = new hash_table<expr_pred_trans_d> (5110);
4983 expression_to_id = new hash_table<pre_expr_d> (num_ssa_names * 3);
4984 FOR_ALL_BB_FN (bb, cfun)
4985 {
4986 EXP_GEN (bb) = bitmap_set_new ();
4987 PHI_GEN (bb) = bitmap_set_new ();
4988 TMP_GEN (bb) = bitmap_set_new ();
4989 AVAIL_OUT (bb) = bitmap_set_new ();
4990 }
4991 }
4992
4993
4994 /* Deallocate data structures used by PRE. */
4995
4996 static void
4997 fini_pre ()
4998 {
4999 value_expressions.release ();
5000 BITMAP_FREE (inserted_exprs);
5001 bitmap_obstack_release (&grand_bitmap_obstack);
5002 bitmap_set_pool.release ();
5003 pre_expr_pool.release ();
5004 delete phi_translate_table;
5005 phi_translate_table = NULL;
5006 delete expression_to_id;
5007 expression_to_id = NULL;
5008 name_to_id.release ();
5009
5010 free_aux_for_blocks ();
5011 }
5012
5013 namespace {
5014
5015 const pass_data pass_data_pre =
5016 {
5017 GIMPLE_PASS, /* type */
5018 "pre", /* name */
5019 OPTGROUP_NONE, /* optinfo_flags */
5020 TV_TREE_PRE, /* tv_id */
5021 ( PROP_cfg | PROP_ssa ), /* properties_required */
5022 0, /* properties_provided */
5023 0, /* properties_destroyed */
5024 TODO_rebuild_alias, /* todo_flags_start */
5025 0, /* todo_flags_finish */
5026 };
5027
5028 class pass_pre : public gimple_opt_pass
5029 {
5030 public:
5031 pass_pre (gcc::context *ctxt)
5032 : gimple_opt_pass (pass_data_pre, ctxt)
5033 {}
5034
5035 /* opt_pass methods: */
5036 virtual bool gate (function *)
5037 { return flag_tree_pre != 0 || flag_code_hoisting != 0; }
5038 virtual unsigned int execute (function *);
5039
5040 }; // class pass_pre
5041
5042 unsigned int
5043 pass_pre::execute (function *fun)
5044 {
5045 unsigned int todo = 0;
5046
5047 do_partial_partial =
5048 flag_tree_partial_pre && optimize_function_for_speed_p (fun);
5049
5050 /* This has to happen before SCCVN runs because
5051 loop_optimizer_init may create new phis, etc. */
5052 loop_optimizer_init (LOOPS_NORMAL);
5053 split_critical_edges ();
5054
5055 run_scc_vn (VN_WALK);
5056
5057 init_pre ();
5058 scev_initialize ();
5059
5060 /* Collect and value number expressions computed in each basic block. */
5061 compute_avail ();
5062
5063 /* Insert can get quite slow on an incredibly large number of basic
5064 blocks due to some quadratic behavior. Until this behavior is
5065 fixed, don't run it when he have an incredibly large number of
5066 bb's. If we aren't going to run insert, there is no point in
5067 computing ANTIC, either, even though it's plenty fast. */
5068 if (n_basic_blocks_for_fn (fun) < 4000)
5069 {
5070 compute_antic ();
5071 insert ();
5072 }
5073
5074 /* Make sure to remove fake edges before committing our inserts.
5075 This makes sure we don't end up with extra critical edges that
5076 we would need to split. */
5077 remove_fake_exit_edges ();
5078 gsi_commit_edge_inserts ();
5079
5080 /* Eliminate folds statements which might (should not...) end up
5081 not keeping virtual operands up-to-date. */
5082 gcc_assert (!need_ssa_update_p (fun));
5083
5084 /* Remove all the redundant expressions. */
5085 todo |= eliminate (true);
5086
5087 statistics_counter_event (fun, "Insertions", pre_stats.insertions);
5088 statistics_counter_event (fun, "PA inserted", pre_stats.pa_insert);
5089 statistics_counter_event (fun, "HOIST inserted", pre_stats.hoist_insert);
5090 statistics_counter_event (fun, "New PHIs", pre_stats.phis);
5091 statistics_counter_event (fun, "Eliminated", pre_stats.eliminations);
5092
5093 clear_expression_ids ();
5094
5095 scev_finalize ();
5096 todo |= fini_eliminate ();
5097 remove_dead_inserted_code ();
5098 fini_pre ();
5099 loop_optimizer_finalize ();
5100
5101 /* Restore SSA info before tail-merging as that resets it as well. */
5102 scc_vn_restore_ssa_info ();
5103
5104 /* TODO: tail_merge_optimize may merge all predecessors of a block, in which
5105 case we can merge the block with the remaining predecessor of the block.
5106 It should either:
5107 - call merge_blocks after each tail merge iteration
5108 - call merge_blocks after all tail merge iterations
5109 - mark TODO_cleanup_cfg when necessary
5110 - share the cfg cleanup with fini_pre. */
5111 todo |= tail_merge_optimize (todo);
5112
5113 free_scc_vn ();
5114
5115 /* Tail merging invalidates the virtual SSA web, together with
5116 cfg-cleanup opportunities exposed by PRE this will wreck the
5117 SSA updating machinery. So make sure to run update-ssa
5118 manually, before eventually scheduling cfg-cleanup as part of
5119 the todo. */
5120 update_ssa (TODO_update_ssa_only_virtuals);
5121
5122 return todo;
5123 }
5124
5125 } // anon namespace
5126
5127 gimple_opt_pass *
5128 make_pass_pre (gcc::context *ctxt)
5129 {
5130 return new pass_pre (ctxt);
5131 }
5132
5133 namespace {
5134
5135 const pass_data pass_data_fre =
5136 {
5137 GIMPLE_PASS, /* type */
5138 "fre", /* name */
5139 OPTGROUP_NONE, /* optinfo_flags */
5140 TV_TREE_FRE, /* tv_id */
5141 ( PROP_cfg | PROP_ssa ), /* properties_required */
5142 0, /* properties_provided */
5143 0, /* properties_destroyed */
5144 0, /* todo_flags_start */
5145 0, /* todo_flags_finish */
5146 };
5147
5148 class pass_fre : public gimple_opt_pass
5149 {
5150 public:
5151 pass_fre (gcc::context *ctxt)
5152 : gimple_opt_pass (pass_data_fre, ctxt)
5153 {}
5154
5155 /* opt_pass methods: */
5156 opt_pass * clone () { return new pass_fre (m_ctxt); }
5157 virtual bool gate (function *) { return flag_tree_fre != 0; }
5158 virtual unsigned int execute (function *);
5159
5160 }; // class pass_fre
5161
5162 unsigned int
5163 pass_fre::execute (function *fun)
5164 {
5165 unsigned int todo = 0;
5166
5167 run_scc_vn (VN_WALKREWRITE);
5168
5169 memset (&pre_stats, 0, sizeof (pre_stats));
5170
5171 /* Remove all the redundant expressions. */
5172 todo |= eliminate (false);
5173
5174 todo |= fini_eliminate ();
5175
5176 scc_vn_restore_ssa_info ();
5177 free_scc_vn ();
5178
5179 statistics_counter_event (fun, "Insertions", pre_stats.insertions);
5180 statistics_counter_event (fun, "Eliminated", pre_stats.eliminations);
5181
5182 return todo;
5183 }
5184
5185 } // anon namespace
5186
5187 gimple_opt_pass *
5188 make_pass_fre (gcc::context *ctxt)
5189 {
5190 return new pass_fre (ctxt);
5191 }