re PR tree-optimization/54245 (incorrect optimisation)
[gcc.git] / gcc / gimple-ssa-strength-reduction.c
1 /* Straight-line strength reduction.
2 Copyright (C) 2012 Free Software Foundation, Inc.
3 Contributed by Bill Schmidt, IBM <wschmidt@linux.ibm.com>
4
5 This file is part of GCC.
6
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
10 version.
11
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
16
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
20
21 /* There are many algorithms for performing strength reduction on
22 loops. This is not one of them. IVOPTS handles strength reduction
23 of induction variables just fine. This pass is intended to pick
24 up the crumbs it leaves behind, by considering opportunities for
25 strength reduction along dominator paths.
26
27 Strength reduction will be implemented in four stages, gradually
28 adding more complex candidates:
29
30 1) Explicit multiplies, known constant multipliers, no
31 conditional increments. (complete)
32 2) Explicit multiplies, unknown constant multipliers,
33 no conditional increments. (complete)
34 3) Implicit multiplies in addressing expressions. (complete)
35 4) Explicit multiplies, conditional increments. (pending)
36
37 It would also be possible to apply strength reduction to divisions
38 and modulos, but such opportunities are relatively uncommon.
39
40 Strength reduction is also currently restricted to integer operations.
41 If desired, it could be extended to floating-point operations under
42 control of something like -funsafe-math-optimizations. */
43
44 #include "config.h"
45 #include "system.h"
46 #include "coretypes.h"
47 #include "tree.h"
48 #include "gimple.h"
49 #include "basic-block.h"
50 #include "tree-pass.h"
51 #include "cfgloop.h"
52 #include "gimple-pretty-print.h"
53 #include "tree-flow.h"
54 #include "domwalk.h"
55 #include "pointer-set.h"
56 #include "expmed.h"
57 \f
58 /* Information about a strength reduction candidate. Each statement
59 in the candidate table represents an expression of one of the
60 following forms (the special case of CAND_REF will be described
61 later):
62
63 (CAND_MULT) S1: X = (B + i) * S
64 (CAND_ADD) S1: X = B + (i * S)
65
66 Here X and B are SSA names, i is an integer constant, and S is
67 either an SSA name or a constant. We call B the "base," i the
68 "index", and S the "stride."
69
70 Any statement S0 that dominates S1 and is of the form:
71
72 (CAND_MULT) S0: Y = (B + i') * S
73 (CAND_ADD) S0: Y = B + (i' * S)
74
75 is called a "basis" for S1. In both cases, S1 may be replaced by
76
77 S1': X = Y + (i - i') * S,
78
79 where (i - i') * S is folded to the extent possible.
80
81 All gimple statements are visited in dominator order, and each
82 statement that may contribute to one of the forms of S1 above is
83 given at least one entry in the candidate table. Such statements
84 include addition, pointer addition, subtraction, multiplication,
85 negation, copies, and nontrivial type casts. If a statement may
86 represent more than one expression of the forms of S1 above,
87 multiple "interpretations" are stored in the table and chained
88 together. Examples:
89
90 * An add of two SSA names may treat either operand as the base.
91 * A multiply of two SSA names, likewise.
92 * A copy or cast may be thought of as either a CAND_MULT with
93 i = 0 and S = 1, or as a CAND_ADD with i = 0 or S = 0.
94
95 Candidate records are allocated from an obstack. They are addressed
96 both from a hash table keyed on S1, and from a vector of candidate
97 pointers arranged in predominator order.
98
99 Opportunity note
100 ----------------
101 Currently we don't recognize:
102
103 S0: Y = (S * i') - B
104 S1: X = (S * i) - B
105
106 as a strength reduction opportunity, even though this S1 would
107 also be replaceable by the S1' above. This can be added if it
108 comes up in practice.
109
110 Strength reduction in addressing
111 --------------------------------
112 There is another kind of candidate known as CAND_REF. A CAND_REF
113 describes a statement containing a memory reference having
114 complex addressing that might benefit from strength reduction.
115 Specifically, we are interested in references for which
116 get_inner_reference returns a base address, offset, and bitpos as
117 follows:
118
119 base: MEM_REF (T1, C1)
120 offset: MULT_EXPR (PLUS_EXPR (T2, C2), C3)
121 bitpos: C4 * BITS_PER_UNIT
122
123 Here T1 and T2 are arbitrary trees, and C1, C2, C3, C4 are
124 arbitrary integer constants. Note that C2 may be zero, in which
125 case the offset will be MULT_EXPR (T2, C3).
126
127 When this pattern is recognized, the original memory reference
128 can be replaced with:
129
130 MEM_REF (POINTER_PLUS_EXPR (T1, MULT_EXPR (T2, C3)),
131 C1 + (C2 * C3) + C4)
132
133 which distributes the multiply to allow constant folding. When
134 two or more addressing expressions can be represented by MEM_REFs
135 of this form, differing only in the constants C1, C2, and C4,
136 making this substitution produces more efficient addressing during
137 the RTL phases. When there are not at least two expressions with
138 the same values of T1, T2, and C3, there is nothing to be gained
139 by the replacement.
140
141 Strength reduction of CAND_REFs uses the same infrastructure as
142 that used by CAND_MULTs and CAND_ADDs. We record T1 in the base (B)
143 field, MULT_EXPR (T2, C3) in the stride (S) field, and
144 C1 + (C2 * C3) + C4 in the index (i) field. A basis for a CAND_REF
145 is thus another CAND_REF with the same B and S values. When at
146 least two CAND_REFs are chained together using the basis relation,
147 each of them is replaced as above, resulting in improved code
148 generation for addressing. */
149
150
151 /* Index into the candidate vector, offset by 1. VECs are zero-based,
152 while cand_idx's are one-based, with zero indicating null. */
153 typedef unsigned cand_idx;
154
155 /* The kind of candidate. */
156 enum cand_kind
157 {
158 CAND_MULT,
159 CAND_ADD,
160 CAND_REF
161 };
162
163 struct slsr_cand_d
164 {
165 /* The candidate statement S1. */
166 gimple cand_stmt;
167
168 /* The base expression B: often an SSA name, but not always. */
169 tree base_expr;
170
171 /* The stride S. */
172 tree stride;
173
174 /* The index constant i. */
175 double_int index;
176
177 /* The type of the candidate. This is normally the type of base_expr,
178 but casts may have occurred when combining feeding instructions.
179 A candidate can only be a basis for candidates of the same final type.
180 (For CAND_REFs, this is the type to be used for operand 1 of the
181 replacement MEM_REF.) */
182 tree cand_type;
183
184 /* The kind of candidate (CAND_MULT, etc.). */
185 enum cand_kind kind;
186
187 /* Index of this candidate in the candidate vector. */
188 cand_idx cand_num;
189
190 /* Index of the next candidate record for the same statement.
191 A statement may be useful in more than one way (e.g., due to
192 commutativity). So we can have multiple "interpretations"
193 of a statement. */
194 cand_idx next_interp;
195
196 /* Index of the basis statement S0, if any, in the candidate vector. */
197 cand_idx basis;
198
199 /* First candidate for which this candidate is a basis, if one exists. */
200 cand_idx dependent;
201
202 /* Next candidate having the same basis as this one. */
203 cand_idx sibling;
204
205 /* If this is a conditional candidate, the defining PHI statement
206 for the base SSA name B. For future use; always NULL for now. */
207 gimple def_phi;
208
209 /* Savings that can be expected from eliminating dead code if this
210 candidate is replaced. */
211 int dead_savings;
212 };
213
214 typedef struct slsr_cand_d slsr_cand, *slsr_cand_t;
215 typedef const struct slsr_cand_d *const_slsr_cand_t;
216
217 /* Pointers to candidates are chained together as part of a mapping
218 from base expressions to the candidates that use them. */
219
220 struct cand_chain_d
221 {
222 /* Base expression for the chain of candidates: often, but not
223 always, an SSA name. */
224 tree base_expr;
225
226 /* Pointer to a candidate. */
227 slsr_cand_t cand;
228
229 /* Chain pointer. */
230 struct cand_chain_d *next;
231
232 };
233
234 typedef struct cand_chain_d cand_chain, *cand_chain_t;
235 typedef const struct cand_chain_d *const_cand_chain_t;
236
237 /* Information about a unique "increment" associated with candidates
238 having an SSA name for a stride. An increment is the difference
239 between the index of the candidate and the index of its basis,
240 i.e., (i - i') as discussed in the module commentary.
241
242 When we are not going to generate address arithmetic we treat
243 increments that differ only in sign as the same, allowing sharing
244 of the cost of initializers. The absolute value of the increment
245 is stored in the incr_info. */
246
247 struct incr_info_d
248 {
249 /* The increment that relates a candidate to its basis. */
250 double_int incr;
251
252 /* How many times the increment occurs in the candidate tree. */
253 unsigned count;
254
255 /* Cost of replacing candidates using this increment. Negative and
256 zero costs indicate replacement should be performed. */
257 int cost;
258
259 /* If this increment is profitable but is not -1, 0, or 1, it requires
260 an initializer T_0 = stride * incr to be found or introduced in the
261 nearest common dominator of all candidates. This field holds T_0
262 for subsequent use. */
263 tree initializer;
264
265 /* If the initializer was found to already exist, this is the block
266 where it was found. */
267 basic_block init_bb;
268 };
269
270 typedef struct incr_info_d incr_info, *incr_info_t;
271
272 /* Candidates are maintained in a vector. If candidate X dominates
273 candidate Y, then X appears before Y in the vector; but the
274 converse does not necessarily hold. */
275 DEF_VEC_P (slsr_cand_t);
276 DEF_VEC_ALLOC_P (slsr_cand_t, heap);
277 static VEC (slsr_cand_t, heap) *cand_vec;
278
279 enum cost_consts
280 {
281 COST_NEUTRAL = 0,
282 COST_INFINITE = 1000
283 };
284
285 /* Pointer map embodying a mapping from statements to candidates. */
286 static struct pointer_map_t *stmt_cand_map;
287
288 /* Obstack for candidates. */
289 static struct obstack cand_obstack;
290
291 /* Hash table embodying a mapping from base exprs to chains of candidates. */
292 static htab_t base_cand_map;
293
294 /* Obstack for candidate chains. */
295 static struct obstack chain_obstack;
296
297 /* An array INCR_VEC of incr_infos is used during analysis of related
298 candidates having an SSA name for a stride. INCR_VEC_LEN describes
299 its current length. */
300 static incr_info_t incr_vec;
301 static unsigned incr_vec_len;
302
303 /* For a chain of candidates with unknown stride, indicates whether or not
304 we must generate pointer arithmetic when replacing statements. */
305 static bool address_arithmetic_p;
306 \f
307 /* Produce a pointer to the IDX'th candidate in the candidate vector. */
308
309 static slsr_cand_t
310 lookup_cand (cand_idx idx)
311 {
312 return VEC_index (slsr_cand_t, cand_vec, idx - 1);
313 }
314
315 /* Callback to produce a hash value for a candidate chain header. */
316
317 static hashval_t
318 base_cand_hash (const void *p)
319 {
320 tree base_expr = ((const_cand_chain_t) p)->base_expr;
321 return iterative_hash_expr (base_expr, 0);
322 }
323
324 /* Callback when an element is removed from the hash table.
325 We never remove entries until the entire table is released. */
326
327 static void
328 base_cand_free (void *p ATTRIBUTE_UNUSED)
329 {
330 }
331
332 /* Callback to return true if two candidate chain headers are equal. */
333
334 static int
335 base_cand_eq (const void *p1, const void *p2)
336 {
337 const_cand_chain_t const chain1 = (const_cand_chain_t) p1;
338 const_cand_chain_t const chain2 = (const_cand_chain_t) p2;
339 return operand_equal_p (chain1->base_expr, chain2->base_expr, 0);
340 }
341 \f
342 /* Use the base expr from candidate C to look for possible candidates
343 that can serve as a basis for C. Each potential basis must also
344 appear in a block that dominates the candidate statement and have
345 the same stride and type. If more than one possible basis exists,
346 the one with highest index in the vector is chosen; this will be
347 the most immediately dominating basis. */
348
349 static int
350 find_basis_for_candidate (slsr_cand_t c)
351 {
352 cand_chain mapping_key;
353 cand_chain_t chain;
354 slsr_cand_t basis = NULL;
355
356 mapping_key.base_expr = c->base_expr;
357 chain = (cand_chain_t) htab_find (base_cand_map, &mapping_key);
358
359 for (; chain; chain = chain->next)
360 {
361 slsr_cand_t one_basis = chain->cand;
362
363 if (one_basis->kind != c->kind
364 || !operand_equal_p (one_basis->stride, c->stride, 0)
365 || !types_compatible_p (one_basis->cand_type, c->cand_type)
366 || !dominated_by_p (CDI_DOMINATORS,
367 gimple_bb (c->cand_stmt),
368 gimple_bb (one_basis->cand_stmt)))
369 continue;
370
371 if (!basis || basis->cand_num < one_basis->cand_num)
372 basis = one_basis;
373 }
374
375 if (basis)
376 {
377 c->sibling = basis->dependent;
378 basis->dependent = c->cand_num;
379 return basis->cand_num;
380 }
381
382 return 0;
383 }
384
385 /* Record a mapping from the base expression of C to C itself, indicating that
386 C may potentially serve as a basis using that base expression. */
387
388 static void
389 record_potential_basis (slsr_cand_t c)
390 {
391 cand_chain_t node;
392 void **slot;
393
394 node = (cand_chain_t) obstack_alloc (&chain_obstack, sizeof (cand_chain));
395 node->base_expr = c->base_expr;
396 node->cand = c;
397 node->next = NULL;
398 slot = htab_find_slot (base_cand_map, node, INSERT);
399
400 if (*slot)
401 {
402 cand_chain_t head = (cand_chain_t) (*slot);
403 node->next = head->next;
404 head->next = node;
405 }
406 else
407 *slot = node;
408 }
409
410 /* Allocate storage for a new candidate and initialize its fields.
411 Attempt to find a basis for the candidate. */
412
413 static slsr_cand_t
414 alloc_cand_and_find_basis (enum cand_kind kind, gimple gs, tree base,
415 double_int index, tree stride, tree ctype,
416 unsigned savings)
417 {
418 slsr_cand_t c = (slsr_cand_t) obstack_alloc (&cand_obstack,
419 sizeof (slsr_cand));
420 c->cand_stmt = gs;
421 c->base_expr = base;
422 c->stride = stride;
423 c->index = index;
424 c->cand_type = ctype;
425 c->kind = kind;
426 c->cand_num = VEC_length (slsr_cand_t, cand_vec) + 1;
427 c->next_interp = 0;
428 c->dependent = 0;
429 c->sibling = 0;
430 c->def_phi = NULL;
431 c->dead_savings = savings;
432
433 VEC_safe_push (slsr_cand_t, heap, cand_vec, c);
434 c->basis = find_basis_for_candidate (c);
435 record_potential_basis (c);
436
437 return c;
438 }
439
440 /* Determine the target cost of statement GS when compiling according
441 to SPEED. */
442
443 static int
444 stmt_cost (gimple gs, bool speed)
445 {
446 tree lhs, rhs1, rhs2;
447 enum machine_mode lhs_mode;
448
449 gcc_assert (is_gimple_assign (gs));
450 lhs = gimple_assign_lhs (gs);
451 rhs1 = gimple_assign_rhs1 (gs);
452 lhs_mode = TYPE_MODE (TREE_TYPE (lhs));
453
454 switch (gimple_assign_rhs_code (gs))
455 {
456 case MULT_EXPR:
457 rhs2 = gimple_assign_rhs2 (gs);
458
459 if (host_integerp (rhs2, 0))
460 return mult_by_coeff_cost (TREE_INT_CST_LOW (rhs2), lhs_mode, speed);
461
462 gcc_assert (TREE_CODE (rhs1) != INTEGER_CST);
463 return mul_cost (speed, lhs_mode);
464
465 case PLUS_EXPR:
466 case POINTER_PLUS_EXPR:
467 case MINUS_EXPR:
468 return add_cost (speed, lhs_mode);
469
470 case NEGATE_EXPR:
471 return neg_cost (speed, lhs_mode);
472
473 case NOP_EXPR:
474 return convert_cost (lhs_mode, TYPE_MODE (TREE_TYPE (rhs1)), speed);
475
476 /* Note that we don't assign costs to copies that in most cases
477 will go away. */
478 default:
479 ;
480 }
481
482 gcc_unreachable ();
483 return 0;
484 }
485
486 /* Look up the defining statement for BASE_IN and return a pointer
487 to its candidate in the candidate table, if any; otherwise NULL.
488 Only CAND_ADD and CAND_MULT candidates are returned. */
489
490 static slsr_cand_t
491 base_cand_from_table (tree base_in)
492 {
493 slsr_cand_t *result;
494
495 gimple def = SSA_NAME_DEF_STMT (base_in);
496 if (!def)
497 return (slsr_cand_t) NULL;
498
499 result = (slsr_cand_t *) pointer_map_contains (stmt_cand_map, def);
500
501 if (result && (*result)->kind != CAND_REF)
502 return *result;
503
504 return (slsr_cand_t) NULL;
505 }
506
507 /* Add an entry to the statement-to-candidate mapping. */
508
509 static void
510 add_cand_for_stmt (gimple gs, slsr_cand_t c)
511 {
512 void **slot = pointer_map_insert (stmt_cand_map, gs);
513 gcc_assert (!*slot);
514 *slot = c;
515 }
516 \f
517 /* Look for the following pattern:
518
519 *PBASE: MEM_REF (T1, C1)
520
521 *POFFSET: MULT_EXPR (T2, C3) [C2 is zero]
522 or
523 MULT_EXPR (PLUS_EXPR (T2, C2), C3)
524 or
525 MULT_EXPR (MINUS_EXPR (T2, -C2), C3)
526
527 *PINDEX: C4 * BITS_PER_UNIT
528
529 If not present, leave the input values unchanged and return FALSE.
530 Otherwise, modify the input values as follows and return TRUE:
531
532 *PBASE: T1
533 *POFFSET: MULT_EXPR (T2, C3)
534 *PINDEX: C1 + (C2 * C3) + C4 */
535
536 static bool
537 restructure_reference (tree *pbase, tree *poffset, double_int *pindex,
538 tree *ptype)
539 {
540 tree base = *pbase, offset = *poffset;
541 double_int index = *pindex;
542 double_int bpu = uhwi_to_double_int (BITS_PER_UNIT);
543 tree mult_op0, mult_op1, t1, t2, type;
544 double_int c1, c2, c3, c4;
545
546 if (!base
547 || !offset
548 || TREE_CODE (base) != MEM_REF
549 || TREE_CODE (offset) != MULT_EXPR
550 || TREE_CODE (TREE_OPERAND (offset, 1)) != INTEGER_CST
551 || !double_int_zero_p (double_int_umod (index, bpu, FLOOR_MOD_EXPR)))
552 return false;
553
554 t1 = TREE_OPERAND (base, 0);
555 c1 = mem_ref_offset (base);
556 type = TREE_TYPE (TREE_OPERAND (base, 1));
557
558 mult_op0 = TREE_OPERAND (offset, 0);
559 mult_op1 = TREE_OPERAND (offset, 1);
560
561 c3 = tree_to_double_int (mult_op1);
562
563 if (TREE_CODE (mult_op0) == PLUS_EXPR)
564
565 if (TREE_CODE (TREE_OPERAND (mult_op0, 1)) == INTEGER_CST)
566 {
567 t2 = TREE_OPERAND (mult_op0, 0);
568 c2 = tree_to_double_int (TREE_OPERAND (mult_op0, 1));
569 }
570 else
571 return false;
572
573 else if (TREE_CODE (mult_op0) == MINUS_EXPR)
574
575 if (TREE_CODE (TREE_OPERAND (mult_op0, 1)) == INTEGER_CST)
576 {
577 t2 = TREE_OPERAND (mult_op0, 0);
578 c2 = double_int_neg (tree_to_double_int (TREE_OPERAND (mult_op0, 1)));
579 }
580 else
581 return false;
582
583 else
584 {
585 t2 = mult_op0;
586 c2 = double_int_zero;
587 }
588
589 c4 = double_int_udiv (index, bpu, FLOOR_DIV_EXPR);
590
591 *pbase = t1;
592 *poffset = fold_build2 (MULT_EXPR, sizetype, t2,
593 double_int_to_tree (sizetype, c3));
594 *pindex = double_int_add (double_int_add (c1, double_int_mul (c2, c3)), c4);
595 *ptype = type;
596
597 return true;
598 }
599
600 /* Given GS which contains a data reference, create a CAND_REF entry in
601 the candidate table and attempt to find a basis. */
602
603 static void
604 slsr_process_ref (gimple gs)
605 {
606 tree ref_expr, base, offset, type;
607 HOST_WIDE_INT bitsize, bitpos;
608 enum machine_mode mode;
609 int unsignedp, volatilep;
610 double_int index;
611 slsr_cand_t c;
612
613 if (gimple_vdef (gs))
614 ref_expr = gimple_assign_lhs (gs);
615 else
616 ref_expr = gimple_assign_rhs1 (gs);
617
618 if (!handled_component_p (ref_expr)
619 || TREE_CODE (ref_expr) == BIT_FIELD_REF
620 || (TREE_CODE (ref_expr) == COMPONENT_REF
621 && DECL_BIT_FIELD (TREE_OPERAND (ref_expr, 1))))
622 return;
623
624 base = get_inner_reference (ref_expr, &bitsize, &bitpos, &offset, &mode,
625 &unsignedp, &volatilep, false);
626 index = uhwi_to_double_int (bitpos);
627
628 if (!restructure_reference (&base, &offset, &index, &type))
629 return;
630
631 c = alloc_cand_and_find_basis (CAND_REF, gs, base, index, offset,
632 type, 0);
633
634 /* Add the candidate to the statement-candidate mapping. */
635 add_cand_for_stmt (gs, c);
636 }
637
638 /* Create a candidate entry for a statement GS, where GS multiplies
639 two SSA names BASE_IN and STRIDE_IN. Propagate any known information
640 about the two SSA names into the new candidate. Return the new
641 candidate. */
642
643 static slsr_cand_t
644 create_mul_ssa_cand (gimple gs, tree base_in, tree stride_in, bool speed)
645 {
646 tree base = NULL_TREE, stride = NULL_TREE, ctype = NULL_TREE;
647 double_int index;
648 unsigned savings = 0;
649 slsr_cand_t c;
650 slsr_cand_t base_cand = base_cand_from_table (base_in);
651
652 /* Look at all interpretations of the base candidate, if necessary,
653 to find information to propagate into this candidate. */
654 while (base_cand && !base)
655 {
656
657 if (base_cand->kind == CAND_MULT
658 && operand_equal_p (base_cand->stride, integer_one_node, 0))
659 {
660 /* Y = (B + i') * 1
661 X = Y * Z
662 ================
663 X = (B + i') * Z */
664 base = base_cand->base_expr;
665 index = base_cand->index;
666 stride = stride_in;
667 ctype = base_cand->cand_type;
668 if (has_single_use (base_in))
669 savings = (base_cand->dead_savings
670 + stmt_cost (base_cand->cand_stmt, speed));
671 }
672 else if (base_cand->kind == CAND_ADD
673 && TREE_CODE (base_cand->stride) == INTEGER_CST)
674 {
675 /* Y = B + (i' * S), S constant
676 X = Y * Z
677 ============================
678 X = B + ((i' * S) * Z) */
679 base = base_cand->base_expr;
680 index = double_int_mul (base_cand->index,
681 tree_to_double_int (base_cand->stride));
682 stride = stride_in;
683 ctype = base_cand->cand_type;
684 if (has_single_use (base_in))
685 savings = (base_cand->dead_savings
686 + stmt_cost (base_cand->cand_stmt, speed));
687 }
688
689 if (base_cand->next_interp)
690 base_cand = lookup_cand (base_cand->next_interp);
691 else
692 base_cand = NULL;
693 }
694
695 if (!base)
696 {
697 /* No interpretations had anything useful to propagate, so
698 produce X = (Y + 0) * Z. */
699 base = base_in;
700 index = double_int_zero;
701 stride = stride_in;
702 ctype = TREE_TYPE (base_in);
703 }
704
705 c = alloc_cand_and_find_basis (CAND_MULT, gs, base, index, stride,
706 ctype, savings);
707 return c;
708 }
709
710 /* Create a candidate entry for a statement GS, where GS multiplies
711 SSA name BASE_IN by constant STRIDE_IN. Propagate any known
712 information about BASE_IN into the new candidate. Return the new
713 candidate. */
714
715 static slsr_cand_t
716 create_mul_imm_cand (gimple gs, tree base_in, tree stride_in, bool speed)
717 {
718 tree base = NULL_TREE, stride = NULL_TREE, ctype = NULL_TREE;
719 double_int index, temp;
720 unsigned savings = 0;
721 slsr_cand_t c;
722 slsr_cand_t base_cand = base_cand_from_table (base_in);
723
724 /* Look at all interpretations of the base candidate, if necessary,
725 to find information to propagate into this candidate. */
726 while (base_cand && !base)
727 {
728 if (base_cand->kind == CAND_MULT
729 && TREE_CODE (base_cand->stride) == INTEGER_CST)
730 {
731 /* Y = (B + i') * S, S constant
732 X = Y * c
733 ============================
734 X = (B + i') * (S * c) */
735 base = base_cand->base_expr;
736 index = base_cand->index;
737 temp = double_int_mul (tree_to_double_int (base_cand->stride),
738 tree_to_double_int (stride_in));
739 stride = double_int_to_tree (TREE_TYPE (stride_in), temp);
740 ctype = base_cand->cand_type;
741 if (has_single_use (base_in))
742 savings = (base_cand->dead_savings
743 + stmt_cost (base_cand->cand_stmt, speed));
744 }
745 else if (base_cand->kind == CAND_ADD
746 && operand_equal_p (base_cand->stride, integer_one_node, 0))
747 {
748 /* Y = B + (i' * 1)
749 X = Y * c
750 ===========================
751 X = (B + i') * c */
752 base = base_cand->base_expr;
753 index = base_cand->index;
754 stride = stride_in;
755 ctype = base_cand->cand_type;
756 if (has_single_use (base_in))
757 savings = (base_cand->dead_savings
758 + stmt_cost (base_cand->cand_stmt, speed));
759 }
760 else if (base_cand->kind == CAND_ADD
761 && double_int_one_p (base_cand->index)
762 && TREE_CODE (base_cand->stride) == INTEGER_CST)
763 {
764 /* Y = B + (1 * S), S constant
765 X = Y * c
766 ===========================
767 X = (B + S) * c */
768 base = base_cand->base_expr;
769 index = tree_to_double_int (base_cand->stride);
770 stride = stride_in;
771 ctype = base_cand->cand_type;
772 if (has_single_use (base_in))
773 savings = (base_cand->dead_savings
774 + stmt_cost (base_cand->cand_stmt, speed));
775 }
776
777 if (base_cand->next_interp)
778 base_cand = lookup_cand (base_cand->next_interp);
779 else
780 base_cand = NULL;
781 }
782
783 if (!base)
784 {
785 /* No interpretations had anything useful to propagate, so
786 produce X = (Y + 0) * c. */
787 base = base_in;
788 index = double_int_zero;
789 stride = stride_in;
790 ctype = TREE_TYPE (base_in);
791 }
792
793 c = alloc_cand_and_find_basis (CAND_MULT, gs, base, index, stride,
794 ctype, savings);
795 return c;
796 }
797
798 /* Given GS which is a multiply of scalar integers, make an appropriate
799 entry in the candidate table. If this is a multiply of two SSA names,
800 create two CAND_MULT interpretations and attempt to find a basis for
801 each of them. Otherwise, create a single CAND_MULT and attempt to
802 find a basis. */
803
804 static void
805 slsr_process_mul (gimple gs, tree rhs1, tree rhs2, bool speed)
806 {
807 slsr_cand_t c, c2;
808
809 /* If this is a multiply of an SSA name with itself, it is highly
810 unlikely that we will get a strength reduction opportunity, so
811 don't record it as a candidate. This simplifies the logic for
812 finding a basis, so if this is removed that must be considered. */
813 if (rhs1 == rhs2)
814 return;
815
816 if (TREE_CODE (rhs2) == SSA_NAME)
817 {
818 /* Record an interpretation of this statement in the candidate table
819 assuming RHS1 is the base expression and RHS2 is the stride. */
820 c = create_mul_ssa_cand (gs, rhs1, rhs2, speed);
821
822 /* Add the first interpretation to the statement-candidate mapping. */
823 add_cand_for_stmt (gs, c);
824
825 /* Record another interpretation of this statement assuming RHS1
826 is the stride and RHS2 is the base expression. */
827 c2 = create_mul_ssa_cand (gs, rhs2, rhs1, speed);
828 c->next_interp = c2->cand_num;
829 }
830 else
831 {
832 /* Record an interpretation for the multiply-immediate. */
833 c = create_mul_imm_cand (gs, rhs1, rhs2, speed);
834
835 /* Add the interpretation to the statement-candidate mapping. */
836 add_cand_for_stmt (gs, c);
837 }
838 }
839
840 /* Create a candidate entry for a statement GS, where GS adds two
841 SSA names BASE_IN and ADDEND_IN if SUBTRACT_P is false, and
842 subtracts ADDEND_IN from BASE_IN otherwise. Propagate any known
843 information about the two SSA names into the new candidate.
844 Return the new candidate. */
845
846 static slsr_cand_t
847 create_add_ssa_cand (gimple gs, tree base_in, tree addend_in,
848 bool subtract_p, bool speed)
849 {
850 tree base = NULL_TREE, stride = NULL_TREE, ctype = NULL;
851 double_int index;
852 unsigned savings = 0;
853 slsr_cand_t c;
854 slsr_cand_t base_cand = base_cand_from_table (base_in);
855 slsr_cand_t addend_cand = base_cand_from_table (addend_in);
856
857 /* The most useful transformation is a multiply-immediate feeding
858 an add or subtract. Look for that first. */
859 while (addend_cand && !base)
860 {
861 if (addend_cand->kind == CAND_MULT
862 && double_int_zero_p (addend_cand->index)
863 && TREE_CODE (addend_cand->stride) == INTEGER_CST)
864 {
865 /* Z = (B + 0) * S, S constant
866 X = Y +/- Z
867 ===========================
868 X = Y + ((+/-1 * S) * B) */
869 base = base_in;
870 index = tree_to_double_int (addend_cand->stride);
871 if (subtract_p)
872 index = double_int_neg (index);
873 stride = addend_cand->base_expr;
874 ctype = TREE_TYPE (base_in);
875 if (has_single_use (addend_in))
876 savings = (addend_cand->dead_savings
877 + stmt_cost (addend_cand->cand_stmt, speed));
878 }
879
880 if (addend_cand->next_interp)
881 addend_cand = lookup_cand (addend_cand->next_interp);
882 else
883 addend_cand = NULL;
884 }
885
886 while (base_cand && !base)
887 {
888 if (base_cand->kind == CAND_ADD
889 && (double_int_zero_p (base_cand->index)
890 || operand_equal_p (base_cand->stride,
891 integer_zero_node, 0)))
892 {
893 /* Y = B + (i' * S), i' * S = 0
894 X = Y +/- Z
895 ============================
896 X = B + (+/-1 * Z) */
897 base = base_cand->base_expr;
898 index = subtract_p ? double_int_minus_one : double_int_one;
899 stride = addend_in;
900 ctype = base_cand->cand_type;
901 if (has_single_use (base_in))
902 savings = (base_cand->dead_savings
903 + stmt_cost (base_cand->cand_stmt, speed));
904 }
905 else if (subtract_p)
906 {
907 slsr_cand_t subtrahend_cand = base_cand_from_table (addend_in);
908
909 while (subtrahend_cand && !base)
910 {
911 if (subtrahend_cand->kind == CAND_MULT
912 && double_int_zero_p (subtrahend_cand->index)
913 && TREE_CODE (subtrahend_cand->stride) == INTEGER_CST)
914 {
915 /* Z = (B + 0) * S, S constant
916 X = Y - Z
917 ===========================
918 Value: X = Y + ((-1 * S) * B) */
919 base = base_in;
920 index = tree_to_double_int (subtrahend_cand->stride);
921 index = double_int_neg (index);
922 stride = subtrahend_cand->base_expr;
923 ctype = TREE_TYPE (base_in);
924 if (has_single_use (addend_in))
925 savings = (subtrahend_cand->dead_savings
926 + stmt_cost (subtrahend_cand->cand_stmt, speed));
927 }
928
929 if (subtrahend_cand->next_interp)
930 subtrahend_cand = lookup_cand (subtrahend_cand->next_interp);
931 else
932 subtrahend_cand = NULL;
933 }
934 }
935
936 if (base_cand->next_interp)
937 base_cand = lookup_cand (base_cand->next_interp);
938 else
939 base_cand = NULL;
940 }
941
942 if (!base)
943 {
944 /* No interpretations had anything useful to propagate, so
945 produce X = Y + (1 * Z). */
946 base = base_in;
947 index = subtract_p ? double_int_minus_one : double_int_one;
948 stride = addend_in;
949 ctype = TREE_TYPE (base_in);
950 }
951
952 c = alloc_cand_and_find_basis (CAND_ADD, gs, base, index, stride,
953 ctype, savings);
954 return c;
955 }
956
957 /* Create a candidate entry for a statement GS, where GS adds SSA
958 name BASE_IN to constant INDEX_IN. Propagate any known information
959 about BASE_IN into the new candidate. Return the new candidate. */
960
961 static slsr_cand_t
962 create_add_imm_cand (gimple gs, tree base_in, double_int index_in, bool speed)
963 {
964 enum cand_kind kind = CAND_ADD;
965 tree base = NULL_TREE, stride = NULL_TREE, ctype = NULL_TREE;
966 double_int index, multiple;
967 unsigned savings = 0;
968 slsr_cand_t c;
969 slsr_cand_t base_cand = base_cand_from_table (base_in);
970
971 while (base_cand && !base)
972 {
973 bool unsigned_p = TYPE_UNSIGNED (TREE_TYPE (base_cand->stride));
974
975 if (TREE_CODE (base_cand->stride) == INTEGER_CST
976 && double_int_multiple_of (index_in,
977 tree_to_double_int (base_cand->stride),
978 unsigned_p,
979 &multiple))
980 {
981 /* Y = (B + i') * S, S constant, c = kS for some integer k
982 X = Y + c
983 ============================
984 X = (B + (i'+ k)) * S
985 OR
986 Y = B + (i' * S), S constant, c = kS for some integer k
987 X = Y + c
988 ============================
989 X = (B + (i'+ k)) * S */
990 kind = base_cand->kind;
991 base = base_cand->base_expr;
992 index = double_int_add (base_cand->index, multiple);
993 stride = base_cand->stride;
994 ctype = base_cand->cand_type;
995 if (has_single_use (base_in))
996 savings = (base_cand->dead_savings
997 + stmt_cost (base_cand->cand_stmt, speed));
998 }
999
1000 if (base_cand->next_interp)
1001 base_cand = lookup_cand (base_cand->next_interp);
1002 else
1003 base_cand = NULL;
1004 }
1005
1006 if (!base)
1007 {
1008 /* No interpretations had anything useful to propagate, so
1009 produce X = Y + (c * 1). */
1010 kind = CAND_ADD;
1011 base = base_in;
1012 index = index_in;
1013 stride = integer_one_node;
1014 ctype = TREE_TYPE (base_in);
1015 }
1016
1017 c = alloc_cand_and_find_basis (kind, gs, base, index, stride,
1018 ctype, savings);
1019 return c;
1020 }
1021
1022 /* Given GS which is an add or subtract of scalar integers or pointers,
1023 make at least one appropriate entry in the candidate table. */
1024
1025 static void
1026 slsr_process_add (gimple gs, tree rhs1, tree rhs2, bool speed)
1027 {
1028 bool subtract_p = gimple_assign_rhs_code (gs) == MINUS_EXPR;
1029 slsr_cand_t c = NULL, c2;
1030
1031 if (TREE_CODE (rhs2) == SSA_NAME)
1032 {
1033 /* First record an interpretation assuming RHS1 is the base expression
1034 and RHS2 is the stride. But it doesn't make sense for the
1035 stride to be a pointer, so don't record a candidate in that case. */
1036 if (!POINTER_TYPE_P (TREE_TYPE (rhs2)))
1037 {
1038 c = create_add_ssa_cand (gs, rhs1, rhs2, subtract_p, speed);
1039
1040 /* Add the first interpretation to the statement-candidate
1041 mapping. */
1042 add_cand_for_stmt (gs, c);
1043 }
1044
1045 /* If the two RHS operands are identical, or this is a subtract,
1046 we're done. */
1047 if (operand_equal_p (rhs1, rhs2, 0) || subtract_p)
1048 return;
1049
1050 /* Otherwise, record another interpretation assuming RHS2 is the
1051 base expression and RHS1 is the stride, again provided that the
1052 stride is not a pointer. */
1053 if (!POINTER_TYPE_P (TREE_TYPE (rhs1)))
1054 {
1055 c2 = create_add_ssa_cand (gs, rhs2, rhs1, false, speed);
1056 if (c)
1057 c->next_interp = c2->cand_num;
1058 else
1059 add_cand_for_stmt (gs, c2);
1060 }
1061 }
1062 else
1063 {
1064 double_int index;
1065
1066 /* Record an interpretation for the add-immediate. */
1067 index = tree_to_double_int (rhs2);
1068 if (subtract_p)
1069 index = double_int_neg (index);
1070
1071 c = create_add_imm_cand (gs, rhs1, index, speed);
1072
1073 /* Add the interpretation to the statement-candidate mapping. */
1074 add_cand_for_stmt (gs, c);
1075 }
1076 }
1077
1078 /* Given GS which is a negate of a scalar integer, make an appropriate
1079 entry in the candidate table. A negate is equivalent to a multiply
1080 by -1. */
1081
1082 static void
1083 slsr_process_neg (gimple gs, tree rhs1, bool speed)
1084 {
1085 /* Record a CAND_MULT interpretation for the multiply by -1. */
1086 slsr_cand_t c = create_mul_imm_cand (gs, rhs1, integer_minus_one_node, speed);
1087
1088 /* Add the interpretation to the statement-candidate mapping. */
1089 add_cand_for_stmt (gs, c);
1090 }
1091
1092 /* Help function for legal_cast_p, operating on two trees. Checks
1093 whether it's allowable to cast from RHS to LHS. See legal_cast_p
1094 for more details. */
1095
1096 static bool
1097 legal_cast_p_1 (tree lhs, tree rhs)
1098 {
1099 tree lhs_type, rhs_type;
1100 unsigned lhs_size, rhs_size;
1101 bool lhs_wraps, rhs_wraps;
1102
1103 lhs_type = TREE_TYPE (lhs);
1104 rhs_type = TREE_TYPE (rhs);
1105 lhs_size = TYPE_PRECISION (lhs_type);
1106 rhs_size = TYPE_PRECISION (rhs_type);
1107 lhs_wraps = TYPE_OVERFLOW_WRAPS (lhs_type);
1108 rhs_wraps = TYPE_OVERFLOW_WRAPS (rhs_type);
1109
1110 if (lhs_size < rhs_size
1111 || (rhs_wraps && !lhs_wraps)
1112 || (rhs_wraps && lhs_wraps && rhs_size != lhs_size))
1113 return false;
1114
1115 return true;
1116 }
1117
1118 /* Return TRUE if GS is a statement that defines an SSA name from
1119 a conversion and is legal for us to combine with an add and multiply
1120 in the candidate table. For example, suppose we have:
1121
1122 A = B + i;
1123 C = (type) A;
1124 D = C * S;
1125
1126 Without the type-cast, we would create a CAND_MULT for D with base B,
1127 index i, and stride S. We want to record this candidate only if it
1128 is equivalent to apply the type cast following the multiply:
1129
1130 A = B + i;
1131 E = A * S;
1132 D = (type) E;
1133
1134 We will record the type with the candidate for D. This allows us
1135 to use a similar previous candidate as a basis. If we have earlier seen
1136
1137 A' = B + i';
1138 C' = (type) A';
1139 D' = C' * S;
1140
1141 we can replace D with
1142
1143 D = D' + (i - i') * S;
1144
1145 But if moving the type-cast would change semantics, we mustn't do this.
1146
1147 This is legitimate for casts from a non-wrapping integral type to
1148 any integral type of the same or larger size. It is not legitimate
1149 to convert a wrapping type to a non-wrapping type, or to a wrapping
1150 type of a different size. I.e., with a wrapping type, we must
1151 assume that the addition B + i could wrap, in which case performing
1152 the multiply before or after one of the "illegal" type casts will
1153 have different semantics. */
1154
1155 static bool
1156 legal_cast_p (gimple gs, tree rhs)
1157 {
1158 if (!is_gimple_assign (gs)
1159 || !CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (gs)))
1160 return false;
1161
1162 return legal_cast_p_1 (gimple_assign_lhs (gs), rhs);
1163 }
1164
1165 /* Given GS which is a cast to a scalar integer type, determine whether
1166 the cast is legal for strength reduction. If so, make at least one
1167 appropriate entry in the candidate table. */
1168
1169 static void
1170 slsr_process_cast (gimple gs, tree rhs1, bool speed)
1171 {
1172 tree lhs, ctype;
1173 slsr_cand_t base_cand, c, c2;
1174 unsigned savings = 0;
1175
1176 if (!legal_cast_p (gs, rhs1))
1177 return;
1178
1179 lhs = gimple_assign_lhs (gs);
1180 base_cand = base_cand_from_table (rhs1);
1181 ctype = TREE_TYPE (lhs);
1182
1183 if (base_cand)
1184 {
1185 while (base_cand)
1186 {
1187 /* Propagate all data from the base candidate except the type,
1188 which comes from the cast, and the base candidate's cast,
1189 which is no longer applicable. */
1190 if (has_single_use (rhs1))
1191 savings = (base_cand->dead_savings
1192 + stmt_cost (base_cand->cand_stmt, speed));
1193
1194 c = alloc_cand_and_find_basis (base_cand->kind, gs,
1195 base_cand->base_expr,
1196 base_cand->index, base_cand->stride,
1197 ctype, savings);
1198 if (base_cand->next_interp)
1199 base_cand = lookup_cand (base_cand->next_interp);
1200 else
1201 base_cand = NULL;
1202 }
1203 }
1204 else
1205 {
1206 /* If nothing is known about the RHS, create fresh CAND_ADD and
1207 CAND_MULT interpretations:
1208
1209 X = Y + (0 * 1)
1210 X = (Y + 0) * 1
1211
1212 The first of these is somewhat arbitrary, but the choice of
1213 1 for the stride simplifies the logic for propagating casts
1214 into their uses. */
1215 c = alloc_cand_and_find_basis (CAND_ADD, gs, rhs1, double_int_zero,
1216 integer_one_node, ctype, 0);
1217 c2 = alloc_cand_and_find_basis (CAND_MULT, gs, rhs1, double_int_zero,
1218 integer_one_node, ctype, 0);
1219 c->next_interp = c2->cand_num;
1220 }
1221
1222 /* Add the first (or only) interpretation to the statement-candidate
1223 mapping. */
1224 add_cand_for_stmt (gs, c);
1225 }
1226
1227 /* Given GS which is a copy of a scalar integer type, make at least one
1228 appropriate entry in the candidate table.
1229
1230 This interface is included for completeness, but is unnecessary
1231 if this pass immediately follows a pass that performs copy
1232 propagation, such as DOM. */
1233
1234 static void
1235 slsr_process_copy (gimple gs, tree rhs1, bool speed)
1236 {
1237 slsr_cand_t base_cand, c, c2;
1238 unsigned savings = 0;
1239
1240 base_cand = base_cand_from_table (rhs1);
1241
1242 if (base_cand)
1243 {
1244 while (base_cand)
1245 {
1246 /* Propagate all data from the base candidate. */
1247 if (has_single_use (rhs1))
1248 savings = (base_cand->dead_savings
1249 + stmt_cost (base_cand->cand_stmt, speed));
1250
1251 c = alloc_cand_and_find_basis (base_cand->kind, gs,
1252 base_cand->base_expr,
1253 base_cand->index, base_cand->stride,
1254 base_cand->cand_type, savings);
1255 if (base_cand->next_interp)
1256 base_cand = lookup_cand (base_cand->next_interp);
1257 else
1258 base_cand = NULL;
1259 }
1260 }
1261 else
1262 {
1263 /* If nothing is known about the RHS, create fresh CAND_ADD and
1264 CAND_MULT interpretations:
1265
1266 X = Y + (0 * 1)
1267 X = (Y + 0) * 1
1268
1269 The first of these is somewhat arbitrary, but the choice of
1270 1 for the stride simplifies the logic for propagating casts
1271 into their uses. */
1272 c = alloc_cand_and_find_basis (CAND_ADD, gs, rhs1, double_int_zero,
1273 integer_one_node, TREE_TYPE (rhs1), 0);
1274 c2 = alloc_cand_and_find_basis (CAND_MULT, gs, rhs1, double_int_zero,
1275 integer_one_node, TREE_TYPE (rhs1), 0);
1276 c->next_interp = c2->cand_num;
1277 }
1278
1279 /* Add the first (or only) interpretation to the statement-candidate
1280 mapping. */
1281 add_cand_for_stmt (gs, c);
1282 }
1283 \f
1284 /* Find strength-reduction candidates in block BB. */
1285
1286 static void
1287 find_candidates_in_block (struct dom_walk_data *walk_data ATTRIBUTE_UNUSED,
1288 basic_block bb)
1289 {
1290 bool speed = optimize_bb_for_speed_p (bb);
1291 gimple_stmt_iterator gsi;
1292
1293 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1294 {
1295 gimple gs = gsi_stmt (gsi);
1296
1297 if (gimple_vuse (gs) && gimple_assign_single_p (gs))
1298 slsr_process_ref (gs);
1299
1300 else if (is_gimple_assign (gs)
1301 && SCALAR_INT_MODE_P
1302 (TYPE_MODE (TREE_TYPE (gimple_assign_lhs (gs)))))
1303 {
1304 tree rhs1 = NULL_TREE, rhs2 = NULL_TREE;
1305
1306 switch (gimple_assign_rhs_code (gs))
1307 {
1308 case MULT_EXPR:
1309 case PLUS_EXPR:
1310 rhs1 = gimple_assign_rhs1 (gs);
1311 rhs2 = gimple_assign_rhs2 (gs);
1312 /* Should never happen, but currently some buggy situations
1313 in earlier phases put constants in rhs1. */
1314 if (TREE_CODE (rhs1) != SSA_NAME)
1315 continue;
1316 break;
1317
1318 /* Possible future opportunity: rhs1 of a ptr+ can be
1319 an ADDR_EXPR. */
1320 case POINTER_PLUS_EXPR:
1321 case MINUS_EXPR:
1322 rhs2 = gimple_assign_rhs2 (gs);
1323 /* Fall-through. */
1324
1325 case NOP_EXPR:
1326 case MODIFY_EXPR:
1327 case NEGATE_EXPR:
1328 rhs1 = gimple_assign_rhs1 (gs);
1329 if (TREE_CODE (rhs1) != SSA_NAME)
1330 continue;
1331 break;
1332
1333 default:
1334 ;
1335 }
1336
1337 switch (gimple_assign_rhs_code (gs))
1338 {
1339 case MULT_EXPR:
1340 slsr_process_mul (gs, rhs1, rhs2, speed);
1341 break;
1342
1343 case PLUS_EXPR:
1344 case POINTER_PLUS_EXPR:
1345 case MINUS_EXPR:
1346 slsr_process_add (gs, rhs1, rhs2, speed);
1347 break;
1348
1349 case NEGATE_EXPR:
1350 slsr_process_neg (gs, rhs1, speed);
1351 break;
1352
1353 case NOP_EXPR:
1354 slsr_process_cast (gs, rhs1, speed);
1355 break;
1356
1357 case MODIFY_EXPR:
1358 slsr_process_copy (gs, rhs1, speed);
1359 break;
1360
1361 default:
1362 ;
1363 }
1364 }
1365 }
1366 }
1367 \f
1368 /* Dump a candidate for debug. */
1369
1370 static void
1371 dump_candidate (slsr_cand_t c)
1372 {
1373 fprintf (dump_file, "%3d [%d] ", c->cand_num,
1374 gimple_bb (c->cand_stmt)->index);
1375 print_gimple_stmt (dump_file, c->cand_stmt, 0, 0);
1376 switch (c->kind)
1377 {
1378 case CAND_MULT:
1379 fputs (" MULT : (", dump_file);
1380 print_generic_expr (dump_file, c->base_expr, 0);
1381 fputs (" + ", dump_file);
1382 dump_double_int (dump_file, c->index, false);
1383 fputs (") * ", dump_file);
1384 print_generic_expr (dump_file, c->stride, 0);
1385 fputs (" : ", dump_file);
1386 break;
1387 case CAND_ADD:
1388 fputs (" ADD : ", dump_file);
1389 print_generic_expr (dump_file, c->base_expr, 0);
1390 fputs (" + (", dump_file);
1391 dump_double_int (dump_file, c->index, false);
1392 fputs (" * ", dump_file);
1393 print_generic_expr (dump_file, c->stride, 0);
1394 fputs (") : ", dump_file);
1395 break;
1396 case CAND_REF:
1397 fputs (" REF : ", dump_file);
1398 print_generic_expr (dump_file, c->base_expr, 0);
1399 fputs (" + (", dump_file);
1400 print_generic_expr (dump_file, c->stride, 0);
1401 fputs (") + ", dump_file);
1402 dump_double_int (dump_file, c->index, false);
1403 fputs (" : ", dump_file);
1404 break;
1405 default:
1406 gcc_unreachable ();
1407 }
1408 print_generic_expr (dump_file, c->cand_type, 0);
1409 fprintf (dump_file, "\n basis: %d dependent: %d sibling: %d\n",
1410 c->basis, c->dependent, c->sibling);
1411 fprintf (dump_file, " next-interp: %d dead-savings: %d\n",
1412 c->next_interp, c->dead_savings);
1413 if (c->def_phi)
1414 {
1415 fputs (" phi: ", dump_file);
1416 print_gimple_stmt (dump_file, c->def_phi, 0, 0);
1417 }
1418 fputs ("\n", dump_file);
1419 }
1420
1421 /* Dump the candidate vector for debug. */
1422
1423 static void
1424 dump_cand_vec (void)
1425 {
1426 unsigned i;
1427 slsr_cand_t c;
1428
1429 fprintf (dump_file, "\nStrength reduction candidate vector:\n\n");
1430
1431 FOR_EACH_VEC_ELT (slsr_cand_t, cand_vec, i, c)
1432 dump_candidate (c);
1433 }
1434
1435 /* Callback used to dump the candidate chains hash table. */
1436
1437 static int
1438 base_cand_dump_callback (void **slot, void *ignored ATTRIBUTE_UNUSED)
1439 {
1440 const_cand_chain_t chain = *((const_cand_chain_t *) slot);
1441 cand_chain_t p;
1442
1443 print_generic_expr (dump_file, chain->base_expr, 0);
1444 fprintf (dump_file, " -> %d", chain->cand->cand_num);
1445
1446 for (p = chain->next; p; p = p->next)
1447 fprintf (dump_file, " -> %d", p->cand->cand_num);
1448
1449 fputs ("\n", dump_file);
1450 return 1;
1451 }
1452
1453 /* Dump the candidate chains. */
1454
1455 static void
1456 dump_cand_chains (void)
1457 {
1458 fprintf (dump_file, "\nStrength reduction candidate chains:\n\n");
1459 htab_traverse_noresize (base_cand_map, base_cand_dump_callback, NULL);
1460 fputs ("\n", dump_file);
1461 }
1462
1463 /* Dump the increment vector for debug. */
1464
1465 static void
1466 dump_incr_vec (void)
1467 {
1468 if (dump_file && (dump_flags & TDF_DETAILS))
1469 {
1470 unsigned i;
1471
1472 fprintf (dump_file, "\nIncrement vector:\n\n");
1473
1474 for (i = 0; i < incr_vec_len; i++)
1475 {
1476 fprintf (dump_file, "%3d increment: ", i);
1477 dump_double_int (dump_file, incr_vec[i].incr, false);
1478 fprintf (dump_file, "\n count: %d", incr_vec[i].count);
1479 fprintf (dump_file, "\n cost: %d", incr_vec[i].cost);
1480 fputs ("\n initializer: ", dump_file);
1481 print_generic_expr (dump_file, incr_vec[i].initializer, 0);
1482 fputs ("\n\n", dump_file);
1483 }
1484 }
1485 }
1486 \f
1487 /* Recursive helper for unconditional_cands_with_known_stride_p.
1488 Returns TRUE iff C, its siblings, and its dependents are all
1489 unconditional candidates. */
1490
1491 static bool
1492 unconditional_cands (slsr_cand_t c)
1493 {
1494 if (c->def_phi)
1495 return false;
1496
1497 if (c->sibling && !unconditional_cands (lookup_cand (c->sibling)))
1498 return false;
1499
1500 if (c->dependent && !unconditional_cands (lookup_cand (c->dependent)))
1501 return false;
1502
1503 return true;
1504 }
1505
1506 /* Determine whether or not the tree of candidates rooted at
1507 ROOT consists entirely of unconditional increments with
1508 an INTEGER_CST stride. */
1509
1510 static bool
1511 unconditional_cands_with_known_stride_p (slsr_cand_t root)
1512 {
1513 /* The stride is identical for all related candidates, so
1514 check it once. */
1515 if (TREE_CODE (root->stride) != INTEGER_CST)
1516 return false;
1517
1518 return unconditional_cands (lookup_cand (root->dependent));
1519 }
1520
1521 /* Replace *EXPR in candidate C with an equivalent strength-reduced
1522 data reference. */
1523
1524 static void
1525 replace_ref (tree *expr, slsr_cand_t c)
1526 {
1527 tree add_expr = fold_build2 (POINTER_PLUS_EXPR, TREE_TYPE (c->base_expr),
1528 c->base_expr, c->stride);
1529 tree mem_ref = fold_build2 (MEM_REF, TREE_TYPE (*expr), add_expr,
1530 double_int_to_tree (c->cand_type, c->index));
1531
1532 /* Gimplify the base addressing expression for the new MEM_REF tree. */
1533 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
1534 TREE_OPERAND (mem_ref, 0)
1535 = force_gimple_operand_gsi (&gsi, TREE_OPERAND (mem_ref, 0),
1536 /*simple_p=*/true, NULL,
1537 /*before=*/true, GSI_SAME_STMT);
1538 copy_ref_info (mem_ref, *expr);
1539 *expr = mem_ref;
1540 update_stmt (c->cand_stmt);
1541 }
1542
1543 /* Replace CAND_REF candidate C, each sibling of candidate C, and each
1544 dependent of candidate C with an equivalent strength-reduced data
1545 reference. */
1546
1547 static void
1548 replace_refs (slsr_cand_t c)
1549 {
1550 if (gimple_vdef (c->cand_stmt))
1551 {
1552 tree *lhs = gimple_assign_lhs_ptr (c->cand_stmt);
1553 replace_ref (lhs, c);
1554 }
1555 else
1556 {
1557 tree *rhs = gimple_assign_rhs1_ptr (c->cand_stmt);
1558 replace_ref (rhs, c);
1559 }
1560
1561 if (c->sibling)
1562 replace_refs (lookup_cand (c->sibling));
1563
1564 if (c->dependent)
1565 replace_refs (lookup_cand (c->dependent));
1566 }
1567
1568 /* Calculate the increment required for candidate C relative to
1569 its basis. */
1570
1571 static double_int
1572 cand_increment (slsr_cand_t c)
1573 {
1574 slsr_cand_t basis;
1575
1576 /* If the candidate doesn't have a basis, just return its own
1577 index. This is useful in record_increments to help us find
1578 an existing initializer. */
1579 if (!c->basis)
1580 return c->index;
1581
1582 basis = lookup_cand (c->basis);
1583 gcc_assert (operand_equal_p (c->base_expr, basis->base_expr, 0));
1584 return double_int_sub (c->index, basis->index);
1585 }
1586
1587 /* Calculate the increment required for candidate C relative to
1588 its basis. If we aren't going to generate pointer arithmetic
1589 for this candidate, return the absolute value of that increment
1590 instead. */
1591
1592 static inline double_int
1593 cand_abs_increment (slsr_cand_t c)
1594 {
1595 double_int increment = cand_increment (c);
1596
1597 if (!address_arithmetic_p && double_int_negative_p (increment))
1598 increment = double_int_neg (increment);
1599
1600 return increment;
1601 }
1602
1603 /* If *VAR is NULL or is not of a compatible type with TYPE, create a
1604 new temporary reg of type TYPE and store it in *VAR. */
1605
1606 static inline void
1607 lazy_create_slsr_reg (tree *var, tree type)
1608 {
1609 if (!*var || !types_compatible_p (TREE_TYPE (*var), type))
1610 *var = create_tmp_reg (type, "slsr");
1611 }
1612
1613 /* Return TRUE iff candidate C has already been replaced under
1614 another interpretation. */
1615
1616 static inline bool
1617 cand_already_replaced (slsr_cand_t c)
1618 {
1619 return (gimple_bb (c->cand_stmt) == 0);
1620 }
1621
1622 /* Helper routine for replace_dependents, doing the work for a
1623 single candidate C. */
1624
1625 static void
1626 replace_dependent (slsr_cand_t c, enum tree_code cand_code)
1627 {
1628 double_int stride = tree_to_double_int (c->stride);
1629 double_int bump = double_int_mul (cand_increment (c), stride);
1630 gimple stmt_to_print = NULL;
1631 slsr_cand_t basis;
1632 tree basis_name, incr_type, bump_tree;
1633 enum tree_code code;
1634
1635 /* It is highly unlikely, but possible, that the resulting
1636 bump doesn't fit in a HWI. Abandon the replacement
1637 in this case. Restriction to signed HWI is conservative
1638 for unsigned types but allows for safe negation without
1639 twisted logic. */
1640 if (!double_int_fits_in_shwi_p (bump))
1641 return;
1642
1643 basis = lookup_cand (c->basis);
1644 basis_name = gimple_assign_lhs (basis->cand_stmt);
1645 incr_type = TREE_TYPE (gimple_assign_rhs1 (c->cand_stmt));
1646 code = PLUS_EXPR;
1647
1648 if (double_int_negative_p (bump))
1649 {
1650 code = MINUS_EXPR;
1651 bump = double_int_neg (bump);
1652 }
1653
1654 bump_tree = double_int_to_tree (incr_type, bump);
1655
1656 if (dump_file && (dump_flags & TDF_DETAILS))
1657 {
1658 fputs ("Replacing: ", dump_file);
1659 print_gimple_stmt (dump_file, c->cand_stmt, 0, 0);
1660 }
1661
1662 if (double_int_zero_p (bump))
1663 {
1664 tree lhs = gimple_assign_lhs (c->cand_stmt);
1665 gimple copy_stmt = gimple_build_assign (lhs, basis_name);
1666 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
1667 gimple_set_location (copy_stmt, gimple_location (c->cand_stmt));
1668 gsi_replace (&gsi, copy_stmt, false);
1669 if (dump_file && (dump_flags & TDF_DETAILS))
1670 stmt_to_print = copy_stmt;
1671 }
1672 else
1673 {
1674 tree rhs1 = gimple_assign_rhs1 (c->cand_stmt);
1675 tree rhs2 = gimple_assign_rhs2 (c->cand_stmt);
1676 if (cand_code != NEGATE_EXPR
1677 && ((operand_equal_p (rhs1, basis_name, 0)
1678 && operand_equal_p (rhs2, bump_tree, 0))
1679 || (operand_equal_p (rhs1, bump_tree, 0)
1680 && operand_equal_p (rhs2, basis_name, 0))))
1681 {
1682 if (dump_file && (dump_flags & TDF_DETAILS))
1683 {
1684 fputs ("(duplicate, not actually replacing)", dump_file);
1685 stmt_to_print = c->cand_stmt;
1686 }
1687 }
1688 else
1689 {
1690 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
1691 gimple_assign_set_rhs_with_ops (&gsi, code, basis_name, bump_tree);
1692 update_stmt (gsi_stmt (gsi));
1693 if (dump_file && (dump_flags & TDF_DETAILS))
1694 stmt_to_print = gsi_stmt (gsi);
1695 }
1696 }
1697
1698 if (dump_file && (dump_flags & TDF_DETAILS))
1699 {
1700 fputs ("With: ", dump_file);
1701 print_gimple_stmt (dump_file, stmt_to_print, 0, 0);
1702 fputs ("\n", dump_file);
1703 }
1704 }
1705
1706 /* Replace candidate C, each sibling of candidate C, and each
1707 dependent of candidate C with an add or subtract. Note that we
1708 only operate on CAND_MULTs with known strides, so we will never
1709 generate a POINTER_PLUS_EXPR. Each candidate X = (B + i) * S is
1710 replaced by X = Y + ((i - i') * S), as described in the module
1711 commentary. The folded value ((i - i') * S) is referred to here
1712 as the "bump." */
1713
1714 static void
1715 replace_dependents (slsr_cand_t c)
1716 {
1717 enum tree_code cand_code = gimple_assign_rhs_code (c->cand_stmt);
1718
1719 /* It is not useful to replace casts, copies, or adds of an SSA name
1720 and a constant. Also skip candidates that have already been
1721 replaced under another interpretation. */
1722 if (cand_code != MODIFY_EXPR
1723 && cand_code != NOP_EXPR
1724 && c->kind == CAND_MULT
1725 && !cand_already_replaced (c))
1726 replace_dependent (c, cand_code);
1727
1728 if (c->sibling)
1729 replace_dependents (lookup_cand (c->sibling));
1730
1731 if (c->dependent)
1732 replace_dependents (lookup_cand (c->dependent));
1733 }
1734 \f
1735 /* Return the index in the increment vector of the given INCREMENT. */
1736
1737 static inline unsigned
1738 incr_vec_index (double_int increment)
1739 {
1740 unsigned i;
1741
1742 for (i = 0;
1743 i < incr_vec_len && !double_int_equal_p (increment, incr_vec[i].incr);
1744 i++)
1745 ;
1746
1747 gcc_assert (i < incr_vec_len);
1748 return i;
1749 }
1750
1751 /* Count the number of candidates in the tree rooted at C that have
1752 not already been replaced under other interpretations. */
1753
1754 static unsigned
1755 count_candidates (slsr_cand_t c)
1756 {
1757 unsigned count = cand_already_replaced (c) ? 0 : 1;
1758
1759 if (c->sibling)
1760 count += count_candidates (lookup_cand (c->sibling));
1761
1762 if (c->dependent)
1763 count += count_candidates (lookup_cand (c->dependent));
1764
1765 return count;
1766 }
1767
1768 /* Increase the count of INCREMENT by one in the increment vector.
1769 INCREMENT is associated with candidate C. If an initializer
1770 T_0 = stride * I is provided by a candidate that dominates all
1771 candidates with the same increment, also record T_0 for subsequent use. */
1772
1773 static void
1774 record_increment (slsr_cand_t c, double_int increment)
1775 {
1776 bool found = false;
1777 unsigned i;
1778
1779 /* Treat increments that differ only in sign as identical so as to
1780 share initializers, unless we are generating pointer arithmetic. */
1781 if (!address_arithmetic_p && double_int_negative_p (increment))
1782 increment = double_int_neg (increment);
1783
1784 for (i = 0; i < incr_vec_len; i++)
1785 {
1786 if (double_int_equal_p (incr_vec[i].incr, increment))
1787 {
1788 incr_vec[i].count++;
1789 found = true;
1790
1791 /* If we previously recorded an initializer that doesn't
1792 dominate this candidate, it's not going to be useful to
1793 us after all. */
1794 if (incr_vec[i].initializer
1795 && !dominated_by_p (CDI_DOMINATORS,
1796 gimple_bb (c->cand_stmt),
1797 incr_vec[i].init_bb))
1798 {
1799 incr_vec[i].initializer = NULL_TREE;
1800 incr_vec[i].init_bb = NULL;
1801 }
1802
1803 break;
1804 }
1805 }
1806
1807 if (!found)
1808 {
1809 /* The first time we see an increment, create the entry for it.
1810 If this is the root candidate which doesn't have a basis, set
1811 the count to zero. We're only processing it so it can possibly
1812 provide an initializer for other candidates. */
1813 incr_vec[incr_vec_len].incr = increment;
1814 incr_vec[incr_vec_len].count = c->basis ? 1 : 0;
1815 incr_vec[incr_vec_len].cost = COST_INFINITE;
1816
1817 /* Optimistically record the first occurrence of this increment
1818 as providing an initializer (if it does); we will revise this
1819 opinion later if it doesn't dominate all other occurrences.
1820 Exception: increments of -1, 0, 1 never need initializers. */
1821 if (c->kind == CAND_ADD
1822 && double_int_equal_p (c->index, increment)
1823 && (double_int_scmp (increment, double_int_one) > 0
1824 || double_int_scmp (increment, double_int_minus_one) < 0))
1825 {
1826 tree t0;
1827 tree rhs1 = gimple_assign_rhs1 (c->cand_stmt);
1828 tree rhs2 = gimple_assign_rhs2 (c->cand_stmt);
1829 if (operand_equal_p (rhs1, c->base_expr, 0))
1830 t0 = rhs2;
1831 else
1832 t0 = rhs1;
1833 if (SSA_NAME_DEF_STMT (t0) && gimple_bb (SSA_NAME_DEF_STMT (t0)))
1834 {
1835 incr_vec[incr_vec_len].initializer = t0;
1836 incr_vec[incr_vec_len++].init_bb
1837 = gimple_bb (SSA_NAME_DEF_STMT (t0));
1838 }
1839 else
1840 {
1841 incr_vec[incr_vec_len].initializer = NULL_TREE;
1842 incr_vec[incr_vec_len++].init_bb = NULL;
1843 }
1844 }
1845 else
1846 {
1847 incr_vec[incr_vec_len].initializer = NULL_TREE;
1848 incr_vec[incr_vec_len++].init_bb = NULL;
1849 }
1850 }
1851 }
1852
1853 /* Determine how many times each unique increment occurs in the set
1854 of candidates rooted at C's parent, recording the data in the
1855 increment vector. For each unique increment I, if an initializer
1856 T_0 = stride * I is provided by a candidate that dominates all
1857 candidates with the same increment, also record T_0 for subsequent
1858 use. */
1859
1860 static void
1861 record_increments (slsr_cand_t c)
1862 {
1863 if (!cand_already_replaced (c))
1864 record_increment (c, cand_increment (c));
1865
1866 if (c->sibling)
1867 record_increments (lookup_cand (c->sibling));
1868
1869 if (c->dependent)
1870 record_increments (lookup_cand (c->dependent));
1871 }
1872
1873 /* Return the first candidate in the tree rooted at C that has not
1874 already been replaced, favoring siblings over dependents. */
1875
1876 static slsr_cand_t
1877 unreplaced_cand_in_tree (slsr_cand_t c)
1878 {
1879 if (!cand_already_replaced (c))
1880 return c;
1881
1882 if (c->sibling)
1883 {
1884 slsr_cand_t sib = unreplaced_cand_in_tree (lookup_cand (c->sibling));
1885 if (sib)
1886 return sib;
1887 }
1888
1889 if (c->dependent)
1890 {
1891 slsr_cand_t dep = unreplaced_cand_in_tree (lookup_cand (c->dependent));
1892 if (dep)
1893 return dep;
1894 }
1895
1896 return NULL;
1897 }
1898
1899 /* Return TRUE if the candidates in the tree rooted at C should be
1900 optimized for speed, else FALSE. We estimate this based on the block
1901 containing the most dominant candidate in the tree that has not yet
1902 been replaced. */
1903
1904 static bool
1905 optimize_cands_for_speed_p (slsr_cand_t c)
1906 {
1907 slsr_cand_t c2 = unreplaced_cand_in_tree (c);
1908 gcc_assert (c2);
1909 return optimize_bb_for_speed_p (gimple_bb (c2->cand_stmt));
1910 }
1911
1912 /* Add COST_IN to the lowest cost of any dependent path starting at
1913 candidate C or any of its siblings, counting only candidates along
1914 such paths with increment INCR. Assume that replacing a candidate
1915 reduces cost by REPL_SAVINGS. Also account for savings from any
1916 statements that would go dead. */
1917
1918 static int
1919 lowest_cost_path (int cost_in, int repl_savings, slsr_cand_t c, double_int incr)
1920 {
1921 int local_cost, sib_cost;
1922 double_int cand_incr = cand_abs_increment (c);
1923
1924 if (cand_already_replaced (c))
1925 local_cost = cost_in;
1926 else if (double_int_equal_p (incr, cand_incr))
1927 local_cost = cost_in - repl_savings - c->dead_savings;
1928 else
1929 local_cost = cost_in - c->dead_savings;
1930
1931 if (c->dependent)
1932 local_cost = lowest_cost_path (local_cost, repl_savings,
1933 lookup_cand (c->dependent), incr);
1934
1935 if (c->sibling)
1936 {
1937 sib_cost = lowest_cost_path (cost_in, repl_savings,
1938 lookup_cand (c->sibling), incr);
1939 local_cost = MIN (local_cost, sib_cost);
1940 }
1941
1942 return local_cost;
1943 }
1944
1945 /* Compute the total savings that would accrue from all replacements
1946 in the candidate tree rooted at C, counting only candidates with
1947 increment INCR. Assume that replacing a candidate reduces cost
1948 by REPL_SAVINGS. Also account for savings from statements that
1949 would go dead. */
1950
1951 static int
1952 total_savings (int repl_savings, slsr_cand_t c, double_int incr)
1953 {
1954 int savings = 0;
1955 double_int cand_incr = cand_abs_increment (c);
1956
1957 if (double_int_equal_p (incr, cand_incr)
1958 && !cand_already_replaced (c))
1959 savings += repl_savings + c->dead_savings;
1960
1961 if (c->dependent)
1962 savings += total_savings (repl_savings, lookup_cand (c->dependent), incr);
1963
1964 if (c->sibling)
1965 savings += total_savings (repl_savings, lookup_cand (c->sibling), incr);
1966
1967 return savings;
1968 }
1969
1970 /* Use target-specific costs to determine and record which increments
1971 in the current candidate tree are profitable to replace, assuming
1972 MODE and SPEED. FIRST_DEP is the first dependent of the root of
1973 the candidate tree.
1974
1975 One slight limitation here is that we don't account for the possible
1976 introduction of casts in some cases. See replace_one_candidate for
1977 the cases where these are introduced. This should probably be cleaned
1978 up sometime. */
1979
1980 static void
1981 analyze_increments (slsr_cand_t first_dep, enum machine_mode mode, bool speed)
1982 {
1983 unsigned i;
1984
1985 for (i = 0; i < incr_vec_len; i++)
1986 {
1987 HOST_WIDE_INT incr = double_int_to_shwi (incr_vec[i].incr);
1988
1989 /* If somehow this increment is bigger than a HWI, we won't
1990 be optimizing candidates that use it. And if the increment
1991 has a count of zero, nothing will be done with it. */
1992 if (!double_int_fits_in_shwi_p (incr_vec[i].incr)
1993 || !incr_vec[i].count)
1994 incr_vec[i].cost = COST_INFINITE;
1995
1996 /* Increments of 0, 1, and -1 are always profitable to replace,
1997 because they always replace a multiply or add with an add or
1998 copy, and may cause one or more existing instructions to go
1999 dead. Exception: -1 can't be assumed to be profitable for
2000 pointer addition. */
2001 else if (incr == 0
2002 || incr == 1
2003 || (incr == -1
2004 && (gimple_assign_rhs_code (first_dep->cand_stmt)
2005 != POINTER_PLUS_EXPR)))
2006 incr_vec[i].cost = COST_NEUTRAL;
2007
2008 /* FORNOW: If we need to add an initializer, give up if a cast from
2009 the candidate's type to its stride's type can lose precision.
2010 This could eventually be handled better by expressly retaining the
2011 result of a cast to a wider type in the stride. Example:
2012
2013 short int _1;
2014 _2 = (int) _1;
2015 _3 = _2 * 10;
2016 _4 = x + _3; ADD: x + (10 * _1) : int
2017 _5 = _2 * 15;
2018 _6 = x + _3; ADD: x + (15 * _1) : int
2019
2020 Right now replacing _6 would cause insertion of an initializer
2021 of the form "short int T = _1 * 5;" followed by a cast to
2022 int, which could overflow incorrectly. Had we recorded _2 or
2023 (int)_1 as the stride, this wouldn't happen. However, doing
2024 this breaks other opportunities, so this will require some
2025 care. */
2026 else if (!incr_vec[i].initializer
2027 && TREE_CODE (first_dep->stride) != INTEGER_CST
2028 && !legal_cast_p_1 (first_dep->stride,
2029 gimple_assign_lhs (first_dep->cand_stmt)))
2030
2031 incr_vec[i].cost = COST_INFINITE;
2032
2033 /* For any other increment, if this is a multiply candidate, we
2034 must introduce a temporary T and initialize it with
2035 T_0 = stride * increment. When optimizing for speed, walk the
2036 candidate tree to calculate the best cost reduction along any
2037 path; if it offsets the fixed cost of inserting the initializer,
2038 replacing the increment is profitable. When optimizing for
2039 size, instead calculate the total cost reduction from replacing
2040 all candidates with this increment. */
2041 else if (first_dep->kind == CAND_MULT)
2042 {
2043 int cost = mult_by_coeff_cost (incr, mode, speed);
2044 int repl_savings = mul_cost (speed, mode) - add_cost (speed, mode);
2045 if (speed)
2046 cost = lowest_cost_path (cost, repl_savings, first_dep,
2047 incr_vec[i].incr);
2048 else
2049 cost -= total_savings (repl_savings, first_dep, incr_vec[i].incr);
2050
2051 incr_vec[i].cost = cost;
2052 }
2053
2054 /* If this is an add candidate, the initializer may already
2055 exist, so only calculate the cost of the initializer if it
2056 doesn't. We are replacing one add with another here, so the
2057 known replacement savings is zero. We will account for removal
2058 of dead instructions in lowest_cost_path or total_savings. */
2059 else
2060 {
2061 int cost = 0;
2062 if (!incr_vec[i].initializer)
2063 cost = mult_by_coeff_cost (incr, mode, speed);
2064
2065 if (speed)
2066 cost = lowest_cost_path (cost, 0, first_dep, incr_vec[i].incr);
2067 else
2068 cost -= total_savings (0, first_dep, incr_vec[i].incr);
2069
2070 incr_vec[i].cost = cost;
2071 }
2072 }
2073 }
2074
2075 /* Return the nearest common dominator of BB1 and BB2. If the blocks
2076 are identical, return the earlier of C1 and C2 in *WHERE. Otherwise,
2077 if the NCD matches BB1, return C1 in *WHERE; if the NCD matches BB2,
2078 return C2 in *WHERE; and if the NCD matches neither, return NULL in
2079 *WHERE. Note: It is possible for one of C1 and C2 to be NULL. */
2080
2081 static basic_block
2082 ncd_for_two_cands (basic_block bb1, basic_block bb2,
2083 slsr_cand_t c1, slsr_cand_t c2, slsr_cand_t *where)
2084 {
2085 basic_block ncd;
2086
2087 if (!bb1)
2088 {
2089 *where = c2;
2090 return bb2;
2091 }
2092
2093 if (!bb2)
2094 {
2095 *where = c1;
2096 return bb1;
2097 }
2098
2099 ncd = nearest_common_dominator (CDI_DOMINATORS, bb1, bb2);
2100
2101 /* If both candidates are in the same block, the earlier
2102 candidate wins. */
2103 if (bb1 == ncd && bb2 == ncd)
2104 {
2105 if (!c1 || (c2 && c2->cand_num < c1->cand_num))
2106 *where = c2;
2107 else
2108 *where = c1;
2109 }
2110
2111 /* Otherwise, if one of them produced a candidate in the
2112 dominator, that one wins. */
2113 else if (bb1 == ncd)
2114 *where = c1;
2115
2116 else if (bb2 == ncd)
2117 *where = c2;
2118
2119 /* If neither matches the dominator, neither wins. */
2120 else
2121 *where = NULL;
2122
2123 return ncd;
2124 }
2125
2126 /* Consider all candidates in the tree rooted at C for which INCR
2127 represents the required increment of C relative to its basis.
2128 Find and return the basic block that most nearly dominates all
2129 such candidates. If the returned block contains one or more of
2130 the candidates, return the earliest candidate in the block in
2131 *WHERE. */
2132
2133 static basic_block
2134 nearest_common_dominator_for_cands (slsr_cand_t c, double_int incr,
2135 slsr_cand_t *where)
2136 {
2137 basic_block sib_ncd = NULL, dep_ncd = NULL, this_ncd = NULL, ncd;
2138 slsr_cand_t sib_where = NULL, dep_where = NULL, this_where = NULL, new_where;
2139 double_int cand_incr;
2140
2141 /* First find the NCD of all siblings and dependents. */
2142 if (c->sibling)
2143 sib_ncd = nearest_common_dominator_for_cands (lookup_cand (c->sibling),
2144 incr, &sib_where);
2145 if (c->dependent)
2146 dep_ncd = nearest_common_dominator_for_cands (lookup_cand (c->dependent),
2147 incr, &dep_where);
2148 if (!sib_ncd && !dep_ncd)
2149 {
2150 new_where = NULL;
2151 ncd = NULL;
2152 }
2153 else if (sib_ncd && !dep_ncd)
2154 {
2155 new_where = sib_where;
2156 ncd = sib_ncd;
2157 }
2158 else if (dep_ncd && !sib_ncd)
2159 {
2160 new_where = dep_where;
2161 ncd = dep_ncd;
2162 }
2163 else
2164 ncd = ncd_for_two_cands (sib_ncd, dep_ncd, sib_where,
2165 dep_where, &new_where);
2166
2167 /* If the candidate's increment doesn't match the one we're interested
2168 in, then the result depends only on siblings and dependents. */
2169 cand_incr = cand_abs_increment (c);
2170
2171 if (!double_int_equal_p (cand_incr, incr) || cand_already_replaced (c))
2172 {
2173 *where = new_where;
2174 return ncd;
2175 }
2176
2177 /* Otherwise, compare this candidate with the result from all siblings
2178 and dependents. */
2179 this_where = c;
2180 this_ncd = gimple_bb (c->cand_stmt);
2181 ncd = ncd_for_two_cands (ncd, this_ncd, new_where, this_where, where);
2182
2183 return ncd;
2184 }
2185
2186 /* Return TRUE if the increment indexed by INDEX is profitable to replace. */
2187
2188 static inline bool
2189 profitable_increment_p (unsigned index)
2190 {
2191 return (incr_vec[index].cost <= COST_NEUTRAL);
2192 }
2193
2194 /* For each profitable increment in the increment vector not equal to
2195 0 or 1 (or -1, for non-pointer arithmetic), find the nearest common
2196 dominator of all statements in the candidate chain rooted at C
2197 that require that increment, and insert an initializer
2198 T_0 = stride * increment at that location. Record T_0 with the
2199 increment record. */
2200
2201 static void
2202 insert_initializers (slsr_cand_t c)
2203 {
2204 unsigned i;
2205 tree new_var = NULL_TREE;
2206
2207 for (i = 0; i < incr_vec_len; i++)
2208 {
2209 basic_block bb;
2210 slsr_cand_t where = NULL;
2211 gimple init_stmt;
2212 tree stride_type, new_name, incr_tree;
2213 double_int incr = incr_vec[i].incr;
2214
2215 if (!profitable_increment_p (i)
2216 || double_int_one_p (incr)
2217 || (double_int_minus_one_p (incr)
2218 && gimple_assign_rhs_code (c->cand_stmt) != POINTER_PLUS_EXPR)
2219 || double_int_zero_p (incr))
2220 continue;
2221
2222 /* We may have already identified an existing initializer that
2223 will suffice. */
2224 if (incr_vec[i].initializer)
2225 {
2226 if (dump_file && (dump_flags & TDF_DETAILS))
2227 {
2228 fputs ("Using existing initializer: ", dump_file);
2229 print_gimple_stmt (dump_file,
2230 SSA_NAME_DEF_STMT (incr_vec[i].initializer),
2231 0, 0);
2232 }
2233 continue;
2234 }
2235
2236 /* Find the block that most closely dominates all candidates
2237 with this increment. If there is at least one candidate in
2238 that block, the earliest one will be returned in WHERE. */
2239 bb = nearest_common_dominator_for_cands (c, incr, &where);
2240
2241 /* Create a new SSA name to hold the initializer's value. */
2242 stride_type = TREE_TYPE (c->stride);
2243 lazy_create_slsr_reg (&new_var, stride_type);
2244 new_name = make_ssa_name (new_var, NULL);
2245 incr_vec[i].initializer = new_name;
2246
2247 /* Create the initializer and insert it in the latest possible
2248 dominating position. */
2249 incr_tree = double_int_to_tree (stride_type, incr);
2250 init_stmt = gimple_build_assign_with_ops (MULT_EXPR, new_name,
2251 c->stride, incr_tree);
2252 if (where)
2253 {
2254 gimple_stmt_iterator gsi = gsi_for_stmt (where->cand_stmt);
2255 gsi_insert_before (&gsi, init_stmt, GSI_SAME_STMT);
2256 gimple_set_location (init_stmt, gimple_location (where->cand_stmt));
2257 }
2258 else
2259 {
2260 gimple_stmt_iterator gsi = gsi_last_bb (bb);
2261 gimple basis_stmt = lookup_cand (c->basis)->cand_stmt;
2262
2263 if (!gsi_end_p (gsi) && is_ctrl_stmt (gsi_stmt (gsi)))
2264 gsi_insert_before (&gsi, init_stmt, GSI_SAME_STMT);
2265 else
2266 gsi_insert_after (&gsi, init_stmt, GSI_SAME_STMT);
2267
2268 gimple_set_location (init_stmt, gimple_location (basis_stmt));
2269 }
2270
2271 if (dump_file && (dump_flags & TDF_DETAILS))
2272 {
2273 fputs ("Inserting initializer: ", dump_file);
2274 print_gimple_stmt (dump_file, init_stmt, 0, 0);
2275 }
2276 }
2277 }
2278
2279 /* Create a NOP_EXPR that copies FROM_EXPR into a new SSA name of
2280 type TO_TYPE, and insert it in front of the statement represented
2281 by candidate C. Use *NEW_VAR to create the new SSA name. Return
2282 the new SSA name. */
2283
2284 static tree
2285 introduce_cast_before_cand (slsr_cand_t c, tree to_type,
2286 tree from_expr, tree *new_var)
2287 {
2288 tree cast_lhs;
2289 gimple cast_stmt;
2290 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
2291
2292 lazy_create_slsr_reg (new_var, to_type);
2293 cast_lhs = make_ssa_name (*new_var, NULL);
2294 cast_stmt = gimple_build_assign_with_ops (NOP_EXPR, cast_lhs,
2295 from_expr, NULL_TREE);
2296 gimple_set_location (cast_stmt, gimple_location (c->cand_stmt));
2297 gsi_insert_before (&gsi, cast_stmt, GSI_SAME_STMT);
2298
2299 if (dump_file && (dump_flags & TDF_DETAILS))
2300 {
2301 fputs (" Inserting: ", dump_file);
2302 print_gimple_stmt (dump_file, cast_stmt, 0, 0);
2303 }
2304
2305 return cast_lhs;
2306 }
2307
2308 /* Replace the RHS of the statement represented by candidate C with
2309 NEW_CODE, NEW_RHS1, and NEW_RHS2, provided that to do so doesn't
2310 leave C unchanged or just interchange its operands. The original
2311 operation and operands are in OLD_CODE, OLD_RHS1, and OLD_RHS2.
2312 If the replacement was made and we are doing a details dump,
2313 return the revised statement, else NULL. */
2314
2315 static gimple
2316 replace_rhs_if_not_dup (enum tree_code new_code, tree new_rhs1, tree new_rhs2,
2317 enum tree_code old_code, tree old_rhs1, tree old_rhs2,
2318 slsr_cand_t c)
2319 {
2320 if (new_code != old_code
2321 || ((!operand_equal_p (new_rhs1, old_rhs1, 0)
2322 || !operand_equal_p (new_rhs2, old_rhs2, 0))
2323 && (!operand_equal_p (new_rhs1, old_rhs2, 0)
2324 || !operand_equal_p (new_rhs2, old_rhs1, 0))))
2325 {
2326 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
2327 gimple_assign_set_rhs_with_ops (&gsi, new_code, new_rhs1, new_rhs2);
2328 update_stmt (gsi_stmt (gsi));
2329
2330 if (dump_file && (dump_flags & TDF_DETAILS))
2331 return gsi_stmt (gsi);
2332 }
2333
2334 else if (dump_file && (dump_flags & TDF_DETAILS))
2335 fputs (" (duplicate, not actually replacing)\n", dump_file);
2336
2337 return NULL;
2338 }
2339
2340 /* Strength-reduce the statement represented by candidate C by replacing
2341 it with an equivalent addition or subtraction. I is the index into
2342 the increment vector identifying C's increment. NEW_VAR is used to
2343 create a new SSA name if a cast needs to be introduced. BASIS_NAME
2344 is the rhs1 to use in creating the add/subtract. */
2345
2346 static void
2347 replace_one_candidate (slsr_cand_t c, unsigned i, tree *new_var,
2348 tree basis_name)
2349 {
2350 gimple stmt_to_print = NULL;
2351 tree orig_rhs1, orig_rhs2;
2352 tree rhs2;
2353 enum tree_code orig_code, repl_code;
2354 double_int cand_incr;
2355
2356 orig_code = gimple_assign_rhs_code (c->cand_stmt);
2357 orig_rhs1 = gimple_assign_rhs1 (c->cand_stmt);
2358 orig_rhs2 = gimple_assign_rhs2 (c->cand_stmt);
2359 cand_incr = cand_increment (c);
2360
2361 if (dump_file && (dump_flags & TDF_DETAILS))
2362 {
2363 fputs ("Replacing: ", dump_file);
2364 print_gimple_stmt (dump_file, c->cand_stmt, 0, 0);
2365 stmt_to_print = c->cand_stmt;
2366 }
2367
2368 if (address_arithmetic_p)
2369 repl_code = POINTER_PLUS_EXPR;
2370 else
2371 repl_code = PLUS_EXPR;
2372
2373 /* If the increment has an initializer T_0, replace the candidate
2374 statement with an add of the basis name and the initializer. */
2375 if (incr_vec[i].initializer)
2376 {
2377 tree init_type = TREE_TYPE (incr_vec[i].initializer);
2378 tree orig_type = TREE_TYPE (orig_rhs2);
2379
2380 if (types_compatible_p (orig_type, init_type))
2381 rhs2 = incr_vec[i].initializer;
2382 else
2383 rhs2 = introduce_cast_before_cand (c, orig_type,
2384 incr_vec[i].initializer,
2385 new_var);
2386
2387 if (!double_int_equal_p (incr_vec[i].incr, cand_incr))
2388 {
2389 gcc_assert (repl_code == PLUS_EXPR);
2390 repl_code = MINUS_EXPR;
2391 }
2392
2393 stmt_to_print = replace_rhs_if_not_dup (repl_code, basis_name, rhs2,
2394 orig_code, orig_rhs1, orig_rhs2,
2395 c);
2396 }
2397
2398 /* Otherwise, the increment is one of -1, 0, and 1. Replace
2399 with a subtract of the stride from the basis name, a copy
2400 from the basis name, or an add of the stride to the basis
2401 name, respectively. It may be necessary to introduce a
2402 cast (or reuse an existing cast). */
2403 else if (double_int_one_p (cand_incr))
2404 {
2405 tree stride_type = TREE_TYPE (c->stride);
2406 tree orig_type = TREE_TYPE (orig_rhs2);
2407
2408 if (types_compatible_p (orig_type, stride_type))
2409 rhs2 = c->stride;
2410 else
2411 rhs2 = introduce_cast_before_cand (c, orig_type, c->stride, new_var);
2412
2413 stmt_to_print = replace_rhs_if_not_dup (repl_code, basis_name, rhs2,
2414 orig_code, orig_rhs1, orig_rhs2,
2415 c);
2416 }
2417
2418 else if (double_int_minus_one_p (cand_incr))
2419 {
2420 tree stride_type = TREE_TYPE (c->stride);
2421 tree orig_type = TREE_TYPE (orig_rhs2);
2422 gcc_assert (repl_code != POINTER_PLUS_EXPR);
2423
2424 if (types_compatible_p (orig_type, stride_type))
2425 rhs2 = c->stride;
2426 else
2427 rhs2 = introduce_cast_before_cand (c, orig_type, c->stride, new_var);
2428
2429 if (orig_code != MINUS_EXPR
2430 || !operand_equal_p (basis_name, orig_rhs1, 0)
2431 || !operand_equal_p (rhs2, orig_rhs2, 0))
2432 {
2433 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
2434 gimple_assign_set_rhs_with_ops (&gsi, MINUS_EXPR, basis_name, rhs2);
2435 update_stmt (gsi_stmt (gsi));
2436
2437 if (dump_file && (dump_flags & TDF_DETAILS))
2438 stmt_to_print = gsi_stmt (gsi);
2439 }
2440 else if (dump_file && (dump_flags & TDF_DETAILS))
2441 fputs (" (duplicate, not actually replacing)\n", dump_file);
2442 }
2443
2444 else if (double_int_zero_p (cand_incr))
2445 {
2446 tree lhs = gimple_assign_lhs (c->cand_stmt);
2447 tree lhs_type = TREE_TYPE (lhs);
2448 tree basis_type = TREE_TYPE (basis_name);
2449
2450 if (types_compatible_p (lhs_type, basis_type))
2451 {
2452 gimple copy_stmt = gimple_build_assign (lhs, basis_name);
2453 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
2454 gimple_set_location (copy_stmt, gimple_location (c->cand_stmt));
2455 gsi_replace (&gsi, copy_stmt, false);
2456
2457 if (dump_file && (dump_flags & TDF_DETAILS))
2458 stmt_to_print = copy_stmt;
2459 }
2460 else
2461 {
2462 gimple_stmt_iterator gsi = gsi_for_stmt (c->cand_stmt);
2463 gimple cast_stmt = gimple_build_assign_with_ops (NOP_EXPR, lhs,
2464 basis_name,
2465 NULL_TREE);
2466 gimple_set_location (cast_stmt, gimple_location (c->cand_stmt));
2467 gsi_replace (&gsi, cast_stmt, false);
2468
2469 if (dump_file && (dump_flags & TDF_DETAILS))
2470 stmt_to_print = cast_stmt;
2471 }
2472 }
2473 else
2474 gcc_unreachable ();
2475
2476 if (dump_file && (dump_flags & TDF_DETAILS) && stmt_to_print)
2477 {
2478 fputs ("With: ", dump_file);
2479 print_gimple_stmt (dump_file, stmt_to_print, 0, 0);
2480 fputs ("\n", dump_file);
2481 }
2482 }
2483
2484 /* For each candidate in the tree rooted at C, replace it with
2485 an increment if such has been shown to be profitable. */
2486
2487 static void
2488 replace_profitable_candidates (slsr_cand_t c)
2489 {
2490 if (!cand_already_replaced (c))
2491 {
2492 double_int increment = cand_abs_increment (c);
2493 tree new_var = NULL;
2494 enum tree_code orig_code = gimple_assign_rhs_code (c->cand_stmt);
2495 unsigned i;
2496
2497 i = incr_vec_index (increment);
2498
2499 /* Only process profitable increments. Nothing useful can be done
2500 to a cast or copy. */
2501 if (profitable_increment_p (i)
2502 && orig_code != MODIFY_EXPR
2503 && orig_code != NOP_EXPR)
2504 {
2505 slsr_cand_t basis = lookup_cand (c->basis);
2506 tree basis_name = gimple_assign_lhs (basis->cand_stmt);
2507 replace_one_candidate (c, i, &new_var, basis_name);
2508 }
2509 }
2510
2511 if (c->sibling)
2512 replace_profitable_candidates (lookup_cand (c->sibling));
2513
2514 if (c->dependent)
2515 replace_profitable_candidates (lookup_cand (c->dependent));
2516 }
2517 \f
2518 /* Analyze costs of related candidates in the candidate vector,
2519 and make beneficial replacements. */
2520
2521 static void
2522 analyze_candidates_and_replace (void)
2523 {
2524 unsigned i;
2525 slsr_cand_t c;
2526
2527 /* Each candidate that has a null basis and a non-null
2528 dependent is the root of a tree of related statements.
2529 Analyze each tree to determine a subset of those
2530 statements that can be replaced with maximum benefit. */
2531 FOR_EACH_VEC_ELT (slsr_cand_t, cand_vec, i, c)
2532 {
2533 slsr_cand_t first_dep;
2534
2535 if (c->basis != 0 || c->dependent == 0)
2536 continue;
2537
2538 if (dump_file && (dump_flags & TDF_DETAILS))
2539 fprintf (dump_file, "\nProcessing dependency tree rooted at %d.\n",
2540 c->cand_num);
2541
2542 first_dep = lookup_cand (c->dependent);
2543
2544 /* If this is a chain of CAND_REFs, unconditionally replace
2545 each of them with a strength-reduced data reference. */
2546 if (c->kind == CAND_REF)
2547 replace_refs (c);
2548
2549 /* If the common stride of all related candidates is a
2550 known constant, and none of these has a phi-dependence,
2551 then all replacements are considered profitable.
2552 Each replaces a multiply by a single add, with the
2553 possibility that a feeding add also goes dead as a
2554 result. */
2555 else if (unconditional_cands_with_known_stride_p (c))
2556 replace_dependents (first_dep);
2557
2558 /* When the stride is an SSA name, it may still be profitable
2559 to replace some or all of the dependent candidates, depending
2560 on whether the introduced increments can be reused, or are
2561 less expensive to calculate than the replaced statements. */
2562 else
2563 {
2564 unsigned length;
2565 enum machine_mode mode;
2566 bool speed;
2567
2568 /* Determine whether we'll be generating pointer arithmetic
2569 when replacing candidates. */
2570 address_arithmetic_p = (c->kind == CAND_ADD
2571 && POINTER_TYPE_P (c->cand_type));
2572
2573 /* If all candidates have already been replaced under other
2574 interpretations, nothing remains to be done. */
2575 length = count_candidates (c);
2576 if (!length)
2577 continue;
2578
2579 /* Construct an array of increments for this candidate chain. */
2580 incr_vec = XNEWVEC (incr_info, length);
2581 incr_vec_len = 0;
2582 record_increments (c);
2583
2584 /* Determine which increments are profitable to replace. */
2585 mode = TYPE_MODE (TREE_TYPE (gimple_assign_lhs (c->cand_stmt)));
2586 speed = optimize_cands_for_speed_p (c);
2587 analyze_increments (first_dep, mode, speed);
2588
2589 /* Insert initializers of the form T_0 = stride * increment
2590 for use in profitable replacements. */
2591 insert_initializers (first_dep);
2592 dump_incr_vec ();
2593
2594 /* Perform the replacements. */
2595 replace_profitable_candidates (first_dep);
2596 free (incr_vec);
2597 }
2598
2599 /* TODO: When conditional increments occur so that a
2600 candidate is dependent upon a phi-basis, the cost of
2601 introducing a temporary must be accounted for. */
2602 }
2603 }
2604
2605 static unsigned
2606 execute_strength_reduction (void)
2607 {
2608 struct dom_walk_data walk_data;
2609
2610 /* Create the obstack where candidates will reside. */
2611 gcc_obstack_init (&cand_obstack);
2612
2613 /* Allocate the candidate vector. */
2614 cand_vec = VEC_alloc (slsr_cand_t, heap, 128);
2615
2616 /* Allocate the mapping from statements to candidate indices. */
2617 stmt_cand_map = pointer_map_create ();
2618
2619 /* Create the obstack where candidate chains will reside. */
2620 gcc_obstack_init (&chain_obstack);
2621
2622 /* Allocate the mapping from base expressions to candidate chains. */
2623 base_cand_map = htab_create (500, base_cand_hash,
2624 base_cand_eq, base_cand_free);
2625
2626 /* Initialize the loop optimizer. We need to detect flow across
2627 back edges, and this gives us dominator information as well. */
2628 loop_optimizer_init (AVOID_CFG_MODIFICATIONS);
2629
2630 /* Set up callbacks for the generic dominator tree walker. */
2631 walk_data.dom_direction = CDI_DOMINATORS;
2632 walk_data.initialize_block_local_data = NULL;
2633 walk_data.before_dom_children = find_candidates_in_block;
2634 walk_data.after_dom_children = NULL;
2635 walk_data.global_data = NULL;
2636 walk_data.block_local_data_size = 0;
2637 init_walk_dominator_tree (&walk_data);
2638
2639 /* Walk the CFG in predominator order looking for strength reduction
2640 candidates. */
2641 walk_dominator_tree (&walk_data, ENTRY_BLOCK_PTR);
2642
2643 if (dump_file && (dump_flags & TDF_DETAILS))
2644 {
2645 dump_cand_vec ();
2646 dump_cand_chains ();
2647 }
2648
2649 /* Analyze costs and make appropriate replacements. */
2650 analyze_candidates_and_replace ();
2651
2652 /* Free resources. */
2653 fini_walk_dominator_tree (&walk_data);
2654 loop_optimizer_finalize ();
2655 htab_delete (base_cand_map);
2656 obstack_free (&chain_obstack, NULL);
2657 pointer_map_destroy (stmt_cand_map);
2658 VEC_free (slsr_cand_t, heap, cand_vec);
2659 obstack_free (&cand_obstack, NULL);
2660
2661 return 0;
2662 }
2663
2664 static bool
2665 gate_strength_reduction (void)
2666 {
2667 return flag_tree_slsr;
2668 }
2669
2670 struct gimple_opt_pass pass_strength_reduction =
2671 {
2672 {
2673 GIMPLE_PASS,
2674 "slsr", /* name */
2675 gate_strength_reduction, /* gate */
2676 execute_strength_reduction, /* execute */
2677 NULL, /* sub */
2678 NULL, /* next */
2679 0, /* static_pass_number */
2680 TV_GIMPLE_SLSR, /* tv_id */
2681 PROP_cfg | PROP_ssa, /* properties_required */
2682 0, /* properties_provided */
2683 0, /* properties_destroyed */
2684 0, /* todo_flags_start */
2685 TODO_verify_ssa /* todo_flags_finish */
2686 }
2687 };