Index...
[gcc.git] / gcc / tree-sra.c
1 /* Scalar Replacement of Aggregates (SRA) converts some structure
2 references into scalar references, exposing them to the scalar
3 optimizers.
4 Copyright (C) 2008, 2009, 2010, 2011 Free Software Foundation, Inc.
5 Contributed by Martin Jambor <mjambor@suse.cz>
6
7 This file is part of GCC.
8
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
13
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
22
23 /* This file implements Scalar Reduction of Aggregates (SRA). SRA is run
24 twice, once in the early stages of compilation (early SRA) and once in the
25 late stages (late SRA). The aim of both is to turn references to scalar
26 parts of aggregates into uses of independent scalar variables.
27
28 The two passes are nearly identical, the only difference is that early SRA
29 does not scalarize unions which are used as the result in a GIMPLE_RETURN
30 statement because together with inlining this can lead to weird type
31 conversions.
32
33 Both passes operate in four stages:
34
35 1. The declarations that have properties which make them candidates for
36 scalarization are identified in function find_var_candidates(). The
37 candidates are stored in candidate_bitmap.
38
39 2. The function body is scanned. In the process, declarations which are
40 used in a manner that prevent their scalarization are removed from the
41 candidate bitmap. More importantly, for every access into an aggregate,
42 an access structure (struct access) is created by create_access() and
43 stored in a vector associated with the aggregate. Among other
44 information, the aggregate declaration, the offset and size of the access
45 and its type are stored in the structure.
46
47 On a related note, assign_link structures are created for every assign
48 statement between candidate aggregates and attached to the related
49 accesses.
50
51 3. The vectors of accesses are analyzed. They are first sorted according to
52 their offset and size and then scanned for partially overlapping accesses
53 (i.e. those which overlap but one is not entirely within another). Such
54 an access disqualifies the whole aggregate from being scalarized.
55
56 If there is no such inhibiting overlap, a representative access structure
57 is chosen for every unique combination of offset and size. Afterwards,
58 the pass builds a set of trees from these structures, in which children
59 of an access are within their parent (in terms of offset and size).
60
61 Then accesses are propagated whenever possible (i.e. in cases when it
62 does not create a partially overlapping access) across assign_links from
63 the right hand side to the left hand side.
64
65 Then the set of trees for each declaration is traversed again and those
66 accesses which should be replaced by a scalar are identified.
67
68 4. The function is traversed again, and for every reference into an
69 aggregate that has some component which is about to be scalarized,
70 statements are amended and new statements are created as necessary.
71 Finally, if a parameter got scalarized, the scalar replacements are
72 initialized with values from respective parameter aggregates. */
73
74 #include "config.h"
75 #include "system.h"
76 #include "coretypes.h"
77 #include "alloc-pool.h"
78 #include "tm.h"
79 #include "tree.h"
80 #include "gimple.h"
81 #include "cgraph.h"
82 #include "tree-flow.h"
83 #include "ipa-prop.h"
84 #include "tree-pretty-print.h"
85 #include "statistics.h"
86 #include "tree-dump.h"
87 #include "timevar.h"
88 #include "params.h"
89 #include "target.h"
90 #include "flags.h"
91 #include "dbgcnt.h"
92 #include "tree-inline.h"
93 #include "gimple-pretty-print.h"
94 #include "ipa-inline.h"
95
96 /* Enumeration of all aggregate reductions we can do. */
97 enum sra_mode { SRA_MODE_EARLY_IPA, /* early call regularization */
98 SRA_MODE_EARLY_INTRA, /* early intraprocedural SRA */
99 SRA_MODE_INTRA }; /* late intraprocedural SRA */
100
101 /* Global variable describing which aggregate reduction we are performing at
102 the moment. */
103 static enum sra_mode sra_mode;
104
105 struct assign_link;
106
107 /* ACCESS represents each access to an aggregate variable (as a whole or a
108 part). It can also represent a group of accesses that refer to exactly the
109 same fragment of an aggregate (i.e. those that have exactly the same offset
110 and size). Such representatives for a single aggregate, once determined,
111 are linked in a linked list and have the group fields set.
112
113 Moreover, when doing intraprocedural SRA, a tree is built from those
114 representatives (by the means of first_child and next_sibling pointers), in
115 which all items in a subtree are "within" the root, i.e. their offset is
116 greater or equal to offset of the root and offset+size is smaller or equal
117 to offset+size of the root. Children of an access are sorted by offset.
118
119 Note that accesses to parts of vector and complex number types always
120 represented by an access to the whole complex number or a vector. It is a
121 duty of the modifying functions to replace them appropriately. */
122
123 struct access
124 {
125 /* Values returned by `get_ref_base_and_extent' for each component reference
126 If EXPR isn't a component reference just set `BASE = EXPR', `OFFSET = 0',
127 `SIZE = TREE_SIZE (TREE_TYPE (expr))'. */
128 HOST_WIDE_INT offset;
129 HOST_WIDE_INT size;
130 tree base;
131
132 /* Expression. It is context dependent so do not use it to create new
133 expressions to access the original aggregate. See PR 42154 for a
134 testcase. */
135 tree expr;
136 /* Type. */
137 tree type;
138
139 /* The statement this access belongs to. */
140 gimple stmt;
141
142 /* Next group representative for this aggregate. */
143 struct access *next_grp;
144
145 /* Pointer to the group representative. Pointer to itself if the struct is
146 the representative. */
147 struct access *group_representative;
148
149 /* If this access has any children (in terms of the definition above), this
150 points to the first one. */
151 struct access *first_child;
152
153 /* In intraprocedural SRA, pointer to the next sibling in the access tree as
154 described above. In IPA-SRA this is a pointer to the next access
155 belonging to the same group (having the same representative). */
156 struct access *next_sibling;
157
158 /* Pointers to the first and last element in the linked list of assign
159 links. */
160 struct assign_link *first_link, *last_link;
161
162 /* Pointer to the next access in the work queue. */
163 struct access *next_queued;
164
165 /* Replacement variable for this access "region." Never to be accessed
166 directly, always only by the means of get_access_replacement() and only
167 when grp_to_be_replaced flag is set. */
168 tree replacement_decl;
169
170 /* Is this particular access write access? */
171 unsigned write : 1;
172
173 /* Is this access an access to a non-addressable field? */
174 unsigned non_addressable : 1;
175
176 /* Is this access currently in the work queue? */
177 unsigned grp_queued : 1;
178
179 /* Does this group contain a write access? This flag is propagated down the
180 access tree. */
181 unsigned grp_write : 1;
182
183 /* Does this group contain a read access? This flag is propagated down the
184 access tree. */
185 unsigned grp_read : 1;
186
187 /* Does this group contain a read access that comes from an assignment
188 statement? This flag is propagated down the access tree. */
189 unsigned grp_assignment_read : 1;
190
191 /* Does this group contain a write access that comes from an assignment
192 statement? This flag is propagated down the access tree. */
193 unsigned grp_assignment_write : 1;
194
195 /* Does this group contain a read access through a scalar type? This flag is
196 not propagated in the access tree in any direction. */
197 unsigned grp_scalar_read : 1;
198
199 /* Does this group contain a write access through a scalar type? This flag
200 is not propagated in the access tree in any direction. */
201 unsigned grp_scalar_write : 1;
202
203 /* Is this access an artificial one created to scalarize some record
204 entirely? */
205 unsigned grp_total_scalarization : 1;
206
207 /* Other passes of the analysis use this bit to make function
208 analyze_access_subtree create scalar replacements for this group if
209 possible. */
210 unsigned grp_hint : 1;
211
212 /* Is the subtree rooted in this access fully covered by scalar
213 replacements? */
214 unsigned grp_covered : 1;
215
216 /* If set to true, this access and all below it in an access tree must not be
217 scalarized. */
218 unsigned grp_unscalarizable_region : 1;
219
220 /* Whether data have been written to parts of the aggregate covered by this
221 access which is not to be scalarized. This flag is propagated up in the
222 access tree. */
223 unsigned grp_unscalarized_data : 1;
224
225 /* Does this access and/or group contain a write access through a
226 BIT_FIELD_REF? */
227 unsigned grp_partial_lhs : 1;
228
229 /* Set when a scalar replacement should be created for this variable. We do
230 the decision and creation at different places because create_tmp_var
231 cannot be called from within FOR_EACH_REFERENCED_VAR. */
232 unsigned grp_to_be_replaced : 1;
233
234 /* Should TREE_NO_WARNING of a replacement be set? */
235 unsigned grp_no_warning : 1;
236
237 /* Is it possible that the group refers to data which might be (directly or
238 otherwise) modified? */
239 unsigned grp_maybe_modified : 1;
240
241 /* Set when this is a representative of a pointer to scalar (i.e. by
242 reference) parameter which we consider for turning into a plain scalar
243 (i.e. a by value parameter). */
244 unsigned grp_scalar_ptr : 1;
245
246 /* Set when we discover that this pointer is not safe to dereference in the
247 caller. */
248 unsigned grp_not_necessarilly_dereferenced : 1;
249 };
250
251 typedef struct access *access_p;
252
253 DEF_VEC_P (access_p);
254 DEF_VEC_ALLOC_P (access_p, heap);
255
256 /* Alloc pool for allocating access structures. */
257 static alloc_pool access_pool;
258
259 /* A structure linking lhs and rhs accesses from an aggregate assignment. They
260 are used to propagate subaccesses from rhs to lhs as long as they don't
261 conflict with what is already there. */
262 struct assign_link
263 {
264 struct access *lacc, *racc;
265 struct assign_link *next;
266 };
267
268 /* Alloc pool for allocating assign link structures. */
269 static alloc_pool link_pool;
270
271 /* Base (tree) -> Vector (VEC(access_p,heap) *) map. */
272 static struct pointer_map_t *base_access_vec;
273
274 /* Bitmap of candidates. */
275 static bitmap candidate_bitmap;
276
277 /* Bitmap of candidates which we should try to entirely scalarize away and
278 those which cannot be (because they are and need be used as a whole). */
279 static bitmap should_scalarize_away_bitmap, cannot_scalarize_away_bitmap;
280
281 /* Obstack for creation of fancy names. */
282 static struct obstack name_obstack;
283
284 /* Head of a linked list of accesses that need to have its subaccesses
285 propagated to their assignment counterparts. */
286 static struct access *work_queue_head;
287
288 /* Number of parameters of the analyzed function when doing early ipa SRA. */
289 static int func_param_count;
290
291 /* scan_function sets the following to true if it encounters a call to
292 __builtin_apply_args. */
293 static bool encountered_apply_args;
294
295 /* Set by scan_function when it finds a recursive call. */
296 static bool encountered_recursive_call;
297
298 /* Set by scan_function when it finds a recursive call with less actual
299 arguments than formal parameters.. */
300 static bool encountered_unchangable_recursive_call;
301
302 /* This is a table in which for each basic block and parameter there is a
303 distance (offset + size) in that parameter which is dereferenced and
304 accessed in that BB. */
305 static HOST_WIDE_INT *bb_dereferences;
306 /* Bitmap of BBs that can cause the function to "stop" progressing by
307 returning, throwing externally, looping infinitely or calling a function
308 which might abort etc.. */
309 static bitmap final_bbs;
310
311 /* Representative of no accesses at all. */
312 static struct access no_accesses_representant;
313
314 /* Predicate to test the special value. */
315
316 static inline bool
317 no_accesses_p (struct access *access)
318 {
319 return access == &no_accesses_representant;
320 }
321
322 /* Dump contents of ACCESS to file F in a human friendly way. If GRP is true,
323 representative fields are dumped, otherwise those which only describe the
324 individual access are. */
325
326 static struct
327 {
328 /* Number of processed aggregates is readily available in
329 analyze_all_variable_accesses and so is not stored here. */
330
331 /* Number of created scalar replacements. */
332 int replacements;
333
334 /* Number of times sra_modify_expr or sra_modify_assign themselves changed an
335 expression. */
336 int exprs;
337
338 /* Number of statements created by generate_subtree_copies. */
339 int subtree_copies;
340
341 /* Number of statements created by load_assign_lhs_subreplacements. */
342 int subreplacements;
343
344 /* Number of times sra_modify_assign has deleted a statement. */
345 int deleted;
346
347 /* Number of times sra_modify_assign has to deal with subaccesses of LHS and
348 RHS reparately due to type conversions or nonexistent matching
349 references. */
350 int separate_lhs_rhs_handling;
351
352 /* Number of parameters that were removed because they were unused. */
353 int deleted_unused_parameters;
354
355 /* Number of scalars passed as parameters by reference that have been
356 converted to be passed by value. */
357 int scalar_by_ref_to_by_val;
358
359 /* Number of aggregate parameters that were replaced by one or more of their
360 components. */
361 int aggregate_params_reduced;
362
363 /* Numbber of components created when splitting aggregate parameters. */
364 int param_reductions_created;
365 } sra_stats;
366
367 static void
368 dump_access (FILE *f, struct access *access, bool grp)
369 {
370 fprintf (f, "access { ");
371 fprintf (f, "base = (%d)'", DECL_UID (access->base));
372 print_generic_expr (f, access->base, 0);
373 fprintf (f, "', offset = " HOST_WIDE_INT_PRINT_DEC, access->offset);
374 fprintf (f, ", size = " HOST_WIDE_INT_PRINT_DEC, access->size);
375 fprintf (f, ", expr = ");
376 print_generic_expr (f, access->expr, 0);
377 fprintf (f, ", type = ");
378 print_generic_expr (f, access->type, 0);
379 if (grp)
380 fprintf (f, ", grp_read = %d, grp_write = %d, grp_assignment_read = %d, "
381 "grp_assignment_write = %d, grp_scalar_read = %d, "
382 "grp_scalar_write = %d, grp_total_scalarization = %d, "
383 "grp_hint = %d, grp_covered = %d, "
384 "grp_unscalarizable_region = %d, grp_unscalarized_data = %d, "
385 "grp_partial_lhs = %d, grp_to_be_replaced = %d, "
386 "grp_maybe_modified = %d, "
387 "grp_not_necessarilly_dereferenced = %d\n",
388 access->grp_read, access->grp_write, access->grp_assignment_read,
389 access->grp_assignment_write, access->grp_scalar_read,
390 access->grp_scalar_write, access->grp_total_scalarization,
391 access->grp_hint, access->grp_covered,
392 access->grp_unscalarizable_region, access->grp_unscalarized_data,
393 access->grp_partial_lhs, access->grp_to_be_replaced,
394 access->grp_maybe_modified,
395 access->grp_not_necessarilly_dereferenced);
396 else
397 fprintf (f, ", write = %d, grp_total_scalarization = %d, "
398 "grp_partial_lhs = %d\n",
399 access->write, access->grp_total_scalarization,
400 access->grp_partial_lhs);
401 }
402
403 /* Dump a subtree rooted in ACCESS to file F, indent by LEVEL. */
404
405 static void
406 dump_access_tree_1 (FILE *f, struct access *access, int level)
407 {
408 do
409 {
410 int i;
411
412 for (i = 0; i < level; i++)
413 fputs ("* ", dump_file);
414
415 dump_access (f, access, true);
416
417 if (access->first_child)
418 dump_access_tree_1 (f, access->first_child, level + 1);
419
420 access = access->next_sibling;
421 }
422 while (access);
423 }
424
425 /* Dump all access trees for a variable, given the pointer to the first root in
426 ACCESS. */
427
428 static void
429 dump_access_tree (FILE *f, struct access *access)
430 {
431 for (; access; access = access->next_grp)
432 dump_access_tree_1 (f, access, 0);
433 }
434
435 /* Return true iff ACC is non-NULL and has subaccesses. */
436
437 static inline bool
438 access_has_children_p (struct access *acc)
439 {
440 return acc && acc->first_child;
441 }
442
443 /* Return true iff ACC is (partly) covered by at least one replacement. */
444
445 static bool
446 access_has_replacements_p (struct access *acc)
447 {
448 struct access *child;
449 if (acc->grp_to_be_replaced)
450 return true;
451 for (child = acc->first_child; child; child = child->next_sibling)
452 if (access_has_replacements_p (child))
453 return true;
454 return false;
455 }
456
457 /* Return a vector of pointers to accesses for the variable given in BASE or
458 NULL if there is none. */
459
460 static VEC (access_p, heap) *
461 get_base_access_vector (tree base)
462 {
463 void **slot;
464
465 slot = pointer_map_contains (base_access_vec, base);
466 if (!slot)
467 return NULL;
468 else
469 return *(VEC (access_p, heap) **) slot;
470 }
471
472 /* Find an access with required OFFSET and SIZE in a subtree of accesses rooted
473 in ACCESS. Return NULL if it cannot be found. */
474
475 static struct access *
476 find_access_in_subtree (struct access *access, HOST_WIDE_INT offset,
477 HOST_WIDE_INT size)
478 {
479 while (access && (access->offset != offset || access->size != size))
480 {
481 struct access *child = access->first_child;
482
483 while (child && (child->offset + child->size <= offset))
484 child = child->next_sibling;
485 access = child;
486 }
487
488 return access;
489 }
490
491 /* Return the first group representative for DECL or NULL if none exists. */
492
493 static struct access *
494 get_first_repr_for_decl (tree base)
495 {
496 VEC (access_p, heap) *access_vec;
497
498 access_vec = get_base_access_vector (base);
499 if (!access_vec)
500 return NULL;
501
502 return VEC_index (access_p, access_vec, 0);
503 }
504
505 /* Find an access representative for the variable BASE and given OFFSET and
506 SIZE. Requires that access trees have already been built. Return NULL if
507 it cannot be found. */
508
509 static struct access *
510 get_var_base_offset_size_access (tree base, HOST_WIDE_INT offset,
511 HOST_WIDE_INT size)
512 {
513 struct access *access;
514
515 access = get_first_repr_for_decl (base);
516 while (access && (access->offset + access->size <= offset))
517 access = access->next_grp;
518 if (!access)
519 return NULL;
520
521 return find_access_in_subtree (access, offset, size);
522 }
523
524 /* Add LINK to the linked list of assign links of RACC. */
525 static void
526 add_link_to_rhs (struct access *racc, struct assign_link *link)
527 {
528 gcc_assert (link->racc == racc);
529
530 if (!racc->first_link)
531 {
532 gcc_assert (!racc->last_link);
533 racc->first_link = link;
534 }
535 else
536 racc->last_link->next = link;
537
538 racc->last_link = link;
539 link->next = NULL;
540 }
541
542 /* Move all link structures in their linked list in OLD_RACC to the linked list
543 in NEW_RACC. */
544 static void
545 relink_to_new_repr (struct access *new_racc, struct access *old_racc)
546 {
547 if (!old_racc->first_link)
548 {
549 gcc_assert (!old_racc->last_link);
550 return;
551 }
552
553 if (new_racc->first_link)
554 {
555 gcc_assert (!new_racc->last_link->next);
556 gcc_assert (!old_racc->last_link || !old_racc->last_link->next);
557
558 new_racc->last_link->next = old_racc->first_link;
559 new_racc->last_link = old_racc->last_link;
560 }
561 else
562 {
563 gcc_assert (!new_racc->last_link);
564
565 new_racc->first_link = old_racc->first_link;
566 new_racc->last_link = old_racc->last_link;
567 }
568 old_racc->first_link = old_racc->last_link = NULL;
569 }
570
571 /* Add ACCESS to the work queue (which is actually a stack). */
572
573 static void
574 add_access_to_work_queue (struct access *access)
575 {
576 if (!access->grp_queued)
577 {
578 gcc_assert (!access->next_queued);
579 access->next_queued = work_queue_head;
580 access->grp_queued = 1;
581 work_queue_head = access;
582 }
583 }
584
585 /* Pop an access from the work queue, and return it, assuming there is one. */
586
587 static struct access *
588 pop_access_from_work_queue (void)
589 {
590 struct access *access = work_queue_head;
591
592 work_queue_head = access->next_queued;
593 access->next_queued = NULL;
594 access->grp_queued = 0;
595 return access;
596 }
597
598
599 /* Allocate necessary structures. */
600
601 static void
602 sra_initialize (void)
603 {
604 candidate_bitmap = BITMAP_ALLOC (NULL);
605 should_scalarize_away_bitmap = BITMAP_ALLOC (NULL);
606 cannot_scalarize_away_bitmap = BITMAP_ALLOC (NULL);
607 gcc_obstack_init (&name_obstack);
608 access_pool = create_alloc_pool ("SRA accesses", sizeof (struct access), 16);
609 link_pool = create_alloc_pool ("SRA links", sizeof (struct assign_link), 16);
610 base_access_vec = pointer_map_create ();
611 memset (&sra_stats, 0, sizeof (sra_stats));
612 encountered_apply_args = false;
613 encountered_recursive_call = false;
614 encountered_unchangable_recursive_call = false;
615 }
616
617 /* Hook fed to pointer_map_traverse, deallocate stored vectors. */
618
619 static bool
620 delete_base_accesses (const void *key ATTRIBUTE_UNUSED, void **value,
621 void *data ATTRIBUTE_UNUSED)
622 {
623 VEC (access_p, heap) *access_vec;
624 access_vec = (VEC (access_p, heap) *) *value;
625 VEC_free (access_p, heap, access_vec);
626
627 return true;
628 }
629
630 /* Deallocate all general structures. */
631
632 static void
633 sra_deinitialize (void)
634 {
635 BITMAP_FREE (candidate_bitmap);
636 BITMAP_FREE (should_scalarize_away_bitmap);
637 BITMAP_FREE (cannot_scalarize_away_bitmap);
638 free_alloc_pool (access_pool);
639 free_alloc_pool (link_pool);
640 obstack_free (&name_obstack, NULL);
641
642 pointer_map_traverse (base_access_vec, delete_base_accesses, NULL);
643 pointer_map_destroy (base_access_vec);
644 }
645
646 /* Remove DECL from candidates for SRA and write REASON to the dump file if
647 there is one. */
648 static void
649 disqualify_candidate (tree decl, const char *reason)
650 {
651 bitmap_clear_bit (candidate_bitmap, DECL_UID (decl));
652
653 if (dump_file && (dump_flags & TDF_DETAILS))
654 {
655 fprintf (dump_file, "! Disqualifying ");
656 print_generic_expr (dump_file, decl, 0);
657 fprintf (dump_file, " - %s\n", reason);
658 }
659 }
660
661 /* Return true iff the type contains a field or an element which does not allow
662 scalarization. */
663
664 static bool
665 type_internals_preclude_sra_p (tree type, const char **msg)
666 {
667 tree fld;
668 tree et;
669
670 switch (TREE_CODE (type))
671 {
672 case RECORD_TYPE:
673 case UNION_TYPE:
674 case QUAL_UNION_TYPE:
675 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
676 if (TREE_CODE (fld) == FIELD_DECL)
677 {
678 tree ft = TREE_TYPE (fld);
679
680 if (TREE_THIS_VOLATILE (fld))
681 {
682 *msg = "volatile structure field";
683 return true;
684 }
685 if (!DECL_FIELD_OFFSET (fld))
686 {
687 *msg = "no structure field offset";
688 return true;
689 }
690 if (!DECL_SIZE (fld))
691 {
692 *msg = "zero structure field size";
693 return true;
694 }
695 if (!host_integerp (DECL_FIELD_OFFSET (fld), 1))
696 {
697 *msg = "structure field offset not fixed";
698 return true;
699 }
700 if (!host_integerp (DECL_SIZE (fld), 1))
701 {
702 *msg = "structure field size not fixed";
703 return true;
704 }
705 if (AGGREGATE_TYPE_P (ft)
706 && int_bit_position (fld) % BITS_PER_UNIT != 0)
707 {
708 *msg = "structure field is bit field";
709 return true;
710 }
711
712 if (AGGREGATE_TYPE_P (ft) && type_internals_preclude_sra_p (ft, msg))
713 return true;
714 }
715
716 return false;
717
718 case ARRAY_TYPE:
719 et = TREE_TYPE (type);
720
721 if (TYPE_VOLATILE (et))
722 {
723 *msg = "element type is volatile";
724 return true;
725 }
726
727 if (AGGREGATE_TYPE_P (et) && type_internals_preclude_sra_p (et, msg))
728 return true;
729
730 return false;
731
732 default:
733 return false;
734 }
735 }
736
737 /* If T is an SSA_NAME, return NULL if it is not a default def or return its
738 base variable if it is. Return T if it is not an SSA_NAME. */
739
740 static tree
741 get_ssa_base_param (tree t)
742 {
743 if (TREE_CODE (t) == SSA_NAME)
744 {
745 if (SSA_NAME_IS_DEFAULT_DEF (t))
746 return SSA_NAME_VAR (t);
747 else
748 return NULL_TREE;
749 }
750 return t;
751 }
752
753 /* Mark a dereference of BASE of distance DIST in a basic block tht STMT
754 belongs to, unless the BB has already been marked as a potentially
755 final. */
756
757 static void
758 mark_parm_dereference (tree base, HOST_WIDE_INT dist, gimple stmt)
759 {
760 basic_block bb = gimple_bb (stmt);
761 int idx, parm_index = 0;
762 tree parm;
763
764 if (bitmap_bit_p (final_bbs, bb->index))
765 return;
766
767 for (parm = DECL_ARGUMENTS (current_function_decl);
768 parm && parm != base;
769 parm = DECL_CHAIN (parm))
770 parm_index++;
771
772 gcc_assert (parm_index < func_param_count);
773
774 idx = bb->index * func_param_count + parm_index;
775 if (bb_dereferences[idx] < dist)
776 bb_dereferences[idx] = dist;
777 }
778
779 /* Allocate an access structure for BASE, OFFSET and SIZE, clear it, fill in
780 the three fields. Also add it to the vector of accesses corresponding to
781 the base. Finally, return the new access. */
782
783 static struct access *
784 create_access_1 (tree base, HOST_WIDE_INT offset, HOST_WIDE_INT size)
785 {
786 VEC (access_p, heap) *vec;
787 struct access *access;
788 void **slot;
789
790 access = (struct access *) pool_alloc (access_pool);
791 memset (access, 0, sizeof (struct access));
792 access->base = base;
793 access->offset = offset;
794 access->size = size;
795
796 slot = pointer_map_contains (base_access_vec, base);
797 if (slot)
798 vec = (VEC (access_p, heap) *) *slot;
799 else
800 vec = VEC_alloc (access_p, heap, 32);
801
802 VEC_safe_push (access_p, heap, vec, access);
803
804 *((struct VEC (access_p,heap) **)
805 pointer_map_insert (base_access_vec, base)) = vec;
806
807 return access;
808 }
809
810 /* Create and insert access for EXPR. Return created access, or NULL if it is
811 not possible. */
812
813 static struct access *
814 create_access (tree expr, gimple stmt, bool write)
815 {
816 struct access *access;
817 HOST_WIDE_INT offset, size, max_size;
818 tree base = expr;
819 bool ptr, unscalarizable_region = false;
820
821 base = get_ref_base_and_extent (expr, &offset, &size, &max_size);
822
823 if (sra_mode == SRA_MODE_EARLY_IPA
824 && TREE_CODE (base) == MEM_REF)
825 {
826 base = get_ssa_base_param (TREE_OPERAND (base, 0));
827 if (!base)
828 return NULL;
829 ptr = true;
830 }
831 else
832 ptr = false;
833
834 if (!DECL_P (base) || !bitmap_bit_p (candidate_bitmap, DECL_UID (base)))
835 return NULL;
836
837 if (sra_mode == SRA_MODE_EARLY_IPA)
838 {
839 if (size < 0 || size != max_size)
840 {
841 disqualify_candidate (base, "Encountered a variable sized access.");
842 return NULL;
843 }
844 if (TREE_CODE (expr) == COMPONENT_REF
845 && DECL_BIT_FIELD (TREE_OPERAND (expr, 1)))
846 {
847 disqualify_candidate (base, "Encountered a bit-field access.");
848 return NULL;
849 }
850 gcc_checking_assert ((offset % BITS_PER_UNIT) == 0);
851
852 if (ptr)
853 mark_parm_dereference (base, offset + size, stmt);
854 }
855 else
856 {
857 if (size != max_size)
858 {
859 size = max_size;
860 unscalarizable_region = true;
861 }
862 if (size < 0)
863 {
864 disqualify_candidate (base, "Encountered an unconstrained access.");
865 return NULL;
866 }
867 }
868
869 access = create_access_1 (base, offset, size);
870 access->expr = expr;
871 access->type = TREE_TYPE (expr);
872 access->write = write;
873 access->grp_unscalarizable_region = unscalarizable_region;
874 access->stmt = stmt;
875
876 if (TREE_CODE (expr) == COMPONENT_REF
877 && DECL_NONADDRESSABLE_P (TREE_OPERAND (expr, 1)))
878 access->non_addressable = 1;
879
880 return access;
881 }
882
883
884 /* Return true iff TYPE is a RECORD_TYPE with fields that are either of gimple
885 register types or (recursively) records with only these two kinds of fields.
886 It also returns false if any of these records contains a bit-field. */
887
888 static bool
889 type_consists_of_records_p (tree type)
890 {
891 tree fld;
892
893 if (TREE_CODE (type) != RECORD_TYPE)
894 return false;
895
896 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
897 if (TREE_CODE (fld) == FIELD_DECL)
898 {
899 tree ft = TREE_TYPE (fld);
900
901 if (DECL_BIT_FIELD (fld))
902 return false;
903
904 if (!is_gimple_reg_type (ft)
905 && !type_consists_of_records_p (ft))
906 return false;
907 }
908
909 return true;
910 }
911
912 /* Create total_scalarization accesses for all scalar type fields in DECL that
913 must be of a RECORD_TYPE conforming to type_consists_of_records_p. BASE
914 must be the top-most VAR_DECL representing the variable, OFFSET must be the
915 offset of DECL within BASE. REF must be the memory reference expression for
916 the given decl. */
917
918 static void
919 completely_scalarize_record (tree base, tree decl, HOST_WIDE_INT offset,
920 tree ref)
921 {
922 tree fld, decl_type = TREE_TYPE (decl);
923
924 for (fld = TYPE_FIELDS (decl_type); fld; fld = DECL_CHAIN (fld))
925 if (TREE_CODE (fld) == FIELD_DECL)
926 {
927 HOST_WIDE_INT pos = offset + int_bit_position (fld);
928 tree ft = TREE_TYPE (fld);
929 tree nref = build3 (COMPONENT_REF, TREE_TYPE (fld), ref, fld,
930 NULL_TREE);
931
932 if (is_gimple_reg_type (ft))
933 {
934 struct access *access;
935 HOST_WIDE_INT size;
936
937 size = tree_low_cst (DECL_SIZE (fld), 1);
938 access = create_access_1 (base, pos, size);
939 access->expr = nref;
940 access->type = ft;
941 access->grp_total_scalarization = 1;
942 /* Accesses for intraprocedural SRA can have their stmt NULL. */
943 }
944 else
945 completely_scalarize_record (base, fld, pos, nref);
946 }
947 }
948
949 /* Create total_scalarization accesses for all scalar type fields in VAR and
950 for VAR a a whole. VAR must be of a RECORD_TYPE conforming to
951 type_consists_of_records_p. */
952
953 static void
954 completely_scalarize_var (tree var)
955 {
956 HOST_WIDE_INT size = tree_low_cst (DECL_SIZE (var), 1);
957 struct access *access;
958
959 access = create_access_1 (var, 0, size);
960 access->expr = var;
961 access->type = TREE_TYPE (var);
962 access->grp_total_scalarization = 1;
963
964 completely_scalarize_record (var, var, 0, var);
965 }
966
967 /* Search the given tree for a declaration by skipping handled components and
968 exclude it from the candidates. */
969
970 static void
971 disqualify_base_of_expr (tree t, const char *reason)
972 {
973 t = get_base_address (t);
974 if (sra_mode == SRA_MODE_EARLY_IPA
975 && TREE_CODE (t) == MEM_REF)
976 t = get_ssa_base_param (TREE_OPERAND (t, 0));
977
978 if (t && DECL_P (t))
979 disqualify_candidate (t, reason);
980 }
981
982 /* Scan expression EXPR and create access structures for all accesses to
983 candidates for scalarization. Return the created access or NULL if none is
984 created. */
985
986 static struct access *
987 build_access_from_expr_1 (tree expr, gimple stmt, bool write)
988 {
989 struct access *ret = NULL;
990 bool partial_ref;
991
992 if (TREE_CODE (expr) == BIT_FIELD_REF
993 || TREE_CODE (expr) == IMAGPART_EXPR
994 || TREE_CODE (expr) == REALPART_EXPR)
995 {
996 expr = TREE_OPERAND (expr, 0);
997 partial_ref = true;
998 }
999 else
1000 partial_ref = false;
1001
1002 /* We need to dive through V_C_Es in order to get the size of its parameter
1003 and not the result type. Ada produces such statements. We are also
1004 capable of handling the topmost V_C_E but not any of those buried in other
1005 handled components. */
1006 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR)
1007 expr = TREE_OPERAND (expr, 0);
1008
1009 if (contains_view_convert_expr_p (expr))
1010 {
1011 disqualify_base_of_expr (expr, "V_C_E under a different handled "
1012 "component.");
1013 return NULL;
1014 }
1015
1016 switch (TREE_CODE (expr))
1017 {
1018 case MEM_REF:
1019 if (TREE_CODE (TREE_OPERAND (expr, 0)) != ADDR_EXPR
1020 && sra_mode != SRA_MODE_EARLY_IPA)
1021 return NULL;
1022 /* fall through */
1023 case VAR_DECL:
1024 case PARM_DECL:
1025 case RESULT_DECL:
1026 case COMPONENT_REF:
1027 case ARRAY_REF:
1028 case ARRAY_RANGE_REF:
1029 ret = create_access (expr, stmt, write);
1030 break;
1031
1032 default:
1033 break;
1034 }
1035
1036 if (write && partial_ref && ret)
1037 ret->grp_partial_lhs = 1;
1038
1039 return ret;
1040 }
1041
1042 /* Scan expression EXPR and create access structures for all accesses to
1043 candidates for scalarization. Return true if any access has been inserted.
1044 STMT must be the statement from which the expression is taken, WRITE must be
1045 true if the expression is a store and false otherwise. */
1046
1047 static bool
1048 build_access_from_expr (tree expr, gimple stmt, bool write)
1049 {
1050 struct access *access;
1051
1052 access = build_access_from_expr_1 (expr, stmt, write);
1053 if (access)
1054 {
1055 /* This means the aggregate is accesses as a whole in a way other than an
1056 assign statement and thus cannot be removed even if we had a scalar
1057 replacement for everything. */
1058 if (cannot_scalarize_away_bitmap)
1059 bitmap_set_bit (cannot_scalarize_away_bitmap, DECL_UID (access->base));
1060 return true;
1061 }
1062 return false;
1063 }
1064
1065 /* Disqualify LHS and RHS for scalarization if STMT must end its basic block in
1066 modes in which it matters, return true iff they have been disqualified. RHS
1067 may be NULL, in that case ignore it. If we scalarize an aggregate in
1068 intra-SRA we may need to add statements after each statement. This is not
1069 possible if a statement unconditionally has to end the basic block. */
1070 static bool
1071 disqualify_ops_if_throwing_stmt (gimple stmt, tree lhs, tree rhs)
1072 {
1073 if ((sra_mode == SRA_MODE_EARLY_INTRA || sra_mode == SRA_MODE_INTRA)
1074 && (stmt_can_throw_internal (stmt) || stmt_ends_bb_p (stmt)))
1075 {
1076 disqualify_base_of_expr (lhs, "LHS of a throwing stmt.");
1077 if (rhs)
1078 disqualify_base_of_expr (rhs, "RHS of a throwing stmt.");
1079 return true;
1080 }
1081 return false;
1082 }
1083
1084 /* Scan expressions occuring in STMT, create access structures for all accesses
1085 to candidates for scalarization and remove those candidates which occur in
1086 statements or expressions that prevent them from being split apart. Return
1087 true if any access has been inserted. */
1088
1089 static bool
1090 build_accesses_from_assign (gimple stmt)
1091 {
1092 tree lhs, rhs;
1093 struct access *lacc, *racc;
1094
1095 if (!gimple_assign_single_p (stmt)
1096 /* Scope clobbers don't influence scalarization. */
1097 || gimple_clobber_p (stmt))
1098 return false;
1099
1100 lhs = gimple_assign_lhs (stmt);
1101 rhs = gimple_assign_rhs1 (stmt);
1102
1103 if (disqualify_ops_if_throwing_stmt (stmt, lhs, rhs))
1104 return false;
1105
1106 racc = build_access_from_expr_1 (rhs, stmt, false);
1107 lacc = build_access_from_expr_1 (lhs, stmt, true);
1108
1109 if (lacc)
1110 lacc->grp_assignment_write = 1;
1111
1112 if (racc)
1113 {
1114 racc->grp_assignment_read = 1;
1115 if (should_scalarize_away_bitmap && !gimple_has_volatile_ops (stmt)
1116 && !is_gimple_reg_type (racc->type))
1117 bitmap_set_bit (should_scalarize_away_bitmap, DECL_UID (racc->base));
1118 }
1119
1120 if (lacc && racc
1121 && (sra_mode == SRA_MODE_EARLY_INTRA || sra_mode == SRA_MODE_INTRA)
1122 && !lacc->grp_unscalarizable_region
1123 && !racc->grp_unscalarizable_region
1124 && AGGREGATE_TYPE_P (TREE_TYPE (lhs))
1125 && lacc->size == racc->size
1126 && useless_type_conversion_p (lacc->type, racc->type))
1127 {
1128 struct assign_link *link;
1129
1130 link = (struct assign_link *) pool_alloc (link_pool);
1131 memset (link, 0, sizeof (struct assign_link));
1132
1133 link->lacc = lacc;
1134 link->racc = racc;
1135
1136 add_link_to_rhs (racc, link);
1137 }
1138
1139 return lacc || racc;
1140 }
1141
1142 /* Callback of walk_stmt_load_store_addr_ops visit_addr used to determine
1143 GIMPLE_ASM operands with memory constrains which cannot be scalarized. */
1144
1145 static bool
1146 asm_visit_addr (gimple stmt ATTRIBUTE_UNUSED, tree op,
1147 void *data ATTRIBUTE_UNUSED)
1148 {
1149 op = get_base_address (op);
1150 if (op
1151 && DECL_P (op))
1152 disqualify_candidate (op, "Non-scalarizable GIMPLE_ASM operand.");
1153
1154 return false;
1155 }
1156
1157 /* Return true iff callsite CALL has at least as many actual arguments as there
1158 are formal parameters of the function currently processed by IPA-SRA. */
1159
1160 static inline bool
1161 callsite_has_enough_arguments_p (gimple call)
1162 {
1163 return gimple_call_num_args (call) >= (unsigned) func_param_count;
1164 }
1165
1166 /* Scan function and look for interesting expressions and create access
1167 structures for them. Return true iff any access is created. */
1168
1169 static bool
1170 scan_function (void)
1171 {
1172 basic_block bb;
1173 bool ret = false;
1174
1175 FOR_EACH_BB (bb)
1176 {
1177 gimple_stmt_iterator gsi;
1178 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1179 {
1180 gimple stmt = gsi_stmt (gsi);
1181 tree t;
1182 unsigned i;
1183
1184 if (final_bbs && stmt_can_throw_external (stmt))
1185 bitmap_set_bit (final_bbs, bb->index);
1186 switch (gimple_code (stmt))
1187 {
1188 case GIMPLE_RETURN:
1189 t = gimple_return_retval (stmt);
1190 if (t != NULL_TREE)
1191 ret |= build_access_from_expr (t, stmt, false);
1192 if (final_bbs)
1193 bitmap_set_bit (final_bbs, bb->index);
1194 break;
1195
1196 case GIMPLE_ASSIGN:
1197 ret |= build_accesses_from_assign (stmt);
1198 break;
1199
1200 case GIMPLE_CALL:
1201 for (i = 0; i < gimple_call_num_args (stmt); i++)
1202 ret |= build_access_from_expr (gimple_call_arg (stmt, i),
1203 stmt, false);
1204
1205 if (sra_mode == SRA_MODE_EARLY_IPA)
1206 {
1207 tree dest = gimple_call_fndecl (stmt);
1208 int flags = gimple_call_flags (stmt);
1209
1210 if (dest)
1211 {
1212 if (DECL_BUILT_IN_CLASS (dest) == BUILT_IN_NORMAL
1213 && DECL_FUNCTION_CODE (dest) == BUILT_IN_APPLY_ARGS)
1214 encountered_apply_args = true;
1215 if (cgraph_get_node (dest)
1216 == cgraph_get_node (current_function_decl))
1217 {
1218 encountered_recursive_call = true;
1219 if (!callsite_has_enough_arguments_p (stmt))
1220 encountered_unchangable_recursive_call = true;
1221 }
1222 }
1223
1224 if (final_bbs
1225 && (flags & (ECF_CONST | ECF_PURE)) == 0)
1226 bitmap_set_bit (final_bbs, bb->index);
1227 }
1228
1229 t = gimple_call_lhs (stmt);
1230 if (t && !disqualify_ops_if_throwing_stmt (stmt, t, NULL))
1231 ret |= build_access_from_expr (t, stmt, true);
1232 break;
1233
1234 case GIMPLE_ASM:
1235 walk_stmt_load_store_addr_ops (stmt, NULL, NULL, NULL,
1236 asm_visit_addr);
1237 if (final_bbs)
1238 bitmap_set_bit (final_bbs, bb->index);
1239
1240 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
1241 {
1242 t = TREE_VALUE (gimple_asm_input_op (stmt, i));
1243 ret |= build_access_from_expr (t, stmt, false);
1244 }
1245 for (i = 0; i < gimple_asm_noutputs (stmt); i++)
1246 {
1247 t = TREE_VALUE (gimple_asm_output_op (stmt, i));
1248 ret |= build_access_from_expr (t, stmt, true);
1249 }
1250 break;
1251
1252 default:
1253 break;
1254 }
1255 }
1256 }
1257
1258 return ret;
1259 }
1260
1261 /* Helper of QSORT function. There are pointers to accesses in the array. An
1262 access is considered smaller than another if it has smaller offset or if the
1263 offsets are the same but is size is bigger. */
1264
1265 static int
1266 compare_access_positions (const void *a, const void *b)
1267 {
1268 const access_p *fp1 = (const access_p *) a;
1269 const access_p *fp2 = (const access_p *) b;
1270 const access_p f1 = *fp1;
1271 const access_p f2 = *fp2;
1272
1273 if (f1->offset != f2->offset)
1274 return f1->offset < f2->offset ? -1 : 1;
1275
1276 if (f1->size == f2->size)
1277 {
1278 if (f1->type == f2->type)
1279 return 0;
1280 /* Put any non-aggregate type before any aggregate type. */
1281 else if (!is_gimple_reg_type (f1->type)
1282 && is_gimple_reg_type (f2->type))
1283 return 1;
1284 else if (is_gimple_reg_type (f1->type)
1285 && !is_gimple_reg_type (f2->type))
1286 return -1;
1287 /* Put any complex or vector type before any other scalar type. */
1288 else if (TREE_CODE (f1->type) != COMPLEX_TYPE
1289 && TREE_CODE (f1->type) != VECTOR_TYPE
1290 && (TREE_CODE (f2->type) == COMPLEX_TYPE
1291 || TREE_CODE (f2->type) == VECTOR_TYPE))
1292 return 1;
1293 else if ((TREE_CODE (f1->type) == COMPLEX_TYPE
1294 || TREE_CODE (f1->type) == VECTOR_TYPE)
1295 && TREE_CODE (f2->type) != COMPLEX_TYPE
1296 && TREE_CODE (f2->type) != VECTOR_TYPE)
1297 return -1;
1298 /* Put the integral type with the bigger precision first. */
1299 else if (INTEGRAL_TYPE_P (f1->type)
1300 && INTEGRAL_TYPE_P (f2->type))
1301 return TYPE_PRECISION (f2->type) - TYPE_PRECISION (f1->type);
1302 /* Put any integral type with non-full precision last. */
1303 else if (INTEGRAL_TYPE_P (f1->type)
1304 && (TREE_INT_CST_LOW (TYPE_SIZE (f1->type))
1305 != TYPE_PRECISION (f1->type)))
1306 return 1;
1307 else if (INTEGRAL_TYPE_P (f2->type)
1308 && (TREE_INT_CST_LOW (TYPE_SIZE (f2->type))
1309 != TYPE_PRECISION (f2->type)))
1310 return -1;
1311 /* Stabilize the sort. */
1312 return TYPE_UID (f1->type) - TYPE_UID (f2->type);
1313 }
1314
1315 /* We want the bigger accesses first, thus the opposite operator in the next
1316 line: */
1317 return f1->size > f2->size ? -1 : 1;
1318 }
1319
1320
1321 /* Append a name of the declaration to the name obstack. A helper function for
1322 make_fancy_name. */
1323
1324 static void
1325 make_fancy_decl_name (tree decl)
1326 {
1327 char buffer[32];
1328
1329 tree name = DECL_NAME (decl);
1330 if (name)
1331 obstack_grow (&name_obstack, IDENTIFIER_POINTER (name),
1332 IDENTIFIER_LENGTH (name));
1333 else
1334 {
1335 sprintf (buffer, "D%u", DECL_UID (decl));
1336 obstack_grow (&name_obstack, buffer, strlen (buffer));
1337 }
1338 }
1339
1340 /* Helper for make_fancy_name. */
1341
1342 static void
1343 make_fancy_name_1 (tree expr)
1344 {
1345 char buffer[32];
1346 tree index;
1347
1348 if (DECL_P (expr))
1349 {
1350 make_fancy_decl_name (expr);
1351 return;
1352 }
1353
1354 switch (TREE_CODE (expr))
1355 {
1356 case COMPONENT_REF:
1357 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1358 obstack_1grow (&name_obstack, '$');
1359 make_fancy_decl_name (TREE_OPERAND (expr, 1));
1360 break;
1361
1362 case ARRAY_REF:
1363 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1364 obstack_1grow (&name_obstack, '$');
1365 /* Arrays with only one element may not have a constant as their
1366 index. */
1367 index = TREE_OPERAND (expr, 1);
1368 if (TREE_CODE (index) != INTEGER_CST)
1369 break;
1370 sprintf (buffer, HOST_WIDE_INT_PRINT_DEC, TREE_INT_CST_LOW (index));
1371 obstack_grow (&name_obstack, buffer, strlen (buffer));
1372 break;
1373
1374 case ADDR_EXPR:
1375 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1376 break;
1377
1378 case MEM_REF:
1379 make_fancy_name_1 (TREE_OPERAND (expr, 0));
1380 if (!integer_zerop (TREE_OPERAND (expr, 1)))
1381 {
1382 obstack_1grow (&name_obstack, '$');
1383 sprintf (buffer, HOST_WIDE_INT_PRINT_DEC,
1384 TREE_INT_CST_LOW (TREE_OPERAND (expr, 1)));
1385 obstack_grow (&name_obstack, buffer, strlen (buffer));
1386 }
1387 break;
1388
1389 case BIT_FIELD_REF:
1390 case REALPART_EXPR:
1391 case IMAGPART_EXPR:
1392 gcc_unreachable (); /* we treat these as scalars. */
1393 break;
1394 default:
1395 break;
1396 }
1397 }
1398
1399 /* Create a human readable name for replacement variable of ACCESS. */
1400
1401 static char *
1402 make_fancy_name (tree expr)
1403 {
1404 make_fancy_name_1 (expr);
1405 obstack_1grow (&name_obstack, '\0');
1406 return XOBFINISH (&name_obstack, char *);
1407 }
1408
1409 /* Construct a MEM_REF that would reference a part of aggregate BASE of type
1410 EXP_TYPE at the given OFFSET. If BASE is something for which
1411 get_addr_base_and_unit_offset returns NULL, gsi must be non-NULL and is used
1412 to insert new statements either before or below the current one as specified
1413 by INSERT_AFTER. This function is not capable of handling bitfields. */
1414
1415 tree
1416 build_ref_for_offset (location_t loc, tree base, HOST_WIDE_INT offset,
1417 tree exp_type, gimple_stmt_iterator *gsi,
1418 bool insert_after)
1419 {
1420 tree prev_base = base;
1421 tree off;
1422 HOST_WIDE_INT base_offset;
1423 unsigned HOST_WIDE_INT misalign;
1424 unsigned int align;
1425
1426 gcc_checking_assert (offset % BITS_PER_UNIT == 0);
1427
1428 base = get_addr_base_and_unit_offset (base, &base_offset);
1429
1430 /* get_addr_base_and_unit_offset returns NULL for references with a variable
1431 offset such as array[var_index]. */
1432 if (!base)
1433 {
1434 gimple stmt;
1435 tree tmp, addr;
1436
1437 gcc_checking_assert (gsi);
1438 tmp = create_tmp_reg (build_pointer_type (TREE_TYPE (prev_base)), NULL);
1439 add_referenced_var (tmp);
1440 tmp = make_ssa_name (tmp, NULL);
1441 addr = build_fold_addr_expr (unshare_expr (prev_base));
1442 STRIP_USELESS_TYPE_CONVERSION (addr);
1443 stmt = gimple_build_assign (tmp, addr);
1444 gimple_set_location (stmt, loc);
1445 SSA_NAME_DEF_STMT (tmp) = stmt;
1446 if (insert_after)
1447 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
1448 else
1449 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
1450 update_stmt (stmt);
1451
1452 off = build_int_cst (reference_alias_ptr_type (prev_base),
1453 offset / BITS_PER_UNIT);
1454 base = tmp;
1455 }
1456 else if (TREE_CODE (base) == MEM_REF)
1457 {
1458 off = build_int_cst (TREE_TYPE (TREE_OPERAND (base, 1)),
1459 base_offset + offset / BITS_PER_UNIT);
1460 off = int_const_binop (PLUS_EXPR, TREE_OPERAND (base, 1), off);
1461 base = unshare_expr (TREE_OPERAND (base, 0));
1462 }
1463 else
1464 {
1465 off = build_int_cst (reference_alias_ptr_type (base),
1466 base_offset + offset / BITS_PER_UNIT);
1467 base = build_fold_addr_expr (unshare_expr (base));
1468 }
1469
1470 /* If prev_base were always an originally performed access
1471 we can extract more optimistic alignment information
1472 by looking at the access mode. That would constrain the
1473 alignment of base + base_offset which we would need to
1474 adjust according to offset. */
1475 if (!get_pointer_alignment_1 (base, &align, &misalign))
1476 {
1477 gcc_assert (misalign == 0);
1478 if (TREE_CODE (prev_base) == MEM_REF
1479 || TREE_CODE (prev_base) == TARGET_MEM_REF)
1480 align = TYPE_ALIGN (TREE_TYPE (prev_base));
1481 }
1482 misalign += (double_int_sext (tree_to_double_int (off),
1483 TYPE_PRECISION (TREE_TYPE (off))).low
1484 * BITS_PER_UNIT);
1485 misalign = misalign & (align - 1);
1486 if (misalign != 0)
1487 align = (misalign & -misalign);
1488 if (align < TYPE_ALIGN (exp_type))
1489 exp_type = build_aligned_type (exp_type, align);
1490
1491 return fold_build2_loc (loc, MEM_REF, exp_type, base, off);
1492 }
1493
1494 /* Construct a memory reference to a part of an aggregate BASE at the given
1495 OFFSET and of the same type as MODEL. In case this is a reference to a
1496 bit-field, the function will replicate the last component_ref of model's
1497 expr to access it. GSI and INSERT_AFTER have the same meaning as in
1498 build_ref_for_offset. */
1499
1500 static tree
1501 build_ref_for_model (location_t loc, tree base, HOST_WIDE_INT offset,
1502 struct access *model, gimple_stmt_iterator *gsi,
1503 bool insert_after)
1504 {
1505 if (TREE_CODE (model->expr) == COMPONENT_REF
1506 && DECL_BIT_FIELD (TREE_OPERAND (model->expr, 1)))
1507 {
1508 /* This access represents a bit-field. */
1509 tree t, exp_type, fld = TREE_OPERAND (model->expr, 1);
1510
1511 offset -= int_bit_position (fld);
1512 exp_type = TREE_TYPE (TREE_OPERAND (model->expr, 0));
1513 t = build_ref_for_offset (loc, base, offset, exp_type, gsi, insert_after);
1514 return fold_build3_loc (loc, COMPONENT_REF, TREE_TYPE (fld), t, fld,
1515 NULL_TREE);
1516 }
1517 else
1518 return build_ref_for_offset (loc, base, offset, model->type,
1519 gsi, insert_after);
1520 }
1521
1522 /* Construct a memory reference consisting of component_refs and array_refs to
1523 a part of an aggregate *RES (which is of type TYPE). The requested part
1524 should have type EXP_TYPE at be the given OFFSET. This function might not
1525 succeed, it returns true when it does and only then *RES points to something
1526 meaningful. This function should be used only to build expressions that we
1527 might need to present to user (e.g. in warnings). In all other situations,
1528 build_ref_for_model or build_ref_for_offset should be used instead. */
1529
1530 static bool
1531 build_user_friendly_ref_for_offset (tree *res, tree type, HOST_WIDE_INT offset,
1532 tree exp_type)
1533 {
1534 while (1)
1535 {
1536 tree fld;
1537 tree tr_size, index, minidx;
1538 HOST_WIDE_INT el_size;
1539
1540 if (offset == 0 && exp_type
1541 && types_compatible_p (exp_type, type))
1542 return true;
1543
1544 switch (TREE_CODE (type))
1545 {
1546 case UNION_TYPE:
1547 case QUAL_UNION_TYPE:
1548 case RECORD_TYPE:
1549 for (fld = TYPE_FIELDS (type); fld; fld = DECL_CHAIN (fld))
1550 {
1551 HOST_WIDE_INT pos, size;
1552 tree expr, *expr_ptr;
1553
1554 if (TREE_CODE (fld) != FIELD_DECL)
1555 continue;
1556
1557 pos = int_bit_position (fld);
1558 gcc_assert (TREE_CODE (type) == RECORD_TYPE || pos == 0);
1559 tr_size = DECL_SIZE (fld);
1560 if (!tr_size || !host_integerp (tr_size, 1))
1561 continue;
1562 size = tree_low_cst (tr_size, 1);
1563 if (size == 0)
1564 {
1565 if (pos != offset)
1566 continue;
1567 }
1568 else if (pos > offset || (pos + size) <= offset)
1569 continue;
1570
1571 expr = build3 (COMPONENT_REF, TREE_TYPE (fld), *res, fld,
1572 NULL_TREE);
1573 expr_ptr = &expr;
1574 if (build_user_friendly_ref_for_offset (expr_ptr, TREE_TYPE (fld),
1575 offset - pos, exp_type))
1576 {
1577 *res = expr;
1578 return true;
1579 }
1580 }
1581 return false;
1582
1583 case ARRAY_TYPE:
1584 tr_size = TYPE_SIZE (TREE_TYPE (type));
1585 if (!tr_size || !host_integerp (tr_size, 1))
1586 return false;
1587 el_size = tree_low_cst (tr_size, 1);
1588
1589 minidx = TYPE_MIN_VALUE (TYPE_DOMAIN (type));
1590 if (TREE_CODE (minidx) != INTEGER_CST || el_size == 0)
1591 return false;
1592 index = build_int_cst (TYPE_DOMAIN (type), offset / el_size);
1593 if (!integer_zerop (minidx))
1594 index = int_const_binop (PLUS_EXPR, index, minidx);
1595 *res = build4 (ARRAY_REF, TREE_TYPE (type), *res, index,
1596 NULL_TREE, NULL_TREE);
1597 offset = offset % el_size;
1598 type = TREE_TYPE (type);
1599 break;
1600
1601 default:
1602 if (offset != 0)
1603 return false;
1604
1605 if (exp_type)
1606 return false;
1607 else
1608 return true;
1609 }
1610 }
1611 }
1612
1613 /* Return true iff TYPE is stdarg va_list type. */
1614
1615 static inline bool
1616 is_va_list_type (tree type)
1617 {
1618 return TYPE_MAIN_VARIANT (type) == TYPE_MAIN_VARIANT (va_list_type_node);
1619 }
1620
1621 /* Print message to dump file why a variable was rejected. */
1622
1623 static void
1624 reject (tree var, const char *msg)
1625 {
1626 if (dump_file && (dump_flags & TDF_DETAILS))
1627 {
1628 fprintf (dump_file, "Rejected (%d): %s: ", DECL_UID (var), msg);
1629 print_generic_expr (dump_file, var, 0);
1630 fprintf (dump_file, "\n");
1631 }
1632 }
1633
1634 /* The very first phase of intraprocedural SRA. It marks in candidate_bitmap
1635 those with type which is suitable for scalarization. */
1636
1637 static bool
1638 find_var_candidates (void)
1639 {
1640 tree var, type;
1641 referenced_var_iterator rvi;
1642 bool ret = false;
1643 const char *msg;
1644
1645 FOR_EACH_REFERENCED_VAR (cfun, var, rvi)
1646 {
1647 if (TREE_CODE (var) != VAR_DECL && TREE_CODE (var) != PARM_DECL)
1648 continue;
1649 type = TREE_TYPE (var);
1650
1651 if (!AGGREGATE_TYPE_P (type))
1652 {
1653 reject (var, "not aggregate");
1654 continue;
1655 }
1656 if (needs_to_live_in_memory (var))
1657 {
1658 reject (var, "needs to live in memory");
1659 continue;
1660 }
1661 if (TREE_THIS_VOLATILE (var))
1662 {
1663 reject (var, "is volatile");
1664 continue;
1665 }
1666 if (!COMPLETE_TYPE_P (type))
1667 {
1668 reject (var, "has incomplete type");
1669 continue;
1670 }
1671 if (!host_integerp (TYPE_SIZE (type), 1))
1672 {
1673 reject (var, "type size not fixed");
1674 continue;
1675 }
1676 if (tree_low_cst (TYPE_SIZE (type), 1) == 0)
1677 {
1678 reject (var, "type size is zero");
1679 continue;
1680 }
1681 if (type_internals_preclude_sra_p (type, &msg))
1682 {
1683 reject (var, msg);
1684 continue;
1685 }
1686 if (/* Fix for PR 41089. tree-stdarg.c needs to have va_lists intact but
1687 we also want to schedule it rather late. Thus we ignore it in
1688 the early pass. */
1689 (sra_mode == SRA_MODE_EARLY_INTRA
1690 && is_va_list_type (type)))
1691 {
1692 reject (var, "is va_list");
1693 continue;
1694 }
1695
1696 bitmap_set_bit (candidate_bitmap, DECL_UID (var));
1697
1698 if (dump_file && (dump_flags & TDF_DETAILS))
1699 {
1700 fprintf (dump_file, "Candidate (%d): ", DECL_UID (var));
1701 print_generic_expr (dump_file, var, 0);
1702 fprintf (dump_file, "\n");
1703 }
1704 ret = true;
1705 }
1706
1707 return ret;
1708 }
1709
1710 /* Sort all accesses for the given variable, check for partial overlaps and
1711 return NULL if there are any. If there are none, pick a representative for
1712 each combination of offset and size and create a linked list out of them.
1713 Return the pointer to the first representative and make sure it is the first
1714 one in the vector of accesses. */
1715
1716 static struct access *
1717 sort_and_splice_var_accesses (tree var)
1718 {
1719 int i, j, access_count;
1720 struct access *res, **prev_acc_ptr = &res;
1721 VEC (access_p, heap) *access_vec;
1722 bool first = true;
1723 HOST_WIDE_INT low = -1, high = 0;
1724
1725 access_vec = get_base_access_vector (var);
1726 if (!access_vec)
1727 return NULL;
1728 access_count = VEC_length (access_p, access_vec);
1729
1730 /* Sort by <OFFSET, SIZE>. */
1731 VEC_qsort (access_p, access_vec, compare_access_positions);
1732
1733 i = 0;
1734 while (i < access_count)
1735 {
1736 struct access *access = VEC_index (access_p, access_vec, i);
1737 bool grp_write = access->write;
1738 bool grp_read = !access->write;
1739 bool grp_scalar_write = access->write
1740 && is_gimple_reg_type (access->type);
1741 bool grp_scalar_read = !access->write
1742 && is_gimple_reg_type (access->type);
1743 bool grp_assignment_read = access->grp_assignment_read;
1744 bool grp_assignment_write = access->grp_assignment_write;
1745 bool multiple_scalar_reads = false;
1746 bool total_scalarization = access->grp_total_scalarization;
1747 bool grp_partial_lhs = access->grp_partial_lhs;
1748 bool first_scalar = is_gimple_reg_type (access->type);
1749 bool unscalarizable_region = access->grp_unscalarizable_region;
1750
1751 if (first || access->offset >= high)
1752 {
1753 first = false;
1754 low = access->offset;
1755 high = access->offset + access->size;
1756 }
1757 else if (access->offset > low && access->offset + access->size > high)
1758 return NULL;
1759 else
1760 gcc_assert (access->offset >= low
1761 && access->offset + access->size <= high);
1762
1763 j = i + 1;
1764 while (j < access_count)
1765 {
1766 struct access *ac2 = VEC_index (access_p, access_vec, j);
1767 if (ac2->offset != access->offset || ac2->size != access->size)
1768 break;
1769 if (ac2->write)
1770 {
1771 grp_write = true;
1772 grp_scalar_write = (grp_scalar_write
1773 || is_gimple_reg_type (ac2->type));
1774 }
1775 else
1776 {
1777 grp_read = true;
1778 if (is_gimple_reg_type (ac2->type))
1779 {
1780 if (grp_scalar_read)
1781 multiple_scalar_reads = true;
1782 else
1783 grp_scalar_read = true;
1784 }
1785 }
1786 grp_assignment_read |= ac2->grp_assignment_read;
1787 grp_assignment_write |= ac2->grp_assignment_write;
1788 grp_partial_lhs |= ac2->grp_partial_lhs;
1789 unscalarizable_region |= ac2->grp_unscalarizable_region;
1790 total_scalarization |= ac2->grp_total_scalarization;
1791 relink_to_new_repr (access, ac2);
1792
1793 /* If there are both aggregate-type and scalar-type accesses with
1794 this combination of size and offset, the comparison function
1795 should have put the scalars first. */
1796 gcc_assert (first_scalar || !is_gimple_reg_type (ac2->type));
1797 ac2->group_representative = access;
1798 j++;
1799 }
1800
1801 i = j;
1802
1803 access->group_representative = access;
1804 access->grp_write = grp_write;
1805 access->grp_read = grp_read;
1806 access->grp_scalar_read = grp_scalar_read;
1807 access->grp_scalar_write = grp_scalar_write;
1808 access->grp_assignment_read = grp_assignment_read;
1809 access->grp_assignment_write = grp_assignment_write;
1810 access->grp_hint = multiple_scalar_reads || total_scalarization;
1811 access->grp_total_scalarization = total_scalarization;
1812 access->grp_partial_lhs = grp_partial_lhs;
1813 access->grp_unscalarizable_region = unscalarizable_region;
1814 if (access->first_link)
1815 add_access_to_work_queue (access);
1816
1817 *prev_acc_ptr = access;
1818 prev_acc_ptr = &access->next_grp;
1819 }
1820
1821 gcc_assert (res == VEC_index (access_p, access_vec, 0));
1822 return res;
1823 }
1824
1825 /* Create a variable for the given ACCESS which determines the type, name and a
1826 few other properties. Return the variable declaration and store it also to
1827 ACCESS->replacement. */
1828
1829 static tree
1830 create_access_replacement (struct access *access, bool rename)
1831 {
1832 tree repl;
1833
1834 repl = create_tmp_var (access->type, "SR");
1835 add_referenced_var (repl);
1836 if (!access->grp_partial_lhs
1837 && rename)
1838 mark_sym_for_renaming (repl);
1839
1840 if (TREE_CODE (access->type) == COMPLEX_TYPE
1841 || TREE_CODE (access->type) == VECTOR_TYPE)
1842 {
1843 if (!access->grp_partial_lhs)
1844 DECL_GIMPLE_REG_P (repl) = 1;
1845 }
1846 else if (access->grp_partial_lhs
1847 && is_gimple_reg_type (access->type))
1848 TREE_ADDRESSABLE (repl) = 1;
1849
1850 DECL_SOURCE_LOCATION (repl) = DECL_SOURCE_LOCATION (access->base);
1851 DECL_ARTIFICIAL (repl) = 1;
1852 DECL_IGNORED_P (repl) = DECL_IGNORED_P (access->base);
1853
1854 if (DECL_NAME (access->base)
1855 && !DECL_IGNORED_P (access->base)
1856 && !DECL_ARTIFICIAL (access->base))
1857 {
1858 char *pretty_name = make_fancy_name (access->expr);
1859 tree debug_expr = unshare_expr (access->expr), d;
1860
1861 DECL_NAME (repl) = get_identifier (pretty_name);
1862 obstack_free (&name_obstack, pretty_name);
1863
1864 /* Get rid of any SSA_NAMEs embedded in debug_expr,
1865 as DECL_DEBUG_EXPR isn't considered when looking for still
1866 used SSA_NAMEs and thus they could be freed. All debug info
1867 generation cares is whether something is constant or variable
1868 and that get_ref_base_and_extent works properly on the
1869 expression. */
1870 for (d = debug_expr; handled_component_p (d); d = TREE_OPERAND (d, 0))
1871 switch (TREE_CODE (d))
1872 {
1873 case ARRAY_REF:
1874 case ARRAY_RANGE_REF:
1875 if (TREE_OPERAND (d, 1)
1876 && TREE_CODE (TREE_OPERAND (d, 1)) == SSA_NAME)
1877 TREE_OPERAND (d, 1) = SSA_NAME_VAR (TREE_OPERAND (d, 1));
1878 if (TREE_OPERAND (d, 3)
1879 && TREE_CODE (TREE_OPERAND (d, 3)) == SSA_NAME)
1880 TREE_OPERAND (d, 3) = SSA_NAME_VAR (TREE_OPERAND (d, 3));
1881 /* FALLTHRU */
1882 case COMPONENT_REF:
1883 if (TREE_OPERAND (d, 2)
1884 && TREE_CODE (TREE_OPERAND (d, 2)) == SSA_NAME)
1885 TREE_OPERAND (d, 2) = SSA_NAME_VAR (TREE_OPERAND (d, 2));
1886 break;
1887 default:
1888 break;
1889 }
1890 SET_DECL_DEBUG_EXPR (repl, debug_expr);
1891 DECL_DEBUG_EXPR_IS_FROM (repl) = 1;
1892 if (access->grp_no_warning)
1893 TREE_NO_WARNING (repl) = 1;
1894 else
1895 TREE_NO_WARNING (repl) = TREE_NO_WARNING (access->base);
1896 }
1897 else
1898 TREE_NO_WARNING (repl) = 1;
1899
1900 if (dump_file)
1901 {
1902 fprintf (dump_file, "Created a replacement for ");
1903 print_generic_expr (dump_file, access->base, 0);
1904 fprintf (dump_file, " offset: %u, size: %u: ",
1905 (unsigned) access->offset, (unsigned) access->size);
1906 print_generic_expr (dump_file, repl, 0);
1907 fprintf (dump_file, "\n");
1908 }
1909 sra_stats.replacements++;
1910
1911 return repl;
1912 }
1913
1914 /* Return ACCESS scalar replacement, create it if it does not exist yet. */
1915
1916 static inline tree
1917 get_access_replacement (struct access *access)
1918 {
1919 gcc_assert (access->grp_to_be_replaced);
1920
1921 if (!access->replacement_decl)
1922 access->replacement_decl = create_access_replacement (access, true);
1923 return access->replacement_decl;
1924 }
1925
1926 /* Return ACCESS scalar replacement, create it if it does not exist yet but do
1927 not mark it for renaming. */
1928
1929 static inline tree
1930 get_unrenamed_access_replacement (struct access *access)
1931 {
1932 gcc_assert (!access->grp_to_be_replaced);
1933
1934 if (!access->replacement_decl)
1935 access->replacement_decl = create_access_replacement (access, false);
1936 return access->replacement_decl;
1937 }
1938
1939
1940 /* Build a subtree of accesses rooted in *ACCESS, and move the pointer in the
1941 linked list along the way. Stop when *ACCESS is NULL or the access pointed
1942 to it is not "within" the root. Return false iff some accesses partially
1943 overlap. */
1944
1945 static bool
1946 build_access_subtree (struct access **access)
1947 {
1948 struct access *root = *access, *last_child = NULL;
1949 HOST_WIDE_INT limit = root->offset + root->size;
1950
1951 *access = (*access)->next_grp;
1952 while (*access && (*access)->offset + (*access)->size <= limit)
1953 {
1954 if (!last_child)
1955 root->first_child = *access;
1956 else
1957 last_child->next_sibling = *access;
1958 last_child = *access;
1959
1960 if (!build_access_subtree (access))
1961 return false;
1962 }
1963
1964 if (*access && (*access)->offset < limit)
1965 return false;
1966
1967 return true;
1968 }
1969
1970 /* Build a tree of access representatives, ACCESS is the pointer to the first
1971 one, others are linked in a list by the next_grp field. Return false iff
1972 some accesses partially overlap. */
1973
1974 static bool
1975 build_access_trees (struct access *access)
1976 {
1977 while (access)
1978 {
1979 struct access *root = access;
1980
1981 if (!build_access_subtree (&access))
1982 return false;
1983 root->next_grp = access;
1984 }
1985 return true;
1986 }
1987
1988 /* Return true if expr contains some ARRAY_REFs into a variable bounded
1989 array. */
1990
1991 static bool
1992 expr_with_var_bounded_array_refs_p (tree expr)
1993 {
1994 while (handled_component_p (expr))
1995 {
1996 if (TREE_CODE (expr) == ARRAY_REF
1997 && !host_integerp (array_ref_low_bound (expr), 0))
1998 return true;
1999 expr = TREE_OPERAND (expr, 0);
2000 }
2001 return false;
2002 }
2003
2004 /* Analyze the subtree of accesses rooted in ROOT, scheduling replacements when
2005 both seeming beneficial and when ALLOW_REPLACEMENTS allows it. Also set all
2006 sorts of access flags appropriately along the way, notably always set
2007 grp_read and grp_assign_read according to MARK_READ and grp_write when
2008 MARK_WRITE is true.
2009
2010 Creating a replacement for a scalar access is considered beneficial if its
2011 grp_hint is set (this means we are either attempting total scalarization or
2012 there is more than one direct read access) or according to the following
2013 table:
2014
2015 Access written to through a scalar type (once or more times)
2016 |
2017 | Written to in an assignment statement
2018 | |
2019 | | Access read as scalar _once_
2020 | | |
2021 | | | Read in an assignment statement
2022 | | | |
2023 | | | | Scalarize Comment
2024 -----------------------------------------------------------------------------
2025 0 0 0 0 No access for the scalar
2026 0 0 0 1 No access for the scalar
2027 0 0 1 0 No Single read - won't help
2028 0 0 1 1 No The same case
2029 0 1 0 0 No access for the scalar
2030 0 1 0 1 No access for the scalar
2031 0 1 1 0 Yes s = *g; return s.i;
2032 0 1 1 1 Yes The same case as above
2033 1 0 0 0 No Won't help
2034 1 0 0 1 Yes s.i = 1; *g = s;
2035 1 0 1 0 Yes s.i = 5; g = s.i;
2036 1 0 1 1 Yes The same case as above
2037 1 1 0 0 No Won't help.
2038 1 1 0 1 Yes s.i = 1; *g = s;
2039 1 1 1 0 Yes s = *g; return s.i;
2040 1 1 1 1 Yes Any of the above yeses */
2041
2042 static bool
2043 analyze_access_subtree (struct access *root, struct access *parent,
2044 bool allow_replacements)
2045 {
2046 struct access *child;
2047 HOST_WIDE_INT limit = root->offset + root->size;
2048 HOST_WIDE_INT covered_to = root->offset;
2049 bool scalar = is_gimple_reg_type (root->type);
2050 bool hole = false, sth_created = false;
2051
2052 if (parent)
2053 {
2054 if (parent->grp_read)
2055 root->grp_read = 1;
2056 if (parent->grp_assignment_read)
2057 root->grp_assignment_read = 1;
2058 if (parent->grp_write)
2059 root->grp_write = 1;
2060 if (parent->grp_assignment_write)
2061 root->grp_assignment_write = 1;
2062 if (parent->grp_total_scalarization)
2063 root->grp_total_scalarization = 1;
2064 }
2065
2066 if (root->grp_unscalarizable_region)
2067 allow_replacements = false;
2068
2069 if (allow_replacements && expr_with_var_bounded_array_refs_p (root->expr))
2070 allow_replacements = false;
2071
2072 for (child = root->first_child; child; child = child->next_sibling)
2073 {
2074 hole |= covered_to < child->offset;
2075 sth_created |= analyze_access_subtree (child, root,
2076 allow_replacements && !scalar);
2077
2078 root->grp_unscalarized_data |= child->grp_unscalarized_data;
2079 root->grp_total_scalarization &= child->grp_total_scalarization;
2080 if (child->grp_covered)
2081 covered_to += child->size;
2082 else
2083 hole = true;
2084 }
2085
2086 if (allow_replacements && scalar && !root->first_child
2087 && (root->grp_hint
2088 || ((root->grp_scalar_read || root->grp_assignment_read)
2089 && (root->grp_scalar_write || root->grp_assignment_write))))
2090 {
2091 bool new_integer_type;
2092 /* Always create access replacements that cover the whole access.
2093 For integral types this means the precision has to match.
2094 Avoid assumptions based on the integral type kind, too. */
2095 if (INTEGRAL_TYPE_P (root->type)
2096 && (TREE_CODE (root->type) != INTEGER_TYPE
2097 || TYPE_PRECISION (root->type) != root->size)
2098 /* But leave bitfield accesses alone. */
2099 && (TREE_CODE (root->expr) != COMPONENT_REF
2100 || !DECL_BIT_FIELD (TREE_OPERAND (root->expr, 1))))
2101 {
2102 tree rt = root->type;
2103 gcc_assert ((root->offset % BITS_PER_UNIT) == 0
2104 && (root->size % BITS_PER_UNIT) == 0);
2105 root->type = build_nonstandard_integer_type (root->size,
2106 TYPE_UNSIGNED (rt));
2107 root->expr = build_ref_for_offset (UNKNOWN_LOCATION,
2108 root->base, root->offset,
2109 root->type, NULL, false);
2110 new_integer_type = true;
2111 }
2112 else
2113 new_integer_type = false;
2114
2115 if (dump_file && (dump_flags & TDF_DETAILS))
2116 {
2117 fprintf (dump_file, "Marking ");
2118 print_generic_expr (dump_file, root->base, 0);
2119 fprintf (dump_file, " offset: %u, size: %u ",
2120 (unsigned) root->offset, (unsigned) root->size);
2121 fprintf (dump_file, " to be replaced%s.\n",
2122 new_integer_type ? " with an integer": "");
2123 }
2124
2125 root->grp_to_be_replaced = 1;
2126 sth_created = true;
2127 hole = false;
2128 }
2129 else
2130 {
2131 if (covered_to < limit)
2132 hole = true;
2133 if (scalar)
2134 root->grp_total_scalarization = 0;
2135 }
2136
2137 if (sth_created
2138 && (!hole || root->grp_total_scalarization))
2139 {
2140 root->grp_covered = 1;
2141 return true;
2142 }
2143 if (root->grp_write || TREE_CODE (root->base) == PARM_DECL)
2144 root->grp_unscalarized_data = 1; /* not covered and written to */
2145 if (sth_created)
2146 return true;
2147 return false;
2148 }
2149
2150 /* Analyze all access trees linked by next_grp by the means of
2151 analyze_access_subtree. */
2152 static bool
2153 analyze_access_trees (struct access *access)
2154 {
2155 bool ret = false;
2156
2157 while (access)
2158 {
2159 if (analyze_access_subtree (access, NULL, true))
2160 ret = true;
2161 access = access->next_grp;
2162 }
2163
2164 return ret;
2165 }
2166
2167 /* Return true iff a potential new child of LACC at offset OFFSET and with size
2168 SIZE would conflict with an already existing one. If exactly such a child
2169 already exists in LACC, store a pointer to it in EXACT_MATCH. */
2170
2171 static bool
2172 child_would_conflict_in_lacc (struct access *lacc, HOST_WIDE_INT norm_offset,
2173 HOST_WIDE_INT size, struct access **exact_match)
2174 {
2175 struct access *child;
2176
2177 for (child = lacc->first_child; child; child = child->next_sibling)
2178 {
2179 if (child->offset == norm_offset && child->size == size)
2180 {
2181 *exact_match = child;
2182 return true;
2183 }
2184
2185 if (child->offset < norm_offset + size
2186 && child->offset + child->size > norm_offset)
2187 return true;
2188 }
2189
2190 return false;
2191 }
2192
2193 /* Create a new child access of PARENT, with all properties just like MODEL
2194 except for its offset and with its grp_write false and grp_read true.
2195 Return the new access or NULL if it cannot be created. Note that this access
2196 is created long after all splicing and sorting, it's not located in any
2197 access vector and is automatically a representative of its group. */
2198
2199 static struct access *
2200 create_artificial_child_access (struct access *parent, struct access *model,
2201 HOST_WIDE_INT new_offset)
2202 {
2203 struct access *access;
2204 struct access **child;
2205 tree expr = parent->base;
2206
2207 gcc_assert (!model->grp_unscalarizable_region);
2208
2209 access = (struct access *) pool_alloc (access_pool);
2210 memset (access, 0, sizeof (struct access));
2211 if (!build_user_friendly_ref_for_offset (&expr, TREE_TYPE (expr), new_offset,
2212 model->type))
2213 {
2214 access->grp_no_warning = true;
2215 expr = build_ref_for_model (EXPR_LOCATION (parent->base), parent->base,
2216 new_offset, model, NULL, false);
2217 }
2218
2219 access->base = parent->base;
2220 access->expr = expr;
2221 access->offset = new_offset;
2222 access->size = model->size;
2223 access->type = model->type;
2224 access->grp_write = true;
2225 access->grp_read = false;
2226
2227 child = &parent->first_child;
2228 while (*child && (*child)->offset < new_offset)
2229 child = &(*child)->next_sibling;
2230
2231 access->next_sibling = *child;
2232 *child = access;
2233
2234 return access;
2235 }
2236
2237
2238 /* Propagate all subaccesses of RACC across an assignment link to LACC. Return
2239 true if any new subaccess was created. Additionally, if RACC is a scalar
2240 access but LACC is not, change the type of the latter, if possible. */
2241
2242 static bool
2243 propagate_subaccesses_across_link (struct access *lacc, struct access *racc)
2244 {
2245 struct access *rchild;
2246 HOST_WIDE_INT norm_delta = lacc->offset - racc->offset;
2247 bool ret = false;
2248
2249 if (is_gimple_reg_type (lacc->type)
2250 || lacc->grp_unscalarizable_region
2251 || racc->grp_unscalarizable_region)
2252 return false;
2253
2254 if (is_gimple_reg_type (racc->type))
2255 {
2256 if (!lacc->first_child && !racc->first_child)
2257 {
2258 tree t = lacc->base;
2259
2260 lacc->type = racc->type;
2261 if (build_user_friendly_ref_for_offset (&t, TREE_TYPE (t),
2262 lacc->offset, racc->type))
2263 lacc->expr = t;
2264 else
2265 {
2266 lacc->expr = build_ref_for_model (EXPR_LOCATION (lacc->base),
2267 lacc->base, lacc->offset,
2268 racc, NULL, false);
2269 lacc->grp_no_warning = true;
2270 }
2271 }
2272 return false;
2273 }
2274
2275 for (rchild = racc->first_child; rchild; rchild = rchild->next_sibling)
2276 {
2277 struct access *new_acc = NULL;
2278 HOST_WIDE_INT norm_offset = rchild->offset + norm_delta;
2279
2280 if (rchild->grp_unscalarizable_region)
2281 continue;
2282
2283 if (child_would_conflict_in_lacc (lacc, norm_offset, rchild->size,
2284 &new_acc))
2285 {
2286 if (new_acc)
2287 {
2288 rchild->grp_hint = 1;
2289 new_acc->grp_hint |= new_acc->grp_read;
2290 if (rchild->first_child)
2291 ret |= propagate_subaccesses_across_link (new_acc, rchild);
2292 }
2293 continue;
2294 }
2295
2296 rchild->grp_hint = 1;
2297 new_acc = create_artificial_child_access (lacc, rchild, norm_offset);
2298 if (new_acc)
2299 {
2300 ret = true;
2301 if (racc->first_child)
2302 propagate_subaccesses_across_link (new_acc, rchild);
2303 }
2304 }
2305
2306 return ret;
2307 }
2308
2309 /* Propagate all subaccesses across assignment links. */
2310
2311 static void
2312 propagate_all_subaccesses (void)
2313 {
2314 while (work_queue_head)
2315 {
2316 struct access *racc = pop_access_from_work_queue ();
2317 struct assign_link *link;
2318
2319 gcc_assert (racc->first_link);
2320
2321 for (link = racc->first_link; link; link = link->next)
2322 {
2323 struct access *lacc = link->lacc;
2324
2325 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (lacc->base)))
2326 continue;
2327 lacc = lacc->group_representative;
2328 if (propagate_subaccesses_across_link (lacc, racc)
2329 && lacc->first_link)
2330 add_access_to_work_queue (lacc);
2331 }
2332 }
2333 }
2334
2335 /* Go through all accesses collected throughout the (intraprocedural) analysis
2336 stage, exclude overlapping ones, identify representatives and build trees
2337 out of them, making decisions about scalarization on the way. Return true
2338 iff there are any to-be-scalarized variables after this stage. */
2339
2340 static bool
2341 analyze_all_variable_accesses (void)
2342 {
2343 int res = 0;
2344 bitmap tmp = BITMAP_ALLOC (NULL);
2345 bitmap_iterator bi;
2346 unsigned i, max_total_scalarization_size;
2347
2348 max_total_scalarization_size = UNITS_PER_WORD * BITS_PER_UNIT
2349 * MOVE_RATIO (optimize_function_for_speed_p (cfun));
2350
2351 EXECUTE_IF_SET_IN_BITMAP (candidate_bitmap, 0, i, bi)
2352 if (bitmap_bit_p (should_scalarize_away_bitmap, i)
2353 && !bitmap_bit_p (cannot_scalarize_away_bitmap, i))
2354 {
2355 tree var = referenced_var (i);
2356
2357 if (TREE_CODE (var) == VAR_DECL
2358 && type_consists_of_records_p (TREE_TYPE (var)))
2359 {
2360 if ((unsigned) tree_low_cst (TYPE_SIZE (TREE_TYPE (var)), 1)
2361 <= max_total_scalarization_size)
2362 {
2363 completely_scalarize_var (var);
2364 if (dump_file && (dump_flags & TDF_DETAILS))
2365 {
2366 fprintf (dump_file, "Will attempt to totally scalarize ");
2367 print_generic_expr (dump_file, var, 0);
2368 fprintf (dump_file, " (UID: %u): \n", DECL_UID (var));
2369 }
2370 }
2371 else if (dump_file && (dump_flags & TDF_DETAILS))
2372 {
2373 fprintf (dump_file, "Too big to totally scalarize: ");
2374 print_generic_expr (dump_file, var, 0);
2375 fprintf (dump_file, " (UID: %u)\n", DECL_UID (var));
2376 }
2377 }
2378 }
2379
2380 bitmap_copy (tmp, candidate_bitmap);
2381 EXECUTE_IF_SET_IN_BITMAP (tmp, 0, i, bi)
2382 {
2383 tree var = referenced_var (i);
2384 struct access *access;
2385
2386 access = sort_and_splice_var_accesses (var);
2387 if (!access || !build_access_trees (access))
2388 disqualify_candidate (var,
2389 "No or inhibitingly overlapping accesses.");
2390 }
2391
2392 propagate_all_subaccesses ();
2393
2394 bitmap_copy (tmp, candidate_bitmap);
2395 EXECUTE_IF_SET_IN_BITMAP (tmp, 0, i, bi)
2396 {
2397 tree var = referenced_var (i);
2398 struct access *access = get_first_repr_for_decl (var);
2399
2400 if (analyze_access_trees (access))
2401 {
2402 res++;
2403 if (dump_file && (dump_flags & TDF_DETAILS))
2404 {
2405 fprintf (dump_file, "\nAccess trees for ");
2406 print_generic_expr (dump_file, var, 0);
2407 fprintf (dump_file, " (UID: %u): \n", DECL_UID (var));
2408 dump_access_tree (dump_file, access);
2409 fprintf (dump_file, "\n");
2410 }
2411 }
2412 else
2413 disqualify_candidate (var, "No scalar replacements to be created.");
2414 }
2415
2416 BITMAP_FREE (tmp);
2417
2418 if (res)
2419 {
2420 statistics_counter_event (cfun, "Scalarized aggregates", res);
2421 return true;
2422 }
2423 else
2424 return false;
2425 }
2426
2427 /* Generate statements copying scalar replacements of accesses within a subtree
2428 into or out of AGG. ACCESS, all its children, siblings and their children
2429 are to be processed. AGG is an aggregate type expression (can be a
2430 declaration but does not have to be, it can for example also be a mem_ref or
2431 a series of handled components). TOP_OFFSET is the offset of the processed
2432 subtree which has to be subtracted from offsets of individual accesses to
2433 get corresponding offsets for AGG. If CHUNK_SIZE is non-null, copy only
2434 replacements in the interval <start_offset, start_offset + chunk_size>,
2435 otherwise copy all. GSI is a statement iterator used to place the new
2436 statements. WRITE should be true when the statements should write from AGG
2437 to the replacement and false if vice versa. if INSERT_AFTER is true, new
2438 statements will be added after the current statement in GSI, they will be
2439 added before the statement otherwise. */
2440
2441 static void
2442 generate_subtree_copies (struct access *access, tree agg,
2443 HOST_WIDE_INT top_offset,
2444 HOST_WIDE_INT start_offset, HOST_WIDE_INT chunk_size,
2445 gimple_stmt_iterator *gsi, bool write,
2446 bool insert_after, location_t loc)
2447 {
2448 do
2449 {
2450 if (chunk_size && access->offset >= start_offset + chunk_size)
2451 return;
2452
2453 if (access->grp_to_be_replaced
2454 && (chunk_size == 0
2455 || access->offset + access->size > start_offset))
2456 {
2457 tree expr, repl = get_access_replacement (access);
2458 gimple stmt;
2459
2460 expr = build_ref_for_model (loc, agg, access->offset - top_offset,
2461 access, gsi, insert_after);
2462
2463 if (write)
2464 {
2465 if (access->grp_partial_lhs)
2466 expr = force_gimple_operand_gsi (gsi, expr, true, NULL_TREE,
2467 !insert_after,
2468 insert_after ? GSI_NEW_STMT
2469 : GSI_SAME_STMT);
2470 stmt = gimple_build_assign (repl, expr);
2471 }
2472 else
2473 {
2474 TREE_NO_WARNING (repl) = 1;
2475 if (access->grp_partial_lhs)
2476 repl = force_gimple_operand_gsi (gsi, repl, true, NULL_TREE,
2477 !insert_after,
2478 insert_after ? GSI_NEW_STMT
2479 : GSI_SAME_STMT);
2480 stmt = gimple_build_assign (expr, repl);
2481 }
2482 gimple_set_location (stmt, loc);
2483
2484 if (insert_after)
2485 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2486 else
2487 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2488 update_stmt (stmt);
2489 sra_stats.subtree_copies++;
2490 }
2491
2492 if (access->first_child)
2493 generate_subtree_copies (access->first_child, agg, top_offset,
2494 start_offset, chunk_size, gsi,
2495 write, insert_after, loc);
2496
2497 access = access->next_sibling;
2498 }
2499 while (access);
2500 }
2501
2502 /* Assign zero to all scalar replacements in an access subtree. ACCESS is the
2503 the root of the subtree to be processed. GSI is the statement iterator used
2504 for inserting statements which are added after the current statement if
2505 INSERT_AFTER is true or before it otherwise. */
2506
2507 static void
2508 init_subtree_with_zero (struct access *access, gimple_stmt_iterator *gsi,
2509 bool insert_after, location_t loc)
2510
2511 {
2512 struct access *child;
2513
2514 if (access->grp_to_be_replaced)
2515 {
2516 gimple stmt;
2517
2518 stmt = gimple_build_assign (get_access_replacement (access),
2519 build_zero_cst (access->type));
2520 if (insert_after)
2521 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2522 else
2523 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2524 update_stmt (stmt);
2525 gimple_set_location (stmt, loc);
2526 }
2527
2528 for (child = access->first_child; child; child = child->next_sibling)
2529 init_subtree_with_zero (child, gsi, insert_after, loc);
2530 }
2531
2532 /* Search for an access representative for the given expression EXPR and
2533 return it or NULL if it cannot be found. */
2534
2535 static struct access *
2536 get_access_for_expr (tree expr)
2537 {
2538 HOST_WIDE_INT offset, size, max_size;
2539 tree base;
2540
2541 /* FIXME: This should not be necessary but Ada produces V_C_Es with a type of
2542 a different size than the size of its argument and we need the latter
2543 one. */
2544 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR)
2545 expr = TREE_OPERAND (expr, 0);
2546
2547 base = get_ref_base_and_extent (expr, &offset, &size, &max_size);
2548 if (max_size == -1 || !DECL_P (base))
2549 return NULL;
2550
2551 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (base)))
2552 return NULL;
2553
2554 return get_var_base_offset_size_access (base, offset, max_size);
2555 }
2556
2557 /* Replace the expression EXPR with a scalar replacement if there is one and
2558 generate other statements to do type conversion or subtree copying if
2559 necessary. GSI is used to place newly created statements, WRITE is true if
2560 the expression is being written to (it is on a LHS of a statement or output
2561 in an assembly statement). */
2562
2563 static bool
2564 sra_modify_expr (tree *expr, gimple_stmt_iterator *gsi, bool write)
2565 {
2566 location_t loc;
2567 struct access *access;
2568 tree type, bfr;
2569
2570 if (TREE_CODE (*expr) == BIT_FIELD_REF)
2571 {
2572 bfr = *expr;
2573 expr = &TREE_OPERAND (*expr, 0);
2574 }
2575 else
2576 bfr = NULL_TREE;
2577
2578 if (TREE_CODE (*expr) == REALPART_EXPR || TREE_CODE (*expr) == IMAGPART_EXPR)
2579 expr = &TREE_OPERAND (*expr, 0);
2580 access = get_access_for_expr (*expr);
2581 if (!access)
2582 return false;
2583 type = TREE_TYPE (*expr);
2584
2585 loc = gimple_location (gsi_stmt (*gsi));
2586 if (access->grp_to_be_replaced)
2587 {
2588 tree repl = get_access_replacement (access);
2589 /* If we replace a non-register typed access simply use the original
2590 access expression to extract the scalar component afterwards.
2591 This happens if scalarizing a function return value or parameter
2592 like in gcc.c-torture/execute/20041124-1.c, 20050316-1.c and
2593 gcc.c-torture/compile/20011217-1.c.
2594
2595 We also want to use this when accessing a complex or vector which can
2596 be accessed as a different type too, potentially creating a need for
2597 type conversion (see PR42196) and when scalarized unions are involved
2598 in assembler statements (see PR42398). */
2599 if (!useless_type_conversion_p (type, access->type))
2600 {
2601 tree ref;
2602
2603 ref = build_ref_for_model (loc, access->base, access->offset, access,
2604 NULL, false);
2605
2606 if (write)
2607 {
2608 gimple stmt;
2609
2610 if (access->grp_partial_lhs)
2611 ref = force_gimple_operand_gsi (gsi, ref, true, NULL_TREE,
2612 false, GSI_NEW_STMT);
2613 stmt = gimple_build_assign (repl, ref);
2614 gimple_set_location (stmt, loc);
2615 gsi_insert_after (gsi, stmt, GSI_NEW_STMT);
2616 }
2617 else
2618 {
2619 gimple stmt;
2620
2621 if (access->grp_partial_lhs)
2622 repl = force_gimple_operand_gsi (gsi, repl, true, NULL_TREE,
2623 true, GSI_SAME_STMT);
2624 stmt = gimple_build_assign (ref, repl);
2625 gimple_set_location (stmt, loc);
2626 gsi_insert_before (gsi, stmt, GSI_SAME_STMT);
2627 }
2628 }
2629 else
2630 *expr = repl;
2631 sra_stats.exprs++;
2632 }
2633
2634 if (access->first_child)
2635 {
2636 HOST_WIDE_INT start_offset, chunk_size;
2637 if (bfr
2638 && host_integerp (TREE_OPERAND (bfr, 1), 1)
2639 && host_integerp (TREE_OPERAND (bfr, 2), 1))
2640 {
2641 chunk_size = tree_low_cst (TREE_OPERAND (bfr, 1), 1);
2642 start_offset = access->offset
2643 + tree_low_cst (TREE_OPERAND (bfr, 2), 1);
2644 }
2645 else
2646 start_offset = chunk_size = 0;
2647
2648 generate_subtree_copies (access->first_child, access->base, 0,
2649 start_offset, chunk_size, gsi, write, write,
2650 loc);
2651 }
2652 return true;
2653 }
2654
2655 /* Where scalar replacements of the RHS have been written to when a replacement
2656 of a LHS of an assigments cannot be direclty loaded from a replacement of
2657 the RHS. */
2658 enum unscalarized_data_handling { SRA_UDH_NONE, /* Nothing done so far. */
2659 SRA_UDH_RIGHT, /* Data flushed to the RHS. */
2660 SRA_UDH_LEFT }; /* Data flushed to the LHS. */
2661
2662 /* Store all replacements in the access tree rooted in TOP_RACC either to their
2663 base aggregate if there are unscalarized data or directly to LHS of the
2664 statement that is pointed to by GSI otherwise. */
2665
2666 static enum unscalarized_data_handling
2667 handle_unscalarized_data_in_subtree (struct access *top_racc,
2668 gimple_stmt_iterator *gsi)
2669 {
2670 if (top_racc->grp_unscalarized_data)
2671 {
2672 generate_subtree_copies (top_racc->first_child, top_racc->base, 0, 0, 0,
2673 gsi, false, false,
2674 gimple_location (gsi_stmt (*gsi)));
2675 return SRA_UDH_RIGHT;
2676 }
2677 else
2678 {
2679 tree lhs = gimple_assign_lhs (gsi_stmt (*gsi));
2680 generate_subtree_copies (top_racc->first_child, lhs, top_racc->offset,
2681 0, 0, gsi, false, false,
2682 gimple_location (gsi_stmt (*gsi)));
2683 return SRA_UDH_LEFT;
2684 }
2685 }
2686
2687
2688 /* Try to generate statements to load all sub-replacements in an access subtree
2689 formed by children of LACC from scalar replacements in the TOP_RACC subtree.
2690 If that is not possible, refresh the TOP_RACC base aggregate and load the
2691 accesses from it. LEFT_OFFSET is the offset of the left whole subtree being
2692 copied. NEW_GSI is stmt iterator used for statement insertions after the
2693 original assignment, OLD_GSI is used to insert statements before the
2694 assignment. *REFRESHED keeps the information whether we have needed to
2695 refresh replacements of the LHS and from which side of the assignments this
2696 takes place. */
2697
2698 static void
2699 load_assign_lhs_subreplacements (struct access *lacc, struct access *top_racc,
2700 HOST_WIDE_INT left_offset,
2701 gimple_stmt_iterator *old_gsi,
2702 gimple_stmt_iterator *new_gsi,
2703 enum unscalarized_data_handling *refreshed)
2704 {
2705 location_t loc = gimple_location (gsi_stmt (*old_gsi));
2706 for (lacc = lacc->first_child; lacc; lacc = lacc->next_sibling)
2707 {
2708 if (lacc->grp_to_be_replaced)
2709 {
2710 struct access *racc;
2711 HOST_WIDE_INT offset = lacc->offset - left_offset + top_racc->offset;
2712 gimple stmt;
2713 tree rhs;
2714
2715 racc = find_access_in_subtree (top_racc, offset, lacc->size);
2716 if (racc && racc->grp_to_be_replaced)
2717 {
2718 rhs = get_access_replacement (racc);
2719 if (!useless_type_conversion_p (lacc->type, racc->type))
2720 rhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR, lacc->type, rhs);
2721
2722 if (racc->grp_partial_lhs && lacc->grp_partial_lhs)
2723 rhs = force_gimple_operand_gsi (old_gsi, rhs, true, NULL_TREE,
2724 true, GSI_SAME_STMT);
2725 }
2726 else
2727 {
2728 /* No suitable access on the right hand side, need to load from
2729 the aggregate. See if we have to update it first... */
2730 if (*refreshed == SRA_UDH_NONE)
2731 *refreshed = handle_unscalarized_data_in_subtree (top_racc,
2732 old_gsi);
2733
2734 if (*refreshed == SRA_UDH_LEFT)
2735 rhs = build_ref_for_model (loc, lacc->base, lacc->offset, lacc,
2736 new_gsi, true);
2737 else
2738 rhs = build_ref_for_model (loc, top_racc->base, offset, lacc,
2739 new_gsi, true);
2740 if (lacc->grp_partial_lhs)
2741 rhs = force_gimple_operand_gsi (new_gsi, rhs, true, NULL_TREE,
2742 false, GSI_NEW_STMT);
2743 }
2744
2745 stmt = gimple_build_assign (get_access_replacement (lacc), rhs);
2746 gsi_insert_after (new_gsi, stmt, GSI_NEW_STMT);
2747 gimple_set_location (stmt, loc);
2748 update_stmt (stmt);
2749 sra_stats.subreplacements++;
2750 }
2751 else if (*refreshed == SRA_UDH_NONE
2752 && lacc->grp_read && !lacc->grp_covered)
2753 *refreshed = handle_unscalarized_data_in_subtree (top_racc,
2754 old_gsi);
2755
2756 if (lacc->first_child)
2757 load_assign_lhs_subreplacements (lacc, top_racc, left_offset,
2758 old_gsi, new_gsi, refreshed);
2759 }
2760 }
2761
2762 /* Result code for SRA assignment modification. */
2763 enum assignment_mod_result { SRA_AM_NONE, /* nothing done for the stmt */
2764 SRA_AM_MODIFIED, /* stmt changed but not
2765 removed */
2766 SRA_AM_REMOVED }; /* stmt eliminated */
2767
2768 /* Modify assignments with a CONSTRUCTOR on their RHS. STMT contains a pointer
2769 to the assignment and GSI is the statement iterator pointing at it. Returns
2770 the same values as sra_modify_assign. */
2771
2772 static enum assignment_mod_result
2773 sra_modify_constructor_assign (gimple *stmt, gimple_stmt_iterator *gsi)
2774 {
2775 tree lhs = gimple_assign_lhs (*stmt);
2776 struct access *acc;
2777 location_t loc;
2778
2779 acc = get_access_for_expr (lhs);
2780 if (!acc)
2781 return SRA_AM_NONE;
2782
2783 if (gimple_clobber_p (*stmt))
2784 {
2785 /* Remove clobbers of fully scalarized variables, otherwise
2786 do nothing. */
2787 if (acc->grp_covered)
2788 {
2789 unlink_stmt_vdef (*stmt);
2790 gsi_remove (gsi, true);
2791 release_defs (*stmt);
2792 return SRA_AM_REMOVED;
2793 }
2794 else
2795 return SRA_AM_NONE;
2796 }
2797
2798 loc = gimple_location (*stmt);
2799 if (VEC_length (constructor_elt,
2800 CONSTRUCTOR_ELTS (gimple_assign_rhs1 (*stmt))) > 0)
2801 {
2802 /* I have never seen this code path trigger but if it can happen the
2803 following should handle it gracefully. */
2804 if (access_has_children_p (acc))
2805 generate_subtree_copies (acc->first_child, acc->base, 0, 0, 0, gsi,
2806 true, true, loc);
2807 return SRA_AM_MODIFIED;
2808 }
2809
2810 if (acc->grp_covered)
2811 {
2812 init_subtree_with_zero (acc, gsi, false, loc);
2813 unlink_stmt_vdef (*stmt);
2814 gsi_remove (gsi, true);
2815 release_defs (*stmt);
2816 return SRA_AM_REMOVED;
2817 }
2818 else
2819 {
2820 init_subtree_with_zero (acc, gsi, true, loc);
2821 return SRA_AM_MODIFIED;
2822 }
2823 }
2824
2825 /* Create and return a new suitable default definition SSA_NAME for RACC which
2826 is an access describing an uninitialized part of an aggregate that is being
2827 loaded. */
2828
2829 static tree
2830 get_repl_default_def_ssa_name (struct access *racc)
2831 {
2832 tree repl, decl;
2833
2834 decl = get_unrenamed_access_replacement (racc);
2835
2836 repl = gimple_default_def (cfun, decl);
2837 if (!repl)
2838 {
2839 repl = make_ssa_name (decl, gimple_build_nop ());
2840 set_default_def (decl, repl);
2841 }
2842
2843 return repl;
2844 }
2845
2846 /* Return true if REF has a COMPONENT_REF with a bit-field field declaration
2847 somewhere in it. */
2848
2849 static inline bool
2850 contains_bitfld_comp_ref_p (const_tree ref)
2851 {
2852 while (handled_component_p (ref))
2853 {
2854 if (TREE_CODE (ref) == COMPONENT_REF
2855 && DECL_BIT_FIELD (TREE_OPERAND (ref, 1)))
2856 return true;
2857 ref = TREE_OPERAND (ref, 0);
2858 }
2859
2860 return false;
2861 }
2862
2863 /* Return true if REF has an VIEW_CONVERT_EXPR or a COMPONENT_REF with a
2864 bit-field field declaration somewhere in it. */
2865
2866 static inline bool
2867 contains_vce_or_bfcref_p (const_tree ref)
2868 {
2869 while (handled_component_p (ref))
2870 {
2871 if (TREE_CODE (ref) == VIEW_CONVERT_EXPR
2872 || (TREE_CODE (ref) == COMPONENT_REF
2873 && DECL_BIT_FIELD (TREE_OPERAND (ref, 1))))
2874 return true;
2875 ref = TREE_OPERAND (ref, 0);
2876 }
2877
2878 return false;
2879 }
2880
2881 /* Examine both sides of the assignment statement pointed to by STMT, replace
2882 them with a scalare replacement if there is one and generate copying of
2883 replacements if scalarized aggregates have been used in the assignment. GSI
2884 is used to hold generated statements for type conversions and subtree
2885 copying. */
2886
2887 static enum assignment_mod_result
2888 sra_modify_assign (gimple *stmt, gimple_stmt_iterator *gsi)
2889 {
2890 struct access *lacc, *racc;
2891 tree lhs, rhs;
2892 bool modify_this_stmt = false;
2893 bool force_gimple_rhs = false;
2894 location_t loc;
2895 gimple_stmt_iterator orig_gsi = *gsi;
2896
2897 if (!gimple_assign_single_p (*stmt))
2898 return SRA_AM_NONE;
2899 lhs = gimple_assign_lhs (*stmt);
2900 rhs = gimple_assign_rhs1 (*stmt);
2901
2902 if (TREE_CODE (rhs) == CONSTRUCTOR)
2903 return sra_modify_constructor_assign (stmt, gsi);
2904
2905 if (TREE_CODE (rhs) == REALPART_EXPR || TREE_CODE (lhs) == REALPART_EXPR
2906 || TREE_CODE (rhs) == IMAGPART_EXPR || TREE_CODE (lhs) == IMAGPART_EXPR
2907 || TREE_CODE (rhs) == BIT_FIELD_REF || TREE_CODE (lhs) == BIT_FIELD_REF)
2908 {
2909 modify_this_stmt = sra_modify_expr (gimple_assign_rhs1_ptr (*stmt),
2910 gsi, false);
2911 modify_this_stmt |= sra_modify_expr (gimple_assign_lhs_ptr (*stmt),
2912 gsi, true);
2913 return modify_this_stmt ? SRA_AM_MODIFIED : SRA_AM_NONE;
2914 }
2915
2916 lacc = get_access_for_expr (lhs);
2917 racc = get_access_for_expr (rhs);
2918 if (!lacc && !racc)
2919 return SRA_AM_NONE;
2920
2921 loc = gimple_location (*stmt);
2922 if (lacc && lacc->grp_to_be_replaced)
2923 {
2924 lhs = get_access_replacement (lacc);
2925 gimple_assign_set_lhs (*stmt, lhs);
2926 modify_this_stmt = true;
2927 if (lacc->grp_partial_lhs)
2928 force_gimple_rhs = true;
2929 sra_stats.exprs++;
2930 }
2931
2932 if (racc && racc->grp_to_be_replaced)
2933 {
2934 rhs = get_access_replacement (racc);
2935 modify_this_stmt = true;
2936 if (racc->grp_partial_lhs)
2937 force_gimple_rhs = true;
2938 sra_stats.exprs++;
2939 }
2940 else if (racc
2941 && !racc->grp_unscalarized_data
2942 && TREE_CODE (lhs) == SSA_NAME
2943 && !access_has_replacements_p (racc))
2944 {
2945 rhs = get_repl_default_def_ssa_name (racc);
2946 modify_this_stmt = true;
2947 sra_stats.exprs++;
2948 }
2949
2950 if (modify_this_stmt)
2951 {
2952 if (!useless_type_conversion_p (TREE_TYPE (lhs), TREE_TYPE (rhs)))
2953 {
2954 /* If we can avoid creating a VIEW_CONVERT_EXPR do so.
2955 ??? This should move to fold_stmt which we simply should
2956 call after building a VIEW_CONVERT_EXPR here. */
2957 if (AGGREGATE_TYPE_P (TREE_TYPE (lhs))
2958 && !contains_bitfld_comp_ref_p (lhs)
2959 && !access_has_children_p (lacc))
2960 {
2961 lhs = build_ref_for_model (loc, lhs, 0, racc, gsi, false);
2962 gimple_assign_set_lhs (*stmt, lhs);
2963 }
2964 else if (AGGREGATE_TYPE_P (TREE_TYPE (rhs))
2965 && !contains_vce_or_bfcref_p (rhs)
2966 && !access_has_children_p (racc))
2967 rhs = build_ref_for_model (loc, rhs, 0, lacc, gsi, false);
2968
2969 if (!useless_type_conversion_p (TREE_TYPE (lhs), TREE_TYPE (rhs)))
2970 {
2971 rhs = fold_build1_loc (loc, VIEW_CONVERT_EXPR, TREE_TYPE (lhs),
2972 rhs);
2973 if (is_gimple_reg_type (TREE_TYPE (lhs))
2974 && TREE_CODE (lhs) != SSA_NAME)
2975 force_gimple_rhs = true;
2976 }
2977 }
2978 }
2979
2980 /* From this point on, the function deals with assignments in between
2981 aggregates when at least one has scalar reductions of some of its
2982 components. There are three possible scenarios: Both the LHS and RHS have
2983 to-be-scalarized components, 2) only the RHS has or 3) only the LHS has.
2984
2985 In the first case, we would like to load the LHS components from RHS
2986 components whenever possible. If that is not possible, we would like to
2987 read it directly from the RHS (after updating it by storing in it its own
2988 components). If there are some necessary unscalarized data in the LHS,
2989 those will be loaded by the original assignment too. If neither of these
2990 cases happen, the original statement can be removed. Most of this is done
2991 by load_assign_lhs_subreplacements.
2992
2993 In the second case, we would like to store all RHS scalarized components
2994 directly into LHS and if they cover the aggregate completely, remove the
2995 statement too. In the third case, we want the LHS components to be loaded
2996 directly from the RHS (DSE will remove the original statement if it
2997 becomes redundant).
2998
2999 This is a bit complex but manageable when types match and when unions do
3000 not cause confusion in a way that we cannot really load a component of LHS
3001 from the RHS or vice versa (the access representing this level can have
3002 subaccesses that are accessible only through a different union field at a
3003 higher level - different from the one used in the examined expression).
3004 Unions are fun.
3005
3006 Therefore, I specially handle a fourth case, happening when there is a
3007 specific type cast or it is impossible to locate a scalarized subaccess on
3008 the other side of the expression. If that happens, I simply "refresh" the
3009 RHS by storing in it is scalarized components leave the original statement
3010 there to do the copying and then load the scalar replacements of the LHS.
3011 This is what the first branch does. */
3012
3013 if (modify_this_stmt
3014 || gimple_has_volatile_ops (*stmt)
3015 || contains_vce_or_bfcref_p (rhs)
3016 || contains_vce_or_bfcref_p (lhs))
3017 {
3018 if (access_has_children_p (racc))
3019 generate_subtree_copies (racc->first_child, racc->base, 0, 0, 0,
3020 gsi, false, false, loc);
3021 if (access_has_children_p (lacc))
3022 generate_subtree_copies (lacc->first_child, lacc->base, 0, 0, 0,
3023 gsi, true, true, loc);
3024 sra_stats.separate_lhs_rhs_handling++;
3025
3026 /* This gimplification must be done after generate_subtree_copies,
3027 lest we insert the subtree copies in the middle of the gimplified
3028 sequence. */
3029 if (force_gimple_rhs)
3030 rhs = force_gimple_operand_gsi (&orig_gsi, rhs, true, NULL_TREE,
3031 true, GSI_SAME_STMT);
3032 if (gimple_assign_rhs1 (*stmt) != rhs)
3033 {
3034 modify_this_stmt = true;
3035 gimple_assign_set_rhs_from_tree (&orig_gsi, rhs);
3036 gcc_assert (*stmt == gsi_stmt (orig_gsi));
3037 }
3038
3039 return modify_this_stmt ? SRA_AM_MODIFIED : SRA_AM_NONE;
3040 }
3041 else
3042 {
3043 if (access_has_children_p (lacc)
3044 && access_has_children_p (racc)
3045 /* When an access represents an unscalarizable region, it usually
3046 represents accesses with variable offset and thus must not be used
3047 to generate new memory accesses. */
3048 && !lacc->grp_unscalarizable_region
3049 && !racc->grp_unscalarizable_region)
3050 {
3051 gimple_stmt_iterator orig_gsi = *gsi;
3052 enum unscalarized_data_handling refreshed;
3053
3054 if (lacc->grp_read && !lacc->grp_covered)
3055 refreshed = handle_unscalarized_data_in_subtree (racc, gsi);
3056 else
3057 refreshed = SRA_UDH_NONE;
3058
3059 load_assign_lhs_subreplacements (lacc, racc, lacc->offset,
3060 &orig_gsi, gsi, &refreshed);
3061 if (refreshed != SRA_UDH_RIGHT)
3062 {
3063 gsi_next (gsi);
3064 unlink_stmt_vdef (*stmt);
3065 gsi_remove (&orig_gsi, true);
3066 release_defs (*stmt);
3067 sra_stats.deleted++;
3068 return SRA_AM_REMOVED;
3069 }
3070 }
3071 else
3072 {
3073 if (access_has_children_p (racc)
3074 && !racc->grp_unscalarized_data)
3075 {
3076 if (dump_file)
3077 {
3078 fprintf (dump_file, "Removing load: ");
3079 print_gimple_stmt (dump_file, *stmt, 0, 0);
3080 }
3081 generate_subtree_copies (racc->first_child, lhs,
3082 racc->offset, 0, 0, gsi,
3083 false, false, loc);
3084 gcc_assert (*stmt == gsi_stmt (*gsi));
3085 unlink_stmt_vdef (*stmt);
3086 gsi_remove (gsi, true);
3087 release_defs (*stmt);
3088 sra_stats.deleted++;
3089 return SRA_AM_REMOVED;
3090 }
3091 /* Restore the aggregate RHS from its components so the
3092 prevailing aggregate copy does the right thing. */
3093 if (access_has_children_p (racc))
3094 generate_subtree_copies (racc->first_child, racc->base, 0, 0, 0,
3095 gsi, false, false, loc);
3096 /* Re-load the components of the aggregate copy destination.
3097 But use the RHS aggregate to load from to expose more
3098 optimization opportunities. */
3099 if (access_has_children_p (lacc))
3100 generate_subtree_copies (lacc->first_child, rhs, lacc->offset,
3101 0, 0, gsi, true, true, loc);
3102 }
3103
3104 return SRA_AM_NONE;
3105 }
3106 }
3107
3108 /* Traverse the function body and all modifications as decided in
3109 analyze_all_variable_accesses. Return true iff the CFG has been
3110 changed. */
3111
3112 static bool
3113 sra_modify_function_body (void)
3114 {
3115 bool cfg_changed = false;
3116 basic_block bb;
3117
3118 FOR_EACH_BB (bb)
3119 {
3120 gimple_stmt_iterator gsi = gsi_start_bb (bb);
3121 while (!gsi_end_p (gsi))
3122 {
3123 gimple stmt = gsi_stmt (gsi);
3124 enum assignment_mod_result assign_result;
3125 bool modified = false, deleted = false;
3126 tree *t;
3127 unsigned i;
3128
3129 switch (gimple_code (stmt))
3130 {
3131 case GIMPLE_RETURN:
3132 t = gimple_return_retval_ptr (stmt);
3133 if (*t != NULL_TREE)
3134 modified |= sra_modify_expr (t, &gsi, false);
3135 break;
3136
3137 case GIMPLE_ASSIGN:
3138 assign_result = sra_modify_assign (&stmt, &gsi);
3139 modified |= assign_result == SRA_AM_MODIFIED;
3140 deleted = assign_result == SRA_AM_REMOVED;
3141 break;
3142
3143 case GIMPLE_CALL:
3144 /* Operands must be processed before the lhs. */
3145 for (i = 0; i < gimple_call_num_args (stmt); i++)
3146 {
3147 t = gimple_call_arg_ptr (stmt, i);
3148 modified |= sra_modify_expr (t, &gsi, false);
3149 }
3150
3151 if (gimple_call_lhs (stmt))
3152 {
3153 t = gimple_call_lhs_ptr (stmt);
3154 modified |= sra_modify_expr (t, &gsi, true);
3155 }
3156 break;
3157
3158 case GIMPLE_ASM:
3159 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
3160 {
3161 t = &TREE_VALUE (gimple_asm_input_op (stmt, i));
3162 modified |= sra_modify_expr (t, &gsi, false);
3163 }
3164 for (i = 0; i < gimple_asm_noutputs (stmt); i++)
3165 {
3166 t = &TREE_VALUE (gimple_asm_output_op (stmt, i));
3167 modified |= sra_modify_expr (t, &gsi, true);
3168 }
3169 break;
3170
3171 default:
3172 break;
3173 }
3174
3175 if (modified)
3176 {
3177 update_stmt (stmt);
3178 if (maybe_clean_eh_stmt (stmt)
3179 && gimple_purge_dead_eh_edges (gimple_bb (stmt)))
3180 cfg_changed = true;
3181 }
3182 if (!deleted)
3183 gsi_next (&gsi);
3184 }
3185 }
3186
3187 return cfg_changed;
3188 }
3189
3190 /* Generate statements initializing scalar replacements of parts of function
3191 parameters. */
3192
3193 static void
3194 initialize_parameter_reductions (void)
3195 {
3196 gimple_stmt_iterator gsi;
3197 gimple_seq seq = NULL;
3198 tree parm;
3199
3200 gsi = gsi_start (seq);
3201 for (parm = DECL_ARGUMENTS (current_function_decl);
3202 parm;
3203 parm = DECL_CHAIN (parm))
3204 {
3205 VEC (access_p, heap) *access_vec;
3206 struct access *access;
3207
3208 if (!bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
3209 continue;
3210 access_vec = get_base_access_vector (parm);
3211 if (!access_vec)
3212 continue;
3213
3214 for (access = VEC_index (access_p, access_vec, 0);
3215 access;
3216 access = access->next_grp)
3217 generate_subtree_copies (access, parm, 0, 0, 0, &gsi, true, true,
3218 EXPR_LOCATION (parm));
3219 }
3220
3221 seq = gsi_seq (gsi);
3222 if (seq)
3223 gsi_insert_seq_on_edge_immediate (single_succ_edge (ENTRY_BLOCK_PTR), seq);
3224 }
3225
3226 /* The "main" function of intraprocedural SRA passes. Runs the analysis and if
3227 it reveals there are components of some aggregates to be scalarized, it runs
3228 the required transformations. */
3229 static unsigned int
3230 perform_intra_sra (void)
3231 {
3232 int ret = 0;
3233 sra_initialize ();
3234
3235 if (!find_var_candidates ())
3236 goto out;
3237
3238 if (!scan_function ())
3239 goto out;
3240
3241 if (!analyze_all_variable_accesses ())
3242 goto out;
3243
3244 if (sra_modify_function_body ())
3245 ret = TODO_update_ssa | TODO_cleanup_cfg;
3246 else
3247 ret = TODO_update_ssa;
3248 initialize_parameter_reductions ();
3249
3250 statistics_counter_event (cfun, "Scalar replacements created",
3251 sra_stats.replacements);
3252 statistics_counter_event (cfun, "Modified expressions", sra_stats.exprs);
3253 statistics_counter_event (cfun, "Subtree copy stmts",
3254 sra_stats.subtree_copies);
3255 statistics_counter_event (cfun, "Subreplacement stmts",
3256 sra_stats.subreplacements);
3257 statistics_counter_event (cfun, "Deleted stmts", sra_stats.deleted);
3258 statistics_counter_event (cfun, "Separate LHS and RHS handling",
3259 sra_stats.separate_lhs_rhs_handling);
3260
3261 out:
3262 sra_deinitialize ();
3263 return ret;
3264 }
3265
3266 /* Perform early intraprocedural SRA. */
3267 static unsigned int
3268 early_intra_sra (void)
3269 {
3270 sra_mode = SRA_MODE_EARLY_INTRA;
3271 return perform_intra_sra ();
3272 }
3273
3274 /* Perform "late" intraprocedural SRA. */
3275 static unsigned int
3276 late_intra_sra (void)
3277 {
3278 sra_mode = SRA_MODE_INTRA;
3279 return perform_intra_sra ();
3280 }
3281
3282
3283 static bool
3284 gate_intra_sra (void)
3285 {
3286 return flag_tree_sra != 0 && dbg_cnt (tree_sra);
3287 }
3288
3289
3290 struct gimple_opt_pass pass_sra_early =
3291 {
3292 {
3293 GIMPLE_PASS,
3294 "esra", /* name */
3295 gate_intra_sra, /* gate */
3296 early_intra_sra, /* execute */
3297 NULL, /* sub */
3298 NULL, /* next */
3299 0, /* static_pass_number */
3300 TV_TREE_SRA, /* tv_id */
3301 PROP_cfg | PROP_ssa, /* properties_required */
3302 0, /* properties_provided */
3303 0, /* properties_destroyed */
3304 0, /* todo_flags_start */
3305 TODO_update_ssa
3306 | TODO_ggc_collect
3307 | TODO_verify_ssa /* todo_flags_finish */
3308 }
3309 };
3310
3311 struct gimple_opt_pass pass_sra =
3312 {
3313 {
3314 GIMPLE_PASS,
3315 "sra", /* name */
3316 gate_intra_sra, /* gate */
3317 late_intra_sra, /* execute */
3318 NULL, /* sub */
3319 NULL, /* next */
3320 0, /* static_pass_number */
3321 TV_TREE_SRA, /* tv_id */
3322 PROP_cfg | PROP_ssa, /* properties_required */
3323 0, /* properties_provided */
3324 0, /* properties_destroyed */
3325 TODO_update_address_taken, /* todo_flags_start */
3326 TODO_update_ssa
3327 | TODO_ggc_collect
3328 | TODO_verify_ssa /* todo_flags_finish */
3329 }
3330 };
3331
3332
3333 /* Return true iff PARM (which must be a parm_decl) is an unused scalar
3334 parameter. */
3335
3336 static bool
3337 is_unused_scalar_param (tree parm)
3338 {
3339 tree name;
3340 return (is_gimple_reg (parm)
3341 && (!(name = gimple_default_def (cfun, parm))
3342 || has_zero_uses (name)));
3343 }
3344
3345 /* Scan immediate uses of a default definition SSA name of a parameter PARM and
3346 examine whether there are any direct or otherwise infeasible ones. If so,
3347 return true, otherwise return false. PARM must be a gimple register with a
3348 non-NULL default definition. */
3349
3350 static bool
3351 ptr_parm_has_direct_uses (tree parm)
3352 {
3353 imm_use_iterator ui;
3354 gimple stmt;
3355 tree name = gimple_default_def (cfun, parm);
3356 bool ret = false;
3357
3358 FOR_EACH_IMM_USE_STMT (stmt, ui, name)
3359 {
3360 int uses_ok = 0;
3361 use_operand_p use_p;
3362
3363 if (is_gimple_debug (stmt))
3364 continue;
3365
3366 /* Valid uses include dereferences on the lhs and the rhs. */
3367 if (gimple_has_lhs (stmt))
3368 {
3369 tree lhs = gimple_get_lhs (stmt);
3370 while (handled_component_p (lhs))
3371 lhs = TREE_OPERAND (lhs, 0);
3372 if (TREE_CODE (lhs) == MEM_REF
3373 && TREE_OPERAND (lhs, 0) == name
3374 && integer_zerop (TREE_OPERAND (lhs, 1))
3375 && types_compatible_p (TREE_TYPE (lhs),
3376 TREE_TYPE (TREE_TYPE (name)))
3377 && !TREE_THIS_VOLATILE (lhs))
3378 uses_ok++;
3379 }
3380 if (gimple_assign_single_p (stmt))
3381 {
3382 tree rhs = gimple_assign_rhs1 (stmt);
3383 while (handled_component_p (rhs))
3384 rhs = TREE_OPERAND (rhs, 0);
3385 if (TREE_CODE (rhs) == MEM_REF
3386 && TREE_OPERAND (rhs, 0) == name
3387 && integer_zerop (TREE_OPERAND (rhs, 1))
3388 && types_compatible_p (TREE_TYPE (rhs),
3389 TREE_TYPE (TREE_TYPE (name)))
3390 && !TREE_THIS_VOLATILE (rhs))
3391 uses_ok++;
3392 }
3393 else if (is_gimple_call (stmt))
3394 {
3395 unsigned i;
3396 for (i = 0; i < gimple_call_num_args (stmt); ++i)
3397 {
3398 tree arg = gimple_call_arg (stmt, i);
3399 while (handled_component_p (arg))
3400 arg = TREE_OPERAND (arg, 0);
3401 if (TREE_CODE (arg) == MEM_REF
3402 && TREE_OPERAND (arg, 0) == name
3403 && integer_zerop (TREE_OPERAND (arg, 1))
3404 && types_compatible_p (TREE_TYPE (arg),
3405 TREE_TYPE (TREE_TYPE (name)))
3406 && !TREE_THIS_VOLATILE (arg))
3407 uses_ok++;
3408 }
3409 }
3410
3411 /* If the number of valid uses does not match the number of
3412 uses in this stmt there is an unhandled use. */
3413 FOR_EACH_IMM_USE_ON_STMT (use_p, ui)
3414 --uses_ok;
3415
3416 if (uses_ok != 0)
3417 ret = true;
3418
3419 if (ret)
3420 BREAK_FROM_IMM_USE_STMT (ui);
3421 }
3422
3423 return ret;
3424 }
3425
3426 /* Identify candidates for reduction for IPA-SRA based on their type and mark
3427 them in candidate_bitmap. Note that these do not necessarily include
3428 parameter which are unused and thus can be removed. Return true iff any
3429 such candidate has been found. */
3430
3431 static bool
3432 find_param_candidates (void)
3433 {
3434 tree parm;
3435 int count = 0;
3436 bool ret = false;
3437 const char *msg;
3438
3439 for (parm = DECL_ARGUMENTS (current_function_decl);
3440 parm;
3441 parm = DECL_CHAIN (parm))
3442 {
3443 tree type = TREE_TYPE (parm);
3444
3445 count++;
3446
3447 if (TREE_THIS_VOLATILE (parm)
3448 || TREE_ADDRESSABLE (parm)
3449 || (!is_gimple_reg_type (type) && is_va_list_type (type)))
3450 continue;
3451
3452 if (is_unused_scalar_param (parm))
3453 {
3454 ret = true;
3455 continue;
3456 }
3457
3458 if (POINTER_TYPE_P (type))
3459 {
3460 type = TREE_TYPE (type);
3461
3462 if (TREE_CODE (type) == FUNCTION_TYPE
3463 || TYPE_VOLATILE (type)
3464 || (TREE_CODE (type) == ARRAY_TYPE
3465 && TYPE_NONALIASED_COMPONENT (type))
3466 || !is_gimple_reg (parm)
3467 || is_va_list_type (type)
3468 || ptr_parm_has_direct_uses (parm))
3469 continue;
3470 }
3471 else if (!AGGREGATE_TYPE_P (type))
3472 continue;
3473
3474 if (!COMPLETE_TYPE_P (type)
3475 || !host_integerp (TYPE_SIZE (type), 1)
3476 || tree_low_cst (TYPE_SIZE (type), 1) == 0
3477 || (AGGREGATE_TYPE_P (type)
3478 && type_internals_preclude_sra_p (type, &msg)))
3479 continue;
3480
3481 bitmap_set_bit (candidate_bitmap, DECL_UID (parm));
3482 ret = true;
3483 if (dump_file && (dump_flags & TDF_DETAILS))
3484 {
3485 fprintf (dump_file, "Candidate (%d): ", DECL_UID (parm));
3486 print_generic_expr (dump_file, parm, 0);
3487 fprintf (dump_file, "\n");
3488 }
3489 }
3490
3491 func_param_count = count;
3492 return ret;
3493 }
3494
3495 /* Callback of walk_aliased_vdefs, marks the access passed as DATA as
3496 maybe_modified. */
3497
3498 static bool
3499 mark_maybe_modified (ao_ref *ao ATTRIBUTE_UNUSED, tree vdef ATTRIBUTE_UNUSED,
3500 void *data)
3501 {
3502 struct access *repr = (struct access *) data;
3503
3504 repr->grp_maybe_modified = 1;
3505 return true;
3506 }
3507
3508 /* Analyze what representatives (in linked lists accessible from
3509 REPRESENTATIVES) can be modified by side effects of statements in the
3510 current function. */
3511
3512 static void
3513 analyze_modified_params (VEC (access_p, heap) *representatives)
3514 {
3515 int i;
3516
3517 for (i = 0; i < func_param_count; i++)
3518 {
3519 struct access *repr;
3520
3521 for (repr = VEC_index (access_p, representatives, i);
3522 repr;
3523 repr = repr->next_grp)
3524 {
3525 struct access *access;
3526 bitmap visited;
3527 ao_ref ar;
3528
3529 if (no_accesses_p (repr))
3530 continue;
3531 if (!POINTER_TYPE_P (TREE_TYPE (repr->base))
3532 || repr->grp_maybe_modified)
3533 continue;
3534
3535 ao_ref_init (&ar, repr->expr);
3536 visited = BITMAP_ALLOC (NULL);
3537 for (access = repr; access; access = access->next_sibling)
3538 {
3539 /* All accesses are read ones, otherwise grp_maybe_modified would
3540 be trivially set. */
3541 walk_aliased_vdefs (&ar, gimple_vuse (access->stmt),
3542 mark_maybe_modified, repr, &visited);
3543 if (repr->grp_maybe_modified)
3544 break;
3545 }
3546 BITMAP_FREE (visited);
3547 }
3548 }
3549 }
3550
3551 /* Propagate distances in bb_dereferences in the opposite direction than the
3552 control flow edges, in each step storing the maximum of the current value
3553 and the minimum of all successors. These steps are repeated until the table
3554 stabilizes. Note that BBs which might terminate the functions (according to
3555 final_bbs bitmap) never updated in this way. */
3556
3557 static void
3558 propagate_dereference_distances (void)
3559 {
3560 VEC (basic_block, heap) *queue;
3561 basic_block bb;
3562
3563 queue = VEC_alloc (basic_block, heap, last_basic_block_for_function (cfun));
3564 VEC_quick_push (basic_block, queue, ENTRY_BLOCK_PTR);
3565 FOR_EACH_BB (bb)
3566 {
3567 VEC_quick_push (basic_block, queue, bb);
3568 bb->aux = bb;
3569 }
3570
3571 while (!VEC_empty (basic_block, queue))
3572 {
3573 edge_iterator ei;
3574 edge e;
3575 bool change = false;
3576 int i;
3577
3578 bb = VEC_pop (basic_block, queue);
3579 bb->aux = NULL;
3580
3581 if (bitmap_bit_p (final_bbs, bb->index))
3582 continue;
3583
3584 for (i = 0; i < func_param_count; i++)
3585 {
3586 int idx = bb->index * func_param_count + i;
3587 bool first = true;
3588 HOST_WIDE_INT inh = 0;
3589
3590 FOR_EACH_EDGE (e, ei, bb->succs)
3591 {
3592 int succ_idx = e->dest->index * func_param_count + i;
3593
3594 if (e->src == EXIT_BLOCK_PTR)
3595 continue;
3596
3597 if (first)
3598 {
3599 first = false;
3600 inh = bb_dereferences [succ_idx];
3601 }
3602 else if (bb_dereferences [succ_idx] < inh)
3603 inh = bb_dereferences [succ_idx];
3604 }
3605
3606 if (!first && bb_dereferences[idx] < inh)
3607 {
3608 bb_dereferences[idx] = inh;
3609 change = true;
3610 }
3611 }
3612
3613 if (change && !bitmap_bit_p (final_bbs, bb->index))
3614 FOR_EACH_EDGE (e, ei, bb->preds)
3615 {
3616 if (e->src->aux)
3617 continue;
3618
3619 e->src->aux = e->src;
3620 VEC_quick_push (basic_block, queue, e->src);
3621 }
3622 }
3623
3624 VEC_free (basic_block, heap, queue);
3625 }
3626
3627 /* Dump a dereferences TABLE with heading STR to file F. */
3628
3629 static void
3630 dump_dereferences_table (FILE *f, const char *str, HOST_WIDE_INT *table)
3631 {
3632 basic_block bb;
3633
3634 fprintf (dump_file, str);
3635 FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR, EXIT_BLOCK_PTR, next_bb)
3636 {
3637 fprintf (f, "%4i %i ", bb->index, bitmap_bit_p (final_bbs, bb->index));
3638 if (bb != EXIT_BLOCK_PTR)
3639 {
3640 int i;
3641 for (i = 0; i < func_param_count; i++)
3642 {
3643 int idx = bb->index * func_param_count + i;
3644 fprintf (f, " %4" HOST_WIDE_INT_PRINT "d", table[idx]);
3645 }
3646 }
3647 fprintf (f, "\n");
3648 }
3649 fprintf (dump_file, "\n");
3650 }
3651
3652 /* Determine what (parts of) parameters passed by reference that are not
3653 assigned to are not certainly dereferenced in this function and thus the
3654 dereferencing cannot be safely moved to the caller without potentially
3655 introducing a segfault. Mark such REPRESENTATIVES as
3656 grp_not_necessarilly_dereferenced.
3657
3658 The dereferenced maximum "distance," i.e. the offset + size of the accessed
3659 part is calculated rather than simple booleans are calculated for each
3660 pointer parameter to handle cases when only a fraction of the whole
3661 aggregate is allocated (see testsuite/gcc.c-torture/execute/ipa-sra-2.c for
3662 an example).
3663
3664 The maximum dereference distances for each pointer parameter and BB are
3665 already stored in bb_dereference. This routine simply propagates these
3666 values upwards by propagate_dereference_distances and then compares the
3667 distances of individual parameters in the ENTRY BB to the equivalent
3668 distances of each representative of a (fraction of a) parameter. */
3669
3670 static void
3671 analyze_caller_dereference_legality (VEC (access_p, heap) *representatives)
3672 {
3673 int i;
3674
3675 if (dump_file && (dump_flags & TDF_DETAILS))
3676 dump_dereferences_table (dump_file,
3677 "Dereference table before propagation:\n",
3678 bb_dereferences);
3679
3680 propagate_dereference_distances ();
3681
3682 if (dump_file && (dump_flags & TDF_DETAILS))
3683 dump_dereferences_table (dump_file,
3684 "Dereference table after propagation:\n",
3685 bb_dereferences);
3686
3687 for (i = 0; i < func_param_count; i++)
3688 {
3689 struct access *repr = VEC_index (access_p, representatives, i);
3690 int idx = ENTRY_BLOCK_PTR->index * func_param_count + i;
3691
3692 if (!repr || no_accesses_p (repr))
3693 continue;
3694
3695 do
3696 {
3697 if ((repr->offset + repr->size) > bb_dereferences[idx])
3698 repr->grp_not_necessarilly_dereferenced = 1;
3699 repr = repr->next_grp;
3700 }
3701 while (repr);
3702 }
3703 }
3704
3705 /* Return the representative access for the parameter declaration PARM if it is
3706 a scalar passed by reference which is not written to and the pointer value
3707 is not used directly. Thus, if it is legal to dereference it in the caller
3708 and we can rule out modifications through aliases, such parameter should be
3709 turned into one passed by value. Return NULL otherwise. */
3710
3711 static struct access *
3712 unmodified_by_ref_scalar_representative (tree parm)
3713 {
3714 int i, access_count;
3715 struct access *repr;
3716 VEC (access_p, heap) *access_vec;
3717
3718 access_vec = get_base_access_vector (parm);
3719 gcc_assert (access_vec);
3720 repr = VEC_index (access_p, access_vec, 0);
3721 if (repr->write)
3722 return NULL;
3723 repr->group_representative = repr;
3724
3725 access_count = VEC_length (access_p, access_vec);
3726 for (i = 1; i < access_count; i++)
3727 {
3728 struct access *access = VEC_index (access_p, access_vec, i);
3729 if (access->write)
3730 return NULL;
3731 access->group_representative = repr;
3732 access->next_sibling = repr->next_sibling;
3733 repr->next_sibling = access;
3734 }
3735
3736 repr->grp_read = 1;
3737 repr->grp_scalar_ptr = 1;
3738 return repr;
3739 }
3740
3741 /* Return true iff this access precludes IPA-SRA of the parameter it is
3742 associated with. */
3743
3744 static bool
3745 access_precludes_ipa_sra_p (struct access *access)
3746 {
3747 /* Avoid issues such as the second simple testcase in PR 42025. The problem
3748 is incompatible assign in a call statement (and possibly even in asm
3749 statements). This can be relaxed by using a new temporary but only for
3750 non-TREE_ADDRESSABLE types and is probably not worth the complexity. (In
3751 intraprocedural SRA we deal with this by keeping the old aggregate around,
3752 something we cannot do in IPA-SRA.) */
3753 if (access->write
3754 && (is_gimple_call (access->stmt)
3755 || gimple_code (access->stmt) == GIMPLE_ASM))
3756 return true;
3757
3758 return false;
3759 }
3760
3761
3762 /* Sort collected accesses for parameter PARM, identify representatives for
3763 each accessed region and link them together. Return NULL if there are
3764 different but overlapping accesses, return the special ptr value meaning
3765 there are no accesses for this parameter if that is the case and return the
3766 first representative otherwise. Set *RO_GRP if there is a group of accesses
3767 with only read (i.e. no write) accesses. */
3768
3769 static struct access *
3770 splice_param_accesses (tree parm, bool *ro_grp)
3771 {
3772 int i, j, access_count, group_count;
3773 int agg_size, total_size = 0;
3774 struct access *access, *res, **prev_acc_ptr = &res;
3775 VEC (access_p, heap) *access_vec;
3776
3777 access_vec = get_base_access_vector (parm);
3778 if (!access_vec)
3779 return &no_accesses_representant;
3780 access_count = VEC_length (access_p, access_vec);
3781
3782 VEC_qsort (access_p, access_vec, compare_access_positions);
3783
3784 i = 0;
3785 total_size = 0;
3786 group_count = 0;
3787 while (i < access_count)
3788 {
3789 bool modification;
3790 tree a1_alias_type;
3791 access = VEC_index (access_p, access_vec, i);
3792 modification = access->write;
3793 if (access_precludes_ipa_sra_p (access))
3794 return NULL;
3795 a1_alias_type = reference_alias_ptr_type (access->expr);
3796
3797 /* Access is about to become group representative unless we find some
3798 nasty overlap which would preclude us from breaking this parameter
3799 apart. */
3800
3801 j = i + 1;
3802 while (j < access_count)
3803 {
3804 struct access *ac2 = VEC_index (access_p, access_vec, j);
3805 if (ac2->offset != access->offset)
3806 {
3807 /* All or nothing law for parameters. */
3808 if (access->offset + access->size > ac2->offset)
3809 return NULL;
3810 else
3811 break;
3812 }
3813 else if (ac2->size != access->size)
3814 return NULL;
3815
3816 if (access_precludes_ipa_sra_p (ac2)
3817 || (ac2->type != access->type
3818 && (TREE_ADDRESSABLE (ac2->type)
3819 || TREE_ADDRESSABLE (access->type)))
3820 || (reference_alias_ptr_type (ac2->expr) != a1_alias_type))
3821 return NULL;
3822
3823 modification |= ac2->write;
3824 ac2->group_representative = access;
3825 ac2->next_sibling = access->next_sibling;
3826 access->next_sibling = ac2;
3827 j++;
3828 }
3829
3830 group_count++;
3831 access->grp_maybe_modified = modification;
3832 if (!modification)
3833 *ro_grp = true;
3834 *prev_acc_ptr = access;
3835 prev_acc_ptr = &access->next_grp;
3836 total_size += access->size;
3837 i = j;
3838 }
3839
3840 if (POINTER_TYPE_P (TREE_TYPE (parm)))
3841 agg_size = tree_low_cst (TYPE_SIZE (TREE_TYPE (TREE_TYPE (parm))), 1);
3842 else
3843 agg_size = tree_low_cst (TYPE_SIZE (TREE_TYPE (parm)), 1);
3844 if (total_size >= agg_size)
3845 return NULL;
3846
3847 gcc_assert (group_count > 0);
3848 return res;
3849 }
3850
3851 /* Decide whether parameters with representative accesses given by REPR should
3852 be reduced into components. */
3853
3854 static int
3855 decide_one_param_reduction (struct access *repr)
3856 {
3857 int total_size, cur_parm_size, agg_size, new_param_count, parm_size_limit;
3858 bool by_ref;
3859 tree parm;
3860
3861 parm = repr->base;
3862 cur_parm_size = tree_low_cst (TYPE_SIZE (TREE_TYPE (parm)), 1);
3863 gcc_assert (cur_parm_size > 0);
3864
3865 if (POINTER_TYPE_P (TREE_TYPE (parm)))
3866 {
3867 by_ref = true;
3868 agg_size = tree_low_cst (TYPE_SIZE (TREE_TYPE (TREE_TYPE (parm))), 1);
3869 }
3870 else
3871 {
3872 by_ref = false;
3873 agg_size = cur_parm_size;
3874 }
3875
3876 if (dump_file)
3877 {
3878 struct access *acc;
3879 fprintf (dump_file, "Evaluating PARAM group sizes for ");
3880 print_generic_expr (dump_file, parm, 0);
3881 fprintf (dump_file, " (UID: %u): \n", DECL_UID (parm));
3882 for (acc = repr; acc; acc = acc->next_grp)
3883 dump_access (dump_file, acc, true);
3884 }
3885
3886 total_size = 0;
3887 new_param_count = 0;
3888
3889 for (; repr; repr = repr->next_grp)
3890 {
3891 gcc_assert (parm == repr->base);
3892
3893 /* Taking the address of a non-addressable field is verboten. */
3894 if (by_ref && repr->non_addressable)
3895 return 0;
3896
3897 /* Do not decompose a non-BLKmode param in a way that would
3898 create BLKmode params. Especially for by-reference passing
3899 (thus, pointer-type param) this is hardly worthwhile. */
3900 if (DECL_MODE (parm) != BLKmode
3901 && TYPE_MODE (repr->type) == BLKmode)
3902 return 0;
3903
3904 if (!by_ref || (!repr->grp_maybe_modified
3905 && !repr->grp_not_necessarilly_dereferenced))
3906 total_size += repr->size;
3907 else
3908 total_size += cur_parm_size;
3909
3910 new_param_count++;
3911 }
3912
3913 gcc_assert (new_param_count > 0);
3914
3915 if (optimize_function_for_size_p (cfun))
3916 parm_size_limit = cur_parm_size;
3917 else
3918 parm_size_limit = (PARAM_VALUE (PARAM_IPA_SRA_PTR_GROWTH_FACTOR)
3919 * cur_parm_size);
3920
3921 if (total_size < agg_size
3922 && total_size <= parm_size_limit)
3923 {
3924 if (dump_file)
3925 fprintf (dump_file, " ....will be split into %i components\n",
3926 new_param_count);
3927 return new_param_count;
3928 }
3929 else
3930 return 0;
3931 }
3932
3933 /* The order of the following enums is important, we need to do extra work for
3934 UNUSED_PARAMS, BY_VAL_ACCESSES and UNMODIF_BY_REF_ACCESSES. */
3935 enum ipa_splicing_result { NO_GOOD_ACCESS, UNUSED_PARAMS, BY_VAL_ACCESSES,
3936 MODIF_BY_REF_ACCESSES, UNMODIF_BY_REF_ACCESSES };
3937
3938 /* Identify representatives of all accesses to all candidate parameters for
3939 IPA-SRA. Return result based on what representatives have been found. */
3940
3941 static enum ipa_splicing_result
3942 splice_all_param_accesses (VEC (access_p, heap) **representatives)
3943 {
3944 enum ipa_splicing_result result = NO_GOOD_ACCESS;
3945 tree parm;
3946 struct access *repr;
3947
3948 *representatives = VEC_alloc (access_p, heap, func_param_count);
3949
3950 for (parm = DECL_ARGUMENTS (current_function_decl);
3951 parm;
3952 parm = DECL_CHAIN (parm))
3953 {
3954 if (is_unused_scalar_param (parm))
3955 {
3956 VEC_quick_push (access_p, *representatives,
3957 &no_accesses_representant);
3958 if (result == NO_GOOD_ACCESS)
3959 result = UNUSED_PARAMS;
3960 }
3961 else if (POINTER_TYPE_P (TREE_TYPE (parm))
3962 && is_gimple_reg_type (TREE_TYPE (TREE_TYPE (parm)))
3963 && bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
3964 {
3965 repr = unmodified_by_ref_scalar_representative (parm);
3966 VEC_quick_push (access_p, *representatives, repr);
3967 if (repr)
3968 result = UNMODIF_BY_REF_ACCESSES;
3969 }
3970 else if (bitmap_bit_p (candidate_bitmap, DECL_UID (parm)))
3971 {
3972 bool ro_grp = false;
3973 repr = splice_param_accesses (parm, &ro_grp);
3974 VEC_quick_push (access_p, *representatives, repr);
3975
3976 if (repr && !no_accesses_p (repr))
3977 {
3978 if (POINTER_TYPE_P (TREE_TYPE (parm)))
3979 {
3980 if (ro_grp)
3981 result = UNMODIF_BY_REF_ACCESSES;
3982 else if (result < MODIF_BY_REF_ACCESSES)
3983 result = MODIF_BY_REF_ACCESSES;
3984 }
3985 else if (result < BY_VAL_ACCESSES)
3986 result = BY_VAL_ACCESSES;
3987 }
3988 else if (no_accesses_p (repr) && (result == NO_GOOD_ACCESS))
3989 result = UNUSED_PARAMS;
3990 }
3991 else
3992 VEC_quick_push (access_p, *representatives, NULL);
3993 }
3994
3995 if (result == NO_GOOD_ACCESS)
3996 {
3997 VEC_free (access_p, heap, *representatives);
3998 *representatives = NULL;
3999 return NO_GOOD_ACCESS;
4000 }
4001
4002 return result;
4003 }
4004
4005 /* Return the index of BASE in PARMS. Abort if it is not found. */
4006
4007 static inline int
4008 get_param_index (tree base, VEC(tree, heap) *parms)
4009 {
4010 int i, len;
4011
4012 len = VEC_length (tree, parms);
4013 for (i = 0; i < len; i++)
4014 if (VEC_index (tree, parms, i) == base)
4015 return i;
4016 gcc_unreachable ();
4017 }
4018
4019 /* Convert the decisions made at the representative level into compact
4020 parameter adjustments. REPRESENTATIVES are pointers to first
4021 representatives of each param accesses, ADJUSTMENTS_COUNT is the expected
4022 final number of adjustments. */
4023
4024 static ipa_parm_adjustment_vec
4025 turn_representatives_into_adjustments (VEC (access_p, heap) *representatives,
4026 int adjustments_count)
4027 {
4028 VEC (tree, heap) *parms;
4029 ipa_parm_adjustment_vec adjustments;
4030 tree parm;
4031 int i;
4032
4033 gcc_assert (adjustments_count > 0);
4034 parms = ipa_get_vector_of_formal_parms (current_function_decl);
4035 adjustments = VEC_alloc (ipa_parm_adjustment_t, heap, adjustments_count);
4036 parm = DECL_ARGUMENTS (current_function_decl);
4037 for (i = 0; i < func_param_count; i++, parm = DECL_CHAIN (parm))
4038 {
4039 struct access *repr = VEC_index (access_p, representatives, i);
4040
4041 if (!repr || no_accesses_p (repr))
4042 {
4043 struct ipa_parm_adjustment *adj;
4044
4045 adj = VEC_quick_push (ipa_parm_adjustment_t, adjustments, NULL);
4046 memset (adj, 0, sizeof (*adj));
4047 adj->base_index = get_param_index (parm, parms);
4048 adj->base = parm;
4049 if (!repr)
4050 adj->copy_param = 1;
4051 else
4052 adj->remove_param = 1;
4053 }
4054 else
4055 {
4056 struct ipa_parm_adjustment *adj;
4057 int index = get_param_index (parm, parms);
4058
4059 for (; repr; repr = repr->next_grp)
4060 {
4061 adj = VEC_quick_push (ipa_parm_adjustment_t, adjustments, NULL);
4062 memset (adj, 0, sizeof (*adj));
4063 gcc_assert (repr->base == parm);
4064 adj->base_index = index;
4065 adj->base = repr->base;
4066 adj->type = repr->type;
4067 adj->alias_ptr_type = reference_alias_ptr_type (repr->expr);
4068 adj->offset = repr->offset;
4069 adj->by_ref = (POINTER_TYPE_P (TREE_TYPE (repr->base))
4070 && (repr->grp_maybe_modified
4071 || repr->grp_not_necessarilly_dereferenced));
4072
4073 }
4074 }
4075 }
4076 VEC_free (tree, heap, parms);
4077 return adjustments;
4078 }
4079
4080 /* Analyze the collected accesses and produce a plan what to do with the
4081 parameters in the form of adjustments, NULL meaning nothing. */
4082
4083 static ipa_parm_adjustment_vec
4084 analyze_all_param_acesses (void)
4085 {
4086 enum ipa_splicing_result repr_state;
4087 bool proceed = false;
4088 int i, adjustments_count = 0;
4089 VEC (access_p, heap) *representatives;
4090 ipa_parm_adjustment_vec adjustments;
4091
4092 repr_state = splice_all_param_accesses (&representatives);
4093 if (repr_state == NO_GOOD_ACCESS)
4094 return NULL;
4095
4096 /* If there are any parameters passed by reference which are not modified
4097 directly, we need to check whether they can be modified indirectly. */
4098 if (repr_state == UNMODIF_BY_REF_ACCESSES)
4099 {
4100 analyze_caller_dereference_legality (representatives);
4101 analyze_modified_params (representatives);
4102 }
4103
4104 for (i = 0; i < func_param_count; i++)
4105 {
4106 struct access *repr = VEC_index (access_p, representatives, i);
4107
4108 if (repr && !no_accesses_p (repr))
4109 {
4110 if (repr->grp_scalar_ptr)
4111 {
4112 adjustments_count++;
4113 if (repr->grp_not_necessarilly_dereferenced
4114 || repr->grp_maybe_modified)
4115 VEC_replace (access_p, representatives, i, NULL);
4116 else
4117 {
4118 proceed = true;
4119 sra_stats.scalar_by_ref_to_by_val++;
4120 }
4121 }
4122 else
4123 {
4124 int new_components = decide_one_param_reduction (repr);
4125
4126 if (new_components == 0)
4127 {
4128 VEC_replace (access_p, representatives, i, NULL);
4129 adjustments_count++;
4130 }
4131 else
4132 {
4133 adjustments_count += new_components;
4134 sra_stats.aggregate_params_reduced++;
4135 sra_stats.param_reductions_created += new_components;
4136 proceed = true;
4137 }
4138 }
4139 }
4140 else
4141 {
4142 if (no_accesses_p (repr))
4143 {
4144 proceed = true;
4145 sra_stats.deleted_unused_parameters++;
4146 }
4147 adjustments_count++;
4148 }
4149 }
4150
4151 if (!proceed && dump_file)
4152 fprintf (dump_file, "NOT proceeding to change params.\n");
4153
4154 if (proceed)
4155 adjustments = turn_representatives_into_adjustments (representatives,
4156 adjustments_count);
4157 else
4158 adjustments = NULL;
4159
4160 VEC_free (access_p, heap, representatives);
4161 return adjustments;
4162 }
4163
4164 /* If a parameter replacement identified by ADJ does not yet exist in the form
4165 of declaration, create it and record it, otherwise return the previously
4166 created one. */
4167
4168 static tree
4169 get_replaced_param_substitute (struct ipa_parm_adjustment *adj)
4170 {
4171 tree repl;
4172 if (!adj->new_ssa_base)
4173 {
4174 char *pretty_name = make_fancy_name (adj->base);
4175
4176 repl = create_tmp_reg (TREE_TYPE (adj->base), "ISR");
4177 DECL_NAME (repl) = get_identifier (pretty_name);
4178 obstack_free (&name_obstack, pretty_name);
4179
4180 add_referenced_var (repl);
4181 adj->new_ssa_base = repl;
4182 }
4183 else
4184 repl = adj->new_ssa_base;
4185 return repl;
4186 }
4187
4188 /* Find the first adjustment for a particular parameter BASE in a vector of
4189 ADJUSTMENTS which is not a copy_param. Return NULL if there is no such
4190 adjustment. */
4191
4192 static struct ipa_parm_adjustment *
4193 get_adjustment_for_base (ipa_parm_adjustment_vec adjustments, tree base)
4194 {
4195 int i, len;
4196
4197 len = VEC_length (ipa_parm_adjustment_t, adjustments);
4198 for (i = 0; i < len; i++)
4199 {
4200 struct ipa_parm_adjustment *adj;
4201
4202 adj = VEC_index (ipa_parm_adjustment_t, adjustments, i);
4203 if (!adj->copy_param && adj->base == base)
4204 return adj;
4205 }
4206
4207 return NULL;
4208 }
4209
4210 /* If the statement STMT defines an SSA_NAME of a parameter which is to be
4211 removed because its value is not used, replace the SSA_NAME with a one
4212 relating to a created VAR_DECL together all of its uses and return true.
4213 ADJUSTMENTS is a pointer to an adjustments vector. */
4214
4215 static bool
4216 replace_removed_params_ssa_names (gimple stmt,
4217 ipa_parm_adjustment_vec adjustments)
4218 {
4219 struct ipa_parm_adjustment *adj;
4220 tree lhs, decl, repl, name;
4221
4222 if (gimple_code (stmt) == GIMPLE_PHI)
4223 lhs = gimple_phi_result (stmt);
4224 else if (is_gimple_assign (stmt))
4225 lhs = gimple_assign_lhs (stmt);
4226 else if (is_gimple_call (stmt))
4227 lhs = gimple_call_lhs (stmt);
4228 else
4229 gcc_unreachable ();
4230
4231 if (TREE_CODE (lhs) != SSA_NAME)
4232 return false;
4233 decl = SSA_NAME_VAR (lhs);
4234 if (TREE_CODE (decl) != PARM_DECL)
4235 return false;
4236
4237 adj = get_adjustment_for_base (adjustments, decl);
4238 if (!adj)
4239 return false;
4240
4241 repl = get_replaced_param_substitute (adj);
4242 name = make_ssa_name (repl, stmt);
4243
4244 if (dump_file)
4245 {
4246 fprintf (dump_file, "replacing an SSA name of a removed param ");
4247 print_generic_expr (dump_file, lhs, 0);
4248 fprintf (dump_file, " with ");
4249 print_generic_expr (dump_file, name, 0);
4250 fprintf (dump_file, "\n");
4251 }
4252
4253 if (is_gimple_assign (stmt))
4254 gimple_assign_set_lhs (stmt, name);
4255 else if (is_gimple_call (stmt))
4256 gimple_call_set_lhs (stmt, name);
4257 else
4258 gimple_phi_set_result (stmt, name);
4259
4260 replace_uses_by (lhs, name);
4261 release_ssa_name (lhs);
4262 return true;
4263 }
4264
4265 /* If the expression *EXPR should be replaced by a reduction of a parameter, do
4266 so. ADJUSTMENTS is a pointer to a vector of adjustments. CONVERT
4267 specifies whether the function should care about type incompatibility the
4268 current and new expressions. If it is false, the function will leave
4269 incompatibility issues to the caller. Return true iff the expression
4270 was modified. */
4271
4272 static bool
4273 sra_ipa_modify_expr (tree *expr, bool convert,
4274 ipa_parm_adjustment_vec adjustments)
4275 {
4276 int i, len;
4277 struct ipa_parm_adjustment *adj, *cand = NULL;
4278 HOST_WIDE_INT offset, size, max_size;
4279 tree base, src;
4280
4281 len = VEC_length (ipa_parm_adjustment_t, adjustments);
4282
4283 if (TREE_CODE (*expr) == BIT_FIELD_REF
4284 || TREE_CODE (*expr) == IMAGPART_EXPR
4285 || TREE_CODE (*expr) == REALPART_EXPR)
4286 {
4287 expr = &TREE_OPERAND (*expr, 0);
4288 convert = true;
4289 }
4290
4291 base = get_ref_base_and_extent (*expr, &offset, &size, &max_size);
4292 if (!base || size == -1 || max_size == -1)
4293 return false;
4294
4295 if (TREE_CODE (base) == MEM_REF)
4296 {
4297 offset += mem_ref_offset (base).low * BITS_PER_UNIT;
4298 base = TREE_OPERAND (base, 0);
4299 }
4300
4301 base = get_ssa_base_param (base);
4302 if (!base || TREE_CODE (base) != PARM_DECL)
4303 return false;
4304
4305 for (i = 0; i < len; i++)
4306 {
4307 adj = VEC_index (ipa_parm_adjustment_t, adjustments, i);
4308
4309 if (adj->base == base &&
4310 (adj->offset == offset || adj->remove_param))
4311 {
4312 cand = adj;
4313 break;
4314 }
4315 }
4316 if (!cand || cand->copy_param || cand->remove_param)
4317 return false;
4318
4319 if (cand->by_ref)
4320 src = build_simple_mem_ref (cand->reduction);
4321 else
4322 src = cand->reduction;
4323
4324 if (dump_file && (dump_flags & TDF_DETAILS))
4325 {
4326 fprintf (dump_file, "About to replace expr ");
4327 print_generic_expr (dump_file, *expr, 0);
4328 fprintf (dump_file, " with ");
4329 print_generic_expr (dump_file, src, 0);
4330 fprintf (dump_file, "\n");
4331 }
4332
4333 if (convert && !useless_type_conversion_p (TREE_TYPE (*expr), cand->type))
4334 {
4335 tree vce = build1 (VIEW_CONVERT_EXPR, TREE_TYPE (*expr), src);
4336 *expr = vce;
4337 }
4338 else
4339 *expr = src;
4340 return true;
4341 }
4342
4343 /* If the statement pointed to by STMT_PTR contains any expressions that need
4344 to replaced with a different one as noted by ADJUSTMENTS, do so. Handle any
4345 potential type incompatibilities (GSI is used to accommodate conversion
4346 statements and must point to the statement). Return true iff the statement
4347 was modified. */
4348
4349 static bool
4350 sra_ipa_modify_assign (gimple *stmt_ptr, gimple_stmt_iterator *gsi,
4351 ipa_parm_adjustment_vec adjustments)
4352 {
4353 gimple stmt = *stmt_ptr;
4354 tree *lhs_p, *rhs_p;
4355 bool any;
4356
4357 if (!gimple_assign_single_p (stmt))
4358 return false;
4359
4360 rhs_p = gimple_assign_rhs1_ptr (stmt);
4361 lhs_p = gimple_assign_lhs_ptr (stmt);
4362
4363 any = sra_ipa_modify_expr (rhs_p, false, adjustments);
4364 any |= sra_ipa_modify_expr (lhs_p, false, adjustments);
4365 if (any)
4366 {
4367 tree new_rhs = NULL_TREE;
4368
4369 if (!useless_type_conversion_p (TREE_TYPE (*lhs_p), TREE_TYPE (*rhs_p)))
4370 {
4371 if (TREE_CODE (*rhs_p) == CONSTRUCTOR)
4372 {
4373 /* V_C_Es of constructors can cause trouble (PR 42714). */
4374 if (is_gimple_reg_type (TREE_TYPE (*lhs_p)))
4375 *rhs_p = build_zero_cst (TREE_TYPE (*lhs_p));
4376 else
4377 *rhs_p = build_constructor (TREE_TYPE (*lhs_p), 0);
4378 }
4379 else
4380 new_rhs = fold_build1_loc (gimple_location (stmt),
4381 VIEW_CONVERT_EXPR, TREE_TYPE (*lhs_p),
4382 *rhs_p);
4383 }
4384 else if (REFERENCE_CLASS_P (*rhs_p)
4385 && is_gimple_reg_type (TREE_TYPE (*lhs_p))
4386 && !is_gimple_reg (*lhs_p))
4387 /* This can happen when an assignment in between two single field
4388 structures is turned into an assignment in between two pointers to
4389 scalars (PR 42237). */
4390 new_rhs = *rhs_p;
4391
4392 if (new_rhs)
4393 {
4394 tree tmp = force_gimple_operand_gsi (gsi, new_rhs, true, NULL_TREE,
4395 true, GSI_SAME_STMT);
4396
4397 gimple_assign_set_rhs_from_tree (gsi, tmp);
4398 }
4399
4400 return true;
4401 }
4402
4403 return false;
4404 }
4405
4406 /* Traverse the function body and all modifications as described in
4407 ADJUSTMENTS. Return true iff the CFG has been changed. */
4408
4409 static bool
4410 ipa_sra_modify_function_body (ipa_parm_adjustment_vec adjustments)
4411 {
4412 bool cfg_changed = false;
4413 basic_block bb;
4414
4415 FOR_EACH_BB (bb)
4416 {
4417 gimple_stmt_iterator gsi;
4418
4419 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
4420 replace_removed_params_ssa_names (gsi_stmt (gsi), adjustments);
4421
4422 gsi = gsi_start_bb (bb);
4423 while (!gsi_end_p (gsi))
4424 {
4425 gimple stmt = gsi_stmt (gsi);
4426 bool modified = false;
4427 tree *t;
4428 unsigned i;
4429
4430 switch (gimple_code (stmt))
4431 {
4432 case GIMPLE_RETURN:
4433 t = gimple_return_retval_ptr (stmt);
4434 if (*t != NULL_TREE)
4435 modified |= sra_ipa_modify_expr (t, true, adjustments);
4436 break;
4437
4438 case GIMPLE_ASSIGN:
4439 modified |= sra_ipa_modify_assign (&stmt, &gsi, adjustments);
4440 modified |= replace_removed_params_ssa_names (stmt, adjustments);
4441 break;
4442
4443 case GIMPLE_CALL:
4444 /* Operands must be processed before the lhs. */
4445 for (i = 0; i < gimple_call_num_args (stmt); i++)
4446 {
4447 t = gimple_call_arg_ptr (stmt, i);
4448 modified |= sra_ipa_modify_expr (t, true, adjustments);
4449 }
4450
4451 if (gimple_call_lhs (stmt))
4452 {
4453 t = gimple_call_lhs_ptr (stmt);
4454 modified |= sra_ipa_modify_expr (t, false, adjustments);
4455 modified |= replace_removed_params_ssa_names (stmt,
4456 adjustments);
4457 }
4458 break;
4459
4460 case GIMPLE_ASM:
4461 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
4462 {
4463 t = &TREE_VALUE (gimple_asm_input_op (stmt, i));
4464 modified |= sra_ipa_modify_expr (t, true, adjustments);
4465 }
4466 for (i = 0; i < gimple_asm_noutputs (stmt); i++)
4467 {
4468 t = &TREE_VALUE (gimple_asm_output_op (stmt, i));
4469 modified |= sra_ipa_modify_expr (t, false, adjustments);
4470 }
4471 break;
4472
4473 default:
4474 break;
4475 }
4476
4477 if (modified)
4478 {
4479 update_stmt (stmt);
4480 if (maybe_clean_eh_stmt (stmt)
4481 && gimple_purge_dead_eh_edges (gimple_bb (stmt)))
4482 cfg_changed = true;
4483 }
4484 gsi_next (&gsi);
4485 }
4486 }
4487
4488 return cfg_changed;
4489 }
4490
4491 /* Call gimple_debug_bind_reset_value on all debug statements describing
4492 gimple register parameters that are being removed or replaced. */
4493
4494 static void
4495 sra_ipa_reset_debug_stmts (ipa_parm_adjustment_vec adjustments)
4496 {
4497 int i, len;
4498 gimple_stmt_iterator *gsip = NULL, gsi;
4499
4500 if (MAY_HAVE_DEBUG_STMTS && single_succ_p (ENTRY_BLOCK_PTR))
4501 {
4502 gsi = gsi_after_labels (single_succ (ENTRY_BLOCK_PTR));
4503 gsip = &gsi;
4504 }
4505 len = VEC_length (ipa_parm_adjustment_t, adjustments);
4506 for (i = 0; i < len; i++)
4507 {
4508 struct ipa_parm_adjustment *adj;
4509 imm_use_iterator ui;
4510 gimple stmt, def_temp;
4511 tree name, vexpr, copy = NULL_TREE;
4512 use_operand_p use_p;
4513
4514 adj = VEC_index (ipa_parm_adjustment_t, adjustments, i);
4515 if (adj->copy_param || !is_gimple_reg (adj->base))
4516 continue;
4517 name = gimple_default_def (cfun, adj->base);
4518 vexpr = NULL;
4519 if (name)
4520 FOR_EACH_IMM_USE_STMT (stmt, ui, name)
4521 {
4522 /* All other users must have been removed by
4523 ipa_sra_modify_function_body. */
4524 gcc_assert (is_gimple_debug (stmt));
4525 if (vexpr == NULL && gsip != NULL)
4526 {
4527 gcc_assert (TREE_CODE (adj->base) == PARM_DECL);
4528 vexpr = make_node (DEBUG_EXPR_DECL);
4529 def_temp = gimple_build_debug_source_bind (vexpr, adj->base,
4530 NULL);
4531 DECL_ARTIFICIAL (vexpr) = 1;
4532 TREE_TYPE (vexpr) = TREE_TYPE (name);
4533 DECL_MODE (vexpr) = DECL_MODE (adj->base);
4534 gsi_insert_before (gsip, def_temp, GSI_SAME_STMT);
4535 }
4536 if (vexpr)
4537 {
4538 FOR_EACH_IMM_USE_ON_STMT (use_p, ui)
4539 SET_USE (use_p, vexpr);
4540 }
4541 else
4542 gimple_debug_bind_reset_value (stmt);
4543 update_stmt (stmt);
4544 }
4545 /* Create a VAR_DECL for debug info purposes. */
4546 if (!DECL_IGNORED_P (adj->base))
4547 {
4548 copy = build_decl (DECL_SOURCE_LOCATION (current_function_decl),
4549 VAR_DECL, DECL_NAME (adj->base),
4550 TREE_TYPE (adj->base));
4551 if (DECL_PT_UID_SET_P (adj->base))
4552 SET_DECL_PT_UID (copy, DECL_PT_UID (adj->base));
4553 TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (adj->base);
4554 TREE_READONLY (copy) = TREE_READONLY (adj->base);
4555 TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (adj->base);
4556 DECL_GIMPLE_REG_P (copy) = DECL_GIMPLE_REG_P (adj->base);
4557 DECL_ARTIFICIAL (copy) = DECL_ARTIFICIAL (adj->base);
4558 DECL_IGNORED_P (copy) = DECL_IGNORED_P (adj->base);
4559 DECL_ABSTRACT_ORIGIN (copy) = DECL_ORIGIN (adj->base);
4560 DECL_SEEN_IN_BIND_EXPR_P (copy) = 1;
4561 SET_DECL_RTL (copy, 0);
4562 TREE_USED (copy) = 1;
4563 DECL_CONTEXT (copy) = current_function_decl;
4564 add_referenced_var (copy);
4565 add_local_decl (cfun, copy);
4566 DECL_CHAIN (copy) =
4567 BLOCK_VARS (DECL_INITIAL (current_function_decl));
4568 BLOCK_VARS (DECL_INITIAL (current_function_decl)) = copy;
4569 }
4570 if (gsip != NULL && copy && target_for_debug_bind (adj->base))
4571 {
4572 gcc_assert (TREE_CODE (adj->base) == PARM_DECL);
4573 if (vexpr)
4574 def_temp = gimple_build_debug_bind (copy, vexpr, NULL);
4575 else
4576 def_temp = gimple_build_debug_source_bind (copy, adj->base,
4577 NULL);
4578 gsi_insert_before (gsip, def_temp, GSI_SAME_STMT);
4579 }
4580 }
4581 }
4582
4583 /* Return false iff all callers have at least as many actual arguments as there
4584 are formal parameters in the current function. */
4585
4586 static bool
4587 not_all_callers_have_enough_arguments_p (struct cgraph_node *node,
4588 void *data ATTRIBUTE_UNUSED)
4589 {
4590 struct cgraph_edge *cs;
4591 for (cs = node->callers; cs; cs = cs->next_caller)
4592 if (!callsite_has_enough_arguments_p (cs->call_stmt))
4593 return true;
4594
4595 return false;
4596 }
4597
4598 /* Convert all callers of NODE. */
4599
4600 static bool
4601 convert_callers_for_node (struct cgraph_node *node,
4602 void *data)
4603 {
4604 ipa_parm_adjustment_vec adjustments = (ipa_parm_adjustment_vec)data;
4605 bitmap recomputed_callers = BITMAP_ALLOC (NULL);
4606 struct cgraph_edge *cs;
4607
4608 for (cs = node->callers; cs; cs = cs->next_caller)
4609 {
4610 current_function_decl = cs->caller->symbol.decl;
4611 push_cfun (DECL_STRUCT_FUNCTION (cs->caller->symbol.decl));
4612
4613 if (dump_file)
4614 fprintf (dump_file, "Adjusting call (%i -> %i) %s -> %s\n",
4615 cs->caller->uid, cs->callee->uid,
4616 xstrdup (cgraph_node_name (cs->caller)),
4617 xstrdup (cgraph_node_name (cs->callee)));
4618
4619 ipa_modify_call_arguments (cs, cs->call_stmt, adjustments);
4620
4621 pop_cfun ();
4622 }
4623
4624 for (cs = node->callers; cs; cs = cs->next_caller)
4625 if (bitmap_set_bit (recomputed_callers, cs->caller->uid)
4626 && gimple_in_ssa_p (DECL_STRUCT_FUNCTION (cs->caller->symbol.decl)))
4627 compute_inline_parameters (cs->caller, true);
4628 BITMAP_FREE (recomputed_callers);
4629
4630 return true;
4631 }
4632
4633 /* Convert all callers of NODE to pass parameters as given in ADJUSTMENTS. */
4634
4635 static void
4636 convert_callers (struct cgraph_node *node, tree old_decl,
4637 ipa_parm_adjustment_vec adjustments)
4638 {
4639 tree old_cur_fndecl = current_function_decl;
4640 basic_block this_block;
4641
4642 cgraph_for_node_and_aliases (node, convert_callers_for_node,
4643 adjustments, false);
4644
4645 current_function_decl = old_cur_fndecl;
4646
4647 if (!encountered_recursive_call)
4648 return;
4649
4650 FOR_EACH_BB (this_block)
4651 {
4652 gimple_stmt_iterator gsi;
4653
4654 for (gsi = gsi_start_bb (this_block); !gsi_end_p (gsi); gsi_next (&gsi))
4655 {
4656 gimple stmt = gsi_stmt (gsi);
4657 tree call_fndecl;
4658 if (gimple_code (stmt) != GIMPLE_CALL)
4659 continue;
4660 call_fndecl = gimple_call_fndecl (stmt);
4661 if (call_fndecl == old_decl)
4662 {
4663 if (dump_file)
4664 fprintf (dump_file, "Adjusting recursive call");
4665 gimple_call_set_fndecl (stmt, node->symbol.decl);
4666 ipa_modify_call_arguments (NULL, stmt, adjustments);
4667 }
4668 }
4669 }
4670
4671 return;
4672 }
4673
4674 /* Perform all the modification required in IPA-SRA for NODE to have parameters
4675 as given in ADJUSTMENTS. Return true iff the CFG has been changed. */
4676
4677 static bool
4678 modify_function (struct cgraph_node *node, ipa_parm_adjustment_vec adjustments)
4679 {
4680 struct cgraph_node *new_node;
4681 bool cfg_changed;
4682 VEC (cgraph_edge_p, heap) * redirect_callers = collect_callers_of_node (node);
4683
4684 rebuild_cgraph_edges ();
4685 free_dominance_info (CDI_DOMINATORS);
4686 pop_cfun ();
4687 current_function_decl = NULL_TREE;
4688
4689 new_node = cgraph_function_versioning (node, redirect_callers, NULL, NULL,
4690 false, NULL, NULL, "isra");
4691 current_function_decl = new_node->symbol.decl;
4692 push_cfun (DECL_STRUCT_FUNCTION (new_node->symbol.decl));
4693
4694 ipa_modify_formal_parameters (current_function_decl, adjustments, "ISRA");
4695 cfg_changed = ipa_sra_modify_function_body (adjustments);
4696 sra_ipa_reset_debug_stmts (adjustments);
4697 convert_callers (new_node, node->symbol.decl, adjustments);
4698 cgraph_make_node_local (new_node);
4699 return cfg_changed;
4700 }
4701
4702 /* Return false the function is apparently unsuitable for IPA-SRA based on it's
4703 attributes, return true otherwise. NODE is the cgraph node of the current
4704 function. */
4705
4706 static bool
4707 ipa_sra_preliminary_function_checks (struct cgraph_node *node)
4708 {
4709 if (!cgraph_node_can_be_local_p (node))
4710 {
4711 if (dump_file)
4712 fprintf (dump_file, "Function not local to this compilation unit.\n");
4713 return false;
4714 }
4715
4716 if (!node->local.can_change_signature)
4717 {
4718 if (dump_file)
4719 fprintf (dump_file, "Function can not change signature.\n");
4720 return false;
4721 }
4722
4723 if (!tree_versionable_function_p (node->symbol.decl))
4724 {
4725 if (dump_file)
4726 fprintf (dump_file, "Function is not versionable.\n");
4727 return false;
4728 }
4729
4730 if (DECL_VIRTUAL_P (current_function_decl))
4731 {
4732 if (dump_file)
4733 fprintf (dump_file, "Function is a virtual method.\n");
4734 return false;
4735 }
4736
4737 if ((DECL_COMDAT (node->symbol.decl) || DECL_EXTERNAL (node->symbol.decl))
4738 && inline_summary(node)->size >= MAX_INLINE_INSNS_AUTO)
4739 {
4740 if (dump_file)
4741 fprintf (dump_file, "Function too big to be made truly local.\n");
4742 return false;
4743 }
4744
4745 if (!node->callers)
4746 {
4747 if (dump_file)
4748 fprintf (dump_file,
4749 "Function has no callers in this compilation unit.\n");
4750 return false;
4751 }
4752
4753 if (cfun->stdarg)
4754 {
4755 if (dump_file)
4756 fprintf (dump_file, "Function uses stdarg. \n");
4757 return false;
4758 }
4759
4760 if (TYPE_ATTRIBUTES (TREE_TYPE (node->symbol.decl)))
4761 return false;
4762
4763 return true;
4764 }
4765
4766 /* Perform early interprocedural SRA. */
4767
4768 static unsigned int
4769 ipa_early_sra (void)
4770 {
4771 struct cgraph_node *node = cgraph_get_node (current_function_decl);
4772 ipa_parm_adjustment_vec adjustments;
4773 int ret = 0;
4774
4775 if (!ipa_sra_preliminary_function_checks (node))
4776 return 0;
4777
4778 sra_initialize ();
4779 sra_mode = SRA_MODE_EARLY_IPA;
4780
4781 if (!find_param_candidates ())
4782 {
4783 if (dump_file)
4784 fprintf (dump_file, "Function has no IPA-SRA candidates.\n");
4785 goto simple_out;
4786 }
4787
4788 if (cgraph_for_node_and_aliases (node, not_all_callers_have_enough_arguments_p,
4789 NULL, true))
4790 {
4791 if (dump_file)
4792 fprintf (dump_file, "There are callers with insufficient number of "
4793 "arguments.\n");
4794 goto simple_out;
4795 }
4796
4797 bb_dereferences = XCNEWVEC (HOST_WIDE_INT,
4798 func_param_count
4799 * last_basic_block_for_function (cfun));
4800 final_bbs = BITMAP_ALLOC (NULL);
4801
4802 scan_function ();
4803 if (encountered_apply_args)
4804 {
4805 if (dump_file)
4806 fprintf (dump_file, "Function calls __builtin_apply_args().\n");
4807 goto out;
4808 }
4809
4810 if (encountered_unchangable_recursive_call)
4811 {
4812 if (dump_file)
4813 fprintf (dump_file, "Function calls itself with insufficient "
4814 "number of arguments.\n");
4815 goto out;
4816 }
4817
4818 adjustments = analyze_all_param_acesses ();
4819 if (!adjustments)
4820 goto out;
4821 if (dump_file)
4822 ipa_dump_param_adjustments (dump_file, adjustments, current_function_decl);
4823
4824 if (modify_function (node, adjustments))
4825 ret = TODO_update_ssa | TODO_cleanup_cfg;
4826 else
4827 ret = TODO_update_ssa;
4828 VEC_free (ipa_parm_adjustment_t, heap, adjustments);
4829
4830 statistics_counter_event (cfun, "Unused parameters deleted",
4831 sra_stats.deleted_unused_parameters);
4832 statistics_counter_event (cfun, "Scalar parameters converted to by-value",
4833 sra_stats.scalar_by_ref_to_by_val);
4834 statistics_counter_event (cfun, "Aggregate parameters broken up",
4835 sra_stats.aggregate_params_reduced);
4836 statistics_counter_event (cfun, "Aggregate parameter components created",
4837 sra_stats.param_reductions_created);
4838
4839 out:
4840 BITMAP_FREE (final_bbs);
4841 free (bb_dereferences);
4842 simple_out:
4843 sra_deinitialize ();
4844 return ret;
4845 }
4846
4847 /* Return if early ipa sra shall be performed. */
4848 static bool
4849 ipa_early_sra_gate (void)
4850 {
4851 return flag_ipa_sra && dbg_cnt (eipa_sra);
4852 }
4853
4854 struct gimple_opt_pass pass_early_ipa_sra =
4855 {
4856 {
4857 GIMPLE_PASS,
4858 "eipa_sra", /* name */
4859 ipa_early_sra_gate, /* gate */
4860 ipa_early_sra, /* execute */
4861 NULL, /* sub */
4862 NULL, /* next */
4863 0, /* static_pass_number */
4864 TV_IPA_SRA, /* tv_id */
4865 0, /* properties_required */
4866 0, /* properties_provided */
4867 0, /* properties_destroyed */
4868 0, /* todo_flags_start */
4869 TODO_dump_symtab /* todo_flags_finish */
4870 }
4871 };