cgraphunit.c (decide_is_function_needed): Honor -fkeep-inline-functions.
[gcc.git] / gcc / cgraphunit.c
1 /* Callgraph based interprocedural optimizations.
2 Copyright (C) 2003, 2004, 2005, 2006 Free Software Foundation, Inc.
3 Contributed by Jan Hubicka
4
5 This file is part of GCC.
6
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 2, or (at your option) any later
10 version.
11
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
15 for more details.
16
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING. If not, write to the Free
19 Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
20 02110-1301, USA. */
21
22 /* This module implements main driver of compilation process as well as
23 few basic interprocedural optimizers.
24
25 The main scope of this file is to act as an interface in between
26 tree based frontends and the backend (and middle end)
27
28 The front-end is supposed to use following functionality:
29
30 - cgraph_finalize_function
31
32 This function is called once front-end has parsed whole body of function
33 and it is certain that the function body nor the declaration will change.
34
35 (There is one exception needed for implementing GCC extern inline
36 function.)
37
38 - varpool_finalize_variable
39
40 This function has same behavior as the above but is used for static
41 variables.
42
43 - cgraph_finalize_compilation_unit
44
45 This function is called once (source level) compilation unit is finalized
46 and it will no longer change.
47
48 In the unit-at-a-time the call-graph construction and local function
49 analysis takes place here. Bodies of unreachable functions are released
50 to conserve memory usage.
51
52 The function can be called multiple times when multiple source level
53 compilation units are combined (such as in C frontend)
54
55 - cgraph_optimize
56
57 In this unit-at-a-time compilation the intra procedural analysis takes
58 place here. In particular the static functions whose address is never
59 taken are marked as local. Backend can then use this information to
60 modify calling conventions, do better inlining or similar optimizations.
61
62 - cgraph_mark_needed_node
63 - varpool_mark_needed_node
64
65 When function or variable is referenced by some hidden way the call-graph
66 data structure must be updated accordingly by this function.
67 There should be little need to call this function and all the references
68 should be made explicit to cgraph code. At present these functions are
69 used by C++ frontend to explicitly mark the keyed methods.
70
71 - analyze_expr callback
72
73 This function is responsible for lowering tree nodes not understood by
74 generic code into understandable ones or alternatively marking
75 callgraph and varpool nodes referenced by the as needed.
76
77 ??? On the tree-ssa genericizing should take place here and we will avoid
78 need for these hooks (replacing them by genericizing hook)
79
80 - expand_function callback
81
82 This function is used to expand function and pass it into RTL back-end.
83 Front-end should not make any assumptions about when this function can be
84 called. In particular cgraph_assemble_pending_functions,
85 varpool_assemble_pending_variables, cgraph_finalize_function,
86 varpool_finalize_function, cgraph_optimize can cause arbitrarily
87 previously finalized functions to be expanded.
88
89 We implement two compilation modes.
90
91 - unit-at-a-time: In this mode analyzing of all functions is deferred
92 to cgraph_finalize_compilation_unit and expansion into cgraph_optimize.
93
94 In cgraph_finalize_compilation_unit the reachable functions are
95 analyzed. During analysis the call-graph edges from reachable
96 functions are constructed and their destinations are marked as
97 reachable. References to functions and variables are discovered too
98 and variables found to be needed output to the assembly file. Via
99 mark_referenced call in assemble_variable functions referenced by
100 static variables are noticed too.
101
102 The intra-procedural information is produced and its existence
103 indicated by global_info_ready. Once this flag is set it is impossible
104 to change function from !reachable to reachable and thus
105 assemble_variable no longer call mark_referenced.
106
107 Finally the call-graph is topologically sorted and all reachable functions
108 that has not been completely inlined or are not external are output.
109
110 ??? It is possible that reference to function or variable is optimized
111 out. We can not deal with this nicely because topological order is not
112 suitable for it. For tree-ssa we may consider another pass doing
113 optimization and re-discovering reachable functions.
114
115 ??? Reorganize code so variables are output very last and only if they
116 really has been referenced by produced code, so we catch more cases
117 where reference has been optimized out.
118
119 - non-unit-at-a-time
120
121 All functions are variables are output as early as possible to conserve
122 memory consumption. This may or may not result in less memory used but
123 it is still needed for some legacy code that rely on particular ordering
124 of things output from the compiler.
125
126 Varpool data structures are not used and variables are output directly.
127
128 Functions are output early using call of
129 cgraph_assemble_pending_function from cgraph_finalize_function. The
130 decision on whether function is needed is made more conservative so
131 uninlininable static functions are needed too. During the call-graph
132 construction the edge destinations are not marked as reachable and it
133 is completely relied upn assemble_variable to mark them. */
134
135
136 #include "config.h"
137 #include "system.h"
138 #include "coretypes.h"
139 #include "tm.h"
140 #include "tree.h"
141 #include "rtl.h"
142 #include "tree-flow.h"
143 #include "tree-inline.h"
144 #include "langhooks.h"
145 #include "pointer-set.h"
146 #include "toplev.h"
147 #include "flags.h"
148 #include "ggc.h"
149 #include "debug.h"
150 #include "target.h"
151 #include "cgraph.h"
152 #include "diagnostic.h"
153 #include "timevar.h"
154 #include "params.h"
155 #include "fibheap.h"
156 #include "c-common.h"
157 #include "intl.h"
158 #include "function.h"
159 #include "ipa-prop.h"
160 #include "tree-gimple.h"
161 #include "tree-pass.h"
162 #include "output.h"
163
164 static void cgraph_expand_all_functions (void);
165 static void cgraph_mark_functions_to_output (void);
166 static void cgraph_expand_function (struct cgraph_node *);
167 static void cgraph_output_pending_asms (void);
168
169 static FILE *cgraph_dump_file;
170
171 /* Determine if function DECL is needed. That is, visible to something
172 either outside this translation unit, something magic in the system
173 configury, or (if not doing unit-at-a-time) to something we havn't
174 seen yet. */
175
176 static bool
177 decide_is_function_needed (struct cgraph_node *node, tree decl)
178 {
179 tree origin;
180 if (MAIN_NAME_P (DECL_NAME (decl))
181 && TREE_PUBLIC (decl))
182 {
183 node->local.externally_visible = true;
184 return true;
185 }
186
187 /* If the user told us it is used, then it must be so. */
188 if (node->local.externally_visible)
189 return true;
190
191 if (!flag_unit_at_a_time && lookup_attribute ("used", DECL_ATTRIBUTES (decl)))
192 return true;
193
194 /* ??? If the assembler name is set by hand, it is possible to assemble
195 the name later after finalizing the function and the fact is noticed
196 in assemble_name then. This is arguably a bug. */
197 if (DECL_ASSEMBLER_NAME_SET_P (decl)
198 && TREE_SYMBOL_REFERENCED (DECL_ASSEMBLER_NAME (decl)))
199 return true;
200
201 /* With -fkeep-inline-functions we are keeping all inline functions except
202 for extern inline ones. */
203 if (flag_keep_inline_functions
204 && DECL_DECLARED_INLINE_P (decl)
205 && !DECL_EXTERNAL (decl))
206 return true;
207
208 /* If we decided it was needed before, but at the time we didn't have
209 the body of the function available, then it's still needed. We have
210 to go back and re-check its dependencies now. */
211 if (node->needed)
212 return true;
213
214 /* Externally visible functions must be output. The exception is
215 COMDAT functions that must be output only when they are needed.
216
217 When not optimizing, also output the static functions. (see
218 PR24561), but don't do so for always_inline functions, functions
219 declared inline and nested functions. These was optimized out
220 in the original implementation and it is unclear whether we want
221 to change the behavior here. */
222 if (((TREE_PUBLIC (decl)
223 || (!optimize && !node->local.disregard_inline_limits
224 && !DECL_DECLARED_INLINE_P (decl)
225 && !node->origin))
226 && !flag_whole_program)
227 && !DECL_COMDAT (decl) && !DECL_EXTERNAL (decl))
228 return true;
229
230 /* Constructors and destructors are reachable from the runtime by
231 some mechanism. */
232 if (DECL_STATIC_CONSTRUCTOR (decl) || DECL_STATIC_DESTRUCTOR (decl))
233 return true;
234
235 if (flag_unit_at_a_time)
236 return false;
237
238 /* If not doing unit at a time, then we'll only defer this function
239 if its marked for inlining. Otherwise we want to emit it now. */
240
241 /* "extern inline" functions are never output locally. */
242 if (DECL_EXTERNAL (decl))
243 return false;
244 /* Nested functions of extern inline function shall not be emit unless
245 we inlined the origin. */
246 for (origin = decl_function_context (decl); origin;
247 origin = decl_function_context (origin))
248 if (DECL_EXTERNAL (origin))
249 return false;
250 /* We want to emit COMDAT functions only when absolutely necessary. */
251 if (DECL_COMDAT (decl))
252 return false;
253 if (!DECL_INLINE (decl)
254 || (!node->local.disregard_inline_limits
255 /* When declared inline, defer even the uninlinable functions.
256 This allows them to be eliminated when unused. */
257 && !DECL_DECLARED_INLINE_P (decl)
258 && (!node->local.inlinable || !cgraph_default_inline_p (node, NULL))))
259 return true;
260
261 return false;
262 }
263
264 /* Process CGRAPH_NEW_FUNCTIONS and perform actions necessary to add these
265 functions into callgraph in a way so they look like ordinary reachable
266 functions inserted into callgraph already at construction time. */
267
268 bool
269 cgraph_process_new_functions (void)
270 {
271 bool output = false;
272 tree fndecl;
273 struct cgraph_node *node;
274
275 /* Note that this queue may grow as its being processed, as the new
276 functions may generate new ones. */
277 while (cgraph_new_nodes)
278 {
279 node = cgraph_new_nodes;
280 fndecl = node->decl;
281 cgraph_new_nodes = cgraph_new_nodes->next_needed;
282 switch (cgraph_state)
283 {
284 case CGRAPH_STATE_CONSTRUCTION:
285 /* At construction time we just need to finalize function and move
286 it into reachable functions list. */
287
288 node->next_needed = NULL;
289 node->needed = node->reachable = false;
290 cgraph_finalize_function (fndecl, false);
291 cgraph_mark_reachable_node (node);
292 output = true;
293 break;
294
295 case CGRAPH_STATE_IPA:
296 case CGRAPH_STATE_IPA_SSA:
297 /* When IPA optimization already started, do all essential
298 transformations that has been already performed on the whole
299 cgraph but not on this function. */
300
301 tree_register_cfg_hooks ();
302 if (!node->analyzed)
303 cgraph_analyze_function (node);
304 push_cfun (DECL_STRUCT_FUNCTION (fndecl));
305 current_function_decl = fndecl;
306 node->local.inlinable = tree_inlinable_function_p (fndecl);
307 node->local.self_insns = estimate_num_insns (fndecl,
308 &eni_inlining_weights);
309 node->local.disregard_inline_limits
310 = lang_hooks.tree_inlining.disregard_inline_limits (fndecl);
311 /* Inlining characteristics are maintained by the
312 cgraph_mark_inline. */
313 node->global.insns = node->local.self_insns;
314 if (flag_really_no_inline && !node->local.disregard_inline_limits)
315 node->local.inlinable = 0;
316 if ((cgraph_state == CGRAPH_STATE_IPA_SSA
317 && !gimple_in_ssa_p (DECL_STRUCT_FUNCTION (fndecl)))
318 /* When not optimizing, be sure we run early local passes anyway
319 to expand OMP. */
320 || !optimize)
321 execute_pass_list (pass_early_local_passes.sub);
322 free_dominance_info (CDI_POST_DOMINATORS);
323 free_dominance_info (CDI_DOMINATORS);
324 pop_cfun ();
325 current_function_decl = NULL;
326 break;
327
328 case CGRAPH_STATE_EXPANSION:
329 /* Functions created during expansion shall be compiled
330 directly. */
331 node->output = 0;
332 cgraph_expand_function (node);
333 break;
334
335 default:
336 gcc_unreachable ();
337 break;
338 }
339 }
340 return output;
341 }
342
343 /* When not doing unit-at-a-time, output all functions enqueued.
344 Return true when such a functions were found. */
345
346 static bool
347 cgraph_assemble_pending_functions (void)
348 {
349 bool output = false;
350
351 if (flag_unit_at_a_time)
352 return false;
353
354 cgraph_output_pending_asms ();
355
356 while (cgraph_nodes_queue)
357 {
358 struct cgraph_node *n = cgraph_nodes_queue;
359
360 cgraph_nodes_queue = cgraph_nodes_queue->next_needed;
361 n->next_needed = NULL;
362 if (!n->global.inlined_to
363 && !n->alias
364 && !DECL_EXTERNAL (n->decl))
365 {
366 cgraph_expand_function (n);
367 output = true;
368 }
369 output |= cgraph_process_new_functions ();
370 }
371
372 return output;
373 }
374
375
376 /* As an GCC extension we allow redefinition of the function. The
377 semantics when both copies of bodies differ is not well defined.
378 We replace the old body with new body so in unit at a time mode
379 we always use new body, while in normal mode we may end up with
380 old body inlined into some functions and new body expanded and
381 inlined in others.
382
383 ??? It may make more sense to use one body for inlining and other
384 body for expanding the function but this is difficult to do. */
385
386 static void
387 cgraph_reset_node (struct cgraph_node *node)
388 {
389 /* If node->output is set, then this is a unit-at-a-time compilation
390 and we have already begun whole-unit analysis. This is *not*
391 testing for whether we've already emitted the function. That
392 case can be sort-of legitimately seen with real function
393 redefinition errors. I would argue that the front end should
394 never present us with such a case, but don't enforce that for now. */
395 gcc_assert (!node->output);
396
397 /* Reset our data structures so we can analyze the function again. */
398 memset (&node->local, 0, sizeof (node->local));
399 memset (&node->global, 0, sizeof (node->global));
400 memset (&node->rtl, 0, sizeof (node->rtl));
401 node->analyzed = false;
402 node->local.redefined_extern_inline = true;
403 node->local.finalized = false;
404
405 if (!flag_unit_at_a_time)
406 {
407 struct cgraph_node *n, *next;
408
409 for (n = cgraph_nodes; n; n = next)
410 {
411 next = n->next;
412 if (n->global.inlined_to == node)
413 cgraph_remove_node (n);
414 }
415 }
416
417 cgraph_node_remove_callees (node);
418
419 /* We may need to re-queue the node for assembling in case
420 we already proceeded it and ignored as not needed. */
421 if (node->reachable && !flag_unit_at_a_time)
422 {
423 struct cgraph_node *n;
424
425 for (n = cgraph_nodes_queue; n; n = n->next_needed)
426 if (n == node)
427 break;
428 if (!n)
429 node->reachable = 0;
430 }
431 }
432
433 static void
434 cgraph_lower_function (struct cgraph_node *node)
435 {
436 if (node->lowered)
437 return;
438 tree_lowering_passes (node->decl);
439 node->lowered = true;
440 }
441
442 /* DECL has been parsed. Take it, queue it, compile it at the whim of the
443 logic in effect. If NESTED is true, then our caller cannot stand to have
444 the garbage collector run at the moment. We would need to either create
445 a new GC context, or just not compile right now. */
446
447 void
448 cgraph_finalize_function (tree decl, bool nested)
449 {
450 struct cgraph_node *node = cgraph_node (decl);
451
452 if (node->local.finalized)
453 cgraph_reset_node (node);
454
455 node->pid = cgraph_max_pid ++;
456 notice_global_symbol (decl);
457 node->decl = decl;
458 node->local.finalized = true;
459 node->lowered = DECL_STRUCT_FUNCTION (decl)->cfg != NULL;
460 if (node->nested)
461 lower_nested_functions (decl);
462 gcc_assert (!node->nested);
463
464 /* If not unit at a time, then we need to create the call graph
465 now, so that called functions can be queued and emitted now. */
466 if (!flag_unit_at_a_time)
467 cgraph_analyze_function (node);
468
469 if (decide_is_function_needed (node, decl))
470 cgraph_mark_needed_node (node);
471
472 /* Since we reclaim unreachable nodes at the end of every language
473 level unit, we need to be conservative about possible entry points
474 there. */
475 if ((TREE_PUBLIC (decl) && !DECL_COMDAT (decl) && !DECL_EXTERNAL (decl)))
476 cgraph_mark_reachable_node (node);
477
478 /* If not unit at a time, go ahead and emit everything we've found
479 to be reachable at this time. */
480 if (!nested)
481 {
482 if (!cgraph_assemble_pending_functions ())
483 ggc_collect ();
484 }
485
486 /* If we've not yet emitted decl, tell the debug info about it. */
487 if (!TREE_ASM_WRITTEN (decl))
488 (*debug_hooks->deferred_inline_function) (decl);
489
490 /* Possibly warn about unused parameters. */
491 if (warn_unused_parameter)
492 do_warn_unused_parameter (decl);
493 }
494
495 /* Verify cgraph nodes of given cgraph node. */
496 void
497 verify_cgraph_node (struct cgraph_node *node)
498 {
499 struct cgraph_edge *e;
500 struct cgraph_node *main_clone;
501 struct function *this_cfun = DECL_STRUCT_FUNCTION (node->decl);
502 basic_block this_block;
503 block_stmt_iterator bsi;
504 bool error_found = false;
505
506 if (errorcount || sorrycount)
507 return;
508
509 timevar_push (TV_CGRAPH_VERIFY);
510 for (e = node->callees; e; e = e->next_callee)
511 if (e->aux)
512 {
513 error ("aux field set for edge %s->%s",
514 cgraph_node_name (e->caller), cgraph_node_name (e->callee));
515 error_found = true;
516 }
517 if (node->count < 0)
518 {
519 error ("Execution count is negative");
520 error_found = true;
521 }
522 for (e = node->callers; e; e = e->next_caller)
523 {
524 if (e->count < 0)
525 {
526 error ("caller edge count is negative");
527 error_found = true;
528 }
529 if (e->frequency < 0)
530 {
531 error ("caller edge frequency is negative");
532 error_found = true;
533 }
534 if (e->frequency > CGRAPH_FREQ_MAX)
535 {
536 error ("caller edge frequency is too large");
537 error_found = true;
538 }
539 if (!e->inline_failed)
540 {
541 if (node->global.inlined_to
542 != (e->caller->global.inlined_to
543 ? e->caller->global.inlined_to : e->caller))
544 {
545 error ("inlined_to pointer is wrong");
546 error_found = true;
547 }
548 if (node->callers->next_caller)
549 {
550 error ("multiple inline callers");
551 error_found = true;
552 }
553 }
554 else
555 if (node->global.inlined_to)
556 {
557 error ("inlined_to pointer set for noninline callers");
558 error_found = true;
559 }
560 }
561 if (!node->callers && node->global.inlined_to)
562 {
563 error ("inlined_to pointer is set but no predecessors found");
564 error_found = true;
565 }
566 if (node->global.inlined_to == node)
567 {
568 error ("inlined_to pointer refers to itself");
569 error_found = true;
570 }
571
572 for (main_clone = cgraph_node (node->decl); main_clone;
573 main_clone = main_clone->next_clone)
574 if (main_clone == node)
575 break;
576 if (!cgraph_node (node->decl))
577 {
578 error ("node not found in cgraph_hash");
579 error_found = true;
580 }
581
582 if (node->analyzed
583 && DECL_SAVED_TREE (node->decl) && !TREE_ASM_WRITTEN (node->decl)
584 && (!DECL_EXTERNAL (node->decl) || node->global.inlined_to))
585 {
586 if (this_cfun->cfg)
587 {
588 /* The nodes we're interested in are never shared, so walk
589 the tree ignoring duplicates. */
590 struct pointer_set_t *visited_nodes = pointer_set_create ();
591 /* Reach the trees by walking over the CFG, and note the
592 enclosing basic-blocks in the call edges. */
593 FOR_EACH_BB_FN (this_block, this_cfun)
594 for (bsi = bsi_start (this_block); !bsi_end_p (bsi); bsi_next (&bsi))
595 {
596 tree stmt = bsi_stmt (bsi);
597 tree call = get_call_expr_in (stmt);
598 tree decl;
599 if (call && (decl = get_callee_fndecl (call)))
600 {
601 struct cgraph_edge *e = cgraph_edge (node, stmt);
602 if (e)
603 {
604 if (e->aux)
605 {
606 error ("shared call_stmt:");
607 debug_generic_stmt (stmt);
608 error_found = true;
609 }
610 if (e->callee->decl != cgraph_node (decl)->decl
611 && e->inline_failed)
612 {
613 error ("edge points to wrong declaration:");
614 debug_tree (e->callee->decl);
615 fprintf (stderr," Instead of:");
616 debug_tree (decl);
617 }
618 e->aux = (void *)1;
619 }
620 else
621 {
622 error ("missing callgraph edge for call stmt:");
623 debug_generic_stmt (stmt);
624 error_found = true;
625 }
626 }
627 }
628 pointer_set_destroy (visited_nodes);
629 }
630 else
631 /* No CFG available?! */
632 gcc_unreachable ();
633
634 for (e = node->callees; e; e = e->next_callee)
635 {
636 if (!e->aux)
637 {
638 error ("edge %s->%s has no corresponding call_stmt",
639 cgraph_node_name (e->caller),
640 cgraph_node_name (e->callee));
641 debug_generic_stmt (e->call_stmt);
642 error_found = true;
643 }
644 e->aux = 0;
645 }
646 }
647 if (error_found)
648 {
649 dump_cgraph_node (stderr, node);
650 internal_error ("verify_cgraph_node failed");
651 }
652 timevar_pop (TV_CGRAPH_VERIFY);
653 }
654
655 /* Verify whole cgraph structure. */
656 void
657 verify_cgraph (void)
658 {
659 struct cgraph_node *node;
660
661 if (sorrycount || errorcount)
662 return;
663
664 for (node = cgraph_nodes; node; node = node->next)
665 verify_cgraph_node (node);
666 }
667
668 /* Output all asm statements we have stored up to be output. */
669
670 static void
671 cgraph_output_pending_asms (void)
672 {
673 struct cgraph_asm_node *can;
674
675 if (errorcount || sorrycount)
676 return;
677
678 for (can = cgraph_asm_nodes; can; can = can->next)
679 assemble_asm (can->asm_str);
680 cgraph_asm_nodes = NULL;
681 }
682
683 /* Analyze the function scheduled to be output. */
684 void
685 cgraph_analyze_function (struct cgraph_node *node)
686 {
687 tree decl = node->decl;
688
689 current_function_decl = decl;
690 push_cfun (DECL_STRUCT_FUNCTION (decl));
691 cgraph_lower_function (node);
692
693 node->local.estimated_self_stack_size = estimated_stack_frame_size ();
694 node->global.estimated_stack_size = node->local.estimated_self_stack_size;
695 node->global.stack_frame_offset = 0;
696 node->local.inlinable = tree_inlinable_function_p (decl);
697 if (!flag_unit_at_a_time)
698 node->local.self_insns = estimate_num_insns (decl, &eni_inlining_weights);
699 if (node->local.inlinable)
700 node->local.disregard_inline_limits
701 = lang_hooks.tree_inlining.disregard_inline_limits (decl);
702 if (flag_really_no_inline && !node->local.disregard_inline_limits)
703 node->local.inlinable = 0;
704 /* Inlining characteristics are maintained by the cgraph_mark_inline. */
705 node->global.insns = node->local.self_insns;
706 if (!flag_unit_at_a_time)
707 {
708 bitmap_obstack_initialize (NULL);
709 tree_register_cfg_hooks ();
710 execute_pass_list (pass_early_local_passes.sub);
711 free_dominance_info (CDI_POST_DOMINATORS);
712 free_dominance_info (CDI_DOMINATORS);
713 bitmap_obstack_release (NULL);
714 }
715
716 node->analyzed = true;
717 pop_cfun ();
718 current_function_decl = NULL;
719 }
720
721 /* Look for externally_visible and used attributes and mark cgraph nodes
722 accordingly.
723
724 We cannot mark the nodes at the point the attributes are processed (in
725 handle_*_attribute) because the copy of the declarations available at that
726 point may not be canonical. For example, in:
727
728 void f();
729 void f() __attribute__((used));
730
731 the declaration we see in handle_used_attribute will be the second
732 declaration -- but the front end will subsequently merge that declaration
733 with the original declaration and discard the second declaration.
734
735 Furthermore, we can't mark these nodes in cgraph_finalize_function because:
736
737 void f() {}
738 void f() __attribute__((externally_visible));
739
740 is valid.
741
742 So, we walk the nodes at the end of the translation unit, applying the
743 attributes at that point. */
744
745 static void
746 process_function_and_variable_attributes (struct cgraph_node *first,
747 struct varpool_node *first_var)
748 {
749 struct cgraph_node *node;
750 struct varpool_node *vnode;
751
752 for (node = cgraph_nodes; node != first; node = node->next)
753 {
754 tree decl = node->decl;
755 if (lookup_attribute ("used", DECL_ATTRIBUTES (decl)))
756 {
757 mark_decl_referenced (decl);
758 if (node->local.finalized)
759 cgraph_mark_needed_node (node);
760 }
761 if (lookup_attribute ("externally_visible", DECL_ATTRIBUTES (decl)))
762 {
763 if (! TREE_PUBLIC (node->decl))
764 warning (OPT_Wattributes,
765 "%J%<externally_visible%> attribute have effect only on public objects",
766 node->decl);
767 else
768 {
769 if (node->local.finalized)
770 cgraph_mark_needed_node (node);
771 node->local.externally_visible = true;
772 }
773 }
774 }
775 for (vnode = varpool_nodes; vnode != first_var; vnode = vnode->next)
776 {
777 tree decl = vnode->decl;
778 if (lookup_attribute ("used", DECL_ATTRIBUTES (decl)))
779 {
780 mark_decl_referenced (decl);
781 if (vnode->finalized)
782 varpool_mark_needed_node (vnode);
783 }
784 if (lookup_attribute ("externally_visible", DECL_ATTRIBUTES (decl)))
785 {
786 if (! TREE_PUBLIC (vnode->decl))
787 warning (OPT_Wattributes,
788 "%J%<externally_visible%> attribute have effect only on public objects",
789 vnode->decl);
790 else
791 {
792 if (vnode->finalized)
793 varpool_mark_needed_node (vnode);
794 vnode->externally_visible = true;
795 }
796 }
797 }
798 }
799
800 /* Process CGRAPH_NODES_NEEDED queue, analyze each function (and transitively
801 each reachable functions) and build cgraph.
802 The function can be called multiple times after inserting new nodes
803 into beginning of queue. Just the new part of queue is re-scanned then. */
804
805 static void
806 cgraph_analyze_functions (void)
807 {
808 /* Keep track of already processed nodes when called multiple times for
809 intermodule optimization. */
810 static struct cgraph_node *first_analyzed;
811 struct cgraph_node *first_processed = first_analyzed;
812 static struct varpool_node *first_analyzed_var;
813 struct cgraph_node *node, *next;
814
815 process_function_and_variable_attributes (first_processed,
816 first_analyzed_var);
817 first_processed = cgraph_nodes;
818 first_analyzed_var = varpool_nodes;
819 varpool_analyze_pending_decls ();
820 if (cgraph_dump_file)
821 {
822 fprintf (cgraph_dump_file, "Initial entry points:");
823 for (node = cgraph_nodes; node != first_analyzed; node = node->next)
824 if (node->needed && DECL_SAVED_TREE (node->decl))
825 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node));
826 fprintf (cgraph_dump_file, "\n");
827 }
828 cgraph_process_new_functions ();
829
830 /* Propagate reachability flag and lower representation of all reachable
831 functions. In the future, lowering will introduce new functions and
832 new entry points on the way (by template instantiation and virtual
833 method table generation for instance). */
834 while (cgraph_nodes_queue)
835 {
836 struct cgraph_edge *edge;
837 tree decl = cgraph_nodes_queue->decl;
838
839 node = cgraph_nodes_queue;
840 cgraph_nodes_queue = cgraph_nodes_queue->next_needed;
841 node->next_needed = NULL;
842
843 /* ??? It is possible to create extern inline function and later using
844 weak alias attribute to kill its body. See
845 gcc.c-torture/compile/20011119-1.c */
846 if (!DECL_SAVED_TREE (decl))
847 {
848 cgraph_reset_node (node);
849 continue;
850 }
851
852 gcc_assert (!node->analyzed && node->reachable);
853 gcc_assert (DECL_SAVED_TREE (decl));
854
855 cgraph_analyze_function (node);
856
857 for (edge = node->callees; edge; edge = edge->next_callee)
858 if (!edge->callee->reachable)
859 cgraph_mark_reachable_node (edge->callee);
860
861 /* We finalize local static variables during constructing callgraph
862 edges. Process their attributes too. */
863 process_function_and_variable_attributes (first_processed,
864 first_analyzed_var);
865 first_processed = cgraph_nodes;
866 first_analyzed_var = varpool_nodes;
867 varpool_analyze_pending_decls ();
868 cgraph_process_new_functions ();
869 }
870
871 /* Collect entry points to the unit. */
872 if (cgraph_dump_file)
873 {
874 fprintf (cgraph_dump_file, "Unit entry points:");
875 for (node = cgraph_nodes; node != first_analyzed; node = node->next)
876 if (node->needed && DECL_SAVED_TREE (node->decl))
877 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node));
878 fprintf (cgraph_dump_file, "\n\nInitial ");
879 dump_cgraph (cgraph_dump_file);
880 }
881
882 if (cgraph_dump_file)
883 fprintf (cgraph_dump_file, "\nReclaiming functions:");
884
885 for (node = cgraph_nodes; node != first_analyzed; node = next)
886 {
887 tree decl = node->decl;
888 next = node->next;
889
890 if (node->local.finalized && !DECL_SAVED_TREE (decl))
891 cgraph_reset_node (node);
892
893 if (!node->reachable && DECL_SAVED_TREE (decl))
894 {
895 if (cgraph_dump_file)
896 fprintf (cgraph_dump_file, " %s", cgraph_node_name (node));
897 cgraph_remove_node (node);
898 continue;
899 }
900 else
901 node->next_needed = NULL;
902 gcc_assert (!node->local.finalized || DECL_SAVED_TREE (decl));
903 gcc_assert (node->analyzed == node->local.finalized);
904 }
905 if (cgraph_dump_file)
906 {
907 fprintf (cgraph_dump_file, "\n\nReclaimed ");
908 dump_cgraph (cgraph_dump_file);
909 }
910 first_analyzed = cgraph_nodes;
911 ggc_collect ();
912 }
913
914 /* Analyze the whole compilation unit once it is parsed completely. */
915
916 void
917 cgraph_finalize_compilation_unit (void)
918 {
919 if (errorcount || sorrycount)
920 return;
921
922 finish_aliases_1 ();
923
924 if (!flag_unit_at_a_time)
925 {
926 cgraph_output_pending_asms ();
927 cgraph_assemble_pending_functions ();
928 varpool_output_debug_info ();
929 return;
930 }
931
932 if (!quiet_flag)
933 {
934 fprintf (stderr, "\nAnalyzing compilation unit\n");
935 fflush (stderr);
936 }
937
938 timevar_push (TV_CGRAPH);
939 cgraph_analyze_functions ();
940 timevar_pop (TV_CGRAPH);
941 }
942 /* Figure out what functions we want to assemble. */
943
944 static void
945 cgraph_mark_functions_to_output (void)
946 {
947 struct cgraph_node *node;
948
949 for (node = cgraph_nodes; node; node = node->next)
950 {
951 tree decl = node->decl;
952 struct cgraph_edge *e;
953
954 gcc_assert (!node->output);
955
956 for (e = node->callers; e; e = e->next_caller)
957 if (e->inline_failed)
958 break;
959
960 /* We need to output all local functions that are used and not
961 always inlined, as well as those that are reachable from
962 outside the current compilation unit. */
963 if (DECL_SAVED_TREE (decl)
964 && !node->global.inlined_to
965 && (node->needed
966 || (e && node->reachable))
967 && !TREE_ASM_WRITTEN (decl)
968 && !DECL_EXTERNAL (decl))
969 node->output = 1;
970 else
971 {
972 /* We should've reclaimed all functions that are not needed. */
973 #ifdef ENABLE_CHECKING
974 if (!node->global.inlined_to && DECL_SAVED_TREE (decl)
975 && !DECL_EXTERNAL (decl))
976 {
977 dump_cgraph_node (stderr, node);
978 internal_error ("failed to reclaim unneeded function");
979 }
980 #endif
981 gcc_assert (node->global.inlined_to || !DECL_SAVED_TREE (decl)
982 || DECL_EXTERNAL (decl));
983
984 }
985
986 }
987 }
988
989 /* Expand function specified by NODE. */
990
991 static void
992 cgraph_expand_function (struct cgraph_node *node)
993 {
994 enum debug_info_type save_write_symbols = NO_DEBUG;
995 const struct gcc_debug_hooks *save_debug_hooks = NULL;
996 tree decl = node->decl;
997
998 /* We ought to not compile any inline clones. */
999 gcc_assert (!node->global.inlined_to);
1000
1001 if (flag_unit_at_a_time)
1002 announce_function (decl);
1003
1004 gcc_assert (node->lowered);
1005
1006 if (DECL_IGNORED_P (decl))
1007 {
1008 save_write_symbols = write_symbols;
1009 write_symbols = NO_DEBUG;
1010 save_debug_hooks = debug_hooks;
1011 debug_hooks = &do_nothing_debug_hooks;
1012 }
1013
1014 /* Generate RTL for the body of DECL. */
1015 lang_hooks.callgraph.expand_function (decl);
1016
1017 /* Make sure that BE didn't give up on compiling. */
1018 /* ??? Can happen with nested function of extern inline. */
1019 gcc_assert (TREE_ASM_WRITTEN (node->decl));
1020
1021 if (DECL_IGNORED_P (decl))
1022 {
1023 write_symbols = save_write_symbols;
1024 debug_hooks = save_debug_hooks;
1025 }
1026
1027 current_function_decl = NULL;
1028 if (!cgraph_preserve_function_body_p (node->decl))
1029 {
1030 cgraph_release_function_body (node);
1031 /* Eliminate all call edges. This is important so the call_expr no longer
1032 points to the dead function body. */
1033 cgraph_node_remove_callees (node);
1034 }
1035
1036 cgraph_function_flags_ready = true;
1037 }
1038
1039 /* Return true when CALLER_DECL should be inlined into CALLEE_DECL. */
1040
1041 bool
1042 cgraph_inline_p (struct cgraph_edge *e, const char **reason)
1043 {
1044 *reason = e->inline_failed;
1045 return !e->inline_failed;
1046 }
1047
1048
1049
1050 /* Expand all functions that must be output.
1051
1052 Attempt to topologically sort the nodes so function is output when
1053 all called functions are already assembled to allow data to be
1054 propagated across the callgraph. Use a stack to get smaller distance
1055 between a function and its callees (later we may choose to use a more
1056 sophisticated algorithm for function reordering; we will likely want
1057 to use subsections to make the output functions appear in top-down
1058 order). */
1059
1060 static void
1061 cgraph_expand_all_functions (void)
1062 {
1063 struct cgraph_node *node;
1064 struct cgraph_node **order = XCNEWVEC (struct cgraph_node *, cgraph_n_nodes);
1065 int order_pos = 0, new_order_pos = 0;
1066 int i;
1067
1068 order_pos = cgraph_postorder (order);
1069 gcc_assert (order_pos == cgraph_n_nodes);
1070
1071 /* Garbage collector may remove inline clones we eliminate during
1072 optimization. So we must be sure to not reference them. */
1073 for (i = 0; i < order_pos; i++)
1074 if (order[i]->output)
1075 order[new_order_pos++] = order[i];
1076
1077 for (i = new_order_pos - 1; i >= 0; i--)
1078 {
1079 node = order[i];
1080 if (node->output)
1081 {
1082 gcc_assert (node->reachable);
1083 node->output = 0;
1084 cgraph_expand_function (node);
1085 }
1086 }
1087 cgraph_process_new_functions ();
1088
1089 free (order);
1090
1091 }
1092
1093 /* This is used to sort the node types by the cgraph order number. */
1094
1095 struct cgraph_order_sort
1096 {
1097 enum { ORDER_UNDEFINED = 0, ORDER_FUNCTION, ORDER_VAR, ORDER_ASM } kind;
1098 union
1099 {
1100 struct cgraph_node *f;
1101 struct varpool_node *v;
1102 struct cgraph_asm_node *a;
1103 } u;
1104 };
1105
1106 /* Output all functions, variables, and asm statements in the order
1107 according to their order fields, which is the order in which they
1108 appeared in the file. This implements -fno-toplevel-reorder. In
1109 this mode we may output functions and variables which don't really
1110 need to be output. */
1111
1112 static void
1113 cgraph_output_in_order (void)
1114 {
1115 int max;
1116 size_t size;
1117 struct cgraph_order_sort *nodes;
1118 int i;
1119 struct cgraph_node *pf;
1120 struct varpool_node *pv;
1121 struct cgraph_asm_node *pa;
1122
1123 max = cgraph_order;
1124 size = max * sizeof (struct cgraph_order_sort);
1125 nodes = (struct cgraph_order_sort *) alloca (size);
1126 memset (nodes, 0, size);
1127
1128 varpool_analyze_pending_decls ();
1129
1130 for (pf = cgraph_nodes; pf; pf = pf->next)
1131 {
1132 if (pf->output)
1133 {
1134 i = pf->order;
1135 gcc_assert (nodes[i].kind == ORDER_UNDEFINED);
1136 nodes[i].kind = ORDER_FUNCTION;
1137 nodes[i].u.f = pf;
1138 }
1139 }
1140
1141 for (pv = varpool_nodes_queue; pv; pv = pv->next_needed)
1142 {
1143 i = pv->order;
1144 gcc_assert (nodes[i].kind == ORDER_UNDEFINED);
1145 nodes[i].kind = ORDER_VAR;
1146 nodes[i].u.v = pv;
1147 }
1148
1149 for (pa = cgraph_asm_nodes; pa; pa = pa->next)
1150 {
1151 i = pa->order;
1152 gcc_assert (nodes[i].kind == ORDER_UNDEFINED);
1153 nodes[i].kind = ORDER_ASM;
1154 nodes[i].u.a = pa;
1155 }
1156
1157 for (i = 0; i < max; ++i)
1158 {
1159 switch (nodes[i].kind)
1160 {
1161 case ORDER_FUNCTION:
1162 nodes[i].u.f->output = 0;
1163 cgraph_expand_function (nodes[i].u.f);
1164 break;
1165
1166 case ORDER_VAR:
1167 varpool_assemble_decl (nodes[i].u.v);
1168 break;
1169
1170 case ORDER_ASM:
1171 assemble_asm (nodes[i].u.a->asm_str);
1172 break;
1173
1174 case ORDER_UNDEFINED:
1175 break;
1176
1177 default:
1178 gcc_unreachable ();
1179 }
1180 }
1181
1182 cgraph_asm_nodes = NULL;
1183 }
1184
1185 /* Return true when function body of DECL still needs to be kept around
1186 for later re-use. */
1187 bool
1188 cgraph_preserve_function_body_p (tree decl)
1189 {
1190 struct cgraph_node *node;
1191 if (!cgraph_global_info_ready)
1192 return (flag_really_no_inline
1193 ? lang_hooks.tree_inlining.disregard_inline_limits (decl)
1194 : DECL_INLINE (decl));
1195 /* Look if there is any clone around. */
1196 for (node = cgraph_node (decl); node; node = node->next_clone)
1197 if (node->global.inlined_to)
1198 return true;
1199 return false;
1200 }
1201
1202 static void
1203 ipa_passes (void)
1204 {
1205 cfun = NULL;
1206 current_function_decl = NULL;
1207 tree_register_cfg_hooks ();
1208 bitmap_obstack_initialize (NULL);
1209 execute_ipa_pass_list (all_ipa_passes);
1210 bitmap_obstack_release (NULL);
1211 }
1212
1213 /* Perform simple optimizations based on callgraph. */
1214
1215 void
1216 cgraph_optimize (void)
1217 {
1218 if (errorcount || sorrycount)
1219 return;
1220
1221 #ifdef ENABLE_CHECKING
1222 verify_cgraph ();
1223 #endif
1224 if (!flag_unit_at_a_time)
1225 {
1226 cgraph_assemble_pending_functions ();
1227 cgraph_process_new_functions ();
1228 cgraph_state = CGRAPH_STATE_FINISHED;
1229 cgraph_output_pending_asms ();
1230 varpool_assemble_pending_decls ();
1231 varpool_output_debug_info ();
1232 return;
1233 }
1234
1235 /* Frontend may output common variables after the unit has been finalized.
1236 It is safe to deal with them here as they are always zero initialized. */
1237 varpool_analyze_pending_decls ();
1238 cgraph_analyze_functions ();
1239
1240 timevar_push (TV_CGRAPHOPT);
1241 if (pre_ipa_mem_report)
1242 {
1243 fprintf (stderr, "Memory consumption before IPA\n");
1244 dump_memory_report (false);
1245 }
1246 if (!quiet_flag)
1247 fprintf (stderr, "Performing interprocedural optimizations\n");
1248 cgraph_state = CGRAPH_STATE_IPA;
1249
1250 /* Don't run the IPA passes if there was any error or sorry messages. */
1251 if (errorcount == 0 && sorrycount == 0)
1252 ipa_passes ();
1253
1254 /* This pass remove bodies of extern inline functions we never inlined.
1255 Do this later so other IPA passes see what is really going on. */
1256 cgraph_remove_unreachable_nodes (false, dump_file);
1257 cgraph_global_info_ready = true;
1258 if (cgraph_dump_file)
1259 {
1260 fprintf (cgraph_dump_file, "Optimized ");
1261 dump_cgraph (cgraph_dump_file);
1262 dump_varpool (cgraph_dump_file);
1263 }
1264 if (post_ipa_mem_report)
1265 {
1266 fprintf (stderr, "Memory consumption after IPA\n");
1267 dump_memory_report (false);
1268 }
1269 timevar_pop (TV_CGRAPHOPT);
1270
1271 /* Output everything. */
1272 if (!quiet_flag)
1273 fprintf (stderr, "Assembling functions:\n");
1274 #ifdef ENABLE_CHECKING
1275 verify_cgraph ();
1276 #endif
1277
1278 cgraph_mark_functions_to_output ();
1279
1280 cgraph_state = CGRAPH_STATE_EXPANSION;
1281 if (!flag_toplevel_reorder)
1282 cgraph_output_in_order ();
1283 else
1284 {
1285 cgraph_output_pending_asms ();
1286
1287 cgraph_expand_all_functions ();
1288 varpool_remove_unreferenced_decls ();
1289
1290 varpool_assemble_pending_decls ();
1291 varpool_output_debug_info ();
1292 }
1293 cgraph_process_new_functions ();
1294 cgraph_state = CGRAPH_STATE_FINISHED;
1295
1296 if (cgraph_dump_file)
1297 {
1298 fprintf (cgraph_dump_file, "\nFinal ");
1299 dump_cgraph (cgraph_dump_file);
1300 }
1301 #ifdef ENABLE_CHECKING
1302 verify_cgraph ();
1303 /* Double check that all inline clones are gone and that all
1304 function bodies have been released from memory. */
1305 if (flag_unit_at_a_time
1306 && !(sorrycount || errorcount))
1307 {
1308 struct cgraph_node *node;
1309 bool error_found = false;
1310
1311 for (node = cgraph_nodes; node; node = node->next)
1312 if (node->analyzed
1313 && (node->global.inlined_to
1314 || DECL_SAVED_TREE (node->decl)))
1315 {
1316 error_found = true;
1317 dump_cgraph_node (stderr, node);
1318 }
1319 if (error_found)
1320 internal_error ("nodes with no released memory found");
1321 }
1322 #endif
1323 }
1324 /* Generate and emit a static constructor or destructor. WHICH must be
1325 one of 'I' or 'D'. BODY should be a STATEMENT_LIST containing
1326 GENERIC statements. */
1327
1328 void
1329 cgraph_build_static_cdtor (char which, tree body, int priority)
1330 {
1331 static int counter = 0;
1332 char which_buf[16];
1333 tree decl, name, resdecl;
1334
1335 sprintf (which_buf, "%c_%d", which, counter++);
1336 name = get_file_function_name (which_buf);
1337
1338 decl = build_decl (FUNCTION_DECL, name,
1339 build_function_type (void_type_node, void_list_node));
1340 current_function_decl = decl;
1341
1342 resdecl = build_decl (RESULT_DECL, NULL_TREE, void_type_node);
1343 DECL_ARTIFICIAL (resdecl) = 1;
1344 DECL_IGNORED_P (resdecl) = 1;
1345 DECL_RESULT (decl) = resdecl;
1346
1347 allocate_struct_function (decl);
1348
1349 TREE_STATIC (decl) = 1;
1350 TREE_USED (decl) = 1;
1351 DECL_ARTIFICIAL (decl) = 1;
1352 DECL_IGNORED_P (decl) = 1;
1353 DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (decl) = 1;
1354 DECL_SAVED_TREE (decl) = body;
1355 TREE_PUBLIC (decl) = ! targetm.have_ctors_dtors;
1356 DECL_UNINLINABLE (decl) = 1;
1357
1358 DECL_INITIAL (decl) = make_node (BLOCK);
1359 TREE_USED (DECL_INITIAL (decl)) = 1;
1360
1361 DECL_SOURCE_LOCATION (decl) = input_location;
1362 cfun->function_end_locus = input_location;
1363
1364 switch (which)
1365 {
1366 case 'I':
1367 DECL_STATIC_CONSTRUCTOR (decl) = 1;
1368 break;
1369 case 'D':
1370 DECL_STATIC_DESTRUCTOR (decl) = 1;
1371 break;
1372 default:
1373 gcc_unreachable ();
1374 }
1375
1376 gimplify_function_tree (decl);
1377
1378 cgraph_add_new_function (decl, false);
1379 cgraph_mark_needed_node (cgraph_node (decl));
1380
1381 if (targetm.have_ctors_dtors)
1382 {
1383 void (*fn) (rtx, int);
1384
1385 if (which == 'I')
1386 fn = targetm.asm_out.constructor;
1387 else
1388 fn = targetm.asm_out.destructor;
1389 fn (XEXP (DECL_RTL (decl), 0), priority);
1390 }
1391 }
1392
1393 void
1394 init_cgraph (void)
1395 {
1396 cgraph_dump_file = dump_begin (TDI_cgraph, NULL);
1397 }
1398
1399 /* The edges representing the callers of the NEW_VERSION node were
1400 fixed by cgraph_function_versioning (), now the call_expr in their
1401 respective tree code should be updated to call the NEW_VERSION. */
1402
1403 static void
1404 update_call_expr (struct cgraph_node *new_version)
1405 {
1406 struct cgraph_edge *e;
1407
1408 gcc_assert (new_version);
1409 for (e = new_version->callers; e; e = e->next_caller)
1410 /* Update the call expr on the edges
1411 to call the new version. */
1412 TREE_OPERAND (CALL_EXPR_FN (get_call_expr_in (e->call_stmt)), 0) = new_version->decl;
1413 }
1414
1415
1416 /* Create a new cgraph node which is the new version of
1417 OLD_VERSION node. REDIRECT_CALLERS holds the callers
1418 edges which should be redirected to point to
1419 NEW_VERSION. ALL the callees edges of OLD_VERSION
1420 are cloned to the new version node. Return the new
1421 version node. */
1422
1423 static struct cgraph_node *
1424 cgraph_copy_node_for_versioning (struct cgraph_node *old_version,
1425 tree new_decl,
1426 VEC(cgraph_edge_p,heap) *redirect_callers)
1427 {
1428 struct cgraph_node *new_version;
1429 struct cgraph_edge *e, *new_e;
1430 struct cgraph_edge *next_callee;
1431 unsigned i;
1432
1433 gcc_assert (old_version);
1434
1435 new_version = cgraph_node (new_decl);
1436
1437 new_version->analyzed = true;
1438 new_version->local = old_version->local;
1439 new_version->global = old_version->global;
1440 new_version->rtl = new_version->rtl;
1441 new_version->reachable = true;
1442 new_version->count = old_version->count;
1443
1444 /* Clone the old node callees. Recursive calls are
1445 also cloned. */
1446 for (e = old_version->callees;e; e=e->next_callee)
1447 {
1448 new_e = cgraph_clone_edge (e, new_version, e->call_stmt, 0, e->frequency,
1449 e->loop_nest, true);
1450 new_e->count = e->count;
1451 }
1452 /* Fix recursive calls.
1453 If OLD_VERSION has a recursive call after the
1454 previous edge cloning, the new version will have an edge
1455 pointing to the old version, which is wrong;
1456 Redirect it to point to the new version. */
1457 for (e = new_version->callees ; e; e = next_callee)
1458 {
1459 next_callee = e->next_callee;
1460 if (e->callee == old_version)
1461 cgraph_redirect_edge_callee (e, new_version);
1462
1463 if (!next_callee)
1464 break;
1465 }
1466 for (i = 0; VEC_iterate (cgraph_edge_p, redirect_callers, i, e); i++)
1467 {
1468 /* Redirect calls to the old version node to point to its new
1469 version. */
1470 cgraph_redirect_edge_callee (e, new_version);
1471 }
1472
1473 return new_version;
1474 }
1475
1476 /* Perform function versioning.
1477 Function versioning includes copying of the tree and
1478 a callgraph update (creating a new cgraph node and updating
1479 its callees and callers).
1480
1481 REDIRECT_CALLERS varray includes the edges to be redirected
1482 to the new version.
1483
1484 TREE_MAP is a mapping of tree nodes we want to replace with
1485 new ones (according to results of prior analysis).
1486 OLD_VERSION_NODE is the node that is versioned.
1487 It returns the new version's cgraph node. */
1488
1489 struct cgraph_node *
1490 cgraph_function_versioning (struct cgraph_node *old_version_node,
1491 VEC(cgraph_edge_p,heap) *redirect_callers,
1492 varray_type tree_map)
1493 {
1494 tree old_decl = old_version_node->decl;
1495 struct cgraph_node *new_version_node = NULL;
1496 tree new_decl;
1497
1498 if (!tree_versionable_function_p (old_decl))
1499 return NULL;
1500
1501 /* Make a new FUNCTION_DECL tree node for the
1502 new version. */
1503 new_decl = copy_node (old_decl);
1504
1505 /* Create the new version's call-graph node.
1506 and update the edges of the new node. */
1507 new_version_node =
1508 cgraph_copy_node_for_versioning (old_version_node, new_decl,
1509 redirect_callers);
1510
1511 /* Copy the OLD_VERSION_NODE function tree to the new version. */
1512 tree_function_versioning (old_decl, new_decl, tree_map, false);
1513 /* Update the call_expr on the edges to call the new version node. */
1514 update_call_expr (new_version_node);
1515
1516 /* Update the new version's properties.
1517 Make The new version visible only within this translation unit.
1518 ??? We cannot use COMDAT linkage because there is no
1519 ABI support for this. */
1520 DECL_EXTERNAL (new_version_node->decl) = 0;
1521 DECL_ONE_ONLY (new_version_node->decl) = 0;
1522 TREE_PUBLIC (new_version_node->decl) = 0;
1523 DECL_COMDAT (new_version_node->decl) = 0;
1524 new_version_node->local.externally_visible = 0;
1525 new_version_node->local.local = 1;
1526 new_version_node->lowered = true;
1527 return new_version_node;
1528 }
1529
1530 /* Produce separate function body for inline clones so the offline copy can be
1531 modified without affecting them. */
1532 struct cgraph_node *
1533 save_inline_function_body (struct cgraph_node *node)
1534 {
1535 struct cgraph_node *first_clone;
1536
1537 gcc_assert (node == cgraph_node (node->decl));
1538
1539 cgraph_lower_function (node);
1540
1541 /* In non-unit-at-a-time we construct full fledged clone we never output to
1542 assembly file. This clone is pointed out by inline_decl of original function
1543 and inlining infrastructure knows how to deal with this. */
1544 if (!flag_unit_at_a_time)
1545 {
1546 struct cgraph_edge *e;
1547
1548 first_clone = cgraph_clone_node (node, node->count, 0, CGRAPH_FREQ_BASE,
1549 false);
1550 first_clone->needed = 0;
1551 first_clone->reachable = 1;
1552 /* Recursively clone all bodies. */
1553 for (e = first_clone->callees; e; e = e->next_callee)
1554 if (!e->inline_failed)
1555 cgraph_clone_inlined_nodes (e, true, false);
1556 }
1557 else
1558 first_clone = node->next_clone;
1559
1560 first_clone->decl = copy_node (node->decl);
1561 node->next_clone = NULL;
1562 if (!flag_unit_at_a_time)
1563 node->inline_decl = first_clone->decl;
1564 first_clone->prev_clone = NULL;
1565 cgraph_insert_node_to_hashtable (first_clone);
1566 gcc_assert (first_clone == cgraph_node (first_clone->decl));
1567
1568 /* Copy the OLD_VERSION_NODE function tree to the new version. */
1569 tree_function_versioning (node->decl, first_clone->decl, NULL, true);
1570
1571 DECL_EXTERNAL (first_clone->decl) = 0;
1572 DECL_ONE_ONLY (first_clone->decl) = 0;
1573 TREE_PUBLIC (first_clone->decl) = 0;
1574 DECL_COMDAT (first_clone->decl) = 0;
1575
1576 for (node = first_clone->next_clone; node; node = node->next_clone)
1577 node->decl = first_clone->decl;
1578 #ifdef ENABLE_CHECKING
1579 verify_cgraph_node (first_clone);
1580 #endif
1581 return first_clone;
1582 }