avr.c: Move definition of TARGET macros to end of file.
[gcc.git] / gcc / lto-cgraph.c
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
3
4 Copyright 2009, 2010, 2011 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
6
7 This file is part of GCC.
8
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
13
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
22
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "tm.h"
27 #include "tree.h"
28 #include "expr.h"
29 #include "flags.h"
30 #include "params.h"
31 #include "input.h"
32 #include "hashtab.h"
33 #include "langhooks.h"
34 #include "basic-block.h"
35 #include "tree-flow.h"
36 #include "cgraph.h"
37 #include "function.h"
38 #include "ggc.h"
39 #include "diagnostic-core.h"
40 #include "except.h"
41 #include "vec.h"
42 #include "timevar.h"
43 #include "output.h"
44 #include "pointer-set.h"
45 #include "lto-streamer.h"
46 #include "data-streamer.h"
47 #include "tree-streamer.h"
48 #include "gcov-io.h"
49
50 static void output_varpool (cgraph_node_set, varpool_node_set);
51 static void output_cgraph_opt_summary (cgraph_node_set set);
52 static void input_cgraph_opt_summary (VEC (cgraph_node_ptr, heap) * nodes);
53
54 /* Number of LDPR values known to GCC. */
55 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
56
57 /* All node orders are ofsetted by ORDER_BASE. */
58 static int order_base;
59
60 /* Cgraph streaming is organized as set of record whose type
61 is indicated by a tag. */
62 enum LTO_cgraph_tags
63 {
64 /* Must leave 0 for the stopper. */
65
66 /* Cgraph node without body available. */
67 LTO_cgraph_unavail_node = 1,
68 /* Cgraph node with function body. */
69 LTO_cgraph_analyzed_node,
70 /* Cgraph edges. */
71 LTO_cgraph_edge,
72 LTO_cgraph_indirect_edge,
73 LTO_cgraph_last_tag
74 };
75
76 /* Create a new cgraph encoder. */
77
78 lto_cgraph_encoder_t
79 lto_cgraph_encoder_new (void)
80 {
81 lto_cgraph_encoder_t encoder = XCNEW (struct lto_cgraph_encoder_d);
82 encoder->map = pointer_map_create ();
83 encoder->nodes = NULL;
84 encoder->body = pointer_set_create ();
85 return encoder;
86 }
87
88
89 /* Delete ENCODER and its components. */
90
91 void
92 lto_cgraph_encoder_delete (lto_cgraph_encoder_t encoder)
93 {
94 VEC_free (cgraph_node_ptr, heap, encoder->nodes);
95 pointer_map_destroy (encoder->map);
96 pointer_set_destroy (encoder->body);
97 free (encoder);
98 }
99
100
101 /* Return the existing reference number of NODE in the cgraph encoder in
102 output block OB. Assign a new reference if this is the first time
103 NODE is encoded. */
104
105 int
106 lto_cgraph_encoder_encode (lto_cgraph_encoder_t encoder,
107 struct cgraph_node *node)
108 {
109 int ref;
110 void **slot;
111
112 slot = pointer_map_contains (encoder->map, node);
113 if (!slot)
114 {
115 ref = VEC_length (cgraph_node_ptr, encoder->nodes);
116 slot = pointer_map_insert (encoder->map, node);
117 *slot = (void *) (intptr_t) ref;
118 VEC_safe_push (cgraph_node_ptr, heap, encoder->nodes, node);
119 }
120 else
121 ref = (int) (intptr_t) *slot;
122
123 return ref;
124 }
125
126 #define LCC_NOT_FOUND (-1)
127
128 /* Look up NODE in encoder. Return NODE's reference if it has been encoded
129 or LCC_NOT_FOUND if it is not there. */
130
131 int
132 lto_cgraph_encoder_lookup (lto_cgraph_encoder_t encoder,
133 struct cgraph_node *node)
134 {
135 void **slot = pointer_map_contains (encoder->map, node);
136 return (slot ? (int) (intptr_t) *slot : LCC_NOT_FOUND);
137 }
138
139
140 /* Return the cgraph node corresponding to REF using ENCODER. */
141
142 struct cgraph_node *
143 lto_cgraph_encoder_deref (lto_cgraph_encoder_t encoder, int ref)
144 {
145 if (ref == LCC_NOT_FOUND)
146 return NULL;
147
148 return VEC_index (cgraph_node_ptr, encoder->nodes, ref);
149 }
150
151
152 /* Return TRUE if we should encode initializer of NODE (if any). */
153
154 bool
155 lto_cgraph_encoder_encode_body_p (lto_cgraph_encoder_t encoder,
156 struct cgraph_node *node)
157 {
158 return pointer_set_contains (encoder->body, node);
159 }
160
161 /* Return TRUE if we should encode body of NODE (if any). */
162
163 static void
164 lto_set_cgraph_encoder_encode_body (lto_cgraph_encoder_t encoder,
165 struct cgraph_node *node)
166 {
167 pointer_set_insert (encoder->body, node);
168 }
169
170 /* Create a new varpool encoder. */
171
172 lto_varpool_encoder_t
173 lto_varpool_encoder_new (void)
174 {
175 lto_varpool_encoder_t encoder = XCNEW (struct lto_varpool_encoder_d);
176 encoder->map = pointer_map_create ();
177 encoder->initializer = pointer_set_create ();
178 encoder->nodes = NULL;
179 return encoder;
180 }
181
182
183 /* Delete ENCODER and its components. */
184
185 void
186 lto_varpool_encoder_delete (lto_varpool_encoder_t encoder)
187 {
188 VEC_free (varpool_node_ptr, heap, encoder->nodes);
189 pointer_map_destroy (encoder->map);
190 pointer_set_destroy (encoder->initializer);
191 free (encoder);
192 }
193
194
195 /* Return the existing reference number of NODE in the varpool encoder in
196 output block OB. Assign a new reference if this is the first time
197 NODE is encoded. */
198
199 int
200 lto_varpool_encoder_encode (lto_varpool_encoder_t encoder,
201 struct varpool_node *node)
202 {
203 int ref;
204 void **slot;
205
206 slot = pointer_map_contains (encoder->map, node);
207 if (!slot)
208 {
209 ref = VEC_length (varpool_node_ptr, encoder->nodes);
210 slot = pointer_map_insert (encoder->map, node);
211 *slot = (void *) (intptr_t) ref;
212 VEC_safe_push (varpool_node_ptr, heap, encoder->nodes, node);
213 }
214 else
215 ref = (int) (intptr_t) *slot;
216
217 return ref;
218 }
219
220 /* Look up NODE in encoder. Return NODE's reference if it has been encoded
221 or LCC_NOT_FOUND if it is not there. */
222
223 int
224 lto_varpool_encoder_lookup (lto_varpool_encoder_t encoder,
225 struct varpool_node *node)
226 {
227 void **slot = pointer_map_contains (encoder->map, node);
228 return (slot ? (int) (intptr_t) *slot : LCC_NOT_FOUND);
229 }
230
231
232 /* Return the varpool node corresponding to REF using ENCODER. */
233
234 struct varpool_node *
235 lto_varpool_encoder_deref (lto_varpool_encoder_t encoder, int ref)
236 {
237 if (ref == LCC_NOT_FOUND)
238 return NULL;
239
240 return VEC_index (varpool_node_ptr, encoder->nodes, ref);
241 }
242
243
244 /* Return TRUE if we should encode initializer of NODE (if any). */
245
246 bool
247 lto_varpool_encoder_encode_initializer_p (lto_varpool_encoder_t encoder,
248 struct varpool_node *node)
249 {
250 return pointer_set_contains (encoder->initializer, node);
251 }
252
253 /* Return TRUE if we should encode initializer of NODE (if any). */
254
255 static void
256 lto_set_varpool_encoder_encode_initializer (lto_varpool_encoder_t encoder,
257 struct varpool_node *node)
258 {
259 pointer_set_insert (encoder->initializer, node);
260 }
261
262 /* Output the cgraph EDGE to OB using ENCODER. */
263
264 static void
265 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
266 lto_cgraph_encoder_t encoder)
267 {
268 unsigned int uid;
269 intptr_t ref;
270 struct bitpack_d bp;
271
272 if (edge->indirect_unknown_callee)
273 streamer_write_enum (ob->main_stream, LTO_cgraph_tags, LTO_cgraph_last_tag,
274 LTO_cgraph_indirect_edge);
275 else
276 streamer_write_enum (ob->main_stream, LTO_cgraph_tags, LTO_cgraph_last_tag,
277 LTO_cgraph_edge);
278
279 ref = lto_cgraph_encoder_lookup (encoder, edge->caller);
280 gcc_assert (ref != LCC_NOT_FOUND);
281 streamer_write_hwi_stream (ob->main_stream, ref);
282
283 if (!edge->indirect_unknown_callee)
284 {
285 ref = lto_cgraph_encoder_lookup (encoder, edge->callee);
286 gcc_assert (ref != LCC_NOT_FOUND);
287 streamer_write_hwi_stream (ob->main_stream, ref);
288 }
289
290 streamer_write_hwi_stream (ob->main_stream, edge->count);
291
292 bp = bitpack_create (ob->main_stream);
293 uid = (!gimple_has_body_p (edge->caller->decl)
294 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt));
295 bp_pack_enum (&bp, cgraph_inline_failed_enum,
296 CIF_N_REASONS, edge->inline_failed);
297 bp_pack_var_len_unsigned (&bp, uid);
298 bp_pack_var_len_unsigned (&bp, edge->frequency);
299 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
300 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
301 bp_pack_value (&bp, edge->can_throw_external, 1);
302 if (edge->indirect_unknown_callee)
303 {
304 int flags = edge->indirect_info->ecf_flags;
305 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
306 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
307 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
308 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
309 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
310 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
311 /* Flags that should not appear on indirect calls. */
312 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
313 | ECF_MAY_BE_ALLOCA
314 | ECF_SIBCALL
315 | ECF_LEAF
316 | ECF_NOVOPS)));
317 }
318 streamer_write_bitpack (&bp);
319 }
320
321 /* Return if LIST contain references from other partitions. */
322
323 bool
324 referenced_from_other_partition_p (struct ipa_ref_list *list, cgraph_node_set set,
325 varpool_node_set vset)
326 {
327 int i;
328 struct ipa_ref *ref;
329 for (i = 0; ipa_ref_list_refering_iterate (list, i, ref); i++)
330 {
331 if (ref->refering_type == IPA_REF_CGRAPH)
332 {
333 if (ipa_ref_refering_node (ref)->in_other_partition
334 || !cgraph_node_in_set_p (ipa_ref_refering_node (ref), set))
335 return true;
336 }
337 else
338 {
339 if (ipa_ref_refering_varpool_node (ref)->in_other_partition
340 || !varpool_node_in_set_p (ipa_ref_refering_varpool_node (ref),
341 vset))
342 return true;
343 }
344 }
345 return false;
346 }
347
348 /* Return true when node is reachable from other partition. */
349
350 bool
351 reachable_from_other_partition_p (struct cgraph_node *node, cgraph_node_set set)
352 {
353 struct cgraph_edge *e;
354 if (!node->analyzed)
355 return false;
356 if (node->global.inlined_to)
357 return false;
358 for (e = node->callers; e; e = e->next_caller)
359 if (e->caller->in_other_partition
360 || !cgraph_node_in_set_p (e->caller, set))
361 return true;
362 return false;
363 }
364
365 /* Return if LIST contain references from other partitions. */
366
367 bool
368 referenced_from_this_partition_p (struct ipa_ref_list *list, cgraph_node_set set,
369 varpool_node_set vset)
370 {
371 int i;
372 struct ipa_ref *ref;
373 for (i = 0; ipa_ref_list_refering_iterate (list, i, ref); i++)
374 {
375 if (ref->refering_type == IPA_REF_CGRAPH)
376 {
377 if (cgraph_node_in_set_p (ipa_ref_refering_node (ref), set))
378 return true;
379 }
380 else
381 {
382 if (varpool_node_in_set_p (ipa_ref_refering_varpool_node (ref),
383 vset))
384 return true;
385 }
386 }
387 return false;
388 }
389
390 /* Return true when node is reachable from other partition. */
391
392 bool
393 reachable_from_this_partition_p (struct cgraph_node *node, cgraph_node_set set)
394 {
395 struct cgraph_edge *e;
396 for (e = node->callers; e; e = e->next_caller)
397 if (cgraph_node_in_set_p (e->caller, set))
398 return true;
399 return false;
400 }
401
402 /* Output the cgraph NODE to OB. ENCODER is used to find the
403 reference number of NODE->inlined_to. SET is the set of nodes we
404 are writing to the current file. If NODE is not in SET, then NODE
405 is a boundary of a cgraph_node_set and we pretend NODE just has a
406 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
407 that have had their callgraph node written so far. This is used to
408 determine if NODE is a clone of a previously written node. */
409
410 static void
411 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
412 lto_cgraph_encoder_t encoder, cgraph_node_set set,
413 varpool_node_set vset)
414 {
415 unsigned int tag;
416 struct bitpack_d bp;
417 bool boundary_p;
418 intptr_t ref;
419 bool in_other_partition = false;
420 struct cgraph_node *clone_of;
421
422 boundary_p = !cgraph_node_in_set_p (node, set);
423
424 if (node->analyzed && !boundary_p)
425 tag = LTO_cgraph_analyzed_node;
426 else
427 tag = LTO_cgraph_unavail_node;
428
429 streamer_write_enum (ob->main_stream, LTO_cgraph_tags, LTO_cgraph_last_tag,
430 tag);
431 streamer_write_hwi_stream (ob->main_stream, node->order);
432
433 /* In WPA mode, we only output part of the call-graph. Also, we
434 fake cgraph node attributes. There are two cases that we care.
435
436 Boundary nodes: There are nodes that are not part of SET but are
437 called from within SET. We artificially make them look like
438 externally visible nodes with no function body.
439
440 Cherry-picked nodes: These are nodes we pulled from other
441 translation units into SET during IPA-inlining. We make them as
442 local static nodes to prevent clashes with other local statics. */
443 if (boundary_p && node->analyzed)
444 {
445 /* Inline clones can not be part of boundary.
446 gcc_assert (!node->global.inlined_to);
447
448 FIXME: At the moment they can be, when partition contains an inline
449 clone that is clone of inline clone from outside partition. We can
450 reshape the clone tree and make other tree to be the root, but it
451 needs a bit extra work and will be promplty done by cgraph_remove_node
452 after reading back. */
453 in_other_partition = 1;
454 }
455
456 clone_of = node->clone_of;
457 while (clone_of
458 && (ref = lto_cgraph_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND)
459 if (clone_of->prev_sibling_clone)
460 clone_of = clone_of->prev_sibling_clone;
461 else
462 clone_of = clone_of->clone_of;
463
464 if (LTO_cgraph_analyzed_node)
465 gcc_assert (clone_of || !node->clone_of);
466 if (!clone_of)
467 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
468 else
469 streamer_write_hwi_stream (ob->main_stream, ref);
470
471
472 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl);
473 streamer_write_hwi_stream (ob->main_stream, node->count);
474 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
475
476 if (tag == LTO_cgraph_analyzed_node)
477 {
478 if (node->global.inlined_to)
479 {
480 ref = lto_cgraph_encoder_lookup (encoder, node->global.inlined_to);
481 gcc_assert (ref != LCC_NOT_FOUND);
482 }
483 else
484 ref = LCC_NOT_FOUND;
485
486 streamer_write_hwi_stream (ob->main_stream, ref);
487 }
488
489 if (node->same_comdat_group && !boundary_p)
490 {
491 ref = lto_cgraph_encoder_lookup (encoder, node->same_comdat_group);
492 gcc_assert (ref != LCC_NOT_FOUND);
493 }
494 else
495 ref = LCC_NOT_FOUND;
496 streamer_write_hwi_stream (ob->main_stream, ref);
497
498 bp = bitpack_create (ob->main_stream);
499 bp_pack_value (&bp, node->local.local, 1);
500 bp_pack_value (&bp, node->local.externally_visible, 1);
501 bp_pack_value (&bp, node->local.finalized, 1);
502 bp_pack_value (&bp, node->local.versionable, 1);
503 bp_pack_value (&bp, node->local.can_change_signature, 1);
504 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
505 bp_pack_value (&bp, node->needed, 1);
506 bp_pack_value (&bp, node->address_taken, 1);
507 bp_pack_value (&bp, node->abstract_and_needed, 1);
508 bp_pack_value (&bp, tag == LTO_cgraph_analyzed_node
509 && !DECL_EXTERNAL (node->decl)
510 && !DECL_COMDAT (node->decl)
511 && (reachable_from_other_partition_p (node, set)
512 || referenced_from_other_partition_p (&node->ref_list, set, vset)), 1);
513 bp_pack_value (&bp, node->lowered, 1);
514 bp_pack_value (&bp, in_other_partition, 1);
515 /* Real aliases in a boundary become non-aliases. However we still stream
516 alias info on weakrefs.
517 TODO: We lose a bit of information here - when we know that variable is
518 defined in other unit, we may use the info on aliases to resolve
519 symbol1 != symbol2 type tests that we can do only for locally defined objects
520 otherwise. */
521 bp_pack_value (&bp, node->alias && (!boundary_p || DECL_EXTERNAL (node->decl)), 1);
522 bp_pack_value (&bp, node->frequency, 2);
523 bp_pack_value (&bp, node->only_called_at_startup, 1);
524 bp_pack_value (&bp, node->only_called_at_exit, 1);
525 bp_pack_value (&bp, node->tm_clone, 1);
526 bp_pack_value (&bp, node->thunk.thunk_p && !boundary_p, 1);
527 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
528 LDPR_NUM_KNOWN, node->resolution);
529 streamer_write_bitpack (&bp);
530
531 if (node->thunk.thunk_p && !boundary_p)
532 {
533 streamer_write_uhwi_stream
534 (ob->main_stream,
535 1 + (node->thunk.this_adjusting != 0) * 2
536 + (node->thunk.virtual_offset_p != 0) * 4);
537 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
538 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
539 }
540 if ((node->alias || node->thunk.thunk_p)
541 && (!boundary_p || (node->alias && DECL_EXTERNAL (node->decl))))
542 {
543 streamer_write_hwi_in_range (ob->main_stream, 0, 1,
544 node->thunk.alias != NULL);
545 if (node->thunk.alias != NULL)
546 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
547 node->thunk.alias);
548 }
549 }
550
551 /* Output the varpool NODE to OB.
552 If NODE is not in SET, then NODE is a boundary. */
553
554 static void
555 lto_output_varpool_node (struct lto_simple_output_block *ob, struct varpool_node *node,
556 lto_varpool_encoder_t varpool_encoder,
557 cgraph_node_set set, varpool_node_set vset)
558 {
559 bool boundary_p = !varpool_node_in_set_p (node, vset) && node->analyzed;
560 struct bitpack_d bp;
561 int ref;
562
563 streamer_write_hwi_stream (ob->main_stream, node->order);
564 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl);
565 bp = bitpack_create (ob->main_stream);
566 bp_pack_value (&bp, node->externally_visible, 1);
567 bp_pack_value (&bp, node->force_output, 1);
568 bp_pack_value (&bp, node->finalized, 1);
569 bp_pack_value (&bp, node->alias, 1);
570 bp_pack_value (&bp, node->alias_of != NULL, 1);
571 gcc_assert (node->finalized || !node->analyzed);
572 gcc_assert (node->needed);
573 /* Constant pool initializers can be de-unified into individual ltrans units.
574 FIXME: Alternatively at -Os we may want to avoid generating for them the local
575 labels and share them across LTRANS partitions. */
576 if (DECL_IN_CONSTANT_POOL (node->decl)
577 && !DECL_COMDAT (node->decl))
578 {
579 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
580 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
581 }
582 else
583 {
584 bp_pack_value (&bp, node->analyzed
585 && referenced_from_other_partition_p (&node->ref_list,
586 set, vset), 1);
587 bp_pack_value (&bp, boundary_p, 1); /* in_other_partition. */
588 }
589 streamer_write_bitpack (&bp);
590 if (node->alias_of)
591 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->alias_of);
592 if (node->same_comdat_group && !boundary_p)
593 {
594 ref = lto_varpool_encoder_lookup (varpool_encoder, node->same_comdat_group);
595 gcc_assert (ref != LCC_NOT_FOUND);
596 }
597 else
598 ref = LCC_NOT_FOUND;
599 streamer_write_hwi_stream (ob->main_stream, ref);
600 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
601 LDPR_NUM_KNOWN, node->resolution);
602 }
603
604 /* Output the varpool NODE to OB.
605 If NODE is not in SET, then NODE is a boundary. */
606
607 static void
608 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
609 lto_cgraph_encoder_t encoder,
610 lto_varpool_encoder_t varpool_encoder)
611 {
612 struct bitpack_d bp;
613 bp = bitpack_create (ob->main_stream);
614 bp_pack_value (&bp, ref->refered_type, 1);
615 bp_pack_value (&bp, ref->use, 2);
616 streamer_write_bitpack (&bp);
617 if (ref->refered_type == IPA_REF_CGRAPH)
618 {
619 int nref = lto_cgraph_encoder_lookup (encoder, ipa_ref_node (ref));
620 gcc_assert (nref != LCC_NOT_FOUND);
621 streamer_write_hwi_stream (ob->main_stream, nref);
622 }
623 else
624 {
625 int nref = lto_varpool_encoder_lookup (varpool_encoder,
626 ipa_ref_varpool_node (ref));
627 gcc_assert (nref != LCC_NOT_FOUND);
628 streamer_write_hwi_stream (ob->main_stream, nref);
629 }
630 }
631
632 /* Stream out profile_summary to OB. */
633
634 static void
635 output_profile_summary (struct lto_simple_output_block *ob)
636 {
637 if (profile_info)
638 {
639 /* We do not output num, sum_all and run_max, they are not used by
640 GCC profile feedback and they are difficult to merge from multiple
641 units. */
642 gcc_assert (profile_info->runs);
643 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
644 streamer_write_uhwi_stream (ob->main_stream, profile_info->sum_max);
645 }
646 else
647 streamer_write_uhwi_stream (ob->main_stream, 0);
648 }
649
650 /* Add NODE into encoder as well as nodes it is cloned from.
651 Do it in a way so clones appear first. */
652
653 static void
654 add_node_to (lto_cgraph_encoder_t encoder, struct cgraph_node *node,
655 bool include_body)
656 {
657 if (node->clone_of)
658 add_node_to (encoder, node->clone_of, include_body);
659 else if (include_body)
660 lto_set_cgraph_encoder_encode_body (encoder, node);
661 lto_cgraph_encoder_encode (encoder, node);
662 }
663
664 /* Add all references in LIST to encoders. */
665
666 static void
667 add_references (lto_cgraph_encoder_t encoder,
668 lto_varpool_encoder_t varpool_encoder,
669 struct ipa_ref_list *list)
670 {
671 int i;
672 struct ipa_ref *ref;
673 for (i = 0; ipa_ref_list_reference_iterate (list, i, ref); i++)
674 if (ref->refered_type == IPA_REF_CGRAPH)
675 add_node_to (encoder, ipa_ref_node (ref), false);
676 else
677 {
678 struct varpool_node *vnode = ipa_ref_varpool_node (ref);
679 lto_varpool_encoder_encode (varpool_encoder, vnode);
680 }
681 }
682
683 /* Output all callees or indirect outgoing edges. EDGE must be the first such
684 edge. */
685
686 static void
687 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
688 struct lto_simple_output_block *ob,
689 lto_cgraph_encoder_t encoder)
690 {
691 if (!edge)
692 return;
693
694 /* Output edges in backward direction, so the reconstructed callgraph match
695 and it is easy to associate call sites in the IPA pass summaries. */
696 while (edge->next_callee)
697 edge = edge->next_callee;
698 for (; edge; edge = edge->prev_callee)
699 lto_output_edge (ob, edge, encoder);
700 }
701
702 /* Output the part of the cgraph in SET. */
703
704 static void
705 output_refs (cgraph_node_set set, varpool_node_set vset,
706 lto_cgraph_encoder_t encoder,
707 lto_varpool_encoder_t varpool_encoder)
708 {
709 cgraph_node_set_iterator csi;
710 varpool_node_set_iterator vsi;
711 struct lto_simple_output_block *ob;
712 int count;
713 struct ipa_ref *ref;
714 int i;
715
716 ob = lto_create_simple_output_block (LTO_section_refs);
717
718 for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi))
719 {
720 struct cgraph_node *node = csi_node (csi);
721
722 count = ipa_ref_list_nreferences (&node->ref_list);
723 if (count)
724 {
725 streamer_write_uhwi_stream (ob->main_stream, count);
726 streamer_write_uhwi_stream (ob->main_stream,
727 lto_cgraph_encoder_lookup (encoder, node));
728 for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list, i, ref); i++)
729 lto_output_ref (ob, ref, encoder, varpool_encoder);
730 }
731 }
732
733 streamer_write_uhwi_stream (ob->main_stream, 0);
734
735 for (vsi = vsi_start (vset); !vsi_end_p (vsi); vsi_next (&vsi))
736 {
737 struct varpool_node *node = vsi_node (vsi);
738
739 count = ipa_ref_list_nreferences (&node->ref_list);
740 if (count)
741 {
742 streamer_write_uhwi_stream (ob->main_stream, count);
743 streamer_write_uhwi_stream (ob->main_stream,
744 lto_varpool_encoder_lookup (varpool_encoder,
745 node));
746 for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list, i, ref); i++)
747 lto_output_ref (ob, ref, encoder, varpool_encoder);
748 }
749 }
750
751 streamer_write_uhwi_stream (ob->main_stream, 0);
752
753 lto_destroy_simple_output_block (ob);
754 }
755
756 /* Find out all cgraph and varpool nodes we want to encode in current unit
757 and insert them to encoders. */
758 void
759 compute_ltrans_boundary (struct lto_out_decl_state *state,
760 cgraph_node_set set, varpool_node_set vset)
761 {
762 struct cgraph_node *node;
763 cgraph_node_set_iterator csi;
764 varpool_node_set_iterator vsi;
765 struct cgraph_edge *edge;
766 int i;
767 lto_cgraph_encoder_t encoder;
768 lto_varpool_encoder_t varpool_encoder;
769
770 encoder = state->cgraph_node_encoder = lto_cgraph_encoder_new ();
771 varpool_encoder = state->varpool_node_encoder = lto_varpool_encoder_new ();
772
773 /* Go over all the nodes in SET and assign references. */
774 for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi))
775 {
776 node = csi_node (csi);
777 add_node_to (encoder, node, true);
778 add_references (encoder, varpool_encoder, &node->ref_list);
779 }
780 for (vsi = vsi_start (vset); !vsi_end_p (vsi); vsi_next (&vsi))
781 {
782 struct varpool_node *vnode = vsi_node (vsi);
783 gcc_assert (!vnode->alias || vnode->alias_of);
784 lto_varpool_encoder_encode (varpool_encoder, vnode);
785 lto_set_varpool_encoder_encode_initializer (varpool_encoder, vnode);
786 add_references (encoder, varpool_encoder, &vnode->ref_list);
787 }
788 /* Pickle in also the initializer of all referenced readonly variables
789 to help folding. Constant pool variables are not shared, so we must
790 pickle those too. */
791 for (i = 0; i < lto_varpool_encoder_size (varpool_encoder); i++)
792 {
793 struct varpool_node *vnode = lto_varpool_encoder_deref (varpool_encoder, i);
794 if (DECL_INITIAL (vnode->decl)
795 && !lto_varpool_encoder_encode_initializer_p (varpool_encoder,
796 vnode)
797 && const_value_known_p (vnode->decl))
798 {
799 lto_set_varpool_encoder_encode_initializer (varpool_encoder, vnode);
800 add_references (encoder, varpool_encoder, &vnode->ref_list);
801 }
802 }
803
804 /* Go over all the nodes again to include callees that are not in
805 SET. */
806 for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi))
807 {
808 node = csi_node (csi);
809 for (edge = node->callees; edge; edge = edge->next_callee)
810 {
811 struct cgraph_node *callee = edge->callee;
812 if (!cgraph_node_in_set_p (callee, set))
813 {
814 /* We should have moved all the inlines. */
815 gcc_assert (!callee->global.inlined_to);
816 add_node_to (encoder, callee, false);
817 }
818 }
819 }
820 }
821
822 /* Output the part of the cgraph in SET. */
823
824 void
825 output_cgraph (cgraph_node_set set, varpool_node_set vset)
826 {
827 struct cgraph_node *node;
828 struct lto_simple_output_block *ob;
829 cgraph_node_set_iterator csi;
830 int i, n_nodes;
831 lto_cgraph_encoder_t encoder;
832 lto_varpool_encoder_t varpool_encoder;
833 static bool asm_nodes_output = false;
834
835 if (flag_wpa)
836 output_cgraph_opt_summary (set);
837
838 ob = lto_create_simple_output_block (LTO_section_cgraph);
839
840 output_profile_summary (ob);
841
842 /* An encoder for cgraph nodes should have been created by
843 ipa_write_summaries_1. */
844 gcc_assert (ob->decl_state->cgraph_node_encoder);
845 gcc_assert (ob->decl_state->varpool_node_encoder);
846 encoder = ob->decl_state->cgraph_node_encoder;
847 varpool_encoder = ob->decl_state->varpool_node_encoder;
848
849 /* Write out the nodes. We must first output a node and then its clones,
850 otherwise at a time reading back the node there would be nothing to clone
851 from. */
852 n_nodes = lto_cgraph_encoder_size (encoder);
853 for (i = 0; i < n_nodes; i++)
854 {
855 node = lto_cgraph_encoder_deref (encoder, i);
856 lto_output_node (ob, node, encoder, set, vset);
857 }
858
859 /* Go over the nodes in SET again to write edges. */
860 for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi))
861 {
862 node = csi_node (csi);
863 output_outgoing_cgraph_edges (node->callees, ob, encoder);
864 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
865 }
866
867 streamer_write_uhwi_stream (ob->main_stream, 0);
868
869 lto_destroy_simple_output_block (ob);
870
871 /* Emit toplevel asms.
872 When doing WPA we must output every asm just once. Since we do not partition asm
873 nodes at all, output them to first output. This is kind of hack, but should work
874 well. */
875 if (!asm_nodes_output)
876 {
877 asm_nodes_output = true;
878 lto_output_toplevel_asms ();
879 }
880
881 output_varpool (set, vset);
882 output_refs (set, vset, encoder, varpool_encoder);
883 }
884
885 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
886 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
887 NODE or to replace the values in it, for instance because the first
888 time we saw it, the function body was not available but now it
889 is. BP is a bitpack with all the bitflags for NODE read from the
890 stream. */
891
892 static void
893 input_overwrite_node (struct lto_file_decl_data *file_data,
894 struct cgraph_node *node,
895 enum LTO_cgraph_tags tag,
896 struct bitpack_d *bp)
897 {
898 node->aux = (void *) tag;
899 node->local.lto_file_data = file_data;
900
901 node->local.local = bp_unpack_value (bp, 1);
902 node->local.externally_visible = bp_unpack_value (bp, 1);
903 node->local.finalized = bp_unpack_value (bp, 1);
904 node->local.versionable = bp_unpack_value (bp, 1);
905 node->local.can_change_signature = bp_unpack_value (bp, 1);
906 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
907 node->needed = bp_unpack_value (bp, 1);
908 node->address_taken = bp_unpack_value (bp, 1);
909 node->abstract_and_needed = bp_unpack_value (bp, 1);
910 node->reachable_from_other_partition = bp_unpack_value (bp, 1);
911 node->lowered = bp_unpack_value (bp, 1);
912 node->analyzed = tag == LTO_cgraph_analyzed_node;
913 node->in_other_partition = bp_unpack_value (bp, 1);
914 if (node->in_other_partition
915 /* Avoid updating decl when we are seeing just inline clone.
916 When inlining function that has functions already inlined into it,
917 we produce clones of inline clones.
918
919 WPA partitioning might put each clone into different unit and
920 we might end up streaming inline clone from other partition
921 to support clone we are interested in. */
922 && (!node->clone_of
923 || node->clone_of->decl != node->decl))
924 {
925 DECL_EXTERNAL (node->decl) = 1;
926 TREE_STATIC (node->decl) = 0;
927 }
928 node->alias = bp_unpack_value (bp, 1);
929 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
930 node->only_called_at_startup = bp_unpack_value (bp, 1);
931 node->only_called_at_exit = bp_unpack_value (bp, 1);
932 node->tm_clone = bp_unpack_value (bp, 1);
933 node->thunk.thunk_p = bp_unpack_value (bp, 1);
934 node->resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
935 LDPR_NUM_KNOWN);
936 }
937
938 /* Output the part of the cgraph in SET. */
939
940 static void
941 output_varpool (cgraph_node_set set, varpool_node_set vset)
942 {
943 struct lto_simple_output_block *ob = lto_create_simple_output_block (LTO_section_varpool);
944 lto_varpool_encoder_t varpool_encoder = ob->decl_state->varpool_node_encoder;
945 int len = lto_varpool_encoder_size (varpool_encoder), i;
946
947 streamer_write_uhwi_stream (ob->main_stream, len);
948
949 /* Write out the nodes. We must first output a node and then its clones,
950 otherwise at a time reading back the node there would be nothing to clone
951 from. */
952 for (i = 0; i < len; i++)
953 {
954 lto_output_varpool_node (ob, lto_varpool_encoder_deref (varpool_encoder, i),
955 varpool_encoder,
956 set, vset);
957 }
958
959 lto_destroy_simple_output_block (ob);
960 }
961
962 /* Read a node from input_block IB. TAG is the node's tag just read.
963 Return the node read or overwriten. */
964
965 static struct cgraph_node *
966 input_node (struct lto_file_decl_data *file_data,
967 struct lto_input_block *ib,
968 enum LTO_cgraph_tags tag,
969 VEC(cgraph_node_ptr, heap) *nodes)
970 {
971 tree fn_decl;
972 struct cgraph_node *node;
973 struct bitpack_d bp;
974 unsigned decl_index;
975 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
976 int clone_ref;
977 int order;
978
979 order = streamer_read_hwi (ib) + order_base;
980 clone_ref = streamer_read_hwi (ib);
981
982 decl_index = streamer_read_uhwi (ib);
983 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
984
985 if (clone_ref != LCC_NOT_FOUND)
986 {
987 node = cgraph_clone_node (VEC_index (cgraph_node_ptr, nodes, clone_ref), fn_decl,
988 0, CGRAPH_FREQ_BASE, false, NULL, false);
989 }
990 else
991 node = cgraph_get_create_node (fn_decl);
992
993 node->order = order;
994 if (order >= cgraph_order)
995 cgraph_order = order + 1;
996
997 node->count = streamer_read_hwi (ib);
998 node->count_materialization_scale = streamer_read_hwi (ib);
999
1000 if (tag == LTO_cgraph_analyzed_node)
1001 ref = streamer_read_hwi (ib);
1002
1003 ref2 = streamer_read_hwi (ib);
1004
1005 /* Make sure that we have not read this node before. Nodes that
1006 have already been read will have their tag stored in the 'aux'
1007 field. Since built-in functions can be referenced in multiple
1008 functions, they are expected to be read more than once. */
1009 if (node->aux && !DECL_BUILT_IN (node->decl))
1010 internal_error ("bytecode stream: found multiple instances of cgraph "
1011 "node %d", node->uid);
1012
1013 bp = streamer_read_bitpack (ib);
1014 input_overwrite_node (file_data, node, tag, &bp);
1015
1016 /* Store a reference for now, and fix up later to be a pointer. */
1017 node->global.inlined_to = (cgraph_node_ptr) (intptr_t) ref;
1018
1019 /* Store a reference for now, and fix up later to be a pointer. */
1020 node->same_comdat_group = (cgraph_node_ptr) (intptr_t) ref2;
1021
1022 if (node->thunk.thunk_p)
1023 {
1024 int type = streamer_read_uhwi (ib);
1025 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
1026 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
1027
1028 node->thunk.fixed_offset = fixed_offset;
1029 node->thunk.this_adjusting = (type & 2);
1030 node->thunk.virtual_value = virtual_value;
1031 node->thunk.virtual_offset_p = (type & 4);
1032 }
1033 if (node->thunk.thunk_p || node->alias)
1034 {
1035 if (streamer_read_hwi_in_range (ib, "alias nonzero flag", 0, 1))
1036 {
1037 decl_index = streamer_read_uhwi (ib);
1038 node->thunk.alias = lto_file_decl_data_get_fn_decl (file_data,
1039 decl_index);
1040 }
1041 }
1042 return node;
1043 }
1044
1045 /* Read a node from input_block IB. TAG is the node's tag just read.
1046 Return the node read or overwriten. */
1047
1048 static struct varpool_node *
1049 input_varpool_node (struct lto_file_decl_data *file_data,
1050 struct lto_input_block *ib)
1051 {
1052 int decl_index;
1053 tree var_decl;
1054 struct varpool_node *node;
1055 struct bitpack_d bp;
1056 int ref = LCC_NOT_FOUND;
1057 bool non_null_aliasof;
1058 int order;
1059
1060 order = streamer_read_hwi (ib) + order_base;
1061 decl_index = streamer_read_uhwi (ib);
1062 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1063 node = varpool_node (var_decl);
1064 node->order = order;
1065 if (order >= cgraph_order)
1066 cgraph_order = order + 1;
1067 node->lto_file_data = file_data;
1068
1069 bp = streamer_read_bitpack (ib);
1070 node->externally_visible = bp_unpack_value (&bp, 1);
1071 node->force_output = bp_unpack_value (&bp, 1);
1072 node->finalized = bp_unpack_value (&bp, 1);
1073 node->alias = bp_unpack_value (&bp, 1);
1074 non_null_aliasof = bp_unpack_value (&bp, 1);
1075 node->analyzed = node->finalized;
1076 node->used_from_other_partition = bp_unpack_value (&bp, 1);
1077 node->in_other_partition = bp_unpack_value (&bp, 1);
1078 if (node->in_other_partition)
1079 {
1080 DECL_EXTERNAL (node->decl) = 1;
1081 TREE_STATIC (node->decl) = 0;
1082 }
1083 if (node->finalized)
1084 varpool_mark_needed_node (node);
1085 if (non_null_aliasof)
1086 {
1087 decl_index = streamer_read_uhwi (ib);
1088 node->alias_of = lto_file_decl_data_get_var_decl (file_data, decl_index);
1089 }
1090 ref = streamer_read_hwi (ib);
1091 /* Store a reference for now, and fix up later to be a pointer. */
1092 node->same_comdat_group = (struct varpool_node *) (intptr_t) ref;
1093 node->resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1094 LDPR_NUM_KNOWN);
1095
1096 return node;
1097 }
1098
1099 /* Read a node from input_block IB. TAG is the node's tag just read.
1100 Return the node read or overwriten. */
1101
1102 static void
1103 input_ref (struct lto_input_block *ib,
1104 struct cgraph_node *refering_node,
1105 struct varpool_node *refering_varpool_node,
1106 VEC(cgraph_node_ptr, heap) *nodes,
1107 VEC(varpool_node_ptr, heap) *varpool_nodes)
1108 {
1109 struct cgraph_node *node = NULL;
1110 struct varpool_node *varpool_node = NULL;
1111 struct bitpack_d bp;
1112 enum ipa_ref_type type;
1113 enum ipa_ref_use use;
1114
1115 bp = streamer_read_bitpack (ib);
1116 type = (enum ipa_ref_type) bp_unpack_value (&bp, 1);
1117 use = (enum ipa_ref_use) bp_unpack_value (&bp, 2);
1118 if (type == IPA_REF_CGRAPH)
1119 node = VEC_index (cgraph_node_ptr, nodes, streamer_read_hwi (ib));
1120 else
1121 varpool_node = VEC_index (varpool_node_ptr, varpool_nodes,
1122 streamer_read_hwi (ib));
1123 ipa_record_reference (refering_node, refering_varpool_node,
1124 node, varpool_node, use, NULL);
1125 }
1126
1127 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1128 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1129 edge being read is indirect (in the sense that it has
1130 indirect_unknown_callee set). */
1131
1132 static void
1133 input_edge (struct lto_input_block *ib, VEC(cgraph_node_ptr, heap) *nodes,
1134 bool indirect)
1135 {
1136 struct cgraph_node *caller, *callee;
1137 struct cgraph_edge *edge;
1138 unsigned int stmt_id;
1139 gcov_type count;
1140 int freq;
1141 cgraph_inline_failed_t inline_failed;
1142 struct bitpack_d bp;
1143 int ecf_flags = 0;
1144
1145 caller = VEC_index (cgraph_node_ptr, nodes, streamer_read_hwi (ib));
1146 if (caller == NULL || caller->decl == NULL_TREE)
1147 internal_error ("bytecode stream: no caller found while reading edge");
1148
1149 if (!indirect)
1150 {
1151 callee = VEC_index (cgraph_node_ptr, nodes, streamer_read_hwi (ib));
1152 if (callee == NULL || callee->decl == NULL_TREE)
1153 internal_error ("bytecode stream: no callee found while reading edge");
1154 }
1155 else
1156 callee = NULL;
1157
1158 count = (gcov_type) streamer_read_hwi (ib);
1159
1160 bp = streamer_read_bitpack (ib);
1161 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_enum, CIF_N_REASONS);
1162 stmt_id = bp_unpack_var_len_unsigned (&bp);
1163 freq = (int) bp_unpack_var_len_unsigned (&bp);
1164
1165 if (indirect)
1166 edge = cgraph_create_indirect_edge (caller, NULL, 0, count, freq);
1167 else
1168 edge = cgraph_create_edge (caller, callee, NULL, count, freq);
1169
1170 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1171 edge->lto_stmt_uid = stmt_id;
1172 edge->inline_failed = inline_failed;
1173 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1174 edge->can_throw_external = bp_unpack_value (&bp, 1);
1175 if (indirect)
1176 {
1177 if (bp_unpack_value (&bp, 1))
1178 ecf_flags |= ECF_CONST;
1179 if (bp_unpack_value (&bp, 1))
1180 ecf_flags |= ECF_PURE;
1181 if (bp_unpack_value (&bp, 1))
1182 ecf_flags |= ECF_NORETURN;
1183 if (bp_unpack_value (&bp, 1))
1184 ecf_flags |= ECF_MALLOC;
1185 if (bp_unpack_value (&bp, 1))
1186 ecf_flags |= ECF_NOTHROW;
1187 if (bp_unpack_value (&bp, 1))
1188 ecf_flags |= ECF_RETURNS_TWICE;
1189 edge->indirect_info->ecf_flags = ecf_flags;
1190 }
1191 }
1192
1193
1194 /* Read a cgraph from IB using the info in FILE_DATA. */
1195
1196 static VEC(cgraph_node_ptr, heap) *
1197 input_cgraph_1 (struct lto_file_decl_data *file_data,
1198 struct lto_input_block *ib)
1199 {
1200 enum LTO_cgraph_tags tag;
1201 VEC(cgraph_node_ptr, heap) *nodes = NULL;
1202 struct cgraph_node *node;
1203 unsigned i;
1204
1205 tag = streamer_read_enum (ib, LTO_cgraph_tags, LTO_cgraph_last_tag);
1206 order_base = cgraph_order;
1207 while (tag)
1208 {
1209 if (tag == LTO_cgraph_edge)
1210 input_edge (ib, nodes, false);
1211 else if (tag == LTO_cgraph_indirect_edge)
1212 input_edge (ib, nodes, true);
1213 else
1214 {
1215 node = input_node (file_data, ib, tag,nodes);
1216 if (node == NULL || node->decl == NULL_TREE)
1217 internal_error ("bytecode stream: found empty cgraph node");
1218 VEC_safe_push (cgraph_node_ptr, heap, nodes, node);
1219 lto_cgraph_encoder_encode (file_data->cgraph_node_encoder, node);
1220 }
1221
1222 tag = streamer_read_enum (ib, LTO_cgraph_tags, LTO_cgraph_last_tag);
1223 }
1224
1225 lto_input_toplevel_asms (file_data, order_base);
1226
1227 /* AUX pointers should be all non-zero for nodes read from the stream. */
1228 #ifdef ENABLE_CHECKING
1229 FOR_EACH_VEC_ELT (cgraph_node_ptr, nodes, i, node)
1230 gcc_assert (node->aux);
1231 #endif
1232 FOR_EACH_VEC_ELT (cgraph_node_ptr, nodes, i, node)
1233 {
1234 int ref = (int) (intptr_t) node->global.inlined_to;
1235
1236 /* We share declaration of builtins, so we may read same node twice. */
1237 if (!node->aux)
1238 continue;
1239 node->aux = NULL;
1240
1241 /* Fixup inlined_to from reference to pointer. */
1242 if (ref != LCC_NOT_FOUND)
1243 node->global.inlined_to = VEC_index (cgraph_node_ptr, nodes, ref);
1244 else
1245 node->global.inlined_to = NULL;
1246
1247 ref = (int) (intptr_t) node->same_comdat_group;
1248
1249 /* Fixup same_comdat_group from reference to pointer. */
1250 if (ref != LCC_NOT_FOUND)
1251 node->same_comdat_group = VEC_index (cgraph_node_ptr, nodes, ref);
1252 else
1253 node->same_comdat_group = NULL;
1254 }
1255 FOR_EACH_VEC_ELT (cgraph_node_ptr, nodes, i, node)
1256 node->aux = (void *)1;
1257 return nodes;
1258 }
1259
1260 /* Read a varpool from IB using the info in FILE_DATA. */
1261
1262 static VEC(varpool_node_ptr, heap) *
1263 input_varpool_1 (struct lto_file_decl_data *file_data,
1264 struct lto_input_block *ib)
1265 {
1266 unsigned HOST_WIDE_INT len;
1267 VEC(varpool_node_ptr, heap) *varpool = NULL;
1268 int i;
1269 struct varpool_node *node;
1270
1271 len = streamer_read_uhwi (ib);
1272 while (len)
1273 {
1274 VEC_safe_push (varpool_node_ptr, heap, varpool,
1275 input_varpool_node (file_data, ib));
1276 len--;
1277 }
1278 #ifdef ENABLE_CHECKING
1279 FOR_EACH_VEC_ELT (varpool_node_ptr, varpool, i, node)
1280 gcc_assert (!node->aux);
1281 #endif
1282 FOR_EACH_VEC_ELT (varpool_node_ptr, varpool, i, node)
1283 {
1284 int ref = (int) (intptr_t) node->same_comdat_group;
1285 /* We share declaration of builtins, so we may read same node twice. */
1286 if (node->aux)
1287 continue;
1288 node->aux = (void *)1;
1289
1290 /* Fixup same_comdat_group from reference to pointer. */
1291 if (ref != LCC_NOT_FOUND)
1292 node->same_comdat_group = VEC_index (varpool_node_ptr, varpool, ref);
1293 else
1294 node->same_comdat_group = NULL;
1295 }
1296 FOR_EACH_VEC_ELT (varpool_node_ptr, varpool, i, node)
1297 node->aux = NULL;
1298 return varpool;
1299 }
1300
1301 /* Input ipa_refs. */
1302
1303 static void
1304 input_refs (struct lto_input_block *ib,
1305 VEC(cgraph_node_ptr, heap) *nodes,
1306 VEC(varpool_node_ptr, heap) *varpool)
1307 {
1308 int count;
1309 int idx;
1310 while (true)
1311 {
1312 struct cgraph_node *node;
1313 count = streamer_read_uhwi (ib);
1314 if (!count)
1315 break;
1316 idx = streamer_read_uhwi (ib);
1317 node = VEC_index (cgraph_node_ptr, nodes, idx);
1318 while (count)
1319 {
1320 input_ref (ib, node, NULL, nodes, varpool);
1321 count--;
1322 }
1323 }
1324 while (true)
1325 {
1326 struct varpool_node *node;
1327 count = streamer_read_uhwi (ib);
1328 if (!count)
1329 break;
1330 node = VEC_index (varpool_node_ptr, varpool,
1331 streamer_read_uhwi (ib));
1332 while (count)
1333 {
1334 input_ref (ib, NULL, node, nodes, varpool);
1335 count--;
1336 }
1337 }
1338 }
1339
1340
1341 static struct gcov_ctr_summary lto_gcov_summary;
1342
1343 /* Input profile_info from IB. */
1344 static void
1345 input_profile_summary (struct lto_input_block *ib,
1346 struct lto_file_decl_data *file_data)
1347 {
1348 unsigned int runs = streamer_read_uhwi (ib);
1349 if (runs)
1350 {
1351 file_data->profile_info.runs = runs;
1352 file_data->profile_info.sum_max = streamer_read_uhwi (ib);
1353 }
1354
1355 }
1356
1357 /* Rescale profile summaries to the same number of runs in the whole unit. */
1358
1359 static void
1360 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1361 {
1362 struct lto_file_decl_data *file_data;
1363 unsigned int j;
1364 gcov_unsigned_t max_runs = 0;
1365 struct cgraph_node *node;
1366 struct cgraph_edge *edge;
1367
1368 /* Find unit with maximal number of runs. If we ever get serious about
1369 roundoff errors, we might also consider computing smallest common
1370 multiply. */
1371 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1372 if (max_runs < file_data->profile_info.runs)
1373 max_runs = file_data->profile_info.runs;
1374
1375 if (!max_runs)
1376 return;
1377
1378 /* Simple overflow check. We probably don't need to support that many train
1379 runs. Such a large value probably imply data corruption anyway. */
1380 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1381 {
1382 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1383 INT_MAX / REG_BR_PROB_BASE);
1384 return;
1385 }
1386
1387 profile_info = &lto_gcov_summary;
1388 lto_gcov_summary.runs = max_runs;
1389 lto_gcov_summary.sum_max = 0;
1390
1391 /* Rescale all units to the maximal number of runs.
1392 sum_max can not be easily merged, as we have no idea what files come from
1393 the same run. We do not use the info anyway, so leave it 0. */
1394 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1395 if (file_data->profile_info.runs)
1396 {
1397 int scale = ((REG_BR_PROB_BASE * max_runs
1398 + file_data->profile_info.runs / 2)
1399 / file_data->profile_info.runs);
1400 lto_gcov_summary.sum_max = MAX (lto_gcov_summary.sum_max,
1401 (file_data->profile_info.sum_max
1402 * scale
1403 + REG_BR_PROB_BASE / 2)
1404 / REG_BR_PROB_BASE);
1405 }
1406
1407 /* Watch roundoff errors. */
1408 if (lto_gcov_summary.sum_max < max_runs)
1409 lto_gcov_summary.sum_max = max_runs;
1410
1411 /* If merging already happent at WPA time, we are done. */
1412 if (flag_ltrans)
1413 return;
1414
1415 /* Now compute count_materialization_scale of each node.
1416 During LTRANS we already have values of count_materialization_scale
1417 computed, so just update them. */
1418 for (node = cgraph_nodes; node; node = node->next)
1419 if (node->local.lto_file_data
1420 && node->local.lto_file_data->profile_info.runs)
1421 {
1422 int scale;
1423
1424 scale =
1425 ((node->count_materialization_scale * max_runs
1426 + node->local.lto_file_data->profile_info.runs / 2)
1427 / node->local.lto_file_data->profile_info.runs);
1428 node->count_materialization_scale = scale;
1429 if (scale < 0)
1430 fatal_error ("Profile information in %s corrupted",
1431 file_data->file_name);
1432
1433 if (scale == REG_BR_PROB_BASE)
1434 continue;
1435 for (edge = node->callees; edge; edge = edge->next_callee)
1436 edge->count = ((edge->count * scale + REG_BR_PROB_BASE / 2)
1437 / REG_BR_PROB_BASE);
1438 node->count = ((node->count * scale + REG_BR_PROB_BASE / 2)
1439 / REG_BR_PROB_BASE);
1440 }
1441 }
1442
1443 /* Input and merge the cgraph from each of the .o files passed to
1444 lto1. */
1445
1446 void
1447 input_cgraph (void)
1448 {
1449 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1450 struct lto_file_decl_data *file_data;
1451 unsigned int j = 0;
1452 struct cgraph_node *node;
1453
1454 while ((file_data = file_data_vec[j++]))
1455 {
1456 const char *data;
1457 size_t len;
1458 struct lto_input_block *ib;
1459 VEC(cgraph_node_ptr, heap) *nodes;
1460 VEC(varpool_node_ptr, heap) *varpool;
1461
1462 ib = lto_create_simple_input_block (file_data, LTO_section_cgraph,
1463 &data, &len);
1464 if (!ib)
1465 fatal_error ("cannot find LTO cgraph in %s", file_data->file_name);
1466 input_profile_summary (ib, file_data);
1467 file_data->cgraph_node_encoder = lto_cgraph_encoder_new ();
1468 nodes = input_cgraph_1 (file_data, ib);
1469 lto_destroy_simple_input_block (file_data, LTO_section_cgraph,
1470 ib, data, len);
1471
1472 ib = lto_create_simple_input_block (file_data, LTO_section_varpool,
1473 &data, &len);
1474 if (!ib)
1475 fatal_error ("cannot find LTO varpool in %s", file_data->file_name);
1476 varpool = input_varpool_1 (file_data, ib);
1477 lto_destroy_simple_input_block (file_data, LTO_section_varpool,
1478 ib, data, len);
1479
1480 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1481 &data, &len);
1482 if (!ib)
1483 fatal_error("cannot find LTO section refs in %s", file_data->file_name);
1484 input_refs (ib, nodes, varpool);
1485 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1486 ib, data, len);
1487 if (flag_ltrans)
1488 input_cgraph_opt_summary (nodes);
1489 VEC_free (cgraph_node_ptr, heap, nodes);
1490 VEC_free (varpool_node_ptr, heap, varpool);
1491 }
1492
1493 merge_profile_summaries (file_data_vec);
1494
1495 /* Clear out the aux field that was used to store enough state to
1496 tell which nodes should be overwritten. */
1497 for (node = cgraph_nodes; node; node = node->next)
1498 {
1499 /* Some nodes may have been created by cgraph_node. This
1500 happens when the callgraph contains nested functions. If the
1501 node for the parent function was never emitted to the gimple
1502 file, cgraph_node will create a node for it when setting the
1503 context of the nested function. */
1504 if (node->local.lto_file_data)
1505 node->aux = NULL;
1506 }
1507 }
1508
1509 /* True when we need optimization summary for NODE. */
1510
1511 static int
1512 output_cgraph_opt_summary_p (struct cgraph_node *node,
1513 cgraph_node_set set ATTRIBUTE_UNUSED)
1514 {
1515 return (node->clone_of
1516 && (node->clone.tree_map
1517 || node->clone.args_to_skip
1518 || node->clone.combined_args_to_skip));
1519 }
1520
1521 /* Output optimization summary for EDGE to OB. */
1522 static void
1523 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1524 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1525 {
1526 }
1527
1528 /* Output optimization summary for NODE to OB. */
1529
1530 static void
1531 output_node_opt_summary (struct output_block *ob,
1532 struct cgraph_node *node,
1533 cgraph_node_set set)
1534 {
1535 unsigned int index;
1536 bitmap_iterator bi;
1537 struct ipa_replace_map *map;
1538 struct bitpack_d bp;
1539 int i;
1540 struct cgraph_edge *e;
1541
1542 if (node->clone.args_to_skip)
1543 {
1544 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1545 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1546 streamer_write_uhwi (ob, index);
1547 }
1548 else
1549 streamer_write_uhwi (ob, 0);
1550 if (node->clone.combined_args_to_skip)
1551 {
1552 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1553 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1554 streamer_write_uhwi (ob, index);
1555 }
1556 else
1557 streamer_write_uhwi (ob, 0);
1558 streamer_write_uhwi (ob, VEC_length (ipa_replace_map_p,
1559 node->clone.tree_map));
1560 FOR_EACH_VEC_ELT (ipa_replace_map_p, node->clone.tree_map, i, map)
1561 {
1562 int parm_num;
1563 tree parm;
1564
1565 for (parm_num = 0, parm = DECL_ARGUMENTS (node->decl); parm;
1566 parm = DECL_CHAIN (parm), parm_num++)
1567 if (map->old_tree == parm)
1568 break;
1569 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1570 mechanism to store function local declarations into summaries. */
1571 gcc_assert (parm);
1572 streamer_write_uhwi (ob, parm_num);
1573 stream_write_tree (ob, map->new_tree, true);
1574 bp = bitpack_create (ob->main_stream);
1575 bp_pack_value (&bp, map->replace_p, 1);
1576 bp_pack_value (&bp, map->ref_p, 1);
1577 streamer_write_bitpack (&bp);
1578 }
1579
1580 if (cgraph_node_in_set_p (node, set))
1581 {
1582 for (e = node->callees; e; e = e->next_callee)
1583 output_edge_opt_summary (ob, e);
1584 for (e = node->indirect_calls; e; e = e->next_callee)
1585 output_edge_opt_summary (ob, e);
1586 }
1587 }
1588
1589 /* Output optimization summaries stored in callgraph.
1590 At the moment it is the clone info structure. */
1591
1592 static void
1593 output_cgraph_opt_summary (cgraph_node_set set)
1594 {
1595 struct cgraph_node *node;
1596 int i, n_nodes;
1597 lto_cgraph_encoder_t encoder;
1598 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
1599 unsigned count = 0;
1600
1601 ob->cgraph_node = NULL;
1602 encoder = ob->decl_state->cgraph_node_encoder;
1603 n_nodes = lto_cgraph_encoder_size (encoder);
1604 for (i = 0; i < n_nodes; i++)
1605 if (output_cgraph_opt_summary_p (lto_cgraph_encoder_deref (encoder, i),
1606 set))
1607 count++;
1608 streamer_write_uhwi (ob, count);
1609 for (i = 0; i < n_nodes; i++)
1610 {
1611 node = lto_cgraph_encoder_deref (encoder, i);
1612 if (output_cgraph_opt_summary_p (node, set))
1613 {
1614 streamer_write_uhwi (ob, i);
1615 output_node_opt_summary (ob, node, set);
1616 }
1617 }
1618 produce_asm (ob, NULL);
1619 destroy_output_block (ob);
1620 }
1621
1622 /* Input optimisation summary of EDGE. */
1623
1624 static void
1625 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
1626 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
1627 {
1628 }
1629
1630 /* Input optimisation summary of NODE. */
1631
1632 static void
1633 input_node_opt_summary (struct cgraph_node *node,
1634 struct lto_input_block *ib_main,
1635 struct data_in *data_in)
1636 {
1637 int i;
1638 int count;
1639 int bit;
1640 struct bitpack_d bp;
1641 struct cgraph_edge *e;
1642
1643 count = streamer_read_uhwi (ib_main);
1644 if (count)
1645 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
1646 for (i = 0; i < count; i++)
1647 {
1648 bit = streamer_read_uhwi (ib_main);
1649 bitmap_set_bit (node->clone.args_to_skip, bit);
1650 }
1651 count = streamer_read_uhwi (ib_main);
1652 if (count)
1653 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
1654 for (i = 0; i < count; i++)
1655 {
1656 bit = streamer_read_uhwi (ib_main);
1657 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
1658 }
1659 count = streamer_read_uhwi (ib_main);
1660 for (i = 0; i < count; i++)
1661 {
1662 int parm_num;
1663 tree parm;
1664 struct ipa_replace_map *map = ggc_alloc_ipa_replace_map ();
1665
1666 VEC_safe_push (ipa_replace_map_p, gc, node->clone.tree_map, map);
1667 for (parm_num = 0, parm = DECL_ARGUMENTS (node->decl); parm_num;
1668 parm = DECL_CHAIN (parm))
1669 parm_num --;
1670 map->parm_num = streamer_read_uhwi (ib_main);
1671 map->old_tree = NULL;
1672 map->new_tree = stream_read_tree (ib_main, data_in);
1673 bp = streamer_read_bitpack (ib_main);
1674 map->replace_p = bp_unpack_value (&bp, 1);
1675 map->ref_p = bp_unpack_value (&bp, 1);
1676 }
1677 for (e = node->callees; e; e = e->next_callee)
1678 input_edge_opt_summary (e, ib_main);
1679 for (e = node->indirect_calls; e; e = e->next_callee)
1680 input_edge_opt_summary (e, ib_main);
1681 }
1682
1683 /* Read section in file FILE_DATA of length LEN with data DATA. */
1684
1685 static void
1686 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
1687 const char *data, size_t len, VEC (cgraph_node_ptr,
1688 heap) * nodes)
1689 {
1690 const struct lto_function_header *header =
1691 (const struct lto_function_header *) data;
1692 const int cfg_offset = sizeof (struct lto_function_header);
1693 const int main_offset = cfg_offset + header->cfg_size;
1694 const int string_offset = main_offset + header->main_size;
1695 struct data_in *data_in;
1696 struct lto_input_block ib_main;
1697 unsigned int i;
1698 unsigned int count;
1699
1700 LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0,
1701 header->main_size);
1702
1703 data_in =
1704 lto_data_in_create (file_data, (const char *) data + string_offset,
1705 header->string_size, NULL);
1706 count = streamer_read_uhwi (&ib_main);
1707
1708 for (i = 0; i < count; i++)
1709 {
1710 int ref = streamer_read_uhwi (&ib_main);
1711 input_node_opt_summary (VEC_index (cgraph_node_ptr, nodes, ref),
1712 &ib_main, data_in);
1713 }
1714 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
1715 len);
1716 lto_data_in_delete (data_in);
1717 }
1718
1719 /* Input optimization summary of cgraph. */
1720
1721 static void
1722 input_cgraph_opt_summary (VEC (cgraph_node_ptr, heap) * nodes)
1723 {
1724 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1725 struct lto_file_decl_data *file_data;
1726 unsigned int j = 0;
1727
1728 while ((file_data = file_data_vec[j++]))
1729 {
1730 size_t len;
1731 const char *data =
1732 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
1733 &len);
1734
1735 if (data)
1736 input_cgraph_opt_section (file_data, data, len, nodes);
1737 }
1738 }