[multiple changes]
[gcc.git] / gcc / lto-cgraph.c
1 /* Write and read the cgraph to the memory mapped representation of a
2 .o file.
3
4 Copyright (C) 2009-2013 Free Software Foundation, Inc.
5 Contributed by Kenneth Zadeck <zadeck@naturalbridge.com>
6
7 This file is part of GCC.
8
9 GCC is free software; you can redistribute it and/or modify it under
10 the terms of the GNU General Public License as published by the Free
11 Software Foundation; either version 3, or (at your option) any later
12 version.
13
14 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
15 WARRANTY; without even the implied warranty of MERCHANTABILITY or
16 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
22
23 #include "config.h"
24 #include "system.h"
25 #include "coretypes.h"
26 #include "tm.h"
27 #include "tree.h"
28 #include "expr.h"
29 #include "flags.h"
30 #include "params.h"
31 #include "input.h"
32 #include "hashtab.h"
33 #include "langhooks.h"
34 #include "basic-block.h"
35 #include "tree-flow.h"
36 #include "cgraph.h"
37 #include "function.h"
38 #include "ggc.h"
39 #include "diagnostic-core.h"
40 #include "except.h"
41 #include "vec.h"
42 #include "timevar.h"
43 #include "pointer-set.h"
44 #include "lto-streamer.h"
45 #include "data-streamer.h"
46 #include "tree-streamer.h"
47 #include "gcov-io.h"
48 #include "tree-pass.h"
49 #include "profile.h"
50
51 static void output_cgraph_opt_summary (void);
52 static void input_cgraph_opt_summary (vec<symtab_node> nodes);
53
54 /* Number of LDPR values known to GCC. */
55 #define LDPR_NUM_KNOWN (LDPR_PREVAILING_DEF_IRONLY_EXP + 1)
56
57 /* All node orders are ofsetted by ORDER_BASE. */
58 static int order_base;
59
60 /* Cgraph streaming is organized as set of record whose type
61 is indicated by a tag. */
62 enum LTO_symtab_tags
63 {
64 /* Must leave 0 for the stopper. */
65
66 /* Cgraph node without body available. */
67 LTO_symtab_unavail_node = 1,
68 /* Cgraph node with function body. */
69 LTO_symtab_analyzed_node,
70 /* Cgraph edges. */
71 LTO_symtab_edge,
72 LTO_symtab_indirect_edge,
73 LTO_symtab_variable,
74 LTO_symtab_last_tag
75 };
76
77 /* Create a new symtab encoder.
78 if FOR_INPUT, the encoder allocate only datastructures needed
79 to read the symtab. */
80
81 lto_symtab_encoder_t
82 lto_symtab_encoder_new (bool for_input)
83 {
84 lto_symtab_encoder_t encoder = XCNEW (struct lto_symtab_encoder_d);
85
86 if (!for_input)
87 encoder->map = pointer_map_create ();
88 encoder->nodes.create (0);
89 return encoder;
90 }
91
92
93 /* Delete ENCODER and its components. */
94
95 void
96 lto_symtab_encoder_delete (lto_symtab_encoder_t encoder)
97 {
98 encoder->nodes.release ();
99 if (encoder->map)
100 pointer_map_destroy (encoder->map);
101 free (encoder);
102 }
103
104
105 /* Return the existing reference number of NODE in the symtab encoder in
106 output block OB. Assign a new reference if this is the first time
107 NODE is encoded. */
108
109 int
110 lto_symtab_encoder_encode (lto_symtab_encoder_t encoder,
111 symtab_node node)
112 {
113 int ref;
114 void **slot;
115
116 if (!encoder->map)
117 {
118 lto_encoder_entry entry = {node, false, false, false};
119
120 ref = encoder->nodes.length ();
121 encoder->nodes.safe_push (entry);
122 return ref;
123 }
124
125 slot = pointer_map_contains (encoder->map, node);
126 if (!slot || !*slot)
127 {
128 lto_encoder_entry entry = {node, false, false, false};
129 ref = encoder->nodes.length ();
130 if (!slot)
131 slot = pointer_map_insert (encoder->map, node);
132 *slot = (void *) (intptr_t) (ref + 1);
133 encoder->nodes.safe_push (entry);
134 }
135 else
136 ref = (size_t) *slot - 1;
137
138 return ref;
139 }
140
141 /* Remove NODE from encoder. */
142
143 bool
144 lto_symtab_encoder_delete_node (lto_symtab_encoder_t encoder,
145 symtab_node node)
146 {
147 void **slot, **last_slot;
148 int index;
149 lto_encoder_entry last_node;
150
151 slot = pointer_map_contains (encoder->map, node);
152 if (slot == NULL || !*slot)
153 return false;
154
155 index = (size_t) *slot - 1;
156 gcc_checking_assert (encoder->nodes[index].node == node);
157
158 /* Remove from vector. We do this by swapping node with the last element
159 of the vector. */
160 last_node = encoder->nodes.pop ();
161 if (last_node.node != node)
162 {
163 last_slot = pointer_map_contains (encoder->map, last_node.node);
164 gcc_checking_assert (last_slot && *last_slot);
165 *last_slot = (void *)(size_t) (index + 1);
166
167 /* Move the last element to the original spot of NODE. */
168 encoder->nodes[index] = last_node;
169 }
170
171 /* Remove element from hash table. */
172 *slot = NULL;
173 return true;
174 }
175
176
177 /* Return TRUE if we should encode initializer of NODE (if any). */
178
179 bool
180 lto_symtab_encoder_encode_body_p (lto_symtab_encoder_t encoder,
181 struct cgraph_node *node)
182 {
183 int index = lto_symtab_encoder_lookup (encoder, (symtab_node)node);
184 return encoder->nodes[index].body;
185 }
186
187 /* Return TRUE if we should encode body of NODE (if any). */
188
189 static void
190 lto_set_symtab_encoder_encode_body (lto_symtab_encoder_t encoder,
191 struct cgraph_node *node)
192 {
193 int index = lto_symtab_encoder_encode (encoder, (symtab_node)node);
194 gcc_checking_assert (encoder->nodes[index].node == (symtab_node)node);
195 encoder->nodes[index].body = true;
196 }
197
198 /* Return TRUE if we should encode initializer of NODE (if any). */
199
200 bool
201 lto_symtab_encoder_encode_initializer_p (lto_symtab_encoder_t encoder,
202 struct varpool_node *node)
203 {
204 int index = lto_symtab_encoder_lookup (encoder, (symtab_node)node);
205 if (index == LCC_NOT_FOUND)
206 return false;
207 return encoder->nodes[index].initializer;
208 }
209
210 /* Return TRUE if we should encode initializer of NODE (if any). */
211
212 static void
213 lto_set_symtab_encoder_encode_initializer (lto_symtab_encoder_t encoder,
214 struct varpool_node *node)
215 {
216 int index = lto_symtab_encoder_lookup (encoder, (symtab_node)node);
217 encoder->nodes[index].initializer = true;
218 }
219
220 /* Return TRUE if we should encode initializer of NODE (if any). */
221
222 bool
223 lto_symtab_encoder_in_partition_p (lto_symtab_encoder_t encoder,
224 symtab_node node)
225 {
226 int index = lto_symtab_encoder_lookup (encoder, (symtab_node)node);
227 if (index == LCC_NOT_FOUND)
228 return false;
229 return encoder->nodes[index].in_partition;
230 }
231
232 /* Return TRUE if we should encode body of NODE (if any). */
233
234 void
235 lto_set_symtab_encoder_in_partition (lto_symtab_encoder_t encoder,
236 symtab_node node)
237 {
238 int index = lto_symtab_encoder_encode (encoder, (symtab_node)node);
239 encoder->nodes[index].in_partition = true;
240 }
241
242 /* Output the cgraph EDGE to OB using ENCODER. */
243
244 static void
245 lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge,
246 lto_symtab_encoder_t encoder)
247 {
248 unsigned int uid;
249 intptr_t ref;
250 struct bitpack_d bp;
251
252 if (edge->indirect_unknown_callee)
253 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
254 LTO_symtab_indirect_edge);
255 else
256 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
257 LTO_symtab_edge);
258
259 ref = lto_symtab_encoder_lookup (encoder, (symtab_node)edge->caller);
260 gcc_assert (ref != LCC_NOT_FOUND);
261 streamer_write_hwi_stream (ob->main_stream, ref);
262
263 if (!edge->indirect_unknown_callee)
264 {
265 ref = lto_symtab_encoder_lookup (encoder, (symtab_node)edge->callee);
266 gcc_assert (ref != LCC_NOT_FOUND);
267 streamer_write_hwi_stream (ob->main_stream, ref);
268 }
269
270 streamer_write_gcov_count_stream (ob->main_stream, edge->count);
271
272 bp = bitpack_create (ob->main_stream);
273 uid = (!gimple_has_body_p (edge->caller->symbol.decl)
274 ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt));
275 bp_pack_enum (&bp, cgraph_inline_failed_enum,
276 CIF_N_REASONS, edge->inline_failed);
277 bp_pack_var_len_unsigned (&bp, uid);
278 bp_pack_var_len_unsigned (&bp, edge->frequency);
279 bp_pack_value (&bp, edge->indirect_inlining_edge, 1);
280 bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1);
281 bp_pack_value (&bp, edge->can_throw_external, 1);
282 if (edge->indirect_unknown_callee)
283 {
284 int flags = edge->indirect_info->ecf_flags;
285 bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1);
286 bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1);
287 bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1);
288 bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1);
289 bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1);
290 bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1);
291 /* Flags that should not appear on indirect calls. */
292 gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE
293 | ECF_MAY_BE_ALLOCA
294 | ECF_SIBCALL
295 | ECF_LEAF
296 | ECF_NOVOPS)));
297 }
298 streamer_write_bitpack (&bp);
299 }
300
301 /* Return if LIST contain references from other partitions. */
302
303 bool
304 referenced_from_other_partition_p (struct ipa_ref_list *list, lto_symtab_encoder_t encoder)
305 {
306 int i;
307 struct ipa_ref *ref;
308 for (i = 0; ipa_ref_list_referring_iterate (list, i, ref); i++)
309 {
310 if (ref->referring->symbol.in_other_partition
311 || !lto_symtab_encoder_in_partition_p (encoder, ref->referring))
312 return true;
313 }
314 return false;
315 }
316
317 /* Return true when node is reachable from other partition. */
318
319 bool
320 reachable_from_other_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
321 {
322 struct cgraph_edge *e;
323 if (!node->analyzed)
324 return false;
325 if (node->global.inlined_to)
326 return false;
327 for (e = node->callers; e; e = e->next_caller)
328 if (e->caller->symbol.in_other_partition
329 || !lto_symtab_encoder_in_partition_p (encoder, (symtab_node)e->caller))
330 return true;
331 return false;
332 }
333
334 /* Return if LIST contain references from other partitions. */
335
336 bool
337 referenced_from_this_partition_p (struct ipa_ref_list *list,
338 lto_symtab_encoder_t encoder)
339 {
340 int i;
341 struct ipa_ref *ref;
342 for (i = 0; ipa_ref_list_referring_iterate (list, i, ref); i++)
343 if (lto_symtab_encoder_in_partition_p (encoder, ref->referring))
344 return true;
345 return false;
346 }
347
348 /* Return true when node is reachable from other partition. */
349
350 bool
351 reachable_from_this_partition_p (struct cgraph_node *node, lto_symtab_encoder_t encoder)
352 {
353 struct cgraph_edge *e;
354 for (e = node->callers; e; e = e->next_caller)
355 if (lto_symtab_encoder_in_partition_p (encoder, (symtab_node)e->caller))
356 return true;
357 return false;
358 }
359
360 /* Output the cgraph NODE to OB. ENCODER is used to find the
361 reference number of NODE->inlined_to. SET is the set of nodes we
362 are writing to the current file. If NODE is not in SET, then NODE
363 is a boundary of a cgraph_node_set and we pretend NODE just has a
364 decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs
365 that have had their callgraph node written so far. This is used to
366 determine if NODE is a clone of a previously written node. */
367
368 static void
369 lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node,
370 lto_symtab_encoder_t encoder)
371 {
372 unsigned int tag;
373 struct bitpack_d bp;
374 bool boundary_p;
375 intptr_t ref;
376 bool in_other_partition = false;
377 struct cgraph_node *clone_of;
378 struct ipa_opt_pass_d *pass;
379 int i;
380
381 boundary_p = !lto_symtab_encoder_in_partition_p (encoder, (symtab_node)node);
382
383 if (node->analyzed && !boundary_p)
384 tag = LTO_symtab_analyzed_node;
385 else
386 tag = LTO_symtab_unavail_node;
387
388 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
389 tag);
390 streamer_write_hwi_stream (ob->main_stream, node->symbol.order);
391
392 /* In WPA mode, we only output part of the call-graph. Also, we
393 fake cgraph node attributes. There are two cases that we care.
394
395 Boundary nodes: There are nodes that are not part of SET but are
396 called from within SET. We artificially make them look like
397 externally visible nodes with no function body.
398
399 Cherry-picked nodes: These are nodes we pulled from other
400 translation units into SET during IPA-inlining. We make them as
401 local static nodes to prevent clashes with other local statics. */
402 if (boundary_p && node->analyzed && !DECL_EXTERNAL (node->symbol.decl))
403 {
404 /* Inline clones can not be part of boundary.
405 gcc_assert (!node->global.inlined_to);
406
407 FIXME: At the moment they can be, when partition contains an inline
408 clone that is clone of inline clone from outside partition. We can
409 reshape the clone tree and make other tree to be the root, but it
410 needs a bit extra work and will be promplty done by cgraph_remove_node
411 after reading back. */
412 in_other_partition = 1;
413 }
414
415 clone_of = node->clone_of;
416 while (clone_of
417 && (ref = lto_symtab_encoder_lookup (encoder, (symtab_node)clone_of)) == LCC_NOT_FOUND)
418 if (clone_of->prev_sibling_clone)
419 clone_of = clone_of->prev_sibling_clone;
420 else
421 clone_of = clone_of->clone_of;
422
423 if (LTO_symtab_analyzed_node)
424 gcc_assert (clone_of || !node->clone_of);
425 if (!clone_of)
426 streamer_write_hwi_stream (ob->main_stream, LCC_NOT_FOUND);
427 else
428 streamer_write_hwi_stream (ob->main_stream, ref);
429
430
431 lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->symbol.decl);
432 streamer_write_gcov_count_stream (ob->main_stream, node->count);
433 streamer_write_hwi_stream (ob->main_stream, node->count_materialization_scale);
434
435 streamer_write_hwi_stream (ob->main_stream,
436 node->ipa_transforms_to_apply.length ());
437 FOR_EACH_VEC_ELT (node->ipa_transforms_to_apply, i, pass)
438 streamer_write_hwi_stream (ob->main_stream, pass->pass.static_pass_number);
439
440 if (tag == LTO_symtab_analyzed_node)
441 {
442 if (node->global.inlined_to)
443 {
444 ref = lto_symtab_encoder_lookup (encoder, (symtab_node)node->global.inlined_to);
445 gcc_assert (ref != LCC_NOT_FOUND);
446 }
447 else
448 ref = LCC_NOT_FOUND;
449
450 streamer_write_hwi_stream (ob->main_stream, ref);
451 }
452
453 if (node->symbol.same_comdat_group && !boundary_p)
454 {
455 ref = lto_symtab_encoder_lookup (encoder,
456 node->symbol.same_comdat_group);
457 gcc_assert (ref != LCC_NOT_FOUND);
458 }
459 else
460 ref = LCC_NOT_FOUND;
461 streamer_write_hwi_stream (ob->main_stream, ref);
462
463 bp = bitpack_create (ob->main_stream);
464 bp_pack_value (&bp, node->local.local, 1);
465 bp_pack_value (&bp, node->symbol.externally_visible, 1);
466 bp_pack_value (&bp, node->local.finalized, 1);
467 bp_pack_value (&bp, node->local.versionable, 1);
468 bp_pack_value (&bp, node->local.can_change_signature, 1);
469 bp_pack_value (&bp, node->local.redefined_extern_inline, 1);
470 bp_pack_value (&bp, node->symbol.force_output, 1);
471 bp_pack_value (&bp, node->symbol.address_taken, 1);
472 bp_pack_value (&bp, node->abstract_and_needed, 1);
473 bp_pack_value (&bp, tag == LTO_symtab_analyzed_node
474 && !DECL_EXTERNAL (node->symbol.decl)
475 && !DECL_COMDAT (node->symbol.decl)
476 && (reachable_from_other_partition_p (node, encoder)
477 || referenced_from_other_partition_p (&node->symbol.ref_list,
478 encoder)), 1);
479 bp_pack_value (&bp, node->lowered, 1);
480 bp_pack_value (&bp, in_other_partition, 1);
481 /* Real aliases in a boundary become non-aliases. However we still stream
482 alias info on weakrefs.
483 TODO: We lose a bit of information here - when we know that variable is
484 defined in other unit, we may use the info on aliases to resolve
485 symbol1 != symbol2 type tests that we can do only for locally defined objects
486 otherwise. */
487 bp_pack_value (&bp, node->alias && (!boundary_p || DECL_EXTERNAL (node->symbol.decl)), 1);
488 bp_pack_value (&bp, node->frequency, 2);
489 bp_pack_value (&bp, node->only_called_at_startup, 1);
490 bp_pack_value (&bp, node->only_called_at_exit, 1);
491 bp_pack_value (&bp, node->tm_clone, 1);
492 bp_pack_value (&bp, node->thunk.thunk_p && !boundary_p, 1);
493 bp_pack_enum (&bp, ld_plugin_symbol_resolution,
494 LDPR_NUM_KNOWN, node->symbol.resolution);
495 streamer_write_bitpack (&bp);
496
497 if (node->thunk.thunk_p && !boundary_p)
498 {
499 streamer_write_uhwi_stream
500 (ob->main_stream,
501 1 + (node->thunk.this_adjusting != 0) * 2
502 + (node->thunk.virtual_offset_p != 0) * 4);
503 streamer_write_uhwi_stream (ob->main_stream, node->thunk.fixed_offset);
504 streamer_write_uhwi_stream (ob->main_stream, node->thunk.virtual_value);
505 }
506 if ((node->alias || node->thunk.thunk_p)
507 && (!boundary_p || (node->alias && DECL_EXTERNAL (node->symbol.decl))))
508 {
509 streamer_write_hwi_in_range (ob->main_stream, 0, 1,
510 node->thunk.alias != NULL);
511 if (node->thunk.alias != NULL)
512 lto_output_fn_decl_index (ob->decl_state, ob->main_stream,
513 node->thunk.alias);
514 }
515 }
516
517 /* Output the varpool NODE to OB.
518 If NODE is not in SET, then NODE is a boundary. */
519
520 static void
521 lto_output_varpool_node (struct lto_simple_output_block *ob, struct varpool_node *node,
522 lto_symtab_encoder_t encoder)
523 {
524 bool boundary_p = (node->analyzed
525 && !lto_symtab_encoder_in_partition_p (encoder, (symtab_node)node));
526 struct bitpack_d bp;
527 int ref;
528
529 streamer_write_enum (ob->main_stream, LTO_symtab_tags, LTO_symtab_last_tag,
530 LTO_symtab_variable);
531 streamer_write_hwi_stream (ob->main_stream, node->symbol.order);
532 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->symbol.decl);
533 bp = bitpack_create (ob->main_stream);
534 bp_pack_value (&bp, node->symbol.externally_visible, 1);
535 bp_pack_value (&bp, node->symbol.force_output, 1);
536 bp_pack_value (&bp, node->finalized, 1);
537 bp_pack_value (&bp, node->alias, 1);
538 bp_pack_value (&bp, node->alias_of != NULL, 1);
539 gcc_assert (node->finalized || !node->analyzed);
540 /* Constant pool initializers can be de-unified into individual ltrans units.
541 FIXME: Alternatively at -Os we may want to avoid generating for them the local
542 labels and share them across LTRANS partitions. */
543 if (DECL_IN_CONSTANT_POOL (node->symbol.decl)
544 && !DECL_EXTERNAL (node->symbol.decl)
545 && !DECL_COMDAT (node->symbol.decl))
546 {
547 bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */
548 bp_pack_value (&bp, 0, 1); /* in_other_partition. */
549 }
550 else
551 {
552 bp_pack_value (&bp, node->analyzed
553 && referenced_from_other_partition_p (&node->symbol.ref_list,
554 encoder), 1);
555 bp_pack_value (&bp, boundary_p && !DECL_EXTERNAL (node->symbol.decl), 1);
556 /* in_other_partition. */
557 }
558 streamer_write_bitpack (&bp);
559 if (node->alias_of)
560 lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->alias_of);
561 if (node->symbol.same_comdat_group && !boundary_p)
562 {
563 ref = lto_symtab_encoder_lookup (encoder,
564 node->symbol.same_comdat_group);
565 gcc_assert (ref != LCC_NOT_FOUND);
566 }
567 else
568 ref = LCC_NOT_FOUND;
569 streamer_write_hwi_stream (ob->main_stream, ref);
570 streamer_write_enum (ob->main_stream, ld_plugin_symbol_resolution,
571 LDPR_NUM_KNOWN, node->symbol.resolution);
572 }
573
574 /* Output the varpool NODE to OB.
575 If NODE is not in SET, then NODE is a boundary. */
576
577 static void
578 lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref,
579 lto_symtab_encoder_t encoder)
580 {
581 struct bitpack_d bp;
582 int nref;
583
584 bp = bitpack_create (ob->main_stream);
585 bp_pack_value (&bp, ref->use, 2);
586 streamer_write_bitpack (&bp);
587 nref = lto_symtab_encoder_lookup (encoder, ref->referred);
588 gcc_assert (nref != LCC_NOT_FOUND);
589 streamer_write_hwi_stream (ob->main_stream, nref);
590 }
591
592 /* Stream out profile_summary to OB. */
593
594 static void
595 output_profile_summary (struct lto_simple_output_block *ob)
596 {
597 unsigned h_ix;
598 struct bitpack_d bp;
599
600 if (profile_info)
601 {
602 /* We do not output num and run_max, they are not used by
603 GCC profile feedback and they are difficult to merge from multiple
604 units. */
605 gcc_assert (profile_info->runs);
606 streamer_write_uhwi_stream (ob->main_stream, profile_info->runs);
607 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_max);
608
609 /* sum_all is needed for computing the working set with the
610 histogram. */
611 streamer_write_gcov_count_stream (ob->main_stream, profile_info->sum_all);
612
613 /* Create and output a bitpack of non-zero histogram entries indices. */
614 bp = bitpack_create (ob->main_stream);
615 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
616 bp_pack_value (&bp, profile_info->histogram[h_ix].num_counters > 0, 1);
617 streamer_write_bitpack (&bp);
618 /* Now stream out only those non-zero entries. */
619 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
620 {
621 if (!profile_info->histogram[h_ix].num_counters)
622 continue;
623 streamer_write_gcov_count_stream (ob->main_stream,
624 profile_info->histogram[h_ix].num_counters);
625 streamer_write_gcov_count_stream (ob->main_stream,
626 profile_info->histogram[h_ix].min_value);
627 streamer_write_gcov_count_stream (ob->main_stream,
628 profile_info->histogram[h_ix].cum_value);
629 }
630 /* IPA-profile computes hot bb threshold based on cumulated
631 whole program profile. We need to stream it down to ltrans. */
632 if (flag_wpa)
633 streamer_write_gcov_count_stream (ob->main_stream,
634 get_hot_bb_threshold ());
635 }
636 else
637 streamer_write_uhwi_stream (ob->main_stream, 0);
638 }
639
640 /* Output all callees or indirect outgoing edges. EDGE must be the first such
641 edge. */
642
643 static void
644 output_outgoing_cgraph_edges (struct cgraph_edge *edge,
645 struct lto_simple_output_block *ob,
646 lto_symtab_encoder_t encoder)
647 {
648 if (!edge)
649 return;
650
651 /* Output edges in backward direction, so the reconstructed callgraph match
652 and it is easy to associate call sites in the IPA pass summaries. */
653 while (edge->next_callee)
654 edge = edge->next_callee;
655 for (; edge; edge = edge->prev_callee)
656 lto_output_edge (ob, edge, encoder);
657 }
658
659 /* Output the part of the cgraph in SET. */
660
661 static void
662 output_refs (lto_symtab_encoder_t encoder)
663 {
664 lto_symtab_encoder_iterator lsei;
665 struct lto_simple_output_block *ob;
666 int count;
667 struct ipa_ref *ref;
668 int i;
669
670 ob = lto_create_simple_output_block (LTO_section_refs);
671
672 for (lsei = lsei_start_in_partition (encoder); !lsei_end_p (lsei);
673 lsei_next_in_partition (&lsei))
674 {
675 symtab_node node = lsei_node (lsei);
676
677 count = ipa_ref_list_nreferences (&node->symbol.ref_list);
678 if (count)
679 {
680 streamer_write_uhwi_stream (ob->main_stream, count);
681 streamer_write_uhwi_stream (ob->main_stream,
682 lto_symtab_encoder_lookup (encoder, node));
683 for (i = 0; ipa_ref_list_reference_iterate (&node->symbol.ref_list,
684 i, ref); i++)
685 lto_output_ref (ob, ref, encoder);
686 }
687 }
688
689 streamer_write_uhwi_stream (ob->main_stream, 0);
690
691 lto_destroy_simple_output_block (ob);
692 }
693
694 /* Add NODE into encoder as well as nodes it is cloned from.
695 Do it in a way so clones appear first. */
696
697 static void
698 add_node_to (lto_symtab_encoder_t encoder, struct cgraph_node *node,
699 bool include_body)
700 {
701 if (node->clone_of)
702 add_node_to (encoder, node->clone_of, include_body);
703 else if (include_body)
704 lto_set_symtab_encoder_encode_body (encoder, node);
705 lto_symtab_encoder_encode (encoder, (symtab_node)node);
706 }
707
708 /* Add all references in LIST to encoders. */
709
710 static void
711 add_references (lto_symtab_encoder_t encoder,
712 struct ipa_ref_list *list)
713 {
714 int i;
715 struct ipa_ref *ref;
716 for (i = 0; ipa_ref_list_reference_iterate (list, i, ref); i++)
717 if (is_a <cgraph_node> (ref->referred))
718 add_node_to (encoder, ipa_ref_node (ref), false);
719 else
720 lto_symtab_encoder_encode (encoder, ref->referred);
721 }
722
723 /* Find all symbols we want to stream into given partition and insert them
724 to encoders.
725
726 The function actually replaces IN_ENCODER by new one. The reason is that
727 streaming code needs clone's origin to be streamed before clone. This
728 means that we need to insert the nodes in specific order. This order is
729 ignored by the partitioning logic earlier. */
730
731 lto_symtab_encoder_t
732 compute_ltrans_boundary (lto_symtab_encoder_t in_encoder)
733 {
734 struct cgraph_node *node;
735 struct cgraph_edge *edge;
736 int i;
737 lto_symtab_encoder_t encoder;
738 lto_symtab_encoder_iterator lsei;
739
740 encoder = lto_symtab_encoder_new (false);
741
742 /* Go over all entries in the IN_ENCODER and duplicate them to
743 ENCODER. At the same time insert masters of clones so
744 every master appears before clone. */
745 for (lsei = lsei_start_function_in_partition (in_encoder);
746 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
747 {
748 node = lsei_cgraph_node (lsei);
749 add_node_to (encoder, node, true);
750 lto_set_symtab_encoder_in_partition (encoder, (symtab_node)node);
751 add_references (encoder, &node->symbol.ref_list);
752 }
753 for (lsei = lsei_start_variable_in_partition (in_encoder);
754 !lsei_end_p (lsei); lsei_next_variable_in_partition (&lsei))
755 {
756 struct varpool_node *vnode = lsei_varpool_node (lsei);
757 gcc_assert (!vnode->alias || vnode->alias_of);
758 lto_set_symtab_encoder_in_partition (encoder, (symtab_node)vnode);
759 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
760 add_references (encoder, &vnode->symbol.ref_list);
761 }
762 /* Pickle in also the initializer of all referenced readonly variables
763 to help folding. Constant pool variables are not shared, so we must
764 pickle those too. */
765 for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
766 {
767 symtab_node node = lto_symtab_encoder_deref (encoder, i);
768 if (varpool_node *vnode = dyn_cast <varpool_node> (node))
769 {
770 if (DECL_INITIAL (vnode->symbol.decl)
771 && !lto_symtab_encoder_encode_initializer_p (encoder,
772 vnode)
773 && const_value_known_p (vnode->symbol.decl))
774 {
775 lto_set_symtab_encoder_encode_initializer (encoder, vnode);
776 add_references (encoder, &vnode->symbol.ref_list);
777 }
778 }
779 }
780
781 /* Go over all the nodes again to include callees that are not in
782 SET. */
783 for (lsei = lsei_start_function_in_partition (encoder);
784 !lsei_end_p (lsei); lsei_next_function_in_partition (&lsei))
785 {
786 node = lsei_cgraph_node (lsei);
787 for (edge = node->callees; edge; edge = edge->next_callee)
788 {
789 struct cgraph_node *callee = edge->callee;
790 if (!lto_symtab_encoder_in_partition_p (encoder, (symtab_node)callee))
791 {
792 /* We should have moved all the inlines. */
793 gcc_assert (!callee->global.inlined_to);
794 add_node_to (encoder, callee, false);
795 }
796 }
797 }
798 lto_symtab_encoder_delete (in_encoder);
799 return encoder;
800 }
801
802 /* Output the part of the symtab in SET and VSET. */
803
804 void
805 output_symtab (void)
806 {
807 struct cgraph_node *node;
808 struct lto_simple_output_block *ob;
809 lto_symtab_encoder_iterator lsei;
810 int i, n_nodes;
811 lto_symtab_encoder_t encoder;
812 static bool asm_nodes_output = false;
813
814 if (flag_wpa)
815 output_cgraph_opt_summary ();
816
817 ob = lto_create_simple_output_block (LTO_section_symtab_nodes);
818
819 output_profile_summary (ob);
820
821 /* An encoder for cgraph nodes should have been created by
822 ipa_write_summaries_1. */
823 gcc_assert (ob->decl_state->symtab_node_encoder);
824 encoder = ob->decl_state->symtab_node_encoder;
825
826 /* Write out the nodes. We must first output a node and then its clones,
827 otherwise at a time reading back the node there would be nothing to clone
828 from. */
829 n_nodes = lto_symtab_encoder_size (encoder);
830 for (i = 0; i < n_nodes; i++)
831 {
832 symtab_node node = lto_symtab_encoder_deref (encoder, i);
833 if (cgraph_node *cnode = dyn_cast <cgraph_node> (node))
834 lto_output_node (ob, cnode, encoder);
835 else
836 lto_output_varpool_node (ob, varpool (node), encoder);
837
838 }
839
840 /* Go over the nodes in SET again to write edges. */
841 for (lsei = lsei_start_function_in_partition (encoder); !lsei_end_p (lsei);
842 lsei_next_function_in_partition (&lsei))
843 {
844 node = lsei_cgraph_node (lsei);
845 output_outgoing_cgraph_edges (node->callees, ob, encoder);
846 output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder);
847 }
848
849 streamer_write_uhwi_stream (ob->main_stream, 0);
850
851 lto_destroy_simple_output_block (ob);
852
853 /* Emit toplevel asms.
854 When doing WPA we must output every asm just once. Since we do not partition asm
855 nodes at all, output them to first output. This is kind of hack, but should work
856 well. */
857 if (!asm_nodes_output)
858 {
859 asm_nodes_output = true;
860 lto_output_toplevel_asms ();
861 }
862
863 output_refs (encoder);
864 }
865
866 /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS,
867 STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize
868 NODE or to replace the values in it, for instance because the first
869 time we saw it, the function body was not available but now it
870 is. BP is a bitpack with all the bitflags for NODE read from the
871 stream. */
872
873 static void
874 input_overwrite_node (struct lto_file_decl_data *file_data,
875 struct cgraph_node *node,
876 enum LTO_symtab_tags tag,
877 struct bitpack_d *bp)
878 {
879 node->symbol.aux = (void *) tag;
880 node->symbol.lto_file_data = file_data;
881
882 node->local.local = bp_unpack_value (bp, 1);
883 node->symbol.externally_visible = bp_unpack_value (bp, 1);
884 node->local.finalized = bp_unpack_value (bp, 1);
885 node->local.versionable = bp_unpack_value (bp, 1);
886 node->local.can_change_signature = bp_unpack_value (bp, 1);
887 node->local.redefined_extern_inline = bp_unpack_value (bp, 1);
888 node->symbol.force_output = bp_unpack_value (bp, 1);
889 node->symbol.address_taken = bp_unpack_value (bp, 1);
890 node->abstract_and_needed = bp_unpack_value (bp, 1);
891 node->symbol.used_from_other_partition = bp_unpack_value (bp, 1);
892 node->lowered = bp_unpack_value (bp, 1);
893 node->analyzed = tag == LTO_symtab_analyzed_node;
894 node->symbol.in_other_partition = bp_unpack_value (bp, 1);
895 if (node->symbol.in_other_partition
896 /* Avoid updating decl when we are seeing just inline clone.
897 When inlining function that has functions already inlined into it,
898 we produce clones of inline clones.
899
900 WPA partitioning might put each clone into different unit and
901 we might end up streaming inline clone from other partition
902 to support clone we are interested in. */
903 && (!node->clone_of
904 || node->clone_of->symbol.decl != node->symbol.decl))
905 {
906 DECL_EXTERNAL (node->symbol.decl) = 1;
907 TREE_STATIC (node->symbol.decl) = 0;
908 }
909 node->alias = bp_unpack_value (bp, 1);
910 node->frequency = (enum node_frequency)bp_unpack_value (bp, 2);
911 node->only_called_at_startup = bp_unpack_value (bp, 1);
912 node->only_called_at_exit = bp_unpack_value (bp, 1);
913 node->tm_clone = bp_unpack_value (bp, 1);
914 node->thunk.thunk_p = bp_unpack_value (bp, 1);
915 node->symbol.resolution = bp_unpack_enum (bp, ld_plugin_symbol_resolution,
916 LDPR_NUM_KNOWN);
917 }
918
919 /* Read a node from input_block IB. TAG is the node's tag just read.
920 Return the node read or overwriten. */
921
922 static struct cgraph_node *
923 input_node (struct lto_file_decl_data *file_data,
924 struct lto_input_block *ib,
925 enum LTO_symtab_tags tag,
926 vec<symtab_node> nodes)
927 {
928 tree fn_decl;
929 struct cgraph_node *node;
930 struct bitpack_d bp;
931 unsigned decl_index;
932 int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND;
933 int clone_ref;
934 int order;
935 int i, count;
936
937 order = streamer_read_hwi (ib) + order_base;
938 clone_ref = streamer_read_hwi (ib);
939
940 decl_index = streamer_read_uhwi (ib);
941 fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index);
942
943 if (clone_ref != LCC_NOT_FOUND)
944 {
945 node = cgraph_clone_node (cgraph (nodes[clone_ref]), fn_decl,
946 0, CGRAPH_FREQ_BASE, false,
947 vNULL, false);
948 }
949 else
950 node = cgraph_get_create_node (fn_decl);
951
952 node->symbol.order = order;
953 if (order >= symtab_order)
954 symtab_order = order + 1;
955
956 node->count = streamer_read_gcov_count (ib);
957 node->count_materialization_scale = streamer_read_hwi (ib);
958
959 count = streamer_read_hwi (ib);
960 node->ipa_transforms_to_apply = vNULL;
961 for (i = 0; i < count; i++)
962 {
963 struct opt_pass *pass;
964 int pid = streamer_read_hwi (ib);
965
966 gcc_assert (pid < passes_by_id_size);
967 pass = passes_by_id[pid];
968 node->ipa_transforms_to_apply.safe_push ((struct ipa_opt_pass_d *) pass);
969 }
970
971 if (tag == LTO_symtab_analyzed_node)
972 ref = streamer_read_hwi (ib);
973
974 ref2 = streamer_read_hwi (ib);
975
976 /* Make sure that we have not read this node before. Nodes that
977 have already been read will have their tag stored in the 'aux'
978 field. Since built-in functions can be referenced in multiple
979 functions, they are expected to be read more than once. */
980 if (node->symbol.aux && !DECL_BUILT_IN (node->symbol.decl))
981 internal_error ("bytecode stream: found multiple instances of cgraph "
982 "node %d", node->uid);
983
984 bp = streamer_read_bitpack (ib);
985 input_overwrite_node (file_data, node, tag, &bp);
986
987 /* Store a reference for now, and fix up later to be a pointer. */
988 node->global.inlined_to = (cgraph_node_ptr) (intptr_t) ref;
989
990 /* Store a reference for now, and fix up later to be a pointer. */
991 node->symbol.same_comdat_group = (symtab_node) (intptr_t) ref2;
992
993 if (node->thunk.thunk_p)
994 {
995 int type = streamer_read_uhwi (ib);
996 HOST_WIDE_INT fixed_offset = streamer_read_uhwi (ib);
997 HOST_WIDE_INT virtual_value = streamer_read_uhwi (ib);
998
999 node->thunk.fixed_offset = fixed_offset;
1000 node->thunk.this_adjusting = (type & 2);
1001 node->thunk.virtual_value = virtual_value;
1002 node->thunk.virtual_offset_p = (type & 4);
1003 }
1004 if (node->thunk.thunk_p || node->alias)
1005 {
1006 if (streamer_read_hwi_in_range (ib, "alias nonzero flag", 0, 1))
1007 {
1008 decl_index = streamer_read_uhwi (ib);
1009 node->thunk.alias = lto_file_decl_data_get_fn_decl (file_data,
1010 decl_index);
1011 }
1012 }
1013 return node;
1014 }
1015
1016 /* Read a node from input_block IB. TAG is the node's tag just read.
1017 Return the node read or overwriten. */
1018
1019 static struct varpool_node *
1020 input_varpool_node (struct lto_file_decl_data *file_data,
1021 struct lto_input_block *ib)
1022 {
1023 int decl_index;
1024 tree var_decl;
1025 struct varpool_node *node;
1026 struct bitpack_d bp;
1027 int ref = LCC_NOT_FOUND;
1028 bool non_null_aliasof;
1029 int order;
1030
1031 order = streamer_read_hwi (ib) + order_base;
1032 decl_index = streamer_read_uhwi (ib);
1033 var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index);
1034 node = varpool_node_for_decl (var_decl);
1035 node->symbol.order = order;
1036 if (order >= symtab_order)
1037 symtab_order = order + 1;
1038 node->symbol.lto_file_data = file_data;
1039
1040 bp = streamer_read_bitpack (ib);
1041 node->symbol.externally_visible = bp_unpack_value (&bp, 1);
1042 node->symbol.force_output = bp_unpack_value (&bp, 1);
1043 node->finalized = bp_unpack_value (&bp, 1);
1044 node->alias = bp_unpack_value (&bp, 1);
1045 non_null_aliasof = bp_unpack_value (&bp, 1);
1046 node->symbol.used_from_other_partition = bp_unpack_value (&bp, 1);
1047 node->symbol.in_other_partition = bp_unpack_value (&bp, 1);
1048 node->analyzed = (node->finalized && (!node->alias || !node->symbol.in_other_partition));
1049 if (node->symbol.in_other_partition)
1050 {
1051 DECL_EXTERNAL (node->symbol.decl) = 1;
1052 TREE_STATIC (node->symbol.decl) = 0;
1053 }
1054 if (non_null_aliasof)
1055 {
1056 decl_index = streamer_read_uhwi (ib);
1057 node->alias_of = lto_file_decl_data_get_var_decl (file_data, decl_index);
1058 }
1059 ref = streamer_read_hwi (ib);
1060 /* Store a reference for now, and fix up later to be a pointer. */
1061 node->symbol.same_comdat_group = (symtab_node) (intptr_t) ref;
1062 node->symbol.resolution = streamer_read_enum (ib, ld_plugin_symbol_resolution,
1063 LDPR_NUM_KNOWN);
1064
1065 return node;
1066 }
1067
1068 /* Read a node from input_block IB. TAG is the node's tag just read.
1069 Return the node read or overwriten. */
1070
1071 static void
1072 input_ref (struct lto_input_block *ib,
1073 symtab_node referring_node,
1074 vec<symtab_node> nodes)
1075 {
1076 symtab_node node = NULL;
1077 struct bitpack_d bp;
1078 enum ipa_ref_use use;
1079
1080 bp = streamer_read_bitpack (ib);
1081 use = (enum ipa_ref_use) bp_unpack_value (&bp, 2);
1082 node = nodes[streamer_read_hwi (ib)];
1083 ipa_record_reference (referring_node, node, use, NULL);
1084 }
1085
1086 /* Read an edge from IB. NODES points to a vector of previously read nodes for
1087 decoding caller and callee of the edge to be read. If INDIRECT is true, the
1088 edge being read is indirect (in the sense that it has
1089 indirect_unknown_callee set). */
1090
1091 static void
1092 input_edge (struct lto_input_block *ib, vec<symtab_node> nodes,
1093 bool indirect)
1094 {
1095 struct cgraph_node *caller, *callee;
1096 struct cgraph_edge *edge;
1097 unsigned int stmt_id;
1098 gcov_type count;
1099 int freq;
1100 cgraph_inline_failed_t inline_failed;
1101 struct bitpack_d bp;
1102 int ecf_flags = 0;
1103
1104 caller = cgraph (nodes[streamer_read_hwi (ib)]);
1105 if (caller == NULL || caller->symbol.decl == NULL_TREE)
1106 internal_error ("bytecode stream: no caller found while reading edge");
1107
1108 if (!indirect)
1109 {
1110 callee = cgraph (nodes[streamer_read_hwi (ib)]);
1111 if (callee == NULL || callee->symbol.decl == NULL_TREE)
1112 internal_error ("bytecode stream: no callee found while reading edge");
1113 }
1114 else
1115 callee = NULL;
1116
1117 count = streamer_read_gcov_count (ib);
1118
1119 bp = streamer_read_bitpack (ib);
1120 inline_failed = bp_unpack_enum (&bp, cgraph_inline_failed_enum, CIF_N_REASONS);
1121 stmt_id = bp_unpack_var_len_unsigned (&bp);
1122 freq = (int) bp_unpack_var_len_unsigned (&bp);
1123
1124 if (indirect)
1125 edge = cgraph_create_indirect_edge (caller, NULL, 0, count, freq);
1126 else
1127 edge = cgraph_create_edge (caller, callee, NULL, count, freq);
1128
1129 edge->indirect_inlining_edge = bp_unpack_value (&bp, 1);
1130 edge->lto_stmt_uid = stmt_id;
1131 edge->inline_failed = inline_failed;
1132 edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1);
1133 edge->can_throw_external = bp_unpack_value (&bp, 1);
1134 if (indirect)
1135 {
1136 if (bp_unpack_value (&bp, 1))
1137 ecf_flags |= ECF_CONST;
1138 if (bp_unpack_value (&bp, 1))
1139 ecf_flags |= ECF_PURE;
1140 if (bp_unpack_value (&bp, 1))
1141 ecf_flags |= ECF_NORETURN;
1142 if (bp_unpack_value (&bp, 1))
1143 ecf_flags |= ECF_MALLOC;
1144 if (bp_unpack_value (&bp, 1))
1145 ecf_flags |= ECF_NOTHROW;
1146 if (bp_unpack_value (&bp, 1))
1147 ecf_flags |= ECF_RETURNS_TWICE;
1148 edge->indirect_info->ecf_flags = ecf_flags;
1149 }
1150 }
1151
1152
1153 /* Read a cgraph from IB using the info in FILE_DATA. */
1154
1155 static vec<symtab_node>
1156 input_cgraph_1 (struct lto_file_decl_data *file_data,
1157 struct lto_input_block *ib)
1158 {
1159 enum LTO_symtab_tags tag;
1160 vec<symtab_node> nodes = vNULL;
1161 symtab_node node;
1162 unsigned i;
1163
1164 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1165 order_base = symtab_order;
1166 while (tag)
1167 {
1168 if (tag == LTO_symtab_edge)
1169 input_edge (ib, nodes, false);
1170 else if (tag == LTO_symtab_indirect_edge)
1171 input_edge (ib, nodes, true);
1172 else if (tag == LTO_symtab_variable)
1173 {
1174 node = (symtab_node)input_varpool_node (file_data, ib);
1175 nodes.safe_push (node);
1176 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1177 }
1178 else
1179 {
1180 node = (symtab_node)input_node (file_data, ib, tag, nodes);
1181 if (node == NULL || node->symbol.decl == NULL_TREE)
1182 internal_error ("bytecode stream: found empty cgraph node");
1183 nodes.safe_push (node);
1184 lto_symtab_encoder_encode (file_data->symtab_node_encoder, node);
1185 }
1186
1187 tag = streamer_read_enum (ib, LTO_symtab_tags, LTO_symtab_last_tag);
1188 }
1189
1190 lto_input_toplevel_asms (file_data, order_base);
1191
1192 /* AUX pointers should be all non-zero for function nodes read from the stream. */
1193 #ifdef ENABLE_CHECKING
1194 FOR_EACH_VEC_ELT (nodes, i, node)
1195 gcc_assert (node->symbol.aux || !is_a <cgraph_node> (node));
1196 #endif
1197 FOR_EACH_VEC_ELT (nodes, i, node)
1198 {
1199 int ref;
1200 if (cgraph_node *cnode = dyn_cast <cgraph_node> (node))
1201 {
1202 ref = (int) (intptr_t) cnode->global.inlined_to;
1203
1204 /* We share declaration of builtins, so we may read same node twice. */
1205 if (!node->symbol.aux)
1206 continue;
1207 node->symbol.aux = NULL;
1208
1209 /* Fixup inlined_to from reference to pointer. */
1210 if (ref != LCC_NOT_FOUND)
1211 cgraph (node)->global.inlined_to = cgraph (nodes[ref]);
1212 else
1213 cnode->global.inlined_to = NULL;
1214 }
1215
1216 ref = (int) (intptr_t) node->symbol.same_comdat_group;
1217
1218 /* Fixup same_comdat_group from reference to pointer. */
1219 if (ref != LCC_NOT_FOUND)
1220 node->symbol.same_comdat_group = nodes[ref];
1221 else
1222 node->symbol.same_comdat_group = NULL;
1223 }
1224 FOR_EACH_VEC_ELT (nodes, i, node)
1225 node->symbol.aux = is_a <cgraph_node> (node) ? (void *)1 : NULL;
1226 return nodes;
1227 }
1228
1229 /* Input ipa_refs. */
1230
1231 static void
1232 input_refs (struct lto_input_block *ib,
1233 vec<symtab_node> nodes)
1234 {
1235 int count;
1236 int idx;
1237 while (true)
1238 {
1239 symtab_node node;
1240 count = streamer_read_uhwi (ib);
1241 if (!count)
1242 break;
1243 idx = streamer_read_uhwi (ib);
1244 node = nodes[idx];
1245 while (count)
1246 {
1247 input_ref (ib, node, nodes);
1248 count--;
1249 }
1250 }
1251 }
1252
1253
1254 static struct gcov_ctr_summary lto_gcov_summary;
1255
1256 /* Input profile_info from IB. */
1257 static void
1258 input_profile_summary (struct lto_input_block *ib,
1259 struct lto_file_decl_data *file_data)
1260 {
1261 unsigned h_ix;
1262 struct bitpack_d bp;
1263 unsigned int runs = streamer_read_uhwi (ib);
1264 if (runs)
1265 {
1266 file_data->profile_info.runs = runs;
1267 file_data->profile_info.sum_max = streamer_read_gcov_count (ib);
1268 file_data->profile_info.sum_all = streamer_read_gcov_count (ib);
1269
1270 memset (file_data->profile_info.histogram, 0,
1271 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1272 /* Input the bitpack of non-zero histogram indices. */
1273 bp = streamer_read_bitpack (ib);
1274 /* Read in and unpack the full bitpack, flagging non-zero
1275 histogram entries by setting the num_counters non-zero. */
1276 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1277 {
1278 file_data->profile_info.histogram[h_ix].num_counters
1279 = bp_unpack_value (&bp, 1);
1280 }
1281 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1282 {
1283 if (!file_data->profile_info.histogram[h_ix].num_counters)
1284 continue;
1285
1286 file_data->profile_info.histogram[h_ix].num_counters
1287 = streamer_read_gcov_count (ib);
1288 file_data->profile_info.histogram[h_ix].min_value
1289 = streamer_read_gcov_count (ib);
1290 file_data->profile_info.histogram[h_ix].cum_value
1291 = streamer_read_gcov_count (ib);
1292 }
1293 /* IPA-profile computes hot bb threshold based on cumulated
1294 whole program profile. We need to stream it down to ltrans. */
1295 if (flag_ltrans)
1296 set_hot_bb_threshold (streamer_read_gcov_count (ib));
1297 }
1298
1299 }
1300
1301 /* Rescale profile summaries to the same number of runs in the whole unit. */
1302
1303 static void
1304 merge_profile_summaries (struct lto_file_decl_data **file_data_vec)
1305 {
1306 struct lto_file_decl_data *file_data;
1307 unsigned int j, h_ix;
1308 gcov_unsigned_t max_runs = 0;
1309 struct cgraph_node *node;
1310 struct cgraph_edge *edge;
1311 gcov_type saved_sum_all = 0;
1312 gcov_ctr_summary *saved_profile_info = 0;
1313 int saved_scale = 0;
1314
1315 /* Find unit with maximal number of runs. If we ever get serious about
1316 roundoff errors, we might also consider computing smallest common
1317 multiply. */
1318 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1319 if (max_runs < file_data->profile_info.runs)
1320 max_runs = file_data->profile_info.runs;
1321
1322 if (!max_runs)
1323 return;
1324
1325 /* Simple overflow check. We probably don't need to support that many train
1326 runs. Such a large value probably imply data corruption anyway. */
1327 if (max_runs > INT_MAX / REG_BR_PROB_BASE)
1328 {
1329 sorry ("At most %i profile runs is supported. Perhaps corrupted profile?",
1330 INT_MAX / REG_BR_PROB_BASE);
1331 return;
1332 }
1333
1334 profile_info = &lto_gcov_summary;
1335 lto_gcov_summary.runs = max_runs;
1336 lto_gcov_summary.sum_max = 0;
1337 memset (lto_gcov_summary.histogram, 0,
1338 sizeof (gcov_bucket_type) * GCOV_HISTOGRAM_SIZE);
1339
1340 /* Rescale all units to the maximal number of runs.
1341 sum_max can not be easily merged, as we have no idea what files come from
1342 the same run. We do not use the info anyway, so leave it 0. */
1343 for (j = 0; (file_data = file_data_vec[j]) != NULL; j++)
1344 if (file_data->profile_info.runs)
1345 {
1346 int scale = GCOV_COMPUTE_SCALE (max_runs,
1347 file_data->profile_info.runs);
1348 lto_gcov_summary.sum_max
1349 = MAX (lto_gcov_summary.sum_max,
1350 apply_probability (file_data->profile_info.sum_max, scale));
1351 lto_gcov_summary.sum_all
1352 = MAX (lto_gcov_summary.sum_all,
1353 apply_probability (file_data->profile_info.sum_all, scale));
1354 /* Save a pointer to the profile_info with the largest
1355 scaled sum_all and the scale for use in merging the
1356 histogram. */
1357 if (!saved_profile_info
1358 || lto_gcov_summary.sum_all > saved_sum_all)
1359 {
1360 saved_profile_info = &file_data->profile_info;
1361 saved_sum_all = lto_gcov_summary.sum_all;
1362 saved_scale = scale;
1363 }
1364 }
1365
1366 gcc_assert (saved_profile_info);
1367
1368 /* Scale up the histogram from the profile that had the largest
1369 scaled sum_all above. */
1370 for (h_ix = 0; h_ix < GCOV_HISTOGRAM_SIZE; h_ix++)
1371 {
1372 /* Scale up the min value as we did the corresponding sum_all
1373 above. Use that to find the new histogram index. */
1374 gcov_type scaled_min
1375 = apply_probability (saved_profile_info->histogram[h_ix].min_value,
1376 saved_scale);
1377 /* The new index may be shared with another scaled histogram entry,
1378 so we need to account for a non-zero histogram entry at new_ix. */
1379 unsigned new_ix = gcov_histo_index (scaled_min);
1380 lto_gcov_summary.histogram[new_ix].min_value
1381 = (lto_gcov_summary.histogram[new_ix].num_counters
1382 ? MIN (lto_gcov_summary.histogram[new_ix].min_value, scaled_min)
1383 : scaled_min);
1384 /* Some of the scaled counter values would ostensibly need to be placed
1385 into different (larger) histogram buckets, but we keep things simple
1386 here and place the scaled cumulative counter value in the bucket
1387 corresponding to the scaled minimum counter value. */
1388 lto_gcov_summary.histogram[new_ix].cum_value
1389 += apply_probability (saved_profile_info->histogram[h_ix].cum_value,
1390 saved_scale);
1391 lto_gcov_summary.histogram[new_ix].num_counters
1392 += saved_profile_info->histogram[h_ix].num_counters;
1393 }
1394
1395 /* Watch roundoff errors. */
1396 if (lto_gcov_summary.sum_max < max_runs)
1397 lto_gcov_summary.sum_max = max_runs;
1398
1399 /* If merging already happent at WPA time, we are done. */
1400 if (flag_ltrans)
1401 return;
1402
1403 /* Now compute count_materialization_scale of each node.
1404 During LTRANS we already have values of count_materialization_scale
1405 computed, so just update them. */
1406 FOR_EACH_FUNCTION (node)
1407 if (node->symbol.lto_file_data
1408 && node->symbol.lto_file_data->profile_info.runs)
1409 {
1410 int scale;
1411
1412 scale = RDIV (node->count_materialization_scale * max_runs,
1413 node->symbol.lto_file_data->profile_info.runs);
1414 node->count_materialization_scale = scale;
1415 if (scale < 0)
1416 fatal_error ("Profile information in %s corrupted",
1417 file_data->file_name);
1418
1419 if (scale == REG_BR_PROB_BASE)
1420 continue;
1421 for (edge = node->callees; edge; edge = edge->next_callee)
1422 edge->count = apply_probability (edge->count, scale);
1423 node->count = apply_probability (node->count, scale);
1424 }
1425 }
1426
1427 /* Input and merge the symtab from each of the .o files passed to
1428 lto1. */
1429
1430 void
1431 input_symtab (void)
1432 {
1433 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1434 struct lto_file_decl_data *file_data;
1435 unsigned int j = 0;
1436 struct cgraph_node *node;
1437
1438 cgraph_state = CGRAPH_STATE_IPA_SSA;
1439
1440 while ((file_data = file_data_vec[j++]))
1441 {
1442 const char *data;
1443 size_t len;
1444 struct lto_input_block *ib;
1445 vec<symtab_node> nodes;
1446
1447 ib = lto_create_simple_input_block (file_data, LTO_section_symtab_nodes,
1448 &data, &len);
1449 if (!ib)
1450 fatal_error ("cannot find LTO cgraph in %s", file_data->file_name);
1451 input_profile_summary (ib, file_data);
1452 file_data->symtab_node_encoder = lto_symtab_encoder_new (true);
1453 nodes = input_cgraph_1 (file_data, ib);
1454 lto_destroy_simple_input_block (file_data, LTO_section_symtab_nodes,
1455 ib, data, len);
1456
1457 ib = lto_create_simple_input_block (file_data, LTO_section_refs,
1458 &data, &len);
1459 if (!ib)
1460 fatal_error("cannot find LTO section refs in %s", file_data->file_name);
1461 input_refs (ib, nodes);
1462 lto_destroy_simple_input_block (file_data, LTO_section_refs,
1463 ib, data, len);
1464 if (flag_ltrans)
1465 input_cgraph_opt_summary (nodes);
1466 nodes.release ();
1467 }
1468
1469 merge_profile_summaries (file_data_vec);
1470 get_working_sets ();
1471
1472
1473 /* Clear out the aux field that was used to store enough state to
1474 tell which nodes should be overwritten. */
1475 FOR_EACH_FUNCTION (node)
1476 {
1477 /* Some nodes may have been created by cgraph_node. This
1478 happens when the callgraph contains nested functions. If the
1479 node for the parent function was never emitted to the gimple
1480 file, cgraph_node will create a node for it when setting the
1481 context of the nested function. */
1482 if (node->symbol.lto_file_data)
1483 node->symbol.aux = NULL;
1484 }
1485 }
1486
1487 /* True when we need optimization summary for NODE. */
1488
1489 static int
1490 output_cgraph_opt_summary_p (struct cgraph_node *node)
1491 {
1492 return (node->clone_of
1493 && (node->clone.tree_map
1494 || node->clone.args_to_skip
1495 || node->clone.combined_args_to_skip));
1496 }
1497
1498 /* Output optimization summary for EDGE to OB. */
1499 static void
1500 output_edge_opt_summary (struct output_block *ob ATTRIBUTE_UNUSED,
1501 struct cgraph_edge *edge ATTRIBUTE_UNUSED)
1502 {
1503 }
1504
1505 /* Output optimization summary for NODE to OB. */
1506
1507 static void
1508 output_node_opt_summary (struct output_block *ob,
1509 struct cgraph_node *node,
1510 lto_symtab_encoder_t encoder)
1511 {
1512 unsigned int index;
1513 bitmap_iterator bi;
1514 struct ipa_replace_map *map;
1515 struct bitpack_d bp;
1516 int i;
1517 struct cgraph_edge *e;
1518
1519 if (node->clone.args_to_skip)
1520 {
1521 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.args_to_skip));
1522 EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi)
1523 streamer_write_uhwi (ob, index);
1524 }
1525 else
1526 streamer_write_uhwi (ob, 0);
1527 if (node->clone.combined_args_to_skip)
1528 {
1529 streamer_write_uhwi (ob, bitmap_count_bits (node->clone.combined_args_to_skip));
1530 EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi)
1531 streamer_write_uhwi (ob, index);
1532 }
1533 else
1534 streamer_write_uhwi (ob, 0);
1535 streamer_write_uhwi (ob, vec_safe_length (node->clone.tree_map));
1536 FOR_EACH_VEC_SAFE_ELT (node->clone.tree_map, i, map)
1537 {
1538 int parm_num;
1539 tree parm;
1540
1541 for (parm_num = 0, parm = DECL_ARGUMENTS (node->symbol.decl); parm;
1542 parm = DECL_CHAIN (parm), parm_num++)
1543 if (map->old_tree == parm)
1544 break;
1545 /* At the moment we assume all old trees to be PARM_DECLs, because we have no
1546 mechanism to store function local declarations into summaries. */
1547 gcc_assert (parm);
1548 streamer_write_uhwi (ob, parm_num);
1549 gcc_assert (EXPR_LOCATION (map->new_tree) == UNKNOWN_LOCATION);
1550 stream_write_tree (ob, map->new_tree, true);
1551 bp = bitpack_create (ob->main_stream);
1552 bp_pack_value (&bp, map->replace_p, 1);
1553 bp_pack_value (&bp, map->ref_p, 1);
1554 streamer_write_bitpack (&bp);
1555 }
1556
1557 if (lto_symtab_encoder_in_partition_p (encoder, (symtab_node) node))
1558 {
1559 for (e = node->callees; e; e = e->next_callee)
1560 output_edge_opt_summary (ob, e);
1561 for (e = node->indirect_calls; e; e = e->next_callee)
1562 output_edge_opt_summary (ob, e);
1563 }
1564 }
1565
1566 /* Output optimization summaries stored in callgraph.
1567 At the moment it is the clone info structure. */
1568
1569 static void
1570 output_cgraph_opt_summary (void)
1571 {
1572 int i, n_nodes;
1573 lto_symtab_encoder_t encoder;
1574 struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum);
1575 unsigned count = 0;
1576
1577 ob->cgraph_node = NULL;
1578 encoder = ob->decl_state->symtab_node_encoder;
1579 n_nodes = lto_symtab_encoder_size (encoder);
1580 for (i = 0; i < n_nodes; i++)
1581 {
1582 symtab_node node = lto_symtab_encoder_deref (encoder, i);
1583 cgraph_node *cnode = dyn_cast <cgraph_node> (node);
1584 if (cnode && output_cgraph_opt_summary_p (cnode))
1585 count++;
1586 }
1587 streamer_write_uhwi (ob, count);
1588 for (i = 0; i < n_nodes; i++)
1589 {
1590 symtab_node node = lto_symtab_encoder_deref (encoder, i);
1591 cgraph_node *cnode = dyn_cast <cgraph_node> (node);
1592 if (cnode && output_cgraph_opt_summary_p (cnode))
1593 {
1594 streamer_write_uhwi (ob, i);
1595 output_node_opt_summary (ob, cnode, encoder);
1596 }
1597 }
1598 produce_asm (ob, NULL);
1599 destroy_output_block (ob);
1600 }
1601
1602 /* Input optimisation summary of EDGE. */
1603
1604 static void
1605 input_edge_opt_summary (struct cgraph_edge *edge ATTRIBUTE_UNUSED,
1606 struct lto_input_block *ib_main ATTRIBUTE_UNUSED)
1607 {
1608 }
1609
1610 /* Input optimisation summary of NODE. */
1611
1612 static void
1613 input_node_opt_summary (struct cgraph_node *node,
1614 struct lto_input_block *ib_main,
1615 struct data_in *data_in)
1616 {
1617 int i;
1618 int count;
1619 int bit;
1620 struct bitpack_d bp;
1621 struct cgraph_edge *e;
1622
1623 count = streamer_read_uhwi (ib_main);
1624 if (count)
1625 node->clone.args_to_skip = BITMAP_GGC_ALLOC ();
1626 for (i = 0; i < count; i++)
1627 {
1628 bit = streamer_read_uhwi (ib_main);
1629 bitmap_set_bit (node->clone.args_to_skip, bit);
1630 }
1631 count = streamer_read_uhwi (ib_main);
1632 if (count)
1633 node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC ();
1634 for (i = 0; i < count; i++)
1635 {
1636 bit = streamer_read_uhwi (ib_main);
1637 bitmap_set_bit (node->clone.combined_args_to_skip, bit);
1638 }
1639 count = streamer_read_uhwi (ib_main);
1640 for (i = 0; i < count; i++)
1641 {
1642 struct ipa_replace_map *map = ggc_alloc_ipa_replace_map ();
1643
1644 vec_safe_push (node->clone.tree_map, map);
1645 map->parm_num = streamer_read_uhwi (ib_main);
1646 map->old_tree = NULL;
1647 map->new_tree = stream_read_tree (ib_main, data_in);
1648 bp = streamer_read_bitpack (ib_main);
1649 map->replace_p = bp_unpack_value (&bp, 1);
1650 map->ref_p = bp_unpack_value (&bp, 1);
1651 }
1652 for (e = node->callees; e; e = e->next_callee)
1653 input_edge_opt_summary (e, ib_main);
1654 for (e = node->indirect_calls; e; e = e->next_callee)
1655 input_edge_opt_summary (e, ib_main);
1656 }
1657
1658 /* Read section in file FILE_DATA of length LEN with data DATA. */
1659
1660 static void
1661 input_cgraph_opt_section (struct lto_file_decl_data *file_data,
1662 const char *data, size_t len,
1663 vec<symtab_node> nodes)
1664 {
1665 const struct lto_function_header *header =
1666 (const struct lto_function_header *) data;
1667 const int cfg_offset = sizeof (struct lto_function_header);
1668 const int main_offset = cfg_offset + header->cfg_size;
1669 const int string_offset = main_offset + header->main_size;
1670 struct data_in *data_in;
1671 struct lto_input_block ib_main;
1672 unsigned int i;
1673 unsigned int count;
1674
1675 LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0,
1676 header->main_size);
1677
1678 data_in =
1679 lto_data_in_create (file_data, (const char *) data + string_offset,
1680 header->string_size, vNULL);
1681 count = streamer_read_uhwi (&ib_main);
1682
1683 for (i = 0; i < count; i++)
1684 {
1685 int ref = streamer_read_uhwi (&ib_main);
1686 input_node_opt_summary (cgraph (nodes[ref]),
1687 &ib_main, data_in);
1688 }
1689 lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data,
1690 len);
1691 lto_data_in_delete (data_in);
1692 }
1693
1694 /* Input optimization summary of cgraph. */
1695
1696 static void
1697 input_cgraph_opt_summary (vec<symtab_node> nodes)
1698 {
1699 struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
1700 struct lto_file_decl_data *file_data;
1701 unsigned int j = 0;
1702
1703 while ((file_data = file_data_vec[j++]))
1704 {
1705 size_t len;
1706 const char *data =
1707 lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL,
1708 &len);
1709
1710 if (data)
1711 input_cgraph_opt_section (file_data, data, len, nodes);
1712 }
1713 }