/* Write and read the cgraph to the memory mapped representation of a .o file. Copyright 2009, 2010 Free Software Foundation, Inc. Contributed by Kenneth Zadeck This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see . */ #include "config.h" #include "system.h" #include "coretypes.h" #include "tm.h" #include "tree.h" #include "expr.h" #include "flags.h" #include "params.h" #include "input.h" #include "hashtab.h" #include "langhooks.h" #include "basic-block.h" #include "tree-flow.h" #include "cgraph.h" #include "function.h" #include "ggc.h" #include "diagnostic-core.h" #include "except.h" #include "vec.h" #include "timevar.h" #include "output.h" #include "pointer-set.h" #include "lto-streamer.h" #include "gcov-io.h" static void output_varpool (cgraph_node_set, varpool_node_set); static void output_cgraph_opt_summary (cgraph_node_set set); static void input_cgraph_opt_summary (VEC (cgraph_node_ptr, heap) * nodes); /* Cgraph streaming is organized as set of record whose type is indicated by a tag. */ enum LTO_cgraph_tags { /* Must leave 0 for the stopper. */ /* Cgraph node without body available. */ LTO_cgraph_unavail_node = 1, /* Cgraph node with function body. */ LTO_cgraph_analyzed_node, /* Cgraph edges. */ LTO_cgraph_edge, LTO_cgraph_indirect_edge }; /* Create a new cgraph encoder. */ lto_cgraph_encoder_t lto_cgraph_encoder_new (void) { lto_cgraph_encoder_t encoder = XCNEW (struct lto_cgraph_encoder_d); encoder->map = pointer_map_create (); encoder->nodes = NULL; encoder->body = pointer_set_create (); return encoder; } /* Delete ENCODER and its components. */ void lto_cgraph_encoder_delete (lto_cgraph_encoder_t encoder) { VEC_free (cgraph_node_ptr, heap, encoder->nodes); pointer_map_destroy (encoder->map); pointer_set_destroy (encoder->body); free (encoder); } /* Return the existing reference number of NODE in the cgraph encoder in output block OB. Assign a new reference if this is the first time NODE is encoded. */ int lto_cgraph_encoder_encode (lto_cgraph_encoder_t encoder, struct cgraph_node *node) { int ref; void **slot; slot = pointer_map_contains (encoder->map, node); if (!slot) { ref = VEC_length (cgraph_node_ptr, encoder->nodes); slot = pointer_map_insert (encoder->map, node); *slot = (void *) (intptr_t) ref; VEC_safe_push (cgraph_node_ptr, heap, encoder->nodes, node); } else ref = (int) (intptr_t) *slot; return ref; } #define LCC_NOT_FOUND (-1) /* Look up NODE in encoder. Return NODE's reference if it has been encoded or LCC_NOT_FOUND if it is not there. */ int lto_cgraph_encoder_lookup (lto_cgraph_encoder_t encoder, struct cgraph_node *node) { void **slot = pointer_map_contains (encoder->map, node); return (slot ? (int) (intptr_t) *slot : LCC_NOT_FOUND); } /* Return the cgraph node corresponding to REF using ENCODER. */ struct cgraph_node * lto_cgraph_encoder_deref (lto_cgraph_encoder_t encoder, int ref) { if (ref == LCC_NOT_FOUND) return NULL; return VEC_index (cgraph_node_ptr, encoder->nodes, ref); } /* Return TRUE if we should encode initializer of NODE (if any). */ bool lto_cgraph_encoder_encode_body_p (lto_cgraph_encoder_t encoder, struct cgraph_node *node) { return pointer_set_contains (encoder->body, node); } /* Return TRUE if we should encode body of NODE (if any). */ static void lto_set_cgraph_encoder_encode_body (lto_cgraph_encoder_t encoder, struct cgraph_node *node) { pointer_set_insert (encoder->body, node); } /* Create a new varpool encoder. */ lto_varpool_encoder_t lto_varpool_encoder_new (void) { lto_varpool_encoder_t encoder = XCNEW (struct lto_varpool_encoder_d); encoder->map = pointer_map_create (); encoder->initializer = pointer_set_create (); encoder->nodes = NULL; return encoder; } /* Delete ENCODER and its components. */ void lto_varpool_encoder_delete (lto_varpool_encoder_t encoder) { VEC_free (varpool_node_ptr, heap, encoder->nodes); pointer_map_destroy (encoder->map); pointer_set_destroy (encoder->initializer); free (encoder); } /* Return the existing reference number of NODE in the varpool encoder in output block OB. Assign a new reference if this is the first time NODE is encoded. */ int lto_varpool_encoder_encode (lto_varpool_encoder_t encoder, struct varpool_node *node) { int ref; void **slot; slot = pointer_map_contains (encoder->map, node); if (!slot) { ref = VEC_length (varpool_node_ptr, encoder->nodes); slot = pointer_map_insert (encoder->map, node); *slot = (void *) (intptr_t) ref; VEC_safe_push (varpool_node_ptr, heap, encoder->nodes, node); } else ref = (int) (intptr_t) *slot; return ref; } /* Look up NODE in encoder. Return NODE's reference if it has been encoded or LCC_NOT_FOUND if it is not there. */ int lto_varpool_encoder_lookup (lto_varpool_encoder_t encoder, struct varpool_node *node) { void **slot = pointer_map_contains (encoder->map, node); return (slot ? (int) (intptr_t) *slot : LCC_NOT_FOUND); } /* Return the varpool node corresponding to REF using ENCODER. */ struct varpool_node * lto_varpool_encoder_deref (lto_varpool_encoder_t encoder, int ref) { if (ref == LCC_NOT_FOUND) return NULL; return VEC_index (varpool_node_ptr, encoder->nodes, ref); } /* Return TRUE if we should encode initializer of NODE (if any). */ bool lto_varpool_encoder_encode_initializer_p (lto_varpool_encoder_t encoder, struct varpool_node *node) { return pointer_set_contains (encoder->initializer, node); } /* Return TRUE if we should encode initializer of NODE (if any). */ static void lto_set_varpool_encoder_encode_initializer (lto_varpool_encoder_t encoder, struct varpool_node *node) { pointer_set_insert (encoder->initializer, node); } /* Output the cgraph EDGE to OB using ENCODER. */ static void lto_output_edge (struct lto_simple_output_block *ob, struct cgraph_edge *edge, lto_cgraph_encoder_t encoder) { unsigned int uid; intptr_t ref; struct bitpack_d bp; if (edge->indirect_unknown_callee) lto_output_uleb128_stream (ob->main_stream, LTO_cgraph_indirect_edge); else lto_output_uleb128_stream (ob->main_stream, LTO_cgraph_edge); ref = lto_cgraph_encoder_lookup (encoder, edge->caller); gcc_assert (ref != LCC_NOT_FOUND); lto_output_sleb128_stream (ob->main_stream, ref); if (!edge->indirect_unknown_callee) { ref = lto_cgraph_encoder_lookup (encoder, edge->callee); gcc_assert (ref != LCC_NOT_FOUND); lto_output_sleb128_stream (ob->main_stream, ref); } lto_output_sleb128_stream (ob->main_stream, edge->count); bp = bitpack_create (ob->main_stream); uid = (!gimple_has_body_p (edge->caller->decl) ? edge->lto_stmt_uid : gimple_uid (edge->call_stmt)); bp_pack_value (&bp, uid, HOST_BITS_PER_INT); bp_pack_value (&bp, edge->inline_failed, HOST_BITS_PER_INT); bp_pack_value (&bp, edge->frequency, HOST_BITS_PER_INT); bp_pack_value (&bp, edge->loop_nest, 30); bp_pack_value (&bp, edge->indirect_inlining_edge, 1); bp_pack_value (&bp, edge->call_stmt_cannot_inline_p, 1); bp_pack_value (&bp, edge->can_throw_external, 1); if (edge->indirect_unknown_callee) { int flags = edge->indirect_info->ecf_flags; bp_pack_value (&bp, (flags & ECF_CONST) != 0, 1); bp_pack_value (&bp, (flags & ECF_PURE) != 0, 1); bp_pack_value (&bp, (flags & ECF_NORETURN) != 0, 1); bp_pack_value (&bp, (flags & ECF_MALLOC) != 0, 1); bp_pack_value (&bp, (flags & ECF_NOTHROW) != 0, 1); bp_pack_value (&bp, (flags & ECF_RETURNS_TWICE) != 0, 1); /* Flags that should not appear on indirect calls. */ gcc_assert (!(flags & (ECF_LOOPING_CONST_OR_PURE | ECF_MAY_BE_ALLOCA | ECF_SIBCALL | ECF_LEAF | ECF_NOVOPS))); } lto_output_bitpack (&bp); } /* Return if LIST contain references from other partitions. */ bool referenced_from_other_partition_p (struct ipa_ref_list *list, cgraph_node_set set, varpool_node_set vset) { int i; struct ipa_ref *ref; for (i = 0; ipa_ref_list_refering_iterate (list, i, ref); i++) { if (ref->refering_type == IPA_REF_CGRAPH) { if (ipa_ref_refering_node (ref)->in_other_partition || !cgraph_node_in_set_p (ipa_ref_refering_node (ref), set)) return true; } else { if (ipa_ref_refering_varpool_node (ref)->in_other_partition || !varpool_node_in_set_p (ipa_ref_refering_varpool_node (ref), vset)) return true; } } return false; } /* Return true when node is reachable from other partition. */ bool reachable_from_other_partition_p (struct cgraph_node *node, cgraph_node_set set) { struct cgraph_edge *e; if (!node->analyzed) return false; if (node->global.inlined_to) return false; for (e = node->callers; e; e = e->next_caller) if (e->caller->in_other_partition || !cgraph_node_in_set_p (e->caller, set)) return true; return false; } /* Return if LIST contain references from other partitions. */ bool referenced_from_this_partition_p (struct ipa_ref_list *list, cgraph_node_set set, varpool_node_set vset) { int i; struct ipa_ref *ref; for (i = 0; ipa_ref_list_refering_iterate (list, i, ref); i++) { if (ref->refering_type == IPA_REF_CGRAPH) { if (cgraph_node_in_set_p (ipa_ref_refering_node (ref), set)) return true; } else { if (varpool_node_in_set_p (ipa_ref_refering_varpool_node (ref), vset)) return true; } } return false; } /* Return true when node is reachable from other partition. */ bool reachable_from_this_partition_p (struct cgraph_node *node, cgraph_node_set set) { struct cgraph_edge *e; for (e = node->callers; e; e = e->next_caller) if (cgraph_node_in_set_p (e->caller, set)) return true; return false; } /* Output the cgraph NODE to OB. ENCODER is used to find the reference number of NODE->inlined_to. SET is the set of nodes we are writing to the current file. If NODE is not in SET, then NODE is a boundary of a cgraph_node_set and we pretend NODE just has a decl and no callees. WRITTEN_DECLS is the set of FUNCTION_DECLs that have had their callgraph node written so far. This is used to determine if NODE is a clone of a previously written node. */ static void lto_output_node (struct lto_simple_output_block *ob, struct cgraph_node *node, lto_cgraph_encoder_t encoder, cgraph_node_set set, varpool_node_set vset) { unsigned int tag; struct bitpack_d bp; bool boundary_p; intptr_t ref; bool in_other_partition = false; struct cgraph_node *clone_of; boundary_p = !cgraph_node_in_set_p (node, set); if (node->analyzed && !boundary_p) tag = LTO_cgraph_analyzed_node; else tag = LTO_cgraph_unavail_node; lto_output_uleb128_stream (ob->main_stream, tag); /* In WPA mode, we only output part of the call-graph. Also, we fake cgraph node attributes. There are two cases that we care. Boundary nodes: There are nodes that are not part of SET but are called from within SET. We artificially make them look like externally visible nodes with no function body. Cherry-picked nodes: These are nodes we pulled from other translation units into SET during IPA-inlining. We make them as local static nodes to prevent clashes with other local statics. */ if (boundary_p && node->analyzed) { /* Inline clones can not be part of boundary. gcc_assert (!node->global.inlined_to); FIXME: At the moment they can be, when partition contains an inline clone that is clone of inline clone from outside partition. We can reshape the clone tree and make other tree to be the root, but it needs a bit extra work and will be promplty done by cgraph_remove_node after reading back. */ in_other_partition = 1; } clone_of = node->clone_of; while (clone_of && (ref = lto_cgraph_encoder_lookup (encoder, clone_of)) == LCC_NOT_FOUND) if (clone_of->prev_sibling_clone) clone_of = clone_of->prev_sibling_clone; else clone_of = clone_of->clone_of; if (LTO_cgraph_analyzed_node) gcc_assert (clone_of || !node->clone_of); if (!clone_of) lto_output_sleb128_stream (ob->main_stream, LCC_NOT_FOUND); else lto_output_sleb128_stream (ob->main_stream, ref); lto_output_fn_decl_index (ob->decl_state, ob->main_stream, node->decl); lto_output_sleb128_stream (ob->main_stream, node->count); lto_output_sleb128_stream (ob->main_stream, node->count_materialization_scale); if (tag == LTO_cgraph_analyzed_node) { lto_output_sleb128_stream (ob->main_stream, node->local.inline_summary.estimated_self_stack_size); lto_output_sleb128_stream (ob->main_stream, node->local.inline_summary.self_size); lto_output_sleb128_stream (ob->main_stream, node->local.inline_summary.size_inlining_benefit); lto_output_sleb128_stream (ob->main_stream, node->local.inline_summary.self_time); lto_output_sleb128_stream (ob->main_stream, node->local.inline_summary.time_inlining_benefit); if (node->global.inlined_to) { ref = lto_cgraph_encoder_lookup (encoder, node->global.inlined_to); gcc_assert (ref != LCC_NOT_FOUND); } else ref = LCC_NOT_FOUND; lto_output_sleb128_stream (ob->main_stream, ref); } if (node->same_comdat_group && !boundary_p) { ref = lto_cgraph_encoder_lookup (encoder, node->same_comdat_group); gcc_assert (ref != LCC_NOT_FOUND); } else ref = LCC_NOT_FOUND; lto_output_sleb128_stream (ob->main_stream, ref); bp = bitpack_create (ob->main_stream); bp_pack_value (&bp, node->local.local, 1); bp_pack_value (&bp, node->local.externally_visible, 1); bp_pack_value (&bp, node->local.finalized, 1); bp_pack_value (&bp, node->local.inlinable, 1); bp_pack_value (&bp, node->local.versionable, 1); bp_pack_value (&bp, node->local.can_change_signature, 1); bp_pack_value (&bp, node->local.disregard_inline_limits, 1); bp_pack_value (&bp, node->local.redefined_extern_inline, 1); bp_pack_value (&bp, node->local.vtable_method, 1); bp_pack_value (&bp, node->needed, 1); bp_pack_value (&bp, node->address_taken, 1); bp_pack_value (&bp, node->abstract_and_needed, 1); bp_pack_value (&bp, tag == LTO_cgraph_analyzed_node && !DECL_EXTERNAL (node->decl) && !DECL_COMDAT (node->decl) && (reachable_from_other_partition_p (node, set) || referenced_from_other_partition_p (&node->ref_list, set, vset)), 1); bp_pack_value (&bp, node->lowered, 1); bp_pack_value (&bp, in_other_partition, 1); bp_pack_value (&bp, node->alias, 1); bp_pack_value (&bp, node->finalized_by_frontend, 1); bp_pack_value (&bp, node->frequency, 2); bp_pack_value (&bp, node->only_called_at_startup, 1); bp_pack_value (&bp, node->only_called_at_exit, 1); lto_output_bitpack (&bp); lto_output_uleb128_stream (ob->main_stream, node->resolution); if (node->same_body) { struct cgraph_node *alias; unsigned long alias_count = 1; for (alias = node->same_body; alias->next; alias = alias->next) alias_count++; lto_output_uleb128_stream (ob->main_stream, alias_count); do { lto_output_fn_decl_index (ob->decl_state, ob->main_stream, alias->decl); if (alias->thunk.thunk_p) { lto_output_uleb128_stream (ob->main_stream, 1 + (alias->thunk.this_adjusting != 0) * 2 + (alias->thunk.virtual_offset_p != 0) * 4); lto_output_uleb128_stream (ob->main_stream, alias->thunk.fixed_offset); lto_output_uleb128_stream (ob->main_stream, alias->thunk.virtual_value); lto_output_fn_decl_index (ob->decl_state, ob->main_stream, alias->thunk.alias); } else { lto_output_uleb128_stream (ob->main_stream, 0); lto_output_fn_decl_index (ob->decl_state, ob->main_stream, alias->thunk.alias); } gcc_assert (cgraph_get_node (alias->thunk.alias) == node); lto_output_uleb128_stream (ob->main_stream, alias->resolution); alias = alias->previous; } while (alias); } else lto_output_uleb128_stream (ob->main_stream, 0); } /* Output the varpool NODE to OB. If NODE is not in SET, then NODE is a boundary. */ static void lto_output_varpool_node (struct lto_simple_output_block *ob, struct varpool_node *node, lto_varpool_encoder_t varpool_encoder, cgraph_node_set set, varpool_node_set vset) { bool boundary_p = !varpool_node_in_set_p (node, vset) && node->analyzed; struct bitpack_d bp; struct varpool_node *alias; int count = 0; int ref; lto_output_var_decl_index (ob->decl_state, ob->main_stream, node->decl); bp = bitpack_create (ob->main_stream); bp_pack_value (&bp, node->externally_visible, 1); bp_pack_value (&bp, node->force_output, 1); bp_pack_value (&bp, node->finalized, 1); bp_pack_value (&bp, node->alias, 1); gcc_assert (!node->alias || !node->extra_name); gcc_assert (node->finalized || !node->analyzed); gcc_assert (node->needed); /* Constant pool initializers can be de-unified into individual ltrans units. FIXME: Alternatively at -Os we may want to avoid generating for them the local labels and share them across LTRANS partitions. */ if (DECL_IN_CONSTANT_POOL (node->decl) && !DECL_COMDAT (node->decl)) { bp_pack_value (&bp, 0, 1); /* used_from_other_parition. */ bp_pack_value (&bp, 0, 1); /* in_other_partition. */ } else { bp_pack_value (&bp, node->analyzed && referenced_from_other_partition_p (&node->ref_list, set, vset), 1); bp_pack_value (&bp, boundary_p, 1); /* in_other_partition. */ } /* Also emit any extra name aliases. */ for (alias = node->extra_name; alias; alias = alias->next) count++; bp_pack_value (&bp, count != 0, 1); lto_output_bitpack (&bp); if (node->same_comdat_group && !boundary_p) { ref = lto_varpool_encoder_lookup (varpool_encoder, node->same_comdat_group); gcc_assert (ref != LCC_NOT_FOUND); } else ref = LCC_NOT_FOUND; lto_output_sleb128_stream (ob->main_stream, ref); lto_output_uleb128_stream (ob->main_stream, node->resolution); if (count) { lto_output_uleb128_stream (ob->main_stream, count); for (alias = node->extra_name; alias; alias = alias->next) { lto_output_var_decl_index (ob->decl_state, ob->main_stream, alias->decl); lto_output_uleb128_stream (ob->main_stream, alias->resolution); } } } /* Output the varpool NODE to OB. If NODE is not in SET, then NODE is a boundary. */ static void lto_output_ref (struct lto_simple_output_block *ob, struct ipa_ref *ref, lto_cgraph_encoder_t encoder, lto_varpool_encoder_t varpool_encoder) { struct bitpack_d bp; bp = bitpack_create (ob->main_stream); bp_pack_value (&bp, ref->refered_type, 1); bp_pack_value (&bp, ref->use, 2); lto_output_bitpack (&bp); if (ref->refered_type == IPA_REF_CGRAPH) { int nref = lto_cgraph_encoder_lookup (encoder, ipa_ref_node (ref)); gcc_assert (nref != LCC_NOT_FOUND); lto_output_sleb128_stream (ob->main_stream, nref); } else { int nref = lto_varpool_encoder_lookup (varpool_encoder, ipa_ref_varpool_node (ref)); gcc_assert (nref != LCC_NOT_FOUND); lto_output_sleb128_stream (ob->main_stream, nref); } } /* Stream out profile_summary to OB. */ static void output_profile_summary (struct lto_simple_output_block *ob) { if (profile_info) { /* We do not output num, sum_all and run_max, they are not used by GCC profile feedback and they are difficult to merge from multiple units. */ gcc_assert (profile_info->runs); lto_output_uleb128_stream (ob->main_stream, profile_info->runs); lto_output_uleb128_stream (ob->main_stream, profile_info->sum_max); } else lto_output_uleb128_stream (ob->main_stream, 0); } /* Add NODE into encoder as well as nodes it is cloned from. Do it in a way so clones appear first. */ static void add_node_to (lto_cgraph_encoder_t encoder, struct cgraph_node *node, bool include_body) { if (node->clone_of) add_node_to (encoder, node->clone_of, include_body); else if (include_body) lto_set_cgraph_encoder_encode_body (encoder, node); lto_cgraph_encoder_encode (encoder, node); } /* Add all references in LIST to encoders. */ static void add_references (lto_cgraph_encoder_t encoder, lto_varpool_encoder_t varpool_encoder, struct ipa_ref_list *list) { int i; struct ipa_ref *ref; for (i = 0; ipa_ref_list_reference_iterate (list, i, ref); i++) if (ref->refered_type == IPA_REF_CGRAPH) add_node_to (encoder, ipa_ref_node (ref), false); else { struct varpool_node *vnode = ipa_ref_varpool_node (ref); lto_varpool_encoder_encode (varpool_encoder, vnode); } } /* Output all callees or indirect outgoing edges. EDGE must be the first such edge. */ static void output_outgoing_cgraph_edges (struct cgraph_edge *edge, struct lto_simple_output_block *ob, lto_cgraph_encoder_t encoder) { if (!edge) return; /* Output edges in backward direction, so the reconstructed callgraph match and it is easy to associate call sites in the IPA pass summaries. */ while (edge->next_callee) edge = edge->next_callee; for (; edge; edge = edge->prev_callee) lto_output_edge (ob, edge, encoder); } /* Output the part of the cgraph in SET. */ static void output_refs (cgraph_node_set set, varpool_node_set vset, lto_cgraph_encoder_t encoder, lto_varpool_encoder_t varpool_encoder) { cgraph_node_set_iterator csi; varpool_node_set_iterator vsi; struct lto_simple_output_block *ob; int count; struct ipa_ref *ref; int i; ob = lto_create_simple_output_block (LTO_section_refs); for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi)) { struct cgraph_node *node = csi_node (csi); count = ipa_ref_list_nreferences (&node->ref_list); if (count) { lto_output_uleb128_stream (ob->main_stream, count); lto_output_uleb128_stream (ob->main_stream, lto_cgraph_encoder_lookup (encoder, node)); for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list, i, ref); i++) lto_output_ref (ob, ref, encoder, varpool_encoder); } } lto_output_uleb128_stream (ob->main_stream, 0); for (vsi = vsi_start (vset); !vsi_end_p (vsi); vsi_next (&vsi)) { struct varpool_node *node = vsi_node (vsi); count = ipa_ref_list_nreferences (&node->ref_list); if (count) { lto_output_uleb128_stream (ob->main_stream, count); lto_output_uleb128_stream (ob->main_stream, lto_varpool_encoder_lookup (varpool_encoder, node)); for (i = 0; ipa_ref_list_reference_iterate (&node->ref_list, i, ref); i++) lto_output_ref (ob, ref, encoder, varpool_encoder); } } lto_output_uleb128_stream (ob->main_stream, 0); lto_destroy_simple_output_block (ob); } /* Find out all cgraph and varpool nodes we want to encode in current unit and insert them to encoders. */ void compute_ltrans_boundary (struct lto_out_decl_state *state, cgraph_node_set set, varpool_node_set vset) { struct cgraph_node *node; cgraph_node_set_iterator csi; varpool_node_set_iterator vsi; struct cgraph_edge *edge; int i; lto_cgraph_encoder_t encoder; lto_varpool_encoder_t varpool_encoder; encoder = state->cgraph_node_encoder = lto_cgraph_encoder_new (); varpool_encoder = state->varpool_node_encoder = lto_varpool_encoder_new (); /* Go over all the nodes in SET and assign references. */ for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi)) { node = csi_node (csi); add_node_to (encoder, node, true); add_references (encoder, varpool_encoder, &node->ref_list); } for (vsi = vsi_start (vset); !vsi_end_p (vsi); vsi_next (&vsi)) { struct varpool_node *vnode = vsi_node (vsi); gcc_assert (!vnode->alias); lto_varpool_encoder_encode (varpool_encoder, vnode); lto_set_varpool_encoder_encode_initializer (varpool_encoder, vnode); add_references (encoder, varpool_encoder, &vnode->ref_list); } /* Pickle in also the initializer of all referenced readonly variables to help folding. Constant pool variables are not shared, so we must pickle those too. */ for (i = 0; i < lto_varpool_encoder_size (varpool_encoder); i++) { struct varpool_node *vnode = lto_varpool_encoder_deref (varpool_encoder, i); if (DECL_INITIAL (vnode->decl) && !lto_varpool_encoder_encode_initializer_p (varpool_encoder, vnode) && const_value_known_p (vnode->decl)) { lto_set_varpool_encoder_encode_initializer (varpool_encoder, vnode); add_references (encoder, varpool_encoder, &vnode->ref_list); } } /* Go over all the nodes again to include callees that are not in SET. */ for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi)) { node = csi_node (csi); for (edge = node->callees; edge; edge = edge->next_callee) { struct cgraph_node *callee = edge->callee; if (!cgraph_node_in_set_p (callee, set)) { /* We should have moved all the inlines. */ gcc_assert (!callee->global.inlined_to); add_node_to (encoder, callee, false); } } } } /* Output the part of the cgraph in SET. */ void output_cgraph (cgraph_node_set set, varpool_node_set vset) { struct cgraph_node *node; struct lto_simple_output_block *ob; cgraph_node_set_iterator csi; int i, n_nodes; lto_cgraph_encoder_t encoder; lto_varpool_encoder_t varpool_encoder; struct cgraph_asm_node *can; static bool asm_nodes_output = false; if (flag_wpa) output_cgraph_opt_summary (set); ob = lto_create_simple_output_block (LTO_section_cgraph); output_profile_summary (ob); /* An encoder for cgraph nodes should have been created by ipa_write_summaries_1. */ gcc_assert (ob->decl_state->cgraph_node_encoder); gcc_assert (ob->decl_state->varpool_node_encoder); encoder = ob->decl_state->cgraph_node_encoder; varpool_encoder = ob->decl_state->varpool_node_encoder; /* Write out the nodes. We must first output a node and then its clones, otherwise at a time reading back the node there would be nothing to clone from. */ n_nodes = lto_cgraph_encoder_size (encoder); for (i = 0; i < n_nodes; i++) { node = lto_cgraph_encoder_deref (encoder, i); lto_output_node (ob, node, encoder, set, vset); } /* Go over the nodes in SET again to write edges. */ for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi)) { node = csi_node (csi); output_outgoing_cgraph_edges (node->callees, ob, encoder); output_outgoing_cgraph_edges (node->indirect_calls, ob, encoder); } lto_output_uleb128_stream (ob->main_stream, 0); /* Emit toplevel asms. When doing WPA we must output every asm just once. Since we do not partition asm nodes at all, output them to first output. This is kind of hack, but should work well. */ if (!asm_nodes_output) { asm_nodes_output = true; for (can = cgraph_asm_nodes; can; can = can->next) { int len = TREE_STRING_LENGTH (can->asm_str); lto_output_uleb128_stream (ob->main_stream, len); for (i = 0; i < len; ++i) lto_output_1_stream (ob->main_stream, TREE_STRING_POINTER (can->asm_str)[i]); } } lto_output_uleb128_stream (ob->main_stream, 0); lto_destroy_simple_output_block (ob); output_varpool (set, vset); output_refs (set, vset, encoder, varpool_encoder); } /* Overwrite the information in NODE based on FILE_DATA, TAG, FLAGS, STACK_SIZE, SELF_TIME and SELF_SIZE. This is called either to initialize NODE or to replace the values in it, for instance because the first time we saw it, the function body was not available but now it is. BP is a bitpack with all the bitflags for NODE read from the stream. */ static void input_overwrite_node (struct lto_file_decl_data *file_data, struct cgraph_node *node, enum LTO_cgraph_tags tag, struct bitpack_d *bp, unsigned int stack_size, unsigned int self_time, unsigned int time_inlining_benefit, unsigned int self_size, unsigned int size_inlining_benefit, enum ld_plugin_symbol_resolution resolution) { node->aux = (void *) tag; node->local.inline_summary.estimated_self_stack_size = stack_size; node->local.inline_summary.self_time = self_time; node->local.inline_summary.time_inlining_benefit = time_inlining_benefit; node->local.inline_summary.self_size = self_size; node->local.inline_summary.size_inlining_benefit = size_inlining_benefit; node->global.time = self_time; node->global.size = self_size; node->global.estimated_stack_size = stack_size; node->global.estimated_growth = INT_MIN; node->local.lto_file_data = file_data; node->local.local = bp_unpack_value (bp, 1); node->local.externally_visible = bp_unpack_value (bp, 1); node->local.finalized = bp_unpack_value (bp, 1); node->local.inlinable = bp_unpack_value (bp, 1); node->local.versionable = bp_unpack_value (bp, 1); node->local.can_change_signature = bp_unpack_value (bp, 1); node->local.disregard_inline_limits = bp_unpack_value (bp, 1); node->local.redefined_extern_inline = bp_unpack_value (bp, 1); node->local.vtable_method = bp_unpack_value (bp, 1); node->needed = bp_unpack_value (bp, 1); node->address_taken = bp_unpack_value (bp, 1); node->abstract_and_needed = bp_unpack_value (bp, 1); node->reachable_from_other_partition = bp_unpack_value (bp, 1); node->lowered = bp_unpack_value (bp, 1); node->analyzed = tag == LTO_cgraph_analyzed_node; node->in_other_partition = bp_unpack_value (bp, 1); if (node->in_other_partition /* Avoid updating decl when we are seeing just inline clone. When inlining function that has functions already inlined into it, we produce clones of inline clones. WPA partitioning might put each clone into different unit and we might end up streaming inline clone from other partition to support clone we are interested in. */ && (!node->clone_of || node->clone_of->decl != node->decl)) { DECL_EXTERNAL (node->decl) = 1; TREE_STATIC (node->decl) = 0; } node->alias = bp_unpack_value (bp, 1); node->finalized_by_frontend = bp_unpack_value (bp, 1); node->frequency = (enum node_frequency)bp_unpack_value (bp, 2); node->only_called_at_startup = bp_unpack_value (bp, 1); node->only_called_at_exit = bp_unpack_value (bp, 1); node->resolution = resolution; } /* Output the part of the cgraph in SET. */ static void output_varpool (cgraph_node_set set, varpool_node_set vset) { struct lto_simple_output_block *ob = lto_create_simple_output_block (LTO_section_varpool); lto_varpool_encoder_t varpool_encoder = ob->decl_state->varpool_node_encoder; int len = lto_varpool_encoder_size (varpool_encoder), i; lto_output_uleb128_stream (ob->main_stream, len); /* Write out the nodes. We must first output a node and then its clones, otherwise at a time reading back the node there would be nothing to clone from. */ for (i = 0; i < len; i++) { lto_output_varpool_node (ob, lto_varpool_encoder_deref (varpool_encoder, i), varpool_encoder, set, vset); } lto_destroy_simple_output_block (ob); } /* Read a node from input_block IB. TAG is the node's tag just read. Return the node read or overwriten. */ static struct cgraph_node * input_node (struct lto_file_decl_data *file_data, struct lto_input_block *ib, enum LTO_cgraph_tags tag, VEC(cgraph_node_ptr, heap) *nodes) { tree fn_decl; struct cgraph_node *node; struct bitpack_d bp; int stack_size = 0; unsigned decl_index; int ref = LCC_NOT_FOUND, ref2 = LCC_NOT_FOUND; int self_time = 0; int self_size = 0; int time_inlining_benefit = 0; int size_inlining_benefit = 0; unsigned long same_body_count = 0; int clone_ref; enum ld_plugin_symbol_resolution resolution; clone_ref = lto_input_sleb128 (ib); decl_index = lto_input_uleb128 (ib); fn_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index); if (clone_ref != LCC_NOT_FOUND) { node = cgraph_clone_node (VEC_index (cgraph_node_ptr, nodes, clone_ref), fn_decl, 0, CGRAPH_FREQ_BASE, 0, false, NULL); } else node = cgraph_node (fn_decl); node->count = lto_input_sleb128 (ib); node->count_materialization_scale = lto_input_sleb128 (ib); if (tag == LTO_cgraph_analyzed_node) { stack_size = lto_input_sleb128 (ib); self_size = lto_input_sleb128 (ib); size_inlining_benefit = lto_input_sleb128 (ib); self_time = lto_input_sleb128 (ib); time_inlining_benefit = lto_input_sleb128 (ib); ref = lto_input_sleb128 (ib); } ref2 = lto_input_sleb128 (ib); /* Make sure that we have not read this node before. Nodes that have already been read will have their tag stored in the 'aux' field. Since built-in functions can be referenced in multiple functions, they are expected to be read more than once. */ if (node->aux && !DECL_IS_BUILTIN (node->decl)) internal_error ("bytecode stream: found multiple instances of cgraph " "node %d", node->uid); bp = lto_input_bitpack (ib); resolution = (enum ld_plugin_symbol_resolution)lto_input_uleb128 (ib); input_overwrite_node (file_data, node, tag, &bp, stack_size, self_time, time_inlining_benefit, self_size, size_inlining_benefit, resolution); /* Store a reference for now, and fix up later to be a pointer. */ node->global.inlined_to = (cgraph_node_ptr) (intptr_t) ref; /* Store a reference for now, and fix up later to be a pointer. */ node->same_comdat_group = (cgraph_node_ptr) (intptr_t) ref2; same_body_count = lto_input_uleb128 (ib); while (same_body_count-- > 0) { tree alias_decl; int type; struct cgraph_node *alias; decl_index = lto_input_uleb128 (ib); alias_decl = lto_file_decl_data_get_fn_decl (file_data, decl_index); type = lto_input_uleb128 (ib); if (!type) { tree real_alias; decl_index = lto_input_uleb128 (ib); real_alias = lto_file_decl_data_get_fn_decl (file_data, decl_index); alias = cgraph_same_body_alias (node, alias_decl, real_alias); } else { HOST_WIDE_INT fixed_offset = lto_input_uleb128 (ib); HOST_WIDE_INT virtual_value = lto_input_uleb128 (ib); tree real_alias; decl_index = lto_input_uleb128 (ib); real_alias = lto_file_decl_data_get_fn_decl (file_data, decl_index); alias = cgraph_add_thunk (node, alias_decl, fn_decl, type & 2, fixed_offset, virtual_value, (type & 4) ? size_int (virtual_value) : NULL_TREE, real_alias); } gcc_assert (alias); alias->resolution = (enum ld_plugin_symbol_resolution)lto_input_uleb128 (ib); } return node; } /* Read a node from input_block IB. TAG is the node's tag just read. Return the node read or overwriten. */ static struct varpool_node * input_varpool_node (struct lto_file_decl_data *file_data, struct lto_input_block *ib) { int decl_index; tree var_decl; struct varpool_node *node; struct bitpack_d bp; bool aliases_p; int count; int ref = LCC_NOT_FOUND; decl_index = lto_input_uleb128 (ib); var_decl = lto_file_decl_data_get_var_decl (file_data, decl_index); node = varpool_node (var_decl); node->lto_file_data = file_data; bp = lto_input_bitpack (ib); node->externally_visible = bp_unpack_value (&bp, 1); node->force_output = bp_unpack_value (&bp, 1); node->finalized = bp_unpack_value (&bp, 1); node->alias = bp_unpack_value (&bp, 1); node->analyzed = node->finalized; node->used_from_other_partition = bp_unpack_value (&bp, 1); node->in_other_partition = bp_unpack_value (&bp, 1); if (node->in_other_partition) { DECL_EXTERNAL (node->decl) = 1; TREE_STATIC (node->decl) = 0; } aliases_p = bp_unpack_value (&bp, 1); if (node->finalized) varpool_mark_needed_node (node); ref = lto_input_sleb128 (ib); /* Store a reference for now, and fix up later to be a pointer. */ node->same_comdat_group = (struct varpool_node *) (intptr_t) ref; node->resolution = (enum ld_plugin_symbol_resolution)lto_input_uleb128 (ib); if (aliases_p) { count = lto_input_uleb128 (ib); for (; count > 0; count --) { tree decl = lto_file_decl_data_get_var_decl (file_data, lto_input_uleb128 (ib)); struct varpool_node *alias; alias = varpool_extra_name_alias (decl, var_decl); alias->resolution = (enum ld_plugin_symbol_resolution)lto_input_uleb128 (ib); } } return node; } /* Read a node from input_block IB. TAG is the node's tag just read. Return the node read or overwriten. */ static void input_ref (struct lto_input_block *ib, struct cgraph_node *refering_node, struct varpool_node *refering_varpool_node, VEC(cgraph_node_ptr, heap) *nodes, VEC(varpool_node_ptr, heap) *varpool_nodes) { struct cgraph_node *node = NULL; struct varpool_node *varpool_node = NULL; struct bitpack_d bp; enum ipa_ref_type type; enum ipa_ref_use use; bp = lto_input_bitpack (ib); type = (enum ipa_ref_type) bp_unpack_value (&bp, 1); use = (enum ipa_ref_use) bp_unpack_value (&bp, 2); if (type == IPA_REF_CGRAPH) node = VEC_index (cgraph_node_ptr, nodes, lto_input_sleb128 (ib)); else varpool_node = VEC_index (varpool_node_ptr, varpool_nodes, lto_input_sleb128 (ib)); ipa_record_reference (refering_node, refering_varpool_node, node, varpool_node, use, NULL); } /* Read an edge from IB. NODES points to a vector of previously read nodes for decoding caller and callee of the edge to be read. If INDIRECT is true, the edge being read is indirect (in the sense that it has indirect_unknown_callee set). */ static void input_edge (struct lto_input_block *ib, VEC(cgraph_node_ptr, heap) *nodes, bool indirect) { struct cgraph_node *caller, *callee; struct cgraph_edge *edge; unsigned int stmt_id; gcov_type count; int freq; unsigned int nest; cgraph_inline_failed_t inline_failed; struct bitpack_d bp; int ecf_flags = 0; caller = VEC_index (cgraph_node_ptr, nodes, lto_input_sleb128 (ib)); if (caller == NULL || caller->decl == NULL_TREE) internal_error ("bytecode stream: no caller found while reading edge"); if (!indirect) { callee = VEC_index (cgraph_node_ptr, nodes, lto_input_sleb128 (ib)); if (callee == NULL || callee->decl == NULL_TREE) internal_error ("bytecode stream: no callee found while reading edge"); } else callee = NULL; count = (gcov_type) lto_input_sleb128 (ib); bp = lto_input_bitpack (ib); stmt_id = (unsigned int) bp_unpack_value (&bp, HOST_BITS_PER_INT); inline_failed = (cgraph_inline_failed_t) bp_unpack_value (&bp, HOST_BITS_PER_INT); freq = (int) bp_unpack_value (&bp, HOST_BITS_PER_INT); nest = (unsigned) bp_unpack_value (&bp, 30); if (indirect) edge = cgraph_create_indirect_edge (caller, NULL, 0, count, freq, nest); else edge = cgraph_create_edge (caller, callee, NULL, count, freq, nest); edge->indirect_inlining_edge = bp_unpack_value (&bp, 1); edge->lto_stmt_uid = stmt_id; edge->inline_failed = inline_failed; edge->call_stmt_cannot_inline_p = bp_unpack_value (&bp, 1); edge->can_throw_external = bp_unpack_value (&bp, 1); if (indirect) { if (bp_unpack_value (&bp, 1)) ecf_flags |= ECF_CONST; if (bp_unpack_value (&bp, 1)) ecf_flags |= ECF_PURE; if (bp_unpack_value (&bp, 1)) ecf_flags |= ECF_NORETURN; if (bp_unpack_value (&bp, 1)) ecf_flags |= ECF_MALLOC; if (bp_unpack_value (&bp, 1)) ecf_flags |= ECF_NOTHROW; if (bp_unpack_value (&bp, 1)) ecf_flags |= ECF_RETURNS_TWICE; edge->indirect_info->ecf_flags = ecf_flags; } } /* Read a cgraph from IB using the info in FILE_DATA. */ static VEC(cgraph_node_ptr, heap) * input_cgraph_1 (struct lto_file_decl_data *file_data, struct lto_input_block *ib) { enum LTO_cgraph_tags tag; VEC(cgraph_node_ptr, heap) *nodes = NULL; struct cgraph_node *node; unsigned i; unsigned HOST_WIDE_INT len; tag = (enum LTO_cgraph_tags) lto_input_uleb128 (ib); while (tag) { if (tag == LTO_cgraph_edge) input_edge (ib, nodes, false); else if (tag == LTO_cgraph_indirect_edge) input_edge (ib, nodes, true); else { node = input_node (file_data, ib, tag,nodes); if (node == NULL || node->decl == NULL_TREE) internal_error ("bytecode stream: found empty cgraph node"); VEC_safe_push (cgraph_node_ptr, heap, nodes, node); lto_cgraph_encoder_encode (file_data->cgraph_node_encoder, node); } tag = (enum LTO_cgraph_tags) lto_input_uleb128 (ib); } /* Input toplevel asms. */ len = lto_input_uleb128 (ib); while (len) { char *str = (char *)xmalloc (len + 1); for (i = 0; i < len; ++i) str[i] = lto_input_1_unsigned (ib); cgraph_add_asm_node (build_string (len, str)); free (str); len = lto_input_uleb128 (ib); } /* AUX pointers should be all non-zero for nodes read from the stream. */ #ifdef ENABLE_CHECKING FOR_EACH_VEC_ELT (cgraph_node_ptr, nodes, i, node) gcc_assert (node->aux); #endif FOR_EACH_VEC_ELT (cgraph_node_ptr, nodes, i, node) { int ref = (int) (intptr_t) node->global.inlined_to; /* We share declaration of builtins, so we may read same node twice. */ if (!node->aux) continue; node->aux = NULL; /* Fixup inlined_to from reference to pointer. */ if (ref != LCC_NOT_FOUND) node->global.inlined_to = VEC_index (cgraph_node_ptr, nodes, ref); else node->global.inlined_to = NULL; ref = (int) (intptr_t) node->same_comdat_group; /* Fixup same_comdat_group from reference to pointer. */ if (ref != LCC_NOT_FOUND) node->same_comdat_group = VEC_index (cgraph_node_ptr, nodes, ref); else node->same_comdat_group = NULL; } FOR_EACH_VEC_ELT (cgraph_node_ptr, nodes, i, node) node->aux = (void *)1; return nodes; } /* Read a varpool from IB using the info in FILE_DATA. */ static VEC(varpool_node_ptr, heap) * input_varpool_1 (struct lto_file_decl_data *file_data, struct lto_input_block *ib) { unsigned HOST_WIDE_INT len; VEC(varpool_node_ptr, heap) *varpool = NULL; int i; struct varpool_node *node; len = lto_input_uleb128 (ib); while (len) { VEC_safe_push (varpool_node_ptr, heap, varpool, input_varpool_node (file_data, ib)); len--; } #ifdef ENABLE_CHECKING FOR_EACH_VEC_ELT (varpool_node_ptr, varpool, i, node) gcc_assert (!node->aux); #endif FOR_EACH_VEC_ELT (varpool_node_ptr, varpool, i, node) { int ref = (int) (intptr_t) node->same_comdat_group; /* We share declaration of builtins, so we may read same node twice. */ if (node->aux) continue; node->aux = (void *)1; /* Fixup same_comdat_group from reference to pointer. */ if (ref != LCC_NOT_FOUND) node->same_comdat_group = VEC_index (varpool_node_ptr, varpool, ref); else node->same_comdat_group = NULL; } FOR_EACH_VEC_ELT (varpool_node_ptr, varpool, i, node) node->aux = NULL; return varpool; } /* Input ipa_refs. */ static void input_refs (struct lto_input_block *ib, VEC(cgraph_node_ptr, heap) *nodes, VEC(varpool_node_ptr, heap) *varpool) { int count; int idx; while (true) { struct cgraph_node *node; count = lto_input_uleb128 (ib); if (!count) break; idx = lto_input_uleb128 (ib); node = VEC_index (cgraph_node_ptr, nodes, idx); while (count) { input_ref (ib, node, NULL, nodes, varpool); count--; } } while (true) { struct varpool_node *node; count = lto_input_uleb128 (ib); if (!count) break; node = VEC_index (varpool_node_ptr, varpool, lto_input_uleb128 (ib)); while (count) { input_ref (ib, NULL, node, nodes, varpool); count--; } } } static struct gcov_ctr_summary lto_gcov_summary; /* Input profile_info from IB. */ static void input_profile_summary (struct lto_input_block *ib, struct lto_file_decl_data *file_data) { unsigned int runs = lto_input_uleb128 (ib); if (runs) { file_data->profile_info.runs = runs; file_data->profile_info.sum_max = lto_input_uleb128 (ib); } } /* Rescale profile summaries to the same number of runs in the whole unit. */ static void merge_profile_summaries (struct lto_file_decl_data **file_data_vec) { struct lto_file_decl_data *file_data; unsigned int j; gcov_unsigned_t max_runs = 0; struct cgraph_node *node; struct cgraph_edge *edge; /* Find unit with maximal number of runs. If we ever get serious about roundoff errors, we might also consider computing smallest common multiply. */ for (j = 0; (file_data = file_data_vec[j]) != NULL; j++) if (max_runs < file_data->profile_info.runs) max_runs = file_data->profile_info.runs; if (!max_runs) return; /* Simple overflow check. We probably don't need to support that many train runs. Such a large value probably imply data corruption anyway. */ if (max_runs > INT_MAX / REG_BR_PROB_BASE) { sorry ("At most %i profile runs is supported. Perhaps corrupted profile?", INT_MAX / REG_BR_PROB_BASE); return; } profile_info = <o_gcov_summary; lto_gcov_summary.runs = max_runs; lto_gcov_summary.sum_max = 0; /* Rescale all units to the maximal number of runs. sum_max can not be easily merged, as we have no idea what files come from the same run. We do not use the info anyway, so leave it 0. */ for (j = 0; (file_data = file_data_vec[j]) != NULL; j++) if (file_data->profile_info.runs) { int scale = ((REG_BR_PROB_BASE * max_runs + file_data->profile_info.runs / 2) / file_data->profile_info.runs); lto_gcov_summary.sum_max = MAX (lto_gcov_summary.sum_max, (file_data->profile_info.sum_max * scale + REG_BR_PROB_BASE / 2) / REG_BR_PROB_BASE); } /* Watch roundoff errors. */ if (lto_gcov_summary.sum_max < max_runs) lto_gcov_summary.sum_max = max_runs; /* If merging already happent at WPA time, we are done. */ if (flag_ltrans) return; /* Now compute count_materialization_scale of each node. During LTRANS we already have values of count_materialization_scale computed, so just update them. */ for (node = cgraph_nodes; node; node = node->next) if (node->local.lto_file_data && node->local.lto_file_data->profile_info.runs) { int scale; scale = ((node->count_materialization_scale * max_runs + node->local.lto_file_data->profile_info.runs / 2) / node->local.lto_file_data->profile_info.runs); node->count_materialization_scale = scale; if (scale < 0) fatal_error ("Profile information in %s corrupted", file_data->file_name); if (scale == REG_BR_PROB_BASE) continue; for (edge = node->callees; edge; edge = edge->next_callee) edge->count = ((edge->count * scale + REG_BR_PROB_BASE / 2) / REG_BR_PROB_BASE); node->count = ((node->count * scale + REG_BR_PROB_BASE / 2) / REG_BR_PROB_BASE); } } /* Input and merge the cgraph from each of the .o files passed to lto1. */ void input_cgraph (void) { struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data (); struct lto_file_decl_data *file_data; unsigned int j = 0; struct cgraph_node *node; while ((file_data = file_data_vec[j++])) { const char *data; size_t len; struct lto_input_block *ib; VEC(cgraph_node_ptr, heap) *nodes; VEC(varpool_node_ptr, heap) *varpool; ib = lto_create_simple_input_block (file_data, LTO_section_cgraph, &data, &len); if (!ib) fatal_error ("cannot find LTO cgraph in %s", file_data->file_name); input_profile_summary (ib, file_data); file_data->cgraph_node_encoder = lto_cgraph_encoder_new (); nodes = input_cgraph_1 (file_data, ib); lto_destroy_simple_input_block (file_data, LTO_section_cgraph, ib, data, len); ib = lto_create_simple_input_block (file_data, LTO_section_varpool, &data, &len); if (!ib) fatal_error ("cannot find LTO varpool in %s", file_data->file_name); varpool = input_varpool_1 (file_data, ib); lto_destroy_simple_input_block (file_data, LTO_section_varpool, ib, data, len); ib = lto_create_simple_input_block (file_data, LTO_section_refs, &data, &len); if (!ib) fatal_error("cannot find LTO section refs in %s", file_data->file_name); input_refs (ib, nodes, varpool); lto_destroy_simple_input_block (file_data, LTO_section_refs, ib, data, len); if (flag_ltrans) input_cgraph_opt_summary (nodes); VEC_free (cgraph_node_ptr, heap, nodes); VEC_free (varpool_node_ptr, heap, varpool); } merge_profile_summaries (file_data_vec); /* Clear out the aux field that was used to store enough state to tell which nodes should be overwritten. */ for (node = cgraph_nodes; node; node = node->next) { /* Some nodes may have been created by cgraph_node. This happens when the callgraph contains nested functions. If the node for the parent function was never emitted to the gimple file, cgraph_node will create a node for it when setting the context of the nested function. */ if (node->local.lto_file_data) node->aux = NULL; } } /* True when we need optimization summary for NODE. */ static int output_cgraph_opt_summary_p (struct cgraph_node *node, cgraph_node_set set) { struct cgraph_edge *e; if (cgraph_node_in_set_p (node, set)) { for (e = node->callees; e; e = e->next_callee) if (e->indirect_info && e->indirect_info->thunk_delta != 0) return true; for (e = node->indirect_calls; e; e = e->next_callee) if (e->indirect_info->thunk_delta != 0) return true; } return (node->clone_of && (node->clone.tree_map || node->clone.args_to_skip || node->clone.combined_args_to_skip)); } /* Output optimization summary for EDGE to OB. */ static void output_edge_opt_summary (struct output_block *ob, struct cgraph_edge *edge) { if (edge->indirect_info) lto_output_sleb128_stream (ob->main_stream, edge->indirect_info->thunk_delta); else lto_output_sleb128_stream (ob->main_stream, 0); } /* Output optimization summary for NODE to OB. */ static void output_node_opt_summary (struct output_block *ob, struct cgraph_node *node, cgraph_node_set set) { unsigned int index; bitmap_iterator bi; struct ipa_replace_map *map; struct bitpack_d bp; int i; struct cgraph_edge *e; lto_output_uleb128_stream (ob->main_stream, bitmap_count_bits (node->clone.args_to_skip)); EXECUTE_IF_SET_IN_BITMAP (node->clone.args_to_skip, 0, index, bi) lto_output_uleb128_stream (ob->main_stream, index); lto_output_uleb128_stream (ob->main_stream, bitmap_count_bits (node->clone.combined_args_to_skip)); EXECUTE_IF_SET_IN_BITMAP (node->clone.combined_args_to_skip, 0, index, bi) lto_output_uleb128_stream (ob->main_stream, index); lto_output_uleb128_stream (ob->main_stream, VEC_length (ipa_replace_map_p, node->clone.tree_map)); FOR_EACH_VEC_ELT (ipa_replace_map_p, node->clone.tree_map, i, map) { int parm_num; tree parm; for (parm_num = 0, parm = DECL_ARGUMENTS (node->decl); parm; parm = DECL_CHAIN (parm), parm_num++) if (map->old_tree == parm) break; /* At the moment we assume all old trees to be PARM_DECLs, because we have no mechanism to store function local declarations into summaries. */ gcc_assert (parm); lto_output_uleb128_stream (ob->main_stream, parm_num); lto_output_tree (ob, map->new_tree, true); bp = bitpack_create (ob->main_stream); bp_pack_value (&bp, map->replace_p, 1); bp_pack_value (&bp, map->ref_p, 1); lto_output_bitpack (&bp); } if (cgraph_node_in_set_p (node, set)) { for (e = node->callees; e; e = e->next_callee) output_edge_opt_summary (ob, e); for (e = node->indirect_calls; e; e = e->next_callee) output_edge_opt_summary (ob, e); } } /* Output optimization summaries stored in callgraph. At the moment it is the clone info structure. */ static void output_cgraph_opt_summary (cgraph_node_set set) { struct cgraph_node *node; int i, n_nodes; lto_cgraph_encoder_t encoder; struct output_block *ob = create_output_block (LTO_section_cgraph_opt_sum); unsigned count = 0; ob->cgraph_node = NULL; encoder = ob->decl_state->cgraph_node_encoder; n_nodes = lto_cgraph_encoder_size (encoder); for (i = 0; i < n_nodes; i++) if (output_cgraph_opt_summary_p (lto_cgraph_encoder_deref (encoder, i), set)) count++; lto_output_uleb128_stream (ob->main_stream, count); for (i = 0; i < n_nodes; i++) { node = lto_cgraph_encoder_deref (encoder, i); if (output_cgraph_opt_summary_p (node, set)) { lto_output_uleb128_stream (ob->main_stream, i); output_node_opt_summary (ob, node, set); } } produce_asm (ob, NULL); destroy_output_block (ob); } /* Input optimisation summary of EDGE. */ static void input_edge_opt_summary (struct cgraph_edge *edge, struct lto_input_block *ib_main) { HOST_WIDE_INT thunk_delta; thunk_delta = lto_input_sleb128 (ib_main); if (thunk_delta != 0) { gcc_assert (!edge->indirect_info); edge->indirect_info = cgraph_allocate_init_indirect_info (); edge->indirect_info->thunk_delta = thunk_delta; } } /* Input optimisation summary of NODE. */ static void input_node_opt_summary (struct cgraph_node *node, struct lto_input_block *ib_main, struct data_in *data_in) { int i; int count; int bit; struct bitpack_d bp; struct cgraph_edge *e; count = lto_input_uleb128 (ib_main); if (count) node->clone.args_to_skip = BITMAP_GGC_ALLOC (); for (i = 0; i < count; i++) { bit = lto_input_uleb128 (ib_main); bitmap_set_bit (node->clone.args_to_skip, bit); } count = lto_input_uleb128 (ib_main); if (count) node->clone.combined_args_to_skip = BITMAP_GGC_ALLOC (); for (i = 0; i < count; i++) { bit = lto_input_uleb128 (ib_main); bitmap_set_bit (node->clone.combined_args_to_skip, bit); } count = lto_input_uleb128 (ib_main); for (i = 0; i < count; i++) { int parm_num; tree parm; struct ipa_replace_map *map = ggc_alloc_ipa_replace_map (); VEC_safe_push (ipa_replace_map_p, gc, node->clone.tree_map, map); for (parm_num = 0, parm = DECL_ARGUMENTS (node->decl); parm_num; parm = DECL_CHAIN (parm)) parm_num --; map->parm_num = lto_input_uleb128 (ib_main); map->old_tree = NULL; map->new_tree = lto_input_tree (ib_main, data_in); bp = lto_input_bitpack (ib_main); map->replace_p = bp_unpack_value (&bp, 1); map->ref_p = bp_unpack_value (&bp, 1); } for (e = node->callees; e; e = e->next_callee) input_edge_opt_summary (e, ib_main); for (e = node->indirect_calls; e; e = e->next_callee) input_edge_opt_summary (e, ib_main); } /* Read section in file FILE_DATA of length LEN with data DATA. */ static void input_cgraph_opt_section (struct lto_file_decl_data *file_data, const char *data, size_t len, VEC (cgraph_node_ptr, heap) * nodes) { const struct lto_function_header *header = (const struct lto_function_header *) data; const int cfg_offset = sizeof (struct lto_function_header); const int main_offset = cfg_offset + header->cfg_size; const int string_offset = main_offset + header->main_size; struct data_in *data_in; struct lto_input_block ib_main; unsigned int i; unsigned int count; LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0, header->main_size); data_in = lto_data_in_create (file_data, (const char *) data + string_offset, header->string_size, NULL); count = lto_input_uleb128 (&ib_main); for (i = 0; i < count; i++) { int ref = lto_input_uleb128 (&ib_main); input_node_opt_summary (VEC_index (cgraph_node_ptr, nodes, ref), &ib_main, data_in); } lto_free_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, data, len); lto_data_in_delete (data_in); } /* Input optimization summary of cgraph. */ static void input_cgraph_opt_summary (VEC (cgraph_node_ptr, heap) * nodes) { struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data (); struct lto_file_decl_data *file_data; unsigned int j = 0; while ((file_data = file_data_vec[j++])) { size_t len; const char *data = lto_get_section_data (file_data, LTO_section_cgraph_opt_sum, NULL, &len); if (data) input_cgraph_opt_section (file_data, data, len, nodes); } }