diff options
Diffstat (limited to 'gcc-4.8/gcc/lto/lto.c')
-rw-r--r-- | gcc-4.8/gcc/lto/lto.c | 3419 |
1 files changed, 3419 insertions, 0 deletions
diff --git a/gcc-4.8/gcc/lto/lto.c b/gcc-4.8/gcc/lto/lto.c new file mode 100644 index 000000000..983fa03aa --- /dev/null +++ b/gcc-4.8/gcc/lto/lto.c @@ -0,0 +1,3419 @@ +/* Top-level LTO routines. + Copyright (C) 2009-2013 Free Software Foundation, Inc. + Contributed by CodeSourcery, Inc. + +This file is part of GCC. + +GCC is free software; you can redistribute it and/or modify it under +the terms of the GNU General Public License as published by the Free +Software Foundation; either version 3, or (at your option) any later +version. + +GCC is distributed in the hope that it will be useful, but WITHOUT ANY +WARRANTY; without even the implied warranty of MERCHANTABILITY or +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License +for more details. + +You should have received a copy of the GNU General Public License +along with GCC; see the file COPYING3. If not see +<http://www.gnu.org/licenses/>. */ + +#include "config.h" +#include "system.h" +#include "coretypes.h" +#include "opts.h" +#include "toplev.h" +#include "tree.h" +#include "tree-flow.h" +#include "diagnostic-core.h" +#include "tm.h" +#include "cgraph.h" +#include "ggc.h" +#include "tree-ssa-operands.h" +#include "tree-pass.h" +#include "langhooks.h" +#include "vec.h" +#include "bitmap.h" +#include "pointer-set.h" +#include "ipa-prop.h" +#include "common.h" +#include "debug.h" +#include "gimple.h" +#include "lto.h" +#include "lto-tree.h" +#include "lto-streamer.h" +#include "tree-streamer.h" +#include "splay-tree.h" +#include "lto-partition.h" + +static GTY(()) tree first_personality_decl; + +/* Returns a hash code for P. */ + +static hashval_t +hash_name (const void *p) +{ + const struct lto_section_slot *ds = (const struct lto_section_slot *) p; + return (hashval_t) htab_hash_string (ds->name); +} + + +/* Returns nonzero if P1 and P2 are equal. */ + +static int +eq_name (const void *p1, const void *p2) +{ + const struct lto_section_slot *s1 = + (const struct lto_section_slot *) p1; + const struct lto_section_slot *s2 = + (const struct lto_section_slot *) p2; + + return strcmp (s1->name, s2->name) == 0; +} + +/* Free lto_section_slot */ + +static void +free_with_string (void *arg) +{ + struct lto_section_slot *s = (struct lto_section_slot *)arg; + + free (CONST_CAST (char *, s->name)); + free (arg); +} + +/* Create section hash table */ + +htab_t +lto_obj_create_section_hash_table (void) +{ + return htab_create (37, hash_name, eq_name, free_with_string); +} + +/* Delete an allocated integer KEY in the splay tree. */ + +static void +lto_splay_tree_delete_id (splay_tree_key key) +{ + free ((void *) key); +} + +/* Compare splay tree node ids A and B. */ + +static int +lto_splay_tree_compare_ids (splay_tree_key a, splay_tree_key b) +{ + unsigned HOST_WIDE_INT ai; + unsigned HOST_WIDE_INT bi; + + ai = *(unsigned HOST_WIDE_INT *) a; + bi = *(unsigned HOST_WIDE_INT *) b; + + if (ai < bi) + return -1; + else if (ai > bi) + return 1; + return 0; +} + +/* Look up splay tree node by ID in splay tree T. */ + +static splay_tree_node +lto_splay_tree_lookup (splay_tree t, unsigned HOST_WIDE_INT id) +{ + return splay_tree_lookup (t, (splay_tree_key) &id); +} + +/* Check if KEY has ID. */ + +static bool +lto_splay_tree_id_equal_p (splay_tree_key key, unsigned HOST_WIDE_INT id) +{ + return *(unsigned HOST_WIDE_INT *) key == id; +} + +/* Insert a splay tree node into tree T with ID as key and FILE_DATA as value. + The ID is allocated separately because we need HOST_WIDE_INTs which may + be wider than a splay_tree_key. */ + +static void +lto_splay_tree_insert (splay_tree t, unsigned HOST_WIDE_INT id, + struct lto_file_decl_data *file_data) +{ + unsigned HOST_WIDE_INT *idp = XCNEW (unsigned HOST_WIDE_INT); + *idp = id; + splay_tree_insert (t, (splay_tree_key) idp, (splay_tree_value) file_data); +} + +/* Create a splay tree. */ + +static splay_tree +lto_splay_tree_new (void) +{ + return splay_tree_new (lto_splay_tree_compare_ids, + lto_splay_tree_delete_id, + NULL); +} + +/* Return true when NODE has a clone that is analyzed (i.e. we need + to load its body even if the node itself is not needed). */ + +static bool +has_analyzed_clone_p (struct cgraph_node *node) +{ + struct cgraph_node *orig = node; + node = node->clones; + if (node) + while (node != orig) + { + if (node->analyzed) + return true; + if (node->clones) + node = node->clones; + else if (node->next_sibling_clone) + node = node->next_sibling_clone; + else + { + while (node != orig && !node->next_sibling_clone) + node = node->clone_of; + if (node != orig) + node = node->next_sibling_clone; + } + } + return false; +} + +/* Read the function body for the function associated with NODE. */ + +static void +lto_materialize_function (struct cgraph_node *node) +{ + tree decl; + struct lto_file_decl_data *file_data; + const char *data, *name; + size_t len; + + decl = node->symbol.decl; + /* Read in functions with body (analyzed nodes) + and also functions that are needed to produce virtual clones. */ + if (cgraph_function_with_gimple_body_p (node) || has_analyzed_clone_p (node)) + { + /* Clones don't need to be read. */ + if (node->clone_of) + return; + + /* Load the function body only if not operating in WPA mode. In + WPA mode, the body of the function is not needed. */ + if (!flag_wpa) + { + file_data = node->symbol.lto_file_data; + name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)); + + /* We may have renamed the declaration, e.g., a static function. */ + name = lto_get_decl_name_mapping (file_data, name); + + data = lto_get_section_data (file_data, LTO_section_function_body, + name, &len); + if (!data) + fatal_error ("%s: section %s is missing", + file_data->file_name, + name); + + gcc_assert (DECL_STRUCT_FUNCTION (decl) == NULL); + + push_struct_function (decl); + announce_function (decl); + lto_input_function_body (file_data, decl, data); + if (DECL_FUNCTION_PERSONALITY (decl) && !first_personality_decl) + first_personality_decl = DECL_FUNCTION_PERSONALITY (decl); + lto_stats.num_function_bodies++; + lto_free_section_data (file_data, LTO_section_function_body, name, + data, len); + pop_cfun (); + ggc_collect (); + } + } + + /* Let the middle end know about the function. */ + rest_of_decl_compilation (decl, 1, 0); +} + + +/* Decode the content of memory pointed to by DATA in the in decl + state object STATE. DATA_IN points to a data_in structure for + decoding. Return the address after the decoded object in the + input. */ + +static const uint32_t * +lto_read_in_decl_state (struct data_in *data_in, const uint32_t *data, + struct lto_in_decl_state *state) +{ + uint32_t ix; + tree decl; + uint32_t i, j; + + ix = *data++; + decl = streamer_tree_cache_get (data_in->reader_cache, ix); + if (TREE_CODE (decl) != FUNCTION_DECL) + { + gcc_assert (decl == void_type_node); + decl = NULL_TREE; + } + state->fn_decl = decl; + + for (i = 0; i < LTO_N_DECL_STREAMS; i++) + { + uint32_t size = *data++; + tree *decls = ggc_alloc_vec_tree (size); + + for (j = 0; j < size; j++) + decls[j] = streamer_tree_cache_get (data_in->reader_cache, data[j]); + + state->streams[i].size = size; + state->streams[i].trees = decls; + data += size; + } + + return data; +} + + + +/* Global type table. FIXME, it should be possible to re-use some + of the type hashing routines in tree.c (type_hash_canon, type_hash_lookup, + etc), but those assume that types were built with the various + build_*_type routines which is not the case with the streamer. */ +static GTY((if_marked ("ggc_marked_p"), param_is (union tree_node))) + htab_t gimple_types; +static GTY((if_marked ("tree_int_map_marked_p"), param_is (struct tree_int_map))) + htab_t type_hash_cache; + +static hashval_t gimple_type_hash (const void *); + +/* Structure used to maintain a cache of some type pairs compared by + gimple_types_compatible_p when comparing aggregate types. There are + three possible values for SAME_P: + + -2: The pair (T1, T2) has just been inserted in the table. + 0: T1 and T2 are different types. + 1: T1 and T2 are the same type. */ + +struct type_pair_d +{ + unsigned int uid1; + unsigned int uid2; + signed char same_p; +}; +typedef struct type_pair_d *type_pair_t; + +#define GIMPLE_TYPE_PAIR_SIZE 16381 +struct type_pair_d *type_pair_cache; + + +/* Lookup the pair of types T1 and T2 in *VISITED_P. Insert a new + entry if none existed. */ + +static inline type_pair_t +lookup_type_pair (tree t1, tree t2) +{ + unsigned int index; + unsigned int uid1, uid2; + + if (TYPE_UID (t1) < TYPE_UID (t2)) + { + uid1 = TYPE_UID (t1); + uid2 = TYPE_UID (t2); + } + else + { + uid1 = TYPE_UID (t2); + uid2 = TYPE_UID (t1); + } + gcc_checking_assert (uid1 != uid2); + + /* iterative_hash_hashval_t imply an function calls. + We know that UIDS are in limited range. */ + index = ((((unsigned HOST_WIDE_INT)uid1 << HOST_BITS_PER_WIDE_INT / 2) + uid2) + % GIMPLE_TYPE_PAIR_SIZE); + if (type_pair_cache [index].uid1 == uid1 + && type_pair_cache [index].uid2 == uid2) + return &type_pair_cache[index]; + + type_pair_cache [index].uid1 = uid1; + type_pair_cache [index].uid2 = uid2; + type_pair_cache [index].same_p = -2; + + return &type_pair_cache[index]; +} + +/* Per pointer state for the SCC finding. The on_sccstack flag + is not strictly required, it is true when there is no hash value + recorded for the type and false otherwise. But querying that + is slower. */ + +struct sccs +{ + unsigned int dfsnum; + unsigned int low; + bool on_sccstack; + union { + hashval_t hash; + signed char same_p; + } u; +}; + +static unsigned int next_dfs_num; +static unsigned int gtc_next_dfs_num; + +/* GIMPLE type merging cache. A direct-mapped cache based on TYPE_UID. */ + +typedef struct GTY(()) gimple_type_leader_entry_s { + tree type; + tree leader; +} gimple_type_leader_entry; + +#define GIMPLE_TYPE_LEADER_SIZE 16381 +static GTY((length("GIMPLE_TYPE_LEADER_SIZE"))) + gimple_type_leader_entry *gimple_type_leader; + +/* Lookup an existing leader for T and return it or NULL_TREE, if + there is none in the cache. */ + +static inline tree +gimple_lookup_type_leader (tree t) +{ + gimple_type_leader_entry *leader; + + leader = &gimple_type_leader[TYPE_UID (t) % GIMPLE_TYPE_LEADER_SIZE]; + if (leader->type != t) + return NULL_TREE; + + return leader->leader; +} + + +/* Return true if T1 and T2 have the same name. If FOR_COMPLETION_P is + true then if any type has no name return false, otherwise return + true if both types have no names. */ + +static bool +compare_type_names_p (tree t1, tree t2) +{ + tree name1 = TYPE_NAME (t1); + tree name2 = TYPE_NAME (t2); + + if ((name1 != NULL_TREE) != (name2 != NULL_TREE)) + return false; + + if (name1 == NULL_TREE) + return true; + + /* Either both should be a TYPE_DECL or both an IDENTIFIER_NODE. */ + if (TREE_CODE (name1) != TREE_CODE (name2)) + return false; + + if (TREE_CODE (name1) == TYPE_DECL) + name1 = DECL_NAME (name1); + gcc_checking_assert (!name1 || TREE_CODE (name1) == IDENTIFIER_NODE); + + if (TREE_CODE (name2) == TYPE_DECL) + name2 = DECL_NAME (name2); + gcc_checking_assert (!name2 || TREE_CODE (name2) == IDENTIFIER_NODE); + + /* Identifiers can be compared with pointer equality rather + than a string comparison. */ + if (name1 == name2) + return true; + + return false; +} + +static bool +gimple_types_compatible_p_1 (tree, tree, type_pair_t, + vec<type_pair_t> *, + struct pointer_map_t *, struct obstack *); + +/* DFS visit the edge from the callers type pair with state *STATE to + the pair T1, T2 while operating in FOR_MERGING_P mode. + Update the merging status if it is not part of the SCC containing the + callers pair and return it. + SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. */ + +static bool +gtc_visit (tree t1, tree t2, + struct sccs *state, + vec<type_pair_t> *sccstack, + struct pointer_map_t *sccstate, + struct obstack *sccstate_obstack) +{ + struct sccs *cstate = NULL; + type_pair_t p; + void **slot; + tree leader1, leader2; + + /* Check first for the obvious case of pointer identity. */ + if (t1 == t2) + return true; + + /* Check that we have two types to compare. */ + if (t1 == NULL_TREE || t2 == NULL_TREE) + return false; + + /* Can't be the same type if the types don't have the same code. */ + if (TREE_CODE (t1) != TREE_CODE (t2)) + return false; + + /* Can't be the same type if they have different CV qualifiers. */ + if (TYPE_QUALS (t1) != TYPE_QUALS (t2)) + return false; + + if (TREE_ADDRESSABLE (t1) != TREE_ADDRESSABLE (t2)) + return false; + + /* Void types and nullptr types are always the same. */ + if (TREE_CODE (t1) == VOID_TYPE + || TREE_CODE (t1) == NULLPTR_TYPE) + return true; + + /* Can't be the same type if they have different alignment or mode. */ + if (TYPE_ALIGN (t1) != TYPE_ALIGN (t2) + || TYPE_MODE (t1) != TYPE_MODE (t2)) + return false; + + /* Do some simple checks before doing three hashtable queries. */ + if (INTEGRAL_TYPE_P (t1) + || SCALAR_FLOAT_TYPE_P (t1) + || FIXED_POINT_TYPE_P (t1) + || TREE_CODE (t1) == VECTOR_TYPE + || TREE_CODE (t1) == COMPLEX_TYPE + || TREE_CODE (t1) == OFFSET_TYPE + || POINTER_TYPE_P (t1)) + { + /* Can't be the same type if they have different sign or precision. */ + if (TYPE_PRECISION (t1) != TYPE_PRECISION (t2) + || TYPE_UNSIGNED (t1) != TYPE_UNSIGNED (t2)) + return false; + + if (TREE_CODE (t1) == INTEGER_TYPE + && TYPE_STRING_FLAG (t1) != TYPE_STRING_FLAG (t2)) + return false; + + /* That's all we need to check for float and fixed-point types. */ + if (SCALAR_FLOAT_TYPE_P (t1) + || FIXED_POINT_TYPE_P (t1)) + return true; + + /* For other types fall through to more complex checks. */ + } + + /* If the types have been previously registered and found equal + they still are. */ + leader1 = gimple_lookup_type_leader (t1); + leader2 = gimple_lookup_type_leader (t2); + if (leader1 == t2 + || t1 == leader2 + || (leader1 && leader1 == leader2)) + return true; + + /* If the hash values of t1 and t2 are different the types can't + possibly be the same. This helps keeping the type-pair hashtable + small, only tracking comparisons for hash collisions. */ + if (gimple_type_hash (t1) != gimple_type_hash (t2)) + return false; + + /* Allocate a new cache entry for this comparison. */ + p = lookup_type_pair (t1, t2); + if (p->same_p == 0 || p->same_p == 1) + { + /* We have already decided whether T1 and T2 are the + same, return the cached result. */ + return p->same_p == 1; + } + + if ((slot = pointer_map_contains (sccstate, p)) != NULL) + cstate = (struct sccs *)*slot; + /* Not yet visited. DFS recurse. */ + if (!cstate) + { + gimple_types_compatible_p_1 (t1, t2, p, + sccstack, sccstate, sccstate_obstack); + cstate = (struct sccs *)* pointer_map_contains (sccstate, p); + state->low = MIN (state->low, cstate->low); + } + /* If the type is still on the SCC stack adjust the parents low. */ + if (cstate->dfsnum < state->dfsnum + && cstate->on_sccstack) + state->low = MIN (cstate->dfsnum, state->low); + + /* Return the current lattice value. We start with an equality + assumption so types part of a SCC will be optimistically + treated equal unless proven otherwise. */ + return cstate->u.same_p; +} + +/* Worker for gimple_types_compatible. + SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. */ + +static bool +gimple_types_compatible_p_1 (tree t1, tree t2, type_pair_t p, + vec<type_pair_t> *sccstack, + struct pointer_map_t *sccstate, + struct obstack *sccstate_obstack) +{ + struct sccs *state; + + gcc_assert (p->same_p == -2); + + state = XOBNEW (sccstate_obstack, struct sccs); + *pointer_map_insert (sccstate, p) = state; + + sccstack->safe_push (p); + state->dfsnum = gtc_next_dfs_num++; + state->low = state->dfsnum; + state->on_sccstack = true; + /* Start with an equality assumption. As we DFS recurse into child + SCCs this assumption may get revisited. */ + state->u.same_p = 1; + + /* The struct tags shall compare equal. */ + if (!compare_type_names_p (t1, t2)) + goto different_types; + + /* The main variant of both types should compare equal. */ + if (TYPE_MAIN_VARIANT (t1) != t1 + || TYPE_MAIN_VARIANT (t2) != t2) + { + if (!gtc_visit (TYPE_MAIN_VARIANT (t1), TYPE_MAIN_VARIANT (t2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + } + + /* We may not merge typedef types to the same type in different + contexts. */ + if (TYPE_NAME (t1) + && TREE_CODE (TYPE_NAME (t1)) == TYPE_DECL + && DECL_CONTEXT (TYPE_NAME (t1)) + && TYPE_P (DECL_CONTEXT (TYPE_NAME (t1)))) + { + if (!gtc_visit (DECL_CONTEXT (TYPE_NAME (t1)), + DECL_CONTEXT (TYPE_NAME (t2)), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + } + + /* If their attributes are not the same they can't be the same type. */ + if (!attribute_list_equal (TYPE_ATTRIBUTES (t1), TYPE_ATTRIBUTES (t2))) + goto different_types; + + /* Do type-specific comparisons. */ + switch (TREE_CODE (t1)) + { + case VECTOR_TYPE: + case COMPLEX_TYPE: + if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + goto same_types; + + case ARRAY_TYPE: + /* Array types are the same if the element types are the same and + the number of elements are the same. */ + if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2), + state, sccstack, sccstate, sccstate_obstack) + || TYPE_STRING_FLAG (t1) != TYPE_STRING_FLAG (t2) + || TYPE_NONALIASED_COMPONENT (t1) != TYPE_NONALIASED_COMPONENT (t2)) + goto different_types; + else + { + tree i1 = TYPE_DOMAIN (t1); + tree i2 = TYPE_DOMAIN (t2); + + /* For an incomplete external array, the type domain can be + NULL_TREE. Check this condition also. */ + if (i1 == NULL_TREE && i2 == NULL_TREE) + goto same_types; + else if (i1 == NULL_TREE || i2 == NULL_TREE) + goto different_types; + else + { + tree min1 = TYPE_MIN_VALUE (i1); + tree min2 = TYPE_MIN_VALUE (i2); + tree max1 = TYPE_MAX_VALUE (i1); + tree max2 = TYPE_MAX_VALUE (i2); + + /* The minimum/maximum values have to be the same. */ + if ((min1 == min2 + || (min1 && min2 + && ((TREE_CODE (min1) == PLACEHOLDER_EXPR + && TREE_CODE (min2) == PLACEHOLDER_EXPR) + || operand_equal_p (min1, min2, 0)))) + && (max1 == max2 + || (max1 && max2 + && ((TREE_CODE (max1) == PLACEHOLDER_EXPR + && TREE_CODE (max2) == PLACEHOLDER_EXPR) + || operand_equal_p (max1, max2, 0))))) + goto same_types; + else + goto different_types; + } + } + + case METHOD_TYPE: + /* Method types should belong to the same class. */ + if (!gtc_visit (TYPE_METHOD_BASETYPE (t1), TYPE_METHOD_BASETYPE (t2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + + /* Fallthru */ + + case FUNCTION_TYPE: + /* Function types are the same if the return type and arguments types + are the same. */ + if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + + if (!comp_type_attributes (t1, t2)) + goto different_types; + + if (TYPE_ARG_TYPES (t1) == TYPE_ARG_TYPES (t2)) + goto same_types; + else + { + tree parms1, parms2; + + for (parms1 = TYPE_ARG_TYPES (t1), parms2 = TYPE_ARG_TYPES (t2); + parms1 && parms2; + parms1 = TREE_CHAIN (parms1), parms2 = TREE_CHAIN (parms2)) + { + if (!gtc_visit (TREE_VALUE (parms1), TREE_VALUE (parms2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + } + + if (parms1 || parms2) + goto different_types; + + goto same_types; + } + + case OFFSET_TYPE: + { + if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2), + state, sccstack, sccstate, sccstate_obstack) + || !gtc_visit (TYPE_OFFSET_BASETYPE (t1), + TYPE_OFFSET_BASETYPE (t2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + + goto same_types; + } + + case POINTER_TYPE: + case REFERENCE_TYPE: + { + /* If the two pointers have different ref-all attributes, + they can't be the same type. */ + if (TYPE_REF_CAN_ALIAS_ALL (t1) != TYPE_REF_CAN_ALIAS_ALL (t2)) + goto different_types; + + /* Otherwise, pointer and reference types are the same if the + pointed-to types are the same. */ + if (gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2), + state, sccstack, sccstate, sccstate_obstack)) + goto same_types; + + goto different_types; + } + + case INTEGER_TYPE: + case BOOLEAN_TYPE: + { + tree min1 = TYPE_MIN_VALUE (t1); + tree max1 = TYPE_MAX_VALUE (t1); + tree min2 = TYPE_MIN_VALUE (t2); + tree max2 = TYPE_MAX_VALUE (t2); + bool min_equal_p = false; + bool max_equal_p = false; + + /* If either type has a minimum value, the other type must + have the same. */ + if (min1 == NULL_TREE && min2 == NULL_TREE) + min_equal_p = true; + else if (min1 && min2 && operand_equal_p (min1, min2, 0)) + min_equal_p = true; + + /* Likewise, if either type has a maximum value, the other + type must have the same. */ + if (max1 == NULL_TREE && max2 == NULL_TREE) + max_equal_p = true; + else if (max1 && max2 && operand_equal_p (max1, max2, 0)) + max_equal_p = true; + + if (!min_equal_p || !max_equal_p) + goto different_types; + + goto same_types; + } + + case ENUMERAL_TYPE: + { + /* FIXME lto, we cannot check bounds on enumeral types because + different front ends will produce different values. + In C, enumeral types are integers, while in C++ each element + will have its own symbolic value. We should decide how enums + are to be represented in GIMPLE and have each front end lower + to that. */ + tree v1, v2; + + /* For enumeral types, all the values must be the same. */ + if (TYPE_VALUES (t1) == TYPE_VALUES (t2)) + goto same_types; + + for (v1 = TYPE_VALUES (t1), v2 = TYPE_VALUES (t2); + v1 && v2; + v1 = TREE_CHAIN (v1), v2 = TREE_CHAIN (v2)) + { + tree c1 = TREE_VALUE (v1); + tree c2 = TREE_VALUE (v2); + + if (TREE_CODE (c1) == CONST_DECL) + c1 = DECL_INITIAL (c1); + + if (TREE_CODE (c2) == CONST_DECL) + c2 = DECL_INITIAL (c2); + + if (tree_int_cst_equal (c1, c2) != 1) + goto different_types; + + if (TREE_PURPOSE (v1) != TREE_PURPOSE (v2)) + goto different_types; + } + + /* If one enumeration has more values than the other, they + are not the same. */ + if (v1 || v2) + goto different_types; + + goto same_types; + } + + case RECORD_TYPE: + case UNION_TYPE: + case QUAL_UNION_TYPE: + { + tree f1, f2; + + /* For aggregate types, all the fields must be the same. */ + for (f1 = TYPE_FIELDS (t1), f2 = TYPE_FIELDS (t2); + f1 && f2; + f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2)) + { + /* Different field kinds are not compatible. */ + if (TREE_CODE (f1) != TREE_CODE (f2)) + goto different_types; + /* Field decls must have the same name and offset. */ + if (TREE_CODE (f1) == FIELD_DECL + && (DECL_NONADDRESSABLE_P (f1) != DECL_NONADDRESSABLE_P (f2) + || !gimple_compare_field_offset (f1, f2))) + goto different_types; + /* All entities should have the same name and type. */ + if (DECL_NAME (f1) != DECL_NAME (f2) + || !gtc_visit (TREE_TYPE (f1), TREE_TYPE (f2), + state, sccstack, sccstate, sccstate_obstack)) + goto different_types; + } + + /* If one aggregate has more fields than the other, they + are not the same. */ + if (f1 || f2) + goto different_types; + + goto same_types; + } + + default: + gcc_unreachable (); + } + + /* Common exit path for types that are not compatible. */ +different_types: + state->u.same_p = 0; + goto pop; + + /* Common exit path for types that are compatible. */ +same_types: + gcc_assert (state->u.same_p == 1); + +pop: + if (state->low == state->dfsnum) + { + type_pair_t x; + + /* Pop off the SCC and set its cache values to the final + comparison result. */ + do + { + struct sccs *cstate; + x = sccstack->pop (); + cstate = (struct sccs *)*pointer_map_contains (sccstate, x); + cstate->on_sccstack = false; + x->same_p = state->u.same_p; + } + while (x != p); + } + + return state->u.same_p; +} + +/* Return true iff T1 and T2 are structurally identical. When + FOR_MERGING_P is true the an incomplete type and a complete type + are considered different, otherwise they are considered compatible. */ + +static bool +gimple_types_compatible_p (tree t1, tree t2) +{ + vec<type_pair_t> sccstack = vNULL; + struct pointer_map_t *sccstate; + struct obstack sccstate_obstack; + type_pair_t p = NULL; + bool res; + tree leader1, leader2; + + /* Before starting to set up the SCC machinery handle simple cases. */ + + /* Check first for the obvious case of pointer identity. */ + if (t1 == t2) + return true; + + /* Check that we have two types to compare. */ + if (t1 == NULL_TREE || t2 == NULL_TREE) + return false; + + /* Can't be the same type if the types don't have the same code. */ + if (TREE_CODE (t1) != TREE_CODE (t2)) + return false; + + /* Can't be the same type if they have different CV qualifiers. */ + if (TYPE_QUALS (t1) != TYPE_QUALS (t2)) + return false; + + if (TREE_ADDRESSABLE (t1) != TREE_ADDRESSABLE (t2)) + return false; + + /* Void types and nullptr types are always the same. */ + if (TREE_CODE (t1) == VOID_TYPE + || TREE_CODE (t1) == NULLPTR_TYPE) + return true; + + /* Can't be the same type if they have different alignment or mode. */ + if (TYPE_ALIGN (t1) != TYPE_ALIGN (t2) + || TYPE_MODE (t1) != TYPE_MODE (t2)) + return false; + + /* Do some simple checks before doing three hashtable queries. */ + if (INTEGRAL_TYPE_P (t1) + || SCALAR_FLOAT_TYPE_P (t1) + || FIXED_POINT_TYPE_P (t1) + || TREE_CODE (t1) == VECTOR_TYPE + || TREE_CODE (t1) == COMPLEX_TYPE + || TREE_CODE (t1) == OFFSET_TYPE + || POINTER_TYPE_P (t1)) + { + /* Can't be the same type if they have different sign or precision. */ + if (TYPE_PRECISION (t1) != TYPE_PRECISION (t2) + || TYPE_UNSIGNED (t1) != TYPE_UNSIGNED (t2)) + return false; + + if (TREE_CODE (t1) == INTEGER_TYPE + && TYPE_STRING_FLAG (t1) != TYPE_STRING_FLAG (t2)) + return false; + + /* That's all we need to check for float and fixed-point types. */ + if (SCALAR_FLOAT_TYPE_P (t1) + || FIXED_POINT_TYPE_P (t1)) + return true; + + /* For other types fall through to more complex checks. */ + } + + /* If the types have been previously registered and found equal + they still are. */ + leader1 = gimple_lookup_type_leader (t1); + leader2 = gimple_lookup_type_leader (t2); + if (leader1 == t2 + || t1 == leader2 + || (leader1 && leader1 == leader2)) + return true; + + /* If the hash values of t1 and t2 are different the types can't + possibly be the same. This helps keeping the type-pair hashtable + small, only tracking comparisons for hash collisions. */ + if (gimple_type_hash (t1) != gimple_type_hash (t2)) + return false; + + /* If we've visited this type pair before (in the case of aggregates + with self-referential types), and we made a decision, return it. */ + p = lookup_type_pair (t1, t2); + if (p->same_p == 0 || p->same_p == 1) + { + /* We have already decided whether T1 and T2 are the + same, return the cached result. */ + return p->same_p == 1; + } + + /* Now set up the SCC machinery for the comparison. */ + gtc_next_dfs_num = 1; + sccstate = pointer_map_create (); + gcc_obstack_init (&sccstate_obstack); + res = gimple_types_compatible_p_1 (t1, t2, p, + &sccstack, sccstate, &sccstate_obstack); + sccstack.release (); + pointer_map_destroy (sccstate); + obstack_free (&sccstate_obstack, NULL); + + return res; +} + +static hashval_t +iterative_hash_gimple_type (tree, hashval_t, vec<tree> *, + struct pointer_map_t *, struct obstack *); + +/* DFS visit the edge from the callers type with state *STATE to T. + Update the callers type hash V with the hash for T if it is not part + of the SCC containing the callers type and return it. + SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. */ + +static hashval_t +visit (tree t, struct sccs *state, hashval_t v, + vec<tree> *sccstack, + struct pointer_map_t *sccstate, + struct obstack *sccstate_obstack) +{ + struct sccs *cstate = NULL; + struct tree_int_map m; + void **slot; + + /* If there is a hash value recorded for this type then it can't + possibly be part of our parent SCC. Simply mix in its hash. */ + m.base.from = t; + if ((slot = htab_find_slot (type_hash_cache, &m, NO_INSERT)) + && *slot) + return iterative_hash_hashval_t (((struct tree_int_map *) *slot)->to, v); + + if ((slot = pointer_map_contains (sccstate, t)) != NULL) + cstate = (struct sccs *)*slot; + if (!cstate) + { + hashval_t tem; + /* Not yet visited. DFS recurse. */ + tem = iterative_hash_gimple_type (t, v, + sccstack, sccstate, sccstate_obstack); + if (!cstate) + cstate = (struct sccs *)* pointer_map_contains (sccstate, t); + state->low = MIN (state->low, cstate->low); + /* If the type is no longer on the SCC stack and thus is not part + of the parents SCC mix in its hash value. Otherwise we will + ignore the type for hashing purposes and return the unaltered + hash value. */ + if (!cstate->on_sccstack) + return tem; + } + if (cstate->dfsnum < state->dfsnum + && cstate->on_sccstack) + state->low = MIN (cstate->dfsnum, state->low); + + /* We are part of our parents SCC, skip this type during hashing + and return the unaltered hash value. */ + return v; +} + +/* Hash NAME with the previous hash value V and return it. */ + +static hashval_t +iterative_hash_name (tree name, hashval_t v) +{ + if (!name) + return v; + v = iterative_hash_hashval_t (TREE_CODE (name), v); + if (TREE_CODE (name) == TYPE_DECL) + name = DECL_NAME (name); + if (!name) + return v; + gcc_assert (TREE_CODE (name) == IDENTIFIER_NODE); + return iterative_hash_object (IDENTIFIER_HASH_VALUE (name), v); +} + +/* A type, hashvalue pair for sorting SCC members. */ + +struct type_hash_pair { + tree type; + hashval_t hash; +}; + +/* Compare two type, hashvalue pairs. */ + +static int +type_hash_pair_compare (const void *p1_, const void *p2_) +{ + const struct type_hash_pair *p1 = (const struct type_hash_pair *) p1_; + const struct type_hash_pair *p2 = (const struct type_hash_pair *) p2_; + if (p1->hash < p2->hash) + return -1; + else if (p1->hash > p2->hash) + return 1; + return 0; +} + +/* Returning a hash value for gimple type TYPE combined with VAL. + SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. + + To hash a type we end up hashing in types that are reachable. + Through pointers we can end up with cycles which messes up the + required property that we need to compute the same hash value + for structurally equivalent types. To avoid this we have to + hash all types in a cycle (the SCC) in a commutative way. The + easiest way is to not mix in the hashes of the SCC members at + all. To make this work we have to delay setting the hash + values of the SCC until it is complete. */ + +static hashval_t +iterative_hash_gimple_type (tree type, hashval_t val, + vec<tree> *sccstack, + struct pointer_map_t *sccstate, + struct obstack *sccstate_obstack) +{ + hashval_t v; + void **slot; + struct sccs *state; + + /* Not visited during this DFS walk. */ + gcc_checking_assert (!pointer_map_contains (sccstate, type)); + state = XOBNEW (sccstate_obstack, struct sccs); + *pointer_map_insert (sccstate, type) = state; + + sccstack->safe_push (type); + state->dfsnum = next_dfs_num++; + state->low = state->dfsnum; + state->on_sccstack = true; + + /* Combine a few common features of types so that types are grouped into + smaller sets; when searching for existing matching types to merge, + only existing types having the same features as the new type will be + checked. */ + v = iterative_hash_name (TYPE_NAME (type), 0); + if (TYPE_NAME (type) + && TREE_CODE (TYPE_NAME (type)) == TYPE_DECL + && DECL_CONTEXT (TYPE_NAME (type)) + && TYPE_P (DECL_CONTEXT (TYPE_NAME (type)))) + v = visit (DECL_CONTEXT (TYPE_NAME (type)), state, v, + sccstack, sccstate, sccstate_obstack); + + /* Factor in the variant structure. */ + if (TYPE_MAIN_VARIANT (type) != type) + v = visit (TYPE_MAIN_VARIANT (type), state, v, + sccstack, sccstate, sccstate_obstack); + + v = iterative_hash_hashval_t (TREE_CODE (type), v); + v = iterative_hash_hashval_t (TYPE_QUALS (type), v); + v = iterative_hash_hashval_t (TREE_ADDRESSABLE (type), v); + + /* Do not hash the types size as this will cause differences in + hash values for the complete vs. the incomplete type variant. */ + + /* Incorporate common features of numerical types. */ + if (INTEGRAL_TYPE_P (type) + || SCALAR_FLOAT_TYPE_P (type) + || FIXED_POINT_TYPE_P (type)) + { + v = iterative_hash_hashval_t (TYPE_PRECISION (type), v); + v = iterative_hash_hashval_t (TYPE_MODE (type), v); + v = iterative_hash_hashval_t (TYPE_UNSIGNED (type), v); + } + + /* For pointer and reference types, fold in information about the type + pointed to. */ + if (POINTER_TYPE_P (type)) + v = visit (TREE_TYPE (type), state, v, + sccstack, sccstate, sccstate_obstack); + + /* For integer types hash the types min/max values and the string flag. */ + if (TREE_CODE (type) == INTEGER_TYPE) + { + /* OMP lowering can introduce error_mark_node in place of + random local decls in types. */ + if (TYPE_MIN_VALUE (type) != error_mark_node) + v = iterative_hash_expr (TYPE_MIN_VALUE (type), v); + if (TYPE_MAX_VALUE (type) != error_mark_node) + v = iterative_hash_expr (TYPE_MAX_VALUE (type), v); + v = iterative_hash_hashval_t (TYPE_STRING_FLAG (type), v); + } + + /* For array types hash the domain and the string flag. */ + if (TREE_CODE (type) == ARRAY_TYPE && TYPE_DOMAIN (type)) + { + v = iterative_hash_hashval_t (TYPE_STRING_FLAG (type), v); + v = visit (TYPE_DOMAIN (type), state, v, + sccstack, sccstate, sccstate_obstack); + } + + /* Recurse for aggregates with a single element type. */ + if (TREE_CODE (type) == ARRAY_TYPE + || TREE_CODE (type) == COMPLEX_TYPE + || TREE_CODE (type) == VECTOR_TYPE) + v = visit (TREE_TYPE (type), state, v, + sccstack, sccstate, sccstate_obstack); + + /* Incorporate function return and argument types. */ + if (TREE_CODE (type) == FUNCTION_TYPE || TREE_CODE (type) == METHOD_TYPE) + { + unsigned na; + tree p; + + /* For method types also incorporate their parent class. */ + if (TREE_CODE (type) == METHOD_TYPE) + v = visit (TYPE_METHOD_BASETYPE (type), state, v, + sccstack, sccstate, sccstate_obstack); + + /* Check result and argument types. */ + v = visit (TREE_TYPE (type), state, v, + sccstack, sccstate, sccstate_obstack); + for (p = TYPE_ARG_TYPES (type), na = 0; p; p = TREE_CHAIN (p)) + { + v = visit (TREE_VALUE (p), state, v, + sccstack, sccstate, sccstate_obstack); + na++; + } + + v = iterative_hash_hashval_t (na, v); + } + + if (RECORD_OR_UNION_TYPE_P (type)) + { + unsigned nf; + tree f; + + for (f = TYPE_FIELDS (type), nf = 0; f; f = TREE_CHAIN (f)) + { + v = iterative_hash_name (DECL_NAME (f), v); + v = visit (TREE_TYPE (f), state, v, + sccstack, sccstate, sccstate_obstack); + nf++; + } + + v = iterative_hash_hashval_t (nf, v); + } + + /* Record hash for us. */ + state->u.hash = v; + + /* See if we found an SCC. */ + if (state->low == state->dfsnum) + { + tree x; + struct tree_int_map *m; + + /* Pop off the SCC and set its hash values. */ + x = sccstack->pop (); + /* Optimize SCC size one. */ + if (x == type) + { + state->on_sccstack = false; + m = ggc_alloc_cleared_tree_int_map (); + m->base.from = x; + m->to = v; + slot = htab_find_slot (type_hash_cache, m, INSERT); + gcc_assert (!*slot); + *slot = (void *) m; + } + else + { + struct sccs *cstate; + unsigned first, i, size, j; + struct type_hash_pair *pairs; + /* Pop off the SCC and build an array of type, hash pairs. */ + first = sccstack->length () - 1; + while ((*sccstack)[first] != type) + --first; + size = sccstack->length () - first + 1; + pairs = XALLOCAVEC (struct type_hash_pair, size); + i = 0; + cstate = (struct sccs *)*pointer_map_contains (sccstate, x); + cstate->on_sccstack = false; + pairs[i].type = x; + pairs[i].hash = cstate->u.hash; + do + { + x = sccstack->pop (); + cstate = (struct sccs *)*pointer_map_contains (sccstate, x); + cstate->on_sccstack = false; + ++i; + pairs[i].type = x; + pairs[i].hash = cstate->u.hash; + } + while (x != type); + gcc_assert (i + 1 == size); + /* Sort the arrays of type, hash pairs so that when we mix in + all members of the SCC the hash value becomes independent on + the order we visited the SCC. Disregard hashes equal to + the hash of the type we mix into because we cannot guarantee + a stable sort for those across different TUs. */ + qsort (pairs, size, sizeof (struct type_hash_pair), + type_hash_pair_compare); + for (i = 0; i < size; ++i) + { + hashval_t hash; + m = ggc_alloc_cleared_tree_int_map (); + m->base.from = pairs[i].type; + hash = pairs[i].hash; + /* Skip same hashes. */ + for (j = i + 1; j < size && pairs[j].hash == pairs[i].hash; ++j) + ; + for (; j < size; ++j) + hash = iterative_hash_hashval_t (pairs[j].hash, hash); + for (j = 0; pairs[j].hash != pairs[i].hash; ++j) + hash = iterative_hash_hashval_t (pairs[j].hash, hash); + m->to = hash; + if (pairs[i].type == type) + v = hash; + slot = htab_find_slot (type_hash_cache, m, INSERT); + gcc_assert (!*slot); + *slot = (void *) m; + } + } + } + + return iterative_hash_hashval_t (v, val); +} + +/* Returns a hash value for P (assumed to be a type). The hash value + is computed using some distinguishing features of the type. Note + that we cannot use pointer hashing here as we may be dealing with + two distinct instances of the same type. + + This function should produce the same hash value for two compatible + types according to gimple_types_compatible_p. */ + +static hashval_t +gimple_type_hash (const void *p) +{ + const_tree t = (const_tree) p; + vec<tree> sccstack = vNULL; + struct pointer_map_t *sccstate; + struct obstack sccstate_obstack; + hashval_t val; + void **slot; + struct tree_int_map m; + + m.base.from = CONST_CAST_TREE (t); + if ((slot = htab_find_slot (type_hash_cache, &m, NO_INSERT)) + && *slot) + return iterative_hash_hashval_t (((struct tree_int_map *) *slot)->to, 0); + + /* Perform a DFS walk and pre-hash all reachable types. */ + next_dfs_num = 1; + sccstate = pointer_map_create (); + gcc_obstack_init (&sccstate_obstack); + val = iterative_hash_gimple_type (CONST_CAST_TREE (t), 0, + &sccstack, sccstate, &sccstate_obstack); + sccstack.release (); + pointer_map_destroy (sccstate); + obstack_free (&sccstate_obstack, NULL); + + return val; +} + +/* Returns nonzero if P1 and P2 are equal. */ + +static int +gimple_type_eq (const void *p1, const void *p2) +{ + const_tree t1 = (const_tree) p1; + const_tree t2 = (const_tree) p2; + return gimple_types_compatible_p (CONST_CAST_TREE (t1), + CONST_CAST_TREE (t2)); +} + + +/* Worker for gimple_register_type. + Register type T in the global type table gimple_types. + When REGISTERING_MV is false first recurse for the main variant of T. */ + +static tree +gimple_register_type_1 (tree t, bool registering_mv) +{ + void **slot; + gimple_type_leader_entry *leader; + + /* If we registered this type before return the cached result. */ + leader = &gimple_type_leader[TYPE_UID (t) % GIMPLE_TYPE_LEADER_SIZE]; + if (leader->type == t) + return leader->leader; + + /* Always register the main variant first. This is important so we + pick up the non-typedef variants as canonical, otherwise we'll end + up taking typedef ids for structure tags during comparison. + It also makes sure that main variants will be merged to main variants. + As we are operating on a possibly partially fixed up type graph + do not bother to recurse more than once, otherwise we may end up + walking in circles. + If we are registering a main variant it will either remain its + own main variant or it will be merged to something else in which + case we do not care for the main variant leader. */ + if (!registering_mv + && TYPE_MAIN_VARIANT (t) != t) + gimple_register_type_1 (TYPE_MAIN_VARIANT (t), true); + + /* See if we already have an equivalent type registered. */ + slot = htab_find_slot (gimple_types, t, INSERT); + if (*slot + && *(tree *)slot != t) + { + tree new_type = (tree) *((tree *) slot); + leader->type = t; + leader->leader = new_type; + return new_type; + } + + /* If not, insert it to the cache and the hash. */ + leader->type = t; + leader->leader = t; + *slot = (void *) t; + return t; +} + +/* Register type T in the global type table gimple_types. + If another type T', compatible with T, already existed in + gimple_types then return T', otherwise return T. This is used by + LTO to merge identical types read from different TUs. */ + +static tree +gimple_register_type (tree t) +{ + gcc_assert (TYPE_P (t)); + return gimple_register_type_1 (t, false); +} + +#define GIMPLE_REGISTER_TYPE(tt) \ + (TREE_VISITED (tt) ? gimple_register_type (tt) : tt) + + + +/* A hashtable of trees that potentially refer to variables or functions + that must be replaced with their prevailing variant. */ +static GTY((if_marked ("ggc_marked_p"), param_is (union tree_node))) htab_t + tree_with_vars; + +/* Remember that T is a tree that (potentially) refers to a variable + or function decl that may be replaced with its prevailing variant. */ +static void +remember_with_vars (tree t) +{ + *(tree *) htab_find_slot (tree_with_vars, t, INSERT) = t; +} + +#define LTO_FIXUP_TREE(tt) \ + do \ + { \ + if (tt) \ + { \ + if (TYPE_P (tt)) \ + (tt) = GIMPLE_REGISTER_TYPE (tt); \ + if (VAR_OR_FUNCTION_DECL_P (tt) && TREE_PUBLIC (tt)) \ + remember_with_vars (t); \ + if (TREE_CODE (tt) == INTEGER_CST) \ + (tt) = fixup_integer_cst (tt); \ + } \ + } while (0) + +static void lto_fixup_types (tree); + +/* Return integer_cst T with updated type. */ + +static tree +fixup_integer_cst (tree t) +{ + tree type = GIMPLE_REGISTER_TYPE (TREE_TYPE (t)); + + if (type == TREE_TYPE (t)) + return t; + + /* If overflow was set, streamer_read_integer_cst + produced local copy of T. */ + if (TREE_OVERFLOW (t)) + { + TREE_TYPE (t) = type; + return t; + } + else + /* Otherwise produce new shared node for the new type. */ + return build_int_cst_wide (type, TREE_INT_CST_LOW (t), + TREE_INT_CST_HIGH (t)); +} + +/* Fix up fields of a tree_typed T. */ + +static void +lto_ft_typed (tree t) +{ + LTO_FIXUP_TREE (TREE_TYPE (t)); +} + +/* Fix up fields of a tree_common T. */ + +static void +lto_ft_common (tree t) +{ + lto_ft_typed (t); + LTO_FIXUP_TREE (TREE_CHAIN (t)); +} + +/* Fix up fields of a decl_minimal T. */ + +static void +lto_ft_decl_minimal (tree t) +{ + lto_ft_common (t); + LTO_FIXUP_TREE (DECL_NAME (t)); + LTO_FIXUP_TREE (DECL_CONTEXT (t)); +} + +/* Fix up fields of a decl_common T. */ + +static void +lto_ft_decl_common (tree t) +{ + lto_ft_decl_minimal (t); + LTO_FIXUP_TREE (DECL_SIZE (t)); + LTO_FIXUP_TREE (DECL_SIZE_UNIT (t)); + LTO_FIXUP_TREE (DECL_INITIAL (t)); + LTO_FIXUP_TREE (DECL_ATTRIBUTES (t)); + LTO_FIXUP_TREE (DECL_ABSTRACT_ORIGIN (t)); +} + +/* Fix up fields of a decl_with_vis T. */ + +static void +lto_ft_decl_with_vis (tree t) +{ + lto_ft_decl_common (t); + + /* Accessor macro has side-effects, use field-name here. */ + LTO_FIXUP_TREE (t->decl_with_vis.assembler_name); + LTO_FIXUP_TREE (DECL_SECTION_NAME (t)); +} + +/* Fix up fields of a decl_non_common T. */ + +static void +lto_ft_decl_non_common (tree t) +{ + lto_ft_decl_with_vis (t); + LTO_FIXUP_TREE (DECL_ARGUMENT_FLD (t)); + LTO_FIXUP_TREE (DECL_RESULT_FLD (t)); + LTO_FIXUP_TREE (DECL_VINDEX (t)); + /* The C frontends may create exact duplicates for DECL_ORIGINAL_TYPE + like for 'typedef enum foo foo'. We have no way of avoiding to + merge them and dwarf2out.c cannot deal with this, + so fix this up by clearing DECL_ORIGINAL_TYPE in this case. */ + if (TREE_CODE (t) == TYPE_DECL + && DECL_ORIGINAL_TYPE (t) == TREE_TYPE (t)) + DECL_ORIGINAL_TYPE (t) = NULL_TREE; +} + +/* Fix up fields of a decl_non_common T. */ + +static void +lto_ft_function (tree t) +{ + lto_ft_decl_non_common (t); + LTO_FIXUP_TREE (DECL_FUNCTION_PERSONALITY (t)); +} + +/* Fix up fields of a field_decl T. */ + +static void +lto_ft_field_decl (tree t) +{ + lto_ft_decl_common (t); + LTO_FIXUP_TREE (DECL_FIELD_OFFSET (t)); + LTO_FIXUP_TREE (DECL_BIT_FIELD_TYPE (t)); + LTO_FIXUP_TREE (DECL_QUALIFIER (t)); + LTO_FIXUP_TREE (DECL_FIELD_BIT_OFFSET (t)); + LTO_FIXUP_TREE (DECL_FCONTEXT (t)); +} + +/* Fix up fields of a type T. */ + +static void +lto_ft_type (tree t) +{ + lto_ft_common (t); + LTO_FIXUP_TREE (TYPE_CACHED_VALUES (t)); + LTO_FIXUP_TREE (TYPE_SIZE (t)); + LTO_FIXUP_TREE (TYPE_SIZE_UNIT (t)); + LTO_FIXUP_TREE (TYPE_ATTRIBUTES (t)); + LTO_FIXUP_TREE (TYPE_NAME (t)); + + /* Accessors are for derived node types only. */ + if (!POINTER_TYPE_P (t)) + LTO_FIXUP_TREE (TYPE_MINVAL (t)); + LTO_FIXUP_TREE (TYPE_MAXVAL (t)); + + /* Accessor is for derived node types only. */ + LTO_FIXUP_TREE (t->type_non_common.binfo); + + LTO_FIXUP_TREE (TYPE_CONTEXT (t)); +} + +/* Fix up fields of a BINFO T. */ + +static void +lto_ft_binfo (tree t) +{ + unsigned HOST_WIDE_INT i, n; + tree base, saved_base; + + lto_ft_common (t); + LTO_FIXUP_TREE (BINFO_VTABLE (t)); + LTO_FIXUP_TREE (BINFO_OFFSET (t)); + LTO_FIXUP_TREE (BINFO_VIRTUALS (t)); + LTO_FIXUP_TREE (BINFO_VPTR_FIELD (t)); + n = vec_safe_length (BINFO_BASE_ACCESSES (t)); + for (i = 0; i < n; i++) + { + saved_base = base = BINFO_BASE_ACCESS (t, i); + LTO_FIXUP_TREE (base); + if (base != saved_base) + (*BINFO_BASE_ACCESSES (t))[i] = base; + } + LTO_FIXUP_TREE (BINFO_INHERITANCE_CHAIN (t)); + LTO_FIXUP_TREE (BINFO_SUBVTT_INDEX (t)); + LTO_FIXUP_TREE (BINFO_VPTR_INDEX (t)); + n = BINFO_N_BASE_BINFOS (t); + for (i = 0; i < n; i++) + { + saved_base = base = BINFO_BASE_BINFO (t, i); + LTO_FIXUP_TREE (base); + if (base != saved_base) + (*BINFO_BASE_BINFOS (t))[i] = base; + } +} + +/* Fix up fields of a CONSTRUCTOR T. */ + +static void +lto_ft_constructor (tree t) +{ + unsigned HOST_WIDE_INT idx; + constructor_elt *ce; + + lto_ft_typed (t); + + for (idx = 0; vec_safe_iterate (CONSTRUCTOR_ELTS (t), idx, &ce); idx++) + { + LTO_FIXUP_TREE (ce->index); + LTO_FIXUP_TREE (ce->value); + } +} + +/* Fix up fields of an expression tree T. */ + +static void +lto_ft_expr (tree t) +{ + int i; + lto_ft_typed (t); + for (i = TREE_OPERAND_LENGTH (t) - 1; i >= 0; --i) + LTO_FIXUP_TREE (TREE_OPERAND (t, i)); +} + +/* Given a tree T fixup fields of T by replacing types with their merged + variant and other entities by an equal entity from an earlier compilation + unit, or an entity being canonical in a different way. This includes + for instance integer or string constants. */ + +static void +lto_fixup_types (tree t) +{ + switch (TREE_CODE (t)) + { + case IDENTIFIER_NODE: + break; + + case TREE_LIST: + LTO_FIXUP_TREE (TREE_VALUE (t)); + LTO_FIXUP_TREE (TREE_PURPOSE (t)); + LTO_FIXUP_TREE (TREE_CHAIN (t)); + break; + + case FIELD_DECL: + lto_ft_field_decl (t); + break; + + case LABEL_DECL: + case CONST_DECL: + case PARM_DECL: + case RESULT_DECL: + case IMPORTED_DECL: + lto_ft_decl_common (t); + break; + + case VAR_DECL: + lto_ft_decl_with_vis (t); + break; + + case TYPE_DECL: + lto_ft_decl_non_common (t); + break; + + case FUNCTION_DECL: + lto_ft_function (t); + break; + + case TREE_BINFO: + lto_ft_binfo (t); + break; + + case PLACEHOLDER_EXPR: + lto_ft_common (t); + break; + + case BLOCK: + case TRANSLATION_UNIT_DECL: + case OPTIMIZATION_NODE: + case TARGET_OPTION_NODE: + break; + + default: + if (TYPE_P (t)) + lto_ft_type (t); + else if (TREE_CODE (t) == CONSTRUCTOR) + lto_ft_constructor (t); + else if (CONSTANT_CLASS_P (t)) + LTO_FIXUP_TREE (TREE_TYPE (t)); + else if (EXPR_P (t)) + { + lto_ft_expr (t); + } + else + { + remember_with_vars (t); + } + } +} + + +/* Return the resolution for the decl with index INDEX from DATA_IN. */ + +static enum ld_plugin_symbol_resolution +get_resolution (struct data_in *data_in, unsigned index) +{ + if (data_in->globals_resolution.exists ()) + { + ld_plugin_symbol_resolution_t ret; + /* We can have references to not emitted functions in + DECL_FUNCTION_PERSONALITY at least. So we can and have + to indeed return LDPR_UNKNOWN in some cases. */ + if (data_in->globals_resolution.length () <= index) + return LDPR_UNKNOWN; + ret = data_in->globals_resolution[index]; + return ret; + } + else + /* Delay resolution finding until decl merging. */ + return LDPR_UNKNOWN; +} + +/* Map assigning declarations their resolutions. */ +static pointer_map_t *resolution_map; + +/* We need to record resolutions until symbol table is read. */ +static void +register_resolution (tree decl, enum ld_plugin_symbol_resolution resolution) +{ + if (resolution == LDPR_UNKNOWN) + return; + if (!resolution_map) + resolution_map = pointer_map_create (); + *pointer_map_insert (resolution_map, decl) = (void *)(size_t)resolution; +} + +/* Register DECL with the global symbol table and change its + name if necessary to avoid name clashes for static globals across + different files. */ + +static void +lto_register_var_decl_in_symtab (struct data_in *data_in, tree decl) +{ + tree context; + + /* Variable has file scope, not local. Need to ensure static variables + between different files don't clash unexpectedly. */ + if (!TREE_PUBLIC (decl) + && !((context = decl_function_context (decl)) + && auto_var_in_fn_p (decl, context))) + { + /* ??? We normally pre-mangle names before we serialize them + out. Here, in lto1, we do not know the language, and + thus cannot do the mangling again. Instead, we just + append a suffix to the mangled name. The resulting name, + however, is not a properly-formed mangled name, and will + confuse any attempt to unmangle it. */ + const char *name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)); + char *label; + + ASM_FORMAT_PRIVATE_NAME (label, name, DECL_UID (decl)); + SET_DECL_ASSEMBLER_NAME (decl, get_identifier (label)); + rest_of_decl_compilation (decl, 1, 0); + } + + /* If this variable has already been declared, queue the + declaration for merging. */ + if (TREE_PUBLIC (decl)) + { + unsigned ix; + if (!streamer_tree_cache_lookup (data_in->reader_cache, decl, &ix)) + gcc_unreachable (); + register_resolution (decl, get_resolution (data_in, ix)); + } +} + + +/* Register DECL with the global symbol table and change its + name if necessary to avoid name clashes for static globals across + different files. DATA_IN contains descriptors and tables for the + file being read. */ + +static void +lto_register_function_decl_in_symtab (struct data_in *data_in, tree decl) +{ + /* Need to ensure static entities between different files + don't clash unexpectedly. */ + if (!TREE_PUBLIC (decl)) + { + /* We must not use the DECL_ASSEMBLER_NAME macro here, as it + may set the assembler name where it was previously empty. */ + tree old_assembler_name = decl->decl_with_vis.assembler_name; + + /* FIXME lto: We normally pre-mangle names before we serialize + them out. Here, in lto1, we do not know the language, and + thus cannot do the mangling again. Instead, we just append a + suffix to the mangled name. The resulting name, however, is + not a properly-formed mangled name, and will confuse any + attempt to unmangle it. */ + const char *name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)); + char *label; + + ASM_FORMAT_PRIVATE_NAME (label, name, DECL_UID (decl)); + SET_DECL_ASSEMBLER_NAME (decl, get_identifier (label)); + + /* We may arrive here with the old assembler name not set + if the function body is not needed, e.g., it has been + inlined away and does not appear in the cgraph. */ + if (old_assembler_name) + { + tree new_assembler_name = DECL_ASSEMBLER_NAME (decl); + + /* Make the original assembler name available for later use. + We may have used it to indicate the section within its + object file where the function body may be found. + FIXME lto: Find a better way to maintain the function decl + to body section mapping so we don't need this hack. */ + lto_record_renamed_decl (data_in->file_data, + IDENTIFIER_POINTER (old_assembler_name), + IDENTIFIER_POINTER (new_assembler_name)); + } + } + + /* If this variable has already been declared, queue the + declaration for merging. */ + if (TREE_PUBLIC (decl) && !DECL_ABSTRACT (decl)) + { + unsigned ix; + if (!streamer_tree_cache_lookup (data_in->reader_cache, decl, &ix)) + gcc_unreachable (); + register_resolution (decl, get_resolution (data_in, ix)); + } +} + + +/* Given a streamer cache structure DATA_IN (holding a sequence of trees + for one compilation unit) go over all trees starting at index FROM until the + end of the sequence and replace fields of those trees, and the trees + themself with their canonical variants as per gimple_register_type. */ + +static void +uniquify_nodes (struct data_in *data_in, unsigned from) +{ + struct streamer_tree_cache_d *cache = data_in->reader_cache; + unsigned len = cache->nodes.length (); + unsigned i; + + /* Go backwards because children streamed for the first time come + as part of their parents, and hence are created after them. */ + + /* First register all the types in the cache. This makes sure to + have the original structure in the type cycles when registering + them and computing hashes. */ + for (i = len; i-- > from;) + { + tree t = cache->nodes[i]; + if (t && TYPE_P (t)) + { + tree newt = gimple_register_type (t); + /* Mark non-prevailing types so we fix them up. No need + to reset that flag afterwards - nothing that refers + to those types is left and they are collected. */ + if (newt != t) + TREE_VISITED (t) = 1; + } + } + + /* Second fixup all trees in the new cache entries. */ + for (i = len; i-- > from;) + { + tree t = cache->nodes[i]; + tree oldt = t; + if (!t) + continue; + + /* First fixup the fields of T. */ + lto_fixup_types (t); + + if (!TYPE_P (t)) + continue; + + /* Now try to find a canonical variant of T itself. */ + t = GIMPLE_REGISTER_TYPE (t); + + if (t == oldt) + { + /* The following re-creates proper variant lists while fixing up + the variant leaders. We do not stream TYPE_NEXT_VARIANT so the + variant list state before fixup is broken. */ + tree tem, mv; + +#ifdef ENABLE_CHECKING + /* Remove us from our main variant list if we are not the + variant leader. */ + if (TYPE_MAIN_VARIANT (t) != t) + { + tem = TYPE_MAIN_VARIANT (t); + while (tem && TYPE_NEXT_VARIANT (tem) != t) + tem = TYPE_NEXT_VARIANT (tem); + gcc_assert (!tem && !TYPE_NEXT_VARIANT (t)); + } +#endif + + /* Query our new main variant. */ + mv = GIMPLE_REGISTER_TYPE (TYPE_MAIN_VARIANT (t)); + + /* If we were the variant leader and we get replaced ourselves drop + all variants from our list. */ + if (TYPE_MAIN_VARIANT (t) == t + && mv != t) + { + tem = t; + while (tem) + { + tree tem2 = TYPE_NEXT_VARIANT (tem); + TYPE_NEXT_VARIANT (tem) = NULL_TREE; + tem = tem2; + } + } + + /* If we are not our own variant leader link us into our new leaders + variant list. */ + if (mv != t) + { + TYPE_NEXT_VARIANT (t) = TYPE_NEXT_VARIANT (mv); + TYPE_NEXT_VARIANT (mv) = t; + if (RECORD_OR_UNION_TYPE_P (t)) + TYPE_BINFO (t) = TYPE_BINFO (mv); + /* Preserve the invariant that type variants share their + TYPE_FIELDS. */ + if (RECORD_OR_UNION_TYPE_P (t) + && TYPE_FIELDS (mv) != TYPE_FIELDS (t)) + { + tree f1, f2; + for (f1 = TYPE_FIELDS (mv), f2 = TYPE_FIELDS (t); + f1 && f2; f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2)) + { + unsigned ix; + gcc_assert (f1 != f2 + && DECL_NAME (f1) == DECL_NAME (f2)); + if (!streamer_tree_cache_lookup (cache, f2, &ix)) + gcc_unreachable (); + /* If we're going to replace an element which we'd + still visit in the next iterations, we wouldn't + handle it, so do it here. We do have to handle it + even though the field_decl itself will be removed, + as it could refer to e.g. integer_cst which we + wouldn't reach via any other way, hence they + (and their type) would stay uncollected. */ + /* ??? We should rather make sure to replace all + references to f2 with f1. That means handling + COMPONENT_REFs and CONSTRUCTOR elements in + lto_fixup_types and special-case the field-decl + operand handling. */ + /* ??? Not sure the above is all relevant in this + path canonicalizing TYPE_FIELDS to that of the + main variant. */ + if (ix < i) + lto_fixup_types (f2); + streamer_tree_cache_insert_at (cache, f1, ix); + } + TYPE_FIELDS (t) = TYPE_FIELDS (mv); + } + } + + /* Finally adjust our main variant and fix it up. */ + TYPE_MAIN_VARIANT (t) = mv; + + /* The following reconstructs the pointer chains + of the new pointed-to type if we are a main variant. We do + not stream those so they are broken before fixup. */ + if (TREE_CODE (t) == POINTER_TYPE + && TYPE_MAIN_VARIANT (t) == t) + { + TYPE_NEXT_PTR_TO (t) = TYPE_POINTER_TO (TREE_TYPE (t)); + TYPE_POINTER_TO (TREE_TYPE (t)) = t; + } + else if (TREE_CODE (t) == REFERENCE_TYPE + && TYPE_MAIN_VARIANT (t) == t) + { + TYPE_NEXT_REF_TO (t) = TYPE_REFERENCE_TO (TREE_TYPE (t)); + TYPE_REFERENCE_TO (TREE_TYPE (t)) = t; + } + } + + else + { + if (RECORD_OR_UNION_TYPE_P (t)) + { + tree f1, f2; + if (TYPE_FIELDS (t) != TYPE_FIELDS (oldt)) + for (f1 = TYPE_FIELDS (t), f2 = TYPE_FIELDS (oldt); + f1 && f2; f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2)) + { + unsigned ix; + gcc_assert (f1 != f2 && DECL_NAME (f1) == DECL_NAME (f2)); + if (!streamer_tree_cache_lookup (cache, f2, &ix)) + gcc_unreachable (); + /* If we're going to replace an element which we'd + still visit in the next iterations, we wouldn't + handle it, so do it here. We do have to handle it + even though the field_decl itself will be removed, + as it could refer to e.g. integer_cst which we + wouldn't reach via any other way, hence they + (and their type) would stay uncollected. */ + /* ??? We should rather make sure to replace all + references to f2 with f1. That means handling + COMPONENT_REFs and CONSTRUCTOR elements in + lto_fixup_types and special-case the field-decl + operand handling. */ + if (ix < i) + lto_fixup_types (f2); + streamer_tree_cache_insert_at (cache, f1, ix); + } + } + + /* If we found a tree that is equal to oldt replace it in the + cache, so that further users (in the various LTO sections) + make use of it. */ + streamer_tree_cache_insert_at (cache, t, i); + } + } + + /* Finally compute the canonical type of all TREE_TYPEs and register + VAR_DECL and FUNCTION_DECL nodes in the symbol table. + From this point there are no longer any types with + TYPE_STRUCTURAL_EQUALITY_P and its type-based alias problems. + This step requires the TYPE_POINTER_TO lists being present, so + make sure it is done last. */ + for (i = len; i-- > from;) + { + tree t = cache->nodes[i]; + if (t == NULL_TREE) + continue; + + if (TREE_CODE (t) == VAR_DECL) + lto_register_var_decl_in_symtab (data_in, t); + else if (TREE_CODE (t) == FUNCTION_DECL && !DECL_BUILT_IN (t)) + lto_register_function_decl_in_symtab (data_in, t); + else if (!flag_wpa + && TREE_CODE (t) == TYPE_DECL) + debug_hooks->type_decl (t, !DECL_FILE_SCOPE_P (t)); + else if (TYPE_P (t) && !TYPE_CANONICAL (t)) + TYPE_CANONICAL (t) = gimple_register_canonical_type (t); + } +} + + +/* Read all the symbols from buffer DATA, using descriptors in DECL_DATA. + RESOLUTIONS is the set of symbols picked by the linker (read from the + resolution file when the linker plugin is being used). */ + +static void +lto_read_decls (struct lto_file_decl_data *decl_data, const void *data, + vec<ld_plugin_symbol_resolution_t> resolutions) +{ + const struct lto_decl_header *header = (const struct lto_decl_header *) data; + const int decl_offset = sizeof (struct lto_decl_header); + const int main_offset = decl_offset + header->decl_state_size; + const int string_offset = main_offset + header->main_size; + struct lto_input_block ib_main; + struct data_in *data_in; + unsigned int i; + const uint32_t *data_ptr, *data_end; + uint32_t num_decl_states; + + LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0, + header->main_size); + + data_in = lto_data_in_create (decl_data, (const char *) data + string_offset, + header->string_size, resolutions); + + /* We do not uniquify the pre-loaded cache entries, those are middle-end + internal types that should not be merged. */ + + /* Read the global declarations and types. */ + while (ib_main.p < ib_main.len) + { + tree t; + unsigned from = data_in->reader_cache->nodes.length (); + t = stream_read_tree (&ib_main, data_in); + gcc_assert (t && ib_main.p <= ib_main.len); + uniquify_nodes (data_in, from); + } + + /* Read in lto_in_decl_state objects. */ + data_ptr = (const uint32_t *) ((const char*) data + decl_offset); + data_end = + (const uint32_t *) ((const char*) data_ptr + header->decl_state_size); + num_decl_states = *data_ptr++; + + gcc_assert (num_decl_states > 0); + decl_data->global_decl_state = lto_new_in_decl_state (); + data_ptr = lto_read_in_decl_state (data_in, data_ptr, + decl_data->global_decl_state); + + /* Read in per-function decl states and enter them in hash table. */ + decl_data->function_decl_states = + htab_create_ggc (37, lto_hash_in_decl_state, lto_eq_in_decl_state, NULL); + + for (i = 1; i < num_decl_states; i++) + { + struct lto_in_decl_state *state = lto_new_in_decl_state (); + void **slot; + + data_ptr = lto_read_in_decl_state (data_in, data_ptr, state); + slot = htab_find_slot (decl_data->function_decl_states, state, INSERT); + gcc_assert (*slot == NULL); + *slot = state; + } + + if (data_ptr != data_end) + internal_error ("bytecode stream: garbage at the end of symbols section"); + + /* Set the current decl state to be the global state. */ + decl_data->current_decl_state = decl_data->global_decl_state; + + lto_data_in_delete (data_in); +} + +/* Custom version of strtoll, which is not portable. */ + +static HOST_WIDEST_INT +lto_parse_hex (const char *p) +{ + HOST_WIDEST_INT ret = 0; + + for (; *p != '\0'; ++p) + { + char c = *p; + unsigned char part; + ret <<= 4; + if (c >= '0' && c <= '9') + part = c - '0'; + else if (c >= 'a' && c <= 'f') + part = c - 'a' + 10; + else if (c >= 'A' && c <= 'F') + part = c - 'A' + 10; + else + internal_error ("could not parse hex number"); + ret |= part; + } + + return ret; +} + +/* Read resolution for file named FILE_NAME. The resolution is read from + RESOLUTION. */ + +static void +lto_resolution_read (splay_tree file_ids, FILE *resolution, lto_file *file) +{ + /* We require that objects in the resolution file are in the same + order as the lto1 command line. */ + unsigned int name_len; + char *obj_name; + unsigned int num_symbols; + unsigned int i; + struct lto_file_decl_data *file_data; + splay_tree_node nd = NULL; + + if (!resolution) + return; + + name_len = strlen (file->filename); + obj_name = XNEWVEC (char, name_len + 1); + fscanf (resolution, " "); /* Read white space. */ + + fread (obj_name, sizeof (char), name_len, resolution); + obj_name[name_len] = '\0'; + if (filename_cmp (obj_name, file->filename) != 0) + internal_error ("unexpected file name %s in linker resolution file. " + "Expected %s", obj_name, file->filename); + if (file->offset != 0) + { + int t; + char offset_p[17]; + HOST_WIDEST_INT offset; + t = fscanf (resolution, "@0x%16s", offset_p); + if (t != 1) + internal_error ("could not parse file offset"); + offset = lto_parse_hex (offset_p); + if (offset != file->offset) + internal_error ("unexpected offset"); + } + + free (obj_name); + + fscanf (resolution, "%u", &num_symbols); + + for (i = 0; i < num_symbols; i++) + { + int t; + unsigned index; + unsigned HOST_WIDE_INT id; + char r_str[27]; + enum ld_plugin_symbol_resolution r = (enum ld_plugin_symbol_resolution) 0; + unsigned int j; + unsigned int lto_resolution_str_len = + sizeof (lto_resolution_str) / sizeof (char *); + res_pair rp; + + t = fscanf (resolution, "%u " HOST_WIDE_INT_PRINT_HEX_PURE " %26s %*[^\n]\n", + &index, &id, r_str); + if (t != 3) + internal_error ("invalid line in the resolution file"); + + for (j = 0; j < lto_resolution_str_len; j++) + { + if (strcmp (lto_resolution_str[j], r_str) == 0) + { + r = (enum ld_plugin_symbol_resolution) j; + break; + } + } + if (j == lto_resolution_str_len) + internal_error ("invalid resolution in the resolution file"); + + if (!(nd && lto_splay_tree_id_equal_p (nd->key, id))) + { + nd = lto_splay_tree_lookup (file_ids, id); + if (nd == NULL) + internal_error ("resolution sub id %wx not in object file", id); + } + + file_data = (struct lto_file_decl_data *)nd->value; + /* The indexes are very sparse. To save memory save them in a compact + format that is only unpacked later when the subfile is processed. */ + rp.res = r; + rp.index = index; + file_data->respairs.safe_push (rp); + if (file_data->max_index < index) + file_data->max_index = index; + } +} + +/* List of file_decl_datas */ +struct file_data_list + { + struct lto_file_decl_data *first, *last; + }; + +/* Is the name for a id'ed LTO section? */ + +static int +lto_section_with_id (const char *name, unsigned HOST_WIDE_INT *id) +{ + const char *s; + + if (strncmp (name, LTO_SECTION_NAME_PREFIX, strlen (LTO_SECTION_NAME_PREFIX))) + return 0; + s = strrchr (name, '.'); + return s && sscanf (s, "." HOST_WIDE_INT_PRINT_HEX_PURE, id) == 1; +} + +/* Create file_data of each sub file id */ + +static int +create_subid_section_table (struct lto_section_slot *ls, splay_tree file_ids, + struct file_data_list *list) +{ + struct lto_section_slot s_slot, *new_slot; + unsigned HOST_WIDE_INT id; + splay_tree_node nd; + void **hash_slot; + char *new_name; + struct lto_file_decl_data *file_data; + + if (!lto_section_with_id (ls->name, &id)) + return 1; + + /* Find hash table of sub module id */ + nd = lto_splay_tree_lookup (file_ids, id); + if (nd != NULL) + { + file_data = (struct lto_file_decl_data *)nd->value; + } + else + { + file_data = ggc_alloc_lto_file_decl_data (); + memset(file_data, 0, sizeof (struct lto_file_decl_data)); + file_data->id = id; + file_data->section_hash_table = lto_obj_create_section_hash_table ();; + lto_splay_tree_insert (file_ids, id, file_data); + + /* Maintain list in linker order */ + if (!list->first) + list->first = file_data; + if (list->last) + list->last->next = file_data; + list->last = file_data; + } + + /* Copy section into sub module hash table */ + new_name = XDUPVEC (char, ls->name, strlen (ls->name) + 1); + s_slot.name = new_name; + hash_slot = htab_find_slot (file_data->section_hash_table, &s_slot, INSERT); + gcc_assert (*hash_slot == NULL); + + new_slot = XDUP (struct lto_section_slot, ls); + new_slot->name = new_name; + *hash_slot = new_slot; + return 1; +} + +/* Read declarations and other initializations for a FILE_DATA. */ + +static void +lto_file_finalize (struct lto_file_decl_data *file_data, lto_file *file) +{ + const char *data; + size_t len; + vec<ld_plugin_symbol_resolution_t> + resolutions = vNULL; + int i; + res_pair *rp; + + /* Create vector for fast access of resolution. We do this lazily + to save memory. */ + resolutions.safe_grow_cleared (file_data->max_index + 1); + for (i = 0; file_data->respairs.iterate (i, &rp); i++) + resolutions[rp->index] = rp->res; + file_data->respairs.release (); + + file_data->renaming_hash_table = lto_create_renaming_table (); + file_data->file_name = file->filename; + data = lto_get_section_data (file_data, LTO_section_decls, NULL, &len); + if (data == NULL) + { + internal_error ("cannot read LTO decls from %s", file_data->file_name); + return; + } + /* Frees resolutions */ + lto_read_decls (file_data, data, resolutions); + lto_free_section_data (file_data, LTO_section_decls, NULL, data, len); +} + +/* Finalize FILE_DATA in FILE and increase COUNT. */ + +static int +lto_create_files_from_ids (lto_file *file, struct lto_file_decl_data *file_data, + int *count) +{ + lto_file_finalize (file_data, file); + if (cgraph_dump_file) + fprintf (cgraph_dump_file, "Creating file %s with sub id " HOST_WIDE_INT_PRINT_HEX "\n", + file_data->file_name, file_data->id); + (*count)++; + return 0; +} + +/* Generate a TREE representation for all types and external decls + entities in FILE. + + Read all of the globals out of the file. Then read the cgraph + and process the .o index into the cgraph nodes so that it can open + the .o file to load the functions and ipa information. */ + +static struct lto_file_decl_data * +lto_file_read (lto_file *file, FILE *resolution_file, int *count) +{ + struct lto_file_decl_data *file_data = NULL; + splay_tree file_ids; + htab_t section_hash_table; + struct lto_section_slot *section; + struct file_data_list file_list; + struct lto_section_list section_list; + + memset (§ion_list, 0, sizeof (struct lto_section_list)); + section_hash_table = lto_obj_build_section_table (file, §ion_list); + + /* Find all sub modules in the object and put their sections into new hash + tables in a splay tree. */ + file_ids = lto_splay_tree_new (); + memset (&file_list, 0, sizeof (struct file_data_list)); + for (section = section_list.first; section != NULL; section = section->next) + create_subid_section_table (section, file_ids, &file_list); + + /* Add resolutions to file ids */ + lto_resolution_read (file_ids, resolution_file, file); + + /* Finalize each lto file for each submodule in the merged object */ + for (file_data = file_list.first; file_data != NULL; file_data = file_data->next) + lto_create_files_from_ids (file, file_data, count); + + splay_tree_delete (file_ids); + htab_delete (section_hash_table); + + return file_list.first; +} + +#if HAVE_MMAP_FILE && HAVE_SYSCONF && defined _SC_PAGE_SIZE +#define LTO_MMAP_IO 1 +#endif + +#if LTO_MMAP_IO +/* Page size of machine is used for mmap and munmap calls. */ +static size_t page_mask; +#endif + +/* Get the section data of length LEN from FILENAME starting at + OFFSET. The data segment must be freed by the caller when the + caller is finished. Returns NULL if all was not well. */ + +static char * +lto_read_section_data (struct lto_file_decl_data *file_data, + intptr_t offset, size_t len) +{ + char *result; + static int fd = -1; + static char *fd_name; +#if LTO_MMAP_IO + intptr_t computed_len; + intptr_t computed_offset; + intptr_t diff; +#endif + + /* Keep a single-entry file-descriptor cache. The last file we + touched will get closed at exit. + ??? Eventually we want to add a more sophisticated larger cache + or rather fix function body streaming to not stream them in + practically random order. */ + if (fd != -1 + && filename_cmp (fd_name, file_data->file_name) != 0) + { + free (fd_name); + close (fd); + fd = -1; + } + if (fd == -1) + { + fd = open (file_data->file_name, O_RDONLY|O_BINARY); + if (fd == -1) + { + fatal_error ("Cannot open %s", file_data->file_name); + return NULL; + } + fd_name = xstrdup (file_data->file_name); + } + +#if LTO_MMAP_IO + if (!page_mask) + { + size_t page_size = sysconf (_SC_PAGE_SIZE); + page_mask = ~(page_size - 1); + } + + computed_offset = offset & page_mask; + diff = offset - computed_offset; + computed_len = len + diff; + + result = (char *) mmap (NULL, computed_len, PROT_READ, MAP_PRIVATE, + fd, computed_offset); + if (result == MAP_FAILED) + { + fatal_error ("Cannot map %s", file_data->file_name); + return NULL; + } + + return result + diff; +#else + result = (char *) xmalloc (len); + if (lseek (fd, offset, SEEK_SET) != offset + || read (fd, result, len) != (ssize_t) len) + { + free (result); + fatal_error ("Cannot read %s", file_data->file_name); + result = NULL; + } +#ifdef __MINGW32__ + /* Native windows doesn't supports delayed unlink on opened file. So + we close file here again. This produces higher I/O load, but at least + it prevents to have dangling file handles preventing unlink. */ + free (fd_name); + fd_name = NULL; + close (fd); + fd = -1; +#endif + return result; +#endif +} + + +/* Get the section data from FILE_DATA of SECTION_TYPE with NAME. + NAME will be NULL unless the section type is for a function + body. */ + +static const char * +get_section_data (struct lto_file_decl_data *file_data, + enum lto_section_type section_type, + const char *name, + size_t *len) +{ + htab_t section_hash_table = file_data->section_hash_table; + struct lto_section_slot *f_slot; + struct lto_section_slot s_slot; + const char *section_name = lto_get_section_name (section_type, name, file_data); + char *data = NULL; + + *len = 0; + s_slot.name = section_name; + f_slot = (struct lto_section_slot *) htab_find (section_hash_table, &s_slot); + if (f_slot) + { + data = lto_read_section_data (file_data, f_slot->start, f_slot->len); + *len = f_slot->len; + } + + free (CONST_CAST (char *, section_name)); + return data; +} + + +/* Free the section data from FILE_DATA of SECTION_TYPE with NAME that + starts at OFFSET and has LEN bytes. */ + +static void +free_section_data (struct lto_file_decl_data *file_data ATTRIBUTE_UNUSED, + enum lto_section_type section_type ATTRIBUTE_UNUSED, + const char *name ATTRIBUTE_UNUSED, + const char *offset, size_t len ATTRIBUTE_UNUSED) +{ +#if LTO_MMAP_IO + intptr_t computed_len; + intptr_t computed_offset; + intptr_t diff; +#endif + +#if LTO_MMAP_IO + computed_offset = ((intptr_t) offset) & page_mask; + diff = (intptr_t) offset - computed_offset; + computed_len = len + diff; + + munmap ((caddr_t) computed_offset, computed_len); +#else + free (CONST_CAST(char *, offset)); +#endif +} + +static lto_file *current_lto_file; + +/* Helper for qsort; compare partitions and return one with smaller size. + We sort from greatest to smallest so parallel build doesn't stale on the + longest compilation being executed too late. */ + +static int +cmp_partitions_size (const void *a, const void *b) +{ + const struct ltrans_partition_def *pa + = *(struct ltrans_partition_def *const *)a; + const struct ltrans_partition_def *pb + = *(struct ltrans_partition_def *const *)b; + return pb->insns - pa->insns; +} + +/* Helper for qsort; compare partitions and return one with smaller order. */ + +static int +cmp_partitions_order (const void *a, const void *b) +{ + const struct ltrans_partition_def *pa + = *(struct ltrans_partition_def *const *)a; + const struct ltrans_partition_def *pb + = *(struct ltrans_partition_def *const *)b; + int ordera = -1, orderb = -1; + + if (lto_symtab_encoder_size (pa->encoder)) + ordera = lto_symtab_encoder_deref (pa->encoder, 0)->symbol.order; + if (lto_symtab_encoder_size (pb->encoder)) + orderb = lto_symtab_encoder_deref (pb->encoder, 0)->symbol.order; + return orderb - ordera; +} + +/* Write all output files in WPA mode and the file with the list of + LTRANS units. */ + +static void +lto_wpa_write_files (void) +{ + unsigned i, n_sets; + lto_file *file; + ltrans_partition part; + FILE *ltrans_output_list_stream; + char *temp_filename; + size_t blen; + + /* Open the LTRANS output list. */ + if (!ltrans_output_list) + fatal_error ("no LTRANS output list filename provided"); + ltrans_output_list_stream = fopen (ltrans_output_list, "w"); + if (ltrans_output_list_stream == NULL) + fatal_error ("opening LTRANS output list %s: %m", ltrans_output_list); + + timevar_push (TV_WHOPR_WPA); + + FOR_EACH_VEC_ELT (ltrans_partitions, i, part) + lto_stats.num_output_symtab_nodes += lto_symtab_encoder_size (part->encoder); + + /* Find out statics that need to be promoted + to globals with hidden visibility because they are accessed from multiple + partitions. */ + lto_promote_cross_file_statics (); + + timevar_pop (TV_WHOPR_WPA); + + timevar_push (TV_WHOPR_WPA_IO); + + /* Generate a prefix for the LTRANS unit files. */ + blen = strlen (ltrans_output_list); + temp_filename = (char *) xmalloc (blen + sizeof ("2147483648.o")); + strcpy (temp_filename, ltrans_output_list); + if (blen > sizeof (".out") + && strcmp (temp_filename + blen - sizeof (".out") + 1, + ".out") == 0) + temp_filename[blen - sizeof (".out") + 1] = '\0'; + blen = strlen (temp_filename); + + n_sets = ltrans_partitions.length (); + + /* Sort partitions by size so small ones are compiled last. + FIXME: Even when not reordering we may want to output one list for parallel make + and other for final link command. */ + ltrans_partitions.qsort (flag_toplevel_reorder + ? cmp_partitions_size + : cmp_partitions_order); + for (i = 0; i < n_sets; i++) + { + size_t len; + ltrans_partition part = ltrans_partitions[i]; + + /* Write all the nodes in SET. */ + sprintf (temp_filename + blen, "%u.o", i); + file = lto_obj_file_open (temp_filename, true); + if (!file) + fatal_error ("lto_obj_file_open() failed"); + + if (!quiet_flag) + fprintf (stderr, " %s (%s %i insns)", temp_filename, part->name, part->insns); + if (cgraph_dump_file) + { + lto_symtab_encoder_iterator lsei; + + fprintf (cgraph_dump_file, "Writing partition %s to file %s, %i insns\n", + part->name, temp_filename, part->insns); + fprintf (cgraph_dump_file, " Symbols in partition: "); + for (lsei = lsei_start_in_partition (part->encoder); !lsei_end_p (lsei); + lsei_next_in_partition (&lsei)) + { + symtab_node node = lsei_node (lsei); + fprintf (cgraph_dump_file, "%s ", symtab_node_asm_name (node)); + } + fprintf (cgraph_dump_file, "\n Symbols in boundary: "); + for (lsei = lsei_start (part->encoder); !lsei_end_p (lsei); + lsei_next (&lsei)) + { + symtab_node node = lsei_node (lsei); + if (!lto_symtab_encoder_in_partition_p (part->encoder, node)) + { + fprintf (cgraph_dump_file, "%s ", symtab_node_asm_name (node)); + cgraph_node *cnode = dyn_cast <cgraph_node> (node); + if (cnode + && lto_symtab_encoder_encode_body_p (part->encoder, cnode)) + fprintf (cgraph_dump_file, "(body included)"); + else + { + varpool_node *vnode = dyn_cast <varpool_node> (node); + if (vnode + && lto_symtab_encoder_encode_initializer_p (part->encoder, vnode)) + fprintf (cgraph_dump_file, "(initializer included)"); + } + } + } + fprintf (cgraph_dump_file, "\n"); + } + gcc_checking_assert (lto_symtab_encoder_size (part->encoder) || !i); + + lto_set_current_out_file (file); + + ipa_write_optimization_summaries (part->encoder); + + lto_set_current_out_file (NULL); + lto_obj_file_close (file); + free (file); + part->encoder = NULL; + + len = strlen (temp_filename); + if (fwrite (temp_filename, 1, len, ltrans_output_list_stream) < len + || fwrite ("\n", 1, 1, ltrans_output_list_stream) < 1) + fatal_error ("writing to LTRANS output list %s: %m", + ltrans_output_list); + } + + lto_stats.num_output_files += n_sets; + + /* Close the LTRANS output list. */ + if (fclose (ltrans_output_list_stream)) + fatal_error ("closing LTRANS output list %s: %m", ltrans_output_list); + + free_ltrans_partitions(); + free (temp_filename); + + timevar_pop (TV_WHOPR_WPA_IO); +} + + +/* If TT is a variable or function decl replace it with its + prevailing variant. */ +#define LTO_SET_PREVAIL(tt) \ + do {\ + if ((tt) && VAR_OR_FUNCTION_DECL_P (tt)) \ + tt = lto_symtab_prevailing_decl (tt); \ + } while (0) + +/* Ensure that TT isn't a replacable var of function decl. */ +#define LTO_NO_PREVAIL(tt) \ + gcc_assert (!(tt) || !VAR_OR_FUNCTION_DECL_P (tt)) + +/* Given a tree T replace all fields referring to variables or functions + with their prevailing variant. */ +static void +lto_fixup_prevailing_decls (tree t) +{ + enum tree_code code = TREE_CODE (t); + LTO_NO_PREVAIL (TREE_TYPE (t)); + if (CODE_CONTAINS_STRUCT (code, TS_COMMON)) + LTO_NO_PREVAIL (TREE_CHAIN (t)); + if (DECL_P (t)) + { + LTO_NO_PREVAIL (DECL_NAME (t)); + LTO_SET_PREVAIL (DECL_CONTEXT (t)); + if (CODE_CONTAINS_STRUCT (code, TS_DECL_COMMON)) + { + LTO_SET_PREVAIL (DECL_SIZE (t)); + LTO_SET_PREVAIL (DECL_SIZE_UNIT (t)); + LTO_SET_PREVAIL (DECL_INITIAL (t)); + LTO_NO_PREVAIL (DECL_ATTRIBUTES (t)); + LTO_SET_PREVAIL (DECL_ABSTRACT_ORIGIN (t)); + } + if (CODE_CONTAINS_STRUCT (code, TS_DECL_WITH_VIS)) + { + LTO_NO_PREVAIL (t->decl_with_vis.assembler_name); + LTO_NO_PREVAIL (DECL_SECTION_NAME (t)); + } + if (CODE_CONTAINS_STRUCT (code, TS_DECL_NON_COMMON)) + { + LTO_NO_PREVAIL (DECL_ARGUMENT_FLD (t)); + LTO_NO_PREVAIL (DECL_RESULT_FLD (t)); + LTO_NO_PREVAIL (DECL_VINDEX (t)); + } + if (CODE_CONTAINS_STRUCT (code, TS_FUNCTION_DECL)) + LTO_SET_PREVAIL (DECL_FUNCTION_PERSONALITY (t)); + if (CODE_CONTAINS_STRUCT (code, TS_FIELD_DECL)) + { + LTO_NO_PREVAIL (DECL_FIELD_OFFSET (t)); + LTO_NO_PREVAIL (DECL_BIT_FIELD_TYPE (t)); + LTO_NO_PREVAIL (DECL_QUALIFIER (t)); + LTO_NO_PREVAIL (DECL_FIELD_BIT_OFFSET (t)); + LTO_NO_PREVAIL (DECL_FCONTEXT (t)); + } + } + else if (TYPE_P (t)) + { + LTO_NO_PREVAIL (TYPE_CACHED_VALUES (t)); + LTO_SET_PREVAIL (TYPE_SIZE (t)); + LTO_SET_PREVAIL (TYPE_SIZE_UNIT (t)); + LTO_NO_PREVAIL (TYPE_ATTRIBUTES (t)); + LTO_NO_PREVAIL (TYPE_NAME (t)); + + LTO_SET_PREVAIL (TYPE_MINVAL (t)); + LTO_SET_PREVAIL (TYPE_MAXVAL (t)); + LTO_SET_PREVAIL (t->type_non_common.binfo); + + LTO_SET_PREVAIL (TYPE_CONTEXT (t)); + + LTO_NO_PREVAIL (TYPE_CANONICAL (t)); + LTO_NO_PREVAIL (TYPE_MAIN_VARIANT (t)); + LTO_NO_PREVAIL (TYPE_NEXT_VARIANT (t)); + } + else if (EXPR_P (t)) + { + int i; + for (i = TREE_OPERAND_LENGTH (t) - 1; i >= 0; --i) + LTO_SET_PREVAIL (TREE_OPERAND (t, i)); + } + else + { + switch (code) + { + case TREE_LIST: + LTO_SET_PREVAIL (TREE_VALUE (t)); + LTO_SET_PREVAIL (TREE_PURPOSE (t)); + break; + default: + gcc_unreachable (); + } + } +} +#undef LTO_SET_PREVAIL +#undef LTO_NO_PREVAIL + +/* Helper function of lto_fixup_decls. Walks the var and fn streams in STATE, + replaces var and function decls with the corresponding prevailing def. */ + +static void +lto_fixup_state (struct lto_in_decl_state *state) +{ + unsigned i, si; + struct lto_tree_ref_table *table; + + /* Although we only want to replace FUNCTION_DECLs and VAR_DECLs, + we still need to walk from all DECLs to find the reachable + FUNCTION_DECLs and VAR_DECLs. */ + for (si = 0; si < LTO_N_DECL_STREAMS; si++) + { + table = &state->streams[si]; + for (i = 0; i < table->size; i++) + { + tree *tp = table->trees + i; + if (VAR_OR_FUNCTION_DECL_P (*tp)) + *tp = lto_symtab_prevailing_decl (*tp); + } + } +} + +/* A callback of htab_traverse. Just extracts a state from SLOT + and calls lto_fixup_state. */ + +static int +lto_fixup_state_aux (void **slot, void *aux ATTRIBUTE_UNUSED) +{ + struct lto_in_decl_state *state = (struct lto_in_decl_state *) *slot; + lto_fixup_state (state); + return 1; +} + +/* Fix the decls from all FILES. Replaces each decl with the corresponding + prevailing one. */ + +static void +lto_fixup_decls (struct lto_file_decl_data **files) +{ + unsigned int i; + htab_iterator hi; + tree t; + + FOR_EACH_HTAB_ELEMENT (tree_with_vars, t, tree, hi) + lto_fixup_prevailing_decls (t); + + for (i = 0; files[i]; i++) + { + struct lto_file_decl_data *file = files[i]; + struct lto_in_decl_state *state = file->global_decl_state; + lto_fixup_state (state); + + htab_traverse (file->function_decl_states, lto_fixup_state_aux, NULL); + } +} + +static GTY((length ("lto_stats.num_input_files + 1"))) struct lto_file_decl_data **all_file_decl_data; + +/* Turn file datas for sub files into a single array, so that they look + like separate files for further passes. */ + +static void +lto_flatten_files (struct lto_file_decl_data **orig, int count, int last_file_ix) +{ + struct lto_file_decl_data *n, *next; + int i, k; + + lto_stats.num_input_files = count; + all_file_decl_data + = ggc_alloc_cleared_vec_lto_file_decl_data_ptr (count + 1); + /* Set the hooks so that all of the ipa passes can read in their data. */ + lto_set_in_hooks (all_file_decl_data, get_section_data, free_section_data); + for (i = 0, k = 0; i < last_file_ix; i++) + { + for (n = orig[i]; n != NULL; n = next) + { + all_file_decl_data[k++] = n; + next = n->next; + n->next = NULL; + } + } + all_file_decl_data[k] = NULL; + gcc_assert (k == count); +} + +/* Input file data before flattening (i.e. splitting them to subfiles to support + incremental linking. */ +static int real_file_count; +static GTY((length ("real_file_count + 1"))) struct lto_file_decl_data **real_file_decl_data; + +/* Read all the symbols from the input files FNAMES. NFILES is the + number of files requested in the command line. Instantiate a + global call graph by aggregating all the sub-graphs found in each + file. */ + +static void +read_cgraph_and_symbols (unsigned nfiles, const char **fnames) +{ + unsigned int i, last_file_ix; + FILE *resolution; + struct cgraph_node *node; + int count = 0; + struct lto_file_decl_data **decl_data; + + init_cgraph (); + + timevar_push (TV_IPA_LTO_DECL_IN); + + real_file_decl_data + = decl_data = ggc_alloc_cleared_vec_lto_file_decl_data_ptr (nfiles + 1); + real_file_count = nfiles; + + /* Read the resolution file. */ + resolution = NULL; + if (resolution_file_name) + { + int t; + unsigned num_objects; + + resolution = fopen (resolution_file_name, "r"); + if (resolution == NULL) + fatal_error ("could not open symbol resolution file: %m"); + + t = fscanf (resolution, "%u", &num_objects); + gcc_assert (t == 1); + + /* True, since the plugin splits the archives. */ + gcc_assert (num_objects == nfiles); + } + + tree_with_vars = htab_create_ggc (101, htab_hash_pointer, htab_eq_pointer, + NULL); + type_hash_cache = htab_create_ggc (512, tree_int_map_hash, + tree_int_map_eq, NULL); + type_pair_cache = XCNEWVEC (struct type_pair_d, GIMPLE_TYPE_PAIR_SIZE); + gimple_type_leader = ggc_alloc_cleared_vec_gimple_type_leader_entry_s + (GIMPLE_TYPE_LEADER_SIZE); + gimple_types = htab_create_ggc (16381, gimple_type_hash, gimple_type_eq, 0); + + if (!quiet_flag) + fprintf (stderr, "Reading object files:"); + + /* Read all of the object files specified on the command line. */ + for (i = 0, last_file_ix = 0; i < nfiles; ++i) + { + struct lto_file_decl_data *file_data = NULL; + if (!quiet_flag) + { + fprintf (stderr, " %s", fnames[i]); + fflush (stderr); + } + + current_lto_file = lto_obj_file_open (fnames[i], false); + if (!current_lto_file) + break; + + file_data = lto_file_read (current_lto_file, resolution, &count); + if (!file_data) + { + lto_obj_file_close (current_lto_file); + free (current_lto_file); + current_lto_file = NULL; + break; + } + + decl_data[last_file_ix++] = file_data; + + lto_obj_file_close (current_lto_file); + free (current_lto_file); + current_lto_file = NULL; + ggc_collect (); + } + + lto_flatten_files (decl_data, count, last_file_ix); + lto_stats.num_input_files = count; + ggc_free(decl_data); + real_file_decl_data = NULL; + + if (resolution_file_name) + fclose (resolution); + + /* Free gimple type merging datastructures. */ + htab_delete (gimple_types); + gimple_types = NULL; + htab_delete (type_hash_cache); + type_hash_cache = NULL; + free (type_pair_cache); + type_pair_cache = NULL; + gimple_type_leader = NULL; + free_gimple_type_tables (); + ggc_collect (); + + /* Set the hooks so that all of the ipa passes can read in their data. */ + lto_set_in_hooks (all_file_decl_data, get_section_data, free_section_data); + + timevar_pop (TV_IPA_LTO_DECL_IN); + + if (!quiet_flag) + fprintf (stderr, "\nReading the callgraph\n"); + + timevar_push (TV_IPA_LTO_CGRAPH_IO); + /* Read the symtab. */ + input_symtab (); + + /* Store resolutions into the symbol table. */ + if (resolution_map) + { + void **res; + symtab_node snode; + + FOR_EACH_SYMBOL (snode) + if (symtab_real_symbol_p (snode) + && (res = pointer_map_contains (resolution_map, + snode->symbol.decl))) + snode->symbol.resolution + = (enum ld_plugin_symbol_resolution)(size_t)*res; + + pointer_map_destroy (resolution_map); + resolution_map = NULL; + } + + timevar_pop (TV_IPA_LTO_CGRAPH_IO); + + if (!quiet_flag) + fprintf (stderr, "Merging declarations\n"); + + timevar_push (TV_IPA_LTO_DECL_MERGE); + /* Merge global decls. In ltrans mode we read merged cgraph, we do not + need to care about resolving symbols again, we only need to replace + duplicated declarations read from the callgraph and from function + sections. */ + if (!flag_ltrans) + { + lto_symtab_merge_decls (); + + /* If there were errors during symbol merging bail out, we have no + good way to recover here. */ + if (seen_error ()) + fatal_error ("errors during merging of translation units"); + + /* Fixup all decls. */ + lto_fixup_decls (all_file_decl_data); + } + htab_delete (tree_with_vars); + tree_with_vars = NULL; + ggc_collect (); + + timevar_pop (TV_IPA_LTO_DECL_MERGE); + /* Each pass will set the appropriate timer. */ + + if (!quiet_flag) + fprintf (stderr, "Reading summaries\n"); + + /* Read the IPA summary data. */ + if (flag_ltrans) + ipa_read_optimization_summaries (); + else + ipa_read_summaries (); + + for (i = 0; all_file_decl_data[i]; i++) + { + gcc_assert (all_file_decl_data[i]->symtab_node_encoder); + lto_symtab_encoder_delete (all_file_decl_data[i]->symtab_node_encoder); + all_file_decl_data[i]->symtab_node_encoder = NULL; + } + + /* Finally merge the cgraph according to the decl merging decisions. */ + timevar_push (TV_IPA_LTO_CGRAPH_MERGE); + if (cgraph_dump_file) + { + fprintf (cgraph_dump_file, "Before merging:\n"); + dump_cgraph (cgraph_dump_file); + dump_varpool (cgraph_dump_file); + } + lto_symtab_merge_cgraph_nodes (); + ggc_collect (); + + /* FIXME: ipa_transforms_to_apply holds list of passes that have optimization + summaries computed and needs to apply changes. At the moment WHOPR only + supports inlining, so we can push it here by hand. In future we need to stream + this field into ltrans compilation. */ + if (flag_ltrans) + FOR_EACH_DEFINED_FUNCTION (node) + node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass)&pass_ipa_inline); + + timevar_pop (TV_IPA_LTO_CGRAPH_MERGE); + + timevar_push (TV_IPA_LTO_DECL_INIT_IO); + + /* Indicate that the cgraph is built and ready. */ + cgraph_function_flags_ready = true; + + timevar_pop (TV_IPA_LTO_DECL_INIT_IO); + ggc_free (all_file_decl_data); + all_file_decl_data = NULL; +} + + +/* Materialize all the bodies for all the nodes in the callgraph. */ + +static void +materialize_cgraph (void) +{ + tree decl; + struct cgraph_node *node; + unsigned i; + timevar_id_t lto_timer; + + if (!quiet_flag) + fprintf (stderr, + flag_wpa ? "Materializing decls:" : "Reading function bodies:"); + + /* Now that we have input the cgraph, we need to clear all of the aux + nodes and read the functions if we are not running in WPA mode. */ + timevar_push (TV_IPA_LTO_GIMPLE_IN); + + FOR_EACH_FUNCTION (node) + { + if (node->symbol.lto_file_data) + { + lto_materialize_function (node); + lto_stats.num_input_cgraph_nodes++; + } + } + + timevar_pop (TV_IPA_LTO_GIMPLE_IN); + + /* Start the appropriate timer depending on the mode that we are + operating in. */ + lto_timer = (flag_wpa) ? TV_WHOPR_WPA + : (flag_ltrans) ? TV_WHOPR_LTRANS + : TV_LTO; + timevar_push (lto_timer); + + current_function_decl = NULL; + set_cfun (NULL); + + /* Inform the middle end about the global variables we have seen. */ + FOR_EACH_VEC_ELT (*lto_global_var_decls, i, decl) + rest_of_decl_compilation (decl, 1, 0); + + if (!quiet_flag) + fprintf (stderr, "\n"); + + timevar_pop (lto_timer); +} + + +/* Show various memory usage statistics related to LTO. */ +static void +print_lto_report_1 (void) +{ + const char *pfx = (flag_lto) ? "LTO" : (flag_wpa) ? "WPA" : "LTRANS"; + fprintf (stderr, "%s statistics\n", pfx); + + if (gimple_types) + fprintf (stderr, "[%s] GIMPLE type table: size %ld, %ld elements, " + "%ld searches, %ld collisions (ratio: %f)\n", pfx, + (long) htab_size (gimple_types), + (long) htab_elements (gimple_types), + (long) gimple_types->searches, + (long) gimple_types->collisions, + htab_collisions (gimple_types)); + else + fprintf (stderr, "[%s] GIMPLE type table is empty\n", pfx); + if (type_hash_cache) + fprintf (stderr, "[%s] GIMPLE type hash table: size %ld, %ld elements, " + "%ld searches, %ld collisions (ratio: %f)\n", pfx, + (long) htab_size (type_hash_cache), + (long) htab_elements (type_hash_cache), + (long) type_hash_cache->searches, + (long) type_hash_cache->collisions, + htab_collisions (type_hash_cache)); + else + fprintf (stderr, "[%s] GIMPLE type hash table is empty\n", pfx); + + print_gimple_types_stats (pfx); + print_lto_report (pfx); +} + +/* Perform whole program analysis (WPA) on the callgraph and write out the + optimization plan. */ + +static void +do_whole_program_analysis (void) +{ + symtab_node node; + + timevar_start (TV_PHASE_OPT_GEN); + + /* Note that since we are in WPA mode, materialize_cgraph will not + actually read in all the function bodies. It only materializes + the decls and cgraph nodes so that analysis can be performed. */ + materialize_cgraph (); + + /* Reading in the cgraph uses different timers, start timing WPA now. */ + timevar_push (TV_WHOPR_WPA); + + if (pre_ipa_mem_report) + { + fprintf (stderr, "Memory consumption before IPA\n"); + dump_memory_report (false); + } + + cgraph_function_flags_ready = true; + + if (cgraph_dump_file) + { + dump_cgraph (cgraph_dump_file); + dump_varpool (cgraph_dump_file); + } + bitmap_obstack_initialize (NULL); + cgraph_state = CGRAPH_STATE_IPA_SSA; + + execute_ipa_pass_list (all_regular_ipa_passes); + symtab_remove_unreachable_nodes (false, dump_file); + + if (cgraph_dump_file) + { + fprintf (cgraph_dump_file, "Optimized "); + dump_cgraph (cgraph_dump_file); + dump_varpool (cgraph_dump_file); + } +#ifdef ENABLE_CHECKING + verify_cgraph (); +#endif + bitmap_obstack_release (NULL); + + /* We are about to launch the final LTRANS phase, stop the WPA timer. */ + timevar_pop (TV_WHOPR_WPA); + + timevar_push (TV_WHOPR_PARTITIONING); + if (flag_lto_partition_1to1) + lto_1_to_1_map (); + else if (flag_lto_partition_max) + lto_max_map (); + else + lto_balanced_map (); + + /* AUX pointers are used by partitioning code to bookkeep number of + partitions symbol is in. This is no longer needed. */ + FOR_EACH_SYMBOL (node) + node->symbol.aux = NULL; + + lto_stats.num_cgraph_partitions += ltrans_partitions.length (); + timevar_pop (TV_WHOPR_PARTITIONING); + + timevar_stop (TV_PHASE_OPT_GEN); + timevar_start (TV_PHASE_STREAM_OUT); + + if (!quiet_flag) + { + fprintf (stderr, "\nStreaming out"); + fflush (stderr); + } + lto_wpa_write_files (); + if (!quiet_flag) + fprintf (stderr, "\n"); + + timevar_stop (TV_PHASE_STREAM_OUT); + + ggc_collect (); + if (post_ipa_mem_report) + { + fprintf (stderr, "Memory consumption after IPA\n"); + dump_memory_report (false); + } + + /* Show the LTO report before launching LTRANS. */ + if (flag_lto_report) + print_lto_report_1 (); + if (mem_report_wpa) + dump_memory_report (true); +} + + +static GTY(()) tree lto_eh_personality_decl; + +/* Return the LTO personality function decl. */ + +tree +lto_eh_personality (void) +{ + if (!lto_eh_personality_decl) + { + /* Use the first personality DECL for our personality if we don't + support multiple ones. This ensures that we don't artificially + create the need for them in a single-language program. */ + if (first_personality_decl && !dwarf2out_do_cfi_asm ()) + lto_eh_personality_decl = first_personality_decl; + else + lto_eh_personality_decl = lhd_gcc_personality (); + } + + return lto_eh_personality_decl; +} + +/* Set the process name based on the LTO mode. */ + +static void +lto_process_name (void) +{ + if (flag_lto) + setproctitle ("lto1-lto"); + if (flag_wpa) + setproctitle ("lto1-wpa"); + if (flag_ltrans) + setproctitle ("lto1-ltrans"); +} + + +/* Initialize the LTO front end. */ + +static void +lto_init (void) +{ + lto_process_name (); + lto_streamer_hooks_init (); + lto_reader_init (); + lto_set_in_hooks (NULL, get_section_data, free_section_data); + memset (<o_stats, 0, sizeof (lto_stats)); + bitmap_obstack_initialize (NULL); + gimple_register_cfg_hooks (); +} + + +/* Main entry point for the GIMPLE front end. This front end has + three main personalities: + + - LTO (-flto). All the object files on the command line are + loaded in memory and processed as a single translation unit. + This is the traditional link-time optimization behavior. + + - WPA (-fwpa). Only the callgraph and summary information for + files in the command file are loaded. A single callgraph + (without function bodies) is instantiated for the whole set of + files. IPA passes are only allowed to analyze the call graph + and make transformation decisions. The callgraph is + partitioned, each partition is written to a new object file + together with the transformation decisions. + + - LTRANS (-fltrans). Similar to -flto but it prevents the IPA + summary files from running again. Since WPA computed summary + information and decided what transformations to apply, LTRANS + simply applies them. */ + +void +lto_main (void) +{ + /* LTO is called as a front end, even though it is not a front end. + Because it is called as a front end, TV_PHASE_PARSING and + TV_PARSE_GLOBAL are active, and we need to turn them off while + doing LTO. Later we turn them back on so they are active up in + toplev.c. */ + timevar_pop (TV_PARSE_GLOBAL); + timevar_stop (TV_PHASE_PARSING); + + timevar_start (TV_PHASE_SETUP); + + /* Initialize the LTO front end. */ + lto_init (); + + timevar_stop (TV_PHASE_SETUP); + timevar_start (TV_PHASE_STREAM_IN); + + /* Read all the symbols and call graph from all the files in the + command line. */ + read_cgraph_and_symbols (num_in_fnames, in_fnames); + + timevar_stop (TV_PHASE_STREAM_IN); + + if (!seen_error ()) + { + /* If WPA is enabled analyze the whole call graph and create an + optimization plan. Otherwise, read in all the function + bodies and continue with optimization. */ + if (flag_wpa) + do_whole_program_analysis (); + else + { + struct varpool_node *vnode; + + timevar_start (TV_PHASE_OPT_GEN); + + materialize_cgraph (); + + /* Let the middle end know that we have read and merged all of + the input files. */ + compile (); + + timevar_stop (TV_PHASE_OPT_GEN); + + /* FIXME lto, if the processes spawned by WPA fail, we miss + the chance to print WPA's report, so WPA will call + print_lto_report before launching LTRANS. If LTRANS was + launched directly by the driver we would not need to do + this. */ + if (flag_lto_report) + print_lto_report_1 (); + + /* Record the global variables. */ + FOR_EACH_DEFINED_VARIABLE (vnode) + vec_safe_push (lto_global_var_decls, vnode->symbol.decl); + } + } + + /* Here we make LTO pretend to be a parser. */ + timevar_start (TV_PHASE_PARSING); + timevar_push (TV_PARSE_GLOBAL); +} + +#include "gt-lto-lto.h" |