mirror of
https://github.com/autc04/Retro68.git
synced 2024-06-06 21:29:40 +00:00
3111 lines
92 KiB
C++
3111 lines
92 KiB
C++
/* Top-level LTO routines.
|
|
Copyright (C) 2009-2022 Free Software Foundation, Inc.
|
|
Contributed by CodeSourcery, Inc.
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify it under
|
|
the terms of the GNU General Public License as published by the Free
|
|
Software Foundation; either version 3, or (at your option) any later
|
|
version.
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GCC; see the file COPYING3. If not see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "coretypes.h"
|
|
#include "tm.h"
|
|
#include "function.h"
|
|
#include "bitmap.h"
|
|
#include "basic-block.h"
|
|
#include "tree.h"
|
|
#include "gimple.h"
|
|
#include "cfghooks.h"
|
|
#include "alloc-pool.h"
|
|
#include "tree-pass.h"
|
|
#include "tree-streamer.h"
|
|
#include "cgraph.h"
|
|
#include "opts.h"
|
|
#include "toplev.h"
|
|
#include "stor-layout.h"
|
|
#include "symbol-summary.h"
|
|
#include "tree-vrp.h"
|
|
#include "ipa-prop.h"
|
|
#include "common.h"
|
|
#include "debug.h"
|
|
#include "lto.h"
|
|
#include "lto-section-names.h"
|
|
#include "splay-tree.h"
|
|
#include "lto-partition.h"
|
|
#include "context.h"
|
|
#include "pass_manager.h"
|
|
#include "ipa-fnsummary.h"
|
|
#include "ipa-utils.h"
|
|
#include "gomp-constants.h"
|
|
#include "lto-symtab.h"
|
|
#include "stringpool.h"
|
|
#include "fold-const.h"
|
|
#include "attribs.h"
|
|
#include "builtins.h"
|
|
#include "lto-common.h"
|
|
#include "tree-pretty-print.h"
|
|
#include "print-tree.h"
|
|
|
|
/* True when no new types are going to be streamd from the global stream. */
|
|
|
|
static bool type_streaming_finished = false;
|
|
|
|
GTY(()) tree first_personality_decl;
|
|
|
|
GTY(()) const unsigned char *lto_mode_identity_table;
|
|
|
|
/* Returns a hash code for P. */
|
|
|
|
static hashval_t
|
|
hash_name (const void *p)
|
|
{
|
|
const struct lto_section_slot *ds = (const struct lto_section_slot *) p;
|
|
return (hashval_t) htab_hash_string (ds->name);
|
|
}
|
|
|
|
|
|
/* Returns nonzero if P1 and P2 are equal. */
|
|
|
|
static int
|
|
eq_name (const void *p1, const void *p2)
|
|
{
|
|
const struct lto_section_slot *s1
|
|
= (const struct lto_section_slot *) p1;
|
|
const struct lto_section_slot *s2
|
|
= (const struct lto_section_slot *) p2;
|
|
|
|
return strcmp (s1->name, s2->name) == 0;
|
|
}
|
|
|
|
/* Free lto_section_slot. */
|
|
|
|
static void
|
|
free_with_string (void *arg)
|
|
{
|
|
struct lto_section_slot *s = (struct lto_section_slot *)arg;
|
|
|
|
free (CONST_CAST (char *, s->name));
|
|
free (arg);
|
|
}
|
|
|
|
/* Create section hash table. */
|
|
|
|
htab_t
|
|
lto_obj_create_section_hash_table (void)
|
|
{
|
|
return htab_create (37, hash_name, eq_name, free_with_string);
|
|
}
|
|
|
|
/* Delete an allocated integer KEY in the splay tree. */
|
|
|
|
static void
|
|
lto_splay_tree_delete_id (splay_tree_key key)
|
|
{
|
|
free ((void *) key);
|
|
}
|
|
|
|
/* Compare splay tree node ids A and B. */
|
|
|
|
static int
|
|
lto_splay_tree_compare_ids (splay_tree_key a, splay_tree_key b)
|
|
{
|
|
unsigned HOST_WIDE_INT ai;
|
|
unsigned HOST_WIDE_INT bi;
|
|
|
|
ai = *(unsigned HOST_WIDE_INT *) a;
|
|
bi = *(unsigned HOST_WIDE_INT *) b;
|
|
|
|
if (ai < bi)
|
|
return -1;
|
|
else if (ai > bi)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
/* Look up splay tree node by ID in splay tree T. */
|
|
|
|
static splay_tree_node
|
|
lto_splay_tree_lookup (splay_tree t, unsigned HOST_WIDE_INT id)
|
|
{
|
|
return splay_tree_lookup (t, (splay_tree_key) &id);
|
|
}
|
|
|
|
/* Check if KEY has ID. */
|
|
|
|
static bool
|
|
lto_splay_tree_id_equal_p (splay_tree_key key, unsigned HOST_WIDE_INT id)
|
|
{
|
|
return *(unsigned HOST_WIDE_INT *) key == id;
|
|
}
|
|
|
|
/* Insert a splay tree node into tree T with ID as key and FILE_DATA as value.
|
|
The ID is allocated separately because we need HOST_WIDE_INTs which may
|
|
be wider than a splay_tree_key. */
|
|
|
|
static void
|
|
lto_splay_tree_insert (splay_tree t, unsigned HOST_WIDE_INT id,
|
|
struct lto_file_decl_data *file_data)
|
|
{
|
|
unsigned HOST_WIDE_INT *idp = XCNEW (unsigned HOST_WIDE_INT);
|
|
*idp = id;
|
|
splay_tree_insert (t, (splay_tree_key) idp, (splay_tree_value) file_data);
|
|
}
|
|
|
|
/* Create a splay tree. */
|
|
|
|
static splay_tree
|
|
lto_splay_tree_new (void)
|
|
{
|
|
return splay_tree_new (lto_splay_tree_compare_ids,
|
|
lto_splay_tree_delete_id,
|
|
NULL);
|
|
}
|
|
|
|
/* Decode the content of memory pointed to by DATA in the in decl
|
|
state object STATE. DATA_IN points to a data_in structure for
|
|
decoding. Return the address after the decoded object in the
|
|
input. */
|
|
|
|
static const uint32_t *
|
|
lto_read_in_decl_state (class data_in *data_in, const uint32_t *data,
|
|
struct lto_in_decl_state *state)
|
|
{
|
|
uint32_t ix;
|
|
tree decl;
|
|
uint32_t i, j;
|
|
|
|
ix = *data++;
|
|
state->compressed = ix & 1;
|
|
ix /= 2;
|
|
decl = streamer_tree_cache_get_tree (data_in->reader_cache, ix);
|
|
if (!VAR_OR_FUNCTION_DECL_P (decl))
|
|
{
|
|
gcc_assert (decl == void_type_node);
|
|
decl = NULL_TREE;
|
|
}
|
|
state->fn_decl = decl;
|
|
|
|
for (i = 0; i < LTO_N_DECL_STREAMS; i++)
|
|
{
|
|
uint32_t size = *data++;
|
|
vec<tree, va_gc> *decls = NULL;
|
|
vec_alloc (decls, size);
|
|
|
|
for (j = 0; j < size; j++)
|
|
vec_safe_push (decls,
|
|
streamer_tree_cache_get_tree (data_in->reader_cache,
|
|
data[j]));
|
|
|
|
state->streams[i] = decls;
|
|
data += size;
|
|
}
|
|
|
|
return data;
|
|
}
|
|
|
|
|
|
/* Global canonical type table. */
|
|
static htab_t gimple_canonical_types;
|
|
static hash_map<const_tree, hashval_t> *canonical_type_hash_cache;
|
|
static unsigned long num_canonical_type_hash_entries;
|
|
static unsigned long num_canonical_type_hash_queries;
|
|
|
|
/* Types postponed for registration to the canonical type table.
|
|
During streaming we postpone all TYPE_CXX_ODR_P types so we can alter
|
|
decide whether there is conflict with non-ODR type or not. */
|
|
static GTY(()) vec<tree, va_gc> *types_to_register = NULL;
|
|
|
|
static void iterative_hash_canonical_type (tree type, inchash::hash &hstate);
|
|
static hashval_t gimple_canonical_type_hash (const void *p);
|
|
static hashval_t gimple_register_canonical_type_1 (tree t, hashval_t hash);
|
|
|
|
/* Returning a hash value for gimple type TYPE.
|
|
|
|
The hash value returned is equal for types considered compatible
|
|
by gimple_canonical_types_compatible_p. */
|
|
|
|
static hashval_t
|
|
hash_canonical_type (tree type)
|
|
{
|
|
inchash::hash hstate;
|
|
enum tree_code code;
|
|
|
|
/* We compute alias sets only for types that needs them.
|
|
Be sure we do not recurse to something else as we cannot hash incomplete
|
|
types in a way they would have same hash value as compatible complete
|
|
types. */
|
|
gcc_checking_assert (type_with_alias_set_p (type));
|
|
|
|
/* Combine a few common features of types so that types are grouped into
|
|
smaller sets; when searching for existing matching types to merge,
|
|
only existing types having the same features as the new type will be
|
|
checked. */
|
|
code = tree_code_for_canonical_type_merging (TREE_CODE (type));
|
|
hstate.add_int (code);
|
|
hstate.add_int (TYPE_MODE (type));
|
|
|
|
/* Incorporate common features of numerical types. */
|
|
if (INTEGRAL_TYPE_P (type)
|
|
|| SCALAR_FLOAT_TYPE_P (type)
|
|
|| FIXED_POINT_TYPE_P (type)
|
|
|| TREE_CODE (type) == OFFSET_TYPE
|
|
|| POINTER_TYPE_P (type))
|
|
{
|
|
hstate.add_int (TYPE_PRECISION (type));
|
|
if (!type_with_interoperable_signedness (type))
|
|
hstate.add_int (TYPE_UNSIGNED (type));
|
|
}
|
|
|
|
if (VECTOR_TYPE_P (type))
|
|
{
|
|
hstate.add_poly_int (TYPE_VECTOR_SUBPARTS (type));
|
|
hstate.add_int (TYPE_UNSIGNED (type));
|
|
}
|
|
|
|
if (TREE_CODE (type) == COMPLEX_TYPE)
|
|
hstate.add_int (TYPE_UNSIGNED (type));
|
|
|
|
/* Fortran's C_SIGNED_CHAR is !TYPE_STRING_FLAG but needs to be
|
|
interoperable with "signed char". Unless all frontends are revisited to
|
|
agree on these types, we must ignore the flag completely. */
|
|
|
|
/* Fortran standard define C_PTR type that is compatible with every
|
|
C pointer. For this reason we need to glob all pointers into one.
|
|
Still pointers in different address spaces are not compatible. */
|
|
if (POINTER_TYPE_P (type))
|
|
hstate.add_int (TYPE_ADDR_SPACE (TREE_TYPE (type)));
|
|
|
|
/* For array types hash the domain bounds and the string flag. */
|
|
if (TREE_CODE (type) == ARRAY_TYPE && TYPE_DOMAIN (type))
|
|
{
|
|
hstate.add_int (TYPE_STRING_FLAG (type));
|
|
/* OMP lowering can introduce error_mark_node in place of
|
|
random local decls in types. */
|
|
if (TYPE_MIN_VALUE (TYPE_DOMAIN (type)) != error_mark_node)
|
|
inchash::add_expr (TYPE_MIN_VALUE (TYPE_DOMAIN (type)), hstate);
|
|
if (TYPE_MAX_VALUE (TYPE_DOMAIN (type)) != error_mark_node)
|
|
inchash::add_expr (TYPE_MAX_VALUE (TYPE_DOMAIN (type)), hstate);
|
|
}
|
|
|
|
/* Recurse for aggregates with a single element type. */
|
|
if (TREE_CODE (type) == ARRAY_TYPE
|
|
|| TREE_CODE (type) == COMPLEX_TYPE
|
|
|| TREE_CODE (type) == VECTOR_TYPE)
|
|
iterative_hash_canonical_type (TREE_TYPE (type), hstate);
|
|
|
|
/* Incorporate function return and argument types. */
|
|
if (TREE_CODE (type) == FUNCTION_TYPE || TREE_CODE (type) == METHOD_TYPE)
|
|
{
|
|
unsigned na;
|
|
tree p;
|
|
|
|
iterative_hash_canonical_type (TREE_TYPE (type), hstate);
|
|
|
|
for (p = TYPE_ARG_TYPES (type), na = 0; p; p = TREE_CHAIN (p))
|
|
{
|
|
iterative_hash_canonical_type (TREE_VALUE (p), hstate);
|
|
na++;
|
|
}
|
|
|
|
hstate.add_int (na);
|
|
}
|
|
|
|
if (RECORD_OR_UNION_TYPE_P (type))
|
|
{
|
|
unsigned nf;
|
|
tree f;
|
|
|
|
for (f = TYPE_FIELDS (type), nf = 0; f; f = TREE_CHAIN (f))
|
|
if (TREE_CODE (f) == FIELD_DECL
|
|
&& (! DECL_SIZE (f)
|
|
|| ! integer_zerop (DECL_SIZE (f))))
|
|
{
|
|
iterative_hash_canonical_type (TREE_TYPE (f), hstate);
|
|
nf++;
|
|
}
|
|
|
|
hstate.add_int (nf);
|
|
}
|
|
|
|
return hstate.end();
|
|
}
|
|
|
|
/* Returning a hash value for gimple type TYPE combined with VAL. */
|
|
|
|
static void
|
|
iterative_hash_canonical_type (tree type, inchash::hash &hstate)
|
|
{
|
|
hashval_t v;
|
|
|
|
/* All type variants have same TYPE_CANONICAL. */
|
|
type = TYPE_MAIN_VARIANT (type);
|
|
|
|
if (!canonical_type_used_p (type))
|
|
v = hash_canonical_type (type);
|
|
/* An already processed type. */
|
|
else if (TYPE_CANONICAL (type))
|
|
{
|
|
type = TYPE_CANONICAL (type);
|
|
v = gimple_canonical_type_hash (type);
|
|
}
|
|
else
|
|
{
|
|
/* Canonical types should not be able to form SCCs by design, this
|
|
recursion is just because we do not register canonical types in
|
|
optimal order. To avoid quadratic behavior also register the
|
|
type here. */
|
|
v = hash_canonical_type (type);
|
|
v = gimple_register_canonical_type_1 (type, v);
|
|
}
|
|
hstate.merge_hash (v);
|
|
}
|
|
|
|
/* Returns the hash for a canonical type P. */
|
|
|
|
static hashval_t
|
|
gimple_canonical_type_hash (const void *p)
|
|
{
|
|
num_canonical_type_hash_queries++;
|
|
hashval_t *slot = canonical_type_hash_cache->get ((const_tree) p);
|
|
gcc_assert (slot != NULL);
|
|
return *slot;
|
|
}
|
|
|
|
|
|
|
|
/* Returns nonzero if P1 and P2 are equal. */
|
|
|
|
static int
|
|
gimple_canonical_type_eq (const void *p1, const void *p2)
|
|
{
|
|
const_tree t1 = (const_tree) p1;
|
|
const_tree t2 = (const_tree) p2;
|
|
return gimple_canonical_types_compatible_p (CONST_CAST_TREE (t1),
|
|
CONST_CAST_TREE (t2));
|
|
}
|
|
|
|
/* Main worker for gimple_register_canonical_type. */
|
|
|
|
static hashval_t
|
|
gimple_register_canonical_type_1 (tree t, hashval_t hash)
|
|
{
|
|
void **slot;
|
|
|
|
gcc_checking_assert (TYPE_P (t) && !TYPE_CANONICAL (t)
|
|
&& type_with_alias_set_p (t)
|
|
&& canonical_type_used_p (t));
|
|
|
|
/* ODR types for which there is no ODR violation and we did not record
|
|
structurally equivalent non-ODR type can be treated as unique by their
|
|
name.
|
|
|
|
hash passed to gimple_register_canonical_type_1 is a structural hash
|
|
that we can use to lookup structurally equivalent non-ODR type.
|
|
In case we decide to treat type as unique ODR type we recompute hash based
|
|
on name and let TBAA machinery know about our decision. */
|
|
if (RECORD_OR_UNION_TYPE_P (t) && odr_type_p (t)
|
|
&& TYPE_CXX_ODR_P (t) && !odr_type_violation_reported_p (t))
|
|
{
|
|
/* Anonymous namespace types never conflict with non-C++ types. */
|
|
if (type_with_linkage_p (t) && type_in_anonymous_namespace_p (t))
|
|
slot = NULL;
|
|
else
|
|
{
|
|
/* Here we rely on fact that all non-ODR types was inserted into
|
|
canonical type hash and thus we can safely detect conflicts between
|
|
ODR types and interoperable non-ODR types. */
|
|
gcc_checking_assert (type_streaming_finished
|
|
&& TYPE_MAIN_VARIANT (t) == t);
|
|
slot = htab_find_slot_with_hash (gimple_canonical_types, t, hash,
|
|
NO_INSERT);
|
|
}
|
|
if (slot && !TYPE_CXX_ODR_P (*(tree *)slot))
|
|
{
|
|
tree nonodr = *(tree *)slot;
|
|
gcc_checking_assert (!flag_ltrans);
|
|
if (symtab->dump_file)
|
|
{
|
|
fprintf (symtab->dump_file,
|
|
"ODR and non-ODR type conflict: ");
|
|
print_generic_expr (symtab->dump_file, t);
|
|
fprintf (symtab->dump_file, " and ");
|
|
print_generic_expr (symtab->dump_file, nonodr);
|
|
fprintf (symtab->dump_file, " mangled:%s\n",
|
|
IDENTIFIER_POINTER
|
|
(DECL_ASSEMBLER_NAME (TYPE_NAME (t))));
|
|
}
|
|
/* Set canonical for T and all other ODR equivalent duplicates
|
|
including incomplete structures. */
|
|
set_type_canonical_for_odr_type (t, nonodr);
|
|
}
|
|
else
|
|
{
|
|
tree prevail = prevailing_odr_type (t);
|
|
|
|
if (symtab->dump_file)
|
|
{
|
|
fprintf (symtab->dump_file,
|
|
"New canonical ODR type: ");
|
|
print_generic_expr (symtab->dump_file, t);
|
|
fprintf (symtab->dump_file, " mangled:%s\n",
|
|
IDENTIFIER_POINTER
|
|
(DECL_ASSEMBLER_NAME (TYPE_NAME (t))));
|
|
}
|
|
/* Set canonical for T and all other ODR equivalent duplicates
|
|
including incomplete structures. */
|
|
set_type_canonical_for_odr_type (t, prevail);
|
|
enable_odr_based_tbaa (t);
|
|
if (!type_in_anonymous_namespace_p (t))
|
|
hash = htab_hash_string (IDENTIFIER_POINTER
|
|
(DECL_ASSEMBLER_NAME
|
|
(TYPE_NAME (t))));
|
|
else
|
|
hash = TYPE_UID (t);
|
|
|
|
/* All variants of t now have TYPE_CANONICAL set to prevail.
|
|
Update canonical type hash cache accordingly. */
|
|
num_canonical_type_hash_entries++;
|
|
bool existed_p = canonical_type_hash_cache->put (prevail, hash);
|
|
gcc_checking_assert (!existed_p);
|
|
}
|
|
return hash;
|
|
}
|
|
|
|
slot = htab_find_slot_with_hash (gimple_canonical_types, t, hash, INSERT);
|
|
if (*slot)
|
|
{
|
|
tree new_type = (tree)(*slot);
|
|
gcc_checking_assert (new_type != t);
|
|
TYPE_CANONICAL (t) = new_type;
|
|
}
|
|
else
|
|
{
|
|
TYPE_CANONICAL (t) = t;
|
|
*slot = (void *) t;
|
|
/* Cache the just computed hash value. */
|
|
num_canonical_type_hash_entries++;
|
|
bool existed_p = canonical_type_hash_cache->put (t, hash);
|
|
gcc_assert (!existed_p);
|
|
}
|
|
return hash;
|
|
}
|
|
|
|
/* Register type T in the global type table gimple_types and set
|
|
TYPE_CANONICAL of T accordingly.
|
|
This is used by LTO to merge structurally equivalent types for
|
|
type-based aliasing purposes across different TUs and languages.
|
|
|
|
??? This merging does not exactly match how the tree.cc middle-end
|
|
functions will assign TYPE_CANONICAL when new types are created
|
|
during optimization (which at least happens for pointer and array
|
|
types). */
|
|
|
|
static void
|
|
gimple_register_canonical_type (tree t)
|
|
{
|
|
if (TYPE_CANONICAL (t) || !type_with_alias_set_p (t)
|
|
|| !canonical_type_used_p (t))
|
|
return;
|
|
|
|
/* Canonical types are same among all complete variants. */
|
|
if (TYPE_CANONICAL (TYPE_MAIN_VARIANT (t)))
|
|
TYPE_CANONICAL (t) = TYPE_CANONICAL (TYPE_MAIN_VARIANT (t));
|
|
else
|
|
{
|
|
hashval_t h = hash_canonical_type (TYPE_MAIN_VARIANT (t));
|
|
gimple_register_canonical_type_1 (TYPE_MAIN_VARIANT (t), h);
|
|
TYPE_CANONICAL (t) = TYPE_CANONICAL (TYPE_MAIN_VARIANT (t));
|
|
}
|
|
}
|
|
|
|
/* Re-compute TYPE_CANONICAL for NODE and related types. */
|
|
|
|
static void
|
|
lto_register_canonical_types (tree node, bool first_p)
|
|
{
|
|
if (!node
|
|
|| !TYPE_P (node))
|
|
return;
|
|
|
|
if (first_p)
|
|
TYPE_CANONICAL (node) = NULL_TREE;
|
|
|
|
if (POINTER_TYPE_P (node)
|
|
|| TREE_CODE (node) == COMPLEX_TYPE
|
|
|| TREE_CODE (node) == ARRAY_TYPE)
|
|
lto_register_canonical_types (TREE_TYPE (node), first_p);
|
|
|
|
if (!first_p)
|
|
gimple_register_canonical_type (node);
|
|
}
|
|
|
|
/* Finish canonical type calculation: after all units has been streamed in we
|
|
can check if given ODR type structurally conflicts with a non-ODR type. In
|
|
the first case we set type canonical according to the canonical type hash.
|
|
In the second case we use type names. */
|
|
|
|
static void
|
|
lto_register_canonical_types_for_odr_types ()
|
|
{
|
|
tree t;
|
|
unsigned int i;
|
|
|
|
if (!types_to_register)
|
|
return;
|
|
|
|
type_streaming_finished = true;
|
|
|
|
/* Be sure that no types derived from ODR types was
|
|
not inserted into the hash table. */
|
|
if (flag_checking)
|
|
FOR_EACH_VEC_ELT (*types_to_register, i, t)
|
|
gcc_assert (!TYPE_CANONICAL (t));
|
|
|
|
/* Register all remaining types. */
|
|
FOR_EACH_VEC_ELT (*types_to_register, i, t)
|
|
{
|
|
/* For pre-streamed types like va-arg it is possible that main variant
|
|
is !CXX_ODR_P while the variant (which is streamed) is.
|
|
Copy CXX_ODR_P to make type verifier happy. This is safe because
|
|
in canonical type calculation we only consider main variants.
|
|
However we can not change this flag before streaming is finished
|
|
to not affect tree merging. */
|
|
TYPE_CXX_ODR_P (t) = TYPE_CXX_ODR_P (TYPE_MAIN_VARIANT (t));
|
|
if (!TYPE_CANONICAL (t))
|
|
gimple_register_canonical_type (t);
|
|
}
|
|
}
|
|
|
|
|
|
/* Remember trees that contains references to declarations. */
|
|
vec <tree, va_gc> *tree_with_vars;
|
|
|
|
#define CHECK_VAR(tt) \
|
|
do \
|
|
{ \
|
|
if ((tt) && VAR_OR_FUNCTION_DECL_P (tt) \
|
|
&& (TREE_PUBLIC (tt) || DECL_EXTERNAL (tt))) \
|
|
return true; \
|
|
} while (0)
|
|
|
|
#define CHECK_NO_VAR(tt) \
|
|
gcc_checking_assert (!(tt) || !VAR_OR_FUNCTION_DECL_P (tt))
|
|
|
|
/* Check presence of pointers to decls in fields of a tree_typed T. */
|
|
|
|
static inline bool
|
|
mentions_vars_p_typed (tree t)
|
|
{
|
|
CHECK_NO_VAR (TREE_TYPE (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a tree_common T. */
|
|
|
|
static inline bool
|
|
mentions_vars_p_common (tree t)
|
|
{
|
|
if (mentions_vars_p_typed (t))
|
|
return true;
|
|
CHECK_NO_VAR (TREE_CHAIN (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a decl_minimal T. */
|
|
|
|
static inline bool
|
|
mentions_vars_p_decl_minimal (tree t)
|
|
{
|
|
if (mentions_vars_p_common (t))
|
|
return true;
|
|
CHECK_NO_VAR (DECL_NAME (t));
|
|
CHECK_VAR (DECL_CONTEXT (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a decl_common T. */
|
|
|
|
static inline bool
|
|
mentions_vars_p_decl_common (tree t)
|
|
{
|
|
if (mentions_vars_p_decl_minimal (t))
|
|
return true;
|
|
CHECK_VAR (DECL_SIZE (t));
|
|
CHECK_VAR (DECL_SIZE_UNIT (t));
|
|
CHECK_VAR (DECL_INITIAL (t));
|
|
CHECK_NO_VAR (DECL_ATTRIBUTES (t));
|
|
CHECK_VAR (DECL_ABSTRACT_ORIGIN (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a decl_with_vis T. */
|
|
|
|
static inline bool
|
|
mentions_vars_p_decl_with_vis (tree t)
|
|
{
|
|
if (mentions_vars_p_decl_common (t))
|
|
return true;
|
|
|
|
/* Accessor macro has side-effects, use field-name here. */
|
|
CHECK_NO_VAR (DECL_ASSEMBLER_NAME_RAW (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a decl_non_common T. */
|
|
|
|
static inline bool
|
|
mentions_vars_p_decl_non_common (tree t)
|
|
{
|
|
if (mentions_vars_p_decl_with_vis (t))
|
|
return true;
|
|
CHECK_NO_VAR (DECL_RESULT_FLD (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a decl_non_common T. */
|
|
|
|
static bool
|
|
mentions_vars_p_function (tree t)
|
|
{
|
|
if (mentions_vars_p_decl_non_common (t))
|
|
return true;
|
|
CHECK_NO_VAR (DECL_ARGUMENTS (t));
|
|
CHECK_NO_VAR (DECL_VINDEX (t));
|
|
CHECK_VAR (DECL_FUNCTION_PERSONALITY (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a field_decl T. */
|
|
|
|
static bool
|
|
mentions_vars_p_field_decl (tree t)
|
|
{
|
|
if (mentions_vars_p_decl_common (t))
|
|
return true;
|
|
CHECK_VAR (DECL_FIELD_OFFSET (t));
|
|
CHECK_NO_VAR (DECL_BIT_FIELD_TYPE (t));
|
|
CHECK_NO_VAR (DECL_QUALIFIER (t));
|
|
CHECK_NO_VAR (DECL_FIELD_BIT_OFFSET (t));
|
|
CHECK_NO_VAR (DECL_FCONTEXT (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a type T. */
|
|
|
|
static bool
|
|
mentions_vars_p_type (tree t)
|
|
{
|
|
if (mentions_vars_p_common (t))
|
|
return true;
|
|
CHECK_NO_VAR (TYPE_CACHED_VALUES (t));
|
|
CHECK_VAR (TYPE_SIZE (t));
|
|
CHECK_VAR (TYPE_SIZE_UNIT (t));
|
|
CHECK_NO_VAR (TYPE_ATTRIBUTES (t));
|
|
CHECK_NO_VAR (TYPE_NAME (t));
|
|
|
|
CHECK_VAR (TYPE_MIN_VALUE_RAW (t));
|
|
CHECK_VAR (TYPE_MAX_VALUE_RAW (t));
|
|
|
|
/* Accessor is for derived node types only. */
|
|
CHECK_NO_VAR (TYPE_LANG_SLOT_1 (t));
|
|
|
|
CHECK_VAR (TYPE_CONTEXT (t));
|
|
CHECK_NO_VAR (TYPE_CANONICAL (t));
|
|
CHECK_NO_VAR (TYPE_MAIN_VARIANT (t));
|
|
CHECK_NO_VAR (TYPE_NEXT_VARIANT (t));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a BINFO T. */
|
|
|
|
static bool
|
|
mentions_vars_p_binfo (tree t)
|
|
{
|
|
unsigned HOST_WIDE_INT i, n;
|
|
|
|
if (mentions_vars_p_common (t))
|
|
return true;
|
|
CHECK_VAR (BINFO_VTABLE (t));
|
|
CHECK_NO_VAR (BINFO_OFFSET (t));
|
|
CHECK_NO_VAR (BINFO_VIRTUALS (t));
|
|
CHECK_NO_VAR (BINFO_VPTR_FIELD (t));
|
|
n = vec_safe_length (BINFO_BASE_ACCESSES (t));
|
|
for (i = 0; i < n; i++)
|
|
CHECK_NO_VAR (BINFO_BASE_ACCESS (t, i));
|
|
/* Do not walk BINFO_INHERITANCE_CHAIN, BINFO_SUBVTT_INDEX
|
|
and BINFO_VPTR_INDEX; these are used by C++ FE only. */
|
|
n = BINFO_N_BASE_BINFOS (t);
|
|
for (i = 0; i < n; i++)
|
|
CHECK_NO_VAR (BINFO_BASE_BINFO (t, i));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of a CONSTRUCTOR T. */
|
|
|
|
static bool
|
|
mentions_vars_p_constructor (tree t)
|
|
{
|
|
unsigned HOST_WIDE_INT idx;
|
|
constructor_elt *ce;
|
|
|
|
if (mentions_vars_p_typed (t))
|
|
return true;
|
|
|
|
for (idx = 0; vec_safe_iterate (CONSTRUCTOR_ELTS (t), idx, &ce); idx++)
|
|
{
|
|
CHECK_NO_VAR (ce->index);
|
|
CHECK_VAR (ce->value);
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of an expression tree T. */
|
|
|
|
static bool
|
|
mentions_vars_p_expr (tree t)
|
|
{
|
|
int i;
|
|
if (mentions_vars_p_typed (t))
|
|
return true;
|
|
for (i = TREE_OPERAND_LENGTH (t) - 1; i >= 0; --i)
|
|
CHECK_VAR (TREE_OPERAND (t, i));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls in fields of an OMP_CLAUSE T. */
|
|
|
|
static bool
|
|
mentions_vars_p_omp_clause (tree t)
|
|
{
|
|
int i;
|
|
if (mentions_vars_p_common (t))
|
|
return true;
|
|
for (i = omp_clause_num_ops[OMP_CLAUSE_CODE (t)] - 1; i >= 0; --i)
|
|
CHECK_VAR (OMP_CLAUSE_OPERAND (t, i));
|
|
return false;
|
|
}
|
|
|
|
/* Check presence of pointers to decls that needs later fixup in T. */
|
|
|
|
static bool
|
|
mentions_vars_p (tree t)
|
|
{
|
|
switch (TREE_CODE (t))
|
|
{
|
|
case IDENTIFIER_NODE:
|
|
break;
|
|
|
|
case TREE_LIST:
|
|
CHECK_VAR (TREE_VALUE (t));
|
|
CHECK_VAR (TREE_PURPOSE (t));
|
|
CHECK_NO_VAR (TREE_CHAIN (t));
|
|
break;
|
|
|
|
case FIELD_DECL:
|
|
return mentions_vars_p_field_decl (t);
|
|
|
|
case LABEL_DECL:
|
|
case CONST_DECL:
|
|
case PARM_DECL:
|
|
case RESULT_DECL:
|
|
case IMPORTED_DECL:
|
|
case NAMESPACE_DECL:
|
|
case NAMELIST_DECL:
|
|
return mentions_vars_p_decl_common (t);
|
|
|
|
case VAR_DECL:
|
|
return mentions_vars_p_decl_with_vis (t);
|
|
|
|
case TYPE_DECL:
|
|
return mentions_vars_p_decl_non_common (t);
|
|
|
|
case FUNCTION_DECL:
|
|
return mentions_vars_p_function (t);
|
|
|
|
case TREE_BINFO:
|
|
return mentions_vars_p_binfo (t);
|
|
|
|
case PLACEHOLDER_EXPR:
|
|
return mentions_vars_p_common (t);
|
|
|
|
case BLOCK:
|
|
case TRANSLATION_UNIT_DECL:
|
|
case OPTIMIZATION_NODE:
|
|
case TARGET_OPTION_NODE:
|
|
break;
|
|
|
|
case CONSTRUCTOR:
|
|
return mentions_vars_p_constructor (t);
|
|
|
|
case OMP_CLAUSE:
|
|
return mentions_vars_p_omp_clause (t);
|
|
|
|
default:
|
|
if (TYPE_P (t))
|
|
{
|
|
if (mentions_vars_p_type (t))
|
|
return true;
|
|
}
|
|
else if (EXPR_P (t))
|
|
{
|
|
if (mentions_vars_p_expr (t))
|
|
return true;
|
|
}
|
|
else if (CONSTANT_CLASS_P (t))
|
|
CHECK_NO_VAR (TREE_TYPE (t));
|
|
else
|
|
gcc_unreachable ();
|
|
}
|
|
return false;
|
|
}
|
|
|
|
|
|
/* Return the resolution for the decl with index INDEX from DATA_IN. */
|
|
|
|
static enum ld_plugin_symbol_resolution
|
|
get_resolution (class data_in *data_in, unsigned index)
|
|
{
|
|
if (data_in->globals_resolution.exists ())
|
|
{
|
|
ld_plugin_symbol_resolution_t ret;
|
|
/* We can have references to not emitted functions in
|
|
DECL_FUNCTION_PERSONALITY at least. So we can and have
|
|
to indeed return LDPR_UNKNOWN in some cases. */
|
|
if (data_in->globals_resolution.length () <= index)
|
|
return LDPR_UNKNOWN;
|
|
ret = data_in->globals_resolution[index];
|
|
return ret;
|
|
}
|
|
else
|
|
/* Delay resolution finding until decl merging. */
|
|
return LDPR_UNKNOWN;
|
|
}
|
|
|
|
/* We need to record resolutions until symbol table is read. */
|
|
static void
|
|
register_resolution (struct lto_file_decl_data *file_data, tree decl,
|
|
enum ld_plugin_symbol_resolution resolution)
|
|
{
|
|
bool existed;
|
|
if (resolution == LDPR_UNKNOWN)
|
|
return;
|
|
if (!file_data->resolution_map)
|
|
file_data->resolution_map
|
|
= new hash_map<tree, ld_plugin_symbol_resolution>;
|
|
ld_plugin_symbol_resolution_t &res
|
|
= file_data->resolution_map->get_or_insert (decl, &existed);
|
|
if (!existed
|
|
|| resolution == LDPR_PREVAILING_DEF_IRONLY
|
|
|| resolution == LDPR_PREVAILING_DEF
|
|
|| resolution == LDPR_PREVAILING_DEF_IRONLY_EXP)
|
|
res = resolution;
|
|
}
|
|
|
|
/* Register DECL with the global symbol table and change its
|
|
name if necessary to avoid name clashes for static globals across
|
|
different files. */
|
|
|
|
static void
|
|
lto_register_var_decl_in_symtab (class data_in *data_in, tree decl,
|
|
unsigned ix)
|
|
{
|
|
tree context;
|
|
|
|
/* Variable has file scope, not local. */
|
|
if (!TREE_PUBLIC (decl)
|
|
&& !((context = decl_function_context (decl))
|
|
&& auto_var_in_fn_p (decl, context)))
|
|
rest_of_decl_compilation (decl, 1, 0);
|
|
|
|
/* If this variable has already been declared, queue the
|
|
declaration for merging. */
|
|
if (TREE_PUBLIC (decl))
|
|
register_resolution (data_in->file_data,
|
|
decl, get_resolution (data_in, ix));
|
|
}
|
|
|
|
|
|
/* Register DECL with the global symbol table and change its
|
|
name if necessary to avoid name clashes for static globals across
|
|
different files. DATA_IN contains descriptors and tables for the
|
|
file being read. */
|
|
|
|
static void
|
|
lto_register_function_decl_in_symtab (class data_in *data_in, tree decl,
|
|
unsigned ix)
|
|
{
|
|
/* If this variable has already been declared, queue the
|
|
declaration for merging. */
|
|
if (TREE_PUBLIC (decl) && !DECL_ABSTRACT_P (decl))
|
|
register_resolution (data_in->file_data,
|
|
decl, get_resolution (data_in, ix));
|
|
}
|
|
|
|
/* Check if T is a decl and needs register its resolution info. */
|
|
|
|
static void
|
|
lto_maybe_register_decl (class data_in *data_in, tree t, unsigned ix)
|
|
{
|
|
if (TREE_CODE (t) == VAR_DECL)
|
|
lto_register_var_decl_in_symtab (data_in, t, ix);
|
|
else if (TREE_CODE (t) == FUNCTION_DECL
|
|
&& !fndecl_built_in_p (t))
|
|
lto_register_function_decl_in_symtab (data_in, t, ix);
|
|
}
|
|
|
|
|
|
/* For the type T re-materialize it in the type variant list and
|
|
the pointer/reference-to chains. */
|
|
|
|
static void
|
|
lto_fixup_prevailing_type (tree t)
|
|
{
|
|
/* The following re-creates proper variant lists while fixing up
|
|
the variant leaders. We do not stream TYPE_NEXT_VARIANT so the
|
|
variant list state before fixup is broken. */
|
|
|
|
/* If we are not our own variant leader link us into our new leaders
|
|
variant list. */
|
|
if (TYPE_MAIN_VARIANT (t) != t)
|
|
{
|
|
tree mv = TYPE_MAIN_VARIANT (t);
|
|
TYPE_NEXT_VARIANT (t) = TYPE_NEXT_VARIANT (mv);
|
|
TYPE_NEXT_VARIANT (mv) = t;
|
|
}
|
|
|
|
/* The following reconstructs the pointer chains
|
|
of the new pointed-to type if we are a main variant. We do
|
|
not stream those so they are broken before fixup. */
|
|
if (TREE_CODE (t) == POINTER_TYPE
|
|
&& TYPE_MAIN_VARIANT (t) == t)
|
|
{
|
|
TYPE_NEXT_PTR_TO (t) = TYPE_POINTER_TO (TREE_TYPE (t));
|
|
TYPE_POINTER_TO (TREE_TYPE (t)) = t;
|
|
}
|
|
else if (TREE_CODE (t) == REFERENCE_TYPE
|
|
&& TYPE_MAIN_VARIANT (t) == t)
|
|
{
|
|
TYPE_NEXT_REF_TO (t) = TYPE_REFERENCE_TO (TREE_TYPE (t));
|
|
TYPE_REFERENCE_TO (TREE_TYPE (t)) = t;
|
|
}
|
|
}
|
|
|
|
|
|
/* We keep prevailing tree SCCs in a hashtable with manual collision
|
|
handling (in case all hashes compare the same) and keep the colliding
|
|
entries in the tree_scc->next chain. */
|
|
|
|
struct tree_scc
|
|
{
|
|
tree_scc *next;
|
|
/* Hash of the whole SCC. */
|
|
hashval_t hash;
|
|
/* Number of trees in the SCC. */
|
|
unsigned len;
|
|
/* Number of possible entries into the SCC (tree nodes [0..entry_len-1]
|
|
which share the same individual tree hash). */
|
|
unsigned entry_len;
|
|
/* The members of the SCC.
|
|
We only need to remember the first entry node candidate for prevailing
|
|
SCCs (but of course have access to all entries for SCCs we are
|
|
processing).
|
|
??? For prevailing SCCs we really only need hash and the first
|
|
entry candidate, but that's too awkward to implement. */
|
|
tree entries[1];
|
|
};
|
|
|
|
struct tree_scc_hasher : nofree_ptr_hash <tree_scc>
|
|
{
|
|
static inline hashval_t hash (const tree_scc *);
|
|
static inline bool equal (const tree_scc *, const tree_scc *);
|
|
};
|
|
|
|
hashval_t
|
|
tree_scc_hasher::hash (const tree_scc *scc)
|
|
{
|
|
return scc->hash;
|
|
}
|
|
|
|
bool
|
|
tree_scc_hasher::equal (const tree_scc *scc1, const tree_scc *scc2)
|
|
{
|
|
if (scc1->hash != scc2->hash
|
|
|| scc1->len != scc2->len
|
|
|| scc1->entry_len != scc2->entry_len)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
static hash_table<tree_scc_hasher> *tree_scc_hash;
|
|
static struct obstack tree_scc_hash_obstack;
|
|
|
|
static unsigned long num_merged_types;
|
|
static unsigned long num_prevailing_types;
|
|
static unsigned long num_type_scc_trees;
|
|
static unsigned long total_scc_size;
|
|
static unsigned long num_sccs_read;
|
|
static unsigned long num_unshared_trees_read;
|
|
static unsigned long total_scc_size_merged;
|
|
static unsigned long num_sccs_merged;
|
|
static unsigned long num_scc_compares;
|
|
static unsigned long num_scc_compare_collisions;
|
|
|
|
|
|
/* Compare the two entries T1 and T2 of two SCCs that are possibly equal,
|
|
recursing through in-SCC tree edges. Returns true if the SCCs entered
|
|
through T1 and T2 are equal and fills in *MAP with the pairs of
|
|
SCC entries we visited, starting with (*MAP)[0] = T1 and (*MAP)[1] = T2. */
|
|
|
|
static bool
|
|
compare_tree_sccs_1 (tree t1, tree t2, tree **map)
|
|
{
|
|
enum tree_code code;
|
|
|
|
/* Mark already visited nodes. */
|
|
TREE_ASM_WRITTEN (t2) = 1;
|
|
|
|
/* Push the pair onto map. */
|
|
(*map)[0] = t1;
|
|
(*map)[1] = t2;
|
|
*map = *map + 2;
|
|
|
|
/* Compare value-fields. */
|
|
#define compare_values(X) \
|
|
do { \
|
|
if (X(t1) != X(t2)) \
|
|
return false; \
|
|
} while (0)
|
|
|
|
compare_values (TREE_CODE);
|
|
code = TREE_CODE (t1);
|
|
|
|
/* If we end up comparing translation unit decls we either forgot to mark
|
|
some SCC as local or we compare too much. */
|
|
gcc_checking_assert (code != TRANSLATION_UNIT_DECL);
|
|
|
|
if (!TYPE_P (t1))
|
|
{
|
|
compare_values (TREE_SIDE_EFFECTS);
|
|
compare_values (TREE_CONSTANT);
|
|
compare_values (TREE_READONLY);
|
|
compare_values (TREE_PUBLIC);
|
|
}
|
|
compare_values (TREE_ADDRESSABLE);
|
|
compare_values (TREE_THIS_VOLATILE);
|
|
if (DECL_P (t1))
|
|
compare_values (DECL_UNSIGNED);
|
|
else if (TYPE_P (t1))
|
|
compare_values (TYPE_UNSIGNED);
|
|
if (TYPE_P (t1))
|
|
compare_values (TYPE_ARTIFICIAL);
|
|
else
|
|
compare_values (TREE_NO_WARNING);
|
|
compare_values (TREE_NOTHROW);
|
|
compare_values (TREE_STATIC);
|
|
if (code != TREE_BINFO)
|
|
compare_values (TREE_PRIVATE);
|
|
compare_values (TREE_PROTECTED);
|
|
compare_values (TREE_DEPRECATED);
|
|
if (TYPE_P (t1))
|
|
{
|
|
if (AGGREGATE_TYPE_P (t1))
|
|
compare_values (TYPE_REVERSE_STORAGE_ORDER);
|
|
else
|
|
compare_values (TYPE_SATURATING);
|
|
compare_values (TYPE_ADDR_SPACE);
|
|
}
|
|
else if (code == SSA_NAME)
|
|
compare_values (SSA_NAME_IS_DEFAULT_DEF);
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_INT_CST))
|
|
{
|
|
if (wi::to_wide (t1) != wi::to_wide (t2))
|
|
return false;
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_REAL_CST))
|
|
{
|
|
/* ??? No suitable compare routine available. */
|
|
REAL_VALUE_TYPE r1 = TREE_REAL_CST (t1);
|
|
REAL_VALUE_TYPE r2 = TREE_REAL_CST (t2);
|
|
if (r1.cl != r2.cl
|
|
|| r1.decimal != r2.decimal
|
|
|| r1.sign != r2.sign
|
|
|| r1.signalling != r2.signalling
|
|
|| r1.canonical != r2.canonical
|
|
|| r1.uexp != r2.uexp)
|
|
return false;
|
|
for (unsigned i = 0; i < SIGSZ; ++i)
|
|
if (r1.sig[i] != r2.sig[i])
|
|
return false;
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FIXED_CST))
|
|
if (!fixed_compare (EQ_EXPR,
|
|
TREE_FIXED_CST_PTR (t1), TREE_FIXED_CST_PTR (t2)))
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_VECTOR))
|
|
{
|
|
compare_values (VECTOR_CST_LOG2_NPATTERNS);
|
|
compare_values (VECTOR_CST_NELTS_PER_PATTERN);
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_COMMON))
|
|
{
|
|
compare_values (DECL_MODE);
|
|
compare_values (DECL_NONLOCAL);
|
|
compare_values (DECL_VIRTUAL_P);
|
|
compare_values (DECL_IGNORED_P);
|
|
compare_values (DECL_ABSTRACT_P);
|
|
compare_values (DECL_ARTIFICIAL);
|
|
compare_values (DECL_USER_ALIGN);
|
|
compare_values (DECL_PRESERVE_P);
|
|
compare_values (DECL_EXTERNAL);
|
|
compare_values (DECL_NOT_GIMPLE_REG_P);
|
|
compare_values (DECL_ALIGN);
|
|
if (code == LABEL_DECL)
|
|
{
|
|
compare_values (EH_LANDING_PAD_NR);
|
|
compare_values (LABEL_DECL_UID);
|
|
}
|
|
else if (code == FIELD_DECL)
|
|
{
|
|
compare_values (DECL_PACKED);
|
|
compare_values (DECL_NONADDRESSABLE_P);
|
|
compare_values (DECL_PADDING_P);
|
|
compare_values (DECL_FIELD_ABI_IGNORED);
|
|
compare_values (DECL_FIELD_CXX_ZERO_WIDTH_BIT_FIELD);
|
|
compare_values (DECL_OFFSET_ALIGN);
|
|
}
|
|
else if (code == VAR_DECL)
|
|
{
|
|
compare_values (DECL_HAS_DEBUG_EXPR_P);
|
|
compare_values (DECL_NONLOCAL_FRAME);
|
|
}
|
|
if (code == RESULT_DECL
|
|
|| code == PARM_DECL
|
|
|| code == VAR_DECL)
|
|
{
|
|
compare_values (DECL_BY_REFERENCE);
|
|
if (code == VAR_DECL
|
|
|| code == PARM_DECL)
|
|
compare_values (DECL_HAS_VALUE_EXPR_P);
|
|
}
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_WRTL))
|
|
compare_values (DECL_REGISTER);
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_WITH_VIS))
|
|
{
|
|
compare_values (DECL_COMMON);
|
|
compare_values (DECL_DLLIMPORT_P);
|
|
compare_values (DECL_WEAK);
|
|
compare_values (DECL_SEEN_IN_BIND_EXPR_P);
|
|
compare_values (DECL_COMDAT);
|
|
compare_values (DECL_VISIBILITY);
|
|
compare_values (DECL_VISIBILITY_SPECIFIED);
|
|
if (code == VAR_DECL)
|
|
{
|
|
compare_values (DECL_HARD_REGISTER);
|
|
/* DECL_IN_TEXT_SECTION is set during final asm output only. */
|
|
compare_values (DECL_IN_CONSTANT_POOL);
|
|
}
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FUNCTION_DECL))
|
|
{
|
|
compare_values (DECL_BUILT_IN_CLASS);
|
|
compare_values (DECL_STATIC_CONSTRUCTOR);
|
|
compare_values (DECL_STATIC_DESTRUCTOR);
|
|
compare_values (DECL_UNINLINABLE);
|
|
compare_values (DECL_POSSIBLY_INLINED);
|
|
compare_values (DECL_IS_NOVOPS);
|
|
compare_values (DECL_IS_RETURNS_TWICE);
|
|
compare_values (DECL_IS_MALLOC);
|
|
compare_values (FUNCTION_DECL_DECL_TYPE);
|
|
compare_values (DECL_DECLARED_INLINE_P);
|
|
compare_values (DECL_STATIC_CHAIN);
|
|
compare_values (DECL_NO_INLINE_WARNING_P);
|
|
compare_values (DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT);
|
|
compare_values (DECL_NO_LIMIT_STACK);
|
|
compare_values (DECL_DISREGARD_INLINE_LIMITS);
|
|
compare_values (DECL_PURE_P);
|
|
compare_values (DECL_LOOPING_CONST_OR_PURE_P);
|
|
compare_values (DECL_IS_REPLACEABLE_OPERATOR);
|
|
compare_values (DECL_FINAL_P);
|
|
compare_values (DECL_CXX_CONSTRUCTOR_P);
|
|
compare_values (DECL_CXX_DESTRUCTOR_P);
|
|
if (DECL_BUILT_IN_CLASS (t1) != NOT_BUILT_IN)
|
|
compare_values (DECL_UNCHECKED_FUNCTION_CODE);
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_TYPE_COMMON))
|
|
{
|
|
compare_values (TYPE_MODE);
|
|
compare_values (TYPE_NEEDS_CONSTRUCTING);
|
|
if (RECORD_OR_UNION_TYPE_P (t1))
|
|
{
|
|
compare_values (TYPE_TRANSPARENT_AGGR);
|
|
compare_values (TYPE_FINAL_P);
|
|
compare_values (TYPE_CXX_ODR_P);
|
|
}
|
|
else if (code == ARRAY_TYPE)
|
|
compare_values (TYPE_NONALIASED_COMPONENT);
|
|
if (code == ARRAY_TYPE || code == INTEGER_TYPE)
|
|
compare_values (TYPE_STRING_FLAG);
|
|
if (AGGREGATE_TYPE_P (t1))
|
|
compare_values (TYPE_TYPELESS_STORAGE);
|
|
compare_values (TYPE_EMPTY_P);
|
|
compare_values (TYPE_PACKED);
|
|
compare_values (TYPE_RESTRICT);
|
|
compare_values (TYPE_USER_ALIGN);
|
|
compare_values (TYPE_READONLY);
|
|
compare_values (TYPE_PRECISION);
|
|
compare_values (TYPE_ALIGN);
|
|
/* Do not compare TYPE_ALIAS_SET. Doing so introduce ordering issues
|
|
with calls to get_alias_set which may initialize it for streamed
|
|
in types. */
|
|
}
|
|
|
|
/* We don't want to compare locations, so there is nothing do compare
|
|
for TS_EXP. */
|
|
|
|
/* BLOCKs are function local and we don't merge anything there, so
|
|
simply refuse to merge. */
|
|
if (CODE_CONTAINS_STRUCT (code, TS_BLOCK))
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_TRANSLATION_UNIT_DECL))
|
|
if (strcmp (TRANSLATION_UNIT_LANGUAGE (t1),
|
|
TRANSLATION_UNIT_LANGUAGE (t2)) != 0)
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_TARGET_OPTION))
|
|
if (!cl_target_option_eq (TREE_TARGET_OPTION (t1), TREE_TARGET_OPTION (t2)))
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_OPTIMIZATION))
|
|
if (!cl_optimization_option_eq (TREE_OPTIMIZATION (t1),
|
|
TREE_OPTIMIZATION (t2)))
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_BINFO))
|
|
if (vec_safe_length (BINFO_BASE_ACCESSES (t1))
|
|
!= vec_safe_length (BINFO_BASE_ACCESSES (t2)))
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_CONSTRUCTOR))
|
|
{
|
|
compare_values (CLOBBER_KIND);
|
|
compare_values (CONSTRUCTOR_NELTS);
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_IDENTIFIER))
|
|
if (IDENTIFIER_LENGTH (t1) != IDENTIFIER_LENGTH (t2)
|
|
|| memcmp (IDENTIFIER_POINTER (t1), IDENTIFIER_POINTER (t2),
|
|
IDENTIFIER_LENGTH (t1)) != 0)
|
|
return false;
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_STRING))
|
|
if (TREE_STRING_LENGTH (t1) != TREE_STRING_LENGTH (t2)
|
|
|| memcmp (TREE_STRING_POINTER (t1), TREE_STRING_POINTER (t2),
|
|
TREE_STRING_LENGTH (t1)) != 0)
|
|
return false;
|
|
|
|
if (code == OMP_CLAUSE)
|
|
{
|
|
compare_values (OMP_CLAUSE_CODE);
|
|
switch (OMP_CLAUSE_CODE (t1))
|
|
{
|
|
case OMP_CLAUSE_DEFAULT:
|
|
compare_values (OMP_CLAUSE_DEFAULT_KIND);
|
|
break;
|
|
case OMP_CLAUSE_SCHEDULE:
|
|
compare_values (OMP_CLAUSE_SCHEDULE_KIND);
|
|
break;
|
|
case OMP_CLAUSE_DEPEND:
|
|
compare_values (OMP_CLAUSE_DEPEND_KIND);
|
|
break;
|
|
case OMP_CLAUSE_MAP:
|
|
compare_values (OMP_CLAUSE_MAP_KIND);
|
|
break;
|
|
case OMP_CLAUSE_PROC_BIND:
|
|
compare_values (OMP_CLAUSE_PROC_BIND_KIND);
|
|
break;
|
|
case OMP_CLAUSE_REDUCTION:
|
|
compare_values (OMP_CLAUSE_REDUCTION_CODE);
|
|
compare_values (OMP_CLAUSE_REDUCTION_GIMPLE_INIT);
|
|
compare_values (OMP_CLAUSE_REDUCTION_GIMPLE_MERGE);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
#undef compare_values
|
|
|
|
|
|
/* Compare pointer fields. */
|
|
|
|
/* Recurse. Search & Replaced from DFS_write_tree_body.
|
|
Folding the early checks into the compare_tree_edges recursion
|
|
macro makes debugging way quicker as you are able to break on
|
|
compare_tree_sccs_1 and simply finish until a call returns false
|
|
to spot the SCC members with the difference. */
|
|
#define compare_tree_edges(E1, E2) \
|
|
do { \
|
|
tree t1_ = (E1), t2_ = (E2); \
|
|
if (t1_ != t2_ \
|
|
&& (!t1_ || !t2_ \
|
|
|| !TREE_VISITED (t2_) \
|
|
|| (!TREE_ASM_WRITTEN (t2_) \
|
|
&& !compare_tree_sccs_1 (t1_, t2_, map)))) \
|
|
return false; \
|
|
/* Only non-NULL trees outside of the SCC may compare equal. */ \
|
|
gcc_checking_assert (t1_ != t2_ || (!t2_ || !TREE_VISITED (t2_))); \
|
|
} while (0)
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_TYPED))
|
|
{
|
|
if (code != IDENTIFIER_NODE)
|
|
compare_tree_edges (TREE_TYPE (t1), TREE_TYPE (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_VECTOR))
|
|
{
|
|
/* Note that the number of elements for EXPR has already been emitted
|
|
in EXPR's header (see streamer_write_tree_header). */
|
|
unsigned int count = vector_cst_encoded_nelts (t1);
|
|
for (unsigned int i = 0; i < count; ++i)
|
|
compare_tree_edges (VECTOR_CST_ENCODED_ELT (t1, i),
|
|
VECTOR_CST_ENCODED_ELT (t2, i));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_COMPLEX))
|
|
{
|
|
compare_tree_edges (TREE_REALPART (t1), TREE_REALPART (t2));
|
|
compare_tree_edges (TREE_IMAGPART (t1), TREE_IMAGPART (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_MINIMAL))
|
|
{
|
|
compare_tree_edges (DECL_NAME (t1), DECL_NAME (t2));
|
|
/* ??? Global decls from different TUs have non-matching
|
|
TRANSLATION_UNIT_DECLs. Only consider a small set of
|
|
decls equivalent, we should not end up merging others. */
|
|
if ((code == TYPE_DECL
|
|
|| code == NAMESPACE_DECL
|
|
|| code == IMPORTED_DECL
|
|
|| code == CONST_DECL
|
|
|| (VAR_OR_FUNCTION_DECL_P (t1)
|
|
&& (TREE_PUBLIC (t1) || DECL_EXTERNAL (t1))))
|
|
&& DECL_FILE_SCOPE_P (t1) && DECL_FILE_SCOPE_P (t2))
|
|
;
|
|
else
|
|
compare_tree_edges (DECL_CONTEXT (t1), DECL_CONTEXT (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_COMMON))
|
|
{
|
|
compare_tree_edges (DECL_SIZE (t1), DECL_SIZE (t2));
|
|
compare_tree_edges (DECL_SIZE_UNIT (t1), DECL_SIZE_UNIT (t2));
|
|
compare_tree_edges (DECL_ATTRIBUTES (t1), DECL_ATTRIBUTES (t2));
|
|
compare_tree_edges (DECL_ABSTRACT_ORIGIN (t1), DECL_ABSTRACT_ORIGIN (t2));
|
|
if ((code == VAR_DECL
|
|
|| code == PARM_DECL)
|
|
&& DECL_HAS_VALUE_EXPR_P (t1))
|
|
compare_tree_edges (DECL_VALUE_EXPR (t1), DECL_VALUE_EXPR (t2));
|
|
if (code == VAR_DECL
|
|
&& DECL_HAS_DEBUG_EXPR_P (t1))
|
|
compare_tree_edges (DECL_DEBUG_EXPR (t1), DECL_DEBUG_EXPR (t2));
|
|
/* LTO specific edges. */
|
|
if (code != FUNCTION_DECL
|
|
&& code != TRANSLATION_UNIT_DECL)
|
|
compare_tree_edges (DECL_INITIAL (t1), DECL_INITIAL (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_NON_COMMON))
|
|
{
|
|
if (code == FUNCTION_DECL)
|
|
{
|
|
tree a1, a2;
|
|
for (a1 = DECL_ARGUMENTS (t1), a2 = DECL_ARGUMENTS (t2);
|
|
a1 || a2;
|
|
a1 = TREE_CHAIN (a1), a2 = TREE_CHAIN (a2))
|
|
compare_tree_edges (a1, a2);
|
|
compare_tree_edges (DECL_RESULT (t1), DECL_RESULT (t2));
|
|
}
|
|
else if (code == TYPE_DECL)
|
|
compare_tree_edges (DECL_ORIGINAL_TYPE (t1), DECL_ORIGINAL_TYPE (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_WITH_VIS))
|
|
{
|
|
/* Make sure we don't inadvertently set the assembler name. */
|
|
if (DECL_ASSEMBLER_NAME_SET_P (t1))
|
|
compare_tree_edges (DECL_ASSEMBLER_NAME (t1),
|
|
DECL_ASSEMBLER_NAME (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FIELD_DECL))
|
|
{
|
|
compare_tree_edges (DECL_FIELD_OFFSET (t1), DECL_FIELD_OFFSET (t2));
|
|
compare_tree_edges (DECL_BIT_FIELD_TYPE (t1), DECL_BIT_FIELD_TYPE (t2));
|
|
compare_tree_edges (DECL_BIT_FIELD_REPRESENTATIVE (t1),
|
|
DECL_BIT_FIELD_REPRESENTATIVE (t2));
|
|
compare_tree_edges (DECL_FIELD_BIT_OFFSET (t1),
|
|
DECL_FIELD_BIT_OFFSET (t2));
|
|
compare_tree_edges (DECL_FCONTEXT (t1), DECL_FCONTEXT (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FUNCTION_DECL))
|
|
{
|
|
compare_tree_edges (DECL_FUNCTION_PERSONALITY (t1),
|
|
DECL_FUNCTION_PERSONALITY (t2));
|
|
compare_tree_edges (DECL_VINDEX (t1), DECL_VINDEX (t2));
|
|
compare_tree_edges (DECL_FUNCTION_SPECIFIC_TARGET (t1),
|
|
DECL_FUNCTION_SPECIFIC_TARGET (t2));
|
|
compare_tree_edges (DECL_FUNCTION_SPECIFIC_OPTIMIZATION (t1),
|
|
DECL_FUNCTION_SPECIFIC_OPTIMIZATION (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_TYPE_COMMON))
|
|
{
|
|
compare_tree_edges (TYPE_SIZE (t1), TYPE_SIZE (t2));
|
|
compare_tree_edges (TYPE_SIZE_UNIT (t1), TYPE_SIZE_UNIT (t2));
|
|
compare_tree_edges (TYPE_ATTRIBUTES (t1), TYPE_ATTRIBUTES (t2));
|
|
compare_tree_edges (TYPE_NAME (t1), TYPE_NAME (t2));
|
|
/* Do not compare TYPE_POINTER_TO or TYPE_REFERENCE_TO. They will be
|
|
reconstructed during fixup. */
|
|
/* Do not compare TYPE_NEXT_VARIANT, we reconstruct the variant lists
|
|
during fixup. */
|
|
compare_tree_edges (TYPE_MAIN_VARIANT (t1), TYPE_MAIN_VARIANT (t2));
|
|
/* ??? Global types from different TUs have non-matching
|
|
TRANSLATION_UNIT_DECLs. Still merge them if they are otherwise
|
|
equal. */
|
|
if (TYPE_FILE_SCOPE_P (t1) && TYPE_FILE_SCOPE_P (t2))
|
|
;
|
|
else
|
|
compare_tree_edges (TYPE_CONTEXT (t1), TYPE_CONTEXT (t2));
|
|
/* TYPE_CANONICAL is re-computed during type merging, so do not
|
|
compare it here. */
|
|
compare_tree_edges (TYPE_STUB_DECL (t1), TYPE_STUB_DECL (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_TYPE_NON_COMMON))
|
|
{
|
|
if (code == ARRAY_TYPE)
|
|
compare_tree_edges (TYPE_DOMAIN (t1), TYPE_DOMAIN (t2));
|
|
else if (RECORD_OR_UNION_TYPE_P (t1))
|
|
{
|
|
tree f1, f2;
|
|
for (f1 = TYPE_FIELDS (t1), f2 = TYPE_FIELDS (t2);
|
|
f1 || f2;
|
|
f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2))
|
|
compare_tree_edges (f1, f2);
|
|
}
|
|
else if (code == FUNCTION_TYPE
|
|
|| code == METHOD_TYPE)
|
|
compare_tree_edges (TYPE_ARG_TYPES (t1), TYPE_ARG_TYPES (t2));
|
|
|
|
if (!POINTER_TYPE_P (t1))
|
|
compare_tree_edges (TYPE_MIN_VALUE_RAW (t1), TYPE_MIN_VALUE_RAW (t2));
|
|
compare_tree_edges (TYPE_MAX_VALUE_RAW (t1), TYPE_MAX_VALUE_RAW (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_LIST))
|
|
{
|
|
compare_tree_edges (TREE_PURPOSE (t1), TREE_PURPOSE (t2));
|
|
compare_tree_edges (TREE_VALUE (t1), TREE_VALUE (t2));
|
|
compare_tree_edges (TREE_CHAIN (t1), TREE_CHAIN (t2));
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_VEC))
|
|
for (int i = 0; i < TREE_VEC_LENGTH (t1); i++)
|
|
compare_tree_edges (TREE_VEC_ELT (t1, i), TREE_VEC_ELT (t2, i));
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_EXP))
|
|
{
|
|
for (int i = 0; i < TREE_OPERAND_LENGTH (t1); i++)
|
|
compare_tree_edges (TREE_OPERAND (t1, i),
|
|
TREE_OPERAND (t2, i));
|
|
|
|
/* BLOCKs are function local and we don't merge anything there. */
|
|
if (TREE_BLOCK (t1) || TREE_BLOCK (t2))
|
|
return false;
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_BINFO))
|
|
{
|
|
unsigned i;
|
|
tree t;
|
|
/* Lengths have already been compared above. */
|
|
FOR_EACH_VEC_ELT (*BINFO_BASE_BINFOS (t1), i, t)
|
|
compare_tree_edges (t, BINFO_BASE_BINFO (t2, i));
|
|
FOR_EACH_VEC_SAFE_ELT (BINFO_BASE_ACCESSES (t1), i, t)
|
|
compare_tree_edges (t, BINFO_BASE_ACCESS (t2, i));
|
|
compare_tree_edges (BINFO_OFFSET (t1), BINFO_OFFSET (t2));
|
|
compare_tree_edges (BINFO_VTABLE (t1), BINFO_VTABLE (t2));
|
|
compare_tree_edges (BINFO_VPTR_FIELD (t1), BINFO_VPTR_FIELD (t2));
|
|
/* Do not walk BINFO_INHERITANCE_CHAIN, BINFO_SUBVTT_INDEX
|
|
and BINFO_VPTR_INDEX; these are used by C++ FE only. */
|
|
}
|
|
|
|
if (CODE_CONTAINS_STRUCT (code, TS_CONSTRUCTOR))
|
|
{
|
|
unsigned i;
|
|
tree index, value;
|
|
/* Lengths have already been compared above. */
|
|
FOR_EACH_CONSTRUCTOR_ELT (CONSTRUCTOR_ELTS (t1), i, index, value)
|
|
{
|
|
compare_tree_edges (index, CONSTRUCTOR_ELT (t2, i)->index);
|
|
compare_tree_edges (value, CONSTRUCTOR_ELT (t2, i)->value);
|
|
}
|
|
}
|
|
|
|
if (code == OMP_CLAUSE)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < omp_clause_num_ops[OMP_CLAUSE_CODE (t1)]; i++)
|
|
compare_tree_edges (OMP_CLAUSE_OPERAND (t1, i),
|
|
OMP_CLAUSE_OPERAND (t2, i));
|
|
compare_tree_edges (OMP_CLAUSE_CHAIN (t1), OMP_CLAUSE_CHAIN (t2));
|
|
}
|
|
|
|
#undef compare_tree_edges
|
|
|
|
return true;
|
|
}
|
|
|
|
/* Compare the tree scc SCC to the prevailing candidate PSCC, filling
|
|
out MAP if they are equal. */
|
|
|
|
static bool
|
|
compare_tree_sccs (tree_scc *pscc, tree_scc *scc,
|
|
tree *map)
|
|
{
|
|
/* Assume SCC entry hashes are sorted after their cardinality. Which
|
|
means we can simply take the first n-tuple of equal hashes
|
|
(which is recorded as entry_len) and do n SCC entry candidate
|
|
comparisons. */
|
|
for (unsigned i = 0; i < pscc->entry_len; ++i)
|
|
{
|
|
tree *mapp = map;
|
|
num_scc_compare_collisions++;
|
|
if (compare_tree_sccs_1 (pscc->entries[0], scc->entries[i], &mapp))
|
|
{
|
|
/* Equal - no need to reset TREE_VISITED or TREE_ASM_WRITTEN
|
|
on the scc as all trees will be freed. */
|
|
return true;
|
|
}
|
|
/* Reset TREE_ASM_WRITTEN on scc for the next compare or in case
|
|
the SCC prevails. */
|
|
for (unsigned j = 0; j < scc->len; ++j)
|
|
TREE_ASM_WRITTEN (scc->entries[j]) = 0;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/* QSort sort function to sort a map of two pointers after the 2nd
|
|
pointer. */
|
|
|
|
static int
|
|
cmp_tree (const void *p1_, const void *p2_)
|
|
{
|
|
tree *p1 = (tree *)(const_cast<void *>(p1_));
|
|
tree *p2 = (tree *)(const_cast<void *>(p2_));
|
|
if (p1[1] == p2[1])
|
|
return 0;
|
|
return ((uintptr_t)p1[1] < (uintptr_t)p2[1]) ? -1 : 1;
|
|
}
|
|
|
|
/* New scc of size 1 containing T was streamed in from DATA_IN and not merged.
|
|
Register it to reader cache at index FROM. */
|
|
|
|
static void
|
|
process_dref (class data_in *data_in, tree t, unsigned from)
|
|
{
|
|
struct streamer_tree_cache_d *cache = data_in->reader_cache;
|
|
/* If we got a debug reference queued, see if the prevailing
|
|
tree has a debug reference and if not, register the one
|
|
for the tree we are about to throw away. */
|
|
if (dref_queue.length () == 1)
|
|
{
|
|
dref_entry e = dref_queue.pop ();
|
|
gcc_assert (e.decl
|
|
== streamer_tree_cache_get_tree (cache, from));
|
|
const char *sym;
|
|
unsigned HOST_WIDE_INT off;
|
|
if (!debug_hooks->die_ref_for_decl (t, &sym, &off))
|
|
debug_hooks->register_external_die (t, e.sym, e.off);
|
|
}
|
|
}
|
|
|
|
/* Try to unify the SCC with nodes FROM to FROM + LEN in CACHE and
|
|
hash value SCC_HASH with an already recorded SCC. Return true if
|
|
that was successful, otherwise return false. */
|
|
|
|
static bool
|
|
unify_scc (class data_in *data_in, unsigned from,
|
|
unsigned len, unsigned scc_entry_len, hashval_t scc_hash)
|
|
{
|
|
bool unified_p = false;
|
|
struct streamer_tree_cache_d *cache = data_in->reader_cache;
|
|
tree_scc *scc
|
|
= (tree_scc *) alloca (sizeof (tree_scc) + (len - 1) * sizeof (tree));
|
|
scc->next = NULL;
|
|
scc->hash = scc_hash;
|
|
scc->len = len;
|
|
scc->entry_len = scc_entry_len;
|
|
for (unsigned i = 0; i < len; ++i)
|
|
{
|
|
tree t = streamer_tree_cache_get_tree (cache, from + i);
|
|
scc->entries[i] = t;
|
|
/* These types should be streamed as unshared. */
|
|
gcc_checking_assert
|
|
(!(TREE_CODE (t) == TRANSLATION_UNIT_DECL
|
|
|| (VAR_OR_FUNCTION_DECL_P (t)
|
|
&& !(TREE_PUBLIC (t) || DECL_EXTERNAL (t)))
|
|
|| TREE_CODE (t) == LABEL_DECL
|
|
|| (TREE_CODE (t) == NAMESPACE_DECL && !DECL_NAME (t))
|
|
|| (TYPE_P (t)
|
|
&& type_with_linkage_p (TYPE_MAIN_VARIANT (t))
|
|
&& type_in_anonymous_namespace_p (TYPE_MAIN_VARIANT (t)))));
|
|
}
|
|
|
|
/* Look for the list of candidate SCCs to compare against. */
|
|
tree_scc **slot;
|
|
slot = tree_scc_hash->find_slot_with_hash (scc, scc_hash, INSERT);
|
|
if (*slot)
|
|
{
|
|
/* Try unifying against each candidate. */
|
|
num_scc_compares++;
|
|
|
|
/* Set TREE_VISITED on the scc so we can easily identify tree nodes
|
|
outside of the scc when following tree edges. Make sure
|
|
that TREE_ASM_WRITTEN is unset so we can use it as 2nd bit
|
|
to track whether we visited the SCC member during the compare.
|
|
We cannot use TREE_VISITED on the pscc members as the extended
|
|
scc and pscc can overlap. */
|
|
for (unsigned i = 0; i < scc->len; ++i)
|
|
{
|
|
TREE_VISITED (scc->entries[i]) = 1;
|
|
gcc_checking_assert (!TREE_ASM_WRITTEN (scc->entries[i]));
|
|
}
|
|
|
|
tree *map = XALLOCAVEC (tree, 2 * len);
|
|
for (tree_scc *pscc = *slot; pscc; pscc = pscc->next)
|
|
{
|
|
if (!compare_tree_sccs (pscc, scc, map))
|
|
continue;
|
|
|
|
/* Found an equal SCC. */
|
|
unified_p = true;
|
|
num_scc_compare_collisions--;
|
|
num_sccs_merged++;
|
|
total_scc_size_merged += len;
|
|
|
|
if (flag_checking)
|
|
for (unsigned i = 0; i < len; ++i)
|
|
{
|
|
tree t = map[2*i+1];
|
|
enum tree_code code = TREE_CODE (t);
|
|
/* IDENTIFIER_NODEs should be singletons and are merged by the
|
|
streamer. The others should be singletons, too, and we
|
|
should not merge them in any way. */
|
|
gcc_assert (code != TRANSLATION_UNIT_DECL
|
|
&& code != IDENTIFIER_NODE);
|
|
}
|
|
|
|
/* Fixup the streamer cache with the prevailing nodes according
|
|
to the tree node mapping computed by compare_tree_sccs. */
|
|
if (len == 1)
|
|
{
|
|
process_dref (data_in, pscc->entries[0], from);
|
|
lto_maybe_register_decl (data_in, pscc->entries[0], from);
|
|
streamer_tree_cache_replace_tree (cache, pscc->entries[0], from);
|
|
}
|
|
else
|
|
{
|
|
tree *map2 = XALLOCAVEC (tree, 2 * len);
|
|
for (unsigned i = 0; i < len; ++i)
|
|
{
|
|
map2[i*2] = (tree)(uintptr_t)(from + i);
|
|
map2[i*2+1] = scc->entries[i];
|
|
}
|
|
qsort (map2, len, 2 * sizeof (tree), cmp_tree);
|
|
qsort (map, len, 2 * sizeof (tree), cmp_tree);
|
|
for (unsigned i = 0; i < len; ++i)
|
|
{
|
|
lto_maybe_register_decl (data_in, map[2*i],
|
|
(uintptr_t)map2[2*i]);
|
|
streamer_tree_cache_replace_tree (cache, map[2*i],
|
|
(uintptr_t)map2[2*i]);
|
|
}
|
|
}
|
|
|
|
/* Free the tree nodes from the read SCC. */
|
|
data_in->location_cache.revert_location_cache ();
|
|
for (unsigned i = 0; i < len; ++i)
|
|
{
|
|
if (TYPE_P (scc->entries[i]))
|
|
num_merged_types++;
|
|
free_node (scc->entries[i]);
|
|
}
|
|
|
|
/* Drop DIE references.
|
|
??? Do as in the size-one SCC case which involves sorting
|
|
the queue. */
|
|
dref_queue.truncate (0);
|
|
|
|
break;
|
|
}
|
|
|
|
/* Reset TREE_VISITED if we didn't unify the SCC with another. */
|
|
if (!unified_p)
|
|
for (unsigned i = 0; i < scc->len; ++i)
|
|
TREE_VISITED (scc->entries[i]) = 0;
|
|
}
|
|
|
|
/* If we didn't unify it to any candidate duplicate the relevant
|
|
pieces to permanent storage and link it into the chain. */
|
|
if (!unified_p)
|
|
{
|
|
tree_scc *pscc
|
|
= XOBNEWVAR (&tree_scc_hash_obstack, tree_scc, sizeof (tree_scc));
|
|
memcpy (pscc, scc, sizeof (tree_scc));
|
|
pscc->next = (*slot);
|
|
*slot = pscc;
|
|
}
|
|
return unified_p;
|
|
}
|
|
|
|
typedef int_hash<unsigned, 0, UINT_MAX> code_id_hash;
|
|
|
|
/* Do registering necessary once new tree fully streamed in (including all
|
|
trees it reffers to). */
|
|
|
|
static void
|
|
process_new_tree (tree t, hash_map <code_id_hash, unsigned> *hm,
|
|
unsigned index, unsigned *total, class data_in *data_in)
|
|
{
|
|
/* Reconstruct the type variant and pointer-to/reference-to
|
|
chains. */
|
|
if (TYPE_P (t))
|
|
{
|
|
/* Map the tree types to their frequencies. */
|
|
if (flag_lto_dump_type_stats)
|
|
{
|
|
unsigned key = (unsigned) TREE_CODE (t);
|
|
unsigned *countp = hm->get (key);
|
|
hm->put (key, countp ? (*countp) + 1 : 1);
|
|
(*total)++;
|
|
}
|
|
|
|
num_prevailing_types++;
|
|
lto_fixup_prevailing_type (t);
|
|
|
|
/* Compute the canonical type of all non-ODR types.
|
|
Delay ODR types for the end of merging process - the canonical
|
|
type for those can be computed using the (unique) name however
|
|
we want to do this only if units in other languages do not
|
|
contain structurally equivalent type.
|
|
|
|
Because SCC components are streamed in random (hash) order
|
|
we may have encountered the type before while registering
|
|
type canonical of a derived type in the same SCC. */
|
|
if (!TYPE_CANONICAL (t))
|
|
{
|
|
if (!RECORD_OR_UNION_TYPE_P (t)
|
|
|| !TYPE_CXX_ODR_P (t))
|
|
gimple_register_canonical_type (t);
|
|
else if (COMPLETE_TYPE_P (t))
|
|
vec_safe_push (types_to_register, t);
|
|
}
|
|
if (TYPE_MAIN_VARIANT (t) == t && odr_type_p (t))
|
|
register_odr_type (t);
|
|
}
|
|
/* Link shared INTEGER_CSTs into TYPE_CACHED_VALUEs of its
|
|
type which is also member of this SCC. */
|
|
if (TREE_CODE (t) == INTEGER_CST
|
|
&& !TREE_OVERFLOW (t))
|
|
cache_integer_cst (t);
|
|
if (!flag_ltrans)
|
|
{
|
|
lto_maybe_register_decl (data_in, t, index);
|
|
/* Scan the tree for references to global functions or
|
|
variables and record those for later fixup. */
|
|
if (mentions_vars_p (t))
|
|
vec_safe_push (tree_with_vars, t);
|
|
}
|
|
}
|
|
|
|
/* Read all the symbols from buffer DATA, using descriptors in DECL_DATA.
|
|
RESOLUTIONS is the set of symbols picked by the linker (read from the
|
|
resolution file when the linker plugin is being used). */
|
|
|
|
static void
|
|
lto_read_decls (struct lto_file_decl_data *decl_data, const void *data,
|
|
vec<ld_plugin_symbol_resolution_t> resolutions)
|
|
{
|
|
const struct lto_decl_header *header = (const struct lto_decl_header *) data;
|
|
const int decl_offset = sizeof (struct lto_decl_header);
|
|
const int main_offset = decl_offset + header->decl_state_size;
|
|
const int string_offset = main_offset + header->main_size;
|
|
class data_in *data_in;
|
|
unsigned int i;
|
|
const uint32_t *data_ptr, *data_end;
|
|
uint32_t num_decl_states;
|
|
|
|
lto_input_block ib_main ((const char *) data + main_offset,
|
|
header->main_size, decl_data->mode_table);
|
|
|
|
data_in = lto_data_in_create (decl_data, (const char *) data + string_offset,
|
|
header->string_size, resolutions);
|
|
|
|
/* We do not uniquify the pre-loaded cache entries, those are middle-end
|
|
internal types that should not be merged. */
|
|
|
|
hash_map <code_id_hash, unsigned> hm;
|
|
unsigned total = 0;
|
|
|
|
/* Read the global declarations and types. */
|
|
while (ib_main.p < ib_main.len)
|
|
{
|
|
tree t;
|
|
unsigned from = data_in->reader_cache->nodes.length ();
|
|
/* Read and uniquify SCCs as in the input stream. */
|
|
enum LTO_tags tag = streamer_read_record_start (&ib_main);
|
|
if (tag == LTO_tree_scc || tag == LTO_trees)
|
|
{
|
|
unsigned len_;
|
|
unsigned scc_entry_len;
|
|
|
|
/* Because we stream in SCC order we know that all unshared trees
|
|
are now fully streamed. Process them. */
|
|
hashval_t scc_hash = lto_input_scc (&ib_main, data_in, &len_,
|
|
&scc_entry_len,
|
|
tag == LTO_tree_scc);
|
|
unsigned len = data_in->reader_cache->nodes.length () - from;
|
|
gcc_assert (len == len_);
|
|
|
|
if (tag == LTO_tree_scc)
|
|
{
|
|
total_scc_size += len;
|
|
num_sccs_read++;
|
|
}
|
|
else
|
|
num_unshared_trees_read += len;
|
|
|
|
/* We have the special case of size-1 SCCs that are pre-merged
|
|
by means of identifier and string sharing for example.
|
|
??? Maybe we should avoid streaming those as SCCs. */
|
|
tree first = streamer_tree_cache_get_tree (data_in->reader_cache,
|
|
from);
|
|
/* Identifier and integers are shared specially, they should never
|
|
go by the tree merging path. */
|
|
gcc_checking_assert ((TREE_CODE (first) != IDENTIFIER_NODE
|
|
&& (TREE_CODE (first) != INTEGER_CST
|
|
|| TREE_OVERFLOW (first)))
|
|
|| len != 1);
|
|
|
|
/* Try to unify the SCC with already existing ones. */
|
|
if (!flag_ltrans && tag != LTO_trees
|
|
&& unify_scc (data_in, from,
|
|
len, scc_entry_len, scc_hash))
|
|
continue;
|
|
|
|
/* Tree merging failed, mark entries in location cache as
|
|
permanent. */
|
|
data_in->location_cache.accept_location_cache ();
|
|
|
|
bool seen_type = false;
|
|
for (unsigned i = 0; i < len; ++i)
|
|
{
|
|
tree t = streamer_tree_cache_get_tree (data_in->reader_cache,
|
|
from + i);
|
|
process_new_tree (t, &hm, from + i, &total, data_in);
|
|
if (TYPE_P (t))
|
|
seen_type = true;
|
|
}
|
|
|
|
/* Register DECLs with the debuginfo machinery. */
|
|
while (!dref_queue.is_empty ())
|
|
{
|
|
dref_entry e = dref_queue.pop ();
|
|
debug_hooks->register_external_die (e.decl, e.sym, e.off);
|
|
}
|
|
|
|
if (seen_type)
|
|
num_type_scc_trees += len;
|
|
}
|
|
else
|
|
{
|
|
t = lto_input_tree_1 (&ib_main, data_in, tag, 0);
|
|
gcc_assert (data_in->reader_cache->nodes.length () == from + 1);
|
|
num_unshared_trees_read++;
|
|
data_in->location_cache.accept_location_cache ();
|
|
process_dref (data_in, t, from);
|
|
if (TREE_CODE (t) == IDENTIFIER_NODE
|
|
|| (TREE_CODE (t) == INTEGER_CST
|
|
&& !TREE_OVERFLOW (t)))
|
|
;
|
|
else
|
|
{
|
|
lto_maybe_register_decl (data_in, t, from);
|
|
process_new_tree (t, &hm, from, &total, data_in);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Dump type statistics. */
|
|
if (flag_lto_dump_type_stats)
|
|
{
|
|
fprintf (stdout, " Type Frequency Percentage\n\n");
|
|
for (hash_map<code_id_hash, unsigned>::iterator itr = hm.begin ();
|
|
itr != hm.end ();
|
|
++itr)
|
|
{
|
|
std::pair<unsigned, unsigned> p = *itr;
|
|
enum tree_code code = (enum tree_code) p.first;
|
|
fprintf (stdout, "%14s %6d %12.2f\n", get_tree_code_name (code),
|
|
p.second, float (p.second)/total*100);
|
|
}
|
|
}
|
|
|
|
data_in->location_cache.apply_location_cache ();
|
|
|
|
/* Read in lto_in_decl_state objects. */
|
|
data_ptr = (const uint32_t *) ((const char*) data + decl_offset);
|
|
data_end
|
|
= (const uint32_t *) ((const char*) data_ptr + header->decl_state_size);
|
|
num_decl_states = *data_ptr++;
|
|
|
|
gcc_assert (num_decl_states > 0);
|
|
decl_data->global_decl_state = lto_new_in_decl_state ();
|
|
data_ptr = lto_read_in_decl_state (data_in, data_ptr,
|
|
decl_data->global_decl_state);
|
|
|
|
/* Read in per-function decl states and enter them in hash table. */
|
|
decl_data->function_decl_states
|
|
= hash_table<decl_state_hasher>::create_ggc (37);
|
|
|
|
for (i = 1; i < num_decl_states; i++)
|
|
{
|
|
struct lto_in_decl_state *state = lto_new_in_decl_state ();
|
|
|
|
data_ptr = lto_read_in_decl_state (data_in, data_ptr, state);
|
|
lto_in_decl_state **slot
|
|
= decl_data->function_decl_states->find_slot (state, INSERT);
|
|
gcc_assert (*slot == NULL);
|
|
*slot = state;
|
|
}
|
|
|
|
if (data_ptr != data_end)
|
|
internal_error ("bytecode stream: garbage at the end of symbols section");
|
|
|
|
/* Set the current decl state to be the global state. */
|
|
decl_data->current_decl_state = decl_data->global_decl_state;
|
|
|
|
lto_data_in_delete (data_in);
|
|
}
|
|
|
|
/* Custom version of strtoll, which is not portable. */
|
|
|
|
static int64_t
|
|
lto_parse_hex (const char *p)
|
|
{
|
|
int64_t ret = 0;
|
|
|
|
for (; *p != '\0'; ++p)
|
|
{
|
|
char c = *p;
|
|
unsigned char part;
|
|
ret <<= 4;
|
|
if (c >= '0' && c <= '9')
|
|
part = c - '0';
|
|
else if (c >= 'a' && c <= 'f')
|
|
part = c - 'a' + 10;
|
|
else if (c >= 'A' && c <= 'F')
|
|
part = c - 'A' + 10;
|
|
else
|
|
internal_error ("could not parse hex number");
|
|
ret |= part;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* Read resolution for file named FILE_NAME. The resolution is read from
|
|
RESOLUTION. */
|
|
|
|
static void
|
|
lto_resolution_read (splay_tree file_ids, FILE *resolution, lto_file *file)
|
|
{
|
|
/* We require that objects in the resolution file are in the same
|
|
order as the lto1 command line. */
|
|
unsigned int name_len;
|
|
char *obj_name;
|
|
unsigned int num_symbols;
|
|
unsigned int i;
|
|
struct lto_file_decl_data *file_data;
|
|
splay_tree_node nd = NULL;
|
|
|
|
if (!resolution)
|
|
return;
|
|
|
|
name_len = strlen (file->filename);
|
|
obj_name = XNEWVEC (char, name_len + 1);
|
|
fscanf (resolution, " "); /* Read white space. */
|
|
|
|
fread (obj_name, sizeof (char), name_len, resolution);
|
|
obj_name[name_len] = '\0';
|
|
if (filename_cmp (obj_name, file->filename) != 0)
|
|
internal_error ("unexpected file name %s in linker resolution file. "
|
|
"Expected %s", obj_name, file->filename);
|
|
if (file->offset != 0)
|
|
{
|
|
int t;
|
|
char offset_p[17];
|
|
int64_t offset;
|
|
t = fscanf (resolution, "@0x%16s", offset_p);
|
|
if (t != 1)
|
|
internal_error ("could not parse file offset");
|
|
offset = lto_parse_hex (offset_p);
|
|
if (offset != file->offset)
|
|
internal_error ("unexpected offset");
|
|
}
|
|
|
|
free (obj_name);
|
|
|
|
fscanf (resolution, "%u", &num_symbols);
|
|
|
|
for (i = 0; i < num_symbols; i++)
|
|
{
|
|
int t;
|
|
unsigned index;
|
|
unsigned HOST_WIDE_INT id;
|
|
char r_str[27];
|
|
enum ld_plugin_symbol_resolution r = (enum ld_plugin_symbol_resolution) 0;
|
|
unsigned int j;
|
|
unsigned int lto_resolution_str_len
|
|
= sizeof (lto_resolution_str) / sizeof (char *);
|
|
res_pair rp;
|
|
|
|
t = fscanf (resolution, "%u " HOST_WIDE_INT_PRINT_HEX_PURE
|
|
" %26s %*[^\n]\n", &index, &id, r_str);
|
|
if (t != 3)
|
|
internal_error ("invalid line in the resolution file");
|
|
|
|
for (j = 0; j < lto_resolution_str_len; j++)
|
|
{
|
|
if (strcmp (lto_resolution_str[j], r_str) == 0)
|
|
{
|
|
r = (enum ld_plugin_symbol_resolution) j;
|
|
break;
|
|
}
|
|
}
|
|
if (j == lto_resolution_str_len)
|
|
internal_error ("invalid resolution in the resolution file");
|
|
|
|
if (!(nd && lto_splay_tree_id_equal_p (nd->key, id)))
|
|
{
|
|
nd = lto_splay_tree_lookup (file_ids, id);
|
|
if (nd == NULL)
|
|
internal_error ("resolution sub id %wx not in object file", id);
|
|
}
|
|
|
|
file_data = (struct lto_file_decl_data *)nd->value;
|
|
/* The indexes are very sparse. To save memory save them in a compact
|
|
format that is only unpacked later when the subfile is processed. */
|
|
rp.res = r;
|
|
rp.index = index;
|
|
file_data->respairs.safe_push (rp);
|
|
if (file_data->max_index < index)
|
|
file_data->max_index = index;
|
|
}
|
|
}
|
|
|
|
/* List of file_decl_datas. */
|
|
struct file_data_list
|
|
{
|
|
struct lto_file_decl_data *first, *last;
|
|
};
|
|
|
|
/* Is the name for a id'ed LTO section? */
|
|
|
|
static int
|
|
lto_section_with_id (const char *name, unsigned HOST_WIDE_INT *id)
|
|
{
|
|
const char *s;
|
|
|
|
if (strncmp (name, section_name_prefix, strlen (section_name_prefix)))
|
|
return 0;
|
|
s = strrchr (name, '.');
|
|
if (!s)
|
|
return 0;
|
|
/* If the section is not suffixed with an ID return. */
|
|
if ((size_t)(s - name) == strlen (section_name_prefix))
|
|
return 0;
|
|
return sscanf (s, "." HOST_WIDE_INT_PRINT_HEX_PURE, id) == 1;
|
|
}
|
|
|
|
/* Create file_data of each sub file id. */
|
|
|
|
static int
|
|
create_subid_section_table (struct lto_section_slot *ls, splay_tree file_ids,
|
|
struct file_data_list *list)
|
|
{
|
|
struct lto_section_slot s_slot, *new_slot;
|
|
unsigned HOST_WIDE_INT id;
|
|
splay_tree_node nd;
|
|
void **hash_slot;
|
|
char *new_name;
|
|
struct lto_file_decl_data *file_data;
|
|
|
|
if (!lto_section_with_id (ls->name, &id))
|
|
return 1;
|
|
|
|
/* Find hash table of sub module id. */
|
|
nd = lto_splay_tree_lookup (file_ids, id);
|
|
if (nd != NULL)
|
|
{
|
|
file_data = (struct lto_file_decl_data *)nd->value;
|
|
}
|
|
else
|
|
{
|
|
file_data = ggc_alloc<lto_file_decl_data> ();
|
|
memset(file_data, 0, sizeof (struct lto_file_decl_data));
|
|
file_data->id = id;
|
|
file_data->section_hash_table = lto_obj_create_section_hash_table ();
|
|
lto_splay_tree_insert (file_ids, id, file_data);
|
|
|
|
/* Maintain list in linker order. */
|
|
if (!list->first)
|
|
list->first = file_data;
|
|
if (list->last)
|
|
list->last->next = file_data;
|
|
|
|
list->last = file_data;
|
|
}
|
|
|
|
/* Copy section into sub module hash table. */
|
|
new_name = XDUPVEC (char, ls->name, strlen (ls->name) + 1);
|
|
s_slot.name = new_name;
|
|
hash_slot = htab_find_slot (file_data->section_hash_table, &s_slot, INSERT);
|
|
gcc_assert (*hash_slot == NULL);
|
|
|
|
new_slot = XDUP (struct lto_section_slot, ls);
|
|
new_slot->name = new_name;
|
|
*hash_slot = new_slot;
|
|
return 1;
|
|
}
|
|
|
|
/* Read declarations and other initializations for a FILE_DATA. */
|
|
|
|
static void
|
|
lto_file_finalize (struct lto_file_decl_data *file_data, lto_file *file,
|
|
int order)
|
|
{
|
|
const char *data;
|
|
size_t len;
|
|
vec<ld_plugin_symbol_resolution_t>
|
|
resolutions = vNULL;
|
|
int i;
|
|
res_pair *rp;
|
|
|
|
/* Create vector for fast access of resolution. We do this lazily
|
|
to save memory. */
|
|
resolutions.safe_grow_cleared (file_data->max_index + 1, true);
|
|
for (i = 0; file_data->respairs.iterate (i, &rp); i++)
|
|
resolutions[rp->index] = rp->res;
|
|
file_data->respairs.release ();
|
|
|
|
file_data->renaming_hash_table = lto_create_renaming_table ();
|
|
file_data->file_name = file->filename;
|
|
file_data->order = order;
|
|
|
|
/* Read and verify LTO section. */
|
|
data = lto_get_summary_section_data (file_data, LTO_section_lto, &len);
|
|
if (data == NULL)
|
|
{
|
|
fatal_error (input_location, "bytecode stream in file %qs generated "
|
|
"with GCC compiler older than 10.0", file_data->file_name);
|
|
return;
|
|
}
|
|
|
|
memcpy (&file_data->lto_section_header, data, sizeof (lto_section));
|
|
lto_check_version (file_data->lto_section_header.major_version,
|
|
file_data->lto_section_header.minor_version,
|
|
file_data->file_name);
|
|
|
|
#ifdef ACCEL_COMPILER
|
|
lto_input_mode_table (file_data);
|
|
#else
|
|
file_data->mode_table = lto_mode_identity_table;
|
|
#endif
|
|
|
|
data = lto_get_summary_section_data (file_data, LTO_section_decls, &len);
|
|
if (data == NULL)
|
|
{
|
|
internal_error ("cannot read %<LTO_section_decls%> from %s",
|
|
file_data->file_name);
|
|
return;
|
|
}
|
|
/* Frees resolutions. */
|
|
lto_read_decls (file_data, data, resolutions);
|
|
lto_free_section_data (file_data, LTO_section_decls, NULL, data, len);
|
|
}
|
|
|
|
/* Finalize FILE_DATA in FILE and increase COUNT. */
|
|
|
|
static int
|
|
lto_create_files_from_ids (lto_file *file, struct lto_file_decl_data *file_data,
|
|
int *count, int order)
|
|
{
|
|
lto_file_finalize (file_data, file, order);
|
|
if (symtab->dump_file)
|
|
fprintf (symtab->dump_file,
|
|
"Creating file %s with sub id " HOST_WIDE_INT_PRINT_HEX "\n",
|
|
file_data->file_name, file_data->id);
|
|
(*count)++;
|
|
return 0;
|
|
}
|
|
|
|
/* Generate a TREE representation for all types and external decls
|
|
entities in FILE.
|
|
|
|
Read all of the globals out of the file. Then read the cgraph
|
|
and process the .o index into the cgraph nodes so that it can open
|
|
the .o file to load the functions and ipa information. */
|
|
|
|
static struct lto_file_decl_data *
|
|
lto_file_read (lto_file *file, FILE *resolution_file, int *count)
|
|
{
|
|
struct lto_file_decl_data *file_data = NULL;
|
|
splay_tree file_ids;
|
|
htab_t section_hash_table;
|
|
struct lto_section_slot *section;
|
|
struct file_data_list file_list;
|
|
struct lto_section_list section_list;
|
|
|
|
memset (§ion_list, 0, sizeof (struct lto_section_list));
|
|
section_hash_table = lto_obj_build_section_table (file, §ion_list);
|
|
|
|
/* Dump the details of LTO objects. */
|
|
if (flag_lto_dump_objects)
|
|
{
|
|
int i=0;
|
|
fprintf (stdout, "\n LTO Object Name: %s\n", file->filename);
|
|
fprintf (stdout, "\nNo. Offset Size Section Name\n\n");
|
|
for (section = section_list.first; section != NULL; section = section->next)
|
|
fprintf (stdout, "%2d %8" PRId64 " %8" PRIu64 " %s\n",
|
|
++i, (int64_t) section->start, (uint64_t) section->len,
|
|
section->name);
|
|
}
|
|
|
|
/* Find all sub modules in the object and put their sections into new hash
|
|
tables in a splay tree. */
|
|
file_ids = lto_splay_tree_new ();
|
|
memset (&file_list, 0, sizeof (struct file_data_list));
|
|
for (section = section_list.first; section != NULL; section = section->next)
|
|
create_subid_section_table (section, file_ids, &file_list);
|
|
|
|
/* Add resolutions to file ids. */
|
|
lto_resolution_read (file_ids, resolution_file, file);
|
|
|
|
/* Finalize each lto file for each submodule in the merged object. */
|
|
int order = 0;
|
|
for (file_data = file_list.first; file_data != NULL;
|
|
file_data = file_data->next)
|
|
lto_create_files_from_ids (file, file_data, count, order++);
|
|
|
|
splay_tree_delete (file_ids);
|
|
htab_delete (section_hash_table);
|
|
|
|
return file_list.first;
|
|
}
|
|
|
|
#if HAVE_MMAP_FILE && HAVE_SYSCONF && defined _SC_PAGE_SIZE
|
|
#define LTO_MMAP_IO 1
|
|
#endif
|
|
|
|
#if LTO_MMAP_IO
|
|
/* Page size of machine is used for mmap and munmap calls. */
|
|
static size_t page_mask;
|
|
#endif
|
|
|
|
/* Get the section data of length LEN from FILENAME starting at
|
|
OFFSET. The data segment must be freed by the caller when the
|
|
caller is finished. Returns NULL if all was not well. */
|
|
|
|
static char *
|
|
lto_read_section_data (struct lto_file_decl_data *file_data,
|
|
intptr_t offset, size_t len)
|
|
{
|
|
char *result;
|
|
static int fd = -1;
|
|
static char *fd_name;
|
|
#if LTO_MMAP_IO
|
|
intptr_t computed_len;
|
|
intptr_t computed_offset;
|
|
intptr_t diff;
|
|
#endif
|
|
|
|
/* Keep a single-entry file-descriptor cache. The last file we
|
|
touched will get closed at exit.
|
|
??? Eventually we want to add a more sophisticated larger cache
|
|
or rather fix function body streaming to not stream them in
|
|
practically random order. */
|
|
if (fd != -1
|
|
&& filename_cmp (fd_name, file_data->file_name) != 0)
|
|
{
|
|
free (fd_name);
|
|
close (fd);
|
|
fd = -1;
|
|
}
|
|
if (fd == -1)
|
|
{
|
|
fd = open (file_data->file_name, O_RDONLY|O_BINARY);
|
|
if (fd == -1)
|
|
{
|
|
fatal_error (input_location, "Cannot open %s", file_data->file_name);
|
|
return NULL;
|
|
}
|
|
fd_name = xstrdup (file_data->file_name);
|
|
}
|
|
|
|
#if LTO_MMAP_IO
|
|
if (!page_mask)
|
|
{
|
|
size_t page_size = sysconf (_SC_PAGE_SIZE);
|
|
page_mask = ~(page_size - 1);
|
|
}
|
|
|
|
computed_offset = offset & page_mask;
|
|
diff = offset - computed_offset;
|
|
computed_len = len + diff;
|
|
|
|
result = (char *) mmap (NULL, computed_len, PROT_READ, MAP_PRIVATE,
|
|
fd, computed_offset);
|
|
if (result == MAP_FAILED)
|
|
{
|
|
fatal_error (input_location, "Cannot map %s", file_data->file_name);
|
|
return NULL;
|
|
}
|
|
|
|
return result + diff;
|
|
#else
|
|
result = (char *) xmalloc (len);
|
|
if (lseek (fd, offset, SEEK_SET) != offset
|
|
|| read (fd, result, len) != (ssize_t) len)
|
|
{
|
|
free (result);
|
|
fatal_error (input_location, "Cannot read %s", file_data->file_name);
|
|
result = NULL;
|
|
}
|
|
#ifdef __MINGW32__
|
|
/* Native windows doesn't supports delayed unlink on opened file. So
|
|
we close file here again. This produces higher I/O load, but at least
|
|
it prevents to have dangling file handles preventing unlink. */
|
|
free (fd_name);
|
|
fd_name = NULL;
|
|
close (fd);
|
|
fd = -1;
|
|
#endif
|
|
return result;
|
|
#endif
|
|
}
|
|
|
|
|
|
/* Get the section data from FILE_DATA of SECTION_TYPE with NAME.
|
|
NAME will be NULL unless the section type is for a function
|
|
body. */
|
|
|
|
static const char *
|
|
get_section_data (struct lto_file_decl_data *file_data,
|
|
enum lto_section_type section_type,
|
|
const char *name, int order,
|
|
size_t *len)
|
|
{
|
|
htab_t section_hash_table = file_data->section_hash_table;
|
|
struct lto_section_slot *f_slot;
|
|
struct lto_section_slot s_slot;
|
|
const char *section_name = lto_get_section_name (section_type, name,
|
|
order, file_data);
|
|
char *data = NULL;
|
|
|
|
*len = 0;
|
|
s_slot.name = section_name;
|
|
f_slot = (struct lto_section_slot *) htab_find (section_hash_table, &s_slot);
|
|
if (f_slot)
|
|
{
|
|
data = lto_read_section_data (file_data, f_slot->start, f_slot->len);
|
|
*len = f_slot->len;
|
|
}
|
|
|
|
free (CONST_CAST (char *, section_name));
|
|
return data;
|
|
}
|
|
|
|
|
|
/* Free the section data from FILE_DATA of SECTION_TYPE with NAME that
|
|
starts at OFFSET and has LEN bytes. */
|
|
|
|
static void
|
|
free_section_data (struct lto_file_decl_data *file_data ATTRIBUTE_UNUSED,
|
|
enum lto_section_type section_type ATTRIBUTE_UNUSED,
|
|
const char *name ATTRIBUTE_UNUSED,
|
|
const char *offset, size_t len ATTRIBUTE_UNUSED)
|
|
{
|
|
#if LTO_MMAP_IO
|
|
intptr_t computed_len;
|
|
intptr_t computed_offset;
|
|
intptr_t diff;
|
|
#endif
|
|
|
|
#if LTO_MMAP_IO
|
|
computed_offset = ((intptr_t) offset) & page_mask;
|
|
diff = (intptr_t) offset - computed_offset;
|
|
computed_len = len + diff;
|
|
|
|
munmap ((caddr_t) computed_offset, computed_len);
|
|
#else
|
|
free (CONST_CAST(char *, offset));
|
|
#endif
|
|
}
|
|
|
|
static lto_file *current_lto_file;
|
|
|
|
/* If TT is a variable or function decl replace it with its
|
|
prevailing variant. */
|
|
#define LTO_SET_PREVAIL(tt) \
|
|
do {\
|
|
if ((tt) && VAR_OR_FUNCTION_DECL_P (tt) \
|
|
&& (TREE_PUBLIC (tt) || DECL_EXTERNAL (tt))) \
|
|
{ \
|
|
tt = lto_symtab_prevailing_decl (tt); \
|
|
fixed = true; \
|
|
} \
|
|
} while (0)
|
|
|
|
/* Ensure that TT isn't a replacable var of function decl. */
|
|
#define LTO_NO_PREVAIL(tt) \
|
|
gcc_checking_assert (!(tt) || !VAR_OR_FUNCTION_DECL_P (tt))
|
|
|
|
/* Given a tree T replace all fields referring to variables or functions
|
|
with their prevailing variant. */
|
|
static void
|
|
lto_fixup_prevailing_decls (tree t)
|
|
{
|
|
enum tree_code code = TREE_CODE (t);
|
|
bool fixed = false;
|
|
|
|
gcc_checking_assert (code != TREE_BINFO);
|
|
LTO_NO_PREVAIL (TREE_TYPE (t));
|
|
if (CODE_CONTAINS_STRUCT (code, TS_COMMON)
|
|
/* lto_symtab_prevail_decl use TREE_CHAIN to link to the prevailing decl.
|
|
in the case T is a prevailed declaration we would ICE here. */
|
|
&& !VAR_OR_FUNCTION_DECL_P (t))
|
|
LTO_NO_PREVAIL (TREE_CHAIN (t));
|
|
if (DECL_P (t))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_NAME (t));
|
|
LTO_SET_PREVAIL (DECL_CONTEXT (t));
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_COMMON))
|
|
{
|
|
LTO_SET_PREVAIL (DECL_SIZE (t));
|
|
LTO_SET_PREVAIL (DECL_SIZE_UNIT (t));
|
|
LTO_SET_PREVAIL (DECL_INITIAL (t));
|
|
LTO_NO_PREVAIL (DECL_ATTRIBUTES (t));
|
|
LTO_SET_PREVAIL (DECL_ABSTRACT_ORIGIN (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_WITH_VIS))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_ASSEMBLER_NAME_RAW (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_NON_COMMON))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_RESULT_FLD (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FUNCTION_DECL))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_ARGUMENTS (t));
|
|
LTO_SET_PREVAIL (DECL_FUNCTION_PERSONALITY (t));
|
|
LTO_NO_PREVAIL (DECL_VINDEX (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FIELD_DECL))
|
|
{
|
|
LTO_SET_PREVAIL (DECL_FIELD_OFFSET (t));
|
|
LTO_NO_PREVAIL (DECL_BIT_FIELD_TYPE (t));
|
|
LTO_NO_PREVAIL (DECL_QUALIFIER (t));
|
|
LTO_NO_PREVAIL (DECL_FIELD_BIT_OFFSET (t));
|
|
LTO_NO_PREVAIL (DECL_FCONTEXT (t));
|
|
}
|
|
}
|
|
else if (TYPE_P (t))
|
|
{
|
|
LTO_NO_PREVAIL (TYPE_CACHED_VALUES (t));
|
|
LTO_SET_PREVAIL (TYPE_SIZE (t));
|
|
LTO_SET_PREVAIL (TYPE_SIZE_UNIT (t));
|
|
LTO_NO_PREVAIL (TYPE_ATTRIBUTES (t));
|
|
LTO_NO_PREVAIL (TYPE_NAME (t));
|
|
|
|
LTO_SET_PREVAIL (TYPE_MIN_VALUE_RAW (t));
|
|
LTO_SET_PREVAIL (TYPE_MAX_VALUE_RAW (t));
|
|
LTO_NO_PREVAIL (TYPE_LANG_SLOT_1 (t));
|
|
|
|
LTO_SET_PREVAIL (TYPE_CONTEXT (t));
|
|
|
|
LTO_NO_PREVAIL (TYPE_CANONICAL (t));
|
|
LTO_NO_PREVAIL (TYPE_MAIN_VARIANT (t));
|
|
LTO_NO_PREVAIL (TYPE_NEXT_VARIANT (t));
|
|
}
|
|
else if (EXPR_P (t))
|
|
{
|
|
int i;
|
|
for (i = TREE_OPERAND_LENGTH (t) - 1; i >= 0; --i)
|
|
LTO_SET_PREVAIL (TREE_OPERAND (t, i));
|
|
}
|
|
else if (TREE_CODE (t) == CONSTRUCTOR)
|
|
{
|
|
unsigned i;
|
|
tree val;
|
|
FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (t), i, val)
|
|
LTO_SET_PREVAIL (val);
|
|
}
|
|
else
|
|
{
|
|
switch (code)
|
|
{
|
|
case TREE_LIST:
|
|
LTO_SET_PREVAIL (TREE_VALUE (t));
|
|
LTO_SET_PREVAIL (TREE_PURPOSE (t));
|
|
break;
|
|
default:
|
|
gcc_unreachable ();
|
|
}
|
|
}
|
|
/* If we fixed nothing, then we missed something seen by
|
|
mentions_vars_p. */
|
|
gcc_checking_assert (fixed);
|
|
}
|
|
#undef LTO_SET_PREVAIL
|
|
#undef LTO_NO_PREVAIL
|
|
|
|
/* Helper function of lto_fixup_decls. Walks the var and fn streams in STATE,
|
|
replaces var and function decls with the corresponding prevailing def. */
|
|
|
|
static void
|
|
lto_fixup_state (struct lto_in_decl_state *state)
|
|
{
|
|
unsigned i, si;
|
|
|
|
/* Although we only want to replace FUNCTION_DECLs and VAR_DECLs,
|
|
we still need to walk from all DECLs to find the reachable
|
|
FUNCTION_DECLs and VAR_DECLs. */
|
|
for (si = 0; si < LTO_N_DECL_STREAMS; si++)
|
|
{
|
|
vec<tree, va_gc> *trees = state->streams[si];
|
|
for (i = 0; i < vec_safe_length (trees); i++)
|
|
{
|
|
tree t = (*trees)[i];
|
|
if (flag_checking && TYPE_P (t))
|
|
verify_type (t);
|
|
if (VAR_OR_FUNCTION_DECL_P (t)
|
|
&& (TREE_PUBLIC (t) || DECL_EXTERNAL (t)))
|
|
(*trees)[i] = lto_symtab_prevailing_decl (t);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Fix the decls from all FILES. Replaces each decl with the corresponding
|
|
prevailing one. */
|
|
|
|
static void
|
|
lto_fixup_decls (struct lto_file_decl_data **files)
|
|
{
|
|
unsigned int i;
|
|
tree t;
|
|
|
|
if (tree_with_vars)
|
|
FOR_EACH_VEC_ELT ((*tree_with_vars), i, t)
|
|
lto_fixup_prevailing_decls (t);
|
|
|
|
for (i = 0; files[i]; i++)
|
|
{
|
|
struct lto_file_decl_data *file = files[i];
|
|
struct lto_in_decl_state *state = file->global_decl_state;
|
|
lto_fixup_state (state);
|
|
|
|
hash_table<decl_state_hasher>::iterator iter;
|
|
lto_in_decl_state *elt;
|
|
FOR_EACH_HASH_TABLE_ELEMENT (*file->function_decl_states, elt,
|
|
lto_in_decl_state *, iter)
|
|
lto_fixup_state (elt);
|
|
}
|
|
}
|
|
|
|
static GTY((length ("lto_stats.num_input_files + 1"))) struct lto_file_decl_data **all_file_decl_data;
|
|
|
|
/* Turn file datas for sub files into a single array, so that they look
|
|
like separate files for further passes. */
|
|
|
|
static void
|
|
lto_flatten_files (struct lto_file_decl_data **orig, int count,
|
|
int last_file_ix)
|
|
{
|
|
struct lto_file_decl_data *n, *next;
|
|
int i, k;
|
|
|
|
lto_stats.num_input_files = count;
|
|
all_file_decl_data
|
|
= ggc_cleared_vec_alloc<lto_file_decl_data_ptr> (count + 1);
|
|
/* Set the hooks so that all of the ipa passes can read in their data. */
|
|
lto_set_in_hooks (all_file_decl_data, get_section_data, free_section_data);
|
|
for (i = 0, k = 0; i < last_file_ix; i++)
|
|
{
|
|
for (n = orig[i]; n != NULL; n = next)
|
|
{
|
|
all_file_decl_data[k++] = n;
|
|
next = n->next;
|
|
n->next = NULL;
|
|
}
|
|
}
|
|
all_file_decl_data[k] = NULL;
|
|
gcc_assert (k == count);
|
|
}
|
|
|
|
/* Input file data before flattening (i.e. splitting them to subfiles to support
|
|
incremental linking. */
|
|
static int real_file_count;
|
|
static GTY((length ("real_file_count + 1"))) struct lto_file_decl_data **real_file_decl_data;
|
|
|
|
/* Read all the symbols from the input files FNAMES. NFILES is the
|
|
number of files requested in the command line. Instantiate a
|
|
global call graph by aggregating all the sub-graphs found in each
|
|
file. */
|
|
|
|
void
|
|
read_cgraph_and_symbols (unsigned nfiles, const char **fnames)
|
|
{
|
|
unsigned int i, last_file_ix;
|
|
FILE *resolution;
|
|
unsigned resolution_objects = 0;
|
|
int count = 0;
|
|
struct lto_file_decl_data **decl_data;
|
|
symtab_node *snode;
|
|
|
|
symtab->initialize ();
|
|
|
|
timevar_push (TV_IPA_LTO_DECL_IN);
|
|
|
|
#ifdef ACCEL_COMPILER
|
|
section_name_prefix = OFFLOAD_SECTION_NAME_PREFIX;
|
|
lto_stream_offload_p = true;
|
|
#endif
|
|
|
|
real_file_decl_data
|
|
= decl_data = ggc_cleared_vec_alloc<lto_file_decl_data_ptr> (nfiles + 1);
|
|
real_file_count = nfiles;
|
|
|
|
/* Read the resolution file. */
|
|
resolution = NULL;
|
|
if (resolution_file_name)
|
|
{
|
|
int t;
|
|
|
|
resolution = fopen (resolution_file_name, "r");
|
|
if (resolution == NULL)
|
|
fatal_error (input_location,
|
|
"could not open symbol resolution file: %m");
|
|
|
|
t = fscanf (resolution, "%u", &resolution_objects);
|
|
gcc_assert (t == 1);
|
|
}
|
|
symtab->state = LTO_STREAMING;
|
|
|
|
canonical_type_hash_cache = new hash_map<const_tree, hashval_t> (251);
|
|
gimple_canonical_types = htab_create (16381, gimple_canonical_type_hash,
|
|
gimple_canonical_type_eq, NULL);
|
|
gcc_obstack_init (&tree_scc_hash_obstack);
|
|
tree_scc_hash = new hash_table<tree_scc_hasher> (4096);
|
|
|
|
/* Register the common node types with the canonical type machinery so
|
|
we properly share alias-sets across languages and TUs. Do not
|
|
expose the common nodes as type merge target - those that should be
|
|
are already exposed so by pre-loading the LTO streamer caches.
|
|
Do two passes - first clear TYPE_CANONICAL and then re-compute it. */
|
|
for (i = 0; i < itk_none; ++i)
|
|
lto_register_canonical_types (integer_types[i], true);
|
|
for (i = 0; i < stk_type_kind_last; ++i)
|
|
lto_register_canonical_types (sizetype_tab[i], true);
|
|
for (i = 0; i < TI_MAX; ++i)
|
|
lto_register_canonical_types (global_trees[i], true);
|
|
for (i = 0; i < itk_none; ++i)
|
|
lto_register_canonical_types (integer_types[i], false);
|
|
for (i = 0; i < stk_type_kind_last; ++i)
|
|
lto_register_canonical_types (sizetype_tab[i], false);
|
|
for (i = 0; i < TI_MAX; ++i)
|
|
lto_register_canonical_types (global_trees[i], false);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "Reading object files:");
|
|
|
|
/* Read all of the object files specified on the command line. */
|
|
for (i = 0, last_file_ix = 0; i < nfiles; ++i)
|
|
{
|
|
struct lto_file_decl_data *file_data = NULL;
|
|
if (!quiet_flag)
|
|
{
|
|
fprintf (stderr, " %s", fnames[i]);
|
|
fflush (stderr);
|
|
}
|
|
|
|
current_lto_file = lto_obj_file_open (fnames[i], false);
|
|
if (!current_lto_file)
|
|
break;
|
|
|
|
file_data = lto_file_read (current_lto_file, resolution, &count);
|
|
if (!file_data)
|
|
{
|
|
lto_obj_file_close (current_lto_file);
|
|
free (current_lto_file);
|
|
current_lto_file = NULL;
|
|
break;
|
|
}
|
|
|
|
decl_data[last_file_ix++] = file_data;
|
|
|
|
lto_obj_file_close (current_lto_file);
|
|
free (current_lto_file);
|
|
current_lto_file = NULL;
|
|
}
|
|
|
|
lto_flatten_files (decl_data, count, last_file_ix);
|
|
lto_stats.num_input_files = count;
|
|
ggc_free(decl_data);
|
|
real_file_decl_data = NULL;
|
|
|
|
lto_register_canonical_types_for_odr_types ();
|
|
|
|
if (resolution_file_name)
|
|
{
|
|
/* True, since the plugin splits the archives. */
|
|
gcc_assert (resolution_objects == nfiles);
|
|
fclose (resolution);
|
|
}
|
|
|
|
/* Show the LTO report before launching LTRANS. */
|
|
if (flag_lto_report || (flag_wpa && flag_lto_report_wpa))
|
|
print_lto_report_1 ();
|
|
|
|
/* Free gimple type merging datastructures. */
|
|
delete tree_scc_hash;
|
|
tree_scc_hash = NULL;
|
|
obstack_free (&tree_scc_hash_obstack, NULL);
|
|
htab_delete (gimple_canonical_types);
|
|
gimple_canonical_types = NULL;
|
|
delete canonical_type_hash_cache;
|
|
canonical_type_hash_cache = NULL;
|
|
|
|
/* At this stage we know that majority of GGC memory is reachable.
|
|
Growing the limits prevents unnecesary invocation of GGC. */
|
|
ggc_grow ();
|
|
report_heap_memory_use ();
|
|
|
|
/* Set the hooks so that all of the ipa passes can read in their data. */
|
|
lto_set_in_hooks (all_file_decl_data, get_section_data, free_section_data);
|
|
|
|
timevar_pop (TV_IPA_LTO_DECL_IN);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\nReading the symbol table:");
|
|
|
|
timevar_push (TV_IPA_LTO_CGRAPH_IO);
|
|
/* Read the symtab. */
|
|
input_symtab ();
|
|
|
|
input_offload_tables (!flag_ltrans);
|
|
|
|
/* Store resolutions into the symbol table. */
|
|
|
|
FOR_EACH_SYMBOL (snode)
|
|
if (snode->externally_visible && snode->real_symbol_p ()
|
|
&& snode->lto_file_data && snode->lto_file_data->resolution_map
|
|
&& !(TREE_CODE (snode->decl) == FUNCTION_DECL
|
|
&& fndecl_built_in_p (snode->decl))
|
|
&& !(VAR_P (snode->decl) && DECL_HARD_REGISTER (snode->decl)))
|
|
{
|
|
ld_plugin_symbol_resolution_t *res;
|
|
|
|
res = snode->lto_file_data->resolution_map->get (snode->decl);
|
|
if (!res || *res == LDPR_UNKNOWN)
|
|
{
|
|
if (snode->output_to_lto_symbol_table_p ())
|
|
fatal_error (input_location, "missing resolution data for %s",
|
|
IDENTIFIER_POINTER
|
|
(DECL_ASSEMBLER_NAME (snode->decl)));
|
|
}
|
|
/* Symbol versions are always used externally, but linker does not
|
|
report that correctly.
|
|
This is binutils PR25924. */
|
|
else if (snode->symver && *res == LDPR_PREVAILING_DEF_IRONLY)
|
|
snode->resolution = LDPR_PREVAILING_DEF_IRONLY_EXP;
|
|
else
|
|
snode->resolution = *res;
|
|
}
|
|
for (i = 0; all_file_decl_data[i]; i++)
|
|
if (all_file_decl_data[i]->resolution_map)
|
|
{
|
|
delete all_file_decl_data[i]->resolution_map;
|
|
all_file_decl_data[i]->resolution_map = NULL;
|
|
}
|
|
|
|
timevar_pop (TV_IPA_LTO_CGRAPH_IO);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\nMerging declarations:");
|
|
|
|
timevar_push (TV_IPA_LTO_DECL_MERGE);
|
|
/* Merge global decls. In ltrans mode we read merged cgraph, we do not
|
|
need to care about resolving symbols again, we only need to replace
|
|
duplicated declarations read from the callgraph and from function
|
|
sections. */
|
|
if (!flag_ltrans)
|
|
{
|
|
lto_symtab_merge_decls ();
|
|
|
|
/* If there were errors during symbol merging bail out, we have no
|
|
good way to recover here. */
|
|
if (seen_error ())
|
|
fatal_error (input_location,
|
|
"errors during merging of translation units");
|
|
|
|
/* Fixup all decls. */
|
|
lto_fixup_decls (all_file_decl_data);
|
|
}
|
|
if (tree_with_vars)
|
|
ggc_free (tree_with_vars);
|
|
tree_with_vars = NULL;
|
|
/* During WPA we want to prevent ggc collecting by default. Grow limits
|
|
until after the IPA summaries are streamed in. Basically all IPA memory
|
|
is explcitly managed by ggc_free and ggc collect is not useful.
|
|
Exception are the merged declarations. */
|
|
ggc_grow ();
|
|
report_heap_memory_use ();
|
|
|
|
timevar_pop (TV_IPA_LTO_DECL_MERGE);
|
|
/* Each pass will set the appropriate timer. */
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\nReading summaries:");
|
|
|
|
/* Read the IPA summary data. */
|
|
if (flag_ltrans)
|
|
ipa_read_optimization_summaries ();
|
|
else
|
|
ipa_read_summaries ();
|
|
|
|
ggc_grow ();
|
|
|
|
for (i = 0; all_file_decl_data[i]; i++)
|
|
{
|
|
gcc_assert (all_file_decl_data[i]->symtab_node_encoder);
|
|
lto_symtab_encoder_delete (all_file_decl_data[i]->symtab_node_encoder);
|
|
all_file_decl_data[i]->symtab_node_encoder = NULL;
|
|
lto_in_decl_state *global_decl_state
|
|
= all_file_decl_data[i]->global_decl_state;
|
|
lto_free_function_in_decl_state (global_decl_state);
|
|
all_file_decl_data[i]->global_decl_state = NULL;
|
|
all_file_decl_data[i]->current_decl_state = NULL;
|
|
}
|
|
|
|
if (!flag_ltrans)
|
|
{
|
|
/* Finally merge the cgraph according to the decl merging decisions. */
|
|
timevar_push (TV_IPA_LTO_CGRAPH_MERGE);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\nMerging symbols:");
|
|
|
|
gcc_assert (!dump_file);
|
|
dump_file = dump_begin (lto_link_dump_id, NULL);
|
|
|
|
if (dump_file)
|
|
{
|
|
fprintf (dump_file, "Before merging:\n");
|
|
symtab->dump (dump_file);
|
|
}
|
|
lto_symtab_merge_symbols ();
|
|
/* Removal of unreachable symbols is needed to make verify_symtab to pass;
|
|
we are still having duplicated comdat groups containing local statics.
|
|
We could also just remove them while merging. */
|
|
symtab->remove_unreachable_nodes (dump_file);
|
|
ggc_collect ();
|
|
report_heap_memory_use ();
|
|
|
|
if (dump_file)
|
|
dump_end (lto_link_dump_id, dump_file);
|
|
dump_file = NULL;
|
|
timevar_pop (TV_IPA_LTO_CGRAPH_MERGE);
|
|
}
|
|
symtab->state = IPA_SSA;
|
|
/* All node removals happening here are useless, because
|
|
WPA should not stream them. Still always perform remove_unreachable_nodes
|
|
because we may reshape clone tree, get rid of dead masters of inline
|
|
clones and remove symbol entries for read-only variables we keep around
|
|
only to be able to constant fold them. */
|
|
if (flag_ltrans)
|
|
{
|
|
if (symtab->dump_file)
|
|
symtab->dump (symtab->dump_file);
|
|
symtab->remove_unreachable_nodes (symtab->dump_file);
|
|
}
|
|
|
|
/* Indicate that the cgraph is built and ready. */
|
|
symtab->function_flags_ready = true;
|
|
|
|
ggc_free (all_file_decl_data);
|
|
all_file_decl_data = NULL;
|
|
}
|
|
|
|
|
|
|
|
/* Show various memory usage statistics related to LTO. */
|
|
void
|
|
print_lto_report_1 (void)
|
|
{
|
|
const char *pfx = (flag_lto) ? "LTO" : (flag_wpa) ? "WPA" : "LTRANS";
|
|
fprintf (stderr, "%s statistics\n", pfx);
|
|
|
|
fprintf (stderr, "[%s] read %lu unshared trees\n",
|
|
pfx, num_unshared_trees_read);
|
|
fprintf (stderr, "[%s] read %lu mergeable SCCs of average size %f\n",
|
|
pfx, num_sccs_read, total_scc_size / (double)num_sccs_read);
|
|
fprintf (stderr, "[%s] %lu tree bodies read in total\n", pfx,
|
|
total_scc_size + num_unshared_trees_read);
|
|
if (flag_wpa && tree_scc_hash && num_sccs_read)
|
|
{
|
|
fprintf (stderr, "[%s] tree SCC table: size %ld, %ld elements, "
|
|
"collision ratio: %f\n", pfx,
|
|
(long) tree_scc_hash->size (),
|
|
(long) tree_scc_hash->elements (),
|
|
tree_scc_hash->collisions ());
|
|
hash_table<tree_scc_hasher>::iterator hiter;
|
|
tree_scc *scc, *max_scc = NULL;
|
|
unsigned max_length = 0;
|
|
FOR_EACH_HASH_TABLE_ELEMENT (*tree_scc_hash, scc, x, hiter)
|
|
{
|
|
unsigned length = 0;
|
|
tree_scc *s = scc;
|
|
for (; s; s = s->next)
|
|
length++;
|
|
if (length > max_length)
|
|
{
|
|
max_length = length;
|
|
max_scc = scc;
|
|
}
|
|
}
|
|
fprintf (stderr, "[%s] tree SCC max chain length %u (size %u)\n",
|
|
pfx, max_length, max_scc->len);
|
|
fprintf (stderr, "[%s] Compared %lu SCCs, %lu collisions (%f)\n", pfx,
|
|
num_scc_compares, num_scc_compare_collisions,
|
|
num_scc_compare_collisions / (double) num_scc_compares);
|
|
fprintf (stderr, "[%s] Merged %lu SCCs\n", pfx, num_sccs_merged);
|
|
fprintf (stderr, "[%s] Merged %lu tree bodies\n", pfx,
|
|
total_scc_size_merged);
|
|
fprintf (stderr, "[%s] Merged %lu types\n", pfx, num_merged_types);
|
|
fprintf (stderr, "[%s] %lu types prevailed (%lu associated trees)\n",
|
|
pfx, num_prevailing_types, num_type_scc_trees);
|
|
fprintf (stderr, "[%s] GIMPLE canonical type table: size %ld, "
|
|
"%ld elements, %ld searches, %ld collisions (ratio: %f)\n", pfx,
|
|
(long) htab_size (gimple_canonical_types),
|
|
(long) htab_elements (gimple_canonical_types),
|
|
(long) gimple_canonical_types->searches,
|
|
(long) gimple_canonical_types->collisions,
|
|
htab_collisions (gimple_canonical_types));
|
|
fprintf (stderr, "[%s] GIMPLE canonical type pointer-map: "
|
|
"%lu elements, %ld searches\n", pfx,
|
|
num_canonical_type_hash_entries,
|
|
num_canonical_type_hash_queries);
|
|
}
|
|
|
|
print_lto_report (pfx);
|
|
}
|
|
|
|
GTY(()) tree lto_eh_personality_decl;
|
|
|
|
/* Return the LTO personality function decl. */
|
|
|
|
tree
|
|
lto_eh_personality (void)
|
|
{
|
|
if (!lto_eh_personality_decl)
|
|
{
|
|
/* Use the first personality DECL for our personality if we don't
|
|
support multiple ones. This ensures that we don't artificially
|
|
create the need for them in a single-language program. */
|
|
if (first_personality_decl && !dwarf2out_do_cfi_asm ())
|
|
lto_eh_personality_decl = first_personality_decl;
|
|
else
|
|
lto_eh_personality_decl = lhd_gcc_personality ();
|
|
}
|
|
|
|
return lto_eh_personality_decl;
|
|
}
|
|
|
|
/* Set the process name based on the LTO mode. */
|
|
|
|
static void
|
|
lto_process_name (void)
|
|
{
|
|
if (flag_lto)
|
|
setproctitle (flag_incremental_link == INCREMENTAL_LINK_LTO
|
|
? "lto1-inclink" : "lto1-lto");
|
|
if (flag_wpa)
|
|
setproctitle ("lto1-wpa");
|
|
if (flag_ltrans)
|
|
setproctitle ("lto1-ltrans");
|
|
}
|
|
|
|
|
|
/* Initialize the LTO front end. */
|
|
|
|
void
|
|
lto_fe_init (void)
|
|
{
|
|
lto_process_name ();
|
|
lto_streamer_hooks_init ();
|
|
lto_reader_init ();
|
|
lto_set_in_hooks (NULL, get_section_data, free_section_data);
|
|
memset (<o_stats, 0, sizeof (lto_stats));
|
|
bitmap_obstack_initialize (NULL);
|
|
gimple_register_cfg_hooks ();
|
|
#ifndef ACCEL_COMPILER
|
|
unsigned char *table
|
|
= ggc_vec_alloc<unsigned char> (MAX_MACHINE_MODE);
|
|
for (int m = 0; m < MAX_MACHINE_MODE; m++)
|
|
table[m] = m;
|
|
lto_mode_identity_table = table;
|
|
#endif
|
|
}
|
|
|
|
#include "gt-lto-lto-common.h"
|