Mercurial > hg > CbC > CbC_gcc
diff gcc/tree-streamer.c @ 111:04ced10e8804
gcc 7
author | kono |
---|---|
date | Fri, 27 Oct 2017 22:46:09 +0900 |
parents | |
children | 84e7813d76e9 |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/gcc/tree-streamer.c Fri Oct 27 22:46:09 2017 +0900 @@ -0,0 +1,430 @@ +/* Miscellaneous utilities for tree streaming. Things that are used + in both input and output are here. + + Copyright (C) 2011-2017 Free Software Foundation, Inc. + Contributed by Diego Novillo <dnovillo@google.com> + +This file is part of GCC. + +GCC is free software; you can redistribute it and/or modify it under +the terms of the GNU General Public License as published by the Free +Software Foundation; either version 3, or (at your option) any later +version. + +GCC is distributed in the hope that it will be useful, but WITHOUT ANY +WARRANTY; without even the implied warranty of MERCHANTABILITY or +FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License +for more details. + +You should have received a copy of the GNU General Public License +along with GCC; see the file COPYING3. If not see +<http://www.gnu.org/licenses/>. */ + +#include "config.h" +#include "system.h" +#include "coretypes.h" +#include "backend.h" +#include "tree.h" +#include "gimple.h" +#include "tree-streamer.h" +#include "cgraph.h" + +/* Table indexed by machine_mode, used for 2 different purposes. + During streaming out we record there non-zero value for all modes + that were streamed out. + During streaming in, we translate the on the disk mode using this + table. For normal LTO it is set to identity, for ACCEL_COMPILER + depending on the mode_table content. */ +unsigned char streamer_mode_table[1 << 8]; + +/* Check that all the TS_* structures handled by the streamer_write_* and + streamer_read_* routines are exactly ALL the structures defined in + treestruct.def. */ + +void +streamer_check_handled_ts_structures (void) +{ + bool handled_p[LAST_TS_ENUM]; + unsigned i; + + memset (&handled_p, 0, sizeof (handled_p)); + + /* These are the TS_* structures that are either handled or + explicitly ignored by the streamer routines. */ + handled_p[TS_BASE] = true; + handled_p[TS_TYPED] = true; + handled_p[TS_COMMON] = true; + handled_p[TS_INT_CST] = true; + handled_p[TS_REAL_CST] = true; + handled_p[TS_FIXED_CST] = true; + handled_p[TS_VECTOR] = true; + handled_p[TS_STRING] = true; + handled_p[TS_COMPLEX] = true; + handled_p[TS_IDENTIFIER] = true; + handled_p[TS_DECL_MINIMAL] = true; + handled_p[TS_DECL_COMMON] = true; + handled_p[TS_DECL_WRTL] = true; + handled_p[TS_DECL_NON_COMMON] = true; + handled_p[TS_DECL_WITH_VIS] = true; + handled_p[TS_FIELD_DECL] = true; + handled_p[TS_VAR_DECL] = true; + handled_p[TS_PARM_DECL] = true; + handled_p[TS_LABEL_DECL] = true; + handled_p[TS_RESULT_DECL] = true; + handled_p[TS_CONST_DECL] = true; + handled_p[TS_TYPE_DECL] = true; + handled_p[TS_FUNCTION_DECL] = true; + handled_p[TS_TYPE_COMMON] = true; + handled_p[TS_TYPE_WITH_LANG_SPECIFIC] = true; + handled_p[TS_TYPE_NON_COMMON] = true; + handled_p[TS_LIST] = true; + handled_p[TS_VEC] = true; + handled_p[TS_EXP] = true; + handled_p[TS_SSA_NAME] = true; + handled_p[TS_BLOCK] = true; + handled_p[TS_BINFO] = true; + handled_p[TS_STATEMENT_LIST] = true; + handled_p[TS_CONSTRUCTOR] = true; + handled_p[TS_OMP_CLAUSE] = true; + handled_p[TS_OPTIMIZATION] = true; + handled_p[TS_TARGET_OPTION] = true; + handled_p[TS_TRANSLATION_UNIT_DECL] = true; + + /* Anything not marked above will trigger the following assertion. + If this assertion triggers, it means that there is a new TS_* + structure that should be handled by the streamer. */ + for (i = 0; i < LAST_TS_ENUM; i++) + gcc_assert (handled_p[i]); +} + + +/* Helper for streamer_tree_cache_insert_1. Add T to CACHE->NODES at + slot IX. */ + +static void +streamer_tree_cache_add_to_node_array (struct streamer_tree_cache_d *cache, + unsigned ix, tree t, hashval_t hash) +{ + /* We're either replacing an old element or appending consecutively. */ + if (cache->nodes.exists ()) + { + if (cache->nodes.length () == ix) + cache->nodes.safe_push (t); + else + cache->nodes[ix] = t; + } + if (cache->hashes.exists ()) + { + if (cache->hashes.length () == ix) + cache->hashes.safe_push (hash); + else + cache->hashes[ix] = hash; + } +} + + +/* Helper for streamer_tree_cache_insert and streamer_tree_cache_insert_at. + CACHE, T, and IX_P are as in streamer_tree_cache_insert. + + If INSERT_AT_NEXT_SLOT_P is true, T is inserted at the next available + slot in the cache. Otherwise, T is inserted at the position indicated + in *IX_P. + + If T already existed in CACHE, return true. Otherwise, + return false. */ + +static bool +streamer_tree_cache_insert_1 (struct streamer_tree_cache_d *cache, + tree t, hashval_t hash, unsigned *ix_p, + bool insert_at_next_slot_p) +{ + bool existed_p; + + gcc_assert (t); + + unsigned int &ix = cache->node_map->get_or_insert (t, &existed_p); + if (!existed_p) + { + /* Determine the next slot to use in the cache. */ + if (insert_at_next_slot_p) + ix = cache->next_idx++; + else + ix = *ix_p; + + streamer_tree_cache_add_to_node_array (cache, ix, t, hash); + } + else + { + if (!insert_at_next_slot_p && ix != *ix_p) + { + /* If the caller wants to insert T at a specific slot + location, and ENTRY->TO does not match *IX_P, add T to + the requested location slot. */ + ix = *ix_p; + streamer_tree_cache_add_to_node_array (cache, ix, t, hash); + } + } + + if (ix_p) + *ix_p = ix; + + return existed_p; +} + + +/* Insert tree node T in CACHE. If T already existed in the cache + return true. Otherwise, return false. + + If IX_P is non-null, update it with the index into the cache where + T has been stored. */ + +bool +streamer_tree_cache_insert (struct streamer_tree_cache_d *cache, tree t, + hashval_t hash, unsigned *ix_p) +{ + return streamer_tree_cache_insert_1 (cache, t, hash, ix_p, true); +} + + +/* Replace the tree node with T in CACHE at slot IX. */ + +void +streamer_tree_cache_replace_tree (struct streamer_tree_cache_d *cache, + tree t, unsigned ix) +{ + hashval_t hash = 0; + if (cache->hashes.exists ()) + hash = streamer_tree_cache_get_hash (cache, ix); + if (!cache->node_map) + streamer_tree_cache_add_to_node_array (cache, ix, t, hash); + else + streamer_tree_cache_insert_1 (cache, t, hash, &ix, false); +} + + +/* Appends tree node T to CACHE, even if T already existed in it. */ + +void +streamer_tree_cache_append (struct streamer_tree_cache_d *cache, + tree t, hashval_t hash) +{ + unsigned ix = cache->next_idx++; + if (!cache->node_map) + streamer_tree_cache_add_to_node_array (cache, ix, t, hash); + else + streamer_tree_cache_insert_1 (cache, t, hash, &ix, false); +} + +/* Return true if tree node T exists in CACHE, otherwise false. If IX_P is + not NULL, write to *IX_P the index into the cache where T is stored + ((unsigned)-1 if T is not found). */ + +bool +streamer_tree_cache_lookup (struct streamer_tree_cache_d *cache, tree t, + unsigned *ix_p) +{ + unsigned *slot; + bool retval; + unsigned ix; + + gcc_assert (t); + + slot = cache->node_map->get (t); + if (slot == NULL) + { + retval = false; + ix = -1; + } + else + { + retval = true; + ix = *slot; + } + + if (ix_p) + *ix_p = ix; + + return retval; +} + + +/* Verify that NODE is in CACHE. */ + +static void +verify_common_node_recorded (struct streamer_tree_cache_d *cache, tree node) +{ + /* Restrict this to flag_checking only because in general violating it is + harmless plus we never know what happens on all targets/frontend/flag(!) + combinations. */ + if (!flag_checking) + return; + + if (cache->node_map) + gcc_assert (streamer_tree_cache_lookup (cache, node, NULL)); + else + { + bool found = false; + gcc_assert (cache->nodes.exists ()); + /* Linear search... */ + for (unsigned i = 0; !found && i < cache->nodes.length (); ++i) + if (cache->nodes[i] == node) + found = true; + gcc_assert (found); + } +} + + +/* Record NODE in CACHE. */ + +static void +record_common_node (struct streamer_tree_cache_d *cache, tree node) +{ + /* If we recursively end up at nodes we do not want to preload simply don't. + ??? We'd want to verify that this doesn't happen, or alternatively + do not recurse at all. */ + if (node == char_type_node) + return; + + gcc_checking_assert (node != boolean_type_node + && node != boolean_true_node + && node != boolean_false_node); + + /* We have to make sure to fill exactly the same number of + elements for all frontends. That can include NULL trees. + As our hash table can't deal with zero entries we'll simply stream + a random other tree. A NULL tree never will be looked up so it + doesn't matter which tree we replace it with, just to be sure + use error_mark_node. */ + if (!node) + node = error_mark_node; + + /* ??? FIXME, devise a better hash value. But the hash needs to be equal + for all frontend and lto1 invocations. So just use the position + in the cache as hash value. */ + streamer_tree_cache_append (cache, node, cache->nodes.length ()); + + switch (TREE_CODE (node)) + { + case ERROR_MARK: + case FIELD_DECL: + case FIXED_POINT_TYPE: + case IDENTIFIER_NODE: + case INTEGER_CST: + case INTEGER_TYPE: + case POINTER_BOUNDS_TYPE: + case REAL_TYPE: + case TREE_LIST: + case VOID_CST: + case VOID_TYPE: + /* No recursive trees. */ + break; + case ARRAY_TYPE: + case POINTER_TYPE: + case REFERENCE_TYPE: + record_common_node (cache, TREE_TYPE (node)); + break; + case COMPLEX_TYPE: + /* Verify that a complex type's component type (node_type) has been + handled already (and we thus don't need to recurse here). */ + verify_common_node_recorded (cache, TREE_TYPE (node)); + break; + case RECORD_TYPE: + /* The FIELD_DECLs of structures should be shared, so that every + COMPONENT_REF uses the same tree node when referencing a field. + Pointer equality between FIELD_DECLs is used by the alias + machinery to compute overlapping component references (see + nonoverlapping_component_refs_p and + nonoverlapping_component_refs_of_decl_p). */ + for (tree f = TYPE_FIELDS (node); f; f = TREE_CHAIN (f)) + record_common_node (cache, f); + break; + default: + /* Unexpected tree code. */ + gcc_unreachable (); + } +} + + +/* Preload common nodes into CACHE and make sure they are merged + properly according to the gimple type table. */ + +static void +preload_common_nodes (struct streamer_tree_cache_d *cache) +{ + unsigned i; + + for (i = 0; i < itk_none; i++) + /* Skip itk_char. char_type_node is dependent on -f[un]signed-char. */ + if (i != itk_char) + record_common_node (cache, integer_types[i]); + + for (i = 0; i < stk_type_kind_last; i++) + record_common_node (cache, sizetype_tab[i]); + + for (i = 0; i < TI_MAX; i++) + /* Skip boolean type and constants, they are frontend dependent. */ + if (i != TI_BOOLEAN_TYPE + && i != TI_BOOLEAN_FALSE + && i != TI_BOOLEAN_TRUE + /* MAIN_IDENTIFIER is not always initialized by Fortran FE. */ + && i != TI_MAIN_IDENTIFIER + /* PID_TYPE is initialized only by C family front-ends. */ + && i != TI_PID_TYPE + /* Skip optimization and target option nodes; they depend on flags. */ + && i != TI_OPTIMIZATION_DEFAULT + && i != TI_OPTIMIZATION_CURRENT + && i != TI_TARGET_OPTION_DEFAULT + && i != TI_TARGET_OPTION_CURRENT + && i != TI_CURRENT_TARGET_PRAGMA + && i != TI_CURRENT_OPTIMIZE_PRAGMA + /* Skip va_list* related nodes if offloading. For native LTO + we want them to be merged for the stdarg pass, for offloading + they might not be identical between host and offloading target. */ + && (!lto_stream_offload_p + || (i != TI_VA_LIST_TYPE + && i != TI_VA_LIST_GPR_COUNTER_FIELD + && i != TI_VA_LIST_FPR_COUNTER_FIELD))) + record_common_node (cache, global_trees[i]); +} + + +/* Create a cache of pickled nodes. */ + +struct streamer_tree_cache_d * +streamer_tree_cache_create (bool with_hashes, bool with_map, bool with_vec) +{ + struct streamer_tree_cache_d *cache; + + cache = XCNEW (struct streamer_tree_cache_d); + + if (with_map) + cache->node_map = new hash_map<tree, unsigned> (251); + cache->next_idx = 0; + if (with_vec) + cache->nodes.create (165); + if (with_hashes) + cache->hashes.create (165); + + /* Load all the well-known tree nodes that are always created by + the compiler on startup. This prevents writing them out + unnecessarily. */ + preload_common_nodes (cache); + + return cache; +} + + +/* Delete the streamer cache C. */ + +void +streamer_tree_cache_delete (struct streamer_tree_cache_d *c) +{ + if (c == NULL) + return; + + delete c->node_map; + c->node_map = NULL; + c->nodes.release (); + c->hashes.release (); + free (c); +}