/* Miscellaneous utilities for tree streaming. Things that are used in both input and output are here. Copyright (C) 2011-2018 Free Software Foundation, Inc. Contributed by Diego Novillo This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see . */ #include "config.h" #include "system.h" #include "coretypes.h" #include "backend.h" #include "tree.h" #include "gimple.h" #include "tree-streamer.h" #include "cgraph.h" /* Table indexed by machine_mode, used for 2 different purposes. During streaming out we record there non-zero value for all modes that were streamed out. During streaming in, we translate the on the disk mode using this table. For normal LTO it is set to identity, for ACCEL_COMPILER depending on the mode_table content. */ unsigned char streamer_mode_table[1 << 8]; /* Check that all the TS_* structures handled by the streamer_write_* and streamer_read_* routines are exactly ALL the structures defined in treestruct.def. */ void streamer_check_handled_ts_structures (void) { bool handled_p[LAST_TS_ENUM]; unsigned i; memset (&handled_p, 0, sizeof (handled_p)); /* These are the TS_* structures that are either handled or explicitly ignored by the streamer routines. */ handled_p[TS_BASE] = true; handled_p[TS_TYPED] = true; handled_p[TS_COMMON] = true; handled_p[TS_INT_CST] = true; handled_p[TS_POLY_INT_CST] = true; handled_p[TS_REAL_CST] = true; handled_p[TS_FIXED_CST] = true; handled_p[TS_VECTOR] = true; handled_p[TS_STRING] = true; handled_p[TS_COMPLEX] = true; handled_p[TS_IDENTIFIER] = true; handled_p[TS_DECL_MINIMAL] = true; handled_p[TS_DECL_COMMON] = true; handled_p[TS_DECL_WRTL] = true; handled_p[TS_DECL_NON_COMMON] = true; handled_p[TS_DECL_WITH_VIS] = true; handled_p[TS_FIELD_DECL] = true; handled_p[TS_VAR_DECL] = true; handled_p[TS_PARM_DECL] = true; handled_p[TS_LABEL_DECL] = true; handled_p[TS_RESULT_DECL] = true; handled_p[TS_CONST_DECL] = true; handled_p[TS_TYPE_DECL] = true; handled_p[TS_FUNCTION_DECL] = true; handled_p[TS_TYPE_COMMON] = true; handled_p[TS_TYPE_WITH_LANG_SPECIFIC] = true; handled_p[TS_TYPE_NON_COMMON] = true; handled_p[TS_LIST] = true; handled_p[TS_VEC] = true; handled_p[TS_EXP] = true; handled_p[TS_SSA_NAME] = true; handled_p[TS_BLOCK] = true; handled_p[TS_BINFO] = true; handled_p[TS_STATEMENT_LIST] = true; handled_p[TS_CONSTRUCTOR] = true; handled_p[TS_OMP_CLAUSE] = true; handled_p[TS_OPTIMIZATION] = true; handled_p[TS_TARGET_OPTION] = true; handled_p[TS_TRANSLATION_UNIT_DECL] = true; /* Anything not marked above will trigger the following assertion. If this assertion triggers, it means that there is a new TS_* structure that should be handled by the streamer. */ for (i = 0; i < LAST_TS_ENUM; i++) gcc_assert (handled_p[i]); } /* Helper for streamer_tree_cache_insert_1. Add T to CACHE->NODES at slot IX. */ static void streamer_tree_cache_add_to_node_array (struct streamer_tree_cache_d *cache, unsigned ix, tree t, hashval_t hash) { /* We're either replacing an old element or appending consecutively. */ if (cache->nodes.exists ()) { if (cache->nodes.length () == ix) cache->nodes.safe_push (t); else cache->nodes[ix] = t; } if (cache->hashes.exists ()) { if (cache->hashes.length () == ix) cache->hashes.safe_push (hash); else cache->hashes[ix] = hash; } } /* Helper for streamer_tree_cache_insert and streamer_tree_cache_insert_at. CACHE, T, and IX_P are as in streamer_tree_cache_insert. If INSERT_AT_NEXT_SLOT_P is true, T is inserted at the next available slot in the cache. Otherwise, T is inserted at the position indicated in *IX_P. If T already existed in CACHE, return true. Otherwise, return false. */ static bool streamer_tree_cache_insert_1 (struct streamer_tree_cache_d *cache, tree t, hashval_t hash, unsigned *ix_p, bool insert_at_next_slot_p) { bool existed_p; gcc_assert (t); unsigned int &ix = cache->node_map->get_or_insert (t, &existed_p); if (!existed_p) { /* Determine the next slot to use in the cache. */ if (insert_at_next_slot_p) ix = cache->next_idx++; else ix = *ix_p; streamer_tree_cache_add_to_node_array (cache, ix, t, hash); } else { if (!insert_at_next_slot_p && ix != *ix_p) { /* If the caller wants to insert T at a specific slot location, and ENTRY->TO does not match *IX_P, add T to the requested location slot. */ ix = *ix_p; streamer_tree_cache_add_to_node_array (cache, ix, t, hash); } } if (ix_p) *ix_p = ix; return existed_p; } /* Insert tree node T in CACHE. If T already existed in the cache return true. Otherwise, return false. If IX_P is non-null, update it with the index into the cache where T has been stored. */ bool streamer_tree_cache_insert (struct streamer_tree_cache_d *cache, tree t, hashval_t hash, unsigned *ix_p) { return streamer_tree_cache_insert_1 (cache, t, hash, ix_p, true); } /* Replace the tree node with T in CACHE at slot IX. */ void streamer_tree_cache_replace_tree (struct streamer_tree_cache_d *cache, tree t, unsigned ix) { hashval_t hash = 0; if (cache->hashes.exists ()) hash = streamer_tree_cache_get_hash (cache, ix); if (!cache->node_map) streamer_tree_cache_add_to_node_array (cache, ix, t, hash); else streamer_tree_cache_insert_1 (cache, t, hash, &ix, false); } /* Appends tree node T to CACHE, even if T already existed in it. */ void streamer_tree_cache_append (struct streamer_tree_cache_d *cache, tree t, hashval_t hash) { unsigned ix = cache->next_idx++; if (!cache->node_map) streamer_tree_cache_add_to_node_array (cache, ix, t, hash); else streamer_tree_cache_insert_1 (cache, t, hash, &ix, false); } /* Return true if tree node T exists in CACHE, otherwise false. If IX_P is not NULL, write to *IX_P the index into the cache where T is stored ((unsigned)-1 if T is not found). */ bool streamer_tree_cache_lookup (struct streamer_tree_cache_d *cache, tree t, unsigned *ix_p) { unsigned *slot; bool retval; unsigned ix; gcc_assert (t); slot = cache->node_map->get (t); if (slot == NULL) { retval = false; ix = -1; } else { retval = true; ix = *slot; } if (ix_p) *ix_p = ix; return retval; } /* Verify that NODE is in CACHE. */ static void verify_common_node_recorded (struct streamer_tree_cache_d *cache, tree node) { /* Restrict this to flag_checking only because in general violating it is harmless plus we never know what happens on all targets/frontend/flag(!) combinations. */ if (!flag_checking) return; if (cache->node_map) gcc_assert (streamer_tree_cache_lookup (cache, node, NULL)); else { bool found = false; gcc_assert (cache->nodes.exists ()); /* Linear search... */ for (unsigned i = 0; !found && i < cache->nodes.length (); ++i) if (cache->nodes[i] == node) found = true; gcc_assert (found); } } /* Record NODE in CACHE. */ static void record_common_node (struct streamer_tree_cache_d *cache, tree node) { /* If we recursively end up at nodes we do not want to preload simply don't. ??? We'd want to verify that this doesn't happen, or alternatively do not recurse at all. */ if (node == char_type_node) return; gcc_checking_assert (node != boolean_type_node && node != boolean_true_node && node != boolean_false_node); /* We have to make sure to fill exactly the same number of elements for all frontends. That can include NULL trees. As our hash table can't deal with zero entries we'll simply stream a random other tree. A NULL tree never will be looked up so it doesn't matter which tree we replace it with, just to be sure use error_mark_node. */ if (!node) node = error_mark_node; /* ??? FIXME, devise a better hash value. But the hash needs to be equal for all frontend and lto1 invocations. So just use the position in the cache as hash value. */ streamer_tree_cache_append (cache, node, cache->nodes.length ()); switch (TREE_CODE (node)) { case ERROR_MARK: case FIELD_DECL: case FIXED_POINT_TYPE: case IDENTIFIER_NODE: case INTEGER_CST: case INTEGER_TYPE: case POINTER_BOUNDS_TYPE: case REAL_TYPE: case TREE_LIST: case VOID_CST: case VOID_TYPE: /* No recursive trees. */ break; case ARRAY_TYPE: case POINTER_TYPE: case REFERENCE_TYPE: record_common_node (cache, TREE_TYPE (node)); break; case COMPLEX_TYPE: /* Verify that a complex type's component type (node_type) has been handled already (and we thus don't need to recurse here). */ verify_common_node_recorded (cache, TREE_TYPE (node)); break; case RECORD_TYPE: /* The FIELD_DECLs of structures should be shared, so that every COMPONENT_REF uses the same tree node when referencing a field. Pointer equality between FIELD_DECLs is used by the alias machinery to compute overlapping component references (see nonoverlapping_component_refs_p and nonoverlapping_component_refs_of_decl_p). */ for (tree f = TYPE_FIELDS (node); f; f = TREE_CHAIN (f)) record_common_node (cache, f); break; default: /* Unexpected tree code. */ gcc_unreachable (); } } /* Preload common nodes into CACHE and make sure they are merged properly according to the gimple type table. */ static void preload_common_nodes (struct streamer_tree_cache_d *cache) { unsigned i; for (i = 0; i < itk_none; i++) /* Skip itk_char. char_type_node is dependent on -f[un]signed-char. */ if (i != itk_char) record_common_node (cache, integer_types[i]); for (i = 0; i < stk_type_kind_last; i++) record_common_node (cache, sizetype_tab[i]); for (i = 0; i < TI_MAX; i++) /* Skip boolean type and constants, they are frontend dependent. */ if (i != TI_BOOLEAN_TYPE && i != TI_BOOLEAN_FALSE && i != TI_BOOLEAN_TRUE /* MAIN_IDENTIFIER is not always initialized by Fortran FE. */ && i != TI_MAIN_IDENTIFIER /* PID_TYPE is initialized only by C family front-ends. */ && i != TI_PID_TYPE /* Skip optimization and target option nodes; they depend on flags. */ && i != TI_OPTIMIZATION_DEFAULT && i != TI_OPTIMIZATION_CURRENT && i != TI_TARGET_OPTION_DEFAULT && i != TI_TARGET_OPTION_CURRENT && i != TI_CURRENT_TARGET_PRAGMA && i != TI_CURRENT_OPTIMIZE_PRAGMA /* Skip va_list* related nodes if offloading. For native LTO we want them to be merged for the stdarg pass, for offloading they might not be identical between host and offloading target. */ && (!lto_stream_offload_p || (i != TI_VA_LIST_TYPE && i != TI_VA_LIST_GPR_COUNTER_FIELD && i != TI_VA_LIST_FPR_COUNTER_FIELD))) record_common_node (cache, global_trees[i]); } /* Create a cache of pickled nodes. */ struct streamer_tree_cache_d * streamer_tree_cache_create (bool with_hashes, bool with_map, bool with_vec) { struct streamer_tree_cache_d *cache; cache = XCNEW (struct streamer_tree_cache_d); if (with_map) cache->node_map = new hash_map (251); cache->next_idx = 0; if (with_vec) cache->nodes.create (165); if (with_hashes) cache->hashes.create (165); /* Load all the well-known tree nodes that are always created by the compiler on startup. This prevents writing them out unnecessarily. */ preload_common_nodes (cache); return cache; } /* Delete the streamer cache C. */ void streamer_tree_cache_delete (struct streamer_tree_cache_d *c) { if (c == NULL) return; delete c->node_map; c->node_map = NULL; c->nodes.release (); c->hashes.release (); free (c); }