| 1 | /* Tree inlining. |
| 2 | Copyright 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009 |
| 3 | Free Software Foundation, Inc. |
| 4 | Contributed by Alexandre Oliva <aoliva@redhat.com> |
| 5 | |
| 6 | This file is part of GCC. |
| 7 | |
| 8 | GCC is free software; you can redistribute it and/or modify |
| 9 | it under the terms of the GNU General Public License as published by |
| 10 | the Free Software Foundation; either version 3, or (at your option) |
| 11 | any later version. |
| 12 | |
| 13 | GCC is distributed in the hope that it will be useful, |
| 14 | but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 15 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| 16 | GNU General Public License for more details. |
| 17 | |
| 18 | You should have received a copy of the GNU General Public License |
| 19 | along with GCC; see the file COPYING3. If not see |
| 20 | <http://www.gnu.org/licenses/>. */ |
| 21 | |
| 22 | #include "config.h" |
| 23 | #include "system.h" |
| 24 | #include "coretypes.h" |
| 25 | #include "tm.h" |
| 26 | #include "toplev.h" |
| 27 | #include "tree.h" |
| 28 | #include "tree-inline.h" |
| 29 | #include "rtl.h" |
| 30 | #include "expr.h" |
| 31 | #include "flags.h" |
| 32 | #include "params.h" |
| 33 | #include "input.h" |
| 34 | #include "insn-config.h" |
| 35 | #include "varray.h" |
| 36 | #include "hashtab.h" |
| 37 | #include "langhooks.h" |
| 38 | #include "basic-block.h" |
| 39 | #include "tree-iterator.h" |
| 40 | #include "cgraph.h" |
| 41 | #include "intl.h" |
| 42 | #include "tree-mudflap.h" |
| 43 | #include "tree-flow.h" |
| 44 | #include "function.h" |
| 45 | #include "ggc.h" |
| 46 | #include "tree-flow.h" |
| 47 | #include "diagnostic.h" |
| 48 | #include "except.h" |
| 49 | #include "debug.h" |
| 50 | #include "pointer-set.h" |
| 51 | #include "ipa-prop.h" |
| 52 | #include "value-prof.h" |
| 53 | #include "tree-pass.h" |
| 54 | #include "target.h" |
| 55 | #include "integrate.h" |
| 56 | |
| 57 | /* I'm not real happy about this, but we need to handle gimple and |
| 58 | non-gimple trees. */ |
| 59 | #include "gimple.h" |
| 60 | |
| 61 | /* Inlining, Cloning, Versioning, Parallelization |
| 62 | |
| 63 | Inlining: a function body is duplicated, but the PARM_DECLs are |
| 64 | remapped into VAR_DECLs, and non-void RETURN_EXPRs become |
| 65 | MODIFY_EXPRs that store to a dedicated returned-value variable. |
| 66 | The duplicated eh_region info of the copy will later be appended |
| 67 | to the info for the caller; the eh_region info in copied throwing |
| 68 | statements and RESX_EXPRs is adjusted accordingly. |
| 69 | |
| 70 | Cloning: (only in C++) We have one body for a con/de/structor, and |
| 71 | multiple function decls, each with a unique parameter list. |
| 72 | Duplicate the body, using the given splay tree; some parameters |
| 73 | will become constants (like 0 or 1). |
| 74 | |
| 75 | Versioning: a function body is duplicated and the result is a new |
| 76 | function rather than into blocks of an existing function as with |
| 77 | inlining. Some parameters will become constants. |
| 78 | |
| 79 | Parallelization: a region of a function is duplicated resulting in |
| 80 | a new function. Variables may be replaced with complex expressions |
| 81 | to enable shared variable semantics. |
| 82 | |
| 83 | All of these will simultaneously lookup any callgraph edges. If |
| 84 | we're going to inline the duplicated function body, and the given |
| 85 | function has some cloned callgraph nodes (one for each place this |
| 86 | function will be inlined) those callgraph edges will be duplicated. |
| 87 | If we're cloning the body, those callgraph edges will be |
| 88 | updated to point into the new body. (Note that the original |
| 89 | callgraph node and edge list will not be altered.) |
| 90 | |
| 91 | See the CALL_EXPR handling case in copy_tree_body_r (). */ |
| 92 | |
| 93 | /* To Do: |
| 94 | |
| 95 | o In order to make inlining-on-trees work, we pessimized |
| 96 | function-local static constants. In particular, they are now |
| 97 | always output, even when not addressed. Fix this by treating |
| 98 | function-local static constants just like global static |
| 99 | constants; the back-end already knows not to output them if they |
| 100 | are not needed. |
| 101 | |
| 102 | o Provide heuristics to clamp inlining of recursive template |
| 103 | calls? */ |
| 104 | |
| 105 | |
| 106 | /* Weights that estimate_num_insns uses for heuristics in inlining. */ |
| 107 | |
| 108 | eni_weights eni_inlining_weights; |
| 109 | |
| 110 | /* Weights that estimate_num_insns uses to estimate the size of the |
| 111 | produced code. */ |
| 112 | |
| 113 | eni_weights eni_size_weights; |
| 114 | |
| 115 | /* Weights that estimate_num_insns uses to estimate the time necessary |
| 116 | to execute the produced code. */ |
| 117 | |
| 118 | eni_weights eni_time_weights; |
| 119 | |
| 120 | /* Prototypes. */ |
| 121 | |
| 122 | static tree declare_return_variable (copy_body_data *, tree, tree, tree *); |
| 123 | static bool inlinable_function_p (tree); |
| 124 | static void remap_block (tree *, copy_body_data *); |
| 125 | static void copy_bind_expr (tree *, int *, copy_body_data *); |
| 126 | static tree mark_local_for_remap_r (tree *, int *, void *); |
| 127 | static void unsave_expr_1 (tree); |
| 128 | static tree unsave_r (tree *, int *, void *); |
| 129 | static void declare_inline_vars (tree, tree); |
| 130 | static void remap_save_expr (tree *, void *, int *); |
| 131 | static void prepend_lexical_block (tree current_block, tree new_block); |
| 132 | static tree copy_decl_to_var (tree, copy_body_data *); |
| 133 | static tree copy_result_decl_to_var (tree, copy_body_data *); |
| 134 | static tree copy_decl_maybe_to_var (tree, copy_body_data *); |
| 135 | static gimple remap_gimple_stmt (gimple, copy_body_data *); |
| 136 | |
| 137 | /* Insert a tree->tree mapping for ID. Despite the name suggests |
| 138 | that the trees should be variables, it is used for more than that. */ |
| 139 | |
| 140 | void |
| 141 | insert_decl_map (copy_body_data *id, tree key, tree value) |
| 142 | { |
| 143 | *pointer_map_insert (id->decl_map, key) = value; |
| 144 | |
| 145 | /* Always insert an identity map as well. If we see this same new |
| 146 | node again, we won't want to duplicate it a second time. */ |
| 147 | if (key != value) |
| 148 | *pointer_map_insert (id->decl_map, value) = value; |
| 149 | } |
| 150 | |
| 151 | /* Construct new SSA name for old NAME. ID is the inline context. */ |
| 152 | |
| 153 | static tree |
| 154 | remap_ssa_name (tree name, copy_body_data *id) |
| 155 | { |
| 156 | tree new_tree; |
| 157 | tree *n; |
| 158 | |
| 159 | gcc_assert (TREE_CODE (name) == SSA_NAME); |
| 160 | |
| 161 | n = (tree *) pointer_map_contains (id->decl_map, name); |
| 162 | if (n) |
| 163 | return unshare_expr (*n); |
| 164 | |
| 165 | /* Do not set DEF_STMT yet as statement is not copied yet. We do that |
| 166 | in copy_bb. */ |
| 167 | new_tree = remap_decl (SSA_NAME_VAR (name), id); |
| 168 | |
| 169 | /* We might've substituted constant or another SSA_NAME for |
| 170 | the variable. |
| 171 | |
| 172 | Replace the SSA name representing RESULT_DECL by variable during |
| 173 | inlining: this saves us from need to introduce PHI node in a case |
| 174 | return value is just partly initialized. */ |
| 175 | if ((TREE_CODE (new_tree) == VAR_DECL || TREE_CODE (new_tree) == PARM_DECL) |
| 176 | && (TREE_CODE (SSA_NAME_VAR (name)) != RESULT_DECL |
| 177 | || !id->transform_return_to_modify)) |
| 178 | { |
| 179 | new_tree = make_ssa_name (new_tree, NULL); |
| 180 | insert_decl_map (id, name, new_tree); |
| 181 | SSA_NAME_OCCURS_IN_ABNORMAL_PHI (new_tree) |
| 182 | = SSA_NAME_OCCURS_IN_ABNORMAL_PHI (name); |
| 183 | TREE_TYPE (new_tree) = TREE_TYPE (SSA_NAME_VAR (new_tree)); |
| 184 | if (gimple_nop_p (SSA_NAME_DEF_STMT (name))) |
| 185 | { |
| 186 | /* By inlining function having uninitialized variable, we might |
| 187 | extend the lifetime (variable might get reused). This cause |
| 188 | ICE in the case we end up extending lifetime of SSA name across |
| 189 | abnormal edge, but also increase register pressure. |
| 190 | |
| 191 | We simply initialize all uninitialized vars by 0 except |
| 192 | for case we are inlining to very first BB. We can avoid |
| 193 | this for all BBs that are not inside strongly connected |
| 194 | regions of the CFG, but this is expensive to test. */ |
| 195 | if (id->entry_bb |
| 196 | && is_gimple_reg (SSA_NAME_VAR (name)) |
| 197 | && TREE_CODE (SSA_NAME_VAR (name)) != PARM_DECL |
| 198 | && (id->entry_bb != EDGE_SUCC (ENTRY_BLOCK_PTR, 0)->dest |
| 199 | || EDGE_COUNT (id->entry_bb->preds) != 1)) |
| 200 | { |
| 201 | gimple_stmt_iterator gsi = gsi_last_bb (id->entry_bb); |
| 202 | gimple init_stmt; |
| 203 | |
| 204 | init_stmt = gimple_build_assign (new_tree, |
| 205 | fold_convert (TREE_TYPE (new_tree), |
| 206 | integer_zero_node)); |
| 207 | gsi_insert_after (&gsi, init_stmt, GSI_NEW_STMT); |
| 208 | SSA_NAME_IS_DEFAULT_DEF (new_tree) = 0; |
| 209 | } |
| 210 | else |
| 211 | { |
| 212 | SSA_NAME_DEF_STMT (new_tree) = gimple_build_nop (); |
| 213 | if (gimple_default_def (id->src_cfun, SSA_NAME_VAR (name)) |
| 214 | == name) |
| 215 | set_default_def (SSA_NAME_VAR (new_tree), new_tree); |
| 216 | } |
| 217 | } |
| 218 | } |
| 219 | else |
| 220 | insert_decl_map (id, name, new_tree); |
| 221 | return new_tree; |
| 222 | } |
| 223 | |
| 224 | /* Remap DECL during the copying of the BLOCK tree for the function. */ |
| 225 | |
| 226 | tree |
| 227 | remap_decl (tree decl, copy_body_data *id) |
| 228 | { |
| 229 | tree *n; |
| 230 | tree fn; |
| 231 | |
| 232 | /* We only remap local variables in the current function. */ |
| 233 | fn = id->src_fn; |
| 234 | |
| 235 | /* See if we have remapped this declaration. */ |
| 236 | |
| 237 | n = (tree *) pointer_map_contains (id->decl_map, decl); |
| 238 | |
| 239 | /* If we didn't already have an equivalent for this declaration, |
| 240 | create one now. */ |
| 241 | if (!n) |
| 242 | { |
| 243 | /* Make a copy of the variable or label. */ |
| 244 | tree t = id->copy_decl (decl, id); |
| 245 | |
| 246 | /* Remember it, so that if we encounter this local entity again |
| 247 | we can reuse this copy. Do this early because remap_type may |
| 248 | need this decl for TYPE_STUB_DECL. */ |
| 249 | insert_decl_map (id, decl, t); |
| 250 | |
| 251 | if (!DECL_P (t)) |
| 252 | return t; |
| 253 | |
| 254 | /* Remap types, if necessary. */ |
| 255 | TREE_TYPE (t) = remap_type (TREE_TYPE (t), id); |
| 256 | if (TREE_CODE (t) == TYPE_DECL) |
| 257 | DECL_ORIGINAL_TYPE (t) = remap_type (DECL_ORIGINAL_TYPE (t), id); |
| 258 | |
| 259 | /* Remap sizes as necessary. */ |
| 260 | walk_tree (&DECL_SIZE (t), copy_tree_body_r, id, NULL); |
| 261 | walk_tree (&DECL_SIZE_UNIT (t), copy_tree_body_r, id, NULL); |
| 262 | |
| 263 | /* If fields, do likewise for offset and qualifier. */ |
| 264 | if (TREE_CODE (t) == FIELD_DECL) |
| 265 | { |
| 266 | walk_tree (&DECL_FIELD_OFFSET (t), copy_tree_body_r, id, NULL); |
| 267 | if (TREE_CODE (DECL_CONTEXT (t)) == QUAL_UNION_TYPE) |
| 268 | walk_tree (&DECL_QUALIFIER (t), copy_tree_body_r, id, NULL); |
| 269 | } |
| 270 | |
| 271 | if (cfun && gimple_in_ssa_p (cfun) |
| 272 | && (TREE_CODE (t) == VAR_DECL |
| 273 | || TREE_CODE (t) == RESULT_DECL || TREE_CODE (t) == PARM_DECL)) |
| 274 | { |
| 275 | tree def = gimple_default_def (id->src_cfun, decl); |
| 276 | get_var_ann (t); |
| 277 | if (TREE_CODE (decl) != PARM_DECL && def) |
| 278 | { |
| 279 | tree map = remap_ssa_name (def, id); |
| 280 | /* Watch out RESULT_DECLs whose SSA names map directly |
| 281 | to them. */ |
| 282 | if (TREE_CODE (map) == SSA_NAME |
| 283 | && gimple_nop_p (SSA_NAME_DEF_STMT (map))) |
| 284 | set_default_def (t, map); |
| 285 | } |
| 286 | add_referenced_var (t); |
| 287 | } |
| 288 | return t; |
| 289 | } |
| 290 | |
| 291 | return unshare_expr (*n); |
| 292 | } |
| 293 | |
| 294 | static tree |
| 295 | remap_type_1 (tree type, copy_body_data *id) |
| 296 | { |
| 297 | tree new_tree, t; |
| 298 | |
| 299 | /* We do need a copy. build and register it now. If this is a pointer or |
| 300 | reference type, remap the designated type and make a new pointer or |
| 301 | reference type. */ |
| 302 | if (TREE_CODE (type) == POINTER_TYPE) |
| 303 | { |
| 304 | new_tree = build_pointer_type_for_mode (remap_type (TREE_TYPE (type), id), |
| 305 | TYPE_MODE (type), |
| 306 | TYPE_REF_CAN_ALIAS_ALL (type)); |
| 307 | insert_decl_map (id, type, new_tree); |
| 308 | return new_tree; |
| 309 | } |
| 310 | else if (TREE_CODE (type) == REFERENCE_TYPE) |
| 311 | { |
| 312 | new_tree = build_reference_type_for_mode (remap_type (TREE_TYPE (type), id), |
| 313 | TYPE_MODE (type), |
| 314 | TYPE_REF_CAN_ALIAS_ALL (type)); |
| 315 | insert_decl_map (id, type, new_tree); |
| 316 | return new_tree; |
| 317 | } |
| 318 | else |
| 319 | new_tree = copy_node (type); |
| 320 | |
| 321 | insert_decl_map (id, type, new_tree); |
| 322 | |
| 323 | /* This is a new type, not a copy of an old type. Need to reassociate |
| 324 | variants. We can handle everything except the main variant lazily. */ |
| 325 | t = TYPE_MAIN_VARIANT (type); |
| 326 | if (type != t) |
| 327 | { |
| 328 | t = remap_type (t, id); |
| 329 | TYPE_MAIN_VARIANT (new_tree) = t; |
| 330 | TYPE_NEXT_VARIANT (new_tree) = TYPE_NEXT_VARIANT (t); |
| 331 | TYPE_NEXT_VARIANT (t) = new_tree; |
| 332 | } |
| 333 | else |
| 334 | { |
| 335 | TYPE_MAIN_VARIANT (new_tree) = new_tree; |
| 336 | TYPE_NEXT_VARIANT (new_tree) = NULL; |
| 337 | } |
| 338 | |
| 339 | if (TYPE_STUB_DECL (type)) |
| 340 | TYPE_STUB_DECL (new_tree) = remap_decl (TYPE_STUB_DECL (type), id); |
| 341 | |
| 342 | /* Lazily create pointer and reference types. */ |
| 343 | TYPE_POINTER_TO (new_tree) = NULL; |
| 344 | TYPE_REFERENCE_TO (new_tree) = NULL; |
| 345 | |
| 346 | switch (TREE_CODE (new_tree)) |
| 347 | { |
| 348 | case INTEGER_TYPE: |
| 349 | case REAL_TYPE: |
| 350 | case FIXED_POINT_TYPE: |
| 351 | case ENUMERAL_TYPE: |
| 352 | case BOOLEAN_TYPE: |
| 353 | t = TYPE_MIN_VALUE (new_tree); |
| 354 | if (t && TREE_CODE (t) != INTEGER_CST) |
| 355 | walk_tree (&TYPE_MIN_VALUE (new_tree), copy_tree_body_r, id, NULL); |
| 356 | |
| 357 | t = TYPE_MAX_VALUE (new_tree); |
| 358 | if (t && TREE_CODE (t) != INTEGER_CST) |
| 359 | walk_tree (&TYPE_MAX_VALUE (new_tree), copy_tree_body_r, id, NULL); |
| 360 | return new_tree; |
| 361 | |
| 362 | case FUNCTION_TYPE: |
| 363 | TREE_TYPE (new_tree) = remap_type (TREE_TYPE (new_tree), id); |
| 364 | walk_tree (&TYPE_ARG_TYPES (new_tree), copy_tree_body_r, id, NULL); |
| 365 | return new_tree; |
| 366 | |
| 367 | case ARRAY_TYPE: |
| 368 | TREE_TYPE (new_tree) = remap_type (TREE_TYPE (new_tree), id); |
| 369 | TYPE_DOMAIN (new_tree) = remap_type (TYPE_DOMAIN (new_tree), id); |
| 370 | break; |
| 371 | |
| 372 | case RECORD_TYPE: |
| 373 | case UNION_TYPE: |
| 374 | case QUAL_UNION_TYPE: |
| 375 | { |
| 376 | tree f, nf = NULL; |
| 377 | |
| 378 | for (f = TYPE_FIELDS (new_tree); f ; f = TREE_CHAIN (f)) |
| 379 | { |
| 380 | t = remap_decl (f, id); |
| 381 | DECL_CONTEXT (t) = new_tree; |
| 382 | TREE_CHAIN (t) = nf; |
| 383 | nf = t; |
| 384 | } |
| 385 | TYPE_FIELDS (new_tree) = nreverse (nf); |
| 386 | } |
| 387 | break; |
| 388 | |
| 389 | case OFFSET_TYPE: |
| 390 | default: |
| 391 | /* Shouldn't have been thought variable sized. */ |
| 392 | gcc_unreachable (); |
| 393 | } |
| 394 | |
| 395 | walk_tree (&TYPE_SIZE (new_tree), copy_tree_body_r, id, NULL); |
| 396 | walk_tree (&TYPE_SIZE_UNIT (new_tree), copy_tree_body_r, id, NULL); |
| 397 | |
| 398 | return new_tree; |
| 399 | } |
| 400 | |
| 401 | tree |
| 402 | remap_type (tree type, copy_body_data *id) |
| 403 | { |
| 404 | tree *node; |
| 405 | tree tmp; |
| 406 | |
| 407 | if (type == NULL) |
| 408 | return type; |
| 409 | |
| 410 | /* See if we have remapped this type. */ |
| 411 | node = (tree *) pointer_map_contains (id->decl_map, type); |
| 412 | if (node) |
| 413 | return *node; |
| 414 | |
| 415 | /* The type only needs remapping if it's variably modified. */ |
| 416 | if (! variably_modified_type_p (type, id->src_fn)) |
| 417 | { |
| 418 | insert_decl_map (id, type, type); |
| 419 | return type; |
| 420 | } |
| 421 | |
| 422 | id->remapping_type_depth++; |
| 423 | tmp = remap_type_1 (type, id); |
| 424 | id->remapping_type_depth--; |
| 425 | |
| 426 | return tmp; |
| 427 | } |
| 428 | |
| 429 | /* Return previously remapped type of TYPE in ID. Return NULL if TYPE |
| 430 | is NULL or TYPE has not been remapped before. */ |
| 431 | |
| 432 | static tree |
| 433 | remapped_type (tree type, copy_body_data *id) |
| 434 | { |
| 435 | tree *node; |
| 436 | |
| 437 | if (type == NULL) |
| 438 | return type; |
| 439 | |
| 440 | /* See if we have remapped this type. */ |
| 441 | node = (tree *) pointer_map_contains (id->decl_map, type); |
| 442 | if (node) |
| 443 | return *node; |
| 444 | else |
| 445 | return NULL; |
| 446 | } |
| 447 | |
| 448 | /* The type only needs remapping if it's variably modified. */ |
| 449 | /* Decide if DECL can be put into BLOCK_NONLOCAL_VARs. */ |
| 450 | |
| 451 | static bool |
| 452 | can_be_nonlocal (tree decl, copy_body_data *id) |
| 453 | { |
| 454 | /* We can not duplicate function decls. */ |
| 455 | if (TREE_CODE (decl) == FUNCTION_DECL) |
| 456 | return true; |
| 457 | |
| 458 | /* Local static vars must be non-local or we get multiple declaration |
| 459 | problems. */ |
| 460 | if (TREE_CODE (decl) == VAR_DECL |
| 461 | && !auto_var_in_fn_p (decl, id->src_fn)) |
| 462 | return true; |
| 463 | |
| 464 | /* At the moment dwarf2out can handle only these types of nodes. We |
| 465 | can support more later. */ |
| 466 | if (TREE_CODE (decl) != VAR_DECL && TREE_CODE (decl) != PARM_DECL) |
| 467 | return false; |
| 468 | |
| 469 | /* We must use global type. We call remapped_type instead of |
| 470 | remap_type since we don't want to remap this type here if it |
| 471 | hasn't been remapped before. */ |
| 472 | if (TREE_TYPE (decl) != remapped_type (TREE_TYPE (decl), id)) |
| 473 | return false; |
| 474 | |
| 475 | /* Wihtout SSA we can't tell if variable is used. */ |
| 476 | if (!gimple_in_ssa_p (cfun)) |
| 477 | return false; |
| 478 | |
| 479 | /* Live variables must be copied so we can attach DECL_RTL. */ |
| 480 | if (var_ann (decl)) |
| 481 | return false; |
| 482 | |
| 483 | return true; |
| 484 | } |
| 485 | |
| 486 | static tree |
| 487 | remap_decls (tree decls, VEC(tree,gc) **nonlocalized_list, copy_body_data *id) |
| 488 | { |
| 489 | tree old_var; |
| 490 | tree new_decls = NULL_TREE; |
| 491 | |
| 492 | /* Remap its variables. */ |
| 493 | for (old_var = decls; old_var; old_var = TREE_CHAIN (old_var)) |
| 494 | { |
| 495 | tree new_var; |
| 496 | tree origin_var = DECL_ORIGIN (old_var); |
| 497 | |
| 498 | if (can_be_nonlocal (old_var, id)) |
| 499 | { |
| 500 | if (TREE_CODE (old_var) == VAR_DECL |
| 501 | && (var_ann (old_var) || !gimple_in_ssa_p (cfun))) |
| 502 | cfun->local_decls = tree_cons (NULL_TREE, old_var, |
| 503 | cfun->local_decls); |
| 504 | if ((!optimize || debug_info_level > DINFO_LEVEL_TERSE) |
| 505 | && !DECL_IGNORED_P (old_var) |
| 506 | && nonlocalized_list) |
| 507 | VEC_safe_push (tree, gc, *nonlocalized_list, origin_var); |
| 508 | continue; |
| 509 | } |
| 510 | |
| 511 | /* Remap the variable. */ |
| 512 | new_var = remap_decl (old_var, id); |
| 513 | |
| 514 | /* If we didn't remap this variable, we can't mess with its |
| 515 | TREE_CHAIN. If we remapped this variable to the return slot, it's |
| 516 | already declared somewhere else, so don't declare it here. */ |
| 517 | |
| 518 | if (new_var == id->retvar) |
| 519 | ; |
| 520 | else if (!new_var) |
| 521 | { |
| 522 | if ((!optimize || debug_info_level > DINFO_LEVEL_TERSE) |
| 523 | && !DECL_IGNORED_P (old_var) |
| 524 | && nonlocalized_list) |
| 525 | VEC_safe_push (tree, gc, *nonlocalized_list, origin_var); |
| 526 | } |
| 527 | else |
| 528 | { |
| 529 | gcc_assert (DECL_P (new_var)); |
| 530 | TREE_CHAIN (new_var) = new_decls; |
| 531 | new_decls = new_var; |
| 532 | } |
| 533 | } |
| 534 | |
| 535 | return nreverse (new_decls); |
| 536 | } |
| 537 | |
| 538 | /* Copy the BLOCK to contain remapped versions of the variables |
| 539 | therein. And hook the new block into the block-tree. */ |
| 540 | |
| 541 | static void |
| 542 | remap_block (tree *block, copy_body_data *id) |
| 543 | { |
| 544 | tree old_block; |
| 545 | tree new_block; |
| 546 | tree fn; |
| 547 | |
| 548 | /* Make the new block. */ |
| 549 | old_block = *block; |
| 550 | new_block = make_node (BLOCK); |
| 551 | TREE_USED (new_block) = TREE_USED (old_block); |
| 552 | BLOCK_ABSTRACT_ORIGIN (new_block) = old_block; |
| 553 | BLOCK_SOURCE_LOCATION (new_block) = BLOCK_SOURCE_LOCATION (old_block); |
| 554 | BLOCK_NONLOCALIZED_VARS (new_block) |
| 555 | = VEC_copy (tree, gc, BLOCK_NONLOCALIZED_VARS (old_block)); |
| 556 | *block = new_block; |
| 557 | |
| 558 | /* Remap its variables. */ |
| 559 | BLOCK_VARS (new_block) = remap_decls (BLOCK_VARS (old_block), |
| 560 | &BLOCK_NONLOCALIZED_VARS (new_block), |
| 561 | id); |
| 562 | |
| 563 | fn = id->dst_fn; |
| 564 | |
| 565 | if (id->transform_lang_insert_block) |
| 566 | id->transform_lang_insert_block (new_block); |
| 567 | |
| 568 | /* Remember the remapped block. */ |
| 569 | insert_decl_map (id, old_block, new_block); |
| 570 | } |
| 571 | |
| 572 | /* Copy the whole block tree and root it in id->block. */ |
| 573 | static tree |
| 574 | remap_blocks (tree block, copy_body_data *id) |
| 575 | { |
| 576 | tree t; |
| 577 | tree new_tree = block; |
| 578 | |
| 579 | if (!block) |
| 580 | return NULL; |
| 581 | |
| 582 | remap_block (&new_tree, id); |
| 583 | gcc_assert (new_tree != block); |
| 584 | for (t = BLOCK_SUBBLOCKS (block); t ; t = BLOCK_CHAIN (t)) |
| 585 | prepend_lexical_block (new_tree, remap_blocks (t, id)); |
| 586 | /* Blocks are in arbitrary order, but make things slightly prettier and do |
| 587 | not swap order when producing a copy. */ |
| 588 | BLOCK_SUBBLOCKS (new_tree) = blocks_nreverse (BLOCK_SUBBLOCKS (new_tree)); |
| 589 | return new_tree; |
| 590 | } |
| 591 | |
| 592 | static void |
| 593 | copy_statement_list (tree *tp) |
| 594 | { |
| 595 | tree_stmt_iterator oi, ni; |
| 596 | tree new_tree; |
| 597 | |
| 598 | new_tree = alloc_stmt_list (); |
| 599 | ni = tsi_start (new_tree); |
| 600 | oi = tsi_start (*tp); |
| 601 | *tp = new_tree; |
| 602 | |
| 603 | for (; !tsi_end_p (oi); tsi_next (&oi)) |
| 604 | tsi_link_after (&ni, tsi_stmt (oi), TSI_NEW_STMT); |
| 605 | } |
| 606 | |
| 607 | static void |
| 608 | copy_bind_expr (tree *tp, int *walk_subtrees, copy_body_data *id) |
| 609 | { |
| 610 | tree block = BIND_EXPR_BLOCK (*tp); |
| 611 | /* Copy (and replace) the statement. */ |
| 612 | copy_tree_r (tp, walk_subtrees, NULL); |
| 613 | if (block) |
| 614 | { |
| 615 | remap_block (&block, id); |
| 616 | BIND_EXPR_BLOCK (*tp) = block; |
| 617 | } |
| 618 | |
| 619 | if (BIND_EXPR_VARS (*tp)) |
| 620 | /* This will remap a lot of the same decls again, but this should be |
| 621 | harmless. */ |
| 622 | BIND_EXPR_VARS (*tp) = remap_decls (BIND_EXPR_VARS (*tp), NULL, id); |
| 623 | } |
| 624 | |
| 625 | |
| 626 | /* Create a new gimple_seq by remapping all the statements in BODY |
| 627 | using the inlining information in ID. */ |
| 628 | |
| 629 | gimple_seq |
| 630 | remap_gimple_seq (gimple_seq body, copy_body_data *id) |
| 631 | { |
| 632 | gimple_stmt_iterator si; |
| 633 | gimple_seq new_body = NULL; |
| 634 | |
| 635 | for (si = gsi_start (body); !gsi_end_p (si); gsi_next (&si)) |
| 636 | { |
| 637 | gimple new_stmt = remap_gimple_stmt (gsi_stmt (si), id); |
| 638 | gimple_seq_add_stmt (&new_body, new_stmt); |
| 639 | } |
| 640 | |
| 641 | return new_body; |
| 642 | } |
| 643 | |
| 644 | |
| 645 | /* Copy a GIMPLE_BIND statement STMT, remapping all the symbols in its |
| 646 | block using the mapping information in ID. */ |
| 647 | |
| 648 | static gimple |
| 649 | copy_gimple_bind (gimple stmt, copy_body_data *id) |
| 650 | { |
| 651 | gimple new_bind; |
| 652 | tree new_block, new_vars; |
| 653 | gimple_seq body, new_body; |
| 654 | |
| 655 | /* Copy the statement. Note that we purposely don't use copy_stmt |
| 656 | here because we need to remap statements as we copy. */ |
| 657 | body = gimple_bind_body (stmt); |
| 658 | new_body = remap_gimple_seq (body, id); |
| 659 | |
| 660 | new_block = gimple_bind_block (stmt); |
| 661 | if (new_block) |
| 662 | remap_block (&new_block, id); |
| 663 | |
| 664 | /* This will remap a lot of the same decls again, but this should be |
| 665 | harmless. */ |
| 666 | new_vars = gimple_bind_vars (stmt); |
| 667 | if (new_vars) |
| 668 | new_vars = remap_decls (new_vars, NULL, id); |
| 669 | |
| 670 | new_bind = gimple_build_bind (new_vars, new_body, new_block); |
| 671 | |
| 672 | return new_bind; |
| 673 | } |
| 674 | |
| 675 | |
| 676 | /* Remap the GIMPLE operand pointed to by *TP. DATA is really a |
| 677 | 'struct walk_stmt_info *'. DATA->INFO is a 'copy_body_data *'. |
| 678 | WALK_SUBTREES is used to indicate walk_gimple_op whether to keep |
| 679 | recursing into the children nodes of *TP. */ |
| 680 | |
| 681 | static tree |
| 682 | remap_gimple_op_r (tree *tp, int *walk_subtrees, void *data) |
| 683 | { |
| 684 | struct walk_stmt_info *wi_p = (struct walk_stmt_info *) data; |
| 685 | copy_body_data *id = (copy_body_data *) wi_p->info; |
| 686 | tree fn = id->src_fn; |
| 687 | |
| 688 | if (TREE_CODE (*tp) == SSA_NAME) |
| 689 | { |
| 690 | *tp = remap_ssa_name (*tp, id); |
| 691 | *walk_subtrees = 0; |
| 692 | return NULL; |
| 693 | } |
| 694 | else if (auto_var_in_fn_p (*tp, fn)) |
| 695 | { |
| 696 | /* Local variables and labels need to be replaced by equivalent |
| 697 | variables. We don't want to copy static variables; there's |
| 698 | only one of those, no matter how many times we inline the |
| 699 | containing function. Similarly for globals from an outer |
| 700 | function. */ |
| 701 | tree new_decl; |
| 702 | |
| 703 | /* Remap the declaration. */ |
| 704 | new_decl = remap_decl (*tp, id); |
| 705 | gcc_assert (new_decl); |
| 706 | /* Replace this variable with the copy. */ |
| 707 | STRIP_TYPE_NOPS (new_decl); |
| 708 | *tp = new_decl; |
| 709 | *walk_subtrees = 0; |
| 710 | } |
| 711 | else if (TREE_CODE (*tp) == STATEMENT_LIST) |
| 712 | gcc_unreachable (); |
| 713 | else if (TREE_CODE (*tp) == SAVE_EXPR) |
| 714 | gcc_unreachable (); |
| 715 | else if (TREE_CODE (*tp) == LABEL_DECL |
| 716 | && (!DECL_CONTEXT (*tp) |
| 717 | || decl_function_context (*tp) == id->src_fn)) |
| 718 | /* These may need to be remapped for EH handling. */ |
| 719 | *tp = remap_decl (*tp, id); |
| 720 | else if (TYPE_P (*tp)) |
| 721 | /* Types may need remapping as well. */ |
| 722 | *tp = remap_type (*tp, id); |
| 723 | else if (CONSTANT_CLASS_P (*tp)) |
| 724 | { |
| 725 | /* If this is a constant, we have to copy the node iff the type |
| 726 | will be remapped. copy_tree_r will not copy a constant. */ |
| 727 | tree new_type = remap_type (TREE_TYPE (*tp), id); |
| 728 | |
| 729 | if (new_type == TREE_TYPE (*tp)) |
| 730 | *walk_subtrees = 0; |
| 731 | |
| 732 | else if (TREE_CODE (*tp) == INTEGER_CST) |
| 733 | *tp = build_int_cst_wide (new_type, TREE_INT_CST_LOW (*tp), |
| 734 | TREE_INT_CST_HIGH (*tp)); |
| 735 | else |
| 736 | { |
| 737 | *tp = copy_node (*tp); |
| 738 | TREE_TYPE (*tp) = new_type; |
| 739 | } |
| 740 | } |
| 741 | else |
| 742 | { |
| 743 | /* Otherwise, just copy the node. Note that copy_tree_r already |
| 744 | knows not to copy VAR_DECLs, etc., so this is safe. */ |
| 745 | if (TREE_CODE (*tp) == INDIRECT_REF) |
| 746 | { |
| 747 | /* Get rid of *& from inline substitutions that can happen when a |
| 748 | pointer argument is an ADDR_EXPR. */ |
| 749 | tree decl = TREE_OPERAND (*tp, 0); |
| 750 | tree *n; |
| 751 | |
| 752 | n = (tree *) pointer_map_contains (id->decl_map, decl); |
| 753 | if (n) |
| 754 | { |
| 755 | tree type, new_tree, old; |
| 756 | |
| 757 | /* If we happen to get an ADDR_EXPR in n->value, strip |
| 758 | it manually here as we'll eventually get ADDR_EXPRs |
| 759 | which lie about their types pointed to. In this case |
| 760 | build_fold_indirect_ref wouldn't strip the |
| 761 | INDIRECT_REF, but we absolutely rely on that. As |
| 762 | fold_indirect_ref does other useful transformations, |
| 763 | try that first, though. */ |
| 764 | type = TREE_TYPE (TREE_TYPE (*n)); |
| 765 | new_tree = unshare_expr (*n); |
| 766 | old = *tp; |
| 767 | *tp = gimple_fold_indirect_ref (new_tree); |
| 768 | if (!*tp) |
| 769 | { |
| 770 | if (TREE_CODE (new_tree) == ADDR_EXPR) |
| 771 | { |
| 772 | *tp = fold_indirect_ref_1 (type, new_tree); |
| 773 | /* ??? We should either assert here or build |
| 774 | a VIEW_CONVERT_EXPR instead of blindly leaking |
| 775 | incompatible types to our IL. */ |
| 776 | if (! *tp) |
| 777 | *tp = TREE_OPERAND (new_tree, 0); |
| 778 | } |
| 779 | else |
| 780 | { |
| 781 | *tp = build1 (INDIRECT_REF, type, new_tree); |
| 782 | TREE_THIS_VOLATILE (*tp) = TREE_THIS_VOLATILE (old); |
| 783 | TREE_NO_WARNING (*tp) = TREE_NO_WARNING (old); |
| 784 | } |
| 785 | } |
| 786 | *walk_subtrees = 0; |
| 787 | return NULL; |
| 788 | } |
| 789 | } |
| 790 | |
| 791 | /* Here is the "usual case". Copy this tree node, and then |
| 792 | tweak some special cases. */ |
| 793 | copy_tree_r (tp, walk_subtrees, NULL); |
| 794 | |
| 795 | /* Global variables we haven't seen yet need to go into referenced |
| 796 | vars. If not referenced from types only. */ |
| 797 | if (gimple_in_ssa_p (cfun) |
| 798 | && TREE_CODE (*tp) == VAR_DECL |
| 799 | && id->remapping_type_depth == 0) |
| 800 | add_referenced_var (*tp); |
| 801 | |
| 802 | /* We should never have TREE_BLOCK set on non-statements. */ |
| 803 | if (EXPR_P (*tp)) |
| 804 | gcc_assert (!TREE_BLOCK (*tp)); |
| 805 | |
| 806 | if (TREE_CODE (*tp) != OMP_CLAUSE) |
| 807 | TREE_TYPE (*tp) = remap_type (TREE_TYPE (*tp), id); |
| 808 | |
| 809 | if (TREE_CODE (*tp) == TARGET_EXPR && TREE_OPERAND (*tp, 3)) |
| 810 | { |
| 811 | /* The copied TARGET_EXPR has never been expanded, even if the |
| 812 | original node was expanded already. */ |
| 813 | TREE_OPERAND (*tp, 1) = TREE_OPERAND (*tp, 3); |
| 814 | TREE_OPERAND (*tp, 3) = NULL_TREE; |
| 815 | } |
| 816 | else if (TREE_CODE (*tp) == ADDR_EXPR) |
| 817 | { |
| 818 | /* Variable substitution need not be simple. In particular, |
| 819 | the INDIRECT_REF substitution above. Make sure that |
| 820 | TREE_CONSTANT and friends are up-to-date. But make sure |
| 821 | to not improperly set TREE_BLOCK on some sub-expressions. */ |
| 822 | int invariant = is_gimple_min_invariant (*tp); |
| 823 | tree block = id->block; |
| 824 | id->block = NULL_TREE; |
| 825 | walk_tree (&TREE_OPERAND (*tp, 0), copy_tree_body_r, id, NULL); |
| 826 | id->block = block; |
| 827 | |
| 828 | /* Handle the case where we substituted an INDIRECT_REF |
| 829 | into the operand of the ADDR_EXPR. */ |
| 830 | if (TREE_CODE (TREE_OPERAND (*tp, 0)) == INDIRECT_REF) |
| 831 | *tp = TREE_OPERAND (TREE_OPERAND (*tp, 0), 0); |
| 832 | else |
| 833 | recompute_tree_invariant_for_addr_expr (*tp); |
| 834 | |
| 835 | /* If this used to be invariant, but is not any longer, |
| 836 | then regimplification is probably needed. */ |
| 837 | if (invariant && !is_gimple_min_invariant (*tp)) |
| 838 | id->regimplify = true; |
| 839 | |
| 840 | *walk_subtrees = 0; |
| 841 | } |
| 842 | } |
| 843 | |
| 844 | /* Keep iterating. */ |
| 845 | return NULL_TREE; |
| 846 | } |
| 847 | |
| 848 | |
| 849 | /* Called from copy_body_id via walk_tree. DATA is really a |
| 850 | `copy_body_data *'. */ |
| 851 | |
| 852 | tree |
| 853 | copy_tree_body_r (tree *tp, int *walk_subtrees, void *data) |
| 854 | { |
| 855 | copy_body_data *id = (copy_body_data *) data; |
| 856 | tree fn = id->src_fn; |
| 857 | tree new_block; |
| 858 | |
| 859 | /* Begin by recognizing trees that we'll completely rewrite for the |
| 860 | inlining context. Our output for these trees is completely |
| 861 | different from out input (e.g. RETURN_EXPR is deleted, and morphs |
| 862 | into an edge). Further down, we'll handle trees that get |
| 863 | duplicated and/or tweaked. */ |
| 864 | |
| 865 | /* When requested, RETURN_EXPRs should be transformed to just the |
| 866 | contained MODIFY_EXPR. The branch semantics of the return will |
| 867 | be handled elsewhere by manipulating the CFG rather than a statement. */ |
| 868 | if (TREE_CODE (*tp) == RETURN_EXPR && id->transform_return_to_modify) |
| 869 | { |
| 870 | tree assignment = TREE_OPERAND (*tp, 0); |
| 871 | |
| 872 | /* If we're returning something, just turn that into an |
| 873 | assignment into the equivalent of the original RESULT_DECL. |
| 874 | If the "assignment" is just the result decl, the result |
| 875 | decl has already been set (e.g. a recent "foo (&result_decl, |
| 876 | ...)"); just toss the entire RETURN_EXPR. */ |
| 877 | if (assignment && TREE_CODE (assignment) == MODIFY_EXPR) |
| 878 | { |
| 879 | /* Replace the RETURN_EXPR with (a copy of) the |
| 880 | MODIFY_EXPR hanging underneath. */ |
| 881 | *tp = copy_node (assignment); |
| 882 | } |
| 883 | else /* Else the RETURN_EXPR returns no value. */ |
| 884 | { |
| 885 | *tp = NULL; |
| 886 | return (tree) (void *)1; |
| 887 | } |
| 888 | } |
| 889 | else if (TREE_CODE (*tp) == SSA_NAME) |
| 890 | { |
| 891 | *tp = remap_ssa_name (*tp, id); |
| 892 | *walk_subtrees = 0; |
| 893 | return NULL; |
| 894 | } |
| 895 | |
| 896 | /* Local variables and labels need to be replaced by equivalent |
| 897 | variables. We don't want to copy static variables; there's only |
| 898 | one of those, no matter how many times we inline the containing |
| 899 | function. Similarly for globals from an outer function. */ |
| 900 | else if (auto_var_in_fn_p (*tp, fn)) |
| 901 | { |
| 902 | tree new_decl; |
| 903 | |
| 904 | /* Remap the declaration. */ |
| 905 | new_decl = remap_decl (*tp, id); |
| 906 | gcc_assert (new_decl); |
| 907 | /* Replace this variable with the copy. */ |
| 908 | STRIP_TYPE_NOPS (new_decl); |
| 909 | *tp = new_decl; |
| 910 | *walk_subtrees = 0; |
| 911 | } |
| 912 | else if (TREE_CODE (*tp) == STATEMENT_LIST) |
| 913 | copy_statement_list (tp); |
| 914 | else if (TREE_CODE (*tp) == SAVE_EXPR) |
| 915 | remap_save_expr (tp, id->decl_map, walk_subtrees); |
| 916 | else if (TREE_CODE (*tp) == LABEL_DECL |
| 917 | && (! DECL_CONTEXT (*tp) |
| 918 | || decl_function_context (*tp) == id->src_fn)) |
| 919 | /* These may need to be remapped for EH handling. */ |
| 920 | *tp = remap_decl (*tp, id); |
| 921 | else if (TREE_CODE (*tp) == BIND_EXPR) |
| 922 | copy_bind_expr (tp, walk_subtrees, id); |
| 923 | /* Types may need remapping as well. */ |
| 924 | else if (TYPE_P (*tp)) |
| 925 | *tp = remap_type (*tp, id); |
| 926 | |
| 927 | /* If this is a constant, we have to copy the node iff the type will be |
| 928 | remapped. copy_tree_r will not copy a constant. */ |
| 929 | else if (CONSTANT_CLASS_P (*tp)) |
| 930 | { |
| 931 | tree new_type = remap_type (TREE_TYPE (*tp), id); |
| 932 | |
| 933 | if (new_type == TREE_TYPE (*tp)) |
| 934 | *walk_subtrees = 0; |
| 935 | |
| 936 | else if (TREE_CODE (*tp) == INTEGER_CST) |
| 937 | *tp = build_int_cst_wide (new_type, TREE_INT_CST_LOW (*tp), |
| 938 | TREE_INT_CST_HIGH (*tp)); |
| 939 | else |
| 940 | { |
| 941 | *tp = copy_node (*tp); |
| 942 | TREE_TYPE (*tp) = new_type; |
| 943 | } |
| 944 | } |
| 945 | |
| 946 | /* Otherwise, just copy the node. Note that copy_tree_r already |
| 947 | knows not to copy VAR_DECLs, etc., so this is safe. */ |
| 948 | else |
| 949 | { |
| 950 | /* Here we handle trees that are not completely rewritten. |
| 951 | First we detect some inlining-induced bogosities for |
| 952 | discarding. */ |
| 953 | if (TREE_CODE (*tp) == MODIFY_EXPR |
| 954 | && TREE_OPERAND (*tp, 0) == TREE_OPERAND (*tp, 1) |
| 955 | && (auto_var_in_fn_p (TREE_OPERAND (*tp, 0), fn))) |
| 956 | { |
| 957 | /* Some assignments VAR = VAR; don't generate any rtl code |
| 958 | and thus don't count as variable modification. Avoid |
| 959 | keeping bogosities like 0 = 0. */ |
| 960 | tree decl = TREE_OPERAND (*tp, 0), value; |
| 961 | tree *n; |
| 962 | |
| 963 | n = (tree *) pointer_map_contains (id->decl_map, decl); |
| 964 | if (n) |
| 965 | { |
| 966 | value = *n; |
| 967 | STRIP_TYPE_NOPS (value); |
| 968 | if (TREE_CONSTANT (value) || TREE_READONLY (value)) |
| 969 | { |
| 970 | *tp = build_empty_stmt (); |
| 971 | return copy_tree_body_r (tp, walk_subtrees, data); |
| 972 | } |
| 973 | } |
| 974 | } |
| 975 | else if (TREE_CODE (*tp) == INDIRECT_REF) |
| 976 | { |
| 977 | /* Get rid of *& from inline substitutions that can happen when a |
| 978 | pointer argument is an ADDR_EXPR. */ |
| 979 | tree decl = TREE_OPERAND (*tp, 0); |
| 980 | tree *n; |
| 981 | |
| 982 | n = (tree *) pointer_map_contains (id->decl_map, decl); |
| 983 | if (n) |
| 984 | { |
| 985 | tree new_tree; |
| 986 | tree old; |
| 987 | /* If we happen to get an ADDR_EXPR in n->value, strip |
| 988 | it manually here as we'll eventually get ADDR_EXPRs |
| 989 | which lie about their types pointed to. In this case |
| 990 | build_fold_indirect_ref wouldn't strip the INDIRECT_REF, |
| 991 | but we absolutely rely on that. As fold_indirect_ref |
| 992 | does other useful transformations, try that first, though. */ |
| 993 | tree type = TREE_TYPE (TREE_TYPE (*n)); |
| 994 | new_tree = unshare_expr (*n); |
| 995 | old = *tp; |
| 996 | *tp = gimple_fold_indirect_ref (new_tree); |
| 997 | if (! *tp) |
| 998 | { |
| 999 | if (TREE_CODE (new_tree) == ADDR_EXPR) |
| 1000 | { |
| 1001 | *tp = fold_indirect_ref_1 (type, new_tree); |
| 1002 | /* ??? We should either assert here or build |
| 1003 | a VIEW_CONVERT_EXPR instead of blindly leaking |
| 1004 | incompatible types to our IL. */ |
| 1005 | if (! *tp) |
| 1006 | *tp = TREE_OPERAND (new_tree, 0); |
| 1007 | } |
| 1008 | else |
| 1009 | { |
| 1010 | *tp = build1 (INDIRECT_REF, type, new_tree); |
| 1011 | TREE_THIS_VOLATILE (*tp) = TREE_THIS_VOLATILE (old); |
| 1012 | TREE_SIDE_EFFECTS (*tp) = TREE_SIDE_EFFECTS (old); |
| 1013 | } |
| 1014 | } |
| 1015 | *walk_subtrees = 0; |
| 1016 | return NULL; |
| 1017 | } |
| 1018 | } |
| 1019 | |
| 1020 | /* Here is the "usual case". Copy this tree node, and then |
| 1021 | tweak some special cases. */ |
| 1022 | copy_tree_r (tp, walk_subtrees, NULL); |
| 1023 | |
| 1024 | /* Global variables we haven't seen yet needs to go into referenced |
| 1025 | vars. If not referenced from types only. */ |
| 1026 | if (gimple_in_ssa_p (cfun) |
| 1027 | && TREE_CODE (*tp) == VAR_DECL |
| 1028 | && id->remapping_type_depth == 0) |
| 1029 | add_referenced_var (*tp); |
| 1030 | |
| 1031 | /* If EXPR has block defined, map it to newly constructed block. |
| 1032 | When inlining we want EXPRs without block appear in the block |
| 1033 | of function call. */ |
| 1034 | if (EXPR_P (*tp)) |
| 1035 | { |
| 1036 | new_block = id->block; |
| 1037 | if (TREE_BLOCK (*tp)) |
| 1038 | { |
| 1039 | tree *n; |
| 1040 | n = (tree *) pointer_map_contains (id->decl_map, |
| 1041 | TREE_BLOCK (*tp)); |
| 1042 | gcc_assert (n); |
| 1043 | new_block = *n; |
| 1044 | } |
| 1045 | TREE_BLOCK (*tp) = new_block; |
| 1046 | } |
| 1047 | |
| 1048 | if (TREE_CODE (*tp) == RESX_EXPR && id->eh_region_offset) |
| 1049 | TREE_OPERAND (*tp, 0) = |
| 1050 | build_int_cst (NULL_TREE, |
| 1051 | id->eh_region_offset |
| 1052 | + TREE_INT_CST_LOW (TREE_OPERAND (*tp, 0))); |
| 1053 | |
| 1054 | if (TREE_CODE (*tp) != OMP_CLAUSE) |
| 1055 | TREE_TYPE (*tp) = remap_type (TREE_TYPE (*tp), id); |
| 1056 | |
| 1057 | /* The copied TARGET_EXPR has never been expanded, even if the |
| 1058 | original node was expanded already. */ |
| 1059 | if (TREE_CODE (*tp) == TARGET_EXPR && TREE_OPERAND (*tp, 3)) |
| 1060 | { |
| 1061 | TREE_OPERAND (*tp, 1) = TREE_OPERAND (*tp, 3); |
| 1062 | TREE_OPERAND (*tp, 3) = NULL_TREE; |
| 1063 | } |
| 1064 | |
| 1065 | /* Variable substitution need not be simple. In particular, the |
| 1066 | INDIRECT_REF substitution above. Make sure that TREE_CONSTANT |
| 1067 | and friends are up-to-date. */ |
| 1068 | else if (TREE_CODE (*tp) == ADDR_EXPR) |
| 1069 | { |
| 1070 | int invariant = is_gimple_min_invariant (*tp); |
| 1071 | walk_tree (&TREE_OPERAND (*tp, 0), copy_tree_body_r, id, NULL); |
| 1072 | |
| 1073 | /* Handle the case where we substituted an INDIRECT_REF |
| 1074 | into the operand of the ADDR_EXPR. */ |
| 1075 | if (TREE_CODE (TREE_OPERAND (*tp, 0)) == INDIRECT_REF) |
| 1076 | *tp = TREE_OPERAND (TREE_OPERAND (*tp, 0), 0); |
| 1077 | else |
| 1078 | recompute_tree_invariant_for_addr_expr (*tp); |
| 1079 | |
| 1080 | /* If this used to be invariant, but is not any longer, |
| 1081 | then regimplification is probably needed. */ |
| 1082 | if (invariant && !is_gimple_min_invariant (*tp)) |
| 1083 | id->regimplify = true; |
| 1084 | |
| 1085 | *walk_subtrees = 0; |
| 1086 | } |
| 1087 | } |
| 1088 | |
| 1089 | /* Keep iterating. */ |
| 1090 | return NULL_TREE; |
| 1091 | } |
| 1092 | |
| 1093 | |
| 1094 | /* Helper for copy_bb. Remap statement STMT using the inlining |
| 1095 | information in ID. Return the new statement copy. */ |
| 1096 | |
| 1097 | static gimple |
| 1098 | remap_gimple_stmt (gimple stmt, copy_body_data *id) |
| 1099 | { |
| 1100 | gimple copy = NULL; |
| 1101 | struct walk_stmt_info wi; |
| 1102 | tree new_block; |
| 1103 | bool skip_first = false; |
| 1104 | |
| 1105 | /* Begin by recognizing trees that we'll completely rewrite for the |
| 1106 | inlining context. Our output for these trees is completely |
| 1107 | different from out input (e.g. RETURN_EXPR is deleted, and morphs |
| 1108 | into an edge). Further down, we'll handle trees that get |
| 1109 | duplicated and/or tweaked. */ |
| 1110 | |
| 1111 | /* When requested, GIMPLE_RETURNs should be transformed to just the |
| 1112 | contained GIMPLE_ASSIGN. The branch semantics of the return will |
| 1113 | be handled elsewhere by manipulating the CFG rather than the |
| 1114 | statement. */ |
| 1115 | if (gimple_code (stmt) == GIMPLE_RETURN && id->transform_return_to_modify) |
| 1116 | { |
| 1117 | tree retval = gimple_return_retval (stmt); |
| 1118 | |
| 1119 | /* If we're returning something, just turn that into an |
| 1120 | assignment into the equivalent of the original RESULT_DECL. |
| 1121 | If RETVAL is just the result decl, the result decl has |
| 1122 | already been set (e.g. a recent "foo (&result_decl, ...)"); |
| 1123 | just toss the entire GIMPLE_RETURN. */ |
| 1124 | if (retval && TREE_CODE (retval) != RESULT_DECL) |
| 1125 | { |
| 1126 | copy = gimple_build_assign (id->retvar, retval); |
| 1127 | /* id->retvar is already substituted. Skip it on later remapping. */ |
| 1128 | skip_first = true; |
| 1129 | } |
| 1130 | else |
| 1131 | return gimple_build_nop (); |
| 1132 | } |
| 1133 | else if (gimple_has_substatements (stmt)) |
| 1134 | { |
| 1135 | gimple_seq s1, s2; |
| 1136 | |
| 1137 | /* When cloning bodies from the C++ front end, we will be handed bodies |
| 1138 | in High GIMPLE form. Handle here all the High GIMPLE statements that |
| 1139 | have embedded statements. */ |
| 1140 | switch (gimple_code (stmt)) |
| 1141 | { |
| 1142 | case GIMPLE_BIND: |
| 1143 | copy = copy_gimple_bind (stmt, id); |
| 1144 | break; |
| 1145 | |
| 1146 | case GIMPLE_CATCH: |
| 1147 | s1 = remap_gimple_seq (gimple_catch_handler (stmt), id); |
| 1148 | copy = gimple_build_catch (gimple_catch_types (stmt), s1); |
| 1149 | break; |
| 1150 | |
| 1151 | case GIMPLE_EH_FILTER: |
| 1152 | s1 = remap_gimple_seq (gimple_eh_filter_failure (stmt), id); |
| 1153 | copy = gimple_build_eh_filter (gimple_eh_filter_types (stmt), s1); |
| 1154 | break; |
| 1155 | |
| 1156 | case GIMPLE_TRY: |
| 1157 | s1 = remap_gimple_seq (gimple_try_eval (stmt), id); |
| 1158 | s2 = remap_gimple_seq (gimple_try_cleanup (stmt), id); |
| 1159 | copy = gimple_build_try (s1, s2, gimple_try_kind (stmt)); |
| 1160 | break; |
| 1161 | |
| 1162 | case GIMPLE_WITH_CLEANUP_EXPR: |
| 1163 | s1 = remap_gimple_seq (gimple_wce_cleanup (stmt), id); |
| 1164 | copy = gimple_build_wce (s1); |
| 1165 | break; |
| 1166 | |
| 1167 | case GIMPLE_OMP_PARALLEL: |
| 1168 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1169 | copy = gimple_build_omp_parallel |
| 1170 | (s1, |
| 1171 | gimple_omp_parallel_clauses (stmt), |
| 1172 | gimple_omp_parallel_child_fn (stmt), |
| 1173 | gimple_omp_parallel_data_arg (stmt)); |
| 1174 | break; |
| 1175 | |
| 1176 | case GIMPLE_OMP_TASK: |
| 1177 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1178 | copy = gimple_build_omp_task |
| 1179 | (s1, |
| 1180 | gimple_omp_task_clauses (stmt), |
| 1181 | gimple_omp_task_child_fn (stmt), |
| 1182 | gimple_omp_task_data_arg (stmt), |
| 1183 | gimple_omp_task_copy_fn (stmt), |
| 1184 | gimple_omp_task_arg_size (stmt), |
| 1185 | gimple_omp_task_arg_align (stmt)); |
| 1186 | break; |
| 1187 | |
| 1188 | case GIMPLE_OMP_FOR: |
| 1189 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1190 | s2 = remap_gimple_seq (gimple_omp_for_pre_body (stmt), id); |
| 1191 | copy = gimple_build_omp_for (s1, gimple_omp_for_clauses (stmt), |
| 1192 | gimple_omp_for_collapse (stmt), s2); |
| 1193 | { |
| 1194 | size_t i; |
| 1195 | for (i = 0; i < gimple_omp_for_collapse (stmt); i++) |
| 1196 | { |
| 1197 | gimple_omp_for_set_index (copy, i, |
| 1198 | gimple_omp_for_index (stmt, i)); |
| 1199 | gimple_omp_for_set_initial (copy, i, |
| 1200 | gimple_omp_for_initial (stmt, i)); |
| 1201 | gimple_omp_for_set_final (copy, i, |
| 1202 | gimple_omp_for_final (stmt, i)); |
| 1203 | gimple_omp_for_set_incr (copy, i, |
| 1204 | gimple_omp_for_incr (stmt, i)); |
| 1205 | gimple_omp_for_set_cond (copy, i, |
| 1206 | gimple_omp_for_cond (stmt, i)); |
| 1207 | } |
| 1208 | } |
| 1209 | break; |
| 1210 | |
| 1211 | case GIMPLE_OMP_MASTER: |
| 1212 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1213 | copy = gimple_build_omp_master (s1); |
| 1214 | break; |
| 1215 | |
| 1216 | case GIMPLE_OMP_ORDERED: |
| 1217 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1218 | copy = gimple_build_omp_ordered (s1); |
| 1219 | break; |
| 1220 | |
| 1221 | case GIMPLE_OMP_SECTION: |
| 1222 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1223 | copy = gimple_build_omp_section (s1); |
| 1224 | break; |
| 1225 | |
| 1226 | case GIMPLE_OMP_SECTIONS: |
| 1227 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1228 | copy = gimple_build_omp_sections |
| 1229 | (s1, gimple_omp_sections_clauses (stmt)); |
| 1230 | break; |
| 1231 | |
| 1232 | case GIMPLE_OMP_SINGLE: |
| 1233 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1234 | copy = gimple_build_omp_single |
| 1235 | (s1, gimple_omp_single_clauses (stmt)); |
| 1236 | break; |
| 1237 | |
| 1238 | case GIMPLE_OMP_CRITICAL: |
| 1239 | s1 = remap_gimple_seq (gimple_omp_body (stmt), id); |
| 1240 | copy |
| 1241 | = gimple_build_omp_critical (s1, gimple_omp_critical_name (stmt)); |
| 1242 | break; |
| 1243 | |
| 1244 | default: |
| 1245 | gcc_unreachable (); |
| 1246 | } |
| 1247 | } |
| 1248 | else |
| 1249 | { |
| 1250 | if (gimple_assign_copy_p (stmt) |
| 1251 | && gimple_assign_lhs (stmt) == gimple_assign_rhs1 (stmt) |
| 1252 | && auto_var_in_fn_p (gimple_assign_lhs (stmt), id->src_fn)) |
| 1253 | { |
| 1254 | /* Here we handle statements that are not completely rewritten. |
| 1255 | First we detect some inlining-induced bogosities for |
| 1256 | discarding. */ |
| 1257 | |
| 1258 | /* Some assignments VAR = VAR; don't generate any rtl code |
| 1259 | and thus don't count as variable modification. Avoid |
| 1260 | keeping bogosities like 0 = 0. */ |
| 1261 | tree decl = gimple_assign_lhs (stmt), value; |
| 1262 | tree *n; |
| 1263 | |
| 1264 | n = (tree *) pointer_map_contains (id->decl_map, decl); |
| 1265 | if (n) |
| 1266 | { |
| 1267 | value = *n; |
| 1268 | STRIP_TYPE_NOPS (value); |
| 1269 | if (TREE_CONSTANT (value) || TREE_READONLY (value)) |
| 1270 | return gimple_build_nop (); |
| 1271 | } |
| 1272 | } |
| 1273 | |
| 1274 | /* Create a new deep copy of the statement. */ |
| 1275 | copy = gimple_copy (stmt); |
| 1276 | } |
| 1277 | |
| 1278 | /* If STMT has a block defined, map it to the newly constructed |
| 1279 | block. When inlining we want statements without a block to |
| 1280 | appear in the block of the function call. */ |
| 1281 | new_block = id->block; |
| 1282 | if (gimple_block (copy)) |
| 1283 | { |
| 1284 | tree *n; |
| 1285 | n = (tree *) pointer_map_contains (id->decl_map, gimple_block (copy)); |
| 1286 | gcc_assert (n); |
| 1287 | new_block = *n; |
| 1288 | } |
| 1289 | |
| 1290 | gimple_set_block (copy, new_block); |
| 1291 | |
| 1292 | /* Remap all the operands in COPY. */ |
| 1293 | memset (&wi, 0, sizeof (wi)); |
| 1294 | wi.info = id; |
| 1295 | if (skip_first) |
| 1296 | walk_tree (gimple_op_ptr (copy, 1), remap_gimple_op_r, &wi, NULL); |
| 1297 | else |
| 1298 | walk_gimple_op (copy, remap_gimple_op_r, &wi); |
| 1299 | |
| 1300 | /* We have to handle EH region remapping of GIMPLE_RESX specially because |
| 1301 | the region number is not an operand. */ |
| 1302 | if (gimple_code (stmt) == GIMPLE_RESX && id->eh_region_offset) |
| 1303 | { |
| 1304 | gimple_resx_set_region (copy, gimple_resx_region (stmt) + id->eh_region_offset); |
| 1305 | } |
| 1306 | return copy; |
| 1307 | } |
| 1308 | |
| 1309 | |
| 1310 | /* Copy basic block, scale profile accordingly. Edges will be taken care of |
| 1311 | later */ |
| 1312 | |
| 1313 | static basic_block |
| 1314 | copy_bb (copy_body_data *id, basic_block bb, int frequency_scale, |
| 1315 | gcov_type count_scale) |
| 1316 | { |
| 1317 | gimple_stmt_iterator gsi, copy_gsi, seq_gsi; |
| 1318 | basic_block copy_basic_block; |
| 1319 | tree decl; |
| 1320 | |
| 1321 | /* create_basic_block() will append every new block to |
| 1322 | basic_block_info automatically. */ |
| 1323 | copy_basic_block = create_basic_block (NULL, (void *) 0, |
| 1324 | (basic_block) bb->prev_bb->aux); |
| 1325 | copy_basic_block->count = bb->count * count_scale / REG_BR_PROB_BASE; |
| 1326 | |
| 1327 | /* We are going to rebuild frequencies from scratch. These values |
| 1328 | have just small importance to drive canonicalize_loop_headers. */ |
| 1329 | copy_basic_block->frequency = ((gcov_type)bb->frequency |
| 1330 | * frequency_scale / REG_BR_PROB_BASE); |
| 1331 | |
| 1332 | if (copy_basic_block->frequency > BB_FREQ_MAX) |
| 1333 | copy_basic_block->frequency = BB_FREQ_MAX; |
| 1334 | |
| 1335 | copy_gsi = gsi_start_bb (copy_basic_block); |
| 1336 | |
| 1337 | for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) |
| 1338 | { |
| 1339 | gimple stmt = gsi_stmt (gsi); |
| 1340 | gimple orig_stmt = stmt; |
| 1341 | |
| 1342 | id->regimplify = false; |
| 1343 | stmt = remap_gimple_stmt (stmt, id); |
| 1344 | if (gimple_nop_p (stmt)) |
| 1345 | continue; |
| 1346 | |
| 1347 | gimple_duplicate_stmt_histograms (cfun, stmt, id->src_cfun, orig_stmt); |
| 1348 | seq_gsi = copy_gsi; |
| 1349 | |
| 1350 | /* With return slot optimization we can end up with |
| 1351 | non-gimple (foo *)&this->m, fix that here. */ |
| 1352 | if (is_gimple_assign (stmt) |
| 1353 | && gimple_assign_rhs_code (stmt) == NOP_EXPR |
| 1354 | && !is_gimple_val (gimple_assign_rhs1 (stmt))) |
| 1355 | { |
| 1356 | tree new_rhs; |
| 1357 | new_rhs = force_gimple_operand_gsi (&seq_gsi, |
| 1358 | gimple_assign_rhs1 (stmt), |
| 1359 | true, NULL, false, GSI_NEW_STMT); |
| 1360 | gimple_assign_set_rhs1 (stmt, new_rhs); |
| 1361 | id->regimplify = false; |
| 1362 | } |
| 1363 | |
| 1364 | gsi_insert_after (&seq_gsi, stmt, GSI_NEW_STMT); |
| 1365 | |
| 1366 | if (id->regimplify) |
| 1367 | gimple_regimplify_operands (stmt, &seq_gsi); |
| 1368 | |
| 1369 | /* If copy_basic_block has been empty at the start of this iteration, |
| 1370 | call gsi_start_bb again to get at the newly added statements. */ |
| 1371 | if (gsi_end_p (copy_gsi)) |
| 1372 | copy_gsi = gsi_start_bb (copy_basic_block); |
| 1373 | else |
| 1374 | gsi_next (©_gsi); |
| 1375 | |
| 1376 | /* Process the new statement. The call to gimple_regimplify_operands |
| 1377 | possibly turned the statement into multiple statements, we |
| 1378 | need to process all of them. */ |
| 1379 | do |
| 1380 | { |
| 1381 | stmt = gsi_stmt (copy_gsi); |
| 1382 | if (is_gimple_call (stmt) |
| 1383 | && gimple_call_va_arg_pack_p (stmt) |
| 1384 | && id->gimple_call) |
| 1385 | { |
| 1386 | /* __builtin_va_arg_pack () should be replaced by |
| 1387 | all arguments corresponding to ... in the caller. */ |
| 1388 | tree p; |
| 1389 | gimple new_call; |
| 1390 | VEC(tree, heap) *argarray; |
| 1391 | size_t nargs = gimple_call_num_args (id->gimple_call); |
| 1392 | size_t n; |
| 1393 | |
| 1394 | for (p = DECL_ARGUMENTS (id->src_fn); p; p = TREE_CHAIN (p)) |
| 1395 | nargs--; |
| 1396 | |
| 1397 | /* Create the new array of arguments. */ |
| 1398 | n = nargs + gimple_call_num_args (stmt); |
| 1399 | argarray = VEC_alloc (tree, heap, n); |
| 1400 | VEC_safe_grow (tree, heap, argarray, n); |
| 1401 | |
| 1402 | /* Copy all the arguments before '...' */ |
| 1403 | memcpy (VEC_address (tree, argarray), |
| 1404 | gimple_call_arg_ptr (stmt, 0), |
| 1405 | gimple_call_num_args (stmt) * sizeof (tree)); |
| 1406 | |
| 1407 | /* Append the arguments passed in '...' */ |
| 1408 | memcpy (VEC_address(tree, argarray) + gimple_call_num_args (stmt), |
| 1409 | gimple_call_arg_ptr (id->gimple_call, 0) |
| 1410 | + (gimple_call_num_args (id->gimple_call) - nargs), |
| 1411 | nargs * sizeof (tree)); |
| 1412 | |
| 1413 | new_call = gimple_build_call_vec (gimple_call_fn (stmt), |
| 1414 | argarray); |
| 1415 | |
| 1416 | VEC_free (tree, heap, argarray); |
| 1417 | |
| 1418 | /* Copy all GIMPLE_CALL flags, location and block, except |
| 1419 | GF_CALL_VA_ARG_PACK. */ |
| 1420 | gimple_call_copy_flags (new_call, stmt); |
| 1421 | gimple_call_set_va_arg_pack (new_call, false); |
| 1422 | gimple_set_location (new_call, gimple_location (stmt)); |
| 1423 | gimple_set_block (new_call, gimple_block (stmt)); |
| 1424 | gimple_call_set_lhs (new_call, gimple_call_lhs (stmt)); |
| 1425 | |
| 1426 | gsi_replace (©_gsi, new_call, false); |
| 1427 | gimple_set_bb (stmt, NULL); |
| 1428 | stmt = new_call; |
| 1429 | } |
| 1430 | else if (is_gimple_call (stmt) |
| 1431 | && id->gimple_call |
| 1432 | && (decl = gimple_call_fndecl (stmt)) |
| 1433 | && DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL |
| 1434 | && DECL_FUNCTION_CODE (decl) == BUILT_IN_VA_ARG_PACK_LEN) |
| 1435 | { |
| 1436 | /* __builtin_va_arg_pack_len () should be replaced by |
| 1437 | the number of anonymous arguments. */ |
| 1438 | size_t nargs = gimple_call_num_args (id->gimple_call); |
| 1439 | tree count, p; |
| 1440 | gimple new_stmt; |
| 1441 | |
| 1442 | for (p = DECL_ARGUMENTS (id->src_fn); p; p = TREE_CHAIN (p)) |
| 1443 | nargs--; |
| 1444 | |
| 1445 | count = build_int_cst (integer_type_node, nargs); |
| 1446 | new_stmt = gimple_build_assign (gimple_call_lhs (stmt), count); |
| 1447 | gsi_replace (©_gsi, new_stmt, false); |
| 1448 | stmt = new_stmt; |
| 1449 | } |
| 1450 | |
| 1451 | /* Statements produced by inlining can be unfolded, especially |
| 1452 | when we constant propagated some operands. We can't fold |
| 1453 | them right now for two reasons: |
| 1454 | 1) folding require SSA_NAME_DEF_STMTs to be correct |
| 1455 | 2) we can't change function calls to builtins. |
| 1456 | So we just mark statement for later folding. We mark |
| 1457 | all new statements, instead just statements that has changed |
| 1458 | by some nontrivial substitution so even statements made |
| 1459 | foldable indirectly are updated. If this turns out to be |
| 1460 | expensive, copy_body can be told to watch for nontrivial |
| 1461 | changes. */ |
| 1462 | if (id->statements_to_fold) |
| 1463 | pointer_set_insert (id->statements_to_fold, stmt); |
| 1464 | |
| 1465 | /* We're duplicating a CALL_EXPR. Find any corresponding |
| 1466 | callgraph edges and update or duplicate them. */ |
| 1467 | if (is_gimple_call (stmt)) |
| 1468 | { |
| 1469 | struct cgraph_node *node; |
| 1470 | struct cgraph_edge *edge; |
| 1471 | int flags; |
| 1472 | |
| 1473 | switch (id->transform_call_graph_edges) |
| 1474 | { |
| 1475 | case CB_CGE_DUPLICATE: |
| 1476 | edge = cgraph_edge (id->src_node, orig_stmt); |
| 1477 | if (edge) |
| 1478 | cgraph_clone_edge (edge, id->dst_node, stmt, |
| 1479 | REG_BR_PROB_BASE, 1, |
| 1480 | edge->frequency, true); |
| 1481 | break; |
| 1482 | |
| 1483 | case CB_CGE_MOVE_CLONES: |
| 1484 | for (node = id->dst_node->next_clone; |
| 1485 | node; |
| 1486 | node = node->next_clone) |
| 1487 | { |
| 1488 | edge = cgraph_edge (node, orig_stmt); |
| 1489 | if (edge) |
| 1490 | cgraph_set_call_stmt (edge, stmt); |
| 1491 | } |
| 1492 | /* FALLTHRU */ |
| 1493 | |
| 1494 | case CB_CGE_MOVE: |
| 1495 | edge = cgraph_edge (id->dst_node, orig_stmt); |
| 1496 | if (edge) |
| 1497 | cgraph_set_call_stmt (edge, stmt); |
| 1498 | break; |
| 1499 | |
| 1500 | default: |
| 1501 | gcc_unreachable (); |
| 1502 | } |
| 1503 | |
| 1504 | flags = gimple_call_flags (stmt); |
| 1505 | |
| 1506 | if (flags & ECF_MAY_BE_ALLOCA) |
| 1507 | cfun->calls_alloca = true; |
| 1508 | if (flags & ECF_RETURNS_TWICE) |
| 1509 | cfun->calls_setjmp = true; |
| 1510 | } |
| 1511 | |
| 1512 | /* If you think we can abort here, you are wrong. |
| 1513 | There is no region 0 in gimple. */ |
| 1514 | gcc_assert (lookup_stmt_eh_region_fn (id->src_cfun, orig_stmt) != 0); |
| 1515 | |
| 1516 | if (stmt_could_throw_p (stmt) |
| 1517 | /* When we are cloning for inlining, we are supposed to |
| 1518 | construct a clone that calls precisely the same functions |
| 1519 | as original. However IPA optimizers might've proved |
| 1520 | earlier some function calls as non-trapping that might |
| 1521 | render some basic blocks dead that might become |
| 1522 | unreachable. |
| 1523 | |
| 1524 | We can't update SSA with unreachable blocks in CFG and thus |
| 1525 | we prevent the scenario by preserving even the "dead" eh |
| 1526 | edges until the point they are later removed by |
| 1527 | fixup_cfg pass. */ |
| 1528 | || (id->transform_call_graph_edges == CB_CGE_MOVE_CLONES |
| 1529 | && lookup_stmt_eh_region_fn (id->src_cfun, orig_stmt) > 0)) |
| 1530 | { |
| 1531 | int region = lookup_stmt_eh_region_fn (id->src_cfun, orig_stmt); |
| 1532 | |
| 1533 | /* Add an entry for the copied tree in the EH hashtable. |
| 1534 | When cloning or versioning, use the hashtable in |
| 1535 | cfun, and just copy the EH number. When inlining, use the |
| 1536 | hashtable in the caller, and adjust the region number. */ |
| 1537 | if (region > 0) |
| 1538 | add_stmt_to_eh_region (stmt, region + id->eh_region_offset); |
| 1539 | |
| 1540 | /* If this tree doesn't have a region associated with it, |
| 1541 | and there is a "current region," |
| 1542 | then associate this tree with the current region |
| 1543 | and add edges associated with this region. */ |
| 1544 | if (lookup_stmt_eh_region_fn (id->src_cfun, orig_stmt) <= 0 |
| 1545 | && id->eh_region > 0 |
| 1546 | && stmt_could_throw_p (stmt)) |
| 1547 | add_stmt_to_eh_region (stmt, id->eh_region); |
| 1548 | } |
| 1549 | |
| 1550 | if (gimple_in_ssa_p (cfun)) |
| 1551 | { |
| 1552 | ssa_op_iter i; |
| 1553 | tree def; |
| 1554 | |
| 1555 | find_new_referenced_vars (gsi_stmt (copy_gsi)); |
| 1556 | FOR_EACH_SSA_TREE_OPERAND (def, stmt, i, SSA_OP_DEF) |
| 1557 | if (TREE_CODE (def) == SSA_NAME) |
| 1558 | SSA_NAME_DEF_STMT (def) = stmt; |
| 1559 | } |
| 1560 | |
| 1561 | gsi_next (©_gsi); |
| 1562 | } |
| 1563 | while (!gsi_end_p (copy_gsi)); |
| 1564 | |
| 1565 | copy_gsi = gsi_last_bb (copy_basic_block); |
| 1566 | } |
| 1567 | |
| 1568 | return copy_basic_block; |
| 1569 | } |
| 1570 | |
| 1571 | /* Inserting Single Entry Multiple Exit region in SSA form into code in SSA |
| 1572 | form is quite easy, since dominator relationship for old basic blocks does |
| 1573 | not change. |
| 1574 | |
| 1575 | There is however exception where inlining might change dominator relation |
| 1576 | across EH edges from basic block within inlined functions destinating |
| 1577 | to landing pads in function we inline into. |
| 1578 | |
| 1579 | The function fills in PHI_RESULTs of such PHI nodes if they refer |
| 1580 | to gimple regs. Otherwise, the function mark PHI_RESULT of such |
| 1581 | PHI nodes for renaming. For non-gimple regs, renaming is safe: the |
| 1582 | EH edges are abnormal and SSA_NAME_OCCURS_IN_ABNORMAL_PHI must be |
| 1583 | set, and this means that there will be no overlapping live ranges |
| 1584 | for the underlying symbol. |
| 1585 | |
| 1586 | This might change in future if we allow redirecting of EH edges and |
| 1587 | we might want to change way build CFG pre-inlining to include |
| 1588 | all the possible edges then. */ |
| 1589 | static void |
| 1590 | update_ssa_across_abnormal_edges (basic_block bb, basic_block ret_bb, |
| 1591 | bool can_throw, bool nonlocal_goto) |
| 1592 | { |
| 1593 | edge e; |
| 1594 | edge_iterator ei; |
| 1595 | |
| 1596 | FOR_EACH_EDGE (e, ei, bb->succs) |
| 1597 | if (!e->dest->aux |
| 1598 | || ((basic_block)e->dest->aux)->index == ENTRY_BLOCK) |
| 1599 | { |
| 1600 | gimple phi; |
| 1601 | gimple_stmt_iterator si; |
| 1602 | |
| 1603 | gcc_assert (e->flags & EDGE_ABNORMAL); |
| 1604 | |
| 1605 | if (!nonlocal_goto) |
| 1606 | gcc_assert (e->flags & EDGE_EH); |
| 1607 | |
| 1608 | if (!can_throw) |
| 1609 | gcc_assert (!(e->flags & EDGE_EH)); |
| 1610 | |
| 1611 | for (si = gsi_start_phis (e->dest); !gsi_end_p (si); gsi_next (&si)) |
| 1612 | { |
| 1613 | edge re; |
| 1614 | |
| 1615 | phi = gsi_stmt (si); |
| 1616 | |
| 1617 | /* There shouldn't be any PHI nodes in the ENTRY_BLOCK. */ |
| 1618 | gcc_assert (!e->dest->aux); |
| 1619 | |
| 1620 | gcc_assert (SSA_NAME_OCCURS_IN_ABNORMAL_PHI (PHI_RESULT (phi))); |
| 1621 | |
| 1622 | if (!is_gimple_reg (PHI_RESULT (phi))) |
| 1623 | { |
| 1624 | mark_sym_for_renaming (SSA_NAME_VAR (PHI_RESULT (phi))); |
| 1625 | continue; |
| 1626 | } |
| 1627 | |
| 1628 | re = find_edge (ret_bb, e->dest); |
| 1629 | gcc_assert (re); |
| 1630 | gcc_assert ((re->flags & (EDGE_EH | EDGE_ABNORMAL)) |
| 1631 | == (e->flags & (EDGE_EH | EDGE_ABNORMAL))); |
| 1632 | |
| 1633 | SET_USE (PHI_ARG_DEF_PTR_FROM_EDGE (phi, e), |
| 1634 | USE_FROM_PTR (PHI_ARG_DEF_PTR_FROM_EDGE (phi, re))); |
| 1635 | } |
| 1636 | } |
| 1637 | } |
| 1638 | |
| 1639 | |
| 1640 | /* Copy edges from BB into its copy constructed earlier, scale profile |
| 1641 | accordingly. Edges will be taken care of later. Assume aux |
| 1642 | pointers to point to the copies of each BB. */ |
| 1643 | |
| 1644 | static void |
| 1645 | copy_edges_for_bb (basic_block bb, gcov_type count_scale, basic_block ret_bb) |
| 1646 | { |
| 1647 | basic_block new_bb = (basic_block) bb->aux; |
| 1648 | edge_iterator ei; |
| 1649 | edge old_edge; |
| 1650 | gimple_stmt_iterator si; |
| 1651 | int flags; |
| 1652 | |
| 1653 | /* Use the indices from the original blocks to create edges for the |
| 1654 | new ones. */ |
| 1655 | FOR_EACH_EDGE (old_edge, ei, bb->succs) |
| 1656 | if (!(old_edge->flags & EDGE_EH)) |
| 1657 | { |
| 1658 | edge new_edge; |
| 1659 | |
| 1660 | flags = old_edge->flags; |
| 1661 | |
| 1662 | /* Return edges do get a FALLTHRU flag when the get inlined. */ |
| 1663 | if (old_edge->dest->index == EXIT_BLOCK && !old_edge->flags |
| 1664 | && old_edge->dest->aux != EXIT_BLOCK_PTR) |
| 1665 | flags |= EDGE_FALLTHRU; |
| 1666 | new_edge = make_edge (new_bb, (basic_block) old_edge->dest->aux, flags); |
| 1667 | new_edge->count = old_edge->count * count_scale / REG_BR_PROB_BASE; |
| 1668 | new_edge->probability = old_edge->probability; |
| 1669 | } |
| 1670 | |
| 1671 | if (bb->index == ENTRY_BLOCK || bb->index == EXIT_BLOCK) |
| 1672 | return; |
| 1673 | |
| 1674 | for (si = gsi_start_bb (new_bb); !gsi_end_p (si);) |
| 1675 | { |
| 1676 | gimple copy_stmt; |
| 1677 | bool can_throw, nonlocal_goto; |
| 1678 | |
| 1679 | copy_stmt = gsi_stmt (si); |
| 1680 | update_stmt (copy_stmt); |
| 1681 | if (gimple_in_ssa_p (cfun)) |
| 1682 | mark_symbols_for_renaming (copy_stmt); |
| 1683 | |
| 1684 | /* Do this before the possible split_block. */ |
| 1685 | gsi_next (&si); |
| 1686 | |
| 1687 | /* If this tree could throw an exception, there are two |
| 1688 | cases where we need to add abnormal edge(s): the |
| 1689 | tree wasn't in a region and there is a "current |
| 1690 | region" in the caller; or the original tree had |
| 1691 | EH edges. In both cases split the block after the tree, |
| 1692 | and add abnormal edge(s) as needed; we need both |
| 1693 | those from the callee and the caller. |
| 1694 | We check whether the copy can throw, because the const |
| 1695 | propagation can change an INDIRECT_REF which throws |
| 1696 | into a COMPONENT_REF which doesn't. If the copy |
| 1697 | can throw, the original could also throw. */ |
| 1698 | can_throw = stmt_can_throw_internal (copy_stmt); |
| 1699 | nonlocal_goto = stmt_can_make_abnormal_goto (copy_stmt); |
| 1700 | |
| 1701 | if (can_throw || nonlocal_goto) |
| 1702 | { |
| 1703 | if (!gsi_end_p (si)) |
| 1704 | /* Note that bb's predecessor edges aren't necessarily |
| 1705 | right at this point; split_block doesn't care. */ |
| 1706 | { |
| 1707 | edge e = split_block (new_bb, copy_stmt); |
| 1708 | |
| 1709 | new_bb = e->dest; |
| 1710 | new_bb->aux = e->src->aux; |
| 1711 | si = gsi_start_bb (new_bb); |
| 1712 | } |
| 1713 | } |
| 1714 | |
| 1715 | if (can_throw) |
| 1716 | make_eh_edges (copy_stmt); |
| 1717 | |
| 1718 | if (nonlocal_goto) |
| 1719 | make_abnormal_goto_edges (gimple_bb (copy_stmt), true); |
| 1720 | |
| 1721 | if ((can_throw || nonlocal_goto) |
| 1722 | && gimple_in_ssa_p (cfun)) |
| 1723 | update_ssa_across_abnormal_edges (gimple_bb (copy_stmt), ret_bb, |
| 1724 | can_throw, nonlocal_goto); |
| 1725 | } |
| 1726 | } |
| 1727 | |
| 1728 | /* Copy the PHIs. All blocks and edges are copied, some blocks |
| 1729 | was possibly split and new outgoing EH edges inserted. |
| 1730 | BB points to the block of original function and AUX pointers links |
| 1731 | the original and newly copied blocks. */ |
| 1732 | |
| 1733 | static void |
| 1734 | copy_phis_for_bb (basic_block bb, copy_body_data *id) |
| 1735 | { |
| 1736 | basic_block const new_bb = (basic_block) bb->aux; |
| 1737 | edge_iterator ei; |
| 1738 | gimple phi; |
| 1739 | gimple_stmt_iterator si; |
| 1740 | edge new_edge; |
| 1741 | bool inserted = false; |
| 1742 | |
| 1743 | for (si = gsi_start (phi_nodes (bb)); !gsi_end_p (si); gsi_next (&si)) |
| 1744 | { |
| 1745 | tree res, new_res; |
| 1746 | gimple new_phi; |
| 1747 | |
| 1748 | phi = gsi_stmt (si); |
| 1749 | res = PHI_RESULT (phi); |
| 1750 | new_res = res; |
| 1751 | if (is_gimple_reg (res)) |
| 1752 | { |
| 1753 | walk_tree (&new_res, copy_tree_body_r, id, NULL); |
| 1754 | SSA_NAME_DEF_STMT (new_res) |
| 1755 | = new_phi = create_phi_node (new_res, new_bb); |
| 1756 | FOR_EACH_EDGE (new_edge, ei, new_bb->preds) |
| 1757 | { |
| 1758 | edge const old_edge |
| 1759 | = find_edge ((basic_block) new_edge->src->aux, bb); |
| 1760 | tree arg = PHI_ARG_DEF_FROM_EDGE (phi, old_edge); |
| 1761 | tree new_arg = arg; |
| 1762 | tree block = id->block; |
| 1763 | id->block = NULL_TREE; |
| 1764 | walk_tree (&new_arg, copy_tree_body_r, id, NULL); |
| 1765 | id->block = block; |
| 1766 | gcc_assert (new_arg); |
| 1767 | /* With return slot optimization we can end up with |
| 1768 | non-gimple (foo *)&this->m, fix that here. */ |
| 1769 | if (TREE_CODE (new_arg) != SSA_NAME |
| 1770 | && TREE_CODE (new_arg) != FUNCTION_DECL |
| 1771 | && !is_gimple_val (new_arg)) |
| 1772 | { |
| 1773 | gimple_seq stmts = NULL; |
| 1774 | new_arg = force_gimple_operand (new_arg, &stmts, true, NULL); |
| 1775 | gsi_insert_seq_on_edge (new_edge, stmts); |
| 1776 | inserted = true; |
| 1777 | } |
| 1778 | add_phi_arg (new_phi, new_arg, new_edge); |
| 1779 | } |
| 1780 | } |
| 1781 | } |
| 1782 | |
| 1783 | /* Commit the delayed edge insertions. */ |
| 1784 | if (inserted) |
| 1785 | FOR_EACH_EDGE (new_edge, ei, new_bb->preds) |
| 1786 | gsi_commit_one_edge_insert (new_edge, NULL); |
| 1787 | } |
| 1788 | |
| 1789 | |
| 1790 | /* Wrapper for remap_decl so it can be used as a callback. */ |
| 1791 | |
| 1792 | static tree |
| 1793 | remap_decl_1 (tree decl, void *data) |
| 1794 | { |
| 1795 | return remap_decl (decl, (copy_body_data *) data); |
| 1796 | } |
| 1797 | |
| 1798 | /* Build struct function and associated datastructures for the new clone |
| 1799 | NEW_FNDECL to be build. CALLEE_FNDECL is the original */ |
| 1800 | |
| 1801 | static void |
| 1802 | initialize_cfun (tree new_fndecl, tree callee_fndecl, gcov_type count, |
| 1803 | int frequency) |
| 1804 | { |
| 1805 | struct function *src_cfun = DECL_STRUCT_FUNCTION (callee_fndecl); |
| 1806 | gcov_type count_scale, frequency_scale; |
| 1807 | |
| 1808 | if (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count) |
| 1809 | count_scale = (REG_BR_PROB_BASE * count |
| 1810 | / ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count); |
| 1811 | else |
| 1812 | count_scale = 1; |
| 1813 | |
| 1814 | if (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency) |
| 1815 | frequency_scale = (REG_BR_PROB_BASE * frequency |
| 1816 | / |
| 1817 | ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency); |
| 1818 | else |
| 1819 | frequency_scale = count_scale; |
| 1820 | |
| 1821 | /* Register specific tree functions. */ |
| 1822 | gimple_register_cfg_hooks (); |
| 1823 | |
| 1824 | /* Get clean struct function. */ |
| 1825 | push_struct_function (new_fndecl); |
| 1826 | |
| 1827 | /* We will rebuild these, so just sanity check that they are empty. */ |
| 1828 | gcc_assert (VALUE_HISTOGRAMS (cfun) == NULL); |
| 1829 | gcc_assert (cfun->local_decls == NULL); |
| 1830 | gcc_assert (cfun->cfg == NULL); |
| 1831 | gcc_assert (cfun->decl == new_fndecl); |
| 1832 | |
| 1833 | /* Copy items we preserve during clonning. */ |
| 1834 | cfun->static_chain_decl = src_cfun->static_chain_decl; |
| 1835 | cfun->nonlocal_goto_save_area = src_cfun->nonlocal_goto_save_area; |
| 1836 | cfun->function_end_locus = src_cfun->function_end_locus; |
| 1837 | cfun->curr_properties = src_cfun->curr_properties; |
| 1838 | cfun->last_verified = src_cfun->last_verified; |
| 1839 | if (src_cfun->ipa_transforms_to_apply) |
| 1840 | cfun->ipa_transforms_to_apply = VEC_copy (ipa_opt_pass, heap, |
| 1841 | src_cfun->ipa_transforms_to_apply); |
| 1842 | cfun->va_list_gpr_size = src_cfun->va_list_gpr_size; |
| 1843 | cfun->va_list_fpr_size = src_cfun->va_list_fpr_size; |
| 1844 | cfun->function_frequency = src_cfun->function_frequency; |
| 1845 | cfun->has_nonlocal_label = src_cfun->has_nonlocal_label; |
| 1846 | cfun->stdarg = src_cfun->stdarg; |
| 1847 | cfun->dont_save_pending_sizes_p = src_cfun->dont_save_pending_sizes_p; |
| 1848 | cfun->after_inlining = src_cfun->after_inlining; |
| 1849 | cfun->returns_struct = src_cfun->returns_struct; |
| 1850 | cfun->returns_pcc_struct = src_cfun->returns_pcc_struct; |
| 1851 | cfun->after_tree_profile = src_cfun->after_tree_profile; |
| 1852 | |
| 1853 | init_empty_tree_cfg (); |
| 1854 | |
| 1855 | ENTRY_BLOCK_PTR->count = |
| 1856 | (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count * count_scale / |
| 1857 | REG_BR_PROB_BASE); |
| 1858 | ENTRY_BLOCK_PTR->frequency = |
| 1859 | (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency * |
| 1860 | frequency_scale / REG_BR_PROB_BASE); |
| 1861 | EXIT_BLOCK_PTR->count = |
| 1862 | (EXIT_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count * count_scale / |
| 1863 | REG_BR_PROB_BASE); |
| 1864 | EXIT_BLOCK_PTR->frequency = |
| 1865 | (EXIT_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency * |
| 1866 | frequency_scale / REG_BR_PROB_BASE); |
| 1867 | if (src_cfun->eh) |
| 1868 | init_eh_for_function (); |
| 1869 | |
| 1870 | if (src_cfun->gimple_df) |
| 1871 | { |
| 1872 | init_tree_ssa (cfun); |
| 1873 | cfun->gimple_df->in_ssa_p = true; |
| 1874 | init_ssa_operands (); |
| 1875 | } |
| 1876 | pop_cfun (); |
| 1877 | } |
| 1878 | |
| 1879 | /* Make a copy of the body of FN so that it can be inserted inline in |
| 1880 | another function. Walks FN via CFG, returns new fndecl. */ |
| 1881 | |
| 1882 | static tree |
| 1883 | copy_cfg_body (copy_body_data * id, gcov_type count, int frequency, |
| 1884 | basic_block entry_block_map, basic_block exit_block_map) |
| 1885 | { |
| 1886 | tree callee_fndecl = id->src_fn; |
| 1887 | /* Original cfun for the callee, doesn't change. */ |
| 1888 | struct function *src_cfun = DECL_STRUCT_FUNCTION (callee_fndecl); |
| 1889 | struct function *cfun_to_copy; |
| 1890 | basic_block bb; |
| 1891 | tree new_fndecl = NULL; |
| 1892 | gcov_type count_scale, frequency_scale; |
| 1893 | int last; |
| 1894 | |
| 1895 | if (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count) |
| 1896 | count_scale = (REG_BR_PROB_BASE * count |
| 1897 | / ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->count); |
| 1898 | else |
| 1899 | count_scale = 1; |
| 1900 | |
| 1901 | if (ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency) |
| 1902 | frequency_scale = (REG_BR_PROB_BASE * frequency |
| 1903 | / |
| 1904 | ENTRY_BLOCK_PTR_FOR_FUNCTION (src_cfun)->frequency); |
| 1905 | else |
| 1906 | frequency_scale = count_scale; |
| 1907 | |
| 1908 | /* Register specific tree functions. */ |
| 1909 | gimple_register_cfg_hooks (); |
| 1910 | |
| 1911 | /* Must have a CFG here at this point. */ |
| 1912 | gcc_assert (ENTRY_BLOCK_PTR_FOR_FUNCTION |
| 1913 | (DECL_STRUCT_FUNCTION (callee_fndecl))); |
| 1914 | |
| 1915 | cfun_to_copy = id->src_cfun = DECL_STRUCT_FUNCTION (callee_fndecl); |
| 1916 | |
| 1917 | ENTRY_BLOCK_PTR_FOR_FUNCTION (cfun_to_copy)->aux = entry_block_map; |
| 1918 | EXIT_BLOCK_PTR_FOR_FUNCTION (cfun_to_copy)->aux = exit_block_map; |
| 1919 | entry_block_map->aux = ENTRY_BLOCK_PTR_FOR_FUNCTION (cfun_to_copy); |
| 1920 | exit_block_map->aux = EXIT_BLOCK_PTR_FOR_FUNCTION (cfun_to_copy); |
| 1921 | |
| 1922 | /* Duplicate any exception-handling regions. */ |
| 1923 | if (cfun->eh) |
| 1924 | { |
| 1925 | id->eh_region_offset |
| 1926 | = duplicate_eh_regions (cfun_to_copy, remap_decl_1, id, |
| 1927 | 0, id->eh_region); |
| 1928 | } |
| 1929 | |
| 1930 | /* Use aux pointers to map the original blocks to copy. */ |
| 1931 | FOR_EACH_BB_FN (bb, cfun_to_copy) |
| 1932 | { |
| 1933 | basic_block new_bb = copy_bb (id, bb, frequency_scale, count_scale); |
| 1934 | bb->aux = new_bb; |
| 1935 | new_bb->aux = bb; |
| 1936 | } |
| 1937 | |
| 1938 | last = last_basic_block; |
| 1939 | |
| 1940 | /* Now that we've duplicated the blocks, duplicate their edges. */ |
| 1941 | FOR_ALL_BB_FN (bb, cfun_to_copy) |
| 1942 | copy_edges_for_bb (bb, count_scale, exit_block_map); |
| 1943 | |
| 1944 | if (gimple_in_ssa_p (cfun)) |
| 1945 | FOR_ALL_BB_FN (bb, cfun_to_copy) |
| 1946 | copy_phis_for_bb (bb, id); |
| 1947 | |
| 1948 | FOR_ALL_BB_FN (bb, cfun_to_copy) |
| 1949 | { |
| 1950 | ((basic_block)bb->aux)->aux = NULL; |
| 1951 | bb->aux = NULL; |
| 1952 | } |
| 1953 | |
| 1954 | /* Zero out AUX fields of newly created block during EH edge |
| 1955 | insertion. */ |
| 1956 | for (; last < last_basic_block; last++) |
| 1957 | BASIC_BLOCK (last)->aux = NULL; |
| 1958 | entry_block_map->aux = NULL; |
| 1959 | exit_block_map->aux = NULL; |
| 1960 | |
| 1961 | return new_fndecl; |
| 1962 | } |
| 1963 | |
| 1964 | static tree |
| 1965 | copy_body (copy_body_data *id, gcov_type count, int frequency, |
| 1966 | basic_block entry_block_map, basic_block exit_block_map) |
| 1967 | { |
| 1968 | tree fndecl = id->src_fn; |
| 1969 | tree body; |
| 1970 | |
| 1971 | /* If this body has a CFG, walk CFG and copy. */ |
| 1972 | gcc_assert (ENTRY_BLOCK_PTR_FOR_FUNCTION (DECL_STRUCT_FUNCTION (fndecl))); |
| 1973 | body = copy_cfg_body (id, count, frequency, entry_block_map, exit_block_map); |
| 1974 | |
| 1975 | return body; |
| 1976 | } |
| 1977 | |
| 1978 | /* Return true if VALUE is an ADDR_EXPR of an automatic variable |
| 1979 | defined in function FN, or of a data member thereof. */ |
| 1980 | |
| 1981 | static bool |
| 1982 | self_inlining_addr_expr (tree value, tree fn) |
| 1983 | { |
| 1984 | tree var; |
| 1985 | |
| 1986 | if (TREE_CODE (value) != ADDR_EXPR) |
| 1987 | return false; |
| 1988 | |
| 1989 | var = get_base_address (TREE_OPERAND (value, 0)); |
| 1990 | |
| 1991 | return var && auto_var_in_fn_p (var, fn); |
| 1992 | } |
| 1993 | |
| 1994 | static void |
| 1995 | insert_init_stmt (basic_block bb, gimple init_stmt) |
| 1996 | { |
| 1997 | /* If VAR represents a zero-sized variable, it's possible that the |
| 1998 | assignment statement may result in no gimple statements. */ |
| 1999 | if (init_stmt) |
| 2000 | { |
| 2001 | gimple_stmt_iterator si = gsi_last_bb (bb); |
| 2002 | |
| 2003 | /* We can end up with init statements that store to a non-register |
| 2004 | from a rhs with a conversion. Handle that here by forcing the |
| 2005 | rhs into a temporary. gimple_regimplify_operands is not |
| 2006 | prepared to do this for us. */ |
| 2007 | if (!is_gimple_reg (gimple_assign_lhs (init_stmt)) |
| 2008 | && is_gimple_reg_type (TREE_TYPE (gimple_assign_lhs (init_stmt))) |
| 2009 | && gimple_assign_rhs_class (init_stmt) == GIMPLE_UNARY_RHS) |
| 2010 | { |
| 2011 | tree rhs = build1 (gimple_assign_rhs_code (init_stmt), |
| 2012 | gimple_expr_type (init_stmt), |
| 2013 | gimple_assign_rhs1 (init_stmt)); |
| 2014 | rhs = force_gimple_operand_gsi (&si, rhs, true, NULL_TREE, false, |
| 2015 | GSI_NEW_STMT); |
| 2016 | gimple_assign_set_rhs_code (init_stmt, TREE_CODE (rhs)); |
| 2017 | gimple_assign_set_rhs1 (init_stmt, rhs); |
| 2018 | } |
| 2019 | gsi_insert_after (&si, init_stmt, GSI_NEW_STMT); |
| 2020 | gimple_regimplify_operands (init_stmt, &si); |
| 2021 | mark_symbols_for_renaming (init_stmt); |
| 2022 | } |
| 2023 | } |
| 2024 | |
| 2025 | /* Initialize parameter P with VALUE. If needed, produce init statement |
| 2026 | at the end of BB. When BB is NULL, we return init statement to be |
| 2027 | output later. */ |
| 2028 | static gimple |
| 2029 | setup_one_parameter (copy_body_data *id, tree p, tree value, tree fn, |
| 2030 | basic_block bb, tree *vars) |
| 2031 | { |
| 2032 | gimple init_stmt = NULL; |
| 2033 | tree var; |
| 2034 | tree rhs = value; |
| 2035 | tree def = (gimple_in_ssa_p (cfun) |
| 2036 | ? gimple_default_def (id->src_cfun, p) : NULL); |
| 2037 | |
| 2038 | if (value |
| 2039 | && value != error_mark_node |
| 2040 | && !useless_type_conversion_p (TREE_TYPE (p), TREE_TYPE (value))) |
| 2041 | { |
| 2042 | if (fold_convertible_p (TREE_TYPE (p), value)) |
| 2043 | rhs = fold_build1 (NOP_EXPR, TREE_TYPE (p), value); |
| 2044 | else |
| 2045 | /* ??? For valid (GIMPLE) programs we should not end up here. |
| 2046 | Still if something has gone wrong and we end up with truly |
| 2047 | mismatched types here, fall back to using a VIEW_CONVERT_EXPR |
| 2048 | to not leak invalid GIMPLE to the following passes. */ |
| 2049 | rhs = fold_build1 (VIEW_CONVERT_EXPR, TREE_TYPE (p), value); |
| 2050 | } |
| 2051 | |
| 2052 | /* If the parameter is never assigned to, has no SSA_NAMEs created, |
| 2053 | we may not need to create a new variable here at all. Instead, we may |
| 2054 | be able to just use the argument value. */ |
| 2055 | if (TREE_READONLY (p) |
| 2056 | && !TREE_ADDRESSABLE (p) |
| 2057 | && value && !TREE_SIDE_EFFECTS (value) |
| 2058 | && !def) |
| 2059 | { |
| 2060 | /* We may produce non-gimple trees by adding NOPs or introduce |
| 2061 | invalid sharing when operand is not really constant. |
| 2062 | It is not big deal to prohibit constant propagation here as |
| 2063 | we will constant propagate in DOM1 pass anyway. */ |
| 2064 | if (is_gimple_min_invariant (value) |
| 2065 | && useless_type_conversion_p (TREE_TYPE (p), |
| 2066 | TREE_TYPE (value)) |
| 2067 | /* We have to be very careful about ADDR_EXPR. Make sure |
| 2068 | the base variable isn't a local variable of the inlined |
| 2069 | function, e.g., when doing recursive inlining, direct or |
| 2070 | mutually-recursive or whatever, which is why we don't |
| 2071 | just test whether fn == current_function_decl. */ |
| 2072 | && ! self_inlining_addr_expr (value, fn)) |
| 2073 | { |
| 2074 | insert_decl_map (id, p, value); |
| 2075 | return NULL; |
| 2076 | } |
| 2077 | } |
| 2078 | |
| 2079 | /* Make an equivalent VAR_DECL. Note that we must NOT remap the type |
| 2080 | here since the type of this decl must be visible to the calling |
| 2081 | function. */ |
| 2082 | var = copy_decl_to_var (p, id); |
| 2083 | if (gimple_in_ssa_p (cfun) && TREE_CODE (var) == VAR_DECL) |
| 2084 | { |
| 2085 | get_var_ann (var); |
| 2086 | add_referenced_var (var); |
| 2087 | } |
| 2088 | |
| 2089 | /* Register the VAR_DECL as the equivalent for the PARM_DECL; |
| 2090 | that way, when the PARM_DECL is encountered, it will be |
| 2091 | automatically replaced by the VAR_DECL. */ |
| 2092 | insert_decl_map (id, p, var); |
| 2093 | |
| 2094 | /* Declare this new variable. */ |
| 2095 | TREE_CHAIN (var) = *vars; |
| 2096 | *vars = var; |
| 2097 | |
| 2098 | /* Make gimplifier happy about this variable. */ |
| 2099 | DECL_SEEN_IN_BIND_EXPR_P (var) = 1; |
| 2100 | |
| 2101 | /* Even if P was TREE_READONLY, the new VAR should not be. |
| 2102 | In the original code, we would have constructed a |
| 2103 | temporary, and then the function body would have never |
| 2104 | changed the value of P. However, now, we will be |
| 2105 | constructing VAR directly. The constructor body may |
| 2106 | change its value multiple times as it is being |
| 2107 | constructed. Therefore, it must not be TREE_READONLY; |
| 2108 | the back-end assumes that TREE_READONLY variable is |
| 2109 | assigned to only once. */ |
| 2110 | if (TYPE_NEEDS_CONSTRUCTING (TREE_TYPE (p))) |
| 2111 | TREE_READONLY (var) = 0; |
| 2112 | |
| 2113 | /* If there is no setup required and we are in SSA, take the easy route |
| 2114 | replacing all SSA names representing the function parameter by the |
| 2115 | SSA name passed to function. |
| 2116 | |
| 2117 | We need to construct map for the variable anyway as it might be used |
| 2118 | in different SSA names when parameter is set in function. |
| 2119 | |
| 2120 | Do replacement at -O0 for const arguments replaced by constant. |
| 2121 | This is important for builtin_constant_p and other construct requiring |
| 2122 | constant argument to be visible in inlined function body. |
| 2123 | |
| 2124 | FIXME: This usually kills the last connection in between inlined |
| 2125 | function parameter and the actual value in debug info. Can we do |
| 2126 | better here? If we just inserted the statement, copy propagation |
| 2127 | would kill it anyway as it always did in older versions of GCC. |
| 2128 | |
| 2129 | We might want to introduce a notion that single SSA_NAME might |
| 2130 | represent multiple variables for purposes of debugging. */ |
| 2131 | if (gimple_in_ssa_p (cfun) && rhs && def && is_gimple_reg (p) |
| 2132 | && (optimize |
| 2133 | || (TREE_READONLY (p) |
| 2134 | && is_gimple_min_invariant (rhs))) |
| 2135 | && (TREE_CODE (rhs) == SSA_NAME |
| 2136 | || is_gimple_min_invariant (rhs)) |
| 2137 | && !SSA_NAME_OCCURS_IN_ABNORMAL_PHI (def)) |
| 2138 | { |
| 2139 | insert_decl_map (id, def, rhs); |
| 2140 | return NULL; |
| 2141 | } |
| 2142 | |
| 2143 | /* If the value of argument is never used, don't care about initializing |
| 2144 | it. */ |
| 2145 | if (optimize && gimple_in_ssa_p (cfun) && !def && is_gimple_reg (p)) |
| 2146 | { |
| 2147 | gcc_assert (!value || !TREE_SIDE_EFFECTS (value)); |
| 2148 | return NULL; |
| 2149 | } |
| 2150 | |
| 2151 | /* Initialize this VAR_DECL from the equivalent argument. Convert |
| 2152 | the argument to the proper type in case it was promoted. */ |
| 2153 | if (value) |
| 2154 | { |
| 2155 | if (rhs == error_mark_node) |
| 2156 | { |
| 2157 | insert_decl_map (id, p, var); |
| 2158 | return NULL; |
| 2159 | } |
| 2160 | |
| 2161 | STRIP_USELESS_TYPE_CONVERSION (rhs); |
| 2162 | |
| 2163 | /* We want to use MODIFY_EXPR, not INIT_EXPR here so that we |
| 2164 | keep our trees in gimple form. */ |
| 2165 | if (def && gimple_in_ssa_p (cfun) && is_gimple_reg (p)) |
| 2166 | { |
| 2167 | def = remap_ssa_name (def, id); |
| 2168 | init_stmt = gimple_build_assign (def, rhs); |
| 2169 | SSA_NAME_IS_DEFAULT_DEF (def) = 0; |
| 2170 | set_default_def (var, NULL); |
| 2171 | } |
| 2172 | else |
| 2173 | init_stmt = gimple_build_assign (var, rhs); |
| 2174 | |
| 2175 | if (bb && init_stmt) |
| 2176 | insert_init_stmt (bb, init_stmt); |
| 2177 | } |
| 2178 | return init_stmt; |
| 2179 | } |
| 2180 | |
| 2181 | /* Generate code to initialize the parameters of the function at the |
| 2182 | top of the stack in ID from the GIMPLE_CALL STMT. */ |
| 2183 | |
| 2184 | static void |
| 2185 | initialize_inlined_parameters (copy_body_data *id, gimple stmt, |
| 2186 | tree fn, basic_block bb) |
| 2187 | { |
| 2188 | tree parms; |
| 2189 | size_t i; |
| 2190 | tree p; |
| 2191 | tree vars = NULL_TREE; |
| 2192 | tree static_chain = gimple_call_chain (stmt); |
| 2193 | |
| 2194 | /* Figure out what the parameters are. */ |
| 2195 | parms = DECL_ARGUMENTS (fn); |
| 2196 | |
| 2197 | /* Loop through the parameter declarations, replacing each with an |
| 2198 | equivalent VAR_DECL, appropriately initialized. */ |
| 2199 | for (p = parms, i = 0; p; p = TREE_CHAIN (p), i++) |
| 2200 | { |
| 2201 | tree val; |
| 2202 | val = i < gimple_call_num_args (stmt) ? gimple_call_arg (stmt, i) : NULL; |
| 2203 | setup_one_parameter (id, p, val, fn, bb, &vars); |
| 2204 | } |
| 2205 | |
| 2206 | /* Initialize the static chain. */ |
| 2207 | p = DECL_STRUCT_FUNCTION (fn)->static_chain_decl; |
| 2208 | gcc_assert (fn != current_function_decl); |
| 2209 | if (p) |
| 2210 | { |
| 2211 | /* No static chain? Seems like a bug in tree-nested.c. */ |
| 2212 | gcc_assert (static_chain); |
| 2213 | |
| 2214 | setup_one_parameter (id, p, static_chain, fn, bb, &vars); |
| 2215 | } |
| 2216 | |
| 2217 | declare_inline_vars (id->block, vars); |
| 2218 | } |
| 2219 | |
| 2220 | |
| 2221 | /* Declare a return variable to replace the RESULT_DECL for the |
| 2222 | function we are calling. An appropriate DECL_STMT is returned. |
| 2223 | The USE_STMT is filled to contain a use of the declaration to |
| 2224 | indicate the return value of the function. |
| 2225 | |
| 2226 | RETURN_SLOT, if non-null is place where to store the result. It |
| 2227 | is set only for CALL_EXPR_RETURN_SLOT_OPT. MODIFY_DEST, if non-null, |
| 2228 | was the LHS of the MODIFY_EXPR to which this call is the RHS. |
| 2229 | |
| 2230 | The return value is a (possibly null) value that is the result of the |
| 2231 | function as seen by the callee. *USE_P is a (possibly null) value that |
| 2232 | holds the result as seen by the caller. */ |
| 2233 | |
| 2234 | static tree |
| 2235 | declare_return_variable (copy_body_data *id, tree return_slot, tree modify_dest, |
| 2236 | tree *use_p) |
| 2237 | { |
| 2238 | tree callee = id->src_fn; |
| 2239 | tree caller = id->dst_fn; |
| 2240 | tree result = DECL_RESULT (callee); |
| 2241 | tree callee_type = TREE_TYPE (result); |
| 2242 | tree caller_type = TREE_TYPE (TREE_TYPE (callee)); |
| 2243 | tree var, use; |
| 2244 | |
| 2245 | /* We don't need to do anything for functions that don't return |
| 2246 | anything. */ |
| 2247 | if (!result || VOID_TYPE_P (callee_type)) |
| 2248 | { |
| 2249 | *use_p = NULL_TREE; |
| 2250 | return NULL_TREE; |
| 2251 | } |
| 2252 | |
| 2253 | /* If there was a return slot, then the return value is the |
| 2254 | dereferenced address of that object. */ |
| 2255 | if (return_slot) |
| 2256 | { |
| 2257 | /* The front end shouldn't have used both return_slot and |
| 2258 | a modify expression. */ |
| 2259 | gcc_assert (!modify_dest); |
| 2260 | if (DECL_BY_REFERENCE (result)) |
| 2261 | { |
| 2262 | tree return_slot_addr = build_fold_addr_expr (return_slot); |
| 2263 | STRIP_USELESS_TYPE_CONVERSION (return_slot_addr); |
| 2264 | |
| 2265 | /* We are going to construct *&return_slot and we can't do that |
| 2266 | for variables believed to be not addressable. |
| 2267 | |
| 2268 | FIXME: This check possibly can match, because values returned |
| 2269 | via return slot optimization are not believed to have address |
| 2270 | taken by alias analysis. */ |
| 2271 | gcc_assert (TREE_CODE (return_slot) != SSA_NAME); |
| 2272 | if (gimple_in_ssa_p (cfun)) |
| 2273 | { |
| 2274 | HOST_WIDE_INT bitsize; |
| 2275 | HOST_WIDE_INT bitpos; |
| 2276 | tree offset; |
| 2277 | enum machine_mode mode; |
| 2278 | int unsignedp; |
| 2279 | int volatilep; |
| 2280 | tree base; |
| 2281 | base = get_inner_reference (return_slot, &bitsize, &bitpos, |
| 2282 | &offset, |
| 2283 | &mode, &unsignedp, &volatilep, |
| 2284 | false); |
| 2285 | if (TREE_CODE (base) == INDIRECT_REF) |
| 2286 | base = TREE_OPERAND (base, 0); |
| 2287 | if (TREE_CODE (base) == SSA_NAME) |
| 2288 | base = SSA_NAME_VAR (base); |
| 2289 | mark_sym_for_renaming (base); |
| 2290 | } |
| 2291 | var = return_slot_addr; |
| 2292 | } |
| 2293 | else |
| 2294 | { |
| 2295 | var = return_slot; |
| 2296 | gcc_assert (TREE_CODE (var) != SSA_NAME); |
| 2297 | TREE_ADDRESSABLE (var) |= TREE_ADDRESSABLE (result); |
| 2298 | } |
| 2299 | if ((TREE_CODE (TREE_TYPE (result)) == COMPLEX_TYPE |
| 2300 | || TREE_CODE (TREE_TYPE (result)) == VECTOR_TYPE) |
| 2301 | && !DECL_GIMPLE_REG_P (result) |
| 2302 | && DECL_P (var)) |
| 2303 | DECL_GIMPLE_REG_P (var) = 0; |
| 2304 | use = NULL; |
| 2305 | goto done; |
| 2306 | } |
| 2307 | |
| 2308 | /* All types requiring non-trivial constructors should have been handled. */ |
| 2309 | gcc_assert (!TREE_ADDRESSABLE (callee_type)); |
| 2310 | |
| 2311 | /* Attempt to avoid creating a new temporary variable. */ |
| 2312 | if (modify_dest |
| 2313 | && TREE_CODE (modify_dest) != SSA_NAME) |
| 2314 | { |
| 2315 | bool use_it = false; |
| 2316 | |
| 2317 | /* We can't use MODIFY_DEST if there's type promotion involved. */ |
| 2318 | if (!useless_type_conversion_p (callee_type, caller_type)) |
| 2319 | use_it = false; |
| 2320 | |
| 2321 | /* ??? If we're assigning to a variable sized type, then we must |
| 2322 | reuse the destination variable, because we've no good way to |
| 2323 | create variable sized temporaries at this point. */ |
| 2324 | else if (TREE_CODE (TYPE_SIZE_UNIT (caller_type)) != INTEGER_CST) |
| 2325 | use_it = true; |
| 2326 | |
| 2327 | /* If the callee cannot possibly modify MODIFY_DEST, then we can |
| 2328 | reuse it as the result of the call directly. Don't do this if |
| 2329 | it would promote MODIFY_DEST to addressable. */ |
| 2330 | else if (TREE_ADDRESSABLE (result)) |
| 2331 | use_it = false; |
| 2332 | else |
| 2333 | { |
| 2334 | tree base_m = get_base_address (modify_dest); |
| 2335 | |
| 2336 | /* If the base isn't a decl, then it's a pointer, and we don't |
| 2337 | know where that's going to go. */ |
| 2338 | if (!DECL_P (base_m)) |
| 2339 | use_it = false; |
| 2340 | else if (is_global_var (base_m)) |
| 2341 | use_it = false; |
| 2342 | else if ((TREE_CODE (TREE_TYPE (result)) == COMPLEX_TYPE |
| 2343 | || TREE_CODE (TREE_TYPE (result)) == VECTOR_TYPE) |
| 2344 | && !DECL_GIMPLE_REG_P (result) |
| 2345 | && DECL_GIMPLE_REG_P (base_m)) |
| 2346 | use_it = false; |
| 2347 | else if (!TREE_ADDRESSABLE (base_m)) |
| 2348 | use_it = true; |
| 2349 | } |
| 2350 | |
| 2351 | if (use_it) |
| 2352 | { |
| 2353 | var = modify_dest; |
| 2354 | use = NULL; |
| 2355 | goto done; |
| 2356 | } |
| 2357 | } |
| 2358 | |
| 2359 | gcc_assert (TREE_CODE (TYPE_SIZE_UNIT (callee_type)) == INTEGER_CST); |
| 2360 | |
| 2361 | var = copy_result_decl_to_var (result, id); |
| 2362 | if (gimple_in_ssa_p (cfun)) |
| 2363 | { |
| 2364 | get_var_ann (var); |
| 2365 | add_referenced_var (var); |
| 2366 | } |
| 2367 | |
| 2368 | DECL_SEEN_IN_BIND_EXPR_P (var) = 1; |
| 2369 | DECL_STRUCT_FUNCTION (caller)->local_decls |
| 2370 | = tree_cons (NULL_TREE, var, |
| 2371 | DECL_STRUCT_FUNCTION (caller)->local_decls); |
| 2372 | |
| 2373 | /* Do not have the rest of GCC warn about this variable as it should |
| 2374 | not be visible to the user. */ |
| 2375 | TREE_NO_WARNING (var) = 1; |
| 2376 | |
| 2377 | declare_inline_vars (id->block, var); |
| 2378 | |
| 2379 | /* Build the use expr. If the return type of the function was |
| 2380 | promoted, convert it back to the expected type. */ |
| 2381 | use = var; |
| 2382 | if (!useless_type_conversion_p (caller_type, TREE_TYPE (var))) |
| 2383 | use = fold_convert (caller_type, var); |
| 2384 | |
| 2385 | STRIP_USELESS_TYPE_CONVERSION (use); |
| 2386 | |
| 2387 | if (DECL_BY_REFERENCE (result)) |
| 2388 | var = build_fold_addr_expr (var); |
| 2389 | |
| 2390 | done: |
| 2391 | /* Register the VAR_DECL as the equivalent for the RESULT_DECL; that |
| 2392 | way, when the RESULT_DECL is encountered, it will be |
| 2393 | automatically replaced by the VAR_DECL. */ |
| 2394 | insert_decl_map (id, result, var); |
| 2395 | |
| 2396 | /* Remember this so we can ignore it in remap_decls. */ |
| 2397 | id->retvar = var; |
| 2398 | |
| 2399 | *use_p = use; |
| 2400 | return var; |
| 2401 | } |
| 2402 | |
| 2403 | /* Returns nonzero if a function can be inlined as a tree. */ |
| 2404 | |
| 2405 | bool |
| 2406 | tree_inlinable_function_p (tree fn) |
| 2407 | { |
| 2408 | return inlinable_function_p (fn); |
| 2409 | } |
| 2410 | |
| 2411 | static const char *inline_forbidden_reason; |
| 2412 | |
| 2413 | /* A callback for walk_gimple_seq to handle tree operands. Returns |
| 2414 | NULL_TREE if a function can be inlined, otherwise sets the reason |
| 2415 | why not and returns a tree representing the offending operand. */ |
| 2416 | |
| 2417 | static tree |
| 2418 | inline_forbidden_p_op (tree *nodep, int *walk_subtrees ATTRIBUTE_UNUSED, |
| 2419 | void *fnp ATTRIBUTE_UNUSED) |
| 2420 | { |
| 2421 | tree node = *nodep; |
| 2422 | tree t; |
| 2423 | |
| 2424 | if (TREE_CODE (node) == RECORD_TYPE || TREE_CODE (node) == UNION_TYPE) |
| 2425 | { |
| 2426 | /* We cannot inline a function of the form |
| 2427 | |
| 2428 | void F (int i) { struct S { int ar[i]; } s; } |
| 2429 | |
| 2430 | Attempting to do so produces a catch-22. |
| 2431 | If walk_tree examines the TYPE_FIELDS chain of RECORD_TYPE/ |
| 2432 | UNION_TYPE nodes, then it goes into infinite recursion on a |
| 2433 | structure containing a pointer to its own type. If it doesn't, |
| 2434 | then the type node for S doesn't get adjusted properly when |
| 2435 | F is inlined. |
| 2436 | |
| 2437 | ??? This is likely no longer true, but it's too late in the 4.0 |
| 2438 | cycle to try to find out. This should be checked for 4.1. */ |
| 2439 | for (t = TYPE_FIELDS (node); t; t = TREE_CHAIN (t)) |
| 2440 | if (variably_modified_type_p (TREE_TYPE (t), NULL)) |
| 2441 | { |
| 2442 | inline_forbidden_reason |
| 2443 | = G_("function %q+F can never be inlined " |
| 2444 | "because it uses variable sized variables"); |
| 2445 | return node; |
| 2446 | } |
| 2447 | } |
| 2448 | |
| 2449 | return NULL_TREE; |
| 2450 | } |
| 2451 | |
| 2452 | |
| 2453 | /* A callback for walk_gimple_seq to handle statements. Returns |
| 2454 | non-NULL iff a function can not be inlined. Also sets the reason |
| 2455 | why. */ |
| 2456 | |
| 2457 | static tree |
| 2458 | inline_forbidden_p_stmt (gimple_stmt_iterator *gsi, bool *handled_ops_p, |
| 2459 | struct walk_stmt_info *wip) |
| 2460 | { |
| 2461 | tree fn = (tree) wip->info; |
| 2462 | tree t; |
| 2463 | gimple stmt = gsi_stmt (*gsi); |
| 2464 | |
| 2465 | switch (gimple_code (stmt)) |
| 2466 | { |
| 2467 | case GIMPLE_CALL: |
| 2468 | /* Refuse to inline alloca call unless user explicitly forced so as |
| 2469 | this may change program's memory overhead drastically when the |
| 2470 | function using alloca is called in loop. In GCC present in |
| 2471 | SPEC2000 inlining into schedule_block cause it to require 2GB of |
| 2472 | RAM instead of 256MB. */ |
| 2473 | if (gimple_alloca_call_p (stmt) |
| 2474 | && !lookup_attribute ("always_inline", DECL_ATTRIBUTES (fn))) |
| 2475 | { |
| 2476 | inline_forbidden_reason |
| 2477 | = G_("function %q+F can never be inlined because it uses " |
| 2478 | "alloca (override using the always_inline attribute)"); |
| 2479 | *handled_ops_p = true; |
| 2480 | return fn; |
| 2481 | } |
| 2482 | |
| 2483 | t = gimple_call_fndecl (stmt); |
| 2484 | if (t == NULL_TREE) |
| 2485 | break; |
| 2486 | |
| 2487 | /* We cannot inline functions that call setjmp. */ |
| 2488 | if (setjmp_call_p (t)) |
| 2489 | { |
| 2490 | inline_forbidden_reason |
| 2491 | = G_("function %q+F can never be inlined because it uses setjmp"); |
| 2492 | *handled_ops_p = true; |
| 2493 | return t; |
| 2494 | } |
| 2495 | |
| 2496 | if (DECL_BUILT_IN_CLASS (t) == BUILT_IN_NORMAL) |
| 2497 | switch (DECL_FUNCTION_CODE (t)) |
| 2498 | { |
| 2499 | /* We cannot inline functions that take a variable number of |
| 2500 | arguments. */ |
| 2501 | case BUILT_IN_VA_START: |
| 2502 | case BUILT_IN_NEXT_ARG: |
| 2503 | case BUILT_IN_VA_END: |
| 2504 | inline_forbidden_reason |
| 2505 | = G_("function %q+F can never be inlined because it " |
| 2506 | "uses variable argument lists"); |
| 2507 | *handled_ops_p = true; |
| 2508 | return t; |
| 2509 | |
| 2510 | case BUILT_IN_LONGJMP: |
| 2511 | /* We can't inline functions that call __builtin_longjmp at |
| 2512 | all. The non-local goto machinery really requires the |
| 2513 | destination be in a different function. If we allow the |
| 2514 | function calling __builtin_longjmp to be inlined into the |
| 2515 | function calling __builtin_setjmp, Things will Go Awry. */ |
| 2516 | inline_forbidden_reason |
| 2517 | = G_("function %q+F can never be inlined because " |
| 2518 | "it uses setjmp-longjmp exception handling"); |
| 2519 | *handled_ops_p = true; |
| 2520 | return t; |
| 2521 | |
| 2522 | case BUILT_IN_NONLOCAL_GOTO: |
| 2523 | /* Similarly. */ |
| 2524 | inline_forbidden_reason |
| 2525 | = G_("function %q+F can never be inlined because " |
| 2526 | "it uses non-local goto"); |
| 2527 | *handled_ops_p = true; |
| 2528 | return t; |
| 2529 | |
| 2530 | case BUILT_IN_RETURN: |
| 2531 | case BUILT_IN_APPLY_ARGS: |
| 2532 | /* If a __builtin_apply_args caller would be inlined, |
| 2533 | it would be saving arguments of the function it has |
| 2534 | been inlined into. Similarly __builtin_return would |
| 2535 | return from the function the inline has been inlined into. */ |
| 2536 | inline_forbidden_reason |
| 2537 | = G_("function %q+F can never be inlined because " |
| 2538 | "it uses __builtin_return or __builtin_apply_args"); |
| 2539 | *handled_ops_p = true; |
| 2540 | return t; |
| 2541 | |
| 2542 | default: |
| 2543 | break; |
| 2544 | } |
| 2545 | break; |
| 2546 | |
| 2547 | case GIMPLE_GOTO: |
| 2548 | t = gimple_goto_dest (stmt); |
| 2549 | |
| 2550 | /* We will not inline a function which uses computed goto. The |
| 2551 | addresses of its local labels, which may be tucked into |
| 2552 | global storage, are of course not constant across |
| 2553 | instantiations, which causes unexpected behavior. */ |
| 2554 | if (TREE_CODE (t) != LABEL_DECL) |
| 2555 | { |
| 2556 | inline_forbidden_reason |
| 2557 | = G_("function %q+F can never be inlined " |
| 2558 | "because it contains a computed goto"); |
| 2559 | *handled_ops_p = true; |
| 2560 | return t; |
| 2561 | } |
| 2562 | break; |
| 2563 | |
| 2564 | case GIMPLE_LABEL: |
| 2565 | t = gimple_label_label (stmt); |
| 2566 | if (DECL_NONLOCAL (t)) |
| 2567 | { |
| 2568 | /* We cannot inline a function that receives a non-local goto |
| 2569 | because we cannot remap the destination label used in the |
| 2570 | function that is performing the non-local goto. */ |
| 2571 | inline_forbidden_reason |
| 2572 | = G_("function %q+F can never be inlined " |
| 2573 | "because it receives a non-local goto"); |
| 2574 | *handled_ops_p = true; |
| 2575 | return t; |
| 2576 | } |
| 2577 | break; |
| 2578 | |
| 2579 | default: |
| 2580 | break; |
| 2581 | } |
| 2582 | |
| 2583 | *handled_ops_p = false; |
| 2584 | return NULL_TREE; |
| 2585 | } |
| 2586 | |
| 2587 | |
| 2588 | static tree |
| 2589 | inline_forbidden_p_2 (tree *nodep, int *walk_subtrees, |
| 2590 | void *fnp) |
| 2591 | { |
| 2592 | tree node = *nodep; |
| 2593 | tree fn = (tree) fnp; |
| 2594 | |
| 2595 | if (TREE_CODE (node) == LABEL_DECL && DECL_CONTEXT (node) == fn) |
| 2596 | { |
| 2597 | inline_forbidden_reason |
| 2598 | = G_("function %q+F can never be inlined " |
| 2599 | "because it saves address of local label in a static variable"); |
| 2600 | return node; |
| 2601 | } |
| 2602 | |
| 2603 | if (TYPE_P (node)) |
| 2604 | *walk_subtrees = 0; |
| 2605 | |
| 2606 | return NULL_TREE; |
| 2607 | } |
| 2608 | |
| 2609 | /* Return true if FNDECL is a function that cannot be inlined into |
| 2610 | another one. */ |
| 2611 | |
| 2612 | static bool |
| 2613 | inline_forbidden_p (tree fndecl) |
| 2614 | { |
| 2615 | location_t saved_loc = input_location; |
| 2616 | struct function *fun = DECL_STRUCT_FUNCTION (fndecl); |
| 2617 | tree step; |
| 2618 | struct walk_stmt_info wi; |
| 2619 | struct pointer_set_t *visited_nodes; |
| 2620 | basic_block bb; |
| 2621 | bool forbidden_p = false; |
| 2622 | |
| 2623 | visited_nodes = pointer_set_create (); |
| 2624 | memset (&wi, 0, sizeof (wi)); |
| 2625 | wi.info = (void *) fndecl; |
| 2626 | wi.pset = visited_nodes; |
| 2627 | |
| 2628 | FOR_EACH_BB_FN (bb, fun) |
| 2629 | { |
| 2630 | gimple ret; |
| 2631 | gimple_seq seq = bb_seq (bb); |
| 2632 | ret = walk_gimple_seq (seq, inline_forbidden_p_stmt, |
| 2633 | inline_forbidden_p_op, &wi); |
| 2634 | forbidden_p = (ret != NULL); |
| 2635 | if (forbidden_p) |
| 2636 | goto egress; |
| 2637 | } |
| 2638 | |
| 2639 | for (step = fun->local_decls; step; step = TREE_CHAIN (step)) |
| 2640 | { |
| 2641 | tree decl = TREE_VALUE (step); |
| 2642 | if (TREE_CODE (decl) == VAR_DECL |
| 2643 | && TREE_STATIC (decl) |
| 2644 | && !DECL_EXTERNAL (decl) |
| 2645 | && DECL_INITIAL (decl)) |
| 2646 | { |
| 2647 | tree ret; |
| 2648 | ret = walk_tree_without_duplicates (&DECL_INITIAL (decl), |
| 2649 | inline_forbidden_p_2, fndecl); |
| 2650 | forbidden_p = (ret != NULL); |
| 2651 | if (forbidden_p) |
| 2652 | goto egress; |
| 2653 | } |
| 2654 | } |
| 2655 | |
| 2656 | egress: |
| 2657 | pointer_set_destroy (visited_nodes); |
| 2658 | input_location = saved_loc; |
| 2659 | return forbidden_p; |
| 2660 | } |
| 2661 | |
| 2662 | /* Returns nonzero if FN is a function that does not have any |
| 2663 | fundamental inline blocking properties. */ |
| 2664 | |
| 2665 | static bool |
| 2666 | inlinable_function_p (tree fn) |
| 2667 | { |
| 2668 | bool inlinable = true; |
| 2669 | bool do_warning; |
| 2670 | tree always_inline; |
| 2671 | |
| 2672 | /* If we've already decided this function shouldn't be inlined, |
| 2673 | there's no need to check again. */ |
| 2674 | if (DECL_UNINLINABLE (fn)) |
| 2675 | return false; |
| 2676 | |
| 2677 | /* We only warn for functions declared `inline' by the user. */ |
| 2678 | do_warning = (warn_inline |
| 2679 | && DECL_DECLARED_INLINE_P (fn) |
| 2680 | && !DECL_NO_INLINE_WARNING_P (fn) |
| 2681 | && !DECL_IN_SYSTEM_HEADER (fn)); |
| 2682 | |
| 2683 | always_inline = lookup_attribute ("always_inline", DECL_ATTRIBUTES (fn)); |
| 2684 | |
| 2685 | if (flag_no_inline |
| 2686 | && always_inline == NULL) |
| 2687 | { |
| 2688 | if (do_warning) |
| 2689 | warning (OPT_Winline, "function %q+F can never be inlined because it " |
| 2690 | "is suppressed using -fno-inline", fn); |
| 2691 | inlinable = false; |
| 2692 | } |
| 2693 | |
| 2694 | /* Don't auto-inline anything that might not be bound within |
| 2695 | this unit of translation. */ |
| 2696 | else if (!DECL_DECLARED_INLINE_P (fn) |
| 2697 | && DECL_REPLACEABLE_P (fn)) |
| 2698 | inlinable = false; |
| 2699 | |
| 2700 | else if (!function_attribute_inlinable_p (fn)) |
| 2701 | { |
| 2702 | if (do_warning) |
| 2703 | warning (OPT_Winline, "function %q+F can never be inlined because it " |
| 2704 | "uses attributes conflicting with inlining", fn); |
| 2705 | inlinable = false; |
| 2706 | } |
| 2707 | |
| 2708 | else if (inline_forbidden_p (fn)) |
| 2709 | { |
| 2710 | /* See if we should warn about uninlinable functions. Previously, |
| 2711 | some of these warnings would be issued while trying to expand |
| 2712 | the function inline, but that would cause multiple warnings |
| 2713 | about functions that would for example call alloca. But since |
| 2714 | this a property of the function, just one warning is enough. |
| 2715 | As a bonus we can now give more details about the reason why a |
| 2716 | function is not inlinable. */ |
| 2717 | if (always_inline) |
| 2718 | sorry (inline_forbidden_reason, fn); |
| 2719 | else if (do_warning) |
| 2720 | warning (OPT_Winline, inline_forbidden_reason, fn); |
| 2721 | |
| 2722 | inlinable = false; |
| 2723 | } |
| 2724 | |
| 2725 | /* Squirrel away the result so that we don't have to check again. */ |
| 2726 | DECL_UNINLINABLE (fn) = !inlinable; |
| 2727 | |
| 2728 | return inlinable; |
| 2729 | } |
| 2730 | |
| 2731 | /* Estimate the cost of a memory move. Use machine dependent |
| 2732 | word size and take possible memcpy call into account. */ |
| 2733 | |
| 2734 | int |
| 2735 | estimate_move_cost (tree type) |
| 2736 | { |
| 2737 | HOST_WIDE_INT size; |
| 2738 | |
| 2739 | gcc_assert (!VOID_TYPE_P (type)); |
| 2740 | |
| 2741 | size = int_size_in_bytes (type); |
| 2742 | |
| 2743 | if (size < 0 || size > MOVE_MAX_PIECES * MOVE_RATIO (!optimize_size)) |
| 2744 | /* Cost of a memcpy call, 3 arguments and the call. */ |
| 2745 | return 4; |
| 2746 | else |
| 2747 | return ((size + MOVE_MAX_PIECES - 1) / MOVE_MAX_PIECES); |
| 2748 | } |
| 2749 | |
| 2750 | /* Returns cost of operation CODE, according to WEIGHTS */ |
| 2751 | |
| 2752 | static int |
| 2753 | estimate_operator_cost (enum tree_code code, eni_weights *weights) |
| 2754 | { |
| 2755 | switch (code) |
| 2756 | { |
| 2757 | /* These are "free" conversions, or their presumed cost |
| 2758 | is folded into other operations. */ |
| 2759 | case RANGE_EXPR: |
| 2760 | CASE_CONVERT: |
| 2761 | case COMPLEX_EXPR: |
| 2762 | case PAREN_EXPR: |
| 2763 | return 0; |
| 2764 | |
| 2765 | /* Assign cost of 1 to usual operations. |
| 2766 | ??? We may consider mapping RTL costs to this. */ |
| 2767 | case COND_EXPR: |
| 2768 | case VEC_COND_EXPR: |
| 2769 | |
| 2770 | case PLUS_EXPR: |
| 2771 | case POINTER_PLUS_EXPR: |
| 2772 | case MINUS_EXPR: |
| 2773 | case MULT_EXPR: |
| 2774 | |
| 2775 | case FIXED_CONVERT_EXPR: |
| 2776 | case FIX_TRUNC_EXPR: |
| 2777 | |
| 2778 | case NEGATE_EXPR: |
| 2779 | case FLOAT_EXPR: |
| 2780 | case MIN_EXPR: |
| 2781 | case MAX_EXPR: |
| 2782 | case ABS_EXPR: |
| 2783 | |
| 2784 | case LSHIFT_EXPR: |
| 2785 | case RSHIFT_EXPR: |
| 2786 | case LROTATE_EXPR: |
| 2787 | case RROTATE_EXPR: |
| 2788 | case VEC_LSHIFT_EXPR: |
| 2789 | case VEC_RSHIFT_EXPR: |
| 2790 | |
| 2791 | case BIT_IOR_EXPR: |
| 2792 | case BIT_XOR_EXPR: |
| 2793 | case BIT_AND_EXPR: |
| 2794 | case BIT_NOT_EXPR: |
| 2795 | |
| 2796 | case TRUTH_ANDIF_EXPR: |
| 2797 | case TRUTH_ORIF_EXPR: |
| 2798 | case TRUTH_AND_EXPR: |
| 2799 | case TRUTH_OR_EXPR: |
| 2800 | case TRUTH_XOR_EXPR: |
| 2801 | case TRUTH_NOT_EXPR: |
| 2802 | |
| 2803 | case LT_EXPR: |
| 2804 | case LE_EXPR: |
| 2805 | case GT_EXPR: |
| 2806 | case GE_EXPR: |
| 2807 | case EQ_EXPR: |
| 2808 | case NE_EXPR: |
| 2809 | case ORDERED_EXPR: |
| 2810 | case UNORDERED_EXPR: |
| 2811 | |
| 2812 | case UNLT_EXPR: |
| 2813 | case UNLE_EXPR: |
| 2814 | case UNGT_EXPR: |
| 2815 | case UNGE_EXPR: |
| 2816 | case UNEQ_EXPR: |
| 2817 | case LTGT_EXPR: |
| 2818 | |
| 2819 | case CONJ_EXPR: |
| 2820 | |
| 2821 | case PREDECREMENT_EXPR: |
| 2822 | case PREINCREMENT_EXPR: |
| 2823 | case POSTDECREMENT_EXPR: |
| 2824 | case POSTINCREMENT_EXPR: |
| 2825 | |
| 2826 | case REALIGN_LOAD_EXPR: |
| 2827 | |
| 2828 | case REDUC_MAX_EXPR: |
| 2829 | case REDUC_MIN_EXPR: |
| 2830 | case REDUC_PLUS_EXPR: |
| 2831 | case WIDEN_SUM_EXPR: |
| 2832 | case WIDEN_MULT_EXPR: |
| 2833 | case DOT_PROD_EXPR: |
| 2834 | |
| 2835 | case VEC_WIDEN_MULT_HI_EXPR: |
| 2836 | case VEC_WIDEN_MULT_LO_EXPR: |
| 2837 | case VEC_UNPACK_HI_EXPR: |
| 2838 | case VEC_UNPACK_LO_EXPR: |
| 2839 | case VEC_UNPACK_FLOAT_HI_EXPR: |
| 2840 | case VEC_UNPACK_FLOAT_LO_EXPR: |
| 2841 | case VEC_PACK_TRUNC_EXPR: |
| 2842 | case VEC_PACK_SAT_EXPR: |
| 2843 | case VEC_PACK_FIX_TRUNC_EXPR: |
| 2844 | case VEC_EXTRACT_EVEN_EXPR: |
| 2845 | case VEC_EXTRACT_ODD_EXPR: |
| 2846 | case VEC_INTERLEAVE_HIGH_EXPR: |
| 2847 | case VEC_INTERLEAVE_LOW_EXPR: |
| 2848 | |
| 2849 | return 1; |
| 2850 | |
| 2851 | /* Few special cases of expensive operations. This is useful |
| 2852 | to avoid inlining on functions having too many of these. */ |
| 2853 | case TRUNC_DIV_EXPR: |
| 2854 | case CEIL_DIV_EXPR: |
| 2855 | case FLOOR_DIV_EXPR: |
| 2856 | case ROUND_DIV_EXPR: |
| 2857 | case EXACT_DIV_EXPR: |
| 2858 | case TRUNC_MOD_EXPR: |
| 2859 | case CEIL_MOD_EXPR: |
| 2860 | case FLOOR_MOD_EXPR: |
| 2861 | case ROUND_MOD_EXPR: |
| 2862 | case RDIV_EXPR: |
| 2863 | return weights->div_mod_cost; |
| 2864 | |
| 2865 | default: |
| 2866 | /* We expect a copy assignment with no operator. */ |
| 2867 | gcc_assert (get_gimple_rhs_class (code) == GIMPLE_SINGLE_RHS); |
| 2868 | return 0; |
| 2869 | } |
| 2870 | } |
| 2871 | |
| 2872 | |
| 2873 | /* Estimate number of instructions that will be created by expanding |
| 2874 | the statements in the statement sequence STMTS. |
| 2875 | WEIGHTS contains weights attributed to various constructs. */ |
| 2876 | |
| 2877 | static |
| 2878 | int estimate_num_insns_seq (gimple_seq stmts, eni_weights *weights) |
| 2879 | { |
| 2880 | int cost; |
| 2881 | gimple_stmt_iterator gsi; |
| 2882 | |
| 2883 | cost = 0; |
| 2884 | for (gsi = gsi_start (stmts); !gsi_end_p (gsi); gsi_next (&gsi)) |
| 2885 | cost += estimate_num_insns (gsi_stmt (gsi), weights); |
| 2886 | |
| 2887 | return cost; |
| 2888 | } |
| 2889 | |
| 2890 | |
| 2891 | /* Estimate number of instructions that will be created by expanding STMT. |
| 2892 | WEIGHTS contains weights attributed to various constructs. */ |
| 2893 | |
| 2894 | int |
| 2895 | estimate_num_insns (gimple stmt, eni_weights *weights) |
| 2896 | { |
| 2897 | unsigned cost, i; |
| 2898 | enum gimple_code code = gimple_code (stmt); |
| 2899 | tree lhs; |
| 2900 | |
| 2901 | switch (code) |
| 2902 | { |
| 2903 | case GIMPLE_ASSIGN: |
| 2904 | /* Try to estimate the cost of assignments. We have three cases to |
| 2905 | deal with: |
| 2906 | 1) Simple assignments to registers; |
| 2907 | 2) Stores to things that must live in memory. This includes |
| 2908 | "normal" stores to scalars, but also assignments of large |
| 2909 | structures, or constructors of big arrays; |
| 2910 | |
| 2911 | Let us look at the first two cases, assuming we have "a = b + C": |
| 2912 | <GIMPLE_ASSIGN <var_decl "a"> |
| 2913 | <plus_expr <var_decl "b"> <constant C>> |
| 2914 | If "a" is a GIMPLE register, the assignment to it is free on almost |
| 2915 | any target, because "a" usually ends up in a real register. Hence |
| 2916 | the only cost of this expression comes from the PLUS_EXPR, and we |
| 2917 | can ignore the GIMPLE_ASSIGN. |
| 2918 | If "a" is not a GIMPLE register, the assignment to "a" will most |
| 2919 | likely be a real store, so the cost of the GIMPLE_ASSIGN is the cost |
| 2920 | of moving something into "a", which we compute using the function |
| 2921 | estimate_move_cost. */ |
| 2922 | lhs = gimple_assign_lhs (stmt); |
| 2923 | if (is_gimple_reg (lhs)) |
| 2924 | cost = 0; |
| 2925 | else |
| 2926 | cost = estimate_move_cost (TREE_TYPE (lhs)); |
| 2927 | |
| 2928 | cost += estimate_operator_cost (gimple_assign_rhs_code (stmt), weights); |
| 2929 | break; |
| 2930 | |
| 2931 | case GIMPLE_COND: |
| 2932 | cost = 1 + estimate_operator_cost (gimple_cond_code (stmt), weights); |
| 2933 | break; |
| 2934 | |
| 2935 | case GIMPLE_SWITCH: |
| 2936 | /* Take into account cost of the switch + guess 2 conditional jumps for |
| 2937 | each case label. |
| 2938 | |
| 2939 | TODO: once the switch expansion logic is sufficiently separated, we can |
| 2940 | do better job on estimating cost of the switch. */ |
| 2941 | cost = gimple_switch_num_labels (stmt) * 2; |
| 2942 | break; |
| 2943 | |
| 2944 | case GIMPLE_CALL: |
| 2945 | { |
| 2946 | tree decl = gimple_call_fndecl (stmt); |
| 2947 | tree addr = gimple_call_fn (stmt); |
| 2948 | tree funtype = TREE_TYPE (addr); |
| 2949 | |
| 2950 | if (POINTER_TYPE_P (funtype)) |
| 2951 | funtype = TREE_TYPE (funtype); |
| 2952 | |
| 2953 | if (decl && DECL_BUILT_IN_CLASS (decl) == BUILT_IN_MD) |
| 2954 | cost = weights->target_builtin_call_cost; |
| 2955 | else |
| 2956 | cost = weights->call_cost; |
| 2957 | |
| 2958 | if (decl && DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL) |
| 2959 | switch (DECL_FUNCTION_CODE (decl)) |
| 2960 | { |
| 2961 | case BUILT_IN_CONSTANT_P: |
| 2962 | return 0; |
| 2963 | case BUILT_IN_EXPECT: |
| 2964 | cost = 0; |
| 2965 | break; |
| 2966 | |
| 2967 | /* Prefetch instruction is not expensive. */ |
| 2968 | case BUILT_IN_PREFETCH: |
| 2969 | cost = weights->target_builtin_call_cost; |
| 2970 | break; |
| 2971 | |
| 2972 | default: |
| 2973 | break; |
| 2974 | } |
| 2975 | |
| 2976 | if (decl) |
| 2977 | funtype = TREE_TYPE (decl); |
| 2978 | |
| 2979 | /* Our cost must be kept in sync with |
| 2980 | cgraph_estimate_size_after_inlining that does use function |
| 2981 | declaration to figure out the arguments. */ |
| 2982 | if (decl && DECL_ARGUMENTS (decl)) |
| 2983 | { |
| 2984 | tree arg; |
| 2985 | for (arg = DECL_ARGUMENTS (decl); arg; arg = TREE_CHAIN (arg)) |
| 2986 | cost += estimate_move_cost (TREE_TYPE (arg)); |
| 2987 | } |
| 2988 | else if (funtype && prototype_p (funtype)) |
| 2989 | { |
| 2990 | tree t; |
| 2991 | for (t = TYPE_ARG_TYPES (funtype); t; t = TREE_CHAIN (t)) |
| 2992 | if (!VOID_TYPE_P (TREE_VALUE (t))) |
| 2993 | cost += estimate_move_cost (TREE_VALUE (t)); |
| 2994 | } |
| 2995 | else |
| 2996 | { |
| 2997 | for (i = 0; i < gimple_call_num_args (stmt); i++) |
| 2998 | { |
| 2999 | tree arg = gimple_call_arg (stmt, i); |
| 3000 | cost += estimate_move_cost (TREE_TYPE (arg)); |
| 3001 | } |
| 3002 | } |
| 3003 | |
| 3004 | break; |
| 3005 | } |
| 3006 | |
| 3007 | case GIMPLE_GOTO: |
| 3008 | case GIMPLE_LABEL: |
| 3009 | case GIMPLE_NOP: |
| 3010 | case GIMPLE_PHI: |
| 3011 | case GIMPLE_RETURN: |
| 3012 | case GIMPLE_CHANGE_DYNAMIC_TYPE: |
| 3013 | case GIMPLE_PREDICT: |
| 3014 | return 0; |
| 3015 | |
| 3016 | case GIMPLE_ASM: |
| 3017 | case GIMPLE_RESX: |
| 3018 | return 1; |
| 3019 | |
| 3020 | case GIMPLE_BIND: |
| 3021 | return estimate_num_insns_seq (gimple_bind_body (stmt), weights); |
| 3022 | |
| 3023 | case GIMPLE_EH_FILTER: |
| 3024 | return estimate_num_insns_seq (gimple_eh_filter_failure (stmt), weights); |
| 3025 | |
| 3026 | case GIMPLE_CATCH: |
| 3027 | return estimate_num_insns_seq (gimple_catch_handler (stmt), weights); |
| 3028 | |
| 3029 | case GIMPLE_TRY: |
| 3030 | return (estimate_num_insns_seq (gimple_try_eval (stmt), weights) |
| 3031 | + estimate_num_insns_seq (gimple_try_cleanup (stmt), weights)); |
| 3032 | |
| 3033 | /* OpenMP directives are generally very expensive. */ |
| 3034 | |
| 3035 | case GIMPLE_OMP_RETURN: |
| 3036 | case GIMPLE_OMP_SECTIONS_SWITCH: |
| 3037 | case GIMPLE_OMP_ATOMIC_STORE: |
| 3038 | case GIMPLE_OMP_CONTINUE: |
| 3039 | /* ...except these, which are cheap. */ |
| 3040 | return 0; |
| 3041 | |
| 3042 | case GIMPLE_OMP_ATOMIC_LOAD: |
| 3043 | return weights->omp_cost; |
| 3044 | |
| 3045 | case GIMPLE_OMP_FOR: |
| 3046 | return (weights->omp_cost |
| 3047 | + estimate_num_insns_seq (gimple_omp_body (stmt), weights) |
| 3048 | + estimate_num_insns_seq (gimple_omp_for_pre_body (stmt), weights)); |
| 3049 | |
| 3050 | case GIMPLE_OMP_PARALLEL: |
| 3051 | case GIMPLE_OMP_TASK: |
| 3052 | case GIMPLE_OMP_CRITICAL: |
| 3053 | case GIMPLE_OMP_MASTER: |
| 3054 | case GIMPLE_OMP_ORDERED: |
| 3055 | case GIMPLE_OMP_SECTION: |
| 3056 | case GIMPLE_OMP_SECTIONS: |
| 3057 | case GIMPLE_OMP_SINGLE: |
| 3058 | return (weights->omp_cost |
| 3059 | + estimate_num_insns_seq (gimple_omp_body (stmt), weights)); |
| 3060 | |
| 3061 | default: |
| 3062 | gcc_unreachable (); |
| 3063 | } |
| 3064 | |
| 3065 | return cost; |
| 3066 | } |
| 3067 | |
| 3068 | /* Estimate number of instructions that will be created by expanding |
| 3069 | function FNDECL. WEIGHTS contains weights attributed to various |
| 3070 | constructs. */ |
| 3071 | |
| 3072 | int |
| 3073 | estimate_num_insns_fn (tree fndecl, eni_weights *weights) |
| 3074 | { |
| 3075 | struct function *my_function = DECL_STRUCT_FUNCTION (fndecl); |
| 3076 | gimple_stmt_iterator bsi; |
| 3077 | basic_block bb; |
| 3078 | int n = 0; |
| 3079 | |
| 3080 | gcc_assert (my_function && my_function->cfg); |
| 3081 | FOR_EACH_BB_FN (bb, my_function) |
| 3082 | { |
| 3083 | for (bsi = gsi_start_bb (bb); !gsi_end_p (bsi); gsi_next (&bsi)) |
| 3084 | n += estimate_num_insns (gsi_stmt (bsi), weights); |
| 3085 | } |
| 3086 | |
| 3087 | return n; |
| 3088 | } |
| 3089 | |
| 3090 | |
| 3091 | /* Initializes weights used by estimate_num_insns. */ |
| 3092 | |
| 3093 | void |
| 3094 | init_inline_once (void) |
| 3095 | { |
| 3096 | eni_inlining_weights.call_cost = PARAM_VALUE (PARAM_INLINE_CALL_COST); |
| 3097 | eni_inlining_weights.target_builtin_call_cost = 1; |
| 3098 | eni_inlining_weights.div_mod_cost = 10; |
| 3099 | eni_inlining_weights.omp_cost = 40; |
| 3100 | |
| 3101 | eni_size_weights.call_cost = 1; |
| 3102 | eni_size_weights.target_builtin_call_cost = 1; |
| 3103 | eni_size_weights.div_mod_cost = 1; |
| 3104 | eni_size_weights.omp_cost = 40; |
| 3105 | |
| 3106 | /* Estimating time for call is difficult, since we have no idea what the |
| 3107 | called function does. In the current uses of eni_time_weights, |
| 3108 | underestimating the cost does less harm than overestimating it, so |
| 3109 | we choose a rather small value here. */ |
| 3110 | eni_time_weights.call_cost = 10; |
| 3111 | eni_time_weights.target_builtin_call_cost = 10; |
| 3112 | eni_time_weights.div_mod_cost = 10; |
| 3113 | eni_time_weights.omp_cost = 40; |
| 3114 | } |
| 3115 | |
| 3116 | /* Estimate the number of instructions in a gimple_seq. */ |
| 3117 | |
| 3118 | int |
| 3119 | count_insns_seq (gimple_seq seq, eni_weights *weights) |
| 3120 | { |
| 3121 | gimple_stmt_iterator gsi; |
| 3122 | int n = 0; |
| 3123 | for (gsi = gsi_start (seq); !gsi_end_p (gsi); gsi_next (&gsi)) |
| 3124 | n += estimate_num_insns (gsi_stmt (gsi), weights); |
| 3125 | |
| 3126 | return n; |
| 3127 | } |
| 3128 | |
| 3129 | |
| 3130 | /* Install new lexical TREE_BLOCK underneath 'current_block'. */ |
| 3131 | |
| 3132 | static void |
| 3133 | prepend_lexical_block (tree current_block, tree new_block) |
| 3134 | { |
| 3135 | BLOCK_CHAIN (new_block) = BLOCK_SUBBLOCKS (current_block); |
| 3136 | BLOCK_SUBBLOCKS (current_block) = new_block; |
| 3137 | BLOCK_SUPERCONTEXT (new_block) = current_block; |
| 3138 | } |
| 3139 | |
| 3140 | /* Fetch callee declaration from the call graph edge going from NODE and |
| 3141 | associated with STMR call statement. Return NULL_TREE if not found. */ |
| 3142 | static tree |
| 3143 | get_indirect_callee_fndecl (struct cgraph_node *node, gimple stmt) |
| 3144 | { |
| 3145 | struct cgraph_edge *cs; |
| 3146 | |
| 3147 | cs = cgraph_edge (node, stmt); |
| 3148 | if (cs) |
| 3149 | return cs->callee->decl; |
| 3150 | |
| 3151 | return NULL_TREE; |
| 3152 | } |
| 3153 | |
| 3154 | /* If STMT is a GIMPLE_CALL, replace it with its inline expansion. */ |
| 3155 | |
| 3156 | static bool |
| 3157 | expand_call_inline (basic_block bb, gimple stmt, copy_body_data *id) |
| 3158 | { |
| 3159 | tree retvar, use_retvar; |
| 3160 | tree fn; |
| 3161 | struct pointer_map_t *st; |
| 3162 | tree return_slot; |
| 3163 | tree modify_dest; |
| 3164 | location_t saved_location; |
| 3165 | struct cgraph_edge *cg_edge; |
| 3166 | const char *reason; |
| 3167 | basic_block return_block; |
| 3168 | edge e; |
| 3169 | gimple_stmt_iterator gsi, stmt_gsi; |
| 3170 | bool successfully_inlined = FALSE; |
| 3171 | bool purge_dead_abnormal_edges; |
| 3172 | tree t_step; |
| 3173 | tree var; |
| 3174 | |
| 3175 | /* Set input_location here so we get the right instantiation context |
| 3176 | if we call instantiate_decl from inlinable_function_p. */ |
| 3177 | saved_location = input_location; |
| 3178 | if (gimple_has_location (stmt)) |
| 3179 | input_location = gimple_location (stmt); |
| 3180 | |
| 3181 | /* From here on, we're only interested in CALL_EXPRs. */ |
| 3182 | if (gimple_code (stmt) != GIMPLE_CALL) |
| 3183 | goto egress; |
| 3184 | |
| 3185 | /* First, see if we can figure out what function is being called. |
| 3186 | If we cannot, then there is no hope of inlining the function. */ |
| 3187 | fn = gimple_call_fndecl (stmt); |
| 3188 | if (!fn) |
| 3189 | { |
| 3190 | fn = get_indirect_callee_fndecl (id->dst_node, stmt); |
| 3191 | if (!fn) |
| 3192 | goto egress; |
| 3193 | } |
| 3194 | |
| 3195 | /* Turn forward declarations into real ones. */ |
| 3196 | fn = cgraph_node (fn)->decl; |
| 3197 | |
| 3198 | /* If FN is a declaration of a function in a nested scope that was |
| 3199 | globally declared inline, we don't set its DECL_INITIAL. |
| 3200 | However, we can't blindly follow DECL_ABSTRACT_ORIGIN because the |
| 3201 | C++ front-end uses it for cdtors to refer to their internal |
| 3202 | declarations, that are not real functions. Fortunately those |
| 3203 | don't have trees to be saved, so we can tell by checking their |
| 3204 | gimple_body. */ |
| 3205 | if (!DECL_INITIAL (fn) |
| 3206 | && DECL_ABSTRACT_ORIGIN (fn) |
| 3207 | && gimple_has_body_p (DECL_ABSTRACT_ORIGIN (fn))) |
| 3208 | fn = DECL_ABSTRACT_ORIGIN (fn); |
| 3209 | |
| 3210 | /* Objective C and fortran still calls tree_rest_of_compilation directly. |
| 3211 | Kill this check once this is fixed. */ |
| 3212 | if (!id->dst_node->analyzed) |
| 3213 | goto egress; |
| 3214 | |
| 3215 | cg_edge = cgraph_edge (id->dst_node, stmt); |
| 3216 | |
| 3217 | /* Constant propagation on argument done during previous inlining |
| 3218 | may create new direct call. Produce an edge for it. */ |
| 3219 | if (!cg_edge) |
| 3220 | { |
| 3221 | struct cgraph_node *dest = cgraph_node (fn); |
| 3222 | |
| 3223 | /* We have missing edge in the callgraph. This can happen in one case |
| 3224 | where previous inlining turned indirect call into direct call by |
| 3225 | constant propagating arguments. In all other cases we hit a bug |
| 3226 | (incorrect node sharing is most common reason for missing edges. */ |
| 3227 | gcc_assert (dest->needed); |
| 3228 | cgraph_create_edge (id->dst_node, dest, stmt, |
| 3229 | bb->count, CGRAPH_FREQ_BASE, |
| 3230 | bb->loop_depth)->inline_failed |
| 3231 | = N_("originally indirect function call not considered for inlining"); |
| 3232 | if (dump_file) |
| 3233 | { |
| 3234 | fprintf (dump_file, "Created new direct edge to %s", |
| 3235 | cgraph_node_name (dest)); |
| 3236 | } |
| 3237 | goto egress; |
| 3238 | } |
| 3239 | |
| 3240 | /* Don't try to inline functions that are not well-suited to |
| 3241 | inlining. */ |
| 3242 | if (!cgraph_inline_p (cg_edge, &reason)) |
| 3243 | { |
| 3244 | /* If this call was originally indirect, we do not want to emit any |
| 3245 | inlining related warnings or sorry messages because there are no |
| 3246 | guarantees regarding those. */ |
| 3247 | if (cg_edge->indirect_call) |
| 3248 | goto egress; |
| 3249 | |
| 3250 | if (lookup_attribute ("always_inline", DECL_ATTRIBUTES (fn)) |
| 3251 | /* Avoid warnings during early inline pass. */ |
| 3252 | && cgraph_global_info_ready) |
| 3253 | { |
| 3254 | sorry ("inlining failed in call to %q+F: %s", fn, reason); |
| 3255 | sorry ("called from here"); |
| 3256 | } |
| 3257 | else if (warn_inline && DECL_DECLARED_INLINE_P (fn) |
| 3258 | && !DECL_IN_SYSTEM_HEADER (fn) |
| 3259 | && strlen (reason) |
| 3260 | && !lookup_attribute ("noinline", DECL_ATTRIBUTES (fn)) |
| 3261 | /* Avoid warnings during early inline pass. */ |
| 3262 | && cgraph_global_info_ready |
| 3263 | && strcmp(reason, "call is unlikely and code size would grow")) |
| 3264 | { |
| 3265 | warning (OPT_Winline, "inlining failed in call to %q+F: %s", |
| 3266 | fn, reason); |
| 3267 | warning (OPT_Winline, "called from here"); |
| 3268 | } |
| 3269 | goto egress; |
| 3270 | } |
| 3271 | fn = cg_edge->callee->decl; |
| 3272 | |
| 3273 | #ifdef ENABLE_CHECKING |
| 3274 | if (cg_edge->callee->decl != id->dst_node->decl) |
| 3275 | verify_cgraph_node (cg_edge->callee); |
| 3276 | #endif |
| 3277 | |
| 3278 | /* We will be inlining this callee. */ |
| 3279 | id->eh_region = lookup_stmt_eh_region (stmt); |
| 3280 | |
| 3281 | /* Split the block holding the GIMPLE_CALL. */ |
| 3282 | e = split_block (bb, stmt); |
| 3283 | bb = e->src; |
| 3284 | return_block = e->dest; |
| 3285 | remove_edge (e); |
| 3286 | |
| 3287 | /* split_block splits after the statement; work around this by |
| 3288 | moving the call into the second block manually. Not pretty, |
| 3289 | but seems easier than doing the CFG manipulation by hand |
| 3290 | when the GIMPLE_CALL is in the last statement of BB. */ |
| 3291 | stmt_gsi = gsi_last_bb (bb); |
| 3292 | gsi_remove (&stmt_gsi, false); |
| 3293 | |
| 3294 | /* If the GIMPLE_CALL was in the last statement of BB, it may have |
| 3295 | been the source of abnormal edges. In this case, schedule |
| 3296 | the removal of dead abnormal edges. */ |
| 3297 | gsi = gsi_start_bb (return_block); |
| 3298 | if (gsi_end_p (gsi)) |
| 3299 | { |
| 3300 | gsi_insert_after (&gsi, stmt, GSI_NEW_STMT); |
| 3301 | purge_dead_abnormal_edges = true; |
| 3302 | } |
| 3303 | else |
| 3304 | { |
| 3305 | gsi_insert_before (&gsi, stmt, GSI_NEW_STMT); |
| 3306 | purge_dead_abnormal_edges = false; |
| 3307 | } |
| 3308 | |
| 3309 | stmt_gsi = gsi_start_bb (return_block); |
| 3310 | |
| 3311 | /* Build a block containing code to initialize the arguments, the |
| 3312 | actual inline expansion of the body, and a label for the return |
| 3313 | statements within the function to jump to. The type of the |
| 3314 | statement expression is the return type of the function call. */ |
| 3315 | id->block = make_node (BLOCK); |
| 3316 | BLOCK_ABSTRACT_ORIGIN (id->block) = fn; |
| 3317 | BLOCK_SOURCE_LOCATION (id->block) = input_location; |
| 3318 | prepend_lexical_block (gimple_block (stmt), id->block); |
| 3319 | |
| 3320 | /* Local declarations will be replaced by their equivalents in this |
| 3321 | map. */ |
| 3322 | st = id->decl_map; |
| 3323 | id->decl_map = pointer_map_create (); |
| 3324 | |
| 3325 | /* Record the function we are about to inline. */ |
| 3326 | id->src_fn = fn; |
| 3327 | id->src_node = cg_edge->callee; |
| 3328 | id->src_cfun = DECL_STRUCT_FUNCTION (fn); |
| 3329 | id->gimple_call = stmt; |
| 3330 | |
| 3331 | gcc_assert (!id->src_cfun->after_inlining); |
| 3332 | |
| 3333 | id->entry_bb = bb; |
| 3334 | if (lookup_attribute ("cold", DECL_ATTRIBUTES (fn))) |
| 3335 | { |
| 3336 | gimple_stmt_iterator si = gsi_last_bb (bb); |
| 3337 | gsi_insert_after (&si, gimple_build_predict (PRED_COLD_FUNCTION, |
| 3338 | NOT_TAKEN), |
| 3339 | GSI_NEW_STMT); |
| 3340 | } |
| 3341 | initialize_inlined_parameters (id, stmt, fn, bb); |
| 3342 | |
| 3343 | if (DECL_INITIAL (fn)) |
| 3344 | prepend_lexical_block (id->block, remap_blocks (DECL_INITIAL (fn), id)); |
| 3345 | |
| 3346 | /* Return statements in the function body will be replaced by jumps |
| 3347 | to the RET_LABEL. */ |
| 3348 | gcc_assert (DECL_INITIAL (fn)); |
| 3349 | gcc_assert (TREE_CODE (DECL_INITIAL (fn)) == BLOCK); |
| 3350 | |
| 3351 | /* Find the LHS to which the result of this call is assigned. */ |
| 3352 | return_slot = NULL; |
| 3353 | if (gimple_call_lhs (stmt)) |
| 3354 | { |
| 3355 | modify_dest = gimple_call_lhs (stmt); |
| 3356 | |
| 3357 | /* The function which we are inlining might not return a value, |
| 3358 | in which case we should issue a warning that the function |
| 3359 | does not return a value. In that case the optimizers will |
| 3360 | see that the variable to which the value is assigned was not |
| 3361 | initialized. We do not want to issue a warning about that |
| 3362 | uninitialized variable. */ |
| 3363 | if (DECL_P (modify_dest)) |
| 3364 | TREE_NO_WARNING (modify_dest) = 1; |
| 3365 | |
| 3366 | if (gimple_call_return_slot_opt_p (stmt)) |
| 3367 | { |
| 3368 | return_slot = modify_dest; |
| 3369 | modify_dest = NULL; |
| 3370 | } |
| 3371 | } |
| 3372 | else |
| 3373 | modify_dest = NULL; |
| 3374 | |
| 3375 | /* If we are inlining a call to the C++ operator new, we don't want |
| 3376 | to use type based alias analysis on the return value. Otherwise |
| 3377 | we may get confused if the compiler sees that the inlined new |
| 3378 | function returns a pointer which was just deleted. See bug |
| 3379 | 33407. */ |
| 3380 | if (DECL_IS_OPERATOR_NEW (fn)) |
| 3381 | { |
| 3382 | return_slot = NULL; |
| 3383 | modify_dest = NULL; |
| 3384 | } |
| 3385 | |
| 3386 | /* Declare the return variable for the function. */ |
| 3387 | retvar = declare_return_variable (id, return_slot, modify_dest, &use_retvar); |
| 3388 | |
| 3389 | if (DECL_IS_OPERATOR_NEW (fn)) |
| 3390 | { |
| 3391 | gcc_assert (TREE_CODE (retvar) == VAR_DECL |
| 3392 | && POINTER_TYPE_P (TREE_TYPE (retvar))); |
| 3393 | DECL_NO_TBAA_P (retvar) = 1; |
| 3394 | } |
| 3395 | |
| 3396 | /* Add local vars in this inlined callee to caller. */ |
| 3397 | t_step = id->src_cfun->local_decls; |
| 3398 | for (; t_step; t_step = TREE_CHAIN (t_step)) |
| 3399 | { |
| 3400 | var = TREE_VALUE (t_step); |
| 3401 | if (TREE_STATIC (var) && !TREE_ASM_WRITTEN (var)) |
| 3402 | { |
| 3403 | if (var_ann (var) && add_referenced_var (var)) |
| 3404 | cfun->local_decls = tree_cons (NULL_TREE, var, |
| 3405 | cfun->local_decls); |
| 3406 | } |
| 3407 | else if (!can_be_nonlocal (var, id)) |
| 3408 | cfun->local_decls = tree_cons (NULL_TREE, remap_decl (var, id), |
| 3409 | cfun->local_decls); |
| 3410 | } |
| 3411 | |
| 3412 | /* This is it. Duplicate the callee body. Assume callee is |
| 3413 | pre-gimplified. Note that we must not alter the caller |
| 3414 | function in any way before this point, as this CALL_EXPR may be |
| 3415 | a self-referential call; if we're calling ourselves, we need to |
| 3416 | duplicate our body before altering anything. */ |
| 3417 | copy_body (id, bb->count, bb->frequency, bb, return_block); |
| 3418 | |
| 3419 | /* Clean up. */ |
| 3420 | pointer_map_destroy (id->decl_map); |
| 3421 | id->decl_map = st; |
| 3422 | |
| 3423 | /* If the inlined function returns a result that we care about, |
| 3424 | substitute the GIMPLE_CALL with an assignment of the return |
| 3425 | variable to the LHS of the call. That is, if STMT was |
| 3426 | 'a = foo (...)', substitute the call with 'a = USE_RETVAR'. */ |
| 3427 | if (use_retvar && gimple_call_lhs (stmt)) |
| 3428 | { |
| 3429 | gimple old_stmt = stmt; |
| 3430 | stmt = gimple_build_assign (gimple_call_lhs (stmt), use_retvar); |
| 3431 | gsi_replace (&stmt_gsi, stmt, false); |
| 3432 | if (gimple_in_ssa_p (cfun)) |
| 3433 | { |
| 3434 | update_stmt (stmt); |
| 3435 | mark_symbols_for_renaming (stmt); |
| 3436 | } |
| 3437 | maybe_clean_or_replace_eh_stmt (old_stmt, stmt); |
| 3438 | } |
| 3439 | else |
| 3440 | { |
| 3441 | /* Handle the case of inlining a function with no return |
| 3442 | statement, which causes the return value to become undefined. */ |
| 3443 | if (gimple_call_lhs (stmt) |
| 3444 | && TREE_CODE (gimple_call_lhs (stmt)) == SSA_NAME) |
| 3445 | { |
| 3446 | tree name = gimple_call_lhs (stmt); |
| 3447 | tree var = SSA_NAME_VAR (name); |
| 3448 | tree def = gimple_default_def (cfun, var); |
| 3449 | |
| 3450 | if (def) |
| 3451 | { |
| 3452 | /* If the variable is used undefined, make this name |
| 3453 | undefined via a move. */ |
| 3454 | stmt = gimple_build_assign (gimple_call_lhs (stmt), def); |
| 3455 | gsi_replace (&stmt_gsi, stmt, true); |
| 3456 | update_stmt (stmt); |
| 3457 | } |
| 3458 | else |
| 3459 | { |
| 3460 | /* Otherwise make this variable undefined. */ |
| 3461 | gsi_remove (&stmt_gsi, true); |
| 3462 | set_default_def (var, name); |
| 3463 | SSA_NAME_DEF_STMT (name) = gimple_build_nop (); |
| 3464 | } |
| 3465 | } |
| 3466 | else |
| 3467 | gsi_remove (&stmt_gsi, true); |
| 3468 | } |
| 3469 | |
| 3470 | if (purge_dead_abnormal_edges) |
| 3471 | gimple_purge_dead_abnormal_call_edges (return_block); |
| 3472 | |
| 3473 | /* If the value of the new expression is ignored, that's OK. We |
| 3474 | don't warn about this for CALL_EXPRs, so we shouldn't warn about |
| 3475 | the equivalent inlined version either. */ |
| 3476 | if (is_gimple_assign (stmt)) |
| 3477 | { |
| 3478 | gcc_assert (gimple_assign_single_p (stmt) |
| 3479 | || CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (stmt))); |
| 3480 | TREE_USED (gimple_assign_rhs1 (stmt)) = 1; |
| 3481 | } |
| 3482 | |
| 3483 | /* Output the inlining info for this abstract function, since it has been |
| 3484 | inlined. If we don't do this now, we can lose the information about the |
| 3485 | variables in the function when the blocks get blown away as soon as we |
| 3486 | remove the cgraph node. */ |
| 3487 | (*debug_hooks->outlining_inline_function) (cg_edge->callee->decl); |
| 3488 | |
| 3489 | /* Update callgraph if needed. */ |
| 3490 | cgraph_remove_node (cg_edge->callee); |
| 3491 | |
| 3492 | id->block = NULL_TREE; |
| 3493 | successfully_inlined = TRUE; |
| 3494 | |
| 3495 | egress: |
| 3496 | input_location = saved_location; |
| 3497 | return successfully_inlined; |
| 3498 | } |
| 3499 | |
| 3500 | /* Expand call statements reachable from STMT_P. |
| 3501 | We can only have CALL_EXPRs as the "toplevel" tree code or nested |
| 3502 | in a MODIFY_EXPR. See tree-gimple.c:get_call_expr_in(). We can |
| 3503 | unfortunately not use that function here because we need a pointer |
| 3504 | to the CALL_EXPR, not the tree itself. */ |
| 3505 | |
| 3506 | static bool |
| 3507 | gimple_expand_calls_inline (basic_block bb, copy_body_data *id) |
| 3508 | { |
| 3509 | gimple_stmt_iterator gsi; |
| 3510 | |
| 3511 | for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) |
| 3512 | { |
| 3513 | gimple stmt = gsi_stmt (gsi); |
| 3514 | |
| 3515 | if (is_gimple_call (stmt) |
| 3516 | && expand_call_inline (bb, stmt, id)) |
| 3517 | return true; |
| 3518 | } |
| 3519 | |
| 3520 | return false; |
| 3521 | } |
| 3522 | |
| 3523 | |
| 3524 | /* Walk all basic blocks created after FIRST and try to fold every statement |
| 3525 | in the STATEMENTS pointer set. */ |
| 3526 | |
| 3527 | static void |
| 3528 | fold_marked_statements (int first, struct pointer_set_t *statements) |
| 3529 | { |
| 3530 | for (; first < n_basic_blocks; first++) |
| 3531 | if (BASIC_BLOCK (first)) |
| 3532 | { |
| 3533 | gimple_stmt_iterator gsi; |
| 3534 | |
| 3535 | for (gsi = gsi_start_bb (BASIC_BLOCK (first)); |
| 3536 | !gsi_end_p (gsi); |
| 3537 | gsi_next (&gsi)) |
| 3538 | if (pointer_set_contains (statements, gsi_stmt (gsi))) |
| 3539 | { |
| 3540 | gimple old_stmt = gsi_stmt (gsi); |
| 3541 | |
| 3542 | if (fold_stmt (&gsi)) |
| 3543 | { |
| 3544 | /* Re-read the statement from GSI as fold_stmt() may |
| 3545 | have changed it. */ |
| 3546 | gimple new_stmt = gsi_stmt (gsi); |
| 3547 | update_stmt (new_stmt); |
| 3548 | |
| 3549 | if (is_gimple_call (old_stmt)) |
| 3550 | cgraph_update_edges_for_call_stmt (old_stmt, new_stmt); |
| 3551 | |
| 3552 | if (maybe_clean_or_replace_eh_stmt (old_stmt, new_stmt)) |
| 3553 | gimple_purge_dead_eh_edges (BASIC_BLOCK (first)); |
| 3554 | } |
| 3555 | } |
| 3556 | } |
| 3557 | } |
| 3558 | |
| 3559 | /* Return true if BB has at least one abnormal outgoing edge. */ |
| 3560 | |
| 3561 | static inline bool |
| 3562 | has_abnormal_outgoing_edge_p (basic_block bb) |
| 3563 | { |
| 3564 | edge e; |
| 3565 | edge_iterator ei; |
| 3566 | |
| 3567 | FOR_EACH_EDGE (e, ei, bb->succs) |
| 3568 | if (e->flags & EDGE_ABNORMAL) |
| 3569 | return true; |
| 3570 | |
| 3571 | return false; |
| 3572 | } |
| 3573 | |
| 3574 | /* Expand calls to inline functions in the body of FN. */ |
| 3575 | |
| 3576 | unsigned int |
| 3577 | optimize_inline_calls (tree fn) |
| 3578 | { |
| 3579 | copy_body_data id; |
| 3580 | tree prev_fn; |
| 3581 | basic_block bb; |
| 3582 | int last = n_basic_blocks; |
| 3583 | struct gimplify_ctx gctx; |
| 3584 | |
| 3585 | /* There is no point in performing inlining if errors have already |
| 3586 | occurred -- and we might crash if we try to inline invalid |
| 3587 | code. */ |
| 3588 | if (errorcount || sorrycount) |
| 3589 | return 0; |
| 3590 | |
| 3591 | /* Clear out ID. */ |
| 3592 | memset (&id, 0, sizeof (id)); |
| 3593 | |
| 3594 | id.src_node = id.dst_node = cgraph_node (fn); |
| 3595 | id.dst_fn = fn; |
| 3596 | /* Or any functions that aren't finished yet. */ |
| 3597 | prev_fn = NULL_TREE; |
| 3598 | if (current_function_decl) |
| 3599 | { |
| 3600 | id.dst_fn = current_function_decl; |
| 3601 | prev_fn = current_function_decl; |
| 3602 | } |
| 3603 | |
| 3604 | id.copy_decl = copy_decl_maybe_to_var; |
| 3605 | id.transform_call_graph_edges = CB_CGE_DUPLICATE; |
| 3606 | id.transform_new_cfg = false; |
| 3607 | id.transform_return_to_modify = true; |
| 3608 | id.transform_lang_insert_block = NULL; |
| 3609 | id.statements_to_fold = pointer_set_create (); |
| 3610 | |
| 3611 | push_gimplify_context (&gctx); |
| 3612 | |
| 3613 | /* We make no attempts to keep dominance info up-to-date. */ |
| 3614 | free_dominance_info (CDI_DOMINATORS); |
| 3615 | free_dominance_info (CDI_POST_DOMINATORS); |
| 3616 | |
| 3617 | /* Register specific gimple functions. */ |
| 3618 | gimple_register_cfg_hooks (); |
| 3619 | |
| 3620 | /* Reach the trees by walking over the CFG, and note the |
| 3621 | enclosing basic-blocks in the call edges. */ |
| 3622 | /* We walk the blocks going forward, because inlined function bodies |
| 3623 | will split id->current_basic_block, and the new blocks will |
| 3624 | follow it; we'll trudge through them, processing their CALL_EXPRs |
| 3625 | along the way. */ |
| 3626 | FOR_EACH_BB (bb) |
| 3627 | gimple_expand_calls_inline (bb, &id); |
| 3628 | |
| 3629 | pop_gimplify_context (NULL); |
| 3630 | |
| 3631 | #ifdef ENABLE_CHECKING |
| 3632 | { |
| 3633 | struct cgraph_edge *e; |
| 3634 | |
| 3635 | verify_cgraph_node (id.dst_node); |
| 3636 | |
| 3637 | /* Double check that we inlined everything we are supposed to inline. */ |
| 3638 | for (e = id.dst_node->callees; e; e = e->next_callee) |
| 3639 | gcc_assert (e->inline_failed); |
| 3640 | } |
| 3641 | #endif |
| 3642 | |
| 3643 | /* Fold the statements before compacting/renumbering the basic blocks. */ |
| 3644 | fold_marked_statements (last, id.statements_to_fold); |
| 3645 | pointer_set_destroy (id.statements_to_fold); |
| 3646 | |
| 3647 | /* Renumber the (code) basic_blocks consecutively. */ |
| 3648 | compact_blocks (); |
| 3649 | /* Renumber the lexical scoping (non-code) blocks consecutively. */ |
| 3650 | number_blocks (fn); |
| 3651 | |
| 3652 | /* We are not going to maintain the cgraph edges up to date. |
| 3653 | Kill it so it won't confuse us. */ |
| 3654 | cgraph_node_remove_callees (id.dst_node); |
| 3655 | |
| 3656 | fold_cond_expr_cond (); |
| 3657 | |
| 3658 | /* It would be nice to check SSA/CFG/statement consistency here, but it is |
| 3659 | not possible yet - the IPA passes might make various functions to not |
| 3660 | throw and they don't care to proactively update local EH info. This is |
| 3661 | done later in fixup_cfg pass that also execute the verification. */ |
| 3662 | return (TODO_update_ssa |
| 3663 | | TODO_cleanup_cfg |
| 3664 | | (gimple_in_ssa_p (cfun) ? TODO_remove_unused_locals : 0) |
| 3665 | | (profile_status != PROFILE_ABSENT ? TODO_rebuild_frequencies : 0)); |
| 3666 | } |
| 3667 | |
| 3668 | /* Passed to walk_tree. Copies the node pointed to, if appropriate. */ |
| 3669 | |
| 3670 | tree |
| 3671 | copy_tree_r (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED) |
| 3672 | { |
| 3673 | enum tree_code code = TREE_CODE (*tp); |
| 3674 | enum tree_code_class cl = TREE_CODE_CLASS (code); |
| 3675 | |
| 3676 | /* We make copies of most nodes. */ |
| 3677 | if (IS_EXPR_CODE_CLASS (cl) |
| 3678 | || code == TREE_LIST |
| 3679 | || code == TREE_VEC |
| 3680 | || code == TYPE_DECL |
| 3681 | || code == OMP_CLAUSE) |
| 3682 | { |
| 3683 | /* Because the chain gets clobbered when we make a copy, we save it |
| 3684 | here. */ |
| 3685 | tree chain = NULL_TREE, new_tree; |
| 3686 | |
| 3687 | chain = TREE_CHAIN (*tp); |
| 3688 | |
| 3689 | /* Copy the node. */ |
| 3690 | new_tree = copy_node (*tp); |
| 3691 | |
| 3692 | /* Propagate mudflap marked-ness. */ |
| 3693 | if (flag_mudflap && mf_marked_p (*tp)) |
| 3694 | mf_mark (new_tree); |
| 3695 | |
| 3696 | *tp = new_tree; |
| 3697 | |
| 3698 | /* Now, restore the chain, if appropriate. That will cause |
| 3699 | walk_tree to walk into the chain as well. */ |
| 3700 | if (code == PARM_DECL |
| 3701 | || code == TREE_LIST |
| 3702 | || code == OMP_CLAUSE) |
| 3703 | TREE_CHAIN (*tp) = chain; |
| 3704 | |
| 3705 | /* For now, we don't update BLOCKs when we make copies. So, we |
| 3706 | have to nullify all BIND_EXPRs. */ |
| 3707 | if (TREE_CODE (*tp) == BIND_EXPR) |
| 3708 | BIND_EXPR_BLOCK (*tp) = NULL_TREE; |
| 3709 | } |
| 3710 | else if (code == CONSTRUCTOR) |
| 3711 | { |
| 3712 | /* CONSTRUCTOR nodes need special handling because |
| 3713 | we need to duplicate the vector of elements. */ |
| 3714 | tree new_tree; |
| 3715 | |
| 3716 | new_tree = copy_node (*tp); |
| 3717 | |
| 3718 | /* Propagate mudflap marked-ness. */ |
| 3719 | if (flag_mudflap && mf_marked_p (*tp)) |
| 3720 | mf_mark (new_tree); |
| 3721 | |
| 3722 | CONSTRUCTOR_ELTS (new_tree) = VEC_copy (constructor_elt, gc, |
| 3723 | CONSTRUCTOR_ELTS (*tp)); |
| 3724 | *tp = new_tree; |
| 3725 | } |
| 3726 | else if (TREE_CODE_CLASS (code) == tcc_type) |
| 3727 | *walk_subtrees = 0; |
| 3728 | else if (TREE_CODE_CLASS (code) == tcc_declaration) |
| 3729 | *walk_subtrees = 0; |
| 3730 | else if (TREE_CODE_CLASS (code) == tcc_constant) |
| 3731 | *walk_subtrees = 0; |
| 3732 | else |
| 3733 | gcc_assert (code != STATEMENT_LIST); |
| 3734 | return NULL_TREE; |
| 3735 | } |
| 3736 | |
| 3737 | /* The SAVE_EXPR pointed to by TP is being copied. If ST contains |
| 3738 | information indicating to what new SAVE_EXPR this one should be mapped, |
| 3739 | use that one. Otherwise, create a new node and enter it in ST. FN is |
| 3740 | the function into which the copy will be placed. */ |
| 3741 | |
| 3742 | static void |
| 3743 | remap_save_expr (tree *tp, void *st_, int *walk_subtrees) |
| 3744 | { |
| 3745 | struct pointer_map_t *st = (struct pointer_map_t *) st_; |
| 3746 | tree *n; |
| 3747 | tree t; |
| 3748 | |
| 3749 | /* See if we already encountered this SAVE_EXPR. */ |
| 3750 | n = (tree *) pointer_map_contains (st, *tp); |
| 3751 | |
| 3752 | /* If we didn't already remap this SAVE_EXPR, do so now. */ |
| 3753 | if (!n) |
| 3754 | { |
| 3755 | t = copy_node (*tp); |
| 3756 | |
| 3757 | /* Remember this SAVE_EXPR. */ |
| 3758 | *pointer_map_insert (st, *tp) = t; |
| 3759 | /* Make sure we don't remap an already-remapped SAVE_EXPR. */ |
| 3760 | *pointer_map_insert (st, t) = t; |
| 3761 | } |
| 3762 | else |
| 3763 | { |
| 3764 | /* We've already walked into this SAVE_EXPR; don't do it again. */ |
| 3765 | *walk_subtrees = 0; |
| 3766 | t = *n; |
| 3767 | } |
| 3768 | |
| 3769 | /* Replace this SAVE_EXPR with the copy. */ |
| 3770 | *tp = t; |
| 3771 | } |
| 3772 | |
| 3773 | /* Called via walk_tree. If *TP points to a DECL_STMT for a local label, |
| 3774 | copies the declaration and enters it in the splay_tree in DATA (which is |
| 3775 | really an `copy_body_data *'). */ |
| 3776 | |
| 3777 | static tree |
| 3778 | mark_local_for_remap_r (tree *tp, int *walk_subtrees ATTRIBUTE_UNUSED, |
| 3779 | void *data) |
| 3780 | { |
| 3781 | copy_body_data *id = (copy_body_data *) data; |
| 3782 | |
| 3783 | /* Don't walk into types. */ |
| 3784 | if (TYPE_P (*tp)) |
| 3785 | *walk_subtrees = 0; |
| 3786 | |
| 3787 | else if (TREE_CODE (*tp) == LABEL_EXPR) |
| 3788 | { |
| 3789 | tree decl = TREE_OPERAND (*tp, 0); |
| 3790 | |
| 3791 | /* Copy the decl and remember the copy. */ |
| 3792 | insert_decl_map (id, decl, id->copy_decl (decl, id)); |
| 3793 | } |
| 3794 | |
| 3795 | return NULL_TREE; |
| 3796 | } |
| 3797 | |
| 3798 | /* Perform any modifications to EXPR required when it is unsaved. Does |
| 3799 | not recurse into EXPR's subtrees. */ |
| 3800 | |
| 3801 | static void |
| 3802 | unsave_expr_1 (tree expr) |
| 3803 | { |
| 3804 | switch (TREE_CODE (expr)) |
| 3805 | { |
| 3806 | case TARGET_EXPR: |
| 3807 | /* Don't mess with a TARGET_EXPR that hasn't been expanded. |
| 3808 | It's OK for this to happen if it was part of a subtree that |
| 3809 | isn't immediately expanded, such as operand 2 of another |
| 3810 | TARGET_EXPR. */ |
| 3811 | if (TREE_OPERAND (expr, 1)) |
| 3812 | break; |
| 3813 | |
| 3814 | TREE_OPERAND (expr, 1) = TREE_OPERAND (expr, 3); |
| 3815 | TREE_OPERAND (expr, 3) = NULL_TREE; |
| 3816 | break; |
| 3817 | |
| 3818 | default: |
| 3819 | break; |
| 3820 | } |
| 3821 | } |
| 3822 | |
| 3823 | /* Called via walk_tree when an expression is unsaved. Using the |
| 3824 | splay_tree pointed to by ST (which is really a `splay_tree'), |
| 3825 | remaps all local declarations to appropriate replacements. */ |
| 3826 | |
| 3827 | static tree |
| 3828 | unsave_r (tree *tp, int *walk_subtrees, void *data) |
| 3829 | { |
| 3830 | copy_body_data *id = (copy_body_data *) data; |
| 3831 | struct pointer_map_t *st = id->decl_map; |
| 3832 | tree *n; |
| 3833 | |
| 3834 | /* Only a local declaration (variable or label). */ |
| 3835 | if ((TREE_CODE (*tp) == VAR_DECL && !TREE_STATIC (*tp)) |
| 3836 | || TREE_CODE (*tp) == LABEL_DECL) |
| 3837 | { |
| 3838 | /* Lookup the declaration. */ |
| 3839 | n = (tree *) pointer_map_contains (st, *tp); |
| 3840 | |
| 3841 | /* If it's there, remap it. */ |
| 3842 | if (n) |
| 3843 | *tp = *n; |
| 3844 | } |
| 3845 | |
| 3846 | else if (TREE_CODE (*tp) == STATEMENT_LIST) |
| 3847 | gcc_unreachable (); |
| 3848 | else if (TREE_CODE (*tp) == BIND_EXPR) |
| 3849 | copy_bind_expr (tp, walk_subtrees, id); |
| 3850 | else if (TREE_CODE (*tp) == SAVE_EXPR) |
| 3851 | remap_save_expr (tp, st, walk_subtrees); |
| 3852 | else |
| 3853 | { |
| 3854 | copy_tree_r (tp, walk_subtrees, NULL); |
| 3855 | |
| 3856 | /* Do whatever unsaving is required. */ |
| 3857 | unsave_expr_1 (*tp); |
| 3858 | } |
| 3859 | |
| 3860 | /* Keep iterating. */ |
| 3861 | return NULL_TREE; |
| 3862 | } |
| 3863 | |
| 3864 | /* Copies everything in EXPR and replaces variables, labels |
| 3865 | and SAVE_EXPRs local to EXPR. */ |
| 3866 | |
| 3867 | tree |
| 3868 | unsave_expr_now (tree expr) |
| 3869 | { |
| 3870 | copy_body_data id; |
| 3871 | |
| 3872 | /* There's nothing to do for NULL_TREE. */ |
| 3873 | if (expr == 0) |
| 3874 | return expr; |
| 3875 | |
| 3876 | /* Set up ID. */ |
| 3877 | memset (&id, 0, sizeof (id)); |
| 3878 | id.src_fn = current_function_decl; |
| 3879 | id.dst_fn = current_function_decl; |
| 3880 | id.decl_map = pointer_map_create (); |
| 3881 | |
| 3882 | id.copy_decl = copy_decl_no_change; |
| 3883 | id.transform_call_graph_edges = CB_CGE_DUPLICATE; |
| 3884 | id.transform_new_cfg = false; |
| 3885 | id.transform_return_to_modify = false; |
| 3886 | id.transform_lang_insert_block = NULL; |
| 3887 | |
| 3888 | /* Walk the tree once to find local labels. */ |
| 3889 | walk_tree_without_duplicates (&expr, mark_local_for_remap_r, &id); |
| 3890 | |
| 3891 | /* Walk the tree again, copying, remapping, and unsaving. */ |
| 3892 | walk_tree (&expr, unsave_r, &id, NULL); |
| 3893 | |
| 3894 | /* Clean up. */ |
| 3895 | pointer_map_destroy (id.decl_map); |
| 3896 | |
| 3897 | return expr; |
| 3898 | } |
| 3899 | |
| 3900 | /* Called via walk_gimple_seq. If *GSIP points to a GIMPLE_LABEL for a local |
| 3901 | label, copies the declaration and enters it in the splay_tree in DATA (which |
| 3902 | is really a 'copy_body_data *'. */ |
| 3903 | |
| 3904 | static tree |
| 3905 | mark_local_labels_stmt (gimple_stmt_iterator *gsip, |
| 3906 | bool *handled_ops_p ATTRIBUTE_UNUSED, |
| 3907 | struct walk_stmt_info *wi) |
| 3908 | { |
| 3909 | copy_body_data *id = (copy_body_data *) wi->info; |
| 3910 | gimple stmt = gsi_stmt (*gsip); |
| 3911 | |
| 3912 | if (gimple_code (stmt) == GIMPLE_LABEL) |
| 3913 | { |
| 3914 | tree decl = gimple_label_label (stmt); |
| 3915 | |
| 3916 | /* Copy the decl and remember the copy. */ |
| 3917 | insert_decl_map (id, decl, id->copy_decl (decl, id)); |
| 3918 | } |
| 3919 | |
| 3920 | return NULL_TREE; |
| 3921 | } |
| 3922 | |
| 3923 | |
| 3924 | /* Called via walk_gimple_seq by copy_gimple_seq_and_replace_local. |
| 3925 | Using the splay_tree pointed to by ST (which is really a `splay_tree'), |
| 3926 | remaps all local declarations to appropriate replacements in gimple |
| 3927 | operands. */ |
| 3928 | |
| 3929 | static tree |
| 3930 | replace_locals_op (tree *tp, int *walk_subtrees, void *data) |
| 3931 | { |
| 3932 | struct walk_stmt_info *wi = (struct walk_stmt_info*) data; |
| 3933 | copy_body_data *id = (copy_body_data *) wi->info; |
| 3934 | struct pointer_map_t *st = id->decl_map; |
| 3935 | tree *n; |
| 3936 | tree expr = *tp; |
| 3937 | |
| 3938 | /* Only a local declaration (variable or label). */ |
| 3939 | if ((TREE_CODE (expr) == VAR_DECL |
| 3940 | && !TREE_STATIC (expr)) |
| 3941 | || TREE_CODE (expr) == LABEL_DECL) |
| 3942 | { |
| 3943 | /* Lookup the declaration. */ |
| 3944 | n = (tree *) pointer_map_contains (st, expr); |
| 3945 | |
| 3946 | /* If it's there, remap it. */ |
| 3947 | if (n) |
| 3948 | *tp = *n; |
| 3949 | *walk_subtrees = 0; |
| 3950 | } |
| 3951 | else if (TREE_CODE (expr) == STATEMENT_LIST |
| 3952 | || TREE_CODE (expr) == BIND_EXPR |
| 3953 | || TREE_CODE (expr) == SAVE_EXPR) |
| 3954 | gcc_unreachable (); |
| 3955 | else if (TREE_CODE (expr) == TARGET_EXPR) |
| 3956 | { |
| 3957 | /* Don't mess with a TARGET_EXPR that hasn't been expanded. |
| 3958 | It's OK for this to happen if it was part of a subtree that |
| 3959 | isn't immediately expanded, such as operand 2 of another |
| 3960 | TARGET_EXPR. */ |
| 3961 | if (!TREE_OPERAND (expr, 1)) |
| 3962 | { |
| 3963 | TREE_OPERAND (expr, 1) = TREE_OPERAND (expr, 3); |
| 3964 | TREE_OPERAND (expr, 3) = NULL_TREE; |
| 3965 | } |
| 3966 | } |
| 3967 | |
| 3968 | /* Keep iterating. */ |
| 3969 | return NULL_TREE; |
| 3970 | } |
| 3971 | |
| 3972 | |
| 3973 | /* Called via walk_gimple_seq by copy_gimple_seq_and_replace_local. |
| 3974 | Using the splay_tree pointed to by ST (which is really a `splay_tree'), |
| 3975 | remaps all local declarations to appropriate replacements in gimple |
| 3976 | statements. */ |
| 3977 | |
| 3978 | static tree |
| 3979 | replace_locals_stmt (gimple_stmt_iterator *gsip, |
| 3980 | bool *handled_ops_p ATTRIBUTE_UNUSED, |
| 3981 | struct walk_stmt_info *wi) |
| 3982 | { |
| 3983 | copy_body_data *id = (copy_body_data *) wi->info; |
| 3984 | gimple stmt = gsi_stmt (*gsip); |
| 3985 | |
| 3986 | if (gimple_code (stmt) == GIMPLE_BIND) |
| 3987 | { |
| 3988 | tree block = gimple_bind_block (stmt); |
| 3989 | |
| 3990 | if (block) |
| 3991 | { |
| 3992 | remap_block (&block, id); |
| 3993 | gimple_bind_set_block (stmt, block); |
| 3994 | } |
| 3995 | |
| 3996 | /* This will remap a lot of the same decls again, but this should be |
| 3997 | harmless. */ |
| 3998 | if (gimple_bind_vars (stmt)) |
| 3999 | gimple_bind_set_vars (stmt, remap_decls (gimple_bind_vars (stmt), NULL, id)); |
| 4000 | } |
| 4001 | |
| 4002 | /* Keep iterating. */ |
| 4003 | return NULL_TREE; |
| 4004 | } |
| 4005 | |
| 4006 | |
| 4007 | /* Copies everything in SEQ and replaces variables and labels local to |
| 4008 | current_function_decl. */ |
| 4009 | |
| 4010 | gimple_seq |
| 4011 | copy_gimple_seq_and_replace_locals (gimple_seq seq) |
| 4012 | { |
| 4013 | copy_body_data id; |
| 4014 | struct walk_stmt_info wi; |
| 4015 | struct pointer_set_t *visited; |
| 4016 | gimple_seq copy; |
| 4017 | |
| 4018 | /* There's nothing to do for NULL_TREE. */ |
| 4019 | if (seq == NULL) |
| 4020 | return seq; |
| 4021 | |
| 4022 | /* Set up ID. */ |
| 4023 | memset (&id, 0, sizeof (id)); |
| 4024 | id.src_fn = current_function_decl; |
| 4025 | id.dst_fn = current_function_decl; |
| 4026 | id.decl_map = pointer_map_create (); |
| 4027 | |
| 4028 | id.copy_decl = copy_decl_no_change; |
| 4029 | id.transform_call_graph_edges = CB_CGE_DUPLICATE; |
| 4030 | id.transform_new_cfg = false; |
| 4031 | id.transform_return_to_modify = false; |
| 4032 | id.transform_lang_insert_block = NULL; |
| 4033 | |
| 4034 | /* Walk the tree once to find local labels. */ |
| 4035 | memset (&wi, 0, sizeof (wi)); |
| 4036 | visited = pointer_set_create (); |
| 4037 | wi.info = &id; |
| 4038 | wi.pset = visited; |
| 4039 | walk_gimple_seq (seq, mark_local_labels_stmt, NULL, &wi); |
| 4040 | pointer_set_destroy (visited); |
| 4041 | |
| 4042 | copy = gimple_seq_copy (seq); |
| 4043 | |
| 4044 | /* Walk the copy, remapping decls. */ |
| 4045 | memset (&wi, 0, sizeof (wi)); |
| 4046 | wi.info = &id; |
| 4047 | walk_gimple_seq (copy, replace_locals_stmt, replace_locals_op, &wi); |
| 4048 | |
| 4049 | /* Clean up. */ |
| 4050 | pointer_map_destroy (id.decl_map); |
| 4051 | |
| 4052 | return copy; |
| 4053 | } |
| 4054 | |
| 4055 | |
| 4056 | /* Allow someone to determine if SEARCH is a child of TOP from gdb. */ |
| 4057 | |
| 4058 | static tree |
| 4059 | debug_find_tree_1 (tree *tp, int *walk_subtrees ATTRIBUTE_UNUSED, void *data) |
| 4060 | { |
| 4061 | if (*tp == data) |
| 4062 | return (tree) data; |
| 4063 | else |
| 4064 | return NULL; |
| 4065 | } |
| 4066 | |
| 4067 | bool |
| 4068 | debug_find_tree (tree top, tree search) |
| 4069 | { |
| 4070 | return walk_tree_without_duplicates (&top, debug_find_tree_1, search) != 0; |
| 4071 | } |
| 4072 | |
| 4073 | |
| 4074 | /* Declare the variables created by the inliner. Add all the variables in |
| 4075 | VARS to BIND_EXPR. */ |
| 4076 | |
| 4077 | static void |
| 4078 | declare_inline_vars (tree block, tree vars) |
| 4079 | { |
| 4080 | tree t; |
| 4081 | for (t = vars; t; t = TREE_CHAIN (t)) |
| 4082 | { |
| 4083 | DECL_SEEN_IN_BIND_EXPR_P (t) = 1; |
| 4084 | gcc_assert (!TREE_STATIC (t) && !TREE_ASM_WRITTEN (t)); |
| 4085 | cfun->local_decls = tree_cons (NULL_TREE, t, cfun->local_decls); |
| 4086 | } |
| 4087 | |
| 4088 | if (block) |
| 4089 | BLOCK_VARS (block) = chainon (BLOCK_VARS (block), vars); |
| 4090 | } |
| 4091 | |
| 4092 | /* Copy NODE (which must be a DECL). The DECL originally was in the FROM_FN, |
| 4093 | but now it will be in the TO_FN. PARM_TO_VAR means enable PARM_DECL to |
| 4094 | VAR_DECL translation. */ |
| 4095 | |
| 4096 | static tree |
| 4097 | copy_decl_for_dup_finish (copy_body_data *id, tree decl, tree copy) |
| 4098 | { |
| 4099 | /* Don't generate debug information for the copy if we wouldn't have |
| 4100 | generated it for the copy either. */ |
| 4101 | DECL_ARTIFICIAL (copy) = DECL_ARTIFICIAL (decl); |
| 4102 | DECL_IGNORED_P (copy) = DECL_IGNORED_P (decl); |
| 4103 | |
| 4104 | /* Set the DECL_ABSTRACT_ORIGIN so the debugging routines know what |
| 4105 | declaration inspired this copy. */ |
| 4106 | DECL_ABSTRACT_ORIGIN (copy) = DECL_ORIGIN (decl); |
| 4107 | |
| 4108 | /* The new variable/label has no RTL, yet. */ |
| 4109 | if (CODE_CONTAINS_STRUCT (TREE_CODE (copy), TS_DECL_WRTL) |
| 4110 | && !TREE_STATIC (copy) && !DECL_EXTERNAL (copy)) |
| 4111 | SET_DECL_RTL (copy, NULL_RTX); |
| 4112 | |
| 4113 | /* These args would always appear unused, if not for this. */ |
| 4114 | TREE_USED (copy) = 1; |
| 4115 | |
| 4116 | /* Set the context for the new declaration. */ |
| 4117 | if (!DECL_CONTEXT (decl)) |
| 4118 | /* Globals stay global. */ |
| 4119 | ; |
| 4120 | else if (DECL_CONTEXT (decl) != id->src_fn) |
| 4121 | /* Things that weren't in the scope of the function we're inlining |
| 4122 | from aren't in the scope we're inlining to, either. */ |
| 4123 | ; |
| 4124 | else if (TREE_STATIC (decl)) |
| 4125 | /* Function-scoped static variables should stay in the original |
| 4126 | function. */ |
| 4127 | ; |
| 4128 | else |
| 4129 | /* Ordinary automatic local variables are now in the scope of the |
| 4130 | new function. */ |
| 4131 | DECL_CONTEXT (copy) = id->dst_fn; |
| 4132 | |
| 4133 | return copy; |
| 4134 | } |
| 4135 | |
| 4136 | static tree |
| 4137 | copy_decl_to_var (tree decl, copy_body_data *id) |
| 4138 | { |
| 4139 | tree copy, type; |
| 4140 | |
| 4141 | gcc_assert (TREE_CODE (decl) == PARM_DECL |
| 4142 | || TREE_CODE (decl) == RESULT_DECL); |
| 4143 | |
| 4144 | type = TREE_TYPE (decl); |
| 4145 | |
| 4146 | copy = build_decl (VAR_DECL, DECL_NAME (decl), type); |
| 4147 | TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (decl); |
| 4148 | TREE_READONLY (copy) = TREE_READONLY (decl); |
| 4149 | TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (decl); |
| 4150 | DECL_GIMPLE_REG_P (copy) = DECL_GIMPLE_REG_P (decl); |
| 4151 | DECL_NO_TBAA_P (copy) = DECL_NO_TBAA_P (decl); |
| 4152 | |
| 4153 | return copy_decl_for_dup_finish (id, decl, copy); |
| 4154 | } |
| 4155 | |
| 4156 | /* Like copy_decl_to_var, but create a return slot object instead of a |
| 4157 | pointer variable for return by invisible reference. */ |
| 4158 | |
| 4159 | static tree |
| 4160 | copy_result_decl_to_var (tree decl, copy_body_data *id) |
| 4161 | { |
| 4162 | tree copy, type; |
| 4163 | |
| 4164 | gcc_assert (TREE_CODE (decl) == PARM_DECL |
| 4165 | || TREE_CODE (decl) == RESULT_DECL); |
| 4166 | |
| 4167 | type = TREE_TYPE (decl); |
| 4168 | if (DECL_BY_REFERENCE (decl)) |
| 4169 | type = TREE_TYPE (type); |
| 4170 | |
| 4171 | copy = build_decl (VAR_DECL, DECL_NAME (decl), type); |
| 4172 | TREE_READONLY (copy) = TREE_READONLY (decl); |
| 4173 | TREE_THIS_VOLATILE (copy) = TREE_THIS_VOLATILE (decl); |
| 4174 | if (!DECL_BY_REFERENCE (decl)) |
| 4175 | { |
| 4176 | TREE_ADDRESSABLE (copy) = TREE_ADDRESSABLE (decl); |
| 4177 | DECL_GIMPLE_REG_P (copy) = DECL_GIMPLE_REG_P (decl); |
| 4178 | DECL_NO_TBAA_P (copy) = DECL_NO_TBAA_P (decl); |
| 4179 | } |
| 4180 | |
| 4181 | return copy_decl_for_dup_finish (id, decl, copy); |
| 4182 | } |
| 4183 | |
| 4184 | tree |
| 4185 | copy_decl_no_change (tree decl, copy_body_data *id) |
| 4186 | { |
| 4187 | tree copy; |
| 4188 | |
| 4189 | copy = copy_node (decl); |
| 4190 | |
| 4191 | /* The COPY is not abstract; it will be generated in DST_FN. */ |
| 4192 | DECL_ABSTRACT (copy) = 0; |
| 4193 | lang_hooks.dup_lang_specific_decl (copy); |
| 4194 | |
| 4195 | /* TREE_ADDRESSABLE isn't used to indicate that a label's address has |
| 4196 | been taken; it's for internal bookkeeping in expand_goto_internal. */ |
| 4197 | if (TREE_CODE (copy) == LABEL_DECL) |
| 4198 | { |
| 4199 | TREE_ADDRESSABLE (copy) = 0; |
| 4200 | LABEL_DECL_UID (copy) = -1; |
| 4201 | } |
| 4202 | |
| 4203 | return copy_decl_for_dup_finish (id, decl, copy); |
| 4204 | } |
| 4205 | |
| 4206 | static tree |
| 4207 | copy_decl_maybe_to_var (tree decl, copy_body_data *id) |
| 4208 | { |
| 4209 | if (TREE_CODE (decl) == PARM_DECL || TREE_CODE (decl) == RESULT_DECL) |
| 4210 | return copy_decl_to_var (decl, id); |
| 4211 | else |
| 4212 | return copy_decl_no_change (decl, id); |
| 4213 | } |
| 4214 | |
| 4215 | /* Return a copy of the function's argument tree. */ |
| 4216 | static tree |
| 4217 | copy_arguments_for_versioning (tree orig_parm, copy_body_data * id, |
| 4218 | bitmap args_to_skip, tree *vars) |
| 4219 | { |
| 4220 | tree arg, *parg; |
| 4221 | tree new_parm = NULL; |
| 4222 | int i = 0; |
| 4223 | |
| 4224 | parg = &new_parm; |
| 4225 | |
| 4226 | for (arg = orig_parm; arg; arg = TREE_CHAIN (arg), i++) |
| 4227 | if (!args_to_skip || !bitmap_bit_p (args_to_skip, i)) |
| 4228 | { |
| 4229 | tree new_tree = remap_decl (arg, id); |
| 4230 | lang_hooks.dup_lang_specific_decl (new_tree); |
| 4231 | *parg = new_tree; |
| 4232 | parg = &TREE_CHAIN (new_tree); |
| 4233 | } |
| 4234 | else if (!pointer_map_contains (id->decl_map, arg)) |
| 4235 | { |
| 4236 | /* Make an equivalent VAR_DECL. If the argument was used |
| 4237 | as temporary variable later in function, the uses will be |
| 4238 | replaced by local variable. */ |
| 4239 | tree var = copy_decl_to_var (arg, id); |
| 4240 | get_var_ann (var); |
| 4241 | add_referenced_var (var); |
| 4242 | insert_decl_map (id, arg, var); |
| 4243 | /* Declare this new variable. */ |
| 4244 | TREE_CHAIN (var) = *vars; |
| 4245 | *vars = var; |
| 4246 | } |
| 4247 | return new_parm; |
| 4248 | } |
| 4249 | |
| 4250 | /* Return a copy of the function's static chain. */ |
| 4251 | static tree |
| 4252 | copy_static_chain (tree static_chain, copy_body_data * id) |
| 4253 | { |
| 4254 | tree *chain_copy, *pvar; |
| 4255 | |
| 4256 | chain_copy = &static_chain; |
| 4257 | for (pvar = chain_copy; *pvar; pvar = &TREE_CHAIN (*pvar)) |
| 4258 | { |
| 4259 | tree new_tree = remap_decl (*pvar, id); |
| 4260 | lang_hooks.dup_lang_specific_decl (new_tree); |
| 4261 | TREE_CHAIN (new_tree) = TREE_CHAIN (*pvar); |
| 4262 | *pvar = new_tree; |
| 4263 | } |
| 4264 | return static_chain; |
| 4265 | } |
| 4266 | |
| 4267 | /* Return true if the function is allowed to be versioned. |
| 4268 | This is a guard for the versioning functionality. */ |
| 4269 | bool |
| 4270 | tree_versionable_function_p (tree fndecl) |
| 4271 | { |
| 4272 | if (fndecl == NULL_TREE) |
| 4273 | return false; |
| 4274 | /* ??? There are cases where a function is |
| 4275 | uninlinable but can be versioned. */ |
| 4276 | if (!tree_inlinable_function_p (fndecl)) |
| 4277 | return false; |
| 4278 | |
| 4279 | return true; |
| 4280 | } |
| 4281 | |
| 4282 | /* Delete all unreachable basic blocks and update callgraph. |
| 4283 | Doing so is somewhat nontrivial because we need to update all clones and |
| 4284 | remove inline function that become unreachable. */ |
| 4285 | |
| 4286 | static bool |
| 4287 | delete_unreachable_blocks_update_callgraph (copy_body_data *id) |
| 4288 | { |
| 4289 | bool changed = false; |
| 4290 | basic_block b, next_bb; |
| 4291 | |
| 4292 | find_unreachable_blocks (); |
| 4293 | |
| 4294 | /* Delete all unreachable basic blocks. */ |
| 4295 | |
| 4296 | for (b = ENTRY_BLOCK_PTR->next_bb; b != EXIT_BLOCK_PTR; b = next_bb) |
| 4297 | { |
| 4298 | next_bb = b->next_bb; |
| 4299 | |
| 4300 | if (!(b->flags & BB_REACHABLE)) |
| 4301 | { |
| 4302 | gimple_stmt_iterator bsi; |
| 4303 | |
| 4304 | for (bsi = gsi_start_bb (b); !gsi_end_p (bsi); gsi_next (&bsi)) |
| 4305 | if (gimple_code (gsi_stmt (bsi)) == GIMPLE_CALL) |
| 4306 | { |
| 4307 | struct cgraph_edge *e; |
| 4308 | |
| 4309 | if ((e = cgraph_edge (id->dst_node, gsi_stmt (bsi))) != NULL) |
| 4310 | cgraph_remove_edge (e); |
| 4311 | } |
| 4312 | delete_basic_block (b); |
| 4313 | changed = true; |
| 4314 | } |
| 4315 | } |
| 4316 | |
| 4317 | if (changed) |
| 4318 | tidy_fallthru_edges (); |
| 4319 | return changed; |
| 4320 | } |
| 4321 | |
| 4322 | /* Create a copy of a function's tree. |
| 4323 | OLD_DECL and NEW_DECL are FUNCTION_DECL tree nodes |
| 4324 | of the original function and the new copied function |
| 4325 | respectively. In case we want to replace a DECL |
| 4326 | tree with another tree while duplicating the function's |
| 4327 | body, TREE_MAP represents the mapping between these |
| 4328 | trees. If UPDATE_CLONES is set, the call_stmt fields |
| 4329 | of edges of clones of the function will be updated. */ |
| 4330 | void |
| 4331 | tree_function_versioning (tree old_decl, tree new_decl, varray_type tree_map, |
| 4332 | bool update_clones, bitmap args_to_skip) |
| 4333 | { |
| 4334 | struct cgraph_node *old_version_node; |
| 4335 | struct cgraph_node *new_version_node; |
| 4336 | copy_body_data id; |
| 4337 | tree p; |
| 4338 | unsigned i; |
| 4339 | struct ipa_replace_map *replace_info; |
| 4340 | basic_block old_entry_block; |
| 4341 | VEC (gimple, heap) *init_stmts = VEC_alloc (gimple, heap, 10); |
| 4342 | |
| 4343 | tree t_step; |
| 4344 | tree old_current_function_decl = current_function_decl; |
| 4345 | tree vars = NULL_TREE; |
| 4346 | |
| 4347 | gcc_assert (TREE_CODE (old_decl) == FUNCTION_DECL |
| 4348 | && TREE_CODE (new_decl) == FUNCTION_DECL); |
| 4349 | DECL_POSSIBLY_INLINED (old_decl) = 1; |
| 4350 | |
| 4351 | old_version_node = cgraph_node (old_decl); |
| 4352 | new_version_node = cgraph_node (new_decl); |
| 4353 | |
| 4354 | /* Output the inlining info for this abstract function, since it has been |
| 4355 | inlined. If we don't do this now, we can lose the information about the |
| 4356 | variables in the function when the blocks get blown away as soon as we |
| 4357 | remove the cgraph node. */ |
| 4358 | (*debug_hooks->outlining_inline_function) (old_decl); |
| 4359 | |
| 4360 | DECL_ARTIFICIAL (new_decl) = 1; |
| 4361 | DECL_ABSTRACT_ORIGIN (new_decl) = DECL_ORIGIN (old_decl); |
| 4362 | |
| 4363 | /* Prepare the data structures for the tree copy. */ |
| 4364 | memset (&id, 0, sizeof (id)); |
| 4365 | |
| 4366 | /* Generate a new name for the new version. */ |
| 4367 | if (!update_clones) |
| 4368 | { |
| 4369 | DECL_NAME (new_decl) = create_tmp_var_name (NULL); |
| 4370 | SET_DECL_ASSEMBLER_NAME (new_decl, DECL_NAME (new_decl)); |
| 4371 | SET_DECL_RTL (new_decl, NULL_RTX); |
| 4372 | id.statements_to_fold = pointer_set_create (); |
| 4373 | } |
| 4374 | |
| 4375 | id.decl_map = pointer_map_create (); |
| 4376 | id.src_fn = old_decl; |
| 4377 | id.dst_fn = new_decl; |
| 4378 | id.src_node = old_version_node; |
| 4379 | id.dst_node = new_version_node; |
| 4380 | id.src_cfun = DECL_STRUCT_FUNCTION (old_decl); |
| 4381 | |
| 4382 | id.copy_decl = copy_decl_no_change; |
| 4383 | id.transform_call_graph_edges |
| 4384 | = update_clones ? CB_CGE_MOVE_CLONES : CB_CGE_MOVE; |
| 4385 | id.transform_new_cfg = true; |
| 4386 | id.transform_return_to_modify = false; |
| 4387 | id.transform_lang_insert_block = NULL; |
| 4388 | |
| 4389 | current_function_decl = new_decl; |
| 4390 | old_entry_block = ENTRY_BLOCK_PTR_FOR_FUNCTION |
| 4391 | (DECL_STRUCT_FUNCTION (old_decl)); |
| 4392 | initialize_cfun (new_decl, old_decl, |
| 4393 | old_entry_block->count, |
| 4394 | old_entry_block->frequency); |
| 4395 | push_cfun (DECL_STRUCT_FUNCTION (new_decl)); |
| 4396 | |
| 4397 | /* Copy the function's static chain. */ |
| 4398 | p = DECL_STRUCT_FUNCTION (old_decl)->static_chain_decl; |
| 4399 | if (p) |
| 4400 | DECL_STRUCT_FUNCTION (new_decl)->static_chain_decl = |
| 4401 | copy_static_chain (DECL_STRUCT_FUNCTION (old_decl)->static_chain_decl, |
| 4402 | &id); |
| 4403 | |
| 4404 | /* If there's a tree_map, prepare for substitution. */ |
| 4405 | if (tree_map) |
| 4406 | for (i = 0; i < VARRAY_ACTIVE_SIZE (tree_map); i++) |
| 4407 | { |
| 4408 | gimple init; |
| 4409 | replace_info |
| 4410 | = (struct ipa_replace_map *) VARRAY_GENERIC_PTR (tree_map, i); |
| 4411 | if (replace_info->replace_p) |
| 4412 | { |
| 4413 | tree op = replace_info->new_tree; |
| 4414 | |
| 4415 | STRIP_NOPS (op); |
| 4416 | |
| 4417 | if (TREE_CODE (op) == VIEW_CONVERT_EXPR) |
| 4418 | op = TREE_OPERAND (op, 0); |
| 4419 | |
| 4420 | if (TREE_CODE (op) == ADDR_EXPR) |
| 4421 | { |
| 4422 | op = TREE_OPERAND (op, 0); |
| 4423 | while (handled_component_p (op)) |
| 4424 | op = TREE_OPERAND (op, 0); |
| 4425 | if (TREE_CODE (op) == VAR_DECL) |
| 4426 | add_referenced_var (op); |
| 4427 | } |
| 4428 | gcc_assert (TREE_CODE (replace_info->old_tree) == PARM_DECL); |
| 4429 | init = setup_one_parameter (&id, replace_info->old_tree, |
| 4430 | replace_info->new_tree, id.src_fn, |
| 4431 | NULL, |
| 4432 | &vars); |
| 4433 | if (init) |
| 4434 | VEC_safe_push (gimple, heap, init_stmts, init); |
| 4435 | } |
| 4436 | } |
| 4437 | /* Copy the function's arguments. */ |
| 4438 | if (DECL_ARGUMENTS (old_decl) != NULL_TREE) |
| 4439 | DECL_ARGUMENTS (new_decl) = |
| 4440 | copy_arguments_for_versioning (DECL_ARGUMENTS (old_decl), &id, |
| 4441 | args_to_skip, &vars); |
| 4442 | |
| 4443 | DECL_INITIAL (new_decl) = remap_blocks (DECL_INITIAL (id.src_fn), &id); |
| 4444 | |
| 4445 | /* Renumber the lexical scoping (non-code) blocks consecutively. */ |
| 4446 | number_blocks (id.dst_fn); |
| 4447 | |
| 4448 | declare_inline_vars (DECL_INITIAL (new_decl), vars); |
| 4449 | if (DECL_STRUCT_FUNCTION (old_decl)->local_decls != NULL_TREE) |
| 4450 | /* Add local vars. */ |
| 4451 | for (t_step = DECL_STRUCT_FUNCTION (old_decl)->local_decls; |
| 4452 | t_step; t_step = TREE_CHAIN (t_step)) |
| 4453 | { |
| 4454 | tree var = TREE_VALUE (t_step); |
| 4455 | if (TREE_STATIC (var) && !TREE_ASM_WRITTEN (var)) |
| 4456 | cfun->local_decls = tree_cons (NULL_TREE, var, cfun->local_decls); |
| 4457 | else if (!can_be_nonlocal (var, &id)) |
| 4458 | cfun->local_decls = |
| 4459 | tree_cons (NULL_TREE, remap_decl (var, &id), |
| 4460 | cfun->local_decls); |
| 4461 | } |
| 4462 | |
| 4463 | /* Copy the Function's body. */ |
| 4464 | copy_body (&id, old_entry_block->count, old_entry_block->frequency, ENTRY_BLOCK_PTR, EXIT_BLOCK_PTR); |
| 4465 | |
| 4466 | if (DECL_RESULT (old_decl) != NULL_TREE) |
| 4467 | { |
| 4468 | tree *res_decl = &DECL_RESULT (old_decl); |
| 4469 | DECL_RESULT (new_decl) = remap_decl (*res_decl, &id); |
| 4470 | lang_hooks.dup_lang_specific_decl (DECL_RESULT (new_decl)); |
| 4471 | } |
| 4472 | |
| 4473 | /* Renumber the lexical scoping (non-code) blocks consecutively. */ |
| 4474 | number_blocks (new_decl); |
| 4475 | |
| 4476 | if (VEC_length (gimple, init_stmts)) |
| 4477 | { |
| 4478 | basic_block bb = split_edge (single_succ_edge (ENTRY_BLOCK_PTR)); |
| 4479 | while (VEC_length (gimple, init_stmts)) |
| 4480 | insert_init_stmt (bb, VEC_pop (gimple, init_stmts)); |
| 4481 | } |
| 4482 | |
| 4483 | /* Clean up. */ |
| 4484 | pointer_map_destroy (id.decl_map); |
| 4485 | if (!update_clones) |
| 4486 | { |
| 4487 | fold_marked_statements (0, id.statements_to_fold); |
| 4488 | pointer_set_destroy (id.statements_to_fold); |
| 4489 | fold_cond_expr_cond (); |
| 4490 | } |
| 4491 | if (gimple_in_ssa_p (cfun)) |
| 4492 | { |
| 4493 | free_dominance_info (CDI_DOMINATORS); |
| 4494 | free_dominance_info (CDI_POST_DOMINATORS); |
| 4495 | if (!update_clones) |
| 4496 | delete_unreachable_blocks_update_callgraph (&id); |
| 4497 | update_ssa (TODO_update_ssa); |
| 4498 | if (!update_clones) |
| 4499 | { |
| 4500 | fold_cond_expr_cond (); |
| 4501 | if (need_ssa_update_p ()) |
| 4502 | update_ssa (TODO_update_ssa); |
| 4503 | } |
| 4504 | } |
| 4505 | free_dominance_info (CDI_DOMINATORS); |
| 4506 | free_dominance_info (CDI_POST_DOMINATORS); |
| 4507 | VEC_free (gimple, heap, init_stmts); |
| 4508 | pop_cfun (); |
| 4509 | current_function_decl = old_current_function_decl; |
| 4510 | gcc_assert (!current_function_decl |
| 4511 | || DECL_STRUCT_FUNCTION (current_function_decl) == cfun); |
| 4512 | return; |
| 4513 | } |
| 4514 | |
| 4515 | /* Duplicate a type, fields and all. */ |
| 4516 | |
| 4517 | tree |
| 4518 | build_duplicate_type (tree type) |
| 4519 | { |
| 4520 | struct copy_body_data id; |
| 4521 | |
| 4522 | memset (&id, 0, sizeof (id)); |
| 4523 | id.src_fn = current_function_decl; |
| 4524 | id.dst_fn = current_function_decl; |
| 4525 | id.src_cfun = cfun; |
| 4526 | id.decl_map = pointer_map_create (); |
| 4527 | id.copy_decl = copy_decl_no_change; |
| 4528 | |
| 4529 | type = remap_type_1 (type, &id); |
| 4530 | |
| 4531 | pointer_map_destroy (id.decl_map); |
| 4532 | |
| 4533 | TYPE_CANONICAL (type) = type; |
| 4534 | |
| 4535 | return type; |
| 4536 | } |
| 4537 | |
| 4538 | /* Return whether it is safe to inline a function because it used different |
| 4539 | target specific options or different optimization options. */ |
| 4540 | bool |
| 4541 | tree_can_inline_p (tree caller, tree callee) |
| 4542 | { |
| 4543 | #if 0 |
| 4544 | /* This causes a regression in SPEC in that it prevents a cold function from |
| 4545 | inlining a hot function. Perhaps this should only apply to functions |
| 4546 | that the user declares hot/cold/optimize explicitly. */ |
| 4547 | |
| 4548 | /* Don't inline a function with a higher optimization level than the |
| 4549 | caller, or with different space constraints (hot/cold functions). */ |
| 4550 | tree caller_tree = DECL_FUNCTION_SPECIFIC_OPTIMIZATION (caller); |
| 4551 | tree callee_tree = DECL_FUNCTION_SPECIFIC_OPTIMIZATION (callee); |
| 4552 | |
| 4553 | if (caller_tree != callee_tree) |
| 4554 | { |
| 4555 | struct cl_optimization *caller_opt |
| 4556 | = TREE_OPTIMIZATION ((caller_tree) |
| 4557 | ? caller_tree |
| 4558 | : optimization_default_node); |
| 4559 | |
| 4560 | struct cl_optimization *callee_opt |
| 4561 | = TREE_OPTIMIZATION ((callee_tree) |
| 4562 | ? callee_tree |
| 4563 | : optimization_default_node); |
| 4564 | |
| 4565 | if ((caller_opt->optimize > callee_opt->optimize) |
| 4566 | || (caller_opt->optimize_size != callee_opt->optimize_size)) |
| 4567 | return false; |
| 4568 | } |
| 4569 | #endif |
| 4570 | |
| 4571 | /* Allow the backend to decide if inlining is ok. */ |
| 4572 | return targetm.target_option.can_inline_p (caller, callee); |
| 4573 | } |