comparison gcc/tree-outof-ssa.c @ 0:a06113de4d67

first commit
author kent <kent@cr.ie.u-ryukyu.ac.jp>
date Fri, 17 Jul 2009 14:47:48 +0900
parents
children 77e2b8dfacca
comparison
equal deleted inserted replaced
-1:000000000000 0:a06113de4d67
1 /* Convert a program in SSA form into Normal form.
2 Copyright (C) 2004, 2005, 2006, 2007, 2008 Free Software Foundation, Inc.
3 Contributed by Andrew Macleod <amacleod@redhat.com>
4
5 This file is part of GCC.
6
7 GCC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
10 any later version.
11
12 GCC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
16
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
20
21 #include "config.h"
22 #include "system.h"
23 #include "coretypes.h"
24 #include "tm.h"
25 #include "tree.h"
26 #include "ggc.h"
27 #include "basic-block.h"
28 #include "diagnostic.h"
29 #include "bitmap.h"
30 #include "tree-flow.h"
31 #include "timevar.h"
32 #include "tree-dump.h"
33 #include "tree-ssa-live.h"
34 #include "tree-pass.h"
35 #include "toplev.h"
36
37
38 /* Used to hold all the components required to do SSA PHI elimination.
39 The node and pred/succ list is a simple linear list of nodes and
40 edges represented as pairs of nodes.
41
42 The predecessor and successor list: Nodes are entered in pairs, where
43 [0] ->PRED, [1]->SUCC. All the even indexes in the array represent
44 predecessors, all the odd elements are successors.
45
46 Rationale:
47 When implemented as bitmaps, very large programs SSA->Normal times were
48 being dominated by clearing the interference graph.
49
50 Typically this list of edges is extremely small since it only includes
51 PHI results and uses from a single edge which have not coalesced with
52 each other. This means that no virtual PHI nodes are included, and
53 empirical evidence suggests that the number of edges rarely exceed
54 3, and in a bootstrap of GCC, the maximum size encountered was 7.
55 This also limits the number of possible nodes that are involved to
56 rarely more than 6, and in the bootstrap of gcc, the maximum number
57 of nodes encountered was 12. */
58
59 typedef struct _elim_graph {
60 /* Size of the elimination vectors. */
61 int size;
62
63 /* List of nodes in the elimination graph. */
64 VEC(tree,heap) *nodes;
65
66 /* The predecessor and successor edge list. */
67 VEC(int,heap) *edge_list;
68
69 /* Visited vector. */
70 sbitmap visited;
71
72 /* Stack for visited nodes. */
73 VEC(int,heap) *stack;
74
75 /* The variable partition map. */
76 var_map map;
77
78 /* Edge being eliminated by this graph. */
79 edge e;
80
81 /* List of constant copies to emit. These are pushed on in pairs. */
82 VEC(tree,heap) *const_copies;
83 } *elim_graph;
84
85
86 /* Create a temporary variable based on the type of variable T. Use T's name
87 as the prefix. */
88
89 static tree
90 create_temp (tree t)
91 {
92 tree tmp;
93 const char *name = NULL;
94 tree type;
95
96 if (TREE_CODE (t) == SSA_NAME)
97 t = SSA_NAME_VAR (t);
98
99 gcc_assert (TREE_CODE (t) == VAR_DECL || TREE_CODE (t) == PARM_DECL);
100
101 type = TREE_TYPE (t);
102 tmp = DECL_NAME (t);
103 if (tmp)
104 name = IDENTIFIER_POINTER (tmp);
105
106 if (name == NULL)
107 name = "temp";
108 tmp = create_tmp_var (type, name);
109
110 if (DECL_DEBUG_EXPR_IS_FROM (t) && DECL_DEBUG_EXPR (t))
111 {
112 SET_DECL_DEBUG_EXPR (tmp, DECL_DEBUG_EXPR (t));
113 DECL_DEBUG_EXPR_IS_FROM (tmp) = 1;
114 }
115 else if (!DECL_IGNORED_P (t))
116 {
117 SET_DECL_DEBUG_EXPR (tmp, t);
118 DECL_DEBUG_EXPR_IS_FROM (tmp) = 1;
119 }
120 DECL_ARTIFICIAL (tmp) = DECL_ARTIFICIAL (t);
121 DECL_IGNORED_P (tmp) = DECL_IGNORED_P (t);
122 DECL_GIMPLE_REG_P (tmp) = DECL_GIMPLE_REG_P (t);
123 add_referenced_var (tmp);
124
125 /* add_referenced_var will create the annotation and set up some
126 of the flags in the annotation. However, some flags we need to
127 inherit from our original variable. */
128 set_symbol_mem_tag (tmp, symbol_mem_tag (t));
129 if (is_call_clobbered (t))
130 mark_call_clobbered (tmp, var_ann (t)->escape_mask);
131 if (bitmap_bit_p (gimple_call_used_vars (cfun), DECL_UID (t)))
132 bitmap_set_bit (gimple_call_used_vars (cfun), DECL_UID (tmp));
133
134 return tmp;
135 }
136
137
138 /* This helper function fill insert a copy from a constant or variable SRC to
139 variable DEST on edge E. */
140
141 static void
142 insert_copy_on_edge (edge e, tree dest, tree src)
143 {
144 gimple copy;
145
146 copy = gimple_build_assign (dest, src);
147 set_is_used (dest);
148
149 if (TREE_CODE (src) == ADDR_EXPR)
150 src = TREE_OPERAND (src, 0);
151 if (TREE_CODE (src) == VAR_DECL || TREE_CODE (src) == PARM_DECL)
152 set_is_used (src);
153
154 if (dump_file && (dump_flags & TDF_DETAILS))
155 {
156 fprintf (dump_file,
157 "Inserting a copy on edge BB%d->BB%d :",
158 e->src->index,
159 e->dest->index);
160 print_gimple_stmt (dump_file, copy, 0, dump_flags);
161 fprintf (dump_file, "\n");
162 }
163
164 gsi_insert_on_edge (e, copy);
165 }
166
167
168 /* Create an elimination graph with SIZE nodes and associated data
169 structures. */
170
171 static elim_graph
172 new_elim_graph (int size)
173 {
174 elim_graph g = (elim_graph) xmalloc (sizeof (struct _elim_graph));
175
176 g->nodes = VEC_alloc (tree, heap, 30);
177 g->const_copies = VEC_alloc (tree, heap, 20);
178 g->edge_list = VEC_alloc (int, heap, 20);
179 g->stack = VEC_alloc (int, heap, 30);
180
181 g->visited = sbitmap_alloc (size);
182
183 return g;
184 }
185
186
187 /* Empty elimination graph G. */
188
189 static inline void
190 clear_elim_graph (elim_graph g)
191 {
192 VEC_truncate (tree, g->nodes, 0);
193 VEC_truncate (int, g->edge_list, 0);
194 }
195
196
197 /* Delete elimination graph G. */
198
199 static inline void
200 delete_elim_graph (elim_graph g)
201 {
202 sbitmap_free (g->visited);
203 VEC_free (int, heap, g->stack);
204 VEC_free (int, heap, g->edge_list);
205 VEC_free (tree, heap, g->const_copies);
206 VEC_free (tree, heap, g->nodes);
207 free (g);
208 }
209
210
211 /* Return the number of nodes in graph G. */
212
213 static inline int
214 elim_graph_size (elim_graph g)
215 {
216 return VEC_length (tree, g->nodes);
217 }
218
219
220 /* Add NODE to graph G, if it doesn't exist already. */
221
222 static inline void
223 elim_graph_add_node (elim_graph g, tree node)
224 {
225 int x;
226 tree t;
227
228 for (x = 0; VEC_iterate (tree, g->nodes, x, t); x++)
229 if (t == node)
230 return;
231 VEC_safe_push (tree, heap, g->nodes, node);
232 }
233
234
235 /* Add the edge PRED->SUCC to graph G. */
236
237 static inline void
238 elim_graph_add_edge (elim_graph g, int pred, int succ)
239 {
240 VEC_safe_push (int, heap, g->edge_list, pred);
241 VEC_safe_push (int, heap, g->edge_list, succ);
242 }
243
244
245 /* Remove an edge from graph G for which NODE is the predecessor, and
246 return the successor node. -1 is returned if there is no such edge. */
247
248 static inline int
249 elim_graph_remove_succ_edge (elim_graph g, int node)
250 {
251 int y;
252 unsigned x;
253 for (x = 0; x < VEC_length (int, g->edge_list); x += 2)
254 if (VEC_index (int, g->edge_list, x) == node)
255 {
256 VEC_replace (int, g->edge_list, x, -1);
257 y = VEC_index (int, g->edge_list, x + 1);
258 VEC_replace (int, g->edge_list, x + 1, -1);
259 return y;
260 }
261 return -1;
262 }
263
264
265 /* Find all the nodes in GRAPH which are successors to NODE in the
266 edge list. VAR will hold the partition number found. CODE is the
267 code fragment executed for every node found. */
268
269 #define FOR_EACH_ELIM_GRAPH_SUCC(GRAPH, NODE, VAR, CODE) \
270 do { \
271 unsigned x_; \
272 int y_; \
273 for (x_ = 0; x_ < VEC_length (int, (GRAPH)->edge_list); x_ += 2) \
274 { \
275 y_ = VEC_index (int, (GRAPH)->edge_list, x_); \
276 if (y_ != (NODE)) \
277 continue; \
278 (VAR) = VEC_index (int, (GRAPH)->edge_list, x_ + 1); \
279 CODE; \
280 } \
281 } while (0)
282
283
284 /* Find all the nodes which are predecessors of NODE in the edge list for
285 GRAPH. VAR will hold the partition number found. CODE is the
286 code fragment executed for every node found. */
287
288 #define FOR_EACH_ELIM_GRAPH_PRED(GRAPH, NODE, VAR, CODE) \
289 do { \
290 unsigned x_; \
291 int y_; \
292 for (x_ = 0; x_ < VEC_length (int, (GRAPH)->edge_list); x_ += 2) \
293 { \
294 y_ = VEC_index (int, (GRAPH)->edge_list, x_ + 1); \
295 if (y_ != (NODE)) \
296 continue; \
297 (VAR) = VEC_index (int, (GRAPH)->edge_list, x_); \
298 CODE; \
299 } \
300 } while (0)
301
302
303 /* Add T to elimination graph G. */
304
305 static inline void
306 eliminate_name (elim_graph g, tree T)
307 {
308 elim_graph_add_node (g, T);
309 }
310
311
312 /* Build elimination graph G for basic block BB on incoming PHI edge
313 G->e. */
314
315 static void
316 eliminate_build (elim_graph g, basic_block B)
317 {
318 tree T0, Ti;
319 int p0, pi;
320 gimple_stmt_iterator gsi;
321
322 clear_elim_graph (g);
323
324 for (gsi = gsi_start_phis (B); !gsi_end_p (gsi); gsi_next (&gsi))
325 {
326 gimple phi = gsi_stmt (gsi);
327
328 T0 = var_to_partition_to_var (g->map, gimple_phi_result (phi));
329
330 /* Ignore results which are not in partitions. */
331 if (T0 == NULL_TREE)
332 continue;
333
334 Ti = PHI_ARG_DEF (phi, g->e->dest_idx);
335
336 /* If this argument is a constant, or a SSA_NAME which is being
337 left in SSA form, just queue a copy to be emitted on this
338 edge. */
339 if (!phi_ssa_name_p (Ti)
340 || (TREE_CODE (Ti) == SSA_NAME
341 && var_to_partition (g->map, Ti) == NO_PARTITION))
342 {
343 /* Save constant copies until all other copies have been emitted
344 on this edge. */
345 VEC_safe_push (tree, heap, g->const_copies, T0);
346 VEC_safe_push (tree, heap, g->const_copies, Ti);
347 }
348 else
349 {
350 Ti = var_to_partition_to_var (g->map, Ti);
351 if (T0 != Ti)
352 {
353 eliminate_name (g, T0);
354 eliminate_name (g, Ti);
355 p0 = var_to_partition (g->map, T0);
356 pi = var_to_partition (g->map, Ti);
357 elim_graph_add_edge (g, p0, pi);
358 }
359 }
360 }
361 }
362
363
364 /* Push successors of T onto the elimination stack for G. */
365
366 static void
367 elim_forward (elim_graph g, int T)
368 {
369 int S;
370 SET_BIT (g->visited, T);
371 FOR_EACH_ELIM_GRAPH_SUCC (g, T, S,
372 {
373 if (!TEST_BIT (g->visited, S))
374 elim_forward (g, S);
375 });
376 VEC_safe_push (int, heap, g->stack, T);
377 }
378
379
380 /* Return 1 if there unvisited predecessors of T in graph G. */
381
382 static int
383 elim_unvisited_predecessor (elim_graph g, int T)
384 {
385 int P;
386 FOR_EACH_ELIM_GRAPH_PRED (g, T, P,
387 {
388 if (!TEST_BIT (g->visited, P))
389 return 1;
390 });
391 return 0;
392 }
393
394 /* Process predecessors first, and insert a copy. */
395
396 static void
397 elim_backward (elim_graph g, int T)
398 {
399 int P;
400 SET_BIT (g->visited, T);
401 FOR_EACH_ELIM_GRAPH_PRED (g, T, P,
402 {
403 if (!TEST_BIT (g->visited, P))
404 {
405 elim_backward (g, P);
406 insert_copy_on_edge (g->e,
407 partition_to_var (g->map, P),
408 partition_to_var (g->map, T));
409 }
410 });
411 }
412
413 /* Insert required copies for T in graph G. Check for a strongly connected
414 region, and create a temporary to break the cycle if one is found. */
415
416 static void
417 elim_create (elim_graph g, int T)
418 {
419 tree U;
420 int P, S;
421
422 if (elim_unvisited_predecessor (g, T))
423 {
424 U = create_temp (partition_to_var (g->map, T));
425 insert_copy_on_edge (g->e, U, partition_to_var (g->map, T));
426 FOR_EACH_ELIM_GRAPH_PRED (g, T, P,
427 {
428 if (!TEST_BIT (g->visited, P))
429 {
430 elim_backward (g, P);
431 insert_copy_on_edge (g->e, partition_to_var (g->map, P), U);
432 }
433 });
434 }
435 else
436 {
437 S = elim_graph_remove_succ_edge (g, T);
438 if (S != -1)
439 {
440 SET_BIT (g->visited, T);
441 insert_copy_on_edge (g->e,
442 partition_to_var (g->map, T),
443 partition_to_var (g->map, S));
444 }
445 }
446
447 }
448
449
450 /* Eliminate all the phi nodes on edge E in graph G. */
451
452 static void
453 eliminate_phi (edge e, elim_graph g)
454 {
455 int x;
456 basic_block B = e->dest;
457
458 gcc_assert (VEC_length (tree, g->const_copies) == 0);
459
460 /* Abnormal edges already have everything coalesced. */
461 if (e->flags & EDGE_ABNORMAL)
462 return;
463
464 g->e = e;
465
466 eliminate_build (g, B);
467
468 if (elim_graph_size (g) != 0)
469 {
470 tree var;
471
472 sbitmap_zero (g->visited);
473 VEC_truncate (int, g->stack, 0);
474
475 for (x = 0; VEC_iterate (tree, g->nodes, x, var); x++)
476 {
477 int p = var_to_partition (g->map, var);
478 if (!TEST_BIT (g->visited, p))
479 elim_forward (g, p);
480 }
481
482 sbitmap_zero (g->visited);
483 while (VEC_length (int, g->stack) > 0)
484 {
485 x = VEC_pop (int, g->stack);
486 if (!TEST_BIT (g->visited, x))
487 elim_create (g, x);
488 }
489 }
490
491 /* If there are any pending constant copies, issue them now. */
492 while (VEC_length (tree, g->const_copies) > 0)
493 {
494 tree src, dest;
495 src = VEC_pop (tree, g->const_copies);
496 dest = VEC_pop (tree, g->const_copies);
497 insert_copy_on_edge (e, dest, src);
498 }
499 }
500
501
502 /* Take the ssa-name var_map MAP, and assign real variables to each
503 partition. */
504
505 static void
506 assign_vars (var_map map)
507 {
508 int x, num;
509 tree var, root;
510 var_ann_t ann;
511
512 num = num_var_partitions (map);
513 for (x = 0; x < num; x++)
514 {
515 var = partition_to_var (map, x);
516 if (TREE_CODE (var) != SSA_NAME)
517 {
518 ann = var_ann (var);
519 /* It must already be coalesced. */
520 gcc_assert (ann->out_of_ssa_tag == 1);
521 if (dump_file && (dump_flags & TDF_DETAILS))
522 {
523 fprintf (dump_file, "partition %d already has variable ", x);
524 print_generic_expr (dump_file, var, TDF_SLIM);
525 fprintf (dump_file, " assigned to it.\n");
526 }
527 }
528 else
529 {
530 root = SSA_NAME_VAR (var);
531 ann = var_ann (root);
532 /* If ROOT is already associated, create a new one. */
533 if (ann->out_of_ssa_tag)
534 {
535 root = create_temp (root);
536 ann = var_ann (root);
537 }
538 /* ROOT has not been coalesced yet, so use it. */
539 if (dump_file && (dump_flags & TDF_DETAILS))
540 {
541 fprintf (dump_file, "Partition %d is assigned to var ", x);
542 print_generic_stmt (dump_file, root, TDF_SLIM);
543 }
544 change_partition_var (map, root, x);
545 }
546 }
547 }
548
549
550 /* Replace use operand P with whatever variable it has been rewritten to based
551 on the partitions in MAP. EXPR is an optional expression vector over SSA
552 versions which is used to replace P with an expression instead of a variable.
553 If the stmt is changed, return true. */
554
555 static inline bool
556 replace_use_variable (var_map map, use_operand_p p, gimple *expr)
557 {
558 tree new_var;
559 tree var = USE_FROM_PTR (p);
560
561 /* Check if we are replacing this variable with an expression. */
562 if (expr)
563 {
564 int version = SSA_NAME_VERSION (var);
565 if (expr[version])
566 {
567 SET_USE (p, gimple_assign_rhs_to_tree (expr[version]));
568 return true;
569 }
570 }
571
572 new_var = var_to_partition_to_var (map, var);
573 if (new_var)
574 {
575 SET_USE (p, new_var);
576 set_is_used (new_var);
577 return true;
578 }
579 return false;
580 }
581
582
583 /* Replace def operand DEF_P with whatever variable it has been rewritten to
584 based on the partitions in MAP. EXPR is an optional expression vector over
585 SSA versions which is used to replace DEF_P with an expression instead of a
586 variable. If the stmt is changed, return true. */
587
588 static inline bool
589 replace_def_variable (var_map map, def_operand_p def_p, tree *expr)
590 {
591 tree new_var;
592 tree var = DEF_FROM_PTR (def_p);
593
594 /* Do nothing if we are replacing this variable with an expression. */
595 if (expr && expr[SSA_NAME_VERSION (var)])
596 return true;
597
598 new_var = var_to_partition_to_var (map, var);
599 if (new_var)
600 {
601 SET_DEF (def_p, new_var);
602 set_is_used (new_var);
603 return true;
604 }
605 return false;
606 }
607
608
609 /* Remove each argument from PHI. If an arg was the last use of an SSA_NAME,
610 check to see if this allows another PHI node to be removed. */
611
612 static void
613 remove_gimple_phi_args (gimple phi)
614 {
615 use_operand_p arg_p;
616 ssa_op_iter iter;
617
618 if (dump_file && (dump_flags & TDF_DETAILS))
619 {
620 fprintf (dump_file, "Removing Dead PHI definition: ");
621 print_gimple_stmt (dump_file, phi, 0, TDF_SLIM);
622 }
623
624 FOR_EACH_PHI_ARG (arg_p, phi, iter, SSA_OP_USE)
625 {
626 tree arg = USE_FROM_PTR (arg_p);
627 if (TREE_CODE (arg) == SSA_NAME)
628 {
629 /* Remove the reference to the existing argument. */
630 SET_USE (arg_p, NULL_TREE);
631 if (has_zero_uses (arg))
632 {
633 gimple stmt;
634 gimple_stmt_iterator gsi;
635
636 stmt = SSA_NAME_DEF_STMT (arg);
637
638 /* Also remove the def if it is a PHI node. */
639 if (gimple_code (stmt) == GIMPLE_PHI)
640 {
641 remove_gimple_phi_args (stmt);
642 gsi = gsi_for_stmt (stmt);
643 remove_phi_node (&gsi, true);
644 }
645
646 }
647 }
648 }
649 }
650
651 /* Remove any PHI node which is a virtual PHI, or a PHI with no uses. */
652
653 static void
654 eliminate_useless_phis (void)
655 {
656 basic_block bb;
657 gimple_stmt_iterator gsi;
658 tree result;
659
660 FOR_EACH_BB (bb)
661 {
662 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); )
663 {
664 gimple phi = gsi_stmt (gsi);
665 result = gimple_phi_result (phi);
666 if (!is_gimple_reg (SSA_NAME_VAR (result)))
667 {
668 #ifdef ENABLE_CHECKING
669 size_t i;
670 /* There should be no arguments which are not virtual, or the
671 results will be incorrect. */
672 for (i = 0; i < gimple_phi_num_args (phi); i++)
673 {
674 tree arg = PHI_ARG_DEF (phi, i);
675 if (TREE_CODE (arg) == SSA_NAME
676 && is_gimple_reg (SSA_NAME_VAR (arg)))
677 {
678 fprintf (stderr, "Argument of PHI is not virtual (");
679 print_generic_expr (stderr, arg, TDF_SLIM);
680 fprintf (stderr, "), but the result is :");
681 print_gimple_stmt (stderr, phi, 0, TDF_SLIM);
682 internal_error ("SSA corruption");
683 }
684 }
685 #endif
686 remove_phi_node (&gsi, true);
687 }
688 else
689 {
690 /* Also remove real PHIs with no uses. */
691 if (has_zero_uses (result))
692 {
693 remove_gimple_phi_args (phi);
694 remove_phi_node (&gsi, true);
695 }
696 else
697 gsi_next (&gsi);
698 }
699 }
700 }
701 }
702
703
704 /* This function will rewrite the current program using the variable mapping
705 found in MAP. If the replacement vector VALUES is provided, any
706 occurrences of partitions with non-null entries in the vector will be
707 replaced with the expression in the vector instead of its mapped
708 variable. */
709
710 static void
711 rewrite_trees (var_map map, gimple *values)
712 {
713 elim_graph g;
714 basic_block bb;
715 gimple_stmt_iterator gsi;
716 edge e;
717 gimple_seq phi;
718 bool changed;
719
720 #ifdef ENABLE_CHECKING
721 /* Search for PHIs where the destination has no partition, but one
722 or more arguments has a partition. This should not happen and can
723 create incorrect code. */
724 FOR_EACH_BB (bb)
725 {
726 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
727 {
728 gimple phi = gsi_stmt (gsi);
729 tree T0 = var_to_partition_to_var (map, gimple_phi_result (phi));
730 if (T0 == NULL_TREE)
731 {
732 size_t i;
733 for (i = 0; i < gimple_phi_num_args (phi); i++)
734 {
735 tree arg = PHI_ARG_DEF (phi, i);
736
737 if (TREE_CODE (arg) == SSA_NAME
738 && var_to_partition (map, arg) != NO_PARTITION)
739 {
740 fprintf (stderr, "Argument of PHI is in a partition :(");
741 print_generic_expr (stderr, arg, TDF_SLIM);
742 fprintf (stderr, "), but the result is not :");
743 print_gimple_stmt (stderr, phi, 0, TDF_SLIM);
744 internal_error ("SSA corruption");
745 }
746 }
747 }
748 }
749 }
750 #endif
751
752 /* Replace PHI nodes with any required copies. */
753 g = new_elim_graph (map->num_partitions);
754 g->map = map;
755 FOR_EACH_BB (bb)
756 {
757 for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); )
758 {
759 gimple stmt = gsi_stmt (gsi);
760 use_operand_p use_p, copy_use_p;
761 def_operand_p def_p;
762 bool remove = false, is_copy = false;
763 int num_uses = 0;
764 ssa_op_iter iter;
765
766 changed = false;
767
768 if (gimple_assign_copy_p (stmt))
769 is_copy = true;
770
771 copy_use_p = NULL_USE_OPERAND_P;
772 FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE)
773 {
774 if (replace_use_variable (map, use_p, values))
775 changed = true;
776 copy_use_p = use_p;
777 num_uses++;
778 }
779
780 if (num_uses != 1)
781 is_copy = false;
782
783 def_p = SINGLE_SSA_DEF_OPERAND (stmt, SSA_OP_DEF);
784
785 if (def_p != NULL)
786 {
787 /* Mark this stmt for removal if it is the list of replaceable
788 expressions. */
789 if (values && values[SSA_NAME_VERSION (DEF_FROM_PTR (def_p))])
790 remove = true;
791 else
792 {
793 if (replace_def_variable (map, def_p, NULL))
794 changed = true;
795 /* If both SSA_NAMEs coalesce to the same variable,
796 mark the now redundant copy for removal. */
797 if (is_copy)
798 {
799 gcc_assert (copy_use_p != NULL_USE_OPERAND_P);
800 if (DEF_FROM_PTR (def_p) == USE_FROM_PTR (copy_use_p))
801 remove = true;
802 }
803 }
804 }
805 else
806 FOR_EACH_SSA_DEF_OPERAND (def_p, stmt, iter, SSA_OP_DEF)
807 if (replace_def_variable (map, def_p, NULL))
808 changed = true;
809
810 /* Remove any stmts marked for removal. */
811 if (remove)
812 gsi_remove (&gsi, true);
813 else
814 {
815 if (changed)
816 if (maybe_clean_or_replace_eh_stmt (stmt, stmt))
817 gimple_purge_dead_eh_edges (bb);
818 gsi_next (&gsi);
819 }
820 }
821
822 phi = phi_nodes (bb);
823 if (phi)
824 {
825 edge_iterator ei;
826 FOR_EACH_EDGE (e, ei, bb->preds)
827 eliminate_phi (e, g);
828 }
829 }
830
831 delete_elim_graph (g);
832 }
833
834 /* These are the local work structures used to determine the best place to
835 insert the copies that were placed on edges by the SSA->normal pass.. */
836 static VEC(edge,heap) *edge_leader;
837 static VEC(gimple_seq,heap) *stmt_list;
838 static bitmap leader_has_match = NULL;
839 static edge leader_match = NULL;
840
841
842 /* Pass this function to make_forwarder_block so that all the edges with
843 matching PENDING_STMT lists to 'curr_stmt_list' get redirected. E is the
844 edge to test for a match. */
845
846 static inline bool
847 same_stmt_list_p (edge e)
848 {
849 return (e->aux == (PTR) leader_match) ? true : false;
850 }
851
852
853 /* Return TRUE if S1 and S2 are equivalent copies. */
854
855 static inline bool
856 identical_copies_p (const_gimple s1, const_gimple s2)
857 {
858 #ifdef ENABLE_CHECKING
859 gcc_assert (is_gimple_assign (s1));
860 gcc_assert (is_gimple_assign (s2));
861 gcc_assert (DECL_P (gimple_assign_lhs (s1)));
862 gcc_assert (DECL_P (gimple_assign_lhs (s2)));
863 #endif
864
865 if (gimple_assign_lhs (s1) != gimple_assign_lhs (s2))
866 return false;
867
868 if (gimple_assign_rhs1 (s1) != gimple_assign_rhs1 (s2))
869 return false;
870
871 return true;
872 }
873
874
875 /* Compare the PENDING_STMT list for edges E1 and E2. Return true if the lists
876 contain the same sequence of copies. */
877
878 static inline bool
879 identical_stmt_lists_p (const_edge e1, const_edge e2)
880 {
881 gimple_seq t1 = PENDING_STMT (e1);
882 gimple_seq t2 = PENDING_STMT (e2);
883 gimple_stmt_iterator gsi1, gsi2;
884
885 for (gsi1 = gsi_start (t1), gsi2 = gsi_start (t2);
886 !gsi_end_p (gsi1) && !gsi_end_p (gsi2);
887 gsi_next (&gsi1), gsi_next (&gsi2))
888 {
889 if (!identical_copies_p (gsi_stmt (gsi1), gsi_stmt (gsi2)))
890 break;
891 }
892
893 if (!gsi_end_p (gsi1) || !gsi_end_p (gsi2))
894 return false;
895
896 return true;
897 }
898
899
900 /* Allocate data structures used in analyze_edges_for_bb. */
901
902 static void
903 init_analyze_edges_for_bb (void)
904 {
905 edge_leader = VEC_alloc (edge, heap, 25);
906 stmt_list = VEC_alloc (gimple_seq, heap, 25);
907 leader_has_match = BITMAP_ALLOC (NULL);
908 }
909
910
911 /* Free data structures used in analyze_edges_for_bb. */
912
913 static void
914 fini_analyze_edges_for_bb (void)
915 {
916 VEC_free (edge, heap, edge_leader);
917 VEC_free (gimple_seq, heap, stmt_list);
918 BITMAP_FREE (leader_has_match);
919 }
920
921 /* A helper function to be called via walk_tree. Return DATA if it is
922 contained in subtree TP. */
923
924 static tree
925 contains_tree_r (tree * tp, int *walk_subtrees, void *data)
926 {
927 if (*tp == data)
928 {
929 *walk_subtrees = 0;
930 return (tree) data;
931 }
932 else
933 return NULL_TREE;
934 }
935
936 /* A threshold for the number of insns contained in the latch block.
937 It is used to prevent blowing the loop with too many copies from
938 the latch. */
939 #define MAX_STMTS_IN_LATCH 2
940
941 /* Return TRUE if the stmts on SINGLE-EDGE can be moved to the
942 body of the loop. This should be permitted only if SINGLE-EDGE is a
943 single-basic-block latch edge and thus cleaning the latch will help
944 to create a single-basic-block loop. Otherwise return FALSE. */
945
946 static bool
947 process_single_block_loop_latch (edge single_edge)
948 {
949 gimple_seq stmts;
950 basic_block b_exit, b_pheader, b_loop = single_edge->src;
951 edge_iterator ei;
952 edge e;
953 gimple_stmt_iterator gsi, gsi_exit;
954 gimple_stmt_iterator tsi;
955 tree expr;
956 gimple stmt;
957 unsigned int count = 0;
958
959 if (single_edge == NULL || (single_edge->dest != single_edge->src)
960 || (EDGE_COUNT (b_loop->succs) != 2)
961 || (EDGE_COUNT (b_loop->preds) != 2))
962 return false;
963
964 /* Get the stmts on the latch edge. */
965 stmts = PENDING_STMT (single_edge);
966
967 /* Find the successor edge which is not the latch edge. */
968 FOR_EACH_EDGE (e, ei, b_loop->succs)
969 if (e->dest != b_loop)
970 break;
971
972 b_exit = e->dest;
973
974 /* Check that the exit block has only the loop as a predecessor,
975 and that there are no pending stmts on that edge as well. */
976 if (EDGE_COUNT (b_exit->preds) != 1 || PENDING_STMT (e))
977 return false;
978
979 /* Find the predecessor edge which is not the latch edge. */
980 FOR_EACH_EDGE (e, ei, b_loop->preds)
981 if (e->src != b_loop)
982 break;
983
984 b_pheader = e->src;
985
986 if (b_exit == b_pheader || b_exit == b_loop || b_pheader == b_loop)
987 return false;
988
989 gsi_exit = gsi_after_labels (b_exit);
990
991 /* Get the last stmt in the loop body. */
992 gsi = gsi_last_bb (single_edge->src);
993 stmt = gsi_stmt (gsi);
994
995 if (gimple_code (stmt) != GIMPLE_COND)
996 return false;
997
998
999 expr = build2 (gimple_cond_code (stmt), boolean_type_node,
1000 gimple_cond_lhs (stmt), gimple_cond_rhs (stmt));
1001 /* Iterate over the insns on the latch and count them. */
1002 for (tsi = gsi_start (stmts); !gsi_end_p (tsi); gsi_next (&tsi))
1003 {
1004 gimple stmt1 = gsi_stmt (tsi);
1005 tree var;
1006
1007 count++;
1008 /* Check that the condition does not contain any new definition
1009 created in the latch as the stmts from the latch intended
1010 to precede it. */
1011 if (gimple_code (stmt1) != GIMPLE_ASSIGN)
1012 return false;
1013 var = gimple_assign_lhs (stmt1);
1014 if (TREE_THIS_VOLATILE (var)
1015 || TYPE_VOLATILE (TREE_TYPE (var))
1016 || walk_tree (&expr, contains_tree_r, var, NULL))
1017 return false;
1018 }
1019 /* Check that the latch does not contain more than MAX_STMTS_IN_LATCH
1020 insns. The purpose of this restriction is to prevent blowing the
1021 loop with too many copies from the latch. */
1022 if (count > MAX_STMTS_IN_LATCH)
1023 return false;
1024
1025 /* Apply the transformation - clean up the latch block:
1026
1027 var = something;
1028 L1:
1029 x1 = expr;
1030 if (cond) goto L2 else goto L3;
1031 L2:
1032 var = x1;
1033 goto L1
1034 L3:
1035 ...
1036
1037 ==>
1038
1039 var = something;
1040 L1:
1041 x1 = expr;
1042 tmp_var = var;
1043 var = x1;
1044 if (cond) goto L1 else goto L2;
1045 L2:
1046 var = tmp_var;
1047 ...
1048 */
1049 for (tsi = gsi_start (stmts); !gsi_end_p (tsi); gsi_next (&tsi))
1050 {
1051 gimple stmt1 = gsi_stmt (tsi);
1052 tree var, tmp_var;
1053 gimple copy;
1054
1055 /* Create a new variable to load back the value of var in case
1056 we exit the loop. */
1057 var = gimple_assign_lhs (stmt1);
1058 tmp_var = create_temp (var);
1059 copy = gimple_build_assign (tmp_var, var);
1060 set_is_used (tmp_var);
1061 gsi_insert_before (&gsi, copy, GSI_SAME_STMT);
1062 copy = gimple_build_assign (var, tmp_var);
1063 gsi_insert_before (&gsi_exit, copy, GSI_SAME_STMT);
1064 }
1065
1066 PENDING_STMT (single_edge) = 0;
1067 /* Insert the new stmts to the loop body. */
1068 gsi_insert_seq_before (&gsi, stmts, GSI_NEW_STMT);
1069
1070 if (dump_file)
1071 fprintf (dump_file,
1072 "\nCleaned-up latch block of loop with single BB: %d\n\n",
1073 single_edge->dest->index);
1074
1075 return true;
1076 }
1077
1078 /* Look at all the incoming edges to block BB, and decide where the best place
1079 to insert the stmts on each edge are, and perform those insertions. */
1080
1081 static void
1082 analyze_edges_for_bb (basic_block bb)
1083 {
1084 edge e;
1085 edge_iterator ei;
1086 int count;
1087 unsigned int x;
1088 bool have_opportunity;
1089 gimple_stmt_iterator gsi;
1090 gimple stmt;
1091 edge single_edge = NULL;
1092 bool is_label;
1093 edge leader;
1094
1095 count = 0;
1096
1097 /* Blocks which contain at least one abnormal edge cannot use
1098 make_forwarder_block. Look for these blocks, and commit any PENDING_STMTs
1099 found on edges in these block. */
1100 have_opportunity = true;
1101 FOR_EACH_EDGE (e, ei, bb->preds)
1102 if (e->flags & EDGE_ABNORMAL)
1103 {
1104 have_opportunity = false;
1105 break;
1106 }
1107
1108 if (!have_opportunity)
1109 {
1110 FOR_EACH_EDGE (e, ei, bb->preds)
1111 if (PENDING_STMT (e))
1112 gsi_commit_one_edge_insert (e, NULL);
1113 return;
1114 }
1115
1116 /* Find out how many edges there are with interesting pending stmts on them.
1117 Commit the stmts on edges we are not interested in. */
1118 FOR_EACH_EDGE (e, ei, bb->preds)
1119 {
1120 if (PENDING_STMT (e))
1121 {
1122 gcc_assert (!(e->flags & EDGE_ABNORMAL));
1123 if (e->flags & EDGE_FALLTHRU)
1124 {
1125 gsi = gsi_start_bb (e->src);
1126 if (!gsi_end_p (gsi))
1127 {
1128 stmt = gsi_stmt (gsi);
1129 gsi_next (&gsi);
1130 gcc_assert (stmt != NULL);
1131 is_label = (gimple_code (stmt) == GIMPLE_LABEL);
1132 /* Punt if it has non-label stmts, or isn't local. */
1133 if (!is_label
1134 || DECL_NONLOCAL (gimple_label_label (stmt))
1135 || !gsi_end_p (gsi))
1136 {
1137 gsi_commit_one_edge_insert (e, NULL);
1138 continue;
1139 }
1140 }
1141 }
1142 single_edge = e;
1143 count++;
1144 }
1145 }
1146
1147 /* If there aren't at least 2 edges, no sharing will happen. */
1148 if (count < 2)
1149 {
1150 if (single_edge)
1151 {
1152 /* Add stmts to the edge unless processed specially as a
1153 single-block loop latch edge. */
1154 if (!process_single_block_loop_latch (single_edge))
1155 gsi_commit_one_edge_insert (single_edge, NULL);
1156 }
1157 return;
1158 }
1159
1160 /* Ensure that we have empty worklists. */
1161 #ifdef ENABLE_CHECKING
1162 gcc_assert (VEC_length (edge, edge_leader) == 0);
1163 gcc_assert (VEC_length (gimple_seq, stmt_list) == 0);
1164 gcc_assert (bitmap_empty_p (leader_has_match));
1165 #endif
1166
1167 /* Find the "leader" block for each set of unique stmt lists. Preference is
1168 given to FALLTHRU blocks since they would need a GOTO to arrive at another
1169 block. The leader edge destination is the block which all the other edges
1170 with the same stmt list will be redirected to. */
1171 have_opportunity = false;
1172 FOR_EACH_EDGE (e, ei, bb->preds)
1173 {
1174 if (PENDING_STMT (e))
1175 {
1176 bool found = false;
1177
1178 /* Look for the same stmt list in edge leaders list. */
1179 for (x = 0; VEC_iterate (edge, edge_leader, x, leader); x++)
1180 {
1181 if (identical_stmt_lists_p (leader, e))
1182 {
1183 /* Give this edge the same stmt list pointer. */
1184 PENDING_STMT (e) = NULL;
1185 e->aux = leader;
1186 bitmap_set_bit (leader_has_match, x);
1187 have_opportunity = found = true;
1188 break;
1189 }
1190 }
1191
1192 /* If no similar stmt list, add this edge to the leader list. */
1193 if (!found)
1194 {
1195 VEC_safe_push (edge, heap, edge_leader, e);
1196 VEC_safe_push (gimple_seq, heap, stmt_list, PENDING_STMT (e));
1197 }
1198 }
1199 }
1200
1201 /* If there are no similar lists, just issue the stmts. */
1202 if (!have_opportunity)
1203 {
1204 for (x = 0; VEC_iterate (edge, edge_leader, x, leader); x++)
1205 gsi_commit_one_edge_insert (leader, NULL);
1206 VEC_truncate (edge, edge_leader, 0);
1207 VEC_truncate (gimple_seq, stmt_list, 0);
1208 bitmap_clear (leader_has_match);
1209 return;
1210 }
1211
1212 if (dump_file)
1213 fprintf (dump_file, "\nOpportunities in BB %d for stmt/block reduction:\n",
1214 bb->index);
1215
1216 /* For each common list, create a forwarding block and issue the stmt's
1217 in that block. */
1218 for (x = 0; VEC_iterate (edge, edge_leader, x, leader); x++)
1219 if (bitmap_bit_p (leader_has_match, x))
1220 {
1221 edge new_edge;
1222 gimple_stmt_iterator gsi;
1223 gimple_seq curr_stmt_list;
1224
1225 leader_match = leader;
1226
1227 /* The tree_* cfg manipulation routines use the PENDING_EDGE field
1228 for various PHI manipulations, so it gets cleared when calls are
1229 made to make_forwarder_block(). So make sure the edge is clear,
1230 and use the saved stmt list. */
1231 PENDING_STMT (leader) = NULL;
1232 leader->aux = leader;
1233 curr_stmt_list = VEC_index (gimple_seq, stmt_list, x);
1234
1235 new_edge = make_forwarder_block (leader->dest, same_stmt_list_p,
1236 NULL);
1237 bb = new_edge->dest;
1238 if (dump_file)
1239 {
1240 fprintf (dump_file, "Splitting BB %d for Common stmt list. ",
1241 leader->dest->index);
1242 fprintf (dump_file, "Original block is now BB%d.\n", bb->index);
1243 print_gimple_seq (dump_file, curr_stmt_list, 0, TDF_VOPS);
1244 }
1245
1246 FOR_EACH_EDGE (e, ei, new_edge->src->preds)
1247 {
1248 e->aux = NULL;
1249 if (dump_file)
1250 fprintf (dump_file, " Edge (%d->%d) lands here.\n",
1251 e->src->index, e->dest->index);
1252 }
1253
1254 gsi = gsi_last_bb (leader->dest);
1255 gsi_insert_seq_after (&gsi, curr_stmt_list, GSI_NEW_STMT);
1256
1257 leader_match = NULL;
1258 /* We should never get a new block now. */
1259 }
1260 else
1261 {
1262 PENDING_STMT (leader) = VEC_index (gimple_seq, stmt_list, x);
1263 gsi_commit_one_edge_insert (leader, NULL);
1264 }
1265
1266
1267 /* Clear the working data structures. */
1268 VEC_truncate (edge, edge_leader, 0);
1269 VEC_truncate (gimple_seq, stmt_list, 0);
1270 bitmap_clear (leader_has_match);
1271 }
1272
1273
1274 /* This function will analyze the insertions which were performed on edges,
1275 and decide whether they should be left on that edge, or whether it is more
1276 efficient to emit some subset of them in a single block. All stmts are
1277 inserted somewhere. */
1278
1279 static void
1280 perform_edge_inserts (void)
1281 {
1282 basic_block bb;
1283
1284 if (dump_file)
1285 fprintf(dump_file, "Analyzing Edge Insertions.\n");
1286
1287 /* analyze_edges_for_bb calls make_forwarder_block, which tries to
1288 incrementally update the dominator information. Since we don't
1289 need dominator information after this pass, go ahead and free the
1290 dominator information. */
1291 free_dominance_info (CDI_DOMINATORS);
1292 free_dominance_info (CDI_POST_DOMINATORS);
1293
1294 /* Allocate data structures used in analyze_edges_for_bb. */
1295 init_analyze_edges_for_bb ();
1296
1297 FOR_EACH_BB (bb)
1298 analyze_edges_for_bb (bb);
1299
1300 analyze_edges_for_bb (EXIT_BLOCK_PTR);
1301
1302 /* Free data structures used in analyze_edges_for_bb. */
1303 fini_analyze_edges_for_bb ();
1304
1305 #ifdef ENABLE_CHECKING
1306 {
1307 edge_iterator ei;
1308 edge e;
1309 FOR_EACH_BB (bb)
1310 {
1311 FOR_EACH_EDGE (e, ei, bb->preds)
1312 {
1313 if (PENDING_STMT (e))
1314 error (" Pending stmts not issued on PRED edge (%d, %d)\n",
1315 e->src->index, e->dest->index);
1316 }
1317 FOR_EACH_EDGE (e, ei, bb->succs)
1318 {
1319 if (PENDING_STMT (e))
1320 error (" Pending stmts not issued on SUCC edge (%d, %d)\n",
1321 e->src->index, e->dest->index);
1322 }
1323 }
1324 FOR_EACH_EDGE (e, ei, ENTRY_BLOCK_PTR->succs)
1325 {
1326 if (PENDING_STMT (e))
1327 error (" Pending stmts not issued on ENTRY edge (%d, %d)\n",
1328 e->src->index, e->dest->index);
1329 }
1330 FOR_EACH_EDGE (e, ei, EXIT_BLOCK_PTR->preds)
1331 {
1332 if (PENDING_STMT (e))
1333 error (" Pending stmts not issued on EXIT edge (%d, %d)\n",
1334 e->src->index, e->dest->index);
1335 }
1336 }
1337 #endif
1338 }
1339
1340
1341 /* Remove the ssa-names in the current function and translate them into normal
1342 compiler variables. PERFORM_TER is true if Temporary Expression Replacement
1343 should also be used. */
1344
1345 static void
1346 remove_ssa_form (bool perform_ter)
1347 {
1348 basic_block bb;
1349 gimple *values = NULL;
1350 var_map map;
1351 gimple_stmt_iterator gsi;
1352
1353 map = coalesce_ssa_name ();
1354
1355 /* Return to viewing the variable list as just all reference variables after
1356 coalescing has been performed. */
1357 partition_view_normal (map, false);
1358
1359 if (dump_file && (dump_flags & TDF_DETAILS))
1360 {
1361 fprintf (dump_file, "After Coalescing:\n");
1362 dump_var_map (dump_file, map);
1363 }
1364
1365 if (perform_ter)
1366 {
1367 values = find_replaceable_exprs (map);
1368 if (values && dump_file && (dump_flags & TDF_DETAILS))
1369 dump_replaceable_exprs (dump_file, values);
1370 }
1371
1372 /* Assign real variables to the partitions now. */
1373 assign_vars (map);
1374
1375 if (dump_file && (dump_flags & TDF_DETAILS))
1376 {
1377 fprintf (dump_file, "After Base variable replacement:\n");
1378 dump_var_map (dump_file, map);
1379 }
1380
1381 rewrite_trees (map, values);
1382
1383 if (values)
1384 free (values);
1385
1386 /* Remove PHI nodes which have been translated back to real variables. */
1387 FOR_EACH_BB (bb)
1388 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi);)
1389 remove_phi_node (&gsi, true);
1390
1391 /* If any copies were inserted on edges, analyze and insert them now. */
1392 perform_edge_inserts ();
1393
1394 delete_var_map (map);
1395 }
1396
1397
1398 /* Search every PHI node for arguments associated with backedges which
1399 we can trivially determine will need a copy (the argument is either
1400 not an SSA_NAME or the argument has a different underlying variable
1401 than the PHI result).
1402
1403 Insert a copy from the PHI argument to a new destination at the
1404 end of the block with the backedge to the top of the loop. Update
1405 the PHI argument to reference this new destination. */
1406
1407 static void
1408 insert_backedge_copies (void)
1409 {
1410 basic_block bb;
1411 gimple_stmt_iterator gsi;
1412
1413 FOR_EACH_BB (bb)
1414 {
1415 for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
1416 {
1417 gimple phi = gsi_stmt (gsi);
1418 tree result = gimple_phi_result (phi);
1419 tree result_var;
1420 size_t i;
1421
1422 if (!is_gimple_reg (result))
1423 continue;
1424
1425 result_var = SSA_NAME_VAR (result);
1426 for (i = 0; i < gimple_phi_num_args (phi); i++)
1427 {
1428 tree arg = gimple_phi_arg_def (phi, i);
1429 edge e = gimple_phi_arg_edge (phi, i);
1430
1431 /* If the argument is not an SSA_NAME, then we will need a
1432 constant initialization. If the argument is an SSA_NAME with
1433 a different underlying variable then a copy statement will be
1434 needed. */
1435 if ((e->flags & EDGE_DFS_BACK)
1436 && (TREE_CODE (arg) != SSA_NAME
1437 || SSA_NAME_VAR (arg) != result_var))
1438 {
1439 tree name;
1440 gimple stmt, last = NULL;
1441 gimple_stmt_iterator gsi2;
1442
1443 gsi2 = gsi_last_bb (gimple_phi_arg_edge (phi, i)->src);
1444 if (!gsi_end_p (gsi2))
1445 last = gsi_stmt (gsi2);
1446
1447 /* In theory the only way we ought to get back to the
1448 start of a loop should be with a COND_EXPR or GOTO_EXPR.
1449 However, better safe than sorry.
1450 If the block ends with a control statement or
1451 something that might throw, then we have to
1452 insert this assignment before the last
1453 statement. Else insert it after the last statement. */
1454 if (last && stmt_ends_bb_p (last))
1455 {
1456 /* If the last statement in the block is the definition
1457 site of the PHI argument, then we can't insert
1458 anything after it. */
1459 if (TREE_CODE (arg) == SSA_NAME
1460 && SSA_NAME_DEF_STMT (arg) == last)
1461 continue;
1462 }
1463
1464 /* Create a new instance of the underlying variable of the
1465 PHI result. */
1466 stmt = gimple_build_assign (result_var,
1467 gimple_phi_arg_def (phi, i));
1468 name = make_ssa_name (result_var, stmt);
1469 gimple_assign_set_lhs (stmt, name);
1470
1471 /* Insert the new statement into the block and update
1472 the PHI node. */
1473 if (last && stmt_ends_bb_p (last))
1474 gsi_insert_before (&gsi2, stmt, GSI_NEW_STMT);
1475 else
1476 gsi_insert_after (&gsi2, stmt, GSI_NEW_STMT);
1477 SET_PHI_ARG_DEF (phi, i, name);
1478 }
1479 }
1480 }
1481 }
1482 }
1483
1484 /* Take the current function out of SSA form, translating PHIs as described in
1485 R. Morgan, ``Building an Optimizing Compiler'',
1486 Butterworth-Heinemann, Boston, MA, 1998. pp 176-186. */
1487
1488 static unsigned int
1489 rewrite_out_of_ssa (void)
1490 {
1491 /* If elimination of a PHI requires inserting a copy on a backedge,
1492 then we will have to split the backedge which has numerous
1493 undesirable performance effects.
1494
1495 A significant number of such cases can be handled here by inserting
1496 copies into the loop itself. */
1497 insert_backedge_copies ();
1498
1499
1500 /* Eliminate PHIs which are of no use, such as virtual or dead phis. */
1501 eliminate_useless_phis ();
1502
1503 if (dump_file && (dump_flags & TDF_DETAILS))
1504 gimple_dump_cfg (dump_file, dump_flags & ~TDF_DETAILS);
1505
1506 remove_ssa_form (flag_tree_ter && !flag_mudflap);
1507
1508 if (dump_file && (dump_flags & TDF_DETAILS))
1509 gimple_dump_cfg (dump_file, dump_flags & ~TDF_DETAILS);
1510
1511 cfun->gimple_df->in_ssa_p = false;
1512 return 0;
1513 }
1514
1515
1516 /* Define the parameters of the out of SSA pass. */
1517
1518 struct gimple_opt_pass pass_del_ssa =
1519 {
1520 {
1521 GIMPLE_PASS,
1522 "optimized", /* name */
1523 NULL, /* gate */
1524 rewrite_out_of_ssa, /* execute */
1525 NULL, /* sub */
1526 NULL, /* next */
1527 0, /* static_pass_number */
1528 TV_TREE_SSA_TO_NORMAL, /* tv_id */
1529 PROP_cfg | PROP_ssa, /* properties_required */
1530 0, /* properties_provided */
1531 /* ??? If TER is enabled, we also kill gimple. */
1532 PROP_ssa, /* properties_destroyed */
1533 TODO_verify_ssa | TODO_verify_flow
1534 | TODO_verify_stmts, /* todo_flags_start */
1535 TODO_dump_func
1536 | TODO_ggc_collect
1537 | TODO_remove_unused_locals /* todo_flags_finish */
1538 }
1539 };