0
|
1 /* Rtl-level induction variable analysis.
|
|
2 Copyright (C) 2004, 2005, 2006, 2007, 2008, 2009
|
|
3 Free Software Foundation, Inc.
|
|
4
|
|
5 This file is part of GCC.
|
|
6
|
|
7 GCC is free software; you can redistribute it and/or modify it
|
|
8 under the terms of the GNU General Public License as published by the
|
|
9 Free Software Foundation; either version 3, or (at your option) any
|
|
10 later version.
|
|
11
|
|
12 GCC is distributed in the hope that it will be useful, but WITHOUT
|
|
13 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
15 for more details.
|
|
16
|
|
17 You should have received a copy of the GNU General Public License
|
|
18 along with GCC; see the file COPYING3. If not see
|
|
19 <http://www.gnu.org/licenses/>. */
|
|
20
|
|
21 /* This is a simple analysis of induction variables of the loop. The major use
|
|
22 is for determining the number of iterations of a loop for loop unrolling,
|
|
23 doloop optimization and branch prediction. The iv information is computed
|
|
24 on demand.
|
|
25
|
|
26 Induction variables are analyzed by walking the use-def chains. When
|
|
27 a basic induction variable (biv) is found, it is cached in the bivs
|
|
28 hash table. When register is proved to be a biv, its description
|
|
29 is stored to DF_REF_DATA of the def reference.
|
|
30
|
|
31 The analysis works always with one loop -- you must call
|
|
32 iv_analysis_loop_init (loop) for it. All the other functions then work with
|
|
33 this loop. When you need to work with another loop, just call
|
|
34 iv_analysis_loop_init for it. When you no longer need iv analysis, call
|
|
35 iv_analysis_done () to clean up the memory.
|
|
36
|
|
37 The available functions are:
|
|
38
|
|
39 iv_analyze (insn, reg, iv): Stores the description of the induction variable
|
|
40 corresponding to the use of register REG in INSN to IV. Returns true if
|
|
41 REG is an induction variable in INSN. false otherwise.
|
|
42 If use of REG is not found in INSN, following insns are scanned (so that
|
|
43 we may call this function on insn returned by get_condition).
|
|
44 iv_analyze_result (insn, def, iv): Stores to IV the description of the iv
|
|
45 corresponding to DEF, which is a register defined in INSN.
|
|
46 iv_analyze_expr (insn, rhs, mode, iv): Stores to IV the description of iv
|
|
47 corresponding to expression EXPR evaluated at INSN. All registers used bu
|
|
48 EXPR must also be used in INSN.
|
|
49 */
|
|
50
|
|
51 #include "config.h"
|
|
52 #include "system.h"
|
|
53 #include "coretypes.h"
|
|
54 #include "tm.h"
|
|
55 #include "rtl.h"
|
|
56 #include "hard-reg-set.h"
|
|
57 #include "obstack.h"
|
|
58 #include "basic-block.h"
|
|
59 #include "cfgloop.h"
|
|
60 #include "expr.h"
|
|
61 #include "intl.h"
|
|
62 #include "output.h"
|
|
63 #include "toplev.h"
|
|
64 #include "df.h"
|
|
65 #include "hashtab.h"
|
|
66
|
|
67 /* Possible return values of iv_get_reaching_def. */
|
|
68
|
|
69 enum iv_grd_result
|
|
70 {
|
|
71 /* More than one reaching def, or reaching def that does not
|
|
72 dominate the use. */
|
|
73 GRD_INVALID,
|
|
74
|
|
75 /* The use is trivial invariant of the loop, i.e. is not changed
|
|
76 inside the loop. */
|
|
77 GRD_INVARIANT,
|
|
78
|
|
79 /* The use is reached by initial value and a value from the
|
|
80 previous iteration. */
|
|
81 GRD_MAYBE_BIV,
|
|
82
|
|
83 /* The use has single dominating def. */
|
|
84 GRD_SINGLE_DOM
|
|
85 };
|
|
86
|
|
87 /* Information about a biv. */
|
|
88
|
|
89 struct biv_entry
|
|
90 {
|
|
91 unsigned regno; /* The register of the biv. */
|
|
92 struct rtx_iv iv; /* Value of the biv. */
|
|
93 };
|
|
94
|
|
95 static bool clean_slate = true;
|
|
96
|
|
97 static unsigned int iv_ref_table_size = 0;
|
|
98
|
|
99 /* Table of rtx_ivs indexed by the df_ref uid field. */
|
|
100 static struct rtx_iv ** iv_ref_table;
|
|
101
|
|
102 /* Induction variable stored at the reference. */
|
|
103 #define DF_REF_IV(REF) iv_ref_table[DF_REF_ID(REF)]
|
|
104 #define DF_REF_IV_SET(REF, IV) iv_ref_table[DF_REF_ID(REF)] = (IV)
|
|
105
|
|
106 /* The current loop. */
|
|
107
|
|
108 static struct loop *current_loop;
|
|
109
|
|
110 /* Bivs of the current loop. */
|
|
111
|
|
112 static htab_t bivs;
|
|
113
|
|
114 static bool iv_analyze_op (rtx, rtx, struct rtx_iv *);
|
|
115
|
|
116 /* Dumps information about IV to FILE. */
|
|
117
|
|
118 extern void dump_iv_info (FILE *, struct rtx_iv *);
|
|
119 void
|
|
120 dump_iv_info (FILE *file, struct rtx_iv *iv)
|
|
121 {
|
|
122 if (!iv->base)
|
|
123 {
|
|
124 fprintf (file, "not simple");
|
|
125 return;
|
|
126 }
|
|
127
|
|
128 if (iv->step == const0_rtx
|
|
129 && !iv->first_special)
|
|
130 fprintf (file, "invariant ");
|
|
131
|
|
132 print_rtl (file, iv->base);
|
|
133 if (iv->step != const0_rtx)
|
|
134 {
|
|
135 fprintf (file, " + ");
|
|
136 print_rtl (file, iv->step);
|
|
137 fprintf (file, " * iteration");
|
|
138 }
|
|
139 fprintf (file, " (in %s)", GET_MODE_NAME (iv->mode));
|
|
140
|
|
141 if (iv->mode != iv->extend_mode)
|
|
142 fprintf (file, " %s to %s",
|
|
143 rtx_name[iv->extend],
|
|
144 GET_MODE_NAME (iv->extend_mode));
|
|
145
|
|
146 if (iv->mult != const1_rtx)
|
|
147 {
|
|
148 fprintf (file, " * ");
|
|
149 print_rtl (file, iv->mult);
|
|
150 }
|
|
151 if (iv->delta != const0_rtx)
|
|
152 {
|
|
153 fprintf (file, " + ");
|
|
154 print_rtl (file, iv->delta);
|
|
155 }
|
|
156 if (iv->first_special)
|
|
157 fprintf (file, " (first special)");
|
|
158 }
|
|
159
|
|
160 /* Generates a subreg to get the least significant part of EXPR (in mode
|
|
161 INNER_MODE) to OUTER_MODE. */
|
|
162
|
|
163 rtx
|
|
164 lowpart_subreg (enum machine_mode outer_mode, rtx expr,
|
|
165 enum machine_mode inner_mode)
|
|
166 {
|
|
167 return simplify_gen_subreg (outer_mode, expr, inner_mode,
|
|
168 subreg_lowpart_offset (outer_mode, inner_mode));
|
|
169 }
|
|
170
|
|
171 static void
|
|
172 check_iv_ref_table_size (void)
|
|
173 {
|
|
174 if (iv_ref_table_size < DF_DEFS_TABLE_SIZE())
|
|
175 {
|
|
176 unsigned int new_size = DF_DEFS_TABLE_SIZE () + (DF_DEFS_TABLE_SIZE () / 4);
|
|
177 iv_ref_table = XRESIZEVEC (struct rtx_iv *, iv_ref_table, new_size);
|
|
178 memset (&iv_ref_table[iv_ref_table_size], 0,
|
|
179 (new_size - iv_ref_table_size) * sizeof (struct rtx_iv *));
|
|
180 iv_ref_table_size = new_size;
|
|
181 }
|
|
182 }
|
|
183
|
|
184
|
|
185 /* Checks whether REG is a well-behaved register. */
|
|
186
|
|
187 static bool
|
|
188 simple_reg_p (rtx reg)
|
|
189 {
|
|
190 unsigned r;
|
|
191
|
|
192 if (GET_CODE (reg) == SUBREG)
|
|
193 {
|
|
194 if (!subreg_lowpart_p (reg))
|
|
195 return false;
|
|
196 reg = SUBREG_REG (reg);
|
|
197 }
|
|
198
|
|
199 if (!REG_P (reg))
|
|
200 return false;
|
|
201
|
|
202 r = REGNO (reg);
|
|
203 if (HARD_REGISTER_NUM_P (r))
|
|
204 return false;
|
|
205
|
|
206 if (GET_MODE_CLASS (GET_MODE (reg)) != MODE_INT)
|
|
207 return false;
|
|
208
|
|
209 return true;
|
|
210 }
|
|
211
|
|
212 /* Clears the information about ivs stored in df. */
|
|
213
|
|
214 static void
|
|
215 clear_iv_info (void)
|
|
216 {
|
|
217 unsigned i, n_defs = DF_DEFS_TABLE_SIZE ();
|
|
218 struct rtx_iv *iv;
|
|
219
|
|
220 check_iv_ref_table_size ();
|
|
221 for (i = 0; i < n_defs; i++)
|
|
222 {
|
|
223 iv = iv_ref_table[i];
|
|
224 if (iv)
|
|
225 {
|
|
226 free (iv);
|
|
227 iv_ref_table[i] = NULL;
|
|
228 }
|
|
229 }
|
|
230
|
|
231 htab_empty (bivs);
|
|
232 }
|
|
233
|
|
234 /* Returns hash value for biv B. */
|
|
235
|
|
236 static hashval_t
|
|
237 biv_hash (const void *b)
|
|
238 {
|
|
239 return ((const struct biv_entry *) b)->regno;
|
|
240 }
|
|
241
|
|
242 /* Compares biv B and register R. */
|
|
243
|
|
244 static int
|
|
245 biv_eq (const void *b, const void *r)
|
|
246 {
|
|
247 return ((const struct biv_entry *) b)->regno == REGNO ((const_rtx) r);
|
|
248 }
|
|
249
|
|
250 /* Prepare the data for an induction variable analysis of a LOOP. */
|
|
251
|
|
252 void
|
|
253 iv_analysis_loop_init (struct loop *loop)
|
|
254 {
|
|
255 basic_block *body = get_loop_body_in_dom_order (loop), bb;
|
|
256 bitmap blocks = BITMAP_ALLOC (NULL);
|
|
257 unsigned i;
|
|
258
|
|
259 current_loop = loop;
|
|
260
|
|
261 /* Clear the information from the analysis of the previous loop. */
|
|
262 if (clean_slate)
|
|
263 {
|
|
264 df_set_flags (DF_EQ_NOTES + DF_DEFER_INSN_RESCAN);
|
|
265 bivs = htab_create (10, biv_hash, biv_eq, free);
|
|
266 clean_slate = false;
|
|
267 }
|
|
268 else
|
|
269 clear_iv_info ();
|
|
270
|
|
271 for (i = 0; i < loop->num_nodes; i++)
|
|
272 {
|
|
273 bb = body[i];
|
|
274 bitmap_set_bit (blocks, bb->index);
|
|
275 }
|
|
276 /* Get rid of the ud chains before processing the rescans. Then add
|
|
277 the problem back. */
|
|
278 df_remove_problem (df_chain);
|
|
279 df_process_deferred_rescans ();
|
|
280 df_chain_add_problem (DF_UD_CHAIN);
|
|
281 df_set_blocks (blocks);
|
|
282 df_analyze ();
|
|
283 if (dump_file)
|
|
284 df_dump_region (dump_file);
|
|
285
|
|
286 check_iv_ref_table_size ();
|
|
287 BITMAP_FREE (blocks);
|
|
288 free (body);
|
|
289 }
|
|
290
|
|
291 /* Finds the definition of REG that dominates loop latch and stores
|
|
292 it to DEF. Returns false if there is not a single definition
|
|
293 dominating the latch. If REG has no definition in loop, DEF
|
|
294 is set to NULL and true is returned. */
|
|
295
|
|
296 static bool
|
|
297 latch_dominating_def (rtx reg, df_ref *def)
|
|
298 {
|
|
299 df_ref single_rd = NULL, adef;
|
|
300 unsigned regno = REGNO (reg);
|
|
301 struct df_rd_bb_info *bb_info = DF_RD_BB_INFO (current_loop->latch);
|
|
302
|
|
303 for (adef = DF_REG_DEF_CHAIN (regno); adef; adef = DF_REF_NEXT_REG (adef))
|
|
304 {
|
|
305 if (!bitmap_bit_p (df->blocks_to_analyze, DF_REF_BBNO (adef))
|
|
306 || !bitmap_bit_p (bb_info->out, DF_REF_ID (adef)))
|
|
307 continue;
|
|
308
|
|
309 /* More than one reaching definition. */
|
|
310 if (single_rd)
|
|
311 return false;
|
|
312
|
|
313 if (!just_once_each_iteration_p (current_loop, DF_REF_BB (adef)))
|
|
314 return false;
|
|
315
|
|
316 single_rd = adef;
|
|
317 }
|
|
318
|
|
319 *def = single_rd;
|
|
320 return true;
|
|
321 }
|
|
322
|
|
323 /* Gets definition of REG reaching its use in INSN and stores it to DEF. */
|
|
324
|
|
325 static enum iv_grd_result
|
|
326 iv_get_reaching_def (rtx insn, rtx reg, df_ref *def)
|
|
327 {
|
|
328 df_ref use, adef;
|
|
329 basic_block def_bb, use_bb;
|
|
330 rtx def_insn;
|
|
331 bool dom_p;
|
|
332
|
|
333 *def = NULL;
|
|
334 if (!simple_reg_p (reg))
|
|
335 return GRD_INVALID;
|
|
336 if (GET_CODE (reg) == SUBREG)
|
|
337 reg = SUBREG_REG (reg);
|
|
338 gcc_assert (REG_P (reg));
|
|
339
|
|
340 use = df_find_use (insn, reg);
|
|
341 gcc_assert (use != NULL);
|
|
342
|
|
343 if (!DF_REF_CHAIN (use))
|
|
344 return GRD_INVARIANT;
|
|
345
|
|
346 /* More than one reaching def. */
|
|
347 if (DF_REF_CHAIN (use)->next)
|
|
348 return GRD_INVALID;
|
|
349
|
|
350 adef = DF_REF_CHAIN (use)->ref;
|
|
351
|
|
352 /* We do not handle setting only part of the register. */
|
|
353 if (DF_REF_FLAGS (adef) & DF_REF_READ_WRITE)
|
|
354 return GRD_INVALID;
|
|
355
|
|
356 def_insn = DF_REF_INSN (adef);
|
|
357 def_bb = DF_REF_BB (adef);
|
|
358 use_bb = BLOCK_FOR_INSN (insn);
|
|
359
|
|
360 if (use_bb == def_bb)
|
|
361 dom_p = (DF_INSN_LUID (def_insn) < DF_INSN_LUID (insn));
|
|
362 else
|
|
363 dom_p = dominated_by_p (CDI_DOMINATORS, use_bb, def_bb);
|
|
364
|
|
365 if (dom_p)
|
|
366 {
|
|
367 *def = adef;
|
|
368 return GRD_SINGLE_DOM;
|
|
369 }
|
|
370
|
|
371 /* The definition does not dominate the use. This is still OK if
|
|
372 this may be a use of a biv, i.e. if the def_bb dominates loop
|
|
373 latch. */
|
|
374 if (just_once_each_iteration_p (current_loop, def_bb))
|
|
375 return GRD_MAYBE_BIV;
|
|
376
|
|
377 return GRD_INVALID;
|
|
378 }
|
|
379
|
|
380 /* Sets IV to invariant CST in MODE. Always returns true (just for
|
|
381 consistency with other iv manipulation functions that may fail). */
|
|
382
|
|
383 static bool
|
|
384 iv_constant (struct rtx_iv *iv, rtx cst, enum machine_mode mode)
|
|
385 {
|
|
386 if (mode == VOIDmode)
|
|
387 mode = GET_MODE (cst);
|
|
388
|
|
389 iv->mode = mode;
|
|
390 iv->base = cst;
|
|
391 iv->step = const0_rtx;
|
|
392 iv->first_special = false;
|
|
393 iv->extend = UNKNOWN;
|
|
394 iv->extend_mode = iv->mode;
|
|
395 iv->delta = const0_rtx;
|
|
396 iv->mult = const1_rtx;
|
|
397
|
|
398 return true;
|
|
399 }
|
|
400
|
|
401 /* Evaluates application of subreg to MODE on IV. */
|
|
402
|
|
403 static bool
|
|
404 iv_subreg (struct rtx_iv *iv, enum machine_mode mode)
|
|
405 {
|
|
406 /* If iv is invariant, just calculate the new value. */
|
|
407 if (iv->step == const0_rtx
|
|
408 && !iv->first_special)
|
|
409 {
|
|
410 rtx val = get_iv_value (iv, const0_rtx);
|
|
411 val = lowpart_subreg (mode, val, iv->extend_mode);
|
|
412
|
|
413 iv->base = val;
|
|
414 iv->extend = UNKNOWN;
|
|
415 iv->mode = iv->extend_mode = mode;
|
|
416 iv->delta = const0_rtx;
|
|
417 iv->mult = const1_rtx;
|
|
418 return true;
|
|
419 }
|
|
420
|
|
421 if (iv->extend_mode == mode)
|
|
422 return true;
|
|
423
|
|
424 if (GET_MODE_BITSIZE (mode) > GET_MODE_BITSIZE (iv->mode))
|
|
425 return false;
|
|
426
|
|
427 iv->extend = UNKNOWN;
|
|
428 iv->mode = mode;
|
|
429
|
|
430 iv->base = simplify_gen_binary (PLUS, iv->extend_mode, iv->delta,
|
|
431 simplify_gen_binary (MULT, iv->extend_mode,
|
|
432 iv->base, iv->mult));
|
|
433 iv->step = simplify_gen_binary (MULT, iv->extend_mode, iv->step, iv->mult);
|
|
434 iv->mult = const1_rtx;
|
|
435 iv->delta = const0_rtx;
|
|
436 iv->first_special = false;
|
|
437
|
|
438 return true;
|
|
439 }
|
|
440
|
|
441 /* Evaluates application of EXTEND to MODE on IV. */
|
|
442
|
|
443 static bool
|
|
444 iv_extend (struct rtx_iv *iv, enum rtx_code extend, enum machine_mode mode)
|
|
445 {
|
|
446 /* If iv is invariant, just calculate the new value. */
|
|
447 if (iv->step == const0_rtx
|
|
448 && !iv->first_special)
|
|
449 {
|
|
450 rtx val = get_iv_value (iv, const0_rtx);
|
|
451 val = simplify_gen_unary (extend, mode, val, iv->extend_mode);
|
|
452
|
|
453 iv->base = val;
|
|
454 iv->extend = UNKNOWN;
|
|
455 iv->mode = iv->extend_mode = mode;
|
|
456 iv->delta = const0_rtx;
|
|
457 iv->mult = const1_rtx;
|
|
458 return true;
|
|
459 }
|
|
460
|
|
461 if (mode != iv->extend_mode)
|
|
462 return false;
|
|
463
|
|
464 if (iv->extend != UNKNOWN
|
|
465 && iv->extend != extend)
|
|
466 return false;
|
|
467
|
|
468 iv->extend = extend;
|
|
469
|
|
470 return true;
|
|
471 }
|
|
472
|
|
473 /* Evaluates negation of IV. */
|
|
474
|
|
475 static bool
|
|
476 iv_neg (struct rtx_iv *iv)
|
|
477 {
|
|
478 if (iv->extend == UNKNOWN)
|
|
479 {
|
|
480 iv->base = simplify_gen_unary (NEG, iv->extend_mode,
|
|
481 iv->base, iv->extend_mode);
|
|
482 iv->step = simplify_gen_unary (NEG, iv->extend_mode,
|
|
483 iv->step, iv->extend_mode);
|
|
484 }
|
|
485 else
|
|
486 {
|
|
487 iv->delta = simplify_gen_unary (NEG, iv->extend_mode,
|
|
488 iv->delta, iv->extend_mode);
|
|
489 iv->mult = simplify_gen_unary (NEG, iv->extend_mode,
|
|
490 iv->mult, iv->extend_mode);
|
|
491 }
|
|
492
|
|
493 return true;
|
|
494 }
|
|
495
|
|
496 /* Evaluates addition or subtraction (according to OP) of IV1 to IV0. */
|
|
497
|
|
498 static bool
|
|
499 iv_add (struct rtx_iv *iv0, struct rtx_iv *iv1, enum rtx_code op)
|
|
500 {
|
|
501 enum machine_mode mode;
|
|
502 rtx arg;
|
|
503
|
|
504 /* Extend the constant to extend_mode of the other operand if necessary. */
|
|
505 if (iv0->extend == UNKNOWN
|
|
506 && iv0->mode == iv0->extend_mode
|
|
507 && iv0->step == const0_rtx
|
|
508 && GET_MODE_SIZE (iv0->extend_mode) < GET_MODE_SIZE (iv1->extend_mode))
|
|
509 {
|
|
510 iv0->extend_mode = iv1->extend_mode;
|
|
511 iv0->base = simplify_gen_unary (ZERO_EXTEND, iv0->extend_mode,
|
|
512 iv0->base, iv0->mode);
|
|
513 }
|
|
514 if (iv1->extend == UNKNOWN
|
|
515 && iv1->mode == iv1->extend_mode
|
|
516 && iv1->step == const0_rtx
|
|
517 && GET_MODE_SIZE (iv1->extend_mode) < GET_MODE_SIZE (iv0->extend_mode))
|
|
518 {
|
|
519 iv1->extend_mode = iv0->extend_mode;
|
|
520 iv1->base = simplify_gen_unary (ZERO_EXTEND, iv1->extend_mode,
|
|
521 iv1->base, iv1->mode);
|
|
522 }
|
|
523
|
|
524 mode = iv0->extend_mode;
|
|
525 if (mode != iv1->extend_mode)
|
|
526 return false;
|
|
527
|
|
528 if (iv0->extend == UNKNOWN && iv1->extend == UNKNOWN)
|
|
529 {
|
|
530 if (iv0->mode != iv1->mode)
|
|
531 return false;
|
|
532
|
|
533 iv0->base = simplify_gen_binary (op, mode, iv0->base, iv1->base);
|
|
534 iv0->step = simplify_gen_binary (op, mode, iv0->step, iv1->step);
|
|
535
|
|
536 return true;
|
|
537 }
|
|
538
|
|
539 /* Handle addition of constant. */
|
|
540 if (iv1->extend == UNKNOWN
|
|
541 && iv1->mode == mode
|
|
542 && iv1->step == const0_rtx)
|
|
543 {
|
|
544 iv0->delta = simplify_gen_binary (op, mode, iv0->delta, iv1->base);
|
|
545 return true;
|
|
546 }
|
|
547
|
|
548 if (iv0->extend == UNKNOWN
|
|
549 && iv0->mode == mode
|
|
550 && iv0->step == const0_rtx)
|
|
551 {
|
|
552 arg = iv0->base;
|
|
553 *iv0 = *iv1;
|
|
554 if (op == MINUS
|
|
555 && !iv_neg (iv0))
|
|
556 return false;
|
|
557
|
|
558 iv0->delta = simplify_gen_binary (PLUS, mode, iv0->delta, arg);
|
|
559 return true;
|
|
560 }
|
|
561
|
|
562 return false;
|
|
563 }
|
|
564
|
|
565 /* Evaluates multiplication of IV by constant CST. */
|
|
566
|
|
567 static bool
|
|
568 iv_mult (struct rtx_iv *iv, rtx mby)
|
|
569 {
|
|
570 enum machine_mode mode = iv->extend_mode;
|
|
571
|
|
572 if (GET_MODE (mby) != VOIDmode
|
|
573 && GET_MODE (mby) != mode)
|
|
574 return false;
|
|
575
|
|
576 if (iv->extend == UNKNOWN)
|
|
577 {
|
|
578 iv->base = simplify_gen_binary (MULT, mode, iv->base, mby);
|
|
579 iv->step = simplify_gen_binary (MULT, mode, iv->step, mby);
|
|
580 }
|
|
581 else
|
|
582 {
|
|
583 iv->delta = simplify_gen_binary (MULT, mode, iv->delta, mby);
|
|
584 iv->mult = simplify_gen_binary (MULT, mode, iv->mult, mby);
|
|
585 }
|
|
586
|
|
587 return true;
|
|
588 }
|
|
589
|
|
590 /* Evaluates shift of IV by constant CST. */
|
|
591
|
|
592 static bool
|
|
593 iv_shift (struct rtx_iv *iv, rtx mby)
|
|
594 {
|
|
595 enum machine_mode mode = iv->extend_mode;
|
|
596
|
|
597 if (GET_MODE (mby) != VOIDmode
|
|
598 && GET_MODE (mby) != mode)
|
|
599 return false;
|
|
600
|
|
601 if (iv->extend == UNKNOWN)
|
|
602 {
|
|
603 iv->base = simplify_gen_binary (ASHIFT, mode, iv->base, mby);
|
|
604 iv->step = simplify_gen_binary (ASHIFT, mode, iv->step, mby);
|
|
605 }
|
|
606 else
|
|
607 {
|
|
608 iv->delta = simplify_gen_binary (ASHIFT, mode, iv->delta, mby);
|
|
609 iv->mult = simplify_gen_binary (ASHIFT, mode, iv->mult, mby);
|
|
610 }
|
|
611
|
|
612 return true;
|
|
613 }
|
|
614
|
|
615 /* The recursive part of get_biv_step. Gets the value of the single value
|
|
616 defined by DEF wrto initial value of REG inside loop, in shape described
|
|
617 at get_biv_step. */
|
|
618
|
|
619 static bool
|
|
620 get_biv_step_1 (df_ref def, rtx reg,
|
|
621 rtx *inner_step, enum machine_mode *inner_mode,
|
|
622 enum rtx_code *extend, enum machine_mode outer_mode,
|
|
623 rtx *outer_step)
|
|
624 {
|
|
625 rtx set, rhs, op0 = NULL_RTX, op1 = NULL_RTX;
|
|
626 rtx next, nextr, tmp;
|
|
627 enum rtx_code code;
|
|
628 rtx insn = DF_REF_INSN (def);
|
|
629 df_ref next_def;
|
|
630 enum iv_grd_result res;
|
|
631
|
|
632 set = single_set (insn);
|
|
633 if (!set)
|
|
634 return false;
|
|
635
|
|
636 rhs = find_reg_equal_equiv_note (insn);
|
|
637 if (rhs)
|
|
638 rhs = XEXP (rhs, 0);
|
|
639 else
|
|
640 rhs = SET_SRC (set);
|
|
641
|
|
642 code = GET_CODE (rhs);
|
|
643 switch (code)
|
|
644 {
|
|
645 case SUBREG:
|
|
646 case REG:
|
|
647 next = rhs;
|
|
648 break;
|
|
649
|
|
650 case PLUS:
|
|
651 case MINUS:
|
|
652 op0 = XEXP (rhs, 0);
|
|
653 op1 = XEXP (rhs, 1);
|
|
654
|
|
655 if (code == PLUS && CONSTANT_P (op0))
|
|
656 {
|
|
657 tmp = op0; op0 = op1; op1 = tmp;
|
|
658 }
|
|
659
|
|
660 if (!simple_reg_p (op0)
|
|
661 || !CONSTANT_P (op1))
|
|
662 return false;
|
|
663
|
|
664 if (GET_MODE (rhs) != outer_mode)
|
|
665 {
|
|
666 /* ppc64 uses expressions like
|
|
667
|
|
668 (set x:SI (plus:SI (subreg:SI y:DI) 1)).
|
|
669
|
|
670 this is equivalent to
|
|
671
|
|
672 (set x':DI (plus:DI y:DI 1))
|
|
673 (set x:SI (subreg:SI (x':DI)). */
|
|
674 if (GET_CODE (op0) != SUBREG)
|
|
675 return false;
|
|
676 if (GET_MODE (SUBREG_REG (op0)) != outer_mode)
|
|
677 return false;
|
|
678 }
|
|
679
|
|
680 next = op0;
|
|
681 break;
|
|
682
|
|
683 case SIGN_EXTEND:
|
|
684 case ZERO_EXTEND:
|
|
685 if (GET_MODE (rhs) != outer_mode)
|
|
686 return false;
|
|
687
|
|
688 op0 = XEXP (rhs, 0);
|
|
689 if (!simple_reg_p (op0))
|
|
690 return false;
|
|
691
|
|
692 next = op0;
|
|
693 break;
|
|
694
|
|
695 default:
|
|
696 return false;
|
|
697 }
|
|
698
|
|
699 if (GET_CODE (next) == SUBREG)
|
|
700 {
|
|
701 if (!subreg_lowpart_p (next))
|
|
702 return false;
|
|
703
|
|
704 nextr = SUBREG_REG (next);
|
|
705 if (GET_MODE (nextr) != outer_mode)
|
|
706 return false;
|
|
707 }
|
|
708 else
|
|
709 nextr = next;
|
|
710
|
|
711 res = iv_get_reaching_def (insn, nextr, &next_def);
|
|
712
|
|
713 if (res == GRD_INVALID || res == GRD_INVARIANT)
|
|
714 return false;
|
|
715
|
|
716 if (res == GRD_MAYBE_BIV)
|
|
717 {
|
|
718 if (!rtx_equal_p (nextr, reg))
|
|
719 return false;
|
|
720
|
|
721 *inner_step = const0_rtx;
|
|
722 *extend = UNKNOWN;
|
|
723 *inner_mode = outer_mode;
|
|
724 *outer_step = const0_rtx;
|
|
725 }
|
|
726 else if (!get_biv_step_1 (next_def, reg,
|
|
727 inner_step, inner_mode, extend, outer_mode,
|
|
728 outer_step))
|
|
729 return false;
|
|
730
|
|
731 if (GET_CODE (next) == SUBREG)
|
|
732 {
|
|
733 enum machine_mode amode = GET_MODE (next);
|
|
734
|
|
735 if (GET_MODE_SIZE (amode) > GET_MODE_SIZE (*inner_mode))
|
|
736 return false;
|
|
737
|
|
738 *inner_mode = amode;
|
|
739 *inner_step = simplify_gen_binary (PLUS, outer_mode,
|
|
740 *inner_step, *outer_step);
|
|
741 *outer_step = const0_rtx;
|
|
742 *extend = UNKNOWN;
|
|
743 }
|
|
744
|
|
745 switch (code)
|
|
746 {
|
|
747 case REG:
|
|
748 case SUBREG:
|
|
749 break;
|
|
750
|
|
751 case PLUS:
|
|
752 case MINUS:
|
|
753 if (*inner_mode == outer_mode
|
|
754 /* See comment in previous switch. */
|
|
755 || GET_MODE (rhs) != outer_mode)
|
|
756 *inner_step = simplify_gen_binary (code, outer_mode,
|
|
757 *inner_step, op1);
|
|
758 else
|
|
759 *outer_step = simplify_gen_binary (code, outer_mode,
|
|
760 *outer_step, op1);
|
|
761 break;
|
|
762
|
|
763 case SIGN_EXTEND:
|
|
764 case ZERO_EXTEND:
|
|
765 gcc_assert (GET_MODE (op0) == *inner_mode
|
|
766 && *extend == UNKNOWN
|
|
767 && *outer_step == const0_rtx);
|
|
768
|
|
769 *extend = code;
|
|
770 break;
|
|
771
|
|
772 default:
|
|
773 return false;
|
|
774 }
|
|
775
|
|
776 return true;
|
|
777 }
|
|
778
|
|
779 /* Gets the operation on register REG inside loop, in shape
|
|
780
|
|
781 OUTER_STEP + EXTEND_{OUTER_MODE} (SUBREG_{INNER_MODE} (REG + INNER_STEP))
|
|
782
|
|
783 If the operation cannot be described in this shape, return false.
|
|
784 LAST_DEF is the definition of REG that dominates loop latch. */
|
|
785
|
|
786 static bool
|
|
787 get_biv_step (df_ref last_def, rtx reg, rtx *inner_step,
|
|
788 enum machine_mode *inner_mode, enum rtx_code *extend,
|
|
789 enum machine_mode *outer_mode, rtx *outer_step)
|
|
790 {
|
|
791 *outer_mode = GET_MODE (reg);
|
|
792
|
|
793 if (!get_biv_step_1 (last_def, reg,
|
|
794 inner_step, inner_mode, extend, *outer_mode,
|
|
795 outer_step))
|
|
796 return false;
|
|
797
|
|
798 gcc_assert ((*inner_mode == *outer_mode) != (*extend != UNKNOWN));
|
|
799 gcc_assert (*inner_mode != *outer_mode || *outer_step == const0_rtx);
|
|
800
|
|
801 return true;
|
|
802 }
|
|
803
|
|
804 /* Records information that DEF is induction variable IV. */
|
|
805
|
|
806 static void
|
|
807 record_iv (df_ref def, struct rtx_iv *iv)
|
|
808 {
|
|
809 struct rtx_iv *recorded_iv = XNEW (struct rtx_iv);
|
|
810
|
|
811 *recorded_iv = *iv;
|
|
812 check_iv_ref_table_size ();
|
|
813 DF_REF_IV_SET (def, recorded_iv);
|
|
814 }
|
|
815
|
|
816 /* If DEF was already analyzed for bivness, store the description of the biv to
|
|
817 IV and return true. Otherwise return false. */
|
|
818
|
|
819 static bool
|
|
820 analyzed_for_bivness_p (rtx def, struct rtx_iv *iv)
|
|
821 {
|
|
822 struct biv_entry *biv =
|
|
823 (struct biv_entry *) htab_find_with_hash (bivs, def, REGNO (def));
|
|
824
|
|
825 if (!biv)
|
|
826 return false;
|
|
827
|
|
828 *iv = biv->iv;
|
|
829 return true;
|
|
830 }
|
|
831
|
|
832 static void
|
|
833 record_biv (rtx def, struct rtx_iv *iv)
|
|
834 {
|
|
835 struct biv_entry *biv = XNEW (struct biv_entry);
|
|
836 void **slot = htab_find_slot_with_hash (bivs, def, REGNO (def), INSERT);
|
|
837
|
|
838 biv->regno = REGNO (def);
|
|
839 biv->iv = *iv;
|
|
840 gcc_assert (!*slot);
|
|
841 *slot = biv;
|
|
842 }
|
|
843
|
|
844 /* Determines whether DEF is a biv and if so, stores its description
|
|
845 to *IV. */
|
|
846
|
|
847 static bool
|
|
848 iv_analyze_biv (rtx def, struct rtx_iv *iv)
|
|
849 {
|
|
850 rtx inner_step, outer_step;
|
|
851 enum machine_mode inner_mode, outer_mode;
|
|
852 enum rtx_code extend;
|
|
853 df_ref last_def;
|
|
854
|
|
855 if (dump_file)
|
|
856 {
|
|
857 fprintf (dump_file, "Analyzing ");
|
|
858 print_rtl (dump_file, def);
|
|
859 fprintf (dump_file, " for bivness.\n");
|
|
860 }
|
|
861
|
|
862 if (!REG_P (def))
|
|
863 {
|
|
864 if (!CONSTANT_P (def))
|
|
865 return false;
|
|
866
|
|
867 return iv_constant (iv, def, VOIDmode);
|
|
868 }
|
|
869
|
|
870 if (!latch_dominating_def (def, &last_def))
|
|
871 {
|
|
872 if (dump_file)
|
|
873 fprintf (dump_file, " not simple.\n");
|
|
874 return false;
|
|
875 }
|
|
876
|
|
877 if (!last_def)
|
|
878 return iv_constant (iv, def, VOIDmode);
|
|
879
|
|
880 if (analyzed_for_bivness_p (def, iv))
|
|
881 {
|
|
882 if (dump_file)
|
|
883 fprintf (dump_file, " already analysed.\n");
|
|
884 return iv->base != NULL_RTX;
|
|
885 }
|
|
886
|
|
887 if (!get_biv_step (last_def, def, &inner_step, &inner_mode, &extend,
|
|
888 &outer_mode, &outer_step))
|
|
889 {
|
|
890 iv->base = NULL_RTX;
|
|
891 goto end;
|
|
892 }
|
|
893
|
|
894 /* Loop transforms base to es (base + inner_step) + outer_step,
|
|
895 where es means extend of subreg between inner_mode and outer_mode.
|
|
896 The corresponding induction variable is
|
|
897
|
|
898 es ((base - outer_step) + i * (inner_step + outer_step)) + outer_step */
|
|
899
|
|
900 iv->base = simplify_gen_binary (MINUS, outer_mode, def, outer_step);
|
|
901 iv->step = simplify_gen_binary (PLUS, outer_mode, inner_step, outer_step);
|
|
902 iv->mode = inner_mode;
|
|
903 iv->extend_mode = outer_mode;
|
|
904 iv->extend = extend;
|
|
905 iv->mult = const1_rtx;
|
|
906 iv->delta = outer_step;
|
|
907 iv->first_special = inner_mode != outer_mode;
|
|
908
|
|
909 end:
|
|
910 if (dump_file)
|
|
911 {
|
|
912 fprintf (dump_file, " ");
|
|
913 dump_iv_info (dump_file, iv);
|
|
914 fprintf (dump_file, "\n");
|
|
915 }
|
|
916
|
|
917 record_biv (def, iv);
|
|
918 return iv->base != NULL_RTX;
|
|
919 }
|
|
920
|
|
921 /* Analyzes expression RHS used at INSN and stores the result to *IV.
|
|
922 The mode of the induction variable is MODE. */
|
|
923
|
|
924 bool
|
|
925 iv_analyze_expr (rtx insn, rtx rhs, enum machine_mode mode, struct rtx_iv *iv)
|
|
926 {
|
|
927 rtx mby = NULL_RTX, tmp;
|
|
928 rtx op0 = NULL_RTX, op1 = NULL_RTX;
|
|
929 struct rtx_iv iv0, iv1;
|
|
930 enum rtx_code code = GET_CODE (rhs);
|
|
931 enum machine_mode omode = mode;
|
|
932
|
|
933 iv->mode = VOIDmode;
|
|
934 iv->base = NULL_RTX;
|
|
935 iv->step = NULL_RTX;
|
|
936
|
|
937 gcc_assert (GET_MODE (rhs) == mode || GET_MODE (rhs) == VOIDmode);
|
|
938
|
|
939 if (CONSTANT_P (rhs)
|
|
940 || REG_P (rhs)
|
|
941 || code == SUBREG)
|
|
942 {
|
|
943 if (!iv_analyze_op (insn, rhs, iv))
|
|
944 return false;
|
|
945
|
|
946 if (iv->mode == VOIDmode)
|
|
947 {
|
|
948 iv->mode = mode;
|
|
949 iv->extend_mode = mode;
|
|
950 }
|
|
951
|
|
952 return true;
|
|
953 }
|
|
954
|
|
955 switch (code)
|
|
956 {
|
|
957 case REG:
|
|
958 op0 = rhs;
|
|
959 break;
|
|
960
|
|
961 case SIGN_EXTEND:
|
|
962 case ZERO_EXTEND:
|
|
963 case NEG:
|
|
964 op0 = XEXP (rhs, 0);
|
|
965 omode = GET_MODE (op0);
|
|
966 break;
|
|
967
|
|
968 case PLUS:
|
|
969 case MINUS:
|
|
970 op0 = XEXP (rhs, 0);
|
|
971 op1 = XEXP (rhs, 1);
|
|
972 break;
|
|
973
|
|
974 case MULT:
|
|
975 op0 = XEXP (rhs, 0);
|
|
976 mby = XEXP (rhs, 1);
|
|
977 if (!CONSTANT_P (mby))
|
|
978 {
|
|
979 tmp = op0;
|
|
980 op0 = mby;
|
|
981 mby = tmp;
|
|
982 }
|
|
983 if (!CONSTANT_P (mby))
|
|
984 return false;
|
|
985 break;
|
|
986
|
|
987 case ASHIFT:
|
|
988 op0 = XEXP (rhs, 0);
|
|
989 mby = XEXP (rhs, 1);
|
|
990 if (!CONSTANT_P (mby))
|
|
991 return false;
|
|
992 break;
|
|
993
|
|
994 default:
|
|
995 return false;
|
|
996 }
|
|
997
|
|
998 if (op0
|
|
999 && !iv_analyze_expr (insn, op0, omode, &iv0))
|
|
1000 return false;
|
|
1001
|
|
1002 if (op1
|
|
1003 && !iv_analyze_expr (insn, op1, omode, &iv1))
|
|
1004 return false;
|
|
1005
|
|
1006 switch (code)
|
|
1007 {
|
|
1008 case SIGN_EXTEND:
|
|
1009 case ZERO_EXTEND:
|
|
1010 if (!iv_extend (&iv0, code, mode))
|
|
1011 return false;
|
|
1012 break;
|
|
1013
|
|
1014 case NEG:
|
|
1015 if (!iv_neg (&iv0))
|
|
1016 return false;
|
|
1017 break;
|
|
1018
|
|
1019 case PLUS:
|
|
1020 case MINUS:
|
|
1021 if (!iv_add (&iv0, &iv1, code))
|
|
1022 return false;
|
|
1023 break;
|
|
1024
|
|
1025 case MULT:
|
|
1026 if (!iv_mult (&iv0, mby))
|
|
1027 return false;
|
|
1028 break;
|
|
1029
|
|
1030 case ASHIFT:
|
|
1031 if (!iv_shift (&iv0, mby))
|
|
1032 return false;
|
|
1033 break;
|
|
1034
|
|
1035 default:
|
|
1036 break;
|
|
1037 }
|
|
1038
|
|
1039 *iv = iv0;
|
|
1040 return iv->base != NULL_RTX;
|
|
1041 }
|
|
1042
|
|
1043 /* Analyzes iv DEF and stores the result to *IV. */
|
|
1044
|
|
1045 static bool
|
|
1046 iv_analyze_def (df_ref def, struct rtx_iv *iv)
|
|
1047 {
|
|
1048 rtx insn = DF_REF_INSN (def);
|
|
1049 rtx reg = DF_REF_REG (def);
|
|
1050 rtx set, rhs;
|
|
1051
|
|
1052 if (dump_file)
|
|
1053 {
|
|
1054 fprintf (dump_file, "Analyzing def of ");
|
|
1055 print_rtl (dump_file, reg);
|
|
1056 fprintf (dump_file, " in insn ");
|
|
1057 print_rtl_single (dump_file, insn);
|
|
1058 }
|
|
1059
|
|
1060 check_iv_ref_table_size ();
|
|
1061 if (DF_REF_IV (def))
|
|
1062 {
|
|
1063 if (dump_file)
|
|
1064 fprintf (dump_file, " already analysed.\n");
|
|
1065 *iv = *DF_REF_IV (def);
|
|
1066 return iv->base != NULL_RTX;
|
|
1067 }
|
|
1068
|
|
1069 iv->mode = VOIDmode;
|
|
1070 iv->base = NULL_RTX;
|
|
1071 iv->step = NULL_RTX;
|
|
1072
|
|
1073 if (!REG_P (reg))
|
|
1074 return false;
|
|
1075
|
|
1076 set = single_set (insn);
|
|
1077 if (!set)
|
|
1078 return false;
|
|
1079
|
|
1080 if (!REG_P (SET_DEST (set)))
|
|
1081 return false;
|
|
1082
|
|
1083 gcc_assert (SET_DEST (set) == reg);
|
|
1084 rhs = find_reg_equal_equiv_note (insn);
|
|
1085 if (rhs)
|
|
1086 rhs = XEXP (rhs, 0);
|
|
1087 else
|
|
1088 rhs = SET_SRC (set);
|
|
1089
|
|
1090 iv_analyze_expr (insn, rhs, GET_MODE (reg), iv);
|
|
1091 record_iv (def, iv);
|
|
1092
|
|
1093 if (dump_file)
|
|
1094 {
|
|
1095 print_rtl (dump_file, reg);
|
|
1096 fprintf (dump_file, " in insn ");
|
|
1097 print_rtl_single (dump_file, insn);
|
|
1098 fprintf (dump_file, " is ");
|
|
1099 dump_iv_info (dump_file, iv);
|
|
1100 fprintf (dump_file, "\n");
|
|
1101 }
|
|
1102
|
|
1103 return iv->base != NULL_RTX;
|
|
1104 }
|
|
1105
|
|
1106 /* Analyzes operand OP of INSN and stores the result to *IV. */
|
|
1107
|
|
1108 static bool
|
|
1109 iv_analyze_op (rtx insn, rtx op, struct rtx_iv *iv)
|
|
1110 {
|
|
1111 df_ref def = NULL;
|
|
1112 enum iv_grd_result res;
|
|
1113
|
|
1114 if (dump_file)
|
|
1115 {
|
|
1116 fprintf (dump_file, "Analyzing operand ");
|
|
1117 print_rtl (dump_file, op);
|
|
1118 fprintf (dump_file, " of insn ");
|
|
1119 print_rtl_single (dump_file, insn);
|
|
1120 }
|
|
1121
|
|
1122 if (CONSTANT_P (op))
|
|
1123 res = GRD_INVARIANT;
|
|
1124 else if (GET_CODE (op) == SUBREG)
|
|
1125 {
|
|
1126 if (!subreg_lowpart_p (op))
|
|
1127 return false;
|
|
1128
|
|
1129 if (!iv_analyze_op (insn, SUBREG_REG (op), iv))
|
|
1130 return false;
|
|
1131
|
|
1132 return iv_subreg (iv, GET_MODE (op));
|
|
1133 }
|
|
1134 else
|
|
1135 {
|
|
1136 res = iv_get_reaching_def (insn, op, &def);
|
|
1137 if (res == GRD_INVALID)
|
|
1138 {
|
|
1139 if (dump_file)
|
|
1140 fprintf (dump_file, " not simple.\n");
|
|
1141 return false;
|
|
1142 }
|
|
1143 }
|
|
1144
|
|
1145 if (res == GRD_INVARIANT)
|
|
1146 {
|
|
1147 iv_constant (iv, op, VOIDmode);
|
|
1148
|
|
1149 if (dump_file)
|
|
1150 {
|
|
1151 fprintf (dump_file, " ");
|
|
1152 dump_iv_info (dump_file, iv);
|
|
1153 fprintf (dump_file, "\n");
|
|
1154 }
|
|
1155 return true;
|
|
1156 }
|
|
1157
|
|
1158 if (res == GRD_MAYBE_BIV)
|
|
1159 return iv_analyze_biv (op, iv);
|
|
1160
|
|
1161 return iv_analyze_def (def, iv);
|
|
1162 }
|
|
1163
|
|
1164 /* Analyzes value VAL at INSN and stores the result to *IV. */
|
|
1165
|
|
1166 bool
|
|
1167 iv_analyze (rtx insn, rtx val, struct rtx_iv *iv)
|
|
1168 {
|
|
1169 rtx reg;
|
|
1170
|
|
1171 /* We must find the insn in that val is used, so that we get to UD chains.
|
|
1172 Since the function is sometimes called on result of get_condition,
|
|
1173 this does not necessarily have to be directly INSN; scan also the
|
|
1174 following insns. */
|
|
1175 if (simple_reg_p (val))
|
|
1176 {
|
|
1177 if (GET_CODE (val) == SUBREG)
|
|
1178 reg = SUBREG_REG (val);
|
|
1179 else
|
|
1180 reg = val;
|
|
1181
|
|
1182 while (!df_find_use (insn, reg))
|
|
1183 insn = NEXT_INSN (insn);
|
|
1184 }
|
|
1185
|
|
1186 return iv_analyze_op (insn, val, iv);
|
|
1187 }
|
|
1188
|
|
1189 /* Analyzes definition of DEF in INSN and stores the result to IV. */
|
|
1190
|
|
1191 bool
|
|
1192 iv_analyze_result (rtx insn, rtx def, struct rtx_iv *iv)
|
|
1193 {
|
|
1194 df_ref adef;
|
|
1195
|
|
1196 adef = df_find_def (insn, def);
|
|
1197 if (!adef)
|
|
1198 return false;
|
|
1199
|
|
1200 return iv_analyze_def (adef, iv);
|
|
1201 }
|
|
1202
|
|
1203 /* Checks whether definition of register REG in INSN is a basic induction
|
|
1204 variable. IV analysis must have been initialized (via a call to
|
|
1205 iv_analysis_loop_init) for this function to produce a result. */
|
|
1206
|
|
1207 bool
|
|
1208 biv_p (rtx insn, rtx reg)
|
|
1209 {
|
|
1210 struct rtx_iv iv;
|
|
1211 df_ref def, last_def;
|
|
1212
|
|
1213 if (!simple_reg_p (reg))
|
|
1214 return false;
|
|
1215
|
|
1216 def = df_find_def (insn, reg);
|
|
1217 gcc_assert (def != NULL);
|
|
1218 if (!latch_dominating_def (reg, &last_def))
|
|
1219 return false;
|
|
1220 if (last_def != def)
|
|
1221 return false;
|
|
1222
|
|
1223 if (!iv_analyze_biv (reg, &iv))
|
|
1224 return false;
|
|
1225
|
|
1226 return iv.step != const0_rtx;
|
|
1227 }
|
|
1228
|
|
1229 /* Calculates value of IV at ITERATION-th iteration. */
|
|
1230
|
|
1231 rtx
|
|
1232 get_iv_value (struct rtx_iv *iv, rtx iteration)
|
|
1233 {
|
|
1234 rtx val;
|
|
1235
|
|
1236 /* We would need to generate some if_then_else patterns, and so far
|
|
1237 it is not needed anywhere. */
|
|
1238 gcc_assert (!iv->first_special);
|
|
1239
|
|
1240 if (iv->step != const0_rtx && iteration != const0_rtx)
|
|
1241 val = simplify_gen_binary (PLUS, iv->extend_mode, iv->base,
|
|
1242 simplify_gen_binary (MULT, iv->extend_mode,
|
|
1243 iv->step, iteration));
|
|
1244 else
|
|
1245 val = iv->base;
|
|
1246
|
|
1247 if (iv->extend_mode == iv->mode)
|
|
1248 return val;
|
|
1249
|
|
1250 val = lowpart_subreg (iv->mode, val, iv->extend_mode);
|
|
1251
|
|
1252 if (iv->extend == UNKNOWN)
|
|
1253 return val;
|
|
1254
|
|
1255 val = simplify_gen_unary (iv->extend, iv->extend_mode, val, iv->mode);
|
|
1256 val = simplify_gen_binary (PLUS, iv->extend_mode, iv->delta,
|
|
1257 simplify_gen_binary (MULT, iv->extend_mode,
|
|
1258 iv->mult, val));
|
|
1259
|
|
1260 return val;
|
|
1261 }
|
|
1262
|
|
1263 /* Free the data for an induction variable analysis. */
|
|
1264
|
|
1265 void
|
|
1266 iv_analysis_done (void)
|
|
1267 {
|
|
1268 if (!clean_slate)
|
|
1269 {
|
|
1270 clear_iv_info ();
|
|
1271 clean_slate = true;
|
|
1272 df_finish_pass (true);
|
|
1273 htab_delete (bivs);
|
|
1274 free (iv_ref_table);
|
|
1275 iv_ref_table = NULL;
|
|
1276 iv_ref_table_size = 0;
|
|
1277 bivs = NULL;
|
|
1278 }
|
|
1279 }
|
|
1280
|
|
1281 /* Computes inverse to X modulo (1 << MOD). */
|
|
1282
|
|
1283 static unsigned HOST_WIDEST_INT
|
|
1284 inverse (unsigned HOST_WIDEST_INT x, int mod)
|
|
1285 {
|
|
1286 unsigned HOST_WIDEST_INT mask =
|
|
1287 ((unsigned HOST_WIDEST_INT) 1 << (mod - 1) << 1) - 1;
|
|
1288 unsigned HOST_WIDEST_INT rslt = 1;
|
|
1289 int i;
|
|
1290
|
|
1291 for (i = 0; i < mod - 1; i++)
|
|
1292 {
|
|
1293 rslt = (rslt * x) & mask;
|
|
1294 x = (x * x) & mask;
|
|
1295 }
|
|
1296
|
|
1297 return rslt;
|
|
1298 }
|
|
1299
|
|
1300 /* Checks whether register *REG is in set ALT. Callback for for_each_rtx. */
|
|
1301
|
|
1302 static int
|
|
1303 altered_reg_used (rtx *reg, void *alt)
|
|
1304 {
|
|
1305 if (!REG_P (*reg))
|
|
1306 return 0;
|
|
1307
|
|
1308 return REGNO_REG_SET_P ((bitmap) alt, REGNO (*reg));
|
|
1309 }
|
|
1310
|
|
1311 /* Marks registers altered by EXPR in set ALT. */
|
|
1312
|
|
1313 static void
|
|
1314 mark_altered (rtx expr, const_rtx by ATTRIBUTE_UNUSED, void *alt)
|
|
1315 {
|
|
1316 if (GET_CODE (expr) == SUBREG)
|
|
1317 expr = SUBREG_REG (expr);
|
|
1318 if (!REG_P (expr))
|
|
1319 return;
|
|
1320
|
|
1321 SET_REGNO_REG_SET ((bitmap) alt, REGNO (expr));
|
|
1322 }
|
|
1323
|
|
1324 /* Checks whether RHS is simple enough to process. */
|
|
1325
|
|
1326 static bool
|
|
1327 simple_rhs_p (rtx rhs)
|
|
1328 {
|
|
1329 rtx op0, op1;
|
|
1330
|
|
1331 if (CONSTANT_P (rhs)
|
|
1332 || (REG_P (rhs) && !HARD_REGISTER_P (rhs)))
|
|
1333 return true;
|
|
1334
|
|
1335 switch (GET_CODE (rhs))
|
|
1336 {
|
|
1337 case PLUS:
|
|
1338 case MINUS:
|
|
1339 op0 = XEXP (rhs, 0);
|
|
1340 op1 = XEXP (rhs, 1);
|
|
1341 /* Allow reg + const and reg + reg. */
|
|
1342 if (!(REG_P (op0) && !HARD_REGISTER_P (op0))
|
|
1343 && !CONSTANT_P (op0))
|
|
1344 return false;
|
|
1345 if (!(REG_P (op1) && !HARD_REGISTER_P (op1))
|
|
1346 && !CONSTANT_P (op1))
|
|
1347 return false;
|
|
1348
|
|
1349 return true;
|
|
1350
|
|
1351 case ASHIFT:
|
|
1352 op0 = XEXP (rhs, 0);
|
|
1353 op1 = XEXP (rhs, 1);
|
|
1354 /* Allow reg << const. */
|
|
1355 if (!(REG_P (op0) && !HARD_REGISTER_P (op0)))
|
|
1356 return false;
|
|
1357 if (!CONSTANT_P (op1))
|
|
1358 return false;
|
|
1359
|
|
1360 return true;
|
|
1361
|
|
1362 default:
|
|
1363 return false;
|
|
1364 }
|
|
1365 }
|
|
1366
|
|
1367 /* Simplifies *EXPR using assignment in INSN. ALTERED is the set of registers
|
|
1368 altered so far. */
|
|
1369
|
|
1370 static void
|
|
1371 simplify_using_assignment (rtx insn, rtx *expr, regset altered)
|
|
1372 {
|
|
1373 rtx set = single_set (insn);
|
|
1374 rtx lhs = NULL_RTX, rhs;
|
|
1375 bool ret = false;
|
|
1376
|
|
1377 if (set)
|
|
1378 {
|
|
1379 lhs = SET_DEST (set);
|
|
1380 if (!REG_P (lhs)
|
|
1381 || altered_reg_used (&lhs, altered))
|
|
1382 ret = true;
|
|
1383 }
|
|
1384 else
|
|
1385 ret = true;
|
|
1386
|
|
1387 note_stores (PATTERN (insn), mark_altered, altered);
|
|
1388 if (CALL_P (insn))
|
|
1389 {
|
|
1390 int i;
|
|
1391
|
|
1392 /* Kill all call clobbered registers. */
|
|
1393 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
|
|
1394 if (TEST_HARD_REG_BIT (regs_invalidated_by_call, i))
|
|
1395 SET_REGNO_REG_SET (altered, i);
|
|
1396 }
|
|
1397
|
|
1398 if (ret)
|
|
1399 return;
|
|
1400
|
|
1401 rhs = find_reg_equal_equiv_note (insn);
|
|
1402 if (rhs)
|
|
1403 rhs = XEXP (rhs, 0);
|
|
1404 else
|
|
1405 rhs = SET_SRC (set);
|
|
1406
|
|
1407 if (!simple_rhs_p (rhs))
|
|
1408 return;
|
|
1409
|
|
1410 if (for_each_rtx (&rhs, altered_reg_used, altered))
|
|
1411 return;
|
|
1412
|
|
1413 *expr = simplify_replace_rtx (*expr, lhs, rhs);
|
|
1414 }
|
|
1415
|
|
1416 /* Checks whether A implies B. */
|
|
1417
|
|
1418 static bool
|
|
1419 implies_p (rtx a, rtx b)
|
|
1420 {
|
|
1421 rtx op0, op1, opb0, opb1, r;
|
|
1422 enum machine_mode mode;
|
|
1423
|
|
1424 if (GET_CODE (a) == EQ)
|
|
1425 {
|
|
1426 op0 = XEXP (a, 0);
|
|
1427 op1 = XEXP (a, 1);
|
|
1428
|
|
1429 if (REG_P (op0))
|
|
1430 {
|
|
1431 r = simplify_replace_rtx (b, op0, op1);
|
|
1432 if (r == const_true_rtx)
|
|
1433 return true;
|
|
1434 }
|
|
1435
|
|
1436 if (REG_P (op1))
|
|
1437 {
|
|
1438 r = simplify_replace_rtx (b, op1, op0);
|
|
1439 if (r == const_true_rtx)
|
|
1440 return true;
|
|
1441 }
|
|
1442 }
|
|
1443
|
|
1444 if (b == const_true_rtx)
|
|
1445 return true;
|
|
1446
|
|
1447 if ((GET_RTX_CLASS (GET_CODE (a)) != RTX_COMM_COMPARE
|
|
1448 && GET_RTX_CLASS (GET_CODE (a)) != RTX_COMPARE)
|
|
1449 || (GET_RTX_CLASS (GET_CODE (b)) != RTX_COMM_COMPARE
|
|
1450 && GET_RTX_CLASS (GET_CODE (b)) != RTX_COMPARE))
|
|
1451 return false;
|
|
1452
|
|
1453 op0 = XEXP (a, 0);
|
|
1454 op1 = XEXP (a, 1);
|
|
1455 opb0 = XEXP (b, 0);
|
|
1456 opb1 = XEXP (b, 1);
|
|
1457
|
|
1458 mode = GET_MODE (op0);
|
|
1459 if (mode != GET_MODE (opb0))
|
|
1460 mode = VOIDmode;
|
|
1461 else if (mode == VOIDmode)
|
|
1462 {
|
|
1463 mode = GET_MODE (op1);
|
|
1464 if (mode != GET_MODE (opb1))
|
|
1465 mode = VOIDmode;
|
|
1466 }
|
|
1467
|
|
1468 /* A < B implies A + 1 <= B. */
|
|
1469 if ((GET_CODE (a) == GT || GET_CODE (a) == LT)
|
|
1470 && (GET_CODE (b) == GE || GET_CODE (b) == LE))
|
|
1471 {
|
|
1472
|
|
1473 if (GET_CODE (a) == GT)
|
|
1474 {
|
|
1475 r = op0;
|
|
1476 op0 = op1;
|
|
1477 op1 = r;
|
|
1478 }
|
|
1479
|
|
1480 if (GET_CODE (b) == GE)
|
|
1481 {
|
|
1482 r = opb0;
|
|
1483 opb0 = opb1;
|
|
1484 opb1 = r;
|
|
1485 }
|
|
1486
|
|
1487 if (SCALAR_INT_MODE_P (mode)
|
|
1488 && rtx_equal_p (op1, opb1)
|
|
1489 && simplify_gen_binary (MINUS, mode, opb0, op0) == const1_rtx)
|
|
1490 return true;
|
|
1491 return false;
|
|
1492 }
|
|
1493
|
|
1494 /* A < B or A > B imply A != B. TODO: Likewise
|
|
1495 A + n < B implies A != B + n if neither wraps. */
|
|
1496 if (GET_CODE (b) == NE
|
|
1497 && (GET_CODE (a) == GT || GET_CODE (a) == GTU
|
|
1498 || GET_CODE (a) == LT || GET_CODE (a) == LTU))
|
|
1499 {
|
|
1500 if (rtx_equal_p (op0, opb0)
|
|
1501 && rtx_equal_p (op1, opb1))
|
|
1502 return true;
|
|
1503 }
|
|
1504
|
|
1505 /* For unsigned comparisons, A != 0 implies A > 0 and A >= 1. */
|
|
1506 if (GET_CODE (a) == NE
|
|
1507 && op1 == const0_rtx)
|
|
1508 {
|
|
1509 if ((GET_CODE (b) == GTU
|
|
1510 && opb1 == const0_rtx)
|
|
1511 || (GET_CODE (b) == GEU
|
|
1512 && opb1 == const1_rtx))
|
|
1513 return rtx_equal_p (op0, opb0);
|
|
1514 }
|
|
1515
|
|
1516 /* A != N is equivalent to A - (N + 1) <u -1. */
|
|
1517 if (GET_CODE (a) == NE
|
|
1518 && GET_CODE (op1) == CONST_INT
|
|
1519 && GET_CODE (b) == LTU
|
|
1520 && opb1 == constm1_rtx
|
|
1521 && GET_CODE (opb0) == PLUS
|
|
1522 && GET_CODE (XEXP (opb0, 1)) == CONST_INT
|
|
1523 /* Avoid overflows. */
|
|
1524 && ((unsigned HOST_WIDE_INT) INTVAL (XEXP (opb0, 1))
|
|
1525 != ((unsigned HOST_WIDE_INT)1
|
|
1526 << (HOST_BITS_PER_WIDE_INT - 1)) - 1)
|
|
1527 && INTVAL (XEXP (opb0, 1)) + 1 == -INTVAL (op1))
|
|
1528 return rtx_equal_p (op0, XEXP (opb0, 0));
|
|
1529
|
|
1530 /* Likewise, A != N implies A - N > 0. */
|
|
1531 if (GET_CODE (a) == NE
|
|
1532 && GET_CODE (op1) == CONST_INT)
|
|
1533 {
|
|
1534 if (GET_CODE (b) == GTU
|
|
1535 && GET_CODE (opb0) == PLUS
|
|
1536 && opb1 == const0_rtx
|
|
1537 && GET_CODE (XEXP (opb0, 1)) == CONST_INT
|
|
1538 /* Avoid overflows. */
|
|
1539 && ((unsigned HOST_WIDE_INT) INTVAL (XEXP (opb0, 1))
|
|
1540 != ((unsigned HOST_WIDE_INT) 1 << (HOST_BITS_PER_WIDE_INT - 1)))
|
|
1541 && rtx_equal_p (XEXP (opb0, 0), op0))
|
|
1542 return INTVAL (op1) == -INTVAL (XEXP (opb0, 1));
|
|
1543 if (GET_CODE (b) == GEU
|
|
1544 && GET_CODE (opb0) == PLUS
|
|
1545 && opb1 == const1_rtx
|
|
1546 && GET_CODE (XEXP (opb0, 1)) == CONST_INT
|
|
1547 /* Avoid overflows. */
|
|
1548 && ((unsigned HOST_WIDE_INT) INTVAL (XEXP (opb0, 1))
|
|
1549 != ((unsigned HOST_WIDE_INT) 1 << (HOST_BITS_PER_WIDE_INT - 1)))
|
|
1550 && rtx_equal_p (XEXP (opb0, 0), op0))
|
|
1551 return INTVAL (op1) == -INTVAL (XEXP (opb0, 1));
|
|
1552 }
|
|
1553
|
|
1554 /* A >s X, where X is positive, implies A <u Y, if Y is negative. */
|
|
1555 if ((GET_CODE (a) == GT || GET_CODE (a) == GE)
|
|
1556 && GET_CODE (op1) == CONST_INT
|
|
1557 && ((GET_CODE (a) == GT && op1 == constm1_rtx)
|
|
1558 || INTVAL (op1) >= 0)
|
|
1559 && GET_CODE (b) == LTU
|
|
1560 && GET_CODE (opb1) == CONST_INT
|
|
1561 && rtx_equal_p (op0, opb0))
|
|
1562 return INTVAL (opb1) < 0;
|
|
1563
|
|
1564 return false;
|
|
1565 }
|
|
1566
|
|
1567 /* Canonicalizes COND so that
|
|
1568
|
|
1569 (1) Ensure that operands are ordered according to
|
|
1570 swap_commutative_operands_p.
|
|
1571 (2) (LE x const) will be replaced with (LT x <const+1>) and similarly
|
|
1572 for GE, GEU, and LEU. */
|
|
1573
|
|
1574 rtx
|
|
1575 canon_condition (rtx cond)
|
|
1576 {
|
|
1577 rtx tem;
|
|
1578 rtx op0, op1;
|
|
1579 enum rtx_code code;
|
|
1580 enum machine_mode mode;
|
|
1581
|
|
1582 code = GET_CODE (cond);
|
|
1583 op0 = XEXP (cond, 0);
|
|
1584 op1 = XEXP (cond, 1);
|
|
1585
|
|
1586 if (swap_commutative_operands_p (op0, op1))
|
|
1587 {
|
|
1588 code = swap_condition (code);
|
|
1589 tem = op0;
|
|
1590 op0 = op1;
|
|
1591 op1 = tem;
|
|
1592 }
|
|
1593
|
|
1594 mode = GET_MODE (op0);
|
|
1595 if (mode == VOIDmode)
|
|
1596 mode = GET_MODE (op1);
|
|
1597 gcc_assert (mode != VOIDmode);
|
|
1598
|
|
1599 if (GET_CODE (op1) == CONST_INT
|
|
1600 && GET_MODE_CLASS (mode) != MODE_CC
|
|
1601 && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT)
|
|
1602 {
|
|
1603 HOST_WIDE_INT const_val = INTVAL (op1);
|
|
1604 unsigned HOST_WIDE_INT uconst_val = const_val;
|
|
1605 unsigned HOST_WIDE_INT max_val
|
|
1606 = (unsigned HOST_WIDE_INT) GET_MODE_MASK (mode);
|
|
1607
|
|
1608 switch (code)
|
|
1609 {
|
|
1610 case LE:
|
|
1611 if ((unsigned HOST_WIDE_INT) const_val != max_val >> 1)
|
|
1612 code = LT, op1 = gen_int_mode (const_val + 1, GET_MODE (op0));
|
|
1613 break;
|
|
1614
|
|
1615 /* When cross-compiling, const_val might be sign-extended from
|
|
1616 BITS_PER_WORD to HOST_BITS_PER_WIDE_INT */
|
|
1617 case GE:
|
|
1618 if ((HOST_WIDE_INT) (const_val & max_val)
|
|
1619 != (((HOST_WIDE_INT) 1
|
|
1620 << (GET_MODE_BITSIZE (GET_MODE (op0)) - 1))))
|
|
1621 code = GT, op1 = gen_int_mode (const_val - 1, mode);
|
|
1622 break;
|
|
1623
|
|
1624 case LEU:
|
|
1625 if (uconst_val < max_val)
|
|
1626 code = LTU, op1 = gen_int_mode (uconst_val + 1, mode);
|
|
1627 break;
|
|
1628
|
|
1629 case GEU:
|
|
1630 if (uconst_val != 0)
|
|
1631 code = GTU, op1 = gen_int_mode (uconst_val - 1, mode);
|
|
1632 break;
|
|
1633
|
|
1634 default:
|
|
1635 break;
|
|
1636 }
|
|
1637 }
|
|
1638
|
|
1639 if (op0 != XEXP (cond, 0)
|
|
1640 || op1 != XEXP (cond, 1)
|
|
1641 || code != GET_CODE (cond)
|
|
1642 || GET_MODE (cond) != SImode)
|
|
1643 cond = gen_rtx_fmt_ee (code, SImode, op0, op1);
|
|
1644
|
|
1645 return cond;
|
|
1646 }
|
|
1647
|
|
1648 /* Tries to use the fact that COND holds to simplify EXPR. ALTERED is the
|
|
1649 set of altered regs. */
|
|
1650
|
|
1651 void
|
|
1652 simplify_using_condition (rtx cond, rtx *expr, regset altered)
|
|
1653 {
|
|
1654 rtx rev, reve, exp = *expr;
|
|
1655
|
|
1656 if (!COMPARISON_P (exp))
|
|
1657 return;
|
|
1658
|
|
1659 /* If some register gets altered later, we do not really speak about its
|
|
1660 value at the time of comparison. */
|
|
1661 if (altered
|
|
1662 && for_each_rtx (&cond, altered_reg_used, altered))
|
|
1663 return;
|
|
1664
|
|
1665 rev = reversed_condition (cond);
|
|
1666 reve = reversed_condition (exp);
|
|
1667
|
|
1668 cond = canon_condition (cond);
|
|
1669 exp = canon_condition (exp);
|
|
1670 if (rev)
|
|
1671 rev = canon_condition (rev);
|
|
1672 if (reve)
|
|
1673 reve = canon_condition (reve);
|
|
1674
|
|
1675 if (rtx_equal_p (exp, cond))
|
|
1676 {
|
|
1677 *expr = const_true_rtx;
|
|
1678 return;
|
|
1679 }
|
|
1680
|
|
1681
|
|
1682 if (rev && rtx_equal_p (exp, rev))
|
|
1683 {
|
|
1684 *expr = const0_rtx;
|
|
1685 return;
|
|
1686 }
|
|
1687
|
|
1688 if (implies_p (cond, exp))
|
|
1689 {
|
|
1690 *expr = const_true_rtx;
|
|
1691 return;
|
|
1692 }
|
|
1693
|
|
1694 if (reve && implies_p (cond, reve))
|
|
1695 {
|
|
1696 *expr = const0_rtx;
|
|
1697 return;
|
|
1698 }
|
|
1699
|
|
1700 /* A proof by contradiction. If *EXPR implies (not cond), *EXPR must
|
|
1701 be false. */
|
|
1702 if (rev && implies_p (exp, rev))
|
|
1703 {
|
|
1704 *expr = const0_rtx;
|
|
1705 return;
|
|
1706 }
|
|
1707
|
|
1708 /* Similarly, If (not *EXPR) implies (not cond), *EXPR must be true. */
|
|
1709 if (rev && reve && implies_p (reve, rev))
|
|
1710 {
|
|
1711 *expr = const_true_rtx;
|
|
1712 return;
|
|
1713 }
|
|
1714
|
|
1715 /* We would like to have some other tests here. TODO. */
|
|
1716
|
|
1717 return;
|
|
1718 }
|
|
1719
|
|
1720 /* Use relationship between A and *B to eventually eliminate *B.
|
|
1721 OP is the operation we consider. */
|
|
1722
|
|
1723 static void
|
|
1724 eliminate_implied_condition (enum rtx_code op, rtx a, rtx *b)
|
|
1725 {
|
|
1726 switch (op)
|
|
1727 {
|
|
1728 case AND:
|
|
1729 /* If A implies *B, we may replace *B by true. */
|
|
1730 if (implies_p (a, *b))
|
|
1731 *b = const_true_rtx;
|
|
1732 break;
|
|
1733
|
|
1734 case IOR:
|
|
1735 /* If *B implies A, we may replace *B by false. */
|
|
1736 if (implies_p (*b, a))
|
|
1737 *b = const0_rtx;
|
|
1738 break;
|
|
1739
|
|
1740 default:
|
|
1741 gcc_unreachable ();
|
|
1742 }
|
|
1743 }
|
|
1744
|
|
1745 /* Eliminates the conditions in TAIL that are implied by HEAD. OP is the
|
|
1746 operation we consider. */
|
|
1747
|
|
1748 static void
|
|
1749 eliminate_implied_conditions (enum rtx_code op, rtx *head, rtx tail)
|
|
1750 {
|
|
1751 rtx elt;
|
|
1752
|
|
1753 for (elt = tail; elt; elt = XEXP (elt, 1))
|
|
1754 eliminate_implied_condition (op, *head, &XEXP (elt, 0));
|
|
1755 for (elt = tail; elt; elt = XEXP (elt, 1))
|
|
1756 eliminate_implied_condition (op, XEXP (elt, 0), head);
|
|
1757 }
|
|
1758
|
|
1759 /* Simplifies *EXPR using initial values at the start of the LOOP. If *EXPR
|
|
1760 is a list, its elements are assumed to be combined using OP. */
|
|
1761
|
|
1762 static void
|
|
1763 simplify_using_initial_values (struct loop *loop, enum rtx_code op, rtx *expr)
|
|
1764 {
|
|
1765 rtx head, tail, insn;
|
|
1766 rtx neutral, aggr;
|
|
1767 regset altered;
|
|
1768 edge e;
|
|
1769
|
|
1770 if (!*expr)
|
|
1771 return;
|
|
1772
|
|
1773 if (CONSTANT_P (*expr))
|
|
1774 return;
|
|
1775
|
|
1776 if (GET_CODE (*expr) == EXPR_LIST)
|
|
1777 {
|
|
1778 head = XEXP (*expr, 0);
|
|
1779 tail = XEXP (*expr, 1);
|
|
1780
|
|
1781 eliminate_implied_conditions (op, &head, tail);
|
|
1782
|
|
1783 switch (op)
|
|
1784 {
|
|
1785 case AND:
|
|
1786 neutral = const_true_rtx;
|
|
1787 aggr = const0_rtx;
|
|
1788 break;
|
|
1789
|
|
1790 case IOR:
|
|
1791 neutral = const0_rtx;
|
|
1792 aggr = const_true_rtx;
|
|
1793 break;
|
|
1794
|
|
1795 default:
|
|
1796 gcc_unreachable ();
|
|
1797 }
|
|
1798
|
|
1799 simplify_using_initial_values (loop, UNKNOWN, &head);
|
|
1800 if (head == aggr)
|
|
1801 {
|
|
1802 XEXP (*expr, 0) = aggr;
|
|
1803 XEXP (*expr, 1) = NULL_RTX;
|
|
1804 return;
|
|
1805 }
|
|
1806 else if (head == neutral)
|
|
1807 {
|
|
1808 *expr = tail;
|
|
1809 simplify_using_initial_values (loop, op, expr);
|
|
1810 return;
|
|
1811 }
|
|
1812 simplify_using_initial_values (loop, op, &tail);
|
|
1813
|
|
1814 if (tail && XEXP (tail, 0) == aggr)
|
|
1815 {
|
|
1816 *expr = tail;
|
|
1817 return;
|
|
1818 }
|
|
1819
|
|
1820 XEXP (*expr, 0) = head;
|
|
1821 XEXP (*expr, 1) = tail;
|
|
1822 return;
|
|
1823 }
|
|
1824
|
|
1825 gcc_assert (op == UNKNOWN);
|
|
1826
|
|
1827 e = loop_preheader_edge (loop);
|
|
1828 if (e->src == ENTRY_BLOCK_PTR)
|
|
1829 return;
|
|
1830
|
|
1831 altered = ALLOC_REG_SET (®_obstack);
|
|
1832
|
|
1833 while (1)
|
|
1834 {
|
|
1835 insn = BB_END (e->src);
|
|
1836 if (any_condjump_p (insn))
|
|
1837 {
|
|
1838 rtx cond = get_condition (BB_END (e->src), NULL, false, true);
|
|
1839
|
|
1840 if (cond && (e->flags & EDGE_FALLTHRU))
|
|
1841 cond = reversed_condition (cond);
|
|
1842 if (cond)
|
|
1843 {
|
|
1844 simplify_using_condition (cond, expr, altered);
|
|
1845 if (CONSTANT_P (*expr))
|
|
1846 {
|
|
1847 FREE_REG_SET (altered);
|
|
1848 return;
|
|
1849 }
|
|
1850 }
|
|
1851 }
|
|
1852
|
|
1853 FOR_BB_INSNS_REVERSE (e->src, insn)
|
|
1854 {
|
|
1855 if (!INSN_P (insn))
|
|
1856 continue;
|
|
1857
|
|
1858 simplify_using_assignment (insn, expr, altered);
|
|
1859 if (CONSTANT_P (*expr))
|
|
1860 {
|
|
1861 FREE_REG_SET (altered);
|
|
1862 return;
|
|
1863 }
|
|
1864 if (for_each_rtx (expr, altered_reg_used, altered))
|
|
1865 {
|
|
1866 FREE_REG_SET (altered);
|
|
1867 return;
|
|
1868 }
|
|
1869 }
|
|
1870
|
|
1871 if (!single_pred_p (e->src)
|
|
1872 || single_pred (e->src) == ENTRY_BLOCK_PTR)
|
|
1873 break;
|
|
1874 e = single_pred_edge (e->src);
|
|
1875 }
|
|
1876
|
|
1877 FREE_REG_SET (altered);
|
|
1878 }
|
|
1879
|
|
1880 /* Transforms invariant IV into MODE. Adds assumptions based on the fact
|
|
1881 that IV occurs as left operands of comparison COND and its signedness
|
|
1882 is SIGNED_P to DESC. */
|
|
1883
|
|
1884 static void
|
|
1885 shorten_into_mode (struct rtx_iv *iv, enum machine_mode mode,
|
|
1886 enum rtx_code cond, bool signed_p, struct niter_desc *desc)
|
|
1887 {
|
|
1888 rtx mmin, mmax, cond_over, cond_under;
|
|
1889
|
|
1890 get_mode_bounds (mode, signed_p, iv->extend_mode, &mmin, &mmax);
|
|
1891 cond_under = simplify_gen_relational (LT, SImode, iv->extend_mode,
|
|
1892 iv->base, mmin);
|
|
1893 cond_over = simplify_gen_relational (GT, SImode, iv->extend_mode,
|
|
1894 iv->base, mmax);
|
|
1895
|
|
1896 switch (cond)
|
|
1897 {
|
|
1898 case LE:
|
|
1899 case LT:
|
|
1900 case LEU:
|
|
1901 case LTU:
|
|
1902 if (cond_under != const0_rtx)
|
|
1903 desc->infinite =
|
|
1904 alloc_EXPR_LIST (0, cond_under, desc->infinite);
|
|
1905 if (cond_over != const0_rtx)
|
|
1906 desc->noloop_assumptions =
|
|
1907 alloc_EXPR_LIST (0, cond_over, desc->noloop_assumptions);
|
|
1908 break;
|
|
1909
|
|
1910 case GE:
|
|
1911 case GT:
|
|
1912 case GEU:
|
|
1913 case GTU:
|
|
1914 if (cond_over != const0_rtx)
|
|
1915 desc->infinite =
|
|
1916 alloc_EXPR_LIST (0, cond_over, desc->infinite);
|
|
1917 if (cond_under != const0_rtx)
|
|
1918 desc->noloop_assumptions =
|
|
1919 alloc_EXPR_LIST (0, cond_under, desc->noloop_assumptions);
|
|
1920 break;
|
|
1921
|
|
1922 case NE:
|
|
1923 if (cond_over != const0_rtx)
|
|
1924 desc->infinite =
|
|
1925 alloc_EXPR_LIST (0, cond_over, desc->infinite);
|
|
1926 if (cond_under != const0_rtx)
|
|
1927 desc->infinite =
|
|
1928 alloc_EXPR_LIST (0, cond_under, desc->infinite);
|
|
1929 break;
|
|
1930
|
|
1931 default:
|
|
1932 gcc_unreachable ();
|
|
1933 }
|
|
1934
|
|
1935 iv->mode = mode;
|
|
1936 iv->extend = signed_p ? SIGN_EXTEND : ZERO_EXTEND;
|
|
1937 }
|
|
1938
|
|
1939 /* Transforms IV0 and IV1 compared by COND so that they are both compared as
|
|
1940 subregs of the same mode if possible (sometimes it is necessary to add
|
|
1941 some assumptions to DESC). */
|
|
1942
|
|
1943 static bool
|
|
1944 canonicalize_iv_subregs (struct rtx_iv *iv0, struct rtx_iv *iv1,
|
|
1945 enum rtx_code cond, struct niter_desc *desc)
|
|
1946 {
|
|
1947 enum machine_mode comp_mode;
|
|
1948 bool signed_p;
|
|
1949
|
|
1950 /* If the ivs behave specially in the first iteration, or are
|
|
1951 added/multiplied after extending, we ignore them. */
|
|
1952 if (iv0->first_special || iv0->mult != const1_rtx || iv0->delta != const0_rtx)
|
|
1953 return false;
|
|
1954 if (iv1->first_special || iv1->mult != const1_rtx || iv1->delta != const0_rtx)
|
|
1955 return false;
|
|
1956
|
|
1957 /* If there is some extend, it must match signedness of the comparison. */
|
|
1958 switch (cond)
|
|
1959 {
|
|
1960 case LE:
|
|
1961 case LT:
|
|
1962 if (iv0->extend == ZERO_EXTEND
|
|
1963 || iv1->extend == ZERO_EXTEND)
|
|
1964 return false;
|
|
1965 signed_p = true;
|
|
1966 break;
|
|
1967
|
|
1968 case LEU:
|
|
1969 case LTU:
|
|
1970 if (iv0->extend == SIGN_EXTEND
|
|
1971 || iv1->extend == SIGN_EXTEND)
|
|
1972 return false;
|
|
1973 signed_p = false;
|
|
1974 break;
|
|
1975
|
|
1976 case NE:
|
|
1977 if (iv0->extend != UNKNOWN
|
|
1978 && iv1->extend != UNKNOWN
|
|
1979 && iv0->extend != iv1->extend)
|
|
1980 return false;
|
|
1981
|
|
1982 signed_p = false;
|
|
1983 if (iv0->extend != UNKNOWN)
|
|
1984 signed_p = iv0->extend == SIGN_EXTEND;
|
|
1985 if (iv1->extend != UNKNOWN)
|
|
1986 signed_p = iv1->extend == SIGN_EXTEND;
|
|
1987 break;
|
|
1988
|
|
1989 default:
|
|
1990 gcc_unreachable ();
|
|
1991 }
|
|
1992
|
|
1993 /* Values of both variables should be computed in the same mode. These
|
|
1994 might indeed be different, if we have comparison like
|
|
1995
|
|
1996 (compare (subreg:SI (iv0)) (subreg:SI (iv1)))
|
|
1997
|
|
1998 and iv0 and iv1 are both ivs iterating in SI mode, but calculated
|
|
1999 in different modes. This does not seem impossible to handle, but
|
|
2000 it hardly ever occurs in practice.
|
|
2001
|
|
2002 The only exception is the case when one of operands is invariant.
|
|
2003 For example pentium 3 generates comparisons like
|
|
2004 (lt (subreg:HI (reg:SI)) 100). Here we assign HImode to 100, but we
|
|
2005 definitely do not want this prevent the optimization. */
|
|
2006 comp_mode = iv0->extend_mode;
|
|
2007 if (GET_MODE_BITSIZE (comp_mode) < GET_MODE_BITSIZE (iv1->extend_mode))
|
|
2008 comp_mode = iv1->extend_mode;
|
|
2009
|
|
2010 if (iv0->extend_mode != comp_mode)
|
|
2011 {
|
|
2012 if (iv0->mode != iv0->extend_mode
|
|
2013 || iv0->step != const0_rtx)
|
|
2014 return false;
|
|
2015
|
|
2016 iv0->base = simplify_gen_unary (signed_p ? SIGN_EXTEND : ZERO_EXTEND,
|
|
2017 comp_mode, iv0->base, iv0->mode);
|
|
2018 iv0->extend_mode = comp_mode;
|
|
2019 }
|
|
2020
|
|
2021 if (iv1->extend_mode != comp_mode)
|
|
2022 {
|
|
2023 if (iv1->mode != iv1->extend_mode
|
|
2024 || iv1->step != const0_rtx)
|
|
2025 return false;
|
|
2026
|
|
2027 iv1->base = simplify_gen_unary (signed_p ? SIGN_EXTEND : ZERO_EXTEND,
|
|
2028 comp_mode, iv1->base, iv1->mode);
|
|
2029 iv1->extend_mode = comp_mode;
|
|
2030 }
|
|
2031
|
|
2032 /* Check that both ivs belong to a range of a single mode. If one of the
|
|
2033 operands is an invariant, we may need to shorten it into the common
|
|
2034 mode. */
|
|
2035 if (iv0->mode == iv0->extend_mode
|
|
2036 && iv0->step == const0_rtx
|
|
2037 && iv0->mode != iv1->mode)
|
|
2038 shorten_into_mode (iv0, iv1->mode, cond, signed_p, desc);
|
|
2039
|
|
2040 if (iv1->mode == iv1->extend_mode
|
|
2041 && iv1->step == const0_rtx
|
|
2042 && iv0->mode != iv1->mode)
|
|
2043 shorten_into_mode (iv1, iv0->mode, swap_condition (cond), signed_p, desc);
|
|
2044
|
|
2045 if (iv0->mode != iv1->mode)
|
|
2046 return false;
|
|
2047
|
|
2048 desc->mode = iv0->mode;
|
|
2049 desc->signed_p = signed_p;
|
|
2050
|
|
2051 return true;
|
|
2052 }
|
|
2053
|
|
2054 /* Tries to estimate the maximum number of iterations. */
|
|
2055
|
|
2056 static unsigned HOST_WIDEST_INT
|
|
2057 determine_max_iter (struct loop *loop, struct niter_desc *desc)
|
|
2058 {
|
|
2059 rtx niter = desc->niter_expr;
|
|
2060 rtx mmin, mmax, cmp;
|
|
2061 unsigned HOST_WIDEST_INT nmax, inc;
|
|
2062
|
|
2063 if (GET_CODE (niter) == AND
|
|
2064 && GET_CODE (XEXP (niter, 0)) == CONST_INT)
|
|
2065 {
|
|
2066 nmax = INTVAL (XEXP (niter, 0));
|
|
2067 if (!(nmax & (nmax + 1)))
|
|
2068 {
|
|
2069 desc->niter_max = nmax;
|
|
2070 return nmax;
|
|
2071 }
|
|
2072 }
|
|
2073
|
|
2074 get_mode_bounds (desc->mode, desc->signed_p, desc->mode, &mmin, &mmax);
|
|
2075 nmax = INTVAL (mmax) - INTVAL (mmin);
|
|
2076
|
|
2077 if (GET_CODE (niter) == UDIV)
|
|
2078 {
|
|
2079 if (GET_CODE (XEXP (niter, 1)) != CONST_INT)
|
|
2080 {
|
|
2081 desc->niter_max = nmax;
|
|
2082 return nmax;
|
|
2083 }
|
|
2084 inc = INTVAL (XEXP (niter, 1));
|
|
2085 niter = XEXP (niter, 0);
|
|
2086 }
|
|
2087 else
|
|
2088 inc = 1;
|
|
2089
|
|
2090 /* We could use a binary search here, but for now improving the upper
|
|
2091 bound by just one eliminates one important corner case. */
|
|
2092 cmp = gen_rtx_fmt_ee (desc->signed_p ? LT : LTU, VOIDmode, niter, mmax);
|
|
2093 simplify_using_initial_values (loop, UNKNOWN, &cmp);
|
|
2094 if (cmp == const_true_rtx)
|
|
2095 {
|
|
2096 nmax--;
|
|
2097
|
|
2098 if (dump_file)
|
|
2099 fprintf (dump_file, ";; improved upper bound by one.\n");
|
|
2100 }
|
|
2101 desc->niter_max = nmax / inc;
|
|
2102 return nmax / inc;
|
|
2103 }
|
|
2104
|
|
2105 /* Computes number of iterations of the CONDITION in INSN in LOOP and stores
|
|
2106 the result into DESC. Very similar to determine_number_of_iterations
|
|
2107 (basically its rtl version), complicated by things like subregs. */
|
|
2108
|
|
2109 static void
|
|
2110 iv_number_of_iterations (struct loop *loop, rtx insn, rtx condition,
|
|
2111 struct niter_desc *desc)
|
|
2112 {
|
|
2113 rtx op0, op1, delta, step, bound, may_xform, tmp, tmp0, tmp1;
|
|
2114 struct rtx_iv iv0, iv1, tmp_iv;
|
|
2115 rtx assumption, may_not_xform;
|
|
2116 enum rtx_code cond;
|
|
2117 enum machine_mode mode, comp_mode;
|
|
2118 rtx mmin, mmax, mode_mmin, mode_mmax;
|
|
2119 unsigned HOST_WIDEST_INT s, size, d, inv;
|
|
2120 HOST_WIDEST_INT up, down, inc, step_val;
|
|
2121 int was_sharp = false;
|
|
2122 rtx old_niter;
|
|
2123 bool step_is_pow2;
|
|
2124
|
|
2125 /* The meaning of these assumptions is this:
|
|
2126 if !assumptions
|
|
2127 then the rest of information does not have to be valid
|
|
2128 if noloop_assumptions then the loop does not roll
|
|
2129 if infinite then this exit is never used */
|
|
2130
|
|
2131 desc->assumptions = NULL_RTX;
|
|
2132 desc->noloop_assumptions = NULL_RTX;
|
|
2133 desc->infinite = NULL_RTX;
|
|
2134 desc->simple_p = true;
|
|
2135
|
|
2136 desc->const_iter = false;
|
|
2137 desc->niter_expr = NULL_RTX;
|
|
2138 desc->niter_max = 0;
|
|
2139
|
|
2140 cond = GET_CODE (condition);
|
|
2141 gcc_assert (COMPARISON_P (condition));
|
|
2142
|
|
2143 mode = GET_MODE (XEXP (condition, 0));
|
|
2144 if (mode == VOIDmode)
|
|
2145 mode = GET_MODE (XEXP (condition, 1));
|
|
2146 /* The constant comparisons should be folded. */
|
|
2147 gcc_assert (mode != VOIDmode);
|
|
2148
|
|
2149 /* We only handle integers or pointers. */
|
|
2150 if (GET_MODE_CLASS (mode) != MODE_INT
|
|
2151 && GET_MODE_CLASS (mode) != MODE_PARTIAL_INT)
|
|
2152 goto fail;
|
|
2153
|
|
2154 op0 = XEXP (condition, 0);
|
|
2155 if (!iv_analyze (insn, op0, &iv0))
|
|
2156 goto fail;
|
|
2157 if (iv0.extend_mode == VOIDmode)
|
|
2158 iv0.mode = iv0.extend_mode = mode;
|
|
2159
|
|
2160 op1 = XEXP (condition, 1);
|
|
2161 if (!iv_analyze (insn, op1, &iv1))
|
|
2162 goto fail;
|
|
2163 if (iv1.extend_mode == VOIDmode)
|
|
2164 iv1.mode = iv1.extend_mode = mode;
|
|
2165
|
|
2166 if (GET_MODE_BITSIZE (iv0.extend_mode) > HOST_BITS_PER_WIDE_INT
|
|
2167 || GET_MODE_BITSIZE (iv1.extend_mode) > HOST_BITS_PER_WIDE_INT)
|
|
2168 goto fail;
|
|
2169
|
|
2170 /* Check condition and normalize it. */
|
|
2171
|
|
2172 switch (cond)
|
|
2173 {
|
|
2174 case GE:
|
|
2175 case GT:
|
|
2176 case GEU:
|
|
2177 case GTU:
|
|
2178 tmp_iv = iv0; iv0 = iv1; iv1 = tmp_iv;
|
|
2179 cond = swap_condition (cond);
|
|
2180 break;
|
|
2181 case NE:
|
|
2182 case LE:
|
|
2183 case LEU:
|
|
2184 case LT:
|
|
2185 case LTU:
|
|
2186 break;
|
|
2187 default:
|
|
2188 goto fail;
|
|
2189 }
|
|
2190
|
|
2191 /* Handle extends. This is relatively nontrivial, so we only try in some
|
|
2192 easy cases, when we can canonicalize the ivs (possibly by adding some
|
|
2193 assumptions) to shape subreg (base + i * step). This function also fills
|
|
2194 in desc->mode and desc->signed_p. */
|
|
2195
|
|
2196 if (!canonicalize_iv_subregs (&iv0, &iv1, cond, desc))
|
|
2197 goto fail;
|
|
2198
|
|
2199 comp_mode = iv0.extend_mode;
|
|
2200 mode = iv0.mode;
|
|
2201 size = GET_MODE_BITSIZE (mode);
|
|
2202 get_mode_bounds (mode, (cond == LE || cond == LT), comp_mode, &mmin, &mmax);
|
|
2203 mode_mmin = lowpart_subreg (mode, mmin, comp_mode);
|
|
2204 mode_mmax = lowpart_subreg (mode, mmax, comp_mode);
|
|
2205
|
|
2206 if (GET_CODE (iv0.step) != CONST_INT || GET_CODE (iv1.step) != CONST_INT)
|
|
2207 goto fail;
|
|
2208
|
|
2209 /* We can take care of the case of two induction variables chasing each other
|
|
2210 if the test is NE. I have never seen a loop using it, but still it is
|
|
2211 cool. */
|
|
2212 if (iv0.step != const0_rtx && iv1.step != const0_rtx)
|
|
2213 {
|
|
2214 if (cond != NE)
|
|
2215 goto fail;
|
|
2216
|
|
2217 iv0.step = simplify_gen_binary (MINUS, comp_mode, iv0.step, iv1.step);
|
|
2218 iv1.step = const0_rtx;
|
|
2219 }
|
|
2220
|
|
2221 /* This is either infinite loop or the one that ends immediately, depending
|
|
2222 on initial values. Unswitching should remove this kind of conditions. */
|
|
2223 if (iv0.step == const0_rtx && iv1.step == const0_rtx)
|
|
2224 goto fail;
|
|
2225
|
|
2226 if (cond != NE)
|
|
2227 {
|
|
2228 if (iv0.step == const0_rtx)
|
|
2229 step_val = -INTVAL (iv1.step);
|
|
2230 else
|
|
2231 step_val = INTVAL (iv0.step);
|
|
2232
|
|
2233 /* Ignore loops of while (i-- < 10) type. */
|
|
2234 if (step_val < 0)
|
|
2235 goto fail;
|
|
2236
|
|
2237 step_is_pow2 = !(step_val & (step_val - 1));
|
|
2238 }
|
|
2239 else
|
|
2240 {
|
|
2241 /* We do not care about whether the step is power of two in this
|
|
2242 case. */
|
|
2243 step_is_pow2 = false;
|
|
2244 step_val = 0;
|
|
2245 }
|
|
2246
|
|
2247 /* Some more condition normalization. We must record some assumptions
|
|
2248 due to overflows. */
|
|
2249 switch (cond)
|
|
2250 {
|
|
2251 case LT:
|
|
2252 case LTU:
|
|
2253 /* We want to take care only of non-sharp relationals; this is easy,
|
|
2254 as in cases the overflow would make the transformation unsafe
|
|
2255 the loop does not roll. Seemingly it would make more sense to want
|
|
2256 to take care of sharp relationals instead, as NE is more similar to
|
|
2257 them, but the problem is that here the transformation would be more
|
|
2258 difficult due to possibly infinite loops. */
|
|
2259 if (iv0.step == const0_rtx)
|
|
2260 {
|
|
2261 tmp = lowpart_subreg (mode, iv0.base, comp_mode);
|
|
2262 assumption = simplify_gen_relational (EQ, SImode, mode, tmp,
|
|
2263 mode_mmax);
|
|
2264 if (assumption == const_true_rtx)
|
|
2265 goto zero_iter_simplify;
|
|
2266 iv0.base = simplify_gen_binary (PLUS, comp_mode,
|
|
2267 iv0.base, const1_rtx);
|
|
2268 }
|
|
2269 else
|
|
2270 {
|
|
2271 tmp = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2272 assumption = simplify_gen_relational (EQ, SImode, mode, tmp,
|
|
2273 mode_mmin);
|
|
2274 if (assumption == const_true_rtx)
|
|
2275 goto zero_iter_simplify;
|
|
2276 iv1.base = simplify_gen_binary (PLUS, comp_mode,
|
|
2277 iv1.base, constm1_rtx);
|
|
2278 }
|
|
2279
|
|
2280 if (assumption != const0_rtx)
|
|
2281 desc->noloop_assumptions =
|
|
2282 alloc_EXPR_LIST (0, assumption, desc->noloop_assumptions);
|
|
2283 cond = (cond == LT) ? LE : LEU;
|
|
2284
|
|
2285 /* It will be useful to be able to tell the difference once more in
|
|
2286 LE -> NE reduction. */
|
|
2287 was_sharp = true;
|
|
2288 break;
|
|
2289 default: ;
|
|
2290 }
|
|
2291
|
|
2292 /* Take care of trivially infinite loops. */
|
|
2293 if (cond != NE)
|
|
2294 {
|
|
2295 if (iv0.step == const0_rtx)
|
|
2296 {
|
|
2297 tmp = lowpart_subreg (mode, iv0.base, comp_mode);
|
|
2298 if (rtx_equal_p (tmp, mode_mmin))
|
|
2299 {
|
|
2300 desc->infinite =
|
|
2301 alloc_EXPR_LIST (0, const_true_rtx, NULL_RTX);
|
|
2302 /* Fill in the remaining fields somehow. */
|
|
2303 goto zero_iter_simplify;
|
|
2304 }
|
|
2305 }
|
|
2306 else
|
|
2307 {
|
|
2308 tmp = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2309 if (rtx_equal_p (tmp, mode_mmax))
|
|
2310 {
|
|
2311 desc->infinite =
|
|
2312 alloc_EXPR_LIST (0, const_true_rtx, NULL_RTX);
|
|
2313 /* Fill in the remaining fields somehow. */
|
|
2314 goto zero_iter_simplify;
|
|
2315 }
|
|
2316 }
|
|
2317 }
|
|
2318
|
|
2319 /* If we can we want to take care of NE conditions instead of size
|
|
2320 comparisons, as they are much more friendly (most importantly
|
|
2321 this takes care of special handling of loops with step 1). We can
|
|
2322 do it if we first check that upper bound is greater or equal to
|
|
2323 lower bound, their difference is constant c modulo step and that
|
|
2324 there is not an overflow. */
|
|
2325 if (cond != NE)
|
|
2326 {
|
|
2327 if (iv0.step == const0_rtx)
|
|
2328 step = simplify_gen_unary (NEG, comp_mode, iv1.step, comp_mode);
|
|
2329 else
|
|
2330 step = iv0.step;
|
|
2331 delta = simplify_gen_binary (MINUS, comp_mode, iv1.base, iv0.base);
|
|
2332 delta = lowpart_subreg (mode, delta, comp_mode);
|
|
2333 delta = simplify_gen_binary (UMOD, mode, delta, step);
|
|
2334 may_xform = const0_rtx;
|
|
2335 may_not_xform = const_true_rtx;
|
|
2336
|
|
2337 if (GET_CODE (delta) == CONST_INT)
|
|
2338 {
|
|
2339 if (was_sharp && INTVAL (delta) == INTVAL (step) - 1)
|
|
2340 {
|
|
2341 /* A special case. We have transformed condition of type
|
|
2342 for (i = 0; i < 4; i += 4)
|
|
2343 into
|
|
2344 for (i = 0; i <= 3; i += 4)
|
|
2345 obviously if the test for overflow during that transformation
|
|
2346 passed, we cannot overflow here. Most importantly any
|
|
2347 loop with sharp end condition and step 1 falls into this
|
|
2348 category, so handling this case specially is definitely
|
|
2349 worth the troubles. */
|
|
2350 may_xform = const_true_rtx;
|
|
2351 }
|
|
2352 else if (iv0.step == const0_rtx)
|
|
2353 {
|
|
2354 bound = simplify_gen_binary (PLUS, comp_mode, mmin, step);
|
|
2355 bound = simplify_gen_binary (MINUS, comp_mode, bound, delta);
|
|
2356 bound = lowpart_subreg (mode, bound, comp_mode);
|
|
2357 tmp = lowpart_subreg (mode, iv0.base, comp_mode);
|
|
2358 may_xform = simplify_gen_relational (cond, SImode, mode,
|
|
2359 bound, tmp);
|
|
2360 may_not_xform = simplify_gen_relational (reverse_condition (cond),
|
|
2361 SImode, mode,
|
|
2362 bound, tmp);
|
|
2363 }
|
|
2364 else
|
|
2365 {
|
|
2366 bound = simplify_gen_binary (MINUS, comp_mode, mmax, step);
|
|
2367 bound = simplify_gen_binary (PLUS, comp_mode, bound, delta);
|
|
2368 bound = lowpart_subreg (mode, bound, comp_mode);
|
|
2369 tmp = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2370 may_xform = simplify_gen_relational (cond, SImode, mode,
|
|
2371 tmp, bound);
|
|
2372 may_not_xform = simplify_gen_relational (reverse_condition (cond),
|
|
2373 SImode, mode,
|
|
2374 tmp, bound);
|
|
2375 }
|
|
2376 }
|
|
2377
|
|
2378 if (may_xform != const0_rtx)
|
|
2379 {
|
|
2380 /* We perform the transformation always provided that it is not
|
|
2381 completely senseless. This is OK, as we would need this assumption
|
|
2382 to determine the number of iterations anyway. */
|
|
2383 if (may_xform != const_true_rtx)
|
|
2384 {
|
|
2385 /* If the step is a power of two and the final value we have
|
|
2386 computed overflows, the cycle is infinite. Otherwise it
|
|
2387 is nontrivial to compute the number of iterations. */
|
|
2388 if (step_is_pow2)
|
|
2389 desc->infinite = alloc_EXPR_LIST (0, may_not_xform,
|
|
2390 desc->infinite);
|
|
2391 else
|
|
2392 desc->assumptions = alloc_EXPR_LIST (0, may_xform,
|
|
2393 desc->assumptions);
|
|
2394 }
|
|
2395
|
|
2396 /* We are going to lose some information about upper bound on
|
|
2397 number of iterations in this step, so record the information
|
|
2398 here. */
|
|
2399 inc = INTVAL (iv0.step) - INTVAL (iv1.step);
|
|
2400 if (GET_CODE (iv1.base) == CONST_INT)
|
|
2401 up = INTVAL (iv1.base);
|
|
2402 else
|
|
2403 up = INTVAL (mode_mmax) - inc;
|
|
2404 down = INTVAL (GET_CODE (iv0.base) == CONST_INT
|
|
2405 ? iv0.base
|
|
2406 : mode_mmin);
|
|
2407 desc->niter_max = (up - down) / inc + 1;
|
|
2408
|
|
2409 if (iv0.step == const0_rtx)
|
|
2410 {
|
|
2411 iv0.base = simplify_gen_binary (PLUS, comp_mode, iv0.base, delta);
|
|
2412 iv0.base = simplify_gen_binary (MINUS, comp_mode, iv0.base, step);
|
|
2413 }
|
|
2414 else
|
|
2415 {
|
|
2416 iv1.base = simplify_gen_binary (MINUS, comp_mode, iv1.base, delta);
|
|
2417 iv1.base = simplify_gen_binary (PLUS, comp_mode, iv1.base, step);
|
|
2418 }
|
|
2419
|
|
2420 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode);
|
|
2421 tmp1 = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2422 assumption = simplify_gen_relational (reverse_condition (cond),
|
|
2423 SImode, mode, tmp0, tmp1);
|
|
2424 if (assumption == const_true_rtx)
|
|
2425 goto zero_iter_simplify;
|
|
2426 else if (assumption != const0_rtx)
|
|
2427 desc->noloop_assumptions =
|
|
2428 alloc_EXPR_LIST (0, assumption, desc->noloop_assumptions);
|
|
2429 cond = NE;
|
|
2430 }
|
|
2431 }
|
|
2432
|
|
2433 /* Count the number of iterations. */
|
|
2434 if (cond == NE)
|
|
2435 {
|
|
2436 /* Everything we do here is just arithmetics modulo size of mode. This
|
|
2437 makes us able to do more involved computations of number of iterations
|
|
2438 than in other cases. First transform the condition into shape
|
|
2439 s * i <> c, with s positive. */
|
|
2440 iv1.base = simplify_gen_binary (MINUS, comp_mode, iv1.base, iv0.base);
|
|
2441 iv0.base = const0_rtx;
|
|
2442 iv0.step = simplify_gen_binary (MINUS, comp_mode, iv0.step, iv1.step);
|
|
2443 iv1.step = const0_rtx;
|
|
2444 if (INTVAL (iv0.step) < 0)
|
|
2445 {
|
|
2446 iv0.step = simplify_gen_unary (NEG, comp_mode, iv0.step, mode);
|
|
2447 iv1.base = simplify_gen_unary (NEG, comp_mode, iv1.base, mode);
|
|
2448 }
|
|
2449 iv0.step = lowpart_subreg (mode, iv0.step, comp_mode);
|
|
2450
|
|
2451 /* Let nsd (s, size of mode) = d. If d does not divide c, the loop
|
|
2452 is infinite. Otherwise, the number of iterations is
|
|
2453 (inverse(s/d) * (c/d)) mod (size of mode/d). */
|
|
2454 s = INTVAL (iv0.step); d = 1;
|
|
2455 while (s % 2 != 1)
|
|
2456 {
|
|
2457 s /= 2;
|
|
2458 d *= 2;
|
|
2459 size--;
|
|
2460 }
|
|
2461 bound = GEN_INT (((unsigned HOST_WIDEST_INT) 1 << (size - 1 ) << 1) - 1);
|
|
2462
|
|
2463 tmp1 = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2464 tmp = simplify_gen_binary (UMOD, mode, tmp1, GEN_INT (d));
|
|
2465 assumption = simplify_gen_relational (NE, SImode, mode, tmp, const0_rtx);
|
|
2466 desc->infinite = alloc_EXPR_LIST (0, assumption, desc->infinite);
|
|
2467
|
|
2468 tmp = simplify_gen_binary (UDIV, mode, tmp1, GEN_INT (d));
|
|
2469 inv = inverse (s, size);
|
|
2470 tmp = simplify_gen_binary (MULT, mode, tmp, gen_int_mode (inv, mode));
|
|
2471 desc->niter_expr = simplify_gen_binary (AND, mode, tmp, bound);
|
|
2472 }
|
|
2473 else
|
|
2474 {
|
|
2475 if (iv1.step == const0_rtx)
|
|
2476 /* Condition in shape a + s * i <= b
|
|
2477 We must know that b + s does not overflow and a <= b + s and then we
|
|
2478 can compute number of iterations as (b + s - a) / s. (It might
|
|
2479 seem that we in fact could be more clever about testing the b + s
|
|
2480 overflow condition using some information about b - a mod s,
|
|
2481 but it was already taken into account during LE -> NE transform). */
|
|
2482 {
|
|
2483 step = iv0.step;
|
|
2484 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode);
|
|
2485 tmp1 = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2486
|
|
2487 bound = simplify_gen_binary (MINUS, mode, mode_mmax,
|
|
2488 lowpart_subreg (mode, step,
|
|
2489 comp_mode));
|
|
2490 if (step_is_pow2)
|
|
2491 {
|
|
2492 rtx t0, t1;
|
|
2493
|
|
2494 /* If s is power of 2, we know that the loop is infinite if
|
|
2495 a % s <= b % s and b + s overflows. */
|
|
2496 assumption = simplify_gen_relational (reverse_condition (cond),
|
|
2497 SImode, mode,
|
|
2498 tmp1, bound);
|
|
2499
|
|
2500 t0 = simplify_gen_binary (UMOD, mode, copy_rtx (tmp0), step);
|
|
2501 t1 = simplify_gen_binary (UMOD, mode, copy_rtx (tmp1), step);
|
|
2502 tmp = simplify_gen_relational (cond, SImode, mode, t0, t1);
|
|
2503 assumption = simplify_gen_binary (AND, SImode, assumption, tmp);
|
|
2504 desc->infinite =
|
|
2505 alloc_EXPR_LIST (0, assumption, desc->infinite);
|
|
2506 }
|
|
2507 else
|
|
2508 {
|
|
2509 assumption = simplify_gen_relational (cond, SImode, mode,
|
|
2510 tmp1, bound);
|
|
2511 desc->assumptions =
|
|
2512 alloc_EXPR_LIST (0, assumption, desc->assumptions);
|
|
2513 }
|
|
2514
|
|
2515 tmp = simplify_gen_binary (PLUS, comp_mode, iv1.base, iv0.step);
|
|
2516 tmp = lowpart_subreg (mode, tmp, comp_mode);
|
|
2517 assumption = simplify_gen_relational (reverse_condition (cond),
|
|
2518 SImode, mode, tmp0, tmp);
|
|
2519
|
|
2520 delta = simplify_gen_binary (PLUS, mode, tmp1, step);
|
|
2521 delta = simplify_gen_binary (MINUS, mode, delta, tmp0);
|
|
2522 }
|
|
2523 else
|
|
2524 {
|
|
2525 /* Condition in shape a <= b - s * i
|
|
2526 We must know that a - s does not overflow and a - s <= b and then
|
|
2527 we can again compute number of iterations as (b - (a - s)) / s. */
|
|
2528 step = simplify_gen_unary (NEG, mode, iv1.step, mode);
|
|
2529 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode);
|
|
2530 tmp1 = lowpart_subreg (mode, iv1.base, comp_mode);
|
|
2531
|
|
2532 bound = simplify_gen_binary (PLUS, mode, mode_mmin,
|
|
2533 lowpart_subreg (mode, step, comp_mode));
|
|
2534 if (step_is_pow2)
|
|
2535 {
|
|
2536 rtx t0, t1;
|
|
2537
|
|
2538 /* If s is power of 2, we know that the loop is infinite if
|
|
2539 a % s <= b % s and a - s overflows. */
|
|
2540 assumption = simplify_gen_relational (reverse_condition (cond),
|
|
2541 SImode, mode,
|
|
2542 bound, tmp0);
|
|
2543
|
|
2544 t0 = simplify_gen_binary (UMOD, mode, copy_rtx (tmp0), step);
|
|
2545 t1 = simplify_gen_binary (UMOD, mode, copy_rtx (tmp1), step);
|
|
2546 tmp = simplify_gen_relational (cond, SImode, mode, t0, t1);
|
|
2547 assumption = simplify_gen_binary (AND, SImode, assumption, tmp);
|
|
2548 desc->infinite =
|
|
2549 alloc_EXPR_LIST (0, assumption, desc->infinite);
|
|
2550 }
|
|
2551 else
|
|
2552 {
|
|
2553 assumption = simplify_gen_relational (cond, SImode, mode,
|
|
2554 bound, tmp0);
|
|
2555 desc->assumptions =
|
|
2556 alloc_EXPR_LIST (0, assumption, desc->assumptions);
|
|
2557 }
|
|
2558
|
|
2559 tmp = simplify_gen_binary (PLUS, comp_mode, iv0.base, iv1.step);
|
|
2560 tmp = lowpart_subreg (mode, tmp, comp_mode);
|
|
2561 assumption = simplify_gen_relational (reverse_condition (cond),
|
|
2562 SImode, mode,
|
|
2563 tmp, tmp1);
|
|
2564 delta = simplify_gen_binary (MINUS, mode, tmp0, step);
|
|
2565 delta = simplify_gen_binary (MINUS, mode, tmp1, delta);
|
|
2566 }
|
|
2567 if (assumption == const_true_rtx)
|
|
2568 goto zero_iter_simplify;
|
|
2569 else if (assumption != const0_rtx)
|
|
2570 desc->noloop_assumptions =
|
|
2571 alloc_EXPR_LIST (0, assumption, desc->noloop_assumptions);
|
|
2572 delta = simplify_gen_binary (UDIV, mode, delta, step);
|
|
2573 desc->niter_expr = delta;
|
|
2574 }
|
|
2575
|
|
2576 old_niter = desc->niter_expr;
|
|
2577
|
|
2578 simplify_using_initial_values (loop, AND, &desc->assumptions);
|
|
2579 if (desc->assumptions
|
|
2580 && XEXP (desc->assumptions, 0) == const0_rtx)
|
|
2581 goto fail;
|
|
2582 simplify_using_initial_values (loop, IOR, &desc->noloop_assumptions);
|
|
2583 simplify_using_initial_values (loop, IOR, &desc->infinite);
|
|
2584 simplify_using_initial_values (loop, UNKNOWN, &desc->niter_expr);
|
|
2585
|
|
2586 /* Rerun the simplification. Consider code (created by copying loop headers)
|
|
2587
|
|
2588 i = 0;
|
|
2589
|
|
2590 if (0 < n)
|
|
2591 {
|
|
2592 do
|
|
2593 {
|
|
2594 i++;
|
|
2595 } while (i < n);
|
|
2596 }
|
|
2597
|
|
2598 The first pass determines that i = 0, the second pass uses it to eliminate
|
|
2599 noloop assumption. */
|
|
2600
|
|
2601 simplify_using_initial_values (loop, AND, &desc->assumptions);
|
|
2602 if (desc->assumptions
|
|
2603 && XEXP (desc->assumptions, 0) == const0_rtx)
|
|
2604 goto fail;
|
|
2605 simplify_using_initial_values (loop, IOR, &desc->noloop_assumptions);
|
|
2606 simplify_using_initial_values (loop, IOR, &desc->infinite);
|
|
2607 simplify_using_initial_values (loop, UNKNOWN, &desc->niter_expr);
|
|
2608
|
|
2609 if (desc->noloop_assumptions
|
|
2610 && XEXP (desc->noloop_assumptions, 0) == const_true_rtx)
|
|
2611 goto zero_iter;
|
|
2612
|
|
2613 if (GET_CODE (desc->niter_expr) == CONST_INT)
|
|
2614 {
|
|
2615 unsigned HOST_WIDEST_INT val = INTVAL (desc->niter_expr);
|
|
2616
|
|
2617 desc->const_iter = true;
|
|
2618 desc->niter_max = desc->niter = val & GET_MODE_MASK (desc->mode);
|
|
2619 }
|
|
2620 else
|
|
2621 {
|
|
2622 if (!desc->niter_max)
|
|
2623 desc->niter_max = determine_max_iter (loop, desc);
|
|
2624
|
|
2625 /* simplify_using_initial_values does a copy propagation on the registers
|
|
2626 in the expression for the number of iterations. This prolongs life
|
|
2627 ranges of registers and increases register pressure, and usually
|
|
2628 brings no gain (and if it happens to do, the cse pass will take care
|
|
2629 of it anyway). So prevent this behavior, unless it enabled us to
|
|
2630 derive that the number of iterations is a constant. */
|
|
2631 desc->niter_expr = old_niter;
|
|
2632 }
|
|
2633
|
|
2634 return;
|
|
2635
|
|
2636 zero_iter_simplify:
|
|
2637 /* Simplify the assumptions. */
|
|
2638 simplify_using_initial_values (loop, AND, &desc->assumptions);
|
|
2639 if (desc->assumptions
|
|
2640 && XEXP (desc->assumptions, 0) == const0_rtx)
|
|
2641 goto fail;
|
|
2642 simplify_using_initial_values (loop, IOR, &desc->infinite);
|
|
2643
|
|
2644 /* Fallthru. */
|
|
2645 zero_iter:
|
|
2646 desc->const_iter = true;
|
|
2647 desc->niter = 0;
|
|
2648 desc->niter_max = 0;
|
|
2649 desc->noloop_assumptions = NULL_RTX;
|
|
2650 desc->niter_expr = const0_rtx;
|
|
2651 return;
|
|
2652
|
|
2653 fail:
|
|
2654 desc->simple_p = false;
|
|
2655 return;
|
|
2656 }
|
|
2657
|
|
2658 /* Checks whether E is a simple exit from LOOP and stores its description
|
|
2659 into DESC. */
|
|
2660
|
|
2661 static void
|
|
2662 check_simple_exit (struct loop *loop, edge e, struct niter_desc *desc)
|
|
2663 {
|
|
2664 basic_block exit_bb;
|
|
2665 rtx condition, at;
|
|
2666 edge ein;
|
|
2667
|
|
2668 exit_bb = e->src;
|
|
2669 desc->simple_p = false;
|
|
2670
|
|
2671 /* It must belong directly to the loop. */
|
|
2672 if (exit_bb->loop_father != loop)
|
|
2673 return;
|
|
2674
|
|
2675 /* It must be tested (at least) once during any iteration. */
|
|
2676 if (!dominated_by_p (CDI_DOMINATORS, loop->latch, exit_bb))
|
|
2677 return;
|
|
2678
|
|
2679 /* It must end in a simple conditional jump. */
|
|
2680 if (!any_condjump_p (BB_END (exit_bb)))
|
|
2681 return;
|
|
2682
|
|
2683 ein = EDGE_SUCC (exit_bb, 0);
|
|
2684 if (ein == e)
|
|
2685 ein = EDGE_SUCC (exit_bb, 1);
|
|
2686
|
|
2687 desc->out_edge = e;
|
|
2688 desc->in_edge = ein;
|
|
2689
|
|
2690 /* Test whether the condition is suitable. */
|
|
2691 if (!(condition = get_condition (BB_END (ein->src), &at, false, false)))
|
|
2692 return;
|
|
2693
|
|
2694 if (ein->flags & EDGE_FALLTHRU)
|
|
2695 {
|
|
2696 condition = reversed_condition (condition);
|
|
2697 if (!condition)
|
|
2698 return;
|
|
2699 }
|
|
2700
|
|
2701 /* Check that we are able to determine number of iterations and fill
|
|
2702 in information about it. */
|
|
2703 iv_number_of_iterations (loop, at, condition, desc);
|
|
2704 }
|
|
2705
|
|
2706 /* Finds a simple exit of LOOP and stores its description into DESC. */
|
|
2707
|
|
2708 void
|
|
2709 find_simple_exit (struct loop *loop, struct niter_desc *desc)
|
|
2710 {
|
|
2711 unsigned i;
|
|
2712 basic_block *body;
|
|
2713 edge e;
|
|
2714 struct niter_desc act;
|
|
2715 bool any = false;
|
|
2716 edge_iterator ei;
|
|
2717
|
|
2718 desc->simple_p = false;
|
|
2719 body = get_loop_body (loop);
|
|
2720
|
|
2721 for (i = 0; i < loop->num_nodes; i++)
|
|
2722 {
|
|
2723 FOR_EACH_EDGE (e, ei, body[i]->succs)
|
|
2724 {
|
|
2725 if (flow_bb_inside_loop_p (loop, e->dest))
|
|
2726 continue;
|
|
2727
|
|
2728 check_simple_exit (loop, e, &act);
|
|
2729 if (!act.simple_p)
|
|
2730 continue;
|
|
2731
|
|
2732 if (!any)
|
|
2733 any = true;
|
|
2734 else
|
|
2735 {
|
|
2736 /* Prefer constant iterations; the less the better. */
|
|
2737 if (!act.const_iter
|
|
2738 || (desc->const_iter && act.niter >= desc->niter))
|
|
2739 continue;
|
|
2740
|
|
2741 /* Also if the actual exit may be infinite, while the old one
|
|
2742 not, prefer the old one. */
|
|
2743 if (act.infinite && !desc->infinite)
|
|
2744 continue;
|
|
2745 }
|
|
2746
|
|
2747 *desc = act;
|
|
2748 }
|
|
2749 }
|
|
2750
|
|
2751 if (dump_file)
|
|
2752 {
|
|
2753 if (desc->simple_p)
|
|
2754 {
|
|
2755 fprintf (dump_file, "Loop %d is simple:\n", loop->num);
|
|
2756 fprintf (dump_file, " simple exit %d -> %d\n",
|
|
2757 desc->out_edge->src->index,
|
|
2758 desc->out_edge->dest->index);
|
|
2759 if (desc->assumptions)
|
|
2760 {
|
|
2761 fprintf (dump_file, " assumptions: ");
|
|
2762 print_rtl (dump_file, desc->assumptions);
|
|
2763 fprintf (dump_file, "\n");
|
|
2764 }
|
|
2765 if (desc->noloop_assumptions)
|
|
2766 {
|
|
2767 fprintf (dump_file, " does not roll if: ");
|
|
2768 print_rtl (dump_file, desc->noloop_assumptions);
|
|
2769 fprintf (dump_file, "\n");
|
|
2770 }
|
|
2771 if (desc->infinite)
|
|
2772 {
|
|
2773 fprintf (dump_file, " infinite if: ");
|
|
2774 print_rtl (dump_file, desc->infinite);
|
|
2775 fprintf (dump_file, "\n");
|
|
2776 }
|
|
2777
|
|
2778 fprintf (dump_file, " number of iterations: ");
|
|
2779 print_rtl (dump_file, desc->niter_expr);
|
|
2780 fprintf (dump_file, "\n");
|
|
2781
|
|
2782 fprintf (dump_file, " upper bound: ");
|
|
2783 fprintf (dump_file, HOST_WIDEST_INT_PRINT_DEC, desc->niter_max);
|
|
2784 fprintf (dump_file, "\n");
|
|
2785 }
|
|
2786 else
|
|
2787 fprintf (dump_file, "Loop %d is not simple.\n", loop->num);
|
|
2788 }
|
|
2789
|
|
2790 free (body);
|
|
2791 }
|
|
2792
|
|
2793 /* Creates a simple loop description of LOOP if it was not computed
|
|
2794 already. */
|
|
2795
|
|
2796 struct niter_desc *
|
|
2797 get_simple_loop_desc (struct loop *loop)
|
|
2798 {
|
|
2799 struct niter_desc *desc = simple_loop_desc (loop);
|
|
2800
|
|
2801 if (desc)
|
|
2802 return desc;
|
|
2803
|
|
2804 /* At least desc->infinite is not always initialized by
|
|
2805 find_simple_loop_exit. */
|
|
2806 desc = XCNEW (struct niter_desc);
|
|
2807 iv_analysis_loop_init (loop);
|
|
2808 find_simple_exit (loop, desc);
|
|
2809 loop->aux = desc;
|
|
2810
|
|
2811 if (desc->simple_p && (desc->assumptions || desc->infinite))
|
|
2812 {
|
|
2813 const char *wording;
|
|
2814
|
|
2815 /* Assume that no overflow happens and that the loop is finite.
|
|
2816 We already warned at the tree level if we ran optimizations there. */
|
|
2817 if (!flag_tree_loop_optimize && warn_unsafe_loop_optimizations)
|
|
2818 {
|
|
2819 if (desc->infinite)
|
|
2820 {
|
|
2821 wording =
|
|
2822 flag_unsafe_loop_optimizations
|
|
2823 ? N_("assuming that the loop is not infinite")
|
|
2824 : N_("cannot optimize possibly infinite loops");
|
|
2825 warning (OPT_Wunsafe_loop_optimizations, "%s",
|
|
2826 gettext (wording));
|
|
2827 }
|
|
2828 if (desc->assumptions)
|
|
2829 {
|
|
2830 wording =
|
|
2831 flag_unsafe_loop_optimizations
|
|
2832 ? N_("assuming that the loop counter does not overflow")
|
|
2833 : N_("cannot optimize loop, the loop counter may overflow");
|
|
2834 warning (OPT_Wunsafe_loop_optimizations, "%s",
|
|
2835 gettext (wording));
|
|
2836 }
|
|
2837 }
|
|
2838
|
|
2839 if (flag_unsafe_loop_optimizations)
|
|
2840 {
|
|
2841 desc->assumptions = NULL_RTX;
|
|
2842 desc->infinite = NULL_RTX;
|
|
2843 }
|
|
2844 }
|
|
2845
|
|
2846 return desc;
|
|
2847 }
|
|
2848
|
|
2849 /* Releases simple loop description for LOOP. */
|
|
2850
|
|
2851 void
|
|
2852 free_simple_loop_desc (struct loop *loop)
|
|
2853 {
|
|
2854 struct niter_desc *desc = simple_loop_desc (loop);
|
|
2855
|
|
2856 if (!desc)
|
|
2857 return;
|
|
2858
|
|
2859 free (desc);
|
|
2860 loop->aux = NULL;
|
|
2861 }
|