111
|
1 /* Helper routines for memory move and comparison insns.
|
145
|
2 Copyright (C) 2013-2020 Free Software Foundation, Inc.
|
111
|
3
|
|
4 This file is part of GCC.
|
|
5
|
|
6 GCC is free software; you can redistribute it and/or modify
|
|
7 it under the terms of the GNU General Public License as published by
|
|
8 the Free Software Foundation; either version 3, or (at your option)
|
|
9 any later version.
|
|
10
|
|
11 GCC is distributed in the hope that it will be useful,
|
|
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
14 GNU General Public License for more details.
|
|
15
|
|
16 You should have received a copy of the GNU General Public License
|
|
17 along with GCC; see the file COPYING3. If not see
|
|
18 <http://www.gnu.org/licenses/>. */
|
|
19
|
131
|
20 #define IN_TARGET_CODE 1
|
|
21
|
111
|
22 #include "config.h"
|
|
23 #include "system.h"
|
|
24 #include "coretypes.h"
|
|
25 #include "tm.h"
|
|
26 #include "function.h"
|
|
27 #include "basic-block.h"
|
|
28 #include "rtl.h"
|
|
29 #include "tree.h"
|
|
30 #include "memmodel.h"
|
|
31 #include "tm_p.h"
|
|
32 #include "emit-rtl.h"
|
|
33 #include "explow.h"
|
|
34 #include "expr.h"
|
|
35
|
|
36 /* Like force_operand, but guarantees that VALUE ends up in TARGET. */
|
|
37 static void
|
|
38 force_into (rtx value, rtx target)
|
|
39 {
|
|
40 value = force_operand (value, target);
|
|
41 if (! rtx_equal_p (value, target))
|
|
42 emit_insn (gen_move_insn (target, value));
|
|
43 }
|
|
44
|
|
45 /* Emit code to perform a block move. Choose the best method.
|
|
46
|
|
47 OPERANDS[0] is the destination.
|
|
48 OPERANDS[1] is the source.
|
|
49 OPERANDS[2] is the size.
|
|
50 OPERANDS[3] is the alignment safe to use. */
|
|
51 bool
|
|
52 expand_block_move (rtx *operands)
|
|
53 {
|
|
54 int align = INTVAL (operands[3]);
|
|
55 int constp = (CONST_INT_P (operands[2]));
|
|
56 int bytes = (constp ? INTVAL (operands[2]) : 0);
|
|
57
|
|
58 if (! constp)
|
|
59 return false;
|
|
60
|
|
61 /* If we could use mov.l to move words and dest is word-aligned, we
|
|
62 can use movua.l for loads and still generate a relatively short
|
|
63 and efficient sequence. */
|
|
64 if (TARGET_SH4A && align < 4
|
|
65 && MEM_ALIGN (operands[0]) >= 32
|
|
66 && can_move_by_pieces (bytes, 32))
|
|
67 {
|
|
68 rtx dest = copy_rtx (operands[0]);
|
|
69 rtx src = copy_rtx (operands[1]);
|
|
70 /* We could use different pseudos for each copied word, but
|
|
71 since movua can only load into r0, it's kind of
|
|
72 pointless. */
|
|
73 rtx temp = gen_reg_rtx (SImode);
|
|
74 rtx src_addr = copy_addr_to_reg (XEXP (src, 0));
|
|
75 int copied = 0;
|
|
76
|
|
77 while (copied + 4 <= bytes)
|
|
78 {
|
|
79 rtx to = adjust_address (dest, SImode, copied);
|
|
80 rtx from = adjust_automodify_address (src, BLKmode,
|
|
81 src_addr, copied);
|
|
82
|
|
83 set_mem_size (from, 4);
|
|
84 emit_insn (gen_movua (temp, from));
|
|
85 emit_move_insn (src_addr, plus_constant (Pmode, src_addr, 4));
|
|
86 emit_move_insn (to, temp);
|
|
87 copied += 4;
|
|
88 }
|
|
89
|
|
90 if (copied < bytes)
|
|
91 move_by_pieces (adjust_address (dest, BLKmode, copied),
|
|
92 adjust_automodify_address (src, BLKmode,
|
|
93 src_addr, copied),
|
145
|
94 bytes - copied, align, RETURN_BEGIN);
|
111
|
95
|
|
96 return true;
|
|
97 }
|
|
98
|
|
99 /* If it isn't a constant number of bytes, or if it doesn't have 4 byte
|
|
100 alignment, or if it isn't a multiple of 4 bytes, then fail. */
|
|
101 if (align < 4 || (bytes % 4 != 0))
|
|
102 return false;
|
|
103
|
|
104 if (TARGET_HARD_SH4)
|
|
105 {
|
|
106 if (bytes < 12)
|
|
107 return false;
|
|
108 else if (bytes == 12)
|
|
109 {
|
|
110 rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
|
111 rtx r4 = gen_rtx_REG (SImode, 4);
|
|
112 rtx r5 = gen_rtx_REG (SImode, 5);
|
|
113
|
|
114 rtx lab = function_symbol (func_addr_rtx, "__movmemSI12_i4",
|
|
115 SFUNC_STATIC).lab;
|
|
116 force_into (XEXP (operands[0], 0), r4);
|
|
117 force_into (XEXP (operands[1], 0), r5);
|
|
118 emit_insn (gen_block_move_real_i4 (func_addr_rtx, lab));
|
|
119 return true;
|
|
120 }
|
|
121 else if (! optimize_size)
|
|
122 {
|
|
123 rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
|
124 rtx r4 = gen_rtx_REG (SImode, 4);
|
|
125 rtx r5 = gen_rtx_REG (SImode, 5);
|
|
126 rtx r6 = gen_rtx_REG (SImode, 6);
|
|
127
|
|
128 rtx lab = function_symbol (func_addr_rtx, bytes & 4
|
|
129 ? "__movmem_i4_odd"
|
|
130 : "__movmem_i4_even",
|
|
131 SFUNC_STATIC).lab;
|
|
132 force_into (XEXP (operands[0], 0), r4);
|
|
133 force_into (XEXP (operands[1], 0), r5);
|
|
134
|
|
135 int dwords = bytes >> 3;
|
|
136 emit_insn (gen_move_insn (r6, GEN_INT (dwords - 1)));
|
|
137 emit_insn (gen_block_lump_real_i4 (func_addr_rtx, lab));
|
|
138 return true;
|
|
139 }
|
|
140 else
|
|
141 return false;
|
|
142 }
|
|
143 if (bytes < 64)
|
|
144 {
|
|
145 char entry[30];
|
|
146 rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
|
147 rtx r4 = gen_rtx_REG (SImode, 4);
|
|
148 rtx r5 = gen_rtx_REG (SImode, 5);
|
|
149
|
|
150 sprintf (entry, "__movmemSI%d", bytes);
|
|
151 rtx lab = function_symbol (func_addr_rtx, entry, SFUNC_STATIC).lab;
|
|
152 force_into (XEXP (operands[0], 0), r4);
|
|
153 force_into (XEXP (operands[1], 0), r5);
|
|
154 emit_insn (gen_block_move_real (func_addr_rtx, lab));
|
|
155 return true;
|
|
156 }
|
|
157
|
|
158 /* This is the same number of bytes as a memcpy call, but to a different
|
|
159 less common function name, so this will occasionally use more space. */
|
|
160 if (! optimize_size)
|
|
161 {
|
|
162 rtx func_addr_rtx = gen_reg_rtx (Pmode);
|
|
163 int final_switch, while_loop;
|
|
164 rtx r4 = gen_rtx_REG (SImode, 4);
|
|
165 rtx r5 = gen_rtx_REG (SImode, 5);
|
|
166 rtx r6 = gen_rtx_REG (SImode, 6);
|
|
167
|
|
168 rtx lab = function_symbol (func_addr_rtx, "__movmem", SFUNC_STATIC).lab;
|
|
169 force_into (XEXP (operands[0], 0), r4);
|
|
170 force_into (XEXP (operands[1], 0), r5);
|
|
171
|
|
172 /* r6 controls the size of the move. 16 is decremented from it
|
|
173 for each 64 bytes moved. Then the negative bit left over is used
|
|
174 as an index into a list of move instructions. e.g., a 72 byte move
|
|
175 would be set up with size(r6) = 14, for one iteration through the
|
|
176 big while loop, and a switch of -2 for the last part. */
|
|
177
|
|
178 final_switch = 16 - ((bytes / 4) % 16);
|
|
179 while_loop = ((bytes / 4) / 16 - 1) * 16;
|
|
180 emit_insn (gen_move_insn (r6, GEN_INT (while_loop + final_switch)));
|
|
181 emit_insn (gen_block_lump_real (func_addr_rtx, lab));
|
|
182 return true;
|
|
183 }
|
|
184
|
|
185 return false;
|
|
186 }
|
|
187
|
131
|
188 static const int prob_unlikely
|
|
189 = profile_probability::from_reg_br_prob_base (REG_BR_PROB_BASE / 10)
|
|
190 .to_reg_br_prob_note ();
|
|
191 static const int prob_likely
|
|
192 = profile_probability::from_reg_br_prob_base (REG_BR_PROB_BASE / 4)
|
|
193 .to_reg_br_prob_note ();
|
111
|
194
|
|
195 /* Emit code to perform a strcmp.
|
|
196
|
|
197 OPERANDS[0] is the destination.
|
|
198 OPERANDS[1] is the first string.
|
|
199 OPERANDS[2] is the second string.
|
|
200 OPERANDS[3] is the known alignment. */
|
|
201 bool
|
|
202 sh_expand_cmpstr (rtx *operands)
|
|
203 {
|
|
204 rtx addr1 = operands[1];
|
|
205 rtx addr2 = operands[2];
|
|
206 rtx s1_addr = copy_addr_to_reg (XEXP (addr1, 0));
|
|
207 rtx s2_addr = copy_addr_to_reg (XEXP (addr2, 0));
|
|
208 rtx tmp0 = gen_reg_rtx (SImode);
|
|
209 rtx tmp1 = gen_reg_rtx (SImode);
|
|
210 rtx tmp2 = gen_reg_rtx (SImode);
|
|
211 rtx tmp3 = gen_reg_rtx (SImode);
|
|
212
|
|
213 rtx_insn *jump;
|
|
214 rtx_code_label *L_return = gen_label_rtx ();
|
|
215 rtx_code_label *L_loop_byte = gen_label_rtx ();
|
|
216 rtx_code_label *L_end_loop_byte = gen_label_rtx ();
|
|
217 rtx_code_label *L_loop_long = gen_label_rtx ();
|
|
218 rtx_code_label *L_end_loop_long = gen_label_rtx ();
|
|
219
|
|
220 const unsigned int addr1_alignment = MEM_ALIGN (operands[1]) / BITS_PER_UNIT;
|
|
221 const unsigned int addr2_alignment = MEM_ALIGN (operands[2]) / BITS_PER_UNIT;
|
|
222
|
|
223 if (addr1_alignment < 4 && addr2_alignment < 4)
|
|
224 {
|
|
225 emit_insn (gen_iorsi3 (tmp1, s1_addr, s2_addr));
|
|
226 emit_insn (gen_tstsi_t (tmp1, GEN_INT (3)));
|
|
227 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
228 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
229 }
|
|
230 else if (addr1_alignment < 4 && addr2_alignment >= 4)
|
|
231 {
|
|
232 emit_insn (gen_tstsi_t (s1_addr, GEN_INT (3)));
|
|
233 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
234 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
235 }
|
|
236 else if (addr1_alignment >= 4 && addr2_alignment < 4)
|
|
237 {
|
|
238 emit_insn (gen_tstsi_t (s2_addr, GEN_INT (3)));
|
|
239 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
240 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
241 }
|
|
242
|
|
243 addr1 = adjust_automodify_address (addr1, SImode, s1_addr, 0);
|
|
244 addr2 = adjust_automodify_address (addr2, SImode, s2_addr, 0);
|
|
245
|
|
246 /* tmp2 is aligned, OK to load. */
|
|
247 emit_move_insn (tmp3, addr2);
|
|
248 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, 4));
|
|
249
|
|
250 /* start long loop. */
|
|
251 emit_label (L_loop_long);
|
|
252
|
|
253 emit_move_insn (tmp2, tmp3);
|
|
254
|
|
255 /* tmp1 is aligned, OK to load. */
|
|
256 emit_move_insn (tmp1, addr1);
|
|
257 emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, 4));
|
|
258
|
|
259 /* Is there a 0 byte ? */
|
|
260 emit_insn (gen_andsi3 (tmp3, tmp3, tmp1));
|
|
261
|
|
262 emit_insn (gen_cmpstr_t (tmp0, tmp3));
|
|
263 jump = emit_jump_insn (gen_branch_true (L_end_loop_long));
|
|
264 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
265
|
|
266 emit_insn (gen_cmpeqsi_t (tmp1, tmp2));
|
|
267
|
|
268 /* tmp2 is aligned, OK to load. */
|
|
269 emit_move_insn (tmp3, addr2);
|
|
270 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, 4));
|
|
271
|
|
272 jump = emit_jump_insn (gen_branch_true (L_loop_long));
|
|
273 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
274 /* end loop. */
|
|
275
|
|
276 /* Fallthu, substract words. */
|
|
277 if (TARGET_LITTLE_ENDIAN)
|
|
278 {
|
|
279 rtx low_1 = gen_lowpart (HImode, tmp1);
|
|
280 rtx low_2 = gen_lowpart (HImode, tmp2);
|
|
281
|
|
282 emit_insn (gen_rotlhi3_8 (low_1, low_1));
|
|
283 emit_insn (gen_rotlhi3_8 (low_2, low_2));
|
|
284 emit_insn (gen_rotlsi3_16 (tmp1, tmp1));
|
|
285 emit_insn (gen_rotlsi3_16 (tmp2, tmp2));
|
|
286 emit_insn (gen_rotlhi3_8 (low_1, low_1));
|
|
287 emit_insn (gen_rotlhi3_8 (low_2, low_2));
|
|
288 }
|
|
289
|
|
290 jump = emit_jump_insn (gen_jump_compact (L_return));
|
|
291 emit_barrier_after (jump);
|
|
292
|
|
293 emit_label (L_end_loop_long);
|
|
294
|
|
295 emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, -4));
|
|
296 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, -4));
|
|
297
|
|
298 /* start byte loop. */
|
|
299 addr1 = adjust_address (addr1, QImode, 0);
|
|
300 addr2 = adjust_address (addr2, QImode, 0);
|
|
301
|
|
302 emit_label (L_loop_byte);
|
|
303
|
|
304 emit_insn (gen_extendqisi2 (tmp2, addr2));
|
|
305 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, 1));
|
|
306
|
|
307 emit_insn (gen_extendqisi2 (tmp1, addr1));
|
|
308 emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, 1));
|
|
309
|
|
310 emit_insn (gen_cmpeqsi_t (tmp2, const0_rtx));
|
|
311 jump = emit_jump_insn (gen_branch_true (L_end_loop_byte));
|
|
312 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
313
|
|
314 emit_insn (gen_cmpeqsi_t (tmp1, tmp2));
|
|
315 if (flag_delayed_branch)
|
|
316 emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2)));
|
|
317 jump = emit_jump_insn (gen_branch_true (L_loop_byte));
|
|
318 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
319 /* end loop. */
|
|
320
|
|
321 emit_label (L_end_loop_byte);
|
|
322
|
|
323 if (! flag_delayed_branch)
|
|
324 emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2)));
|
|
325 emit_insn (gen_zero_extendqisi2 (tmp1, gen_lowpart (QImode, tmp1)));
|
|
326
|
|
327 emit_label (L_return);
|
|
328
|
|
329 emit_insn (gen_subsi3 (operands[0], tmp1, tmp2));
|
|
330
|
|
331 return true;
|
|
332 }
|
|
333
|
|
334 /* Emit code to perform a strncmp.
|
|
335
|
|
336 OPERANDS[0] is the destination.
|
|
337 OPERANDS[1] is the first string.
|
|
338 OPERANDS[2] is the second string.
|
|
339 OPERANDS[3] is the length.
|
|
340 OPERANDS[4] is the known alignment. */
|
|
341 bool
|
|
342 sh_expand_cmpnstr (rtx *operands)
|
|
343 {
|
|
344 rtx addr1 = operands[1];
|
|
345 rtx addr2 = operands[2];
|
|
346 rtx s1_addr = copy_addr_to_reg (XEXP (addr1, 0));
|
|
347 rtx s2_addr = copy_addr_to_reg (XEXP (addr2, 0));
|
|
348 rtx tmp1 = gen_reg_rtx (SImode);
|
|
349 rtx tmp2 = gen_reg_rtx (SImode);
|
|
350
|
|
351 rtx_insn *jump;
|
|
352 rtx_code_label *L_return = gen_label_rtx ();
|
|
353 rtx_code_label *L_loop_byte = gen_label_rtx ();
|
|
354 rtx_code_label *L_end_loop_byte = gen_label_rtx ();
|
|
355
|
|
356 rtx len = copy_to_mode_reg (SImode, operands[3]);
|
|
357 int constp = CONST_INT_P (operands[3]);
|
|
358 HOST_WIDE_INT bytes = constp ? INTVAL (operands[3]) : 0;
|
|
359
|
|
360 const unsigned int addr1_alignment = MEM_ALIGN (operands[1]) / BITS_PER_UNIT;
|
|
361 const unsigned int addr2_alignment = MEM_ALIGN (operands[2]) / BITS_PER_UNIT;
|
|
362
|
|
363 /* Loop on a register count. */
|
|
364 if (constp && bytes >= 0 && bytes < 32)
|
|
365 {
|
|
366 rtx tmp0 = gen_reg_rtx (SImode);
|
|
367 rtx tmp3 = gen_reg_rtx (SImode);
|
|
368 rtx lenw = gen_reg_rtx (SImode);
|
|
369
|
|
370 rtx_code_label *L_loop_long = gen_label_rtx ();
|
|
371 rtx_code_label *L_end_loop_long = gen_label_rtx ();
|
|
372
|
|
373 int witers = bytes / 4;
|
|
374
|
|
375 if (witers > 1)
|
|
376 {
|
|
377 addr1 = adjust_automodify_address (addr1, SImode, s1_addr, 0);
|
|
378 addr2 = adjust_automodify_address (addr2, SImode, s2_addr, 0);
|
|
379
|
|
380 emit_move_insn (tmp0, const0_rtx);
|
|
381
|
|
382 if (addr1_alignment < 4 && addr2_alignment < 4)
|
|
383 {
|
|
384 emit_insn (gen_iorsi3 (tmp1, s1_addr, s2_addr));
|
|
385 emit_insn (gen_tstsi_t (tmp1, GEN_INT (3)));
|
|
386 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
387 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
388 }
|
|
389 else if (addr1_alignment < 4 && addr2_alignment >= 4)
|
|
390 {
|
|
391 emit_insn (gen_tstsi_t (s1_addr, GEN_INT (3)));
|
|
392 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
393 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
394 }
|
|
395 else if (addr1_alignment >= 4 && addr2_alignment < 4)
|
|
396 {
|
|
397 emit_insn (gen_tstsi_t (s2_addr, GEN_INT (3)));
|
|
398 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
399 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
400 }
|
|
401
|
|
402 /* word count. Do we have iterations ? */
|
|
403 emit_insn (gen_lshrsi3 (lenw, len, GEN_INT (2)));
|
|
404
|
|
405 /* start long loop. */
|
|
406 emit_label (L_loop_long);
|
|
407
|
|
408 /* tmp2 is aligned, OK to load. */
|
|
409 emit_move_insn (tmp2, addr2);
|
|
410 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr,
|
|
411 GET_MODE_SIZE (SImode)));
|
|
412
|
|
413 /* tmp1 is aligned, OK to load. */
|
|
414 emit_move_insn (tmp1, addr1);
|
|
415 emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr,
|
|
416 GET_MODE_SIZE (SImode)));
|
|
417
|
|
418 /* Is there a 0 byte ? */
|
|
419 emit_insn (gen_andsi3 (tmp3, tmp2, tmp1));
|
|
420
|
|
421 emit_insn (gen_cmpstr_t (tmp0, tmp3));
|
|
422 jump = emit_jump_insn (gen_branch_true (L_end_loop_long));
|
|
423 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
424
|
|
425 emit_insn (gen_cmpeqsi_t (tmp1, tmp2));
|
|
426 jump = emit_jump_insn (gen_branch_false (L_end_loop_long));
|
|
427 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
428
|
|
429 if (TARGET_SH2)
|
|
430 emit_insn (gen_dect (lenw, lenw));
|
|
431 else
|
|
432 {
|
|
433 emit_insn (gen_addsi3 (lenw, lenw, GEN_INT (-1)));
|
|
434 emit_insn (gen_tstsi_t (lenw, lenw));
|
|
435 }
|
|
436
|
|
437 jump = emit_jump_insn (gen_branch_false (L_loop_long));
|
|
438 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
439
|
|
440 int sbytes = bytes % 4;
|
|
441
|
|
442 /* end loop. Reached max iterations. */
|
|
443 if (sbytes == 0)
|
|
444 {
|
|
445 emit_insn (gen_subsi3 (operands[0], tmp1, tmp2));
|
|
446 jump = emit_jump_insn (gen_jump_compact (L_return));
|
|
447 emit_barrier_after (jump);
|
|
448 }
|
|
449 else
|
|
450 {
|
|
451 /* Remaining bytes to check. */
|
|
452
|
|
453 addr1 = adjust_automodify_address (addr1, QImode, s1_addr, 0);
|
|
454 addr2 = adjust_automodify_address (addr2, QImode, s2_addr, 0);
|
|
455
|
|
456 while (sbytes--)
|
|
457 {
|
|
458 emit_insn (gen_extendqisi2 (tmp1, addr1));
|
|
459 emit_insn (gen_extendqisi2 (tmp2, addr2));
|
|
460
|
|
461 emit_insn (gen_cmpeqsi_t (tmp2, const0_rtx));
|
|
462 jump = emit_jump_insn (gen_branch_true (L_end_loop_byte));
|
|
463 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
464
|
|
465 emit_insn (gen_cmpeqsi_t (tmp1, tmp2));
|
|
466 if (flag_delayed_branch)
|
|
467 emit_insn (gen_zero_extendqisi2 (tmp2,
|
|
468 gen_lowpart (QImode,
|
|
469 tmp2)));
|
|
470 jump = emit_jump_insn (gen_branch_false (L_end_loop_byte));
|
|
471 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
472
|
|
473 addr1 = adjust_address (addr1, QImode,
|
|
474 GET_MODE_SIZE (QImode));
|
|
475 addr2 = adjust_address (addr2, QImode,
|
|
476 GET_MODE_SIZE (QImode));
|
|
477 }
|
|
478
|
|
479 jump = emit_jump_insn (gen_jump_compact( L_end_loop_byte));
|
|
480 emit_barrier_after (jump);
|
|
481 }
|
|
482
|
|
483 emit_label (L_end_loop_long);
|
|
484
|
|
485 /* Found last word. Restart it byte per byte. */
|
|
486
|
|
487 emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr,
|
|
488 -GET_MODE_SIZE (SImode)));
|
|
489 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr,
|
|
490 -GET_MODE_SIZE (SImode)));
|
|
491
|
|
492 /* fall thru. */
|
|
493 }
|
|
494
|
|
495 addr1 = adjust_automodify_address (addr1, QImode, s1_addr, 0);
|
|
496 addr2 = adjust_automodify_address (addr2, QImode, s2_addr, 0);
|
|
497
|
|
498 while (bytes--)
|
|
499 {
|
|
500 emit_insn (gen_extendqisi2 (tmp1, addr1));
|
|
501 emit_insn (gen_extendqisi2 (tmp2, addr2));
|
|
502
|
|
503 emit_insn (gen_cmpeqsi_t (tmp2, const0_rtx));
|
|
504 jump = emit_jump_insn (gen_branch_true (L_end_loop_byte));
|
|
505 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
506
|
|
507 emit_insn (gen_cmpeqsi_t (tmp1, tmp2));
|
|
508 if (flag_delayed_branch)
|
|
509 emit_insn (gen_zero_extendqisi2 (tmp2,
|
|
510 gen_lowpart (QImode, tmp2)));
|
|
511 jump = emit_jump_insn (gen_branch_false (L_end_loop_byte));
|
|
512 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
513
|
|
514 addr1 = adjust_address (addr1, QImode, GET_MODE_SIZE (QImode));
|
|
515 addr2 = adjust_address (addr2, QImode, GET_MODE_SIZE (QImode));
|
|
516 }
|
|
517
|
|
518 jump = emit_jump_insn (gen_jump_compact( L_end_loop_byte));
|
|
519 emit_barrier_after (jump);
|
|
520 }
|
|
521 else
|
|
522 {
|
|
523 emit_insn (gen_cmpeqsi_t (len, const0_rtx));
|
|
524 emit_move_insn (operands[0], const0_rtx);
|
|
525 jump = emit_jump_insn (gen_branch_true (L_return));
|
|
526 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
527 }
|
|
528
|
|
529 addr1 = adjust_automodify_address (addr1, QImode, s1_addr, 0);
|
|
530 addr2 = adjust_automodify_address (addr2, QImode, s2_addr, 0);
|
|
531
|
|
532 emit_label (L_loop_byte);
|
|
533
|
|
534 emit_insn (gen_extendqisi2 (tmp2, addr2));
|
|
535 emit_move_insn (s2_addr, plus_constant (Pmode, s2_addr, 1));
|
|
536
|
|
537 emit_insn (gen_extendqisi2 (tmp1, addr1));
|
|
538 emit_move_insn (s1_addr, plus_constant (Pmode, s1_addr, 1));
|
|
539
|
|
540 emit_insn (gen_cmpeqsi_t (tmp2, const0_rtx));
|
|
541 jump = emit_jump_insn (gen_branch_true (L_end_loop_byte));
|
|
542 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
543
|
|
544 emit_insn (gen_cmpeqsi_t (tmp1, tmp2));
|
|
545 if (flag_delayed_branch)
|
|
546 emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2)));
|
|
547 jump = emit_jump_insn (gen_branch_false (L_end_loop_byte));
|
|
548 add_int_reg_note (jump, REG_BR_PROB, prob_unlikely);
|
|
549
|
|
550 if (TARGET_SH2)
|
|
551 emit_insn (gen_dect (len, len));
|
|
552 else
|
|
553 {
|
|
554 emit_insn (gen_addsi3 (len, len, GEN_INT (-1)));
|
|
555 emit_insn (gen_tstsi_t (len, len));
|
|
556 }
|
|
557
|
|
558 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
559 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
560 /* end byte loop. */
|
|
561
|
|
562 emit_label (L_end_loop_byte);
|
|
563
|
|
564 if (! flag_delayed_branch)
|
|
565 emit_insn (gen_zero_extendqisi2 (tmp2, gen_lowpart (QImode, tmp2)));
|
|
566 emit_insn (gen_zero_extendqisi2 (tmp1, gen_lowpart (QImode, tmp1)));
|
|
567
|
|
568 emit_insn (gen_subsi3 (operands[0], tmp1, tmp2));
|
|
569
|
|
570 emit_label (L_return);
|
|
571
|
|
572 return true;
|
|
573 }
|
|
574
|
|
575 /* Emit code to perform a strlen.
|
|
576
|
|
577 OPERANDS[0] is the destination.
|
|
578 OPERANDS[1] is the string.
|
|
579 OPERANDS[2] is the char to search.
|
|
580 OPERANDS[3] is the alignment. */
|
|
581 bool
|
|
582 sh_expand_strlen (rtx *operands)
|
|
583 {
|
|
584 rtx addr1 = operands[1];
|
|
585 rtx current_addr = copy_addr_to_reg (XEXP (addr1, 0));
|
|
586 rtx start_addr = gen_reg_rtx (Pmode);
|
|
587 rtx tmp0 = gen_reg_rtx (SImode);
|
|
588 rtx tmp1 = gen_reg_rtx (SImode);
|
|
589 rtx_code_label *L_return = gen_label_rtx ();
|
|
590 rtx_code_label *L_loop_byte = gen_label_rtx ();
|
|
591
|
|
592 rtx_insn *jump;
|
|
593 rtx_code_label *L_loop_long = gen_label_rtx ();
|
|
594 rtx_code_label *L_end_loop_long = gen_label_rtx ();
|
|
595
|
|
596 int align = INTVAL (operands[3]);
|
|
597
|
|
598 emit_move_insn (operands[0], GEN_INT (-1));
|
|
599
|
|
600 /* remember start of string. */
|
|
601 emit_move_insn (start_addr, current_addr);
|
|
602
|
|
603 if (align < 4)
|
|
604 {
|
|
605 emit_insn (gen_tstsi_t (current_addr, GEN_INT (3)));
|
|
606 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
607 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
608 }
|
|
609
|
|
610 emit_move_insn (tmp0, operands[2]);
|
|
611
|
|
612 addr1 = adjust_automodify_address (addr1, SImode, current_addr, 0);
|
|
613
|
|
614 /* start long loop. */
|
|
615 emit_label (L_loop_long);
|
|
616
|
|
617 /* tmp1 is aligned, OK to load. */
|
|
618 emit_move_insn (tmp1, addr1);
|
|
619 emit_move_insn (current_addr, plus_constant (Pmode, current_addr, 4));
|
|
620
|
|
621 /* Is there a 0 byte ? */
|
|
622 emit_insn (gen_cmpstr_t (tmp0, tmp1));
|
|
623
|
|
624 jump = emit_jump_insn (gen_branch_false (L_loop_long));
|
|
625 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
626 /* end loop. */
|
|
627
|
|
628 emit_label (L_end_loop_long);
|
|
629
|
|
630 emit_move_insn (current_addr, plus_constant (Pmode, current_addr, -4));
|
|
631
|
|
632 addr1 = adjust_address (addr1, QImode, 0);
|
|
633
|
|
634 /* unroll remaining bytes. */
|
|
635 for (int i = 0; i < 4; ++i)
|
|
636 {
|
|
637 emit_insn (gen_extendqisi2 (tmp1, addr1));
|
|
638 emit_move_insn (current_addr, plus_constant (Pmode, current_addr, 1));
|
|
639 emit_insn (gen_cmpeqsi_t (tmp1, const0_rtx));
|
|
640 jump = emit_jump_insn (gen_branch_true (L_return));
|
|
641 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
642 }
|
|
643
|
|
644 emit_barrier_after (jump);
|
|
645
|
|
646 /* start byte loop. */
|
|
647 emit_label (L_loop_byte);
|
|
648
|
|
649 emit_insn (gen_extendqisi2 (tmp1, addr1));
|
|
650 emit_move_insn (current_addr, plus_constant (Pmode, current_addr, 1));
|
|
651
|
|
652 emit_insn (gen_cmpeqsi_t (tmp1, const0_rtx));
|
|
653 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
654 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
655
|
|
656 /* end loop. */
|
|
657
|
|
658 emit_label (L_return);
|
|
659
|
|
660 emit_insn (gen_addsi3 (start_addr, start_addr, GEN_INT (1)));
|
|
661 emit_insn (gen_subsi3 (operands[0], current_addr, start_addr));
|
|
662
|
|
663 return true;
|
|
664 }
|
|
665
|
|
666 /* Emit code to perform a memset.
|
|
667
|
|
668 OPERANDS[0] is the destination.
|
|
669 OPERANDS[1] is the size;
|
|
670 OPERANDS[2] is the char to search.
|
|
671 OPERANDS[3] is the alignment. */
|
|
672 void
|
|
673 sh_expand_setmem (rtx *operands)
|
|
674 {
|
|
675 rtx_code_label *L_loop_byte = gen_label_rtx ();
|
|
676 rtx_code_label *L_loop_word = gen_label_rtx ();
|
|
677 rtx_code_label *L_return = gen_label_rtx ();
|
|
678 rtx_insn *jump;
|
|
679 rtx dest = copy_rtx (operands[0]);
|
|
680 rtx dest_addr = copy_addr_to_reg (XEXP (dest, 0));
|
|
681 rtx val = copy_to_mode_reg (SImode, operands[2]);
|
|
682 int align = INTVAL (operands[3]);
|
|
683 rtx len = copy_to_mode_reg (SImode, operands[1]);
|
|
684
|
|
685 if (! CONST_INT_P (operands[1]))
|
|
686 return;
|
|
687
|
|
688 int count = INTVAL (operands[1]);
|
|
689
|
|
690 if (CONST_INT_P (operands[2])
|
|
691 && (INTVAL (operands[2]) == 0 || INTVAL (operands[2]) == -1) && count > 8)
|
|
692 {
|
|
693 rtx lenw = gen_reg_rtx (SImode);
|
|
694
|
|
695 if (align < 4)
|
|
696 {
|
|
697 emit_insn (gen_tstsi_t (dest_addr, GEN_INT (3)));
|
|
698 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
699 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
700 }
|
|
701
|
|
702 /* word count. Do we have iterations ? */
|
|
703 emit_insn (gen_lshrsi3 (lenw, len, GEN_INT (2)));
|
|
704
|
|
705 dest = adjust_automodify_address (dest, SImode, dest_addr, 0);
|
|
706
|
|
707 /* start loop. */
|
|
708 emit_label (L_loop_word);
|
|
709
|
|
710 if (TARGET_SH2)
|
|
711 emit_insn (gen_dect (lenw, lenw));
|
|
712 else
|
|
713 {
|
|
714 emit_insn (gen_addsi3 (lenw, lenw, GEN_INT (-1)));
|
|
715 emit_insn (gen_tstsi_t (lenw, lenw));
|
|
716 }
|
|
717
|
|
718 emit_move_insn (dest, val);
|
|
719 emit_move_insn (dest_addr, plus_constant (Pmode, dest_addr,
|
|
720 GET_MODE_SIZE (SImode)));
|
|
721
|
|
722
|
|
723 jump = emit_jump_insn (gen_branch_false (L_loop_word));
|
|
724 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
725 count = count % 4;
|
|
726
|
|
727 dest = adjust_address (dest, QImode, 0);
|
|
728
|
|
729 val = gen_lowpart (QImode, val);
|
|
730
|
|
731 while (count--)
|
|
732 {
|
|
733 emit_move_insn (dest, val);
|
|
734 emit_move_insn (dest_addr, plus_constant (Pmode, dest_addr,
|
|
735 GET_MODE_SIZE (QImode)));
|
|
736 }
|
|
737
|
|
738 jump = emit_jump_insn (gen_jump_compact (L_return));
|
|
739 emit_barrier_after (jump);
|
|
740 }
|
|
741
|
|
742 dest = adjust_automodify_address (dest, QImode, dest_addr, 0);
|
|
743
|
|
744 /* start loop. */
|
|
745 emit_label (L_loop_byte);
|
|
746
|
|
747 if (TARGET_SH2)
|
|
748 emit_insn (gen_dect (len, len));
|
|
749 else
|
|
750 {
|
|
751 emit_insn (gen_addsi3 (len, len, GEN_INT (-1)));
|
|
752 emit_insn (gen_tstsi_t (len, len));
|
|
753 }
|
|
754
|
|
755 val = gen_lowpart (QImode, val);
|
|
756 emit_move_insn (dest, val);
|
|
757 emit_move_insn (dest_addr, plus_constant (Pmode, dest_addr,
|
|
758 GET_MODE_SIZE (QImode)));
|
|
759
|
|
760 jump = emit_jump_insn (gen_branch_false (L_loop_byte));
|
|
761 add_int_reg_note (jump, REG_BR_PROB, prob_likely);
|
|
762
|
|
763 emit_label (L_return);
|
|
764 }
|