rs6000.md (fseldfsf4): Add TARGET_SINGLE_FLOAT condition.
[gcc.git] / gcc / local-alloc.c
1 /* Allocate registers within a basic block, for GNU compiler.
2 Copyright (C) 1987, 1988, 1991, 1993, 1994, 1995, 1996, 1997, 1998,
3 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007
4 Free Software Foundation, Inc.
5
6 This file is part of GCC.
7
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
11 version.
12
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
16 for more details.
17
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
21
22 /* Allocation of hard register numbers to pseudo registers is done in
23 two passes. In this pass we consider only regs that are born and
24 die once within one basic block. We do this one basic block at a
25 time. Then the next pass allocates the registers that remain.
26 Two passes are used because this pass uses methods that work only
27 on linear code, but that do a better job than the general methods
28 used in global_alloc, and more quickly too.
29
30 The assignments made are recorded in the vector reg_renumber
31 whose space is allocated here. The rtl code itself is not altered.
32
33 We assign each instruction in the basic block a number
34 which is its order from the beginning of the block.
35 Then we can represent the lifetime of a pseudo register with
36 a pair of numbers, and check for conflicts easily.
37 We can record the availability of hard registers with a
38 HARD_REG_SET for each instruction. The HARD_REG_SET
39 contains 0 or 1 for each hard reg.
40
41 To avoid register shuffling, we tie registers together when one
42 dies by being copied into another, or dies in an instruction that
43 does arithmetic to produce another. The tied registers are
44 allocated as one. Registers with different reg class preferences
45 can never be tied unless the class preferred by one is a subclass
46 of the one preferred by the other.
47
48 Tying is represented with "quantity numbers".
49 A non-tied register is given a new quantity number.
50 Tied registers have the same quantity number.
51
52 We have provision to exempt registers, even when they are contained
53 within the block, that can be tied to others that are not contained in it.
54 This is so that global_alloc could process them both and tie them then.
55 But this is currently disabled since tying in global_alloc is not
56 yet implemented. */
57
58 /* Pseudos allocated here can be reallocated by global.c if the hard register
59 is used as a spill register. Currently we don't allocate such pseudos
60 here if their preferred class is likely to be used by spills. */
61
62 #include "config.h"
63 #include "system.h"
64 #include "coretypes.h"
65 #include "tm.h"
66 #include "hard-reg-set.h"
67 #include "rtl.h"
68 #include "tm_p.h"
69 #include "flags.h"
70 #include "regs.h"
71 #include "function.h"
72 #include "insn-config.h"
73 #include "insn-attr.h"
74 #include "recog.h"
75 #include "output.h"
76 #include "toplev.h"
77 #include "except.h"
78 #include "integrate.h"
79 #include "reload.h"
80 #include "ggc.h"
81 #include "timevar.h"
82 #include "tree-pass.h"
83 #include "df.h"
84 #include "dbgcnt.h"
85
86 \f
87 /* Next quantity number available for allocation. */
88
89 static int next_qty;
90
91 /* Information we maintain about each quantity. */
92 struct qty
93 {
94 /* The number of refs to quantity Q. */
95
96 int n_refs;
97
98 /* The frequency of uses of quantity Q. */
99
100 int freq;
101
102 /* Insn number (counting from head of basic block)
103 where quantity Q was born. -1 if birth has not been recorded. */
104
105 int birth;
106
107 /* Insn number (counting from head of basic block)
108 where given quantity died. Due to the way tying is done,
109 and the fact that we consider in this pass only regs that die but once,
110 a quantity can die only once. Each quantity's life span
111 is a set of consecutive insns. -1 if death has not been recorded. */
112
113 int death;
114
115 /* Number of words needed to hold the data in given quantity.
116 This depends on its machine mode. It is used for these purposes:
117 1. It is used in computing the relative importance of qtys,
118 which determines the order in which we look for regs for them.
119 2. It is used in rules that prevent tying several registers of
120 different sizes in a way that is geometrically impossible
121 (see combine_regs). */
122
123 int size;
124
125 /* Number of times a reg tied to given qty lives across a CALL_INSN. */
126
127 int n_calls_crossed;
128
129 /* Number of times a reg tied to given qty lives across a CALL_INSN. */
130
131 int freq_calls_crossed;
132
133 /* Number of times a reg tied to given qty lives across a CALL_INSN
134 that might throw. */
135
136 int n_throwing_calls_crossed;
137
138 /* The register number of one pseudo register whose reg_qty value is Q.
139 This register should be the head of the chain
140 maintained in reg_next_in_qty. */
141
142 int first_reg;
143
144 /* Reg class contained in (smaller than) the preferred classes of all
145 the pseudo regs that are tied in given quantity.
146 This is the preferred class for allocating that quantity. */
147
148 enum reg_class min_class;
149
150 /* Register class within which we allocate given qty if we can't get
151 its preferred class. */
152
153 enum reg_class alternate_class;
154
155 /* This holds the mode of the registers that are tied to given qty,
156 or VOIDmode if registers with differing modes are tied together. */
157
158 enum machine_mode mode;
159
160 /* the hard reg number chosen for given quantity,
161 or -1 if none was found. */
162
163 short phys_reg;
164 };
165
166 static struct qty *qty;
167
168 /* These fields are kept separately to speedup their clearing. */
169
170 /* We maintain two hard register sets that indicate suggested hard registers
171 for each quantity. The first, phys_copy_sugg, contains hard registers
172 that are tied to the quantity by a simple copy. The second contains all
173 hard registers that are tied to the quantity via an arithmetic operation.
174
175 The former register set is given priority for allocation. This tends to
176 eliminate copy insns. */
177
178 /* Element Q is a set of hard registers that are suggested for quantity Q by
179 copy insns. */
180
181 static HARD_REG_SET *qty_phys_copy_sugg;
182
183 /* Element Q is a set of hard registers that are suggested for quantity Q by
184 arithmetic insns. */
185
186 static HARD_REG_SET *qty_phys_sugg;
187
188 /* Element Q is the number of suggested registers in qty_phys_copy_sugg. */
189
190 static short *qty_phys_num_copy_sugg;
191
192 /* Element Q is the number of suggested registers in qty_phys_sugg. */
193
194 static short *qty_phys_num_sugg;
195
196 /* If (REG N) has been assigned a quantity number, is a register number
197 of another register assigned the same quantity number, or -1 for the
198 end of the chain. qty->first_reg point to the head of this chain. */
199
200 static int *reg_next_in_qty;
201
202 /* reg_qty[N] (where N is a pseudo reg number) is the qty number of that reg
203 if it is >= 0,
204 of -1 if this register cannot be allocated by local-alloc,
205 or -2 if not known yet.
206
207 Note that if we see a use or death of pseudo register N with
208 reg_qty[N] == -2, register N must be local to the current block. If
209 it were used in more than one block, we would have reg_qty[N] == -1.
210 This relies on the fact that if reg_basic_block[N] is >= 0, register N
211 will not appear in any other block. We save a considerable number of
212 tests by exploiting this.
213
214 If N is < FIRST_PSEUDO_REGISTER, reg_qty[N] is undefined and should not
215 be referenced. */
216
217 static int *reg_qty;
218
219 /* The offset (in words) of register N within its quantity.
220 This can be nonzero if register N is SImode, and has been tied
221 to a subreg of a DImode register. */
222
223 static char *reg_offset;
224
225 /* Vector of substitutions of register numbers,
226 used to map pseudo regs into hardware regs.
227 This is set up as a result of register allocation.
228 Element N is the hard reg assigned to pseudo reg N,
229 or is -1 if no hard reg was assigned.
230 If N is a hard reg number, element N is N. */
231
232 short *reg_renumber;
233
234 /* Set of hard registers live at the current point in the scan
235 of the instructions in a basic block. */
236
237 static HARD_REG_SET regs_live;
238
239 /* Each set of hard registers indicates registers live at a particular
240 point in the basic block. For N even, regs_live_at[N] says which
241 hard registers are needed *after* insn N/2 (i.e., they may not
242 conflict with the outputs of insn N/2 or the inputs of insn N/2 + 1.
243
244 If an object is to conflict with the inputs of insn J but not the
245 outputs of insn J + 1, we say it is born at index J*2 - 1. Similarly,
246 if it is to conflict with the outputs of insn J but not the inputs of
247 insn J + 1, it is said to die at index J*2 + 1. */
248
249 static HARD_REG_SET *regs_live_at;
250
251 /* Communicate local vars `insn_number' and `insn'
252 from `block_alloc' to `reg_is_set', `wipe_dead_reg', and `alloc_qty'. */
253 static int this_insn_number;
254 static rtx this_insn;
255
256 struct equivalence
257 {
258 /* Set when an attempt should be made to replace a register
259 with the associated src_p entry. */
260
261 char replace;
262
263 /* Set when a REG_EQUIV note is found or created. Use to
264 keep track of what memory accesses might be created later,
265 e.g. by reload. */
266
267 rtx replacement;
268
269 rtx *src_p;
270
271 /* Loop depth is used to recognize equivalences which appear
272 to be present within the same loop (or in an inner loop). */
273
274 int loop_depth;
275
276 /* The list of each instruction which initializes this register. */
277
278 rtx init_insns;
279
280 /* Nonzero if this had a preexisting REG_EQUIV note. */
281
282 int is_arg_equivalence;
283 };
284
285 /* reg_equiv[N] (where N is a pseudo reg number) is the equivalence
286 structure for that register. */
287
288 static struct equivalence *reg_equiv;
289
290 /* Nonzero if we recorded an equivalence for a LABEL_REF. */
291 static int recorded_label_ref;
292
293 static void alloc_qty (int, enum machine_mode, int, int);
294 static void validate_equiv_mem_from_store (rtx, const_rtx, void *);
295 static int validate_equiv_mem (rtx, rtx, rtx);
296 static int equiv_init_varies_p (rtx);
297 static int equiv_init_movable_p (rtx, int);
298 static int contains_replace_regs (rtx);
299 static int memref_referenced_p (rtx, rtx);
300 static int memref_used_between_p (rtx, rtx, rtx);
301 static void no_equiv (rtx, const_rtx, void *);
302 static void block_alloc (basic_block);
303 static int qty_sugg_compare (int, int);
304 static int qty_sugg_compare_1 (const void *, const void *);
305 static int qty_compare (int, int);
306 static int qty_compare_1 (const void *, const void *);
307 static int combine_regs (rtx, rtx, int, int, rtx);
308 static int reg_meets_class_p (int, enum reg_class);
309 static void update_qty_class (int, int);
310 static void reg_is_set (rtx, const_rtx, void *);
311 static void reg_is_born (rtx, int);
312 static void wipe_dead_reg (rtx, int);
313 static int find_free_reg (enum reg_class, enum machine_mode, int, int, int,
314 int, int, basic_block);
315 static void mark_life (int, enum machine_mode, int);
316 static void post_mark_life (int, enum machine_mode, int, int, int);
317 static int requires_inout (const char *);
318 \f
319 /* Allocate a new quantity (new within current basic block)
320 for register number REGNO which is born at index BIRTH
321 within the block. MODE and SIZE are info on reg REGNO. */
322
323 static void
324 alloc_qty (int regno, enum machine_mode mode, int size, int birth)
325 {
326 int qtyno = next_qty++;
327
328 reg_qty[regno] = qtyno;
329 reg_offset[regno] = 0;
330 reg_next_in_qty[regno] = -1;
331
332 qty[qtyno].first_reg = regno;
333 qty[qtyno].size = size;
334 qty[qtyno].mode = mode;
335 qty[qtyno].birth = birth;
336 qty[qtyno].n_calls_crossed = REG_N_CALLS_CROSSED (regno);
337 qty[qtyno].freq_calls_crossed = REG_FREQ_CALLS_CROSSED (regno);
338 qty[qtyno].n_throwing_calls_crossed = REG_N_THROWING_CALLS_CROSSED (regno);
339 qty[qtyno].min_class = reg_preferred_class (regno);
340 qty[qtyno].alternate_class = reg_alternate_class (regno);
341 qty[qtyno].n_refs = REG_N_REFS (regno);
342 qty[qtyno].freq = REG_FREQ (regno);
343 }
344 \f
345 /* Main entry point of this file. */
346
347 static int
348 local_alloc (void)
349 {
350 int i;
351 int max_qty;
352 basic_block b;
353
354 /* We need to keep track of whether or not we recorded a LABEL_REF so
355 that we know if the jump optimizer needs to be rerun. */
356 recorded_label_ref = 0;
357
358 /* Leaf functions and non-leaf functions have different needs.
359 If defined, let the machine say what kind of ordering we
360 should use. */
361 #ifdef ORDER_REGS_FOR_LOCAL_ALLOC
362 ORDER_REGS_FOR_LOCAL_ALLOC;
363 #endif
364
365 /* Promote REG_EQUAL notes to REG_EQUIV notes and adjust status of affected
366 registers. */
367 update_equiv_regs ();
368
369 /* This sets the maximum number of quantities we can have. Quantity
370 numbers start at zero and we can have one for each pseudo. */
371 max_qty = (max_regno - FIRST_PSEUDO_REGISTER);
372
373 /* Allocate vectors of temporary data.
374 See the declarations of these variables, above,
375 for what they mean. */
376
377 qty = XNEWVEC (struct qty, max_qty);
378 qty_phys_copy_sugg = XNEWVEC (HARD_REG_SET, max_qty);
379 qty_phys_num_copy_sugg = XNEWVEC (short, max_qty);
380 qty_phys_sugg = XNEWVEC (HARD_REG_SET, max_qty);
381 qty_phys_num_sugg = XNEWVEC (short, max_qty);
382
383 reg_qty = XNEWVEC (int, max_regno);
384 reg_offset = XNEWVEC (char, max_regno);
385 reg_next_in_qty = XNEWVEC (int, max_regno);
386
387 /* Determine which pseudo-registers can be allocated by local-alloc.
388 In general, these are the registers used only in a single block and
389 which only die once.
390
391 We need not be concerned with which block actually uses the register
392 since we will never see it outside that block. */
393
394 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
395 {
396 if (REG_BASIC_BLOCK (i) >= NUM_FIXED_BLOCKS && REG_N_DEATHS (i) == 1)
397 reg_qty[i] = -2;
398 else
399 reg_qty[i] = -1;
400 }
401
402 /* Force loop below to initialize entire quantity array. */
403 next_qty = max_qty;
404
405 /* Allocate each block's local registers, block by block. */
406
407 FOR_EACH_BB (b)
408 {
409 /* NEXT_QTY indicates which elements of the `qty_...'
410 vectors might need to be initialized because they were used
411 for the previous block; it is set to the entire array before
412 block 0. Initialize those, with explicit loop if there are few,
413 else with bzero and bcopy. Do not initialize vectors that are
414 explicit set by `alloc_qty'. */
415
416 if (next_qty < 6)
417 {
418 for (i = 0; i < next_qty; i++)
419 {
420 CLEAR_HARD_REG_SET (qty_phys_copy_sugg[i]);
421 qty_phys_num_copy_sugg[i] = 0;
422 CLEAR_HARD_REG_SET (qty_phys_sugg[i]);
423 qty_phys_num_sugg[i] = 0;
424 }
425 }
426 else
427 {
428 #define CLEAR(vector) \
429 memset ((vector), 0, (sizeof (*(vector))) * next_qty);
430
431 CLEAR (qty_phys_copy_sugg);
432 CLEAR (qty_phys_num_copy_sugg);
433 CLEAR (qty_phys_sugg);
434 CLEAR (qty_phys_num_sugg);
435 }
436
437 next_qty = 0;
438
439 block_alloc (b);
440 }
441
442 free (qty);
443 free (qty_phys_copy_sugg);
444 free (qty_phys_num_copy_sugg);
445 free (qty_phys_sugg);
446 free (qty_phys_num_sugg);
447
448 free (reg_qty);
449 free (reg_offset);
450 free (reg_next_in_qty);
451
452 return recorded_label_ref;
453 }
454 \f
455 /* Used for communication between the following two functions: contains
456 a MEM that we wish to ensure remains unchanged. */
457 static rtx equiv_mem;
458
459 /* Set nonzero if EQUIV_MEM is modified. */
460 static int equiv_mem_modified;
461
462 /* If EQUIV_MEM is modified by modifying DEST, indicate that it is modified.
463 Called via note_stores. */
464
465 static void
466 validate_equiv_mem_from_store (rtx dest, const_rtx set ATTRIBUTE_UNUSED,
467 void *data ATTRIBUTE_UNUSED)
468 {
469 if ((REG_P (dest)
470 && reg_overlap_mentioned_p (dest, equiv_mem))
471 || (MEM_P (dest)
472 && true_dependence (dest, VOIDmode, equiv_mem, rtx_varies_p)))
473 equiv_mem_modified = 1;
474 }
475
476 /* Verify that no store between START and the death of REG invalidates
477 MEMREF. MEMREF is invalidated by modifying a register used in MEMREF,
478 by storing into an overlapping memory location, or with a non-const
479 CALL_INSN.
480
481 Return 1 if MEMREF remains valid. */
482
483 static int
484 validate_equiv_mem (rtx start, rtx reg, rtx memref)
485 {
486 rtx insn;
487 rtx note;
488
489 equiv_mem = memref;
490 equiv_mem_modified = 0;
491
492 /* If the memory reference has side effects or is volatile, it isn't a
493 valid equivalence. */
494 if (side_effects_p (memref))
495 return 0;
496
497 for (insn = start; insn && ! equiv_mem_modified; insn = NEXT_INSN (insn))
498 {
499 if (! INSN_P (insn))
500 continue;
501
502 if (find_reg_note (insn, REG_DEAD, reg))
503 return 1;
504
505 if (CALL_P (insn) && ! MEM_READONLY_P (memref)
506 && ! RTL_CONST_OR_PURE_CALL_P (insn))
507 return 0;
508
509 note_stores (PATTERN (insn), validate_equiv_mem_from_store, NULL);
510
511 /* If a register mentioned in MEMREF is modified via an
512 auto-increment, we lose the equivalence. Do the same if one
513 dies; although we could extend the life, it doesn't seem worth
514 the trouble. */
515
516 for (note = REG_NOTES (insn); note; note = XEXP (note, 1))
517 if ((REG_NOTE_KIND (note) == REG_INC
518 || REG_NOTE_KIND (note) == REG_DEAD)
519 && REG_P (XEXP (note, 0))
520 && reg_overlap_mentioned_p (XEXP (note, 0), memref))
521 return 0;
522 }
523
524 return 0;
525 }
526
527 /* Returns zero if X is known to be invariant. */
528
529 static int
530 equiv_init_varies_p (rtx x)
531 {
532 RTX_CODE code = GET_CODE (x);
533 int i;
534 const char *fmt;
535
536 switch (code)
537 {
538 case MEM:
539 return !MEM_READONLY_P (x) || equiv_init_varies_p (XEXP (x, 0));
540
541 case CONST:
542 case CONST_INT:
543 case CONST_DOUBLE:
544 case CONST_FIXED:
545 case CONST_VECTOR:
546 case SYMBOL_REF:
547 case LABEL_REF:
548 return 0;
549
550 case REG:
551 return reg_equiv[REGNO (x)].replace == 0 && rtx_varies_p (x, 0);
552
553 case ASM_OPERANDS:
554 if (MEM_VOLATILE_P (x))
555 return 1;
556
557 /* Fall through. */
558
559 default:
560 break;
561 }
562
563 fmt = GET_RTX_FORMAT (code);
564 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
565 if (fmt[i] == 'e')
566 {
567 if (equiv_init_varies_p (XEXP (x, i)))
568 return 1;
569 }
570 else if (fmt[i] == 'E')
571 {
572 int j;
573 for (j = 0; j < XVECLEN (x, i); j++)
574 if (equiv_init_varies_p (XVECEXP (x, i, j)))
575 return 1;
576 }
577
578 return 0;
579 }
580
581 /* Returns nonzero if X (used to initialize register REGNO) is movable.
582 X is only movable if the registers it uses have equivalent initializations
583 which appear to be within the same loop (or in an inner loop) and movable
584 or if they are not candidates for local_alloc and don't vary. */
585
586 static int
587 equiv_init_movable_p (rtx x, int regno)
588 {
589 int i, j;
590 const char *fmt;
591 enum rtx_code code = GET_CODE (x);
592
593 switch (code)
594 {
595 case SET:
596 return equiv_init_movable_p (SET_SRC (x), regno);
597
598 case CC0:
599 case CLOBBER:
600 return 0;
601
602 case PRE_INC:
603 case PRE_DEC:
604 case POST_INC:
605 case POST_DEC:
606 case PRE_MODIFY:
607 case POST_MODIFY:
608 return 0;
609
610 case REG:
611 return (reg_equiv[REGNO (x)].loop_depth >= reg_equiv[regno].loop_depth
612 && reg_equiv[REGNO (x)].replace)
613 || (REG_BASIC_BLOCK (REGNO (x)) < NUM_FIXED_BLOCKS && ! rtx_varies_p (x, 0));
614
615 case UNSPEC_VOLATILE:
616 return 0;
617
618 case ASM_OPERANDS:
619 if (MEM_VOLATILE_P (x))
620 return 0;
621
622 /* Fall through. */
623
624 default:
625 break;
626 }
627
628 fmt = GET_RTX_FORMAT (code);
629 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
630 switch (fmt[i])
631 {
632 case 'e':
633 if (! equiv_init_movable_p (XEXP (x, i), regno))
634 return 0;
635 break;
636 case 'E':
637 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
638 if (! equiv_init_movable_p (XVECEXP (x, i, j), regno))
639 return 0;
640 break;
641 }
642
643 return 1;
644 }
645
646 /* TRUE if X uses any registers for which reg_equiv[REGNO].replace is true. */
647
648 static int
649 contains_replace_regs (rtx x)
650 {
651 int i, j;
652 const char *fmt;
653 enum rtx_code code = GET_CODE (x);
654
655 switch (code)
656 {
657 case CONST_INT:
658 case CONST:
659 case LABEL_REF:
660 case SYMBOL_REF:
661 case CONST_DOUBLE:
662 case CONST_FIXED:
663 case CONST_VECTOR:
664 case PC:
665 case CC0:
666 case HIGH:
667 return 0;
668
669 case REG:
670 return reg_equiv[REGNO (x)].replace;
671
672 default:
673 break;
674 }
675
676 fmt = GET_RTX_FORMAT (code);
677 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
678 switch (fmt[i])
679 {
680 case 'e':
681 if (contains_replace_regs (XEXP (x, i)))
682 return 1;
683 break;
684 case 'E':
685 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
686 if (contains_replace_regs (XVECEXP (x, i, j)))
687 return 1;
688 break;
689 }
690
691 return 0;
692 }
693 \f
694 /* TRUE if X references a memory location that would be affected by a store
695 to MEMREF. */
696
697 static int
698 memref_referenced_p (rtx memref, rtx x)
699 {
700 int i, j;
701 const char *fmt;
702 enum rtx_code code = GET_CODE (x);
703
704 switch (code)
705 {
706 case CONST_INT:
707 case CONST:
708 case LABEL_REF:
709 case SYMBOL_REF:
710 case CONST_DOUBLE:
711 case CONST_FIXED:
712 case CONST_VECTOR:
713 case PC:
714 case CC0:
715 case HIGH:
716 case LO_SUM:
717 return 0;
718
719 case REG:
720 return (reg_equiv[REGNO (x)].replacement
721 && memref_referenced_p (memref,
722 reg_equiv[REGNO (x)].replacement));
723
724 case MEM:
725 if (true_dependence (memref, VOIDmode, x, rtx_varies_p))
726 return 1;
727 break;
728
729 case SET:
730 /* If we are setting a MEM, it doesn't count (its address does), but any
731 other SET_DEST that has a MEM in it is referencing the MEM. */
732 if (MEM_P (SET_DEST (x)))
733 {
734 if (memref_referenced_p (memref, XEXP (SET_DEST (x), 0)))
735 return 1;
736 }
737 else if (memref_referenced_p (memref, SET_DEST (x)))
738 return 1;
739
740 return memref_referenced_p (memref, SET_SRC (x));
741
742 default:
743 break;
744 }
745
746 fmt = GET_RTX_FORMAT (code);
747 for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
748 switch (fmt[i])
749 {
750 case 'e':
751 if (memref_referenced_p (memref, XEXP (x, i)))
752 return 1;
753 break;
754 case 'E':
755 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
756 if (memref_referenced_p (memref, XVECEXP (x, i, j)))
757 return 1;
758 break;
759 }
760
761 return 0;
762 }
763
764 /* TRUE if some insn in the range (START, END] references a memory location
765 that would be affected by a store to MEMREF. */
766
767 static int
768 memref_used_between_p (rtx memref, rtx start, rtx end)
769 {
770 rtx insn;
771
772 for (insn = NEXT_INSN (start); insn != NEXT_INSN (end);
773 insn = NEXT_INSN (insn))
774 {
775 if (!INSN_P (insn))
776 continue;
777
778 if (memref_referenced_p (memref, PATTERN (insn)))
779 return 1;
780
781 /* Nonconst functions may access memory. */
782 if (CALL_P (insn) && (! RTL_CONST_CALL_P (insn)))
783 return 1;
784 }
785
786 return 0;
787 }
788 \f
789 /* Find registers that are equivalent to a single value throughout the
790 compilation (either because they can be referenced in memory or are set once
791 from a single constant). Lower their priority for a register.
792
793 If such a register is only referenced once, try substituting its value
794 into the using insn. If it succeeds, we can eliminate the register
795 completely.
796
797 Initialize the REG_EQUIV_INIT array of initializing insns.
798
799 Return non-zero if jump label rebuilding should be done. */
800
801 int
802 update_equiv_regs (void)
803 {
804 rtx insn;
805 basic_block bb;
806 int loop_depth;
807 bitmap cleared_regs;
808
809 reg_equiv = XCNEWVEC (struct equivalence, max_regno);
810 reg_equiv_init = GGC_CNEWVEC (rtx, max_regno);
811 reg_equiv_init_size = max_regno;
812
813 init_alias_analysis ();
814
815 /* Scan the insns and find which registers have equivalences. Do this
816 in a separate scan of the insns because (due to -fcse-follow-jumps)
817 a register can be set below its use. */
818 FOR_EACH_BB (bb)
819 {
820 loop_depth = bb->loop_depth;
821
822 for (insn = BB_HEAD (bb);
823 insn != NEXT_INSN (BB_END (bb));
824 insn = NEXT_INSN (insn))
825 {
826 rtx note;
827 rtx set;
828 rtx dest, src;
829 int regno;
830
831 if (! INSN_P (insn))
832 continue;
833
834 for (note = REG_NOTES (insn); note; note = XEXP (note, 1))
835 if (REG_NOTE_KIND (note) == REG_INC)
836 no_equiv (XEXP (note, 0), note, NULL);
837
838 set = single_set (insn);
839
840 /* If this insn contains more (or less) than a single SET,
841 only mark all destinations as having no known equivalence. */
842 if (set == 0)
843 {
844 note_stores (PATTERN (insn), no_equiv, NULL);
845 continue;
846 }
847 else if (GET_CODE (PATTERN (insn)) == PARALLEL)
848 {
849 int i;
850
851 for (i = XVECLEN (PATTERN (insn), 0) - 1; i >= 0; i--)
852 {
853 rtx part = XVECEXP (PATTERN (insn), 0, i);
854 if (part != set)
855 note_stores (part, no_equiv, NULL);
856 }
857 }
858
859 dest = SET_DEST (set);
860 src = SET_SRC (set);
861
862 /* See if this is setting up the equivalence between an argument
863 register and its stack slot. */
864 note = find_reg_note (insn, REG_EQUIV, NULL_RTX);
865 if (note)
866 {
867 gcc_assert (REG_P (dest));
868 regno = REGNO (dest);
869
870 /* Note that we don't want to clear reg_equiv_init even if there
871 are multiple sets of this register. */
872 reg_equiv[regno].is_arg_equivalence = 1;
873
874 /* Record for reload that this is an equivalencing insn. */
875 if (rtx_equal_p (src, XEXP (note, 0)))
876 reg_equiv_init[regno]
877 = gen_rtx_INSN_LIST (VOIDmode, insn, reg_equiv_init[regno]);
878
879 /* Continue normally in case this is a candidate for
880 replacements. */
881 }
882
883 if (!optimize)
884 continue;
885
886 /* We only handle the case of a pseudo register being set
887 once, or always to the same value. */
888 /* ??? The mn10200 port breaks if we add equivalences for
889 values that need an ADDRESS_REGS register and set them equivalent
890 to a MEM of a pseudo. The actual problem is in the over-conservative
891 handling of INPADDR_ADDRESS / INPUT_ADDRESS / INPUT triples in
892 calculate_needs, but we traditionally work around this problem
893 here by rejecting equivalences when the destination is in a register
894 that's likely spilled. This is fragile, of course, since the
895 preferred class of a pseudo depends on all instructions that set
896 or use it. */
897
898 if (!REG_P (dest)
899 || (regno = REGNO (dest)) < FIRST_PSEUDO_REGISTER
900 || reg_equiv[regno].init_insns == const0_rtx
901 || (CLASS_LIKELY_SPILLED_P (reg_preferred_class (regno))
902 && MEM_P (src) && ! reg_equiv[regno].is_arg_equivalence))
903 {
904 /* This might be setting a SUBREG of a pseudo, a pseudo that is
905 also set somewhere else to a constant. */
906 note_stores (set, no_equiv, NULL);
907 continue;
908 }
909
910 note = find_reg_note (insn, REG_EQUAL, NULL_RTX);
911
912 /* cse sometimes generates function invariants, but doesn't put a
913 REG_EQUAL note on the insn. Since this note would be redundant,
914 there's no point creating it earlier than here. */
915 if (! note && ! rtx_varies_p (src, 0))
916 note = set_unique_reg_note (insn, REG_EQUAL, copy_rtx (src));
917
918 /* Don't bother considering a REG_EQUAL note containing an EXPR_LIST
919 since it represents a function call */
920 if (note && GET_CODE (XEXP (note, 0)) == EXPR_LIST)
921 note = NULL_RTX;
922
923 if (DF_REG_DEF_COUNT (regno) != 1
924 && (! note
925 || rtx_varies_p (XEXP (note, 0), 0)
926 || (reg_equiv[regno].replacement
927 && ! rtx_equal_p (XEXP (note, 0),
928 reg_equiv[regno].replacement))))
929 {
930 no_equiv (dest, set, NULL);
931 continue;
932 }
933 /* Record this insn as initializing this register. */
934 reg_equiv[regno].init_insns
935 = gen_rtx_INSN_LIST (VOIDmode, insn, reg_equiv[regno].init_insns);
936
937 /* If this register is known to be equal to a constant, record that
938 it is always equivalent to the constant. */
939 if (DF_REG_DEF_COUNT (regno) == 1
940 && note && ! rtx_varies_p (XEXP (note, 0), 0))
941 {
942 rtx note_value = XEXP (note, 0);
943 remove_note (insn, note);
944 set_unique_reg_note (insn, REG_EQUIV, note_value);
945 }
946
947 /* If this insn introduces a "constant" register, decrease the priority
948 of that register. Record this insn if the register is only used once
949 more and the equivalence value is the same as our source.
950
951 The latter condition is checked for two reasons: First, it is an
952 indication that it may be more efficient to actually emit the insn
953 as written (if no registers are available, reload will substitute
954 the equivalence). Secondly, it avoids problems with any registers
955 dying in this insn whose death notes would be missed.
956
957 If we don't have a REG_EQUIV note, see if this insn is loading
958 a register used only in one basic block from a MEM. If so, and the
959 MEM remains unchanged for the life of the register, add a REG_EQUIV
960 note. */
961
962 note = find_reg_note (insn, REG_EQUIV, NULL_RTX);
963
964 if (note == 0 && REG_BASIC_BLOCK (regno) >= NUM_FIXED_BLOCKS
965 && MEM_P (SET_SRC (set))
966 && validate_equiv_mem (insn, dest, SET_SRC (set)))
967 note = set_unique_reg_note (insn, REG_EQUIV, copy_rtx (SET_SRC (set)));
968
969 if (note)
970 {
971 int regno = REGNO (dest);
972 rtx x = XEXP (note, 0);
973
974 /* If we haven't done so, record for reload that this is an
975 equivalencing insn. */
976 if (!reg_equiv[regno].is_arg_equivalence)
977 reg_equiv_init[regno]
978 = gen_rtx_INSN_LIST (VOIDmode, insn, reg_equiv_init[regno]);
979
980 /* Record whether or not we created a REG_EQUIV note for a LABEL_REF.
981 We might end up substituting the LABEL_REF for uses of the
982 pseudo here or later. That kind of transformation may turn an
983 indirect jump into a direct jump, in which case we must rerun the
984 jump optimizer to ensure that the JUMP_LABEL fields are valid. */
985 if (GET_CODE (x) == LABEL_REF
986 || (GET_CODE (x) == CONST
987 && GET_CODE (XEXP (x, 0)) == PLUS
988 && (GET_CODE (XEXP (XEXP (x, 0), 0)) == LABEL_REF)))
989 recorded_label_ref = 1;
990
991 reg_equiv[regno].replacement = x;
992 reg_equiv[regno].src_p = &SET_SRC (set);
993 reg_equiv[regno].loop_depth = loop_depth;
994
995 /* Don't mess with things live during setjmp. */
996 if (REG_LIVE_LENGTH (regno) >= 0 && optimize)
997 {
998 /* Note that the statement below does not affect the priority
999 in local-alloc! */
1000 REG_LIVE_LENGTH (regno) *= 2;
1001
1002 /* If the register is referenced exactly twice, meaning it is
1003 set once and used once, indicate that the reference may be
1004 replaced by the equivalence we computed above. Do this
1005 even if the register is only used in one block so that
1006 dependencies can be handled where the last register is
1007 used in a different block (i.e. HIGH / LO_SUM sequences)
1008 and to reduce the number of registers alive across
1009 calls. */
1010
1011 if (REG_N_REFS (regno) == 2
1012 && (rtx_equal_p (x, src)
1013 || ! equiv_init_varies_p (src))
1014 && NONJUMP_INSN_P (insn)
1015 && equiv_init_movable_p (PATTERN (insn), regno))
1016 reg_equiv[regno].replace = 1;
1017 }
1018 }
1019 }
1020 }
1021
1022 if (!optimize)
1023 goto out;
1024
1025 /* A second pass, to gather additional equivalences with memory. This needs
1026 to be done after we know which registers we are going to replace. */
1027
1028 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
1029 {
1030 rtx set, src, dest;
1031 unsigned regno;
1032
1033 if (! INSN_P (insn))
1034 continue;
1035
1036 set = single_set (insn);
1037 if (! set)
1038 continue;
1039
1040 dest = SET_DEST (set);
1041 src = SET_SRC (set);
1042
1043 /* If this sets a MEM to the contents of a REG that is only used
1044 in a single basic block, see if the register is always equivalent
1045 to that memory location and if moving the store from INSN to the
1046 insn that set REG is safe. If so, put a REG_EQUIV note on the
1047 initializing insn.
1048
1049 Don't add a REG_EQUIV note if the insn already has one. The existing
1050 REG_EQUIV is likely more useful than the one we are adding.
1051
1052 If one of the regs in the address has reg_equiv[REGNO].replace set,
1053 then we can't add this REG_EQUIV note. The reg_equiv[REGNO].replace
1054 optimization may move the set of this register immediately before
1055 insn, which puts it after reg_equiv[REGNO].init_insns, and hence
1056 the mention in the REG_EQUIV note would be to an uninitialized
1057 pseudo. */
1058
1059 if (MEM_P (dest) && REG_P (src)
1060 && (regno = REGNO (src)) >= FIRST_PSEUDO_REGISTER
1061 && REG_BASIC_BLOCK (regno) >= NUM_FIXED_BLOCKS
1062 && DF_REG_DEF_COUNT (regno) == 1
1063 && reg_equiv[regno].init_insns != 0
1064 && reg_equiv[regno].init_insns != const0_rtx
1065 && ! find_reg_note (XEXP (reg_equiv[regno].init_insns, 0),
1066 REG_EQUIV, NULL_RTX)
1067 && ! contains_replace_regs (XEXP (dest, 0)))
1068 {
1069 rtx init_insn = XEXP (reg_equiv[regno].init_insns, 0);
1070 if (validate_equiv_mem (init_insn, src, dest)
1071 && ! memref_used_between_p (dest, init_insn, insn)
1072 /* Attaching a REG_EQUIV note will fail if INIT_INSN has
1073 multiple sets. */
1074 && set_unique_reg_note (init_insn, REG_EQUIV, copy_rtx (dest)))
1075 {
1076 /* This insn makes the equivalence, not the one initializing
1077 the register. */
1078 reg_equiv_init[regno]
1079 = gen_rtx_INSN_LIST (VOIDmode, insn, NULL_RTX);
1080 df_notes_rescan (init_insn);
1081 }
1082 }
1083 }
1084
1085 cleared_regs = BITMAP_ALLOC (NULL);
1086 /* Now scan all regs killed in an insn to see if any of them are
1087 registers only used that once. If so, see if we can replace the
1088 reference with the equivalent form. If we can, delete the
1089 initializing reference and this register will go away. If we
1090 can't replace the reference, and the initializing reference is
1091 within the same loop (or in an inner loop), then move the register
1092 initialization just before the use, so that they are in the same
1093 basic block. */
1094 FOR_EACH_BB_REVERSE (bb)
1095 {
1096 loop_depth = bb->loop_depth;
1097 for (insn = BB_END (bb);
1098 insn != PREV_INSN (BB_HEAD (bb));
1099 insn = PREV_INSN (insn))
1100 {
1101 rtx link;
1102
1103 if (! INSN_P (insn))
1104 continue;
1105
1106 /* Don't substitute into a non-local goto, this confuses CFG. */
1107 if (JUMP_P (insn)
1108 && find_reg_note (insn, REG_NON_LOCAL_GOTO, NULL_RTX))
1109 continue;
1110
1111 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1112 {
1113 if (REG_NOTE_KIND (link) == REG_DEAD
1114 /* Make sure this insn still refers to the register. */
1115 && reg_mentioned_p (XEXP (link, 0), PATTERN (insn)))
1116 {
1117 int regno = REGNO (XEXP (link, 0));
1118 rtx equiv_insn;
1119
1120 if (! reg_equiv[regno].replace
1121 || reg_equiv[regno].loop_depth < loop_depth)
1122 continue;
1123
1124 /* reg_equiv[REGNO].replace gets set only when
1125 REG_N_REFS[REGNO] is 2, i.e. the register is set
1126 once and used once. (If it were only set, but not used,
1127 flow would have deleted the setting insns.) Hence
1128 there can only be one insn in reg_equiv[REGNO].init_insns. */
1129 gcc_assert (reg_equiv[regno].init_insns
1130 && !XEXP (reg_equiv[regno].init_insns, 1));
1131 equiv_insn = XEXP (reg_equiv[regno].init_insns, 0);
1132
1133 /* We may not move instructions that can throw, since
1134 that changes basic block boundaries and we are not
1135 prepared to adjust the CFG to match. */
1136 if (can_throw_internal (equiv_insn))
1137 continue;
1138
1139 if (asm_noperands (PATTERN (equiv_insn)) < 0
1140 && validate_replace_rtx (regno_reg_rtx[regno],
1141 *(reg_equiv[regno].src_p), insn))
1142 {
1143 rtx equiv_link;
1144 rtx last_link;
1145 rtx note;
1146
1147 /* Find the last note. */
1148 for (last_link = link; XEXP (last_link, 1);
1149 last_link = XEXP (last_link, 1))
1150 ;
1151
1152 /* Append the REG_DEAD notes from equiv_insn. */
1153 equiv_link = REG_NOTES (equiv_insn);
1154 while (equiv_link)
1155 {
1156 note = equiv_link;
1157 equiv_link = XEXP (equiv_link, 1);
1158 if (REG_NOTE_KIND (note) == REG_DEAD)
1159 {
1160 remove_note (equiv_insn, note);
1161 XEXP (last_link, 1) = note;
1162 XEXP (note, 1) = NULL_RTX;
1163 last_link = note;
1164 }
1165 }
1166
1167 remove_death (regno, insn);
1168 SET_REG_N_REFS (regno, 0);
1169 REG_FREQ (regno) = 0;
1170 delete_insn (equiv_insn);
1171
1172 reg_equiv[regno].init_insns
1173 = XEXP (reg_equiv[regno].init_insns, 1);
1174
1175 reg_equiv_init[regno] = NULL_RTX;
1176 bitmap_set_bit (cleared_regs, regno);
1177 }
1178 /* Move the initialization of the register to just before
1179 INSN. Update the flow information. */
1180 else if (PREV_INSN (insn) != equiv_insn)
1181 {
1182 rtx new_insn;
1183
1184 new_insn = emit_insn_before (PATTERN (equiv_insn), insn);
1185 REG_NOTES (new_insn) = REG_NOTES (equiv_insn);
1186 REG_NOTES (equiv_insn) = 0;
1187 /* Rescan it to process the notes. */
1188 df_insn_rescan (new_insn);
1189
1190 /* Make sure this insn is recognized before
1191 reload begins, otherwise
1192 eliminate_regs_in_insn will die. */
1193 INSN_CODE (new_insn) = INSN_CODE (equiv_insn);
1194
1195 delete_insn (equiv_insn);
1196
1197 XEXP (reg_equiv[regno].init_insns, 0) = new_insn;
1198
1199 REG_BASIC_BLOCK (regno) = bb->index;
1200 REG_N_CALLS_CROSSED (regno) = 0;
1201 REG_FREQ_CALLS_CROSSED (regno) = 0;
1202 REG_N_THROWING_CALLS_CROSSED (regno) = 0;
1203 REG_LIVE_LENGTH (regno) = 2;
1204
1205 if (insn == BB_HEAD (bb))
1206 BB_HEAD (bb) = PREV_INSN (insn);
1207
1208 reg_equiv_init[regno]
1209 = gen_rtx_INSN_LIST (VOIDmode, new_insn, NULL_RTX);
1210 bitmap_set_bit (cleared_regs, regno);
1211 }
1212 }
1213 }
1214 }
1215 }
1216
1217 if (!bitmap_empty_p (cleared_regs))
1218 FOR_EACH_BB (bb)
1219 {
1220 bitmap_and_compl_into (DF_LIVE_IN (bb), cleared_regs);
1221 bitmap_and_compl_into (DF_LIVE_OUT (bb), cleared_regs);
1222 bitmap_and_compl_into (DF_LR_IN (bb), cleared_regs);
1223 bitmap_and_compl_into (DF_LR_OUT (bb), cleared_regs);
1224 }
1225
1226 BITMAP_FREE (cleared_regs);
1227
1228 out:
1229 /* Clean up. */
1230
1231 end_alias_analysis ();
1232 free (reg_equiv);
1233 return recorded_label_ref;
1234 }
1235
1236 /* Mark REG as having no known equivalence.
1237 Some instructions might have been processed before and furnished
1238 with REG_EQUIV notes for this register; these notes will have to be
1239 removed.
1240 STORE is the piece of RTL that does the non-constant / conflicting
1241 assignment - a SET, CLOBBER or REG_INC note. It is currently not used,
1242 but needs to be there because this function is called from note_stores. */
1243 static void
1244 no_equiv (rtx reg, const_rtx store ATTRIBUTE_UNUSED, void *data ATTRIBUTE_UNUSED)
1245 {
1246 int regno;
1247 rtx list;
1248
1249 if (!REG_P (reg))
1250 return;
1251 regno = REGNO (reg);
1252 list = reg_equiv[regno].init_insns;
1253 if (list == const0_rtx)
1254 return;
1255 reg_equiv[regno].init_insns = const0_rtx;
1256 reg_equiv[regno].replacement = NULL_RTX;
1257 /* This doesn't matter for equivalences made for argument registers, we
1258 should keep their initialization insns. */
1259 if (reg_equiv[regno].is_arg_equivalence)
1260 return;
1261 reg_equiv_init[regno] = NULL_RTX;
1262 for (; list; list = XEXP (list, 1))
1263 {
1264 rtx insn = XEXP (list, 0);
1265 remove_note (insn, find_reg_note (insn, REG_EQUIV, NULL_RTX));
1266 }
1267 }
1268 \f
1269 /* Allocate hard regs to the pseudo regs used only within block number B.
1270 Only the pseudos that die but once can be handled. */
1271
1272 static void
1273 block_alloc (basic_block b)
1274 {
1275 int i, q;
1276 rtx insn;
1277 rtx hard_reg;
1278 int insn_number = 0;
1279 int insn_count = 0;
1280 int max_uid = get_max_uid ();
1281 int *qty_order;
1282 struct df_ref ** def_rec;
1283
1284 /* Count the instructions in the basic block. */
1285
1286 insn = BB_END (b);
1287 while (1)
1288 {
1289 if (!NOTE_P (insn))
1290 {
1291 ++insn_count;
1292 gcc_assert (insn_count <= max_uid);
1293 }
1294 if (insn == BB_HEAD (b))
1295 break;
1296 insn = PREV_INSN (insn);
1297 }
1298
1299 /* +2 to leave room for a post_mark_life at the last insn and for
1300 the birth of a CLOBBER in the first insn. */
1301 regs_live_at = XCNEWVEC (HARD_REG_SET, 2 * insn_count + 2);
1302
1303 /* Initialize table of hardware registers currently live. */
1304
1305 REG_SET_TO_HARD_REG_SET (regs_live, DF_LR_IN (b));
1306
1307 /* This is conservative, as this would include registers that are
1308 artificial-def'ed-but-not-used. However, artificial-defs are
1309 rare, and such uninitialized use is rarer still, and the chance
1310 of this having any performance impact is even less, while the
1311 benefit is not having to compute and keep the TOP set around. */
1312 for (def_rec = df_get_artificial_defs (b->index); *def_rec; def_rec++)
1313 {
1314 int regno = DF_REF_REGNO (*def_rec);
1315 if (regno < FIRST_PSEUDO_REGISTER)
1316 SET_HARD_REG_BIT (regs_live, regno);
1317 }
1318
1319 /* This loop scans the instructions of the basic block
1320 and assigns quantities to registers.
1321 It computes which registers to tie. */
1322
1323 insn = BB_HEAD (b);
1324 while (1)
1325 {
1326 if (!NOTE_P (insn))
1327 insn_number++;
1328
1329 if (INSN_P (insn))
1330 {
1331 rtx link;
1332 int win = 0;
1333 rtx r0, r1 = NULL_RTX;
1334 int combined_regno = -1;
1335 int i;
1336
1337 this_insn_number = insn_number;
1338 this_insn = insn;
1339
1340 extract_insn (insn);
1341 which_alternative = -1;
1342
1343 /* Is this insn suitable for tying two registers?
1344 If so, try doing that.
1345 Suitable insns are those with at least two operands and where
1346 operand 0 is an output that is a register that is not
1347 earlyclobber.
1348
1349 We can tie operand 0 with some operand that dies in this insn.
1350 First look for operands that are required to be in the same
1351 register as operand 0. If we find such, only try tying that
1352 operand or one that can be put into that operand if the
1353 operation is commutative. If we don't find an operand
1354 that is required to be in the same register as operand 0,
1355 we can tie with any operand.
1356
1357 Subregs in place of regs are also ok.
1358
1359 If tying is done, WIN is set nonzero. */
1360
1361 if (optimize
1362 && recog_data.n_operands > 1
1363 && recog_data.constraints[0][0] == '='
1364 && recog_data.constraints[0][1] != '&')
1365 {
1366 /* If non-negative, is an operand that must match operand 0. */
1367 int must_match_0 = -1;
1368 /* Counts number of alternatives that require a match with
1369 operand 0. */
1370 int n_matching_alts = 0;
1371
1372 for (i = 1; i < recog_data.n_operands; i++)
1373 {
1374 const char *p = recog_data.constraints[i];
1375 int this_match = requires_inout (p);
1376
1377 n_matching_alts += this_match;
1378 if (this_match == recog_data.n_alternatives)
1379 must_match_0 = i;
1380 }
1381
1382 r0 = recog_data.operand[0];
1383 for (i = 1; i < recog_data.n_operands; i++)
1384 {
1385 /* Skip this operand if we found an operand that
1386 must match operand 0 and this operand isn't it
1387 and can't be made to be it by commutativity. */
1388
1389 if (must_match_0 >= 0 && i != must_match_0
1390 && ! (i == must_match_0 + 1
1391 && recog_data.constraints[i-1][0] == '%')
1392 && ! (i == must_match_0 - 1
1393 && recog_data.constraints[i][0] == '%'))
1394 continue;
1395
1396 /* Likewise if each alternative has some operand that
1397 must match operand zero. In that case, skip any
1398 operand that doesn't list operand 0 since we know that
1399 the operand always conflicts with operand 0. We
1400 ignore commutativity in this case to keep things simple. */
1401 if (n_matching_alts == recog_data.n_alternatives
1402 && 0 == requires_inout (recog_data.constraints[i]))
1403 continue;
1404
1405 r1 = recog_data.operand[i];
1406
1407 /* If the operand is an address, find a register in it.
1408 There may be more than one register, but we only try one
1409 of them. */
1410 if (recog_data.constraints[i][0] == 'p'
1411 || EXTRA_ADDRESS_CONSTRAINT (recog_data.constraints[i][0],
1412 recog_data.constraints[i]))
1413 while (GET_CODE (r1) == PLUS || GET_CODE (r1) == MULT)
1414 r1 = XEXP (r1, 0);
1415
1416 /* Avoid making a call-saved register unnecessarily
1417 clobbered. */
1418 hard_reg = get_hard_reg_initial_reg (r1);
1419 if (hard_reg != NULL_RTX)
1420 {
1421 if (REG_P (hard_reg)
1422 && REGNO (hard_reg) < FIRST_PSEUDO_REGISTER
1423 && !call_used_regs[REGNO (hard_reg)])
1424 continue;
1425 }
1426
1427 if (REG_P (r0) || GET_CODE (r0) == SUBREG)
1428 {
1429 /* We have two priorities for hard register preferences.
1430 If we have a move insn or an insn whose first input
1431 can only be in the same register as the output, give
1432 priority to an equivalence found from that insn. */
1433 int may_save_copy
1434 = (r1 == recog_data.operand[i] && must_match_0 >= 0);
1435
1436 if (REG_P (r1) || GET_CODE (r1) == SUBREG)
1437 win = combine_regs (r1, r0, may_save_copy,
1438 insn_number, insn);
1439 }
1440 if (win)
1441 break;
1442 }
1443 }
1444
1445 /* If registers were just tied, set COMBINED_REGNO
1446 to the number of the register used in this insn
1447 that was tied to the register set in this insn.
1448 This register's qty should not be "killed". */
1449
1450 if (win)
1451 {
1452 while (GET_CODE (r1) == SUBREG)
1453 r1 = SUBREG_REG (r1);
1454 combined_regno = REGNO (r1);
1455 }
1456
1457 /* Mark the death of everything that dies in this instruction. */
1458
1459 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1460 if (REG_NOTE_KIND (link) == REG_DEAD
1461 && REG_P (XEXP (link, 0))
1462 && combined_regno != (int) REGNO (XEXP (link, 0)))
1463 wipe_dead_reg (XEXP (link, 0), 0);
1464
1465 /* Allocate qty numbers for all registers local to this block
1466 that are born (set) in this instruction.
1467 A pseudo that already has a qty is not changed. */
1468
1469 note_stores (PATTERN (insn), reg_is_set, NULL);
1470
1471 /* If anything is set in this insn and then unused, mark it as dying
1472 after this insn, so it will conflict with our outputs. This
1473 can't match with something that combined, and it doesn't matter
1474 if it did. Do this after the calls to reg_is_set since these
1475 die after, not during, the current insn. */
1476
1477 for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
1478 if (REG_NOTE_KIND (link) == REG_UNUSED
1479 && REG_P (XEXP (link, 0)))
1480 wipe_dead_reg (XEXP (link, 0), 1);
1481 }
1482
1483 /* Set the registers live after INSN_NUMBER. Note that we never
1484 record the registers live before the block's first insn, since no
1485 pseudos we care about are live before that insn. */
1486
1487 IOR_HARD_REG_SET (regs_live_at[2 * insn_number], regs_live);
1488 IOR_HARD_REG_SET (regs_live_at[2 * insn_number + 1], regs_live);
1489
1490 if (insn == BB_END (b))
1491 break;
1492
1493 insn = NEXT_INSN (insn);
1494 }
1495
1496 /* Now every register that is local to this basic block
1497 should have been given a quantity, or else -1 meaning ignore it.
1498 Every quantity should have a known birth and death.
1499
1500 Order the qtys so we assign them registers in order of the
1501 number of suggested registers they need so we allocate those with
1502 the most restrictive needs first. */
1503
1504 qty_order = XNEWVEC (int, next_qty);
1505 for (i = 0; i < next_qty; i++)
1506 qty_order[i] = i;
1507
1508 #define EXCHANGE(I1, I2) \
1509 { i = qty_order[I1]; qty_order[I1] = qty_order[I2]; qty_order[I2] = i; }
1510
1511 switch (next_qty)
1512 {
1513 case 3:
1514 /* Make qty_order[2] be the one to allocate last. */
1515 if (qty_sugg_compare (0, 1) > 0)
1516 EXCHANGE (0, 1);
1517 if (qty_sugg_compare (1, 2) > 0)
1518 EXCHANGE (2, 1);
1519
1520 /* ... Fall through ... */
1521 case 2:
1522 /* Put the best one to allocate in qty_order[0]. */
1523 if (qty_sugg_compare (0, 1) > 0)
1524 EXCHANGE (0, 1);
1525
1526 /* ... Fall through ... */
1527
1528 case 1:
1529 case 0:
1530 /* Nothing to do here. */
1531 break;
1532
1533 default:
1534 qsort (qty_order, next_qty, sizeof (int), qty_sugg_compare_1);
1535 }
1536
1537 /* Try to put each quantity in a suggested physical register, if it has one.
1538 This may cause registers to be allocated that otherwise wouldn't be, but
1539 this seems acceptable in local allocation (unlike global allocation). */
1540 for (i = 0; i < next_qty; i++)
1541 {
1542 q = qty_order[i];
1543 if (qty_phys_num_sugg[q] != 0 || qty_phys_num_copy_sugg[q] != 0)
1544 qty[q].phys_reg = find_free_reg (qty[q].min_class, qty[q].mode, q,
1545 0, 1, qty[q].birth, qty[q].death, b);
1546 else
1547 qty[q].phys_reg = -1;
1548 }
1549
1550 /* Order the qtys so we assign them registers in order of
1551 decreasing length of life. Normally call qsort, but if we
1552 have only a very small number of quantities, sort them ourselves. */
1553
1554 for (i = 0; i < next_qty; i++)
1555 qty_order[i] = i;
1556
1557 #define EXCHANGE(I1, I2) \
1558 { i = qty_order[I1]; qty_order[I1] = qty_order[I2]; qty_order[I2] = i; }
1559
1560 switch (next_qty)
1561 {
1562 case 3:
1563 /* Make qty_order[2] be the one to allocate last. */
1564 if (qty_compare (0, 1) > 0)
1565 EXCHANGE (0, 1);
1566 if (qty_compare (1, 2) > 0)
1567 EXCHANGE (2, 1);
1568
1569 /* ... Fall through ... */
1570 case 2:
1571 /* Put the best one to allocate in qty_order[0]. */
1572 if (qty_compare (0, 1) > 0)
1573 EXCHANGE (0, 1);
1574
1575 /* ... Fall through ... */
1576
1577 case 1:
1578 case 0:
1579 /* Nothing to do here. */
1580 break;
1581
1582 default:
1583 qsort (qty_order, next_qty, sizeof (int), qty_compare_1);
1584 }
1585
1586 /* Now for each qty that is not a hardware register,
1587 look for a hardware register to put it in.
1588 First try the register class that is cheapest for this qty,
1589 if there is more than one class. */
1590
1591 for (i = 0; i < next_qty; i++)
1592 {
1593 q = qty_order[i];
1594 if (qty[q].phys_reg < 0)
1595 {
1596 #ifdef INSN_SCHEDULING
1597 /* These values represent the adjusted lifetime of a qty so
1598 that it conflicts with qtys which appear near the start/end
1599 of this qty's lifetime.
1600
1601 The purpose behind extending the lifetime of this qty is to
1602 discourage the register allocator from creating false
1603 dependencies.
1604
1605 The adjustment value is chosen to indicate that this qty
1606 conflicts with all the qtys in the instructions immediately
1607 before and after the lifetime of this qty.
1608
1609 Experiments have shown that higher values tend to hurt
1610 overall code performance.
1611
1612 If allocation using the extended lifetime fails we will try
1613 again with the qty's unadjusted lifetime. */
1614 int fake_birth = MAX (0, qty[q].birth - 2 + qty[q].birth % 2);
1615 int fake_death = MIN (insn_number * 2 + 1,
1616 qty[q].death + 2 - qty[q].death % 2);
1617 #endif
1618
1619 if (N_REG_CLASSES > 1)
1620 {
1621 #ifdef INSN_SCHEDULING
1622 /* We try to avoid using hard registers allocated to qtys which
1623 are born immediately after this qty or die immediately before
1624 this qty.
1625
1626 This optimization is only appropriate when we will run
1627 a scheduling pass after reload and we are not optimizing
1628 for code size. */
1629 if (flag_schedule_insns_after_reload && dbg_cnt (local_alloc_for_sched)
1630 && optimize_bb_for_speed_p (b)
1631 && !SMALL_REGISTER_CLASSES)
1632 {
1633 qty[q].phys_reg = find_free_reg (qty[q].min_class,
1634 qty[q].mode, q, 0, 0,
1635 fake_birth, fake_death, b);
1636 if (qty[q].phys_reg >= 0)
1637 continue;
1638 }
1639 #endif
1640 qty[q].phys_reg = find_free_reg (qty[q].min_class,
1641 qty[q].mode, q, 0, 0,
1642 qty[q].birth, qty[q].death, b);
1643 if (qty[q].phys_reg >= 0)
1644 continue;
1645 }
1646
1647 #ifdef INSN_SCHEDULING
1648 /* Similarly, avoid false dependencies. */
1649 if (flag_schedule_insns_after_reload && dbg_cnt (local_alloc_for_sched)
1650 && optimize_bb_for_speed_p (b)
1651 && !SMALL_REGISTER_CLASSES
1652 && qty[q].alternate_class != NO_REGS)
1653 qty[q].phys_reg = find_free_reg (qty[q].alternate_class,
1654 qty[q].mode, q, 0, 0,
1655 fake_birth, fake_death, b);
1656 #endif
1657 if (qty[q].alternate_class != NO_REGS)
1658 qty[q].phys_reg = find_free_reg (qty[q].alternate_class,
1659 qty[q].mode, q, 0, 0,
1660 qty[q].birth, qty[q].death, b);
1661 }
1662 }
1663
1664 /* Now propagate the register assignments
1665 to the pseudo regs belonging to the qtys. */
1666
1667 for (q = 0; q < next_qty; q++)
1668 if (qty[q].phys_reg >= 0)
1669 {
1670 for (i = qty[q].first_reg; i >= 0; i = reg_next_in_qty[i])
1671 reg_renumber[i] = qty[q].phys_reg + reg_offset[i];
1672 }
1673
1674 /* Clean up. */
1675 free (regs_live_at);
1676 free (qty_order);
1677 }
1678 \f
1679 /* Compare two quantities' priority for getting real registers.
1680 We give shorter-lived quantities higher priority.
1681 Quantities with more references are also preferred, as are quantities that
1682 require multiple registers. This is the identical prioritization as
1683 done by global-alloc.
1684
1685 We used to give preference to registers with *longer* lives, but using
1686 the same algorithm in both local- and global-alloc can speed up execution
1687 of some programs by as much as a factor of three! */
1688
1689 /* Note that the quotient will never be bigger than
1690 the value of floor_log2 times the maximum number of
1691 times a register can occur in one insn (surely less than 100)
1692 weighted by frequency (max REG_FREQ_MAX).
1693 Multiplying this by 10000/REG_FREQ_MAX can't overflow.
1694 QTY_CMP_PRI is also used by qty_sugg_compare. */
1695
1696 #define QTY_CMP_PRI(q) \
1697 ((int) (((double) (floor_log2 (qty[q].n_refs) * qty[q].freq * qty[q].size) \
1698 / (qty[q].death - qty[q].birth)) * (10000 / REG_FREQ_MAX)))
1699
1700 static int
1701 qty_compare (int q1, int q2)
1702 {
1703 return QTY_CMP_PRI (q2) - QTY_CMP_PRI (q1);
1704 }
1705
1706 static int
1707 qty_compare_1 (const void *q1p, const void *q2p)
1708 {
1709 int q1 = *(const int *) q1p, q2 = *(const int *) q2p;
1710 int tem = QTY_CMP_PRI (q2) - QTY_CMP_PRI (q1);
1711
1712 if (tem != 0)
1713 return tem;
1714
1715 /* If qtys are equally good, sort by qty number,
1716 so that the results of qsort leave nothing to chance. */
1717 return q1 - q2;
1718 }
1719 \f
1720 /* Compare two quantities' priority for getting real registers. This version
1721 is called for quantities that have suggested hard registers. First priority
1722 goes to quantities that have copy preferences, then to those that have
1723 normal preferences. Within those groups, quantities with the lower
1724 number of preferences have the highest priority. Of those, we use the same
1725 algorithm as above. */
1726
1727 #define QTY_CMP_SUGG(q) \
1728 (qty_phys_num_copy_sugg[q] \
1729 ? qty_phys_num_copy_sugg[q] \
1730 : qty_phys_num_sugg[q] * FIRST_PSEUDO_REGISTER)
1731
1732 static int
1733 qty_sugg_compare (int q1, int q2)
1734 {
1735 int tem = QTY_CMP_SUGG (q1) - QTY_CMP_SUGG (q2);
1736
1737 if (tem != 0)
1738 return tem;
1739
1740 return QTY_CMP_PRI (q2) - QTY_CMP_PRI (q1);
1741 }
1742
1743 static int
1744 qty_sugg_compare_1 (const void *q1p, const void *q2p)
1745 {
1746 int q1 = *(const int *) q1p, q2 = *(const int *) q2p;
1747 int tem = QTY_CMP_SUGG (q1) - QTY_CMP_SUGG (q2);
1748
1749 if (tem != 0)
1750 return tem;
1751
1752 tem = QTY_CMP_PRI (q2) - QTY_CMP_PRI (q1);
1753 if (tem != 0)
1754 return tem;
1755
1756 /* If qtys are equally good, sort by qty number,
1757 so that the results of qsort leave nothing to chance. */
1758 return q1 - q2;
1759 }
1760
1761 #undef QTY_CMP_SUGG
1762 #undef QTY_CMP_PRI
1763 \f
1764 /* Attempt to combine the two registers (rtx's) USEDREG and SETREG.
1765 Returns 1 if have done so, or 0 if cannot.
1766
1767 Combining registers means marking them as having the same quantity
1768 and adjusting the offsets within the quantity if either of
1769 them is a SUBREG.
1770
1771 We don't actually combine a hard reg with a pseudo; instead
1772 we just record the hard reg as the suggestion for the pseudo's quantity.
1773 If we really combined them, we could lose if the pseudo lives
1774 across an insn that clobbers the hard reg (eg, movmem).
1775
1776 MAY_SAVE_COPY is nonzero if this insn is simply copying USEDREG to
1777 SETREG or if the input and output must share a register.
1778 In that case, we record a hard reg suggestion in QTY_PHYS_COPY_SUGG.
1779
1780 There are elaborate checks for the validity of combining. */
1781
1782 static int
1783 combine_regs (rtx usedreg, rtx setreg, int may_save_copy, int insn_number,
1784 rtx insn)
1785 {
1786 int ureg, sreg;
1787 int offset = 0;
1788 int usize, ssize;
1789 int sqty;
1790
1791 /* Determine the numbers and sizes of registers being used. If a subreg
1792 is present that does not change the entire register, don't consider
1793 this a copy insn. */
1794
1795 while (GET_CODE (usedreg) == SUBREG)
1796 {
1797 rtx subreg = SUBREG_REG (usedreg);
1798
1799 if (REG_P (subreg))
1800 {
1801 if (GET_MODE_SIZE (GET_MODE (subreg)) > UNITS_PER_WORD)
1802 may_save_copy = 0;
1803
1804 if (REGNO (subreg) < FIRST_PSEUDO_REGISTER)
1805 offset += subreg_regno_offset (REGNO (subreg),
1806 GET_MODE (subreg),
1807 SUBREG_BYTE (usedreg),
1808 GET_MODE (usedreg));
1809 else
1810 offset += (SUBREG_BYTE (usedreg)
1811 / REGMODE_NATURAL_SIZE (GET_MODE (usedreg)));
1812 }
1813
1814 usedreg = subreg;
1815 }
1816
1817 if (!REG_P (usedreg))
1818 return 0;
1819
1820 ureg = REGNO (usedreg);
1821 if (ureg < FIRST_PSEUDO_REGISTER)
1822 usize = hard_regno_nregs[ureg][GET_MODE (usedreg)];
1823 else
1824 usize = ((GET_MODE_SIZE (GET_MODE (usedreg))
1825 + (REGMODE_NATURAL_SIZE (GET_MODE (usedreg)) - 1))
1826 / REGMODE_NATURAL_SIZE (GET_MODE (usedreg)));
1827
1828 while (GET_CODE (setreg) == SUBREG)
1829 {
1830 rtx subreg = SUBREG_REG (setreg);
1831
1832 if (REG_P (subreg))
1833 {
1834 if (GET_MODE_SIZE (GET_MODE (subreg)) > UNITS_PER_WORD)
1835 may_save_copy = 0;
1836
1837 if (REGNO (subreg) < FIRST_PSEUDO_REGISTER)
1838 offset -= subreg_regno_offset (REGNO (subreg),
1839 GET_MODE (subreg),
1840 SUBREG_BYTE (setreg),
1841 GET_MODE (setreg));
1842 else
1843 offset -= (SUBREG_BYTE (setreg)
1844 / REGMODE_NATURAL_SIZE (GET_MODE (setreg)));
1845 }
1846
1847 setreg = subreg;
1848 }
1849
1850 if (!REG_P (setreg))
1851 return 0;
1852
1853 sreg = REGNO (setreg);
1854 if (sreg < FIRST_PSEUDO_REGISTER)
1855 ssize = hard_regno_nregs[sreg][GET_MODE (setreg)];
1856 else
1857 ssize = ((GET_MODE_SIZE (GET_MODE (setreg))
1858 + (REGMODE_NATURAL_SIZE (GET_MODE (setreg)) - 1))
1859 / REGMODE_NATURAL_SIZE (GET_MODE (setreg)));
1860
1861 /* If UREG is a pseudo-register that hasn't already been assigned a
1862 quantity number, it means that it is not local to this block or dies
1863 more than once. In either event, we can't do anything with it. */
1864 if ((ureg >= FIRST_PSEUDO_REGISTER && reg_qty[ureg] < 0)
1865 /* Do not combine registers unless one fits within the other. */
1866 || (offset > 0 && usize + offset > ssize)
1867 || (offset < 0 && usize + offset < ssize)
1868 /* Do not combine with a smaller already-assigned object
1869 if that smaller object is already combined with something bigger. */
1870 || (ssize > usize && ureg >= FIRST_PSEUDO_REGISTER
1871 && usize < qty[reg_qty[ureg]].size)
1872 /* Can't combine if SREG is not a register we can allocate. */
1873 || (sreg >= FIRST_PSEUDO_REGISTER && reg_qty[sreg] == -1)
1874 /* Don't tie something to itself. In most cases it would make no
1875 difference, but it would screw up if the reg being tied to itself
1876 also dies in this insn. */
1877 || ureg == sreg
1878 /* Don't try to connect two different hardware registers. */
1879 || (ureg < FIRST_PSEUDO_REGISTER && sreg < FIRST_PSEUDO_REGISTER)
1880 /* Don't connect two different machine modes if they have different
1881 implications as to which registers may be used. */
1882 || !MODES_TIEABLE_P (GET_MODE (usedreg), GET_MODE (setreg)))
1883 return 0;
1884
1885 /* Now, if UREG is a hard reg and SREG is a pseudo, record the hard reg in
1886 qty_phys_sugg for the pseudo instead of tying them.
1887
1888 Return "failure" so that the lifespan of UREG is terminated here;
1889 that way the two lifespans will be disjoint and nothing will prevent
1890 the pseudo reg from being given this hard reg. */
1891
1892 if (ureg < FIRST_PSEUDO_REGISTER)
1893 {
1894 /* Allocate a quantity number so we have a place to put our
1895 suggestions. */
1896 if (reg_qty[sreg] == -2)
1897 reg_is_born (setreg, 2 * insn_number);
1898
1899 if (reg_qty[sreg] >= 0)
1900 {
1901 if (may_save_copy
1902 && ! TEST_HARD_REG_BIT (qty_phys_copy_sugg[reg_qty[sreg]], ureg))
1903 {
1904 SET_HARD_REG_BIT (qty_phys_copy_sugg[reg_qty[sreg]], ureg);
1905 qty_phys_num_copy_sugg[reg_qty[sreg]]++;
1906 }
1907 else if (! TEST_HARD_REG_BIT (qty_phys_sugg[reg_qty[sreg]], ureg))
1908 {
1909 SET_HARD_REG_BIT (qty_phys_sugg[reg_qty[sreg]], ureg);
1910 qty_phys_num_sugg[reg_qty[sreg]]++;
1911 }
1912 }
1913 return 0;
1914 }
1915
1916 /* Similarly for SREG a hard register and UREG a pseudo register. */
1917
1918 if (sreg < FIRST_PSEUDO_REGISTER)
1919 {
1920 if (may_save_copy
1921 && ! TEST_HARD_REG_BIT (qty_phys_copy_sugg[reg_qty[ureg]], sreg))
1922 {
1923 SET_HARD_REG_BIT (qty_phys_copy_sugg[reg_qty[ureg]], sreg);
1924 qty_phys_num_copy_sugg[reg_qty[ureg]]++;
1925 }
1926 else if (! TEST_HARD_REG_BIT (qty_phys_sugg[reg_qty[ureg]], sreg))
1927 {
1928 SET_HARD_REG_BIT (qty_phys_sugg[reg_qty[ureg]], sreg);
1929 qty_phys_num_sugg[reg_qty[ureg]]++;
1930 }
1931 return 0;
1932 }
1933
1934 /* At this point we know that SREG and UREG are both pseudos.
1935 Do nothing if SREG already has a quantity or is a register that we
1936 don't allocate. */
1937 if (reg_qty[sreg] >= -1
1938 /* If we are not going to let any regs live across calls,
1939 don't tie a call-crossing reg to a non-call-crossing reg. */
1940 || (cfun->has_nonlocal_label
1941 && ((REG_N_CALLS_CROSSED (ureg) > 0)
1942 != (REG_N_CALLS_CROSSED (sreg) > 0))))
1943 return 0;
1944
1945 /* We don't already know about SREG, so tie it to UREG
1946 if this is the last use of UREG, provided the classes they want
1947 are compatible. */
1948
1949 if (find_regno_note (insn, REG_DEAD, ureg)
1950 && reg_meets_class_p (sreg, qty[reg_qty[ureg]].min_class))
1951 {
1952 /* Add SREG to UREG's quantity. */
1953 sqty = reg_qty[ureg];
1954 reg_qty[sreg] = sqty;
1955 reg_offset[sreg] = reg_offset[ureg] + offset;
1956 reg_next_in_qty[sreg] = qty[sqty].first_reg;
1957 qty[sqty].first_reg = sreg;
1958
1959 /* If SREG's reg class is smaller, set qty[SQTY].min_class. */
1960 update_qty_class (sqty, sreg);
1961
1962 /* Update info about quantity SQTY. */
1963 qty[sqty].n_calls_crossed += REG_N_CALLS_CROSSED (sreg);
1964 qty[sqty].freq_calls_crossed += REG_FREQ_CALLS_CROSSED (sreg);
1965 qty[sqty].n_throwing_calls_crossed
1966 += REG_N_THROWING_CALLS_CROSSED (sreg);
1967 qty[sqty].n_refs += REG_N_REFS (sreg);
1968 qty[sqty].freq += REG_FREQ (sreg);
1969 if (usize < ssize)
1970 {
1971 int i;
1972
1973 for (i = qty[sqty].first_reg; i >= 0; i = reg_next_in_qty[i])
1974 reg_offset[i] -= offset;
1975
1976 qty[sqty].size = ssize;
1977 qty[sqty].mode = GET_MODE (setreg);
1978 }
1979 }
1980 else
1981 return 0;
1982
1983 return 1;
1984 }
1985 \f
1986 /* Return 1 if the preferred class of REG allows it to be tied
1987 to a quantity or register whose class is CLASS.
1988 True if REG's reg class either contains or is contained in CLASS. */
1989
1990 static int
1991 reg_meets_class_p (int reg, enum reg_class rclass)
1992 {
1993 enum reg_class rclass2 = reg_preferred_class (reg);
1994 return (reg_class_subset_p (rclass2, rclass)
1995 || reg_class_subset_p (rclass, rclass2));
1996 }
1997
1998 /* Update the class of QTYNO assuming that REG is being tied to it. */
1999
2000 static void
2001 update_qty_class (int qtyno, int reg)
2002 {
2003 enum reg_class rclass = reg_preferred_class (reg);
2004 if (reg_class_subset_p (rclass, qty[qtyno].min_class))
2005 qty[qtyno].min_class = rclass;
2006
2007 rclass = reg_alternate_class (reg);
2008 if (reg_class_subset_p (rclass, qty[qtyno].alternate_class))
2009 qty[qtyno].alternate_class = rclass;
2010 }
2011 \f
2012 /* Handle something which alters the value of an rtx REG.
2013
2014 REG is whatever is set or clobbered. SETTER is the rtx that
2015 is modifying the register.
2016
2017 If it is not really a register, we do nothing.
2018 The file-global variables `this_insn' and `this_insn_number'
2019 carry info from `block_alloc'. */
2020
2021 static void
2022 reg_is_set (rtx reg, const_rtx setter, void *data ATTRIBUTE_UNUSED)
2023 {
2024 /* Note that note_stores will only pass us a SUBREG if it is a SUBREG of
2025 a hard register. These may actually not exist any more. */
2026
2027 if (GET_CODE (reg) != SUBREG
2028 && !REG_P (reg))
2029 return;
2030
2031 /* Mark this register as being born. If it is used in a CLOBBER, mark
2032 it as being born halfway between the previous insn and this insn so that
2033 it conflicts with our inputs but not the outputs of the previous insn. */
2034
2035 reg_is_born (reg, 2 * this_insn_number - (GET_CODE (setter) == CLOBBER));
2036 }
2037 \f
2038 /* Handle beginning of the life of register REG.
2039 BIRTH is the index at which this is happening. */
2040
2041 static void
2042 reg_is_born (rtx reg, int birth)
2043 {
2044 int regno;
2045
2046 if (GET_CODE (reg) == SUBREG)
2047 {
2048 regno = REGNO (SUBREG_REG (reg));
2049 if (regno < FIRST_PSEUDO_REGISTER)
2050 regno = subreg_regno (reg);
2051 }
2052 else
2053 regno = REGNO (reg);
2054
2055 if (regno < FIRST_PSEUDO_REGISTER)
2056 {
2057 mark_life (regno, GET_MODE (reg), 1);
2058
2059 /* If the register was to have been born earlier that the present
2060 insn, mark it as live where it is actually born. */
2061 if (birth < 2 * this_insn_number)
2062 post_mark_life (regno, GET_MODE (reg), 1, birth, 2 * this_insn_number);
2063 }
2064 else
2065 {
2066 if (reg_qty[regno] == -2)
2067 alloc_qty (regno, GET_MODE (reg), PSEUDO_REGNO_SIZE (regno), birth);
2068
2069 /* If this register has a quantity number, show that it isn't dead. */
2070 if (reg_qty[regno] >= 0)
2071 qty[reg_qty[regno]].death = -1;
2072 }
2073 }
2074
2075 /* Record the death of REG in the current insn. If OUTPUT_P is nonzero,
2076 REG is an output that is dying (i.e., it is never used), otherwise it
2077 is an input (the normal case).
2078 If OUTPUT_P is 1, then we extend the life past the end of this insn. */
2079
2080 static void
2081 wipe_dead_reg (rtx reg, int output_p)
2082 {
2083 int regno = REGNO (reg);
2084
2085 /* If this insn has multiple results,
2086 and the dead reg is used in one of the results,
2087 extend its life to after this insn,
2088 so it won't get allocated together with any other result of this insn.
2089
2090 It is unsafe to use !single_set here since it will ignore an unused
2091 output. Just because an output is unused does not mean the compiler
2092 can assume the side effect will not occur. Consider if REG appears
2093 in the address of an output and we reload the output. If we allocate
2094 REG to the same hard register as an unused output we could set the hard
2095 register before the output reload insn. */
2096 if (GET_CODE (PATTERN (this_insn)) == PARALLEL
2097 && multiple_sets (this_insn))
2098 {
2099 int i;
2100 for (i = XVECLEN (PATTERN (this_insn), 0) - 1; i >= 0; i--)
2101 {
2102 rtx set = XVECEXP (PATTERN (this_insn), 0, i);
2103 if (GET_CODE (set) == SET
2104 && !REG_P (SET_DEST (set))
2105 && !rtx_equal_p (reg, SET_DEST (set))
2106 && reg_overlap_mentioned_p (reg, SET_DEST (set)))
2107 output_p = 1;
2108 }
2109 }
2110
2111 /* If this register is used in an auto-increment address, then extend its
2112 life to after this insn, so that it won't get allocated together with
2113 the result of this insn. */
2114 if (! output_p && find_regno_note (this_insn, REG_INC, regno))
2115 output_p = 1;
2116
2117 if (regno < FIRST_PSEUDO_REGISTER)
2118 {
2119 mark_life (regno, GET_MODE (reg), 0);
2120
2121 /* If a hard register is dying as an output, mark it as in use at
2122 the beginning of this insn (the above statement would cause this
2123 not to happen). */
2124 if (output_p)
2125 post_mark_life (regno, GET_MODE (reg), 1,
2126 2 * this_insn_number, 2 * this_insn_number + 1);
2127 }
2128
2129 else if (reg_qty[regno] >= 0)
2130 qty[reg_qty[regno]].death = 2 * this_insn_number + output_p;
2131 }
2132 \f
2133 /* Find a block of SIZE words of hard regs in reg_class CLASS
2134 that can hold something of machine-mode MODE
2135 (but actually we test only the first of the block for holding MODE)
2136 and still free between insn BORN_INDEX and insn DEAD_INDEX,
2137 and return the number of the first of them.
2138 Return -1 if such a block cannot be found.
2139 If QTYNO crosses calls, insist on a register preserved by calls,
2140 unless ACCEPT_CALL_CLOBBERED is nonzero.
2141
2142 If JUST_TRY_SUGGESTED is nonzero, only try to see if the suggested
2143 register is available. If not, return -1. */
2144
2145 static int
2146 find_free_reg (enum reg_class rclass, enum machine_mode mode, int qtyno,
2147 int accept_call_clobbered, int just_try_suggested,
2148 int born_index, int dead_index, basic_block bb)
2149 {
2150 int i, ins;
2151 HARD_REG_SET first_used, used;
2152 #ifdef ELIMINABLE_REGS
2153 static const struct {const int from, to; } eliminables[] = ELIMINABLE_REGS;
2154 #endif
2155
2156 /* Validate our parameters. */
2157 gcc_assert (born_index >= 0 && born_index <= dead_index);
2158
2159 /* Don't let a pseudo live in a reg across a function call
2160 if we might get a nonlocal goto. */
2161 if (cfun->has_nonlocal_label
2162 && qty[qtyno].n_calls_crossed > 0)
2163 return -1;
2164
2165 if (accept_call_clobbered)
2166 COPY_HARD_REG_SET (used, call_fixed_reg_set);
2167 else if (qty[qtyno].n_calls_crossed == 0)
2168 COPY_HARD_REG_SET (used, fixed_reg_set);
2169 else
2170 COPY_HARD_REG_SET (used, call_used_reg_set);
2171
2172 if (accept_call_clobbered)
2173 IOR_HARD_REG_SET (used, losing_caller_save_reg_set);
2174
2175 for (ins = born_index; ins < dead_index; ins++)
2176 IOR_HARD_REG_SET (used, regs_live_at[ins]);
2177
2178 IOR_COMPL_HARD_REG_SET (used, reg_class_contents[(int) rclass]);
2179
2180 /* Don't use the frame pointer reg in local-alloc even if
2181 we may omit the frame pointer, because if we do that and then we
2182 need a frame pointer, reload won't know how to move the pseudo
2183 to another hard reg. It can move only regs made by global-alloc.
2184
2185 This is true of any register that can be eliminated. */
2186 #ifdef ELIMINABLE_REGS
2187 for (i = 0; i < (int) ARRAY_SIZE (eliminables); i++)
2188 SET_HARD_REG_BIT (used, eliminables[i].from);
2189 #if FRAME_POINTER_REGNUM != HARD_FRAME_POINTER_REGNUM
2190 /* If FRAME_POINTER_REGNUM is not a real register, then protect the one
2191 that it might be eliminated into. */
2192 SET_HARD_REG_BIT (used, HARD_FRAME_POINTER_REGNUM);
2193 #endif
2194 #else
2195 SET_HARD_REG_BIT (used, FRAME_POINTER_REGNUM);
2196 #endif
2197
2198 #ifdef CANNOT_CHANGE_MODE_CLASS
2199 cannot_change_mode_set_regs (&used, mode, qty[qtyno].first_reg);
2200 #endif
2201
2202 /* Normally, the registers that can be used for the first register in
2203 a multi-register quantity are the same as those that can be used for
2204 subsequent registers. However, if just trying suggested registers,
2205 restrict our consideration to them. If there are copy-suggested
2206 register, try them. Otherwise, try the arithmetic-suggested
2207 registers. */
2208 COPY_HARD_REG_SET (first_used, used);
2209
2210 if (just_try_suggested)
2211 {
2212 if (qty_phys_num_copy_sugg[qtyno] != 0)
2213 IOR_COMPL_HARD_REG_SET (first_used, qty_phys_copy_sugg[qtyno]);
2214 else
2215 IOR_COMPL_HARD_REG_SET (first_used, qty_phys_sugg[qtyno]);
2216 }
2217
2218 /* If at least one would be suitable, test each hard reg. */
2219 if (!hard_reg_set_subset_p (reg_class_contents[(int) ALL_REGS], first_used))
2220 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
2221 {
2222 #ifdef REG_ALLOC_ORDER
2223 int regno = reg_alloc_order[i];
2224 #else
2225 int regno = i;
2226 #endif
2227 if (!TEST_HARD_REG_BIT (first_used, regno)
2228 && HARD_REGNO_MODE_OK (regno, mode)
2229 && (qty[qtyno].n_calls_crossed == 0
2230 || accept_call_clobbered
2231 || !HARD_REGNO_CALL_PART_CLOBBERED (regno, mode)))
2232 {
2233 int j;
2234 int size1 = hard_regno_nregs[regno][mode];
2235 j = 1;
2236 while (j < size1 && !TEST_HARD_REG_BIT (used, regno + j))
2237 j++;
2238 if (j == size1)
2239 {
2240 /* Mark that this register is in use between its birth
2241 and death insns. */
2242 post_mark_life (regno, mode, 1, born_index, dead_index);
2243 return regno;
2244 }
2245 #ifndef REG_ALLOC_ORDER
2246 /* Skip starting points we know will lose. */
2247 i += j;
2248 #endif
2249 }
2250 }
2251
2252 /* If we are just trying suggested register, we have just tried copy-
2253 suggested registers, and there are arithmetic-suggested registers,
2254 try them. */
2255
2256 /* If it would be profitable to allocate a call-clobbered register
2257 and save and restore it around calls, do that. */
2258 if (just_try_suggested && qty_phys_num_copy_sugg[qtyno] != 0
2259 && qty_phys_num_sugg[qtyno] != 0)
2260 {
2261 /* Don't try the copy-suggested regs again. */
2262 qty_phys_num_copy_sugg[qtyno] = 0;
2263 return find_free_reg (rclass, mode, qtyno, accept_call_clobbered, 1,
2264 born_index, dead_index, bb);
2265 }
2266
2267 /* We need not check to see if the current function has nonlocal
2268 labels because we don't put any pseudos that are live over calls in
2269 registers in that case. Avoid putting pseudos crossing calls that
2270 might throw into call used registers. */
2271
2272 if (! accept_call_clobbered
2273 && flag_caller_saves
2274 && ! just_try_suggested
2275 && qty[qtyno].n_calls_crossed != 0
2276 && qty[qtyno].n_throwing_calls_crossed == 0
2277 && CALLER_SAVE_PROFITABLE (optimize_bb_for_size_p (bb) ? qty[qtyno].n_refs
2278 : qty[qtyno].freq,
2279 optimize_bb_for_size_p (bb) ? qty[qtyno].n_calls_crossed
2280 : qty[qtyno].freq_calls_crossed))
2281 {
2282 i = find_free_reg (rclass, mode, qtyno, 1, 0, born_index, dead_index, bb);
2283 if (i >= 0)
2284 caller_save_needed = 1;
2285 return i;
2286 }
2287 return -1;
2288 }
2289 \f
2290 /* Mark that REGNO with machine-mode MODE is live starting from the current
2291 insn (if LIFE is nonzero) or dead starting at the current insn (if LIFE
2292 is zero). */
2293
2294 static void
2295 mark_life (int regno, enum machine_mode mode, int life)
2296 {
2297 if (life)
2298 add_to_hard_reg_set (&regs_live, mode, regno);
2299 else
2300 remove_from_hard_reg_set (&regs_live, mode, regno);
2301 }
2302
2303 /* Mark register number REGNO (with machine-mode MODE) as live (if LIFE
2304 is nonzero) or dead (if LIFE is zero) from insn number BIRTH (inclusive)
2305 to insn number DEATH (exclusive). */
2306
2307 static void
2308 post_mark_life (int regno, enum machine_mode mode, int life, int birth,
2309 int death)
2310 {
2311 HARD_REG_SET this_reg;
2312
2313 CLEAR_HARD_REG_SET (this_reg);
2314 add_to_hard_reg_set (&this_reg, mode, regno);
2315
2316 if (life)
2317 while (birth < death)
2318 {
2319 IOR_HARD_REG_SET (regs_live_at[birth], this_reg);
2320 birth++;
2321 }
2322 else
2323 while (birth < death)
2324 {
2325 AND_COMPL_HARD_REG_SET (regs_live_at[birth], this_reg);
2326 birth++;
2327 }
2328 }
2329 \f
2330 /* Return the number of alternatives for which the constraint string P
2331 indicates that the operand must be equal to operand 0 and that no register
2332 is acceptable. */
2333
2334 static int
2335 requires_inout (const char *p)
2336 {
2337 char c;
2338 int found_zero = 0;
2339 int reg_allowed = 0;
2340 int num_matching_alts = 0;
2341 int len;
2342
2343 for ( ; (c = *p); p += len)
2344 {
2345 len = CONSTRAINT_LEN (c, p);
2346 switch (c)
2347 {
2348 case '=': case '+': case '?':
2349 case '#': case '&': case '!':
2350 case '*': case '%':
2351 case 'm': case '<': case '>': case 'V': case 'o':
2352 case 'E': case 'F': case 'G': case 'H':
2353 case 's': case 'i': case 'n':
2354 case 'I': case 'J': case 'K': case 'L':
2355 case 'M': case 'N': case 'O': case 'P':
2356 case 'X':
2357 /* These don't say anything we care about. */
2358 break;
2359
2360 case ',':
2361 if (found_zero && ! reg_allowed)
2362 num_matching_alts++;
2363
2364 found_zero = reg_allowed = 0;
2365 break;
2366
2367 case '0':
2368 found_zero = 1;
2369 break;
2370
2371 case '1': case '2': case '3': case '4': case '5':
2372 case '6': case '7': case '8': case '9':
2373 /* Skip the balance of the matching constraint. */
2374 do
2375 p++;
2376 while (ISDIGIT (*p));
2377 len = 0;
2378 break;
2379
2380 default:
2381 if (REG_CLASS_FROM_CONSTRAINT (c, p) == NO_REGS
2382 && !EXTRA_ADDRESS_CONSTRAINT (c, p))
2383 break;
2384 /* Fall through. */
2385 case 'p':
2386 case 'g': case 'r':
2387 reg_allowed = 1;
2388 break;
2389 }
2390 }
2391
2392 if (found_zero && ! reg_allowed)
2393 num_matching_alts++;
2394
2395 return num_matching_alts;
2396 }
2397 \f
2398 void
2399 dump_local_alloc (FILE *file)
2400 {
2401 int i;
2402 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
2403 if (reg_renumber[i] != -1)
2404 fprintf (file, ";; Register %d in %d.\n", i, reg_renumber[i]);
2405 }
2406
2407 #ifdef STACK_REGS
2408 static void
2409 find_stack_regs (void)
2410 {
2411 bitmap stack_regs = BITMAP_ALLOC (NULL);
2412 int i;
2413 HARD_REG_SET stack_hard_regs, used;
2414 basic_block bb;
2415
2416 /* Any register that MAY be allocated to a register stack (like the
2417 387) is treated poorly. Each such register is marked as being
2418 live everywhere. This keeps the register allocator and the
2419 subsequent passes from doing anything useful with these values.
2420
2421 FIXME: This seems like an incredibly poor idea. */
2422
2423 CLEAR_HARD_REG_SET (stack_hard_regs);
2424 for (i = FIRST_STACK_REG; i <= LAST_STACK_REG; i++)
2425 SET_HARD_REG_BIT (stack_hard_regs, i);
2426
2427 for (i = FIRST_PSEUDO_REGISTER; i < max_regno; i++)
2428 {
2429 COPY_HARD_REG_SET (used, reg_class_contents[reg_preferred_class (i)]);
2430 IOR_HARD_REG_SET (used, reg_class_contents[reg_alternate_class (i)]);
2431 AND_HARD_REG_SET (used, stack_hard_regs);
2432 if (!hard_reg_set_empty_p (used))
2433 bitmap_set_bit (stack_regs, i);
2434 }
2435
2436 if (dump_file)
2437 bitmap_print (dump_file, stack_regs, "stack regs:", "\n");
2438
2439 FOR_EACH_BB (bb)
2440 {
2441 bitmap_ior_into (DF_LIVE_IN (bb), stack_regs);
2442 bitmap_and_into (DF_LIVE_IN (bb), DF_LR_IN (bb));
2443 bitmap_ior_into (DF_LIVE_OUT (bb), stack_regs);
2444 bitmap_and_into (DF_LIVE_OUT (bb), DF_LR_OUT (bb));
2445 }
2446 BITMAP_FREE (stack_regs);
2447 }
2448 #endif
2449
2450 static bool
2451 gate_handle_local_alloc (void)
2452 {
2453 return ! flag_ira;
2454 }
2455
2456 /* Run old register allocator. Return TRUE if we must exit
2457 rest_of_compilation upon return. */
2458 static unsigned int
2459 rest_of_handle_local_alloc (void)
2460 {
2461 int rebuild_notes;
2462 int max_regno = max_reg_num ();
2463
2464 df_note_add_problem ();
2465
2466 if (optimize == 1)
2467 {
2468 df_live_add_problem ();
2469 df_live_set_all_dirty ();
2470 }
2471 #ifdef ENABLE_CHECKING
2472 df->changeable_flags |= DF_VERIFY_SCHEDULED;
2473 #endif
2474 df_analyze ();
2475 #ifdef STACK_REGS
2476 if (optimize)
2477 find_stack_regs ();
2478 #endif
2479 regstat_init_n_sets_and_refs ();
2480 regstat_compute_ri ();
2481
2482 /* If we are not optimizing, then this is the only place before
2483 register allocation where dataflow is done. And that is needed
2484 to generate these warnings. */
2485 if (warn_clobbered)
2486 generate_setjmp_warnings ();
2487
2488 /* Determine if the current function is a leaf before running reload
2489 since this can impact optimizations done by the prologue and
2490 epilogue thus changing register elimination offsets. */
2491 current_function_is_leaf = leaf_function_p ();
2492
2493 /* And the reg_equiv_memory_loc array. */
2494 VEC_safe_grow (rtx, gc, reg_equiv_memory_loc_vec, max_regno);
2495 memset (VEC_address (rtx, reg_equiv_memory_loc_vec), 0,
2496 sizeof (rtx) * max_regno);
2497 reg_equiv_memory_loc = VEC_address (rtx, reg_equiv_memory_loc_vec);
2498
2499 allocate_initial_values (reg_equiv_memory_loc);
2500
2501 regclass (get_insns (), max_regno);
2502 rebuild_notes = local_alloc ();
2503
2504 /* Local allocation may have turned an indirect jump into a direct
2505 jump. If so, we must rebuild the JUMP_LABEL fields of jumping
2506 instructions. */
2507 if (rebuild_notes)
2508 {
2509 timevar_push (TV_JUMP);
2510
2511 rebuild_jump_labels (get_insns ());
2512 purge_all_dead_edges ();
2513 timevar_pop (TV_JUMP);
2514 }
2515
2516 if (dump_file && (dump_flags & TDF_DETAILS))
2517 {
2518 timevar_push (TV_DUMP);
2519 dump_flow_info (dump_file, dump_flags);
2520 dump_local_alloc (dump_file);
2521 timevar_pop (TV_DUMP);
2522 }
2523 return 0;
2524 }
2525
2526 struct rtl_opt_pass pass_local_alloc =
2527 {
2528 {
2529 RTL_PASS,
2530 "lreg", /* name */
2531 gate_handle_local_alloc, /* gate */
2532 rest_of_handle_local_alloc, /* execute */
2533 NULL, /* sub */
2534 NULL, /* next */
2535 0, /* static_pass_number */
2536 TV_LOCAL_ALLOC, /* tv_id */
2537 0, /* properties_required */
2538 0, /* properties_provided */
2539 0, /* properties_destroyed */
2540 0, /* todo_flags_start */
2541 TODO_dump_func |
2542 TODO_ggc_collect /* todo_flags_finish */
2543 }
2544 };
2545