1 /* Subroutines used for code generation on IBM S/390 and zSeries
2 Copyright (C) 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006,
3 2007, 2008, 2009, 2010 Free Software Foundation, Inc.
4 Contributed by Hartmut Penner (hpenner@de.ibm.com) and
5 Ulrich Weigand (uweigand@de.ibm.com) and
6 Andreas Krebbel (Andreas.Krebbel@de.ibm.com).
8 This file is part of GCC.
10 GCC is free software; you can redistribute it and/or modify it under
11 the terms of the GNU General Public License as published by the Free
12 Software Foundation; either version 3, or (at your option) any later
15 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
16 WARRANTY; without even the implied warranty of MERCHANTABILITY or
17 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
20 You should have received a copy of the GNU General Public License
21 along with GCC; see the file COPYING3. If not see
22 <http://www.gnu.org/licenses/>. */
26 #include "coretypes.h"
32 #include "hard-reg-set.h"
34 #include "insn-config.h"
35 #include "conditions.h"
37 #include "insn-attr.h"
45 #include "basic-block.h"
46 #include "integrate.h"
49 #include "target-def.h"
51 #include "langhooks.h"
59 /* Define the specific costs for a given cpu. */
61 struct processor_costs
64 const int m
; /* cost of an M instruction. */
65 const int mghi
; /* cost of an MGHI instruction. */
66 const int mh
; /* cost of an MH instruction. */
67 const int mhi
; /* cost of an MHI instruction. */
68 const int ml
; /* cost of an ML instruction. */
69 const int mr
; /* cost of an MR instruction. */
70 const int ms
; /* cost of an MS instruction. */
71 const int msg
; /* cost of an MSG instruction. */
72 const int msgf
; /* cost of an MSGF instruction. */
73 const int msgfr
; /* cost of an MSGFR instruction. */
74 const int msgr
; /* cost of an MSGR instruction. */
75 const int msr
; /* cost of an MSR instruction. */
76 const int mult_df
; /* cost of multiplication in DFmode. */
79 const int sqxbr
; /* cost of square root in TFmode. */
80 const int sqdbr
; /* cost of square root in DFmode. */
81 const int sqebr
; /* cost of square root in SFmode. */
82 /* multiply and add */
83 const int madbr
; /* cost of multiply and add in DFmode. */
84 const int maebr
; /* cost of multiply and add in SFmode. */
96 const struct processor_costs
*s390_cost
;
99 struct processor_costs z900_cost
=
101 COSTS_N_INSNS (5), /* M */
102 COSTS_N_INSNS (10), /* MGHI */
103 COSTS_N_INSNS (5), /* MH */
104 COSTS_N_INSNS (4), /* MHI */
105 COSTS_N_INSNS (5), /* ML */
106 COSTS_N_INSNS (5), /* MR */
107 COSTS_N_INSNS (4), /* MS */
108 COSTS_N_INSNS (15), /* MSG */
109 COSTS_N_INSNS (7), /* MSGF */
110 COSTS_N_INSNS (7), /* MSGFR */
111 COSTS_N_INSNS (10), /* MSGR */
112 COSTS_N_INSNS (4), /* MSR */
113 COSTS_N_INSNS (7), /* multiplication in DFmode */
114 COSTS_N_INSNS (13), /* MXBR */
115 COSTS_N_INSNS (136), /* SQXBR */
116 COSTS_N_INSNS (44), /* SQDBR */
117 COSTS_N_INSNS (35), /* SQEBR */
118 COSTS_N_INSNS (18), /* MADBR */
119 COSTS_N_INSNS (13), /* MAEBR */
120 COSTS_N_INSNS (134), /* DXBR */
121 COSTS_N_INSNS (30), /* DDBR */
122 COSTS_N_INSNS (27), /* DEBR */
123 COSTS_N_INSNS (220), /* DLGR */
124 COSTS_N_INSNS (34), /* DLR */
125 COSTS_N_INSNS (34), /* DR */
126 COSTS_N_INSNS (32), /* DSGFR */
127 COSTS_N_INSNS (32), /* DSGR */
131 struct processor_costs z990_cost
=
133 COSTS_N_INSNS (4), /* M */
134 COSTS_N_INSNS (2), /* MGHI */
135 COSTS_N_INSNS (2), /* MH */
136 COSTS_N_INSNS (2), /* MHI */
137 COSTS_N_INSNS (4), /* ML */
138 COSTS_N_INSNS (4), /* MR */
139 COSTS_N_INSNS (5), /* MS */
140 COSTS_N_INSNS (6), /* MSG */
141 COSTS_N_INSNS (4), /* MSGF */
142 COSTS_N_INSNS (4), /* MSGFR */
143 COSTS_N_INSNS (4), /* MSGR */
144 COSTS_N_INSNS (4), /* MSR */
145 COSTS_N_INSNS (1), /* multiplication in DFmode */
146 COSTS_N_INSNS (28), /* MXBR */
147 COSTS_N_INSNS (130), /* SQXBR */
148 COSTS_N_INSNS (66), /* SQDBR */
149 COSTS_N_INSNS (38), /* SQEBR */
150 COSTS_N_INSNS (1), /* MADBR */
151 COSTS_N_INSNS (1), /* MAEBR */
152 COSTS_N_INSNS (60), /* DXBR */
153 COSTS_N_INSNS (40), /* DDBR */
154 COSTS_N_INSNS (26), /* DEBR */
155 COSTS_N_INSNS (176), /* DLGR */
156 COSTS_N_INSNS (31), /* DLR */
157 COSTS_N_INSNS (31), /* DR */
158 COSTS_N_INSNS (31), /* DSGFR */
159 COSTS_N_INSNS (31), /* DSGR */
163 struct processor_costs z9_109_cost
=
165 COSTS_N_INSNS (4), /* M */
166 COSTS_N_INSNS (2), /* MGHI */
167 COSTS_N_INSNS (2), /* MH */
168 COSTS_N_INSNS (2), /* MHI */
169 COSTS_N_INSNS (4), /* ML */
170 COSTS_N_INSNS (4), /* MR */
171 COSTS_N_INSNS (5), /* MS */
172 COSTS_N_INSNS (6), /* MSG */
173 COSTS_N_INSNS (4), /* MSGF */
174 COSTS_N_INSNS (4), /* MSGFR */
175 COSTS_N_INSNS (4), /* MSGR */
176 COSTS_N_INSNS (4), /* MSR */
177 COSTS_N_INSNS (1), /* multiplication in DFmode */
178 COSTS_N_INSNS (28), /* MXBR */
179 COSTS_N_INSNS (130), /* SQXBR */
180 COSTS_N_INSNS (66), /* SQDBR */
181 COSTS_N_INSNS (38), /* SQEBR */
182 COSTS_N_INSNS (1), /* MADBR */
183 COSTS_N_INSNS (1), /* MAEBR */
184 COSTS_N_INSNS (60), /* DXBR */
185 COSTS_N_INSNS (40), /* DDBR */
186 COSTS_N_INSNS (26), /* DEBR */
187 COSTS_N_INSNS (30), /* DLGR */
188 COSTS_N_INSNS (23), /* DLR */
189 COSTS_N_INSNS (23), /* DR */
190 COSTS_N_INSNS (24), /* DSGFR */
191 COSTS_N_INSNS (24), /* DSGR */
195 struct processor_costs z10_cost
=
197 COSTS_N_INSNS (10), /* M */
198 COSTS_N_INSNS (10), /* MGHI */
199 COSTS_N_INSNS (10), /* MH */
200 COSTS_N_INSNS (10), /* MHI */
201 COSTS_N_INSNS (10), /* ML */
202 COSTS_N_INSNS (10), /* MR */
203 COSTS_N_INSNS (10), /* MS */
204 COSTS_N_INSNS (10), /* MSG */
205 COSTS_N_INSNS (10), /* MSGF */
206 COSTS_N_INSNS (10), /* MSGFR */
207 COSTS_N_INSNS (10), /* MSGR */
208 COSTS_N_INSNS (10), /* MSR */
209 COSTS_N_INSNS (1) , /* multiplication in DFmode */
210 COSTS_N_INSNS (50), /* MXBR */
211 COSTS_N_INSNS (120), /* SQXBR */
212 COSTS_N_INSNS (52), /* SQDBR */
213 COSTS_N_INSNS (38), /* SQEBR */
214 COSTS_N_INSNS (1), /* MADBR */
215 COSTS_N_INSNS (1), /* MAEBR */
216 COSTS_N_INSNS (111), /* DXBR */
217 COSTS_N_INSNS (39), /* DDBR */
218 COSTS_N_INSNS (32), /* DEBR */
219 COSTS_N_INSNS (160), /* DLGR */
220 COSTS_N_INSNS (71), /* DLR */
221 COSTS_N_INSNS (71), /* DR */
222 COSTS_N_INSNS (71), /* DSGFR */
223 COSTS_N_INSNS (71), /* DSGR */
226 extern int reload_completed
;
228 /* Kept up to date using the SCHED_VARIABLE_ISSUE hook. */
229 static rtx last_scheduled_insn
;
231 /* Structure used to hold the components of a S/390 memory
232 address. A legitimate address on S/390 is of the general
234 base + index + displacement
235 where any of the components is optional.
237 base and index are registers of the class ADDR_REGS,
238 displacement is an unsigned 12-bit immediate constant. */
249 /* Which cpu are we tuning for. */
250 enum processor_type s390_tune
= PROCESSOR_max
;
252 /* Which instruction set architecture to use. */
253 enum processor_type s390_arch
;
256 HOST_WIDE_INT s390_warn_framesize
= 0;
257 HOST_WIDE_INT s390_stack_size
= 0;
258 HOST_WIDE_INT s390_stack_guard
= 0;
260 /* The following structure is embedded in the machine
261 specific part of struct function. */
263 struct GTY (()) s390_frame_layout
265 /* Offset within stack frame. */
266 HOST_WIDE_INT gprs_offset
;
267 HOST_WIDE_INT f0_offset
;
268 HOST_WIDE_INT f4_offset
;
269 HOST_WIDE_INT f8_offset
;
270 HOST_WIDE_INT backchain_offset
;
272 /* Number of first and last gpr where slots in the register
273 save area are reserved for. */
274 int first_save_gpr_slot
;
275 int last_save_gpr_slot
;
277 /* Number of first and last gpr to be saved, restored. */
279 int first_restore_gpr
;
281 int last_restore_gpr
;
283 /* Bits standing for floating point registers. Set, if the
284 respective register has to be saved. Starting with reg 16 (f0)
285 at the rightmost bit.
286 Bit 15 - 8 7 6 5 4 3 2 1 0
287 fpr 15 - 8 7 5 3 1 6 4 2 0
288 reg 31 - 24 23 22 21 20 19 18 17 16 */
289 unsigned int fpr_bitmap
;
291 /* Number of floating point registers f8-f15 which must be saved. */
294 /* Set if return address needs to be saved.
295 This flag is set by s390_return_addr_rtx if it could not use
296 the initial value of r14 and therefore depends on r14 saved
298 bool save_return_addr_p
;
300 /* Size of stack frame. */
301 HOST_WIDE_INT frame_size
;
304 /* Define the structure for the machine field in struct function. */
306 struct GTY(()) machine_function
308 struct s390_frame_layout frame_layout
;
310 /* Literal pool base register. */
313 /* True if we may need to perform branch splitting. */
314 bool split_branches_pending_p
;
316 /* Some local-dynamic TLS symbol name. */
317 const char *some_ld_name
;
319 bool has_landing_pad_p
;
322 /* Few accessor macros for struct cfun->machine->s390_frame_layout. */
324 #define cfun_frame_layout (cfun->machine->frame_layout)
325 #define cfun_save_high_fprs_p (!!cfun_frame_layout.high_fprs)
326 #define cfun_gprs_save_area_size ((cfun_frame_layout.last_save_gpr_slot - \
327 cfun_frame_layout.first_save_gpr_slot + 1) * UNITS_PER_LONG)
328 #define cfun_set_fpr_bit(BITNUM) (cfun->machine->frame_layout.fpr_bitmap |= \
330 #define cfun_fpr_bit_p(BITNUM) (!!(cfun->machine->frame_layout.fpr_bitmap & \
333 /* Number of GPRs and FPRs used for argument passing. */
334 #define GP_ARG_NUM_REG 5
335 #define FP_ARG_NUM_REG (TARGET_64BIT? 4 : 2)
337 /* A couple of shortcuts. */
338 #define CONST_OK_FOR_J(x) \
339 CONST_OK_FOR_CONSTRAINT_P((x), 'J', "J")
340 #define CONST_OK_FOR_K(x) \
341 CONST_OK_FOR_CONSTRAINT_P((x), 'K', "K")
342 #define CONST_OK_FOR_Os(x) \
343 CONST_OK_FOR_CONSTRAINT_P((x), 'O', "Os")
344 #define CONST_OK_FOR_Op(x) \
345 CONST_OK_FOR_CONSTRAINT_P((x), 'O', "Op")
346 #define CONST_OK_FOR_On(x) \
347 CONST_OK_FOR_CONSTRAINT_P((x), 'O', "On")
349 #define REGNO_PAIR_OK(REGNO, MODE) \
350 (HARD_REGNO_NREGS ((REGNO), (MODE)) == 1 || !((REGNO) & 1))
352 /* That's the read ahead of the dynamic branch prediction unit in
353 bytes on a z10 CPU. */
354 #define Z10_PREDICT_DISTANCE 384
356 static enum machine_mode
357 s390_libgcc_cmp_return_mode (void)
359 return TARGET_64BIT
? DImode
: SImode
;
362 static enum machine_mode
363 s390_libgcc_shift_count_mode (void)
365 return TARGET_64BIT
? DImode
: SImode
;
368 static enum machine_mode
369 s390_unwind_word_mode (void)
371 return TARGET_64BIT
? DImode
: SImode
;
374 /* Return true if the back end supports mode MODE. */
376 s390_scalar_mode_supported_p (enum machine_mode mode
)
378 /* In contrast to the default implementation reject TImode constants on 31bit
379 TARGET_ZARCH for ABI compliance. */
380 if (!TARGET_64BIT
&& TARGET_ZARCH
&& mode
== TImode
)
383 if (DECIMAL_FLOAT_MODE_P (mode
))
384 return default_decimal_float_supported_p ();
386 return default_scalar_mode_supported_p (mode
);
389 /* Set the has_landing_pad_p flag in struct machine_function to VALUE. */
392 s390_set_has_landing_pad_p (bool value
)
394 cfun
->machine
->has_landing_pad_p
= value
;
397 /* If two condition code modes are compatible, return a condition code
398 mode which is compatible with both. Otherwise, return
401 static enum machine_mode
402 s390_cc_modes_compatible (enum machine_mode m1
, enum machine_mode m2
)
410 if (m2
== CCUmode
|| m2
== CCTmode
|| m2
== CCZ1mode
411 || m2
== CCSmode
|| m2
== CCSRmode
|| m2
== CCURmode
)
432 /* Return true if SET either doesn't set the CC register, or else
433 the source and destination have matching CC modes and that
434 CC mode is at least as constrained as REQ_MODE. */
437 s390_match_ccmode_set (rtx set
, enum machine_mode req_mode
)
439 enum machine_mode set_mode
;
441 gcc_assert (GET_CODE (set
) == SET
);
443 if (GET_CODE (SET_DEST (set
)) != REG
|| !CC_REGNO_P (REGNO (SET_DEST (set
))))
446 set_mode
= GET_MODE (SET_DEST (set
));
460 if (req_mode
!= set_mode
)
465 if (req_mode
!= CCSmode
&& req_mode
!= CCUmode
&& req_mode
!= CCTmode
466 && req_mode
!= CCSRmode
&& req_mode
!= CCURmode
)
472 if (req_mode
!= CCAmode
)
480 return (GET_MODE (SET_SRC (set
)) == set_mode
);
483 /* Return true if every SET in INSN that sets the CC register
484 has source and destination with matching CC modes and that
485 CC mode is at least as constrained as REQ_MODE.
486 If REQ_MODE is VOIDmode, always return false. */
489 s390_match_ccmode (rtx insn
, enum machine_mode req_mode
)
493 /* s390_tm_ccmode returns VOIDmode to indicate failure. */
494 if (req_mode
== VOIDmode
)
497 if (GET_CODE (PATTERN (insn
)) == SET
)
498 return s390_match_ccmode_set (PATTERN (insn
), req_mode
);
500 if (GET_CODE (PATTERN (insn
)) == PARALLEL
)
501 for (i
= 0; i
< XVECLEN (PATTERN (insn
), 0); i
++)
503 rtx set
= XVECEXP (PATTERN (insn
), 0, i
);
504 if (GET_CODE (set
) == SET
)
505 if (!s390_match_ccmode_set (set
, req_mode
))
512 /* If a test-under-mask instruction can be used to implement
513 (compare (and ... OP1) OP2), return the CC mode required
514 to do that. Otherwise, return VOIDmode.
515 MIXED is true if the instruction can distinguish between
516 CC1 and CC2 for mixed selected bits (TMxx), it is false
517 if the instruction cannot (TM). */
520 s390_tm_ccmode (rtx op1
, rtx op2
, bool mixed
)
524 /* ??? Fixme: should work on CONST_DOUBLE as well. */
525 if (GET_CODE (op1
) != CONST_INT
|| GET_CODE (op2
) != CONST_INT
)
528 /* Selected bits all zero: CC0.
529 e.g.: int a; if ((a & (16 + 128)) == 0) */
530 if (INTVAL (op2
) == 0)
533 /* Selected bits all one: CC3.
534 e.g.: int a; if ((a & (16 + 128)) == 16 + 128) */
535 if (INTVAL (op2
) == INTVAL (op1
))
538 /* Exactly two bits selected, mixed zeroes and ones: CC1 or CC2. e.g.:
540 if ((a & (16 + 128)) == 16) -> CCT1
541 if ((a & (16 + 128)) == 128) -> CCT2 */
544 bit1
= exact_log2 (INTVAL (op2
));
545 bit0
= exact_log2 (INTVAL (op1
) ^ INTVAL (op2
));
546 if (bit0
!= -1 && bit1
!= -1)
547 return bit0
> bit1
? CCT1mode
: CCT2mode
;
553 /* Given a comparison code OP (EQ, NE, etc.) and the operands
554 OP0 and OP1 of a COMPARE, return the mode to be used for the
558 s390_select_ccmode (enum rtx_code code
, rtx op0
, rtx op1
)
564 if ((GET_CODE (op0
) == NEG
|| GET_CODE (op0
) == ABS
)
565 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
567 if (GET_CODE (op0
) == PLUS
&& GET_CODE (XEXP (op0
, 1)) == CONST_INT
568 && CONST_OK_FOR_K (INTVAL (XEXP (op0
, 1))))
570 if ((GET_CODE (op0
) == PLUS
|| GET_CODE (op0
) == MINUS
571 || GET_CODE (op1
) == NEG
)
572 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
575 if (GET_CODE (op0
) == AND
)
577 /* Check whether we can potentially do it via TM. */
578 enum machine_mode ccmode
;
579 ccmode
= s390_tm_ccmode (XEXP (op0
, 1), op1
, 1);
580 if (ccmode
!= VOIDmode
)
582 /* Relax CCTmode to CCZmode to allow fall-back to AND
583 if that turns out to be beneficial. */
584 return ccmode
== CCTmode
? CCZmode
: ccmode
;
588 if (register_operand (op0
, HImode
)
589 && GET_CODE (op1
) == CONST_INT
590 && (INTVAL (op1
) == -1 || INTVAL (op1
) == 65535))
592 if (register_operand (op0
, QImode
)
593 && GET_CODE (op1
) == CONST_INT
594 && (INTVAL (op1
) == -1 || INTVAL (op1
) == 255))
603 /* The only overflow condition of NEG and ABS happens when
604 -INT_MAX is used as parameter, which stays negative. So
605 we have an overflow from a positive value to a negative.
606 Using CCAP mode the resulting cc can be used for comparisons. */
607 if ((GET_CODE (op0
) == NEG
|| GET_CODE (op0
) == ABS
)
608 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
611 /* If constants are involved in an add instruction it is possible to use
612 the resulting cc for comparisons with zero. Knowing the sign of the
613 constant the overflow behavior gets predictable. e.g.:
614 int a, b; if ((b = a + c) > 0)
615 with c as a constant value: c < 0 -> CCAN and c >= 0 -> CCAP */
616 if (GET_CODE (op0
) == PLUS
&& GET_CODE (XEXP (op0
, 1)) == CONST_INT
617 && CONST_OK_FOR_K (INTVAL (XEXP (op0
, 1))))
619 if (INTVAL (XEXP((op0
), 1)) < 0)
633 if ((GET_CODE (op0
) == SIGN_EXTEND
|| GET_CODE (op0
) == ZERO_EXTEND
)
634 && GET_CODE (op1
) != CONST_INT
)
640 if (GET_CODE (op0
) == PLUS
641 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
644 if ((GET_CODE (op0
) == SIGN_EXTEND
|| GET_CODE (op0
) == ZERO_EXTEND
)
645 && GET_CODE (op1
) != CONST_INT
)
651 if (GET_CODE (op0
) == MINUS
652 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
655 if ((GET_CODE (op0
) == SIGN_EXTEND
|| GET_CODE (op0
) == ZERO_EXTEND
)
656 && GET_CODE (op1
) != CONST_INT
)
665 /* Replace the comparison OP0 CODE OP1 by a semantically equivalent one
666 that we can implement more efficiently. */
669 s390_canonicalize_comparison (enum rtx_code
*code
, rtx
*op0
, rtx
*op1
)
671 /* Convert ZERO_EXTRACT back to AND to enable TM patterns. */
672 if ((*code
== EQ
|| *code
== NE
)
673 && *op1
== const0_rtx
674 && GET_CODE (*op0
) == ZERO_EXTRACT
675 && GET_CODE (XEXP (*op0
, 1)) == CONST_INT
676 && GET_CODE (XEXP (*op0
, 2)) == CONST_INT
677 && SCALAR_INT_MODE_P (GET_MODE (XEXP (*op0
, 0))))
679 rtx inner
= XEXP (*op0
, 0);
680 HOST_WIDE_INT modesize
= GET_MODE_BITSIZE (GET_MODE (inner
));
681 HOST_WIDE_INT len
= INTVAL (XEXP (*op0
, 1));
682 HOST_WIDE_INT pos
= INTVAL (XEXP (*op0
, 2));
684 if (len
> 0 && len
< modesize
685 && pos
>= 0 && pos
+ len
<= modesize
686 && modesize
<= HOST_BITS_PER_WIDE_INT
)
688 unsigned HOST_WIDE_INT block
;
689 block
= ((unsigned HOST_WIDE_INT
) 1 << len
) - 1;
690 block
<<= modesize
- pos
- len
;
692 *op0
= gen_rtx_AND (GET_MODE (inner
), inner
,
693 gen_int_mode (block
, GET_MODE (inner
)));
697 /* Narrow AND of memory against immediate to enable TM. */
698 if ((*code
== EQ
|| *code
== NE
)
699 && *op1
== const0_rtx
700 && GET_CODE (*op0
) == AND
701 && GET_CODE (XEXP (*op0
, 1)) == CONST_INT
702 && SCALAR_INT_MODE_P (GET_MODE (XEXP (*op0
, 0))))
704 rtx inner
= XEXP (*op0
, 0);
705 rtx mask
= XEXP (*op0
, 1);
707 /* Ignore paradoxical SUBREGs if all extra bits are masked out. */
708 if (GET_CODE (inner
) == SUBREG
709 && SCALAR_INT_MODE_P (GET_MODE (SUBREG_REG (inner
)))
710 && (GET_MODE_SIZE (GET_MODE (inner
))
711 >= GET_MODE_SIZE (GET_MODE (SUBREG_REG (inner
))))
713 & GET_MODE_MASK (GET_MODE (inner
))
714 & ~GET_MODE_MASK (GET_MODE (SUBREG_REG (inner
))))
716 inner
= SUBREG_REG (inner
);
718 /* Do not change volatile MEMs. */
719 if (MEM_P (inner
) && !MEM_VOLATILE_P (inner
))
721 int part
= s390_single_part (XEXP (*op0
, 1),
722 GET_MODE (inner
), QImode
, 0);
725 mask
= gen_int_mode (s390_extract_part (mask
, QImode
, 0), QImode
);
726 inner
= adjust_address_nv (inner
, QImode
, part
);
727 *op0
= gen_rtx_AND (QImode
, inner
, mask
);
732 /* Narrow comparisons against 0xffff to HImode if possible. */
733 if ((*code
== EQ
|| *code
== NE
)
734 && GET_CODE (*op1
) == CONST_INT
735 && INTVAL (*op1
) == 0xffff
736 && SCALAR_INT_MODE_P (GET_MODE (*op0
))
737 && (nonzero_bits (*op0
, GET_MODE (*op0
))
738 & ~(unsigned HOST_WIDE_INT
) 0xffff) == 0)
740 *op0
= gen_lowpart (HImode
, *op0
);
744 /* Remove redundant UNSPEC_CCU_TO_INT conversions if possible. */
745 if (GET_CODE (*op0
) == UNSPEC
746 && XINT (*op0
, 1) == UNSPEC_CCU_TO_INT
747 && XVECLEN (*op0
, 0) == 1
748 && GET_MODE (XVECEXP (*op0
, 0, 0)) == CCUmode
749 && GET_CODE (XVECEXP (*op0
, 0, 0)) == REG
750 && REGNO (XVECEXP (*op0
, 0, 0)) == CC_REGNUM
751 && *op1
== const0_rtx
)
753 enum rtx_code new_code
= UNKNOWN
;
756 case EQ
: new_code
= EQ
; break;
757 case NE
: new_code
= NE
; break;
758 case LT
: new_code
= GTU
; break;
759 case GT
: new_code
= LTU
; break;
760 case LE
: new_code
= GEU
; break;
761 case GE
: new_code
= LEU
; break;
765 if (new_code
!= UNKNOWN
)
767 *op0
= XVECEXP (*op0
, 0, 0);
772 /* Remove redundant UNSPEC_CCZ_TO_INT conversions if possible. */
773 if (GET_CODE (*op0
) == UNSPEC
774 && XINT (*op0
, 1) == UNSPEC_CCZ_TO_INT
775 && XVECLEN (*op0
, 0) == 1
776 && GET_MODE (XVECEXP (*op0
, 0, 0)) == CCZmode
777 && GET_CODE (XVECEXP (*op0
, 0, 0)) == REG
778 && REGNO (XVECEXP (*op0
, 0, 0)) == CC_REGNUM
779 && *op1
== const0_rtx
)
781 enum rtx_code new_code
= UNKNOWN
;
784 case EQ
: new_code
= EQ
; break;
785 case NE
: new_code
= NE
; break;
789 if (new_code
!= UNKNOWN
)
791 *op0
= XVECEXP (*op0
, 0, 0);
796 /* Simplify cascaded EQ, NE with const0_rtx. */
797 if ((*code
== NE
|| *code
== EQ
)
798 && (GET_CODE (*op0
) == EQ
|| GET_CODE (*op0
) == NE
)
799 && GET_MODE (*op0
) == SImode
800 && GET_MODE (XEXP (*op0
, 0)) == CCZ1mode
801 && REG_P (XEXP (*op0
, 0))
802 && XEXP (*op0
, 1) == const0_rtx
803 && *op1
== const0_rtx
)
805 if ((*code
== EQ
&& GET_CODE (*op0
) == NE
)
806 || (*code
== NE
&& GET_CODE (*op0
) == EQ
))
810 *op0
= XEXP (*op0
, 0);
813 /* Prefer register over memory as first operand. */
814 if (MEM_P (*op0
) && REG_P (*op1
))
816 rtx tem
= *op0
; *op0
= *op1
; *op1
= tem
;
817 *code
= swap_condition (*code
);
821 /* Emit a compare instruction suitable to implement the comparison
822 OP0 CODE OP1. Return the correct condition RTL to be placed in
823 the IF_THEN_ELSE of the conditional branch testing the result. */
826 s390_emit_compare (enum rtx_code code
, rtx op0
, rtx op1
)
828 enum machine_mode mode
= s390_select_ccmode (code
, op0
, op1
);
831 /* Do not output a redundant compare instruction if a compare_and_swap
832 pattern already computed the result and the machine modes are compatible. */
833 if (GET_MODE_CLASS (GET_MODE (op0
)) == MODE_CC
)
835 gcc_assert (s390_cc_modes_compatible (GET_MODE (op0
), mode
)
841 cc
= gen_rtx_REG (mode
, CC_REGNUM
);
842 emit_insn (gen_rtx_SET (VOIDmode
, cc
, gen_rtx_COMPARE (mode
, op0
, op1
)));
845 return gen_rtx_fmt_ee (code
, VOIDmode
, cc
, const0_rtx
);
848 /* Emit a SImode compare and swap instruction setting MEM to NEW_RTX if OLD
850 Return the correct condition RTL to be placed in the IF_THEN_ELSE of the
851 conditional branch testing the result. */
854 s390_emit_compare_and_swap (enum rtx_code code
, rtx old
, rtx mem
, rtx cmp
, rtx new_rtx
)
856 emit_insn (gen_sync_compare_and_swapsi (old
, mem
, cmp
, new_rtx
));
857 return s390_emit_compare (code
, gen_rtx_REG (CCZ1mode
, CC_REGNUM
), const0_rtx
);
860 /* Emit a jump instruction to TARGET. If COND is NULL_RTX, emit an
861 unconditional jump, else a conditional jump under condition COND. */
864 s390_emit_jump (rtx target
, rtx cond
)
868 target
= gen_rtx_LABEL_REF (VOIDmode
, target
);
870 target
= gen_rtx_IF_THEN_ELSE (VOIDmode
, cond
, target
, pc_rtx
);
872 insn
= gen_rtx_SET (VOIDmode
, pc_rtx
, target
);
873 emit_jump_insn (insn
);
876 /* Return branch condition mask to implement a branch
877 specified by CODE. Return -1 for invalid comparisons. */
880 s390_branch_condition_mask (rtx code
)
882 const int CC0
= 1 << 3;
883 const int CC1
= 1 << 2;
884 const int CC2
= 1 << 1;
885 const int CC3
= 1 << 0;
887 gcc_assert (GET_CODE (XEXP (code
, 0)) == REG
);
888 gcc_assert (REGNO (XEXP (code
, 0)) == CC_REGNUM
);
889 gcc_assert (XEXP (code
, 1) == const0_rtx
);
891 switch (GET_MODE (XEXP (code
, 0)))
895 switch (GET_CODE (code
))
898 case NE
: return CC1
| CC2
| CC3
;
904 switch (GET_CODE (code
))
907 case NE
: return CC0
| CC2
| CC3
;
913 switch (GET_CODE (code
))
916 case NE
: return CC0
| CC1
| CC3
;
922 switch (GET_CODE (code
))
925 case NE
: return CC0
| CC1
| CC2
;
931 switch (GET_CODE (code
))
933 case EQ
: return CC0
| CC2
;
934 case NE
: return CC1
| CC3
;
940 switch (GET_CODE (code
))
942 case LTU
: return CC2
| CC3
; /* carry */
943 case GEU
: return CC0
| CC1
; /* no carry */
949 switch (GET_CODE (code
))
951 case GTU
: return CC0
| CC1
; /* borrow */
952 case LEU
: return CC2
| CC3
; /* no borrow */
958 switch (GET_CODE (code
))
960 case EQ
: return CC0
| CC2
;
961 case NE
: return CC1
| CC3
;
962 case LTU
: return CC1
;
963 case GTU
: return CC3
;
964 case LEU
: return CC1
| CC2
;
965 case GEU
: return CC2
| CC3
;
970 switch (GET_CODE (code
))
973 case NE
: return CC1
| CC2
| CC3
;
974 case LTU
: return CC1
;
975 case GTU
: return CC2
;
976 case LEU
: return CC0
| CC1
;
977 case GEU
: return CC0
| CC2
;
983 switch (GET_CODE (code
))
986 case NE
: return CC2
| CC1
| CC3
;
987 case LTU
: return CC2
;
988 case GTU
: return CC1
;
989 case LEU
: return CC0
| CC2
;
990 case GEU
: return CC0
| CC1
;
996 switch (GET_CODE (code
))
999 case NE
: return CC1
| CC2
| CC3
;
1000 case LT
: return CC1
| CC3
;
1001 case GT
: return CC2
;
1002 case LE
: return CC0
| CC1
| CC3
;
1003 case GE
: return CC0
| CC2
;
1009 switch (GET_CODE (code
))
1011 case EQ
: return CC0
;
1012 case NE
: return CC1
| CC2
| CC3
;
1013 case LT
: return CC1
;
1014 case GT
: return CC2
| CC3
;
1015 case LE
: return CC0
| CC1
;
1016 case GE
: return CC0
| CC2
| CC3
;
1022 switch (GET_CODE (code
))
1024 case EQ
: return CC0
;
1025 case NE
: return CC1
| CC2
| CC3
;
1026 case LT
: return CC1
;
1027 case GT
: return CC2
;
1028 case LE
: return CC0
| CC1
;
1029 case GE
: return CC0
| CC2
;
1030 case UNORDERED
: return CC3
;
1031 case ORDERED
: return CC0
| CC1
| CC2
;
1032 case UNEQ
: return CC0
| CC3
;
1033 case UNLT
: return CC1
| CC3
;
1034 case UNGT
: return CC2
| CC3
;
1035 case UNLE
: return CC0
| CC1
| CC3
;
1036 case UNGE
: return CC0
| CC2
| CC3
;
1037 case LTGT
: return CC1
| CC2
;
1043 switch (GET_CODE (code
))
1045 case EQ
: return CC0
;
1046 case NE
: return CC2
| CC1
| CC3
;
1047 case LT
: return CC2
;
1048 case GT
: return CC1
;
1049 case LE
: return CC0
| CC2
;
1050 case GE
: return CC0
| CC1
;
1051 case UNORDERED
: return CC3
;
1052 case ORDERED
: return CC0
| CC2
| CC1
;
1053 case UNEQ
: return CC0
| CC3
;
1054 case UNLT
: return CC2
| CC3
;
1055 case UNGT
: return CC1
| CC3
;
1056 case UNLE
: return CC0
| CC2
| CC3
;
1057 case UNGE
: return CC0
| CC1
| CC3
;
1058 case LTGT
: return CC2
| CC1
;
1069 /* Return branch condition mask to implement a compare and branch
1070 specified by CODE. Return -1 for invalid comparisons. */
1073 s390_compare_and_branch_condition_mask (rtx code
)
1075 const int CC0
= 1 << 3;
1076 const int CC1
= 1 << 2;
1077 const int CC2
= 1 << 1;
1079 switch (GET_CODE (code
))
1103 /* If INV is false, return assembler mnemonic string to implement
1104 a branch specified by CODE. If INV is true, return mnemonic
1105 for the corresponding inverted branch. */
1108 s390_branch_condition_mnemonic (rtx code
, int inv
)
1112 static const char *const mnemonic
[16] =
1114 NULL
, "o", "h", "nle",
1115 "l", "nhe", "lh", "ne",
1116 "e", "nlh", "he", "nl",
1117 "le", "nh", "no", NULL
1120 if (GET_CODE (XEXP (code
, 0)) == REG
1121 && REGNO (XEXP (code
, 0)) == CC_REGNUM
1122 && XEXP (code
, 1) == const0_rtx
)
1123 mask
= s390_branch_condition_mask (code
);
1125 mask
= s390_compare_and_branch_condition_mask (code
);
1127 gcc_assert (mask
>= 0);
1132 gcc_assert (mask
>= 1 && mask
<= 14);
1134 return mnemonic
[mask
];
1137 /* Return the part of op which has a value different from def.
1138 The size of the part is determined by mode.
1139 Use this function only if you already know that op really
1140 contains such a part. */
1142 unsigned HOST_WIDE_INT
1143 s390_extract_part (rtx op
, enum machine_mode mode
, int def
)
1145 unsigned HOST_WIDE_INT value
= 0;
1146 int max_parts
= HOST_BITS_PER_WIDE_INT
/ GET_MODE_BITSIZE (mode
);
1147 int part_bits
= GET_MODE_BITSIZE (mode
);
1148 unsigned HOST_WIDE_INT part_mask
1149 = ((unsigned HOST_WIDE_INT
)1 << part_bits
) - 1;
1152 for (i
= 0; i
< max_parts
; i
++)
1155 value
= (unsigned HOST_WIDE_INT
) INTVAL (op
);
1157 value
>>= part_bits
;
1159 if ((value
& part_mask
) != (def
& part_mask
))
1160 return value
& part_mask
;
1166 /* If OP is an integer constant of mode MODE with exactly one
1167 part of mode PART_MODE unequal to DEF, return the number of that
1168 part. Otherwise, return -1. */
1171 s390_single_part (rtx op
,
1172 enum machine_mode mode
,
1173 enum machine_mode part_mode
,
1176 unsigned HOST_WIDE_INT value
= 0;
1177 int n_parts
= GET_MODE_SIZE (mode
) / GET_MODE_SIZE (part_mode
);
1178 unsigned HOST_WIDE_INT part_mask
1179 = ((unsigned HOST_WIDE_INT
)1 << GET_MODE_BITSIZE (part_mode
)) - 1;
1182 if (GET_CODE (op
) != CONST_INT
)
1185 for (i
= 0; i
< n_parts
; i
++)
1188 value
= (unsigned HOST_WIDE_INT
) INTVAL (op
);
1190 value
>>= GET_MODE_BITSIZE (part_mode
);
1192 if ((value
& part_mask
) != (def
& part_mask
))
1200 return part
== -1 ? -1 : n_parts
- 1 - part
;
1203 /* Return true if IN contains a contiguous bitfield in the lower SIZE
1204 bits and no other bits are set in IN. POS and LENGTH can be used
1205 to obtain the start position and the length of the bitfield.
1207 POS gives the position of the first bit of the bitfield counting
1208 from the lowest order bit starting with zero. In order to use this
1209 value for S/390 instructions this has to be converted to "bits big
1213 s390_contiguous_bitmask_p (unsigned HOST_WIDE_INT in
, int size
,
1214 int *pos
, int *length
)
1219 unsigned HOST_WIDE_INT mask
= 1ULL;
1220 bool contiguous
= false;
1222 for (i
= 0; i
< size
; mask
<<= 1, i
++)
1246 /* Calculate a mask for all bits beyond the contiguous bits. */
1247 mask
= (-1LL & ~(((1ULL << (tmp_length
+ tmp_pos
- 1)) << 1) - 1));
1252 if (tmp_length
+ tmp_pos
- 1 > size
)
1256 *length
= tmp_length
;
1264 /* Check whether we can (and want to) split a double-word
1265 move in mode MODE from SRC to DST into two single-word
1266 moves, moving the subword FIRST_SUBWORD first. */
1269 s390_split_ok_p (rtx dst
, rtx src
, enum machine_mode mode
, int first_subword
)
1271 /* Floating point registers cannot be split. */
1272 if (FP_REG_P (src
) || FP_REG_P (dst
))
1275 /* We don't need to split if operands are directly accessible. */
1276 if (s_operand (src
, mode
) || s_operand (dst
, mode
))
1279 /* Non-offsettable memory references cannot be split. */
1280 if ((GET_CODE (src
) == MEM
&& !offsettable_memref_p (src
))
1281 || (GET_CODE (dst
) == MEM
&& !offsettable_memref_p (dst
)))
1284 /* Moving the first subword must not clobber a register
1285 needed to move the second subword. */
1286 if (register_operand (dst
, mode
))
1288 rtx subreg
= operand_subword (dst
, first_subword
, 0, mode
);
1289 if (reg_overlap_mentioned_p (subreg
, src
))
1296 /* Return true if it can be proven that [MEM1, MEM1 + SIZE]
1297 and [MEM2, MEM2 + SIZE] do overlap and false
1301 s390_overlap_p (rtx mem1
, rtx mem2
, HOST_WIDE_INT size
)
1303 rtx addr1
, addr2
, addr_delta
;
1304 HOST_WIDE_INT delta
;
1306 if (GET_CODE (mem1
) != MEM
|| GET_CODE (mem2
) != MEM
)
1312 addr1
= XEXP (mem1
, 0);
1313 addr2
= XEXP (mem2
, 0);
1315 addr_delta
= simplify_binary_operation (MINUS
, Pmode
, addr2
, addr1
);
1317 /* This overlapping check is used by peepholes merging memory block operations.
1318 Overlapping operations would otherwise be recognized by the S/390 hardware
1319 and would fall back to a slower implementation. Allowing overlapping
1320 operations would lead to slow code but not to wrong code. Therefore we are
1321 somewhat optimistic if we cannot prove that the memory blocks are
1323 That's why we return false here although this may accept operations on
1324 overlapping memory areas. */
1325 if (!addr_delta
|| GET_CODE (addr_delta
) != CONST_INT
)
1328 delta
= INTVAL (addr_delta
);
1331 || (delta
> 0 && delta
< size
)
1332 || (delta
< 0 && -delta
< size
))
1338 /* Check whether the address of memory reference MEM2 equals exactly
1339 the address of memory reference MEM1 plus DELTA. Return true if
1340 we can prove this to be the case, false otherwise. */
1343 s390_offset_p (rtx mem1
, rtx mem2
, rtx delta
)
1345 rtx addr1
, addr2
, addr_delta
;
1347 if (GET_CODE (mem1
) != MEM
|| GET_CODE (mem2
) != MEM
)
1350 addr1
= XEXP (mem1
, 0);
1351 addr2
= XEXP (mem2
, 0);
1353 addr_delta
= simplify_binary_operation (MINUS
, Pmode
, addr2
, addr1
);
1354 if (!addr_delta
|| !rtx_equal_p (addr_delta
, delta
))
1360 /* Expand logical operator CODE in mode MODE with operands OPERANDS. */
1363 s390_expand_logical_operator (enum rtx_code code
, enum machine_mode mode
,
1366 enum machine_mode wmode
= mode
;
1367 rtx dst
= operands
[0];
1368 rtx src1
= operands
[1];
1369 rtx src2
= operands
[2];
1372 /* If we cannot handle the operation directly, use a temp register. */
1373 if (!s390_logical_operator_ok_p (operands
))
1374 dst
= gen_reg_rtx (mode
);
1376 /* QImode and HImode patterns make sense only if we have a destination
1377 in memory. Otherwise perform the operation in SImode. */
1378 if ((mode
== QImode
|| mode
== HImode
) && GET_CODE (dst
) != MEM
)
1381 /* Widen operands if required. */
1384 if (GET_CODE (dst
) == SUBREG
1385 && (tem
= simplify_subreg (wmode
, dst
, mode
, 0)) != 0)
1387 else if (REG_P (dst
))
1388 dst
= gen_rtx_SUBREG (wmode
, dst
, 0);
1390 dst
= gen_reg_rtx (wmode
);
1392 if (GET_CODE (src1
) == SUBREG
1393 && (tem
= simplify_subreg (wmode
, src1
, mode
, 0)) != 0)
1395 else if (GET_MODE (src1
) != VOIDmode
)
1396 src1
= gen_rtx_SUBREG (wmode
, force_reg (mode
, src1
), 0);
1398 if (GET_CODE (src2
) == SUBREG
1399 && (tem
= simplify_subreg (wmode
, src2
, mode
, 0)) != 0)
1401 else if (GET_MODE (src2
) != VOIDmode
)
1402 src2
= gen_rtx_SUBREG (wmode
, force_reg (mode
, src2
), 0);
1405 /* Emit the instruction. */
1406 op
= gen_rtx_SET (VOIDmode
, dst
, gen_rtx_fmt_ee (code
, wmode
, src1
, src2
));
1407 clob
= gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
1408 emit_insn (gen_rtx_PARALLEL (VOIDmode
, gen_rtvec (2, op
, clob
)));
1410 /* Fix up the destination if needed. */
1411 if (dst
!= operands
[0])
1412 emit_move_insn (operands
[0], gen_lowpart (mode
, dst
));
1415 /* Check whether OPERANDS are OK for a logical operation (AND, IOR, XOR). */
1418 s390_logical_operator_ok_p (rtx
*operands
)
1420 /* If the destination operand is in memory, it needs to coincide
1421 with one of the source operands. After reload, it has to be
1422 the first source operand. */
1423 if (GET_CODE (operands
[0]) == MEM
)
1424 return rtx_equal_p (operands
[0], operands
[1])
1425 || (!reload_completed
&& rtx_equal_p (operands
[0], operands
[2]));
1430 /* Narrow logical operation CODE of memory operand MEMOP with immediate
1431 operand IMMOP to switch from SS to SI type instructions. */
1434 s390_narrow_logical_operator (enum rtx_code code
, rtx
*memop
, rtx
*immop
)
1436 int def
= code
== AND
? -1 : 0;
1440 gcc_assert (GET_CODE (*memop
) == MEM
);
1441 gcc_assert (!MEM_VOLATILE_P (*memop
));
1443 mask
= s390_extract_part (*immop
, QImode
, def
);
1444 part
= s390_single_part (*immop
, GET_MODE (*memop
), QImode
, def
);
1445 gcc_assert (part
>= 0);
1447 *memop
= adjust_address (*memop
, QImode
, part
);
1448 *immop
= gen_int_mode (mask
, QImode
);
1452 /* How to allocate a 'struct machine_function'. */
1454 static struct machine_function
*
1455 s390_init_machine_status (void)
1457 return GGC_CNEW (struct machine_function
);
1460 /* Change optimizations to be performed, depending on the
1463 LEVEL is the optimization level specified; 2 if `-O2' is
1464 specified, 1 if `-O' is specified, and 0 if neither is specified.
1466 SIZE is nonzero if `-Os' is specified and zero otherwise. */
1469 optimization_options (int level ATTRIBUTE_UNUSED
, int size ATTRIBUTE_UNUSED
)
1471 /* ??? There are apparently still problems with -fcaller-saves. */
1472 flag_caller_saves
= 0;
1474 /* By default, always emit DWARF-2 unwind info. This allows debugging
1475 without maintaining a stack frame back-chain. */
1476 flag_asynchronous_unwind_tables
= 1;
1478 if (HAVE_prefetch
|| optimize
>= 3)
1479 flag_prefetch_loop_arrays
= 1;
1481 /* Use MVCLE instructions to decrease code size if requested. */
1483 target_flags
|= MASK_MVCLE
;
1486 /* Return true if ARG is the name of a processor. Set *TYPE and *FLAGS
1487 to the associated processor_type and processor_flags if so. */
1490 s390_handle_arch_option (const char *arg
,
1491 enum processor_type
*type
,
1496 const char *const name
; /* processor name or nickname. */
1497 const enum processor_type processor
;
1498 const int flags
; /* From enum processor_flags. */
1500 const processor_alias_table
[] =
1502 {"g5", PROCESSOR_9672_G5
, PF_IEEE_FLOAT
},
1503 {"g6", PROCESSOR_9672_G6
, PF_IEEE_FLOAT
},
1504 {"z900", PROCESSOR_2064_Z900
, PF_IEEE_FLOAT
| PF_ZARCH
},
1505 {"z990", PROCESSOR_2084_Z990
, PF_IEEE_FLOAT
| PF_ZARCH
1506 | PF_LONG_DISPLACEMENT
},
1507 {"z9-109", PROCESSOR_2094_Z9_109
, PF_IEEE_FLOAT
| PF_ZARCH
1508 | PF_LONG_DISPLACEMENT
| PF_EXTIMM
},
1509 {"z9-ec", PROCESSOR_2094_Z9_109
, PF_IEEE_FLOAT
| PF_ZARCH
1510 | PF_LONG_DISPLACEMENT
| PF_EXTIMM
| PF_DFP
},
1511 {"z10", PROCESSOR_2097_Z10
, PF_IEEE_FLOAT
| PF_ZARCH
1512 | PF_LONG_DISPLACEMENT
| PF_EXTIMM
| PF_DFP
| PF_Z10
},
1516 for (i
= 0; i
< ARRAY_SIZE (processor_alias_table
); i
++)
1517 if (strcmp (arg
, processor_alias_table
[i
].name
) == 0)
1519 *type
= processor_alias_table
[i
].processor
;
1520 *flags
= processor_alias_table
[i
].flags
;
1526 /* Implement TARGET_HANDLE_OPTION. */
1529 s390_handle_option (size_t code
, const char *arg
, int value ATTRIBUTE_UNUSED
)
1534 return s390_handle_arch_option (arg
, &s390_arch
, &s390_arch_flags
);
1536 case OPT_mstack_guard_
:
1537 if (sscanf (arg
, HOST_WIDE_INT_PRINT_DEC
, &s390_stack_guard
) != 1)
1539 if (exact_log2 (s390_stack_guard
) == -1)
1540 error ("stack guard value must be an exact power of 2");
1543 case OPT_mstack_size_
:
1544 if (sscanf (arg
, HOST_WIDE_INT_PRINT_DEC
, &s390_stack_size
) != 1)
1546 if (exact_log2 (s390_stack_size
) == -1)
1547 error ("stack size must be an exact power of 2");
1551 return s390_handle_arch_option (arg
, &s390_tune
, &s390_tune_flags
);
1553 case OPT_mwarn_framesize_
:
1554 return sscanf (arg
, HOST_WIDE_INT_PRINT_DEC
, &s390_warn_framesize
) == 1;
1562 override_options (void)
1564 /* Set up function hooks. */
1565 init_machine_status
= s390_init_machine_status
;
1567 /* Architecture mode defaults according to ABI. */
1568 if (!(target_flags_explicit
& MASK_ZARCH
))
1571 target_flags
|= MASK_ZARCH
;
1573 target_flags
&= ~MASK_ZARCH
;
1576 /* Determine processor architectural level. */
1577 if (!s390_arch_string
)
1579 s390_arch_string
= TARGET_ZARCH
? "z900" : "g5";
1580 s390_handle_arch_option (s390_arch_string
, &s390_arch
, &s390_arch_flags
);
1583 /* Determine processor to tune for. */
1584 if (s390_tune
== PROCESSOR_max
)
1586 s390_tune
= s390_arch
;
1587 s390_tune_flags
= s390_arch_flags
;
1590 /* Sanity checks. */
1591 if (TARGET_ZARCH
&& !TARGET_CPU_ZARCH
)
1592 error ("z/Architecture mode not supported on %s", s390_arch_string
);
1593 if (TARGET_64BIT
&& !TARGET_ZARCH
)
1594 error ("64-bit ABI not supported in ESA/390 mode");
1596 if (TARGET_HARD_DFP
&& !TARGET_DFP
)
1598 if (target_flags_explicit
& MASK_HARD_DFP
)
1600 if (!TARGET_CPU_DFP
)
1601 error ("Hardware decimal floating point instructions"
1602 " not available on %s", s390_arch_string
);
1604 error ("Hardware decimal floating point instructions"
1605 " not available in ESA/390 mode");
1608 target_flags
&= ~MASK_HARD_DFP
;
1611 if ((target_flags_explicit
& MASK_SOFT_FLOAT
) && TARGET_SOFT_FLOAT
)
1613 if ((target_flags_explicit
& MASK_HARD_DFP
) && TARGET_HARD_DFP
)
1614 error ("-mhard-dfp can't be used in conjunction with -msoft-float");
1616 target_flags
&= ~MASK_HARD_DFP
;
1619 /* Set processor cost function. */
1622 case PROCESSOR_2084_Z990
:
1623 s390_cost
= &z990_cost
;
1625 case PROCESSOR_2094_Z9_109
:
1626 s390_cost
= &z9_109_cost
;
1628 case PROCESSOR_2097_Z10
:
1629 s390_cost
= &z10_cost
;
1632 s390_cost
= &z900_cost
;
1635 if (TARGET_BACKCHAIN
&& TARGET_PACKED_STACK
&& TARGET_HARD_FLOAT
)
1636 error ("-mbackchain -mpacked-stack -mhard-float are not supported "
1639 if (s390_stack_size
)
1641 if (s390_stack_guard
>= s390_stack_size
)
1642 error ("stack size must be greater than the stack guard value");
1643 else if (s390_stack_size
> 1 << 16)
1644 error ("stack size must not be greater than 64k");
1646 else if (s390_stack_guard
)
1647 error ("-mstack-guard implies use of -mstack-size");
1649 #ifdef TARGET_DEFAULT_LONG_DOUBLE_128
1650 if (!(target_flags_explicit
& MASK_LONG_DOUBLE_128
))
1651 target_flags
|= MASK_LONG_DOUBLE_128
;
1654 if (s390_tune
== PROCESSOR_2097_Z10
)
1656 if (!PARAM_SET_P (PARAM_MAX_UNROLLED_INSNS
))
1657 set_param_value ("max-unrolled-insns", 100);
1658 if (!PARAM_SET_P (PARAM_MAX_UNROLL_TIMES
))
1659 set_param_value ("max-unroll-times", 32);
1660 if (!PARAM_SET_P (PARAM_MAX_COMPLETELY_PEELED_INSNS
))
1661 set_param_value ("max-completely-peeled-insns", 800);
1662 if (!PARAM_SET_P (PARAM_MAX_COMPLETELY_PEEL_TIMES
))
1663 set_param_value ("max-completely-peel-times", 64);
1666 set_param_value ("max-pending-list-length", 256);
1667 /* values for loop prefetching */
1668 set_param_value ("l1-cache-line-size", 256);
1669 if (!PARAM_SET_P (PARAM_L1_CACHE_SIZE
))
1670 set_param_value ("l1-cache-size", 128);
1671 /* s390 has more than 2 levels and the size is much larger. Since
1672 we are always running virtualized assume that we only get a small
1673 part of the caches above l1. */
1674 if (!PARAM_SET_P (PARAM_L2_CACHE_SIZE
))
1675 set_param_value ("l2-cache-size", 1500);
1676 if (!PARAM_SET_P (PARAM_PREFETCH_MIN_INSN_TO_MEM_RATIO
))
1677 set_param_value ("prefetch-min-insn-to-mem-ratio", 2);
1678 if (!PARAM_SET_P (PARAM_SIMULTANEOUS_PREFETCHES
))
1679 set_param_value ("simultaneous-prefetches", 6);
1682 /* Map for smallest class containing reg regno. */
1684 const enum reg_class regclass_map
[FIRST_PSEUDO_REGISTER
] =
1685 { GENERAL_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1686 ADDR_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1687 ADDR_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1688 ADDR_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1689 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1690 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1691 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1692 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1693 ADDR_REGS
, CC_REGS
, ADDR_REGS
, ADDR_REGS
,
1694 ACCESS_REGS
, ACCESS_REGS
1697 /* Return attribute type of insn. */
1699 static enum attr_type
1700 s390_safe_attr_type (rtx insn
)
1702 if (recog_memoized (insn
) >= 0)
1703 return get_attr_type (insn
);
1708 /* Return true if DISP is a valid short displacement. */
1711 s390_short_displacement (rtx disp
)
1713 /* No displacement is OK. */
1717 /* Without the long displacement facility we don't need to
1718 distingiush between long and short displacement. */
1719 if (!TARGET_LONG_DISPLACEMENT
)
1722 /* Integer displacement in range. */
1723 if (GET_CODE (disp
) == CONST_INT
)
1724 return INTVAL (disp
) >= 0 && INTVAL (disp
) < 4096;
1726 /* GOT offset is not OK, the GOT can be large. */
1727 if (GET_CODE (disp
) == CONST
1728 && GET_CODE (XEXP (disp
, 0)) == UNSPEC
1729 && (XINT (XEXP (disp
, 0), 1) == UNSPEC_GOT
1730 || XINT (XEXP (disp
, 0), 1) == UNSPEC_GOTNTPOFF
))
1733 /* All other symbolic constants are literal pool references,
1734 which are OK as the literal pool must be small. */
1735 if (GET_CODE (disp
) == CONST
)
1741 /* Decompose a RTL expression ADDR for a memory address into
1742 its components, returned in OUT.
1744 Returns false if ADDR is not a valid memory address, true
1745 otherwise. If OUT is NULL, don't return the components,
1746 but check for validity only.
1748 Note: Only addresses in canonical form are recognized.
1749 LEGITIMIZE_ADDRESS should convert non-canonical forms to the
1750 canonical form so that they will be recognized. */
1753 s390_decompose_address (rtx addr
, struct s390_address
*out
)
1755 HOST_WIDE_INT offset
= 0;
1756 rtx base
= NULL_RTX
;
1757 rtx indx
= NULL_RTX
;
1758 rtx disp
= NULL_RTX
;
1760 bool pointer
= false;
1761 bool base_ptr
= false;
1762 bool indx_ptr
= false;
1763 bool literal_pool
= false;
1765 /* We may need to substitute the literal pool base register into the address
1766 below. However, at this point we do not know which register is going to
1767 be used as base, so we substitute the arg pointer register. This is going
1768 to be treated as holding a pointer below -- it shouldn't be used for any
1770 rtx fake_pool_base
= gen_rtx_REG (Pmode
, ARG_POINTER_REGNUM
);
1772 /* Decompose address into base + index + displacement. */
1774 if (GET_CODE (addr
) == REG
|| GET_CODE (addr
) == UNSPEC
)
1777 else if (GET_CODE (addr
) == PLUS
)
1779 rtx op0
= XEXP (addr
, 0);
1780 rtx op1
= XEXP (addr
, 1);
1781 enum rtx_code code0
= GET_CODE (op0
);
1782 enum rtx_code code1
= GET_CODE (op1
);
1784 if (code0
== REG
|| code0
== UNSPEC
)
1786 if (code1
== REG
|| code1
== UNSPEC
)
1788 indx
= op0
; /* index + base */
1794 base
= op0
; /* base + displacement */
1799 else if (code0
== PLUS
)
1801 indx
= XEXP (op0
, 0); /* index + base + disp */
1802 base
= XEXP (op0
, 1);
1813 disp
= addr
; /* displacement */
1815 /* Extract integer part of displacement. */
1819 if (GET_CODE (disp
) == CONST_INT
)
1821 offset
= INTVAL (disp
);
1824 else if (GET_CODE (disp
) == CONST
1825 && GET_CODE (XEXP (disp
, 0)) == PLUS
1826 && GET_CODE (XEXP (XEXP (disp
, 0), 1)) == CONST_INT
)
1828 offset
= INTVAL (XEXP (XEXP (disp
, 0), 1));
1829 disp
= XEXP (XEXP (disp
, 0), 0);
1833 /* Strip off CONST here to avoid special case tests later. */
1834 if (disp
&& GET_CODE (disp
) == CONST
)
1835 disp
= XEXP (disp
, 0);
1837 /* We can convert literal pool addresses to
1838 displacements by basing them off the base register. */
1839 if (disp
&& GET_CODE (disp
) == SYMBOL_REF
&& CONSTANT_POOL_ADDRESS_P (disp
))
1841 /* Either base or index must be free to hold the base register. */
1843 base
= fake_pool_base
, literal_pool
= true;
1845 indx
= fake_pool_base
, literal_pool
= true;
1849 /* Mark up the displacement. */
1850 disp
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, disp
),
1851 UNSPEC_LTREL_OFFSET
);
1854 /* Validate base register. */
1857 if (GET_CODE (base
) == UNSPEC
)
1858 switch (XINT (base
, 1))
1862 disp
= gen_rtx_UNSPEC (Pmode
,
1863 gen_rtvec (1, XVECEXP (base
, 0, 0)),
1864 UNSPEC_LTREL_OFFSET
);
1868 base
= XVECEXP (base
, 0, 1);
1871 case UNSPEC_LTREL_BASE
:
1872 if (XVECLEN (base
, 0) == 1)
1873 base
= fake_pool_base
, literal_pool
= true;
1875 base
= XVECEXP (base
, 0, 1);
1883 || (GET_MODE (base
) != SImode
1884 && GET_MODE (base
) != Pmode
))
1887 if (REGNO (base
) == STACK_POINTER_REGNUM
1888 || REGNO (base
) == FRAME_POINTER_REGNUM
1889 || ((reload_completed
|| reload_in_progress
)
1890 && frame_pointer_needed
1891 && REGNO (base
) == HARD_FRAME_POINTER_REGNUM
)
1892 || REGNO (base
) == ARG_POINTER_REGNUM
1894 && REGNO (base
) == PIC_OFFSET_TABLE_REGNUM
))
1895 pointer
= base_ptr
= true;
1897 if ((reload_completed
|| reload_in_progress
)
1898 && base
== cfun
->machine
->base_reg
)
1899 pointer
= base_ptr
= literal_pool
= true;
1902 /* Validate index register. */
1905 if (GET_CODE (indx
) == UNSPEC
)
1906 switch (XINT (indx
, 1))
1910 disp
= gen_rtx_UNSPEC (Pmode
,
1911 gen_rtvec (1, XVECEXP (indx
, 0, 0)),
1912 UNSPEC_LTREL_OFFSET
);
1916 indx
= XVECEXP (indx
, 0, 1);
1919 case UNSPEC_LTREL_BASE
:
1920 if (XVECLEN (indx
, 0) == 1)
1921 indx
= fake_pool_base
, literal_pool
= true;
1923 indx
= XVECEXP (indx
, 0, 1);
1931 || (GET_MODE (indx
) != SImode
1932 && GET_MODE (indx
) != Pmode
))
1935 if (REGNO (indx
) == STACK_POINTER_REGNUM
1936 || REGNO (indx
) == FRAME_POINTER_REGNUM
1937 || ((reload_completed
|| reload_in_progress
)
1938 && frame_pointer_needed
1939 && REGNO (indx
) == HARD_FRAME_POINTER_REGNUM
)
1940 || REGNO (indx
) == ARG_POINTER_REGNUM
1942 && REGNO (indx
) == PIC_OFFSET_TABLE_REGNUM
))
1943 pointer
= indx_ptr
= true;
1945 if ((reload_completed
|| reload_in_progress
)
1946 && indx
== cfun
->machine
->base_reg
)
1947 pointer
= indx_ptr
= literal_pool
= true;
1950 /* Prefer to use pointer as base, not index. */
1951 if (base
&& indx
&& !base_ptr
1952 && (indx_ptr
|| (!REG_POINTER (base
) && REG_POINTER (indx
))))
1959 /* Validate displacement. */
1962 /* If virtual registers are involved, the displacement will change later
1963 anyway as the virtual registers get eliminated. This could make a
1964 valid displacement invalid, but it is more likely to make an invalid
1965 displacement valid, because we sometimes access the register save area
1966 via negative offsets to one of those registers.
1967 Thus we don't check the displacement for validity here. If after
1968 elimination the displacement turns out to be invalid after all,
1969 this is fixed up by reload in any case. */
1970 if (base
!= arg_pointer_rtx
1971 && indx
!= arg_pointer_rtx
1972 && base
!= return_address_pointer_rtx
1973 && indx
!= return_address_pointer_rtx
1974 && base
!= frame_pointer_rtx
1975 && indx
!= frame_pointer_rtx
1976 && base
!= virtual_stack_vars_rtx
1977 && indx
!= virtual_stack_vars_rtx
)
1978 if (!DISP_IN_RANGE (offset
))
1983 /* All the special cases are pointers. */
1986 /* In the small-PIC case, the linker converts @GOT
1987 and @GOTNTPOFF offsets to possible displacements. */
1988 if (GET_CODE (disp
) == UNSPEC
1989 && (XINT (disp
, 1) == UNSPEC_GOT
1990 || XINT (disp
, 1) == UNSPEC_GOTNTPOFF
)
1996 /* Accept pool label offsets. */
1997 else if (GET_CODE (disp
) == UNSPEC
1998 && XINT (disp
, 1) == UNSPEC_POOL_OFFSET
)
2001 /* Accept literal pool references. */
2002 else if (GET_CODE (disp
) == UNSPEC
2003 && XINT (disp
, 1) == UNSPEC_LTREL_OFFSET
)
2005 orig_disp
= gen_rtx_CONST (Pmode
, disp
);
2008 /* If we have an offset, make sure it does not
2009 exceed the size of the constant pool entry. */
2010 rtx sym
= XVECEXP (disp
, 0, 0);
2011 if (offset
>= GET_MODE_SIZE (get_pool_mode (sym
)))
2014 orig_disp
= plus_constant (orig_disp
, offset
);
2029 out
->disp
= orig_disp
;
2030 out
->pointer
= pointer
;
2031 out
->literal_pool
= literal_pool
;
2037 /* Decompose a RTL expression OP for a shift count into its components,
2038 and return the base register in BASE and the offset in OFFSET.
2040 Return true if OP is a valid shift count, false if not. */
2043 s390_decompose_shift_count (rtx op
, rtx
*base
, HOST_WIDE_INT
*offset
)
2045 HOST_WIDE_INT off
= 0;
2047 /* We can have an integer constant, an address register,
2048 or a sum of the two. */
2049 if (GET_CODE (op
) == CONST_INT
)
2054 if (op
&& GET_CODE (op
) == PLUS
&& GET_CODE (XEXP (op
, 1)) == CONST_INT
)
2056 off
= INTVAL (XEXP (op
, 1));
2059 while (op
&& GET_CODE (op
) == SUBREG
)
2060 op
= SUBREG_REG (op
);
2062 if (op
&& GET_CODE (op
) != REG
)
2074 /* Return true if CODE is a valid address without index. */
2077 s390_legitimate_address_without_index_p (rtx op
)
2079 struct s390_address addr
;
2081 if (!s390_decompose_address (XEXP (op
, 0), &addr
))
2090 /* Return true if ADDR is of kind symbol_ref or symbol_ref + const_int
2091 and return these parts in SYMREF and ADDEND. You can pass NULL in
2092 SYMREF and/or ADDEND if you are not interested in these values. */
2095 s390_symref_operand_p (rtx addr
, rtx
*symref
, HOST_WIDE_INT
*addend
)
2097 HOST_WIDE_INT tmpaddend
= 0;
2099 if (GET_CODE (addr
) == CONST
)
2100 addr
= XEXP (addr
, 0);
2102 if (GET_CODE (addr
) == PLUS
)
2104 if (GET_CODE (XEXP (addr
, 0)) == SYMBOL_REF
2105 && CONST_INT_P (XEXP (addr
, 1)))
2107 tmpaddend
= INTVAL (XEXP (addr
, 1));
2108 addr
= XEXP (addr
, 0);
2114 if (GET_CODE (addr
) != SYMBOL_REF
)
2120 *addend
= tmpaddend
;
2126 /* Return true if the address in OP is valid for constraint letter C
2127 if wrapped in a MEM rtx. Set LIT_POOL_OK to true if it literal
2128 pool MEMs should be accepted. Only the Q, R, S, T constraint
2129 letters are allowed for C. */
2132 s390_check_qrst_address (char c
, rtx op
, bool lit_pool_ok
)
2134 struct s390_address addr
;
2135 bool decomposed
= false;
2137 /* This check makes sure that no symbolic address (except literal
2138 pool references) are accepted by the R or T constraints. */
2139 if (s390_symref_operand_p (op
, NULL
, NULL
))
2143 if (!s390_decompose_address (op
, &addr
))
2145 if (!addr
.literal_pool
)
2152 case 'Q': /* no index short displacement */
2153 if (!decomposed
&& !s390_decompose_address (op
, &addr
))
2157 if (!s390_short_displacement (addr
.disp
))
2161 case 'R': /* with index short displacement */
2162 if (TARGET_LONG_DISPLACEMENT
)
2164 if (!decomposed
&& !s390_decompose_address (op
, &addr
))
2166 if (!s390_short_displacement (addr
.disp
))
2169 /* Any invalid address here will be fixed up by reload,
2170 so accept it for the most generic constraint. */
2173 case 'S': /* no index long displacement */
2174 if (!TARGET_LONG_DISPLACEMENT
)
2176 if (!decomposed
&& !s390_decompose_address (op
, &addr
))
2180 if (s390_short_displacement (addr
.disp
))
2184 case 'T': /* with index long displacement */
2185 if (!TARGET_LONG_DISPLACEMENT
)
2187 /* Any invalid address here will be fixed up by reload,
2188 so accept it for the most generic constraint. */
2189 if ((decomposed
|| s390_decompose_address (op
, &addr
))
2190 && s390_short_displacement (addr
.disp
))
2200 /* Evaluates constraint strings described by the regular expression
2201 ([A|B|Z](Q|R|S|T))|U|W|Y and returns 1 if OP is a valid operand for
2202 the constraint given in STR, or 0 else. */
2205 s390_mem_constraint (const char *str
, rtx op
)
2212 /* Check for offsettable variants of memory constraints. */
2213 if (!MEM_P (op
) || MEM_VOLATILE_P (op
))
2215 if ((reload_completed
|| reload_in_progress
)
2216 ? !offsettable_memref_p (op
) : !offsettable_nonstrict_memref_p (op
))
2218 return s390_check_qrst_address (str
[1], XEXP (op
, 0), true);
2220 /* Check for non-literal-pool variants of memory constraints. */
2223 return s390_check_qrst_address (str
[1], XEXP (op
, 0), false);
2228 if (GET_CODE (op
) != MEM
)
2230 return s390_check_qrst_address (c
, XEXP (op
, 0), true);
2232 return (s390_check_qrst_address ('Q', op
, true)
2233 || s390_check_qrst_address ('R', op
, true));
2235 return (s390_check_qrst_address ('S', op
, true)
2236 || s390_check_qrst_address ('T', op
, true));
2238 /* Simply check for the basic form of a shift count. Reload will
2239 take care of making sure we have a proper base register. */
2240 if (!s390_decompose_shift_count (op
, NULL
, NULL
))
2244 return s390_check_qrst_address (str
[1], op
, true);
2252 /* Evaluates constraint strings starting with letter O. Input
2253 parameter C is the second letter following the "O" in the constraint
2254 string. Returns 1 if VALUE meets the respective constraint and 0
2258 s390_O_constraint_str (const char c
, HOST_WIDE_INT value
)
2266 return trunc_int_for_mode (value
, SImode
) == value
;
2270 || s390_single_part (GEN_INT (value
), DImode
, SImode
, 0) == 1;
2273 return s390_single_part (GEN_INT (value
- 1), DImode
, SImode
, -1) == 1;
2281 /* Evaluates constraint strings starting with letter N. Parameter STR
2282 contains the letters following letter "N" in the constraint string.
2283 Returns true if VALUE matches the constraint. */
2286 s390_N_constraint_str (const char *str
, HOST_WIDE_INT value
)
2288 enum machine_mode mode
, part_mode
;
2290 int part
, part_goal
;
2296 part_goal
= str
[0] - '0';
2340 if (GET_MODE_SIZE (mode
) <= GET_MODE_SIZE (part_mode
))
2343 part
= s390_single_part (GEN_INT (value
), mode
, part_mode
, def
);
2346 if (part_goal
!= -1 && part_goal
!= part
)
2353 /* Returns true if the input parameter VALUE is a float zero. */
2356 s390_float_const_zero_p (rtx value
)
2358 return (GET_MODE_CLASS (GET_MODE (value
)) == MODE_FLOAT
2359 && value
== CONST0_RTX (GET_MODE (value
)));
2363 /* Compute a (partial) cost for rtx X. Return true if the complete
2364 cost has been computed, and false if subexpressions should be
2365 scanned. In either case, *TOTAL contains the cost result.
2366 CODE contains GET_CODE (x), OUTER_CODE contains the code
2367 of the superexpression of x. */
2370 s390_rtx_costs (rtx x
, int code
, int outer_code
, int *total
,
2371 bool speed ATTRIBUTE_UNUSED
)
2394 *total
= COSTS_N_INSNS (1);
2399 /* Check for multiply and add. */
2400 if ((GET_MODE (x
) == DFmode
|| GET_MODE (x
) == SFmode
)
2401 && GET_CODE (XEXP (x
, 0)) == MULT
2402 && TARGET_HARD_FLOAT
&& TARGET_FUSED_MADD
)
2404 /* This is the multiply and add case. */
2405 if (GET_MODE (x
) == DFmode
)
2406 *total
= s390_cost
->madbr
;
2408 *total
= s390_cost
->maebr
;
2409 *total
+= (rtx_cost (XEXP (XEXP (x
, 0), 0), MULT
, speed
)
2410 + rtx_cost (XEXP (XEXP (x
, 0), 1), MULT
, speed
)
2411 + rtx_cost (XEXP (x
, 1), (enum rtx_code
) code
, speed
));
2412 return true; /* Do not do an additional recursive descent. */
2414 *total
= COSTS_N_INSNS (1);
2418 switch (GET_MODE (x
))
2422 rtx left
= XEXP (x
, 0);
2423 rtx right
= XEXP (x
, 1);
2424 if (GET_CODE (right
) == CONST_INT
2425 && CONST_OK_FOR_K (INTVAL (right
)))
2426 *total
= s390_cost
->mhi
;
2427 else if (GET_CODE (left
) == SIGN_EXTEND
)
2428 *total
= s390_cost
->mh
;
2430 *total
= s390_cost
->ms
; /* msr, ms, msy */
2435 rtx left
= XEXP (x
, 0);
2436 rtx right
= XEXP (x
, 1);
2439 if (GET_CODE (right
) == CONST_INT
2440 && CONST_OK_FOR_K (INTVAL (right
)))
2441 *total
= s390_cost
->mghi
;
2442 else if (GET_CODE (left
) == SIGN_EXTEND
)
2443 *total
= s390_cost
->msgf
;
2445 *total
= s390_cost
->msg
; /* msgr, msg */
2447 else /* TARGET_31BIT */
2449 if (GET_CODE (left
) == SIGN_EXTEND
2450 && GET_CODE (right
) == SIGN_EXTEND
)
2451 /* mulsidi case: mr, m */
2452 *total
= s390_cost
->m
;
2453 else if (GET_CODE (left
) == ZERO_EXTEND
2454 && GET_CODE (right
) == ZERO_EXTEND
2455 && TARGET_CPU_ZARCH
)
2456 /* umulsidi case: ml, mlr */
2457 *total
= s390_cost
->ml
;
2459 /* Complex calculation is required. */
2460 *total
= COSTS_N_INSNS (40);
2466 *total
= s390_cost
->mult_df
;
2469 *total
= s390_cost
->mxbr
;
2478 if (GET_MODE (x
) == TImode
) /* 128 bit division */
2479 *total
= s390_cost
->dlgr
;
2480 else if (GET_MODE (x
) == DImode
)
2482 rtx right
= XEXP (x
, 1);
2483 if (GET_CODE (right
) == ZERO_EXTEND
) /* 64 by 32 bit division */
2484 *total
= s390_cost
->dlr
;
2485 else /* 64 by 64 bit division */
2486 *total
= s390_cost
->dlgr
;
2488 else if (GET_MODE (x
) == SImode
) /* 32 bit division */
2489 *total
= s390_cost
->dlr
;
2494 if (GET_MODE (x
) == DImode
)
2496 rtx right
= XEXP (x
, 1);
2497 if (GET_CODE (right
) == ZERO_EXTEND
) /* 64 by 32 bit division */
2499 *total
= s390_cost
->dsgfr
;
2501 *total
= s390_cost
->dr
;
2502 else /* 64 by 64 bit division */
2503 *total
= s390_cost
->dsgr
;
2505 else if (GET_MODE (x
) == SImode
) /* 32 bit division */
2506 *total
= s390_cost
->dlr
;
2507 else if (GET_MODE (x
) == SFmode
)
2509 *total
= s390_cost
->debr
;
2511 else if (GET_MODE (x
) == DFmode
)
2513 *total
= s390_cost
->ddbr
;
2515 else if (GET_MODE (x
) == TFmode
)
2517 *total
= s390_cost
->dxbr
;
2522 if (GET_MODE (x
) == SFmode
)
2523 *total
= s390_cost
->sqebr
;
2524 else if (GET_MODE (x
) == DFmode
)
2525 *total
= s390_cost
->sqdbr
;
2527 *total
= s390_cost
->sqxbr
;
2532 if (outer_code
== MULT
|| outer_code
== DIV
|| outer_code
== MOD
2533 || outer_code
== PLUS
|| outer_code
== MINUS
2534 || outer_code
== COMPARE
)
2539 *total
= COSTS_N_INSNS (1);
2540 if (GET_CODE (XEXP (x
, 0)) == AND
2541 && GET_CODE (XEXP (x
, 1)) == CONST_INT
2542 && GET_CODE (XEXP (XEXP (x
, 0), 1)) == CONST_INT
)
2544 rtx op0
= XEXP (XEXP (x
, 0), 0);
2545 rtx op1
= XEXP (XEXP (x
, 0), 1);
2546 rtx op2
= XEXP (x
, 1);
2548 if (memory_operand (op0
, GET_MODE (op0
))
2549 && s390_tm_ccmode (op1
, op2
, 0) != VOIDmode
)
2551 if (register_operand (op0
, GET_MODE (op0
))
2552 && s390_tm_ccmode (op1
, op2
, 1) != VOIDmode
)
2562 /* Return the cost of an address rtx ADDR. */
2565 s390_address_cost (rtx addr
, bool speed ATTRIBUTE_UNUSED
)
2567 struct s390_address ad
;
2568 if (!s390_decompose_address (addr
, &ad
))
2571 return ad
.indx
? COSTS_N_INSNS (1) + 1 : COSTS_N_INSNS (1);
2574 /* If OP is a SYMBOL_REF of a thread-local symbol, return its TLS mode,
2575 otherwise return 0. */
2578 tls_symbolic_operand (rtx op
)
2580 if (GET_CODE (op
) != SYMBOL_REF
)
2582 return SYMBOL_REF_TLS_MODEL (op
);
2585 /* Split DImode access register reference REG (on 64-bit) into its constituent
2586 low and high parts, and store them into LO and HI. Note that gen_lowpart/
2587 gen_highpart cannot be used as they assume all registers are word-sized,
2588 while our access registers have only half that size. */
2591 s390_split_access_reg (rtx reg
, rtx
*lo
, rtx
*hi
)
2593 gcc_assert (TARGET_64BIT
);
2594 gcc_assert (ACCESS_REG_P (reg
));
2595 gcc_assert (GET_MODE (reg
) == DImode
);
2596 gcc_assert (!(REGNO (reg
) & 1));
2598 *lo
= gen_rtx_REG (SImode
, REGNO (reg
) + 1);
2599 *hi
= gen_rtx_REG (SImode
, REGNO (reg
));
2602 /* Return true if OP contains a symbol reference */
2605 symbolic_reference_mentioned_p (rtx op
)
2610 if (GET_CODE (op
) == SYMBOL_REF
|| GET_CODE (op
) == LABEL_REF
)
2613 fmt
= GET_RTX_FORMAT (GET_CODE (op
));
2614 for (i
= GET_RTX_LENGTH (GET_CODE (op
)) - 1; i
>= 0; i
--)
2620 for (j
= XVECLEN (op
, i
) - 1; j
>= 0; j
--)
2621 if (symbolic_reference_mentioned_p (XVECEXP (op
, i
, j
)))
2625 else if (fmt
[i
] == 'e' && symbolic_reference_mentioned_p (XEXP (op
, i
)))
2632 /* Return true if OP contains a reference to a thread-local symbol. */
2635 tls_symbolic_reference_mentioned_p (rtx op
)
2640 if (GET_CODE (op
) == SYMBOL_REF
)
2641 return tls_symbolic_operand (op
);
2643 fmt
= GET_RTX_FORMAT (GET_CODE (op
));
2644 for (i
= GET_RTX_LENGTH (GET_CODE (op
)) - 1; i
>= 0; i
--)
2650 for (j
= XVECLEN (op
, i
) - 1; j
>= 0; j
--)
2651 if (tls_symbolic_reference_mentioned_p (XVECEXP (op
, i
, j
)))
2655 else if (fmt
[i
] == 'e' && tls_symbolic_reference_mentioned_p (XEXP (op
, i
)))
2663 /* Return true if OP is a legitimate general operand when
2664 generating PIC code. It is given that flag_pic is on
2665 and that OP satisfies CONSTANT_P or is a CONST_DOUBLE. */
2668 legitimate_pic_operand_p (rtx op
)
2670 /* Accept all non-symbolic constants. */
2671 if (!SYMBOLIC_CONST (op
))
2674 /* Reject everything else; must be handled
2675 via emit_symbolic_move. */
2679 /* Returns true if the constant value OP is a legitimate general operand.
2680 It is given that OP satisfies CONSTANT_P or is a CONST_DOUBLE. */
2683 legitimate_constant_p (rtx op
)
2685 /* Accept all non-symbolic constants. */
2686 if (!SYMBOLIC_CONST (op
))
2689 /* Accept immediate LARL operands. */
2690 if (TARGET_CPU_ZARCH
&& larl_operand (op
, VOIDmode
))
2693 /* Thread-local symbols are never legal constants. This is
2694 so that emit_call knows that computing such addresses
2695 might require a function call. */
2696 if (TLS_SYMBOLIC_CONST (op
))
2699 /* In the PIC case, symbolic constants must *not* be
2700 forced into the literal pool. We accept them here,
2701 so that they will be handled by emit_symbolic_move. */
2705 /* All remaining non-PIC symbolic constants are
2706 forced into the literal pool. */
2710 /* Determine if it's legal to put X into the constant pool. This
2711 is not possible if X contains the address of a symbol that is
2712 not constant (TLS) or not known at final link time (PIC). */
2715 s390_cannot_force_const_mem (rtx x
)
2717 switch (GET_CODE (x
))
2721 /* Accept all non-symbolic constants. */
2725 /* Labels are OK iff we are non-PIC. */
2726 return flag_pic
!= 0;
2729 /* 'Naked' TLS symbol references are never OK,
2730 non-TLS symbols are OK iff we are non-PIC. */
2731 if (tls_symbolic_operand (x
))
2734 return flag_pic
!= 0;
2737 return s390_cannot_force_const_mem (XEXP (x
, 0));
2740 return s390_cannot_force_const_mem (XEXP (x
, 0))
2741 || s390_cannot_force_const_mem (XEXP (x
, 1));
2744 switch (XINT (x
, 1))
2746 /* Only lt-relative or GOT-relative UNSPECs are OK. */
2747 case UNSPEC_LTREL_OFFSET
:
2755 case UNSPEC_GOTNTPOFF
:
2756 case UNSPEC_INDNTPOFF
:
2759 /* If the literal pool shares the code section, be put
2760 execute template placeholders into the pool as well. */
2762 return TARGET_CPU_ZARCH
;
2774 /* Returns true if the constant value OP is a legitimate general
2775 operand during and after reload. The difference to
2776 legitimate_constant_p is that this function will not accept
2777 a constant that would need to be forced to the literal pool
2778 before it can be used as operand. */
2781 legitimate_reload_constant_p (rtx op
)
2783 /* Accept la(y) operands. */
2784 if (GET_CODE (op
) == CONST_INT
2785 && DISP_IN_RANGE (INTVAL (op
)))
2788 /* Accept l(g)hi/l(g)fi operands. */
2789 if (GET_CODE (op
) == CONST_INT
2790 && (CONST_OK_FOR_K (INTVAL (op
)) || CONST_OK_FOR_Os (INTVAL (op
))))
2793 /* Accept lliXX operands. */
2795 && GET_CODE (op
) == CONST_INT
2796 && trunc_int_for_mode (INTVAL (op
), word_mode
) == INTVAL (op
)
2797 && s390_single_part (op
, word_mode
, HImode
, 0) >= 0)
2801 && GET_CODE (op
) == CONST_INT
2802 && trunc_int_for_mode (INTVAL (op
), word_mode
) == INTVAL (op
)
2803 && s390_single_part (op
, word_mode
, SImode
, 0) >= 0)
2806 /* Accept larl operands. */
2807 if (TARGET_CPU_ZARCH
2808 && larl_operand (op
, VOIDmode
))
2811 /* Accept lzXX operands. */
2812 if (GET_CODE (op
) == CONST_DOUBLE
2813 && CONST_DOUBLE_OK_FOR_CONSTRAINT_P (op
, 'G', "G"))
2816 /* Accept double-word operands that can be split. */
2817 if (GET_CODE (op
) == CONST_INT
2818 && trunc_int_for_mode (INTVAL (op
), word_mode
) != INTVAL (op
))
2820 enum machine_mode dword_mode
= word_mode
== SImode
? DImode
: TImode
;
2821 rtx hi
= operand_subword (op
, 0, 0, dword_mode
);
2822 rtx lo
= operand_subword (op
, 1, 0, dword_mode
);
2823 return legitimate_reload_constant_p (hi
)
2824 && legitimate_reload_constant_p (lo
);
2827 /* Everything else cannot be handled without reload. */
2831 /* Given an rtx OP being reloaded into a reg required to be in class RCLASS,
2832 return the class of reg to actually use. */
2835 s390_preferred_reload_class (rtx op
, enum reg_class rclass
)
2837 switch (GET_CODE (op
))
2839 /* Constants we cannot reload must be forced into the
2844 if (legitimate_reload_constant_p (op
))
2849 /* If a symbolic constant or a PLUS is reloaded,
2850 it is most likely being used as an address, so
2851 prefer ADDR_REGS. If 'class' is not a superset
2852 of ADDR_REGS, e.g. FP_REGS, reject this reload. */
2857 if (reg_class_subset_p (ADDR_REGS
, rclass
))
2869 /* Return true if ADDR is SYMBOL_REF + addend with addend being a
2870 multiple of ALIGNMENT and the SYMBOL_REF being naturally
2874 s390_check_symref_alignment (rtx addr
, HOST_WIDE_INT alignment
)
2876 HOST_WIDE_INT addend
;
2879 if (!s390_symref_operand_p (addr
, &symref
, &addend
))
2882 return (!SYMBOL_REF_NOT_NATURALLY_ALIGNED_P (symref
)
2883 && !(addend
& (alignment
- 1)));
2886 /* ADDR is moved into REG using larl. If ADDR isn't a valid larl
2887 operand SCRATCH is used to reload the even part of the address and
2891 s390_reload_larl_operand (rtx reg
, rtx addr
, rtx scratch
)
2893 HOST_WIDE_INT addend
;
2896 if (!s390_symref_operand_p (addr
, &symref
, &addend
))
2900 /* Easy case. The addend is even so larl will do fine. */
2901 emit_move_insn (reg
, addr
);
2904 /* We can leave the scratch register untouched if the target
2905 register is a valid base register. */
2906 if (REGNO (reg
) < FIRST_PSEUDO_REGISTER
2907 && REGNO_REG_CLASS (REGNO (reg
)) == ADDR_REGS
)
2910 gcc_assert (REGNO (scratch
) < FIRST_PSEUDO_REGISTER
);
2911 gcc_assert (REGNO_REG_CLASS (REGNO (scratch
)) == ADDR_REGS
);
2914 emit_move_insn (scratch
,
2915 gen_rtx_CONST (Pmode
,
2916 gen_rtx_PLUS (Pmode
, symref
,
2917 GEN_INT (addend
- 1))));
2919 emit_move_insn (scratch
, symref
);
2921 /* Increment the address using la in order to avoid clobbering cc. */
2922 emit_move_insn (reg
, gen_rtx_PLUS (Pmode
, scratch
, const1_rtx
));
2926 /* Generate what is necessary to move between REG and MEM using
2927 SCRATCH. The direction is given by TOMEM. */
2930 s390_reload_symref_address (rtx reg
, rtx mem
, rtx scratch
, bool tomem
)
2932 /* Reload might have pulled a constant out of the literal pool.
2933 Force it back in. */
2934 if (CONST_INT_P (mem
) || GET_CODE (mem
) == CONST_DOUBLE
2935 || GET_CODE (mem
) == CONST
)
2936 mem
= force_const_mem (GET_MODE (reg
), mem
);
2938 gcc_assert (MEM_P (mem
));
2940 /* For a load from memory we can leave the scratch register
2941 untouched if the target register is a valid base register. */
2943 && REGNO (reg
) < FIRST_PSEUDO_REGISTER
2944 && REGNO_REG_CLASS (REGNO (reg
)) == ADDR_REGS
2945 && GET_MODE (reg
) == GET_MODE (scratch
))
2948 /* Load address into scratch register. Since we can't have a
2949 secondary reload for a secondary reload we have to cover the case
2950 where larl would need a secondary reload here as well. */
2951 s390_reload_larl_operand (scratch
, XEXP (mem
, 0), scratch
);
2953 /* Now we can use a standard load/store to do the move. */
2955 emit_move_insn (replace_equiv_address (mem
, scratch
), reg
);
2957 emit_move_insn (reg
, replace_equiv_address (mem
, scratch
));
2960 /* Inform reload about cases where moving X with a mode MODE to a register in
2961 RCLASS requires an extra scratch or immediate register. Return the class
2962 needed for the immediate register. */
2964 static enum reg_class
2965 s390_secondary_reload (bool in_p
, rtx x
, enum reg_class rclass
,
2966 enum machine_mode mode
, secondary_reload_info
*sri
)
2968 /* Intermediate register needed. */
2969 if (reg_classes_intersect_p (CC_REGS
, rclass
))
2970 return GENERAL_REGS
;
2974 /* On z10 several optimizer steps may generate larl operands with
2977 && s390_symref_operand_p (x
, NULL
, NULL
)
2979 && !s390_check_symref_alignment (x
, 2))
2980 sri
->icode
= ((mode
== DImode
) ? CODE_FOR_reloaddi_larl_odd_addend_z10
2981 : CODE_FOR_reloadsi_larl_odd_addend_z10
);
2983 /* On z10 we need a scratch register when moving QI, TI or floating
2984 point mode values from or to a memory location with a SYMBOL_REF
2985 or if the symref addend of a SI or DI move is not aligned to the
2986 width of the access. */
2988 && s390_symref_operand_p (XEXP (x
, 0), NULL
, NULL
)
2989 && (mode
== QImode
|| mode
== TImode
|| FLOAT_MODE_P (mode
)
2990 || (!TARGET_ZARCH
&& mode
== DImode
)
2991 || ((mode
== HImode
|| mode
== SImode
|| mode
== DImode
)
2992 && (!s390_check_symref_alignment (XEXP (x
, 0),
2993 GET_MODE_SIZE (mode
))))))
2995 #define __SECONDARY_RELOAD_CASE(M,m) \
2998 sri->icode = in_p ? CODE_FOR_reload##m##di_toreg_z10 : \
2999 CODE_FOR_reload##m##di_tomem_z10; \
3001 sri->icode = in_p ? CODE_FOR_reload##m##si_toreg_z10 : \
3002 CODE_FOR_reload##m##si_tomem_z10; \
3005 switch (GET_MODE (x
))
3007 __SECONDARY_RELOAD_CASE (QI
, qi
);
3008 __SECONDARY_RELOAD_CASE (HI
, hi
);
3009 __SECONDARY_RELOAD_CASE (SI
, si
);
3010 __SECONDARY_RELOAD_CASE (DI
, di
);
3011 __SECONDARY_RELOAD_CASE (TI
, ti
);
3012 __SECONDARY_RELOAD_CASE (SF
, sf
);
3013 __SECONDARY_RELOAD_CASE (DF
, df
);
3014 __SECONDARY_RELOAD_CASE (TF
, tf
);
3015 __SECONDARY_RELOAD_CASE (SD
, sd
);
3016 __SECONDARY_RELOAD_CASE (DD
, dd
);
3017 __SECONDARY_RELOAD_CASE (TD
, td
);
3022 #undef __SECONDARY_RELOAD_CASE
3026 /* We need a scratch register when loading a PLUS expression which
3027 is not a legitimate operand of the LOAD ADDRESS instruction. */
3028 if (in_p
&& s390_plus_operand (x
, mode
))
3029 sri
->icode
= (TARGET_64BIT
?
3030 CODE_FOR_reloaddi_plus
: CODE_FOR_reloadsi_plus
);
3032 /* Performing a multiword move from or to memory we have to make sure the
3033 second chunk in memory is addressable without causing a displacement
3034 overflow. If that would be the case we calculate the address in
3035 a scratch register. */
3037 && GET_CODE (XEXP (x
, 0)) == PLUS
3038 && GET_CODE (XEXP (XEXP (x
, 0), 1)) == CONST_INT
3039 && !DISP_IN_RANGE (INTVAL (XEXP (XEXP (x
, 0), 1))
3040 + GET_MODE_SIZE (mode
) - 1))
3042 /* For GENERAL_REGS a displacement overflow is no problem if occurring
3043 in a s_operand address since we may fallback to lm/stm. So we only
3044 have to care about overflows in the b+i+d case. */
3045 if ((reg_classes_intersect_p (GENERAL_REGS
, rclass
)
3046 && s390_class_max_nregs (GENERAL_REGS
, mode
) > 1
3047 && GET_CODE (XEXP (XEXP (x
, 0), 0)) == PLUS
)
3048 /* For FP_REGS no lm/stm is available so this check is triggered
3049 for displacement overflows in b+i+d and b+d like addresses. */
3050 || (reg_classes_intersect_p (FP_REGS
, rclass
)
3051 && s390_class_max_nregs (FP_REGS
, mode
) > 1))
3054 sri
->icode
= (TARGET_64BIT
?
3055 CODE_FOR_reloaddi_nonoffmem_in
:
3056 CODE_FOR_reloadsi_nonoffmem_in
);
3058 sri
->icode
= (TARGET_64BIT
?
3059 CODE_FOR_reloaddi_nonoffmem_out
:
3060 CODE_FOR_reloadsi_nonoffmem_out
);
3064 /* A scratch address register is needed when a symbolic constant is
3065 copied to r0 compiling with -fPIC. In other cases the target
3066 register might be used as temporary (see legitimize_pic_address). */
3067 if (in_p
&& SYMBOLIC_CONST (x
) && flag_pic
== 2 && rclass
!= ADDR_REGS
)
3068 sri
->icode
= (TARGET_64BIT
?
3069 CODE_FOR_reloaddi_PIC_addr
:
3070 CODE_FOR_reloadsi_PIC_addr
);
3072 /* Either scratch or no register needed. */
3076 /* Generate code to load SRC, which is PLUS that is not a
3077 legitimate operand for the LA instruction, into TARGET.
3078 SCRATCH may be used as scratch register. */
3081 s390_expand_plus_operand (rtx target
, rtx src
,
3085 struct s390_address ad
;
3087 /* src must be a PLUS; get its two operands. */
3088 gcc_assert (GET_CODE (src
) == PLUS
);
3089 gcc_assert (GET_MODE (src
) == Pmode
);
3091 /* Check if any of the two operands is already scheduled
3092 for replacement by reload. This can happen e.g. when
3093 float registers occur in an address. */
3094 sum1
= find_replacement (&XEXP (src
, 0));
3095 sum2
= find_replacement (&XEXP (src
, 1));
3096 src
= gen_rtx_PLUS (Pmode
, sum1
, sum2
);
3098 /* If the address is already strictly valid, there's nothing to do. */
3099 if (!s390_decompose_address (src
, &ad
)
3100 || (ad
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)))
3101 || (ad
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (ad
.indx
))))
3103 /* Otherwise, one of the operands cannot be an address register;
3104 we reload its value into the scratch register. */
3105 if (true_regnum (sum1
) < 1 || true_regnum (sum1
) > 15)
3107 emit_move_insn (scratch
, sum1
);
3110 if (true_regnum (sum2
) < 1 || true_regnum (sum2
) > 15)
3112 emit_move_insn (scratch
, sum2
);
3116 /* According to the way these invalid addresses are generated
3117 in reload.c, it should never happen (at least on s390) that
3118 *neither* of the PLUS components, after find_replacements
3119 was applied, is an address register. */
3120 if (sum1
== scratch
&& sum2
== scratch
)
3126 src
= gen_rtx_PLUS (Pmode
, sum1
, sum2
);
3129 /* Emit the LOAD ADDRESS pattern. Note that reload of PLUS
3130 is only ever performed on addresses, so we can mark the
3131 sum as legitimate for LA in any case. */
3132 s390_load_address (target
, src
);
3136 /* Return true if ADDR is a valid memory address.
3137 STRICT specifies whether strict register checking applies. */
3140 s390_legitimate_address_p (enum machine_mode mode
, rtx addr
, bool strict
)
3142 struct s390_address ad
;
3145 && larl_operand (addr
, VOIDmode
)
3146 && (mode
== VOIDmode
3147 || s390_check_symref_alignment (addr
, GET_MODE_SIZE (mode
))))
3150 if (!s390_decompose_address (addr
, &ad
))
3155 if (ad
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)))
3158 if (ad
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (ad
.indx
)))
3164 && !(REGNO (ad
.base
) >= FIRST_PSEUDO_REGISTER
3165 || REGNO_REG_CLASS (REGNO (ad
.base
)) == ADDR_REGS
))
3169 && !(REGNO (ad
.indx
) >= FIRST_PSEUDO_REGISTER
3170 || REGNO_REG_CLASS (REGNO (ad
.indx
)) == ADDR_REGS
))
3176 /* Return true if OP is a valid operand for the LA instruction.
3177 In 31-bit, we need to prove that the result is used as an
3178 address, as LA performs only a 31-bit addition. */
3181 legitimate_la_operand_p (rtx op
)
3183 struct s390_address addr
;
3184 if (!s390_decompose_address (op
, &addr
))
3187 return (TARGET_64BIT
|| addr
.pointer
);
3190 /* Return true if it is valid *and* preferable to use LA to
3191 compute the sum of OP1 and OP2. */
3194 preferred_la_operand_p (rtx op1
, rtx op2
)
3196 struct s390_address addr
;
3198 if (op2
!= const0_rtx
)
3199 op1
= gen_rtx_PLUS (Pmode
, op1
, op2
);
3201 if (!s390_decompose_address (op1
, &addr
))
3203 if (addr
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (addr
.base
)))
3205 if (addr
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (addr
.indx
)))
3208 if (!TARGET_64BIT
&& !addr
.pointer
)
3214 if ((addr
.base
&& REG_P (addr
.base
) && REG_POINTER (addr
.base
))
3215 || (addr
.indx
&& REG_P (addr
.indx
) && REG_POINTER (addr
.indx
)))
3221 /* Emit a forced load-address operation to load SRC into DST.
3222 This will use the LOAD ADDRESS instruction even in situations
3223 where legitimate_la_operand_p (SRC) returns false. */
3226 s390_load_address (rtx dst
, rtx src
)
3229 emit_move_insn (dst
, src
);
3231 emit_insn (gen_force_la_31 (dst
, src
));
3234 /* Return a legitimate reference for ORIG (an address) using the
3235 register REG. If REG is 0, a new pseudo is generated.
3237 There are two types of references that must be handled:
3239 1. Global data references must load the address from the GOT, via
3240 the PIC reg. An insn is emitted to do this load, and the reg is
3243 2. Static data references, constant pool addresses, and code labels
3244 compute the address as an offset from the GOT, whose base is in
3245 the PIC reg. Static data objects have SYMBOL_FLAG_LOCAL set to
3246 differentiate them from global data objects. The returned
3247 address is the PIC reg + an unspec constant.
3249 TARGET_LEGITIMIZE_ADDRESS_P rejects symbolic references unless the PIC
3250 reg also appears in the address. */
3253 legitimize_pic_address (rtx orig
, rtx reg
)
3259 gcc_assert (!TLS_SYMBOLIC_CONST (addr
));
3261 if (GET_CODE (addr
) == LABEL_REF
3262 || (GET_CODE (addr
) == SYMBOL_REF
&& SYMBOL_REF_LOCAL_P (addr
)))
3264 /* This is a local symbol. */
3265 if (TARGET_CPU_ZARCH
&& larl_operand (addr
, VOIDmode
))
3267 /* Access local symbols PC-relative via LARL.
3268 This is the same as in the non-PIC case, so it is
3269 handled automatically ... */
3273 /* Access local symbols relative to the GOT. */
3275 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3277 if (reload_in_progress
|| reload_completed
)
3278 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3280 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTOFF
);
3281 addr
= gen_rtx_CONST (Pmode
, addr
);
3282 addr
= force_const_mem (Pmode
, addr
);
3283 emit_move_insn (temp
, addr
);
3285 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3288 s390_load_address (reg
, new_rtx
);
3293 else if (GET_CODE (addr
) == SYMBOL_REF
)
3296 reg
= gen_reg_rtx (Pmode
);
3300 /* Assume GOT offset < 4k. This is handled the same way
3301 in both 31- and 64-bit code (@GOT). */
3303 if (reload_in_progress
|| reload_completed
)
3304 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3306 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOT
);
3307 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3308 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, new_rtx
);
3309 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3310 emit_move_insn (reg
, new_rtx
);
3313 else if (TARGET_CPU_ZARCH
)
3315 /* If the GOT offset might be >= 4k, we determine the position
3316 of the GOT entry via a PC-relative LARL (@GOTENT). */
3318 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3320 gcc_assert (REGNO (temp
) >= FIRST_PSEUDO_REGISTER
3321 || REGNO_REG_CLASS (REGNO (temp
)) == ADDR_REGS
);
3323 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTENT
);
3324 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3325 emit_move_insn (temp
, new_rtx
);
3327 new_rtx
= gen_const_mem (Pmode
, temp
);
3328 emit_move_insn (reg
, new_rtx
);
3333 /* If the GOT offset might be >= 4k, we have to load it
3334 from the literal pool (@GOT). */
3336 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3338 gcc_assert (REGNO (temp
) >= FIRST_PSEUDO_REGISTER
3339 || REGNO_REG_CLASS (REGNO (temp
)) == ADDR_REGS
);
3341 if (reload_in_progress
|| reload_completed
)
3342 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3344 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOT
);
3345 addr
= gen_rtx_CONST (Pmode
, addr
);
3346 addr
= force_const_mem (Pmode
, addr
);
3347 emit_move_insn (temp
, addr
);
3349 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3350 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3351 emit_move_insn (reg
, new_rtx
);
3357 if (GET_CODE (addr
) == CONST
)
3359 addr
= XEXP (addr
, 0);
3360 if (GET_CODE (addr
) == UNSPEC
)
3362 gcc_assert (XVECLEN (addr
, 0) == 1);
3363 switch (XINT (addr
, 1))
3365 /* If someone moved a GOT-relative UNSPEC
3366 out of the literal pool, force them back in. */
3369 new_rtx
= force_const_mem (Pmode
, orig
);
3372 /* @GOT is OK as is if small. */
3375 new_rtx
= force_const_mem (Pmode
, orig
);
3378 /* @GOTENT is OK as is. */
3382 /* @PLT is OK as is on 64-bit, must be converted to
3383 GOT-relative @PLTOFF on 31-bit. */
3385 if (!TARGET_CPU_ZARCH
)
3387 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3389 if (reload_in_progress
|| reload_completed
)
3390 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3392 addr
= XVECEXP (addr
, 0, 0);
3393 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
),
3395 addr
= gen_rtx_CONST (Pmode
, addr
);
3396 addr
= force_const_mem (Pmode
, addr
);
3397 emit_move_insn (temp
, addr
);
3399 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3402 s390_load_address (reg
, new_rtx
);
3408 /* Everything else cannot happen. */
3414 gcc_assert (GET_CODE (addr
) == PLUS
);
3416 if (GET_CODE (addr
) == PLUS
)
3418 rtx op0
= XEXP (addr
, 0), op1
= XEXP (addr
, 1);
3420 gcc_assert (!TLS_SYMBOLIC_CONST (op0
));
3421 gcc_assert (!TLS_SYMBOLIC_CONST (op1
));
3423 /* Check first to see if this is a constant offset
3424 from a local symbol reference. */
3425 if ((GET_CODE (op0
) == LABEL_REF
3426 || (GET_CODE (op0
) == SYMBOL_REF
&& SYMBOL_REF_LOCAL_P (op0
)))
3427 && GET_CODE (op1
) == CONST_INT
)
3429 if (TARGET_CPU_ZARCH
3430 && larl_operand (op0
, VOIDmode
)
3431 && INTVAL (op1
) < (HOST_WIDE_INT
)1 << 31
3432 && INTVAL (op1
) >= -((HOST_WIDE_INT
)1 << 31))
3434 if (INTVAL (op1
) & 1)
3436 /* LARL can't handle odd offsets, so emit a
3437 pair of LARL and LA. */
3438 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3440 if (!DISP_IN_RANGE (INTVAL (op1
)))
3442 HOST_WIDE_INT even
= INTVAL (op1
) - 1;
3443 op0
= gen_rtx_PLUS (Pmode
, op0
, GEN_INT (even
));
3444 op0
= gen_rtx_CONST (Pmode
, op0
);
3448 emit_move_insn (temp
, op0
);
3449 new_rtx
= gen_rtx_PLUS (Pmode
, temp
, op1
);
3453 s390_load_address (reg
, new_rtx
);
3459 /* If the offset is even, we can just use LARL.
3460 This will happen automatically. */
3465 /* Access local symbols relative to the GOT. */
3467 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3469 if (reload_in_progress
|| reload_completed
)
3470 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3472 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, op0
),
3474 addr
= gen_rtx_PLUS (Pmode
, addr
, op1
);
3475 addr
= gen_rtx_CONST (Pmode
, addr
);
3476 addr
= force_const_mem (Pmode
, addr
);
3477 emit_move_insn (temp
, addr
);
3479 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3482 s390_load_address (reg
, new_rtx
);
3488 /* Now, check whether it is a GOT relative symbol plus offset
3489 that was pulled out of the literal pool. Force it back in. */
3491 else if (GET_CODE (op0
) == UNSPEC
3492 && GET_CODE (op1
) == CONST_INT
3493 && XINT (op0
, 1) == UNSPEC_GOTOFF
)
3495 gcc_assert (XVECLEN (op0
, 0) == 1);
3497 new_rtx
= force_const_mem (Pmode
, orig
);
3500 /* Otherwise, compute the sum. */
3503 base
= legitimize_pic_address (XEXP (addr
, 0), reg
);
3504 new_rtx
= legitimize_pic_address (XEXP (addr
, 1),
3505 base
== reg
? NULL_RTX
: reg
);
3506 if (GET_CODE (new_rtx
) == CONST_INT
)
3507 new_rtx
= plus_constant (base
, INTVAL (new_rtx
));
3510 if (GET_CODE (new_rtx
) == PLUS
&& CONSTANT_P (XEXP (new_rtx
, 1)))
3512 base
= gen_rtx_PLUS (Pmode
, base
, XEXP (new_rtx
, 0));
3513 new_rtx
= XEXP (new_rtx
, 1);
3515 new_rtx
= gen_rtx_PLUS (Pmode
, base
, new_rtx
);
3518 if (GET_CODE (new_rtx
) == CONST
)
3519 new_rtx
= XEXP (new_rtx
, 0);
3520 new_rtx
= force_operand (new_rtx
, 0);
3527 /* Load the thread pointer into a register. */
3530 s390_get_thread_pointer (void)
3532 rtx tp
= gen_reg_rtx (Pmode
);
3534 emit_move_insn (tp
, gen_rtx_REG (Pmode
, TP_REGNUM
));
3535 mark_reg_pointer (tp
, BITS_PER_WORD
);
3540 /* Emit a tls call insn. The call target is the SYMBOL_REF stored
3541 in s390_tls_symbol which always refers to __tls_get_offset.
3542 The returned offset is written to RESULT_REG and an USE rtx is
3543 generated for TLS_CALL. */
3545 static GTY(()) rtx s390_tls_symbol
;
3548 s390_emit_tls_call_insn (rtx result_reg
, rtx tls_call
)
3552 gcc_assert (flag_pic
);
3554 if (!s390_tls_symbol
)
3555 s390_tls_symbol
= gen_rtx_SYMBOL_REF (Pmode
, "__tls_get_offset");
3557 insn
= s390_emit_call (s390_tls_symbol
, tls_call
, result_reg
,
3558 gen_rtx_REG (Pmode
, RETURN_REGNUM
));
3560 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), result_reg
);
3561 RTL_CONST_CALL_P (insn
) = 1;
3564 /* ADDR contains a thread-local SYMBOL_REF. Generate code to compute
3565 this (thread-local) address. REG may be used as temporary. */
3568 legitimize_tls_address (rtx addr
, rtx reg
)
3570 rtx new_rtx
, tls_call
, temp
, base
, r2
, insn
;
3572 if (GET_CODE (addr
) == SYMBOL_REF
)
3573 switch (tls_symbolic_operand (addr
))
3575 case TLS_MODEL_GLOBAL_DYNAMIC
:
3577 r2
= gen_rtx_REG (Pmode
, 2);
3578 tls_call
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_TLSGD
);
3579 new_rtx
= gen_rtx_CONST (Pmode
, tls_call
);
3580 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3581 emit_move_insn (r2
, new_rtx
);
3582 s390_emit_tls_call_insn (r2
, tls_call
);
3583 insn
= get_insns ();
3586 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_NTPOFF
);
3587 temp
= gen_reg_rtx (Pmode
);
3588 emit_libcall_block (insn
, temp
, r2
, new_rtx
);
3590 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3593 s390_load_address (reg
, new_rtx
);
3598 case TLS_MODEL_LOCAL_DYNAMIC
:
3600 r2
= gen_rtx_REG (Pmode
, 2);
3601 tls_call
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, const0_rtx
), UNSPEC_TLSLDM
);
3602 new_rtx
= gen_rtx_CONST (Pmode
, tls_call
);
3603 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3604 emit_move_insn (r2
, new_rtx
);
3605 s390_emit_tls_call_insn (r2
, tls_call
);
3606 insn
= get_insns ();
3609 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, const0_rtx
), UNSPEC_TLSLDM_NTPOFF
);
3610 temp
= gen_reg_rtx (Pmode
);
3611 emit_libcall_block (insn
, temp
, r2
, new_rtx
);
3613 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3614 base
= gen_reg_rtx (Pmode
);
3615 s390_load_address (base
, new_rtx
);
3617 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_DTPOFF
);
3618 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3619 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3620 temp
= gen_reg_rtx (Pmode
);
3621 emit_move_insn (temp
, new_rtx
);
3623 new_rtx
= gen_rtx_PLUS (Pmode
, base
, temp
);
3626 s390_load_address (reg
, new_rtx
);
3631 case TLS_MODEL_INITIAL_EXEC
:
3634 /* Assume GOT offset < 4k. This is handled the same way
3635 in both 31- and 64-bit code. */
3637 if (reload_in_progress
|| reload_completed
)
3638 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3640 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTNTPOFF
);
3641 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3642 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, new_rtx
);
3643 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3644 temp
= gen_reg_rtx (Pmode
);
3645 emit_move_insn (temp
, new_rtx
);
3647 else if (TARGET_CPU_ZARCH
)
3649 /* If the GOT offset might be >= 4k, we determine the position
3650 of the GOT entry via a PC-relative LARL. */
3652 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_INDNTPOFF
);
3653 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3654 temp
= gen_reg_rtx (Pmode
);
3655 emit_move_insn (temp
, new_rtx
);
3657 new_rtx
= gen_const_mem (Pmode
, temp
);
3658 temp
= gen_reg_rtx (Pmode
);
3659 emit_move_insn (temp
, new_rtx
);
3663 /* If the GOT offset might be >= 4k, we have to load it
3664 from the literal pool. */
3666 if (reload_in_progress
|| reload_completed
)
3667 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3669 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTNTPOFF
);
3670 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3671 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3672 temp
= gen_reg_rtx (Pmode
);
3673 emit_move_insn (temp
, new_rtx
);
3675 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3676 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3678 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, new_rtx
, addr
), UNSPEC_TLS_LOAD
);
3679 temp
= gen_reg_rtx (Pmode
);
3680 emit_insn (gen_rtx_SET (Pmode
, temp
, new_rtx
));
3684 /* In position-dependent code, load the absolute address of
3685 the GOT entry from the literal pool. */
3687 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_INDNTPOFF
);
3688 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3689 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3690 temp
= gen_reg_rtx (Pmode
);
3691 emit_move_insn (temp
, new_rtx
);
3694 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3695 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, new_rtx
, addr
), UNSPEC_TLS_LOAD
);
3696 temp
= gen_reg_rtx (Pmode
);
3697 emit_insn (gen_rtx_SET (Pmode
, temp
, new_rtx
));
3700 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3703 s390_load_address (reg
, new_rtx
);
3708 case TLS_MODEL_LOCAL_EXEC
:
3709 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_NTPOFF
);
3710 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3711 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3712 temp
= gen_reg_rtx (Pmode
);
3713 emit_move_insn (temp
, new_rtx
);
3715 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3718 s390_load_address (reg
, new_rtx
);
3727 else if (GET_CODE (addr
) == CONST
&& GET_CODE (XEXP (addr
, 0)) == UNSPEC
)
3729 switch (XINT (XEXP (addr
, 0), 1))
3731 case UNSPEC_INDNTPOFF
:
3732 gcc_assert (TARGET_CPU_ZARCH
);
3741 else if (GET_CODE (addr
) == CONST
&& GET_CODE (XEXP (addr
, 0)) == PLUS
3742 && GET_CODE (XEXP (XEXP (addr
, 0), 1)) == CONST_INT
)
3744 new_rtx
= XEXP (XEXP (addr
, 0), 0);
3745 if (GET_CODE (new_rtx
) != SYMBOL_REF
)
3746 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3748 new_rtx
= legitimize_tls_address (new_rtx
, reg
);
3749 new_rtx
= plus_constant (new_rtx
, INTVAL (XEXP (XEXP (addr
, 0), 1)));
3750 new_rtx
= force_operand (new_rtx
, 0);
3754 gcc_unreachable (); /* for now ... */
3759 /* Emit insns making the address in operands[1] valid for a standard
3760 move to operands[0]. operands[1] is replaced by an address which
3761 should be used instead of the former RTX to emit the move
3765 emit_symbolic_move (rtx
*operands
)
3767 rtx temp
= !can_create_pseudo_p () ? operands
[0] : gen_reg_rtx (Pmode
);
3769 if (GET_CODE (operands
[0]) == MEM
)
3770 operands
[1] = force_reg (Pmode
, operands
[1]);
3771 else if (TLS_SYMBOLIC_CONST (operands
[1]))
3772 operands
[1] = legitimize_tls_address (operands
[1], temp
);
3774 operands
[1] = legitimize_pic_address (operands
[1], temp
);
3777 /* Try machine-dependent ways of modifying an illegitimate address X
3778 to be legitimate. If we find one, return the new, valid address.
3780 OLDX is the address as it was before break_out_memory_refs was called.
3781 In some cases it is useful to look at this to decide what needs to be done.
3783 MODE is the mode of the operand pointed to by X.
3785 When -fpic is used, special handling is needed for symbolic references.
3786 See comments by legitimize_pic_address for details. */
3789 s390_legitimize_address (rtx x
, rtx oldx ATTRIBUTE_UNUSED
,
3790 enum machine_mode mode ATTRIBUTE_UNUSED
)
3792 rtx constant_term
= const0_rtx
;
3794 if (TLS_SYMBOLIC_CONST (x
))
3796 x
= legitimize_tls_address (x
, 0);
3798 if (s390_legitimate_address_p (mode
, x
, FALSE
))
3801 else if (GET_CODE (x
) == PLUS
3802 && (TLS_SYMBOLIC_CONST (XEXP (x
, 0))
3803 || TLS_SYMBOLIC_CONST (XEXP (x
, 1))))
3809 if (SYMBOLIC_CONST (x
)
3810 || (GET_CODE (x
) == PLUS
3811 && (SYMBOLIC_CONST (XEXP (x
, 0))
3812 || SYMBOLIC_CONST (XEXP (x
, 1)))))
3813 x
= legitimize_pic_address (x
, 0);
3815 if (s390_legitimate_address_p (mode
, x
, FALSE
))
3819 x
= eliminate_constant_term (x
, &constant_term
);
3821 /* Optimize loading of large displacements by splitting them
3822 into the multiple of 4K and the rest; this allows the
3823 former to be CSE'd if possible.
3825 Don't do this if the displacement is added to a register
3826 pointing into the stack frame, as the offsets will
3827 change later anyway. */
3829 if (GET_CODE (constant_term
) == CONST_INT
3830 && !TARGET_LONG_DISPLACEMENT
3831 && !DISP_IN_RANGE (INTVAL (constant_term
))
3832 && !(REG_P (x
) && REGNO_PTR_FRAME_P (REGNO (x
))))
3834 HOST_WIDE_INT lower
= INTVAL (constant_term
) & 0xfff;
3835 HOST_WIDE_INT upper
= INTVAL (constant_term
) ^ lower
;
3837 rtx temp
= gen_reg_rtx (Pmode
);
3838 rtx val
= force_operand (GEN_INT (upper
), temp
);
3840 emit_move_insn (temp
, val
);
3842 x
= gen_rtx_PLUS (Pmode
, x
, temp
);
3843 constant_term
= GEN_INT (lower
);
3846 if (GET_CODE (x
) == PLUS
)
3848 if (GET_CODE (XEXP (x
, 0)) == REG
)
3850 rtx temp
= gen_reg_rtx (Pmode
);
3851 rtx val
= force_operand (XEXP (x
, 1), temp
);
3853 emit_move_insn (temp
, val
);
3855 x
= gen_rtx_PLUS (Pmode
, XEXP (x
, 0), temp
);
3858 else if (GET_CODE (XEXP (x
, 1)) == REG
)
3860 rtx temp
= gen_reg_rtx (Pmode
);
3861 rtx val
= force_operand (XEXP (x
, 0), temp
);
3863 emit_move_insn (temp
, val
);
3865 x
= gen_rtx_PLUS (Pmode
, temp
, XEXP (x
, 1));
3869 if (constant_term
!= const0_rtx
)
3870 x
= gen_rtx_PLUS (Pmode
, x
, constant_term
);
3875 /* Try a machine-dependent way of reloading an illegitimate address AD
3876 operand. If we find one, push the reload and and return the new address.
3878 MODE is the mode of the enclosing MEM. OPNUM is the operand number
3879 and TYPE is the reload type of the current reload. */
3882 legitimize_reload_address (rtx ad
, enum machine_mode mode ATTRIBUTE_UNUSED
,
3883 int opnum
, int type
)
3885 if (!optimize
|| TARGET_LONG_DISPLACEMENT
)
3888 if (GET_CODE (ad
) == PLUS
)
3890 rtx tem
= simplify_binary_operation (PLUS
, Pmode
,
3891 XEXP (ad
, 0), XEXP (ad
, 1));
3896 if (GET_CODE (ad
) == PLUS
3897 && GET_CODE (XEXP (ad
, 0)) == REG
3898 && GET_CODE (XEXP (ad
, 1)) == CONST_INT
3899 && !DISP_IN_RANGE (INTVAL (XEXP (ad
, 1))))
3901 HOST_WIDE_INT lower
= INTVAL (XEXP (ad
, 1)) & 0xfff;
3902 HOST_WIDE_INT upper
= INTVAL (XEXP (ad
, 1)) ^ lower
;
3903 rtx cst
, tem
, new_rtx
;
3905 cst
= GEN_INT (upper
);
3906 if (!legitimate_reload_constant_p (cst
))
3907 cst
= force_const_mem (Pmode
, cst
);
3909 tem
= gen_rtx_PLUS (Pmode
, XEXP (ad
, 0), cst
);
3910 new_rtx
= gen_rtx_PLUS (Pmode
, tem
, GEN_INT (lower
));
3912 push_reload (XEXP (tem
, 1), 0, &XEXP (tem
, 1), 0,
3913 BASE_REG_CLASS
, Pmode
, VOIDmode
, 0, 0,
3914 opnum
, (enum reload_type
) type
);
3921 /* Emit code to move LEN bytes from DST to SRC. */
3924 s390_expand_movmem (rtx dst
, rtx src
, rtx len
)
3926 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) >= 0 && INTVAL (len
) <= 256)
3928 if (INTVAL (len
) > 0)
3929 emit_insn (gen_movmem_short (dst
, src
, GEN_INT (INTVAL (len
) - 1)));
3932 else if (TARGET_MVCLE
)
3934 emit_insn (gen_movmem_long (dst
, src
, convert_to_mode (Pmode
, len
, 1)));
3939 rtx dst_addr
, src_addr
, count
, blocks
, temp
;
3940 rtx loop_start_label
= gen_label_rtx ();
3941 rtx loop_end_label
= gen_label_rtx ();
3942 rtx end_label
= gen_label_rtx ();
3943 enum machine_mode mode
;
3945 mode
= GET_MODE (len
);
3946 if (mode
== VOIDmode
)
3949 dst_addr
= gen_reg_rtx (Pmode
);
3950 src_addr
= gen_reg_rtx (Pmode
);
3951 count
= gen_reg_rtx (mode
);
3952 blocks
= gen_reg_rtx (mode
);
3954 convert_move (count
, len
, 1);
3955 emit_cmp_and_jump_insns (count
, const0_rtx
,
3956 EQ
, NULL_RTX
, mode
, 1, end_label
);
3958 emit_move_insn (dst_addr
, force_operand (XEXP (dst
, 0), NULL_RTX
));
3959 emit_move_insn (src_addr
, force_operand (XEXP (src
, 0), NULL_RTX
));
3960 dst
= change_address (dst
, VOIDmode
, dst_addr
);
3961 src
= change_address (src
, VOIDmode
, src_addr
);
3963 temp
= expand_binop (mode
, add_optab
, count
, constm1_rtx
, count
, 1,
3966 emit_move_insn (count
, temp
);
3968 temp
= expand_binop (mode
, lshr_optab
, count
, GEN_INT (8), blocks
, 1,
3971 emit_move_insn (blocks
, temp
);
3973 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
3974 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
3976 emit_label (loop_start_label
);
3979 && (GET_CODE (len
) != CONST_INT
|| INTVAL (len
) > 768))
3983 /* Issue a read prefetch for the +3 cache line. */
3984 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, src_addr
, GEN_INT (768)),
3985 const0_rtx
, const0_rtx
);
3986 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
3987 emit_insn (prefetch
);
3989 /* Issue a write prefetch for the +3 cache line. */
3990 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, dst_addr
, GEN_INT (768)),
3991 const1_rtx
, const0_rtx
);
3992 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
3993 emit_insn (prefetch
);
3996 emit_insn (gen_movmem_short (dst
, src
, GEN_INT (255)));
3997 s390_load_address (dst_addr
,
3998 gen_rtx_PLUS (Pmode
, dst_addr
, GEN_INT (256)));
3999 s390_load_address (src_addr
,
4000 gen_rtx_PLUS (Pmode
, src_addr
, GEN_INT (256)));
4002 temp
= expand_binop (mode
, add_optab
, blocks
, constm1_rtx
, blocks
, 1,
4005 emit_move_insn (blocks
, temp
);
4007 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4008 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4010 emit_jump (loop_start_label
);
4011 emit_label (loop_end_label
);
4013 emit_insn (gen_movmem_short (dst
, src
,
4014 convert_to_mode (Pmode
, count
, 1)));
4015 emit_label (end_label
);
4019 /* Emit code to set LEN bytes at DST to VAL.
4020 Make use of clrmem if VAL is zero. */
4023 s390_expand_setmem (rtx dst
, rtx len
, rtx val
)
4025 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) == 0)
4028 gcc_assert (GET_CODE (val
) == CONST_INT
|| GET_MODE (val
) == QImode
);
4030 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) > 0 && INTVAL (len
) <= 257)
4032 if (val
== const0_rtx
&& INTVAL (len
) <= 256)
4033 emit_insn (gen_clrmem_short (dst
, GEN_INT (INTVAL (len
) - 1)));
4036 /* Initialize memory by storing the first byte. */
4037 emit_move_insn (adjust_address (dst
, QImode
, 0), val
);
4039 if (INTVAL (len
) > 1)
4041 /* Initiate 1 byte overlap move.
4042 The first byte of DST is propagated through DSTP1.
4043 Prepare a movmem for: DST+1 = DST (length = LEN - 1).
4044 DST is set to size 1 so the rest of the memory location
4045 does not count as source operand. */
4046 rtx dstp1
= adjust_address (dst
, VOIDmode
, 1);
4047 set_mem_size (dst
, const1_rtx
);
4049 emit_insn (gen_movmem_short (dstp1
, dst
,
4050 GEN_INT (INTVAL (len
) - 2)));
4055 else if (TARGET_MVCLE
)
4057 val
= force_not_mem (convert_modes (Pmode
, QImode
, val
, 1));
4058 emit_insn (gen_setmem_long (dst
, convert_to_mode (Pmode
, len
, 1), val
));
4063 rtx dst_addr
, count
, blocks
, temp
, dstp1
= NULL_RTX
;
4064 rtx loop_start_label
= gen_label_rtx ();
4065 rtx loop_end_label
= gen_label_rtx ();
4066 rtx end_label
= gen_label_rtx ();
4067 enum machine_mode mode
;
4069 mode
= GET_MODE (len
);
4070 if (mode
== VOIDmode
)
4073 dst_addr
= gen_reg_rtx (Pmode
);
4074 count
= gen_reg_rtx (mode
);
4075 blocks
= gen_reg_rtx (mode
);
4077 convert_move (count
, len
, 1);
4078 emit_cmp_and_jump_insns (count
, const0_rtx
,
4079 EQ
, NULL_RTX
, mode
, 1, end_label
);
4081 emit_move_insn (dst_addr
, force_operand (XEXP (dst
, 0), NULL_RTX
));
4082 dst
= change_address (dst
, VOIDmode
, dst_addr
);
4084 if (val
== const0_rtx
)
4085 temp
= expand_binop (mode
, add_optab
, count
, constm1_rtx
, count
, 1,
4089 dstp1
= adjust_address (dst
, VOIDmode
, 1);
4090 set_mem_size (dst
, const1_rtx
);
4092 /* Initialize memory by storing the first byte. */
4093 emit_move_insn (adjust_address (dst
, QImode
, 0), val
);
4095 /* If count is 1 we are done. */
4096 emit_cmp_and_jump_insns (count
, const1_rtx
,
4097 EQ
, NULL_RTX
, mode
, 1, end_label
);
4099 temp
= expand_binop (mode
, add_optab
, count
, GEN_INT (-2), count
, 1,
4103 emit_move_insn (count
, temp
);
4105 temp
= expand_binop (mode
, lshr_optab
, count
, GEN_INT (8), blocks
, 1,
4108 emit_move_insn (blocks
, temp
);
4110 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4111 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4113 emit_label (loop_start_label
);
4116 && (GET_CODE (len
) != CONST_INT
|| INTVAL (len
) > 1024))
4118 /* Issue a write prefetch for the +4 cache line. */
4119 rtx prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, dst_addr
,
4121 const1_rtx
, const0_rtx
);
4122 emit_insn (prefetch
);
4123 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
4126 if (val
== const0_rtx
)
4127 emit_insn (gen_clrmem_short (dst
, GEN_INT (255)));
4129 emit_insn (gen_movmem_short (dstp1
, dst
, GEN_INT (255)));
4130 s390_load_address (dst_addr
,
4131 gen_rtx_PLUS (Pmode
, dst_addr
, GEN_INT (256)));
4133 temp
= expand_binop (mode
, add_optab
, blocks
, constm1_rtx
, blocks
, 1,
4136 emit_move_insn (blocks
, temp
);
4138 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4139 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4141 emit_jump (loop_start_label
);
4142 emit_label (loop_end_label
);
4144 if (val
== const0_rtx
)
4145 emit_insn (gen_clrmem_short (dst
, convert_to_mode (Pmode
, count
, 1)));
4147 emit_insn (gen_movmem_short (dstp1
, dst
, convert_to_mode (Pmode
, count
, 1)));
4148 emit_label (end_label
);
4152 /* Emit code to compare LEN bytes at OP0 with those at OP1,
4153 and return the result in TARGET. */
4156 s390_expand_cmpmem (rtx target
, rtx op0
, rtx op1
, rtx len
)
4158 rtx ccreg
= gen_rtx_REG (CCUmode
, CC_REGNUM
);
4161 /* As the result of CMPINT is inverted compared to what we need,
4162 we have to swap the operands. */
4163 tmp
= op0
; op0
= op1
; op1
= tmp
;
4165 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) >= 0 && INTVAL (len
) <= 256)
4167 if (INTVAL (len
) > 0)
4169 emit_insn (gen_cmpmem_short (op0
, op1
, GEN_INT (INTVAL (len
) - 1)));
4170 emit_insn (gen_cmpint (target
, ccreg
));
4173 emit_move_insn (target
, const0_rtx
);
4175 else if (TARGET_MVCLE
)
4177 emit_insn (gen_cmpmem_long (op0
, op1
, convert_to_mode (Pmode
, len
, 1)));
4178 emit_insn (gen_cmpint (target
, ccreg
));
4182 rtx addr0
, addr1
, count
, blocks
, temp
;
4183 rtx loop_start_label
= gen_label_rtx ();
4184 rtx loop_end_label
= gen_label_rtx ();
4185 rtx end_label
= gen_label_rtx ();
4186 enum machine_mode mode
;
4188 mode
= GET_MODE (len
);
4189 if (mode
== VOIDmode
)
4192 addr0
= gen_reg_rtx (Pmode
);
4193 addr1
= gen_reg_rtx (Pmode
);
4194 count
= gen_reg_rtx (mode
);
4195 blocks
= gen_reg_rtx (mode
);
4197 convert_move (count
, len
, 1);
4198 emit_cmp_and_jump_insns (count
, const0_rtx
,
4199 EQ
, NULL_RTX
, mode
, 1, end_label
);
4201 emit_move_insn (addr0
, force_operand (XEXP (op0
, 0), NULL_RTX
));
4202 emit_move_insn (addr1
, force_operand (XEXP (op1
, 0), NULL_RTX
));
4203 op0
= change_address (op0
, VOIDmode
, addr0
);
4204 op1
= change_address (op1
, VOIDmode
, addr1
);
4206 temp
= expand_binop (mode
, add_optab
, count
, constm1_rtx
, count
, 1,
4209 emit_move_insn (count
, temp
);
4211 temp
= expand_binop (mode
, lshr_optab
, count
, GEN_INT (8), blocks
, 1,
4214 emit_move_insn (blocks
, temp
);
4216 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4217 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4219 emit_label (loop_start_label
);
4222 && (GET_CODE (len
) != CONST_INT
|| INTVAL (len
) > 512))
4226 /* Issue a read prefetch for the +2 cache line of operand 1. */
4227 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, addr0
, GEN_INT (512)),
4228 const0_rtx
, const0_rtx
);
4229 emit_insn (prefetch
);
4230 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
4232 /* Issue a read prefetch for the +2 cache line of operand 2. */
4233 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, addr1
, GEN_INT (512)),
4234 const0_rtx
, const0_rtx
);
4235 emit_insn (prefetch
);
4236 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
4239 emit_insn (gen_cmpmem_short (op0
, op1
, GEN_INT (255)));
4240 temp
= gen_rtx_NE (VOIDmode
, ccreg
, const0_rtx
);
4241 temp
= gen_rtx_IF_THEN_ELSE (VOIDmode
, temp
,
4242 gen_rtx_LABEL_REF (VOIDmode
, end_label
), pc_rtx
);
4243 temp
= gen_rtx_SET (VOIDmode
, pc_rtx
, temp
);
4244 emit_jump_insn (temp
);
4246 s390_load_address (addr0
,
4247 gen_rtx_PLUS (Pmode
, addr0
, GEN_INT (256)));
4248 s390_load_address (addr1
,
4249 gen_rtx_PLUS (Pmode
, addr1
, GEN_INT (256)));
4251 temp
= expand_binop (mode
, add_optab
, blocks
, constm1_rtx
, blocks
, 1,
4254 emit_move_insn (blocks
, temp
);
4256 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4257 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4259 emit_jump (loop_start_label
);
4260 emit_label (loop_end_label
);
4262 emit_insn (gen_cmpmem_short (op0
, op1
,
4263 convert_to_mode (Pmode
, count
, 1)));
4264 emit_label (end_label
);
4266 emit_insn (gen_cmpint (target
, ccreg
));
4271 /* Expand conditional increment or decrement using alc/slb instructions.
4272 Should generate code setting DST to either SRC or SRC + INCREMENT,
4273 depending on the result of the comparison CMP_OP0 CMP_CODE CMP_OP1.
4274 Returns true if successful, false otherwise.
4276 That makes it possible to implement some if-constructs without jumps e.g.:
4277 (borrow = CC0 | CC1 and carry = CC2 | CC3)
4278 unsigned int a, b, c;
4279 if (a < b) c++; -> CCU b > a -> CC2; c += carry;
4280 if (a < b) c--; -> CCL3 a - b -> borrow; c -= borrow;
4281 if (a <= b) c++; -> CCL3 b - a -> borrow; c += carry;
4282 if (a <= b) c--; -> CCU a <= b -> borrow; c -= borrow;
4284 Checks for EQ and NE with a nonzero value need an additional xor e.g.:
4285 if (a == b) c++; -> CCL3 a ^= b; 0 - a -> borrow; c += carry;
4286 if (a == b) c--; -> CCU a ^= b; a <= 0 -> CC0 | CC1; c -= borrow;
4287 if (a != b) c++; -> CCU a ^= b; a > 0 -> CC2; c += carry;
4288 if (a != b) c--; -> CCL3 a ^= b; 0 - a -> borrow; c -= borrow; */
4291 s390_expand_addcc (enum rtx_code cmp_code
, rtx cmp_op0
, rtx cmp_op1
,
4292 rtx dst
, rtx src
, rtx increment
)
4294 enum machine_mode cmp_mode
;
4295 enum machine_mode cc_mode
;
4301 if ((GET_MODE (cmp_op0
) == SImode
|| GET_MODE (cmp_op0
) == VOIDmode
)
4302 && (GET_MODE (cmp_op1
) == SImode
|| GET_MODE (cmp_op1
) == VOIDmode
))
4304 else if ((GET_MODE (cmp_op0
) == DImode
|| GET_MODE (cmp_op0
) == VOIDmode
)
4305 && (GET_MODE (cmp_op1
) == DImode
|| GET_MODE (cmp_op1
) == VOIDmode
))
4310 /* Try ADD LOGICAL WITH CARRY. */
4311 if (increment
== const1_rtx
)
4313 /* Determine CC mode to use. */
4314 if (cmp_code
== EQ
|| cmp_code
== NE
)
4316 if (cmp_op1
!= const0_rtx
)
4318 cmp_op0
= expand_simple_binop (cmp_mode
, XOR
, cmp_op0
, cmp_op1
,
4319 NULL_RTX
, 0, OPTAB_WIDEN
);
4320 cmp_op1
= const0_rtx
;
4323 cmp_code
= cmp_code
== EQ
? LEU
: GTU
;
4326 if (cmp_code
== LTU
|| cmp_code
== LEU
)
4331 cmp_code
= swap_condition (cmp_code
);
4348 /* Emit comparison instruction pattern. */
4349 if (!register_operand (cmp_op0
, cmp_mode
))
4350 cmp_op0
= force_reg (cmp_mode
, cmp_op0
);
4352 insn
= gen_rtx_SET (VOIDmode
, gen_rtx_REG (cc_mode
, CC_REGNUM
),
4353 gen_rtx_COMPARE (cc_mode
, cmp_op0
, cmp_op1
));
4354 /* We use insn_invalid_p here to add clobbers if required. */
4355 ret
= insn_invalid_p (emit_insn (insn
));
4358 /* Emit ALC instruction pattern. */
4359 op_res
= gen_rtx_fmt_ee (cmp_code
, GET_MODE (dst
),
4360 gen_rtx_REG (cc_mode
, CC_REGNUM
),
4363 if (src
!= const0_rtx
)
4365 if (!register_operand (src
, GET_MODE (dst
)))
4366 src
= force_reg (GET_MODE (dst
), src
);
4368 op_res
= gen_rtx_PLUS (GET_MODE (dst
), op_res
, src
);
4369 op_res
= gen_rtx_PLUS (GET_MODE (dst
), op_res
, const0_rtx
);
4372 p
= rtvec_alloc (2);
4374 gen_rtx_SET (VOIDmode
, dst
, op_res
);
4376 gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
4377 emit_insn (gen_rtx_PARALLEL (VOIDmode
, p
));
4382 /* Try SUBTRACT LOGICAL WITH BORROW. */
4383 if (increment
== constm1_rtx
)
4385 /* Determine CC mode to use. */
4386 if (cmp_code
== EQ
|| cmp_code
== NE
)
4388 if (cmp_op1
!= const0_rtx
)
4390 cmp_op0
= expand_simple_binop (cmp_mode
, XOR
, cmp_op0
, cmp_op1
,
4391 NULL_RTX
, 0, OPTAB_WIDEN
);
4392 cmp_op1
= const0_rtx
;
4395 cmp_code
= cmp_code
== EQ
? LEU
: GTU
;
4398 if (cmp_code
== GTU
|| cmp_code
== GEU
)
4403 cmp_code
= swap_condition (cmp_code
);
4420 /* Emit comparison instruction pattern. */
4421 if (!register_operand (cmp_op0
, cmp_mode
))
4422 cmp_op0
= force_reg (cmp_mode
, cmp_op0
);
4424 insn
= gen_rtx_SET (VOIDmode
, gen_rtx_REG (cc_mode
, CC_REGNUM
),
4425 gen_rtx_COMPARE (cc_mode
, cmp_op0
, cmp_op1
));
4426 /* We use insn_invalid_p here to add clobbers if required. */
4427 ret
= insn_invalid_p (emit_insn (insn
));
4430 /* Emit SLB instruction pattern. */
4431 if (!register_operand (src
, GET_MODE (dst
)))
4432 src
= force_reg (GET_MODE (dst
), src
);
4434 op_res
= gen_rtx_MINUS (GET_MODE (dst
),
4435 gen_rtx_MINUS (GET_MODE (dst
), src
, const0_rtx
),
4436 gen_rtx_fmt_ee (cmp_code
, GET_MODE (dst
),
4437 gen_rtx_REG (cc_mode
, CC_REGNUM
),
4439 p
= rtvec_alloc (2);
4441 gen_rtx_SET (VOIDmode
, dst
, op_res
);
4443 gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
4444 emit_insn (gen_rtx_PARALLEL (VOIDmode
, p
));
4452 /* Expand code for the insv template. Return true if successful. */
4455 s390_expand_insv (rtx dest
, rtx op1
, rtx op2
, rtx src
)
4457 int bitsize
= INTVAL (op1
);
4458 int bitpos
= INTVAL (op2
);
4460 /* On z10 we can use the risbg instruction to implement insv. */
4462 && ((GET_MODE (dest
) == DImode
&& GET_MODE (src
) == DImode
)
4463 || (GET_MODE (dest
) == SImode
&& GET_MODE (src
) == SImode
)))
4468 op
= gen_rtx_SET (GET_MODE(src
),
4469 gen_rtx_ZERO_EXTRACT (GET_MODE (dest
), dest
, op1
, op2
),
4471 clobber
= gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
4472 emit_insn (gen_rtx_PARALLEL (VOIDmode
, gen_rtvec (2, op
, clobber
)));
4477 /* We need byte alignment. */
4478 if (bitsize
% BITS_PER_UNIT
)
4482 && memory_operand (dest
, VOIDmode
)
4483 && (register_operand (src
, word_mode
)
4484 || const_int_operand (src
, VOIDmode
)))
4486 /* Emit standard pattern if possible. */
4487 enum machine_mode mode
= smallest_mode_for_size (bitsize
, MODE_INT
);
4488 if (GET_MODE_BITSIZE (mode
) == bitsize
)
4489 emit_move_insn (adjust_address (dest
, mode
, 0), gen_lowpart (mode
, src
));
4491 /* (set (ze (mem)) (const_int)). */
4492 else if (const_int_operand (src
, VOIDmode
))
4494 int size
= bitsize
/ BITS_PER_UNIT
;
4495 rtx src_mem
= adjust_address (force_const_mem (word_mode
, src
), BLKmode
,
4496 GET_MODE_SIZE (word_mode
) - size
);
4498 dest
= adjust_address (dest
, BLKmode
, 0);
4499 set_mem_size (dest
, GEN_INT (size
));
4500 s390_expand_movmem (dest
, src_mem
, GEN_INT (size
));
4503 /* (set (ze (mem)) (reg)). */
4504 else if (register_operand (src
, word_mode
))
4506 if (bitsize
<= GET_MODE_BITSIZE (SImode
))
4507 emit_move_insn (gen_rtx_ZERO_EXTRACT (word_mode
, dest
, op1
,
4511 /* Emit st,stcmh sequence. */
4512 int stcmh_width
= bitsize
- GET_MODE_BITSIZE (SImode
);
4513 int size
= stcmh_width
/ BITS_PER_UNIT
;
4515 emit_move_insn (adjust_address (dest
, SImode
, size
),
4516 gen_lowpart (SImode
, src
));
4517 set_mem_size (dest
, GEN_INT (size
));
4518 emit_move_insn (gen_rtx_ZERO_EXTRACT (word_mode
, dest
, GEN_INT
4519 (stcmh_width
), const0_rtx
),
4520 gen_rtx_LSHIFTRT (word_mode
, src
, GEN_INT
4521 (GET_MODE_BITSIZE (SImode
))));
4530 /* (set (ze (reg)) (const_int)). */
4532 && register_operand (dest
, word_mode
)
4533 && (bitpos
% 16) == 0
4534 && (bitsize
% 16) == 0
4535 && const_int_operand (src
, VOIDmode
))
4537 HOST_WIDE_INT val
= INTVAL (src
);
4538 int regpos
= bitpos
+ bitsize
;
4540 while (regpos
> bitpos
)
4542 enum machine_mode putmode
;
4545 if (TARGET_EXTIMM
&& (regpos
% 32 == 0) && (regpos
>= bitpos
+ 32))
4550 putsize
= GET_MODE_BITSIZE (putmode
);
4552 emit_move_insn (gen_rtx_ZERO_EXTRACT (word_mode
, dest
,
4555 gen_int_mode (val
, putmode
));
4558 gcc_assert (regpos
== bitpos
);
4565 /* A subroutine of s390_expand_cs_hqi and s390_expand_atomic which returns a
4566 register that holds VAL of mode MODE shifted by COUNT bits. */
4569 s390_expand_mask_and_shift (rtx val
, enum machine_mode mode
, rtx count
)
4571 val
= expand_simple_binop (SImode
, AND
, val
, GEN_INT (GET_MODE_MASK (mode
)),
4572 NULL_RTX
, 1, OPTAB_DIRECT
);
4573 return expand_simple_binop (SImode
, ASHIFT
, val
, count
,
4574 NULL_RTX
, 1, OPTAB_DIRECT
);
4577 /* Structure to hold the initial parameters for a compare_and_swap operation
4578 in HImode and QImode. */
4580 struct alignment_context
4582 rtx memsi
; /* SI aligned memory location. */
4583 rtx shift
; /* Bit offset with regard to lsb. */
4584 rtx modemask
; /* Mask of the HQImode shifted by SHIFT bits. */
4585 rtx modemaski
; /* ~modemask */
4586 bool aligned
; /* True if memory is aligned, false else. */
4589 /* A subroutine of s390_expand_cs_hqi and s390_expand_atomic to initialize
4590 structure AC for transparent simplifying, if the memory alignment is known
4591 to be at least 32bit. MEM is the memory location for the actual operation
4592 and MODE its mode. */
4595 init_alignment_context (struct alignment_context
*ac
, rtx mem
,
4596 enum machine_mode mode
)
4598 ac
->shift
= GEN_INT (GET_MODE_SIZE (SImode
) - GET_MODE_SIZE (mode
));
4599 ac
->aligned
= (MEM_ALIGN (mem
) >= GET_MODE_BITSIZE (SImode
));
4602 ac
->memsi
= adjust_address (mem
, SImode
, 0); /* Memory is aligned. */
4605 /* Alignment is unknown. */
4606 rtx byteoffset
, addr
, align
;
4608 /* Force the address into a register. */
4609 addr
= force_reg (Pmode
, XEXP (mem
, 0));
4611 /* Align it to SImode. */
4612 align
= expand_simple_binop (Pmode
, AND
, addr
,
4613 GEN_INT (-GET_MODE_SIZE (SImode
)),
4614 NULL_RTX
, 1, OPTAB_DIRECT
);
4616 ac
->memsi
= gen_rtx_MEM (SImode
, align
);
4617 MEM_VOLATILE_P (ac
->memsi
) = MEM_VOLATILE_P (mem
);
4618 set_mem_alias_set (ac
->memsi
, ALIAS_SET_MEMORY_BARRIER
);
4619 set_mem_align (ac
->memsi
, GET_MODE_BITSIZE (SImode
));
4621 /* Calculate shiftcount. */
4622 byteoffset
= expand_simple_binop (Pmode
, AND
, addr
,
4623 GEN_INT (GET_MODE_SIZE (SImode
) - 1),
4624 NULL_RTX
, 1, OPTAB_DIRECT
);
4625 /* As we already have some offset, evaluate the remaining distance. */
4626 ac
->shift
= expand_simple_binop (SImode
, MINUS
, ac
->shift
, byteoffset
,
4627 NULL_RTX
, 1, OPTAB_DIRECT
);
4630 /* Shift is the byte count, but we need the bitcount. */
4631 ac
->shift
= expand_simple_binop (SImode
, MULT
, ac
->shift
, GEN_INT (BITS_PER_UNIT
),
4632 NULL_RTX
, 1, OPTAB_DIRECT
);
4633 /* Calculate masks. */
4634 ac
->modemask
= expand_simple_binop (SImode
, ASHIFT
,
4635 GEN_INT (GET_MODE_MASK (mode
)), ac
->shift
,
4636 NULL_RTX
, 1, OPTAB_DIRECT
);
4637 ac
->modemaski
= expand_simple_unop (SImode
, NOT
, ac
->modemask
, NULL_RTX
, 1);
4640 /* Expand an atomic compare and swap operation for HImode and QImode. MEM is
4641 the memory location, CMP the old value to compare MEM with and NEW_RTX the value
4642 to set if CMP == MEM.
4643 CMP is never in memory for compare_and_swap_cc because
4644 expand_bool_compare_and_swap puts it into a register for later compare. */
4647 s390_expand_cs_hqi (enum machine_mode mode
, rtx target
, rtx mem
, rtx cmp
, rtx new_rtx
)
4649 struct alignment_context ac
;
4650 rtx cmpv
, newv
, val
, resv
, cc
;
4651 rtx res
= gen_reg_rtx (SImode
);
4652 rtx csloop
= gen_label_rtx ();
4653 rtx csend
= gen_label_rtx ();
4655 gcc_assert (register_operand (target
, VOIDmode
));
4656 gcc_assert (MEM_P (mem
));
4658 init_alignment_context (&ac
, mem
, mode
);
4660 /* Shift the values to the correct bit positions. */
4661 if (!(ac
.aligned
&& MEM_P (cmp
)))
4662 cmp
= s390_expand_mask_and_shift (cmp
, mode
, ac
.shift
);
4663 if (!(ac
.aligned
&& MEM_P (new_rtx
)))
4664 new_rtx
= s390_expand_mask_and_shift (new_rtx
, mode
, ac
.shift
);
4666 /* Load full word. Subsequent loads are performed by CS. */
4667 val
= expand_simple_binop (SImode
, AND
, ac
.memsi
, ac
.modemaski
,
4668 NULL_RTX
, 1, OPTAB_DIRECT
);
4670 /* Start CS loop. */
4671 emit_label (csloop
);
4672 /* val = "<mem>00..0<mem>"
4673 * cmp = "00..0<cmp>00..0"
4674 * new = "00..0<new>00..0"
4677 /* Patch cmp and new with val at correct position. */
4678 if (ac
.aligned
&& MEM_P (cmp
))
4680 cmpv
= force_reg (SImode
, val
);
4681 store_bit_field (cmpv
, GET_MODE_BITSIZE (mode
), 0, SImode
, cmp
);
4684 cmpv
= force_reg (SImode
, expand_simple_binop (SImode
, IOR
, cmp
, val
,
4685 NULL_RTX
, 1, OPTAB_DIRECT
));
4686 if (ac
.aligned
&& MEM_P (new_rtx
))
4688 newv
= force_reg (SImode
, val
);
4689 store_bit_field (newv
, GET_MODE_BITSIZE (mode
), 0, SImode
, new_rtx
);
4692 newv
= force_reg (SImode
, expand_simple_binop (SImode
, IOR
, new_rtx
, val
,
4693 NULL_RTX
, 1, OPTAB_DIRECT
));
4695 /* Jump to end if we're done (likely?). */
4696 s390_emit_jump (csend
, s390_emit_compare_and_swap (EQ
, res
, ac
.memsi
,
4699 /* Check for changes outside mode. */
4700 resv
= expand_simple_binop (SImode
, AND
, res
, ac
.modemaski
,
4701 NULL_RTX
, 1, OPTAB_DIRECT
);
4702 cc
= s390_emit_compare (NE
, resv
, val
);
4703 emit_move_insn (val
, resv
);
4704 /* Loop internal if so. */
4705 s390_emit_jump (csloop
, cc
);
4709 /* Return the correct part of the bitfield. */
4710 convert_move (target
, expand_simple_binop (SImode
, LSHIFTRT
, res
, ac
.shift
,
4711 NULL_RTX
, 1, OPTAB_DIRECT
), 1);
4714 /* Expand an atomic operation CODE of mode MODE. MEM is the memory location
4715 and VAL the value to play with. If AFTER is true then store the value
4716 MEM holds after the operation, if AFTER is false then store the value MEM
4717 holds before the operation. If TARGET is zero then discard that value, else
4718 store it to TARGET. */
4721 s390_expand_atomic (enum machine_mode mode
, enum rtx_code code
,
4722 rtx target
, rtx mem
, rtx val
, bool after
)
4724 struct alignment_context ac
;
4726 rtx new_rtx
= gen_reg_rtx (SImode
);
4727 rtx orig
= gen_reg_rtx (SImode
);
4728 rtx csloop
= gen_label_rtx ();
4730 gcc_assert (!target
|| register_operand (target
, VOIDmode
));
4731 gcc_assert (MEM_P (mem
));
4733 init_alignment_context (&ac
, mem
, mode
);
4735 /* Shift val to the correct bit positions.
4736 Preserve "icm", but prevent "ex icm". */
4737 if (!(ac
.aligned
&& code
== SET
&& MEM_P (val
)))
4738 val
= s390_expand_mask_and_shift (val
, mode
, ac
.shift
);
4740 /* Further preparation insns. */
4741 if (code
== PLUS
|| code
== MINUS
)
4742 emit_move_insn (orig
, val
);
4743 else if (code
== MULT
|| code
== AND
) /* val = "11..1<val>11..1" */
4744 val
= expand_simple_binop (SImode
, XOR
, val
, ac
.modemaski
,
4745 NULL_RTX
, 1, OPTAB_DIRECT
);
4747 /* Load full word. Subsequent loads are performed by CS. */
4748 cmp
= force_reg (SImode
, ac
.memsi
);
4750 /* Start CS loop. */
4751 emit_label (csloop
);
4752 emit_move_insn (new_rtx
, cmp
);
4754 /* Patch new with val at correct position. */
4759 val
= expand_simple_binop (SImode
, code
, new_rtx
, orig
,
4760 NULL_RTX
, 1, OPTAB_DIRECT
);
4761 val
= expand_simple_binop (SImode
, AND
, val
, ac
.modemask
,
4762 NULL_RTX
, 1, OPTAB_DIRECT
);
4765 if (ac
.aligned
&& MEM_P (val
))
4766 store_bit_field (new_rtx
, GET_MODE_BITSIZE (mode
), 0, SImode
, val
);
4769 new_rtx
= expand_simple_binop (SImode
, AND
, new_rtx
, ac
.modemaski
,
4770 NULL_RTX
, 1, OPTAB_DIRECT
);
4771 new_rtx
= expand_simple_binop (SImode
, IOR
, new_rtx
, val
,
4772 NULL_RTX
, 1, OPTAB_DIRECT
);
4778 new_rtx
= expand_simple_binop (SImode
, code
, new_rtx
, val
,
4779 NULL_RTX
, 1, OPTAB_DIRECT
);
4781 case MULT
: /* NAND */
4782 new_rtx
= expand_simple_binop (SImode
, AND
, new_rtx
, val
,
4783 NULL_RTX
, 1, OPTAB_DIRECT
);
4784 new_rtx
= expand_simple_binop (SImode
, XOR
, new_rtx
, ac
.modemask
,
4785 NULL_RTX
, 1, OPTAB_DIRECT
);
4791 s390_emit_jump (csloop
, s390_emit_compare_and_swap (NE
, cmp
,
4792 ac
.memsi
, cmp
, new_rtx
));
4794 /* Return the correct part of the bitfield. */
4796 convert_move (target
, expand_simple_binop (SImode
, LSHIFTRT
,
4797 after
? new_rtx
: cmp
, ac
.shift
,
4798 NULL_RTX
, 1, OPTAB_DIRECT
), 1);
4801 /* This is called from dwarf2out.c via TARGET_ASM_OUTPUT_DWARF_DTPREL.
4802 We need to emit DTP-relative relocations. */
4804 static void s390_output_dwarf_dtprel (FILE *, int, rtx
) ATTRIBUTE_UNUSED
;
4807 s390_output_dwarf_dtprel (FILE *file
, int size
, rtx x
)
4812 fputs ("\t.long\t", file
);
4815 fputs ("\t.quad\t", file
);
4820 output_addr_const (file
, x
);
4821 fputs ("@DTPOFF", file
);
4824 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
4825 /* Implement TARGET_MANGLE_TYPE. */
4828 s390_mangle_type (const_tree type
)
4830 if (TYPE_MAIN_VARIANT (type
) == long_double_type_node
4831 && TARGET_LONG_DOUBLE_128
)
4834 /* For all other types, use normal C++ mangling. */
4839 /* In the name of slightly smaller debug output, and to cater to
4840 general assembler lossage, recognize various UNSPEC sequences
4841 and turn them back into a direct symbol reference. */
4844 s390_delegitimize_address (rtx orig_x
)
4848 orig_x
= delegitimize_mem_from_attrs (orig_x
);
4850 if (GET_CODE (x
) != MEM
)
4854 if (GET_CODE (x
) == PLUS
4855 && GET_CODE (XEXP (x
, 1)) == CONST
4856 && GET_CODE (XEXP (x
, 0)) == REG
4857 && REGNO (XEXP (x
, 0)) == PIC_OFFSET_TABLE_REGNUM
)
4859 y
= XEXP (XEXP (x
, 1), 0);
4860 if (GET_CODE (y
) == UNSPEC
4861 && XINT (y
, 1) == UNSPEC_GOT
)
4862 return XVECEXP (y
, 0, 0);
4866 if (GET_CODE (x
) == CONST
)
4869 if (GET_CODE (y
) == UNSPEC
4870 && XINT (y
, 1) == UNSPEC_GOTENT
)
4871 return XVECEXP (y
, 0, 0);
4878 /* Output operand OP to stdio stream FILE.
4879 OP is an address (register + offset) which is not used to address data;
4880 instead the rightmost bits are interpreted as the value. */
4883 print_shift_count_operand (FILE *file
, rtx op
)
4885 HOST_WIDE_INT offset
;
4888 /* Extract base register and offset. */
4889 if (!s390_decompose_shift_count (op
, &base
, &offset
))
4895 gcc_assert (GET_CODE (base
) == REG
);
4896 gcc_assert (REGNO (base
) < FIRST_PSEUDO_REGISTER
);
4897 gcc_assert (REGNO_REG_CLASS (REGNO (base
)) == ADDR_REGS
);
4900 /* Offsets are constricted to twelve bits. */
4901 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, offset
& ((1 << 12) - 1));
4903 fprintf (file
, "(%s)", reg_names
[REGNO (base
)]);
4906 /* See 'get_some_local_dynamic_name'. */
4909 get_some_local_dynamic_name_1 (rtx
*px
, void *data ATTRIBUTE_UNUSED
)
4913 if (GET_CODE (x
) == SYMBOL_REF
&& CONSTANT_POOL_ADDRESS_P (x
))
4915 x
= get_pool_constant (x
);
4916 return for_each_rtx (&x
, get_some_local_dynamic_name_1
, 0);
4919 if (GET_CODE (x
) == SYMBOL_REF
4920 && tls_symbolic_operand (x
) == TLS_MODEL_LOCAL_DYNAMIC
)
4922 cfun
->machine
->some_ld_name
= XSTR (x
, 0);
4929 /* Locate some local-dynamic symbol still in use by this function
4930 so that we can print its name in local-dynamic base patterns. */
4933 get_some_local_dynamic_name (void)
4937 if (cfun
->machine
->some_ld_name
)
4938 return cfun
->machine
->some_ld_name
;
4940 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
4942 && for_each_rtx (&PATTERN (insn
), get_some_local_dynamic_name_1
, 0))
4943 return cfun
->machine
->some_ld_name
;
4948 /* Output machine-dependent UNSPECs occurring in address constant X
4949 in assembler syntax to stdio stream FILE. Returns true if the
4950 constant X could be recognized, false otherwise. */
4953 s390_output_addr_const_extra (FILE *file
, rtx x
)
4955 if (GET_CODE (x
) == UNSPEC
&& XVECLEN (x
, 0) == 1)
4956 switch (XINT (x
, 1))
4959 output_addr_const (file
, XVECEXP (x
, 0, 0));
4960 fprintf (file
, "@GOTENT");
4963 output_addr_const (file
, XVECEXP (x
, 0, 0));
4964 fprintf (file
, "@GOT");
4967 output_addr_const (file
, XVECEXP (x
, 0, 0));
4968 fprintf (file
, "@GOTOFF");
4971 output_addr_const (file
, XVECEXP (x
, 0, 0));
4972 fprintf (file
, "@PLT");
4975 output_addr_const (file
, XVECEXP (x
, 0, 0));
4976 fprintf (file
, "@PLTOFF");
4979 output_addr_const (file
, XVECEXP (x
, 0, 0));
4980 fprintf (file
, "@TLSGD");
4983 assemble_name (file
, get_some_local_dynamic_name ());
4984 fprintf (file
, "@TLSLDM");
4987 output_addr_const (file
, XVECEXP (x
, 0, 0));
4988 fprintf (file
, "@DTPOFF");
4991 output_addr_const (file
, XVECEXP (x
, 0, 0));
4992 fprintf (file
, "@NTPOFF");
4994 case UNSPEC_GOTNTPOFF
:
4995 output_addr_const (file
, XVECEXP (x
, 0, 0));
4996 fprintf (file
, "@GOTNTPOFF");
4998 case UNSPEC_INDNTPOFF
:
4999 output_addr_const (file
, XVECEXP (x
, 0, 0));
5000 fprintf (file
, "@INDNTPOFF");
5004 if (GET_CODE (x
) == UNSPEC
&& XVECLEN (x
, 0) == 2)
5005 switch (XINT (x
, 1))
5007 case UNSPEC_POOL_OFFSET
:
5008 x
= gen_rtx_MINUS (GET_MODE (x
), XVECEXP (x
, 0, 0), XVECEXP (x
, 0, 1));
5009 output_addr_const (file
, x
);
5015 /* Output address operand ADDR in assembler syntax to
5016 stdio stream FILE. */
5019 print_operand_address (FILE *file
, rtx addr
)
5021 struct s390_address ad
;
5023 if (s390_symref_operand_p (addr
, NULL
, NULL
))
5025 gcc_assert (TARGET_Z10
);
5026 output_addr_const (file
, addr
);
5030 if (!s390_decompose_address (addr
, &ad
)
5031 || (ad
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)))
5032 || (ad
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (ad
.indx
))))
5033 output_operand_lossage ("cannot decompose address");
5036 output_addr_const (file
, ad
.disp
);
5038 fprintf (file
, "0");
5040 if (ad
.base
&& ad
.indx
)
5041 fprintf (file
, "(%s,%s)", reg_names
[REGNO (ad
.indx
)],
5042 reg_names
[REGNO (ad
.base
)]);
5044 fprintf (file
, "(%s)", reg_names
[REGNO (ad
.base
)]);
5047 /* Output operand X in assembler syntax to stdio stream FILE.
5048 CODE specified the format flag. The following format flags
5051 'C': print opcode suffix for branch condition.
5052 'D': print opcode suffix for inverse branch condition.
5053 'E': print opcode suffix for branch on index instruction.
5054 'J': print tls_load/tls_gdcall/tls_ldcall suffix
5055 'G': print the size of the operand in bytes.
5056 'O': print only the displacement of a memory reference.
5057 'R': print only the base register of a memory reference.
5058 'S': print S-type memory reference (base+displacement).
5059 'N': print the second word of a DImode operand.
5060 'M': print the second word of a TImode operand.
5061 'Y': print shift count operand.
5063 'b': print integer X as if it's an unsigned byte.
5064 'c': print integer X as if it's an signed byte.
5065 'x': print integer X as if it's an unsigned halfword.
5066 'h': print integer X as if it's a signed halfword.
5067 'i': print the first nonzero HImode part of X.
5068 'j': print the first HImode part unequal to -1 of X.
5069 'k': print the first nonzero SImode part of X.
5070 'm': print the first SImode part unequal to -1 of X.
5071 'o': print integer X as if it's an unsigned 32bit word. */
5074 print_operand (FILE *file
, rtx x
, int code
)
5079 fprintf (file
, s390_branch_condition_mnemonic (x
, FALSE
));
5083 fprintf (file
, s390_branch_condition_mnemonic (x
, TRUE
));
5087 if (GET_CODE (x
) == LE
)
5088 fprintf (file
, "l");
5089 else if (GET_CODE (x
) == GT
)
5090 fprintf (file
, "h");
5096 if (GET_CODE (x
) == SYMBOL_REF
)
5098 fprintf (file
, "%s", ":tls_load:");
5099 output_addr_const (file
, x
);
5101 else if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_TLSGD
)
5103 fprintf (file
, "%s", ":tls_gdcall:");
5104 output_addr_const (file
, XVECEXP (x
, 0, 0));
5106 else if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_TLSLDM
)
5108 fprintf (file
, "%s", ":tls_ldcall:");
5109 assemble_name (file
, get_some_local_dynamic_name ());
5116 fprintf (file
, "%u", GET_MODE_SIZE (GET_MODE (x
)));
5121 struct s390_address ad
;
5124 gcc_assert (GET_CODE (x
) == MEM
);
5125 ret
= s390_decompose_address (XEXP (x
, 0), &ad
);
5127 gcc_assert (!ad
.base
|| REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)));
5128 gcc_assert (!ad
.indx
);
5131 output_addr_const (file
, ad
.disp
);
5133 fprintf (file
, "0");
5139 struct s390_address ad
;
5142 gcc_assert (GET_CODE (x
) == MEM
);
5143 ret
= s390_decompose_address (XEXP (x
, 0), &ad
);
5145 gcc_assert (!ad
.base
|| REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)));
5146 gcc_assert (!ad
.indx
);
5149 fprintf (file
, "%s", reg_names
[REGNO (ad
.base
)]);
5151 fprintf (file
, "0");
5157 struct s390_address ad
;
5160 gcc_assert (GET_CODE (x
) == MEM
);
5161 ret
= s390_decompose_address (XEXP (x
, 0), &ad
);
5163 gcc_assert (!ad
.base
|| REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)));
5164 gcc_assert (!ad
.indx
);
5167 output_addr_const (file
, ad
.disp
);
5169 fprintf (file
, "0");
5172 fprintf (file
, "(%s)", reg_names
[REGNO (ad
.base
)]);
5177 if (GET_CODE (x
) == REG
)
5178 x
= gen_rtx_REG (GET_MODE (x
), REGNO (x
) + 1);
5179 else if (GET_CODE (x
) == MEM
)
5180 x
= change_address (x
, VOIDmode
, plus_constant (XEXP (x
, 0), 4));
5186 if (GET_CODE (x
) == REG
)
5187 x
= gen_rtx_REG (GET_MODE (x
), REGNO (x
) + 1);
5188 else if (GET_CODE (x
) == MEM
)
5189 x
= change_address (x
, VOIDmode
, plus_constant (XEXP (x
, 0), 8));
5195 print_shift_count_operand (file
, x
);
5199 switch (GET_CODE (x
))
5202 fprintf (file
, "%s", reg_names
[REGNO (x
)]);
5206 output_address (XEXP (x
, 0));
5213 output_addr_const (file
, x
);
5218 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) & 0xff);
5219 else if (code
== 'c')
5220 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ((INTVAL (x
) & 0xff) ^ 0x80) - 0x80);
5221 else if (code
== 'x')
5222 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) & 0xffff);
5223 else if (code
== 'h')
5224 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ((INTVAL (x
) & 0xffff) ^ 0x8000) - 0x8000);
5225 else if (code
== 'i')
5226 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5227 s390_extract_part (x
, HImode
, 0));
5228 else if (code
== 'j')
5229 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5230 s390_extract_part (x
, HImode
, -1));
5231 else if (code
== 'k')
5232 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5233 s390_extract_part (x
, SImode
, 0));
5234 else if (code
== 'm')
5235 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5236 s390_extract_part (x
, SImode
, -1));
5237 else if (code
== 'o')
5238 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) & 0xffffffff);
5240 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
));
5244 gcc_assert (GET_MODE (x
) == VOIDmode
);
5246 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, CONST_DOUBLE_LOW (x
) & 0xff);
5247 else if (code
== 'x')
5248 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, CONST_DOUBLE_LOW (x
) & 0xffff);
5249 else if (code
== 'h')
5250 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ((CONST_DOUBLE_LOW (x
) & 0xffff) ^ 0x8000) - 0x8000);
5256 fatal_insn ("UNKNOWN in print_operand !?", x
);
5261 /* Target hook for assembling integer objects. We need to define it
5262 here to work a round a bug in some versions of GAS, which couldn't
5263 handle values smaller than INT_MIN when printed in decimal. */
5266 s390_assemble_integer (rtx x
, unsigned int size
, int aligned_p
)
5268 if (size
== 8 && aligned_p
5269 && GET_CODE (x
) == CONST_INT
&& INTVAL (x
) < INT_MIN
)
5271 fprintf (asm_out_file
, "\t.quad\t" HOST_WIDE_INT_PRINT_HEX
"\n",
5275 return default_assemble_integer (x
, size
, aligned_p
);
5278 /* Returns true if register REGNO is used for forming
5279 a memory address in expression X. */
5282 reg_used_in_mem_p (int regno
, rtx x
)
5284 enum rtx_code code
= GET_CODE (x
);
5290 if (refers_to_regno_p (regno
, regno
+1,
5294 else if (code
== SET
5295 && GET_CODE (SET_DEST (x
)) == PC
)
5297 if (refers_to_regno_p (regno
, regno
+1,
5302 fmt
= GET_RTX_FORMAT (code
);
5303 for (i
= GET_RTX_LENGTH (code
) - 1; i
>= 0; i
--)
5306 && reg_used_in_mem_p (regno
, XEXP (x
, i
)))
5309 else if (fmt
[i
] == 'E')
5310 for (j
= 0; j
< XVECLEN (x
, i
); j
++)
5311 if (reg_used_in_mem_p (regno
, XVECEXP (x
, i
, j
)))
5317 /* Returns true if expression DEP_RTX sets an address register
5318 used by instruction INSN to address memory. */
5321 addr_generation_dependency_p (rtx dep_rtx
, rtx insn
)
5325 if (GET_CODE (dep_rtx
) == INSN
)
5326 dep_rtx
= PATTERN (dep_rtx
);
5328 if (GET_CODE (dep_rtx
) == SET
)
5330 target
= SET_DEST (dep_rtx
);
5331 if (GET_CODE (target
) == STRICT_LOW_PART
)
5332 target
= XEXP (target
, 0);
5333 while (GET_CODE (target
) == SUBREG
)
5334 target
= SUBREG_REG (target
);
5336 if (GET_CODE (target
) == REG
)
5338 int regno
= REGNO (target
);
5340 if (s390_safe_attr_type (insn
) == TYPE_LA
)
5342 pat
= PATTERN (insn
);
5343 if (GET_CODE (pat
) == PARALLEL
)
5345 gcc_assert (XVECLEN (pat
, 0) == 2);
5346 pat
= XVECEXP (pat
, 0, 0);
5348 gcc_assert (GET_CODE (pat
) == SET
);
5349 return refers_to_regno_p (regno
, regno
+1, SET_SRC (pat
), 0);
5351 else if (get_attr_atype (insn
) == ATYPE_AGEN
)
5352 return reg_used_in_mem_p (regno
, PATTERN (insn
));
5358 /* Return 1, if dep_insn sets register used in insn in the agen unit. */
5361 s390_agen_dep_p (rtx dep_insn
, rtx insn
)
5363 rtx dep_rtx
= PATTERN (dep_insn
);
5366 if (GET_CODE (dep_rtx
) == SET
5367 && addr_generation_dependency_p (dep_rtx
, insn
))
5369 else if (GET_CODE (dep_rtx
) == PARALLEL
)
5371 for (i
= 0; i
< XVECLEN (dep_rtx
, 0); i
++)
5373 if (addr_generation_dependency_p (XVECEXP (dep_rtx
, 0, i
), insn
))
5381 /* A C statement (sans semicolon) to update the integer scheduling priority
5382 INSN_PRIORITY (INSN). Increase the priority to execute the INSN earlier,
5383 reduce the priority to execute INSN later. Do not define this macro if
5384 you do not need to adjust the scheduling priorities of insns.
5386 A STD instruction should be scheduled earlier,
5387 in order to use the bypass. */
5391 s390_adjust_priority (rtx insn ATTRIBUTE_UNUSED
, int priority
)
5393 if (! INSN_P (insn
))
5396 if (s390_tune
!= PROCESSOR_2084_Z990
5397 && s390_tune
!= PROCESSOR_2094_Z9_109
5398 && s390_tune
!= PROCESSOR_2097_Z10
)
5401 switch (s390_safe_attr_type (insn
))
5405 priority
= priority
<< 3;
5409 priority
= priority
<< 1;
5418 /* The number of instructions that can be issued per cycle. */
5421 s390_issue_rate (void)
5425 case PROCESSOR_2084_Z990
:
5426 case PROCESSOR_2094_Z9_109
:
5428 case PROCESSOR_2097_Z10
:
5436 s390_first_cycle_multipass_dfa_lookahead (void)
5441 /* Annotate every literal pool reference in X by an UNSPEC_LTREF expression.
5442 Fix up MEMs as required. */
5445 annotate_constant_pool_refs (rtx
*x
)
5450 gcc_assert (GET_CODE (*x
) != SYMBOL_REF
5451 || !CONSTANT_POOL_ADDRESS_P (*x
));
5453 /* Literal pool references can only occur inside a MEM ... */
5454 if (GET_CODE (*x
) == MEM
)
5456 rtx memref
= XEXP (*x
, 0);
5458 if (GET_CODE (memref
) == SYMBOL_REF
5459 && CONSTANT_POOL_ADDRESS_P (memref
))
5461 rtx base
= cfun
->machine
->base_reg
;
5462 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, memref
, base
),
5465 *x
= replace_equiv_address (*x
, addr
);
5469 if (GET_CODE (memref
) == CONST
5470 && GET_CODE (XEXP (memref
, 0)) == PLUS
5471 && GET_CODE (XEXP (XEXP (memref
, 0), 1)) == CONST_INT
5472 && GET_CODE (XEXP (XEXP (memref
, 0), 0)) == SYMBOL_REF
5473 && CONSTANT_POOL_ADDRESS_P (XEXP (XEXP (memref
, 0), 0)))
5475 HOST_WIDE_INT off
= INTVAL (XEXP (XEXP (memref
, 0), 1));
5476 rtx sym
= XEXP (XEXP (memref
, 0), 0);
5477 rtx base
= cfun
->machine
->base_reg
;
5478 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, sym
, base
),
5481 *x
= replace_equiv_address (*x
, plus_constant (addr
, off
));
5486 /* ... or a load-address type pattern. */
5487 if (GET_CODE (*x
) == SET
)
5489 rtx addrref
= SET_SRC (*x
);
5491 if (GET_CODE (addrref
) == SYMBOL_REF
5492 && CONSTANT_POOL_ADDRESS_P (addrref
))
5494 rtx base
= cfun
->machine
->base_reg
;
5495 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, addrref
, base
),
5498 SET_SRC (*x
) = addr
;
5502 if (GET_CODE (addrref
) == CONST
5503 && GET_CODE (XEXP (addrref
, 0)) == PLUS
5504 && GET_CODE (XEXP (XEXP (addrref
, 0), 1)) == CONST_INT
5505 && GET_CODE (XEXP (XEXP (addrref
, 0), 0)) == SYMBOL_REF
5506 && CONSTANT_POOL_ADDRESS_P (XEXP (XEXP (addrref
, 0), 0)))
5508 HOST_WIDE_INT off
= INTVAL (XEXP (XEXP (addrref
, 0), 1));
5509 rtx sym
= XEXP (XEXP (addrref
, 0), 0);
5510 rtx base
= cfun
->machine
->base_reg
;
5511 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, sym
, base
),
5514 SET_SRC (*x
) = plus_constant (addr
, off
);
5519 /* Annotate LTREL_BASE as well. */
5520 if (GET_CODE (*x
) == UNSPEC
5521 && XINT (*x
, 1) == UNSPEC_LTREL_BASE
)
5523 rtx base
= cfun
->machine
->base_reg
;
5524 *x
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, XVECEXP (*x
, 0, 0), base
),
5529 fmt
= GET_RTX_FORMAT (GET_CODE (*x
));
5530 for (i
= GET_RTX_LENGTH (GET_CODE (*x
)) - 1; i
>= 0; i
--)
5534 annotate_constant_pool_refs (&XEXP (*x
, i
));
5536 else if (fmt
[i
] == 'E')
5538 for (j
= 0; j
< XVECLEN (*x
, i
); j
++)
5539 annotate_constant_pool_refs (&XVECEXP (*x
, i
, j
));
5544 /* Split all branches that exceed the maximum distance.
5545 Returns true if this created a new literal pool entry. */
5548 s390_split_branches (void)
5550 rtx temp_reg
= gen_rtx_REG (Pmode
, RETURN_REGNUM
);
5551 int new_literal
= 0, ret
;
5552 rtx insn
, pat
, tmp
, target
;
5555 /* We need correct insn addresses. */
5557 shorten_branches (get_insns ());
5559 /* Find all branches that exceed 64KB, and split them. */
5561 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
5563 if (GET_CODE (insn
) != JUMP_INSN
)
5566 pat
= PATTERN (insn
);
5567 if (GET_CODE (pat
) == PARALLEL
&& XVECLEN (pat
, 0) > 2)
5568 pat
= XVECEXP (pat
, 0, 0);
5569 if (GET_CODE (pat
) != SET
|| SET_DEST (pat
) != pc_rtx
)
5572 if (GET_CODE (SET_SRC (pat
)) == LABEL_REF
)
5574 label
= &SET_SRC (pat
);
5576 else if (GET_CODE (SET_SRC (pat
)) == IF_THEN_ELSE
)
5578 if (GET_CODE (XEXP (SET_SRC (pat
), 1)) == LABEL_REF
)
5579 label
= &XEXP (SET_SRC (pat
), 1);
5580 else if (GET_CODE (XEXP (SET_SRC (pat
), 2)) == LABEL_REF
)
5581 label
= &XEXP (SET_SRC (pat
), 2);
5588 if (get_attr_length (insn
) <= 4)
5591 /* We are going to use the return register as scratch register,
5592 make sure it will be saved/restored by the prologue/epilogue. */
5593 cfun_frame_layout
.save_return_addr_p
= 1;
5598 tmp
= force_const_mem (Pmode
, *label
);
5599 tmp
= emit_insn_before (gen_rtx_SET (Pmode
, temp_reg
, tmp
), insn
);
5600 INSN_ADDRESSES_NEW (tmp
, -1);
5601 annotate_constant_pool_refs (&PATTERN (tmp
));
5608 target
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, *label
),
5609 UNSPEC_LTREL_OFFSET
);
5610 target
= gen_rtx_CONST (Pmode
, target
);
5611 target
= force_const_mem (Pmode
, target
);
5612 tmp
= emit_insn_before (gen_rtx_SET (Pmode
, temp_reg
, target
), insn
);
5613 INSN_ADDRESSES_NEW (tmp
, -1);
5614 annotate_constant_pool_refs (&PATTERN (tmp
));
5616 target
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, XEXP (target
, 0),
5617 cfun
->machine
->base_reg
),
5619 target
= gen_rtx_PLUS (Pmode
, temp_reg
, target
);
5622 ret
= validate_change (insn
, label
, target
, 0);
5630 /* Find an annotated literal pool symbol referenced in RTX X,
5631 and store it at REF. Will abort if X contains references to
5632 more than one such pool symbol; multiple references to the same
5633 symbol are allowed, however.
5635 The rtx pointed to by REF must be initialized to NULL_RTX
5636 by the caller before calling this routine. */
5639 find_constant_pool_ref (rtx x
, rtx
*ref
)
5644 /* Ignore LTREL_BASE references. */
5645 if (GET_CODE (x
) == UNSPEC
5646 && XINT (x
, 1) == UNSPEC_LTREL_BASE
)
5648 /* Likewise POOL_ENTRY insns. */
5649 if (GET_CODE (x
) == UNSPEC_VOLATILE
5650 && XINT (x
, 1) == UNSPECV_POOL_ENTRY
)
5653 gcc_assert (GET_CODE (x
) != SYMBOL_REF
5654 || !CONSTANT_POOL_ADDRESS_P (x
));
5656 if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_LTREF
)
5658 rtx sym
= XVECEXP (x
, 0, 0);
5659 gcc_assert (GET_CODE (sym
) == SYMBOL_REF
5660 && CONSTANT_POOL_ADDRESS_P (sym
));
5662 if (*ref
== NULL_RTX
)
5665 gcc_assert (*ref
== sym
);
5670 fmt
= GET_RTX_FORMAT (GET_CODE (x
));
5671 for (i
= GET_RTX_LENGTH (GET_CODE (x
)) - 1; i
>= 0; i
--)
5675 find_constant_pool_ref (XEXP (x
, i
), ref
);
5677 else if (fmt
[i
] == 'E')
5679 for (j
= 0; j
< XVECLEN (x
, i
); j
++)
5680 find_constant_pool_ref (XVECEXP (x
, i
, j
), ref
);
5685 /* Replace every reference to the annotated literal pool
5686 symbol REF in X by its base plus OFFSET. */
5689 replace_constant_pool_ref (rtx
*x
, rtx ref
, rtx offset
)
5694 gcc_assert (*x
!= ref
);
5696 if (GET_CODE (*x
) == UNSPEC
5697 && XINT (*x
, 1) == UNSPEC_LTREF
5698 && XVECEXP (*x
, 0, 0) == ref
)
5700 *x
= gen_rtx_PLUS (Pmode
, XVECEXP (*x
, 0, 1), offset
);
5704 if (GET_CODE (*x
) == PLUS
5705 && GET_CODE (XEXP (*x
, 1)) == CONST_INT
5706 && GET_CODE (XEXP (*x
, 0)) == UNSPEC
5707 && XINT (XEXP (*x
, 0), 1) == UNSPEC_LTREF
5708 && XVECEXP (XEXP (*x
, 0), 0, 0) == ref
)
5710 rtx addr
= gen_rtx_PLUS (Pmode
, XVECEXP (XEXP (*x
, 0), 0, 1), offset
);
5711 *x
= plus_constant (addr
, INTVAL (XEXP (*x
, 1)));
5715 fmt
= GET_RTX_FORMAT (GET_CODE (*x
));
5716 for (i
= GET_RTX_LENGTH (GET_CODE (*x
)) - 1; i
>= 0; i
--)
5720 replace_constant_pool_ref (&XEXP (*x
, i
), ref
, offset
);
5722 else if (fmt
[i
] == 'E')
5724 for (j
= 0; j
< XVECLEN (*x
, i
); j
++)
5725 replace_constant_pool_ref (&XVECEXP (*x
, i
, j
), ref
, offset
);
5730 /* Check whether X contains an UNSPEC_LTREL_BASE.
5731 Return its constant pool symbol if found, NULL_RTX otherwise. */
5734 find_ltrel_base (rtx x
)
5739 if (GET_CODE (x
) == UNSPEC
5740 && XINT (x
, 1) == UNSPEC_LTREL_BASE
)
5741 return XVECEXP (x
, 0, 0);
5743 fmt
= GET_RTX_FORMAT (GET_CODE (x
));
5744 for (i
= GET_RTX_LENGTH (GET_CODE (x
)) - 1; i
>= 0; i
--)
5748 rtx fnd
= find_ltrel_base (XEXP (x
, i
));
5752 else if (fmt
[i
] == 'E')
5754 for (j
= 0; j
< XVECLEN (x
, i
); j
++)
5756 rtx fnd
= find_ltrel_base (XVECEXP (x
, i
, j
));
5766 /* Replace any occurrence of UNSPEC_LTREL_BASE in X with its base. */
5769 replace_ltrel_base (rtx
*x
)
5774 if (GET_CODE (*x
) == UNSPEC
5775 && XINT (*x
, 1) == UNSPEC_LTREL_BASE
)
5777 *x
= XVECEXP (*x
, 0, 1);
5781 fmt
= GET_RTX_FORMAT (GET_CODE (*x
));
5782 for (i
= GET_RTX_LENGTH (GET_CODE (*x
)) - 1; i
>= 0; i
--)
5786 replace_ltrel_base (&XEXP (*x
, i
));
5788 else if (fmt
[i
] == 'E')
5790 for (j
= 0; j
< XVECLEN (*x
, i
); j
++)
5791 replace_ltrel_base (&XVECEXP (*x
, i
, j
));
5797 /* We keep a list of constants which we have to add to internal
5798 constant tables in the middle of large functions. */
5800 #define NR_C_MODES 11
5801 enum machine_mode constant_modes
[NR_C_MODES
] =
5803 TFmode
, TImode
, TDmode
,
5804 DFmode
, DImode
, DDmode
,
5805 SFmode
, SImode
, SDmode
,
5812 struct constant
*next
;
5817 struct constant_pool
5819 struct constant_pool
*next
;
5823 rtx emit_pool_after
;
5825 struct constant
*constants
[NR_C_MODES
];
5826 struct constant
*execute
;
5831 /* Allocate new constant_pool structure. */
5833 static struct constant_pool
*
5834 s390_alloc_pool (void)
5836 struct constant_pool
*pool
;
5839 pool
= (struct constant_pool
*) xmalloc (sizeof *pool
);
5841 for (i
= 0; i
< NR_C_MODES
; i
++)
5842 pool
->constants
[i
] = NULL
;
5844 pool
->execute
= NULL
;
5845 pool
->label
= gen_label_rtx ();
5846 pool
->first_insn
= NULL_RTX
;
5847 pool
->pool_insn
= NULL_RTX
;
5848 pool
->insns
= BITMAP_ALLOC (NULL
);
5850 pool
->emit_pool_after
= NULL_RTX
;
5855 /* Create new constant pool covering instructions starting at INSN
5856 and chain it to the end of POOL_LIST. */
5858 static struct constant_pool
*
5859 s390_start_pool (struct constant_pool
**pool_list
, rtx insn
)
5861 struct constant_pool
*pool
, **prev
;
5863 pool
= s390_alloc_pool ();
5864 pool
->first_insn
= insn
;
5866 for (prev
= pool_list
; *prev
; prev
= &(*prev
)->next
)
5873 /* End range of instructions covered by POOL at INSN and emit
5874 placeholder insn representing the pool. */
5877 s390_end_pool (struct constant_pool
*pool
, rtx insn
)
5879 rtx pool_size
= GEN_INT (pool
->size
+ 8 /* alignment slop */);
5882 insn
= get_last_insn ();
5884 pool
->pool_insn
= emit_insn_after (gen_pool (pool_size
), insn
);
5885 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
5888 /* Add INSN to the list of insns covered by POOL. */
5891 s390_add_pool_insn (struct constant_pool
*pool
, rtx insn
)
5893 bitmap_set_bit (pool
->insns
, INSN_UID (insn
));
5896 /* Return pool out of POOL_LIST that covers INSN. */
5898 static struct constant_pool
*
5899 s390_find_pool (struct constant_pool
*pool_list
, rtx insn
)
5901 struct constant_pool
*pool
;
5903 for (pool
= pool_list
; pool
; pool
= pool
->next
)
5904 if (bitmap_bit_p (pool
->insns
, INSN_UID (insn
)))
5910 /* Add constant VAL of mode MODE to the constant pool POOL. */
5913 s390_add_constant (struct constant_pool
*pool
, rtx val
, enum machine_mode mode
)
5918 for (i
= 0; i
< NR_C_MODES
; i
++)
5919 if (constant_modes
[i
] == mode
)
5921 gcc_assert (i
!= NR_C_MODES
);
5923 for (c
= pool
->constants
[i
]; c
!= NULL
; c
= c
->next
)
5924 if (rtx_equal_p (val
, c
->value
))
5929 c
= (struct constant
*) xmalloc (sizeof *c
);
5931 c
->label
= gen_label_rtx ();
5932 c
->next
= pool
->constants
[i
];
5933 pool
->constants
[i
] = c
;
5934 pool
->size
+= GET_MODE_SIZE (mode
);
5938 /* Return an rtx that represents the offset of X from the start of
5942 s390_pool_offset (struct constant_pool
*pool
, rtx x
)
5946 label
= gen_rtx_LABEL_REF (GET_MODE (x
), pool
->label
);
5947 x
= gen_rtx_UNSPEC (GET_MODE (x
), gen_rtvec (2, x
, label
),
5948 UNSPEC_POOL_OFFSET
);
5949 return gen_rtx_CONST (GET_MODE (x
), x
);
5952 /* Find constant VAL of mode MODE in the constant pool POOL.
5953 Return an RTX describing the distance from the start of
5954 the pool to the location of the new constant. */
5957 s390_find_constant (struct constant_pool
*pool
, rtx val
,
5958 enum machine_mode mode
)
5963 for (i
= 0; i
< NR_C_MODES
; i
++)
5964 if (constant_modes
[i
] == mode
)
5966 gcc_assert (i
!= NR_C_MODES
);
5968 for (c
= pool
->constants
[i
]; c
!= NULL
; c
= c
->next
)
5969 if (rtx_equal_p (val
, c
->value
))
5974 return s390_pool_offset (pool
, gen_rtx_LABEL_REF (Pmode
, c
->label
));
5977 /* Check whether INSN is an execute. Return the label_ref to its
5978 execute target template if so, NULL_RTX otherwise. */
5981 s390_execute_label (rtx insn
)
5983 if (GET_CODE (insn
) == INSN
5984 && GET_CODE (PATTERN (insn
)) == PARALLEL
5985 && GET_CODE (XVECEXP (PATTERN (insn
), 0, 0)) == UNSPEC
5986 && XINT (XVECEXP (PATTERN (insn
), 0, 0), 1) == UNSPEC_EXECUTE
)
5987 return XVECEXP (XVECEXP (PATTERN (insn
), 0, 0), 0, 2);
5992 /* Add execute target for INSN to the constant pool POOL. */
5995 s390_add_execute (struct constant_pool
*pool
, rtx insn
)
5999 for (c
= pool
->execute
; c
!= NULL
; c
= c
->next
)
6000 if (INSN_UID (insn
) == INSN_UID (c
->value
))
6005 c
= (struct constant
*) xmalloc (sizeof *c
);
6007 c
->label
= gen_label_rtx ();
6008 c
->next
= pool
->execute
;
6014 /* Find execute target for INSN in the constant pool POOL.
6015 Return an RTX describing the distance from the start of
6016 the pool to the location of the execute target. */
6019 s390_find_execute (struct constant_pool
*pool
, rtx insn
)
6023 for (c
= pool
->execute
; c
!= NULL
; c
= c
->next
)
6024 if (INSN_UID (insn
) == INSN_UID (c
->value
))
6029 return s390_pool_offset (pool
, gen_rtx_LABEL_REF (Pmode
, c
->label
));
6032 /* For an execute INSN, extract the execute target template. */
6035 s390_execute_target (rtx insn
)
6037 rtx pattern
= PATTERN (insn
);
6038 gcc_assert (s390_execute_label (insn
));
6040 if (XVECLEN (pattern
, 0) == 2)
6042 pattern
= copy_rtx (XVECEXP (pattern
, 0, 1));
6046 rtvec vec
= rtvec_alloc (XVECLEN (pattern
, 0) - 1);
6049 for (i
= 0; i
< XVECLEN (pattern
, 0) - 1; i
++)
6050 RTVEC_ELT (vec
, i
) = copy_rtx (XVECEXP (pattern
, 0, i
+ 1));
6052 pattern
= gen_rtx_PARALLEL (VOIDmode
, vec
);
6058 /* Indicate that INSN cannot be duplicated. This is the case for
6059 execute insns that carry a unique label. */
6062 s390_cannot_copy_insn_p (rtx insn
)
6064 rtx label
= s390_execute_label (insn
);
6065 return label
&& label
!= const0_rtx
;
6068 /* Dump out the constants in POOL. If REMOTE_LABEL is true,
6069 do not emit the pool base label. */
6072 s390_dump_pool (struct constant_pool
*pool
, bool remote_label
)
6075 rtx insn
= pool
->pool_insn
;
6078 /* Switch to rodata section. */
6079 if (TARGET_CPU_ZARCH
)
6081 insn
= emit_insn_after (gen_pool_section_start (), insn
);
6082 INSN_ADDRESSES_NEW (insn
, -1);
6085 /* Ensure minimum pool alignment. */
6086 if (TARGET_CPU_ZARCH
)
6087 insn
= emit_insn_after (gen_pool_align (GEN_INT (8)), insn
);
6089 insn
= emit_insn_after (gen_pool_align (GEN_INT (4)), insn
);
6090 INSN_ADDRESSES_NEW (insn
, -1);
6092 /* Emit pool base label. */
6095 insn
= emit_label_after (pool
->label
, insn
);
6096 INSN_ADDRESSES_NEW (insn
, -1);
6099 /* Dump constants in descending alignment requirement order,
6100 ensuring proper alignment for every constant. */
6101 for (i
= 0; i
< NR_C_MODES
; i
++)
6102 for (c
= pool
->constants
[i
]; c
; c
= c
->next
)
6104 /* Convert UNSPEC_LTREL_OFFSET unspecs to pool-relative references. */
6105 rtx value
= copy_rtx (c
->value
);
6106 if (GET_CODE (value
) == CONST
6107 && GET_CODE (XEXP (value
, 0)) == UNSPEC
6108 && XINT (XEXP (value
, 0), 1) == UNSPEC_LTREL_OFFSET
6109 && XVECLEN (XEXP (value
, 0), 0) == 1)
6110 value
= s390_pool_offset (pool
, XVECEXP (XEXP (value
, 0), 0, 0));
6112 insn
= emit_label_after (c
->label
, insn
);
6113 INSN_ADDRESSES_NEW (insn
, -1);
6115 value
= gen_rtx_UNSPEC_VOLATILE (constant_modes
[i
],
6116 gen_rtvec (1, value
),
6117 UNSPECV_POOL_ENTRY
);
6118 insn
= emit_insn_after (value
, insn
);
6119 INSN_ADDRESSES_NEW (insn
, -1);
6122 /* Ensure minimum alignment for instructions. */
6123 insn
= emit_insn_after (gen_pool_align (GEN_INT (2)), insn
);
6124 INSN_ADDRESSES_NEW (insn
, -1);
6126 /* Output in-pool execute template insns. */
6127 for (c
= pool
->execute
; c
; c
= c
->next
)
6129 insn
= emit_label_after (c
->label
, insn
);
6130 INSN_ADDRESSES_NEW (insn
, -1);
6132 insn
= emit_insn_after (s390_execute_target (c
->value
), insn
);
6133 INSN_ADDRESSES_NEW (insn
, -1);
6136 /* Switch back to previous section. */
6137 if (TARGET_CPU_ZARCH
)
6139 insn
= emit_insn_after (gen_pool_section_end (), insn
);
6140 INSN_ADDRESSES_NEW (insn
, -1);
6143 insn
= emit_barrier_after (insn
);
6144 INSN_ADDRESSES_NEW (insn
, -1);
6146 /* Remove placeholder insn. */
6147 remove_insn (pool
->pool_insn
);
6150 /* Free all memory used by POOL. */
6153 s390_free_pool (struct constant_pool
*pool
)
6155 struct constant
*c
, *next
;
6158 for (i
= 0; i
< NR_C_MODES
; i
++)
6159 for (c
= pool
->constants
[i
]; c
; c
= next
)
6165 for (c
= pool
->execute
; c
; c
= next
)
6171 BITMAP_FREE (pool
->insns
);
6176 /* Collect main literal pool. Return NULL on overflow. */
6178 static struct constant_pool
*
6179 s390_mainpool_start (void)
6181 struct constant_pool
*pool
;
6184 pool
= s390_alloc_pool ();
6186 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6188 if (GET_CODE (insn
) == INSN
6189 && GET_CODE (PATTERN (insn
)) == SET
6190 && GET_CODE (SET_SRC (PATTERN (insn
))) == UNSPEC_VOLATILE
6191 && XINT (SET_SRC (PATTERN (insn
)), 1) == UNSPECV_MAIN_POOL
)
6193 gcc_assert (!pool
->pool_insn
);
6194 pool
->pool_insn
= insn
;
6197 if (!TARGET_CPU_ZARCH
&& s390_execute_label (insn
))
6199 s390_add_execute (pool
, insn
);
6201 else if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6203 rtx pool_ref
= NULL_RTX
;
6204 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6207 rtx constant
= get_pool_constant (pool_ref
);
6208 enum machine_mode mode
= get_pool_mode (pool_ref
);
6209 s390_add_constant (pool
, constant
, mode
);
6213 /* If hot/cold partitioning is enabled we have to make sure that
6214 the literal pool is emitted in the same section where the
6215 initialization of the literal pool base pointer takes place.
6216 emit_pool_after is only used in the non-overflow case on non
6217 Z cpus where we can emit the literal pool at the end of the
6218 function body within the text section. */
6220 && NOTE_KIND (insn
) == NOTE_INSN_SWITCH_TEXT_SECTIONS
6221 && !pool
->emit_pool_after
)
6222 pool
->emit_pool_after
= PREV_INSN (insn
);
6225 gcc_assert (pool
->pool_insn
|| pool
->size
== 0);
6227 if (pool
->size
>= 4096)
6229 /* We're going to chunkify the pool, so remove the main
6230 pool placeholder insn. */
6231 remove_insn (pool
->pool_insn
);
6233 s390_free_pool (pool
);
6237 /* If the functions ends with the section where the literal pool
6238 should be emitted set the marker to its end. */
6239 if (pool
&& !pool
->emit_pool_after
)
6240 pool
->emit_pool_after
= get_last_insn ();
6245 /* POOL holds the main literal pool as collected by s390_mainpool_start.
6246 Modify the current function to output the pool constants as well as
6247 the pool register setup instruction. */
6250 s390_mainpool_finish (struct constant_pool
*pool
)
6252 rtx base_reg
= cfun
->machine
->base_reg
;
6255 /* If the pool is empty, we're done. */
6256 if (pool
->size
== 0)
6258 /* We don't actually need a base register after all. */
6259 cfun
->machine
->base_reg
= NULL_RTX
;
6261 if (pool
->pool_insn
)
6262 remove_insn (pool
->pool_insn
);
6263 s390_free_pool (pool
);
6267 /* We need correct insn addresses. */
6268 shorten_branches (get_insns ());
6270 /* On zSeries, we use a LARL to load the pool register. The pool is
6271 located in the .rodata section, so we emit it after the function. */
6272 if (TARGET_CPU_ZARCH
)
6274 insn
= gen_main_base_64 (base_reg
, pool
->label
);
6275 insn
= emit_insn_after (insn
, pool
->pool_insn
);
6276 INSN_ADDRESSES_NEW (insn
, -1);
6277 remove_insn (pool
->pool_insn
);
6279 insn
= get_last_insn ();
6280 pool
->pool_insn
= emit_insn_after (gen_pool (const0_rtx
), insn
);
6281 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
6283 s390_dump_pool (pool
, 0);
6286 /* On S/390, if the total size of the function's code plus literal pool
6287 does not exceed 4096 bytes, we use BASR to set up a function base
6288 pointer, and emit the literal pool at the end of the function. */
6289 else if (INSN_ADDRESSES (INSN_UID (pool
->emit_pool_after
))
6290 + pool
->size
+ 8 /* alignment slop */ < 4096)
6292 insn
= gen_main_base_31_small (base_reg
, pool
->label
);
6293 insn
= emit_insn_after (insn
, pool
->pool_insn
);
6294 INSN_ADDRESSES_NEW (insn
, -1);
6295 remove_insn (pool
->pool_insn
);
6297 insn
= emit_label_after (pool
->label
, insn
);
6298 INSN_ADDRESSES_NEW (insn
, -1);
6300 /* emit_pool_after will be set by s390_mainpool_start to the
6301 last insn of the section where the literal pool should be
6303 insn
= pool
->emit_pool_after
;
6305 pool
->pool_insn
= emit_insn_after (gen_pool (const0_rtx
), insn
);
6306 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
6308 s390_dump_pool (pool
, 1);
6311 /* Otherwise, we emit an inline literal pool and use BASR to branch
6312 over it, setting up the pool register at the same time. */
6315 rtx pool_end
= gen_label_rtx ();
6317 insn
= gen_main_base_31_large (base_reg
, pool
->label
, pool_end
);
6318 insn
= emit_insn_after (insn
, pool
->pool_insn
);
6319 INSN_ADDRESSES_NEW (insn
, -1);
6320 remove_insn (pool
->pool_insn
);
6322 insn
= emit_label_after (pool
->label
, insn
);
6323 INSN_ADDRESSES_NEW (insn
, -1);
6325 pool
->pool_insn
= emit_insn_after (gen_pool (const0_rtx
), insn
);
6326 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
6328 insn
= emit_label_after (pool_end
, pool
->pool_insn
);
6329 INSN_ADDRESSES_NEW (insn
, -1);
6331 s390_dump_pool (pool
, 1);
6335 /* Replace all literal pool references. */
6337 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6340 replace_ltrel_base (&PATTERN (insn
));
6342 if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6344 rtx addr
, pool_ref
= NULL_RTX
;
6345 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6348 if (s390_execute_label (insn
))
6349 addr
= s390_find_execute (pool
, insn
);
6351 addr
= s390_find_constant (pool
, get_pool_constant (pool_ref
),
6352 get_pool_mode (pool_ref
));
6354 replace_constant_pool_ref (&PATTERN (insn
), pool_ref
, addr
);
6355 INSN_CODE (insn
) = -1;
6361 /* Free the pool. */
6362 s390_free_pool (pool
);
6365 /* POOL holds the main literal pool as collected by s390_mainpool_start.
6366 We have decided we cannot use this pool, so revert all changes
6367 to the current function that were done by s390_mainpool_start. */
6369 s390_mainpool_cancel (struct constant_pool
*pool
)
6371 /* We didn't actually change the instruction stream, so simply
6372 free the pool memory. */
6373 s390_free_pool (pool
);
6377 /* Chunkify the literal pool. */
6379 #define S390_POOL_CHUNK_MIN 0xc00
6380 #define S390_POOL_CHUNK_MAX 0xe00
6382 static struct constant_pool
*
6383 s390_chunkify_start (void)
6385 struct constant_pool
*curr_pool
= NULL
, *pool_list
= NULL
;
6388 rtx pending_ltrel
= NULL_RTX
;
6391 rtx (*gen_reload_base
) (rtx
, rtx
) =
6392 TARGET_CPU_ZARCH
? gen_reload_base_64
: gen_reload_base_31
;
6395 /* We need correct insn addresses. */
6397 shorten_branches (get_insns ());
6399 /* Scan all insns and move literals to pool chunks. */
6401 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6403 bool section_switch_p
= false;
6405 /* Check for pending LTREL_BASE. */
6408 rtx ltrel_base
= find_ltrel_base (PATTERN (insn
));
6411 gcc_assert (ltrel_base
== pending_ltrel
);
6412 pending_ltrel
= NULL_RTX
;
6416 if (!TARGET_CPU_ZARCH
&& s390_execute_label (insn
))
6419 curr_pool
= s390_start_pool (&pool_list
, insn
);
6421 s390_add_execute (curr_pool
, insn
);
6422 s390_add_pool_insn (curr_pool
, insn
);
6424 else if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6426 rtx pool_ref
= NULL_RTX
;
6427 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6430 rtx constant
= get_pool_constant (pool_ref
);
6431 enum machine_mode mode
= get_pool_mode (pool_ref
);
6434 curr_pool
= s390_start_pool (&pool_list
, insn
);
6436 s390_add_constant (curr_pool
, constant
, mode
);
6437 s390_add_pool_insn (curr_pool
, insn
);
6439 /* Don't split the pool chunk between a LTREL_OFFSET load
6440 and the corresponding LTREL_BASE. */
6441 if (GET_CODE (constant
) == CONST
6442 && GET_CODE (XEXP (constant
, 0)) == UNSPEC
6443 && XINT (XEXP (constant
, 0), 1) == UNSPEC_LTREL_OFFSET
)
6445 gcc_assert (!pending_ltrel
);
6446 pending_ltrel
= pool_ref
;
6451 if (GET_CODE (insn
) == JUMP_INSN
|| GET_CODE (insn
) == CODE_LABEL
)
6454 s390_add_pool_insn (curr_pool
, insn
);
6455 /* An LTREL_BASE must follow within the same basic block. */
6456 gcc_assert (!pending_ltrel
);
6459 if (NOTE_P (insn
) && NOTE_KIND (insn
) == NOTE_INSN_SWITCH_TEXT_SECTIONS
)
6460 section_switch_p
= true;
6463 || INSN_ADDRESSES_SIZE () <= (size_t) INSN_UID (insn
)
6464 || INSN_ADDRESSES (INSN_UID (insn
)) == -1)
6467 if (TARGET_CPU_ZARCH
)
6469 if (curr_pool
->size
< S390_POOL_CHUNK_MAX
)
6472 s390_end_pool (curr_pool
, NULL_RTX
);
6477 int chunk_size
= INSN_ADDRESSES (INSN_UID (insn
))
6478 - INSN_ADDRESSES (INSN_UID (curr_pool
->first_insn
))
6481 /* We will later have to insert base register reload insns.
6482 Those will have an effect on code size, which we need to
6483 consider here. This calculation makes rather pessimistic
6484 worst-case assumptions. */
6485 if (GET_CODE (insn
) == CODE_LABEL
)
6488 if (chunk_size
< S390_POOL_CHUNK_MIN
6489 && curr_pool
->size
< S390_POOL_CHUNK_MIN
6490 && !section_switch_p
)
6493 /* Pool chunks can only be inserted after BARRIERs ... */
6494 if (GET_CODE (insn
) == BARRIER
)
6496 s390_end_pool (curr_pool
, insn
);
6501 /* ... so if we don't find one in time, create one. */
6502 else if (chunk_size
> S390_POOL_CHUNK_MAX
6503 || curr_pool
->size
> S390_POOL_CHUNK_MAX
6504 || section_switch_p
)
6506 rtx label
, jump
, barrier
;
6508 if (!section_switch_p
)
6510 /* We can insert the barrier only after a 'real' insn. */
6511 if (GET_CODE (insn
) != INSN
&& GET_CODE (insn
) != CALL_INSN
)
6513 if (get_attr_length (insn
) == 0)
6515 /* Don't separate LTREL_BASE from the corresponding
6516 LTREL_OFFSET load. */
6522 gcc_assert (!pending_ltrel
);
6524 /* The old pool has to end before the section switch
6525 note in order to make it part of the current
6527 insn
= PREV_INSN (insn
);
6530 label
= gen_label_rtx ();
6531 jump
= emit_jump_insn_after (gen_jump (label
), insn
);
6532 barrier
= emit_barrier_after (jump
);
6533 insn
= emit_label_after (label
, barrier
);
6534 JUMP_LABEL (jump
) = label
;
6535 LABEL_NUSES (label
) = 1;
6537 INSN_ADDRESSES_NEW (jump
, -1);
6538 INSN_ADDRESSES_NEW (barrier
, -1);
6539 INSN_ADDRESSES_NEW (insn
, -1);
6541 s390_end_pool (curr_pool
, barrier
);
6549 s390_end_pool (curr_pool
, NULL_RTX
);
6550 gcc_assert (!pending_ltrel
);
6552 /* Find all labels that are branched into
6553 from an insn belonging to a different chunk. */
6555 far_labels
= BITMAP_ALLOC (NULL
);
6557 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6559 /* Labels marked with LABEL_PRESERVE_P can be target
6560 of non-local jumps, so we have to mark them.
6561 The same holds for named labels.
6563 Don't do that, however, if it is the label before
6566 if (GET_CODE (insn
) == CODE_LABEL
6567 && (LABEL_PRESERVE_P (insn
) || LABEL_NAME (insn
)))
6569 rtx vec_insn
= next_real_insn (insn
);
6570 rtx vec_pat
= vec_insn
&& GET_CODE (vec_insn
) == JUMP_INSN
?
6571 PATTERN (vec_insn
) : NULL_RTX
;
6573 || !(GET_CODE (vec_pat
) == ADDR_VEC
6574 || GET_CODE (vec_pat
) == ADDR_DIFF_VEC
))
6575 bitmap_set_bit (far_labels
, CODE_LABEL_NUMBER (insn
));
6578 /* If we have a direct jump (conditional or unconditional)
6579 or a casesi jump, check all potential targets. */
6580 else if (GET_CODE (insn
) == JUMP_INSN
)
6582 rtx pat
= PATTERN (insn
);
6583 if (GET_CODE (pat
) == PARALLEL
&& XVECLEN (pat
, 0) > 2)
6584 pat
= XVECEXP (pat
, 0, 0);
6586 if (GET_CODE (pat
) == SET
)
6588 rtx label
= JUMP_LABEL (insn
);
6591 if (s390_find_pool (pool_list
, label
)
6592 != s390_find_pool (pool_list
, insn
))
6593 bitmap_set_bit (far_labels
, CODE_LABEL_NUMBER (label
));
6596 else if (GET_CODE (pat
) == PARALLEL
6597 && XVECLEN (pat
, 0) == 2
6598 && GET_CODE (XVECEXP (pat
, 0, 0)) == SET
6599 && GET_CODE (XVECEXP (pat
, 0, 1)) == USE
6600 && GET_CODE (XEXP (XVECEXP (pat
, 0, 1), 0)) == LABEL_REF
)
6602 /* Find the jump table used by this casesi jump. */
6603 rtx vec_label
= XEXP (XEXP (XVECEXP (pat
, 0, 1), 0), 0);
6604 rtx vec_insn
= next_real_insn (vec_label
);
6605 rtx vec_pat
= vec_insn
&& GET_CODE (vec_insn
) == JUMP_INSN
?
6606 PATTERN (vec_insn
) : NULL_RTX
;
6608 && (GET_CODE (vec_pat
) == ADDR_VEC
6609 || GET_CODE (vec_pat
) == ADDR_DIFF_VEC
))
6611 int i
, diff_p
= GET_CODE (vec_pat
) == ADDR_DIFF_VEC
;
6613 for (i
= 0; i
< XVECLEN (vec_pat
, diff_p
); i
++)
6615 rtx label
= XEXP (XVECEXP (vec_pat
, diff_p
, i
), 0);
6617 if (s390_find_pool (pool_list
, label
)
6618 != s390_find_pool (pool_list
, insn
))
6619 bitmap_set_bit (far_labels
, CODE_LABEL_NUMBER (label
));
6626 /* Insert base register reload insns before every pool. */
6628 for (curr_pool
= pool_list
; curr_pool
; curr_pool
= curr_pool
->next
)
6630 rtx new_insn
= gen_reload_base (cfun
->machine
->base_reg
,
6632 rtx insn
= curr_pool
->first_insn
;
6633 INSN_ADDRESSES_NEW (emit_insn_before (new_insn
, insn
), -1);
6636 /* Insert base register reload insns at every far label. */
6638 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6639 if (GET_CODE (insn
) == CODE_LABEL
6640 && bitmap_bit_p (far_labels
, CODE_LABEL_NUMBER (insn
)))
6642 struct constant_pool
*pool
= s390_find_pool (pool_list
, insn
);
6645 rtx new_insn
= gen_reload_base (cfun
->machine
->base_reg
,
6647 INSN_ADDRESSES_NEW (emit_insn_after (new_insn
, insn
), -1);
6652 BITMAP_FREE (far_labels
);
6655 /* Recompute insn addresses. */
6657 init_insn_lengths ();
6658 shorten_branches (get_insns ());
6663 /* POOL_LIST is a chunk list as prepared by s390_chunkify_start.
6664 After we have decided to use this list, finish implementing
6665 all changes to the current function as required. */
6668 s390_chunkify_finish (struct constant_pool
*pool_list
)
6670 struct constant_pool
*curr_pool
= NULL
;
6674 /* Replace all literal pool references. */
6676 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6679 replace_ltrel_base (&PATTERN (insn
));
6681 curr_pool
= s390_find_pool (pool_list
, insn
);
6685 if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6687 rtx addr
, pool_ref
= NULL_RTX
;
6688 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6691 if (s390_execute_label (insn
))
6692 addr
= s390_find_execute (curr_pool
, insn
);
6694 addr
= s390_find_constant (curr_pool
,
6695 get_pool_constant (pool_ref
),
6696 get_pool_mode (pool_ref
));
6698 replace_constant_pool_ref (&PATTERN (insn
), pool_ref
, addr
);
6699 INSN_CODE (insn
) = -1;
6704 /* Dump out all literal pools. */
6706 for (curr_pool
= pool_list
; curr_pool
; curr_pool
= curr_pool
->next
)
6707 s390_dump_pool (curr_pool
, 0);
6709 /* Free pool list. */
6713 struct constant_pool
*next
= pool_list
->next
;
6714 s390_free_pool (pool_list
);
6719 /* POOL_LIST is a chunk list as prepared by s390_chunkify_start.
6720 We have decided we cannot use this list, so revert all changes
6721 to the current function that were done by s390_chunkify_start. */
6724 s390_chunkify_cancel (struct constant_pool
*pool_list
)
6726 struct constant_pool
*curr_pool
= NULL
;
6729 /* Remove all pool placeholder insns. */
6731 for (curr_pool
= pool_list
; curr_pool
; curr_pool
= curr_pool
->next
)
6733 /* Did we insert an extra barrier? Remove it. */
6734 rtx barrier
= PREV_INSN (curr_pool
->pool_insn
);
6735 rtx jump
= barrier
? PREV_INSN (barrier
) : NULL_RTX
;
6736 rtx label
= NEXT_INSN (curr_pool
->pool_insn
);
6738 if (jump
&& GET_CODE (jump
) == JUMP_INSN
6739 && barrier
&& GET_CODE (barrier
) == BARRIER
6740 && label
&& GET_CODE (label
) == CODE_LABEL
6741 && GET_CODE (PATTERN (jump
)) == SET
6742 && SET_DEST (PATTERN (jump
)) == pc_rtx
6743 && GET_CODE (SET_SRC (PATTERN (jump
))) == LABEL_REF
6744 && XEXP (SET_SRC (PATTERN (jump
)), 0) == label
)
6747 remove_insn (barrier
);
6748 remove_insn (label
);
6751 remove_insn (curr_pool
->pool_insn
);
6754 /* Remove all base register reload insns. */
6756 for (insn
= get_insns (); insn
; )
6758 rtx next_insn
= NEXT_INSN (insn
);
6760 if (GET_CODE (insn
) == INSN
6761 && GET_CODE (PATTERN (insn
)) == SET
6762 && GET_CODE (SET_SRC (PATTERN (insn
))) == UNSPEC
6763 && XINT (SET_SRC (PATTERN (insn
)), 1) == UNSPEC_RELOAD_BASE
)
6769 /* Free pool list. */
6773 struct constant_pool
*next
= pool_list
->next
;
6774 s390_free_pool (pool_list
);
6779 /* Output the constant pool entry EXP in mode MODE with alignment ALIGN. */
6782 s390_output_pool_entry (rtx exp
, enum machine_mode mode
, unsigned int align
)
6786 switch (GET_MODE_CLASS (mode
))
6789 case MODE_DECIMAL_FLOAT
:
6790 gcc_assert (GET_CODE (exp
) == CONST_DOUBLE
);
6792 REAL_VALUE_FROM_CONST_DOUBLE (r
, exp
);
6793 assemble_real (r
, mode
, align
);
6797 assemble_integer (exp
, GET_MODE_SIZE (mode
), align
, 1);
6798 mark_symbol_refs_as_used (exp
);
6807 /* Return an RTL expression representing the value of the return address
6808 for the frame COUNT steps up from the current frame. FRAME is the
6809 frame pointer of that frame. */
6812 s390_return_addr_rtx (int count
, rtx frame ATTRIBUTE_UNUSED
)
6817 /* Without backchain, we fail for all but the current frame. */
6819 if (!TARGET_BACKCHAIN
&& count
> 0)
6822 /* For the current frame, we need to make sure the initial
6823 value of RETURN_REGNUM is actually saved. */
6827 /* On non-z architectures branch splitting could overwrite r14. */
6828 if (TARGET_CPU_ZARCH
)
6829 return get_hard_reg_initial_val (Pmode
, RETURN_REGNUM
);
6832 cfun_frame_layout
.save_return_addr_p
= true;
6833 return gen_rtx_MEM (Pmode
, return_address_pointer_rtx
);
6837 if (TARGET_PACKED_STACK
)
6838 offset
= -2 * UNITS_PER_LONG
;
6840 offset
= RETURN_REGNUM
* UNITS_PER_LONG
;
6842 addr
= plus_constant (frame
, offset
);
6843 addr
= memory_address (Pmode
, addr
);
6844 return gen_rtx_MEM (Pmode
, addr
);
6847 /* Return an RTL expression representing the back chain stored in
6848 the current stack frame. */
6851 s390_back_chain_rtx (void)
6855 gcc_assert (TARGET_BACKCHAIN
);
6857 if (TARGET_PACKED_STACK
)
6858 chain
= plus_constant (stack_pointer_rtx
,
6859 STACK_POINTER_OFFSET
- UNITS_PER_LONG
);
6861 chain
= stack_pointer_rtx
;
6863 chain
= gen_rtx_MEM (Pmode
, chain
);
6867 /* Find first call clobbered register unused in a function.
6868 This could be used as base register in a leaf function
6869 or for holding the return address before epilogue. */
6872 find_unused_clobbered_reg (void)
6875 for (i
= 0; i
< 6; i
++)
6876 if (!df_regs_ever_live_p (i
))
6882 /* Helper function for s390_regs_ever_clobbered. Sets the fields in DATA for all
6883 clobbered hard regs in SETREG. */
6886 s390_reg_clobbered_rtx (rtx setreg
, const_rtx set_insn ATTRIBUTE_UNUSED
, void *data
)
6888 int *regs_ever_clobbered
= (int *)data
;
6889 unsigned int i
, regno
;
6890 enum machine_mode mode
= GET_MODE (setreg
);
6892 if (GET_CODE (setreg
) == SUBREG
)
6894 rtx inner
= SUBREG_REG (setreg
);
6895 if (!GENERAL_REG_P (inner
))
6897 regno
= subreg_regno (setreg
);
6899 else if (GENERAL_REG_P (setreg
))
6900 regno
= REGNO (setreg
);
6905 i
< regno
+ HARD_REGNO_NREGS (regno
, mode
);
6907 regs_ever_clobbered
[i
] = 1;
6910 /* Walks through all basic blocks of the current function looking
6911 for clobbered hard regs using s390_reg_clobbered_rtx. The fields
6912 of the passed integer array REGS_EVER_CLOBBERED are set to one for
6913 each of those regs. */
6916 s390_regs_ever_clobbered (int *regs_ever_clobbered
)
6922 memset (regs_ever_clobbered
, 0, 16 * sizeof (int));
6924 /* For non-leaf functions we have to consider all call clobbered regs to be
6926 if (!current_function_is_leaf
)
6928 for (i
= 0; i
< 16; i
++)
6929 regs_ever_clobbered
[i
] = call_really_used_regs
[i
];
6932 /* Make the "magic" eh_return registers live if necessary. For regs_ever_live
6933 this work is done by liveness analysis (mark_regs_live_at_end).
6934 Special care is needed for functions containing landing pads. Landing pads
6935 may use the eh registers, but the code which sets these registers is not
6936 contained in that function. Hence s390_regs_ever_clobbered is not able to
6937 deal with this automatically. */
6938 if (crtl
->calls_eh_return
|| cfun
->machine
->has_landing_pad_p
)
6939 for (i
= 0; EH_RETURN_DATA_REGNO (i
) != INVALID_REGNUM
; i
++)
6940 if (crtl
->calls_eh_return
6941 || (cfun
->machine
->has_landing_pad_p
6942 && df_regs_ever_live_p (EH_RETURN_DATA_REGNO (i
))))
6943 regs_ever_clobbered
[EH_RETURN_DATA_REGNO (i
)] = 1;
6945 /* For nonlocal gotos all call-saved registers have to be saved.
6946 This flag is also set for the unwinding code in libgcc.
6947 See expand_builtin_unwind_init. For regs_ever_live this is done by
6949 if (cfun
->has_nonlocal_label
)
6950 for (i
= 0; i
< 16; i
++)
6951 if (!call_really_used_regs
[i
])
6952 regs_ever_clobbered
[i
] = 1;
6954 FOR_EACH_BB (cur_bb
)
6956 FOR_BB_INSNS (cur_bb
, cur_insn
)
6958 if (INSN_P (cur_insn
))
6959 note_stores (PATTERN (cur_insn
),
6960 s390_reg_clobbered_rtx
,
6961 regs_ever_clobbered
);
6966 /* Determine the frame area which actually has to be accessed
6967 in the function epilogue. The values are stored at the
6968 given pointers AREA_BOTTOM (address of the lowest used stack
6969 address) and AREA_TOP (address of the first item which does
6970 not belong to the stack frame). */
6973 s390_frame_area (int *area_bottom
, int *area_top
)
6981 if (cfun_frame_layout
.first_restore_gpr
!= -1)
6983 b
= (cfun_frame_layout
.gprs_offset
6984 + cfun_frame_layout
.first_restore_gpr
* UNITS_PER_LONG
);
6985 t
= b
+ (cfun_frame_layout
.last_restore_gpr
6986 - cfun_frame_layout
.first_restore_gpr
+ 1) * UNITS_PER_LONG
;
6989 if (TARGET_64BIT
&& cfun_save_high_fprs_p
)
6991 b
= MIN (b
, cfun_frame_layout
.f8_offset
);
6992 t
= MAX (t
, (cfun_frame_layout
.f8_offset
6993 + cfun_frame_layout
.high_fprs
* 8));
6997 for (i
= 2; i
< 4; i
++)
6998 if (cfun_fpr_bit_p (i
))
7000 b
= MIN (b
, cfun_frame_layout
.f4_offset
+ (i
- 2) * 8);
7001 t
= MAX (t
, cfun_frame_layout
.f4_offset
+ (i
- 1) * 8);
7008 /* Fill cfun->machine with info about register usage of current function.
7009 Return in CLOBBERED_REGS which GPRs are currently considered set. */
7012 s390_register_info (int clobbered_regs
[])
7016 /* fprs 8 - 15 are call saved for 64 Bit ABI. */
7017 cfun_frame_layout
.fpr_bitmap
= 0;
7018 cfun_frame_layout
.high_fprs
= 0;
7020 for (i
= 24; i
< 32; i
++)
7021 if (df_regs_ever_live_p (i
) && !global_regs
[i
])
7023 cfun_set_fpr_bit (i
- 16);
7024 cfun_frame_layout
.high_fprs
++;
7027 /* Find first and last gpr to be saved. We trust regs_ever_live
7028 data, except that we don't save and restore global registers.
7030 Also, all registers with special meaning to the compiler need
7031 to be handled extra. */
7033 s390_regs_ever_clobbered (clobbered_regs
);
7035 for (i
= 0; i
< 16; i
++)
7036 clobbered_regs
[i
] = clobbered_regs
[i
] && !global_regs
[i
] && !fixed_regs
[i
];
7038 if (frame_pointer_needed
)
7039 clobbered_regs
[HARD_FRAME_POINTER_REGNUM
] = 1;
7042 clobbered_regs
[PIC_OFFSET_TABLE_REGNUM
]
7043 |= df_regs_ever_live_p (PIC_OFFSET_TABLE_REGNUM
);
7045 clobbered_regs
[BASE_REGNUM
]
7046 |= (cfun
->machine
->base_reg
7047 && REGNO (cfun
->machine
->base_reg
) == BASE_REGNUM
);
7049 clobbered_regs
[RETURN_REGNUM
]
7050 |= (!current_function_is_leaf
7051 || TARGET_TPF_PROFILING
7052 || cfun
->machine
->split_branches_pending_p
7053 || cfun_frame_layout
.save_return_addr_p
7054 || crtl
->calls_eh_return
7057 clobbered_regs
[STACK_POINTER_REGNUM
]
7058 |= (!current_function_is_leaf
7059 || TARGET_TPF_PROFILING
7060 || cfun_save_high_fprs_p
7061 || get_frame_size () > 0
7062 || cfun
->calls_alloca
7065 for (i
= 6; i
< 16; i
++)
7066 if (df_regs_ever_live_p (i
) || clobbered_regs
[i
])
7068 for (j
= 15; j
> i
; j
--)
7069 if (df_regs_ever_live_p (j
) || clobbered_regs
[j
])
7074 /* Nothing to save/restore. */
7075 cfun_frame_layout
.first_save_gpr_slot
= -1;
7076 cfun_frame_layout
.last_save_gpr_slot
= -1;
7077 cfun_frame_layout
.first_save_gpr
= -1;
7078 cfun_frame_layout
.first_restore_gpr
= -1;
7079 cfun_frame_layout
.last_save_gpr
= -1;
7080 cfun_frame_layout
.last_restore_gpr
= -1;
7084 /* Save slots for gprs from i to j. */
7085 cfun_frame_layout
.first_save_gpr_slot
= i
;
7086 cfun_frame_layout
.last_save_gpr_slot
= j
;
7088 for (i
= cfun_frame_layout
.first_save_gpr_slot
;
7089 i
< cfun_frame_layout
.last_save_gpr_slot
+ 1;
7091 if (clobbered_regs
[i
])
7094 for (j
= cfun_frame_layout
.last_save_gpr_slot
; j
> i
; j
--)
7095 if (clobbered_regs
[j
])
7098 if (i
== cfun_frame_layout
.last_save_gpr_slot
+ 1)
7100 /* Nothing to save/restore. */
7101 cfun_frame_layout
.first_save_gpr
= -1;
7102 cfun_frame_layout
.first_restore_gpr
= -1;
7103 cfun_frame_layout
.last_save_gpr
= -1;
7104 cfun_frame_layout
.last_restore_gpr
= -1;
7108 /* Save / Restore from gpr i to j. */
7109 cfun_frame_layout
.first_save_gpr
= i
;
7110 cfun_frame_layout
.first_restore_gpr
= i
;
7111 cfun_frame_layout
.last_save_gpr
= j
;
7112 cfun_frame_layout
.last_restore_gpr
= j
;
7118 /* Varargs functions need to save gprs 2 to 6. */
7119 if (cfun
->va_list_gpr_size
7120 && crtl
->args
.info
.gprs
< GP_ARG_NUM_REG
)
7122 int min_gpr
= crtl
->args
.info
.gprs
;
7123 int max_gpr
= min_gpr
+ cfun
->va_list_gpr_size
;
7124 if (max_gpr
> GP_ARG_NUM_REG
)
7125 max_gpr
= GP_ARG_NUM_REG
;
7127 if (cfun_frame_layout
.first_save_gpr
== -1
7128 || cfun_frame_layout
.first_save_gpr
> 2 + min_gpr
)
7130 cfun_frame_layout
.first_save_gpr
= 2 + min_gpr
;
7131 cfun_frame_layout
.first_save_gpr_slot
= 2 + min_gpr
;
7134 if (cfun_frame_layout
.last_save_gpr
== -1
7135 || cfun_frame_layout
.last_save_gpr
< 2 + max_gpr
- 1)
7137 cfun_frame_layout
.last_save_gpr
= 2 + max_gpr
- 1;
7138 cfun_frame_layout
.last_save_gpr_slot
= 2 + max_gpr
- 1;
7142 /* Mark f0, f2 for 31 bit and f0-f4 for 64 bit to be saved. */
7143 if (TARGET_HARD_FLOAT
&& cfun
->va_list_fpr_size
7144 && crtl
->args
.info
.fprs
< FP_ARG_NUM_REG
)
7146 int min_fpr
= crtl
->args
.info
.fprs
;
7147 int max_fpr
= min_fpr
+ cfun
->va_list_fpr_size
;
7148 if (max_fpr
> FP_ARG_NUM_REG
)
7149 max_fpr
= FP_ARG_NUM_REG
;
7151 /* ??? This is currently required to ensure proper location
7152 of the fpr save slots within the va_list save area. */
7153 if (TARGET_PACKED_STACK
)
7156 for (i
= min_fpr
; i
< max_fpr
; i
++)
7157 cfun_set_fpr_bit (i
);
7162 for (i
= 2; i
< 4; i
++)
7163 if (df_regs_ever_live_p (i
+ 16) && !global_regs
[i
+ 16])
7164 cfun_set_fpr_bit (i
);
7167 /* Fill cfun->machine with info about frame of current function. */
7170 s390_frame_info (void)
7174 cfun_frame_layout
.frame_size
= get_frame_size ();
7175 if (!TARGET_64BIT
&& cfun_frame_layout
.frame_size
> 0x7fff0000)
7176 fatal_error ("total size of local variables exceeds architecture limit");
7178 if (!TARGET_PACKED_STACK
)
7180 cfun_frame_layout
.backchain_offset
= 0;
7181 cfun_frame_layout
.f0_offset
= 16 * UNITS_PER_LONG
;
7182 cfun_frame_layout
.f4_offset
= cfun_frame_layout
.f0_offset
+ 2 * 8;
7183 cfun_frame_layout
.f8_offset
= -cfun_frame_layout
.high_fprs
* 8;
7184 cfun_frame_layout
.gprs_offset
= (cfun_frame_layout
.first_save_gpr_slot
7187 else if (TARGET_BACKCHAIN
) /* kernel stack layout */
7189 cfun_frame_layout
.backchain_offset
= (STACK_POINTER_OFFSET
7191 cfun_frame_layout
.gprs_offset
7192 = (cfun_frame_layout
.backchain_offset
7193 - (STACK_POINTER_REGNUM
- cfun_frame_layout
.first_save_gpr_slot
+ 1)
7198 cfun_frame_layout
.f4_offset
7199 = (cfun_frame_layout
.gprs_offset
7200 - 8 * (cfun_fpr_bit_p (2) + cfun_fpr_bit_p (3)));
7202 cfun_frame_layout
.f0_offset
7203 = (cfun_frame_layout
.f4_offset
7204 - 8 * (cfun_fpr_bit_p (0) + cfun_fpr_bit_p (1)));
7208 /* On 31 bit we have to care about alignment of the
7209 floating point regs to provide fastest access. */
7210 cfun_frame_layout
.f0_offset
7211 = ((cfun_frame_layout
.gprs_offset
7212 & ~(STACK_BOUNDARY
/ BITS_PER_UNIT
- 1))
7213 - 8 * (cfun_fpr_bit_p (0) + cfun_fpr_bit_p (1)));
7215 cfun_frame_layout
.f4_offset
7216 = (cfun_frame_layout
.f0_offset
7217 - 8 * (cfun_fpr_bit_p (2) + cfun_fpr_bit_p (3)));
7220 else /* no backchain */
7222 cfun_frame_layout
.f4_offset
7223 = (STACK_POINTER_OFFSET
7224 - 8 * (cfun_fpr_bit_p (2) + cfun_fpr_bit_p (3)));
7226 cfun_frame_layout
.f0_offset
7227 = (cfun_frame_layout
.f4_offset
7228 - 8 * (cfun_fpr_bit_p (0) + cfun_fpr_bit_p (1)));
7230 cfun_frame_layout
.gprs_offset
7231 = cfun_frame_layout
.f0_offset
- cfun_gprs_save_area_size
;
7234 if (current_function_is_leaf
7235 && !TARGET_TPF_PROFILING
7236 && cfun_frame_layout
.frame_size
== 0
7237 && !cfun_save_high_fprs_p
7238 && !cfun
->calls_alloca
7242 if (!TARGET_PACKED_STACK
)
7243 cfun_frame_layout
.frame_size
+= (STACK_POINTER_OFFSET
7244 + crtl
->outgoing_args_size
7245 + cfun_frame_layout
.high_fprs
* 8);
7248 if (TARGET_BACKCHAIN
)
7249 cfun_frame_layout
.frame_size
+= UNITS_PER_LONG
;
7251 /* No alignment trouble here because f8-f15 are only saved under
7253 cfun_frame_layout
.f8_offset
= (MIN (MIN (cfun_frame_layout
.f0_offset
,
7254 cfun_frame_layout
.f4_offset
),
7255 cfun_frame_layout
.gprs_offset
)
7256 - cfun_frame_layout
.high_fprs
* 8);
7258 cfun_frame_layout
.frame_size
+= cfun_frame_layout
.high_fprs
* 8;
7260 for (i
= 0; i
< 8; i
++)
7261 if (cfun_fpr_bit_p (i
))
7262 cfun_frame_layout
.frame_size
+= 8;
7264 cfun_frame_layout
.frame_size
+= cfun_gprs_save_area_size
;
7266 /* If under 31 bit an odd number of gprs has to be saved we have to adjust
7267 the frame size to sustain 8 byte alignment of stack frames. */
7268 cfun_frame_layout
.frame_size
= ((cfun_frame_layout
.frame_size
+
7269 STACK_BOUNDARY
/ BITS_PER_UNIT
- 1)
7270 & ~(STACK_BOUNDARY
/ BITS_PER_UNIT
- 1));
7272 cfun_frame_layout
.frame_size
+= crtl
->outgoing_args_size
;
7276 /* Generate frame layout. Fills in register and frame data for the current
7277 function in cfun->machine. This routine can be called multiple times;
7278 it will re-do the complete frame layout every time. */
7281 s390_init_frame_layout (void)
7283 HOST_WIDE_INT frame_size
;
7285 int clobbered_regs
[16];
7287 /* On S/390 machines, we may need to perform branch splitting, which
7288 will require both base and return address register. We have no
7289 choice but to assume we're going to need them until right at the
7290 end of the machine dependent reorg phase. */
7291 if (!TARGET_CPU_ZARCH
)
7292 cfun
->machine
->split_branches_pending_p
= true;
7296 frame_size
= cfun_frame_layout
.frame_size
;
7298 /* Try to predict whether we'll need the base register. */
7299 base_used
= cfun
->machine
->split_branches_pending_p
7300 || crtl
->uses_const_pool
7301 || (!DISP_IN_RANGE (frame_size
)
7302 && !CONST_OK_FOR_K (frame_size
));
7304 /* Decide which register to use as literal pool base. In small
7305 leaf functions, try to use an unused call-clobbered register
7306 as base register to avoid save/restore overhead. */
7308 cfun
->machine
->base_reg
= NULL_RTX
;
7309 else if (current_function_is_leaf
&& !df_regs_ever_live_p (5))
7310 cfun
->machine
->base_reg
= gen_rtx_REG (Pmode
, 5);
7312 cfun
->machine
->base_reg
= gen_rtx_REG (Pmode
, BASE_REGNUM
);
7314 s390_register_info (clobbered_regs
);
7317 while (frame_size
!= cfun_frame_layout
.frame_size
);
7320 /* Update frame layout. Recompute actual register save data based on
7321 current info and update regs_ever_live for the special registers.
7322 May be called multiple times, but may never cause *more* registers
7323 to be saved than s390_init_frame_layout allocated room for. */
7326 s390_update_frame_layout (void)
7328 int clobbered_regs
[16];
7330 s390_register_info (clobbered_regs
);
7332 df_set_regs_ever_live (BASE_REGNUM
,
7333 clobbered_regs
[BASE_REGNUM
] ? true : false);
7334 df_set_regs_ever_live (RETURN_REGNUM
,
7335 clobbered_regs
[RETURN_REGNUM
] ? true : false);
7336 df_set_regs_ever_live (STACK_POINTER_REGNUM
,
7337 clobbered_regs
[STACK_POINTER_REGNUM
] ? true : false);
7339 if (cfun
->machine
->base_reg
)
7340 df_set_regs_ever_live (REGNO (cfun
->machine
->base_reg
), true);
7343 /* Return true if it is legal to put a value with MODE into REGNO. */
7346 s390_hard_regno_mode_ok (unsigned int regno
, enum machine_mode mode
)
7348 switch (REGNO_REG_CLASS (regno
))
7351 if (REGNO_PAIR_OK (regno
, mode
))
7353 if (mode
== SImode
|| mode
== DImode
)
7356 if (FLOAT_MODE_P (mode
) && GET_MODE_CLASS (mode
) != MODE_VECTOR_FLOAT
)
7361 if (FRAME_REGNO_P (regno
) && mode
== Pmode
)
7366 if (REGNO_PAIR_OK (regno
, mode
))
7369 || (mode
!= TFmode
&& mode
!= TCmode
&& mode
!= TDmode
))
7374 if (GET_MODE_CLASS (mode
) == MODE_CC
)
7378 if (REGNO_PAIR_OK (regno
, mode
))
7380 if (mode
== SImode
|| mode
== Pmode
)
7391 /* Return nonzero if register OLD_REG can be renamed to register NEW_REG. */
7394 s390_hard_regno_rename_ok (unsigned int old_reg
, unsigned int new_reg
)
7396 /* Once we've decided upon a register to use as base register, it must
7397 no longer be used for any other purpose. */
7398 if (cfun
->machine
->base_reg
)
7399 if (REGNO (cfun
->machine
->base_reg
) == old_reg
7400 || REGNO (cfun
->machine
->base_reg
) == new_reg
)
7406 /* Maximum number of registers to represent a value of mode MODE
7407 in a register of class RCLASS. */
7410 s390_class_max_nregs (enum reg_class rclass
, enum machine_mode mode
)
7415 if (GET_MODE_CLASS (mode
) == MODE_COMPLEX_FLOAT
)
7416 return 2 * ((GET_MODE_SIZE (mode
) / 2 + 8 - 1) / 8);
7418 return (GET_MODE_SIZE (mode
) + 8 - 1) / 8;
7420 return (GET_MODE_SIZE (mode
) + 4 - 1) / 4;
7424 return (GET_MODE_SIZE (mode
) + UNITS_PER_WORD
- 1) / UNITS_PER_WORD
;
7427 /* Return true if register FROM can be eliminated via register TO. */
7430 s390_can_eliminate (const int from
, const int to
)
7432 /* On zSeries machines, we have not marked the base register as fixed.
7433 Instead, we have an elimination rule BASE_REGNUM -> BASE_REGNUM.
7434 If a function requires the base register, we say here that this
7435 elimination cannot be performed. This will cause reload to free
7436 up the base register (as if it were fixed). On the other hand,
7437 if the current function does *not* require the base register, we
7438 say here the elimination succeeds, which in turn allows reload
7439 to allocate the base register for any other purpose. */
7440 if (from
== BASE_REGNUM
&& to
== BASE_REGNUM
)
7442 if (TARGET_CPU_ZARCH
)
7444 s390_init_frame_layout ();
7445 return cfun
->machine
->base_reg
== NULL_RTX
;
7451 /* Everything else must point into the stack frame. */
7452 gcc_assert (to
== STACK_POINTER_REGNUM
7453 || to
== HARD_FRAME_POINTER_REGNUM
);
7455 gcc_assert (from
== FRAME_POINTER_REGNUM
7456 || from
== ARG_POINTER_REGNUM
7457 || from
== RETURN_ADDRESS_POINTER_REGNUM
);
7459 /* Make sure we actually saved the return address. */
7460 if (from
== RETURN_ADDRESS_POINTER_REGNUM
)
7461 if (!crtl
->calls_eh_return
7463 && !cfun_frame_layout
.save_return_addr_p
)
7469 /* Return offset between register FROM and TO initially after prolog. */
7472 s390_initial_elimination_offset (int from
, int to
)
7474 HOST_WIDE_INT offset
;
7477 /* ??? Why are we called for non-eliminable pairs? */
7478 if (!s390_can_eliminate (from
, to
))
7483 case FRAME_POINTER_REGNUM
:
7484 offset
= (get_frame_size()
7485 + STACK_POINTER_OFFSET
7486 + crtl
->outgoing_args_size
);
7489 case ARG_POINTER_REGNUM
:
7490 s390_init_frame_layout ();
7491 offset
= cfun_frame_layout
.frame_size
+ STACK_POINTER_OFFSET
;
7494 case RETURN_ADDRESS_POINTER_REGNUM
:
7495 s390_init_frame_layout ();
7496 index
= RETURN_REGNUM
- cfun_frame_layout
.first_save_gpr_slot
;
7497 gcc_assert (index
>= 0);
7498 offset
= cfun_frame_layout
.frame_size
+ cfun_frame_layout
.gprs_offset
;
7499 offset
+= index
* UNITS_PER_LONG
;
7513 /* Emit insn to save fpr REGNUM at offset OFFSET relative
7514 to register BASE. Return generated insn. */
7517 save_fpr (rtx base
, int offset
, int regnum
)
7520 addr
= gen_rtx_MEM (DFmode
, plus_constant (base
, offset
));
7522 if (regnum
>= 16 && regnum
<= (16 + FP_ARG_NUM_REG
))
7523 set_mem_alias_set (addr
, get_varargs_alias_set ());
7525 set_mem_alias_set (addr
, get_frame_alias_set ());
7527 return emit_move_insn (addr
, gen_rtx_REG (DFmode
, regnum
));
7530 /* Emit insn to restore fpr REGNUM from offset OFFSET relative
7531 to register BASE. Return generated insn. */
7534 restore_fpr (rtx base
, int offset
, int regnum
)
7537 addr
= gen_rtx_MEM (DFmode
, plus_constant (base
, offset
));
7538 set_mem_alias_set (addr
, get_frame_alias_set ());
7540 return emit_move_insn (gen_rtx_REG (DFmode
, regnum
), addr
);
7543 /* Return true if REGNO is a global register, but not one
7544 of the special ones that need to be saved/restored in anyway. */
7547 global_not_special_regno_p (int regno
)
7549 return (global_regs
[regno
]
7550 /* These registers are special and need to be
7551 restored in any case. */
7552 && !(regno
== STACK_POINTER_REGNUM
7553 || regno
== RETURN_REGNUM
7554 || regno
== BASE_REGNUM
7555 || (flag_pic
&& regno
== (int)PIC_OFFSET_TABLE_REGNUM
)));
7558 /* Generate insn to save registers FIRST to LAST into
7559 the register save area located at offset OFFSET
7560 relative to register BASE. */
7563 save_gprs (rtx base
, int offset
, int first
, int last
)
7565 rtx addr
, insn
, note
;
7568 addr
= plus_constant (base
, offset
);
7569 addr
= gen_rtx_MEM (Pmode
, addr
);
7571 set_mem_alias_set (addr
, get_frame_alias_set ());
7573 /* Special-case single register. */
7577 insn
= gen_movdi (addr
, gen_rtx_REG (Pmode
, first
));
7579 insn
= gen_movsi (addr
, gen_rtx_REG (Pmode
, first
));
7581 if (!global_not_special_regno_p (first
))
7582 RTX_FRAME_RELATED_P (insn
) = 1;
7587 insn
= gen_store_multiple (addr
,
7588 gen_rtx_REG (Pmode
, first
),
7589 GEN_INT (last
- first
+ 1));
7591 if (first
<= 6 && cfun
->stdarg
)
7592 for (i
= 0; i
< XVECLEN (PATTERN (insn
), 0); i
++)
7594 rtx mem
= XEXP (XVECEXP (PATTERN (insn
), 0, i
), 0);
7597 set_mem_alias_set (mem
, get_varargs_alias_set ());
7600 /* We need to set the FRAME_RELATED flag on all SETs
7601 inside the store-multiple pattern.
7603 However, we must not emit DWARF records for registers 2..5
7604 if they are stored for use by variable arguments ...
7606 ??? Unfortunately, it is not enough to simply not the
7607 FRAME_RELATED flags for those SETs, because the first SET
7608 of the PARALLEL is always treated as if it had the flag
7609 set, even if it does not. Therefore we emit a new pattern
7610 without those registers as REG_FRAME_RELATED_EXPR note. */
7612 if (first
>= 6 && !global_not_special_regno_p (first
))
7614 rtx pat
= PATTERN (insn
);
7616 for (i
= 0; i
< XVECLEN (pat
, 0); i
++)
7617 if (GET_CODE (XVECEXP (pat
, 0, i
)) == SET
7618 && !global_not_special_regno_p (REGNO (SET_SRC (XVECEXP (pat
,
7620 RTX_FRAME_RELATED_P (XVECEXP (pat
, 0, i
)) = 1;
7622 RTX_FRAME_RELATED_P (insn
) = 1;
7628 for (start
= first
>= 6 ? first
: 6; start
<= last
; start
++)
7629 if (!global_not_special_regno_p (start
))
7635 addr
= plus_constant (base
, offset
+ (start
- first
) * UNITS_PER_LONG
);
7636 note
= gen_store_multiple (gen_rtx_MEM (Pmode
, addr
),
7637 gen_rtx_REG (Pmode
, start
),
7638 GEN_INT (last
- start
+ 1));
7639 note
= PATTERN (note
);
7641 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
, note
);
7643 for (i
= 0; i
< XVECLEN (note
, 0); i
++)
7644 if (GET_CODE (XVECEXP (note
, 0, i
)) == SET
7645 && !global_not_special_regno_p (REGNO (SET_SRC (XVECEXP (note
,
7647 RTX_FRAME_RELATED_P (XVECEXP (note
, 0, i
)) = 1;
7649 RTX_FRAME_RELATED_P (insn
) = 1;
7655 /* Generate insn to restore registers FIRST to LAST from
7656 the register save area located at offset OFFSET
7657 relative to register BASE. */
7660 restore_gprs (rtx base
, int offset
, int first
, int last
)
7664 addr
= plus_constant (base
, offset
);
7665 addr
= gen_rtx_MEM (Pmode
, addr
);
7666 set_mem_alias_set (addr
, get_frame_alias_set ());
7668 /* Special-case single register. */
7672 insn
= gen_movdi (gen_rtx_REG (Pmode
, first
), addr
);
7674 insn
= gen_movsi (gen_rtx_REG (Pmode
, first
), addr
);
7679 insn
= gen_load_multiple (gen_rtx_REG (Pmode
, first
),
7681 GEN_INT (last
- first
+ 1));
7685 /* Return insn sequence to load the GOT register. */
7687 static GTY(()) rtx got_symbol
;
7689 s390_load_got (void)
7695 got_symbol
= gen_rtx_SYMBOL_REF (Pmode
, "_GLOBAL_OFFSET_TABLE_");
7696 SYMBOL_REF_FLAGS (got_symbol
) = SYMBOL_FLAG_LOCAL
;
7701 if (TARGET_CPU_ZARCH
)
7703 emit_move_insn (pic_offset_table_rtx
, got_symbol
);
7709 offset
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, got_symbol
),
7710 UNSPEC_LTREL_OFFSET
);
7711 offset
= gen_rtx_CONST (Pmode
, offset
);
7712 offset
= force_const_mem (Pmode
, offset
);
7714 emit_move_insn (pic_offset_table_rtx
, offset
);
7716 offset
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, XEXP (offset
, 0)),
7718 offset
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, offset
);
7720 emit_move_insn (pic_offset_table_rtx
, offset
);
7723 insns
= get_insns ();
7728 /* This ties together stack memory (MEM with an alias set of frame_alias_set)
7729 and the change to the stack pointer. */
7732 s390_emit_stack_tie (void)
7734 rtx mem
= gen_frame_mem (BLKmode
,
7735 gen_rtx_REG (Pmode
, STACK_POINTER_REGNUM
));
7737 emit_insn (gen_stack_tie (mem
));
7740 /* Expand the prologue into a bunch of separate insns. */
7743 s390_emit_prologue (void)
7751 /* Complete frame layout. */
7753 s390_update_frame_layout ();
7755 /* Annotate all constant pool references to let the scheduler know
7756 they implicitly use the base register. */
7758 push_topmost_sequence ();
7760 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
7763 annotate_constant_pool_refs (&PATTERN (insn
));
7764 df_insn_rescan (insn
);
7767 pop_topmost_sequence ();
7769 /* Choose best register to use for temp use within prologue.
7770 See below for why TPF must use the register 1. */
7772 if (!has_hard_reg_initial_val (Pmode
, RETURN_REGNUM
)
7773 && !current_function_is_leaf
7774 && !TARGET_TPF_PROFILING
)
7775 temp_reg
= gen_rtx_REG (Pmode
, RETURN_REGNUM
);
7777 temp_reg
= gen_rtx_REG (Pmode
, 1);
7779 /* Save call saved gprs. */
7780 if (cfun_frame_layout
.first_save_gpr
!= -1)
7782 insn
= save_gprs (stack_pointer_rtx
,
7783 cfun_frame_layout
.gprs_offset
+
7784 UNITS_PER_LONG
* (cfun_frame_layout
.first_save_gpr
7785 - cfun_frame_layout
.first_save_gpr_slot
),
7786 cfun_frame_layout
.first_save_gpr
,
7787 cfun_frame_layout
.last_save_gpr
);
7791 /* Dummy insn to mark literal pool slot. */
7793 if (cfun
->machine
->base_reg
)
7794 emit_insn (gen_main_pool (cfun
->machine
->base_reg
));
7796 offset
= cfun_frame_layout
.f0_offset
;
7798 /* Save f0 and f2. */
7799 for (i
= 0; i
< 2; i
++)
7801 if (cfun_fpr_bit_p (i
))
7803 save_fpr (stack_pointer_rtx
, offset
, i
+ 16);
7806 else if (!TARGET_PACKED_STACK
)
7810 /* Save f4 and f6. */
7811 offset
= cfun_frame_layout
.f4_offset
;
7812 for (i
= 2; i
< 4; i
++)
7814 if (cfun_fpr_bit_p (i
))
7816 insn
= save_fpr (stack_pointer_rtx
, offset
, i
+ 16);
7819 /* If f4 and f6 are call clobbered they are saved due to stdargs and
7820 therefore are not frame related. */
7821 if (!call_really_used_regs
[i
+ 16])
7822 RTX_FRAME_RELATED_P (insn
) = 1;
7824 else if (!TARGET_PACKED_STACK
)
7828 if (TARGET_PACKED_STACK
7829 && cfun_save_high_fprs_p
7830 && cfun_frame_layout
.f8_offset
+ cfun_frame_layout
.high_fprs
* 8 > 0)
7832 offset
= (cfun_frame_layout
.f8_offset
7833 + (cfun_frame_layout
.high_fprs
- 1) * 8);
7835 for (i
= 15; i
> 7 && offset
>= 0; i
--)
7836 if (cfun_fpr_bit_p (i
))
7838 insn
= save_fpr (stack_pointer_rtx
, offset
, i
+ 16);
7840 RTX_FRAME_RELATED_P (insn
) = 1;
7843 if (offset
>= cfun_frame_layout
.f8_offset
)
7847 if (!TARGET_PACKED_STACK
)
7848 next_fpr
= cfun_save_high_fprs_p
? 31 : 0;
7850 /* Decrement stack pointer. */
7852 if (cfun_frame_layout
.frame_size
> 0)
7854 rtx frame_off
= GEN_INT (-cfun_frame_layout
.frame_size
);
7857 if (s390_stack_size
)
7859 HOST_WIDE_INT stack_guard
;
7861 if (s390_stack_guard
)
7862 stack_guard
= s390_stack_guard
;
7865 /* If no value for stack guard is provided the smallest power of 2
7866 larger than the current frame size is chosen. */
7868 while (stack_guard
< cfun_frame_layout
.frame_size
)
7872 if (cfun_frame_layout
.frame_size
>= s390_stack_size
)
7874 warning (0, "frame size of function %qs is "
7875 HOST_WIDE_INT_PRINT_DEC
7876 " bytes exceeding user provided stack limit of "
7877 HOST_WIDE_INT_PRINT_DEC
" bytes. "
7878 "An unconditional trap is added.",
7879 current_function_name(), cfun_frame_layout
.frame_size
,
7881 emit_insn (gen_trap ());
7885 /* stack_guard has to be smaller than s390_stack_size.
7886 Otherwise we would emit an AND with zero which would
7887 not match the test under mask pattern. */
7888 if (stack_guard
>= s390_stack_size
)
7890 warning (0, "frame size of function %qs is "
7891 HOST_WIDE_INT_PRINT_DEC
7892 " bytes which is more than half the stack size. "
7893 "The dynamic check would not be reliable. "
7894 "No check emitted for this function.",
7895 current_function_name(),
7896 cfun_frame_layout
.frame_size
);
7900 HOST_WIDE_INT stack_check_mask
= ((s390_stack_size
- 1)
7901 & ~(stack_guard
- 1));
7903 rtx t
= gen_rtx_AND (Pmode
, stack_pointer_rtx
,
7904 GEN_INT (stack_check_mask
));
7906 emit_insn (gen_ctrapdi4 (gen_rtx_EQ (VOIDmode
,
7908 t
, const0_rtx
, const0_rtx
));
7910 emit_insn (gen_ctrapsi4 (gen_rtx_EQ (VOIDmode
,
7912 t
, const0_rtx
, const0_rtx
));
7917 if (s390_warn_framesize
> 0
7918 && cfun_frame_layout
.frame_size
>= s390_warn_framesize
)
7919 warning (0, "frame size of %qs is " HOST_WIDE_INT_PRINT_DEC
" bytes",
7920 current_function_name (), cfun_frame_layout
.frame_size
);
7922 if (s390_warn_dynamicstack_p
&& cfun
->calls_alloca
)
7923 warning (0, "%qs uses dynamic stack allocation", current_function_name ());
7925 /* Save incoming stack pointer into temp reg. */
7926 if (TARGET_BACKCHAIN
|| next_fpr
)
7927 insn
= emit_insn (gen_move_insn (temp_reg
, stack_pointer_rtx
));
7929 /* Subtract frame size from stack pointer. */
7931 if (DISP_IN_RANGE (INTVAL (frame_off
)))
7933 insn
= gen_rtx_SET (VOIDmode
, stack_pointer_rtx
,
7934 gen_rtx_PLUS (Pmode
, stack_pointer_rtx
,
7936 insn
= emit_insn (insn
);
7940 if (!CONST_OK_FOR_K (INTVAL (frame_off
)))
7941 frame_off
= force_const_mem (Pmode
, frame_off
);
7943 insn
= emit_insn (gen_add2_insn (stack_pointer_rtx
, frame_off
));
7944 annotate_constant_pool_refs (&PATTERN (insn
));
7947 RTX_FRAME_RELATED_P (insn
) = 1;
7948 real_frame_off
= GEN_INT (-cfun_frame_layout
.frame_size
);
7949 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
,
7950 gen_rtx_SET (VOIDmode
, stack_pointer_rtx
,
7951 gen_rtx_PLUS (Pmode
, stack_pointer_rtx
,
7954 /* Set backchain. */
7956 if (TARGET_BACKCHAIN
)
7958 if (cfun_frame_layout
.backchain_offset
)
7959 addr
= gen_rtx_MEM (Pmode
,
7960 plus_constant (stack_pointer_rtx
,
7961 cfun_frame_layout
.backchain_offset
));
7963 addr
= gen_rtx_MEM (Pmode
, stack_pointer_rtx
);
7964 set_mem_alias_set (addr
, get_frame_alias_set ());
7965 insn
= emit_insn (gen_move_insn (addr
, temp_reg
));
7968 /* If we support asynchronous exceptions (e.g. for Java),
7969 we need to make sure the backchain pointer is set up
7970 before any possibly trapping memory access. */
7972 if (TARGET_BACKCHAIN
&& flag_non_call_exceptions
)
7974 addr
= gen_rtx_MEM (BLKmode
, gen_rtx_SCRATCH (VOIDmode
));
7975 emit_clobber (addr
);
7979 /* Save fprs 8 - 15 (64 bit ABI). */
7981 if (cfun_save_high_fprs_p
&& next_fpr
)
7983 /* If the stack might be accessed through a different register
7984 we have to make sure that the stack pointer decrement is not
7985 moved below the use of the stack slots. */
7986 s390_emit_stack_tie ();
7988 insn
= emit_insn (gen_add2_insn (temp_reg
,
7989 GEN_INT (cfun_frame_layout
.f8_offset
)));
7993 for (i
= 24; i
<= next_fpr
; i
++)
7994 if (cfun_fpr_bit_p (i
- 16))
7996 rtx addr
= plus_constant (stack_pointer_rtx
,
7997 cfun_frame_layout
.frame_size
7998 + cfun_frame_layout
.f8_offset
8001 insn
= save_fpr (temp_reg
, offset
, i
);
8003 RTX_FRAME_RELATED_P (insn
) = 1;
8004 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
,
8005 gen_rtx_SET (VOIDmode
,
8006 gen_rtx_MEM (DFmode
, addr
),
8007 gen_rtx_REG (DFmode
, i
)));
8011 /* Set frame pointer, if needed. */
8013 if (frame_pointer_needed
)
8015 insn
= emit_move_insn (hard_frame_pointer_rtx
, stack_pointer_rtx
);
8016 RTX_FRAME_RELATED_P (insn
) = 1;
8019 /* Set up got pointer, if needed. */
8021 if (flag_pic
&& df_regs_ever_live_p (PIC_OFFSET_TABLE_REGNUM
))
8023 rtx insns
= s390_load_got ();
8025 for (insn
= insns
; insn
; insn
= NEXT_INSN (insn
))
8026 annotate_constant_pool_refs (&PATTERN (insn
));
8031 if (TARGET_TPF_PROFILING
)
8033 /* Generate a BAS instruction to serve as a function
8034 entry intercept to facilitate the use of tracing
8035 algorithms located at the branch target. */
8036 emit_insn (gen_prologue_tpf ());
8038 /* Emit a blockage here so that all code
8039 lies between the profiling mechanisms. */
8040 emit_insn (gen_blockage ());
8044 /* Expand the epilogue into a bunch of separate insns. */
8047 s390_emit_epilogue (bool sibcall
)
8049 rtx frame_pointer
, return_reg
, cfa_restores
= NULL_RTX
;
8050 int area_bottom
, area_top
, offset
= 0;
8055 if (TARGET_TPF_PROFILING
)
8058 /* Generate a BAS instruction to serve as a function
8059 entry intercept to facilitate the use of tracing
8060 algorithms located at the branch target. */
8062 /* Emit a blockage here so that all code
8063 lies between the profiling mechanisms. */
8064 emit_insn (gen_blockage ());
8066 emit_insn (gen_epilogue_tpf ());
8069 /* Check whether to use frame or stack pointer for restore. */
8071 frame_pointer
= (frame_pointer_needed
8072 ? hard_frame_pointer_rtx
: stack_pointer_rtx
);
8074 s390_frame_area (&area_bottom
, &area_top
);
8076 /* Check whether we can access the register save area.
8077 If not, increment the frame pointer as required. */
8079 if (area_top
<= area_bottom
)
8081 /* Nothing to restore. */
8083 else if (DISP_IN_RANGE (cfun_frame_layout
.frame_size
+ area_bottom
)
8084 && DISP_IN_RANGE (cfun_frame_layout
.frame_size
+ area_top
- 1))
8086 /* Area is in range. */
8087 offset
= cfun_frame_layout
.frame_size
;
8091 rtx insn
, frame_off
, cfa
;
8093 offset
= area_bottom
< 0 ? -area_bottom
: 0;
8094 frame_off
= GEN_INT (cfun_frame_layout
.frame_size
- offset
);
8096 cfa
= gen_rtx_SET (VOIDmode
, frame_pointer
,
8097 gen_rtx_PLUS (Pmode
, frame_pointer
, frame_off
));
8098 if (DISP_IN_RANGE (INTVAL (frame_off
)))
8100 insn
= gen_rtx_SET (VOIDmode
, frame_pointer
,
8101 gen_rtx_PLUS (Pmode
, frame_pointer
, frame_off
));
8102 insn
= emit_insn (insn
);
8106 if (!CONST_OK_FOR_K (INTVAL (frame_off
)))
8107 frame_off
= force_const_mem (Pmode
, frame_off
);
8109 insn
= emit_insn (gen_add2_insn (frame_pointer
, frame_off
));
8110 annotate_constant_pool_refs (&PATTERN (insn
));
8112 add_reg_note (insn
, REG_CFA_ADJUST_CFA
, cfa
);
8113 RTX_FRAME_RELATED_P (insn
) = 1;
8116 /* Restore call saved fprs. */
8120 if (cfun_save_high_fprs_p
)
8122 next_offset
= cfun_frame_layout
.f8_offset
;
8123 for (i
= 24; i
< 32; i
++)
8125 if (cfun_fpr_bit_p (i
- 16))
8127 restore_fpr (frame_pointer
,
8128 offset
+ next_offset
, i
);
8130 = alloc_reg_note (REG_CFA_RESTORE
,
8131 gen_rtx_REG (DFmode
, i
), cfa_restores
);
8140 next_offset
= cfun_frame_layout
.f4_offset
;
8141 for (i
= 18; i
< 20; i
++)
8143 if (cfun_fpr_bit_p (i
- 16))
8145 restore_fpr (frame_pointer
,
8146 offset
+ next_offset
, i
);
8148 = alloc_reg_note (REG_CFA_RESTORE
,
8149 gen_rtx_REG (DFmode
, i
), cfa_restores
);
8152 else if (!TARGET_PACKED_STACK
)
8158 /* Return register. */
8160 return_reg
= gen_rtx_REG (Pmode
, RETURN_REGNUM
);
8162 /* Restore call saved gprs. */
8164 if (cfun_frame_layout
.first_restore_gpr
!= -1)
8169 /* Check for global register and save them
8170 to stack location from where they get restored. */
8172 for (i
= cfun_frame_layout
.first_restore_gpr
;
8173 i
<= cfun_frame_layout
.last_restore_gpr
;
8176 if (global_not_special_regno_p (i
))
8178 addr
= plus_constant (frame_pointer
,
8179 offset
+ cfun_frame_layout
.gprs_offset
8180 + (i
- cfun_frame_layout
.first_save_gpr_slot
)
8182 addr
= gen_rtx_MEM (Pmode
, addr
);
8183 set_mem_alias_set (addr
, get_frame_alias_set ());
8184 emit_move_insn (addr
, gen_rtx_REG (Pmode
, i
));
8188 = alloc_reg_note (REG_CFA_RESTORE
,
8189 gen_rtx_REG (Pmode
, i
), cfa_restores
);
8194 /* Fetch return address from stack before load multiple,
8195 this will do good for scheduling. */
8197 if (cfun_frame_layout
.save_return_addr_p
8198 || (cfun_frame_layout
.first_restore_gpr
< BASE_REGNUM
8199 && cfun_frame_layout
.last_restore_gpr
> RETURN_REGNUM
))
8201 int return_regnum
= find_unused_clobbered_reg();
8204 return_reg
= gen_rtx_REG (Pmode
, return_regnum
);
8206 addr
= plus_constant (frame_pointer
,
8207 offset
+ cfun_frame_layout
.gprs_offset
8209 - cfun_frame_layout
.first_save_gpr_slot
)
8211 addr
= gen_rtx_MEM (Pmode
, addr
);
8212 set_mem_alias_set (addr
, get_frame_alias_set ());
8213 emit_move_insn (return_reg
, addr
);
8217 insn
= restore_gprs (frame_pointer
,
8218 offset
+ cfun_frame_layout
.gprs_offset
8219 + (cfun_frame_layout
.first_restore_gpr
8220 - cfun_frame_layout
.first_save_gpr_slot
)
8222 cfun_frame_layout
.first_restore_gpr
,
8223 cfun_frame_layout
.last_restore_gpr
);
8224 insn
= emit_insn (insn
);
8225 REG_NOTES (insn
) = cfa_restores
;
8226 add_reg_note (insn
, REG_CFA_DEF_CFA
,
8227 plus_constant (stack_pointer_rtx
, STACK_POINTER_OFFSET
));
8228 RTX_FRAME_RELATED_P (insn
) = 1;
8234 /* Return to caller. */
8236 p
= rtvec_alloc (2);
8238 RTVEC_ELT (p
, 0) = gen_rtx_RETURN (VOIDmode
);
8239 RTVEC_ELT (p
, 1) = gen_rtx_USE (VOIDmode
, return_reg
);
8240 emit_jump_insn (gen_rtx_PARALLEL (VOIDmode
, p
));
8245 /* Return the size in bytes of a function argument of
8246 type TYPE and/or mode MODE. At least one of TYPE or
8247 MODE must be specified. */
8250 s390_function_arg_size (enum machine_mode mode
, const_tree type
)
8253 return int_size_in_bytes (type
);
8255 /* No type info available for some library calls ... */
8256 if (mode
!= BLKmode
)
8257 return GET_MODE_SIZE (mode
);
8259 /* If we have neither type nor mode, abort */
8263 /* Return true if a function argument of type TYPE and mode MODE
8264 is to be passed in a floating-point register, if available. */
8267 s390_function_arg_float (enum machine_mode mode
, tree type
)
8269 int size
= s390_function_arg_size (mode
, type
);
8273 /* Soft-float changes the ABI: no floating-point registers are used. */
8274 if (TARGET_SOFT_FLOAT
)
8277 /* No type info available for some library calls ... */
8279 return mode
== SFmode
|| mode
== DFmode
|| mode
== SDmode
|| mode
== DDmode
;
8281 /* The ABI says that record types with a single member are treated
8282 just like that member would be. */
8283 while (TREE_CODE (type
) == RECORD_TYPE
)
8285 tree field
, single
= NULL_TREE
;
8287 for (field
= TYPE_FIELDS (type
); field
; field
= TREE_CHAIN (field
))
8289 if (TREE_CODE (field
) != FIELD_DECL
)
8292 if (single
== NULL_TREE
)
8293 single
= TREE_TYPE (field
);
8298 if (single
== NULL_TREE
)
8304 return TREE_CODE (type
) == REAL_TYPE
;
8307 /* Return true if a function argument of type TYPE and mode MODE
8308 is to be passed in an integer register, or a pair of integer
8309 registers, if available. */
8312 s390_function_arg_integer (enum machine_mode mode
, tree type
)
8314 int size
= s390_function_arg_size (mode
, type
);
8318 /* No type info available for some library calls ... */
8320 return GET_MODE_CLASS (mode
) == MODE_INT
8321 || (TARGET_SOFT_FLOAT
&& SCALAR_FLOAT_MODE_P (mode
));
8323 /* We accept small integral (and similar) types. */
8324 if (INTEGRAL_TYPE_P (type
)
8325 || POINTER_TYPE_P (type
)
8326 || TREE_CODE (type
) == OFFSET_TYPE
8327 || (TARGET_SOFT_FLOAT
&& TREE_CODE (type
) == REAL_TYPE
))
8330 /* We also accept structs of size 1, 2, 4, 8 that are not
8331 passed in floating-point registers. */
8332 if (AGGREGATE_TYPE_P (type
)
8333 && exact_log2 (size
) >= 0
8334 && !s390_function_arg_float (mode
, type
))
8340 /* Return 1 if a function argument of type TYPE and mode MODE
8341 is to be passed by reference. The ABI specifies that only
8342 structures of size 1, 2, 4, or 8 bytes are passed by value,
8343 all other structures (and complex numbers) are passed by
8347 s390_pass_by_reference (CUMULATIVE_ARGS
*ca ATTRIBUTE_UNUSED
,
8348 enum machine_mode mode
, const_tree type
,
8349 bool named ATTRIBUTE_UNUSED
)
8351 int size
= s390_function_arg_size (mode
, type
);
8357 if (AGGREGATE_TYPE_P (type
) && exact_log2 (size
) < 0)
8360 if (TREE_CODE (type
) == COMPLEX_TYPE
8361 || TREE_CODE (type
) == VECTOR_TYPE
)
8368 /* Update the data in CUM to advance over an argument of mode MODE and
8369 data type TYPE. (TYPE is null for libcalls where that information
8370 may not be available.). The boolean NAMED specifies whether the
8371 argument is a named argument (as opposed to an unnamed argument
8372 matching an ellipsis). */
8375 s390_function_arg_advance (CUMULATIVE_ARGS
*cum
, enum machine_mode mode
,
8376 tree type
, int named ATTRIBUTE_UNUSED
)
8378 if (s390_function_arg_float (mode
, type
))
8382 else if (s390_function_arg_integer (mode
, type
))
8384 int size
= s390_function_arg_size (mode
, type
);
8385 cum
->gprs
+= ((size
+ UNITS_PER_LONG
- 1) / UNITS_PER_LONG
);
8391 /* Define where to put the arguments to a function.
8392 Value is zero to push the argument on the stack,
8393 or a hard register in which to store the argument.
8395 MODE is the argument's machine mode.
8396 TYPE is the data type of the argument (as a tree).
8397 This is null for libcalls where that information may
8399 CUM is a variable of type CUMULATIVE_ARGS which gives info about
8400 the preceding args and about the function being called.
8401 NAMED is nonzero if this argument is a named parameter
8402 (otherwise it is an extra parameter matching an ellipsis).
8404 On S/390, we use general purpose registers 2 through 6 to
8405 pass integer, pointer, and certain structure arguments, and
8406 floating point registers 0 and 2 (0, 2, 4, and 6 on 64-bit)
8407 to pass floating point arguments. All remaining arguments
8408 are pushed to the stack. */
8411 s390_function_arg (CUMULATIVE_ARGS
*cum
, enum machine_mode mode
, tree type
,
8412 int named ATTRIBUTE_UNUSED
)
8414 if (s390_function_arg_float (mode
, type
))
8416 if (cum
->fprs
+ 1 > FP_ARG_NUM_REG
)
8419 return gen_rtx_REG (mode
, cum
->fprs
+ 16);
8421 else if (s390_function_arg_integer (mode
, type
))
8423 int size
= s390_function_arg_size (mode
, type
);
8424 int n_gprs
= (size
+ UNITS_PER_LONG
- 1) / UNITS_PER_LONG
;
8426 if (cum
->gprs
+ n_gprs
> GP_ARG_NUM_REG
)
8428 else if (n_gprs
== 1 || UNITS_PER_WORD
== UNITS_PER_LONG
)
8429 return gen_rtx_REG (mode
, cum
->gprs
+ 2);
8430 else if (n_gprs
== 2)
8432 rtvec p
= rtvec_alloc (2);
8435 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, cum
->gprs
+ 2),
8438 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, cum
->gprs
+ 3),
8441 return gen_rtx_PARALLEL (mode
, p
);
8445 /* After the real arguments, expand_call calls us once again
8446 with a void_type_node type. Whatever we return here is
8447 passed as operand 2 to the call expanders.
8449 We don't need this feature ... */
8450 else if (type
== void_type_node
)
8456 /* Return true if return values of type TYPE should be returned
8457 in a memory buffer whose address is passed by the caller as
8458 hidden first argument. */
8461 s390_return_in_memory (const_tree type
, const_tree fundecl ATTRIBUTE_UNUSED
)
8463 /* We accept small integral (and similar) types. */
8464 if (INTEGRAL_TYPE_P (type
)
8465 || POINTER_TYPE_P (type
)
8466 || TREE_CODE (type
) == OFFSET_TYPE
8467 || TREE_CODE (type
) == REAL_TYPE
)
8468 return int_size_in_bytes (type
) > 8;
8470 /* Aggregates and similar constructs are always returned
8472 if (AGGREGATE_TYPE_P (type
)
8473 || TREE_CODE (type
) == COMPLEX_TYPE
8474 || TREE_CODE (type
) == VECTOR_TYPE
)
8477 /* ??? We get called on all sorts of random stuff from
8478 aggregate_value_p. We can't abort, but it's not clear
8479 what's safe to return. Pretend it's a struct I guess. */
8483 /* Function arguments and return values are promoted to word size. */
8485 static enum machine_mode
8486 s390_promote_function_mode (const_tree type
, enum machine_mode mode
,
8488 const_tree fntype ATTRIBUTE_UNUSED
,
8489 int for_return ATTRIBUTE_UNUSED
)
8491 if (INTEGRAL_MODE_P (mode
)
8492 && GET_MODE_SIZE (mode
) < UNITS_PER_LONG
)
8494 if (POINTER_TYPE_P (type
))
8495 *punsignedp
= POINTERS_EXTEND_UNSIGNED
;
8502 /* Define where to return a (scalar) value of type TYPE.
8503 If TYPE is null, define where to return a (scalar)
8504 value of mode MODE from a libcall. */
8507 s390_function_value (const_tree type
, const_tree fn
, enum machine_mode mode
)
8511 int unsignedp
= TYPE_UNSIGNED (type
);
8512 mode
= promote_function_mode (type
, TYPE_MODE (type
), &unsignedp
, fn
, 1);
8515 gcc_assert (GET_MODE_CLASS (mode
) == MODE_INT
|| SCALAR_FLOAT_MODE_P (mode
));
8516 gcc_assert (GET_MODE_SIZE (mode
) <= 8);
8518 if (TARGET_HARD_FLOAT
&& SCALAR_FLOAT_MODE_P (mode
))
8519 return gen_rtx_REG (mode
, 16);
8520 else if (GET_MODE_SIZE (mode
) <= UNITS_PER_LONG
8521 || UNITS_PER_LONG
== UNITS_PER_WORD
)
8522 return gen_rtx_REG (mode
, 2);
8523 else if (GET_MODE_SIZE (mode
) == 2 * UNITS_PER_LONG
)
8525 rtvec p
= rtvec_alloc (2);
8528 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, 2), const0_rtx
);
8530 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, 3), GEN_INT (4));
8532 return gen_rtx_PARALLEL (mode
, p
);
8539 /* Create and return the va_list datatype.
8541 On S/390, va_list is an array type equivalent to
8543 typedef struct __va_list_tag
8547 void *__overflow_arg_area;
8548 void *__reg_save_area;
8551 where __gpr and __fpr hold the number of general purpose
8552 or floating point arguments used up to now, respectively,
8553 __overflow_arg_area points to the stack location of the
8554 next argument passed on the stack, and __reg_save_area
8555 always points to the start of the register area in the
8556 call frame of the current function. The function prologue
8557 saves all registers used for argument passing into this
8558 area if the function uses variable arguments. */
8561 s390_build_builtin_va_list (void)
8563 tree f_gpr
, f_fpr
, f_ovf
, f_sav
, record
, type_decl
;
8565 record
= lang_hooks
.types
.make_type (RECORD_TYPE
);
8568 build_decl (BUILTINS_LOCATION
,
8569 TYPE_DECL
, get_identifier ("__va_list_tag"), record
);
8571 f_gpr
= build_decl (BUILTINS_LOCATION
,
8572 FIELD_DECL
, get_identifier ("__gpr"),
8573 long_integer_type_node
);
8574 f_fpr
= build_decl (BUILTINS_LOCATION
,
8575 FIELD_DECL
, get_identifier ("__fpr"),
8576 long_integer_type_node
);
8577 f_ovf
= build_decl (BUILTINS_LOCATION
,
8578 FIELD_DECL
, get_identifier ("__overflow_arg_area"),
8580 f_sav
= build_decl (BUILTINS_LOCATION
,
8581 FIELD_DECL
, get_identifier ("__reg_save_area"),
8584 va_list_gpr_counter_field
= f_gpr
;
8585 va_list_fpr_counter_field
= f_fpr
;
8587 DECL_FIELD_CONTEXT (f_gpr
) = record
;
8588 DECL_FIELD_CONTEXT (f_fpr
) = record
;
8589 DECL_FIELD_CONTEXT (f_ovf
) = record
;
8590 DECL_FIELD_CONTEXT (f_sav
) = record
;
8592 TREE_CHAIN (record
) = type_decl
;
8593 TYPE_NAME (record
) = type_decl
;
8594 TYPE_FIELDS (record
) = f_gpr
;
8595 TREE_CHAIN (f_gpr
) = f_fpr
;
8596 TREE_CHAIN (f_fpr
) = f_ovf
;
8597 TREE_CHAIN (f_ovf
) = f_sav
;
8599 layout_type (record
);
8601 /* The correct type is an array type of one element. */
8602 return build_array_type (record
, build_index_type (size_zero_node
));
8605 /* Implement va_start by filling the va_list structure VALIST.
8606 STDARG_P is always true, and ignored.
8607 NEXTARG points to the first anonymous stack argument.
8609 The following global variables are used to initialize
8610 the va_list structure:
8613 holds number of gprs and fprs used for named arguments.
8614 crtl->args.arg_offset_rtx:
8615 holds the offset of the first anonymous stack argument
8616 (relative to the virtual arg pointer). */
8619 s390_va_start (tree valist
, rtx nextarg ATTRIBUTE_UNUSED
)
8621 HOST_WIDE_INT n_gpr
, n_fpr
;
8623 tree f_gpr
, f_fpr
, f_ovf
, f_sav
;
8624 tree gpr
, fpr
, ovf
, sav
, t
;
8626 f_gpr
= TYPE_FIELDS (TREE_TYPE (va_list_type_node
));
8627 f_fpr
= TREE_CHAIN (f_gpr
);
8628 f_ovf
= TREE_CHAIN (f_fpr
);
8629 f_sav
= TREE_CHAIN (f_ovf
);
8631 valist
= build_va_arg_indirect_ref (valist
);
8632 gpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_gpr
), valist
, f_gpr
, NULL_TREE
);
8633 fpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_fpr
), valist
, f_fpr
, NULL_TREE
);
8634 ovf
= build3 (COMPONENT_REF
, TREE_TYPE (f_ovf
), valist
, f_ovf
, NULL_TREE
);
8635 sav
= build3 (COMPONENT_REF
, TREE_TYPE (f_sav
), valist
, f_sav
, NULL_TREE
);
8637 /* Count number of gp and fp argument registers used. */
8639 n_gpr
= crtl
->args
.info
.gprs
;
8640 n_fpr
= crtl
->args
.info
.fprs
;
8642 if (cfun
->va_list_gpr_size
)
8644 t
= build2 (MODIFY_EXPR
, TREE_TYPE (gpr
), gpr
,
8645 build_int_cst (NULL_TREE
, n_gpr
));
8646 TREE_SIDE_EFFECTS (t
) = 1;
8647 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8650 if (cfun
->va_list_fpr_size
)
8652 t
= build2 (MODIFY_EXPR
, TREE_TYPE (fpr
), fpr
,
8653 build_int_cst (NULL_TREE
, n_fpr
));
8654 TREE_SIDE_EFFECTS (t
) = 1;
8655 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8658 /* Find the overflow area. */
8659 if (n_gpr
+ cfun
->va_list_gpr_size
> GP_ARG_NUM_REG
8660 || n_fpr
+ cfun
->va_list_fpr_size
> FP_ARG_NUM_REG
)
8662 t
= make_tree (TREE_TYPE (ovf
), virtual_incoming_args_rtx
);
8664 off
= INTVAL (crtl
->args
.arg_offset_rtx
);
8665 off
= off
< 0 ? 0 : off
;
8666 if (TARGET_DEBUG_ARG
)
8667 fprintf (stderr
, "va_start: n_gpr = %d, n_fpr = %d off %d\n",
8668 (int)n_gpr
, (int)n_fpr
, off
);
8670 t
= build2 (POINTER_PLUS_EXPR
, TREE_TYPE (ovf
), t
, size_int (off
));
8672 t
= build2 (MODIFY_EXPR
, TREE_TYPE (ovf
), ovf
, t
);
8673 TREE_SIDE_EFFECTS (t
) = 1;
8674 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8677 /* Find the register save area. */
8678 if ((cfun
->va_list_gpr_size
&& n_gpr
< GP_ARG_NUM_REG
)
8679 || (cfun
->va_list_fpr_size
&& n_fpr
< FP_ARG_NUM_REG
))
8681 t
= make_tree (TREE_TYPE (sav
), return_address_pointer_rtx
);
8682 t
= build2 (POINTER_PLUS_EXPR
, TREE_TYPE (sav
), t
,
8683 size_int (-RETURN_REGNUM
* UNITS_PER_LONG
));
8685 t
= build2 (MODIFY_EXPR
, TREE_TYPE (sav
), sav
, t
);
8686 TREE_SIDE_EFFECTS (t
) = 1;
8687 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8691 /* Implement va_arg by updating the va_list structure
8692 VALIST as required to retrieve an argument of type
8693 TYPE, and returning that argument.
8695 Generates code equivalent to:
8697 if (integral value) {
8698 if (size <= 4 && args.gpr < 5 ||
8699 size > 4 && args.gpr < 4 )
8700 ret = args.reg_save_area[args.gpr+8]
8702 ret = *args.overflow_arg_area++;
8703 } else if (float value) {
8705 ret = args.reg_save_area[args.fpr+64]
8707 ret = *args.overflow_arg_area++;
8708 } else if (aggregate value) {
8710 ret = *args.reg_save_area[args.gpr]
8712 ret = **args.overflow_arg_area++;
8716 s390_gimplify_va_arg (tree valist
, tree type
, gimple_seq
*pre_p
,
8717 gimple_seq
*post_p ATTRIBUTE_UNUSED
)
8719 tree f_gpr
, f_fpr
, f_ovf
, f_sav
;
8720 tree gpr
, fpr
, ovf
, sav
, reg
, t
, u
;
8721 int indirect_p
, size
, n_reg
, sav_ofs
, sav_scale
, max_reg
;
8722 tree lab_false
, lab_over
, addr
;
8724 f_gpr
= TYPE_FIELDS (TREE_TYPE (va_list_type_node
));
8725 f_fpr
= TREE_CHAIN (f_gpr
);
8726 f_ovf
= TREE_CHAIN (f_fpr
);
8727 f_sav
= TREE_CHAIN (f_ovf
);
8729 valist
= build_va_arg_indirect_ref (valist
);
8730 gpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_gpr
), valist
, f_gpr
, NULL_TREE
);
8731 fpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_fpr
), valist
, f_fpr
, NULL_TREE
);
8732 sav
= build3 (COMPONENT_REF
, TREE_TYPE (f_sav
), valist
, f_sav
, NULL_TREE
);
8734 /* The tree for args* cannot be shared between gpr/fpr and ovf since
8735 both appear on a lhs. */
8736 valist
= unshare_expr (valist
);
8737 ovf
= build3 (COMPONENT_REF
, TREE_TYPE (f_ovf
), valist
, f_ovf
, NULL_TREE
);
8739 size
= int_size_in_bytes (type
);
8741 if (pass_by_reference (NULL
, TYPE_MODE (type
), type
, false))
8743 if (TARGET_DEBUG_ARG
)
8745 fprintf (stderr
, "va_arg: aggregate type");
8749 /* Aggregates are passed by reference. */
8754 /* kernel stack layout on 31 bit: It is assumed here that no padding
8755 will be added by s390_frame_info because for va_args always an even
8756 number of gprs has to be saved r15-r2 = 14 regs. */
8757 sav_ofs
= 2 * UNITS_PER_LONG
;
8758 sav_scale
= UNITS_PER_LONG
;
8759 size
= UNITS_PER_LONG
;
8760 max_reg
= GP_ARG_NUM_REG
- n_reg
;
8762 else if (s390_function_arg_float (TYPE_MODE (type
), type
))
8764 if (TARGET_DEBUG_ARG
)
8766 fprintf (stderr
, "va_arg: float type");
8770 /* FP args go in FP registers, if present. */
8774 sav_ofs
= 16 * UNITS_PER_LONG
;
8776 max_reg
= FP_ARG_NUM_REG
- n_reg
;
8780 if (TARGET_DEBUG_ARG
)
8782 fprintf (stderr
, "va_arg: other type");
8786 /* Otherwise into GP registers. */
8789 n_reg
= (size
+ UNITS_PER_LONG
- 1) / UNITS_PER_LONG
;
8791 /* kernel stack layout on 31 bit: It is assumed here that no padding
8792 will be added by s390_frame_info because for va_args always an even
8793 number of gprs has to be saved r15-r2 = 14 regs. */
8794 sav_ofs
= 2 * UNITS_PER_LONG
;
8796 if (size
< UNITS_PER_LONG
)
8797 sav_ofs
+= UNITS_PER_LONG
- size
;
8799 sav_scale
= UNITS_PER_LONG
;
8800 max_reg
= GP_ARG_NUM_REG
- n_reg
;
8803 /* Pull the value out of the saved registers ... */
8805 lab_false
= create_artificial_label (UNKNOWN_LOCATION
);
8806 lab_over
= create_artificial_label (UNKNOWN_LOCATION
);
8807 addr
= create_tmp_var (ptr_type_node
, "addr");
8809 t
= fold_convert (TREE_TYPE (reg
), size_int (max_reg
));
8810 t
= build2 (GT_EXPR
, boolean_type_node
, reg
, t
);
8811 u
= build1 (GOTO_EXPR
, void_type_node
, lab_false
);
8812 t
= build3 (COND_EXPR
, void_type_node
, t
, u
, NULL_TREE
);
8813 gimplify_and_add (t
, pre_p
);
8815 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, sav
,
8816 size_int (sav_ofs
));
8817 u
= build2 (MULT_EXPR
, TREE_TYPE (reg
), reg
,
8818 fold_convert (TREE_TYPE (reg
), size_int (sav_scale
)));
8819 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
, fold_convert (sizetype
, u
));
8821 gimplify_assign (addr
, t
, pre_p
);
8823 gimple_seq_add_stmt (pre_p
, gimple_build_goto (lab_over
));
8825 gimple_seq_add_stmt (pre_p
, gimple_build_label (lab_false
));
8828 /* ... Otherwise out of the overflow area. */
8831 if (size
< UNITS_PER_LONG
)
8832 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
,
8833 size_int (UNITS_PER_LONG
- size
));
8835 gimplify_expr (&t
, pre_p
, NULL
, is_gimple_val
, fb_rvalue
);
8837 gimplify_assign (addr
, t
, pre_p
);
8839 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
,
8841 gimplify_assign (ovf
, t
, pre_p
);
8843 gimple_seq_add_stmt (pre_p
, gimple_build_label (lab_over
));
8846 /* Increment register save count. */
8848 u
= build2 (PREINCREMENT_EXPR
, TREE_TYPE (reg
), reg
,
8849 fold_convert (TREE_TYPE (reg
), size_int (n_reg
)));
8850 gimplify_and_add (u
, pre_p
);
8854 t
= build_pointer_type_for_mode (build_pointer_type (type
),
8856 addr
= fold_convert (t
, addr
);
8857 addr
= build_va_arg_indirect_ref (addr
);
8861 t
= build_pointer_type_for_mode (type
, ptr_mode
, true);
8862 addr
= fold_convert (t
, addr
);
8865 return build_va_arg_indirect_ref (addr
);
8873 S390_BUILTIN_THREAD_POINTER
,
8874 S390_BUILTIN_SET_THREAD_POINTER
,
8879 static enum insn_code
const code_for_builtin_64
[S390_BUILTIN_max
] = {
8884 static enum insn_code
const code_for_builtin_31
[S390_BUILTIN_max
] = {
8890 s390_init_builtins (void)
8894 ftype
= build_function_type (ptr_type_node
, void_list_node
);
8895 add_builtin_function ("__builtin_thread_pointer", ftype
,
8896 S390_BUILTIN_THREAD_POINTER
, BUILT_IN_MD
,
8899 ftype
= build_function_type_list (void_type_node
, ptr_type_node
, NULL_TREE
);
8900 add_builtin_function ("__builtin_set_thread_pointer", ftype
,
8901 S390_BUILTIN_SET_THREAD_POINTER
, BUILT_IN_MD
,
8905 /* Expand an expression EXP that calls a built-in function,
8906 with result going to TARGET if that's convenient
8907 (and in mode MODE if that's convenient).
8908 SUBTARGET may be used as the target for computing one of EXP's operands.
8909 IGNORE is nonzero if the value is to be ignored. */
8912 s390_expand_builtin (tree exp
, rtx target
, rtx subtarget ATTRIBUTE_UNUSED
,
8913 enum machine_mode mode ATTRIBUTE_UNUSED
,
8914 int ignore ATTRIBUTE_UNUSED
)
8918 enum insn_code
const *code_for_builtin
=
8919 TARGET_64BIT
? code_for_builtin_64
: code_for_builtin_31
;
8921 tree fndecl
= TREE_OPERAND (CALL_EXPR_FN (exp
), 0);
8922 unsigned int fcode
= DECL_FUNCTION_CODE (fndecl
);
8923 enum insn_code icode
;
8924 rtx op
[MAX_ARGS
], pat
;
8928 call_expr_arg_iterator iter
;
8930 if (fcode
>= S390_BUILTIN_max
)
8931 internal_error ("bad builtin fcode");
8932 icode
= code_for_builtin
[fcode
];
8934 internal_error ("bad builtin fcode");
8936 nonvoid
= TREE_TYPE (TREE_TYPE (fndecl
)) != void_type_node
;
8939 FOR_EACH_CALL_EXPR_ARG (arg
, iter
, exp
)
8941 const struct insn_operand_data
*insn_op
;
8943 if (arg
== error_mark_node
)
8945 if (arity
> MAX_ARGS
)
8948 insn_op
= &insn_data
[icode
].operand
[arity
+ nonvoid
];
8950 op
[arity
] = expand_expr (arg
, NULL_RTX
, insn_op
->mode
, EXPAND_NORMAL
);
8952 if (!(*insn_op
->predicate
) (op
[arity
], insn_op
->mode
))
8953 op
[arity
] = copy_to_mode_reg (insn_op
->mode
, op
[arity
]);
8959 enum machine_mode tmode
= insn_data
[icode
].operand
[0].mode
;
8961 || GET_MODE (target
) != tmode
8962 || !(*insn_data
[icode
].operand
[0].predicate
) (target
, tmode
))
8963 target
= gen_reg_rtx (tmode
);
8969 pat
= GEN_FCN (icode
) (target
);
8973 pat
= GEN_FCN (icode
) (target
, op
[0]);
8975 pat
= GEN_FCN (icode
) (op
[0]);
8978 pat
= GEN_FCN (icode
) (target
, op
[0], op
[1]);
8994 /* Output assembly code for the trampoline template to
8997 On S/390, we use gpr 1 internally in the trampoline code;
8998 gpr 0 is used to hold the static chain. */
9001 s390_asm_trampoline_template (FILE *file
)
9004 op
[0] = gen_rtx_REG (Pmode
, 0);
9005 op
[1] = gen_rtx_REG (Pmode
, 1);
9009 output_asm_insn ("basr\t%1,0", op
);
9010 output_asm_insn ("lmg\t%0,%1,14(%1)", op
);
9011 output_asm_insn ("br\t%1", op
);
9012 ASM_OUTPUT_SKIP (file
, (HOST_WIDE_INT
)(TRAMPOLINE_SIZE
- 10));
9016 output_asm_insn ("basr\t%1,0", op
);
9017 output_asm_insn ("lm\t%0,%1,6(%1)", op
);
9018 output_asm_insn ("br\t%1", op
);
9019 ASM_OUTPUT_SKIP (file
, (HOST_WIDE_INT
)(TRAMPOLINE_SIZE
- 8));
9023 /* Emit RTL insns to initialize the variable parts of a trampoline.
9024 FNADDR is an RTX for the address of the function's pure code.
9025 CXT is an RTX for the static chain value for the function. */
9028 s390_trampoline_init (rtx m_tramp
, tree fndecl
, rtx cxt
)
9030 rtx fnaddr
= XEXP (DECL_RTL (fndecl
), 0);
9033 emit_block_move (m_tramp
, assemble_trampoline_template (),
9034 GEN_INT (2*UNITS_PER_WORD
), BLOCK_OP_NORMAL
);
9036 mem
= adjust_address (m_tramp
, Pmode
, 2*UNITS_PER_WORD
);
9037 emit_move_insn (mem
, cxt
);
9038 mem
= adjust_address (m_tramp
, Pmode
, 3*UNITS_PER_WORD
);
9039 emit_move_insn (mem
, fnaddr
);
9042 /* Output assembler code to FILE to increment profiler label # LABELNO
9043 for profiling a function entry. */
9046 s390_function_profiler (FILE *file
, int labelno
)
9051 ASM_GENERATE_INTERNAL_LABEL (label
, "LP", labelno
);
9053 fprintf (file
, "# function profiler \n");
9055 op
[0] = gen_rtx_REG (Pmode
, RETURN_REGNUM
);
9056 op
[1] = gen_rtx_REG (Pmode
, STACK_POINTER_REGNUM
);
9057 op
[1] = gen_rtx_MEM (Pmode
, plus_constant (op
[1], UNITS_PER_LONG
));
9059 op
[2] = gen_rtx_REG (Pmode
, 1);
9060 op
[3] = gen_rtx_SYMBOL_REF (Pmode
, label
);
9061 SYMBOL_REF_FLAGS (op
[3]) = SYMBOL_FLAG_LOCAL
;
9063 op
[4] = gen_rtx_SYMBOL_REF (Pmode
, "_mcount");
9066 op
[4] = gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, op
[4]), UNSPEC_PLT
);
9067 op
[4] = gen_rtx_CONST (Pmode
, op
[4]);
9072 output_asm_insn ("stg\t%0,%1", op
);
9073 output_asm_insn ("larl\t%2,%3", op
);
9074 output_asm_insn ("brasl\t%0,%4", op
);
9075 output_asm_insn ("lg\t%0,%1", op
);
9079 op
[6] = gen_label_rtx ();
9081 output_asm_insn ("st\t%0,%1", op
);
9082 output_asm_insn ("bras\t%2,%l6", op
);
9083 output_asm_insn (".long\t%4", op
);
9084 output_asm_insn (".long\t%3", op
);
9085 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[6]));
9086 output_asm_insn ("l\t%0,0(%2)", op
);
9087 output_asm_insn ("l\t%2,4(%2)", op
);
9088 output_asm_insn ("basr\t%0,%0", op
);
9089 output_asm_insn ("l\t%0,%1", op
);
9093 op
[5] = gen_label_rtx ();
9094 op
[6] = gen_label_rtx ();
9096 output_asm_insn ("st\t%0,%1", op
);
9097 output_asm_insn ("bras\t%2,%l6", op
);
9098 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[5]));
9099 output_asm_insn (".long\t%4-%l5", op
);
9100 output_asm_insn (".long\t%3-%l5", op
);
9101 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[6]));
9102 output_asm_insn ("lr\t%0,%2", op
);
9103 output_asm_insn ("a\t%0,0(%2)", op
);
9104 output_asm_insn ("a\t%2,4(%2)", op
);
9105 output_asm_insn ("basr\t%0,%0", op
);
9106 output_asm_insn ("l\t%0,%1", op
);
9110 /* Encode symbol attributes (local vs. global, tls model) of a SYMBOL_REF
9111 into its SYMBOL_REF_FLAGS. */
9114 s390_encode_section_info (tree decl
, rtx rtl
, int first
)
9116 default_encode_section_info (decl
, rtl
, first
);
9118 if (TREE_CODE (decl
) == VAR_DECL
)
9120 /* If a variable has a forced alignment to < 2 bytes, mark it
9121 with SYMBOL_FLAG_ALIGN1 to prevent it from being used as LARL
9123 if (DECL_USER_ALIGN (decl
) && DECL_ALIGN (decl
) < 16)
9124 SYMBOL_REF_FLAGS (XEXP (rtl
, 0)) |= SYMBOL_FLAG_ALIGN1
;
9125 if (!DECL_SIZE (decl
)
9126 || !DECL_ALIGN (decl
)
9127 || !host_integerp (DECL_SIZE (decl
), 0)
9128 || (DECL_ALIGN (decl
) <= 64
9129 && DECL_ALIGN (decl
) != tree_low_cst (DECL_SIZE (decl
), 0)))
9130 SYMBOL_REF_FLAGS (XEXP (rtl
, 0)) |= SYMBOL_FLAG_NOT_NATURALLY_ALIGNED
;
9133 /* Literal pool references don't have a decl so they are handled
9134 differently here. We rely on the information in the MEM_ALIGN
9135 entry to decide upon natural alignment. */
9137 && GET_CODE (XEXP (rtl
, 0)) == SYMBOL_REF
9138 && TREE_CONSTANT_POOL_ADDRESS_P (XEXP (rtl
, 0))
9139 && (MEM_ALIGN (rtl
) == 0
9140 || GET_MODE_BITSIZE (GET_MODE (rtl
)) == 0
9141 || MEM_ALIGN (rtl
) < GET_MODE_BITSIZE (GET_MODE (rtl
))))
9142 SYMBOL_REF_FLAGS (XEXP (rtl
, 0)) |= SYMBOL_FLAG_NOT_NATURALLY_ALIGNED
;
9145 /* Output thunk to FILE that implements a C++ virtual function call (with
9146 multiple inheritance) to FUNCTION. The thunk adjusts the this pointer
9147 by DELTA, and unless VCALL_OFFSET is zero, applies an additional adjustment
9148 stored at VCALL_OFFSET in the vtable whose address is located at offset 0
9149 relative to the resulting this pointer. */
9152 s390_output_mi_thunk (FILE *file
, tree thunk ATTRIBUTE_UNUSED
,
9153 HOST_WIDE_INT delta
, HOST_WIDE_INT vcall_offset
,
9159 /* Make sure unwind info is emitted for the thunk if needed. */
9160 final_start_function (emit_barrier (), file
, 1);
9162 /* Operand 0 is the target function. */
9163 op
[0] = XEXP (DECL_RTL (function
), 0);
9164 if (flag_pic
&& !SYMBOL_REF_LOCAL_P (op
[0]))
9167 op
[0] = gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, op
[0]),
9168 TARGET_64BIT
? UNSPEC_PLT
: UNSPEC_GOT
);
9169 op
[0] = gen_rtx_CONST (Pmode
, op
[0]);
9172 /* Operand 1 is the 'this' pointer. */
9173 if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function
)), function
))
9174 op
[1] = gen_rtx_REG (Pmode
, 3);
9176 op
[1] = gen_rtx_REG (Pmode
, 2);
9178 /* Operand 2 is the delta. */
9179 op
[2] = GEN_INT (delta
);
9181 /* Operand 3 is the vcall_offset. */
9182 op
[3] = GEN_INT (vcall_offset
);
9184 /* Operand 4 is the temporary register. */
9185 op
[4] = gen_rtx_REG (Pmode
, 1);
9187 /* Operands 5 to 8 can be used as labels. */
9193 /* Operand 9 can be used for temporary register. */
9196 /* Generate code. */
9199 /* Setup literal pool pointer if required. */
9200 if ((!DISP_IN_RANGE (delta
)
9201 && !CONST_OK_FOR_K (delta
)
9202 && !CONST_OK_FOR_Os (delta
))
9203 || (!DISP_IN_RANGE (vcall_offset
)
9204 && !CONST_OK_FOR_K (vcall_offset
)
9205 && !CONST_OK_FOR_Os (vcall_offset
)))
9207 op
[5] = gen_label_rtx ();
9208 output_asm_insn ("larl\t%4,%5", op
);
9211 /* Add DELTA to this pointer. */
9214 if (CONST_OK_FOR_J (delta
))
9215 output_asm_insn ("la\t%1,%2(%1)", op
);
9216 else if (DISP_IN_RANGE (delta
))
9217 output_asm_insn ("lay\t%1,%2(%1)", op
);
9218 else if (CONST_OK_FOR_K (delta
))
9219 output_asm_insn ("aghi\t%1,%2", op
);
9220 else if (CONST_OK_FOR_Os (delta
))
9221 output_asm_insn ("agfi\t%1,%2", op
);
9224 op
[6] = gen_label_rtx ();
9225 output_asm_insn ("agf\t%1,%6-%5(%4)", op
);
9229 /* Perform vcall adjustment. */
9232 if (DISP_IN_RANGE (vcall_offset
))
9234 output_asm_insn ("lg\t%4,0(%1)", op
);
9235 output_asm_insn ("ag\t%1,%3(%4)", op
);
9237 else if (CONST_OK_FOR_K (vcall_offset
))
9239 output_asm_insn ("lghi\t%4,%3", op
);
9240 output_asm_insn ("ag\t%4,0(%1)", op
);
9241 output_asm_insn ("ag\t%1,0(%4)", op
);
9243 else if (CONST_OK_FOR_Os (vcall_offset
))
9245 output_asm_insn ("lgfi\t%4,%3", op
);
9246 output_asm_insn ("ag\t%4,0(%1)", op
);
9247 output_asm_insn ("ag\t%1,0(%4)", op
);
9251 op
[7] = gen_label_rtx ();
9252 output_asm_insn ("llgf\t%4,%7-%5(%4)", op
);
9253 output_asm_insn ("ag\t%4,0(%1)", op
);
9254 output_asm_insn ("ag\t%1,0(%4)", op
);
9258 /* Jump to target. */
9259 output_asm_insn ("jg\t%0", op
);
9261 /* Output literal pool if required. */
9264 output_asm_insn (".align\t4", op
);
9265 targetm
.asm_out
.internal_label (file
, "L",
9266 CODE_LABEL_NUMBER (op
[5]));
9270 targetm
.asm_out
.internal_label (file
, "L",
9271 CODE_LABEL_NUMBER (op
[6]));
9272 output_asm_insn (".long\t%2", op
);
9276 targetm
.asm_out
.internal_label (file
, "L",
9277 CODE_LABEL_NUMBER (op
[7]));
9278 output_asm_insn (".long\t%3", op
);
9283 /* Setup base pointer if required. */
9285 || (!DISP_IN_RANGE (delta
)
9286 && !CONST_OK_FOR_K (delta
)
9287 && !CONST_OK_FOR_Os (delta
))
9288 || (!DISP_IN_RANGE (delta
)
9289 && !CONST_OK_FOR_K (vcall_offset
)
9290 && !CONST_OK_FOR_Os (vcall_offset
)))
9292 op
[5] = gen_label_rtx ();
9293 output_asm_insn ("basr\t%4,0", op
);
9294 targetm
.asm_out
.internal_label (file
, "L",
9295 CODE_LABEL_NUMBER (op
[5]));
9298 /* Add DELTA to this pointer. */
9301 if (CONST_OK_FOR_J (delta
))
9302 output_asm_insn ("la\t%1,%2(%1)", op
);
9303 else if (DISP_IN_RANGE (delta
))
9304 output_asm_insn ("lay\t%1,%2(%1)", op
);
9305 else if (CONST_OK_FOR_K (delta
))
9306 output_asm_insn ("ahi\t%1,%2", op
);
9307 else if (CONST_OK_FOR_Os (delta
))
9308 output_asm_insn ("afi\t%1,%2", op
);
9311 op
[6] = gen_label_rtx ();
9312 output_asm_insn ("a\t%1,%6-%5(%4)", op
);
9316 /* Perform vcall adjustment. */
9319 if (CONST_OK_FOR_J (vcall_offset
))
9321 output_asm_insn ("l\t%4,0(%1)", op
);
9322 output_asm_insn ("a\t%1,%3(%4)", op
);
9324 else if (DISP_IN_RANGE (vcall_offset
))
9326 output_asm_insn ("l\t%4,0(%1)", op
);
9327 output_asm_insn ("ay\t%1,%3(%4)", op
);
9329 else if (CONST_OK_FOR_K (vcall_offset
))
9331 output_asm_insn ("lhi\t%4,%3", op
);
9332 output_asm_insn ("a\t%4,0(%1)", op
);
9333 output_asm_insn ("a\t%1,0(%4)", op
);
9335 else if (CONST_OK_FOR_Os (vcall_offset
))
9337 output_asm_insn ("iilf\t%4,%3", op
);
9338 output_asm_insn ("a\t%4,0(%1)", op
);
9339 output_asm_insn ("a\t%1,0(%4)", op
);
9343 op
[7] = gen_label_rtx ();
9344 output_asm_insn ("l\t%4,%7-%5(%4)", op
);
9345 output_asm_insn ("a\t%4,0(%1)", op
);
9346 output_asm_insn ("a\t%1,0(%4)", op
);
9349 /* We had to clobber the base pointer register.
9350 Re-setup the base pointer (with a different base). */
9351 op
[5] = gen_label_rtx ();
9352 output_asm_insn ("basr\t%4,0", op
);
9353 targetm
.asm_out
.internal_label (file
, "L",
9354 CODE_LABEL_NUMBER (op
[5]));
9357 /* Jump to target. */
9358 op
[8] = gen_label_rtx ();
9361 output_asm_insn ("l\t%4,%8-%5(%4)", op
);
9363 output_asm_insn ("a\t%4,%8-%5(%4)", op
);
9364 /* We cannot call through .plt, since .plt requires %r12 loaded. */
9365 else if (flag_pic
== 1)
9367 output_asm_insn ("a\t%4,%8-%5(%4)", op
);
9368 output_asm_insn ("l\t%4,%0(%4)", op
);
9370 else if (flag_pic
== 2)
9372 op
[9] = gen_rtx_REG (Pmode
, 0);
9373 output_asm_insn ("l\t%9,%8-4-%5(%4)", op
);
9374 output_asm_insn ("a\t%4,%8-%5(%4)", op
);
9375 output_asm_insn ("ar\t%4,%9", op
);
9376 output_asm_insn ("l\t%4,0(%4)", op
);
9379 output_asm_insn ("br\t%4", op
);
9381 /* Output literal pool. */
9382 output_asm_insn (".align\t4", op
);
9384 if (nonlocal
&& flag_pic
== 2)
9385 output_asm_insn (".long\t%0", op
);
9388 op
[0] = gen_rtx_SYMBOL_REF (Pmode
, "_GLOBAL_OFFSET_TABLE_");
9389 SYMBOL_REF_FLAGS (op
[0]) = SYMBOL_FLAG_LOCAL
;
9392 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[8]));
9394 output_asm_insn (".long\t%0", op
);
9396 output_asm_insn (".long\t%0-%5", op
);
9400 targetm
.asm_out
.internal_label (file
, "L",
9401 CODE_LABEL_NUMBER (op
[6]));
9402 output_asm_insn (".long\t%2", op
);
9406 targetm
.asm_out
.internal_label (file
, "L",
9407 CODE_LABEL_NUMBER (op
[7]));
9408 output_asm_insn (".long\t%3", op
);
9411 final_end_function ();
9415 s390_valid_pointer_mode (enum machine_mode mode
)
9417 return (mode
== SImode
|| (TARGET_64BIT
&& mode
== DImode
));
9420 /* Checks whether the given CALL_EXPR would use a caller
9421 saved register. This is used to decide whether sibling call
9422 optimization could be performed on the respective function
9426 s390_call_saved_register_used (tree call_expr
)
9428 CUMULATIVE_ARGS cum
;
9430 enum machine_mode mode
;
9435 INIT_CUMULATIVE_ARGS (cum
, NULL
, NULL
, 0, 0);
9437 for (i
= 0; i
< call_expr_nargs (call_expr
); i
++)
9439 parameter
= CALL_EXPR_ARG (call_expr
, i
);
9440 gcc_assert (parameter
);
9442 /* For an undeclared variable passed as parameter we will get
9443 an ERROR_MARK node here. */
9444 if (TREE_CODE (parameter
) == ERROR_MARK
)
9447 type
= TREE_TYPE (parameter
);
9450 mode
= TYPE_MODE (type
);
9453 if (pass_by_reference (&cum
, mode
, type
, true))
9456 type
= build_pointer_type (type
);
9459 parm_rtx
= s390_function_arg (&cum
, mode
, type
, 0);
9461 s390_function_arg_advance (&cum
, mode
, type
, 0);
9466 if (REG_P (parm_rtx
))
9469 reg
< HARD_REGNO_NREGS (REGNO (parm_rtx
), GET_MODE (parm_rtx
));
9471 if (!call_used_regs
[reg
+ REGNO (parm_rtx
)])
9475 if (GET_CODE (parm_rtx
) == PARALLEL
)
9479 for (i
= 0; i
< XVECLEN (parm_rtx
, 0); i
++)
9481 rtx r
= XEXP (XVECEXP (parm_rtx
, 0, i
), 0);
9483 gcc_assert (REG_P (r
));
9486 reg
< HARD_REGNO_NREGS (REGNO (r
), GET_MODE (r
));
9488 if (!call_used_regs
[reg
+ REGNO (r
)])
9497 /* Return true if the given call expression can be
9498 turned into a sibling call.
9499 DECL holds the declaration of the function to be called whereas
9500 EXP is the call expression itself. */
9503 s390_function_ok_for_sibcall (tree decl
, tree exp
)
9505 /* The TPF epilogue uses register 1. */
9506 if (TARGET_TPF_PROFILING
)
9509 /* The 31 bit PLT code uses register 12 (GOT pointer - caller saved)
9510 which would have to be restored before the sibcall. */
9511 if (!TARGET_64BIT
&& flag_pic
&& decl
&& !targetm
.binds_local_p (decl
))
9514 /* Register 6 on s390 is available as an argument register but unfortunately
9515 "caller saved". This makes functions needing this register for arguments
9516 not suitable for sibcalls. */
9517 return !s390_call_saved_register_used (exp
);
9520 /* Return the fixed registers used for condition codes. */
9523 s390_fixed_condition_code_regs (unsigned int *p1
, unsigned int *p2
)
9526 *p2
= INVALID_REGNUM
;
9531 /* This function is used by the call expanders of the machine description.
9532 It emits the call insn itself together with the necessary operations
9533 to adjust the target address and returns the emitted insn.
9534 ADDR_LOCATION is the target address rtx
9535 TLS_CALL the location of the thread-local symbol
9536 RESULT_REG the register where the result of the call should be stored
9537 RETADDR_REG the register where the return address should be stored
9538 If this parameter is NULL_RTX the call is considered
9539 to be a sibling call. */
9542 s390_emit_call (rtx addr_location
, rtx tls_call
, rtx result_reg
,
9545 bool plt_call
= false;
9551 /* Direct function calls need special treatment. */
9552 if (GET_CODE (addr_location
) == SYMBOL_REF
)
9554 /* When calling a global routine in PIC mode, we must
9555 replace the symbol itself with the PLT stub. */
9556 if (flag_pic
&& !SYMBOL_REF_LOCAL_P (addr_location
))
9558 if (retaddr_reg
!= NULL_RTX
)
9560 addr_location
= gen_rtx_UNSPEC (Pmode
,
9561 gen_rtvec (1, addr_location
),
9563 addr_location
= gen_rtx_CONST (Pmode
, addr_location
);
9567 /* For -fpic code the PLT entries might use r12 which is
9568 call-saved. Therefore we cannot do a sibcall when
9569 calling directly using a symbol ref. When reaching
9570 this point we decided (in s390_function_ok_for_sibcall)
9571 to do a sibcall for a function pointer but one of the
9572 optimizers was able to get rid of the function pointer
9573 by propagating the symbol ref into the call. This
9574 optimization is illegal for S/390 so we turn the direct
9575 call into a indirect call again. */
9576 addr_location
= force_reg (Pmode
, addr_location
);
9579 /* Unless we can use the bras(l) insn, force the
9580 routine address into a register. */
9581 if (!TARGET_SMALL_EXEC
&& !TARGET_CPU_ZARCH
)
9584 addr_location
= legitimize_pic_address (addr_location
, 0);
9586 addr_location
= force_reg (Pmode
, addr_location
);
9590 /* If it is already an indirect call or the code above moved the
9591 SYMBOL_REF to somewhere else make sure the address can be found in
9593 if (retaddr_reg
== NULL_RTX
9594 && GET_CODE (addr_location
) != SYMBOL_REF
9597 emit_move_insn (gen_rtx_REG (Pmode
, SIBCALL_REGNUM
), addr_location
);
9598 addr_location
= gen_rtx_REG (Pmode
, SIBCALL_REGNUM
);
9601 addr_location
= gen_rtx_MEM (QImode
, addr_location
);
9602 call
= gen_rtx_CALL (VOIDmode
, addr_location
, const0_rtx
);
9604 if (result_reg
!= NULL_RTX
)
9605 call
= gen_rtx_SET (VOIDmode
, result_reg
, call
);
9607 if (retaddr_reg
!= NULL_RTX
)
9609 clobber
= gen_rtx_CLOBBER (VOIDmode
, retaddr_reg
);
9611 if (tls_call
!= NULL_RTX
)
9612 vec
= gen_rtvec (3, call
, clobber
,
9613 gen_rtx_USE (VOIDmode
, tls_call
));
9615 vec
= gen_rtvec (2, call
, clobber
);
9617 call
= gen_rtx_PARALLEL (VOIDmode
, vec
);
9620 insn
= emit_call_insn (call
);
9622 /* 31-bit PLT stubs and tls calls use the GOT register implicitly. */
9623 if ((!TARGET_64BIT
&& plt_call
) || tls_call
!= NULL_RTX
)
9625 /* s390_function_ok_for_sibcall should
9626 have denied sibcalls in this case. */
9627 gcc_assert (retaddr_reg
!= NULL_RTX
);
9629 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), pic_offset_table_rtx
);
9634 /* Implement CONDITIONAL_REGISTER_USAGE. */
9637 s390_conditional_register_usage (void)
9643 fixed_regs
[PIC_OFFSET_TABLE_REGNUM
] = 1;
9644 call_used_regs
[PIC_OFFSET_TABLE_REGNUM
] = 1;
9646 if (TARGET_CPU_ZARCH
)
9648 fixed_regs
[BASE_REGNUM
] = 0;
9649 call_used_regs
[BASE_REGNUM
] = 0;
9650 fixed_regs
[RETURN_REGNUM
] = 0;
9651 call_used_regs
[RETURN_REGNUM
] = 0;
9655 for (i
= 24; i
< 32; i
++)
9656 call_used_regs
[i
] = call_really_used_regs
[i
] = 0;
9660 for (i
= 18; i
< 20; i
++)
9661 call_used_regs
[i
] = call_really_used_regs
[i
] = 0;
9664 if (TARGET_SOFT_FLOAT
)
9666 for (i
= 16; i
< 32; i
++)
9667 call_used_regs
[i
] = fixed_regs
[i
] = 1;
9671 /* Corresponding function to eh_return expander. */
9673 static GTY(()) rtx s390_tpf_eh_return_symbol
;
9675 s390_emit_tpf_eh_return (rtx target
)
9679 if (!s390_tpf_eh_return_symbol
)
9680 s390_tpf_eh_return_symbol
= gen_rtx_SYMBOL_REF (Pmode
, "__tpf_eh_return");
9682 reg
= gen_rtx_REG (Pmode
, 2);
9684 emit_move_insn (reg
, target
);
9685 insn
= s390_emit_call (s390_tpf_eh_return_symbol
, NULL_RTX
, reg
,
9686 gen_rtx_REG (Pmode
, RETURN_REGNUM
));
9687 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), reg
);
9689 emit_move_insn (EH_RETURN_HANDLER_RTX
, reg
);
9692 /* Rework the prologue/epilogue to avoid saving/restoring
9693 registers unnecessarily. */
9696 s390_optimize_prologue (void)
9698 rtx insn
, new_insn
, next_insn
;
9700 /* Do a final recompute of the frame-related data. */
9702 s390_update_frame_layout ();
9704 /* If all special registers are in fact used, there's nothing we
9705 can do, so no point in walking the insn list. */
9707 if (cfun_frame_layout
.first_save_gpr
<= BASE_REGNUM
9708 && cfun_frame_layout
.last_save_gpr
>= BASE_REGNUM
9709 && (TARGET_CPU_ZARCH
9710 || (cfun_frame_layout
.first_save_gpr
<= RETURN_REGNUM
9711 && cfun_frame_layout
.last_save_gpr
>= RETURN_REGNUM
)))
9714 /* Search for prologue/epilogue insns and replace them. */
9716 for (insn
= get_insns (); insn
; insn
= next_insn
)
9718 int first
, last
, off
;
9719 rtx set
, base
, offset
;
9721 next_insn
= NEXT_INSN (insn
);
9723 if (GET_CODE (insn
) != INSN
)
9726 if (GET_CODE (PATTERN (insn
)) == PARALLEL
9727 && store_multiple_operation (PATTERN (insn
), VOIDmode
))
9729 set
= XVECEXP (PATTERN (insn
), 0, 0);
9730 first
= REGNO (SET_SRC (set
));
9731 last
= first
+ XVECLEN (PATTERN (insn
), 0) - 1;
9732 offset
= const0_rtx
;
9733 base
= eliminate_constant_term (XEXP (SET_DEST (set
), 0), &offset
);
9734 off
= INTVAL (offset
);
9736 if (GET_CODE (base
) != REG
|| off
< 0)
9738 if (cfun_frame_layout
.first_save_gpr
!= -1
9739 && (cfun_frame_layout
.first_save_gpr
< first
9740 || cfun_frame_layout
.last_save_gpr
> last
))
9742 if (REGNO (base
) != STACK_POINTER_REGNUM
9743 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9745 if (first
> BASE_REGNUM
|| last
< BASE_REGNUM
)
9748 if (cfun_frame_layout
.first_save_gpr
!= -1)
9750 new_insn
= save_gprs (base
,
9751 off
+ (cfun_frame_layout
.first_save_gpr
9752 - first
) * UNITS_PER_LONG
,
9753 cfun_frame_layout
.first_save_gpr
,
9754 cfun_frame_layout
.last_save_gpr
);
9755 new_insn
= emit_insn_before (new_insn
, insn
);
9756 INSN_ADDRESSES_NEW (new_insn
, -1);
9763 if (cfun_frame_layout
.first_save_gpr
== -1
9764 && GET_CODE (PATTERN (insn
)) == SET
9765 && GET_CODE (SET_SRC (PATTERN (insn
))) == REG
9766 && (REGNO (SET_SRC (PATTERN (insn
))) == BASE_REGNUM
9767 || (!TARGET_CPU_ZARCH
9768 && REGNO (SET_SRC (PATTERN (insn
))) == RETURN_REGNUM
))
9769 && GET_CODE (SET_DEST (PATTERN (insn
))) == MEM
)
9771 set
= PATTERN (insn
);
9772 first
= REGNO (SET_SRC (set
));
9773 offset
= const0_rtx
;
9774 base
= eliminate_constant_term (XEXP (SET_DEST (set
), 0), &offset
);
9775 off
= INTVAL (offset
);
9777 if (GET_CODE (base
) != REG
|| off
< 0)
9779 if (REGNO (base
) != STACK_POINTER_REGNUM
9780 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9787 if (GET_CODE (PATTERN (insn
)) == PARALLEL
9788 && load_multiple_operation (PATTERN (insn
), VOIDmode
))
9790 set
= XVECEXP (PATTERN (insn
), 0, 0);
9791 first
= REGNO (SET_DEST (set
));
9792 last
= first
+ XVECLEN (PATTERN (insn
), 0) - 1;
9793 offset
= const0_rtx
;
9794 base
= eliminate_constant_term (XEXP (SET_SRC (set
), 0), &offset
);
9795 off
= INTVAL (offset
);
9797 if (GET_CODE (base
) != REG
|| off
< 0)
9799 if (cfun_frame_layout
.first_restore_gpr
!= -1
9800 && (cfun_frame_layout
.first_restore_gpr
< first
9801 || cfun_frame_layout
.last_restore_gpr
> last
))
9803 if (REGNO (base
) != STACK_POINTER_REGNUM
9804 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9806 if (first
> BASE_REGNUM
|| last
< BASE_REGNUM
)
9809 if (cfun_frame_layout
.first_restore_gpr
!= -1)
9811 new_insn
= restore_gprs (base
,
9812 off
+ (cfun_frame_layout
.first_restore_gpr
9813 - first
) * UNITS_PER_LONG
,
9814 cfun_frame_layout
.first_restore_gpr
,
9815 cfun_frame_layout
.last_restore_gpr
);
9816 new_insn
= emit_insn_before (new_insn
, insn
);
9817 INSN_ADDRESSES_NEW (new_insn
, -1);
9824 if (cfun_frame_layout
.first_restore_gpr
== -1
9825 && GET_CODE (PATTERN (insn
)) == SET
9826 && GET_CODE (SET_DEST (PATTERN (insn
))) == REG
9827 && (REGNO (SET_DEST (PATTERN (insn
))) == BASE_REGNUM
9828 || (!TARGET_CPU_ZARCH
9829 && REGNO (SET_DEST (PATTERN (insn
))) == RETURN_REGNUM
))
9830 && GET_CODE (SET_SRC (PATTERN (insn
))) == MEM
)
9832 set
= PATTERN (insn
);
9833 first
= REGNO (SET_DEST (set
));
9834 offset
= const0_rtx
;
9835 base
= eliminate_constant_term (XEXP (SET_SRC (set
), 0), &offset
);
9836 off
= INTVAL (offset
);
9838 if (GET_CODE (base
) != REG
|| off
< 0)
9840 if (REGNO (base
) != STACK_POINTER_REGNUM
9841 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9850 /* On z10 the dynamic branch prediction must see the backward jump in
9851 a window of 384 bytes. If not it falls back to the static
9852 prediction. This function rearranges the loop backward branch in a
9853 way which makes the static prediction always correct. The function
9854 returns true if it added an instruction. */
9856 s390_z10_fix_long_loop_prediction (rtx insn
)
9858 rtx set
= single_set (insn
);
9859 rtx code_label
, label_ref
, new_label
;
9865 /* This will exclude branch on count and branch on index patterns
9866 since these are correctly statically predicted. */
9868 || SET_DEST (set
) != pc_rtx
9869 || GET_CODE (SET_SRC(set
)) != IF_THEN_ELSE
)
9872 label_ref
= (GET_CODE (XEXP (SET_SRC (set
), 1)) == LABEL_REF
?
9873 XEXP (SET_SRC (set
), 1) : XEXP (SET_SRC (set
), 2));
9875 gcc_assert (GET_CODE (label_ref
) == LABEL_REF
);
9877 code_label
= XEXP (label_ref
, 0);
9879 if (INSN_ADDRESSES (INSN_UID (code_label
)) == -1
9880 || INSN_ADDRESSES (INSN_UID (insn
)) == -1
9881 || (INSN_ADDRESSES (INSN_UID (insn
))
9882 - INSN_ADDRESSES (INSN_UID (code_label
)) < Z10_PREDICT_DISTANCE
))
9885 for (distance
= 0, cur_insn
= PREV_INSN (insn
);
9886 distance
< Z10_PREDICT_DISTANCE
- 6;
9887 distance
+= get_attr_length (cur_insn
), cur_insn
= PREV_INSN (cur_insn
))
9888 if (!cur_insn
|| JUMP_P (cur_insn
) || LABEL_P (cur_insn
))
9891 new_label
= gen_label_rtx ();
9892 uncond_jump
= emit_jump_insn_after (
9893 gen_rtx_SET (VOIDmode
, pc_rtx
,
9894 gen_rtx_LABEL_REF (VOIDmode
, code_label
)),
9896 emit_label_after (new_label
, uncond_jump
);
9898 tmp
= XEXP (SET_SRC (set
), 1);
9899 XEXP (SET_SRC (set
), 1) = XEXP (SET_SRC (set
), 2);
9900 XEXP (SET_SRC (set
), 2) = tmp
;
9901 INSN_CODE (insn
) = -1;
9903 XEXP (label_ref
, 0) = new_label
;
9904 JUMP_LABEL (insn
) = new_label
;
9905 JUMP_LABEL (uncond_jump
) = code_label
;
9910 /* Returns 1 if INSN reads the value of REG for purposes not related
9911 to addressing of memory, and 0 otherwise. */
9913 s390_non_addr_reg_read_p (rtx reg
, rtx insn
)
9915 return reg_referenced_p (reg
, PATTERN (insn
))
9916 && !reg_used_in_mem_p (REGNO (reg
), PATTERN (insn
));
9919 /* Starting from INSN find_cond_jump looks downwards in the insn
9920 stream for a single jump insn which is the last user of the
9921 condition code set in INSN. */
9923 find_cond_jump (rtx insn
)
9925 for (; insn
; insn
= NEXT_INSN (insn
))
9934 if (reg_mentioned_p (gen_rtx_REG (CCmode
, CC_REGNUM
), insn
))
9939 /* This will be triggered by a return. */
9940 if (GET_CODE (PATTERN (insn
)) != SET
)
9943 gcc_assert (SET_DEST (PATTERN (insn
)) == pc_rtx
);
9944 ite
= SET_SRC (PATTERN (insn
));
9946 if (GET_CODE (ite
) != IF_THEN_ELSE
)
9949 cc
= XEXP (XEXP (ite
, 0), 0);
9950 if (!REG_P (cc
) || !CC_REGNO_P (REGNO (cc
)))
9953 if (find_reg_note (insn
, REG_DEAD
, cc
))
9961 /* Swap the condition in COND and the operands in OP0 and OP1 so that
9962 the semantics does not change. If NULL_RTX is passed as COND the
9963 function tries to find the conditional jump starting with INSN. */
9965 s390_swap_cmp (rtx cond
, rtx
*op0
, rtx
*op1
, rtx insn
)
9969 if (cond
== NULL_RTX
)
9971 rtx jump
= find_cond_jump (NEXT_INSN (insn
));
9972 jump
= jump
? single_set (jump
) : NULL_RTX
;
9974 if (jump
== NULL_RTX
)
9977 cond
= XEXP (XEXP (jump
, 1), 0);
9982 PUT_CODE (cond
, swap_condition (GET_CODE (cond
)));
9985 /* On z10, instructions of the compare-and-branch family have the
9986 property to access the register occurring as second operand with
9987 its bits complemented. If such a compare is grouped with a second
9988 instruction that accesses the same register non-complemented, and
9989 if that register's value is delivered via a bypass, then the
9990 pipeline recycles, thereby causing significant performance decline.
9991 This function locates such situations and exchanges the two
9992 operands of the compare. The function return true whenever it
9995 s390_z10_optimize_cmp (rtx insn
)
9997 rtx prev_insn
, next_insn
;
9998 bool insn_added_p
= false;
9999 rtx cond
, *op0
, *op1
;
10001 if (GET_CODE (PATTERN (insn
)) == PARALLEL
)
10003 /* Handle compare and branch and branch on count
10005 rtx pattern
= single_set (insn
);
10008 || SET_DEST (pattern
) != pc_rtx
10009 || GET_CODE (SET_SRC (pattern
)) != IF_THEN_ELSE
)
10012 cond
= XEXP (SET_SRC (pattern
), 0);
10013 op0
= &XEXP (cond
, 0);
10014 op1
= &XEXP (cond
, 1);
10016 else if (GET_CODE (PATTERN (insn
)) == SET
)
10020 /* Handle normal compare instructions. */
10021 src
= SET_SRC (PATTERN (insn
));
10022 dest
= SET_DEST (PATTERN (insn
));
10025 || !CC_REGNO_P (REGNO (dest
))
10026 || GET_CODE (src
) != COMPARE
)
10029 /* s390_swap_cmp will try to find the conditional
10030 jump when passing NULL_RTX as condition. */
10032 op0
= &XEXP (src
, 0);
10033 op1
= &XEXP (src
, 1);
10038 if (!REG_P (*op0
) || !REG_P (*op1
))
10041 if (GET_MODE_CLASS (GET_MODE (*op0
)) != MODE_INT
)
10044 /* Swap the COMPARE arguments and its mask if there is a
10045 conflicting access in the previous insn. */
10046 prev_insn
= prev_active_insn (insn
);
10047 if (prev_insn
!= NULL_RTX
&& INSN_P (prev_insn
)
10048 && reg_referenced_p (*op1
, PATTERN (prev_insn
)))
10049 s390_swap_cmp (cond
, op0
, op1
, insn
);
10051 /* Check if there is a conflict with the next insn. If there
10052 was no conflict with the previous insn, then swap the
10053 COMPARE arguments and its mask. If we already swapped
10054 the operands, or if swapping them would cause a conflict
10055 with the previous insn, issue a NOP after the COMPARE in
10056 order to separate the two instuctions. */
10057 next_insn
= next_active_insn (insn
);
10058 if (next_insn
!= NULL_RTX
&& INSN_P (next_insn
)
10059 && s390_non_addr_reg_read_p (*op1
, next_insn
))
10061 if (prev_insn
!= NULL_RTX
&& INSN_P (prev_insn
)
10062 && s390_non_addr_reg_read_p (*op0
, prev_insn
))
10064 if (REGNO (*op1
) == 0)
10065 emit_insn_after (gen_nop1 (), insn
);
10067 emit_insn_after (gen_nop (), insn
);
10068 insn_added_p
= true;
10071 s390_swap_cmp (cond
, op0
, op1
, insn
);
10073 return insn_added_p
;
10076 /* Perform machine-dependent processing. */
10081 bool pool_overflow
= false;
10083 /* Make sure all splits have been performed; splits after
10084 machine_dependent_reorg might confuse insn length counts. */
10085 split_all_insns_noflow ();
10087 /* Install the main literal pool and the associated base
10088 register load insns.
10090 In addition, there are two problematic situations we need
10093 - the literal pool might be > 4096 bytes in size, so that
10094 some of its elements cannot be directly accessed
10096 - a branch target might be > 64K away from the branch, so that
10097 it is not possible to use a PC-relative instruction.
10099 To fix those, we split the single literal pool into multiple
10100 pool chunks, reloading the pool base register at various
10101 points throughout the function to ensure it always points to
10102 the pool chunk the following code expects, and / or replace
10103 PC-relative branches by absolute branches.
10105 However, the two problems are interdependent: splitting the
10106 literal pool can move a branch further away from its target,
10107 causing the 64K limit to overflow, and on the other hand,
10108 replacing a PC-relative branch by an absolute branch means
10109 we need to put the branch target address into the literal
10110 pool, possibly causing it to overflow.
10112 So, we loop trying to fix up both problems until we manage
10113 to satisfy both conditions at the same time. Note that the
10114 loop is guaranteed to terminate as every pass of the loop
10115 strictly decreases the total number of PC-relative branches
10116 in the function. (This is not completely true as there
10117 might be branch-over-pool insns introduced by chunkify_start.
10118 Those never need to be split however.) */
10122 struct constant_pool
*pool
= NULL
;
10124 /* Collect the literal pool. */
10125 if (!pool_overflow
)
10127 pool
= s390_mainpool_start ();
10129 pool_overflow
= true;
10132 /* If literal pool overflowed, start to chunkify it. */
10134 pool
= s390_chunkify_start ();
10136 /* Split out-of-range branches. If this has created new
10137 literal pool entries, cancel current chunk list and
10138 recompute it. zSeries machines have large branch
10139 instructions, so we never need to split a branch. */
10140 if (!TARGET_CPU_ZARCH
&& s390_split_branches ())
10143 s390_chunkify_cancel (pool
);
10145 s390_mainpool_cancel (pool
);
10150 /* If we made it up to here, both conditions are satisfied.
10151 Finish up literal pool related changes. */
10153 s390_chunkify_finish (pool
);
10155 s390_mainpool_finish (pool
);
10157 /* We're done splitting branches. */
10158 cfun
->machine
->split_branches_pending_p
= false;
10162 /* Generate out-of-pool execute target insns. */
10163 if (TARGET_CPU_ZARCH
)
10165 rtx insn
, label
, target
;
10167 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
10169 label
= s390_execute_label (insn
);
10173 gcc_assert (label
!= const0_rtx
);
10175 target
= emit_label (XEXP (label
, 0));
10176 INSN_ADDRESSES_NEW (target
, -1);
10178 target
= emit_insn (s390_execute_target (insn
));
10179 INSN_ADDRESSES_NEW (target
, -1);
10183 /* Try to optimize prologue and epilogue further. */
10184 s390_optimize_prologue ();
10186 /* Walk over the insns and do some z10 specific changes. */
10187 if (s390_tune
== PROCESSOR_2097_Z10
)
10190 bool insn_added_p
= false;
10192 /* The insn lengths and addresses have to be up to date for the
10193 following manipulations. */
10194 shorten_branches (get_insns ());
10196 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
10198 if (!INSN_P (insn
) || INSN_CODE (insn
) <= 0)
10202 insn_added_p
|= s390_z10_fix_long_loop_prediction (insn
);
10204 if (GET_CODE (PATTERN (insn
)) == PARALLEL
10205 || GET_CODE (PATTERN (insn
)) == SET
)
10206 insn_added_p
|= s390_z10_optimize_cmp (insn
);
10209 /* Adjust branches if we added new instructions. */
10211 shorten_branches (get_insns ());
10215 /* Return true if INSN is a fp load insn writing register REGNO. */
10217 s390_fpload_toreg (rtx insn
, unsigned int regno
)
10220 enum attr_type flag
= s390_safe_attr_type (insn
);
10222 if (flag
!= TYPE_FLOADSF
&& flag
!= TYPE_FLOADDF
)
10225 set
= single_set (insn
);
10227 if (set
== NULL_RTX
)
10230 if (!REG_P (SET_DEST (set
)) || !MEM_P (SET_SRC (set
)))
10233 if (REGNO (SET_DEST (set
)) != regno
)
10239 /* This value describes the distance to be avoided between an
10240 aritmetic fp instruction and an fp load writing the same register.
10241 Z10_EARLYLOAD_DISTANCE - 1 as well as Z10_EARLYLOAD_DISTANCE + 1 is
10242 fine but the exact value has to be avoided. Otherwise the FP
10243 pipeline will throw an exception causing a major penalty. */
10244 #define Z10_EARLYLOAD_DISTANCE 7
10246 /* Rearrange the ready list in order to avoid the situation described
10247 for Z10_EARLYLOAD_DISTANCE. A problematic load instruction is
10248 moved to the very end of the ready list. */
10250 s390_z10_prevent_earlyload_conflicts (rtx
*ready
, int *nready_p
)
10252 unsigned int regno
;
10253 int nready
= *nready_p
;
10258 enum attr_type flag
;
10261 /* Skip DISTANCE - 1 active insns. */
10262 for (insn
= last_scheduled_insn
, distance
= Z10_EARLYLOAD_DISTANCE
- 1;
10263 distance
> 0 && insn
!= NULL_RTX
;
10264 distance
--, insn
= prev_active_insn (insn
))
10265 if (CALL_P (insn
) || JUMP_P (insn
))
10268 if (insn
== NULL_RTX
)
10271 set
= single_set (insn
);
10273 if (set
== NULL_RTX
|| !REG_P (SET_DEST (set
))
10274 || GET_MODE_CLASS (GET_MODE (SET_DEST (set
))) != MODE_FLOAT
)
10277 flag
= s390_safe_attr_type (insn
);
10279 if (flag
== TYPE_FLOADSF
|| flag
== TYPE_FLOADDF
)
10282 regno
= REGNO (SET_DEST (set
));
10285 while (!s390_fpload_toreg (ready
[i
], regno
) && i
> 0)
10292 memmove (&ready
[1], &ready
[0], sizeof (rtx
) * i
);
10296 /* This function is called via hook TARGET_SCHED_REORDER before
10297 issueing one insn from list READY which contains *NREADYP entries.
10298 For target z10 it reorders load instructions to avoid early load
10299 conflicts in the floating point pipeline */
10301 s390_sched_reorder (FILE *file ATTRIBUTE_UNUSED
, int verbose ATTRIBUTE_UNUSED
,
10302 rtx
*ready
, int *nreadyp
, int clock ATTRIBUTE_UNUSED
)
10304 if (s390_tune
== PROCESSOR_2097_Z10
)
10305 if (reload_completed
&& *nreadyp
> 1)
10306 s390_z10_prevent_earlyload_conflicts (ready
, nreadyp
);
10308 return s390_issue_rate ();
10311 /* This function is called via hook TARGET_SCHED_VARIABLE_ISSUE after
10312 the scheduler has issued INSN. It stores the last issued insn into
10313 last_scheduled_insn in order to make it available for
10314 s390_sched_reorder. */
10316 s390_sched_variable_issue (FILE *file ATTRIBUTE_UNUSED
,
10317 int verbose ATTRIBUTE_UNUSED
,
10318 rtx insn
, int more
)
10320 last_scheduled_insn
= insn
;
10322 if (GET_CODE (PATTERN (insn
)) != USE
10323 && GET_CODE (PATTERN (insn
)) != CLOBBER
)
10330 s390_sched_init (FILE *file ATTRIBUTE_UNUSED
,
10331 int verbose ATTRIBUTE_UNUSED
,
10332 int max_ready ATTRIBUTE_UNUSED
)
10334 last_scheduled_insn
= NULL_RTX
;
10337 /* This function checks the whole of insn X for memory references. The
10338 function always returns zero because the framework it is called
10339 from would stop recursively analyzing the insn upon a return value
10340 other than zero. The real result of this function is updating
10341 counter variable MEM_COUNT. */
10343 check_dpu (rtx
*x
, unsigned *mem_count
)
10345 if (*x
!= NULL_RTX
&& MEM_P (*x
))
10350 /* This target hook implementation for TARGET_LOOP_UNROLL_ADJUST calculates
10351 a new number struct loop *loop should be unrolled if tuned for the z10
10352 cpu. The loop is analyzed for memory accesses by calling check_dpu for
10353 each rtx of the loop. Depending on the loop_depth and the amount of
10354 memory accesses a new number <=nunroll is returned to improve the
10355 behaviour of the hardware prefetch unit. */
10357 s390_loop_unroll_adjust (unsigned nunroll
, struct loop
*loop
)
10362 unsigned mem_count
= 0;
10364 /* Only z10 needs special handling. */
10365 if (s390_tune
!= PROCESSOR_2097_Z10
)
10368 /* Count the number of memory references within the loop body. */
10369 bbs
= get_loop_body (loop
);
10370 for (i
= 0; i
< loop
->num_nodes
; i
++)
10372 for (insn
= BB_HEAD (bbs
[i
]); insn
!= BB_END (bbs
[i
]); insn
= NEXT_INSN (insn
))
10373 if (INSN_P (insn
) && INSN_CODE (insn
) != -1)
10374 for_each_rtx (&insn
, (rtx_function
) check_dpu
, &mem_count
);
10378 /* Prevent division by zero, and we do not need to adjust nunroll in this case. */
10379 if (mem_count
== 0)
10382 switch (loop_depth(loop
))
10385 return MIN (nunroll
, 28 / mem_count
);
10387 return MIN (nunroll
, 22 / mem_count
);
10389 return MIN (nunroll
, 16 / mem_count
);
10393 /* Initialize GCC target structure. */
10395 #undef TARGET_ASM_ALIGNED_HI_OP
10396 #define TARGET_ASM_ALIGNED_HI_OP "\t.word\t"
10397 #undef TARGET_ASM_ALIGNED_DI_OP
10398 #define TARGET_ASM_ALIGNED_DI_OP "\t.quad\t"
10399 #undef TARGET_ASM_INTEGER
10400 #define TARGET_ASM_INTEGER s390_assemble_integer
10402 #undef TARGET_ASM_OPEN_PAREN
10403 #define TARGET_ASM_OPEN_PAREN ""
10405 #undef TARGET_ASM_CLOSE_PAREN
10406 #define TARGET_ASM_CLOSE_PAREN ""
10408 #undef TARGET_DEFAULT_TARGET_FLAGS
10409 #define TARGET_DEFAULT_TARGET_FLAGS (TARGET_DEFAULT | MASK_FUSED_MADD)
10411 #undef TARGET_HANDLE_OPTION
10412 #define TARGET_HANDLE_OPTION s390_handle_option
10414 #undef TARGET_ENCODE_SECTION_INFO
10415 #define TARGET_ENCODE_SECTION_INFO s390_encode_section_info
10417 #undef TARGET_SCALAR_MODE_SUPPORTED_P
10418 #define TARGET_SCALAR_MODE_SUPPORTED_P s390_scalar_mode_supported_p
10421 #undef TARGET_HAVE_TLS
10422 #define TARGET_HAVE_TLS true
10424 #undef TARGET_CANNOT_FORCE_CONST_MEM
10425 #define TARGET_CANNOT_FORCE_CONST_MEM s390_cannot_force_const_mem
10427 #undef TARGET_DELEGITIMIZE_ADDRESS
10428 #define TARGET_DELEGITIMIZE_ADDRESS s390_delegitimize_address
10430 #undef TARGET_LEGITIMIZE_ADDRESS
10431 #define TARGET_LEGITIMIZE_ADDRESS s390_legitimize_address
10433 #undef TARGET_RETURN_IN_MEMORY
10434 #define TARGET_RETURN_IN_MEMORY s390_return_in_memory
10436 #undef TARGET_INIT_BUILTINS
10437 #define TARGET_INIT_BUILTINS s390_init_builtins
10438 #undef TARGET_EXPAND_BUILTIN
10439 #define TARGET_EXPAND_BUILTIN s390_expand_builtin
10441 #undef TARGET_ASM_OUTPUT_MI_THUNK
10442 #define TARGET_ASM_OUTPUT_MI_THUNK s390_output_mi_thunk
10443 #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
10444 #define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_const_tree_hwi_hwi_const_tree_true
10446 #undef TARGET_SCHED_ADJUST_PRIORITY
10447 #define TARGET_SCHED_ADJUST_PRIORITY s390_adjust_priority
10448 #undef TARGET_SCHED_ISSUE_RATE
10449 #define TARGET_SCHED_ISSUE_RATE s390_issue_rate
10450 #undef TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD
10451 #define TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD s390_first_cycle_multipass_dfa_lookahead
10453 #undef TARGET_SCHED_VARIABLE_ISSUE
10454 #define TARGET_SCHED_VARIABLE_ISSUE s390_sched_variable_issue
10455 #undef TARGET_SCHED_REORDER
10456 #define TARGET_SCHED_REORDER s390_sched_reorder
10457 #undef TARGET_SCHED_INIT
10458 #define TARGET_SCHED_INIT s390_sched_init
10460 #undef TARGET_CANNOT_COPY_INSN_P
10461 #define TARGET_CANNOT_COPY_INSN_P s390_cannot_copy_insn_p
10462 #undef TARGET_RTX_COSTS
10463 #define TARGET_RTX_COSTS s390_rtx_costs
10464 #undef TARGET_ADDRESS_COST
10465 #define TARGET_ADDRESS_COST s390_address_cost
10467 #undef TARGET_MACHINE_DEPENDENT_REORG
10468 #define TARGET_MACHINE_DEPENDENT_REORG s390_reorg
10470 #undef TARGET_VALID_POINTER_MODE
10471 #define TARGET_VALID_POINTER_MODE s390_valid_pointer_mode
10473 #undef TARGET_BUILD_BUILTIN_VA_LIST
10474 #define TARGET_BUILD_BUILTIN_VA_LIST s390_build_builtin_va_list
10475 #undef TARGET_EXPAND_BUILTIN_VA_START
10476 #define TARGET_EXPAND_BUILTIN_VA_START s390_va_start
10477 #undef TARGET_GIMPLIFY_VA_ARG_EXPR
10478 #define TARGET_GIMPLIFY_VA_ARG_EXPR s390_gimplify_va_arg
10480 #undef TARGET_PROMOTE_FUNCTION_MODE
10481 #define TARGET_PROMOTE_FUNCTION_MODE s390_promote_function_mode
10482 #undef TARGET_PASS_BY_REFERENCE
10483 #define TARGET_PASS_BY_REFERENCE s390_pass_by_reference
10485 #undef TARGET_FUNCTION_OK_FOR_SIBCALL
10486 #define TARGET_FUNCTION_OK_FOR_SIBCALL s390_function_ok_for_sibcall
10488 #undef TARGET_FIXED_CONDITION_CODE_REGS
10489 #define TARGET_FIXED_CONDITION_CODE_REGS s390_fixed_condition_code_regs
10491 #undef TARGET_CC_MODES_COMPATIBLE
10492 #define TARGET_CC_MODES_COMPATIBLE s390_cc_modes_compatible
10494 #undef TARGET_INVALID_WITHIN_DOLOOP
10495 #define TARGET_INVALID_WITHIN_DOLOOP hook_constcharptr_const_rtx_null
10498 #undef TARGET_ASM_OUTPUT_DWARF_DTPREL
10499 #define TARGET_ASM_OUTPUT_DWARF_DTPREL s390_output_dwarf_dtprel
10502 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
10503 #undef TARGET_MANGLE_TYPE
10504 #define TARGET_MANGLE_TYPE s390_mangle_type
10507 #undef TARGET_SCALAR_MODE_SUPPORTED_P
10508 #define TARGET_SCALAR_MODE_SUPPORTED_P s390_scalar_mode_supported_p
10510 #undef TARGET_SECONDARY_RELOAD
10511 #define TARGET_SECONDARY_RELOAD s390_secondary_reload
10513 #undef TARGET_LIBGCC_CMP_RETURN_MODE
10514 #define TARGET_LIBGCC_CMP_RETURN_MODE s390_libgcc_cmp_return_mode
10516 #undef TARGET_LIBGCC_SHIFT_COUNT_MODE
10517 #define TARGET_LIBGCC_SHIFT_COUNT_MODE s390_libgcc_shift_count_mode
10519 #undef TARGET_LEGITIMATE_ADDRESS_P
10520 #define TARGET_LEGITIMATE_ADDRESS_P s390_legitimate_address_p
10522 #undef TARGET_CAN_ELIMINATE
10523 #define TARGET_CAN_ELIMINATE s390_can_eliminate
10525 #undef TARGET_LOOP_UNROLL_ADJUST
10526 #define TARGET_LOOP_UNROLL_ADJUST s390_loop_unroll_adjust
10528 #undef TARGET_ASM_TRAMPOLINE_TEMPLATE
10529 #define TARGET_ASM_TRAMPOLINE_TEMPLATE s390_asm_trampoline_template
10530 #undef TARGET_TRAMPOLINE_INIT
10531 #define TARGET_TRAMPOLINE_INIT s390_trampoline_init
10533 #undef TARGET_UNWIND_WORD_MODE
10534 #define TARGET_UNWIND_WORD_MODE s390_unwind_word_mode
10536 struct gcc_target targetm
= TARGET_INITIALIZER
;
10538 #include "gt-s390.h"