1 /* Subroutines used for code generation on IBM S/390 and zSeries
2 Copyright (C) 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006,
3 2007, 2008, 2009, 2010 Free Software Foundation, Inc.
4 Contributed by Hartmut Penner (hpenner@de.ibm.com) and
5 Ulrich Weigand (uweigand@de.ibm.com) and
6 Andreas Krebbel (Andreas.Krebbel@de.ibm.com).
8 This file is part of GCC.
10 GCC is free software; you can redistribute it and/or modify it under
11 the terms of the GNU General Public License as published by the Free
12 Software Foundation; either version 3, or (at your option) any later
15 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
16 WARRANTY; without even the implied warranty of MERCHANTABILITY or
17 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
20 You should have received a copy of the GNU General Public License
21 along with GCC; see the file COPYING3. If not see
22 <http://www.gnu.org/licenses/>. */
26 #include "coretypes.h"
32 #include "hard-reg-set.h"
33 #include "insn-config.h"
34 #include "conditions.h"
36 #include "insn-attr.h"
43 #include "diagnostic-core.h"
45 #include "basic-block.h"
46 #include "integrate.h"
49 #include "target-def.h"
51 #include "langhooks.h"
59 /* Define the specific costs for a given cpu. */
61 struct processor_costs
64 const int m
; /* cost of an M instruction. */
65 const int mghi
; /* cost of an MGHI instruction. */
66 const int mh
; /* cost of an MH instruction. */
67 const int mhi
; /* cost of an MHI instruction. */
68 const int ml
; /* cost of an ML instruction. */
69 const int mr
; /* cost of an MR instruction. */
70 const int ms
; /* cost of an MS instruction. */
71 const int msg
; /* cost of an MSG instruction. */
72 const int msgf
; /* cost of an MSGF instruction. */
73 const int msgfr
; /* cost of an MSGFR instruction. */
74 const int msgr
; /* cost of an MSGR instruction. */
75 const int msr
; /* cost of an MSR instruction. */
76 const int mult_df
; /* cost of multiplication in DFmode. */
79 const int sqxbr
; /* cost of square root in TFmode. */
80 const int sqdbr
; /* cost of square root in DFmode. */
81 const int sqebr
; /* cost of square root in SFmode. */
82 /* multiply and add */
83 const int madbr
; /* cost of multiply and add in DFmode. */
84 const int maebr
; /* cost of multiply and add in SFmode. */
96 const struct processor_costs
*s390_cost
;
99 struct processor_costs z900_cost
=
101 COSTS_N_INSNS (5), /* M */
102 COSTS_N_INSNS (10), /* MGHI */
103 COSTS_N_INSNS (5), /* MH */
104 COSTS_N_INSNS (4), /* MHI */
105 COSTS_N_INSNS (5), /* ML */
106 COSTS_N_INSNS (5), /* MR */
107 COSTS_N_INSNS (4), /* MS */
108 COSTS_N_INSNS (15), /* MSG */
109 COSTS_N_INSNS (7), /* MSGF */
110 COSTS_N_INSNS (7), /* MSGFR */
111 COSTS_N_INSNS (10), /* MSGR */
112 COSTS_N_INSNS (4), /* MSR */
113 COSTS_N_INSNS (7), /* multiplication in DFmode */
114 COSTS_N_INSNS (13), /* MXBR */
115 COSTS_N_INSNS (136), /* SQXBR */
116 COSTS_N_INSNS (44), /* SQDBR */
117 COSTS_N_INSNS (35), /* SQEBR */
118 COSTS_N_INSNS (18), /* MADBR */
119 COSTS_N_INSNS (13), /* MAEBR */
120 COSTS_N_INSNS (134), /* DXBR */
121 COSTS_N_INSNS (30), /* DDBR */
122 COSTS_N_INSNS (27), /* DEBR */
123 COSTS_N_INSNS (220), /* DLGR */
124 COSTS_N_INSNS (34), /* DLR */
125 COSTS_N_INSNS (34), /* DR */
126 COSTS_N_INSNS (32), /* DSGFR */
127 COSTS_N_INSNS (32), /* DSGR */
131 struct processor_costs z990_cost
=
133 COSTS_N_INSNS (4), /* M */
134 COSTS_N_INSNS (2), /* MGHI */
135 COSTS_N_INSNS (2), /* MH */
136 COSTS_N_INSNS (2), /* MHI */
137 COSTS_N_INSNS (4), /* ML */
138 COSTS_N_INSNS (4), /* MR */
139 COSTS_N_INSNS (5), /* MS */
140 COSTS_N_INSNS (6), /* MSG */
141 COSTS_N_INSNS (4), /* MSGF */
142 COSTS_N_INSNS (4), /* MSGFR */
143 COSTS_N_INSNS (4), /* MSGR */
144 COSTS_N_INSNS (4), /* MSR */
145 COSTS_N_INSNS (1), /* multiplication in DFmode */
146 COSTS_N_INSNS (28), /* MXBR */
147 COSTS_N_INSNS (130), /* SQXBR */
148 COSTS_N_INSNS (66), /* SQDBR */
149 COSTS_N_INSNS (38), /* SQEBR */
150 COSTS_N_INSNS (1), /* MADBR */
151 COSTS_N_INSNS (1), /* MAEBR */
152 COSTS_N_INSNS (60), /* DXBR */
153 COSTS_N_INSNS (40), /* DDBR */
154 COSTS_N_INSNS (26), /* DEBR */
155 COSTS_N_INSNS (176), /* DLGR */
156 COSTS_N_INSNS (31), /* DLR */
157 COSTS_N_INSNS (31), /* DR */
158 COSTS_N_INSNS (31), /* DSGFR */
159 COSTS_N_INSNS (31), /* DSGR */
163 struct processor_costs z9_109_cost
=
165 COSTS_N_INSNS (4), /* M */
166 COSTS_N_INSNS (2), /* MGHI */
167 COSTS_N_INSNS (2), /* MH */
168 COSTS_N_INSNS (2), /* MHI */
169 COSTS_N_INSNS (4), /* ML */
170 COSTS_N_INSNS (4), /* MR */
171 COSTS_N_INSNS (5), /* MS */
172 COSTS_N_INSNS (6), /* MSG */
173 COSTS_N_INSNS (4), /* MSGF */
174 COSTS_N_INSNS (4), /* MSGFR */
175 COSTS_N_INSNS (4), /* MSGR */
176 COSTS_N_INSNS (4), /* MSR */
177 COSTS_N_INSNS (1), /* multiplication in DFmode */
178 COSTS_N_INSNS (28), /* MXBR */
179 COSTS_N_INSNS (130), /* SQXBR */
180 COSTS_N_INSNS (66), /* SQDBR */
181 COSTS_N_INSNS (38), /* SQEBR */
182 COSTS_N_INSNS (1), /* MADBR */
183 COSTS_N_INSNS (1), /* MAEBR */
184 COSTS_N_INSNS (60), /* DXBR */
185 COSTS_N_INSNS (40), /* DDBR */
186 COSTS_N_INSNS (26), /* DEBR */
187 COSTS_N_INSNS (30), /* DLGR */
188 COSTS_N_INSNS (23), /* DLR */
189 COSTS_N_INSNS (23), /* DR */
190 COSTS_N_INSNS (24), /* DSGFR */
191 COSTS_N_INSNS (24), /* DSGR */
195 struct processor_costs z10_cost
=
197 COSTS_N_INSNS (10), /* M */
198 COSTS_N_INSNS (10), /* MGHI */
199 COSTS_N_INSNS (10), /* MH */
200 COSTS_N_INSNS (10), /* MHI */
201 COSTS_N_INSNS (10), /* ML */
202 COSTS_N_INSNS (10), /* MR */
203 COSTS_N_INSNS (10), /* MS */
204 COSTS_N_INSNS (10), /* MSG */
205 COSTS_N_INSNS (10), /* MSGF */
206 COSTS_N_INSNS (10), /* MSGFR */
207 COSTS_N_INSNS (10), /* MSGR */
208 COSTS_N_INSNS (10), /* MSR */
209 COSTS_N_INSNS (1) , /* multiplication in DFmode */
210 COSTS_N_INSNS (50), /* MXBR */
211 COSTS_N_INSNS (120), /* SQXBR */
212 COSTS_N_INSNS (52), /* SQDBR */
213 COSTS_N_INSNS (38), /* SQEBR */
214 COSTS_N_INSNS (1), /* MADBR */
215 COSTS_N_INSNS (1), /* MAEBR */
216 COSTS_N_INSNS (111), /* DXBR */
217 COSTS_N_INSNS (39), /* DDBR */
218 COSTS_N_INSNS (32), /* DEBR */
219 COSTS_N_INSNS (160), /* DLGR */
220 COSTS_N_INSNS (71), /* DLR */
221 COSTS_N_INSNS (71), /* DR */
222 COSTS_N_INSNS (71), /* DSGFR */
223 COSTS_N_INSNS (71), /* DSGR */
226 extern int reload_completed
;
228 /* Kept up to date using the SCHED_VARIABLE_ISSUE hook. */
229 static rtx last_scheduled_insn
;
231 /* Structure used to hold the components of a S/390 memory
232 address. A legitimate address on S/390 is of the general
234 base + index + displacement
235 where any of the components is optional.
237 base and index are registers of the class ADDR_REGS,
238 displacement is an unsigned 12-bit immediate constant. */
249 /* Which cpu are we tuning for. */
250 enum processor_type s390_tune
= PROCESSOR_max
;
252 /* Which instruction set architecture to use. */
253 enum processor_type s390_arch
;
256 HOST_WIDE_INT s390_warn_framesize
= 0;
257 HOST_WIDE_INT s390_stack_size
= 0;
258 HOST_WIDE_INT s390_stack_guard
= 0;
260 /* The following structure is embedded in the machine
261 specific part of struct function. */
263 struct GTY (()) s390_frame_layout
265 /* Offset within stack frame. */
266 HOST_WIDE_INT gprs_offset
;
267 HOST_WIDE_INT f0_offset
;
268 HOST_WIDE_INT f4_offset
;
269 HOST_WIDE_INT f8_offset
;
270 HOST_WIDE_INT backchain_offset
;
272 /* Number of first and last gpr where slots in the register
273 save area are reserved for. */
274 int first_save_gpr_slot
;
275 int last_save_gpr_slot
;
277 /* Number of first and last gpr to be saved, restored. */
279 int first_restore_gpr
;
281 int last_restore_gpr
;
283 /* Bits standing for floating point registers. Set, if the
284 respective register has to be saved. Starting with reg 16 (f0)
285 at the rightmost bit.
286 Bit 15 - 8 7 6 5 4 3 2 1 0
287 fpr 15 - 8 7 5 3 1 6 4 2 0
288 reg 31 - 24 23 22 21 20 19 18 17 16 */
289 unsigned int fpr_bitmap
;
291 /* Number of floating point registers f8-f15 which must be saved. */
294 /* Set if return address needs to be saved.
295 This flag is set by s390_return_addr_rtx if it could not use
296 the initial value of r14 and therefore depends on r14 saved
298 bool save_return_addr_p
;
300 /* Size of stack frame. */
301 HOST_WIDE_INT frame_size
;
304 /* Define the structure for the machine field in struct function. */
306 struct GTY(()) machine_function
308 struct s390_frame_layout frame_layout
;
310 /* Literal pool base register. */
313 /* True if we may need to perform branch splitting. */
314 bool split_branches_pending_p
;
316 /* Some local-dynamic TLS symbol name. */
317 const char *some_ld_name
;
319 bool has_landing_pad_p
;
322 /* Few accessor macros for struct cfun->machine->s390_frame_layout. */
324 #define cfun_frame_layout (cfun->machine->frame_layout)
325 #define cfun_save_high_fprs_p (!!cfun_frame_layout.high_fprs)
326 #define cfun_gprs_save_area_size ((cfun_frame_layout.last_save_gpr_slot - \
327 cfun_frame_layout.first_save_gpr_slot + 1) * UNITS_PER_LONG)
328 #define cfun_set_fpr_bit(BITNUM) (cfun->machine->frame_layout.fpr_bitmap |= \
330 #define cfun_fpr_bit_p(BITNUM) (!!(cfun->machine->frame_layout.fpr_bitmap & \
333 /* Number of GPRs and FPRs used for argument passing. */
334 #define GP_ARG_NUM_REG 5
335 #define FP_ARG_NUM_REG (TARGET_64BIT? 4 : 2)
337 /* A couple of shortcuts. */
338 #define CONST_OK_FOR_J(x) \
339 CONST_OK_FOR_CONSTRAINT_P((x), 'J', "J")
340 #define CONST_OK_FOR_K(x) \
341 CONST_OK_FOR_CONSTRAINT_P((x), 'K', "K")
342 #define CONST_OK_FOR_Os(x) \
343 CONST_OK_FOR_CONSTRAINT_P((x), 'O', "Os")
344 #define CONST_OK_FOR_Op(x) \
345 CONST_OK_FOR_CONSTRAINT_P((x), 'O', "Op")
346 #define CONST_OK_FOR_On(x) \
347 CONST_OK_FOR_CONSTRAINT_P((x), 'O', "On")
349 #define REGNO_PAIR_OK(REGNO, MODE) \
350 (HARD_REGNO_NREGS ((REGNO), (MODE)) == 1 || !((REGNO) & 1))
352 /* That's the read ahead of the dynamic branch prediction unit in
353 bytes on a z10 CPU. */
354 #define Z10_PREDICT_DISTANCE 384
356 static enum machine_mode
357 s390_libgcc_cmp_return_mode (void)
359 return TARGET_64BIT
? DImode
: SImode
;
362 static enum machine_mode
363 s390_libgcc_shift_count_mode (void)
365 return TARGET_64BIT
? DImode
: SImode
;
368 static enum machine_mode
369 s390_unwind_word_mode (void)
371 return TARGET_64BIT
? DImode
: SImode
;
374 /* Return true if the back end supports mode MODE. */
376 s390_scalar_mode_supported_p (enum machine_mode mode
)
378 /* In contrast to the default implementation reject TImode constants on 31bit
379 TARGET_ZARCH for ABI compliance. */
380 if (!TARGET_64BIT
&& TARGET_ZARCH
&& mode
== TImode
)
383 if (DECIMAL_FLOAT_MODE_P (mode
))
384 return default_decimal_float_supported_p ();
386 return default_scalar_mode_supported_p (mode
);
389 /* Set the has_landing_pad_p flag in struct machine_function to VALUE. */
392 s390_set_has_landing_pad_p (bool value
)
394 cfun
->machine
->has_landing_pad_p
= value
;
397 /* If two condition code modes are compatible, return a condition code
398 mode which is compatible with both. Otherwise, return
401 static enum machine_mode
402 s390_cc_modes_compatible (enum machine_mode m1
, enum machine_mode m2
)
410 if (m2
== CCUmode
|| m2
== CCTmode
|| m2
== CCZ1mode
411 || m2
== CCSmode
|| m2
== CCSRmode
|| m2
== CCURmode
)
432 /* Return true if SET either doesn't set the CC register, or else
433 the source and destination have matching CC modes and that
434 CC mode is at least as constrained as REQ_MODE. */
437 s390_match_ccmode_set (rtx set
, enum machine_mode req_mode
)
439 enum machine_mode set_mode
;
441 gcc_assert (GET_CODE (set
) == SET
);
443 if (GET_CODE (SET_DEST (set
)) != REG
|| !CC_REGNO_P (REGNO (SET_DEST (set
))))
446 set_mode
= GET_MODE (SET_DEST (set
));
460 if (req_mode
!= set_mode
)
465 if (req_mode
!= CCSmode
&& req_mode
!= CCUmode
&& req_mode
!= CCTmode
466 && req_mode
!= CCSRmode
&& req_mode
!= CCURmode
)
472 if (req_mode
!= CCAmode
)
480 return (GET_MODE (SET_SRC (set
)) == set_mode
);
483 /* Return true if every SET in INSN that sets the CC register
484 has source and destination with matching CC modes and that
485 CC mode is at least as constrained as REQ_MODE.
486 If REQ_MODE is VOIDmode, always return false. */
489 s390_match_ccmode (rtx insn
, enum machine_mode req_mode
)
493 /* s390_tm_ccmode returns VOIDmode to indicate failure. */
494 if (req_mode
== VOIDmode
)
497 if (GET_CODE (PATTERN (insn
)) == SET
)
498 return s390_match_ccmode_set (PATTERN (insn
), req_mode
);
500 if (GET_CODE (PATTERN (insn
)) == PARALLEL
)
501 for (i
= 0; i
< XVECLEN (PATTERN (insn
), 0); i
++)
503 rtx set
= XVECEXP (PATTERN (insn
), 0, i
);
504 if (GET_CODE (set
) == SET
)
505 if (!s390_match_ccmode_set (set
, req_mode
))
512 /* If a test-under-mask instruction can be used to implement
513 (compare (and ... OP1) OP2), return the CC mode required
514 to do that. Otherwise, return VOIDmode.
515 MIXED is true if the instruction can distinguish between
516 CC1 and CC2 for mixed selected bits (TMxx), it is false
517 if the instruction cannot (TM). */
520 s390_tm_ccmode (rtx op1
, rtx op2
, bool mixed
)
524 /* ??? Fixme: should work on CONST_DOUBLE as well. */
525 if (GET_CODE (op1
) != CONST_INT
|| GET_CODE (op2
) != CONST_INT
)
528 /* Selected bits all zero: CC0.
529 e.g.: int a; if ((a & (16 + 128)) == 0) */
530 if (INTVAL (op2
) == 0)
533 /* Selected bits all one: CC3.
534 e.g.: int a; if ((a & (16 + 128)) == 16 + 128) */
535 if (INTVAL (op2
) == INTVAL (op1
))
538 /* Exactly two bits selected, mixed zeroes and ones: CC1 or CC2. e.g.:
540 if ((a & (16 + 128)) == 16) -> CCT1
541 if ((a & (16 + 128)) == 128) -> CCT2 */
544 bit1
= exact_log2 (INTVAL (op2
));
545 bit0
= exact_log2 (INTVAL (op1
) ^ INTVAL (op2
));
546 if (bit0
!= -1 && bit1
!= -1)
547 return bit0
> bit1
? CCT1mode
: CCT2mode
;
553 /* Given a comparison code OP (EQ, NE, etc.) and the operands
554 OP0 and OP1 of a COMPARE, return the mode to be used for the
558 s390_select_ccmode (enum rtx_code code
, rtx op0
, rtx op1
)
564 if ((GET_CODE (op0
) == NEG
|| GET_CODE (op0
) == ABS
)
565 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
567 if (GET_CODE (op0
) == PLUS
&& GET_CODE (XEXP (op0
, 1)) == CONST_INT
568 && CONST_OK_FOR_K (INTVAL (XEXP (op0
, 1))))
570 if ((GET_CODE (op0
) == PLUS
|| GET_CODE (op0
) == MINUS
571 || GET_CODE (op1
) == NEG
)
572 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
575 if (GET_CODE (op0
) == AND
)
577 /* Check whether we can potentially do it via TM. */
578 enum machine_mode ccmode
;
579 ccmode
= s390_tm_ccmode (XEXP (op0
, 1), op1
, 1);
580 if (ccmode
!= VOIDmode
)
582 /* Relax CCTmode to CCZmode to allow fall-back to AND
583 if that turns out to be beneficial. */
584 return ccmode
== CCTmode
? CCZmode
: ccmode
;
588 if (register_operand (op0
, HImode
)
589 && GET_CODE (op1
) == CONST_INT
590 && (INTVAL (op1
) == -1 || INTVAL (op1
) == 65535))
592 if (register_operand (op0
, QImode
)
593 && GET_CODE (op1
) == CONST_INT
594 && (INTVAL (op1
) == -1 || INTVAL (op1
) == 255))
603 /* The only overflow condition of NEG and ABS happens when
604 -INT_MAX is used as parameter, which stays negative. So
605 we have an overflow from a positive value to a negative.
606 Using CCAP mode the resulting cc can be used for comparisons. */
607 if ((GET_CODE (op0
) == NEG
|| GET_CODE (op0
) == ABS
)
608 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
611 /* If constants are involved in an add instruction it is possible to use
612 the resulting cc for comparisons with zero. Knowing the sign of the
613 constant the overflow behavior gets predictable. e.g.:
614 int a, b; if ((b = a + c) > 0)
615 with c as a constant value: c < 0 -> CCAN and c >= 0 -> CCAP */
616 if (GET_CODE (op0
) == PLUS
&& GET_CODE (XEXP (op0
, 1)) == CONST_INT
617 && CONST_OK_FOR_K (INTVAL (XEXP (op0
, 1))))
619 if (INTVAL (XEXP((op0
), 1)) < 0)
633 if ((GET_CODE (op0
) == SIGN_EXTEND
|| GET_CODE (op0
) == ZERO_EXTEND
)
634 && GET_CODE (op1
) != CONST_INT
)
640 if (GET_CODE (op0
) == PLUS
641 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
644 if ((GET_CODE (op0
) == SIGN_EXTEND
|| GET_CODE (op0
) == ZERO_EXTEND
)
645 && GET_CODE (op1
) != CONST_INT
)
651 if (GET_CODE (op0
) == MINUS
652 && GET_MODE_CLASS (GET_MODE (op0
)) == MODE_INT
)
655 if ((GET_CODE (op0
) == SIGN_EXTEND
|| GET_CODE (op0
) == ZERO_EXTEND
)
656 && GET_CODE (op1
) != CONST_INT
)
665 /* Replace the comparison OP0 CODE OP1 by a semantically equivalent one
666 that we can implement more efficiently. */
669 s390_canonicalize_comparison (enum rtx_code
*code
, rtx
*op0
, rtx
*op1
)
671 /* Convert ZERO_EXTRACT back to AND to enable TM patterns. */
672 if ((*code
== EQ
|| *code
== NE
)
673 && *op1
== const0_rtx
674 && GET_CODE (*op0
) == ZERO_EXTRACT
675 && GET_CODE (XEXP (*op0
, 1)) == CONST_INT
676 && GET_CODE (XEXP (*op0
, 2)) == CONST_INT
677 && SCALAR_INT_MODE_P (GET_MODE (XEXP (*op0
, 0))))
679 rtx inner
= XEXP (*op0
, 0);
680 HOST_WIDE_INT modesize
= GET_MODE_BITSIZE (GET_MODE (inner
));
681 HOST_WIDE_INT len
= INTVAL (XEXP (*op0
, 1));
682 HOST_WIDE_INT pos
= INTVAL (XEXP (*op0
, 2));
684 if (len
> 0 && len
< modesize
685 && pos
>= 0 && pos
+ len
<= modesize
686 && modesize
<= HOST_BITS_PER_WIDE_INT
)
688 unsigned HOST_WIDE_INT block
;
689 block
= ((unsigned HOST_WIDE_INT
) 1 << len
) - 1;
690 block
<<= modesize
- pos
- len
;
692 *op0
= gen_rtx_AND (GET_MODE (inner
), inner
,
693 gen_int_mode (block
, GET_MODE (inner
)));
697 /* Narrow AND of memory against immediate to enable TM. */
698 if ((*code
== EQ
|| *code
== NE
)
699 && *op1
== const0_rtx
700 && GET_CODE (*op0
) == AND
701 && GET_CODE (XEXP (*op0
, 1)) == CONST_INT
702 && SCALAR_INT_MODE_P (GET_MODE (XEXP (*op0
, 0))))
704 rtx inner
= XEXP (*op0
, 0);
705 rtx mask
= XEXP (*op0
, 1);
707 /* Ignore paradoxical SUBREGs if all extra bits are masked out. */
708 if (GET_CODE (inner
) == SUBREG
709 && SCALAR_INT_MODE_P (GET_MODE (SUBREG_REG (inner
)))
710 && (GET_MODE_SIZE (GET_MODE (inner
))
711 >= GET_MODE_SIZE (GET_MODE (SUBREG_REG (inner
))))
713 & GET_MODE_MASK (GET_MODE (inner
))
714 & ~GET_MODE_MASK (GET_MODE (SUBREG_REG (inner
))))
716 inner
= SUBREG_REG (inner
);
718 /* Do not change volatile MEMs. */
719 if (MEM_P (inner
) && !MEM_VOLATILE_P (inner
))
721 int part
= s390_single_part (XEXP (*op0
, 1),
722 GET_MODE (inner
), QImode
, 0);
725 mask
= gen_int_mode (s390_extract_part (mask
, QImode
, 0), QImode
);
726 inner
= adjust_address_nv (inner
, QImode
, part
);
727 *op0
= gen_rtx_AND (QImode
, inner
, mask
);
732 /* Narrow comparisons against 0xffff to HImode if possible. */
733 if ((*code
== EQ
|| *code
== NE
)
734 && GET_CODE (*op1
) == CONST_INT
735 && INTVAL (*op1
) == 0xffff
736 && SCALAR_INT_MODE_P (GET_MODE (*op0
))
737 && (nonzero_bits (*op0
, GET_MODE (*op0
))
738 & ~(unsigned HOST_WIDE_INT
) 0xffff) == 0)
740 *op0
= gen_lowpart (HImode
, *op0
);
744 /* Remove redundant UNSPEC_CCU_TO_INT conversions if possible. */
745 if (GET_CODE (*op0
) == UNSPEC
746 && XINT (*op0
, 1) == UNSPEC_CCU_TO_INT
747 && XVECLEN (*op0
, 0) == 1
748 && GET_MODE (XVECEXP (*op0
, 0, 0)) == CCUmode
749 && GET_CODE (XVECEXP (*op0
, 0, 0)) == REG
750 && REGNO (XVECEXP (*op0
, 0, 0)) == CC_REGNUM
751 && *op1
== const0_rtx
)
753 enum rtx_code new_code
= UNKNOWN
;
756 case EQ
: new_code
= EQ
; break;
757 case NE
: new_code
= NE
; break;
758 case LT
: new_code
= GTU
; break;
759 case GT
: new_code
= LTU
; break;
760 case LE
: new_code
= GEU
; break;
761 case GE
: new_code
= LEU
; break;
765 if (new_code
!= UNKNOWN
)
767 *op0
= XVECEXP (*op0
, 0, 0);
772 /* Remove redundant UNSPEC_CCZ_TO_INT conversions if possible. */
773 if (GET_CODE (*op0
) == UNSPEC
774 && XINT (*op0
, 1) == UNSPEC_CCZ_TO_INT
775 && XVECLEN (*op0
, 0) == 1
776 && GET_MODE (XVECEXP (*op0
, 0, 0)) == CCZmode
777 && GET_CODE (XVECEXP (*op0
, 0, 0)) == REG
778 && REGNO (XVECEXP (*op0
, 0, 0)) == CC_REGNUM
779 && *op1
== const0_rtx
)
781 enum rtx_code new_code
= UNKNOWN
;
784 case EQ
: new_code
= EQ
; break;
785 case NE
: new_code
= NE
; break;
789 if (new_code
!= UNKNOWN
)
791 *op0
= XVECEXP (*op0
, 0, 0);
796 /* Simplify cascaded EQ, NE with const0_rtx. */
797 if ((*code
== NE
|| *code
== EQ
)
798 && (GET_CODE (*op0
) == EQ
|| GET_CODE (*op0
) == NE
)
799 && GET_MODE (*op0
) == SImode
800 && GET_MODE (XEXP (*op0
, 0)) == CCZ1mode
801 && REG_P (XEXP (*op0
, 0))
802 && XEXP (*op0
, 1) == const0_rtx
803 && *op1
== const0_rtx
)
805 if ((*code
== EQ
&& GET_CODE (*op0
) == NE
)
806 || (*code
== NE
&& GET_CODE (*op0
) == EQ
))
810 *op0
= XEXP (*op0
, 0);
813 /* Prefer register over memory as first operand. */
814 if (MEM_P (*op0
) && REG_P (*op1
))
816 rtx tem
= *op0
; *op0
= *op1
; *op1
= tem
;
817 *code
= swap_condition (*code
);
821 /* Emit a compare instruction suitable to implement the comparison
822 OP0 CODE OP1. Return the correct condition RTL to be placed in
823 the IF_THEN_ELSE of the conditional branch testing the result. */
826 s390_emit_compare (enum rtx_code code
, rtx op0
, rtx op1
)
828 enum machine_mode mode
= s390_select_ccmode (code
, op0
, op1
);
831 /* Do not output a redundant compare instruction if a compare_and_swap
832 pattern already computed the result and the machine modes are compatible. */
833 if (GET_MODE_CLASS (GET_MODE (op0
)) == MODE_CC
)
835 gcc_assert (s390_cc_modes_compatible (GET_MODE (op0
), mode
)
841 cc
= gen_rtx_REG (mode
, CC_REGNUM
);
842 emit_insn (gen_rtx_SET (VOIDmode
, cc
, gen_rtx_COMPARE (mode
, op0
, op1
)));
845 return gen_rtx_fmt_ee (code
, VOIDmode
, cc
, const0_rtx
);
848 /* Emit a SImode compare and swap instruction setting MEM to NEW_RTX if OLD
850 Return the correct condition RTL to be placed in the IF_THEN_ELSE of the
851 conditional branch testing the result. */
854 s390_emit_compare_and_swap (enum rtx_code code
, rtx old
, rtx mem
, rtx cmp
, rtx new_rtx
)
856 emit_insn (gen_sync_compare_and_swapsi (old
, mem
, cmp
, new_rtx
));
857 return s390_emit_compare (code
, gen_rtx_REG (CCZ1mode
, CC_REGNUM
), const0_rtx
);
860 /* Emit a jump instruction to TARGET. If COND is NULL_RTX, emit an
861 unconditional jump, else a conditional jump under condition COND. */
864 s390_emit_jump (rtx target
, rtx cond
)
868 target
= gen_rtx_LABEL_REF (VOIDmode
, target
);
870 target
= gen_rtx_IF_THEN_ELSE (VOIDmode
, cond
, target
, pc_rtx
);
872 insn
= gen_rtx_SET (VOIDmode
, pc_rtx
, target
);
873 emit_jump_insn (insn
);
876 /* Return branch condition mask to implement a branch
877 specified by CODE. Return -1 for invalid comparisons. */
880 s390_branch_condition_mask (rtx code
)
882 const int CC0
= 1 << 3;
883 const int CC1
= 1 << 2;
884 const int CC2
= 1 << 1;
885 const int CC3
= 1 << 0;
887 gcc_assert (GET_CODE (XEXP (code
, 0)) == REG
);
888 gcc_assert (REGNO (XEXP (code
, 0)) == CC_REGNUM
);
889 gcc_assert (XEXP (code
, 1) == const0_rtx
);
891 switch (GET_MODE (XEXP (code
, 0)))
895 switch (GET_CODE (code
))
898 case NE
: return CC1
| CC2
| CC3
;
904 switch (GET_CODE (code
))
907 case NE
: return CC0
| CC2
| CC3
;
913 switch (GET_CODE (code
))
916 case NE
: return CC0
| CC1
| CC3
;
922 switch (GET_CODE (code
))
925 case NE
: return CC0
| CC1
| CC2
;
931 switch (GET_CODE (code
))
933 case EQ
: return CC0
| CC2
;
934 case NE
: return CC1
| CC3
;
940 switch (GET_CODE (code
))
942 case LTU
: return CC2
| CC3
; /* carry */
943 case GEU
: return CC0
| CC1
; /* no carry */
949 switch (GET_CODE (code
))
951 case GTU
: return CC0
| CC1
; /* borrow */
952 case LEU
: return CC2
| CC3
; /* no borrow */
958 switch (GET_CODE (code
))
960 case EQ
: return CC0
| CC2
;
961 case NE
: return CC1
| CC3
;
962 case LTU
: return CC1
;
963 case GTU
: return CC3
;
964 case LEU
: return CC1
| CC2
;
965 case GEU
: return CC2
| CC3
;
970 switch (GET_CODE (code
))
973 case NE
: return CC1
| CC2
| CC3
;
974 case LTU
: return CC1
;
975 case GTU
: return CC2
;
976 case LEU
: return CC0
| CC1
;
977 case GEU
: return CC0
| CC2
;
983 switch (GET_CODE (code
))
986 case NE
: return CC2
| CC1
| CC3
;
987 case LTU
: return CC2
;
988 case GTU
: return CC1
;
989 case LEU
: return CC0
| CC2
;
990 case GEU
: return CC0
| CC1
;
996 switch (GET_CODE (code
))
999 case NE
: return CC1
| CC2
| CC3
;
1000 case LT
: return CC1
| CC3
;
1001 case GT
: return CC2
;
1002 case LE
: return CC0
| CC1
| CC3
;
1003 case GE
: return CC0
| CC2
;
1009 switch (GET_CODE (code
))
1011 case EQ
: return CC0
;
1012 case NE
: return CC1
| CC2
| CC3
;
1013 case LT
: return CC1
;
1014 case GT
: return CC2
| CC3
;
1015 case LE
: return CC0
| CC1
;
1016 case GE
: return CC0
| CC2
| CC3
;
1022 switch (GET_CODE (code
))
1024 case EQ
: return CC0
;
1025 case NE
: return CC1
| CC2
| CC3
;
1026 case LT
: return CC1
;
1027 case GT
: return CC2
;
1028 case LE
: return CC0
| CC1
;
1029 case GE
: return CC0
| CC2
;
1030 case UNORDERED
: return CC3
;
1031 case ORDERED
: return CC0
| CC1
| CC2
;
1032 case UNEQ
: return CC0
| CC3
;
1033 case UNLT
: return CC1
| CC3
;
1034 case UNGT
: return CC2
| CC3
;
1035 case UNLE
: return CC0
| CC1
| CC3
;
1036 case UNGE
: return CC0
| CC2
| CC3
;
1037 case LTGT
: return CC1
| CC2
;
1043 switch (GET_CODE (code
))
1045 case EQ
: return CC0
;
1046 case NE
: return CC2
| CC1
| CC3
;
1047 case LT
: return CC2
;
1048 case GT
: return CC1
;
1049 case LE
: return CC0
| CC2
;
1050 case GE
: return CC0
| CC1
;
1051 case UNORDERED
: return CC3
;
1052 case ORDERED
: return CC0
| CC2
| CC1
;
1053 case UNEQ
: return CC0
| CC3
;
1054 case UNLT
: return CC2
| CC3
;
1055 case UNGT
: return CC1
| CC3
;
1056 case UNLE
: return CC0
| CC2
| CC3
;
1057 case UNGE
: return CC0
| CC1
| CC3
;
1058 case LTGT
: return CC2
| CC1
;
1069 /* Return branch condition mask to implement a compare and branch
1070 specified by CODE. Return -1 for invalid comparisons. */
1073 s390_compare_and_branch_condition_mask (rtx code
)
1075 const int CC0
= 1 << 3;
1076 const int CC1
= 1 << 2;
1077 const int CC2
= 1 << 1;
1079 switch (GET_CODE (code
))
1103 /* If INV is false, return assembler mnemonic string to implement
1104 a branch specified by CODE. If INV is true, return mnemonic
1105 for the corresponding inverted branch. */
1108 s390_branch_condition_mnemonic (rtx code
, int inv
)
1112 static const char *const mnemonic
[16] =
1114 NULL
, "o", "h", "nle",
1115 "l", "nhe", "lh", "ne",
1116 "e", "nlh", "he", "nl",
1117 "le", "nh", "no", NULL
1120 if (GET_CODE (XEXP (code
, 0)) == REG
1121 && REGNO (XEXP (code
, 0)) == CC_REGNUM
1122 && XEXP (code
, 1) == const0_rtx
)
1123 mask
= s390_branch_condition_mask (code
);
1125 mask
= s390_compare_and_branch_condition_mask (code
);
1127 gcc_assert (mask
>= 0);
1132 gcc_assert (mask
>= 1 && mask
<= 14);
1134 return mnemonic
[mask
];
1137 /* Return the part of op which has a value different from def.
1138 The size of the part is determined by mode.
1139 Use this function only if you already know that op really
1140 contains such a part. */
1142 unsigned HOST_WIDE_INT
1143 s390_extract_part (rtx op
, enum machine_mode mode
, int def
)
1145 unsigned HOST_WIDE_INT value
= 0;
1146 int max_parts
= HOST_BITS_PER_WIDE_INT
/ GET_MODE_BITSIZE (mode
);
1147 int part_bits
= GET_MODE_BITSIZE (mode
);
1148 unsigned HOST_WIDE_INT part_mask
1149 = ((unsigned HOST_WIDE_INT
)1 << part_bits
) - 1;
1152 for (i
= 0; i
< max_parts
; i
++)
1155 value
= (unsigned HOST_WIDE_INT
) INTVAL (op
);
1157 value
>>= part_bits
;
1159 if ((value
& part_mask
) != (def
& part_mask
))
1160 return value
& part_mask
;
1166 /* If OP is an integer constant of mode MODE with exactly one
1167 part of mode PART_MODE unequal to DEF, return the number of that
1168 part. Otherwise, return -1. */
1171 s390_single_part (rtx op
,
1172 enum machine_mode mode
,
1173 enum machine_mode part_mode
,
1176 unsigned HOST_WIDE_INT value
= 0;
1177 int n_parts
= GET_MODE_SIZE (mode
) / GET_MODE_SIZE (part_mode
);
1178 unsigned HOST_WIDE_INT part_mask
1179 = ((unsigned HOST_WIDE_INT
)1 << GET_MODE_BITSIZE (part_mode
)) - 1;
1182 if (GET_CODE (op
) != CONST_INT
)
1185 for (i
= 0; i
< n_parts
; i
++)
1188 value
= (unsigned HOST_WIDE_INT
) INTVAL (op
);
1190 value
>>= GET_MODE_BITSIZE (part_mode
);
1192 if ((value
& part_mask
) != (def
& part_mask
))
1200 return part
== -1 ? -1 : n_parts
- 1 - part
;
1203 /* Return true if IN contains a contiguous bitfield in the lower SIZE
1204 bits and no other bits are set in IN. POS and LENGTH can be used
1205 to obtain the start position and the length of the bitfield.
1207 POS gives the position of the first bit of the bitfield counting
1208 from the lowest order bit starting with zero. In order to use this
1209 value for S/390 instructions this has to be converted to "bits big
1213 s390_contiguous_bitmask_p (unsigned HOST_WIDE_INT in
, int size
,
1214 int *pos
, int *length
)
1219 unsigned HOST_WIDE_INT mask
= 1ULL;
1220 bool contiguous
= false;
1222 for (i
= 0; i
< size
; mask
<<= 1, i
++)
1246 /* Calculate a mask for all bits beyond the contiguous bits. */
1247 mask
= (-1LL & ~(((1ULL << (tmp_length
+ tmp_pos
- 1)) << 1) - 1));
1252 if (tmp_length
+ tmp_pos
- 1 > size
)
1256 *length
= tmp_length
;
1264 /* Check whether we can (and want to) split a double-word
1265 move in mode MODE from SRC to DST into two single-word
1266 moves, moving the subword FIRST_SUBWORD first. */
1269 s390_split_ok_p (rtx dst
, rtx src
, enum machine_mode mode
, int first_subword
)
1271 /* Floating point registers cannot be split. */
1272 if (FP_REG_P (src
) || FP_REG_P (dst
))
1275 /* We don't need to split if operands are directly accessible. */
1276 if (s_operand (src
, mode
) || s_operand (dst
, mode
))
1279 /* Non-offsettable memory references cannot be split. */
1280 if ((GET_CODE (src
) == MEM
&& !offsettable_memref_p (src
))
1281 || (GET_CODE (dst
) == MEM
&& !offsettable_memref_p (dst
)))
1284 /* Moving the first subword must not clobber a register
1285 needed to move the second subword. */
1286 if (register_operand (dst
, mode
))
1288 rtx subreg
= operand_subword (dst
, first_subword
, 0, mode
);
1289 if (reg_overlap_mentioned_p (subreg
, src
))
1296 /* Return true if it can be proven that [MEM1, MEM1 + SIZE]
1297 and [MEM2, MEM2 + SIZE] do overlap and false
1301 s390_overlap_p (rtx mem1
, rtx mem2
, HOST_WIDE_INT size
)
1303 rtx addr1
, addr2
, addr_delta
;
1304 HOST_WIDE_INT delta
;
1306 if (GET_CODE (mem1
) != MEM
|| GET_CODE (mem2
) != MEM
)
1312 addr1
= XEXP (mem1
, 0);
1313 addr2
= XEXP (mem2
, 0);
1315 addr_delta
= simplify_binary_operation (MINUS
, Pmode
, addr2
, addr1
);
1317 /* This overlapping check is used by peepholes merging memory block operations.
1318 Overlapping operations would otherwise be recognized by the S/390 hardware
1319 and would fall back to a slower implementation. Allowing overlapping
1320 operations would lead to slow code but not to wrong code. Therefore we are
1321 somewhat optimistic if we cannot prove that the memory blocks are
1323 That's why we return false here although this may accept operations on
1324 overlapping memory areas. */
1325 if (!addr_delta
|| GET_CODE (addr_delta
) != CONST_INT
)
1328 delta
= INTVAL (addr_delta
);
1331 || (delta
> 0 && delta
< size
)
1332 || (delta
< 0 && -delta
< size
))
1338 /* Check whether the address of memory reference MEM2 equals exactly
1339 the address of memory reference MEM1 plus DELTA. Return true if
1340 we can prove this to be the case, false otherwise. */
1343 s390_offset_p (rtx mem1
, rtx mem2
, rtx delta
)
1345 rtx addr1
, addr2
, addr_delta
;
1347 if (GET_CODE (mem1
) != MEM
|| GET_CODE (mem2
) != MEM
)
1350 addr1
= XEXP (mem1
, 0);
1351 addr2
= XEXP (mem2
, 0);
1353 addr_delta
= simplify_binary_operation (MINUS
, Pmode
, addr2
, addr1
);
1354 if (!addr_delta
|| !rtx_equal_p (addr_delta
, delta
))
1360 /* Expand logical operator CODE in mode MODE with operands OPERANDS. */
1363 s390_expand_logical_operator (enum rtx_code code
, enum machine_mode mode
,
1366 enum machine_mode wmode
= mode
;
1367 rtx dst
= operands
[0];
1368 rtx src1
= operands
[1];
1369 rtx src2
= operands
[2];
1372 /* If we cannot handle the operation directly, use a temp register. */
1373 if (!s390_logical_operator_ok_p (operands
))
1374 dst
= gen_reg_rtx (mode
);
1376 /* QImode and HImode patterns make sense only if we have a destination
1377 in memory. Otherwise perform the operation in SImode. */
1378 if ((mode
== QImode
|| mode
== HImode
) && GET_CODE (dst
) != MEM
)
1381 /* Widen operands if required. */
1384 if (GET_CODE (dst
) == SUBREG
1385 && (tem
= simplify_subreg (wmode
, dst
, mode
, 0)) != 0)
1387 else if (REG_P (dst
))
1388 dst
= gen_rtx_SUBREG (wmode
, dst
, 0);
1390 dst
= gen_reg_rtx (wmode
);
1392 if (GET_CODE (src1
) == SUBREG
1393 && (tem
= simplify_subreg (wmode
, src1
, mode
, 0)) != 0)
1395 else if (GET_MODE (src1
) != VOIDmode
)
1396 src1
= gen_rtx_SUBREG (wmode
, force_reg (mode
, src1
), 0);
1398 if (GET_CODE (src2
) == SUBREG
1399 && (tem
= simplify_subreg (wmode
, src2
, mode
, 0)) != 0)
1401 else if (GET_MODE (src2
) != VOIDmode
)
1402 src2
= gen_rtx_SUBREG (wmode
, force_reg (mode
, src2
), 0);
1405 /* Emit the instruction. */
1406 op
= gen_rtx_SET (VOIDmode
, dst
, gen_rtx_fmt_ee (code
, wmode
, src1
, src2
));
1407 clob
= gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
1408 emit_insn (gen_rtx_PARALLEL (VOIDmode
, gen_rtvec (2, op
, clob
)));
1410 /* Fix up the destination if needed. */
1411 if (dst
!= operands
[0])
1412 emit_move_insn (operands
[0], gen_lowpart (mode
, dst
));
1415 /* Check whether OPERANDS are OK for a logical operation (AND, IOR, XOR). */
1418 s390_logical_operator_ok_p (rtx
*operands
)
1420 /* If the destination operand is in memory, it needs to coincide
1421 with one of the source operands. After reload, it has to be
1422 the first source operand. */
1423 if (GET_CODE (operands
[0]) == MEM
)
1424 return rtx_equal_p (operands
[0], operands
[1])
1425 || (!reload_completed
&& rtx_equal_p (operands
[0], operands
[2]));
1430 /* Narrow logical operation CODE of memory operand MEMOP with immediate
1431 operand IMMOP to switch from SS to SI type instructions. */
1434 s390_narrow_logical_operator (enum rtx_code code
, rtx
*memop
, rtx
*immop
)
1436 int def
= code
== AND
? -1 : 0;
1440 gcc_assert (GET_CODE (*memop
) == MEM
);
1441 gcc_assert (!MEM_VOLATILE_P (*memop
));
1443 mask
= s390_extract_part (*immop
, QImode
, def
);
1444 part
= s390_single_part (*immop
, GET_MODE (*memop
), QImode
, def
);
1445 gcc_assert (part
>= 0);
1447 *memop
= adjust_address (*memop
, QImode
, part
);
1448 *immop
= gen_int_mode (mask
, QImode
);
1452 /* How to allocate a 'struct machine_function'. */
1454 static struct machine_function
*
1455 s390_init_machine_status (void)
1457 return ggc_alloc_cleared_machine_function ();
1460 /* Change optimizations to be performed, depending on the
1463 LEVEL is the optimization level specified; 2 if `-O2' is
1464 specified, 1 if `-O' is specified, and 0 if neither is specified.
1466 SIZE is nonzero if `-Os' is specified and zero otherwise. */
1469 optimization_options (int level ATTRIBUTE_UNUSED
, int size ATTRIBUTE_UNUSED
)
1471 /* ??? There are apparently still problems with -fcaller-saves. */
1472 flag_caller_saves
= 0;
1474 /* By default, always emit DWARF-2 unwind info. This allows debugging
1475 without maintaining a stack frame back-chain. */
1476 flag_asynchronous_unwind_tables
= 1;
1478 /* Use MVCLE instructions to decrease code size if requested. */
1480 target_flags
|= MASK_MVCLE
;
1483 /* Return true if ARG is the name of a processor. Set *TYPE and *FLAGS
1484 to the associated processor_type and processor_flags if so. */
1487 s390_handle_arch_option (const char *arg
,
1488 enum processor_type
*type
,
1493 const char *const name
; /* processor name or nickname. */
1494 const enum processor_type processor
;
1495 const int flags
; /* From enum processor_flags. */
1497 const processor_alias_table
[] =
1499 {"g5", PROCESSOR_9672_G5
, PF_IEEE_FLOAT
},
1500 {"g6", PROCESSOR_9672_G6
, PF_IEEE_FLOAT
},
1501 {"z900", PROCESSOR_2064_Z900
, PF_IEEE_FLOAT
| PF_ZARCH
},
1502 {"z990", PROCESSOR_2084_Z990
, PF_IEEE_FLOAT
| PF_ZARCH
1503 | PF_LONG_DISPLACEMENT
},
1504 {"z9-109", PROCESSOR_2094_Z9_109
, PF_IEEE_FLOAT
| PF_ZARCH
1505 | PF_LONG_DISPLACEMENT
| PF_EXTIMM
},
1506 {"z9-ec", PROCESSOR_2094_Z9_109
, PF_IEEE_FLOAT
| PF_ZARCH
1507 | PF_LONG_DISPLACEMENT
| PF_EXTIMM
| PF_DFP
},
1508 {"z10", PROCESSOR_2097_Z10
, PF_IEEE_FLOAT
| PF_ZARCH
1509 | PF_LONG_DISPLACEMENT
| PF_EXTIMM
| PF_DFP
| PF_Z10
},
1513 for (i
= 0; i
< ARRAY_SIZE (processor_alias_table
); i
++)
1514 if (strcmp (arg
, processor_alias_table
[i
].name
) == 0)
1516 *type
= processor_alias_table
[i
].processor
;
1517 *flags
= processor_alias_table
[i
].flags
;
1523 /* Implement TARGET_HANDLE_OPTION. */
1526 s390_handle_option (size_t code
, const char *arg
, int value ATTRIBUTE_UNUSED
)
1531 return s390_handle_arch_option (arg
, &s390_arch
, &s390_arch_flags
);
1533 case OPT_mstack_guard_
:
1534 if (sscanf (arg
, HOST_WIDE_INT_PRINT_DEC
, &s390_stack_guard
) != 1)
1536 if (exact_log2 (s390_stack_guard
) == -1)
1537 error ("stack guard value must be an exact power of 2");
1540 case OPT_mstack_size_
:
1541 if (sscanf (arg
, HOST_WIDE_INT_PRINT_DEC
, &s390_stack_size
) != 1)
1543 if (exact_log2 (s390_stack_size
) == -1)
1544 error ("stack size must be an exact power of 2");
1548 return s390_handle_arch_option (arg
, &s390_tune
, &s390_tune_flags
);
1550 case OPT_mwarn_framesize_
:
1551 return sscanf (arg
, HOST_WIDE_INT_PRINT_DEC
, &s390_warn_framesize
) == 1;
1559 override_options (void)
1561 /* Set up function hooks. */
1562 init_machine_status
= s390_init_machine_status
;
1564 /* Architecture mode defaults according to ABI. */
1565 if (!(target_flags_explicit
& MASK_ZARCH
))
1568 target_flags
|= MASK_ZARCH
;
1570 target_flags
&= ~MASK_ZARCH
;
1573 /* Determine processor architectural level. */
1574 if (!s390_arch_string
)
1576 s390_arch_string
= TARGET_ZARCH
? "z900" : "g5";
1577 s390_handle_arch_option (s390_arch_string
, &s390_arch
, &s390_arch_flags
);
1580 /* Determine processor to tune for. */
1581 if (s390_tune
== PROCESSOR_max
)
1583 s390_tune
= s390_arch
;
1584 s390_tune_flags
= s390_arch_flags
;
1587 /* Sanity checks. */
1588 if (TARGET_ZARCH
&& !TARGET_CPU_ZARCH
)
1589 error ("z/Architecture mode not supported on %s", s390_arch_string
);
1590 if (TARGET_64BIT
&& !TARGET_ZARCH
)
1591 error ("64-bit ABI not supported in ESA/390 mode");
1593 if (TARGET_HARD_DFP
&& !TARGET_DFP
)
1595 if (target_flags_explicit
& MASK_HARD_DFP
)
1597 if (!TARGET_CPU_DFP
)
1598 error ("Hardware decimal floating point instructions"
1599 " not available on %s", s390_arch_string
);
1601 error ("Hardware decimal floating point instructions"
1602 " not available in ESA/390 mode");
1605 target_flags
&= ~MASK_HARD_DFP
;
1608 if ((target_flags_explicit
& MASK_SOFT_FLOAT
) && TARGET_SOFT_FLOAT
)
1610 if ((target_flags_explicit
& MASK_HARD_DFP
) && TARGET_HARD_DFP
)
1611 error ("-mhard-dfp can't be used in conjunction with -msoft-float");
1613 target_flags
&= ~MASK_HARD_DFP
;
1616 /* Set processor cost function. */
1619 case PROCESSOR_2084_Z990
:
1620 s390_cost
= &z990_cost
;
1622 case PROCESSOR_2094_Z9_109
:
1623 s390_cost
= &z9_109_cost
;
1625 case PROCESSOR_2097_Z10
:
1626 s390_cost
= &z10_cost
;
1629 s390_cost
= &z900_cost
;
1632 if (TARGET_BACKCHAIN
&& TARGET_PACKED_STACK
&& TARGET_HARD_FLOAT
)
1633 error ("-mbackchain -mpacked-stack -mhard-float are not supported "
1636 if (s390_stack_size
)
1638 if (s390_stack_guard
>= s390_stack_size
)
1639 error ("stack size must be greater than the stack guard value");
1640 else if (s390_stack_size
> 1 << 16)
1641 error ("stack size must not be greater than 64k");
1643 else if (s390_stack_guard
)
1644 error ("-mstack-guard implies use of -mstack-size");
1646 #ifdef TARGET_DEFAULT_LONG_DOUBLE_128
1647 if (!(target_flags_explicit
& MASK_LONG_DOUBLE_128
))
1648 target_flags
|= MASK_LONG_DOUBLE_128
;
1651 if (s390_tune
== PROCESSOR_2097_Z10
)
1653 if (!PARAM_SET_P (PARAM_MAX_UNROLLED_INSNS
))
1654 set_param_value ("max-unrolled-insns", 100);
1655 if (!PARAM_SET_P (PARAM_MAX_UNROLL_TIMES
))
1656 set_param_value ("max-unroll-times", 32);
1657 if (!PARAM_SET_P (PARAM_MAX_COMPLETELY_PEELED_INSNS
))
1658 set_param_value ("max-completely-peeled-insns", 2000);
1659 if (!PARAM_SET_P (PARAM_MAX_COMPLETELY_PEEL_TIMES
))
1660 set_param_value ("max-completely-peel-times", 64);
1663 set_param_value ("max-pending-list-length", 256);
1664 /* values for loop prefetching */
1665 set_param_value ("l1-cache-line-size", 256);
1666 if (!PARAM_SET_P (PARAM_L1_CACHE_SIZE
))
1667 set_param_value ("l1-cache-size", 128);
1668 /* s390 has more than 2 levels and the size is much larger. Since
1669 we are always running virtualized assume that we only get a small
1670 part of the caches above l1. */
1671 if (!PARAM_SET_P (PARAM_L2_CACHE_SIZE
))
1672 set_param_value ("l2-cache-size", 1500);
1673 if (!PARAM_SET_P (PARAM_PREFETCH_MIN_INSN_TO_MEM_RATIO
))
1674 set_param_value ("prefetch-min-insn-to-mem-ratio", 2);
1675 if (!PARAM_SET_P (PARAM_SIMULTANEOUS_PREFETCHES
))
1676 set_param_value ("simultaneous-prefetches", 6);
1678 /* This cannot reside in optimization_options since HAVE_prefetch
1679 requires the arch flags to be evaluated already. Since prefetching
1680 is beneficial on s390, we enable it if available. */
1681 if (flag_prefetch_loop_arrays
< 0 && HAVE_prefetch
&& optimize
>= 3)
1682 flag_prefetch_loop_arrays
= 1;
1685 /* Map for smallest class containing reg regno. */
1687 const enum reg_class regclass_map
[FIRST_PSEUDO_REGISTER
] =
1688 { GENERAL_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1689 ADDR_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1690 ADDR_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1691 ADDR_REGS
, ADDR_REGS
, ADDR_REGS
, ADDR_REGS
,
1692 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1693 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1694 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1695 FP_REGS
, FP_REGS
, FP_REGS
, FP_REGS
,
1696 ADDR_REGS
, CC_REGS
, ADDR_REGS
, ADDR_REGS
,
1697 ACCESS_REGS
, ACCESS_REGS
1700 /* Return attribute type of insn. */
1702 static enum attr_type
1703 s390_safe_attr_type (rtx insn
)
1705 if (recog_memoized (insn
) >= 0)
1706 return get_attr_type (insn
);
1711 /* Return true if DISP is a valid short displacement. */
1714 s390_short_displacement (rtx disp
)
1716 /* No displacement is OK. */
1720 /* Without the long displacement facility we don't need to
1721 distingiush between long and short displacement. */
1722 if (!TARGET_LONG_DISPLACEMENT
)
1725 /* Integer displacement in range. */
1726 if (GET_CODE (disp
) == CONST_INT
)
1727 return INTVAL (disp
) >= 0 && INTVAL (disp
) < 4096;
1729 /* GOT offset is not OK, the GOT can be large. */
1730 if (GET_CODE (disp
) == CONST
1731 && GET_CODE (XEXP (disp
, 0)) == UNSPEC
1732 && (XINT (XEXP (disp
, 0), 1) == UNSPEC_GOT
1733 || XINT (XEXP (disp
, 0), 1) == UNSPEC_GOTNTPOFF
))
1736 /* All other symbolic constants are literal pool references,
1737 which are OK as the literal pool must be small. */
1738 if (GET_CODE (disp
) == CONST
)
1744 /* Decompose a RTL expression ADDR for a memory address into
1745 its components, returned in OUT.
1747 Returns false if ADDR is not a valid memory address, true
1748 otherwise. If OUT is NULL, don't return the components,
1749 but check for validity only.
1751 Note: Only addresses in canonical form are recognized.
1752 LEGITIMIZE_ADDRESS should convert non-canonical forms to the
1753 canonical form so that they will be recognized. */
1756 s390_decompose_address (rtx addr
, struct s390_address
*out
)
1758 HOST_WIDE_INT offset
= 0;
1759 rtx base
= NULL_RTX
;
1760 rtx indx
= NULL_RTX
;
1761 rtx disp
= NULL_RTX
;
1763 bool pointer
= false;
1764 bool base_ptr
= false;
1765 bool indx_ptr
= false;
1766 bool literal_pool
= false;
1768 /* We may need to substitute the literal pool base register into the address
1769 below. However, at this point we do not know which register is going to
1770 be used as base, so we substitute the arg pointer register. This is going
1771 to be treated as holding a pointer below -- it shouldn't be used for any
1773 rtx fake_pool_base
= gen_rtx_REG (Pmode
, ARG_POINTER_REGNUM
);
1775 /* Decompose address into base + index + displacement. */
1777 if (GET_CODE (addr
) == REG
|| GET_CODE (addr
) == UNSPEC
)
1780 else if (GET_CODE (addr
) == PLUS
)
1782 rtx op0
= XEXP (addr
, 0);
1783 rtx op1
= XEXP (addr
, 1);
1784 enum rtx_code code0
= GET_CODE (op0
);
1785 enum rtx_code code1
= GET_CODE (op1
);
1787 if (code0
== REG
|| code0
== UNSPEC
)
1789 if (code1
== REG
|| code1
== UNSPEC
)
1791 indx
= op0
; /* index + base */
1797 base
= op0
; /* base + displacement */
1802 else if (code0
== PLUS
)
1804 indx
= XEXP (op0
, 0); /* index + base + disp */
1805 base
= XEXP (op0
, 1);
1816 disp
= addr
; /* displacement */
1818 /* Extract integer part of displacement. */
1822 if (GET_CODE (disp
) == CONST_INT
)
1824 offset
= INTVAL (disp
);
1827 else if (GET_CODE (disp
) == CONST
1828 && GET_CODE (XEXP (disp
, 0)) == PLUS
1829 && GET_CODE (XEXP (XEXP (disp
, 0), 1)) == CONST_INT
)
1831 offset
= INTVAL (XEXP (XEXP (disp
, 0), 1));
1832 disp
= XEXP (XEXP (disp
, 0), 0);
1836 /* Strip off CONST here to avoid special case tests later. */
1837 if (disp
&& GET_CODE (disp
) == CONST
)
1838 disp
= XEXP (disp
, 0);
1840 /* We can convert literal pool addresses to
1841 displacements by basing them off the base register. */
1842 if (disp
&& GET_CODE (disp
) == SYMBOL_REF
&& CONSTANT_POOL_ADDRESS_P (disp
))
1844 /* Either base or index must be free to hold the base register. */
1846 base
= fake_pool_base
, literal_pool
= true;
1848 indx
= fake_pool_base
, literal_pool
= true;
1852 /* Mark up the displacement. */
1853 disp
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, disp
),
1854 UNSPEC_LTREL_OFFSET
);
1857 /* Validate base register. */
1860 if (GET_CODE (base
) == UNSPEC
)
1861 switch (XINT (base
, 1))
1865 disp
= gen_rtx_UNSPEC (Pmode
,
1866 gen_rtvec (1, XVECEXP (base
, 0, 0)),
1867 UNSPEC_LTREL_OFFSET
);
1871 base
= XVECEXP (base
, 0, 1);
1874 case UNSPEC_LTREL_BASE
:
1875 if (XVECLEN (base
, 0) == 1)
1876 base
= fake_pool_base
, literal_pool
= true;
1878 base
= XVECEXP (base
, 0, 1);
1886 || (GET_MODE (base
) != SImode
1887 && GET_MODE (base
) != Pmode
))
1890 if (REGNO (base
) == STACK_POINTER_REGNUM
1891 || REGNO (base
) == FRAME_POINTER_REGNUM
1892 || ((reload_completed
|| reload_in_progress
)
1893 && frame_pointer_needed
1894 && REGNO (base
) == HARD_FRAME_POINTER_REGNUM
)
1895 || REGNO (base
) == ARG_POINTER_REGNUM
1897 && REGNO (base
) == PIC_OFFSET_TABLE_REGNUM
))
1898 pointer
= base_ptr
= true;
1900 if ((reload_completed
|| reload_in_progress
)
1901 && base
== cfun
->machine
->base_reg
)
1902 pointer
= base_ptr
= literal_pool
= true;
1905 /* Validate index register. */
1908 if (GET_CODE (indx
) == UNSPEC
)
1909 switch (XINT (indx
, 1))
1913 disp
= gen_rtx_UNSPEC (Pmode
,
1914 gen_rtvec (1, XVECEXP (indx
, 0, 0)),
1915 UNSPEC_LTREL_OFFSET
);
1919 indx
= XVECEXP (indx
, 0, 1);
1922 case UNSPEC_LTREL_BASE
:
1923 if (XVECLEN (indx
, 0) == 1)
1924 indx
= fake_pool_base
, literal_pool
= true;
1926 indx
= XVECEXP (indx
, 0, 1);
1934 || (GET_MODE (indx
) != SImode
1935 && GET_MODE (indx
) != Pmode
))
1938 if (REGNO (indx
) == STACK_POINTER_REGNUM
1939 || REGNO (indx
) == FRAME_POINTER_REGNUM
1940 || ((reload_completed
|| reload_in_progress
)
1941 && frame_pointer_needed
1942 && REGNO (indx
) == HARD_FRAME_POINTER_REGNUM
)
1943 || REGNO (indx
) == ARG_POINTER_REGNUM
1945 && REGNO (indx
) == PIC_OFFSET_TABLE_REGNUM
))
1946 pointer
= indx_ptr
= true;
1948 if ((reload_completed
|| reload_in_progress
)
1949 && indx
== cfun
->machine
->base_reg
)
1950 pointer
= indx_ptr
= literal_pool
= true;
1953 /* Prefer to use pointer as base, not index. */
1954 if (base
&& indx
&& !base_ptr
1955 && (indx_ptr
|| (!REG_POINTER (base
) && REG_POINTER (indx
))))
1962 /* Validate displacement. */
1965 /* If virtual registers are involved, the displacement will change later
1966 anyway as the virtual registers get eliminated. This could make a
1967 valid displacement invalid, but it is more likely to make an invalid
1968 displacement valid, because we sometimes access the register save area
1969 via negative offsets to one of those registers.
1970 Thus we don't check the displacement for validity here. If after
1971 elimination the displacement turns out to be invalid after all,
1972 this is fixed up by reload in any case. */
1973 if (base
!= arg_pointer_rtx
1974 && indx
!= arg_pointer_rtx
1975 && base
!= return_address_pointer_rtx
1976 && indx
!= return_address_pointer_rtx
1977 && base
!= frame_pointer_rtx
1978 && indx
!= frame_pointer_rtx
1979 && base
!= virtual_stack_vars_rtx
1980 && indx
!= virtual_stack_vars_rtx
)
1981 if (!DISP_IN_RANGE (offset
))
1986 /* All the special cases are pointers. */
1989 /* In the small-PIC case, the linker converts @GOT
1990 and @GOTNTPOFF offsets to possible displacements. */
1991 if (GET_CODE (disp
) == UNSPEC
1992 && (XINT (disp
, 1) == UNSPEC_GOT
1993 || XINT (disp
, 1) == UNSPEC_GOTNTPOFF
)
1999 /* Accept pool label offsets. */
2000 else if (GET_CODE (disp
) == UNSPEC
2001 && XINT (disp
, 1) == UNSPEC_POOL_OFFSET
)
2004 /* Accept literal pool references. */
2005 else if (GET_CODE (disp
) == UNSPEC
2006 && XINT (disp
, 1) == UNSPEC_LTREL_OFFSET
)
2008 orig_disp
= gen_rtx_CONST (Pmode
, disp
);
2011 /* If we have an offset, make sure it does not
2012 exceed the size of the constant pool entry. */
2013 rtx sym
= XVECEXP (disp
, 0, 0);
2014 if (offset
>= GET_MODE_SIZE (get_pool_mode (sym
)))
2017 orig_disp
= plus_constant (orig_disp
, offset
);
2032 out
->disp
= orig_disp
;
2033 out
->pointer
= pointer
;
2034 out
->literal_pool
= literal_pool
;
2040 /* Decompose a RTL expression OP for a shift count into its components,
2041 and return the base register in BASE and the offset in OFFSET.
2043 Return true if OP is a valid shift count, false if not. */
2046 s390_decompose_shift_count (rtx op
, rtx
*base
, HOST_WIDE_INT
*offset
)
2048 HOST_WIDE_INT off
= 0;
2050 /* We can have an integer constant, an address register,
2051 or a sum of the two. */
2052 if (GET_CODE (op
) == CONST_INT
)
2057 if (op
&& GET_CODE (op
) == PLUS
&& GET_CODE (XEXP (op
, 1)) == CONST_INT
)
2059 off
= INTVAL (XEXP (op
, 1));
2062 while (op
&& GET_CODE (op
) == SUBREG
)
2063 op
= SUBREG_REG (op
);
2065 if (op
&& GET_CODE (op
) != REG
)
2077 /* Return true if CODE is a valid address without index. */
2080 s390_legitimate_address_without_index_p (rtx op
)
2082 struct s390_address addr
;
2084 if (!s390_decompose_address (XEXP (op
, 0), &addr
))
2093 /* Return true if ADDR is of kind symbol_ref or symbol_ref + const_int
2094 and return these parts in SYMREF and ADDEND. You can pass NULL in
2095 SYMREF and/or ADDEND if you are not interested in these values. */
2098 s390_symref_operand_p (rtx addr
, rtx
*symref
, HOST_WIDE_INT
*addend
)
2100 HOST_WIDE_INT tmpaddend
= 0;
2102 if (GET_CODE (addr
) == CONST
)
2103 addr
= XEXP (addr
, 0);
2105 if (GET_CODE (addr
) == PLUS
)
2107 if (GET_CODE (XEXP (addr
, 0)) == SYMBOL_REF
2108 && CONST_INT_P (XEXP (addr
, 1)))
2110 tmpaddend
= INTVAL (XEXP (addr
, 1));
2111 addr
= XEXP (addr
, 0);
2117 if (GET_CODE (addr
) != SYMBOL_REF
)
2123 *addend
= tmpaddend
;
2129 /* Return true if the address in OP is valid for constraint letter C
2130 if wrapped in a MEM rtx. Set LIT_POOL_OK to true if it literal
2131 pool MEMs should be accepted. Only the Q, R, S, T constraint
2132 letters are allowed for C. */
2135 s390_check_qrst_address (char c
, rtx op
, bool lit_pool_ok
)
2137 struct s390_address addr
;
2138 bool decomposed
= false;
2140 /* This check makes sure that no symbolic address (except literal
2141 pool references) are accepted by the R or T constraints. */
2142 if (s390_symref_operand_p (op
, NULL
, NULL
))
2146 if (!s390_decompose_address (op
, &addr
))
2148 if (!addr
.literal_pool
)
2155 case 'Q': /* no index short displacement */
2156 if (!decomposed
&& !s390_decompose_address (op
, &addr
))
2160 if (!s390_short_displacement (addr
.disp
))
2164 case 'R': /* with index short displacement */
2165 if (TARGET_LONG_DISPLACEMENT
)
2167 if (!decomposed
&& !s390_decompose_address (op
, &addr
))
2169 if (!s390_short_displacement (addr
.disp
))
2172 /* Any invalid address here will be fixed up by reload,
2173 so accept it for the most generic constraint. */
2176 case 'S': /* no index long displacement */
2177 if (!TARGET_LONG_DISPLACEMENT
)
2179 if (!decomposed
&& !s390_decompose_address (op
, &addr
))
2183 if (s390_short_displacement (addr
.disp
))
2187 case 'T': /* with index long displacement */
2188 if (!TARGET_LONG_DISPLACEMENT
)
2190 /* Any invalid address here will be fixed up by reload,
2191 so accept it for the most generic constraint. */
2192 if ((decomposed
|| s390_decompose_address (op
, &addr
))
2193 && s390_short_displacement (addr
.disp
))
2203 /* Evaluates constraint strings described by the regular expression
2204 ([A|B|Z](Q|R|S|T))|U|W|Y and returns 1 if OP is a valid operand for
2205 the constraint given in STR, or 0 else. */
2208 s390_mem_constraint (const char *str
, rtx op
)
2215 /* Check for offsettable variants of memory constraints. */
2216 if (!MEM_P (op
) || MEM_VOLATILE_P (op
))
2218 if ((reload_completed
|| reload_in_progress
)
2219 ? !offsettable_memref_p (op
) : !offsettable_nonstrict_memref_p (op
))
2221 return s390_check_qrst_address (str
[1], XEXP (op
, 0), true);
2223 /* Check for non-literal-pool variants of memory constraints. */
2226 return s390_check_qrst_address (str
[1], XEXP (op
, 0), false);
2231 if (GET_CODE (op
) != MEM
)
2233 return s390_check_qrst_address (c
, XEXP (op
, 0), true);
2235 return (s390_check_qrst_address ('Q', op
, true)
2236 || s390_check_qrst_address ('R', op
, true));
2238 return (s390_check_qrst_address ('S', op
, true)
2239 || s390_check_qrst_address ('T', op
, true));
2241 /* Simply check for the basic form of a shift count. Reload will
2242 take care of making sure we have a proper base register. */
2243 if (!s390_decompose_shift_count (op
, NULL
, NULL
))
2247 return s390_check_qrst_address (str
[1], op
, true);
2255 /* Evaluates constraint strings starting with letter O. Input
2256 parameter C is the second letter following the "O" in the constraint
2257 string. Returns 1 if VALUE meets the respective constraint and 0
2261 s390_O_constraint_str (const char c
, HOST_WIDE_INT value
)
2269 return trunc_int_for_mode (value
, SImode
) == value
;
2273 || s390_single_part (GEN_INT (value
), DImode
, SImode
, 0) == 1;
2276 return s390_single_part (GEN_INT (value
- 1), DImode
, SImode
, -1) == 1;
2284 /* Evaluates constraint strings starting with letter N. Parameter STR
2285 contains the letters following letter "N" in the constraint string.
2286 Returns true if VALUE matches the constraint. */
2289 s390_N_constraint_str (const char *str
, HOST_WIDE_INT value
)
2291 enum machine_mode mode
, part_mode
;
2293 int part
, part_goal
;
2299 part_goal
= str
[0] - '0';
2343 if (GET_MODE_SIZE (mode
) <= GET_MODE_SIZE (part_mode
))
2346 part
= s390_single_part (GEN_INT (value
), mode
, part_mode
, def
);
2349 if (part_goal
!= -1 && part_goal
!= part
)
2356 /* Returns true if the input parameter VALUE is a float zero. */
2359 s390_float_const_zero_p (rtx value
)
2361 return (GET_MODE_CLASS (GET_MODE (value
)) == MODE_FLOAT
2362 && value
== CONST0_RTX (GET_MODE (value
)));
2366 /* Compute a (partial) cost for rtx X. Return true if the complete
2367 cost has been computed, and false if subexpressions should be
2368 scanned. In either case, *TOTAL contains the cost result.
2369 CODE contains GET_CODE (x), OUTER_CODE contains the code
2370 of the superexpression of x. */
2373 s390_rtx_costs (rtx x
, int code
, int outer_code
, int *total
,
2374 bool speed ATTRIBUTE_UNUSED
)
2397 *total
= COSTS_N_INSNS (1);
2402 /* Check for multiply and add. */
2403 if ((GET_MODE (x
) == DFmode
|| GET_MODE (x
) == SFmode
)
2404 && GET_CODE (XEXP (x
, 0)) == MULT
2405 && TARGET_HARD_FLOAT
&& TARGET_FUSED_MADD
)
2407 /* This is the multiply and add case. */
2408 if (GET_MODE (x
) == DFmode
)
2409 *total
= s390_cost
->madbr
;
2411 *total
= s390_cost
->maebr
;
2412 *total
+= (rtx_cost (XEXP (XEXP (x
, 0), 0), MULT
, speed
)
2413 + rtx_cost (XEXP (XEXP (x
, 0), 1), MULT
, speed
)
2414 + rtx_cost (XEXP (x
, 1), (enum rtx_code
) code
, speed
));
2415 return true; /* Do not do an additional recursive descent. */
2417 *total
= COSTS_N_INSNS (1);
2421 switch (GET_MODE (x
))
2425 rtx left
= XEXP (x
, 0);
2426 rtx right
= XEXP (x
, 1);
2427 if (GET_CODE (right
) == CONST_INT
2428 && CONST_OK_FOR_K (INTVAL (right
)))
2429 *total
= s390_cost
->mhi
;
2430 else if (GET_CODE (left
) == SIGN_EXTEND
)
2431 *total
= s390_cost
->mh
;
2433 *total
= s390_cost
->ms
; /* msr, ms, msy */
2438 rtx left
= XEXP (x
, 0);
2439 rtx right
= XEXP (x
, 1);
2442 if (GET_CODE (right
) == CONST_INT
2443 && CONST_OK_FOR_K (INTVAL (right
)))
2444 *total
= s390_cost
->mghi
;
2445 else if (GET_CODE (left
) == SIGN_EXTEND
)
2446 *total
= s390_cost
->msgf
;
2448 *total
= s390_cost
->msg
; /* msgr, msg */
2450 else /* TARGET_31BIT */
2452 if (GET_CODE (left
) == SIGN_EXTEND
2453 && GET_CODE (right
) == SIGN_EXTEND
)
2454 /* mulsidi case: mr, m */
2455 *total
= s390_cost
->m
;
2456 else if (GET_CODE (left
) == ZERO_EXTEND
2457 && GET_CODE (right
) == ZERO_EXTEND
2458 && TARGET_CPU_ZARCH
)
2459 /* umulsidi case: ml, mlr */
2460 *total
= s390_cost
->ml
;
2462 /* Complex calculation is required. */
2463 *total
= COSTS_N_INSNS (40);
2469 *total
= s390_cost
->mult_df
;
2472 *total
= s390_cost
->mxbr
;
2481 if (GET_MODE (x
) == TImode
) /* 128 bit division */
2482 *total
= s390_cost
->dlgr
;
2483 else if (GET_MODE (x
) == DImode
)
2485 rtx right
= XEXP (x
, 1);
2486 if (GET_CODE (right
) == ZERO_EXTEND
) /* 64 by 32 bit division */
2487 *total
= s390_cost
->dlr
;
2488 else /* 64 by 64 bit division */
2489 *total
= s390_cost
->dlgr
;
2491 else if (GET_MODE (x
) == SImode
) /* 32 bit division */
2492 *total
= s390_cost
->dlr
;
2497 if (GET_MODE (x
) == DImode
)
2499 rtx right
= XEXP (x
, 1);
2500 if (GET_CODE (right
) == ZERO_EXTEND
) /* 64 by 32 bit division */
2502 *total
= s390_cost
->dsgfr
;
2504 *total
= s390_cost
->dr
;
2505 else /* 64 by 64 bit division */
2506 *total
= s390_cost
->dsgr
;
2508 else if (GET_MODE (x
) == SImode
) /* 32 bit division */
2509 *total
= s390_cost
->dlr
;
2510 else if (GET_MODE (x
) == SFmode
)
2512 *total
= s390_cost
->debr
;
2514 else if (GET_MODE (x
) == DFmode
)
2516 *total
= s390_cost
->ddbr
;
2518 else if (GET_MODE (x
) == TFmode
)
2520 *total
= s390_cost
->dxbr
;
2525 if (GET_MODE (x
) == SFmode
)
2526 *total
= s390_cost
->sqebr
;
2527 else if (GET_MODE (x
) == DFmode
)
2528 *total
= s390_cost
->sqdbr
;
2530 *total
= s390_cost
->sqxbr
;
2535 if (outer_code
== MULT
|| outer_code
== DIV
|| outer_code
== MOD
2536 || outer_code
== PLUS
|| outer_code
== MINUS
2537 || outer_code
== COMPARE
)
2542 *total
= COSTS_N_INSNS (1);
2543 if (GET_CODE (XEXP (x
, 0)) == AND
2544 && GET_CODE (XEXP (x
, 1)) == CONST_INT
2545 && GET_CODE (XEXP (XEXP (x
, 0), 1)) == CONST_INT
)
2547 rtx op0
= XEXP (XEXP (x
, 0), 0);
2548 rtx op1
= XEXP (XEXP (x
, 0), 1);
2549 rtx op2
= XEXP (x
, 1);
2551 if (memory_operand (op0
, GET_MODE (op0
))
2552 && s390_tm_ccmode (op1
, op2
, 0) != VOIDmode
)
2554 if (register_operand (op0
, GET_MODE (op0
))
2555 && s390_tm_ccmode (op1
, op2
, 1) != VOIDmode
)
2565 /* Return the cost of an address rtx ADDR. */
2568 s390_address_cost (rtx addr
, bool speed ATTRIBUTE_UNUSED
)
2570 struct s390_address ad
;
2571 if (!s390_decompose_address (addr
, &ad
))
2574 return ad
.indx
? COSTS_N_INSNS (1) + 1 : COSTS_N_INSNS (1);
2577 /* If OP is a SYMBOL_REF of a thread-local symbol, return its TLS mode,
2578 otherwise return 0. */
2581 tls_symbolic_operand (rtx op
)
2583 if (GET_CODE (op
) != SYMBOL_REF
)
2585 return SYMBOL_REF_TLS_MODEL (op
);
2588 /* Split DImode access register reference REG (on 64-bit) into its constituent
2589 low and high parts, and store them into LO and HI. Note that gen_lowpart/
2590 gen_highpart cannot be used as they assume all registers are word-sized,
2591 while our access registers have only half that size. */
2594 s390_split_access_reg (rtx reg
, rtx
*lo
, rtx
*hi
)
2596 gcc_assert (TARGET_64BIT
);
2597 gcc_assert (ACCESS_REG_P (reg
));
2598 gcc_assert (GET_MODE (reg
) == DImode
);
2599 gcc_assert (!(REGNO (reg
) & 1));
2601 *lo
= gen_rtx_REG (SImode
, REGNO (reg
) + 1);
2602 *hi
= gen_rtx_REG (SImode
, REGNO (reg
));
2605 /* Return true if OP contains a symbol reference */
2608 symbolic_reference_mentioned_p (rtx op
)
2613 if (GET_CODE (op
) == SYMBOL_REF
|| GET_CODE (op
) == LABEL_REF
)
2616 fmt
= GET_RTX_FORMAT (GET_CODE (op
));
2617 for (i
= GET_RTX_LENGTH (GET_CODE (op
)) - 1; i
>= 0; i
--)
2623 for (j
= XVECLEN (op
, i
) - 1; j
>= 0; j
--)
2624 if (symbolic_reference_mentioned_p (XVECEXP (op
, i
, j
)))
2628 else if (fmt
[i
] == 'e' && symbolic_reference_mentioned_p (XEXP (op
, i
)))
2635 /* Return true if OP contains a reference to a thread-local symbol. */
2638 tls_symbolic_reference_mentioned_p (rtx op
)
2643 if (GET_CODE (op
) == SYMBOL_REF
)
2644 return tls_symbolic_operand (op
);
2646 fmt
= GET_RTX_FORMAT (GET_CODE (op
));
2647 for (i
= GET_RTX_LENGTH (GET_CODE (op
)) - 1; i
>= 0; i
--)
2653 for (j
= XVECLEN (op
, i
) - 1; j
>= 0; j
--)
2654 if (tls_symbolic_reference_mentioned_p (XVECEXP (op
, i
, j
)))
2658 else if (fmt
[i
] == 'e' && tls_symbolic_reference_mentioned_p (XEXP (op
, i
)))
2666 /* Return true if OP is a legitimate general operand when
2667 generating PIC code. It is given that flag_pic is on
2668 and that OP satisfies CONSTANT_P or is a CONST_DOUBLE. */
2671 legitimate_pic_operand_p (rtx op
)
2673 /* Accept all non-symbolic constants. */
2674 if (!SYMBOLIC_CONST (op
))
2677 /* Reject everything else; must be handled
2678 via emit_symbolic_move. */
2682 /* Returns true if the constant value OP is a legitimate general operand.
2683 It is given that OP satisfies CONSTANT_P or is a CONST_DOUBLE. */
2686 legitimate_constant_p (rtx op
)
2688 /* Accept all non-symbolic constants. */
2689 if (!SYMBOLIC_CONST (op
))
2692 /* Accept immediate LARL operands. */
2693 if (TARGET_CPU_ZARCH
&& larl_operand (op
, VOIDmode
))
2696 /* Thread-local symbols are never legal constants. This is
2697 so that emit_call knows that computing such addresses
2698 might require a function call. */
2699 if (TLS_SYMBOLIC_CONST (op
))
2702 /* In the PIC case, symbolic constants must *not* be
2703 forced into the literal pool. We accept them here,
2704 so that they will be handled by emit_symbolic_move. */
2708 /* All remaining non-PIC symbolic constants are
2709 forced into the literal pool. */
2713 /* Determine if it's legal to put X into the constant pool. This
2714 is not possible if X contains the address of a symbol that is
2715 not constant (TLS) or not known at final link time (PIC). */
2718 s390_cannot_force_const_mem (rtx x
)
2720 switch (GET_CODE (x
))
2724 /* Accept all non-symbolic constants. */
2728 /* Labels are OK iff we are non-PIC. */
2729 return flag_pic
!= 0;
2732 /* 'Naked' TLS symbol references are never OK,
2733 non-TLS symbols are OK iff we are non-PIC. */
2734 if (tls_symbolic_operand (x
))
2737 return flag_pic
!= 0;
2740 return s390_cannot_force_const_mem (XEXP (x
, 0));
2743 return s390_cannot_force_const_mem (XEXP (x
, 0))
2744 || s390_cannot_force_const_mem (XEXP (x
, 1));
2747 switch (XINT (x
, 1))
2749 /* Only lt-relative or GOT-relative UNSPECs are OK. */
2750 case UNSPEC_LTREL_OFFSET
:
2758 case UNSPEC_GOTNTPOFF
:
2759 case UNSPEC_INDNTPOFF
:
2762 /* If the literal pool shares the code section, be put
2763 execute template placeholders into the pool as well. */
2765 return TARGET_CPU_ZARCH
;
2777 /* Returns true if the constant value OP is a legitimate general
2778 operand during and after reload. The difference to
2779 legitimate_constant_p is that this function will not accept
2780 a constant that would need to be forced to the literal pool
2781 before it can be used as operand. */
2784 legitimate_reload_constant_p (rtx op
)
2786 /* Accept la(y) operands. */
2787 if (GET_CODE (op
) == CONST_INT
2788 && DISP_IN_RANGE (INTVAL (op
)))
2791 /* Accept l(g)hi/l(g)fi operands. */
2792 if (GET_CODE (op
) == CONST_INT
2793 && (CONST_OK_FOR_K (INTVAL (op
)) || CONST_OK_FOR_Os (INTVAL (op
))))
2796 /* Accept lliXX operands. */
2798 && GET_CODE (op
) == CONST_INT
2799 && trunc_int_for_mode (INTVAL (op
), word_mode
) == INTVAL (op
)
2800 && s390_single_part (op
, word_mode
, HImode
, 0) >= 0)
2804 && GET_CODE (op
) == CONST_INT
2805 && trunc_int_for_mode (INTVAL (op
), word_mode
) == INTVAL (op
)
2806 && s390_single_part (op
, word_mode
, SImode
, 0) >= 0)
2809 /* Accept larl operands. */
2810 if (TARGET_CPU_ZARCH
2811 && larl_operand (op
, VOIDmode
))
2814 /* Accept double-word operands that can be split. */
2815 if (GET_CODE (op
) == CONST_INT
2816 && trunc_int_for_mode (INTVAL (op
), word_mode
) != INTVAL (op
))
2818 enum machine_mode dword_mode
= word_mode
== SImode
? DImode
: TImode
;
2819 rtx hi
= operand_subword (op
, 0, 0, dword_mode
);
2820 rtx lo
= operand_subword (op
, 1, 0, dword_mode
);
2821 return legitimate_reload_constant_p (hi
)
2822 && legitimate_reload_constant_p (lo
);
2825 /* Everything else cannot be handled without reload. */
2829 /* Given an rtx OP being reloaded into a reg required to be in class RCLASS,
2830 return the class of reg to actually use. */
2833 s390_preferred_reload_class (rtx op
, enum reg_class rclass
)
2835 switch (GET_CODE (op
))
2837 /* Constants we cannot reload must be forced into the
2842 if (legitimate_reload_constant_p (op
))
2847 /* If a symbolic constant or a PLUS is reloaded,
2848 it is most likely being used as an address, so
2849 prefer ADDR_REGS. If 'class' is not a superset
2850 of ADDR_REGS, e.g. FP_REGS, reject this reload. */
2855 if (reg_class_subset_p (ADDR_REGS
, rclass
))
2867 /* Return true if ADDR is SYMBOL_REF + addend with addend being a
2868 multiple of ALIGNMENT and the SYMBOL_REF being naturally
2872 s390_check_symref_alignment (rtx addr
, HOST_WIDE_INT alignment
)
2874 HOST_WIDE_INT addend
;
2877 if (!s390_symref_operand_p (addr
, &symref
, &addend
))
2880 return (!SYMBOL_REF_NOT_NATURALLY_ALIGNED_P (symref
)
2881 && !(addend
& (alignment
- 1)));
2884 /* ADDR is moved into REG using larl. If ADDR isn't a valid larl
2885 operand SCRATCH is used to reload the even part of the address and
2889 s390_reload_larl_operand (rtx reg
, rtx addr
, rtx scratch
)
2891 HOST_WIDE_INT addend
;
2894 if (!s390_symref_operand_p (addr
, &symref
, &addend
))
2898 /* Easy case. The addend is even so larl will do fine. */
2899 emit_move_insn (reg
, addr
);
2902 /* We can leave the scratch register untouched if the target
2903 register is a valid base register. */
2904 if (REGNO (reg
) < FIRST_PSEUDO_REGISTER
2905 && REGNO_REG_CLASS (REGNO (reg
)) == ADDR_REGS
)
2908 gcc_assert (REGNO (scratch
) < FIRST_PSEUDO_REGISTER
);
2909 gcc_assert (REGNO_REG_CLASS (REGNO (scratch
)) == ADDR_REGS
);
2912 emit_move_insn (scratch
,
2913 gen_rtx_CONST (Pmode
,
2914 gen_rtx_PLUS (Pmode
, symref
,
2915 GEN_INT (addend
- 1))));
2917 emit_move_insn (scratch
, symref
);
2919 /* Increment the address using la in order to avoid clobbering cc. */
2920 emit_move_insn (reg
, gen_rtx_PLUS (Pmode
, scratch
, const1_rtx
));
2924 /* Generate what is necessary to move between REG and MEM using
2925 SCRATCH. The direction is given by TOMEM. */
2928 s390_reload_symref_address (rtx reg
, rtx mem
, rtx scratch
, bool tomem
)
2930 /* Reload might have pulled a constant out of the literal pool.
2931 Force it back in. */
2932 if (CONST_INT_P (mem
) || GET_CODE (mem
) == CONST_DOUBLE
2933 || GET_CODE (mem
) == CONST
)
2934 mem
= force_const_mem (GET_MODE (reg
), mem
);
2936 gcc_assert (MEM_P (mem
));
2938 /* For a load from memory we can leave the scratch register
2939 untouched if the target register is a valid base register. */
2941 && REGNO (reg
) < FIRST_PSEUDO_REGISTER
2942 && REGNO_REG_CLASS (REGNO (reg
)) == ADDR_REGS
2943 && GET_MODE (reg
) == GET_MODE (scratch
))
2946 /* Load address into scratch register. Since we can't have a
2947 secondary reload for a secondary reload we have to cover the case
2948 where larl would need a secondary reload here as well. */
2949 s390_reload_larl_operand (scratch
, XEXP (mem
, 0), scratch
);
2951 /* Now we can use a standard load/store to do the move. */
2953 emit_move_insn (replace_equiv_address (mem
, scratch
), reg
);
2955 emit_move_insn (reg
, replace_equiv_address (mem
, scratch
));
2958 /* Inform reload about cases where moving X with a mode MODE to a register in
2959 RCLASS requires an extra scratch or immediate register. Return the class
2960 needed for the immediate register. */
2963 s390_secondary_reload (bool in_p
, rtx x
, reg_class_t rclass_i
,
2964 enum machine_mode mode
, secondary_reload_info
*sri
)
2966 enum reg_class rclass
= (enum reg_class
) rclass_i
;
2968 /* Intermediate register needed. */
2969 if (reg_classes_intersect_p (CC_REGS
, rclass
))
2970 return GENERAL_REGS
;
2974 /* On z10 several optimizer steps may generate larl operands with
2977 && s390_symref_operand_p (x
, NULL
, NULL
)
2979 && !s390_check_symref_alignment (x
, 2))
2980 sri
->icode
= ((mode
== DImode
) ? CODE_FOR_reloaddi_larl_odd_addend_z10
2981 : CODE_FOR_reloadsi_larl_odd_addend_z10
);
2983 /* On z10 we need a scratch register when moving QI, TI or floating
2984 point mode values from or to a memory location with a SYMBOL_REF
2985 or if the symref addend of a SI or DI move is not aligned to the
2986 width of the access. */
2988 && s390_symref_operand_p (XEXP (x
, 0), NULL
, NULL
)
2989 && (mode
== QImode
|| mode
== TImode
|| FLOAT_MODE_P (mode
)
2990 || (!TARGET_ZARCH
&& mode
== DImode
)
2991 || ((mode
== HImode
|| mode
== SImode
|| mode
== DImode
)
2992 && (!s390_check_symref_alignment (XEXP (x
, 0),
2993 GET_MODE_SIZE (mode
))))))
2995 #define __SECONDARY_RELOAD_CASE(M,m) \
2998 sri->icode = in_p ? CODE_FOR_reload##m##di_toreg_z10 : \
2999 CODE_FOR_reload##m##di_tomem_z10; \
3001 sri->icode = in_p ? CODE_FOR_reload##m##si_toreg_z10 : \
3002 CODE_FOR_reload##m##si_tomem_z10; \
3005 switch (GET_MODE (x
))
3007 __SECONDARY_RELOAD_CASE (QI
, qi
);
3008 __SECONDARY_RELOAD_CASE (HI
, hi
);
3009 __SECONDARY_RELOAD_CASE (SI
, si
);
3010 __SECONDARY_RELOAD_CASE (DI
, di
);
3011 __SECONDARY_RELOAD_CASE (TI
, ti
);
3012 __SECONDARY_RELOAD_CASE (SF
, sf
);
3013 __SECONDARY_RELOAD_CASE (DF
, df
);
3014 __SECONDARY_RELOAD_CASE (TF
, tf
);
3015 __SECONDARY_RELOAD_CASE (SD
, sd
);
3016 __SECONDARY_RELOAD_CASE (DD
, dd
);
3017 __SECONDARY_RELOAD_CASE (TD
, td
);
3022 #undef __SECONDARY_RELOAD_CASE
3026 /* We need a scratch register when loading a PLUS expression which
3027 is not a legitimate operand of the LOAD ADDRESS instruction. */
3028 if (in_p
&& s390_plus_operand (x
, mode
))
3029 sri
->icode
= (TARGET_64BIT
?
3030 CODE_FOR_reloaddi_plus
: CODE_FOR_reloadsi_plus
);
3032 /* Performing a multiword move from or to memory we have to make sure the
3033 second chunk in memory is addressable without causing a displacement
3034 overflow. If that would be the case we calculate the address in
3035 a scratch register. */
3037 && GET_CODE (XEXP (x
, 0)) == PLUS
3038 && GET_CODE (XEXP (XEXP (x
, 0), 1)) == CONST_INT
3039 && !DISP_IN_RANGE (INTVAL (XEXP (XEXP (x
, 0), 1))
3040 + GET_MODE_SIZE (mode
) - 1))
3042 /* For GENERAL_REGS a displacement overflow is no problem if occurring
3043 in a s_operand address since we may fallback to lm/stm. So we only
3044 have to care about overflows in the b+i+d case. */
3045 if ((reg_classes_intersect_p (GENERAL_REGS
, rclass
)
3046 && s390_class_max_nregs (GENERAL_REGS
, mode
) > 1
3047 && GET_CODE (XEXP (XEXP (x
, 0), 0)) == PLUS
)
3048 /* For FP_REGS no lm/stm is available so this check is triggered
3049 for displacement overflows in b+i+d and b+d like addresses. */
3050 || (reg_classes_intersect_p (FP_REGS
, rclass
)
3051 && s390_class_max_nregs (FP_REGS
, mode
) > 1))
3054 sri
->icode
= (TARGET_64BIT
?
3055 CODE_FOR_reloaddi_nonoffmem_in
:
3056 CODE_FOR_reloadsi_nonoffmem_in
);
3058 sri
->icode
= (TARGET_64BIT
?
3059 CODE_FOR_reloaddi_nonoffmem_out
:
3060 CODE_FOR_reloadsi_nonoffmem_out
);
3064 /* A scratch address register is needed when a symbolic constant is
3065 copied to r0 compiling with -fPIC. In other cases the target
3066 register might be used as temporary (see legitimize_pic_address). */
3067 if (in_p
&& SYMBOLIC_CONST (x
) && flag_pic
== 2 && rclass
!= ADDR_REGS
)
3068 sri
->icode
= (TARGET_64BIT
?
3069 CODE_FOR_reloaddi_PIC_addr
:
3070 CODE_FOR_reloadsi_PIC_addr
);
3072 /* Either scratch or no register needed. */
3076 /* Generate code to load SRC, which is PLUS that is not a
3077 legitimate operand for the LA instruction, into TARGET.
3078 SCRATCH may be used as scratch register. */
3081 s390_expand_plus_operand (rtx target
, rtx src
,
3085 struct s390_address ad
;
3087 /* src must be a PLUS; get its two operands. */
3088 gcc_assert (GET_CODE (src
) == PLUS
);
3089 gcc_assert (GET_MODE (src
) == Pmode
);
3091 /* Check if any of the two operands is already scheduled
3092 for replacement by reload. This can happen e.g. when
3093 float registers occur in an address. */
3094 sum1
= find_replacement (&XEXP (src
, 0));
3095 sum2
= find_replacement (&XEXP (src
, 1));
3096 src
= gen_rtx_PLUS (Pmode
, sum1
, sum2
);
3098 /* If the address is already strictly valid, there's nothing to do. */
3099 if (!s390_decompose_address (src
, &ad
)
3100 || (ad
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)))
3101 || (ad
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (ad
.indx
))))
3103 /* Otherwise, one of the operands cannot be an address register;
3104 we reload its value into the scratch register. */
3105 if (true_regnum (sum1
) < 1 || true_regnum (sum1
) > 15)
3107 emit_move_insn (scratch
, sum1
);
3110 if (true_regnum (sum2
) < 1 || true_regnum (sum2
) > 15)
3112 emit_move_insn (scratch
, sum2
);
3116 /* According to the way these invalid addresses are generated
3117 in reload.c, it should never happen (at least on s390) that
3118 *neither* of the PLUS components, after find_replacements
3119 was applied, is an address register. */
3120 if (sum1
== scratch
&& sum2
== scratch
)
3126 src
= gen_rtx_PLUS (Pmode
, sum1
, sum2
);
3129 /* Emit the LOAD ADDRESS pattern. Note that reload of PLUS
3130 is only ever performed on addresses, so we can mark the
3131 sum as legitimate for LA in any case. */
3132 s390_load_address (target
, src
);
3136 /* Return true if ADDR is a valid memory address.
3137 STRICT specifies whether strict register checking applies. */
3140 s390_legitimate_address_p (enum machine_mode mode
, rtx addr
, bool strict
)
3142 struct s390_address ad
;
3145 && larl_operand (addr
, VOIDmode
)
3146 && (mode
== VOIDmode
3147 || s390_check_symref_alignment (addr
, GET_MODE_SIZE (mode
))))
3150 if (!s390_decompose_address (addr
, &ad
))
3155 if (ad
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)))
3158 if (ad
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (ad
.indx
)))
3164 && !(REGNO (ad
.base
) >= FIRST_PSEUDO_REGISTER
3165 || REGNO_REG_CLASS (REGNO (ad
.base
)) == ADDR_REGS
))
3169 && !(REGNO (ad
.indx
) >= FIRST_PSEUDO_REGISTER
3170 || REGNO_REG_CLASS (REGNO (ad
.indx
)) == ADDR_REGS
))
3176 /* Return true if OP is a valid operand for the LA instruction.
3177 In 31-bit, we need to prove that the result is used as an
3178 address, as LA performs only a 31-bit addition. */
3181 legitimate_la_operand_p (rtx op
)
3183 struct s390_address addr
;
3184 if (!s390_decompose_address (op
, &addr
))
3187 return (TARGET_64BIT
|| addr
.pointer
);
3190 /* Return true if it is valid *and* preferable to use LA to
3191 compute the sum of OP1 and OP2. */
3194 preferred_la_operand_p (rtx op1
, rtx op2
)
3196 struct s390_address addr
;
3198 if (op2
!= const0_rtx
)
3199 op1
= gen_rtx_PLUS (Pmode
, op1
, op2
);
3201 if (!s390_decompose_address (op1
, &addr
))
3203 if (addr
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (addr
.base
)))
3205 if (addr
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (addr
.indx
)))
3208 if (!TARGET_64BIT
&& !addr
.pointer
)
3214 if ((addr
.base
&& REG_P (addr
.base
) && REG_POINTER (addr
.base
))
3215 || (addr
.indx
&& REG_P (addr
.indx
) && REG_POINTER (addr
.indx
)))
3221 /* Emit a forced load-address operation to load SRC into DST.
3222 This will use the LOAD ADDRESS instruction even in situations
3223 where legitimate_la_operand_p (SRC) returns false. */
3226 s390_load_address (rtx dst
, rtx src
)
3229 emit_move_insn (dst
, src
);
3231 emit_insn (gen_force_la_31 (dst
, src
));
3234 /* Return a legitimate reference for ORIG (an address) using the
3235 register REG. If REG is 0, a new pseudo is generated.
3237 There are two types of references that must be handled:
3239 1. Global data references must load the address from the GOT, via
3240 the PIC reg. An insn is emitted to do this load, and the reg is
3243 2. Static data references, constant pool addresses, and code labels
3244 compute the address as an offset from the GOT, whose base is in
3245 the PIC reg. Static data objects have SYMBOL_FLAG_LOCAL set to
3246 differentiate them from global data objects. The returned
3247 address is the PIC reg + an unspec constant.
3249 TARGET_LEGITIMIZE_ADDRESS_P rejects symbolic references unless the PIC
3250 reg also appears in the address. */
3253 legitimize_pic_address (rtx orig
, rtx reg
)
3259 gcc_assert (!TLS_SYMBOLIC_CONST (addr
));
3261 if (GET_CODE (addr
) == LABEL_REF
3262 || (GET_CODE (addr
) == SYMBOL_REF
&& SYMBOL_REF_LOCAL_P (addr
)))
3264 /* This is a local symbol. */
3265 if (TARGET_CPU_ZARCH
&& larl_operand (addr
, VOIDmode
))
3267 /* Access local symbols PC-relative via LARL.
3268 This is the same as in the non-PIC case, so it is
3269 handled automatically ... */
3273 /* Access local symbols relative to the GOT. */
3275 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3277 if (reload_in_progress
|| reload_completed
)
3278 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3280 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTOFF
);
3281 addr
= gen_rtx_CONST (Pmode
, addr
);
3282 addr
= force_const_mem (Pmode
, addr
);
3283 emit_move_insn (temp
, addr
);
3285 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3288 s390_load_address (reg
, new_rtx
);
3293 else if (GET_CODE (addr
) == SYMBOL_REF
)
3296 reg
= gen_reg_rtx (Pmode
);
3300 /* Assume GOT offset < 4k. This is handled the same way
3301 in both 31- and 64-bit code (@GOT). */
3303 if (reload_in_progress
|| reload_completed
)
3304 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3306 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOT
);
3307 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3308 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, new_rtx
);
3309 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3310 emit_move_insn (reg
, new_rtx
);
3313 else if (TARGET_CPU_ZARCH
)
3315 /* If the GOT offset might be >= 4k, we determine the position
3316 of the GOT entry via a PC-relative LARL (@GOTENT). */
3318 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3320 gcc_assert (REGNO (temp
) >= FIRST_PSEUDO_REGISTER
3321 || REGNO_REG_CLASS (REGNO (temp
)) == ADDR_REGS
);
3323 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTENT
);
3324 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3325 emit_move_insn (temp
, new_rtx
);
3327 new_rtx
= gen_const_mem (Pmode
, temp
);
3328 emit_move_insn (reg
, new_rtx
);
3333 /* If the GOT offset might be >= 4k, we have to load it
3334 from the literal pool (@GOT). */
3336 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3338 gcc_assert (REGNO (temp
) >= FIRST_PSEUDO_REGISTER
3339 || REGNO_REG_CLASS (REGNO (temp
)) == ADDR_REGS
);
3341 if (reload_in_progress
|| reload_completed
)
3342 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3344 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOT
);
3345 addr
= gen_rtx_CONST (Pmode
, addr
);
3346 addr
= force_const_mem (Pmode
, addr
);
3347 emit_move_insn (temp
, addr
);
3349 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3350 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3351 emit_move_insn (reg
, new_rtx
);
3357 if (GET_CODE (addr
) == CONST
)
3359 addr
= XEXP (addr
, 0);
3360 if (GET_CODE (addr
) == UNSPEC
)
3362 gcc_assert (XVECLEN (addr
, 0) == 1);
3363 switch (XINT (addr
, 1))
3365 /* If someone moved a GOT-relative UNSPEC
3366 out of the literal pool, force them back in. */
3369 new_rtx
= force_const_mem (Pmode
, orig
);
3372 /* @GOT is OK as is if small. */
3375 new_rtx
= force_const_mem (Pmode
, orig
);
3378 /* @GOTENT is OK as is. */
3382 /* @PLT is OK as is on 64-bit, must be converted to
3383 GOT-relative @PLTOFF on 31-bit. */
3385 if (!TARGET_CPU_ZARCH
)
3387 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3389 if (reload_in_progress
|| reload_completed
)
3390 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3392 addr
= XVECEXP (addr
, 0, 0);
3393 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
),
3395 addr
= gen_rtx_CONST (Pmode
, addr
);
3396 addr
= force_const_mem (Pmode
, addr
);
3397 emit_move_insn (temp
, addr
);
3399 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3402 s390_load_address (reg
, new_rtx
);
3408 /* Everything else cannot happen. */
3414 gcc_assert (GET_CODE (addr
) == PLUS
);
3416 if (GET_CODE (addr
) == PLUS
)
3418 rtx op0
= XEXP (addr
, 0), op1
= XEXP (addr
, 1);
3420 gcc_assert (!TLS_SYMBOLIC_CONST (op0
));
3421 gcc_assert (!TLS_SYMBOLIC_CONST (op1
));
3423 /* Check first to see if this is a constant offset
3424 from a local symbol reference. */
3425 if ((GET_CODE (op0
) == LABEL_REF
3426 || (GET_CODE (op0
) == SYMBOL_REF
&& SYMBOL_REF_LOCAL_P (op0
)))
3427 && GET_CODE (op1
) == CONST_INT
)
3429 if (TARGET_CPU_ZARCH
3430 && larl_operand (op0
, VOIDmode
)
3431 && INTVAL (op1
) < (HOST_WIDE_INT
)1 << 31
3432 && INTVAL (op1
) >= -((HOST_WIDE_INT
)1 << 31))
3434 if (INTVAL (op1
) & 1)
3436 /* LARL can't handle odd offsets, so emit a
3437 pair of LARL and LA. */
3438 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3440 if (!DISP_IN_RANGE (INTVAL (op1
)))
3442 HOST_WIDE_INT even
= INTVAL (op1
) - 1;
3443 op0
= gen_rtx_PLUS (Pmode
, op0
, GEN_INT (even
));
3444 op0
= gen_rtx_CONST (Pmode
, op0
);
3448 emit_move_insn (temp
, op0
);
3449 new_rtx
= gen_rtx_PLUS (Pmode
, temp
, op1
);
3453 s390_load_address (reg
, new_rtx
);
3459 /* If the offset is even, we can just use LARL.
3460 This will happen automatically. */
3465 /* Access local symbols relative to the GOT. */
3467 rtx temp
= reg
? reg
: gen_reg_rtx (Pmode
);
3469 if (reload_in_progress
|| reload_completed
)
3470 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3472 addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, op0
),
3474 addr
= gen_rtx_PLUS (Pmode
, addr
, op1
);
3475 addr
= gen_rtx_CONST (Pmode
, addr
);
3476 addr
= force_const_mem (Pmode
, addr
);
3477 emit_move_insn (temp
, addr
);
3479 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3482 s390_load_address (reg
, new_rtx
);
3488 /* Now, check whether it is a GOT relative symbol plus offset
3489 that was pulled out of the literal pool. Force it back in. */
3491 else if (GET_CODE (op0
) == UNSPEC
3492 && GET_CODE (op1
) == CONST_INT
3493 && XINT (op0
, 1) == UNSPEC_GOTOFF
)
3495 gcc_assert (XVECLEN (op0
, 0) == 1);
3497 new_rtx
= force_const_mem (Pmode
, orig
);
3500 /* Otherwise, compute the sum. */
3503 base
= legitimize_pic_address (XEXP (addr
, 0), reg
);
3504 new_rtx
= legitimize_pic_address (XEXP (addr
, 1),
3505 base
== reg
? NULL_RTX
: reg
);
3506 if (GET_CODE (new_rtx
) == CONST_INT
)
3507 new_rtx
= plus_constant (base
, INTVAL (new_rtx
));
3510 if (GET_CODE (new_rtx
) == PLUS
&& CONSTANT_P (XEXP (new_rtx
, 1)))
3512 base
= gen_rtx_PLUS (Pmode
, base
, XEXP (new_rtx
, 0));
3513 new_rtx
= XEXP (new_rtx
, 1);
3515 new_rtx
= gen_rtx_PLUS (Pmode
, base
, new_rtx
);
3518 if (GET_CODE (new_rtx
) == CONST
)
3519 new_rtx
= XEXP (new_rtx
, 0);
3520 new_rtx
= force_operand (new_rtx
, 0);
3527 /* Load the thread pointer into a register. */
3530 s390_get_thread_pointer (void)
3532 rtx tp
= gen_reg_rtx (Pmode
);
3534 emit_move_insn (tp
, gen_rtx_REG (Pmode
, TP_REGNUM
));
3535 mark_reg_pointer (tp
, BITS_PER_WORD
);
3540 /* Emit a tls call insn. The call target is the SYMBOL_REF stored
3541 in s390_tls_symbol which always refers to __tls_get_offset.
3542 The returned offset is written to RESULT_REG and an USE rtx is
3543 generated for TLS_CALL. */
3545 static GTY(()) rtx s390_tls_symbol
;
3548 s390_emit_tls_call_insn (rtx result_reg
, rtx tls_call
)
3552 gcc_assert (flag_pic
);
3554 if (!s390_tls_symbol
)
3555 s390_tls_symbol
= gen_rtx_SYMBOL_REF (Pmode
, "__tls_get_offset");
3557 insn
= s390_emit_call (s390_tls_symbol
, tls_call
, result_reg
,
3558 gen_rtx_REG (Pmode
, RETURN_REGNUM
));
3560 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), result_reg
);
3561 RTL_CONST_CALL_P (insn
) = 1;
3564 /* ADDR contains a thread-local SYMBOL_REF. Generate code to compute
3565 this (thread-local) address. REG may be used as temporary. */
3568 legitimize_tls_address (rtx addr
, rtx reg
)
3570 rtx new_rtx
, tls_call
, temp
, base
, r2
, insn
;
3572 if (GET_CODE (addr
) == SYMBOL_REF
)
3573 switch (tls_symbolic_operand (addr
))
3575 case TLS_MODEL_GLOBAL_DYNAMIC
:
3577 r2
= gen_rtx_REG (Pmode
, 2);
3578 tls_call
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_TLSGD
);
3579 new_rtx
= gen_rtx_CONST (Pmode
, tls_call
);
3580 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3581 emit_move_insn (r2
, new_rtx
);
3582 s390_emit_tls_call_insn (r2
, tls_call
);
3583 insn
= get_insns ();
3586 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_NTPOFF
);
3587 temp
= gen_reg_rtx (Pmode
);
3588 emit_libcall_block (insn
, temp
, r2
, new_rtx
);
3590 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3593 s390_load_address (reg
, new_rtx
);
3598 case TLS_MODEL_LOCAL_DYNAMIC
:
3600 r2
= gen_rtx_REG (Pmode
, 2);
3601 tls_call
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, const0_rtx
), UNSPEC_TLSLDM
);
3602 new_rtx
= gen_rtx_CONST (Pmode
, tls_call
);
3603 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3604 emit_move_insn (r2
, new_rtx
);
3605 s390_emit_tls_call_insn (r2
, tls_call
);
3606 insn
= get_insns ();
3609 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, const0_rtx
), UNSPEC_TLSLDM_NTPOFF
);
3610 temp
= gen_reg_rtx (Pmode
);
3611 emit_libcall_block (insn
, temp
, r2
, new_rtx
);
3613 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3614 base
= gen_reg_rtx (Pmode
);
3615 s390_load_address (base
, new_rtx
);
3617 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_DTPOFF
);
3618 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3619 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3620 temp
= gen_reg_rtx (Pmode
);
3621 emit_move_insn (temp
, new_rtx
);
3623 new_rtx
= gen_rtx_PLUS (Pmode
, base
, temp
);
3626 s390_load_address (reg
, new_rtx
);
3631 case TLS_MODEL_INITIAL_EXEC
:
3634 /* Assume GOT offset < 4k. This is handled the same way
3635 in both 31- and 64-bit code. */
3637 if (reload_in_progress
|| reload_completed
)
3638 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3640 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTNTPOFF
);
3641 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3642 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, new_rtx
);
3643 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3644 temp
= gen_reg_rtx (Pmode
);
3645 emit_move_insn (temp
, new_rtx
);
3647 else if (TARGET_CPU_ZARCH
)
3649 /* If the GOT offset might be >= 4k, we determine the position
3650 of the GOT entry via a PC-relative LARL. */
3652 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_INDNTPOFF
);
3653 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3654 temp
= gen_reg_rtx (Pmode
);
3655 emit_move_insn (temp
, new_rtx
);
3657 new_rtx
= gen_const_mem (Pmode
, temp
);
3658 temp
= gen_reg_rtx (Pmode
);
3659 emit_move_insn (temp
, new_rtx
);
3663 /* If the GOT offset might be >= 4k, we have to load it
3664 from the literal pool. */
3666 if (reload_in_progress
|| reload_completed
)
3667 df_set_regs_ever_live (PIC_OFFSET_TABLE_REGNUM
, true);
3669 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_GOTNTPOFF
);
3670 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3671 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3672 temp
= gen_reg_rtx (Pmode
);
3673 emit_move_insn (temp
, new_rtx
);
3675 new_rtx
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, temp
);
3676 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3678 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, new_rtx
, addr
), UNSPEC_TLS_LOAD
);
3679 temp
= gen_reg_rtx (Pmode
);
3680 emit_insn (gen_rtx_SET (Pmode
, temp
, new_rtx
));
3684 /* In position-dependent code, load the absolute address of
3685 the GOT entry from the literal pool. */
3687 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_INDNTPOFF
);
3688 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3689 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3690 temp
= gen_reg_rtx (Pmode
);
3691 emit_move_insn (temp
, new_rtx
);
3694 new_rtx
= gen_const_mem (Pmode
, new_rtx
);
3695 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, new_rtx
, addr
), UNSPEC_TLS_LOAD
);
3696 temp
= gen_reg_rtx (Pmode
);
3697 emit_insn (gen_rtx_SET (Pmode
, temp
, new_rtx
));
3700 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3703 s390_load_address (reg
, new_rtx
);
3708 case TLS_MODEL_LOCAL_EXEC
:
3709 new_rtx
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, addr
), UNSPEC_NTPOFF
);
3710 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3711 new_rtx
= force_const_mem (Pmode
, new_rtx
);
3712 temp
= gen_reg_rtx (Pmode
);
3713 emit_move_insn (temp
, new_rtx
);
3715 new_rtx
= gen_rtx_PLUS (Pmode
, s390_get_thread_pointer (), temp
);
3718 s390_load_address (reg
, new_rtx
);
3727 else if (GET_CODE (addr
) == CONST
&& GET_CODE (XEXP (addr
, 0)) == UNSPEC
)
3729 switch (XINT (XEXP (addr
, 0), 1))
3731 case UNSPEC_INDNTPOFF
:
3732 gcc_assert (TARGET_CPU_ZARCH
);
3741 else if (GET_CODE (addr
) == CONST
&& GET_CODE (XEXP (addr
, 0)) == PLUS
3742 && GET_CODE (XEXP (XEXP (addr
, 0), 1)) == CONST_INT
)
3744 new_rtx
= XEXP (XEXP (addr
, 0), 0);
3745 if (GET_CODE (new_rtx
) != SYMBOL_REF
)
3746 new_rtx
= gen_rtx_CONST (Pmode
, new_rtx
);
3748 new_rtx
= legitimize_tls_address (new_rtx
, reg
);
3749 new_rtx
= plus_constant (new_rtx
, INTVAL (XEXP (XEXP (addr
, 0), 1)));
3750 new_rtx
= force_operand (new_rtx
, 0);
3754 gcc_unreachable (); /* for now ... */
3759 /* Emit insns making the address in operands[1] valid for a standard
3760 move to operands[0]. operands[1] is replaced by an address which
3761 should be used instead of the former RTX to emit the move
3765 emit_symbolic_move (rtx
*operands
)
3767 rtx temp
= !can_create_pseudo_p () ? operands
[0] : gen_reg_rtx (Pmode
);
3769 if (GET_CODE (operands
[0]) == MEM
)
3770 operands
[1] = force_reg (Pmode
, operands
[1]);
3771 else if (TLS_SYMBOLIC_CONST (operands
[1]))
3772 operands
[1] = legitimize_tls_address (operands
[1], temp
);
3774 operands
[1] = legitimize_pic_address (operands
[1], temp
);
3777 /* Try machine-dependent ways of modifying an illegitimate address X
3778 to be legitimate. If we find one, return the new, valid address.
3780 OLDX is the address as it was before break_out_memory_refs was called.
3781 In some cases it is useful to look at this to decide what needs to be done.
3783 MODE is the mode of the operand pointed to by X.
3785 When -fpic is used, special handling is needed for symbolic references.
3786 See comments by legitimize_pic_address for details. */
3789 s390_legitimize_address (rtx x
, rtx oldx ATTRIBUTE_UNUSED
,
3790 enum machine_mode mode ATTRIBUTE_UNUSED
)
3792 rtx constant_term
= const0_rtx
;
3794 if (TLS_SYMBOLIC_CONST (x
))
3796 x
= legitimize_tls_address (x
, 0);
3798 if (s390_legitimate_address_p (mode
, x
, FALSE
))
3801 else if (GET_CODE (x
) == PLUS
3802 && (TLS_SYMBOLIC_CONST (XEXP (x
, 0))
3803 || TLS_SYMBOLIC_CONST (XEXP (x
, 1))))
3809 if (SYMBOLIC_CONST (x
)
3810 || (GET_CODE (x
) == PLUS
3811 && (SYMBOLIC_CONST (XEXP (x
, 0))
3812 || SYMBOLIC_CONST (XEXP (x
, 1)))))
3813 x
= legitimize_pic_address (x
, 0);
3815 if (s390_legitimate_address_p (mode
, x
, FALSE
))
3819 x
= eliminate_constant_term (x
, &constant_term
);
3821 /* Optimize loading of large displacements by splitting them
3822 into the multiple of 4K and the rest; this allows the
3823 former to be CSE'd if possible.
3825 Don't do this if the displacement is added to a register
3826 pointing into the stack frame, as the offsets will
3827 change later anyway. */
3829 if (GET_CODE (constant_term
) == CONST_INT
3830 && !TARGET_LONG_DISPLACEMENT
3831 && !DISP_IN_RANGE (INTVAL (constant_term
))
3832 && !(REG_P (x
) && REGNO_PTR_FRAME_P (REGNO (x
))))
3834 HOST_WIDE_INT lower
= INTVAL (constant_term
) & 0xfff;
3835 HOST_WIDE_INT upper
= INTVAL (constant_term
) ^ lower
;
3837 rtx temp
= gen_reg_rtx (Pmode
);
3838 rtx val
= force_operand (GEN_INT (upper
), temp
);
3840 emit_move_insn (temp
, val
);
3842 x
= gen_rtx_PLUS (Pmode
, x
, temp
);
3843 constant_term
= GEN_INT (lower
);
3846 if (GET_CODE (x
) == PLUS
)
3848 if (GET_CODE (XEXP (x
, 0)) == REG
)
3850 rtx temp
= gen_reg_rtx (Pmode
);
3851 rtx val
= force_operand (XEXP (x
, 1), temp
);
3853 emit_move_insn (temp
, val
);
3855 x
= gen_rtx_PLUS (Pmode
, XEXP (x
, 0), temp
);
3858 else if (GET_CODE (XEXP (x
, 1)) == REG
)
3860 rtx temp
= gen_reg_rtx (Pmode
);
3861 rtx val
= force_operand (XEXP (x
, 0), temp
);
3863 emit_move_insn (temp
, val
);
3865 x
= gen_rtx_PLUS (Pmode
, temp
, XEXP (x
, 1));
3869 if (constant_term
!= const0_rtx
)
3870 x
= gen_rtx_PLUS (Pmode
, x
, constant_term
);
3875 /* Try a machine-dependent way of reloading an illegitimate address AD
3876 operand. If we find one, push the reload and and return the new address.
3878 MODE is the mode of the enclosing MEM. OPNUM is the operand number
3879 and TYPE is the reload type of the current reload. */
3882 legitimize_reload_address (rtx ad
, enum machine_mode mode ATTRIBUTE_UNUSED
,
3883 int opnum
, int type
)
3885 if (!optimize
|| TARGET_LONG_DISPLACEMENT
)
3888 if (GET_CODE (ad
) == PLUS
)
3890 rtx tem
= simplify_binary_operation (PLUS
, Pmode
,
3891 XEXP (ad
, 0), XEXP (ad
, 1));
3896 if (GET_CODE (ad
) == PLUS
3897 && GET_CODE (XEXP (ad
, 0)) == REG
3898 && GET_CODE (XEXP (ad
, 1)) == CONST_INT
3899 && !DISP_IN_RANGE (INTVAL (XEXP (ad
, 1))))
3901 HOST_WIDE_INT lower
= INTVAL (XEXP (ad
, 1)) & 0xfff;
3902 HOST_WIDE_INT upper
= INTVAL (XEXP (ad
, 1)) ^ lower
;
3903 rtx cst
, tem
, new_rtx
;
3905 cst
= GEN_INT (upper
);
3906 if (!legitimate_reload_constant_p (cst
))
3907 cst
= force_const_mem (Pmode
, cst
);
3909 tem
= gen_rtx_PLUS (Pmode
, XEXP (ad
, 0), cst
);
3910 new_rtx
= gen_rtx_PLUS (Pmode
, tem
, GEN_INT (lower
));
3912 push_reload (XEXP (tem
, 1), 0, &XEXP (tem
, 1), 0,
3913 BASE_REG_CLASS
, Pmode
, VOIDmode
, 0, 0,
3914 opnum
, (enum reload_type
) type
);
3921 /* Emit code to move LEN bytes from DST to SRC. */
3924 s390_expand_movmem (rtx dst
, rtx src
, rtx len
)
3926 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) >= 0 && INTVAL (len
) <= 256)
3928 if (INTVAL (len
) > 0)
3929 emit_insn (gen_movmem_short (dst
, src
, GEN_INT (INTVAL (len
) - 1)));
3932 else if (TARGET_MVCLE
)
3934 emit_insn (gen_movmem_long (dst
, src
, convert_to_mode (Pmode
, len
, 1)));
3939 rtx dst_addr
, src_addr
, count
, blocks
, temp
;
3940 rtx loop_start_label
= gen_label_rtx ();
3941 rtx loop_end_label
= gen_label_rtx ();
3942 rtx end_label
= gen_label_rtx ();
3943 enum machine_mode mode
;
3945 mode
= GET_MODE (len
);
3946 if (mode
== VOIDmode
)
3949 dst_addr
= gen_reg_rtx (Pmode
);
3950 src_addr
= gen_reg_rtx (Pmode
);
3951 count
= gen_reg_rtx (mode
);
3952 blocks
= gen_reg_rtx (mode
);
3954 convert_move (count
, len
, 1);
3955 emit_cmp_and_jump_insns (count
, const0_rtx
,
3956 EQ
, NULL_RTX
, mode
, 1, end_label
);
3958 emit_move_insn (dst_addr
, force_operand (XEXP (dst
, 0), NULL_RTX
));
3959 emit_move_insn (src_addr
, force_operand (XEXP (src
, 0), NULL_RTX
));
3960 dst
= change_address (dst
, VOIDmode
, dst_addr
);
3961 src
= change_address (src
, VOIDmode
, src_addr
);
3963 temp
= expand_binop (mode
, add_optab
, count
, constm1_rtx
, count
, 1,
3966 emit_move_insn (count
, temp
);
3968 temp
= expand_binop (mode
, lshr_optab
, count
, GEN_INT (8), blocks
, 1,
3971 emit_move_insn (blocks
, temp
);
3973 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
3974 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
3976 emit_label (loop_start_label
);
3979 && (GET_CODE (len
) != CONST_INT
|| INTVAL (len
) > 768))
3983 /* Issue a read prefetch for the +3 cache line. */
3984 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, src_addr
, GEN_INT (768)),
3985 const0_rtx
, const0_rtx
);
3986 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
3987 emit_insn (prefetch
);
3989 /* Issue a write prefetch for the +3 cache line. */
3990 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, dst_addr
, GEN_INT (768)),
3991 const1_rtx
, const0_rtx
);
3992 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
3993 emit_insn (prefetch
);
3996 emit_insn (gen_movmem_short (dst
, src
, GEN_INT (255)));
3997 s390_load_address (dst_addr
,
3998 gen_rtx_PLUS (Pmode
, dst_addr
, GEN_INT (256)));
3999 s390_load_address (src_addr
,
4000 gen_rtx_PLUS (Pmode
, src_addr
, GEN_INT (256)));
4002 temp
= expand_binop (mode
, add_optab
, blocks
, constm1_rtx
, blocks
, 1,
4005 emit_move_insn (blocks
, temp
);
4007 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4008 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4010 emit_jump (loop_start_label
);
4011 emit_label (loop_end_label
);
4013 emit_insn (gen_movmem_short (dst
, src
,
4014 convert_to_mode (Pmode
, count
, 1)));
4015 emit_label (end_label
);
4019 /* Emit code to set LEN bytes at DST to VAL.
4020 Make use of clrmem if VAL is zero. */
4023 s390_expand_setmem (rtx dst
, rtx len
, rtx val
)
4025 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) == 0)
4028 gcc_assert (GET_CODE (val
) == CONST_INT
|| GET_MODE (val
) == QImode
);
4030 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) > 0 && INTVAL (len
) <= 257)
4032 if (val
== const0_rtx
&& INTVAL (len
) <= 256)
4033 emit_insn (gen_clrmem_short (dst
, GEN_INT (INTVAL (len
) - 1)));
4036 /* Initialize memory by storing the first byte. */
4037 emit_move_insn (adjust_address (dst
, QImode
, 0), val
);
4039 if (INTVAL (len
) > 1)
4041 /* Initiate 1 byte overlap move.
4042 The first byte of DST is propagated through DSTP1.
4043 Prepare a movmem for: DST+1 = DST (length = LEN - 1).
4044 DST is set to size 1 so the rest of the memory location
4045 does not count as source operand. */
4046 rtx dstp1
= adjust_address (dst
, VOIDmode
, 1);
4047 set_mem_size (dst
, const1_rtx
);
4049 emit_insn (gen_movmem_short (dstp1
, dst
,
4050 GEN_INT (INTVAL (len
) - 2)));
4055 else if (TARGET_MVCLE
)
4057 val
= force_not_mem (convert_modes (Pmode
, QImode
, val
, 1));
4058 emit_insn (gen_setmem_long (dst
, convert_to_mode (Pmode
, len
, 1), val
));
4063 rtx dst_addr
, count
, blocks
, temp
, dstp1
= NULL_RTX
;
4064 rtx loop_start_label
= gen_label_rtx ();
4065 rtx loop_end_label
= gen_label_rtx ();
4066 rtx end_label
= gen_label_rtx ();
4067 enum machine_mode mode
;
4069 mode
= GET_MODE (len
);
4070 if (mode
== VOIDmode
)
4073 dst_addr
= gen_reg_rtx (Pmode
);
4074 count
= gen_reg_rtx (mode
);
4075 blocks
= gen_reg_rtx (mode
);
4077 convert_move (count
, len
, 1);
4078 emit_cmp_and_jump_insns (count
, const0_rtx
,
4079 EQ
, NULL_RTX
, mode
, 1, end_label
);
4081 emit_move_insn (dst_addr
, force_operand (XEXP (dst
, 0), NULL_RTX
));
4082 dst
= change_address (dst
, VOIDmode
, dst_addr
);
4084 if (val
== const0_rtx
)
4085 temp
= expand_binop (mode
, add_optab
, count
, constm1_rtx
, count
, 1,
4089 dstp1
= adjust_address (dst
, VOIDmode
, 1);
4090 set_mem_size (dst
, const1_rtx
);
4092 /* Initialize memory by storing the first byte. */
4093 emit_move_insn (adjust_address (dst
, QImode
, 0), val
);
4095 /* If count is 1 we are done. */
4096 emit_cmp_and_jump_insns (count
, const1_rtx
,
4097 EQ
, NULL_RTX
, mode
, 1, end_label
);
4099 temp
= expand_binop (mode
, add_optab
, count
, GEN_INT (-2), count
, 1,
4103 emit_move_insn (count
, temp
);
4105 temp
= expand_binop (mode
, lshr_optab
, count
, GEN_INT (8), blocks
, 1,
4108 emit_move_insn (blocks
, temp
);
4110 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4111 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4113 emit_label (loop_start_label
);
4116 && (GET_CODE (len
) != CONST_INT
|| INTVAL (len
) > 1024))
4118 /* Issue a write prefetch for the +4 cache line. */
4119 rtx prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, dst_addr
,
4121 const1_rtx
, const0_rtx
);
4122 emit_insn (prefetch
);
4123 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
4126 if (val
== const0_rtx
)
4127 emit_insn (gen_clrmem_short (dst
, GEN_INT (255)));
4129 emit_insn (gen_movmem_short (dstp1
, dst
, GEN_INT (255)));
4130 s390_load_address (dst_addr
,
4131 gen_rtx_PLUS (Pmode
, dst_addr
, GEN_INT (256)));
4133 temp
= expand_binop (mode
, add_optab
, blocks
, constm1_rtx
, blocks
, 1,
4136 emit_move_insn (blocks
, temp
);
4138 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4139 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4141 emit_jump (loop_start_label
);
4142 emit_label (loop_end_label
);
4144 if (val
== const0_rtx
)
4145 emit_insn (gen_clrmem_short (dst
, convert_to_mode (Pmode
, count
, 1)));
4147 emit_insn (gen_movmem_short (dstp1
, dst
, convert_to_mode (Pmode
, count
, 1)));
4148 emit_label (end_label
);
4152 /* Emit code to compare LEN bytes at OP0 with those at OP1,
4153 and return the result in TARGET. */
4156 s390_expand_cmpmem (rtx target
, rtx op0
, rtx op1
, rtx len
)
4158 rtx ccreg
= gen_rtx_REG (CCUmode
, CC_REGNUM
);
4161 /* As the result of CMPINT is inverted compared to what we need,
4162 we have to swap the operands. */
4163 tmp
= op0
; op0
= op1
; op1
= tmp
;
4165 if (GET_CODE (len
) == CONST_INT
&& INTVAL (len
) >= 0 && INTVAL (len
) <= 256)
4167 if (INTVAL (len
) > 0)
4169 emit_insn (gen_cmpmem_short (op0
, op1
, GEN_INT (INTVAL (len
) - 1)));
4170 emit_insn (gen_cmpint (target
, ccreg
));
4173 emit_move_insn (target
, const0_rtx
);
4175 else if (TARGET_MVCLE
)
4177 emit_insn (gen_cmpmem_long (op0
, op1
, convert_to_mode (Pmode
, len
, 1)));
4178 emit_insn (gen_cmpint (target
, ccreg
));
4182 rtx addr0
, addr1
, count
, blocks
, temp
;
4183 rtx loop_start_label
= gen_label_rtx ();
4184 rtx loop_end_label
= gen_label_rtx ();
4185 rtx end_label
= gen_label_rtx ();
4186 enum machine_mode mode
;
4188 mode
= GET_MODE (len
);
4189 if (mode
== VOIDmode
)
4192 addr0
= gen_reg_rtx (Pmode
);
4193 addr1
= gen_reg_rtx (Pmode
);
4194 count
= gen_reg_rtx (mode
);
4195 blocks
= gen_reg_rtx (mode
);
4197 convert_move (count
, len
, 1);
4198 emit_cmp_and_jump_insns (count
, const0_rtx
,
4199 EQ
, NULL_RTX
, mode
, 1, end_label
);
4201 emit_move_insn (addr0
, force_operand (XEXP (op0
, 0), NULL_RTX
));
4202 emit_move_insn (addr1
, force_operand (XEXP (op1
, 0), NULL_RTX
));
4203 op0
= change_address (op0
, VOIDmode
, addr0
);
4204 op1
= change_address (op1
, VOIDmode
, addr1
);
4206 temp
= expand_binop (mode
, add_optab
, count
, constm1_rtx
, count
, 1,
4209 emit_move_insn (count
, temp
);
4211 temp
= expand_binop (mode
, lshr_optab
, count
, GEN_INT (8), blocks
, 1,
4214 emit_move_insn (blocks
, temp
);
4216 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4217 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4219 emit_label (loop_start_label
);
4222 && (GET_CODE (len
) != CONST_INT
|| INTVAL (len
) > 512))
4226 /* Issue a read prefetch for the +2 cache line of operand 1. */
4227 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, addr0
, GEN_INT (512)),
4228 const0_rtx
, const0_rtx
);
4229 emit_insn (prefetch
);
4230 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
4232 /* Issue a read prefetch for the +2 cache line of operand 2. */
4233 prefetch
= gen_prefetch (gen_rtx_PLUS (Pmode
, addr1
, GEN_INT (512)),
4234 const0_rtx
, const0_rtx
);
4235 emit_insn (prefetch
);
4236 PREFETCH_SCHEDULE_BARRIER_P (prefetch
) = true;
4239 emit_insn (gen_cmpmem_short (op0
, op1
, GEN_INT (255)));
4240 temp
= gen_rtx_NE (VOIDmode
, ccreg
, const0_rtx
);
4241 temp
= gen_rtx_IF_THEN_ELSE (VOIDmode
, temp
,
4242 gen_rtx_LABEL_REF (VOIDmode
, end_label
), pc_rtx
);
4243 temp
= gen_rtx_SET (VOIDmode
, pc_rtx
, temp
);
4244 emit_jump_insn (temp
);
4246 s390_load_address (addr0
,
4247 gen_rtx_PLUS (Pmode
, addr0
, GEN_INT (256)));
4248 s390_load_address (addr1
,
4249 gen_rtx_PLUS (Pmode
, addr1
, GEN_INT (256)));
4251 temp
= expand_binop (mode
, add_optab
, blocks
, constm1_rtx
, blocks
, 1,
4254 emit_move_insn (blocks
, temp
);
4256 emit_cmp_and_jump_insns (blocks
, const0_rtx
,
4257 EQ
, NULL_RTX
, mode
, 1, loop_end_label
);
4259 emit_jump (loop_start_label
);
4260 emit_label (loop_end_label
);
4262 emit_insn (gen_cmpmem_short (op0
, op1
,
4263 convert_to_mode (Pmode
, count
, 1)));
4264 emit_label (end_label
);
4266 emit_insn (gen_cmpint (target
, ccreg
));
4271 /* Expand conditional increment or decrement using alc/slb instructions.
4272 Should generate code setting DST to either SRC or SRC + INCREMENT,
4273 depending on the result of the comparison CMP_OP0 CMP_CODE CMP_OP1.
4274 Returns true if successful, false otherwise.
4276 That makes it possible to implement some if-constructs without jumps e.g.:
4277 (borrow = CC0 | CC1 and carry = CC2 | CC3)
4278 unsigned int a, b, c;
4279 if (a < b) c++; -> CCU b > a -> CC2; c += carry;
4280 if (a < b) c--; -> CCL3 a - b -> borrow; c -= borrow;
4281 if (a <= b) c++; -> CCL3 b - a -> borrow; c += carry;
4282 if (a <= b) c--; -> CCU a <= b -> borrow; c -= borrow;
4284 Checks for EQ and NE with a nonzero value need an additional xor e.g.:
4285 if (a == b) c++; -> CCL3 a ^= b; 0 - a -> borrow; c += carry;
4286 if (a == b) c--; -> CCU a ^= b; a <= 0 -> CC0 | CC1; c -= borrow;
4287 if (a != b) c++; -> CCU a ^= b; a > 0 -> CC2; c += carry;
4288 if (a != b) c--; -> CCL3 a ^= b; 0 - a -> borrow; c -= borrow; */
4291 s390_expand_addcc (enum rtx_code cmp_code
, rtx cmp_op0
, rtx cmp_op1
,
4292 rtx dst
, rtx src
, rtx increment
)
4294 enum machine_mode cmp_mode
;
4295 enum machine_mode cc_mode
;
4301 if ((GET_MODE (cmp_op0
) == SImode
|| GET_MODE (cmp_op0
) == VOIDmode
)
4302 && (GET_MODE (cmp_op1
) == SImode
|| GET_MODE (cmp_op1
) == VOIDmode
))
4304 else if ((GET_MODE (cmp_op0
) == DImode
|| GET_MODE (cmp_op0
) == VOIDmode
)
4305 && (GET_MODE (cmp_op1
) == DImode
|| GET_MODE (cmp_op1
) == VOIDmode
))
4310 /* Try ADD LOGICAL WITH CARRY. */
4311 if (increment
== const1_rtx
)
4313 /* Determine CC mode to use. */
4314 if (cmp_code
== EQ
|| cmp_code
== NE
)
4316 if (cmp_op1
!= const0_rtx
)
4318 cmp_op0
= expand_simple_binop (cmp_mode
, XOR
, cmp_op0
, cmp_op1
,
4319 NULL_RTX
, 0, OPTAB_WIDEN
);
4320 cmp_op1
= const0_rtx
;
4323 cmp_code
= cmp_code
== EQ
? LEU
: GTU
;
4326 if (cmp_code
== LTU
|| cmp_code
== LEU
)
4331 cmp_code
= swap_condition (cmp_code
);
4348 /* Emit comparison instruction pattern. */
4349 if (!register_operand (cmp_op0
, cmp_mode
))
4350 cmp_op0
= force_reg (cmp_mode
, cmp_op0
);
4352 insn
= gen_rtx_SET (VOIDmode
, gen_rtx_REG (cc_mode
, CC_REGNUM
),
4353 gen_rtx_COMPARE (cc_mode
, cmp_op0
, cmp_op1
));
4354 /* We use insn_invalid_p here to add clobbers if required. */
4355 ret
= insn_invalid_p (emit_insn (insn
));
4358 /* Emit ALC instruction pattern. */
4359 op_res
= gen_rtx_fmt_ee (cmp_code
, GET_MODE (dst
),
4360 gen_rtx_REG (cc_mode
, CC_REGNUM
),
4363 if (src
!= const0_rtx
)
4365 if (!register_operand (src
, GET_MODE (dst
)))
4366 src
= force_reg (GET_MODE (dst
), src
);
4368 op_res
= gen_rtx_PLUS (GET_MODE (dst
), op_res
, src
);
4369 op_res
= gen_rtx_PLUS (GET_MODE (dst
), op_res
, const0_rtx
);
4372 p
= rtvec_alloc (2);
4374 gen_rtx_SET (VOIDmode
, dst
, op_res
);
4376 gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
4377 emit_insn (gen_rtx_PARALLEL (VOIDmode
, p
));
4382 /* Try SUBTRACT LOGICAL WITH BORROW. */
4383 if (increment
== constm1_rtx
)
4385 /* Determine CC mode to use. */
4386 if (cmp_code
== EQ
|| cmp_code
== NE
)
4388 if (cmp_op1
!= const0_rtx
)
4390 cmp_op0
= expand_simple_binop (cmp_mode
, XOR
, cmp_op0
, cmp_op1
,
4391 NULL_RTX
, 0, OPTAB_WIDEN
);
4392 cmp_op1
= const0_rtx
;
4395 cmp_code
= cmp_code
== EQ
? LEU
: GTU
;
4398 if (cmp_code
== GTU
|| cmp_code
== GEU
)
4403 cmp_code
= swap_condition (cmp_code
);
4420 /* Emit comparison instruction pattern. */
4421 if (!register_operand (cmp_op0
, cmp_mode
))
4422 cmp_op0
= force_reg (cmp_mode
, cmp_op0
);
4424 insn
= gen_rtx_SET (VOIDmode
, gen_rtx_REG (cc_mode
, CC_REGNUM
),
4425 gen_rtx_COMPARE (cc_mode
, cmp_op0
, cmp_op1
));
4426 /* We use insn_invalid_p here to add clobbers if required. */
4427 ret
= insn_invalid_p (emit_insn (insn
));
4430 /* Emit SLB instruction pattern. */
4431 if (!register_operand (src
, GET_MODE (dst
)))
4432 src
= force_reg (GET_MODE (dst
), src
);
4434 op_res
= gen_rtx_MINUS (GET_MODE (dst
),
4435 gen_rtx_MINUS (GET_MODE (dst
), src
, const0_rtx
),
4436 gen_rtx_fmt_ee (cmp_code
, GET_MODE (dst
),
4437 gen_rtx_REG (cc_mode
, CC_REGNUM
),
4439 p
= rtvec_alloc (2);
4441 gen_rtx_SET (VOIDmode
, dst
, op_res
);
4443 gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
4444 emit_insn (gen_rtx_PARALLEL (VOIDmode
, p
));
4452 /* Expand code for the insv template. Return true if successful. */
4455 s390_expand_insv (rtx dest
, rtx op1
, rtx op2
, rtx src
)
4457 int bitsize
= INTVAL (op1
);
4458 int bitpos
= INTVAL (op2
);
4460 /* On z10 we can use the risbg instruction to implement insv. */
4462 && ((GET_MODE (dest
) == DImode
&& GET_MODE (src
) == DImode
)
4463 || (GET_MODE (dest
) == SImode
&& GET_MODE (src
) == SImode
)))
4468 op
= gen_rtx_SET (GET_MODE(src
),
4469 gen_rtx_ZERO_EXTRACT (GET_MODE (dest
), dest
, op1
, op2
),
4471 clobber
= gen_rtx_CLOBBER (VOIDmode
, gen_rtx_REG (CCmode
, CC_REGNUM
));
4472 emit_insn (gen_rtx_PARALLEL (VOIDmode
, gen_rtvec (2, op
, clobber
)));
4477 /* We need byte alignment. */
4478 if (bitsize
% BITS_PER_UNIT
)
4482 && memory_operand (dest
, VOIDmode
)
4483 && (register_operand (src
, word_mode
)
4484 || const_int_operand (src
, VOIDmode
)))
4486 /* Emit standard pattern if possible. */
4487 enum machine_mode mode
= smallest_mode_for_size (bitsize
, MODE_INT
);
4488 if (GET_MODE_BITSIZE (mode
) == bitsize
)
4489 emit_move_insn (adjust_address (dest
, mode
, 0), gen_lowpart (mode
, src
));
4491 /* (set (ze (mem)) (const_int)). */
4492 else if (const_int_operand (src
, VOIDmode
))
4494 int size
= bitsize
/ BITS_PER_UNIT
;
4495 rtx src_mem
= adjust_address (force_const_mem (word_mode
, src
), BLKmode
,
4496 GET_MODE_SIZE (word_mode
) - size
);
4498 dest
= adjust_address (dest
, BLKmode
, 0);
4499 set_mem_size (dest
, GEN_INT (size
));
4500 s390_expand_movmem (dest
, src_mem
, GEN_INT (size
));
4503 /* (set (ze (mem)) (reg)). */
4504 else if (register_operand (src
, word_mode
))
4506 if (bitsize
<= GET_MODE_BITSIZE (SImode
))
4507 emit_move_insn (gen_rtx_ZERO_EXTRACT (word_mode
, dest
, op1
,
4511 /* Emit st,stcmh sequence. */
4512 int stcmh_width
= bitsize
- GET_MODE_BITSIZE (SImode
);
4513 int size
= stcmh_width
/ BITS_PER_UNIT
;
4515 emit_move_insn (adjust_address (dest
, SImode
, size
),
4516 gen_lowpart (SImode
, src
));
4517 set_mem_size (dest
, GEN_INT (size
));
4518 emit_move_insn (gen_rtx_ZERO_EXTRACT (word_mode
, dest
, GEN_INT
4519 (stcmh_width
), const0_rtx
),
4520 gen_rtx_LSHIFTRT (word_mode
, src
, GEN_INT
4521 (GET_MODE_BITSIZE (SImode
))));
4530 /* (set (ze (reg)) (const_int)). */
4532 && register_operand (dest
, word_mode
)
4533 && (bitpos
% 16) == 0
4534 && (bitsize
% 16) == 0
4535 && const_int_operand (src
, VOIDmode
))
4537 HOST_WIDE_INT val
= INTVAL (src
);
4538 int regpos
= bitpos
+ bitsize
;
4540 while (regpos
> bitpos
)
4542 enum machine_mode putmode
;
4545 if (TARGET_EXTIMM
&& (regpos
% 32 == 0) && (regpos
>= bitpos
+ 32))
4550 putsize
= GET_MODE_BITSIZE (putmode
);
4552 emit_move_insn (gen_rtx_ZERO_EXTRACT (word_mode
, dest
,
4555 gen_int_mode (val
, putmode
));
4558 gcc_assert (regpos
== bitpos
);
4565 /* A subroutine of s390_expand_cs_hqi and s390_expand_atomic which returns a
4566 register that holds VAL of mode MODE shifted by COUNT bits. */
4569 s390_expand_mask_and_shift (rtx val
, enum machine_mode mode
, rtx count
)
4571 val
= expand_simple_binop (SImode
, AND
, val
, GEN_INT (GET_MODE_MASK (mode
)),
4572 NULL_RTX
, 1, OPTAB_DIRECT
);
4573 return expand_simple_binop (SImode
, ASHIFT
, val
, count
,
4574 NULL_RTX
, 1, OPTAB_DIRECT
);
4577 /* Structure to hold the initial parameters for a compare_and_swap operation
4578 in HImode and QImode. */
4580 struct alignment_context
4582 rtx memsi
; /* SI aligned memory location. */
4583 rtx shift
; /* Bit offset with regard to lsb. */
4584 rtx modemask
; /* Mask of the HQImode shifted by SHIFT bits. */
4585 rtx modemaski
; /* ~modemask */
4586 bool aligned
; /* True if memory is aligned, false else. */
4589 /* A subroutine of s390_expand_cs_hqi and s390_expand_atomic to initialize
4590 structure AC for transparent simplifying, if the memory alignment is known
4591 to be at least 32bit. MEM is the memory location for the actual operation
4592 and MODE its mode. */
4595 init_alignment_context (struct alignment_context
*ac
, rtx mem
,
4596 enum machine_mode mode
)
4598 ac
->shift
= GEN_INT (GET_MODE_SIZE (SImode
) - GET_MODE_SIZE (mode
));
4599 ac
->aligned
= (MEM_ALIGN (mem
) >= GET_MODE_BITSIZE (SImode
));
4602 ac
->memsi
= adjust_address (mem
, SImode
, 0); /* Memory is aligned. */
4605 /* Alignment is unknown. */
4606 rtx byteoffset
, addr
, align
;
4608 /* Force the address into a register. */
4609 addr
= force_reg (Pmode
, XEXP (mem
, 0));
4611 /* Align it to SImode. */
4612 align
= expand_simple_binop (Pmode
, AND
, addr
,
4613 GEN_INT (-GET_MODE_SIZE (SImode
)),
4614 NULL_RTX
, 1, OPTAB_DIRECT
);
4616 ac
->memsi
= gen_rtx_MEM (SImode
, align
);
4617 MEM_VOLATILE_P (ac
->memsi
) = MEM_VOLATILE_P (mem
);
4618 set_mem_alias_set (ac
->memsi
, ALIAS_SET_MEMORY_BARRIER
);
4619 set_mem_align (ac
->memsi
, GET_MODE_BITSIZE (SImode
));
4621 /* Calculate shiftcount. */
4622 byteoffset
= expand_simple_binop (Pmode
, AND
, addr
,
4623 GEN_INT (GET_MODE_SIZE (SImode
) - 1),
4624 NULL_RTX
, 1, OPTAB_DIRECT
);
4625 /* As we already have some offset, evaluate the remaining distance. */
4626 ac
->shift
= expand_simple_binop (SImode
, MINUS
, ac
->shift
, byteoffset
,
4627 NULL_RTX
, 1, OPTAB_DIRECT
);
4630 /* Shift is the byte count, but we need the bitcount. */
4631 ac
->shift
= expand_simple_binop (SImode
, MULT
, ac
->shift
, GEN_INT (BITS_PER_UNIT
),
4632 NULL_RTX
, 1, OPTAB_DIRECT
);
4633 /* Calculate masks. */
4634 ac
->modemask
= expand_simple_binop (SImode
, ASHIFT
,
4635 GEN_INT (GET_MODE_MASK (mode
)), ac
->shift
,
4636 NULL_RTX
, 1, OPTAB_DIRECT
);
4637 ac
->modemaski
= expand_simple_unop (SImode
, NOT
, ac
->modemask
, NULL_RTX
, 1);
4640 /* Expand an atomic compare and swap operation for HImode and QImode. MEM is
4641 the memory location, CMP the old value to compare MEM with and NEW_RTX the value
4642 to set if CMP == MEM.
4643 CMP is never in memory for compare_and_swap_cc because
4644 expand_bool_compare_and_swap puts it into a register for later compare. */
4647 s390_expand_cs_hqi (enum machine_mode mode
, rtx target
, rtx mem
, rtx cmp
, rtx new_rtx
)
4649 struct alignment_context ac
;
4650 rtx cmpv
, newv
, val
, resv
, cc
;
4651 rtx res
= gen_reg_rtx (SImode
);
4652 rtx csloop
= gen_label_rtx ();
4653 rtx csend
= gen_label_rtx ();
4655 gcc_assert (register_operand (target
, VOIDmode
));
4656 gcc_assert (MEM_P (mem
));
4658 init_alignment_context (&ac
, mem
, mode
);
4660 /* Shift the values to the correct bit positions. */
4661 if (!(ac
.aligned
&& MEM_P (cmp
)))
4662 cmp
= s390_expand_mask_and_shift (cmp
, mode
, ac
.shift
);
4663 if (!(ac
.aligned
&& MEM_P (new_rtx
)))
4664 new_rtx
= s390_expand_mask_and_shift (new_rtx
, mode
, ac
.shift
);
4666 /* Load full word. Subsequent loads are performed by CS. */
4667 val
= expand_simple_binop (SImode
, AND
, ac
.memsi
, ac
.modemaski
,
4668 NULL_RTX
, 1, OPTAB_DIRECT
);
4670 /* Start CS loop. */
4671 emit_label (csloop
);
4672 /* val = "<mem>00..0<mem>"
4673 * cmp = "00..0<cmp>00..0"
4674 * new = "00..0<new>00..0"
4677 /* Patch cmp and new with val at correct position. */
4678 if (ac
.aligned
&& MEM_P (cmp
))
4680 cmpv
= force_reg (SImode
, val
);
4681 store_bit_field (cmpv
, GET_MODE_BITSIZE (mode
), 0, SImode
, cmp
);
4684 cmpv
= force_reg (SImode
, expand_simple_binop (SImode
, IOR
, cmp
, val
,
4685 NULL_RTX
, 1, OPTAB_DIRECT
));
4686 if (ac
.aligned
&& MEM_P (new_rtx
))
4688 newv
= force_reg (SImode
, val
);
4689 store_bit_field (newv
, GET_MODE_BITSIZE (mode
), 0, SImode
, new_rtx
);
4692 newv
= force_reg (SImode
, expand_simple_binop (SImode
, IOR
, new_rtx
, val
,
4693 NULL_RTX
, 1, OPTAB_DIRECT
));
4695 /* Jump to end if we're done (likely?). */
4696 s390_emit_jump (csend
, s390_emit_compare_and_swap (EQ
, res
, ac
.memsi
,
4699 /* Check for changes outside mode. */
4700 resv
= expand_simple_binop (SImode
, AND
, res
, ac
.modemaski
,
4701 NULL_RTX
, 1, OPTAB_DIRECT
);
4702 cc
= s390_emit_compare (NE
, resv
, val
);
4703 emit_move_insn (val
, resv
);
4704 /* Loop internal if so. */
4705 s390_emit_jump (csloop
, cc
);
4709 /* Return the correct part of the bitfield. */
4710 convert_move (target
, expand_simple_binop (SImode
, LSHIFTRT
, res
, ac
.shift
,
4711 NULL_RTX
, 1, OPTAB_DIRECT
), 1);
4714 /* Expand an atomic operation CODE of mode MODE. MEM is the memory location
4715 and VAL the value to play with. If AFTER is true then store the value
4716 MEM holds after the operation, if AFTER is false then store the value MEM
4717 holds before the operation. If TARGET is zero then discard that value, else
4718 store it to TARGET. */
4721 s390_expand_atomic (enum machine_mode mode
, enum rtx_code code
,
4722 rtx target
, rtx mem
, rtx val
, bool after
)
4724 struct alignment_context ac
;
4726 rtx new_rtx
= gen_reg_rtx (SImode
);
4727 rtx orig
= gen_reg_rtx (SImode
);
4728 rtx csloop
= gen_label_rtx ();
4730 gcc_assert (!target
|| register_operand (target
, VOIDmode
));
4731 gcc_assert (MEM_P (mem
));
4733 init_alignment_context (&ac
, mem
, mode
);
4735 /* Shift val to the correct bit positions.
4736 Preserve "icm", but prevent "ex icm". */
4737 if (!(ac
.aligned
&& code
== SET
&& MEM_P (val
)))
4738 val
= s390_expand_mask_and_shift (val
, mode
, ac
.shift
);
4740 /* Further preparation insns. */
4741 if (code
== PLUS
|| code
== MINUS
)
4742 emit_move_insn (orig
, val
);
4743 else if (code
== MULT
|| code
== AND
) /* val = "11..1<val>11..1" */
4744 val
= expand_simple_binop (SImode
, XOR
, val
, ac
.modemaski
,
4745 NULL_RTX
, 1, OPTAB_DIRECT
);
4747 /* Load full word. Subsequent loads are performed by CS. */
4748 cmp
= force_reg (SImode
, ac
.memsi
);
4750 /* Start CS loop. */
4751 emit_label (csloop
);
4752 emit_move_insn (new_rtx
, cmp
);
4754 /* Patch new with val at correct position. */
4759 val
= expand_simple_binop (SImode
, code
, new_rtx
, orig
,
4760 NULL_RTX
, 1, OPTAB_DIRECT
);
4761 val
= expand_simple_binop (SImode
, AND
, val
, ac
.modemask
,
4762 NULL_RTX
, 1, OPTAB_DIRECT
);
4765 if (ac
.aligned
&& MEM_P (val
))
4766 store_bit_field (new_rtx
, GET_MODE_BITSIZE (mode
), 0, SImode
, val
);
4769 new_rtx
= expand_simple_binop (SImode
, AND
, new_rtx
, ac
.modemaski
,
4770 NULL_RTX
, 1, OPTAB_DIRECT
);
4771 new_rtx
= expand_simple_binop (SImode
, IOR
, new_rtx
, val
,
4772 NULL_RTX
, 1, OPTAB_DIRECT
);
4778 new_rtx
= expand_simple_binop (SImode
, code
, new_rtx
, val
,
4779 NULL_RTX
, 1, OPTAB_DIRECT
);
4781 case MULT
: /* NAND */
4782 new_rtx
= expand_simple_binop (SImode
, AND
, new_rtx
, val
,
4783 NULL_RTX
, 1, OPTAB_DIRECT
);
4784 new_rtx
= expand_simple_binop (SImode
, XOR
, new_rtx
, ac
.modemask
,
4785 NULL_RTX
, 1, OPTAB_DIRECT
);
4791 s390_emit_jump (csloop
, s390_emit_compare_and_swap (NE
, cmp
,
4792 ac
.memsi
, cmp
, new_rtx
));
4794 /* Return the correct part of the bitfield. */
4796 convert_move (target
, expand_simple_binop (SImode
, LSHIFTRT
,
4797 after
? new_rtx
: cmp
, ac
.shift
,
4798 NULL_RTX
, 1, OPTAB_DIRECT
), 1);
4801 /* This is called from dwarf2out.c via TARGET_ASM_OUTPUT_DWARF_DTPREL.
4802 We need to emit DTP-relative relocations. */
4804 static void s390_output_dwarf_dtprel (FILE *, int, rtx
) ATTRIBUTE_UNUSED
;
4807 s390_output_dwarf_dtprel (FILE *file
, int size
, rtx x
)
4812 fputs ("\t.long\t", file
);
4815 fputs ("\t.quad\t", file
);
4820 output_addr_const (file
, x
);
4821 fputs ("@DTPOFF", file
);
4824 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
4825 /* Implement TARGET_MANGLE_TYPE. */
4828 s390_mangle_type (const_tree type
)
4830 if (TYPE_MAIN_VARIANT (type
) == long_double_type_node
4831 && TARGET_LONG_DOUBLE_128
)
4834 /* For all other types, use normal C++ mangling. */
4839 /* In the name of slightly smaller debug output, and to cater to
4840 general assembler lossage, recognize various UNSPEC sequences
4841 and turn them back into a direct symbol reference. */
4844 s390_delegitimize_address (rtx orig_x
)
4848 orig_x
= delegitimize_mem_from_attrs (orig_x
);
4850 if (GET_CODE (x
) != MEM
)
4854 if (GET_CODE (x
) == PLUS
4855 && GET_CODE (XEXP (x
, 1)) == CONST
4856 && GET_CODE (XEXP (x
, 0)) == REG
4857 && REGNO (XEXP (x
, 0)) == PIC_OFFSET_TABLE_REGNUM
)
4859 y
= XEXP (XEXP (x
, 1), 0);
4860 if (GET_CODE (y
) == UNSPEC
4861 && XINT (y
, 1) == UNSPEC_GOT
)
4862 return XVECEXP (y
, 0, 0);
4866 if (GET_CODE (x
) == CONST
)
4869 if (GET_CODE (y
) == UNSPEC
4870 && XINT (y
, 1) == UNSPEC_GOTENT
)
4871 return XVECEXP (y
, 0, 0);
4878 /* Output operand OP to stdio stream FILE.
4879 OP is an address (register + offset) which is not used to address data;
4880 instead the rightmost bits are interpreted as the value. */
4883 print_shift_count_operand (FILE *file
, rtx op
)
4885 HOST_WIDE_INT offset
;
4888 /* Extract base register and offset. */
4889 if (!s390_decompose_shift_count (op
, &base
, &offset
))
4895 gcc_assert (GET_CODE (base
) == REG
);
4896 gcc_assert (REGNO (base
) < FIRST_PSEUDO_REGISTER
);
4897 gcc_assert (REGNO_REG_CLASS (REGNO (base
)) == ADDR_REGS
);
4900 /* Offsets are constricted to twelve bits. */
4901 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, offset
& ((1 << 12) - 1));
4903 fprintf (file
, "(%s)", reg_names
[REGNO (base
)]);
4906 /* See 'get_some_local_dynamic_name'. */
4909 get_some_local_dynamic_name_1 (rtx
*px
, void *data ATTRIBUTE_UNUSED
)
4913 if (GET_CODE (x
) == SYMBOL_REF
&& CONSTANT_POOL_ADDRESS_P (x
))
4915 x
= get_pool_constant (x
);
4916 return for_each_rtx (&x
, get_some_local_dynamic_name_1
, 0);
4919 if (GET_CODE (x
) == SYMBOL_REF
4920 && tls_symbolic_operand (x
) == TLS_MODEL_LOCAL_DYNAMIC
)
4922 cfun
->machine
->some_ld_name
= XSTR (x
, 0);
4929 /* Locate some local-dynamic symbol still in use by this function
4930 so that we can print its name in local-dynamic base patterns. */
4933 get_some_local_dynamic_name (void)
4937 if (cfun
->machine
->some_ld_name
)
4938 return cfun
->machine
->some_ld_name
;
4940 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
4942 && for_each_rtx (&PATTERN (insn
), get_some_local_dynamic_name_1
, 0))
4943 return cfun
->machine
->some_ld_name
;
4948 /* Output machine-dependent UNSPECs occurring in address constant X
4949 in assembler syntax to stdio stream FILE. Returns true if the
4950 constant X could be recognized, false otherwise. */
4953 s390_output_addr_const_extra (FILE *file
, rtx x
)
4955 if (GET_CODE (x
) == UNSPEC
&& XVECLEN (x
, 0) == 1)
4956 switch (XINT (x
, 1))
4959 output_addr_const (file
, XVECEXP (x
, 0, 0));
4960 fprintf (file
, "@GOTENT");
4963 output_addr_const (file
, XVECEXP (x
, 0, 0));
4964 fprintf (file
, "@GOT");
4967 output_addr_const (file
, XVECEXP (x
, 0, 0));
4968 fprintf (file
, "@GOTOFF");
4971 output_addr_const (file
, XVECEXP (x
, 0, 0));
4972 fprintf (file
, "@PLT");
4975 output_addr_const (file
, XVECEXP (x
, 0, 0));
4976 fprintf (file
, "@PLTOFF");
4979 output_addr_const (file
, XVECEXP (x
, 0, 0));
4980 fprintf (file
, "@TLSGD");
4983 assemble_name (file
, get_some_local_dynamic_name ());
4984 fprintf (file
, "@TLSLDM");
4987 output_addr_const (file
, XVECEXP (x
, 0, 0));
4988 fprintf (file
, "@DTPOFF");
4991 output_addr_const (file
, XVECEXP (x
, 0, 0));
4992 fprintf (file
, "@NTPOFF");
4994 case UNSPEC_GOTNTPOFF
:
4995 output_addr_const (file
, XVECEXP (x
, 0, 0));
4996 fprintf (file
, "@GOTNTPOFF");
4998 case UNSPEC_INDNTPOFF
:
4999 output_addr_const (file
, XVECEXP (x
, 0, 0));
5000 fprintf (file
, "@INDNTPOFF");
5004 if (GET_CODE (x
) == UNSPEC
&& XVECLEN (x
, 0) == 2)
5005 switch (XINT (x
, 1))
5007 case UNSPEC_POOL_OFFSET
:
5008 x
= gen_rtx_MINUS (GET_MODE (x
), XVECEXP (x
, 0, 0), XVECEXP (x
, 0, 1));
5009 output_addr_const (file
, x
);
5015 /* Output address operand ADDR in assembler syntax to
5016 stdio stream FILE. */
5019 print_operand_address (FILE *file
, rtx addr
)
5021 struct s390_address ad
;
5023 if (s390_symref_operand_p (addr
, NULL
, NULL
))
5025 gcc_assert (TARGET_Z10
);
5026 output_addr_const (file
, addr
);
5030 if (!s390_decompose_address (addr
, &ad
)
5031 || (ad
.base
&& !REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)))
5032 || (ad
.indx
&& !REGNO_OK_FOR_INDEX_P (REGNO (ad
.indx
))))
5033 output_operand_lossage ("cannot decompose address");
5036 output_addr_const (file
, ad
.disp
);
5038 fprintf (file
, "0");
5040 if (ad
.base
&& ad
.indx
)
5041 fprintf (file
, "(%s,%s)", reg_names
[REGNO (ad
.indx
)],
5042 reg_names
[REGNO (ad
.base
)]);
5044 fprintf (file
, "(%s)", reg_names
[REGNO (ad
.base
)]);
5047 /* Output operand X in assembler syntax to stdio stream FILE.
5048 CODE specified the format flag. The following format flags
5051 'C': print opcode suffix for branch condition.
5052 'D': print opcode suffix for inverse branch condition.
5053 'E': print opcode suffix for branch on index instruction.
5054 'J': print tls_load/tls_gdcall/tls_ldcall suffix
5055 'G': print the size of the operand in bytes.
5056 'O': print only the displacement of a memory reference.
5057 'R': print only the base register of a memory reference.
5058 'S': print S-type memory reference (base+displacement).
5059 'N': print the second word of a DImode operand.
5060 'M': print the second word of a TImode operand.
5061 'Y': print shift count operand.
5063 'b': print integer X as if it's an unsigned byte.
5064 'c': print integer X as if it's an signed byte.
5065 'x': print integer X as if it's an unsigned halfword.
5066 'h': print integer X as if it's a signed halfword.
5067 'i': print the first nonzero HImode part of X.
5068 'j': print the first HImode part unequal to -1 of X.
5069 'k': print the first nonzero SImode part of X.
5070 'm': print the first SImode part unequal to -1 of X.
5071 'o': print integer X as if it's an unsigned 32bit word. */
5074 print_operand (FILE *file
, rtx x
, int code
)
5079 fprintf (file
, s390_branch_condition_mnemonic (x
, FALSE
));
5083 fprintf (file
, s390_branch_condition_mnemonic (x
, TRUE
));
5087 if (GET_CODE (x
) == LE
)
5088 fprintf (file
, "l");
5089 else if (GET_CODE (x
) == GT
)
5090 fprintf (file
, "h");
5096 if (GET_CODE (x
) == SYMBOL_REF
)
5098 fprintf (file
, "%s", ":tls_load:");
5099 output_addr_const (file
, x
);
5101 else if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_TLSGD
)
5103 fprintf (file
, "%s", ":tls_gdcall:");
5104 output_addr_const (file
, XVECEXP (x
, 0, 0));
5106 else if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_TLSLDM
)
5108 fprintf (file
, "%s", ":tls_ldcall:");
5109 assemble_name (file
, get_some_local_dynamic_name ());
5116 fprintf (file
, "%u", GET_MODE_SIZE (GET_MODE (x
)));
5121 struct s390_address ad
;
5124 gcc_assert (GET_CODE (x
) == MEM
);
5125 ret
= s390_decompose_address (XEXP (x
, 0), &ad
);
5127 gcc_assert (!ad
.base
|| REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)));
5128 gcc_assert (!ad
.indx
);
5131 output_addr_const (file
, ad
.disp
);
5133 fprintf (file
, "0");
5139 struct s390_address ad
;
5142 gcc_assert (GET_CODE (x
) == MEM
);
5143 ret
= s390_decompose_address (XEXP (x
, 0), &ad
);
5145 gcc_assert (!ad
.base
|| REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)));
5146 gcc_assert (!ad
.indx
);
5149 fprintf (file
, "%s", reg_names
[REGNO (ad
.base
)]);
5151 fprintf (file
, "0");
5157 struct s390_address ad
;
5160 gcc_assert (GET_CODE (x
) == MEM
);
5161 ret
= s390_decompose_address (XEXP (x
, 0), &ad
);
5163 gcc_assert (!ad
.base
|| REGNO_OK_FOR_BASE_P (REGNO (ad
.base
)));
5164 gcc_assert (!ad
.indx
);
5167 output_addr_const (file
, ad
.disp
);
5169 fprintf (file
, "0");
5172 fprintf (file
, "(%s)", reg_names
[REGNO (ad
.base
)]);
5177 if (GET_CODE (x
) == REG
)
5178 x
= gen_rtx_REG (GET_MODE (x
), REGNO (x
) + 1);
5179 else if (GET_CODE (x
) == MEM
)
5180 x
= change_address (x
, VOIDmode
, plus_constant (XEXP (x
, 0), 4));
5186 if (GET_CODE (x
) == REG
)
5187 x
= gen_rtx_REG (GET_MODE (x
), REGNO (x
) + 1);
5188 else if (GET_CODE (x
) == MEM
)
5189 x
= change_address (x
, VOIDmode
, plus_constant (XEXP (x
, 0), 8));
5195 print_shift_count_operand (file
, x
);
5199 switch (GET_CODE (x
))
5202 fprintf (file
, "%s", reg_names
[REGNO (x
)]);
5206 output_address (XEXP (x
, 0));
5213 output_addr_const (file
, x
);
5218 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) & 0xff);
5219 else if (code
== 'c')
5220 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ((INTVAL (x
) & 0xff) ^ 0x80) - 0x80);
5221 else if (code
== 'x')
5222 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) & 0xffff);
5223 else if (code
== 'h')
5224 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ((INTVAL (x
) & 0xffff) ^ 0x8000) - 0x8000);
5225 else if (code
== 'i')
5226 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5227 s390_extract_part (x
, HImode
, 0));
5228 else if (code
== 'j')
5229 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5230 s390_extract_part (x
, HImode
, -1));
5231 else if (code
== 'k')
5232 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5233 s390_extract_part (x
, SImode
, 0));
5234 else if (code
== 'm')
5235 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5236 s390_extract_part (x
, SImode
, -1));
5237 else if (code
== 'o')
5238 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) & 0xffffffff);
5240 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
));
5244 gcc_assert (GET_MODE (x
) == VOIDmode
);
5246 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, CONST_DOUBLE_LOW (x
) & 0xff);
5247 else if (code
== 'x')
5248 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, CONST_DOUBLE_LOW (x
) & 0xffff);
5249 else if (code
== 'h')
5250 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ((CONST_DOUBLE_LOW (x
) & 0xffff) ^ 0x8000) - 0x8000);
5256 fatal_insn ("UNKNOWN in print_operand !?", x
);
5261 /* Target hook for assembling integer objects. We need to define it
5262 here to work a round a bug in some versions of GAS, which couldn't
5263 handle values smaller than INT_MIN when printed in decimal. */
5266 s390_assemble_integer (rtx x
, unsigned int size
, int aligned_p
)
5268 if (size
== 8 && aligned_p
5269 && GET_CODE (x
) == CONST_INT
&& INTVAL (x
) < INT_MIN
)
5271 fprintf (asm_out_file
, "\t.quad\t" HOST_WIDE_INT_PRINT_HEX
"\n",
5275 return default_assemble_integer (x
, size
, aligned_p
);
5278 /* Returns true if register REGNO is used for forming
5279 a memory address in expression X. */
5282 reg_used_in_mem_p (int regno
, rtx x
)
5284 enum rtx_code code
= GET_CODE (x
);
5290 if (refers_to_regno_p (regno
, regno
+1,
5294 else if (code
== SET
5295 && GET_CODE (SET_DEST (x
)) == PC
)
5297 if (refers_to_regno_p (regno
, regno
+1,
5302 fmt
= GET_RTX_FORMAT (code
);
5303 for (i
= GET_RTX_LENGTH (code
) - 1; i
>= 0; i
--)
5306 && reg_used_in_mem_p (regno
, XEXP (x
, i
)))
5309 else if (fmt
[i
] == 'E')
5310 for (j
= 0; j
< XVECLEN (x
, i
); j
++)
5311 if (reg_used_in_mem_p (regno
, XVECEXP (x
, i
, j
)))
5317 /* Returns true if expression DEP_RTX sets an address register
5318 used by instruction INSN to address memory. */
5321 addr_generation_dependency_p (rtx dep_rtx
, rtx insn
)
5325 if (GET_CODE (dep_rtx
) == INSN
)
5326 dep_rtx
= PATTERN (dep_rtx
);
5328 if (GET_CODE (dep_rtx
) == SET
)
5330 target
= SET_DEST (dep_rtx
);
5331 if (GET_CODE (target
) == STRICT_LOW_PART
)
5332 target
= XEXP (target
, 0);
5333 while (GET_CODE (target
) == SUBREG
)
5334 target
= SUBREG_REG (target
);
5336 if (GET_CODE (target
) == REG
)
5338 int regno
= REGNO (target
);
5340 if (s390_safe_attr_type (insn
) == TYPE_LA
)
5342 pat
= PATTERN (insn
);
5343 if (GET_CODE (pat
) == PARALLEL
)
5345 gcc_assert (XVECLEN (pat
, 0) == 2);
5346 pat
= XVECEXP (pat
, 0, 0);
5348 gcc_assert (GET_CODE (pat
) == SET
);
5349 return refers_to_regno_p (regno
, regno
+1, SET_SRC (pat
), 0);
5351 else if (get_attr_atype (insn
) == ATYPE_AGEN
)
5352 return reg_used_in_mem_p (regno
, PATTERN (insn
));
5358 /* Return 1, if dep_insn sets register used in insn in the agen unit. */
5361 s390_agen_dep_p (rtx dep_insn
, rtx insn
)
5363 rtx dep_rtx
= PATTERN (dep_insn
);
5366 if (GET_CODE (dep_rtx
) == SET
5367 && addr_generation_dependency_p (dep_rtx
, insn
))
5369 else if (GET_CODE (dep_rtx
) == PARALLEL
)
5371 for (i
= 0; i
< XVECLEN (dep_rtx
, 0); i
++)
5373 if (addr_generation_dependency_p (XVECEXP (dep_rtx
, 0, i
), insn
))
5381 /* A C statement (sans semicolon) to update the integer scheduling priority
5382 INSN_PRIORITY (INSN). Increase the priority to execute the INSN earlier,
5383 reduce the priority to execute INSN later. Do not define this macro if
5384 you do not need to adjust the scheduling priorities of insns.
5386 A STD instruction should be scheduled earlier,
5387 in order to use the bypass. */
5391 s390_adjust_priority (rtx insn ATTRIBUTE_UNUSED
, int priority
)
5393 if (! INSN_P (insn
))
5396 if (s390_tune
!= PROCESSOR_2084_Z990
5397 && s390_tune
!= PROCESSOR_2094_Z9_109
5398 && s390_tune
!= PROCESSOR_2097_Z10
)
5401 switch (s390_safe_attr_type (insn
))
5405 priority
= priority
<< 3;
5409 priority
= priority
<< 1;
5418 /* The number of instructions that can be issued per cycle. */
5421 s390_issue_rate (void)
5425 case PROCESSOR_2084_Z990
:
5426 case PROCESSOR_2094_Z9_109
:
5428 case PROCESSOR_2097_Z10
:
5436 s390_first_cycle_multipass_dfa_lookahead (void)
5441 /* Annotate every literal pool reference in X by an UNSPEC_LTREF expression.
5442 Fix up MEMs as required. */
5445 annotate_constant_pool_refs (rtx
*x
)
5450 gcc_assert (GET_CODE (*x
) != SYMBOL_REF
5451 || !CONSTANT_POOL_ADDRESS_P (*x
));
5453 /* Literal pool references can only occur inside a MEM ... */
5454 if (GET_CODE (*x
) == MEM
)
5456 rtx memref
= XEXP (*x
, 0);
5458 if (GET_CODE (memref
) == SYMBOL_REF
5459 && CONSTANT_POOL_ADDRESS_P (memref
))
5461 rtx base
= cfun
->machine
->base_reg
;
5462 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, memref
, base
),
5465 *x
= replace_equiv_address (*x
, addr
);
5469 if (GET_CODE (memref
) == CONST
5470 && GET_CODE (XEXP (memref
, 0)) == PLUS
5471 && GET_CODE (XEXP (XEXP (memref
, 0), 1)) == CONST_INT
5472 && GET_CODE (XEXP (XEXP (memref
, 0), 0)) == SYMBOL_REF
5473 && CONSTANT_POOL_ADDRESS_P (XEXP (XEXP (memref
, 0), 0)))
5475 HOST_WIDE_INT off
= INTVAL (XEXP (XEXP (memref
, 0), 1));
5476 rtx sym
= XEXP (XEXP (memref
, 0), 0);
5477 rtx base
= cfun
->machine
->base_reg
;
5478 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, sym
, base
),
5481 *x
= replace_equiv_address (*x
, plus_constant (addr
, off
));
5486 /* ... or a load-address type pattern. */
5487 if (GET_CODE (*x
) == SET
)
5489 rtx addrref
= SET_SRC (*x
);
5491 if (GET_CODE (addrref
) == SYMBOL_REF
5492 && CONSTANT_POOL_ADDRESS_P (addrref
))
5494 rtx base
= cfun
->machine
->base_reg
;
5495 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, addrref
, base
),
5498 SET_SRC (*x
) = addr
;
5502 if (GET_CODE (addrref
) == CONST
5503 && GET_CODE (XEXP (addrref
, 0)) == PLUS
5504 && GET_CODE (XEXP (XEXP (addrref
, 0), 1)) == CONST_INT
5505 && GET_CODE (XEXP (XEXP (addrref
, 0), 0)) == SYMBOL_REF
5506 && CONSTANT_POOL_ADDRESS_P (XEXP (XEXP (addrref
, 0), 0)))
5508 HOST_WIDE_INT off
= INTVAL (XEXP (XEXP (addrref
, 0), 1));
5509 rtx sym
= XEXP (XEXP (addrref
, 0), 0);
5510 rtx base
= cfun
->machine
->base_reg
;
5511 rtx addr
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, sym
, base
),
5514 SET_SRC (*x
) = plus_constant (addr
, off
);
5519 /* Annotate LTREL_BASE as well. */
5520 if (GET_CODE (*x
) == UNSPEC
5521 && XINT (*x
, 1) == UNSPEC_LTREL_BASE
)
5523 rtx base
= cfun
->machine
->base_reg
;
5524 *x
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, XVECEXP (*x
, 0, 0), base
),
5529 fmt
= GET_RTX_FORMAT (GET_CODE (*x
));
5530 for (i
= GET_RTX_LENGTH (GET_CODE (*x
)) - 1; i
>= 0; i
--)
5534 annotate_constant_pool_refs (&XEXP (*x
, i
));
5536 else if (fmt
[i
] == 'E')
5538 for (j
= 0; j
< XVECLEN (*x
, i
); j
++)
5539 annotate_constant_pool_refs (&XVECEXP (*x
, i
, j
));
5544 /* Split all branches that exceed the maximum distance.
5545 Returns true if this created a new literal pool entry. */
5548 s390_split_branches (void)
5550 rtx temp_reg
= gen_rtx_REG (Pmode
, RETURN_REGNUM
);
5551 int new_literal
= 0, ret
;
5552 rtx insn
, pat
, tmp
, target
;
5555 /* We need correct insn addresses. */
5557 shorten_branches (get_insns ());
5559 /* Find all branches that exceed 64KB, and split them. */
5561 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
5563 if (GET_CODE (insn
) != JUMP_INSN
)
5566 pat
= PATTERN (insn
);
5567 if (GET_CODE (pat
) == PARALLEL
&& XVECLEN (pat
, 0) > 2)
5568 pat
= XVECEXP (pat
, 0, 0);
5569 if (GET_CODE (pat
) != SET
|| SET_DEST (pat
) != pc_rtx
)
5572 if (GET_CODE (SET_SRC (pat
)) == LABEL_REF
)
5574 label
= &SET_SRC (pat
);
5576 else if (GET_CODE (SET_SRC (pat
)) == IF_THEN_ELSE
)
5578 if (GET_CODE (XEXP (SET_SRC (pat
), 1)) == LABEL_REF
)
5579 label
= &XEXP (SET_SRC (pat
), 1);
5580 else if (GET_CODE (XEXP (SET_SRC (pat
), 2)) == LABEL_REF
)
5581 label
= &XEXP (SET_SRC (pat
), 2);
5588 if (get_attr_length (insn
) <= 4)
5591 /* We are going to use the return register as scratch register,
5592 make sure it will be saved/restored by the prologue/epilogue. */
5593 cfun_frame_layout
.save_return_addr_p
= 1;
5598 tmp
= force_const_mem (Pmode
, *label
);
5599 tmp
= emit_insn_before (gen_rtx_SET (Pmode
, temp_reg
, tmp
), insn
);
5600 INSN_ADDRESSES_NEW (tmp
, -1);
5601 annotate_constant_pool_refs (&PATTERN (tmp
));
5608 target
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, *label
),
5609 UNSPEC_LTREL_OFFSET
);
5610 target
= gen_rtx_CONST (Pmode
, target
);
5611 target
= force_const_mem (Pmode
, target
);
5612 tmp
= emit_insn_before (gen_rtx_SET (Pmode
, temp_reg
, target
), insn
);
5613 INSN_ADDRESSES_NEW (tmp
, -1);
5614 annotate_constant_pool_refs (&PATTERN (tmp
));
5616 target
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (2, XEXP (target
, 0),
5617 cfun
->machine
->base_reg
),
5619 target
= gen_rtx_PLUS (Pmode
, temp_reg
, target
);
5622 ret
= validate_change (insn
, label
, target
, 0);
5630 /* Find an annotated literal pool symbol referenced in RTX X,
5631 and store it at REF. Will abort if X contains references to
5632 more than one such pool symbol; multiple references to the same
5633 symbol are allowed, however.
5635 The rtx pointed to by REF must be initialized to NULL_RTX
5636 by the caller before calling this routine. */
5639 find_constant_pool_ref (rtx x
, rtx
*ref
)
5644 /* Ignore LTREL_BASE references. */
5645 if (GET_CODE (x
) == UNSPEC
5646 && XINT (x
, 1) == UNSPEC_LTREL_BASE
)
5648 /* Likewise POOL_ENTRY insns. */
5649 if (GET_CODE (x
) == UNSPEC_VOLATILE
5650 && XINT (x
, 1) == UNSPECV_POOL_ENTRY
)
5653 gcc_assert (GET_CODE (x
) != SYMBOL_REF
5654 || !CONSTANT_POOL_ADDRESS_P (x
));
5656 if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_LTREF
)
5658 rtx sym
= XVECEXP (x
, 0, 0);
5659 gcc_assert (GET_CODE (sym
) == SYMBOL_REF
5660 && CONSTANT_POOL_ADDRESS_P (sym
));
5662 if (*ref
== NULL_RTX
)
5665 gcc_assert (*ref
== sym
);
5670 fmt
= GET_RTX_FORMAT (GET_CODE (x
));
5671 for (i
= GET_RTX_LENGTH (GET_CODE (x
)) - 1; i
>= 0; i
--)
5675 find_constant_pool_ref (XEXP (x
, i
), ref
);
5677 else if (fmt
[i
] == 'E')
5679 for (j
= 0; j
< XVECLEN (x
, i
); j
++)
5680 find_constant_pool_ref (XVECEXP (x
, i
, j
), ref
);
5685 /* Replace every reference to the annotated literal pool
5686 symbol REF in X by its base plus OFFSET. */
5689 replace_constant_pool_ref (rtx
*x
, rtx ref
, rtx offset
)
5694 gcc_assert (*x
!= ref
);
5696 if (GET_CODE (*x
) == UNSPEC
5697 && XINT (*x
, 1) == UNSPEC_LTREF
5698 && XVECEXP (*x
, 0, 0) == ref
)
5700 *x
= gen_rtx_PLUS (Pmode
, XVECEXP (*x
, 0, 1), offset
);
5704 if (GET_CODE (*x
) == PLUS
5705 && GET_CODE (XEXP (*x
, 1)) == CONST_INT
5706 && GET_CODE (XEXP (*x
, 0)) == UNSPEC
5707 && XINT (XEXP (*x
, 0), 1) == UNSPEC_LTREF
5708 && XVECEXP (XEXP (*x
, 0), 0, 0) == ref
)
5710 rtx addr
= gen_rtx_PLUS (Pmode
, XVECEXP (XEXP (*x
, 0), 0, 1), offset
);
5711 *x
= plus_constant (addr
, INTVAL (XEXP (*x
, 1)));
5715 fmt
= GET_RTX_FORMAT (GET_CODE (*x
));
5716 for (i
= GET_RTX_LENGTH (GET_CODE (*x
)) - 1; i
>= 0; i
--)
5720 replace_constant_pool_ref (&XEXP (*x
, i
), ref
, offset
);
5722 else if (fmt
[i
] == 'E')
5724 for (j
= 0; j
< XVECLEN (*x
, i
); j
++)
5725 replace_constant_pool_ref (&XVECEXP (*x
, i
, j
), ref
, offset
);
5730 /* Check whether X contains an UNSPEC_LTREL_BASE.
5731 Return its constant pool symbol if found, NULL_RTX otherwise. */
5734 find_ltrel_base (rtx x
)
5739 if (GET_CODE (x
) == UNSPEC
5740 && XINT (x
, 1) == UNSPEC_LTREL_BASE
)
5741 return XVECEXP (x
, 0, 0);
5743 fmt
= GET_RTX_FORMAT (GET_CODE (x
));
5744 for (i
= GET_RTX_LENGTH (GET_CODE (x
)) - 1; i
>= 0; i
--)
5748 rtx fnd
= find_ltrel_base (XEXP (x
, i
));
5752 else if (fmt
[i
] == 'E')
5754 for (j
= 0; j
< XVECLEN (x
, i
); j
++)
5756 rtx fnd
= find_ltrel_base (XVECEXP (x
, i
, j
));
5766 /* Replace any occurrence of UNSPEC_LTREL_BASE in X with its base. */
5769 replace_ltrel_base (rtx
*x
)
5774 if (GET_CODE (*x
) == UNSPEC
5775 && XINT (*x
, 1) == UNSPEC_LTREL_BASE
)
5777 *x
= XVECEXP (*x
, 0, 1);
5781 fmt
= GET_RTX_FORMAT (GET_CODE (*x
));
5782 for (i
= GET_RTX_LENGTH (GET_CODE (*x
)) - 1; i
>= 0; i
--)
5786 replace_ltrel_base (&XEXP (*x
, i
));
5788 else if (fmt
[i
] == 'E')
5790 for (j
= 0; j
< XVECLEN (*x
, i
); j
++)
5791 replace_ltrel_base (&XVECEXP (*x
, i
, j
));
5797 /* We keep a list of constants which we have to add to internal
5798 constant tables in the middle of large functions. */
5800 #define NR_C_MODES 11
5801 enum machine_mode constant_modes
[NR_C_MODES
] =
5803 TFmode
, TImode
, TDmode
,
5804 DFmode
, DImode
, DDmode
,
5805 SFmode
, SImode
, SDmode
,
5812 struct constant
*next
;
5817 struct constant_pool
5819 struct constant_pool
*next
;
5823 rtx emit_pool_after
;
5825 struct constant
*constants
[NR_C_MODES
];
5826 struct constant
*execute
;
5831 /* Allocate new constant_pool structure. */
5833 static struct constant_pool
*
5834 s390_alloc_pool (void)
5836 struct constant_pool
*pool
;
5839 pool
= (struct constant_pool
*) xmalloc (sizeof *pool
);
5841 for (i
= 0; i
< NR_C_MODES
; i
++)
5842 pool
->constants
[i
] = NULL
;
5844 pool
->execute
= NULL
;
5845 pool
->label
= gen_label_rtx ();
5846 pool
->first_insn
= NULL_RTX
;
5847 pool
->pool_insn
= NULL_RTX
;
5848 pool
->insns
= BITMAP_ALLOC (NULL
);
5850 pool
->emit_pool_after
= NULL_RTX
;
5855 /* Create new constant pool covering instructions starting at INSN
5856 and chain it to the end of POOL_LIST. */
5858 static struct constant_pool
*
5859 s390_start_pool (struct constant_pool
**pool_list
, rtx insn
)
5861 struct constant_pool
*pool
, **prev
;
5863 pool
= s390_alloc_pool ();
5864 pool
->first_insn
= insn
;
5866 for (prev
= pool_list
; *prev
; prev
= &(*prev
)->next
)
5873 /* End range of instructions covered by POOL at INSN and emit
5874 placeholder insn representing the pool. */
5877 s390_end_pool (struct constant_pool
*pool
, rtx insn
)
5879 rtx pool_size
= GEN_INT (pool
->size
+ 8 /* alignment slop */);
5882 insn
= get_last_insn ();
5884 pool
->pool_insn
= emit_insn_after (gen_pool (pool_size
), insn
);
5885 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
5888 /* Add INSN to the list of insns covered by POOL. */
5891 s390_add_pool_insn (struct constant_pool
*pool
, rtx insn
)
5893 bitmap_set_bit (pool
->insns
, INSN_UID (insn
));
5896 /* Return pool out of POOL_LIST that covers INSN. */
5898 static struct constant_pool
*
5899 s390_find_pool (struct constant_pool
*pool_list
, rtx insn
)
5901 struct constant_pool
*pool
;
5903 for (pool
= pool_list
; pool
; pool
= pool
->next
)
5904 if (bitmap_bit_p (pool
->insns
, INSN_UID (insn
)))
5910 /* Add constant VAL of mode MODE to the constant pool POOL. */
5913 s390_add_constant (struct constant_pool
*pool
, rtx val
, enum machine_mode mode
)
5918 for (i
= 0; i
< NR_C_MODES
; i
++)
5919 if (constant_modes
[i
] == mode
)
5921 gcc_assert (i
!= NR_C_MODES
);
5923 for (c
= pool
->constants
[i
]; c
!= NULL
; c
= c
->next
)
5924 if (rtx_equal_p (val
, c
->value
))
5929 c
= (struct constant
*) xmalloc (sizeof *c
);
5931 c
->label
= gen_label_rtx ();
5932 c
->next
= pool
->constants
[i
];
5933 pool
->constants
[i
] = c
;
5934 pool
->size
+= GET_MODE_SIZE (mode
);
5938 /* Return an rtx that represents the offset of X from the start of
5942 s390_pool_offset (struct constant_pool
*pool
, rtx x
)
5946 label
= gen_rtx_LABEL_REF (GET_MODE (x
), pool
->label
);
5947 x
= gen_rtx_UNSPEC (GET_MODE (x
), gen_rtvec (2, x
, label
),
5948 UNSPEC_POOL_OFFSET
);
5949 return gen_rtx_CONST (GET_MODE (x
), x
);
5952 /* Find constant VAL of mode MODE in the constant pool POOL.
5953 Return an RTX describing the distance from the start of
5954 the pool to the location of the new constant. */
5957 s390_find_constant (struct constant_pool
*pool
, rtx val
,
5958 enum machine_mode mode
)
5963 for (i
= 0; i
< NR_C_MODES
; i
++)
5964 if (constant_modes
[i
] == mode
)
5966 gcc_assert (i
!= NR_C_MODES
);
5968 for (c
= pool
->constants
[i
]; c
!= NULL
; c
= c
->next
)
5969 if (rtx_equal_p (val
, c
->value
))
5974 return s390_pool_offset (pool
, gen_rtx_LABEL_REF (Pmode
, c
->label
));
5977 /* Check whether INSN is an execute. Return the label_ref to its
5978 execute target template if so, NULL_RTX otherwise. */
5981 s390_execute_label (rtx insn
)
5983 if (GET_CODE (insn
) == INSN
5984 && GET_CODE (PATTERN (insn
)) == PARALLEL
5985 && GET_CODE (XVECEXP (PATTERN (insn
), 0, 0)) == UNSPEC
5986 && XINT (XVECEXP (PATTERN (insn
), 0, 0), 1) == UNSPEC_EXECUTE
)
5987 return XVECEXP (XVECEXP (PATTERN (insn
), 0, 0), 0, 2);
5992 /* Add execute target for INSN to the constant pool POOL. */
5995 s390_add_execute (struct constant_pool
*pool
, rtx insn
)
5999 for (c
= pool
->execute
; c
!= NULL
; c
= c
->next
)
6000 if (INSN_UID (insn
) == INSN_UID (c
->value
))
6005 c
= (struct constant
*) xmalloc (sizeof *c
);
6007 c
->label
= gen_label_rtx ();
6008 c
->next
= pool
->execute
;
6014 /* Find execute target for INSN in the constant pool POOL.
6015 Return an RTX describing the distance from the start of
6016 the pool to the location of the execute target. */
6019 s390_find_execute (struct constant_pool
*pool
, rtx insn
)
6023 for (c
= pool
->execute
; c
!= NULL
; c
= c
->next
)
6024 if (INSN_UID (insn
) == INSN_UID (c
->value
))
6029 return s390_pool_offset (pool
, gen_rtx_LABEL_REF (Pmode
, c
->label
));
6032 /* For an execute INSN, extract the execute target template. */
6035 s390_execute_target (rtx insn
)
6037 rtx pattern
= PATTERN (insn
);
6038 gcc_assert (s390_execute_label (insn
));
6040 if (XVECLEN (pattern
, 0) == 2)
6042 pattern
= copy_rtx (XVECEXP (pattern
, 0, 1));
6046 rtvec vec
= rtvec_alloc (XVECLEN (pattern
, 0) - 1);
6049 for (i
= 0; i
< XVECLEN (pattern
, 0) - 1; i
++)
6050 RTVEC_ELT (vec
, i
) = copy_rtx (XVECEXP (pattern
, 0, i
+ 1));
6052 pattern
= gen_rtx_PARALLEL (VOIDmode
, vec
);
6058 /* Indicate that INSN cannot be duplicated. This is the case for
6059 execute insns that carry a unique label. */
6062 s390_cannot_copy_insn_p (rtx insn
)
6064 rtx label
= s390_execute_label (insn
);
6065 return label
&& label
!= const0_rtx
;
6068 /* Dump out the constants in POOL. If REMOTE_LABEL is true,
6069 do not emit the pool base label. */
6072 s390_dump_pool (struct constant_pool
*pool
, bool remote_label
)
6075 rtx insn
= pool
->pool_insn
;
6078 /* Switch to rodata section. */
6079 if (TARGET_CPU_ZARCH
)
6081 insn
= emit_insn_after (gen_pool_section_start (), insn
);
6082 INSN_ADDRESSES_NEW (insn
, -1);
6085 /* Ensure minimum pool alignment. */
6086 if (TARGET_CPU_ZARCH
)
6087 insn
= emit_insn_after (gen_pool_align (GEN_INT (8)), insn
);
6089 insn
= emit_insn_after (gen_pool_align (GEN_INT (4)), insn
);
6090 INSN_ADDRESSES_NEW (insn
, -1);
6092 /* Emit pool base label. */
6095 insn
= emit_label_after (pool
->label
, insn
);
6096 INSN_ADDRESSES_NEW (insn
, -1);
6099 /* Dump constants in descending alignment requirement order,
6100 ensuring proper alignment for every constant. */
6101 for (i
= 0; i
< NR_C_MODES
; i
++)
6102 for (c
= pool
->constants
[i
]; c
; c
= c
->next
)
6104 /* Convert UNSPEC_LTREL_OFFSET unspecs to pool-relative references. */
6105 rtx value
= copy_rtx (c
->value
);
6106 if (GET_CODE (value
) == CONST
6107 && GET_CODE (XEXP (value
, 0)) == UNSPEC
6108 && XINT (XEXP (value
, 0), 1) == UNSPEC_LTREL_OFFSET
6109 && XVECLEN (XEXP (value
, 0), 0) == 1)
6110 value
= s390_pool_offset (pool
, XVECEXP (XEXP (value
, 0), 0, 0));
6112 insn
= emit_label_after (c
->label
, insn
);
6113 INSN_ADDRESSES_NEW (insn
, -1);
6115 value
= gen_rtx_UNSPEC_VOLATILE (constant_modes
[i
],
6116 gen_rtvec (1, value
),
6117 UNSPECV_POOL_ENTRY
);
6118 insn
= emit_insn_after (value
, insn
);
6119 INSN_ADDRESSES_NEW (insn
, -1);
6122 /* Ensure minimum alignment for instructions. */
6123 insn
= emit_insn_after (gen_pool_align (GEN_INT (2)), insn
);
6124 INSN_ADDRESSES_NEW (insn
, -1);
6126 /* Output in-pool execute template insns. */
6127 for (c
= pool
->execute
; c
; c
= c
->next
)
6129 insn
= emit_label_after (c
->label
, insn
);
6130 INSN_ADDRESSES_NEW (insn
, -1);
6132 insn
= emit_insn_after (s390_execute_target (c
->value
), insn
);
6133 INSN_ADDRESSES_NEW (insn
, -1);
6136 /* Switch back to previous section. */
6137 if (TARGET_CPU_ZARCH
)
6139 insn
= emit_insn_after (gen_pool_section_end (), insn
);
6140 INSN_ADDRESSES_NEW (insn
, -1);
6143 insn
= emit_barrier_after (insn
);
6144 INSN_ADDRESSES_NEW (insn
, -1);
6146 /* Remove placeholder insn. */
6147 remove_insn (pool
->pool_insn
);
6150 /* Free all memory used by POOL. */
6153 s390_free_pool (struct constant_pool
*pool
)
6155 struct constant
*c
, *next
;
6158 for (i
= 0; i
< NR_C_MODES
; i
++)
6159 for (c
= pool
->constants
[i
]; c
; c
= next
)
6165 for (c
= pool
->execute
; c
; c
= next
)
6171 BITMAP_FREE (pool
->insns
);
6176 /* Collect main literal pool. Return NULL on overflow. */
6178 static struct constant_pool
*
6179 s390_mainpool_start (void)
6181 struct constant_pool
*pool
;
6184 pool
= s390_alloc_pool ();
6186 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6188 if (GET_CODE (insn
) == INSN
6189 && GET_CODE (PATTERN (insn
)) == SET
6190 && GET_CODE (SET_SRC (PATTERN (insn
))) == UNSPEC_VOLATILE
6191 && XINT (SET_SRC (PATTERN (insn
)), 1) == UNSPECV_MAIN_POOL
)
6193 gcc_assert (!pool
->pool_insn
);
6194 pool
->pool_insn
= insn
;
6197 if (!TARGET_CPU_ZARCH
&& s390_execute_label (insn
))
6199 s390_add_execute (pool
, insn
);
6201 else if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6203 rtx pool_ref
= NULL_RTX
;
6204 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6207 rtx constant
= get_pool_constant (pool_ref
);
6208 enum machine_mode mode
= get_pool_mode (pool_ref
);
6209 s390_add_constant (pool
, constant
, mode
);
6213 /* If hot/cold partitioning is enabled we have to make sure that
6214 the literal pool is emitted in the same section where the
6215 initialization of the literal pool base pointer takes place.
6216 emit_pool_after is only used in the non-overflow case on non
6217 Z cpus where we can emit the literal pool at the end of the
6218 function body within the text section. */
6220 && NOTE_KIND (insn
) == NOTE_INSN_SWITCH_TEXT_SECTIONS
6221 && !pool
->emit_pool_after
)
6222 pool
->emit_pool_after
= PREV_INSN (insn
);
6225 gcc_assert (pool
->pool_insn
|| pool
->size
== 0);
6227 if (pool
->size
>= 4096)
6229 /* We're going to chunkify the pool, so remove the main
6230 pool placeholder insn. */
6231 remove_insn (pool
->pool_insn
);
6233 s390_free_pool (pool
);
6237 /* If the functions ends with the section where the literal pool
6238 should be emitted set the marker to its end. */
6239 if (pool
&& !pool
->emit_pool_after
)
6240 pool
->emit_pool_after
= get_last_insn ();
6245 /* POOL holds the main literal pool as collected by s390_mainpool_start.
6246 Modify the current function to output the pool constants as well as
6247 the pool register setup instruction. */
6250 s390_mainpool_finish (struct constant_pool
*pool
)
6252 rtx base_reg
= cfun
->machine
->base_reg
;
6255 /* If the pool is empty, we're done. */
6256 if (pool
->size
== 0)
6258 /* We don't actually need a base register after all. */
6259 cfun
->machine
->base_reg
= NULL_RTX
;
6261 if (pool
->pool_insn
)
6262 remove_insn (pool
->pool_insn
);
6263 s390_free_pool (pool
);
6267 /* We need correct insn addresses. */
6268 shorten_branches (get_insns ());
6270 /* On zSeries, we use a LARL to load the pool register. The pool is
6271 located in the .rodata section, so we emit it after the function. */
6272 if (TARGET_CPU_ZARCH
)
6274 insn
= gen_main_base_64 (base_reg
, pool
->label
);
6275 insn
= emit_insn_after (insn
, pool
->pool_insn
);
6276 INSN_ADDRESSES_NEW (insn
, -1);
6277 remove_insn (pool
->pool_insn
);
6279 insn
= get_last_insn ();
6280 pool
->pool_insn
= emit_insn_after (gen_pool (const0_rtx
), insn
);
6281 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
6283 s390_dump_pool (pool
, 0);
6286 /* On S/390, if the total size of the function's code plus literal pool
6287 does not exceed 4096 bytes, we use BASR to set up a function base
6288 pointer, and emit the literal pool at the end of the function. */
6289 else if (INSN_ADDRESSES (INSN_UID (pool
->emit_pool_after
))
6290 + pool
->size
+ 8 /* alignment slop */ < 4096)
6292 insn
= gen_main_base_31_small (base_reg
, pool
->label
);
6293 insn
= emit_insn_after (insn
, pool
->pool_insn
);
6294 INSN_ADDRESSES_NEW (insn
, -1);
6295 remove_insn (pool
->pool_insn
);
6297 insn
= emit_label_after (pool
->label
, insn
);
6298 INSN_ADDRESSES_NEW (insn
, -1);
6300 /* emit_pool_after will be set by s390_mainpool_start to the
6301 last insn of the section where the literal pool should be
6303 insn
= pool
->emit_pool_after
;
6305 pool
->pool_insn
= emit_insn_after (gen_pool (const0_rtx
), insn
);
6306 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
6308 s390_dump_pool (pool
, 1);
6311 /* Otherwise, we emit an inline literal pool and use BASR to branch
6312 over it, setting up the pool register at the same time. */
6315 rtx pool_end
= gen_label_rtx ();
6317 insn
= gen_main_base_31_large (base_reg
, pool
->label
, pool_end
);
6318 insn
= emit_insn_after (insn
, pool
->pool_insn
);
6319 INSN_ADDRESSES_NEW (insn
, -1);
6320 remove_insn (pool
->pool_insn
);
6322 insn
= emit_label_after (pool
->label
, insn
);
6323 INSN_ADDRESSES_NEW (insn
, -1);
6325 pool
->pool_insn
= emit_insn_after (gen_pool (const0_rtx
), insn
);
6326 INSN_ADDRESSES_NEW (pool
->pool_insn
, -1);
6328 insn
= emit_label_after (pool_end
, pool
->pool_insn
);
6329 INSN_ADDRESSES_NEW (insn
, -1);
6331 s390_dump_pool (pool
, 1);
6335 /* Replace all literal pool references. */
6337 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6340 replace_ltrel_base (&PATTERN (insn
));
6342 if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6344 rtx addr
, pool_ref
= NULL_RTX
;
6345 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6348 if (s390_execute_label (insn
))
6349 addr
= s390_find_execute (pool
, insn
);
6351 addr
= s390_find_constant (pool
, get_pool_constant (pool_ref
),
6352 get_pool_mode (pool_ref
));
6354 replace_constant_pool_ref (&PATTERN (insn
), pool_ref
, addr
);
6355 INSN_CODE (insn
) = -1;
6361 /* Free the pool. */
6362 s390_free_pool (pool
);
6365 /* POOL holds the main literal pool as collected by s390_mainpool_start.
6366 We have decided we cannot use this pool, so revert all changes
6367 to the current function that were done by s390_mainpool_start. */
6369 s390_mainpool_cancel (struct constant_pool
*pool
)
6371 /* We didn't actually change the instruction stream, so simply
6372 free the pool memory. */
6373 s390_free_pool (pool
);
6377 /* Chunkify the literal pool. */
6379 #define S390_POOL_CHUNK_MIN 0xc00
6380 #define S390_POOL_CHUNK_MAX 0xe00
6382 static struct constant_pool
*
6383 s390_chunkify_start (void)
6385 struct constant_pool
*curr_pool
= NULL
, *pool_list
= NULL
;
6388 rtx pending_ltrel
= NULL_RTX
;
6391 rtx (*gen_reload_base
) (rtx
, rtx
) =
6392 TARGET_CPU_ZARCH
? gen_reload_base_64
: gen_reload_base_31
;
6395 /* We need correct insn addresses. */
6397 shorten_branches (get_insns ());
6399 /* Scan all insns and move literals to pool chunks. */
6401 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6403 bool section_switch_p
= false;
6405 /* Check for pending LTREL_BASE. */
6408 rtx ltrel_base
= find_ltrel_base (PATTERN (insn
));
6411 gcc_assert (ltrel_base
== pending_ltrel
);
6412 pending_ltrel
= NULL_RTX
;
6416 if (!TARGET_CPU_ZARCH
&& s390_execute_label (insn
))
6419 curr_pool
= s390_start_pool (&pool_list
, insn
);
6421 s390_add_execute (curr_pool
, insn
);
6422 s390_add_pool_insn (curr_pool
, insn
);
6424 else if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6426 rtx pool_ref
= NULL_RTX
;
6427 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6430 rtx constant
= get_pool_constant (pool_ref
);
6431 enum machine_mode mode
= get_pool_mode (pool_ref
);
6434 curr_pool
= s390_start_pool (&pool_list
, insn
);
6436 s390_add_constant (curr_pool
, constant
, mode
);
6437 s390_add_pool_insn (curr_pool
, insn
);
6439 /* Don't split the pool chunk between a LTREL_OFFSET load
6440 and the corresponding LTREL_BASE. */
6441 if (GET_CODE (constant
) == CONST
6442 && GET_CODE (XEXP (constant
, 0)) == UNSPEC
6443 && XINT (XEXP (constant
, 0), 1) == UNSPEC_LTREL_OFFSET
)
6445 gcc_assert (!pending_ltrel
);
6446 pending_ltrel
= pool_ref
;
6451 if (GET_CODE (insn
) == JUMP_INSN
|| GET_CODE (insn
) == CODE_LABEL
)
6454 s390_add_pool_insn (curr_pool
, insn
);
6455 /* An LTREL_BASE must follow within the same basic block. */
6456 gcc_assert (!pending_ltrel
);
6459 if (NOTE_P (insn
) && NOTE_KIND (insn
) == NOTE_INSN_SWITCH_TEXT_SECTIONS
)
6460 section_switch_p
= true;
6463 || INSN_ADDRESSES_SIZE () <= (size_t) INSN_UID (insn
)
6464 || INSN_ADDRESSES (INSN_UID (insn
)) == -1)
6467 if (TARGET_CPU_ZARCH
)
6469 if (curr_pool
->size
< S390_POOL_CHUNK_MAX
)
6472 s390_end_pool (curr_pool
, NULL_RTX
);
6477 int chunk_size
= INSN_ADDRESSES (INSN_UID (insn
))
6478 - INSN_ADDRESSES (INSN_UID (curr_pool
->first_insn
))
6481 /* We will later have to insert base register reload insns.
6482 Those will have an effect on code size, which we need to
6483 consider here. This calculation makes rather pessimistic
6484 worst-case assumptions. */
6485 if (GET_CODE (insn
) == CODE_LABEL
)
6488 if (chunk_size
< S390_POOL_CHUNK_MIN
6489 && curr_pool
->size
< S390_POOL_CHUNK_MIN
6490 && !section_switch_p
)
6493 /* Pool chunks can only be inserted after BARRIERs ... */
6494 if (GET_CODE (insn
) == BARRIER
)
6496 s390_end_pool (curr_pool
, insn
);
6501 /* ... so if we don't find one in time, create one. */
6502 else if (chunk_size
> S390_POOL_CHUNK_MAX
6503 || curr_pool
->size
> S390_POOL_CHUNK_MAX
6504 || section_switch_p
)
6506 rtx label
, jump
, barrier
;
6508 if (!section_switch_p
)
6510 /* We can insert the barrier only after a 'real' insn. */
6511 if (GET_CODE (insn
) != INSN
&& GET_CODE (insn
) != CALL_INSN
)
6513 if (get_attr_length (insn
) == 0)
6515 /* Don't separate LTREL_BASE from the corresponding
6516 LTREL_OFFSET load. */
6522 gcc_assert (!pending_ltrel
);
6524 /* The old pool has to end before the section switch
6525 note in order to make it part of the current
6527 insn
= PREV_INSN (insn
);
6530 label
= gen_label_rtx ();
6531 jump
= emit_jump_insn_after (gen_jump (label
), insn
);
6532 barrier
= emit_barrier_after (jump
);
6533 insn
= emit_label_after (label
, barrier
);
6534 JUMP_LABEL (jump
) = label
;
6535 LABEL_NUSES (label
) = 1;
6537 INSN_ADDRESSES_NEW (jump
, -1);
6538 INSN_ADDRESSES_NEW (barrier
, -1);
6539 INSN_ADDRESSES_NEW (insn
, -1);
6541 s390_end_pool (curr_pool
, barrier
);
6549 s390_end_pool (curr_pool
, NULL_RTX
);
6550 gcc_assert (!pending_ltrel
);
6552 /* Find all labels that are branched into
6553 from an insn belonging to a different chunk. */
6555 far_labels
= BITMAP_ALLOC (NULL
);
6557 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6559 /* Labels marked with LABEL_PRESERVE_P can be target
6560 of non-local jumps, so we have to mark them.
6561 The same holds for named labels.
6563 Don't do that, however, if it is the label before
6566 if (GET_CODE (insn
) == CODE_LABEL
6567 && (LABEL_PRESERVE_P (insn
) || LABEL_NAME (insn
)))
6569 rtx vec_insn
= next_real_insn (insn
);
6570 rtx vec_pat
= vec_insn
&& GET_CODE (vec_insn
) == JUMP_INSN
?
6571 PATTERN (vec_insn
) : NULL_RTX
;
6573 || !(GET_CODE (vec_pat
) == ADDR_VEC
6574 || GET_CODE (vec_pat
) == ADDR_DIFF_VEC
))
6575 bitmap_set_bit (far_labels
, CODE_LABEL_NUMBER (insn
));
6578 /* If we have a direct jump (conditional or unconditional)
6579 or a casesi jump, check all potential targets. */
6580 else if (GET_CODE (insn
) == JUMP_INSN
)
6582 rtx pat
= PATTERN (insn
);
6583 if (GET_CODE (pat
) == PARALLEL
&& XVECLEN (pat
, 0) > 2)
6584 pat
= XVECEXP (pat
, 0, 0);
6586 if (GET_CODE (pat
) == SET
)
6588 rtx label
= JUMP_LABEL (insn
);
6591 if (s390_find_pool (pool_list
, label
)
6592 != s390_find_pool (pool_list
, insn
))
6593 bitmap_set_bit (far_labels
, CODE_LABEL_NUMBER (label
));
6596 else if (GET_CODE (pat
) == PARALLEL
6597 && XVECLEN (pat
, 0) == 2
6598 && GET_CODE (XVECEXP (pat
, 0, 0)) == SET
6599 && GET_CODE (XVECEXP (pat
, 0, 1)) == USE
6600 && GET_CODE (XEXP (XVECEXP (pat
, 0, 1), 0)) == LABEL_REF
)
6602 /* Find the jump table used by this casesi jump. */
6603 rtx vec_label
= XEXP (XEXP (XVECEXP (pat
, 0, 1), 0), 0);
6604 rtx vec_insn
= next_real_insn (vec_label
);
6605 rtx vec_pat
= vec_insn
&& GET_CODE (vec_insn
) == JUMP_INSN
?
6606 PATTERN (vec_insn
) : NULL_RTX
;
6608 && (GET_CODE (vec_pat
) == ADDR_VEC
6609 || GET_CODE (vec_pat
) == ADDR_DIFF_VEC
))
6611 int i
, diff_p
= GET_CODE (vec_pat
) == ADDR_DIFF_VEC
;
6613 for (i
= 0; i
< XVECLEN (vec_pat
, diff_p
); i
++)
6615 rtx label
= XEXP (XVECEXP (vec_pat
, diff_p
, i
), 0);
6617 if (s390_find_pool (pool_list
, label
)
6618 != s390_find_pool (pool_list
, insn
))
6619 bitmap_set_bit (far_labels
, CODE_LABEL_NUMBER (label
));
6626 /* Insert base register reload insns before every pool. */
6628 for (curr_pool
= pool_list
; curr_pool
; curr_pool
= curr_pool
->next
)
6630 rtx new_insn
= gen_reload_base (cfun
->machine
->base_reg
,
6632 rtx insn
= curr_pool
->first_insn
;
6633 INSN_ADDRESSES_NEW (emit_insn_before (new_insn
, insn
), -1);
6636 /* Insert base register reload insns at every far label. */
6638 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6639 if (GET_CODE (insn
) == CODE_LABEL
6640 && bitmap_bit_p (far_labels
, CODE_LABEL_NUMBER (insn
)))
6642 struct constant_pool
*pool
= s390_find_pool (pool_list
, insn
);
6645 rtx new_insn
= gen_reload_base (cfun
->machine
->base_reg
,
6647 INSN_ADDRESSES_NEW (emit_insn_after (new_insn
, insn
), -1);
6652 BITMAP_FREE (far_labels
);
6655 /* Recompute insn addresses. */
6657 init_insn_lengths ();
6658 shorten_branches (get_insns ());
6663 /* POOL_LIST is a chunk list as prepared by s390_chunkify_start.
6664 After we have decided to use this list, finish implementing
6665 all changes to the current function as required. */
6668 s390_chunkify_finish (struct constant_pool
*pool_list
)
6670 struct constant_pool
*curr_pool
= NULL
;
6674 /* Replace all literal pool references. */
6676 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
6679 replace_ltrel_base (&PATTERN (insn
));
6681 curr_pool
= s390_find_pool (pool_list
, insn
);
6685 if (GET_CODE (insn
) == INSN
|| GET_CODE (insn
) == CALL_INSN
)
6687 rtx addr
, pool_ref
= NULL_RTX
;
6688 find_constant_pool_ref (PATTERN (insn
), &pool_ref
);
6691 if (s390_execute_label (insn
))
6692 addr
= s390_find_execute (curr_pool
, insn
);
6694 addr
= s390_find_constant (curr_pool
,
6695 get_pool_constant (pool_ref
),
6696 get_pool_mode (pool_ref
));
6698 replace_constant_pool_ref (&PATTERN (insn
), pool_ref
, addr
);
6699 INSN_CODE (insn
) = -1;
6704 /* Dump out all literal pools. */
6706 for (curr_pool
= pool_list
; curr_pool
; curr_pool
= curr_pool
->next
)
6707 s390_dump_pool (curr_pool
, 0);
6709 /* Free pool list. */
6713 struct constant_pool
*next
= pool_list
->next
;
6714 s390_free_pool (pool_list
);
6719 /* POOL_LIST is a chunk list as prepared by s390_chunkify_start.
6720 We have decided we cannot use this list, so revert all changes
6721 to the current function that were done by s390_chunkify_start. */
6724 s390_chunkify_cancel (struct constant_pool
*pool_list
)
6726 struct constant_pool
*curr_pool
= NULL
;
6729 /* Remove all pool placeholder insns. */
6731 for (curr_pool
= pool_list
; curr_pool
; curr_pool
= curr_pool
->next
)
6733 /* Did we insert an extra barrier? Remove it. */
6734 rtx barrier
= PREV_INSN (curr_pool
->pool_insn
);
6735 rtx jump
= barrier
? PREV_INSN (barrier
) : NULL_RTX
;
6736 rtx label
= NEXT_INSN (curr_pool
->pool_insn
);
6738 if (jump
&& GET_CODE (jump
) == JUMP_INSN
6739 && barrier
&& GET_CODE (barrier
) == BARRIER
6740 && label
&& GET_CODE (label
) == CODE_LABEL
6741 && GET_CODE (PATTERN (jump
)) == SET
6742 && SET_DEST (PATTERN (jump
)) == pc_rtx
6743 && GET_CODE (SET_SRC (PATTERN (jump
))) == LABEL_REF
6744 && XEXP (SET_SRC (PATTERN (jump
)), 0) == label
)
6747 remove_insn (barrier
);
6748 remove_insn (label
);
6751 remove_insn (curr_pool
->pool_insn
);
6754 /* Remove all base register reload insns. */
6756 for (insn
= get_insns (); insn
; )
6758 rtx next_insn
= NEXT_INSN (insn
);
6760 if (GET_CODE (insn
) == INSN
6761 && GET_CODE (PATTERN (insn
)) == SET
6762 && GET_CODE (SET_SRC (PATTERN (insn
))) == UNSPEC
6763 && XINT (SET_SRC (PATTERN (insn
)), 1) == UNSPEC_RELOAD_BASE
)
6769 /* Free pool list. */
6773 struct constant_pool
*next
= pool_list
->next
;
6774 s390_free_pool (pool_list
);
6779 /* Output the constant pool entry EXP in mode MODE with alignment ALIGN. */
6782 s390_output_pool_entry (rtx exp
, enum machine_mode mode
, unsigned int align
)
6786 switch (GET_MODE_CLASS (mode
))
6789 case MODE_DECIMAL_FLOAT
:
6790 gcc_assert (GET_CODE (exp
) == CONST_DOUBLE
);
6792 REAL_VALUE_FROM_CONST_DOUBLE (r
, exp
);
6793 assemble_real (r
, mode
, align
);
6797 assemble_integer (exp
, GET_MODE_SIZE (mode
), align
, 1);
6798 mark_symbol_refs_as_used (exp
);
6807 /* Return an RTL expression representing the value of the return address
6808 for the frame COUNT steps up from the current frame. FRAME is the
6809 frame pointer of that frame. */
6812 s390_return_addr_rtx (int count
, rtx frame ATTRIBUTE_UNUSED
)
6817 /* Without backchain, we fail for all but the current frame. */
6819 if (!TARGET_BACKCHAIN
&& count
> 0)
6822 /* For the current frame, we need to make sure the initial
6823 value of RETURN_REGNUM is actually saved. */
6827 /* On non-z architectures branch splitting could overwrite r14. */
6828 if (TARGET_CPU_ZARCH
)
6829 return get_hard_reg_initial_val (Pmode
, RETURN_REGNUM
);
6832 cfun_frame_layout
.save_return_addr_p
= true;
6833 return gen_rtx_MEM (Pmode
, return_address_pointer_rtx
);
6837 if (TARGET_PACKED_STACK
)
6838 offset
= -2 * UNITS_PER_LONG
;
6840 offset
= RETURN_REGNUM
* UNITS_PER_LONG
;
6842 addr
= plus_constant (frame
, offset
);
6843 addr
= memory_address (Pmode
, addr
);
6844 return gen_rtx_MEM (Pmode
, addr
);
6847 /* Return an RTL expression representing the back chain stored in
6848 the current stack frame. */
6851 s390_back_chain_rtx (void)
6855 gcc_assert (TARGET_BACKCHAIN
);
6857 if (TARGET_PACKED_STACK
)
6858 chain
= plus_constant (stack_pointer_rtx
,
6859 STACK_POINTER_OFFSET
- UNITS_PER_LONG
);
6861 chain
= stack_pointer_rtx
;
6863 chain
= gen_rtx_MEM (Pmode
, chain
);
6867 /* Find first call clobbered register unused in a function.
6868 This could be used as base register in a leaf function
6869 or for holding the return address before epilogue. */
6872 find_unused_clobbered_reg (void)
6875 for (i
= 0; i
< 6; i
++)
6876 if (!df_regs_ever_live_p (i
))
6882 /* Helper function for s390_regs_ever_clobbered. Sets the fields in DATA for all
6883 clobbered hard regs in SETREG. */
6886 s390_reg_clobbered_rtx (rtx setreg
, const_rtx set_insn ATTRIBUTE_UNUSED
, void *data
)
6888 int *regs_ever_clobbered
= (int *)data
;
6889 unsigned int i
, regno
;
6890 enum machine_mode mode
= GET_MODE (setreg
);
6892 if (GET_CODE (setreg
) == SUBREG
)
6894 rtx inner
= SUBREG_REG (setreg
);
6895 if (!GENERAL_REG_P (inner
))
6897 regno
= subreg_regno (setreg
);
6899 else if (GENERAL_REG_P (setreg
))
6900 regno
= REGNO (setreg
);
6905 i
< regno
+ HARD_REGNO_NREGS (regno
, mode
);
6907 regs_ever_clobbered
[i
] = 1;
6910 /* Walks through all basic blocks of the current function looking
6911 for clobbered hard regs using s390_reg_clobbered_rtx. The fields
6912 of the passed integer array REGS_EVER_CLOBBERED are set to one for
6913 each of those regs. */
6916 s390_regs_ever_clobbered (int *regs_ever_clobbered
)
6922 memset (regs_ever_clobbered
, 0, 16 * sizeof (int));
6924 /* For non-leaf functions we have to consider all call clobbered regs to be
6926 if (!current_function_is_leaf
)
6928 for (i
= 0; i
< 16; i
++)
6929 regs_ever_clobbered
[i
] = call_really_used_regs
[i
];
6932 /* Make the "magic" eh_return registers live if necessary. For regs_ever_live
6933 this work is done by liveness analysis (mark_regs_live_at_end).
6934 Special care is needed for functions containing landing pads. Landing pads
6935 may use the eh registers, but the code which sets these registers is not
6936 contained in that function. Hence s390_regs_ever_clobbered is not able to
6937 deal with this automatically. */
6938 if (crtl
->calls_eh_return
|| cfun
->machine
->has_landing_pad_p
)
6939 for (i
= 0; EH_RETURN_DATA_REGNO (i
) != INVALID_REGNUM
; i
++)
6940 if (crtl
->calls_eh_return
6941 || (cfun
->machine
->has_landing_pad_p
6942 && df_regs_ever_live_p (EH_RETURN_DATA_REGNO (i
))))
6943 regs_ever_clobbered
[EH_RETURN_DATA_REGNO (i
)] = 1;
6945 /* For nonlocal gotos all call-saved registers have to be saved.
6946 This flag is also set for the unwinding code in libgcc.
6947 See expand_builtin_unwind_init. For regs_ever_live this is done by
6949 if (cfun
->has_nonlocal_label
)
6950 for (i
= 0; i
< 16; i
++)
6951 if (!call_really_used_regs
[i
])
6952 regs_ever_clobbered
[i
] = 1;
6954 FOR_EACH_BB (cur_bb
)
6956 FOR_BB_INSNS (cur_bb
, cur_insn
)
6958 if (INSN_P (cur_insn
))
6959 note_stores (PATTERN (cur_insn
),
6960 s390_reg_clobbered_rtx
,
6961 regs_ever_clobbered
);
6966 /* Determine the frame area which actually has to be accessed
6967 in the function epilogue. The values are stored at the
6968 given pointers AREA_BOTTOM (address of the lowest used stack
6969 address) and AREA_TOP (address of the first item which does
6970 not belong to the stack frame). */
6973 s390_frame_area (int *area_bottom
, int *area_top
)
6981 if (cfun_frame_layout
.first_restore_gpr
!= -1)
6983 b
= (cfun_frame_layout
.gprs_offset
6984 + cfun_frame_layout
.first_restore_gpr
* UNITS_PER_LONG
);
6985 t
= b
+ (cfun_frame_layout
.last_restore_gpr
6986 - cfun_frame_layout
.first_restore_gpr
+ 1) * UNITS_PER_LONG
;
6989 if (TARGET_64BIT
&& cfun_save_high_fprs_p
)
6991 b
= MIN (b
, cfun_frame_layout
.f8_offset
);
6992 t
= MAX (t
, (cfun_frame_layout
.f8_offset
6993 + cfun_frame_layout
.high_fprs
* 8));
6997 for (i
= 2; i
< 4; i
++)
6998 if (cfun_fpr_bit_p (i
))
7000 b
= MIN (b
, cfun_frame_layout
.f4_offset
+ (i
- 2) * 8);
7001 t
= MAX (t
, cfun_frame_layout
.f4_offset
+ (i
- 1) * 8);
7008 /* Fill cfun->machine with info about register usage of current function.
7009 Return in CLOBBERED_REGS which GPRs are currently considered set. */
7012 s390_register_info (int clobbered_regs
[])
7016 /* fprs 8 - 15 are call saved for 64 Bit ABI. */
7017 cfun_frame_layout
.fpr_bitmap
= 0;
7018 cfun_frame_layout
.high_fprs
= 0;
7020 for (i
= 24; i
< 32; i
++)
7021 if (df_regs_ever_live_p (i
) && !global_regs
[i
])
7023 cfun_set_fpr_bit (i
- 16);
7024 cfun_frame_layout
.high_fprs
++;
7027 /* Find first and last gpr to be saved. We trust regs_ever_live
7028 data, except that we don't save and restore global registers.
7030 Also, all registers with special meaning to the compiler need
7031 to be handled extra. */
7033 s390_regs_ever_clobbered (clobbered_regs
);
7035 for (i
= 0; i
< 16; i
++)
7036 clobbered_regs
[i
] = clobbered_regs
[i
] && !global_regs
[i
] && !fixed_regs
[i
];
7038 if (frame_pointer_needed
)
7039 clobbered_regs
[HARD_FRAME_POINTER_REGNUM
] = 1;
7042 clobbered_regs
[PIC_OFFSET_TABLE_REGNUM
]
7043 |= df_regs_ever_live_p (PIC_OFFSET_TABLE_REGNUM
);
7045 clobbered_regs
[BASE_REGNUM
]
7046 |= (cfun
->machine
->base_reg
7047 && REGNO (cfun
->machine
->base_reg
) == BASE_REGNUM
);
7049 clobbered_regs
[RETURN_REGNUM
]
7050 |= (!current_function_is_leaf
7051 || TARGET_TPF_PROFILING
7052 || cfun
->machine
->split_branches_pending_p
7053 || cfun_frame_layout
.save_return_addr_p
7054 || crtl
->calls_eh_return
7057 clobbered_regs
[STACK_POINTER_REGNUM
]
7058 |= (!current_function_is_leaf
7059 || TARGET_TPF_PROFILING
7060 || cfun_save_high_fprs_p
7061 || get_frame_size () > 0
7062 || cfun
->calls_alloca
7065 for (i
= 6; i
< 16; i
++)
7066 if (df_regs_ever_live_p (i
) || clobbered_regs
[i
])
7068 for (j
= 15; j
> i
; j
--)
7069 if (df_regs_ever_live_p (j
) || clobbered_regs
[j
])
7074 /* Nothing to save/restore. */
7075 cfun_frame_layout
.first_save_gpr_slot
= -1;
7076 cfun_frame_layout
.last_save_gpr_slot
= -1;
7077 cfun_frame_layout
.first_save_gpr
= -1;
7078 cfun_frame_layout
.first_restore_gpr
= -1;
7079 cfun_frame_layout
.last_save_gpr
= -1;
7080 cfun_frame_layout
.last_restore_gpr
= -1;
7084 /* Save slots for gprs from i to j. */
7085 cfun_frame_layout
.first_save_gpr_slot
= i
;
7086 cfun_frame_layout
.last_save_gpr_slot
= j
;
7088 for (i
= cfun_frame_layout
.first_save_gpr_slot
;
7089 i
< cfun_frame_layout
.last_save_gpr_slot
+ 1;
7091 if (clobbered_regs
[i
])
7094 for (j
= cfun_frame_layout
.last_save_gpr_slot
; j
> i
; j
--)
7095 if (clobbered_regs
[j
])
7098 if (i
== cfun_frame_layout
.last_save_gpr_slot
+ 1)
7100 /* Nothing to save/restore. */
7101 cfun_frame_layout
.first_save_gpr
= -1;
7102 cfun_frame_layout
.first_restore_gpr
= -1;
7103 cfun_frame_layout
.last_save_gpr
= -1;
7104 cfun_frame_layout
.last_restore_gpr
= -1;
7108 /* Save / Restore from gpr i to j. */
7109 cfun_frame_layout
.first_save_gpr
= i
;
7110 cfun_frame_layout
.first_restore_gpr
= i
;
7111 cfun_frame_layout
.last_save_gpr
= j
;
7112 cfun_frame_layout
.last_restore_gpr
= j
;
7118 /* Varargs functions need to save gprs 2 to 6. */
7119 if (cfun
->va_list_gpr_size
7120 && crtl
->args
.info
.gprs
< GP_ARG_NUM_REG
)
7122 int min_gpr
= crtl
->args
.info
.gprs
;
7123 int max_gpr
= min_gpr
+ cfun
->va_list_gpr_size
;
7124 if (max_gpr
> GP_ARG_NUM_REG
)
7125 max_gpr
= GP_ARG_NUM_REG
;
7127 if (cfun_frame_layout
.first_save_gpr
== -1
7128 || cfun_frame_layout
.first_save_gpr
> 2 + min_gpr
)
7130 cfun_frame_layout
.first_save_gpr
= 2 + min_gpr
;
7131 cfun_frame_layout
.first_save_gpr_slot
= 2 + min_gpr
;
7134 if (cfun_frame_layout
.last_save_gpr
== -1
7135 || cfun_frame_layout
.last_save_gpr
< 2 + max_gpr
- 1)
7137 cfun_frame_layout
.last_save_gpr
= 2 + max_gpr
- 1;
7138 cfun_frame_layout
.last_save_gpr_slot
= 2 + max_gpr
- 1;
7142 /* Mark f0, f2 for 31 bit and f0-f4 for 64 bit to be saved. */
7143 if (TARGET_HARD_FLOAT
&& cfun
->va_list_fpr_size
7144 && crtl
->args
.info
.fprs
< FP_ARG_NUM_REG
)
7146 int min_fpr
= crtl
->args
.info
.fprs
;
7147 int max_fpr
= min_fpr
+ cfun
->va_list_fpr_size
;
7148 if (max_fpr
> FP_ARG_NUM_REG
)
7149 max_fpr
= FP_ARG_NUM_REG
;
7151 /* ??? This is currently required to ensure proper location
7152 of the fpr save slots within the va_list save area. */
7153 if (TARGET_PACKED_STACK
)
7156 for (i
= min_fpr
; i
< max_fpr
; i
++)
7157 cfun_set_fpr_bit (i
);
7162 for (i
= 2; i
< 4; i
++)
7163 if (df_regs_ever_live_p (i
+ 16) && !global_regs
[i
+ 16])
7164 cfun_set_fpr_bit (i
);
7167 /* Fill cfun->machine with info about frame of current function. */
7170 s390_frame_info (void)
7174 cfun_frame_layout
.frame_size
= get_frame_size ();
7175 if (!TARGET_64BIT
&& cfun_frame_layout
.frame_size
> 0x7fff0000)
7176 fatal_error ("total size of local variables exceeds architecture limit");
7178 if (!TARGET_PACKED_STACK
)
7180 cfun_frame_layout
.backchain_offset
= 0;
7181 cfun_frame_layout
.f0_offset
= 16 * UNITS_PER_LONG
;
7182 cfun_frame_layout
.f4_offset
= cfun_frame_layout
.f0_offset
+ 2 * 8;
7183 cfun_frame_layout
.f8_offset
= -cfun_frame_layout
.high_fprs
* 8;
7184 cfun_frame_layout
.gprs_offset
= (cfun_frame_layout
.first_save_gpr_slot
7187 else if (TARGET_BACKCHAIN
) /* kernel stack layout */
7189 cfun_frame_layout
.backchain_offset
= (STACK_POINTER_OFFSET
7191 cfun_frame_layout
.gprs_offset
7192 = (cfun_frame_layout
.backchain_offset
7193 - (STACK_POINTER_REGNUM
- cfun_frame_layout
.first_save_gpr_slot
+ 1)
7198 cfun_frame_layout
.f4_offset
7199 = (cfun_frame_layout
.gprs_offset
7200 - 8 * (cfun_fpr_bit_p (2) + cfun_fpr_bit_p (3)));
7202 cfun_frame_layout
.f0_offset
7203 = (cfun_frame_layout
.f4_offset
7204 - 8 * (cfun_fpr_bit_p (0) + cfun_fpr_bit_p (1)));
7208 /* On 31 bit we have to care about alignment of the
7209 floating point regs to provide fastest access. */
7210 cfun_frame_layout
.f0_offset
7211 = ((cfun_frame_layout
.gprs_offset
7212 & ~(STACK_BOUNDARY
/ BITS_PER_UNIT
- 1))
7213 - 8 * (cfun_fpr_bit_p (0) + cfun_fpr_bit_p (1)));
7215 cfun_frame_layout
.f4_offset
7216 = (cfun_frame_layout
.f0_offset
7217 - 8 * (cfun_fpr_bit_p (2) + cfun_fpr_bit_p (3)));
7220 else /* no backchain */
7222 cfun_frame_layout
.f4_offset
7223 = (STACK_POINTER_OFFSET
7224 - 8 * (cfun_fpr_bit_p (2) + cfun_fpr_bit_p (3)));
7226 cfun_frame_layout
.f0_offset
7227 = (cfun_frame_layout
.f4_offset
7228 - 8 * (cfun_fpr_bit_p (0) + cfun_fpr_bit_p (1)));
7230 cfun_frame_layout
.gprs_offset
7231 = cfun_frame_layout
.f0_offset
- cfun_gprs_save_area_size
;
7234 if (current_function_is_leaf
7235 && !TARGET_TPF_PROFILING
7236 && cfun_frame_layout
.frame_size
== 0
7237 && !cfun_save_high_fprs_p
7238 && !cfun
->calls_alloca
7242 if (!TARGET_PACKED_STACK
)
7243 cfun_frame_layout
.frame_size
+= (STACK_POINTER_OFFSET
7244 + crtl
->outgoing_args_size
7245 + cfun_frame_layout
.high_fprs
* 8);
7248 if (TARGET_BACKCHAIN
)
7249 cfun_frame_layout
.frame_size
+= UNITS_PER_LONG
;
7251 /* No alignment trouble here because f8-f15 are only saved under
7253 cfun_frame_layout
.f8_offset
= (MIN (MIN (cfun_frame_layout
.f0_offset
,
7254 cfun_frame_layout
.f4_offset
),
7255 cfun_frame_layout
.gprs_offset
)
7256 - cfun_frame_layout
.high_fprs
* 8);
7258 cfun_frame_layout
.frame_size
+= cfun_frame_layout
.high_fprs
* 8;
7260 for (i
= 0; i
< 8; i
++)
7261 if (cfun_fpr_bit_p (i
))
7262 cfun_frame_layout
.frame_size
+= 8;
7264 cfun_frame_layout
.frame_size
+= cfun_gprs_save_area_size
;
7266 /* If under 31 bit an odd number of gprs has to be saved we have to adjust
7267 the frame size to sustain 8 byte alignment of stack frames. */
7268 cfun_frame_layout
.frame_size
= ((cfun_frame_layout
.frame_size
+
7269 STACK_BOUNDARY
/ BITS_PER_UNIT
- 1)
7270 & ~(STACK_BOUNDARY
/ BITS_PER_UNIT
- 1));
7272 cfun_frame_layout
.frame_size
+= crtl
->outgoing_args_size
;
7276 /* Generate frame layout. Fills in register and frame data for the current
7277 function in cfun->machine. This routine can be called multiple times;
7278 it will re-do the complete frame layout every time. */
7281 s390_init_frame_layout (void)
7283 HOST_WIDE_INT frame_size
;
7285 int clobbered_regs
[16];
7287 /* On S/390 machines, we may need to perform branch splitting, which
7288 will require both base and return address register. We have no
7289 choice but to assume we're going to need them until right at the
7290 end of the machine dependent reorg phase. */
7291 if (!TARGET_CPU_ZARCH
)
7292 cfun
->machine
->split_branches_pending_p
= true;
7296 frame_size
= cfun_frame_layout
.frame_size
;
7298 /* Try to predict whether we'll need the base register. */
7299 base_used
= cfun
->machine
->split_branches_pending_p
7300 || crtl
->uses_const_pool
7301 || (!DISP_IN_RANGE (frame_size
)
7302 && !CONST_OK_FOR_K (frame_size
));
7304 /* Decide which register to use as literal pool base. In small
7305 leaf functions, try to use an unused call-clobbered register
7306 as base register to avoid save/restore overhead. */
7308 cfun
->machine
->base_reg
= NULL_RTX
;
7309 else if (current_function_is_leaf
&& !df_regs_ever_live_p (5))
7310 cfun
->machine
->base_reg
= gen_rtx_REG (Pmode
, 5);
7312 cfun
->machine
->base_reg
= gen_rtx_REG (Pmode
, BASE_REGNUM
);
7314 s390_register_info (clobbered_regs
);
7317 while (frame_size
!= cfun_frame_layout
.frame_size
);
7320 /* Update frame layout. Recompute actual register save data based on
7321 current info and update regs_ever_live for the special registers.
7322 May be called multiple times, but may never cause *more* registers
7323 to be saved than s390_init_frame_layout allocated room for. */
7326 s390_update_frame_layout (void)
7328 int clobbered_regs
[16];
7330 s390_register_info (clobbered_regs
);
7332 df_set_regs_ever_live (BASE_REGNUM
,
7333 clobbered_regs
[BASE_REGNUM
] ? true : false);
7334 df_set_regs_ever_live (RETURN_REGNUM
,
7335 clobbered_regs
[RETURN_REGNUM
] ? true : false);
7336 df_set_regs_ever_live (STACK_POINTER_REGNUM
,
7337 clobbered_regs
[STACK_POINTER_REGNUM
] ? true : false);
7339 if (cfun
->machine
->base_reg
)
7340 df_set_regs_ever_live (REGNO (cfun
->machine
->base_reg
), true);
7343 /* Return true if it is legal to put a value with MODE into REGNO. */
7346 s390_hard_regno_mode_ok (unsigned int regno
, enum machine_mode mode
)
7348 switch (REGNO_REG_CLASS (regno
))
7351 if (REGNO_PAIR_OK (regno
, mode
))
7353 if (mode
== SImode
|| mode
== DImode
)
7356 if (FLOAT_MODE_P (mode
) && GET_MODE_CLASS (mode
) != MODE_VECTOR_FLOAT
)
7361 if (FRAME_REGNO_P (regno
) && mode
== Pmode
)
7366 if (REGNO_PAIR_OK (regno
, mode
))
7369 || (mode
!= TFmode
&& mode
!= TCmode
&& mode
!= TDmode
))
7374 if (GET_MODE_CLASS (mode
) == MODE_CC
)
7378 if (REGNO_PAIR_OK (regno
, mode
))
7380 if (mode
== SImode
|| mode
== Pmode
)
7391 /* Return nonzero if register OLD_REG can be renamed to register NEW_REG. */
7394 s390_hard_regno_rename_ok (unsigned int old_reg
, unsigned int new_reg
)
7396 /* Once we've decided upon a register to use as base register, it must
7397 no longer be used for any other purpose. */
7398 if (cfun
->machine
->base_reg
)
7399 if (REGNO (cfun
->machine
->base_reg
) == old_reg
7400 || REGNO (cfun
->machine
->base_reg
) == new_reg
)
7406 /* Maximum number of registers to represent a value of mode MODE
7407 in a register of class RCLASS. */
7410 s390_class_max_nregs (enum reg_class rclass
, enum machine_mode mode
)
7415 if (GET_MODE_CLASS (mode
) == MODE_COMPLEX_FLOAT
)
7416 return 2 * ((GET_MODE_SIZE (mode
) / 2 + 8 - 1) / 8);
7418 return (GET_MODE_SIZE (mode
) + 8 - 1) / 8;
7420 return (GET_MODE_SIZE (mode
) + 4 - 1) / 4;
7424 return (GET_MODE_SIZE (mode
) + UNITS_PER_WORD
- 1) / UNITS_PER_WORD
;
7427 /* Return true if register FROM can be eliminated via register TO. */
7430 s390_can_eliminate (const int from
, const int to
)
7432 /* On zSeries machines, we have not marked the base register as fixed.
7433 Instead, we have an elimination rule BASE_REGNUM -> BASE_REGNUM.
7434 If a function requires the base register, we say here that this
7435 elimination cannot be performed. This will cause reload to free
7436 up the base register (as if it were fixed). On the other hand,
7437 if the current function does *not* require the base register, we
7438 say here the elimination succeeds, which in turn allows reload
7439 to allocate the base register for any other purpose. */
7440 if (from
== BASE_REGNUM
&& to
== BASE_REGNUM
)
7442 if (TARGET_CPU_ZARCH
)
7444 s390_init_frame_layout ();
7445 return cfun
->machine
->base_reg
== NULL_RTX
;
7451 /* Everything else must point into the stack frame. */
7452 gcc_assert (to
== STACK_POINTER_REGNUM
7453 || to
== HARD_FRAME_POINTER_REGNUM
);
7455 gcc_assert (from
== FRAME_POINTER_REGNUM
7456 || from
== ARG_POINTER_REGNUM
7457 || from
== RETURN_ADDRESS_POINTER_REGNUM
);
7459 /* Make sure we actually saved the return address. */
7460 if (from
== RETURN_ADDRESS_POINTER_REGNUM
)
7461 if (!crtl
->calls_eh_return
7463 && !cfun_frame_layout
.save_return_addr_p
)
7469 /* Return offset between register FROM and TO initially after prolog. */
7472 s390_initial_elimination_offset (int from
, int to
)
7474 HOST_WIDE_INT offset
;
7477 /* ??? Why are we called for non-eliminable pairs? */
7478 if (!s390_can_eliminate (from
, to
))
7483 case FRAME_POINTER_REGNUM
:
7484 offset
= (get_frame_size()
7485 + STACK_POINTER_OFFSET
7486 + crtl
->outgoing_args_size
);
7489 case ARG_POINTER_REGNUM
:
7490 s390_init_frame_layout ();
7491 offset
= cfun_frame_layout
.frame_size
+ STACK_POINTER_OFFSET
;
7494 case RETURN_ADDRESS_POINTER_REGNUM
:
7495 s390_init_frame_layout ();
7496 index
= RETURN_REGNUM
- cfun_frame_layout
.first_save_gpr_slot
;
7497 gcc_assert (index
>= 0);
7498 offset
= cfun_frame_layout
.frame_size
+ cfun_frame_layout
.gprs_offset
;
7499 offset
+= index
* UNITS_PER_LONG
;
7513 /* Emit insn to save fpr REGNUM at offset OFFSET relative
7514 to register BASE. Return generated insn. */
7517 save_fpr (rtx base
, int offset
, int regnum
)
7520 addr
= gen_rtx_MEM (DFmode
, plus_constant (base
, offset
));
7522 if (regnum
>= 16 && regnum
<= (16 + FP_ARG_NUM_REG
))
7523 set_mem_alias_set (addr
, get_varargs_alias_set ());
7525 set_mem_alias_set (addr
, get_frame_alias_set ());
7527 return emit_move_insn (addr
, gen_rtx_REG (DFmode
, regnum
));
7530 /* Emit insn to restore fpr REGNUM from offset OFFSET relative
7531 to register BASE. Return generated insn. */
7534 restore_fpr (rtx base
, int offset
, int regnum
)
7537 addr
= gen_rtx_MEM (DFmode
, plus_constant (base
, offset
));
7538 set_mem_alias_set (addr
, get_frame_alias_set ());
7540 return emit_move_insn (gen_rtx_REG (DFmode
, regnum
), addr
);
7543 /* Return true if REGNO is a global register, but not one
7544 of the special ones that need to be saved/restored in anyway. */
7547 global_not_special_regno_p (int regno
)
7549 return (global_regs
[regno
]
7550 /* These registers are special and need to be
7551 restored in any case. */
7552 && !(regno
== STACK_POINTER_REGNUM
7553 || regno
== RETURN_REGNUM
7554 || regno
== BASE_REGNUM
7555 || (flag_pic
&& regno
== (int)PIC_OFFSET_TABLE_REGNUM
)));
7558 /* Generate insn to save registers FIRST to LAST into
7559 the register save area located at offset OFFSET
7560 relative to register BASE. */
7563 save_gprs (rtx base
, int offset
, int first
, int last
)
7565 rtx addr
, insn
, note
;
7568 addr
= plus_constant (base
, offset
);
7569 addr
= gen_rtx_MEM (Pmode
, addr
);
7571 set_mem_alias_set (addr
, get_frame_alias_set ());
7573 /* Special-case single register. */
7577 insn
= gen_movdi (addr
, gen_rtx_REG (Pmode
, first
));
7579 insn
= gen_movsi (addr
, gen_rtx_REG (Pmode
, first
));
7581 if (!global_not_special_regno_p (first
))
7582 RTX_FRAME_RELATED_P (insn
) = 1;
7587 insn
= gen_store_multiple (addr
,
7588 gen_rtx_REG (Pmode
, first
),
7589 GEN_INT (last
- first
+ 1));
7591 if (first
<= 6 && cfun
->stdarg
)
7592 for (i
= 0; i
< XVECLEN (PATTERN (insn
), 0); i
++)
7594 rtx mem
= XEXP (XVECEXP (PATTERN (insn
), 0, i
), 0);
7597 set_mem_alias_set (mem
, get_varargs_alias_set ());
7600 /* We need to set the FRAME_RELATED flag on all SETs
7601 inside the store-multiple pattern.
7603 However, we must not emit DWARF records for registers 2..5
7604 if they are stored for use by variable arguments ...
7606 ??? Unfortunately, it is not enough to simply not the
7607 FRAME_RELATED flags for those SETs, because the first SET
7608 of the PARALLEL is always treated as if it had the flag
7609 set, even if it does not. Therefore we emit a new pattern
7610 without those registers as REG_FRAME_RELATED_EXPR note. */
7612 if (first
>= 6 && !global_not_special_regno_p (first
))
7614 rtx pat
= PATTERN (insn
);
7616 for (i
= 0; i
< XVECLEN (pat
, 0); i
++)
7617 if (GET_CODE (XVECEXP (pat
, 0, i
)) == SET
7618 && !global_not_special_regno_p (REGNO (SET_SRC (XVECEXP (pat
,
7620 RTX_FRAME_RELATED_P (XVECEXP (pat
, 0, i
)) = 1;
7622 RTX_FRAME_RELATED_P (insn
) = 1;
7628 for (start
= first
>= 6 ? first
: 6; start
<= last
; start
++)
7629 if (!global_not_special_regno_p (start
))
7635 addr
= plus_constant (base
, offset
+ (start
- first
) * UNITS_PER_LONG
);
7636 note
= gen_store_multiple (gen_rtx_MEM (Pmode
, addr
),
7637 gen_rtx_REG (Pmode
, start
),
7638 GEN_INT (last
- start
+ 1));
7639 note
= PATTERN (note
);
7641 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
, note
);
7643 for (i
= 0; i
< XVECLEN (note
, 0); i
++)
7644 if (GET_CODE (XVECEXP (note
, 0, i
)) == SET
7645 && !global_not_special_regno_p (REGNO (SET_SRC (XVECEXP (note
,
7647 RTX_FRAME_RELATED_P (XVECEXP (note
, 0, i
)) = 1;
7649 RTX_FRAME_RELATED_P (insn
) = 1;
7655 /* Generate insn to restore registers FIRST to LAST from
7656 the register save area located at offset OFFSET
7657 relative to register BASE. */
7660 restore_gprs (rtx base
, int offset
, int first
, int last
)
7664 addr
= plus_constant (base
, offset
);
7665 addr
= gen_rtx_MEM (Pmode
, addr
);
7666 set_mem_alias_set (addr
, get_frame_alias_set ());
7668 /* Special-case single register. */
7672 insn
= gen_movdi (gen_rtx_REG (Pmode
, first
), addr
);
7674 insn
= gen_movsi (gen_rtx_REG (Pmode
, first
), addr
);
7679 insn
= gen_load_multiple (gen_rtx_REG (Pmode
, first
),
7681 GEN_INT (last
- first
+ 1));
7685 /* Return insn sequence to load the GOT register. */
7687 static GTY(()) rtx got_symbol
;
7689 s390_load_got (void)
7695 got_symbol
= gen_rtx_SYMBOL_REF (Pmode
, "_GLOBAL_OFFSET_TABLE_");
7696 SYMBOL_REF_FLAGS (got_symbol
) = SYMBOL_FLAG_LOCAL
;
7701 if (TARGET_CPU_ZARCH
)
7703 emit_move_insn (pic_offset_table_rtx
, got_symbol
);
7709 offset
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, got_symbol
),
7710 UNSPEC_LTREL_OFFSET
);
7711 offset
= gen_rtx_CONST (Pmode
, offset
);
7712 offset
= force_const_mem (Pmode
, offset
);
7714 emit_move_insn (pic_offset_table_rtx
, offset
);
7716 offset
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, XEXP (offset
, 0)),
7718 offset
= gen_rtx_PLUS (Pmode
, pic_offset_table_rtx
, offset
);
7720 emit_move_insn (pic_offset_table_rtx
, offset
);
7723 insns
= get_insns ();
7728 /* This ties together stack memory (MEM with an alias set of frame_alias_set)
7729 and the change to the stack pointer. */
7732 s390_emit_stack_tie (void)
7734 rtx mem
= gen_frame_mem (BLKmode
,
7735 gen_rtx_REG (Pmode
, STACK_POINTER_REGNUM
));
7737 emit_insn (gen_stack_tie (mem
));
7740 /* Expand the prologue into a bunch of separate insns. */
7743 s390_emit_prologue (void)
7751 /* Complete frame layout. */
7753 s390_update_frame_layout ();
7755 /* Annotate all constant pool references to let the scheduler know
7756 they implicitly use the base register. */
7758 push_topmost_sequence ();
7760 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
7763 annotate_constant_pool_refs (&PATTERN (insn
));
7764 df_insn_rescan (insn
);
7767 pop_topmost_sequence ();
7769 /* Choose best register to use for temp use within prologue.
7770 See below for why TPF must use the register 1. */
7772 if (!has_hard_reg_initial_val (Pmode
, RETURN_REGNUM
)
7773 && !current_function_is_leaf
7774 && !TARGET_TPF_PROFILING
)
7775 temp_reg
= gen_rtx_REG (Pmode
, RETURN_REGNUM
);
7777 temp_reg
= gen_rtx_REG (Pmode
, 1);
7779 /* Save call saved gprs. */
7780 if (cfun_frame_layout
.first_save_gpr
!= -1)
7782 insn
= save_gprs (stack_pointer_rtx
,
7783 cfun_frame_layout
.gprs_offset
+
7784 UNITS_PER_LONG
* (cfun_frame_layout
.first_save_gpr
7785 - cfun_frame_layout
.first_save_gpr_slot
),
7786 cfun_frame_layout
.first_save_gpr
,
7787 cfun_frame_layout
.last_save_gpr
);
7791 /* Dummy insn to mark literal pool slot. */
7793 if (cfun
->machine
->base_reg
)
7794 emit_insn (gen_main_pool (cfun
->machine
->base_reg
));
7796 offset
= cfun_frame_layout
.f0_offset
;
7798 /* Save f0 and f2. */
7799 for (i
= 0; i
< 2; i
++)
7801 if (cfun_fpr_bit_p (i
))
7803 save_fpr (stack_pointer_rtx
, offset
, i
+ 16);
7806 else if (!TARGET_PACKED_STACK
)
7810 /* Save f4 and f6. */
7811 offset
= cfun_frame_layout
.f4_offset
;
7812 for (i
= 2; i
< 4; i
++)
7814 if (cfun_fpr_bit_p (i
))
7816 insn
= save_fpr (stack_pointer_rtx
, offset
, i
+ 16);
7819 /* If f4 and f6 are call clobbered they are saved due to stdargs and
7820 therefore are not frame related. */
7821 if (!call_really_used_regs
[i
+ 16])
7822 RTX_FRAME_RELATED_P (insn
) = 1;
7824 else if (!TARGET_PACKED_STACK
)
7828 if (TARGET_PACKED_STACK
7829 && cfun_save_high_fprs_p
7830 && cfun_frame_layout
.f8_offset
+ cfun_frame_layout
.high_fprs
* 8 > 0)
7832 offset
= (cfun_frame_layout
.f8_offset
7833 + (cfun_frame_layout
.high_fprs
- 1) * 8);
7835 for (i
= 15; i
> 7 && offset
>= 0; i
--)
7836 if (cfun_fpr_bit_p (i
))
7838 insn
= save_fpr (stack_pointer_rtx
, offset
, i
+ 16);
7840 RTX_FRAME_RELATED_P (insn
) = 1;
7843 if (offset
>= cfun_frame_layout
.f8_offset
)
7847 if (!TARGET_PACKED_STACK
)
7848 next_fpr
= cfun_save_high_fprs_p
? 31 : 0;
7850 /* Decrement stack pointer. */
7852 if (cfun_frame_layout
.frame_size
> 0)
7854 rtx frame_off
= GEN_INT (-cfun_frame_layout
.frame_size
);
7857 if (s390_stack_size
)
7859 HOST_WIDE_INT stack_guard
;
7861 if (s390_stack_guard
)
7862 stack_guard
= s390_stack_guard
;
7865 /* If no value for stack guard is provided the smallest power of 2
7866 larger than the current frame size is chosen. */
7868 while (stack_guard
< cfun_frame_layout
.frame_size
)
7872 if (cfun_frame_layout
.frame_size
>= s390_stack_size
)
7874 warning (0, "frame size of function %qs is "
7875 HOST_WIDE_INT_PRINT_DEC
7876 " bytes exceeding user provided stack limit of "
7877 HOST_WIDE_INT_PRINT_DEC
" bytes. "
7878 "An unconditional trap is added.",
7879 current_function_name(), cfun_frame_layout
.frame_size
,
7881 emit_insn (gen_trap ());
7885 /* stack_guard has to be smaller than s390_stack_size.
7886 Otherwise we would emit an AND with zero which would
7887 not match the test under mask pattern. */
7888 if (stack_guard
>= s390_stack_size
)
7890 warning (0, "frame size of function %qs is "
7891 HOST_WIDE_INT_PRINT_DEC
7892 " bytes which is more than half the stack size. "
7893 "The dynamic check would not be reliable. "
7894 "No check emitted for this function.",
7895 current_function_name(),
7896 cfun_frame_layout
.frame_size
);
7900 HOST_WIDE_INT stack_check_mask
= ((s390_stack_size
- 1)
7901 & ~(stack_guard
- 1));
7903 rtx t
= gen_rtx_AND (Pmode
, stack_pointer_rtx
,
7904 GEN_INT (stack_check_mask
));
7906 emit_insn (gen_ctrapdi4 (gen_rtx_EQ (VOIDmode
,
7908 t
, const0_rtx
, const0_rtx
));
7910 emit_insn (gen_ctrapsi4 (gen_rtx_EQ (VOIDmode
,
7912 t
, const0_rtx
, const0_rtx
));
7917 if (s390_warn_framesize
> 0
7918 && cfun_frame_layout
.frame_size
>= s390_warn_framesize
)
7919 warning (0, "frame size of %qs is " HOST_WIDE_INT_PRINT_DEC
" bytes",
7920 current_function_name (), cfun_frame_layout
.frame_size
);
7922 if (s390_warn_dynamicstack_p
&& cfun
->calls_alloca
)
7923 warning (0, "%qs uses dynamic stack allocation", current_function_name ());
7925 /* Save incoming stack pointer into temp reg. */
7926 if (TARGET_BACKCHAIN
|| next_fpr
)
7927 insn
= emit_insn (gen_move_insn (temp_reg
, stack_pointer_rtx
));
7929 /* Subtract frame size from stack pointer. */
7931 if (DISP_IN_RANGE (INTVAL (frame_off
)))
7933 insn
= gen_rtx_SET (VOIDmode
, stack_pointer_rtx
,
7934 gen_rtx_PLUS (Pmode
, stack_pointer_rtx
,
7936 insn
= emit_insn (insn
);
7940 if (!CONST_OK_FOR_K (INTVAL (frame_off
)))
7941 frame_off
= force_const_mem (Pmode
, frame_off
);
7943 insn
= emit_insn (gen_add2_insn (stack_pointer_rtx
, frame_off
));
7944 annotate_constant_pool_refs (&PATTERN (insn
));
7947 RTX_FRAME_RELATED_P (insn
) = 1;
7948 real_frame_off
= GEN_INT (-cfun_frame_layout
.frame_size
);
7949 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
,
7950 gen_rtx_SET (VOIDmode
, stack_pointer_rtx
,
7951 gen_rtx_PLUS (Pmode
, stack_pointer_rtx
,
7954 /* Set backchain. */
7956 if (TARGET_BACKCHAIN
)
7958 if (cfun_frame_layout
.backchain_offset
)
7959 addr
= gen_rtx_MEM (Pmode
,
7960 plus_constant (stack_pointer_rtx
,
7961 cfun_frame_layout
.backchain_offset
));
7963 addr
= gen_rtx_MEM (Pmode
, stack_pointer_rtx
);
7964 set_mem_alias_set (addr
, get_frame_alias_set ());
7965 insn
= emit_insn (gen_move_insn (addr
, temp_reg
));
7968 /* If we support non-call exceptions (e.g. for Java),
7969 we need to make sure the backchain pointer is set up
7970 before any possibly trapping memory access. */
7971 if (TARGET_BACKCHAIN
&& cfun
->can_throw_non_call_exceptions
)
7973 addr
= gen_rtx_MEM (BLKmode
, gen_rtx_SCRATCH (VOIDmode
));
7974 emit_clobber (addr
);
7978 /* Save fprs 8 - 15 (64 bit ABI). */
7980 if (cfun_save_high_fprs_p
&& next_fpr
)
7982 /* If the stack might be accessed through a different register
7983 we have to make sure that the stack pointer decrement is not
7984 moved below the use of the stack slots. */
7985 s390_emit_stack_tie ();
7987 insn
= emit_insn (gen_add2_insn (temp_reg
,
7988 GEN_INT (cfun_frame_layout
.f8_offset
)));
7992 for (i
= 24; i
<= next_fpr
; i
++)
7993 if (cfun_fpr_bit_p (i
- 16))
7995 rtx addr
= plus_constant (stack_pointer_rtx
,
7996 cfun_frame_layout
.frame_size
7997 + cfun_frame_layout
.f8_offset
8000 insn
= save_fpr (temp_reg
, offset
, i
);
8002 RTX_FRAME_RELATED_P (insn
) = 1;
8003 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
,
8004 gen_rtx_SET (VOIDmode
,
8005 gen_rtx_MEM (DFmode
, addr
),
8006 gen_rtx_REG (DFmode
, i
)));
8010 /* Set frame pointer, if needed. */
8012 if (frame_pointer_needed
)
8014 insn
= emit_move_insn (hard_frame_pointer_rtx
, stack_pointer_rtx
);
8015 RTX_FRAME_RELATED_P (insn
) = 1;
8018 /* Set up got pointer, if needed. */
8020 if (flag_pic
&& df_regs_ever_live_p (PIC_OFFSET_TABLE_REGNUM
))
8022 rtx insns
= s390_load_got ();
8024 for (insn
= insns
; insn
; insn
= NEXT_INSN (insn
))
8025 annotate_constant_pool_refs (&PATTERN (insn
));
8030 if (TARGET_TPF_PROFILING
)
8032 /* Generate a BAS instruction to serve as a function
8033 entry intercept to facilitate the use of tracing
8034 algorithms located at the branch target. */
8035 emit_insn (gen_prologue_tpf ());
8037 /* Emit a blockage here so that all code
8038 lies between the profiling mechanisms. */
8039 emit_insn (gen_blockage ());
8043 /* Expand the epilogue into a bunch of separate insns. */
8046 s390_emit_epilogue (bool sibcall
)
8048 rtx frame_pointer
, return_reg
, cfa_restores
= NULL_RTX
;
8049 int area_bottom
, area_top
, offset
= 0;
8054 if (TARGET_TPF_PROFILING
)
8057 /* Generate a BAS instruction to serve as a function
8058 entry intercept to facilitate the use of tracing
8059 algorithms located at the branch target. */
8061 /* Emit a blockage here so that all code
8062 lies between the profiling mechanisms. */
8063 emit_insn (gen_blockage ());
8065 emit_insn (gen_epilogue_tpf ());
8068 /* Check whether to use frame or stack pointer for restore. */
8070 frame_pointer
= (frame_pointer_needed
8071 ? hard_frame_pointer_rtx
: stack_pointer_rtx
);
8073 s390_frame_area (&area_bottom
, &area_top
);
8075 /* Check whether we can access the register save area.
8076 If not, increment the frame pointer as required. */
8078 if (area_top
<= area_bottom
)
8080 /* Nothing to restore. */
8082 else if (DISP_IN_RANGE (cfun_frame_layout
.frame_size
+ area_bottom
)
8083 && DISP_IN_RANGE (cfun_frame_layout
.frame_size
+ area_top
- 1))
8085 /* Area is in range. */
8086 offset
= cfun_frame_layout
.frame_size
;
8090 rtx insn
, frame_off
, cfa
;
8092 offset
= area_bottom
< 0 ? -area_bottom
: 0;
8093 frame_off
= GEN_INT (cfun_frame_layout
.frame_size
- offset
);
8095 cfa
= gen_rtx_SET (VOIDmode
, frame_pointer
,
8096 gen_rtx_PLUS (Pmode
, frame_pointer
, frame_off
));
8097 if (DISP_IN_RANGE (INTVAL (frame_off
)))
8099 insn
= gen_rtx_SET (VOIDmode
, frame_pointer
,
8100 gen_rtx_PLUS (Pmode
, frame_pointer
, frame_off
));
8101 insn
= emit_insn (insn
);
8105 if (!CONST_OK_FOR_K (INTVAL (frame_off
)))
8106 frame_off
= force_const_mem (Pmode
, frame_off
);
8108 insn
= emit_insn (gen_add2_insn (frame_pointer
, frame_off
));
8109 annotate_constant_pool_refs (&PATTERN (insn
));
8111 add_reg_note (insn
, REG_CFA_ADJUST_CFA
, cfa
);
8112 RTX_FRAME_RELATED_P (insn
) = 1;
8115 /* Restore call saved fprs. */
8119 if (cfun_save_high_fprs_p
)
8121 next_offset
= cfun_frame_layout
.f8_offset
;
8122 for (i
= 24; i
< 32; i
++)
8124 if (cfun_fpr_bit_p (i
- 16))
8126 restore_fpr (frame_pointer
,
8127 offset
+ next_offset
, i
);
8129 = alloc_reg_note (REG_CFA_RESTORE
,
8130 gen_rtx_REG (DFmode
, i
), cfa_restores
);
8139 next_offset
= cfun_frame_layout
.f4_offset
;
8140 for (i
= 18; i
< 20; i
++)
8142 if (cfun_fpr_bit_p (i
- 16))
8144 restore_fpr (frame_pointer
,
8145 offset
+ next_offset
, i
);
8147 = alloc_reg_note (REG_CFA_RESTORE
,
8148 gen_rtx_REG (DFmode
, i
), cfa_restores
);
8151 else if (!TARGET_PACKED_STACK
)
8157 /* Return register. */
8159 return_reg
= gen_rtx_REG (Pmode
, RETURN_REGNUM
);
8161 /* Restore call saved gprs. */
8163 if (cfun_frame_layout
.first_restore_gpr
!= -1)
8168 /* Check for global register and save them
8169 to stack location from where they get restored. */
8171 for (i
= cfun_frame_layout
.first_restore_gpr
;
8172 i
<= cfun_frame_layout
.last_restore_gpr
;
8175 if (global_not_special_regno_p (i
))
8177 addr
= plus_constant (frame_pointer
,
8178 offset
+ cfun_frame_layout
.gprs_offset
8179 + (i
- cfun_frame_layout
.first_save_gpr_slot
)
8181 addr
= gen_rtx_MEM (Pmode
, addr
);
8182 set_mem_alias_set (addr
, get_frame_alias_set ());
8183 emit_move_insn (addr
, gen_rtx_REG (Pmode
, i
));
8187 = alloc_reg_note (REG_CFA_RESTORE
,
8188 gen_rtx_REG (Pmode
, i
), cfa_restores
);
8193 /* Fetch return address from stack before load multiple,
8194 this will do good for scheduling. */
8196 if (cfun_frame_layout
.save_return_addr_p
8197 || (cfun_frame_layout
.first_restore_gpr
< BASE_REGNUM
8198 && cfun_frame_layout
.last_restore_gpr
> RETURN_REGNUM
))
8200 int return_regnum
= find_unused_clobbered_reg();
8203 return_reg
= gen_rtx_REG (Pmode
, return_regnum
);
8205 addr
= plus_constant (frame_pointer
,
8206 offset
+ cfun_frame_layout
.gprs_offset
8208 - cfun_frame_layout
.first_save_gpr_slot
)
8210 addr
= gen_rtx_MEM (Pmode
, addr
);
8211 set_mem_alias_set (addr
, get_frame_alias_set ());
8212 emit_move_insn (return_reg
, addr
);
8216 insn
= restore_gprs (frame_pointer
,
8217 offset
+ cfun_frame_layout
.gprs_offset
8218 + (cfun_frame_layout
.first_restore_gpr
8219 - cfun_frame_layout
.first_save_gpr_slot
)
8221 cfun_frame_layout
.first_restore_gpr
,
8222 cfun_frame_layout
.last_restore_gpr
);
8223 insn
= emit_insn (insn
);
8224 REG_NOTES (insn
) = cfa_restores
;
8225 add_reg_note (insn
, REG_CFA_DEF_CFA
,
8226 plus_constant (stack_pointer_rtx
, STACK_POINTER_OFFSET
));
8227 RTX_FRAME_RELATED_P (insn
) = 1;
8233 /* Return to caller. */
8235 p
= rtvec_alloc (2);
8237 RTVEC_ELT (p
, 0) = gen_rtx_RETURN (VOIDmode
);
8238 RTVEC_ELT (p
, 1) = gen_rtx_USE (VOIDmode
, return_reg
);
8239 emit_jump_insn (gen_rtx_PARALLEL (VOIDmode
, p
));
8244 /* Return the size in bytes of a function argument of
8245 type TYPE and/or mode MODE. At least one of TYPE or
8246 MODE must be specified. */
8249 s390_function_arg_size (enum machine_mode mode
, const_tree type
)
8252 return int_size_in_bytes (type
);
8254 /* No type info available for some library calls ... */
8255 if (mode
!= BLKmode
)
8256 return GET_MODE_SIZE (mode
);
8258 /* If we have neither type nor mode, abort */
8262 /* Return true if a function argument of type TYPE and mode MODE
8263 is to be passed in a floating-point register, if available. */
8266 s390_function_arg_float (enum machine_mode mode
, tree type
)
8268 int size
= s390_function_arg_size (mode
, type
);
8272 /* Soft-float changes the ABI: no floating-point registers are used. */
8273 if (TARGET_SOFT_FLOAT
)
8276 /* No type info available for some library calls ... */
8278 return mode
== SFmode
|| mode
== DFmode
|| mode
== SDmode
|| mode
== DDmode
;
8280 /* The ABI says that record types with a single member are treated
8281 just like that member would be. */
8282 while (TREE_CODE (type
) == RECORD_TYPE
)
8284 tree field
, single
= NULL_TREE
;
8286 for (field
= TYPE_FIELDS (type
); field
; field
= DECL_CHAIN (field
))
8288 if (TREE_CODE (field
) != FIELD_DECL
)
8291 if (single
== NULL_TREE
)
8292 single
= TREE_TYPE (field
);
8297 if (single
== NULL_TREE
)
8303 return TREE_CODE (type
) == REAL_TYPE
;
8306 /* Return true if a function argument of type TYPE and mode MODE
8307 is to be passed in an integer register, or a pair of integer
8308 registers, if available. */
8311 s390_function_arg_integer (enum machine_mode mode
, tree type
)
8313 int size
= s390_function_arg_size (mode
, type
);
8317 /* No type info available for some library calls ... */
8319 return GET_MODE_CLASS (mode
) == MODE_INT
8320 || (TARGET_SOFT_FLOAT
&& SCALAR_FLOAT_MODE_P (mode
));
8322 /* We accept small integral (and similar) types. */
8323 if (INTEGRAL_TYPE_P (type
)
8324 || POINTER_TYPE_P (type
)
8325 || TREE_CODE (type
) == OFFSET_TYPE
8326 || (TARGET_SOFT_FLOAT
&& TREE_CODE (type
) == REAL_TYPE
))
8329 /* We also accept structs of size 1, 2, 4, 8 that are not
8330 passed in floating-point registers. */
8331 if (AGGREGATE_TYPE_P (type
)
8332 && exact_log2 (size
) >= 0
8333 && !s390_function_arg_float (mode
, type
))
8339 /* Return 1 if a function argument of type TYPE and mode MODE
8340 is to be passed by reference. The ABI specifies that only
8341 structures of size 1, 2, 4, or 8 bytes are passed by value,
8342 all other structures (and complex numbers) are passed by
8346 s390_pass_by_reference (CUMULATIVE_ARGS
*ca ATTRIBUTE_UNUSED
,
8347 enum machine_mode mode
, const_tree type
,
8348 bool named ATTRIBUTE_UNUSED
)
8350 int size
= s390_function_arg_size (mode
, type
);
8356 if (AGGREGATE_TYPE_P (type
) && exact_log2 (size
) < 0)
8359 if (TREE_CODE (type
) == COMPLEX_TYPE
8360 || TREE_CODE (type
) == VECTOR_TYPE
)
8367 /* Update the data in CUM to advance over an argument of mode MODE and
8368 data type TYPE. (TYPE is null for libcalls where that information
8369 may not be available.). The boolean NAMED specifies whether the
8370 argument is a named argument (as opposed to an unnamed argument
8371 matching an ellipsis). */
8374 s390_function_arg_advance (CUMULATIVE_ARGS
*cum
, enum machine_mode mode
,
8375 tree type
, int named ATTRIBUTE_UNUSED
)
8377 if (s390_function_arg_float (mode
, type
))
8381 else if (s390_function_arg_integer (mode
, type
))
8383 int size
= s390_function_arg_size (mode
, type
);
8384 cum
->gprs
+= ((size
+ UNITS_PER_LONG
- 1) / UNITS_PER_LONG
);
8390 /* Define where to put the arguments to a function.
8391 Value is zero to push the argument on the stack,
8392 or a hard register in which to store the argument.
8394 MODE is the argument's machine mode.
8395 TYPE is the data type of the argument (as a tree).
8396 This is null for libcalls where that information may
8398 CUM is a variable of type CUMULATIVE_ARGS which gives info about
8399 the preceding args and about the function being called.
8400 NAMED is nonzero if this argument is a named parameter
8401 (otherwise it is an extra parameter matching an ellipsis).
8403 On S/390, we use general purpose registers 2 through 6 to
8404 pass integer, pointer, and certain structure arguments, and
8405 floating point registers 0 and 2 (0, 2, 4, and 6 on 64-bit)
8406 to pass floating point arguments. All remaining arguments
8407 are pushed to the stack. */
8410 s390_function_arg (CUMULATIVE_ARGS
*cum
, enum machine_mode mode
, tree type
,
8411 int named ATTRIBUTE_UNUSED
)
8413 if (s390_function_arg_float (mode
, type
))
8415 if (cum
->fprs
+ 1 > FP_ARG_NUM_REG
)
8418 return gen_rtx_REG (mode
, cum
->fprs
+ 16);
8420 else if (s390_function_arg_integer (mode
, type
))
8422 int size
= s390_function_arg_size (mode
, type
);
8423 int n_gprs
= (size
+ UNITS_PER_LONG
- 1) / UNITS_PER_LONG
;
8425 if (cum
->gprs
+ n_gprs
> GP_ARG_NUM_REG
)
8427 else if (n_gprs
== 1 || UNITS_PER_WORD
== UNITS_PER_LONG
)
8428 return gen_rtx_REG (mode
, cum
->gprs
+ 2);
8429 else if (n_gprs
== 2)
8431 rtvec p
= rtvec_alloc (2);
8434 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, cum
->gprs
+ 2),
8437 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, cum
->gprs
+ 3),
8440 return gen_rtx_PARALLEL (mode
, p
);
8444 /* After the real arguments, expand_call calls us once again
8445 with a void_type_node type. Whatever we return here is
8446 passed as operand 2 to the call expanders.
8448 We don't need this feature ... */
8449 else if (type
== void_type_node
)
8455 /* Return true if return values of type TYPE should be returned
8456 in a memory buffer whose address is passed by the caller as
8457 hidden first argument. */
8460 s390_return_in_memory (const_tree type
, const_tree fundecl ATTRIBUTE_UNUSED
)
8462 /* We accept small integral (and similar) types. */
8463 if (INTEGRAL_TYPE_P (type
)
8464 || POINTER_TYPE_P (type
)
8465 || TREE_CODE (type
) == OFFSET_TYPE
8466 || TREE_CODE (type
) == REAL_TYPE
)
8467 return int_size_in_bytes (type
) > 8;
8469 /* Aggregates and similar constructs are always returned
8471 if (AGGREGATE_TYPE_P (type
)
8472 || TREE_CODE (type
) == COMPLEX_TYPE
8473 || TREE_CODE (type
) == VECTOR_TYPE
)
8476 /* ??? We get called on all sorts of random stuff from
8477 aggregate_value_p. We can't abort, but it's not clear
8478 what's safe to return. Pretend it's a struct I guess. */
8482 /* Function arguments and return values are promoted to word size. */
8484 static enum machine_mode
8485 s390_promote_function_mode (const_tree type
, enum machine_mode mode
,
8487 const_tree fntype ATTRIBUTE_UNUSED
,
8488 int for_return ATTRIBUTE_UNUSED
)
8490 if (INTEGRAL_MODE_P (mode
)
8491 && GET_MODE_SIZE (mode
) < UNITS_PER_LONG
)
8493 if (POINTER_TYPE_P (type
))
8494 *punsignedp
= POINTERS_EXTEND_UNSIGNED
;
8501 /* Define where to return a (scalar) value of type TYPE.
8502 If TYPE is null, define where to return a (scalar)
8503 value of mode MODE from a libcall. */
8506 s390_function_value (const_tree type
, const_tree fn
, enum machine_mode mode
)
8510 int unsignedp
= TYPE_UNSIGNED (type
);
8511 mode
= promote_function_mode (type
, TYPE_MODE (type
), &unsignedp
, fn
, 1);
8514 gcc_assert (GET_MODE_CLASS (mode
) == MODE_INT
|| SCALAR_FLOAT_MODE_P (mode
));
8515 gcc_assert (GET_MODE_SIZE (mode
) <= 8);
8517 if (TARGET_HARD_FLOAT
&& SCALAR_FLOAT_MODE_P (mode
))
8518 return gen_rtx_REG (mode
, 16);
8519 else if (GET_MODE_SIZE (mode
) <= UNITS_PER_LONG
8520 || UNITS_PER_LONG
== UNITS_PER_WORD
)
8521 return gen_rtx_REG (mode
, 2);
8522 else if (GET_MODE_SIZE (mode
) == 2 * UNITS_PER_LONG
)
8524 rtvec p
= rtvec_alloc (2);
8527 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, 2), const0_rtx
);
8529 = gen_rtx_EXPR_LIST (SImode
, gen_rtx_REG (SImode
, 3), GEN_INT (4));
8531 return gen_rtx_PARALLEL (mode
, p
);
8538 /* Create and return the va_list datatype.
8540 On S/390, va_list is an array type equivalent to
8542 typedef struct __va_list_tag
8546 void *__overflow_arg_area;
8547 void *__reg_save_area;
8550 where __gpr and __fpr hold the number of general purpose
8551 or floating point arguments used up to now, respectively,
8552 __overflow_arg_area points to the stack location of the
8553 next argument passed on the stack, and __reg_save_area
8554 always points to the start of the register area in the
8555 call frame of the current function. The function prologue
8556 saves all registers used for argument passing into this
8557 area if the function uses variable arguments. */
8560 s390_build_builtin_va_list (void)
8562 tree f_gpr
, f_fpr
, f_ovf
, f_sav
, record
, type_decl
;
8564 record
= lang_hooks
.types
.make_type (RECORD_TYPE
);
8567 build_decl (BUILTINS_LOCATION
,
8568 TYPE_DECL
, get_identifier ("__va_list_tag"), record
);
8570 f_gpr
= build_decl (BUILTINS_LOCATION
,
8571 FIELD_DECL
, get_identifier ("__gpr"),
8572 long_integer_type_node
);
8573 f_fpr
= build_decl (BUILTINS_LOCATION
,
8574 FIELD_DECL
, get_identifier ("__fpr"),
8575 long_integer_type_node
);
8576 f_ovf
= build_decl (BUILTINS_LOCATION
,
8577 FIELD_DECL
, get_identifier ("__overflow_arg_area"),
8579 f_sav
= build_decl (BUILTINS_LOCATION
,
8580 FIELD_DECL
, get_identifier ("__reg_save_area"),
8583 va_list_gpr_counter_field
= f_gpr
;
8584 va_list_fpr_counter_field
= f_fpr
;
8586 DECL_FIELD_CONTEXT (f_gpr
) = record
;
8587 DECL_FIELD_CONTEXT (f_fpr
) = record
;
8588 DECL_FIELD_CONTEXT (f_ovf
) = record
;
8589 DECL_FIELD_CONTEXT (f_sav
) = record
;
8591 TREE_CHAIN (record
) = type_decl
;
8592 TYPE_NAME (record
) = type_decl
;
8593 TYPE_FIELDS (record
) = f_gpr
;
8594 DECL_CHAIN (f_gpr
) = f_fpr
;
8595 DECL_CHAIN (f_fpr
) = f_ovf
;
8596 DECL_CHAIN (f_ovf
) = f_sav
;
8598 layout_type (record
);
8600 /* The correct type is an array type of one element. */
8601 return build_array_type (record
, build_index_type (size_zero_node
));
8604 /* Implement va_start by filling the va_list structure VALIST.
8605 STDARG_P is always true, and ignored.
8606 NEXTARG points to the first anonymous stack argument.
8608 The following global variables are used to initialize
8609 the va_list structure:
8612 holds number of gprs and fprs used for named arguments.
8613 crtl->args.arg_offset_rtx:
8614 holds the offset of the first anonymous stack argument
8615 (relative to the virtual arg pointer). */
8618 s390_va_start (tree valist
, rtx nextarg ATTRIBUTE_UNUSED
)
8620 HOST_WIDE_INT n_gpr
, n_fpr
;
8622 tree f_gpr
, f_fpr
, f_ovf
, f_sav
;
8623 tree gpr
, fpr
, ovf
, sav
, t
;
8625 f_gpr
= TYPE_FIELDS (TREE_TYPE (va_list_type_node
));
8626 f_fpr
= DECL_CHAIN (f_gpr
);
8627 f_ovf
= DECL_CHAIN (f_fpr
);
8628 f_sav
= DECL_CHAIN (f_ovf
);
8630 valist
= build_va_arg_indirect_ref (valist
);
8631 gpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_gpr
), valist
, f_gpr
, NULL_TREE
);
8632 fpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_fpr
), valist
, f_fpr
, NULL_TREE
);
8633 ovf
= build3 (COMPONENT_REF
, TREE_TYPE (f_ovf
), valist
, f_ovf
, NULL_TREE
);
8634 sav
= build3 (COMPONENT_REF
, TREE_TYPE (f_sav
), valist
, f_sav
, NULL_TREE
);
8636 /* Count number of gp and fp argument registers used. */
8638 n_gpr
= crtl
->args
.info
.gprs
;
8639 n_fpr
= crtl
->args
.info
.fprs
;
8641 if (cfun
->va_list_gpr_size
)
8643 t
= build2 (MODIFY_EXPR
, TREE_TYPE (gpr
), gpr
,
8644 build_int_cst (NULL_TREE
, n_gpr
));
8645 TREE_SIDE_EFFECTS (t
) = 1;
8646 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8649 if (cfun
->va_list_fpr_size
)
8651 t
= build2 (MODIFY_EXPR
, TREE_TYPE (fpr
), fpr
,
8652 build_int_cst (NULL_TREE
, n_fpr
));
8653 TREE_SIDE_EFFECTS (t
) = 1;
8654 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8657 /* Find the overflow area. */
8658 if (n_gpr
+ cfun
->va_list_gpr_size
> GP_ARG_NUM_REG
8659 || n_fpr
+ cfun
->va_list_fpr_size
> FP_ARG_NUM_REG
)
8661 t
= make_tree (TREE_TYPE (ovf
), virtual_incoming_args_rtx
);
8663 off
= INTVAL (crtl
->args
.arg_offset_rtx
);
8664 off
= off
< 0 ? 0 : off
;
8665 if (TARGET_DEBUG_ARG
)
8666 fprintf (stderr
, "va_start: n_gpr = %d, n_fpr = %d off %d\n",
8667 (int)n_gpr
, (int)n_fpr
, off
);
8669 t
= build2 (POINTER_PLUS_EXPR
, TREE_TYPE (ovf
), t
, size_int (off
));
8671 t
= build2 (MODIFY_EXPR
, TREE_TYPE (ovf
), ovf
, t
);
8672 TREE_SIDE_EFFECTS (t
) = 1;
8673 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8676 /* Find the register save area. */
8677 if ((cfun
->va_list_gpr_size
&& n_gpr
< GP_ARG_NUM_REG
)
8678 || (cfun
->va_list_fpr_size
&& n_fpr
< FP_ARG_NUM_REG
))
8680 t
= make_tree (TREE_TYPE (sav
), return_address_pointer_rtx
);
8681 t
= build2 (POINTER_PLUS_EXPR
, TREE_TYPE (sav
), t
,
8682 size_int (-RETURN_REGNUM
* UNITS_PER_LONG
));
8684 t
= build2 (MODIFY_EXPR
, TREE_TYPE (sav
), sav
, t
);
8685 TREE_SIDE_EFFECTS (t
) = 1;
8686 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
8690 /* Implement va_arg by updating the va_list structure
8691 VALIST as required to retrieve an argument of type
8692 TYPE, and returning that argument.
8694 Generates code equivalent to:
8696 if (integral value) {
8697 if (size <= 4 && args.gpr < 5 ||
8698 size > 4 && args.gpr < 4 )
8699 ret = args.reg_save_area[args.gpr+8]
8701 ret = *args.overflow_arg_area++;
8702 } else if (float value) {
8704 ret = args.reg_save_area[args.fpr+64]
8706 ret = *args.overflow_arg_area++;
8707 } else if (aggregate value) {
8709 ret = *args.reg_save_area[args.gpr]
8711 ret = **args.overflow_arg_area++;
8715 s390_gimplify_va_arg (tree valist
, tree type
, gimple_seq
*pre_p
,
8716 gimple_seq
*post_p ATTRIBUTE_UNUSED
)
8718 tree f_gpr
, f_fpr
, f_ovf
, f_sav
;
8719 tree gpr
, fpr
, ovf
, sav
, reg
, t
, u
;
8720 int indirect_p
, size
, n_reg
, sav_ofs
, sav_scale
, max_reg
;
8721 tree lab_false
, lab_over
, addr
;
8723 f_gpr
= TYPE_FIELDS (TREE_TYPE (va_list_type_node
));
8724 f_fpr
= DECL_CHAIN (f_gpr
);
8725 f_ovf
= DECL_CHAIN (f_fpr
);
8726 f_sav
= DECL_CHAIN (f_ovf
);
8728 valist
= build_va_arg_indirect_ref (valist
);
8729 gpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_gpr
), valist
, f_gpr
, NULL_TREE
);
8730 fpr
= build3 (COMPONENT_REF
, TREE_TYPE (f_fpr
), valist
, f_fpr
, NULL_TREE
);
8731 sav
= build3 (COMPONENT_REF
, TREE_TYPE (f_sav
), valist
, f_sav
, NULL_TREE
);
8733 /* The tree for args* cannot be shared between gpr/fpr and ovf since
8734 both appear on a lhs. */
8735 valist
= unshare_expr (valist
);
8736 ovf
= build3 (COMPONENT_REF
, TREE_TYPE (f_ovf
), valist
, f_ovf
, NULL_TREE
);
8738 size
= int_size_in_bytes (type
);
8740 if (pass_by_reference (NULL
, TYPE_MODE (type
), type
, false))
8742 if (TARGET_DEBUG_ARG
)
8744 fprintf (stderr
, "va_arg: aggregate type");
8748 /* Aggregates are passed by reference. */
8753 /* kernel stack layout on 31 bit: It is assumed here that no padding
8754 will be added by s390_frame_info because for va_args always an even
8755 number of gprs has to be saved r15-r2 = 14 regs. */
8756 sav_ofs
= 2 * UNITS_PER_LONG
;
8757 sav_scale
= UNITS_PER_LONG
;
8758 size
= UNITS_PER_LONG
;
8759 max_reg
= GP_ARG_NUM_REG
- n_reg
;
8761 else if (s390_function_arg_float (TYPE_MODE (type
), type
))
8763 if (TARGET_DEBUG_ARG
)
8765 fprintf (stderr
, "va_arg: float type");
8769 /* FP args go in FP registers, if present. */
8773 sav_ofs
= 16 * UNITS_PER_LONG
;
8775 max_reg
= FP_ARG_NUM_REG
- n_reg
;
8779 if (TARGET_DEBUG_ARG
)
8781 fprintf (stderr
, "va_arg: other type");
8785 /* Otherwise into GP registers. */
8788 n_reg
= (size
+ UNITS_PER_LONG
- 1) / UNITS_PER_LONG
;
8790 /* kernel stack layout on 31 bit: It is assumed here that no padding
8791 will be added by s390_frame_info because for va_args always an even
8792 number of gprs has to be saved r15-r2 = 14 regs. */
8793 sav_ofs
= 2 * UNITS_PER_LONG
;
8795 if (size
< UNITS_PER_LONG
)
8796 sav_ofs
+= UNITS_PER_LONG
- size
;
8798 sav_scale
= UNITS_PER_LONG
;
8799 max_reg
= GP_ARG_NUM_REG
- n_reg
;
8802 /* Pull the value out of the saved registers ... */
8804 lab_false
= create_artificial_label (UNKNOWN_LOCATION
);
8805 lab_over
= create_artificial_label (UNKNOWN_LOCATION
);
8806 addr
= create_tmp_var (ptr_type_node
, "addr");
8808 t
= fold_convert (TREE_TYPE (reg
), size_int (max_reg
));
8809 t
= build2 (GT_EXPR
, boolean_type_node
, reg
, t
);
8810 u
= build1 (GOTO_EXPR
, void_type_node
, lab_false
);
8811 t
= build3 (COND_EXPR
, void_type_node
, t
, u
, NULL_TREE
);
8812 gimplify_and_add (t
, pre_p
);
8814 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, sav
,
8815 size_int (sav_ofs
));
8816 u
= build2 (MULT_EXPR
, TREE_TYPE (reg
), reg
,
8817 fold_convert (TREE_TYPE (reg
), size_int (sav_scale
)));
8818 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
, fold_convert (sizetype
, u
));
8820 gimplify_assign (addr
, t
, pre_p
);
8822 gimple_seq_add_stmt (pre_p
, gimple_build_goto (lab_over
));
8824 gimple_seq_add_stmt (pre_p
, gimple_build_label (lab_false
));
8827 /* ... Otherwise out of the overflow area. */
8830 if (size
< UNITS_PER_LONG
)
8831 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
,
8832 size_int (UNITS_PER_LONG
- size
));
8834 gimplify_expr (&t
, pre_p
, NULL
, is_gimple_val
, fb_rvalue
);
8836 gimplify_assign (addr
, t
, pre_p
);
8838 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
,
8840 gimplify_assign (ovf
, t
, pre_p
);
8842 gimple_seq_add_stmt (pre_p
, gimple_build_label (lab_over
));
8845 /* Increment register save count. */
8847 u
= build2 (PREINCREMENT_EXPR
, TREE_TYPE (reg
), reg
,
8848 fold_convert (TREE_TYPE (reg
), size_int (n_reg
)));
8849 gimplify_and_add (u
, pre_p
);
8853 t
= build_pointer_type_for_mode (build_pointer_type (type
),
8855 addr
= fold_convert (t
, addr
);
8856 addr
= build_va_arg_indirect_ref (addr
);
8860 t
= build_pointer_type_for_mode (type
, ptr_mode
, true);
8861 addr
= fold_convert (t
, addr
);
8864 return build_va_arg_indirect_ref (addr
);
8872 S390_BUILTIN_THREAD_POINTER
,
8873 S390_BUILTIN_SET_THREAD_POINTER
,
8878 static enum insn_code
const code_for_builtin_64
[S390_BUILTIN_max
] = {
8883 static enum insn_code
const code_for_builtin_31
[S390_BUILTIN_max
] = {
8889 s390_init_builtins (void)
8893 ftype
= build_function_type (ptr_type_node
, void_list_node
);
8894 add_builtin_function ("__builtin_thread_pointer", ftype
,
8895 S390_BUILTIN_THREAD_POINTER
, BUILT_IN_MD
,
8898 ftype
= build_function_type_list (void_type_node
, ptr_type_node
, NULL_TREE
);
8899 add_builtin_function ("__builtin_set_thread_pointer", ftype
,
8900 S390_BUILTIN_SET_THREAD_POINTER
, BUILT_IN_MD
,
8904 /* Expand an expression EXP that calls a built-in function,
8905 with result going to TARGET if that's convenient
8906 (and in mode MODE if that's convenient).
8907 SUBTARGET may be used as the target for computing one of EXP's operands.
8908 IGNORE is nonzero if the value is to be ignored. */
8911 s390_expand_builtin (tree exp
, rtx target
, rtx subtarget ATTRIBUTE_UNUSED
,
8912 enum machine_mode mode ATTRIBUTE_UNUSED
,
8913 int ignore ATTRIBUTE_UNUSED
)
8917 enum insn_code
const *code_for_builtin
=
8918 TARGET_64BIT
? code_for_builtin_64
: code_for_builtin_31
;
8920 tree fndecl
= TREE_OPERAND (CALL_EXPR_FN (exp
), 0);
8921 unsigned int fcode
= DECL_FUNCTION_CODE (fndecl
);
8922 enum insn_code icode
;
8923 rtx op
[MAX_ARGS
], pat
;
8927 call_expr_arg_iterator iter
;
8929 if (fcode
>= S390_BUILTIN_max
)
8930 internal_error ("bad builtin fcode");
8931 icode
= code_for_builtin
[fcode
];
8933 internal_error ("bad builtin fcode");
8935 nonvoid
= TREE_TYPE (TREE_TYPE (fndecl
)) != void_type_node
;
8938 FOR_EACH_CALL_EXPR_ARG (arg
, iter
, exp
)
8940 const struct insn_operand_data
*insn_op
;
8942 if (arg
== error_mark_node
)
8944 if (arity
> MAX_ARGS
)
8947 insn_op
= &insn_data
[icode
].operand
[arity
+ nonvoid
];
8949 op
[arity
] = expand_expr (arg
, NULL_RTX
, insn_op
->mode
, EXPAND_NORMAL
);
8951 if (!(*insn_op
->predicate
) (op
[arity
], insn_op
->mode
))
8952 op
[arity
] = copy_to_mode_reg (insn_op
->mode
, op
[arity
]);
8958 enum machine_mode tmode
= insn_data
[icode
].operand
[0].mode
;
8960 || GET_MODE (target
) != tmode
8961 || !(*insn_data
[icode
].operand
[0].predicate
) (target
, tmode
))
8962 target
= gen_reg_rtx (tmode
);
8968 pat
= GEN_FCN (icode
) (target
);
8972 pat
= GEN_FCN (icode
) (target
, op
[0]);
8974 pat
= GEN_FCN (icode
) (op
[0]);
8977 pat
= GEN_FCN (icode
) (target
, op
[0], op
[1]);
8993 /* Output assembly code for the trampoline template to
8996 On S/390, we use gpr 1 internally in the trampoline code;
8997 gpr 0 is used to hold the static chain. */
9000 s390_asm_trampoline_template (FILE *file
)
9003 op
[0] = gen_rtx_REG (Pmode
, 0);
9004 op
[1] = gen_rtx_REG (Pmode
, 1);
9008 output_asm_insn ("basr\t%1,0", op
);
9009 output_asm_insn ("lmg\t%0,%1,14(%1)", op
);
9010 output_asm_insn ("br\t%1", op
);
9011 ASM_OUTPUT_SKIP (file
, (HOST_WIDE_INT
)(TRAMPOLINE_SIZE
- 10));
9015 output_asm_insn ("basr\t%1,0", op
);
9016 output_asm_insn ("lm\t%0,%1,6(%1)", op
);
9017 output_asm_insn ("br\t%1", op
);
9018 ASM_OUTPUT_SKIP (file
, (HOST_WIDE_INT
)(TRAMPOLINE_SIZE
- 8));
9022 /* Emit RTL insns to initialize the variable parts of a trampoline.
9023 FNADDR is an RTX for the address of the function's pure code.
9024 CXT is an RTX for the static chain value for the function. */
9027 s390_trampoline_init (rtx m_tramp
, tree fndecl
, rtx cxt
)
9029 rtx fnaddr
= XEXP (DECL_RTL (fndecl
), 0);
9032 emit_block_move (m_tramp
, assemble_trampoline_template (),
9033 GEN_INT (2*UNITS_PER_WORD
), BLOCK_OP_NORMAL
);
9035 mem
= adjust_address (m_tramp
, Pmode
, 2*UNITS_PER_WORD
);
9036 emit_move_insn (mem
, cxt
);
9037 mem
= adjust_address (m_tramp
, Pmode
, 3*UNITS_PER_WORD
);
9038 emit_move_insn (mem
, fnaddr
);
9041 /* Output assembler code to FILE to increment profiler label # LABELNO
9042 for profiling a function entry. */
9045 s390_function_profiler (FILE *file
, int labelno
)
9050 ASM_GENERATE_INTERNAL_LABEL (label
, "LP", labelno
);
9052 fprintf (file
, "# function profiler \n");
9054 op
[0] = gen_rtx_REG (Pmode
, RETURN_REGNUM
);
9055 op
[1] = gen_rtx_REG (Pmode
, STACK_POINTER_REGNUM
);
9056 op
[1] = gen_rtx_MEM (Pmode
, plus_constant (op
[1], UNITS_PER_LONG
));
9058 op
[2] = gen_rtx_REG (Pmode
, 1);
9059 op
[3] = gen_rtx_SYMBOL_REF (Pmode
, label
);
9060 SYMBOL_REF_FLAGS (op
[3]) = SYMBOL_FLAG_LOCAL
;
9062 op
[4] = gen_rtx_SYMBOL_REF (Pmode
, "_mcount");
9065 op
[4] = gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, op
[4]), UNSPEC_PLT
);
9066 op
[4] = gen_rtx_CONST (Pmode
, op
[4]);
9071 output_asm_insn ("stg\t%0,%1", op
);
9072 output_asm_insn ("larl\t%2,%3", op
);
9073 output_asm_insn ("brasl\t%0,%4", op
);
9074 output_asm_insn ("lg\t%0,%1", op
);
9078 op
[6] = gen_label_rtx ();
9080 output_asm_insn ("st\t%0,%1", op
);
9081 output_asm_insn ("bras\t%2,%l6", op
);
9082 output_asm_insn (".long\t%4", op
);
9083 output_asm_insn (".long\t%3", op
);
9084 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[6]));
9085 output_asm_insn ("l\t%0,0(%2)", op
);
9086 output_asm_insn ("l\t%2,4(%2)", op
);
9087 output_asm_insn ("basr\t%0,%0", op
);
9088 output_asm_insn ("l\t%0,%1", op
);
9092 op
[5] = gen_label_rtx ();
9093 op
[6] = gen_label_rtx ();
9095 output_asm_insn ("st\t%0,%1", op
);
9096 output_asm_insn ("bras\t%2,%l6", op
);
9097 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[5]));
9098 output_asm_insn (".long\t%4-%l5", op
);
9099 output_asm_insn (".long\t%3-%l5", op
);
9100 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[6]));
9101 output_asm_insn ("lr\t%0,%2", op
);
9102 output_asm_insn ("a\t%0,0(%2)", op
);
9103 output_asm_insn ("a\t%2,4(%2)", op
);
9104 output_asm_insn ("basr\t%0,%0", op
);
9105 output_asm_insn ("l\t%0,%1", op
);
9109 /* Encode symbol attributes (local vs. global, tls model) of a SYMBOL_REF
9110 into its SYMBOL_REF_FLAGS. */
9113 s390_encode_section_info (tree decl
, rtx rtl
, int first
)
9115 default_encode_section_info (decl
, rtl
, first
);
9117 if (TREE_CODE (decl
) == VAR_DECL
)
9119 /* If a variable has a forced alignment to < 2 bytes, mark it
9120 with SYMBOL_FLAG_ALIGN1 to prevent it from being used as LARL
9122 if (DECL_USER_ALIGN (decl
) && DECL_ALIGN (decl
) < 16)
9123 SYMBOL_REF_FLAGS (XEXP (rtl
, 0)) |= SYMBOL_FLAG_ALIGN1
;
9124 if (!DECL_SIZE (decl
)
9125 || !DECL_ALIGN (decl
)
9126 || !host_integerp (DECL_SIZE (decl
), 0)
9127 || (DECL_ALIGN (decl
) <= 64
9128 && DECL_ALIGN (decl
) != tree_low_cst (DECL_SIZE (decl
), 0)))
9129 SYMBOL_REF_FLAGS (XEXP (rtl
, 0)) |= SYMBOL_FLAG_NOT_NATURALLY_ALIGNED
;
9132 /* Literal pool references don't have a decl so they are handled
9133 differently here. We rely on the information in the MEM_ALIGN
9134 entry to decide upon natural alignment. */
9136 && GET_CODE (XEXP (rtl
, 0)) == SYMBOL_REF
9137 && TREE_CONSTANT_POOL_ADDRESS_P (XEXP (rtl
, 0))
9138 && (MEM_ALIGN (rtl
) == 0
9139 || GET_MODE_BITSIZE (GET_MODE (rtl
)) == 0
9140 || MEM_ALIGN (rtl
) < GET_MODE_BITSIZE (GET_MODE (rtl
))))
9141 SYMBOL_REF_FLAGS (XEXP (rtl
, 0)) |= SYMBOL_FLAG_NOT_NATURALLY_ALIGNED
;
9144 /* Output thunk to FILE that implements a C++ virtual function call (with
9145 multiple inheritance) to FUNCTION. The thunk adjusts the this pointer
9146 by DELTA, and unless VCALL_OFFSET is zero, applies an additional adjustment
9147 stored at VCALL_OFFSET in the vtable whose address is located at offset 0
9148 relative to the resulting this pointer. */
9151 s390_output_mi_thunk (FILE *file
, tree thunk ATTRIBUTE_UNUSED
,
9152 HOST_WIDE_INT delta
, HOST_WIDE_INT vcall_offset
,
9158 /* Make sure unwind info is emitted for the thunk if needed. */
9159 final_start_function (emit_barrier (), file
, 1);
9161 /* Operand 0 is the target function. */
9162 op
[0] = XEXP (DECL_RTL (function
), 0);
9163 if (flag_pic
&& !SYMBOL_REF_LOCAL_P (op
[0]))
9166 op
[0] = gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, op
[0]),
9167 TARGET_64BIT
? UNSPEC_PLT
: UNSPEC_GOT
);
9168 op
[0] = gen_rtx_CONST (Pmode
, op
[0]);
9171 /* Operand 1 is the 'this' pointer. */
9172 if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function
)), function
))
9173 op
[1] = gen_rtx_REG (Pmode
, 3);
9175 op
[1] = gen_rtx_REG (Pmode
, 2);
9177 /* Operand 2 is the delta. */
9178 op
[2] = GEN_INT (delta
);
9180 /* Operand 3 is the vcall_offset. */
9181 op
[3] = GEN_INT (vcall_offset
);
9183 /* Operand 4 is the temporary register. */
9184 op
[4] = gen_rtx_REG (Pmode
, 1);
9186 /* Operands 5 to 8 can be used as labels. */
9192 /* Operand 9 can be used for temporary register. */
9195 /* Generate code. */
9198 /* Setup literal pool pointer if required. */
9199 if ((!DISP_IN_RANGE (delta
)
9200 && !CONST_OK_FOR_K (delta
)
9201 && !CONST_OK_FOR_Os (delta
))
9202 || (!DISP_IN_RANGE (vcall_offset
)
9203 && !CONST_OK_FOR_K (vcall_offset
)
9204 && !CONST_OK_FOR_Os (vcall_offset
)))
9206 op
[5] = gen_label_rtx ();
9207 output_asm_insn ("larl\t%4,%5", op
);
9210 /* Add DELTA to this pointer. */
9213 if (CONST_OK_FOR_J (delta
))
9214 output_asm_insn ("la\t%1,%2(%1)", op
);
9215 else if (DISP_IN_RANGE (delta
))
9216 output_asm_insn ("lay\t%1,%2(%1)", op
);
9217 else if (CONST_OK_FOR_K (delta
))
9218 output_asm_insn ("aghi\t%1,%2", op
);
9219 else if (CONST_OK_FOR_Os (delta
))
9220 output_asm_insn ("agfi\t%1,%2", op
);
9223 op
[6] = gen_label_rtx ();
9224 output_asm_insn ("agf\t%1,%6-%5(%4)", op
);
9228 /* Perform vcall adjustment. */
9231 if (DISP_IN_RANGE (vcall_offset
))
9233 output_asm_insn ("lg\t%4,0(%1)", op
);
9234 output_asm_insn ("ag\t%1,%3(%4)", op
);
9236 else if (CONST_OK_FOR_K (vcall_offset
))
9238 output_asm_insn ("lghi\t%4,%3", op
);
9239 output_asm_insn ("ag\t%4,0(%1)", op
);
9240 output_asm_insn ("ag\t%1,0(%4)", op
);
9242 else if (CONST_OK_FOR_Os (vcall_offset
))
9244 output_asm_insn ("lgfi\t%4,%3", op
);
9245 output_asm_insn ("ag\t%4,0(%1)", op
);
9246 output_asm_insn ("ag\t%1,0(%4)", op
);
9250 op
[7] = gen_label_rtx ();
9251 output_asm_insn ("llgf\t%4,%7-%5(%4)", op
);
9252 output_asm_insn ("ag\t%4,0(%1)", op
);
9253 output_asm_insn ("ag\t%1,0(%4)", op
);
9257 /* Jump to target. */
9258 output_asm_insn ("jg\t%0", op
);
9260 /* Output literal pool if required. */
9263 output_asm_insn (".align\t4", op
);
9264 targetm
.asm_out
.internal_label (file
, "L",
9265 CODE_LABEL_NUMBER (op
[5]));
9269 targetm
.asm_out
.internal_label (file
, "L",
9270 CODE_LABEL_NUMBER (op
[6]));
9271 output_asm_insn (".long\t%2", op
);
9275 targetm
.asm_out
.internal_label (file
, "L",
9276 CODE_LABEL_NUMBER (op
[7]));
9277 output_asm_insn (".long\t%3", op
);
9282 /* Setup base pointer if required. */
9284 || (!DISP_IN_RANGE (delta
)
9285 && !CONST_OK_FOR_K (delta
)
9286 && !CONST_OK_FOR_Os (delta
))
9287 || (!DISP_IN_RANGE (delta
)
9288 && !CONST_OK_FOR_K (vcall_offset
)
9289 && !CONST_OK_FOR_Os (vcall_offset
)))
9291 op
[5] = gen_label_rtx ();
9292 output_asm_insn ("basr\t%4,0", op
);
9293 targetm
.asm_out
.internal_label (file
, "L",
9294 CODE_LABEL_NUMBER (op
[5]));
9297 /* Add DELTA to this pointer. */
9300 if (CONST_OK_FOR_J (delta
))
9301 output_asm_insn ("la\t%1,%2(%1)", op
);
9302 else if (DISP_IN_RANGE (delta
))
9303 output_asm_insn ("lay\t%1,%2(%1)", op
);
9304 else if (CONST_OK_FOR_K (delta
))
9305 output_asm_insn ("ahi\t%1,%2", op
);
9306 else if (CONST_OK_FOR_Os (delta
))
9307 output_asm_insn ("afi\t%1,%2", op
);
9310 op
[6] = gen_label_rtx ();
9311 output_asm_insn ("a\t%1,%6-%5(%4)", op
);
9315 /* Perform vcall adjustment. */
9318 if (CONST_OK_FOR_J (vcall_offset
))
9320 output_asm_insn ("l\t%4,0(%1)", op
);
9321 output_asm_insn ("a\t%1,%3(%4)", op
);
9323 else if (DISP_IN_RANGE (vcall_offset
))
9325 output_asm_insn ("l\t%4,0(%1)", op
);
9326 output_asm_insn ("ay\t%1,%3(%4)", op
);
9328 else if (CONST_OK_FOR_K (vcall_offset
))
9330 output_asm_insn ("lhi\t%4,%3", op
);
9331 output_asm_insn ("a\t%4,0(%1)", op
);
9332 output_asm_insn ("a\t%1,0(%4)", op
);
9334 else if (CONST_OK_FOR_Os (vcall_offset
))
9336 output_asm_insn ("iilf\t%4,%3", op
);
9337 output_asm_insn ("a\t%4,0(%1)", op
);
9338 output_asm_insn ("a\t%1,0(%4)", op
);
9342 op
[7] = gen_label_rtx ();
9343 output_asm_insn ("l\t%4,%7-%5(%4)", op
);
9344 output_asm_insn ("a\t%4,0(%1)", op
);
9345 output_asm_insn ("a\t%1,0(%4)", op
);
9348 /* We had to clobber the base pointer register.
9349 Re-setup the base pointer (with a different base). */
9350 op
[5] = gen_label_rtx ();
9351 output_asm_insn ("basr\t%4,0", op
);
9352 targetm
.asm_out
.internal_label (file
, "L",
9353 CODE_LABEL_NUMBER (op
[5]));
9356 /* Jump to target. */
9357 op
[8] = gen_label_rtx ();
9360 output_asm_insn ("l\t%4,%8-%5(%4)", op
);
9362 output_asm_insn ("a\t%4,%8-%5(%4)", op
);
9363 /* We cannot call through .plt, since .plt requires %r12 loaded. */
9364 else if (flag_pic
== 1)
9366 output_asm_insn ("a\t%4,%8-%5(%4)", op
);
9367 output_asm_insn ("l\t%4,%0(%4)", op
);
9369 else if (flag_pic
== 2)
9371 op
[9] = gen_rtx_REG (Pmode
, 0);
9372 output_asm_insn ("l\t%9,%8-4-%5(%4)", op
);
9373 output_asm_insn ("a\t%4,%8-%5(%4)", op
);
9374 output_asm_insn ("ar\t%4,%9", op
);
9375 output_asm_insn ("l\t%4,0(%4)", op
);
9378 output_asm_insn ("br\t%4", op
);
9380 /* Output literal pool. */
9381 output_asm_insn (".align\t4", op
);
9383 if (nonlocal
&& flag_pic
== 2)
9384 output_asm_insn (".long\t%0", op
);
9387 op
[0] = gen_rtx_SYMBOL_REF (Pmode
, "_GLOBAL_OFFSET_TABLE_");
9388 SYMBOL_REF_FLAGS (op
[0]) = SYMBOL_FLAG_LOCAL
;
9391 targetm
.asm_out
.internal_label (file
, "L", CODE_LABEL_NUMBER (op
[8]));
9393 output_asm_insn (".long\t%0", op
);
9395 output_asm_insn (".long\t%0-%5", op
);
9399 targetm
.asm_out
.internal_label (file
, "L",
9400 CODE_LABEL_NUMBER (op
[6]));
9401 output_asm_insn (".long\t%2", op
);
9405 targetm
.asm_out
.internal_label (file
, "L",
9406 CODE_LABEL_NUMBER (op
[7]));
9407 output_asm_insn (".long\t%3", op
);
9410 final_end_function ();
9414 s390_valid_pointer_mode (enum machine_mode mode
)
9416 return (mode
== SImode
|| (TARGET_64BIT
&& mode
== DImode
));
9419 /* Checks whether the given CALL_EXPR would use a caller
9420 saved register. This is used to decide whether sibling call
9421 optimization could be performed on the respective function
9425 s390_call_saved_register_used (tree call_expr
)
9427 CUMULATIVE_ARGS cum
;
9429 enum machine_mode mode
;
9434 INIT_CUMULATIVE_ARGS (cum
, NULL
, NULL
, 0, 0);
9436 for (i
= 0; i
< call_expr_nargs (call_expr
); i
++)
9438 parameter
= CALL_EXPR_ARG (call_expr
, i
);
9439 gcc_assert (parameter
);
9441 /* For an undeclared variable passed as parameter we will get
9442 an ERROR_MARK node here. */
9443 if (TREE_CODE (parameter
) == ERROR_MARK
)
9446 type
= TREE_TYPE (parameter
);
9449 mode
= TYPE_MODE (type
);
9452 if (pass_by_reference (&cum
, mode
, type
, true))
9455 type
= build_pointer_type (type
);
9458 parm_rtx
= s390_function_arg (&cum
, mode
, type
, 0);
9460 s390_function_arg_advance (&cum
, mode
, type
, 0);
9465 if (REG_P (parm_rtx
))
9468 reg
< HARD_REGNO_NREGS (REGNO (parm_rtx
), GET_MODE (parm_rtx
));
9470 if (!call_used_regs
[reg
+ REGNO (parm_rtx
)])
9474 if (GET_CODE (parm_rtx
) == PARALLEL
)
9478 for (i
= 0; i
< XVECLEN (parm_rtx
, 0); i
++)
9480 rtx r
= XEXP (XVECEXP (parm_rtx
, 0, i
), 0);
9482 gcc_assert (REG_P (r
));
9485 reg
< HARD_REGNO_NREGS (REGNO (r
), GET_MODE (r
));
9487 if (!call_used_regs
[reg
+ REGNO (r
)])
9496 /* Return true if the given call expression can be
9497 turned into a sibling call.
9498 DECL holds the declaration of the function to be called whereas
9499 EXP is the call expression itself. */
9502 s390_function_ok_for_sibcall (tree decl
, tree exp
)
9504 /* The TPF epilogue uses register 1. */
9505 if (TARGET_TPF_PROFILING
)
9508 /* The 31 bit PLT code uses register 12 (GOT pointer - caller saved)
9509 which would have to be restored before the sibcall. */
9510 if (!TARGET_64BIT
&& flag_pic
&& decl
&& !targetm
.binds_local_p (decl
))
9513 /* Register 6 on s390 is available as an argument register but unfortunately
9514 "caller saved". This makes functions needing this register for arguments
9515 not suitable for sibcalls. */
9516 return !s390_call_saved_register_used (exp
);
9519 /* Return the fixed registers used for condition codes. */
9522 s390_fixed_condition_code_regs (unsigned int *p1
, unsigned int *p2
)
9525 *p2
= INVALID_REGNUM
;
9530 /* This function is used by the call expanders of the machine description.
9531 It emits the call insn itself together with the necessary operations
9532 to adjust the target address and returns the emitted insn.
9533 ADDR_LOCATION is the target address rtx
9534 TLS_CALL the location of the thread-local symbol
9535 RESULT_REG the register where the result of the call should be stored
9536 RETADDR_REG the register where the return address should be stored
9537 If this parameter is NULL_RTX the call is considered
9538 to be a sibling call. */
9541 s390_emit_call (rtx addr_location
, rtx tls_call
, rtx result_reg
,
9544 bool plt_call
= false;
9550 /* Direct function calls need special treatment. */
9551 if (GET_CODE (addr_location
) == SYMBOL_REF
)
9553 /* When calling a global routine in PIC mode, we must
9554 replace the symbol itself with the PLT stub. */
9555 if (flag_pic
&& !SYMBOL_REF_LOCAL_P (addr_location
))
9557 if (retaddr_reg
!= NULL_RTX
)
9559 addr_location
= gen_rtx_UNSPEC (Pmode
,
9560 gen_rtvec (1, addr_location
),
9562 addr_location
= gen_rtx_CONST (Pmode
, addr_location
);
9566 /* For -fpic code the PLT entries might use r12 which is
9567 call-saved. Therefore we cannot do a sibcall when
9568 calling directly using a symbol ref. When reaching
9569 this point we decided (in s390_function_ok_for_sibcall)
9570 to do a sibcall for a function pointer but one of the
9571 optimizers was able to get rid of the function pointer
9572 by propagating the symbol ref into the call. This
9573 optimization is illegal for S/390 so we turn the direct
9574 call into a indirect call again. */
9575 addr_location
= force_reg (Pmode
, addr_location
);
9578 /* Unless we can use the bras(l) insn, force the
9579 routine address into a register. */
9580 if (!TARGET_SMALL_EXEC
&& !TARGET_CPU_ZARCH
)
9583 addr_location
= legitimize_pic_address (addr_location
, 0);
9585 addr_location
= force_reg (Pmode
, addr_location
);
9589 /* If it is already an indirect call or the code above moved the
9590 SYMBOL_REF to somewhere else make sure the address can be found in
9592 if (retaddr_reg
== NULL_RTX
9593 && GET_CODE (addr_location
) != SYMBOL_REF
9596 emit_move_insn (gen_rtx_REG (Pmode
, SIBCALL_REGNUM
), addr_location
);
9597 addr_location
= gen_rtx_REG (Pmode
, SIBCALL_REGNUM
);
9600 addr_location
= gen_rtx_MEM (QImode
, addr_location
);
9601 call
= gen_rtx_CALL (VOIDmode
, addr_location
, const0_rtx
);
9603 if (result_reg
!= NULL_RTX
)
9604 call
= gen_rtx_SET (VOIDmode
, result_reg
, call
);
9606 if (retaddr_reg
!= NULL_RTX
)
9608 clobber
= gen_rtx_CLOBBER (VOIDmode
, retaddr_reg
);
9610 if (tls_call
!= NULL_RTX
)
9611 vec
= gen_rtvec (3, call
, clobber
,
9612 gen_rtx_USE (VOIDmode
, tls_call
));
9614 vec
= gen_rtvec (2, call
, clobber
);
9616 call
= gen_rtx_PARALLEL (VOIDmode
, vec
);
9619 insn
= emit_call_insn (call
);
9621 /* 31-bit PLT stubs and tls calls use the GOT register implicitly. */
9622 if ((!TARGET_64BIT
&& plt_call
) || tls_call
!= NULL_RTX
)
9624 /* s390_function_ok_for_sibcall should
9625 have denied sibcalls in this case. */
9626 gcc_assert (retaddr_reg
!= NULL_RTX
);
9628 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), pic_offset_table_rtx
);
9633 /* Implement CONDITIONAL_REGISTER_USAGE. */
9636 s390_conditional_register_usage (void)
9642 fixed_regs
[PIC_OFFSET_TABLE_REGNUM
] = 1;
9643 call_used_regs
[PIC_OFFSET_TABLE_REGNUM
] = 1;
9645 if (TARGET_CPU_ZARCH
)
9647 fixed_regs
[BASE_REGNUM
] = 0;
9648 call_used_regs
[BASE_REGNUM
] = 0;
9649 fixed_regs
[RETURN_REGNUM
] = 0;
9650 call_used_regs
[RETURN_REGNUM
] = 0;
9654 for (i
= 24; i
< 32; i
++)
9655 call_used_regs
[i
] = call_really_used_regs
[i
] = 0;
9659 for (i
= 18; i
< 20; i
++)
9660 call_used_regs
[i
] = call_really_used_regs
[i
] = 0;
9663 if (TARGET_SOFT_FLOAT
)
9665 for (i
= 16; i
< 32; i
++)
9666 call_used_regs
[i
] = fixed_regs
[i
] = 1;
9670 /* Corresponding function to eh_return expander. */
9672 static GTY(()) rtx s390_tpf_eh_return_symbol
;
9674 s390_emit_tpf_eh_return (rtx target
)
9678 if (!s390_tpf_eh_return_symbol
)
9679 s390_tpf_eh_return_symbol
= gen_rtx_SYMBOL_REF (Pmode
, "__tpf_eh_return");
9681 reg
= gen_rtx_REG (Pmode
, 2);
9683 emit_move_insn (reg
, target
);
9684 insn
= s390_emit_call (s390_tpf_eh_return_symbol
, NULL_RTX
, reg
,
9685 gen_rtx_REG (Pmode
, RETURN_REGNUM
));
9686 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), reg
);
9688 emit_move_insn (EH_RETURN_HANDLER_RTX
, reg
);
9691 /* Rework the prologue/epilogue to avoid saving/restoring
9692 registers unnecessarily. */
9695 s390_optimize_prologue (void)
9697 rtx insn
, new_insn
, next_insn
;
9699 /* Do a final recompute of the frame-related data. */
9701 s390_update_frame_layout ();
9703 /* If all special registers are in fact used, there's nothing we
9704 can do, so no point in walking the insn list. */
9706 if (cfun_frame_layout
.first_save_gpr
<= BASE_REGNUM
9707 && cfun_frame_layout
.last_save_gpr
>= BASE_REGNUM
9708 && (TARGET_CPU_ZARCH
9709 || (cfun_frame_layout
.first_save_gpr
<= RETURN_REGNUM
9710 && cfun_frame_layout
.last_save_gpr
>= RETURN_REGNUM
)))
9713 /* Search for prologue/epilogue insns and replace them. */
9715 for (insn
= get_insns (); insn
; insn
= next_insn
)
9717 int first
, last
, off
;
9718 rtx set
, base
, offset
;
9720 next_insn
= NEXT_INSN (insn
);
9722 if (GET_CODE (insn
) != INSN
)
9725 if (GET_CODE (PATTERN (insn
)) == PARALLEL
9726 && store_multiple_operation (PATTERN (insn
), VOIDmode
))
9728 set
= XVECEXP (PATTERN (insn
), 0, 0);
9729 first
= REGNO (SET_SRC (set
));
9730 last
= first
+ XVECLEN (PATTERN (insn
), 0) - 1;
9731 offset
= const0_rtx
;
9732 base
= eliminate_constant_term (XEXP (SET_DEST (set
), 0), &offset
);
9733 off
= INTVAL (offset
);
9735 if (GET_CODE (base
) != REG
|| off
< 0)
9737 if (cfun_frame_layout
.first_save_gpr
!= -1
9738 && (cfun_frame_layout
.first_save_gpr
< first
9739 || cfun_frame_layout
.last_save_gpr
> last
))
9741 if (REGNO (base
) != STACK_POINTER_REGNUM
9742 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9744 if (first
> BASE_REGNUM
|| last
< BASE_REGNUM
)
9747 if (cfun_frame_layout
.first_save_gpr
!= -1)
9749 new_insn
= save_gprs (base
,
9750 off
+ (cfun_frame_layout
.first_save_gpr
9751 - first
) * UNITS_PER_LONG
,
9752 cfun_frame_layout
.first_save_gpr
,
9753 cfun_frame_layout
.last_save_gpr
);
9754 new_insn
= emit_insn_before (new_insn
, insn
);
9755 INSN_ADDRESSES_NEW (new_insn
, -1);
9762 if (cfun_frame_layout
.first_save_gpr
== -1
9763 && GET_CODE (PATTERN (insn
)) == SET
9764 && GET_CODE (SET_SRC (PATTERN (insn
))) == REG
9765 && (REGNO (SET_SRC (PATTERN (insn
))) == BASE_REGNUM
9766 || (!TARGET_CPU_ZARCH
9767 && REGNO (SET_SRC (PATTERN (insn
))) == RETURN_REGNUM
))
9768 && GET_CODE (SET_DEST (PATTERN (insn
))) == MEM
)
9770 set
= PATTERN (insn
);
9771 first
= REGNO (SET_SRC (set
));
9772 offset
= const0_rtx
;
9773 base
= eliminate_constant_term (XEXP (SET_DEST (set
), 0), &offset
);
9774 off
= INTVAL (offset
);
9776 if (GET_CODE (base
) != REG
|| off
< 0)
9778 if (REGNO (base
) != STACK_POINTER_REGNUM
9779 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9786 if (GET_CODE (PATTERN (insn
)) == PARALLEL
9787 && load_multiple_operation (PATTERN (insn
), VOIDmode
))
9789 set
= XVECEXP (PATTERN (insn
), 0, 0);
9790 first
= REGNO (SET_DEST (set
));
9791 last
= first
+ XVECLEN (PATTERN (insn
), 0) - 1;
9792 offset
= const0_rtx
;
9793 base
= eliminate_constant_term (XEXP (SET_SRC (set
), 0), &offset
);
9794 off
= INTVAL (offset
);
9796 if (GET_CODE (base
) != REG
|| off
< 0)
9798 if (cfun_frame_layout
.first_restore_gpr
!= -1
9799 && (cfun_frame_layout
.first_restore_gpr
< first
9800 || cfun_frame_layout
.last_restore_gpr
> last
))
9802 if (REGNO (base
) != STACK_POINTER_REGNUM
9803 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9805 if (first
> BASE_REGNUM
|| last
< BASE_REGNUM
)
9808 if (cfun_frame_layout
.first_restore_gpr
!= -1)
9810 new_insn
= restore_gprs (base
,
9811 off
+ (cfun_frame_layout
.first_restore_gpr
9812 - first
) * UNITS_PER_LONG
,
9813 cfun_frame_layout
.first_restore_gpr
,
9814 cfun_frame_layout
.last_restore_gpr
);
9815 new_insn
= emit_insn_before (new_insn
, insn
);
9816 INSN_ADDRESSES_NEW (new_insn
, -1);
9823 if (cfun_frame_layout
.first_restore_gpr
== -1
9824 && GET_CODE (PATTERN (insn
)) == SET
9825 && GET_CODE (SET_DEST (PATTERN (insn
))) == REG
9826 && (REGNO (SET_DEST (PATTERN (insn
))) == BASE_REGNUM
9827 || (!TARGET_CPU_ZARCH
9828 && REGNO (SET_DEST (PATTERN (insn
))) == RETURN_REGNUM
))
9829 && GET_CODE (SET_SRC (PATTERN (insn
))) == MEM
)
9831 set
= PATTERN (insn
);
9832 first
= REGNO (SET_DEST (set
));
9833 offset
= const0_rtx
;
9834 base
= eliminate_constant_term (XEXP (SET_SRC (set
), 0), &offset
);
9835 off
= INTVAL (offset
);
9837 if (GET_CODE (base
) != REG
|| off
< 0)
9839 if (REGNO (base
) != STACK_POINTER_REGNUM
9840 && REGNO (base
) != HARD_FRAME_POINTER_REGNUM
)
9849 /* On z10 the dynamic branch prediction must see the backward jump in
9850 a window of 384 bytes. If not it falls back to the static
9851 prediction. This function rearranges the loop backward branch in a
9852 way which makes the static prediction always correct. The function
9853 returns true if it added an instruction. */
9855 s390_z10_fix_long_loop_prediction (rtx insn
)
9857 rtx set
= single_set (insn
);
9858 rtx code_label
, label_ref
, new_label
;
9864 /* This will exclude branch on count and branch on index patterns
9865 since these are correctly statically predicted. */
9867 || SET_DEST (set
) != pc_rtx
9868 || GET_CODE (SET_SRC(set
)) != IF_THEN_ELSE
)
9871 label_ref
= (GET_CODE (XEXP (SET_SRC (set
), 1)) == LABEL_REF
?
9872 XEXP (SET_SRC (set
), 1) : XEXP (SET_SRC (set
), 2));
9874 gcc_assert (GET_CODE (label_ref
) == LABEL_REF
);
9876 code_label
= XEXP (label_ref
, 0);
9878 if (INSN_ADDRESSES (INSN_UID (code_label
)) == -1
9879 || INSN_ADDRESSES (INSN_UID (insn
)) == -1
9880 || (INSN_ADDRESSES (INSN_UID (insn
))
9881 - INSN_ADDRESSES (INSN_UID (code_label
)) < Z10_PREDICT_DISTANCE
))
9884 for (distance
= 0, cur_insn
= PREV_INSN (insn
);
9885 distance
< Z10_PREDICT_DISTANCE
- 6;
9886 distance
+= get_attr_length (cur_insn
), cur_insn
= PREV_INSN (cur_insn
))
9887 if (!cur_insn
|| JUMP_P (cur_insn
) || LABEL_P (cur_insn
))
9890 new_label
= gen_label_rtx ();
9891 uncond_jump
= emit_jump_insn_after (
9892 gen_rtx_SET (VOIDmode
, pc_rtx
,
9893 gen_rtx_LABEL_REF (VOIDmode
, code_label
)),
9895 emit_label_after (new_label
, uncond_jump
);
9897 tmp
= XEXP (SET_SRC (set
), 1);
9898 XEXP (SET_SRC (set
), 1) = XEXP (SET_SRC (set
), 2);
9899 XEXP (SET_SRC (set
), 2) = tmp
;
9900 INSN_CODE (insn
) = -1;
9902 XEXP (label_ref
, 0) = new_label
;
9903 JUMP_LABEL (insn
) = new_label
;
9904 JUMP_LABEL (uncond_jump
) = code_label
;
9909 /* Returns 1 if INSN reads the value of REG for purposes not related
9910 to addressing of memory, and 0 otherwise. */
9912 s390_non_addr_reg_read_p (rtx reg
, rtx insn
)
9914 return reg_referenced_p (reg
, PATTERN (insn
))
9915 && !reg_used_in_mem_p (REGNO (reg
), PATTERN (insn
));
9918 /* Starting from INSN find_cond_jump looks downwards in the insn
9919 stream for a single jump insn which is the last user of the
9920 condition code set in INSN. */
9922 find_cond_jump (rtx insn
)
9924 for (; insn
; insn
= NEXT_INSN (insn
))
9933 if (reg_mentioned_p (gen_rtx_REG (CCmode
, CC_REGNUM
), insn
))
9938 /* This will be triggered by a return. */
9939 if (GET_CODE (PATTERN (insn
)) != SET
)
9942 gcc_assert (SET_DEST (PATTERN (insn
)) == pc_rtx
);
9943 ite
= SET_SRC (PATTERN (insn
));
9945 if (GET_CODE (ite
) != IF_THEN_ELSE
)
9948 cc
= XEXP (XEXP (ite
, 0), 0);
9949 if (!REG_P (cc
) || !CC_REGNO_P (REGNO (cc
)))
9952 if (find_reg_note (insn
, REG_DEAD
, cc
))
9960 /* Swap the condition in COND and the operands in OP0 and OP1 so that
9961 the semantics does not change. If NULL_RTX is passed as COND the
9962 function tries to find the conditional jump starting with INSN. */
9964 s390_swap_cmp (rtx cond
, rtx
*op0
, rtx
*op1
, rtx insn
)
9968 if (cond
== NULL_RTX
)
9970 rtx jump
= find_cond_jump (NEXT_INSN (insn
));
9971 jump
= jump
? single_set (jump
) : NULL_RTX
;
9973 if (jump
== NULL_RTX
)
9976 cond
= XEXP (XEXP (jump
, 1), 0);
9981 PUT_CODE (cond
, swap_condition (GET_CODE (cond
)));
9984 /* On z10, instructions of the compare-and-branch family have the
9985 property to access the register occurring as second operand with
9986 its bits complemented. If such a compare is grouped with a second
9987 instruction that accesses the same register non-complemented, and
9988 if that register's value is delivered via a bypass, then the
9989 pipeline recycles, thereby causing significant performance decline.
9990 This function locates such situations and exchanges the two
9991 operands of the compare. The function return true whenever it
9994 s390_z10_optimize_cmp (rtx insn
)
9996 rtx prev_insn
, next_insn
;
9997 bool insn_added_p
= false;
9998 rtx cond
, *op0
, *op1
;
10000 if (GET_CODE (PATTERN (insn
)) == PARALLEL
)
10002 /* Handle compare and branch and branch on count
10004 rtx pattern
= single_set (insn
);
10007 || SET_DEST (pattern
) != pc_rtx
10008 || GET_CODE (SET_SRC (pattern
)) != IF_THEN_ELSE
)
10011 cond
= XEXP (SET_SRC (pattern
), 0);
10012 op0
= &XEXP (cond
, 0);
10013 op1
= &XEXP (cond
, 1);
10015 else if (GET_CODE (PATTERN (insn
)) == SET
)
10019 /* Handle normal compare instructions. */
10020 src
= SET_SRC (PATTERN (insn
));
10021 dest
= SET_DEST (PATTERN (insn
));
10024 || !CC_REGNO_P (REGNO (dest
))
10025 || GET_CODE (src
) != COMPARE
)
10028 /* s390_swap_cmp will try to find the conditional
10029 jump when passing NULL_RTX as condition. */
10031 op0
= &XEXP (src
, 0);
10032 op1
= &XEXP (src
, 1);
10037 if (!REG_P (*op0
) || !REG_P (*op1
))
10040 if (GET_MODE_CLASS (GET_MODE (*op0
)) != MODE_INT
)
10043 /* Swap the COMPARE arguments and its mask if there is a
10044 conflicting access in the previous insn. */
10045 prev_insn
= prev_active_insn (insn
);
10046 if (prev_insn
!= NULL_RTX
&& INSN_P (prev_insn
)
10047 && reg_referenced_p (*op1
, PATTERN (prev_insn
)))
10048 s390_swap_cmp (cond
, op0
, op1
, insn
);
10050 /* Check if there is a conflict with the next insn. If there
10051 was no conflict with the previous insn, then swap the
10052 COMPARE arguments and its mask. If we already swapped
10053 the operands, or if swapping them would cause a conflict
10054 with the previous insn, issue a NOP after the COMPARE in
10055 order to separate the two instuctions. */
10056 next_insn
= next_active_insn (insn
);
10057 if (next_insn
!= NULL_RTX
&& INSN_P (next_insn
)
10058 && s390_non_addr_reg_read_p (*op1
, next_insn
))
10060 if (prev_insn
!= NULL_RTX
&& INSN_P (prev_insn
)
10061 && s390_non_addr_reg_read_p (*op0
, prev_insn
))
10063 if (REGNO (*op1
) == 0)
10064 emit_insn_after (gen_nop1 (), insn
);
10066 emit_insn_after (gen_nop (), insn
);
10067 insn_added_p
= true;
10070 s390_swap_cmp (cond
, op0
, op1
, insn
);
10072 return insn_added_p
;
10075 /* Perform machine-dependent processing. */
10080 bool pool_overflow
= false;
10082 /* Make sure all splits have been performed; splits after
10083 machine_dependent_reorg might confuse insn length counts. */
10084 split_all_insns_noflow ();
10086 /* Install the main literal pool and the associated base
10087 register load insns.
10089 In addition, there are two problematic situations we need
10092 - the literal pool might be > 4096 bytes in size, so that
10093 some of its elements cannot be directly accessed
10095 - a branch target might be > 64K away from the branch, so that
10096 it is not possible to use a PC-relative instruction.
10098 To fix those, we split the single literal pool into multiple
10099 pool chunks, reloading the pool base register at various
10100 points throughout the function to ensure it always points to
10101 the pool chunk the following code expects, and / or replace
10102 PC-relative branches by absolute branches.
10104 However, the two problems are interdependent: splitting the
10105 literal pool can move a branch further away from its target,
10106 causing the 64K limit to overflow, and on the other hand,
10107 replacing a PC-relative branch by an absolute branch means
10108 we need to put the branch target address into the literal
10109 pool, possibly causing it to overflow.
10111 So, we loop trying to fix up both problems until we manage
10112 to satisfy both conditions at the same time. Note that the
10113 loop is guaranteed to terminate as every pass of the loop
10114 strictly decreases the total number of PC-relative branches
10115 in the function. (This is not completely true as there
10116 might be branch-over-pool insns introduced by chunkify_start.
10117 Those never need to be split however.) */
10121 struct constant_pool
*pool
= NULL
;
10123 /* Collect the literal pool. */
10124 if (!pool_overflow
)
10126 pool
= s390_mainpool_start ();
10128 pool_overflow
= true;
10131 /* If literal pool overflowed, start to chunkify it. */
10133 pool
= s390_chunkify_start ();
10135 /* Split out-of-range branches. If this has created new
10136 literal pool entries, cancel current chunk list and
10137 recompute it. zSeries machines have large branch
10138 instructions, so we never need to split a branch. */
10139 if (!TARGET_CPU_ZARCH
&& s390_split_branches ())
10142 s390_chunkify_cancel (pool
);
10144 s390_mainpool_cancel (pool
);
10149 /* If we made it up to here, both conditions are satisfied.
10150 Finish up literal pool related changes. */
10152 s390_chunkify_finish (pool
);
10154 s390_mainpool_finish (pool
);
10156 /* We're done splitting branches. */
10157 cfun
->machine
->split_branches_pending_p
= false;
10161 /* Generate out-of-pool execute target insns. */
10162 if (TARGET_CPU_ZARCH
)
10164 rtx insn
, label
, target
;
10166 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
10168 label
= s390_execute_label (insn
);
10172 gcc_assert (label
!= const0_rtx
);
10174 target
= emit_label (XEXP (label
, 0));
10175 INSN_ADDRESSES_NEW (target
, -1);
10177 target
= emit_insn (s390_execute_target (insn
));
10178 INSN_ADDRESSES_NEW (target
, -1);
10182 /* Try to optimize prologue and epilogue further. */
10183 s390_optimize_prologue ();
10185 /* Walk over the insns and do some z10 specific changes. */
10186 if (s390_tune
== PROCESSOR_2097_Z10
)
10189 bool insn_added_p
= false;
10191 /* The insn lengths and addresses have to be up to date for the
10192 following manipulations. */
10193 shorten_branches (get_insns ());
10195 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
10197 if (!INSN_P (insn
) || INSN_CODE (insn
) <= 0)
10201 insn_added_p
|= s390_z10_fix_long_loop_prediction (insn
);
10203 if (GET_CODE (PATTERN (insn
)) == PARALLEL
10204 || GET_CODE (PATTERN (insn
)) == SET
)
10205 insn_added_p
|= s390_z10_optimize_cmp (insn
);
10208 /* Adjust branches if we added new instructions. */
10210 shorten_branches (get_insns ());
10214 /* Return true if INSN is a fp load insn writing register REGNO. */
10216 s390_fpload_toreg (rtx insn
, unsigned int regno
)
10219 enum attr_type flag
= s390_safe_attr_type (insn
);
10221 if (flag
!= TYPE_FLOADSF
&& flag
!= TYPE_FLOADDF
)
10224 set
= single_set (insn
);
10226 if (set
== NULL_RTX
)
10229 if (!REG_P (SET_DEST (set
)) || !MEM_P (SET_SRC (set
)))
10232 if (REGNO (SET_DEST (set
)) != regno
)
10238 /* This value describes the distance to be avoided between an
10239 aritmetic fp instruction and an fp load writing the same register.
10240 Z10_EARLYLOAD_DISTANCE - 1 as well as Z10_EARLYLOAD_DISTANCE + 1 is
10241 fine but the exact value has to be avoided. Otherwise the FP
10242 pipeline will throw an exception causing a major penalty. */
10243 #define Z10_EARLYLOAD_DISTANCE 7
10245 /* Rearrange the ready list in order to avoid the situation described
10246 for Z10_EARLYLOAD_DISTANCE. A problematic load instruction is
10247 moved to the very end of the ready list. */
10249 s390_z10_prevent_earlyload_conflicts (rtx
*ready
, int *nready_p
)
10251 unsigned int regno
;
10252 int nready
= *nready_p
;
10257 enum attr_type flag
;
10260 /* Skip DISTANCE - 1 active insns. */
10261 for (insn
= last_scheduled_insn
, distance
= Z10_EARLYLOAD_DISTANCE
- 1;
10262 distance
> 0 && insn
!= NULL_RTX
;
10263 distance
--, insn
= prev_active_insn (insn
))
10264 if (CALL_P (insn
) || JUMP_P (insn
))
10267 if (insn
== NULL_RTX
)
10270 set
= single_set (insn
);
10272 if (set
== NULL_RTX
|| !REG_P (SET_DEST (set
))
10273 || GET_MODE_CLASS (GET_MODE (SET_DEST (set
))) != MODE_FLOAT
)
10276 flag
= s390_safe_attr_type (insn
);
10278 if (flag
== TYPE_FLOADSF
|| flag
== TYPE_FLOADDF
)
10281 regno
= REGNO (SET_DEST (set
));
10284 while (!s390_fpload_toreg (ready
[i
], regno
) && i
> 0)
10291 memmove (&ready
[1], &ready
[0], sizeof (rtx
) * i
);
10295 /* This function is called via hook TARGET_SCHED_REORDER before
10296 issueing one insn from list READY which contains *NREADYP entries.
10297 For target z10 it reorders load instructions to avoid early load
10298 conflicts in the floating point pipeline */
10300 s390_sched_reorder (FILE *file ATTRIBUTE_UNUSED
, int verbose ATTRIBUTE_UNUSED
,
10301 rtx
*ready
, int *nreadyp
, int clock ATTRIBUTE_UNUSED
)
10303 if (s390_tune
== PROCESSOR_2097_Z10
)
10304 if (reload_completed
&& *nreadyp
> 1)
10305 s390_z10_prevent_earlyload_conflicts (ready
, nreadyp
);
10307 return s390_issue_rate ();
10310 /* This function is called via hook TARGET_SCHED_VARIABLE_ISSUE after
10311 the scheduler has issued INSN. It stores the last issued insn into
10312 last_scheduled_insn in order to make it available for
10313 s390_sched_reorder. */
10315 s390_sched_variable_issue (FILE *file ATTRIBUTE_UNUSED
,
10316 int verbose ATTRIBUTE_UNUSED
,
10317 rtx insn
, int more
)
10319 last_scheduled_insn
= insn
;
10321 if (GET_CODE (PATTERN (insn
)) != USE
10322 && GET_CODE (PATTERN (insn
)) != CLOBBER
)
10329 s390_sched_init (FILE *file ATTRIBUTE_UNUSED
,
10330 int verbose ATTRIBUTE_UNUSED
,
10331 int max_ready ATTRIBUTE_UNUSED
)
10333 last_scheduled_insn
= NULL_RTX
;
10336 /* This function checks the whole of insn X for memory references. The
10337 function always returns zero because the framework it is called
10338 from would stop recursively analyzing the insn upon a return value
10339 other than zero. The real result of this function is updating
10340 counter variable MEM_COUNT. */
10342 check_dpu (rtx
*x
, unsigned *mem_count
)
10344 if (*x
!= NULL_RTX
&& MEM_P (*x
))
10349 /* This target hook implementation for TARGET_LOOP_UNROLL_ADJUST calculates
10350 a new number struct loop *loop should be unrolled if tuned for the z10
10351 cpu. The loop is analyzed for memory accesses by calling check_dpu for
10352 each rtx of the loop. Depending on the loop_depth and the amount of
10353 memory accesses a new number <=nunroll is returned to improve the
10354 behaviour of the hardware prefetch unit. */
10356 s390_loop_unroll_adjust (unsigned nunroll
, struct loop
*loop
)
10361 unsigned mem_count
= 0;
10363 /* Only z10 needs special handling. */
10364 if (s390_tune
!= PROCESSOR_2097_Z10
)
10367 /* Count the number of memory references within the loop body. */
10368 bbs
= get_loop_body (loop
);
10369 for (i
= 0; i
< loop
->num_nodes
; i
++)
10371 for (insn
= BB_HEAD (bbs
[i
]); insn
!= BB_END (bbs
[i
]); insn
= NEXT_INSN (insn
))
10372 if (INSN_P (insn
) && INSN_CODE (insn
) != -1)
10373 for_each_rtx (&insn
, (rtx_function
) check_dpu
, &mem_count
);
10377 /* Prevent division by zero, and we do not need to adjust nunroll in this case. */
10378 if (mem_count
== 0)
10381 switch (loop_depth(loop
))
10384 return MIN (nunroll
, 28 / mem_count
);
10386 return MIN (nunroll
, 22 / mem_count
);
10388 return MIN (nunroll
, 16 / mem_count
);
10392 /* Initialize GCC target structure. */
10394 #undef TARGET_ASM_ALIGNED_HI_OP
10395 #define TARGET_ASM_ALIGNED_HI_OP "\t.word\t"
10396 #undef TARGET_ASM_ALIGNED_DI_OP
10397 #define TARGET_ASM_ALIGNED_DI_OP "\t.quad\t"
10398 #undef TARGET_ASM_INTEGER
10399 #define TARGET_ASM_INTEGER s390_assemble_integer
10401 #undef TARGET_ASM_OPEN_PAREN
10402 #define TARGET_ASM_OPEN_PAREN ""
10404 #undef TARGET_ASM_CLOSE_PAREN
10405 #define TARGET_ASM_CLOSE_PAREN ""
10407 #undef TARGET_DEFAULT_TARGET_FLAGS
10408 #define TARGET_DEFAULT_TARGET_FLAGS (TARGET_DEFAULT | MASK_FUSED_MADD)
10410 #undef TARGET_HANDLE_OPTION
10411 #define TARGET_HANDLE_OPTION s390_handle_option
10413 #undef TARGET_ENCODE_SECTION_INFO
10414 #define TARGET_ENCODE_SECTION_INFO s390_encode_section_info
10416 #undef TARGET_SCALAR_MODE_SUPPORTED_P
10417 #define TARGET_SCALAR_MODE_SUPPORTED_P s390_scalar_mode_supported_p
10420 #undef TARGET_HAVE_TLS
10421 #define TARGET_HAVE_TLS true
10423 #undef TARGET_CANNOT_FORCE_CONST_MEM
10424 #define TARGET_CANNOT_FORCE_CONST_MEM s390_cannot_force_const_mem
10426 #undef TARGET_DELEGITIMIZE_ADDRESS
10427 #define TARGET_DELEGITIMIZE_ADDRESS s390_delegitimize_address
10429 #undef TARGET_LEGITIMIZE_ADDRESS
10430 #define TARGET_LEGITIMIZE_ADDRESS s390_legitimize_address
10432 #undef TARGET_RETURN_IN_MEMORY
10433 #define TARGET_RETURN_IN_MEMORY s390_return_in_memory
10435 #undef TARGET_INIT_BUILTINS
10436 #define TARGET_INIT_BUILTINS s390_init_builtins
10437 #undef TARGET_EXPAND_BUILTIN
10438 #define TARGET_EXPAND_BUILTIN s390_expand_builtin
10440 #undef TARGET_ASM_OUTPUT_MI_THUNK
10441 #define TARGET_ASM_OUTPUT_MI_THUNK s390_output_mi_thunk
10442 #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
10443 #define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_const_tree_hwi_hwi_const_tree_true
10445 #undef TARGET_SCHED_ADJUST_PRIORITY
10446 #define TARGET_SCHED_ADJUST_PRIORITY s390_adjust_priority
10447 #undef TARGET_SCHED_ISSUE_RATE
10448 #define TARGET_SCHED_ISSUE_RATE s390_issue_rate
10449 #undef TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD
10450 #define TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD s390_first_cycle_multipass_dfa_lookahead
10452 #undef TARGET_SCHED_VARIABLE_ISSUE
10453 #define TARGET_SCHED_VARIABLE_ISSUE s390_sched_variable_issue
10454 #undef TARGET_SCHED_REORDER
10455 #define TARGET_SCHED_REORDER s390_sched_reorder
10456 #undef TARGET_SCHED_INIT
10457 #define TARGET_SCHED_INIT s390_sched_init
10459 #undef TARGET_CANNOT_COPY_INSN_P
10460 #define TARGET_CANNOT_COPY_INSN_P s390_cannot_copy_insn_p
10461 #undef TARGET_RTX_COSTS
10462 #define TARGET_RTX_COSTS s390_rtx_costs
10463 #undef TARGET_ADDRESS_COST
10464 #define TARGET_ADDRESS_COST s390_address_cost
10466 #undef TARGET_MACHINE_DEPENDENT_REORG
10467 #define TARGET_MACHINE_DEPENDENT_REORG s390_reorg
10469 #undef TARGET_VALID_POINTER_MODE
10470 #define TARGET_VALID_POINTER_MODE s390_valid_pointer_mode
10472 #undef TARGET_BUILD_BUILTIN_VA_LIST
10473 #define TARGET_BUILD_BUILTIN_VA_LIST s390_build_builtin_va_list
10474 #undef TARGET_EXPAND_BUILTIN_VA_START
10475 #define TARGET_EXPAND_BUILTIN_VA_START s390_va_start
10476 #undef TARGET_GIMPLIFY_VA_ARG_EXPR
10477 #define TARGET_GIMPLIFY_VA_ARG_EXPR s390_gimplify_va_arg
10479 #undef TARGET_PROMOTE_FUNCTION_MODE
10480 #define TARGET_PROMOTE_FUNCTION_MODE s390_promote_function_mode
10481 #undef TARGET_PASS_BY_REFERENCE
10482 #define TARGET_PASS_BY_REFERENCE s390_pass_by_reference
10484 #undef TARGET_FUNCTION_OK_FOR_SIBCALL
10485 #define TARGET_FUNCTION_OK_FOR_SIBCALL s390_function_ok_for_sibcall
10487 #undef TARGET_FIXED_CONDITION_CODE_REGS
10488 #define TARGET_FIXED_CONDITION_CODE_REGS s390_fixed_condition_code_regs
10490 #undef TARGET_CC_MODES_COMPATIBLE
10491 #define TARGET_CC_MODES_COMPATIBLE s390_cc_modes_compatible
10493 #undef TARGET_INVALID_WITHIN_DOLOOP
10494 #define TARGET_INVALID_WITHIN_DOLOOP hook_constcharptr_const_rtx_null
10497 #undef TARGET_ASM_OUTPUT_DWARF_DTPREL
10498 #define TARGET_ASM_OUTPUT_DWARF_DTPREL s390_output_dwarf_dtprel
10501 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
10502 #undef TARGET_MANGLE_TYPE
10503 #define TARGET_MANGLE_TYPE s390_mangle_type
10506 #undef TARGET_SCALAR_MODE_SUPPORTED_P
10507 #define TARGET_SCALAR_MODE_SUPPORTED_P s390_scalar_mode_supported_p
10509 #undef TARGET_SECONDARY_RELOAD
10510 #define TARGET_SECONDARY_RELOAD s390_secondary_reload
10512 #undef TARGET_LIBGCC_CMP_RETURN_MODE
10513 #define TARGET_LIBGCC_CMP_RETURN_MODE s390_libgcc_cmp_return_mode
10515 #undef TARGET_LIBGCC_SHIFT_COUNT_MODE
10516 #define TARGET_LIBGCC_SHIFT_COUNT_MODE s390_libgcc_shift_count_mode
10518 #undef TARGET_LEGITIMATE_ADDRESS_P
10519 #define TARGET_LEGITIMATE_ADDRESS_P s390_legitimate_address_p
10521 #undef TARGET_CAN_ELIMINATE
10522 #define TARGET_CAN_ELIMINATE s390_can_eliminate
10524 #undef TARGET_LOOP_UNROLL_ADJUST
10525 #define TARGET_LOOP_UNROLL_ADJUST s390_loop_unroll_adjust
10527 #undef TARGET_ASM_TRAMPOLINE_TEMPLATE
10528 #define TARGET_ASM_TRAMPOLINE_TEMPLATE s390_asm_trampoline_template
10529 #undef TARGET_TRAMPOLINE_INIT
10530 #define TARGET_TRAMPOLINE_INIT s390_trampoline_init
10532 #undef TARGET_UNWIND_WORD_MODE
10533 #define TARGET_UNWIND_WORD_MODE s390_unwind_word_mode
10535 struct gcc_target targetm
= TARGET_INITIALIZER
;
10537 #include "gt-s390.h"