alpha.c: Remove mentions of deprecates macros in comments...
[gcc.git] / gcc / config / alpha / alpha.c
1 /* Subroutines used for code generation on the DEC Alpha.
2 Copyright (C) 1992, 1993, 1994, 1995, 1996, 1997, 1998, 1999,
3 2000, 2001, 2002, 2003 Free Software Foundation, Inc.
4 Contributed by Richard Kenner (kenner@vlsi1.ultra.nyu.edu)
5
6 This file is part of GCC.
7
8 GCC is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 2, or (at your option)
11 any later version.
12
13 GCC is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
17
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING. If not, write to
20 the Free Software Foundation, 59 Temple Place - Suite 330,
21 Boston, MA 02111-1307, USA. */
22
23
24 #include "config.h"
25 #include "system.h"
26 #include "coretypes.h"
27 #include "tm.h"
28 #include "rtl.h"
29 #include "tree.h"
30 #include "regs.h"
31 #include "hard-reg-set.h"
32 #include "real.h"
33 #include "insn-config.h"
34 #include "conditions.h"
35 #include "output.h"
36 #include "insn-attr.h"
37 #include "flags.h"
38 #include "recog.h"
39 #include "expr.h"
40 #include "optabs.h"
41 #include "reload.h"
42 #include "obstack.h"
43 #include "except.h"
44 #include "function.h"
45 #include "toplev.h"
46 #include "ggc.h"
47 #include "integrate.h"
48 #include "tm_p.h"
49 #include "target.h"
50 #include "target-def.h"
51 #include "debug.h"
52 #include "langhooks.h"
53 #include <splay-tree.h>
54 #include "cfglayout.h"
55
56 /* Specify which cpu to schedule for. */
57
58 enum processor_type alpha_cpu;
59 static const char * const alpha_cpu_name[] =
60 {
61 "ev4", "ev5", "ev6"
62 };
63
64 /* Specify how accurate floating-point traps need to be. */
65
66 enum alpha_trap_precision alpha_tp;
67
68 /* Specify the floating-point rounding mode. */
69
70 enum alpha_fp_rounding_mode alpha_fprm;
71
72 /* Specify which things cause traps. */
73
74 enum alpha_fp_trap_mode alpha_fptm;
75
76 /* Specify bit size of immediate TLS offsets. */
77
78 int alpha_tls_size = 32;
79
80 /* Strings decoded into the above options. */
81
82 const char *alpha_cpu_string; /* -mcpu= */
83 const char *alpha_tune_string; /* -mtune= */
84 const char *alpha_tp_string; /* -mtrap-precision=[p|s|i] */
85 const char *alpha_fprm_string; /* -mfp-rounding-mode=[n|m|c|d] */
86 const char *alpha_fptm_string; /* -mfp-trap-mode=[n|u|su|sui] */
87 const char *alpha_mlat_string; /* -mmemory-latency= */
88 const char *alpha_tls_size_string; /* -mtls-size=[16|32|64] */
89
90 /* Save information from a "cmpxx" operation until the branch or scc is
91 emitted. */
92
93 struct alpha_compare alpha_compare;
94
95 /* Nonzero if inside of a function, because the Alpha asm can't
96 handle .files inside of functions. */
97
98 static int inside_function = FALSE;
99
100 /* The number of cycles of latency we should assume on memory reads. */
101
102 int alpha_memory_latency = 3;
103
104 /* Whether the function needs the GP. */
105
106 static int alpha_function_needs_gp;
107
108 /* The alias set for prologue/epilogue register save/restore. */
109
110 static GTY(()) int alpha_sr_alias_set;
111
112 /* The assembler name of the current function. */
113
114 static const char *alpha_fnname;
115
116 /* The next explicit relocation sequence number. */
117 extern GTY(()) int alpha_next_sequence_number;
118 int alpha_next_sequence_number = 1;
119
120 /* The literal and gpdisp sequence numbers for this insn, as printed
121 by %# and %* respectively. */
122 extern GTY(()) int alpha_this_literal_sequence_number;
123 extern GTY(()) int alpha_this_gpdisp_sequence_number;
124 int alpha_this_literal_sequence_number;
125 int alpha_this_gpdisp_sequence_number;
126
127 /* Costs of various operations on the different architectures. */
128
129 struct alpha_rtx_cost_data
130 {
131 unsigned char fp_add;
132 unsigned char fp_mult;
133 unsigned char fp_div_sf;
134 unsigned char fp_div_df;
135 unsigned char int_mult_si;
136 unsigned char int_mult_di;
137 unsigned char int_shift;
138 unsigned char int_cmov;
139 };
140
141 static struct alpha_rtx_cost_data const alpha_rtx_cost_data[PROCESSOR_MAX] =
142 {
143 { /* EV4 */
144 COSTS_N_INSNS (6), /* fp_add */
145 COSTS_N_INSNS (6), /* fp_mult */
146 COSTS_N_INSNS (34), /* fp_div_sf */
147 COSTS_N_INSNS (63), /* fp_div_df */
148 COSTS_N_INSNS (23), /* int_mult_si */
149 COSTS_N_INSNS (23), /* int_mult_di */
150 COSTS_N_INSNS (2), /* int_shift */
151 COSTS_N_INSNS (2), /* int_cmov */
152 },
153 { /* EV5 */
154 COSTS_N_INSNS (4), /* fp_add */
155 COSTS_N_INSNS (4), /* fp_mult */
156 COSTS_N_INSNS (15), /* fp_div_sf */
157 COSTS_N_INSNS (22), /* fp_div_df */
158 COSTS_N_INSNS (8), /* int_mult_si */
159 COSTS_N_INSNS (12), /* int_mult_di */
160 COSTS_N_INSNS (1) + 1, /* int_shift */
161 COSTS_N_INSNS (1), /* int_cmov */
162 },
163 { /* EV6 */
164 COSTS_N_INSNS (4), /* fp_add */
165 COSTS_N_INSNS (4), /* fp_mult */
166 COSTS_N_INSNS (12), /* fp_div_sf */
167 COSTS_N_INSNS (15), /* fp_div_df */
168 COSTS_N_INSNS (7), /* int_mult_si */
169 COSTS_N_INSNS (7), /* int_mult_di */
170 COSTS_N_INSNS (1), /* int_shift */
171 COSTS_N_INSNS (2), /* int_cmov */
172 },
173 };
174
175 /* Get the number of args of a function in one of two ways. */
176 #if TARGET_ABI_OPEN_VMS || TARGET_ABI_UNICOSMK
177 #define NUM_ARGS current_function_args_info.num_args
178 #else
179 #define NUM_ARGS current_function_args_info
180 #endif
181
182 #define REG_PV 27
183 #define REG_RA 26
184
185 /* Declarations of static functions. */
186 static struct machine_function *alpha_init_machine_status (void);
187 static rtx alpha_emit_xfloating_compare (enum rtx_code, rtx, rtx);
188
189 #if TARGET_ABI_OPEN_VMS
190 static void alpha_write_linkage (FILE *, const char *, tree);
191 #endif
192
193 static void unicosmk_output_deferred_case_vectors (FILE *);
194 static void unicosmk_gen_dsib (unsigned long *);
195 static void unicosmk_output_ssib (FILE *, const char *);
196 static int unicosmk_need_dex (rtx);
197 \f
198 /* Parse target option strings. */
199
200 void
201 override_options (void)
202 {
203 int i;
204 static const struct cpu_table {
205 const char *const name;
206 const enum processor_type processor;
207 const int flags;
208 } cpu_table[] = {
209 #define EV5_MASK (MASK_CPU_EV5)
210 #define EV6_MASK (MASK_CPU_EV6|MASK_BWX|MASK_MAX|MASK_FIX)
211 { "ev4", PROCESSOR_EV4, 0 },
212 { "ev45", PROCESSOR_EV4, 0 },
213 { "21064", PROCESSOR_EV4, 0 },
214 { "ev5", PROCESSOR_EV5, EV5_MASK },
215 { "21164", PROCESSOR_EV5, EV5_MASK },
216 { "ev56", PROCESSOR_EV5, EV5_MASK|MASK_BWX },
217 { "21164a", PROCESSOR_EV5, EV5_MASK|MASK_BWX },
218 { "pca56", PROCESSOR_EV5, EV5_MASK|MASK_BWX|MASK_MAX },
219 { "21164PC",PROCESSOR_EV5, EV5_MASK|MASK_BWX|MASK_MAX },
220 { "21164pc",PROCESSOR_EV5, EV5_MASK|MASK_BWX|MASK_MAX },
221 { "ev6", PROCESSOR_EV6, EV6_MASK },
222 { "21264", PROCESSOR_EV6, EV6_MASK },
223 { "ev67", PROCESSOR_EV6, EV6_MASK|MASK_CIX },
224 { "21264a", PROCESSOR_EV6, EV6_MASK|MASK_CIX },
225 { 0, 0, 0 }
226 };
227
228 /* Unicos/Mk doesn't have shared libraries. */
229 if (TARGET_ABI_UNICOSMK && flag_pic)
230 {
231 warning ("-f%s ignored for Unicos/Mk (not supported)",
232 (flag_pic > 1) ? "PIC" : "pic");
233 flag_pic = 0;
234 }
235
236 /* On Unicos/Mk, the native compiler consistently generates /d suffices for
237 floating-point instructions. Make that the default for this target. */
238 if (TARGET_ABI_UNICOSMK)
239 alpha_fprm = ALPHA_FPRM_DYN;
240 else
241 alpha_fprm = ALPHA_FPRM_NORM;
242
243 alpha_tp = ALPHA_TP_PROG;
244 alpha_fptm = ALPHA_FPTM_N;
245
246 /* We cannot use su and sui qualifiers for conversion instructions on
247 Unicos/Mk. I'm not sure if this is due to assembler or hardware
248 limitations. Right now, we issue a warning if -mieee is specified
249 and then ignore it; eventually, we should either get it right or
250 disable the option altogether. */
251
252 if (TARGET_IEEE)
253 {
254 if (TARGET_ABI_UNICOSMK)
255 warning ("-mieee not supported on Unicos/Mk");
256 else
257 {
258 alpha_tp = ALPHA_TP_INSN;
259 alpha_fptm = ALPHA_FPTM_SU;
260 }
261 }
262
263 if (TARGET_IEEE_WITH_INEXACT)
264 {
265 if (TARGET_ABI_UNICOSMK)
266 warning ("-mieee-with-inexact not supported on Unicos/Mk");
267 else
268 {
269 alpha_tp = ALPHA_TP_INSN;
270 alpha_fptm = ALPHA_FPTM_SUI;
271 }
272 }
273
274 if (alpha_tp_string)
275 {
276 if (! strcmp (alpha_tp_string, "p"))
277 alpha_tp = ALPHA_TP_PROG;
278 else if (! strcmp (alpha_tp_string, "f"))
279 alpha_tp = ALPHA_TP_FUNC;
280 else if (! strcmp (alpha_tp_string, "i"))
281 alpha_tp = ALPHA_TP_INSN;
282 else
283 error ("bad value `%s' for -mtrap-precision switch", alpha_tp_string);
284 }
285
286 if (alpha_fprm_string)
287 {
288 if (! strcmp (alpha_fprm_string, "n"))
289 alpha_fprm = ALPHA_FPRM_NORM;
290 else if (! strcmp (alpha_fprm_string, "m"))
291 alpha_fprm = ALPHA_FPRM_MINF;
292 else if (! strcmp (alpha_fprm_string, "c"))
293 alpha_fprm = ALPHA_FPRM_CHOP;
294 else if (! strcmp (alpha_fprm_string,"d"))
295 alpha_fprm = ALPHA_FPRM_DYN;
296 else
297 error ("bad value `%s' for -mfp-rounding-mode switch",
298 alpha_fprm_string);
299 }
300
301 if (alpha_fptm_string)
302 {
303 if (strcmp (alpha_fptm_string, "n") == 0)
304 alpha_fptm = ALPHA_FPTM_N;
305 else if (strcmp (alpha_fptm_string, "u") == 0)
306 alpha_fptm = ALPHA_FPTM_U;
307 else if (strcmp (alpha_fptm_string, "su") == 0)
308 alpha_fptm = ALPHA_FPTM_SU;
309 else if (strcmp (alpha_fptm_string, "sui") == 0)
310 alpha_fptm = ALPHA_FPTM_SUI;
311 else
312 error ("bad value `%s' for -mfp-trap-mode switch", alpha_fptm_string);
313 }
314
315 if (alpha_tls_size_string)
316 {
317 if (strcmp (alpha_tls_size_string, "16") == 0)
318 alpha_tls_size = 16;
319 else if (strcmp (alpha_tls_size_string, "32") == 0)
320 alpha_tls_size = 32;
321 else if (strcmp (alpha_tls_size_string, "64") == 0)
322 alpha_tls_size = 64;
323 else
324 error ("bad value `%s' for -mtls-size switch", alpha_tls_size_string);
325 }
326
327 alpha_cpu
328 = TARGET_CPU_DEFAULT & MASK_CPU_EV6 ? PROCESSOR_EV6
329 : (TARGET_CPU_DEFAULT & MASK_CPU_EV5 ? PROCESSOR_EV5 : PROCESSOR_EV4);
330
331 if (alpha_cpu_string)
332 {
333 for (i = 0; cpu_table [i].name; i++)
334 if (! strcmp (alpha_cpu_string, cpu_table [i].name))
335 {
336 alpha_cpu = cpu_table [i].processor;
337 target_flags &= ~ (MASK_BWX | MASK_MAX | MASK_FIX | MASK_CIX
338 | MASK_CPU_EV5 | MASK_CPU_EV6);
339 target_flags |= cpu_table [i].flags;
340 break;
341 }
342 if (! cpu_table [i].name)
343 error ("bad value `%s' for -mcpu switch", alpha_cpu_string);
344 }
345
346 if (alpha_tune_string)
347 {
348 for (i = 0; cpu_table [i].name; i++)
349 if (! strcmp (alpha_tune_string, cpu_table [i].name))
350 {
351 alpha_cpu = cpu_table [i].processor;
352 break;
353 }
354 if (! cpu_table [i].name)
355 error ("bad value `%s' for -mcpu switch", alpha_tune_string);
356 }
357
358 /* Do some sanity checks on the above options. */
359
360 if (TARGET_ABI_UNICOSMK && alpha_fptm != ALPHA_FPTM_N)
361 {
362 warning ("trap mode not supported on Unicos/Mk");
363 alpha_fptm = ALPHA_FPTM_N;
364 }
365
366 if ((alpha_fptm == ALPHA_FPTM_SU || alpha_fptm == ALPHA_FPTM_SUI)
367 && alpha_tp != ALPHA_TP_INSN && ! TARGET_CPU_EV6)
368 {
369 warning ("fp software completion requires -mtrap-precision=i");
370 alpha_tp = ALPHA_TP_INSN;
371 }
372
373 if (TARGET_CPU_EV6)
374 {
375 /* Except for EV6 pass 1 (not released), we always have precise
376 arithmetic traps. Which means we can do software completion
377 without minding trap shadows. */
378 alpha_tp = ALPHA_TP_PROG;
379 }
380
381 if (TARGET_FLOAT_VAX)
382 {
383 if (alpha_fprm == ALPHA_FPRM_MINF || alpha_fprm == ALPHA_FPRM_DYN)
384 {
385 warning ("rounding mode not supported for VAX floats");
386 alpha_fprm = ALPHA_FPRM_NORM;
387 }
388 if (alpha_fptm == ALPHA_FPTM_SUI)
389 {
390 warning ("trap mode not supported for VAX floats");
391 alpha_fptm = ALPHA_FPTM_SU;
392 }
393 }
394
395 {
396 char *end;
397 int lat;
398
399 if (!alpha_mlat_string)
400 alpha_mlat_string = "L1";
401
402 if (ISDIGIT ((unsigned char)alpha_mlat_string[0])
403 && (lat = strtol (alpha_mlat_string, &end, 10), *end == '\0'))
404 ;
405 else if ((alpha_mlat_string[0] == 'L' || alpha_mlat_string[0] == 'l')
406 && ISDIGIT ((unsigned char)alpha_mlat_string[1])
407 && alpha_mlat_string[2] == '\0')
408 {
409 static int const cache_latency[][4] =
410 {
411 { 3, 30, -1 }, /* ev4 -- Bcache is a guess */
412 { 2, 12, 38 }, /* ev5 -- Bcache from PC164 LMbench numbers */
413 { 3, 12, 30 }, /* ev6 -- Bcache from DS20 LMbench. */
414 };
415
416 lat = alpha_mlat_string[1] - '0';
417 if (lat <= 0 || lat > 3 || cache_latency[alpha_cpu][lat-1] == -1)
418 {
419 warning ("L%d cache latency unknown for %s",
420 lat, alpha_cpu_name[alpha_cpu]);
421 lat = 3;
422 }
423 else
424 lat = cache_latency[alpha_cpu][lat-1];
425 }
426 else if (! strcmp (alpha_mlat_string, "main"))
427 {
428 /* Most current memories have about 370ns latency. This is
429 a reasonable guess for a fast cpu. */
430 lat = 150;
431 }
432 else
433 {
434 warning ("bad value `%s' for -mmemory-latency", alpha_mlat_string);
435 lat = 3;
436 }
437
438 alpha_memory_latency = lat;
439 }
440
441 /* Default the definition of "small data" to 8 bytes. */
442 if (!g_switch_set)
443 g_switch_value = 8;
444
445 /* Infer TARGET_SMALL_DATA from -fpic/-fPIC. */
446 if (flag_pic == 1)
447 target_flags |= MASK_SMALL_DATA;
448 else if (flag_pic == 2)
449 target_flags &= ~MASK_SMALL_DATA;
450
451 /* Align labels and loops for optimal branching. */
452 /* ??? Kludge these by not doing anything if we don't optimize and also if
453 we are writing ECOFF symbols to work around a bug in DEC's assembler. */
454 if (optimize > 0 && write_symbols != SDB_DEBUG)
455 {
456 if (align_loops <= 0)
457 align_loops = 16;
458 if (align_jumps <= 0)
459 align_jumps = 16;
460 }
461 if (align_functions <= 0)
462 align_functions = 16;
463
464 /* Acquire a unique set number for our register saves and restores. */
465 alpha_sr_alias_set = new_alias_set ();
466
467 /* Register variables and functions with the garbage collector. */
468
469 /* Set up function hooks. */
470 init_machine_status = alpha_init_machine_status;
471
472 /* Tell the compiler when we're using VAX floating point. */
473 if (TARGET_FLOAT_VAX)
474 {
475 REAL_MODE_FORMAT (SFmode) = &vax_f_format;
476 REAL_MODE_FORMAT (DFmode) = &vax_g_format;
477 REAL_MODE_FORMAT (TFmode) = NULL;
478 }
479 }
480 \f
481 /* Returns 1 if VALUE is a mask that contains full bytes of zero or ones. */
482
483 int
484 zap_mask (HOST_WIDE_INT value)
485 {
486 int i;
487
488 for (i = 0; i < HOST_BITS_PER_WIDE_INT / HOST_BITS_PER_CHAR;
489 i++, value >>= 8)
490 if ((value & 0xff) != 0 && (value & 0xff) != 0xff)
491 return 0;
492
493 return 1;
494 }
495
496 /* Returns 1 if OP is either the constant zero or a register. If a
497 register, it must be in the proper mode unless MODE is VOIDmode. */
498
499 int
500 reg_or_0_operand (rtx op, enum machine_mode mode)
501 {
502 return op == CONST0_RTX (mode) || register_operand (op, mode);
503 }
504
505 /* Return 1 if OP is a constant in the range of 0-63 (for a shift) or
506 any register. */
507
508 int
509 reg_or_6bit_operand (rtx op, enum machine_mode mode)
510 {
511 return ((GET_CODE (op) == CONST_INT
512 && (unsigned HOST_WIDE_INT) INTVAL (op) < 64)
513 || register_operand (op, mode));
514 }
515
516
517 /* Return 1 if OP is an 8-bit constant or any register. */
518
519 int
520 reg_or_8bit_operand (rtx op, enum machine_mode mode)
521 {
522 return ((GET_CODE (op) == CONST_INT
523 && (unsigned HOST_WIDE_INT) INTVAL (op) < 0x100)
524 || register_operand (op, mode));
525 }
526
527 /* Return 1 if OP is a constant or any register. */
528
529 int
530 reg_or_const_int_operand (rtx op, enum machine_mode mode)
531 {
532 return GET_CODE (op) == CONST_INT || register_operand (op, mode);
533 }
534
535 /* Return 1 if OP is an 8-bit constant. */
536
537 int
538 cint8_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
539 {
540 return ((GET_CODE (op) == CONST_INT
541 && (unsigned HOST_WIDE_INT) INTVAL (op) < 0x100));
542 }
543
544 /* Return 1 if the operand is a valid second operand to an add insn. */
545
546 int
547 add_operand (rtx op, enum machine_mode mode)
548 {
549 if (GET_CODE (op) == CONST_INT)
550 /* Constraints I, J, O and P are covered by K. */
551 return (CONST_OK_FOR_LETTER_P (INTVAL (op), 'K')
552 || CONST_OK_FOR_LETTER_P (INTVAL (op), 'L'));
553
554 return register_operand (op, mode);
555 }
556
557 /* Return 1 if the operand is a valid second operand to a sign-extending
558 add insn. */
559
560 int
561 sext_add_operand (rtx op, enum machine_mode mode)
562 {
563 if (GET_CODE (op) == CONST_INT)
564 return (CONST_OK_FOR_LETTER_P (INTVAL (op), 'I')
565 || CONST_OK_FOR_LETTER_P (INTVAL (op), 'O'));
566
567 return reg_not_elim_operand (op, mode);
568 }
569
570 /* Return 1 if OP is the constant 4 or 8. */
571
572 int
573 const48_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
574 {
575 return (GET_CODE (op) == CONST_INT
576 && (INTVAL (op) == 4 || INTVAL (op) == 8));
577 }
578
579 /* Return 1 if OP is a valid first operand to an AND insn. */
580
581 int
582 and_operand (rtx op, enum machine_mode mode)
583 {
584 if (GET_CODE (op) == CONST_DOUBLE && GET_MODE (op) == VOIDmode)
585 return (zap_mask (CONST_DOUBLE_LOW (op))
586 && zap_mask (CONST_DOUBLE_HIGH (op)));
587
588 if (GET_CODE (op) == CONST_INT)
589 return ((unsigned HOST_WIDE_INT) INTVAL (op) < 0x100
590 || (unsigned HOST_WIDE_INT) ~ INTVAL (op) < 0x100
591 || zap_mask (INTVAL (op)));
592
593 return register_operand (op, mode);
594 }
595
596 /* Return 1 if OP is a valid first operand to an IOR or XOR insn. */
597
598 int
599 or_operand (rtx op, enum machine_mode mode)
600 {
601 if (GET_CODE (op) == CONST_INT)
602 return ((unsigned HOST_WIDE_INT) INTVAL (op) < 0x100
603 || (unsigned HOST_WIDE_INT) ~ INTVAL (op) < 0x100);
604
605 return register_operand (op, mode);
606 }
607
608 /* Return 1 if OP is a constant that is the width, in bits, of an integral
609 mode smaller than DImode. */
610
611 int
612 mode_width_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
613 {
614 return (GET_CODE (op) == CONST_INT
615 && (INTVAL (op) == 8 || INTVAL (op) == 16
616 || INTVAL (op) == 32 || INTVAL (op) == 64));
617 }
618
619 /* Return 1 if OP is a constant that is the width of an integral machine mode
620 smaller than an integer. */
621
622 int
623 mode_mask_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
624 {
625 if (GET_CODE (op) == CONST_INT)
626 {
627 HOST_WIDE_INT value = INTVAL (op);
628
629 if (value == 0xff)
630 return 1;
631 if (value == 0xffff)
632 return 1;
633 if (value == 0xffffffff)
634 return 1;
635 if (value == -1)
636 return 1;
637 }
638 else if (HOST_BITS_PER_WIDE_INT == 32 && GET_CODE (op) == CONST_DOUBLE)
639 {
640 if (CONST_DOUBLE_LOW (op) == 0xffffffff && CONST_DOUBLE_HIGH (op) == 0)
641 return 1;
642 }
643
644 return 0;
645 }
646
647 /* Return 1 if OP is a multiple of 8 less than 64. */
648
649 int
650 mul8_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
651 {
652 return (GET_CODE (op) == CONST_INT
653 && (unsigned HOST_WIDE_INT) INTVAL (op) < 64
654 && (INTVAL (op) & 7) == 0);
655 }
656
657 /* Return 1 if OP is the zero constant for MODE. */
658
659 int
660 const0_operand (rtx op, enum machine_mode mode)
661 {
662 return op == CONST0_RTX (mode);
663 }
664
665 /* Return 1 if OP is a hard floating-point register. */
666
667 int
668 hard_fp_register_operand (rtx op, enum machine_mode mode)
669 {
670 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
671 return 0;
672
673 if (GET_CODE (op) == SUBREG)
674 op = SUBREG_REG (op);
675 return GET_CODE (op) == REG && REGNO_REG_CLASS (REGNO (op)) == FLOAT_REGS;
676 }
677
678 /* Return 1 if OP is a hard general register. */
679
680 int
681 hard_int_register_operand (rtx op, enum machine_mode mode)
682 {
683 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
684 return 0;
685
686 if (GET_CODE (op) == SUBREG)
687 op = SUBREG_REG (op);
688 return GET_CODE (op) == REG && REGNO_REG_CLASS (REGNO (op)) == GENERAL_REGS;
689 }
690
691 /* Return 1 if OP is a register or a constant integer. */
692
693
694 int
695 reg_or_cint_operand (rtx op, enum machine_mode mode)
696 {
697 return (GET_CODE (op) == CONST_INT
698 || register_operand (op, mode));
699 }
700
701 /* Return 1 if OP is something that can be reloaded into a register;
702 if it is a MEM, it need not be valid. */
703
704 int
705 some_operand (rtx op, enum machine_mode mode)
706 {
707 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
708 return 0;
709
710 switch (GET_CODE (op))
711 {
712 case REG:
713 case MEM:
714 case CONST_INT:
715 case CONST_DOUBLE:
716 case CONST_VECTOR:
717 case LABEL_REF:
718 case SYMBOL_REF:
719 case CONST:
720 case HIGH:
721 return 1;
722
723 case SUBREG:
724 return some_operand (SUBREG_REG (op), VOIDmode);
725
726 default:
727 break;
728 }
729
730 return 0;
731 }
732
733 /* Likewise, but don't accept constants. */
734
735 int
736 some_ni_operand (rtx op, enum machine_mode mode)
737 {
738 if (GET_MODE (op) != mode && mode != VOIDmode)
739 return 0;
740
741 if (GET_CODE (op) == SUBREG)
742 op = SUBREG_REG (op);
743
744 return (GET_CODE (op) == REG || GET_CODE (op) == MEM);
745 }
746
747 /* Return 1 if OP is a valid operand for the source of a move insn. */
748
749 int
750 input_operand (rtx op, enum machine_mode mode)
751 {
752 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
753 return 0;
754
755 if (GET_MODE_CLASS (mode) == MODE_FLOAT && GET_MODE (op) != mode)
756 return 0;
757
758 switch (GET_CODE (op))
759 {
760 case LABEL_REF:
761 case SYMBOL_REF:
762 case CONST:
763 if (TARGET_EXPLICIT_RELOCS)
764 {
765 /* We don't split symbolic operands into something unintelligable
766 until after reload, but we do not wish non-small, non-global
767 symbolic operands to be reconstructed from their high/lo_sum
768 form. */
769 return (small_symbolic_operand (op, mode)
770 || global_symbolic_operand (op, mode)
771 || gotdtp_symbolic_operand (op, mode)
772 || gottp_symbolic_operand (op, mode));
773 }
774
775 /* This handles both the Windows/NT and OSF cases. */
776 return mode == ptr_mode || mode == DImode;
777
778 case HIGH:
779 return (TARGET_EXPLICIT_RELOCS
780 && local_symbolic_operand (XEXP (op, 0), mode));
781
782 case REG:
783 case ADDRESSOF:
784 return 1;
785
786 case SUBREG:
787 if (register_operand (op, mode))
788 return 1;
789 /* ... fall through ... */
790 case MEM:
791 return ((TARGET_BWX || (mode != HImode && mode != QImode))
792 && general_operand (op, mode));
793
794 case CONST_DOUBLE:
795 case CONST_VECTOR:
796 return op == CONST0_RTX (mode);
797
798 case CONST_INT:
799 return mode == QImode || mode == HImode || add_operand (op, mode);
800
801 case CONSTANT_P_RTX:
802 return 1;
803
804 default:
805 break;
806 }
807
808 return 0;
809 }
810
811 /* Return 1 if OP is a SYMBOL_REF for a function known to be in this
812 file, and in the same section as the current function. */
813
814 int
815 samegp_function_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
816 {
817 if (GET_CODE (op) != SYMBOL_REF)
818 return false;
819
820 /* Easy test for recursion. */
821 if (op == XEXP (DECL_RTL (current_function_decl), 0))
822 return true;
823
824 /* Functions that are not local can be overridden, and thus may
825 not share the same gp. */
826 if (! SYMBOL_REF_LOCAL_P (op))
827 return false;
828
829 /* If -msmall-data is in effect, assume that there is only one GP
830 for the module, and so any local symbol has this property. We
831 need explicit relocations to be able to enforce this for symbols
832 not defined in this unit of translation, however. */
833 if (TARGET_EXPLICIT_RELOCS && TARGET_SMALL_DATA)
834 return true;
835
836 /* Functions that are not external are defined in this UoT,
837 and thus must share the same gp. */
838 return ! SYMBOL_REF_EXTERNAL_P (op);
839 }
840
841 /* Return 1 if OP is a SYMBOL_REF for which we can make a call via bsr. */
842
843 int
844 direct_call_operand (rtx op, enum machine_mode mode)
845 {
846 tree op_decl, cfun_sec, op_sec;
847
848 /* Must share the same GP. */
849 if (!samegp_function_operand (op, mode))
850 return false;
851
852 /* If profiling is implemented via linker tricks, we can't jump
853 to the nogp alternate entry point. Note that current_function_profile
854 would not be correct, since that doesn't indicate if the target
855 function uses profiling. */
856 /* ??? TARGET_PROFILING_NEEDS_GP isn't really the right test,
857 but is approximately correct for the OSF ABIs. Don't know
858 what to do for VMS, NT, or UMK. */
859 if (!TARGET_PROFILING_NEEDS_GP && profile_flag)
860 return false;
861
862 /* Must be a function. In some cases folks create thunks in static
863 data structures and then make calls to them. If we allow the
864 direct call, we'll get an error from the linker about !samegp reloc
865 against a symbol without a .prologue directive. */
866 if (!SYMBOL_REF_FUNCTION_P (op))
867 return false;
868
869 /* Must be "near" so that the branch is assumed to reach. With
870 -msmall-text, this is assumed true of all local symbols. Since
871 we've already checked samegp, locality is already assured. */
872 if (TARGET_SMALL_TEXT)
873 return true;
874
875 /* Otherwise, a decl is "near" if it is defined in the same section. */
876 if (flag_function_sections)
877 return false;
878
879 op_decl = SYMBOL_REF_DECL (op);
880 if (DECL_ONE_ONLY (current_function_decl)
881 || (op_decl && DECL_ONE_ONLY (op_decl)))
882 return false;
883
884 cfun_sec = DECL_SECTION_NAME (current_function_decl);
885 op_sec = op_decl ? DECL_SECTION_NAME (op_decl) : NULL;
886 return ((!cfun_sec && !op_sec)
887 || (cfun_sec && op_sec
888 && strcmp (TREE_STRING_POINTER (cfun_sec),
889 TREE_STRING_POINTER (op_sec)) == 0));
890 }
891
892 /* Return true if OP is a LABEL_REF, or SYMBOL_REF or CONST referencing
893 a (non-tls) variable known to be defined in this file. */
894
895 int
896 local_symbolic_operand (rtx op, enum machine_mode mode)
897 {
898 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
899 return 0;
900
901 if (GET_CODE (op) == LABEL_REF)
902 return 1;
903
904 if (GET_CODE (op) == CONST
905 && GET_CODE (XEXP (op, 0)) == PLUS
906 && GET_CODE (XEXP (XEXP (op, 0), 1)) == CONST_INT)
907 op = XEXP (XEXP (op, 0), 0);
908
909 if (GET_CODE (op) != SYMBOL_REF)
910 return 0;
911
912 return SYMBOL_REF_LOCAL_P (op) && !SYMBOL_REF_TLS_MODEL (op);
913 }
914
915 /* Return true if OP is a SYMBOL_REF or CONST referencing a variable
916 known to be defined in this file in the small data area. */
917
918 int
919 small_symbolic_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
920 {
921 if (! TARGET_SMALL_DATA)
922 return 0;
923
924 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
925 return 0;
926
927 if (GET_CODE (op) == CONST
928 && GET_CODE (XEXP (op, 0)) == PLUS
929 && GET_CODE (XEXP (XEXP (op, 0), 1)) == CONST_INT)
930 op = XEXP (XEXP (op, 0), 0);
931
932 if (GET_CODE (op) != SYMBOL_REF)
933 return 0;
934
935 /* ??? There's no encode_section_info equivalent for the rtl
936 constant pool, so SYMBOL_FLAG_SMALL never gets set. */
937 if (CONSTANT_POOL_ADDRESS_P (op))
938 return GET_MODE_SIZE (get_pool_mode (op)) <= g_switch_value;
939
940 return (SYMBOL_REF_LOCAL_P (op)
941 && SYMBOL_REF_SMALL_P (op)
942 && SYMBOL_REF_TLS_MODEL (op) == 0);
943 }
944
945 /* Return true if OP is a SYMBOL_REF or CONST referencing a variable
946 not known (or known not) to be defined in this file. */
947
948 int
949 global_symbolic_operand (rtx op, enum machine_mode mode)
950 {
951 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
952 return 0;
953
954 if (GET_CODE (op) == CONST
955 && GET_CODE (XEXP (op, 0)) == PLUS
956 && GET_CODE (XEXP (XEXP (op, 0), 1)) == CONST_INT)
957 op = XEXP (XEXP (op, 0), 0);
958
959 if (GET_CODE (op) != SYMBOL_REF)
960 return 0;
961
962 return !SYMBOL_REF_LOCAL_P (op) && !SYMBOL_REF_TLS_MODEL (op);
963 }
964
965 /* Return 1 if OP is a valid operand for the MEM of a CALL insn. */
966
967 int
968 call_operand (rtx op, enum machine_mode mode)
969 {
970 if (mode != Pmode)
971 return 0;
972
973 if (GET_CODE (op) == REG)
974 {
975 if (TARGET_ABI_OSF)
976 {
977 /* Disallow virtual registers to cope with pathological test cases
978 such as compile/930117-1.c in which the virtual reg decomposes
979 to the frame pointer. Which is a hard reg that is not $27. */
980 return (REGNO (op) == 27 || REGNO (op) > LAST_VIRTUAL_REGISTER);
981 }
982 else
983 return 1;
984 }
985 if (TARGET_ABI_UNICOSMK)
986 return 0;
987 if (GET_CODE (op) == SYMBOL_REF)
988 return 1;
989
990 return 0;
991 }
992
993 /* Returns 1 if OP is a symbolic operand, i.e. a symbol_ref or a label_ref,
994 possibly with an offset. */
995
996 int
997 symbolic_operand (rtx op, enum machine_mode mode)
998 {
999 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
1000 return 0;
1001 if (GET_CODE (op) == SYMBOL_REF || GET_CODE (op) == LABEL_REF)
1002 return 1;
1003 if (GET_CODE (op) == CONST
1004 && GET_CODE (XEXP (op,0)) == PLUS
1005 && GET_CODE (XEXP (XEXP (op,0), 0)) == SYMBOL_REF
1006 && GET_CODE (XEXP (XEXP (op,0), 1)) == CONST_INT)
1007 return 1;
1008 return 0;
1009 }
1010
1011 /* Return true if OP is valid for a particular TLS relocation. */
1012
1013 static int
1014 tls_symbolic_operand_1 (rtx op, enum machine_mode mode, int size, int unspec)
1015 {
1016 if (mode != VOIDmode && GET_MODE (op) != VOIDmode && mode != GET_MODE (op))
1017 return 0;
1018
1019 if (GET_CODE (op) != CONST)
1020 return 0;
1021 op = XEXP (op, 0);
1022
1023 if (GET_CODE (op) != UNSPEC || XINT (op, 1) != unspec)
1024 return 0;
1025 op = XVECEXP (op, 0, 0);
1026
1027 if (GET_CODE (op) != SYMBOL_REF)
1028 return 0;
1029
1030 if (SYMBOL_REF_LOCAL_P (op))
1031 {
1032 if (alpha_tls_size > size)
1033 return 0;
1034 }
1035 else
1036 {
1037 if (size != 64)
1038 return 0;
1039 }
1040
1041 switch (SYMBOL_REF_TLS_MODEL (op))
1042 {
1043 case TLS_MODEL_LOCAL_DYNAMIC:
1044 return unspec == UNSPEC_DTPREL;
1045 case TLS_MODEL_INITIAL_EXEC:
1046 return unspec == UNSPEC_TPREL && size == 64;
1047 case TLS_MODEL_LOCAL_EXEC:
1048 return unspec == UNSPEC_TPREL;
1049 default:
1050 abort ();
1051 }
1052 }
1053
1054 /* Return true if OP is valid for 16-bit DTP relative relocations. */
1055
1056 int
1057 dtp16_symbolic_operand (rtx op, enum machine_mode mode)
1058 {
1059 return tls_symbolic_operand_1 (op, mode, 16, UNSPEC_DTPREL);
1060 }
1061
1062 /* Return true if OP is valid for 32-bit DTP relative relocations. */
1063
1064 int
1065 dtp32_symbolic_operand (rtx op, enum machine_mode mode)
1066 {
1067 return tls_symbolic_operand_1 (op, mode, 32, UNSPEC_DTPREL);
1068 }
1069
1070 /* Return true if OP is valid for 64-bit DTP relative relocations. */
1071
1072 int
1073 gotdtp_symbolic_operand (rtx op, enum machine_mode mode)
1074 {
1075 return tls_symbolic_operand_1 (op, mode, 64, UNSPEC_DTPREL);
1076 }
1077
1078 /* Return true if OP is valid for 16-bit TP relative relocations. */
1079
1080 int
1081 tp16_symbolic_operand (rtx op, enum machine_mode mode)
1082 {
1083 return tls_symbolic_operand_1 (op, mode, 16, UNSPEC_TPREL);
1084 }
1085
1086 /* Return true if OP is valid for 32-bit TP relative relocations. */
1087
1088 int
1089 tp32_symbolic_operand (rtx op, enum machine_mode mode)
1090 {
1091 return tls_symbolic_operand_1 (op, mode, 32, UNSPEC_TPREL);
1092 }
1093
1094 /* Return true if OP is valid for 64-bit TP relative relocations. */
1095
1096 int
1097 gottp_symbolic_operand (rtx op, enum machine_mode mode)
1098 {
1099 return tls_symbolic_operand_1 (op, mode, 64, UNSPEC_TPREL);
1100 }
1101
1102 /* Return 1 if OP is a valid Alpha comparison operator. Here we know which
1103 comparisons are valid in which insn. */
1104
1105 int
1106 alpha_comparison_operator (rtx op, enum machine_mode mode)
1107 {
1108 enum rtx_code code = GET_CODE (op);
1109
1110 if (mode != GET_MODE (op) && mode != VOIDmode)
1111 return 0;
1112
1113 return (code == EQ || code == LE || code == LT
1114 || code == LEU || code == LTU);
1115 }
1116
1117 /* Return 1 if OP is a valid Alpha comparison operator against zero.
1118 Here we know which comparisons are valid in which insn. */
1119
1120 int
1121 alpha_zero_comparison_operator (rtx op, enum machine_mode mode)
1122 {
1123 enum rtx_code code = GET_CODE (op);
1124
1125 if (mode != GET_MODE (op) && mode != VOIDmode)
1126 return 0;
1127
1128 return (code == EQ || code == NE || code == LE || code == LT
1129 || code == LEU || code == LTU);
1130 }
1131
1132 /* Return 1 if OP is a valid Alpha swapped comparison operator. */
1133
1134 int
1135 alpha_swapped_comparison_operator (rtx op, enum machine_mode mode)
1136 {
1137 enum rtx_code code = GET_CODE (op);
1138
1139 if ((mode != GET_MODE (op) && mode != VOIDmode)
1140 || GET_RTX_CLASS (code) != '<')
1141 return 0;
1142
1143 code = swap_condition (code);
1144 return (code == EQ || code == LE || code == LT
1145 || code == LEU || code == LTU);
1146 }
1147
1148 /* Return 1 if OP is a signed comparison operation. */
1149
1150 int
1151 signed_comparison_operator (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
1152 {
1153 enum rtx_code code = GET_CODE (op);
1154
1155 if (mode != GET_MODE (op) && mode != VOIDmode)
1156 return 0;
1157
1158 return (code == EQ || code == NE
1159 || code == LE || code == LT
1160 || code == GE || code == GT);
1161 }
1162
1163 /* Return 1 if OP is a valid Alpha floating point comparison operator.
1164 Here we know which comparisons are valid in which insn. */
1165
1166 int
1167 alpha_fp_comparison_operator (rtx op, enum machine_mode mode)
1168 {
1169 enum rtx_code code = GET_CODE (op);
1170
1171 if (mode != GET_MODE (op) && mode != VOIDmode)
1172 return 0;
1173
1174 return (code == EQ || code == LE || code == LT || code == UNORDERED);
1175 }
1176
1177 /* Return 1 if this is a divide or modulus operator. */
1178
1179 int
1180 divmod_operator (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
1181 {
1182 enum rtx_code code = GET_CODE (op);
1183
1184 return (code == DIV || code == MOD || code == UDIV || code == UMOD);
1185 }
1186
1187 /* Return 1 if this is a float->int conversion operator. */
1188
1189 int
1190 fix_operator (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
1191 {
1192 enum rtx_code code = GET_CODE (op);
1193
1194 return (code == FIX || code == UNSIGNED_FIX);
1195 }
1196
1197 /* Return 1 if this memory address is a known aligned register plus
1198 a constant. It must be a valid address. This means that we can do
1199 this as an aligned reference plus some offset.
1200
1201 Take into account what reload will do. */
1202
1203 int
1204 aligned_memory_operand (rtx op, enum machine_mode mode)
1205 {
1206 rtx base;
1207
1208 if (reload_in_progress)
1209 {
1210 rtx tmp = op;
1211 if (GET_CODE (tmp) == SUBREG)
1212 tmp = SUBREG_REG (tmp);
1213 if (GET_CODE (tmp) == REG
1214 && REGNO (tmp) >= FIRST_PSEUDO_REGISTER)
1215 {
1216 op = reg_equiv_memory_loc[REGNO (tmp)];
1217 if (op == 0)
1218 return 0;
1219 }
1220 }
1221
1222 if (GET_CODE (op) != MEM)
1223 return 0;
1224 if (MEM_ALIGN (op) >= 32)
1225 return 1;
1226 op = XEXP (op, 0);
1227
1228 /* LEGITIMIZE_RELOAD_ADDRESS creates (plus (plus reg const_hi) const_lo)
1229 sorts of constructs. Dig for the real base register. */
1230 if (reload_in_progress
1231 && GET_CODE (op) == PLUS
1232 && GET_CODE (XEXP (op, 0)) == PLUS)
1233 base = XEXP (XEXP (op, 0), 0);
1234 else
1235 {
1236 if (! memory_address_p (mode, op))
1237 return 0;
1238 base = (GET_CODE (op) == PLUS ? XEXP (op, 0) : op);
1239 }
1240
1241 return (GET_CODE (base) == REG && REGNO_POINTER_ALIGN (REGNO (base)) >= 32);
1242 }
1243
1244 /* Similar, but return 1 if OP is a MEM which is not alignable. */
1245
1246 int
1247 unaligned_memory_operand (rtx op, enum machine_mode mode)
1248 {
1249 rtx base;
1250
1251 if (reload_in_progress)
1252 {
1253 rtx tmp = op;
1254 if (GET_CODE (tmp) == SUBREG)
1255 tmp = SUBREG_REG (tmp);
1256 if (GET_CODE (tmp) == REG
1257 && REGNO (tmp) >= FIRST_PSEUDO_REGISTER)
1258 {
1259 op = reg_equiv_memory_loc[REGNO (tmp)];
1260 if (op == 0)
1261 return 0;
1262 }
1263 }
1264
1265 if (GET_CODE (op) != MEM)
1266 return 0;
1267 if (MEM_ALIGN (op) >= 32)
1268 return 0;
1269 op = XEXP (op, 0);
1270
1271 /* LEGITIMIZE_RELOAD_ADDRESS creates (plus (plus reg const_hi) const_lo)
1272 sorts of constructs. Dig for the real base register. */
1273 if (reload_in_progress
1274 && GET_CODE (op) == PLUS
1275 && GET_CODE (XEXP (op, 0)) == PLUS)
1276 base = XEXP (XEXP (op, 0), 0);
1277 else
1278 {
1279 if (! memory_address_p (mode, op))
1280 return 0;
1281 base = (GET_CODE (op) == PLUS ? XEXP (op, 0) : op);
1282 }
1283
1284 return (GET_CODE (base) == REG && REGNO_POINTER_ALIGN (REGNO (base)) < 32);
1285 }
1286
1287 /* Return 1 if OP is either a register or an unaligned memory location. */
1288
1289 int
1290 reg_or_unaligned_mem_operand (rtx op, enum machine_mode mode)
1291 {
1292 return register_operand (op, mode) || unaligned_memory_operand (op, mode);
1293 }
1294
1295 /* Return 1 if OP is any memory location. During reload a pseudo matches. */
1296
1297 int
1298 any_memory_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
1299 {
1300 return (GET_CODE (op) == MEM
1301 || (GET_CODE (op) == SUBREG && GET_CODE (SUBREG_REG (op)) == REG)
1302 || (reload_in_progress && GET_CODE (op) == REG
1303 && REGNO (op) >= FIRST_PSEUDO_REGISTER)
1304 || (reload_in_progress && GET_CODE (op) == SUBREG
1305 && GET_CODE (SUBREG_REG (op)) == REG
1306 && REGNO (SUBREG_REG (op)) >= FIRST_PSEUDO_REGISTER));
1307 }
1308
1309 /* Returns 1 if OP is not an eliminable register.
1310
1311 This exists to cure a pathological abort in the s8addq (et al) patterns,
1312
1313 long foo () { long t; bar(); return (long) &t * 26107; }
1314
1315 which run afoul of a hack in reload to cure a (presumably) similar
1316 problem with lea-type instructions on other targets. But there is
1317 one of us and many of them, so work around the problem by selectively
1318 preventing combine from making the optimization. */
1319
1320 int
1321 reg_not_elim_operand (rtx op, enum machine_mode mode)
1322 {
1323 rtx inner = op;
1324 if (GET_CODE (op) == SUBREG)
1325 inner = SUBREG_REG (op);
1326 if (inner == frame_pointer_rtx || inner == arg_pointer_rtx)
1327 return 0;
1328
1329 return register_operand (op, mode);
1330 }
1331
1332 /* Return 1 is OP is a memory location that is not a reference (using
1333 an AND) to an unaligned location. Take into account what reload
1334 will do. */
1335
1336 int
1337 normal_memory_operand (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
1338 {
1339 if (reload_in_progress)
1340 {
1341 rtx tmp = op;
1342 if (GET_CODE (tmp) == SUBREG)
1343 tmp = SUBREG_REG (tmp);
1344 if (GET_CODE (tmp) == REG
1345 && REGNO (tmp) >= FIRST_PSEUDO_REGISTER)
1346 {
1347 op = reg_equiv_memory_loc[REGNO (tmp)];
1348
1349 /* This may not have been assigned an equivalent address if it will
1350 be eliminated. In that case, it doesn't matter what we do. */
1351 if (op == 0)
1352 return 1;
1353 }
1354 }
1355
1356 return GET_CODE (op) == MEM && GET_CODE (XEXP (op, 0)) != AND;
1357 }
1358
1359 /* Accept a register, but not a subreg of any kind. This allows us to
1360 avoid pathological cases in reload wrt data movement common in
1361 int->fp conversion. */
1362
1363 int
1364 reg_no_subreg_operand (rtx op, enum machine_mode mode)
1365 {
1366 if (GET_CODE (op) != REG)
1367 return 0;
1368 return register_operand (op, mode);
1369 }
1370
1371 /* Recognize an addition operation that includes a constant. Used to
1372 convince reload to canonize (plus (plus reg c1) c2) during register
1373 elimination. */
1374
1375 int
1376 addition_operation (rtx op, enum machine_mode mode)
1377 {
1378 if (GET_MODE (op) != mode && mode != VOIDmode)
1379 return 0;
1380 if (GET_CODE (op) == PLUS
1381 && register_operand (XEXP (op, 0), mode)
1382 && GET_CODE (XEXP (op, 1)) == CONST_INT
1383 && CONST_OK_FOR_LETTER_P (INTVAL (XEXP (op, 1)), 'K'))
1384 return 1;
1385 return 0;
1386 }
1387
1388 /* Implements CONST_OK_FOR_LETTER_P. Return true if the value matches
1389 the range defined for C in [I-P]. */
1390
1391 bool
1392 alpha_const_ok_for_letter_p (HOST_WIDE_INT value, int c)
1393 {
1394 switch (c)
1395 {
1396 case 'I':
1397 /* An unsigned 8 bit constant. */
1398 return (unsigned HOST_WIDE_INT) value < 0x100;
1399 case 'J':
1400 /* The constant zero. */
1401 return value == 0;
1402 case 'K':
1403 /* A signed 16 bit constant. */
1404 return (unsigned HOST_WIDE_INT) (value + 0x8000) < 0x10000;
1405 case 'L':
1406 /* A shifted signed 16 bit constant appropriate for LDAH. */
1407 return ((value & 0xffff) == 0
1408 && ((value) >> 31 == -1 || value >> 31 == 0));
1409 case 'M':
1410 /* A constant that can be AND'ed with using a ZAP insn. */
1411 return zap_mask (value);
1412 case 'N':
1413 /* A complemented unsigned 8 bit constant. */
1414 return (unsigned HOST_WIDE_INT) (~ value) < 0x100;
1415 case 'O':
1416 /* A negated unsigned 8 bit constant. */
1417 return (unsigned HOST_WIDE_INT) (- value) < 0x100;
1418 case 'P':
1419 /* The constant 1, 2 or 3. */
1420 return value == 1 || value == 2 || value == 3;
1421
1422 default:
1423 return false;
1424 }
1425 }
1426
1427 /* Implements CONST_DOUBLE_OK_FOR_LETTER_P. Return true if VALUE
1428 matches for C in [GH]. */
1429
1430 bool
1431 alpha_const_double_ok_for_letter_p (rtx value, int c)
1432 {
1433 switch (c)
1434 {
1435 case 'G':
1436 /* The floating point zero constant. */
1437 return (GET_MODE_CLASS (GET_MODE (value)) == MODE_FLOAT
1438 && value == CONST0_RTX (GET_MODE (value)));
1439
1440 case 'H':
1441 /* A valid operand of a ZAP insn. */
1442 return (GET_MODE (value) == VOIDmode
1443 && zap_mask (CONST_DOUBLE_LOW (value))
1444 && zap_mask (CONST_DOUBLE_HIGH (value)));
1445
1446 default:
1447 return false;
1448 }
1449 }
1450
1451 /* Implements CONST_DOUBLE_OK_FOR_LETTER_P. Return true if VALUE
1452 matches for C. */
1453
1454 bool
1455 alpha_extra_constraint (rtx value, int c)
1456 {
1457 switch (c)
1458 {
1459 case 'Q':
1460 return normal_memory_operand (value, VOIDmode);
1461 case 'R':
1462 return direct_call_operand (value, Pmode);
1463 case 'S':
1464 return (GET_CODE (value) == CONST_INT
1465 && (unsigned HOST_WIDE_INT) INTVAL (value) < 64);
1466 case 'T':
1467 return GET_CODE (value) == HIGH;
1468 case 'U':
1469 return TARGET_ABI_UNICOSMK && symbolic_operand (value, VOIDmode);
1470 case 'W':
1471 return (GET_CODE (value) == CONST_VECTOR
1472 && value == CONST0_RTX (GET_MODE (value)));
1473 default:
1474 return false;
1475 }
1476 }
1477
1478 /* Return 1 if this function can directly return via $26. */
1479
1480 int
1481 direct_return (void)
1482 {
1483 return (! TARGET_ABI_OPEN_VMS && ! TARGET_ABI_UNICOSMK
1484 && reload_completed
1485 && alpha_sa_size () == 0
1486 && get_frame_size () == 0
1487 && current_function_outgoing_args_size == 0
1488 && current_function_pretend_args_size == 0);
1489 }
1490
1491 /* Return the ADDR_VEC associated with a tablejump insn. */
1492
1493 rtx
1494 alpha_tablejump_addr_vec (rtx insn)
1495 {
1496 rtx tmp;
1497
1498 tmp = JUMP_LABEL (insn);
1499 if (!tmp)
1500 return NULL_RTX;
1501 tmp = NEXT_INSN (tmp);
1502 if (!tmp)
1503 return NULL_RTX;
1504 if (GET_CODE (tmp) == JUMP_INSN
1505 && GET_CODE (PATTERN (tmp)) == ADDR_DIFF_VEC)
1506 return PATTERN (tmp);
1507 return NULL_RTX;
1508 }
1509
1510 /* Return the label of the predicted edge, or CONST0_RTX if we don't know. */
1511
1512 rtx
1513 alpha_tablejump_best_label (rtx insn)
1514 {
1515 rtx jump_table = alpha_tablejump_addr_vec (insn);
1516 rtx best_label = NULL_RTX;
1517
1518 /* ??? Once the CFG doesn't keep getting completely rebuilt, look
1519 there for edge frequency counts from profile data. */
1520
1521 if (jump_table)
1522 {
1523 int n_labels = XVECLEN (jump_table, 1);
1524 int best_count = -1;
1525 int i, j;
1526
1527 for (i = 0; i < n_labels; i++)
1528 {
1529 int count = 1;
1530
1531 for (j = i + 1; j < n_labels; j++)
1532 if (XEXP (XVECEXP (jump_table, 1, i), 0)
1533 == XEXP (XVECEXP (jump_table, 1, j), 0))
1534 count++;
1535
1536 if (count > best_count)
1537 best_count = count, best_label = XVECEXP (jump_table, 1, i);
1538 }
1539 }
1540
1541 return best_label ? best_label : const0_rtx;
1542 }
1543
1544 /* Return the TLS model to use for SYMBOL. */
1545
1546 static enum tls_model
1547 tls_symbolic_operand_type (rtx symbol)
1548 {
1549 enum tls_model model;
1550
1551 if (GET_CODE (symbol) != SYMBOL_REF)
1552 return 0;
1553 model = SYMBOL_REF_TLS_MODEL (symbol);
1554
1555 /* Local-exec with a 64-bit size is the same code as initial-exec. */
1556 if (model == TLS_MODEL_LOCAL_EXEC && alpha_tls_size == 64)
1557 model = TLS_MODEL_INITIAL_EXEC;
1558
1559 return model;
1560 }
1561 \f
1562 /* Return true if the function DECL will share the same GP as any
1563 function in the current unit of translation. */
1564
1565 static bool
1566 decl_has_samegp (tree decl)
1567 {
1568 /* Functions that are not local can be overridden, and thus may
1569 not share the same gp. */
1570 if (!(*targetm.binds_local_p) (decl))
1571 return false;
1572
1573 /* If -msmall-data is in effect, assume that there is only one GP
1574 for the module, and so any local symbol has this property. We
1575 need explicit relocations to be able to enforce this for symbols
1576 not defined in this unit of translation, however. */
1577 if (TARGET_EXPLICIT_RELOCS && TARGET_SMALL_DATA)
1578 return true;
1579
1580 /* Functions that are not external are defined in this UoT. */
1581 /* ??? Irritatingly, static functions not yet emitted are still
1582 marked "external". Apply this to non-static functions only. */
1583 return !TREE_PUBLIC (decl) || !DECL_EXTERNAL (decl);
1584 }
1585
1586 /* Return true if EXP should be placed in the small data section. */
1587
1588 static bool
1589 alpha_in_small_data_p (tree exp)
1590 {
1591 /* We want to merge strings, so we never consider them small data. */
1592 if (TREE_CODE (exp) == STRING_CST)
1593 return false;
1594
1595 if (TREE_CODE (exp) == VAR_DECL && DECL_SECTION_NAME (exp))
1596 {
1597 const char *section = TREE_STRING_POINTER (DECL_SECTION_NAME (exp));
1598 if (strcmp (section, ".sdata") == 0
1599 || strcmp (section, ".sbss") == 0)
1600 return true;
1601 }
1602 else
1603 {
1604 HOST_WIDE_INT size = int_size_in_bytes (TREE_TYPE (exp));
1605
1606 /* If this is an incomplete type with size 0, then we can't put it
1607 in sdata because it might be too big when completed. */
1608 if (size > 0 && (unsigned HOST_WIDE_INT) size <= g_switch_value)
1609 return true;
1610 }
1611
1612 return false;
1613 }
1614
1615 #if TARGET_ABI_OPEN_VMS
1616 static bool
1617 alpha_linkage_symbol_p (const char *symname)
1618 {
1619 int symlen = strlen (symname);
1620
1621 if (symlen > 4)
1622 return strcmp (&symname [symlen - 4], "..lk") == 0;
1623
1624 return false;
1625 }
1626
1627 #define LINKAGE_SYMBOL_REF_P(X) \
1628 ((GET_CODE (X) == SYMBOL_REF \
1629 && alpha_linkage_symbol_p (XSTR (X, 0))) \
1630 || (GET_CODE (X) == CONST \
1631 && GET_CODE (XEXP (X, 0)) == PLUS \
1632 && GET_CODE (XEXP (XEXP (X, 0), 0)) == SYMBOL_REF \
1633 && alpha_linkage_symbol_p (XSTR (XEXP (XEXP (X, 0), 0), 0))))
1634 #endif
1635
1636 /* legitimate_address_p recognizes an RTL expression that is a valid
1637 memory address for an instruction. The MODE argument is the
1638 machine mode for the MEM expression that wants to use this address.
1639
1640 For Alpha, we have either a constant address or the sum of a
1641 register and a constant address, or just a register. For DImode,
1642 any of those forms can be surrounded with an AND that clear the
1643 low-order three bits; this is an "unaligned" access. */
1644
1645 bool
1646 alpha_legitimate_address_p (enum machine_mode mode, rtx x, int strict)
1647 {
1648 /* If this is an ldq_u type address, discard the outer AND. */
1649 if (mode == DImode
1650 && GET_CODE (x) == AND
1651 && GET_CODE (XEXP (x, 1)) == CONST_INT
1652 && INTVAL (XEXP (x, 1)) == -8)
1653 x = XEXP (x, 0);
1654
1655 /* Discard non-paradoxical subregs. */
1656 if (GET_CODE (x) == SUBREG
1657 && (GET_MODE_SIZE (GET_MODE (x))
1658 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
1659 x = SUBREG_REG (x);
1660
1661 /* Unadorned general registers are valid. */
1662 if (REG_P (x)
1663 && (strict
1664 ? STRICT_REG_OK_FOR_BASE_P (x)
1665 : NONSTRICT_REG_OK_FOR_BASE_P (x)))
1666 return true;
1667
1668 /* Constant addresses (i.e. +/- 32k) are valid. */
1669 if (CONSTANT_ADDRESS_P (x))
1670 return true;
1671
1672 #if TARGET_ABI_OPEN_VMS
1673 if (LINKAGE_SYMBOL_REF_P (x))
1674 return true;
1675 #endif
1676
1677 /* Register plus a small constant offset is valid. */
1678 if (GET_CODE (x) == PLUS)
1679 {
1680 rtx ofs = XEXP (x, 1);
1681 x = XEXP (x, 0);
1682
1683 /* Discard non-paradoxical subregs. */
1684 if (GET_CODE (x) == SUBREG
1685 && (GET_MODE_SIZE (GET_MODE (x))
1686 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
1687 x = SUBREG_REG (x);
1688
1689 if (REG_P (x))
1690 {
1691 if (! strict
1692 && NONSTRICT_REG_OK_FP_BASE_P (x)
1693 && GET_CODE (ofs) == CONST_INT)
1694 return true;
1695 if ((strict
1696 ? STRICT_REG_OK_FOR_BASE_P (x)
1697 : NONSTRICT_REG_OK_FOR_BASE_P (x))
1698 && CONSTANT_ADDRESS_P (ofs))
1699 return true;
1700 }
1701 else if (GET_CODE (x) == ADDRESSOF
1702 && GET_CODE (ofs) == CONST_INT)
1703 return true;
1704 }
1705
1706 /* If we're managing explicit relocations, LO_SUM is valid, as
1707 are small data symbols. */
1708 else if (TARGET_EXPLICIT_RELOCS)
1709 {
1710 if (small_symbolic_operand (x, Pmode))
1711 return true;
1712
1713 if (GET_CODE (x) == LO_SUM)
1714 {
1715 rtx ofs = XEXP (x, 1);
1716 x = XEXP (x, 0);
1717
1718 /* Discard non-paradoxical subregs. */
1719 if (GET_CODE (x) == SUBREG
1720 && (GET_MODE_SIZE (GET_MODE (x))
1721 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
1722 x = SUBREG_REG (x);
1723
1724 /* Must have a valid base register. */
1725 if (! (REG_P (x)
1726 && (strict
1727 ? STRICT_REG_OK_FOR_BASE_P (x)
1728 : NONSTRICT_REG_OK_FOR_BASE_P (x))))
1729 return false;
1730
1731 /* The symbol must be local. */
1732 if (local_symbolic_operand (ofs, Pmode)
1733 || dtp32_symbolic_operand (ofs, Pmode)
1734 || tp32_symbolic_operand (ofs, Pmode))
1735 return true;
1736 }
1737 }
1738
1739 return false;
1740 }
1741
1742 /* Build the SYMBOL_REF for __tls_get_addr. */
1743
1744 static GTY(()) rtx tls_get_addr_libfunc;
1745
1746 static rtx
1747 get_tls_get_addr (void)
1748 {
1749 if (!tls_get_addr_libfunc)
1750 tls_get_addr_libfunc = init_one_libfunc ("__tls_get_addr");
1751 return tls_get_addr_libfunc;
1752 }
1753
1754 /* Try machine-dependent ways of modifying an illegitimate address
1755 to be legitimate. If we find one, return the new, valid address. */
1756
1757 rtx
1758 alpha_legitimize_address (rtx x, rtx scratch,
1759 enum machine_mode mode ATTRIBUTE_UNUSED)
1760 {
1761 HOST_WIDE_INT addend;
1762
1763 /* If the address is (plus reg const_int) and the CONST_INT is not a
1764 valid offset, compute the high part of the constant and add it to
1765 the register. Then our address is (plus temp low-part-const). */
1766 if (GET_CODE (x) == PLUS
1767 && GET_CODE (XEXP (x, 0)) == REG
1768 && GET_CODE (XEXP (x, 1)) == CONST_INT
1769 && ! CONSTANT_ADDRESS_P (XEXP (x, 1)))
1770 {
1771 addend = INTVAL (XEXP (x, 1));
1772 x = XEXP (x, 0);
1773 goto split_addend;
1774 }
1775
1776 /* If the address is (const (plus FOO const_int)), find the low-order
1777 part of the CONST_INT. Then load FOO plus any high-order part of the
1778 CONST_INT into a register. Our address is (plus reg low-part-const).
1779 This is done to reduce the number of GOT entries. */
1780 if (!no_new_pseudos
1781 && GET_CODE (x) == CONST
1782 && GET_CODE (XEXP (x, 0)) == PLUS
1783 && GET_CODE (XEXP (XEXP (x, 0), 1)) == CONST_INT)
1784 {
1785 addend = INTVAL (XEXP (XEXP (x, 0), 1));
1786 x = force_reg (Pmode, XEXP (XEXP (x, 0), 0));
1787 goto split_addend;
1788 }
1789
1790 /* If we have a (plus reg const), emit the load as in (2), then add
1791 the two registers, and finally generate (plus reg low-part-const) as
1792 our address. */
1793 if (!no_new_pseudos
1794 && GET_CODE (x) == PLUS
1795 && GET_CODE (XEXP (x, 0)) == REG
1796 && GET_CODE (XEXP (x, 1)) == CONST
1797 && GET_CODE (XEXP (XEXP (x, 1), 0)) == PLUS
1798 && GET_CODE (XEXP (XEXP (XEXP (x, 1), 0), 1)) == CONST_INT)
1799 {
1800 addend = INTVAL (XEXP (XEXP (XEXP (x, 1), 0), 1));
1801 x = expand_simple_binop (Pmode, PLUS, XEXP (x, 0),
1802 XEXP (XEXP (XEXP (x, 1), 0), 0),
1803 NULL_RTX, 1, OPTAB_LIB_WIDEN);
1804 goto split_addend;
1805 }
1806
1807 /* If this is a local symbol, split the address into HIGH/LO_SUM parts. */
1808 if (TARGET_EXPLICIT_RELOCS && symbolic_operand (x, Pmode))
1809 {
1810 rtx r0, r16, eqv, tga, tp, insn, dest, seq;
1811
1812 switch (tls_symbolic_operand_type (x))
1813 {
1814 case TLS_MODEL_GLOBAL_DYNAMIC:
1815 start_sequence ();
1816
1817 r0 = gen_rtx_REG (Pmode, 0);
1818 r16 = gen_rtx_REG (Pmode, 16);
1819 tga = get_tls_get_addr ();
1820 dest = gen_reg_rtx (Pmode);
1821 seq = GEN_INT (alpha_next_sequence_number++);
1822
1823 emit_insn (gen_movdi_er_tlsgd (r16, pic_offset_table_rtx, x, seq));
1824 insn = gen_call_value_osf_tlsgd (r0, tga, seq);
1825 insn = emit_call_insn (insn);
1826 CONST_OR_PURE_CALL_P (insn) = 1;
1827 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), r16);
1828
1829 insn = get_insns ();
1830 end_sequence ();
1831
1832 emit_libcall_block (insn, dest, r0, x);
1833 return dest;
1834
1835 case TLS_MODEL_LOCAL_DYNAMIC:
1836 start_sequence ();
1837
1838 r0 = gen_rtx_REG (Pmode, 0);
1839 r16 = gen_rtx_REG (Pmode, 16);
1840 tga = get_tls_get_addr ();
1841 scratch = gen_reg_rtx (Pmode);
1842 seq = GEN_INT (alpha_next_sequence_number++);
1843
1844 emit_insn (gen_movdi_er_tlsldm (r16, pic_offset_table_rtx, seq));
1845 insn = gen_call_value_osf_tlsldm (r0, tga, seq);
1846 insn = emit_call_insn (insn);
1847 CONST_OR_PURE_CALL_P (insn) = 1;
1848 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), r16);
1849
1850 insn = get_insns ();
1851 end_sequence ();
1852
1853 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, const0_rtx),
1854 UNSPEC_TLSLDM_CALL);
1855 emit_libcall_block (insn, scratch, r0, eqv);
1856
1857 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, x), UNSPEC_DTPREL);
1858 eqv = gen_rtx_CONST (Pmode, eqv);
1859
1860 if (alpha_tls_size == 64)
1861 {
1862 dest = gen_reg_rtx (Pmode);
1863 emit_insn (gen_rtx_SET (VOIDmode, dest, eqv));
1864 emit_insn (gen_adddi3 (dest, dest, scratch));
1865 return dest;
1866 }
1867 if (alpha_tls_size == 32)
1868 {
1869 insn = gen_rtx_HIGH (Pmode, eqv);
1870 insn = gen_rtx_PLUS (Pmode, scratch, insn);
1871 scratch = gen_reg_rtx (Pmode);
1872 emit_insn (gen_rtx_SET (VOIDmode, scratch, insn));
1873 }
1874 return gen_rtx_LO_SUM (Pmode, scratch, eqv);
1875
1876 case TLS_MODEL_INITIAL_EXEC:
1877 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, x), UNSPEC_TPREL);
1878 eqv = gen_rtx_CONST (Pmode, eqv);
1879 tp = gen_reg_rtx (Pmode);
1880 scratch = gen_reg_rtx (Pmode);
1881 dest = gen_reg_rtx (Pmode);
1882
1883 emit_insn (gen_load_tp (tp));
1884 emit_insn (gen_rtx_SET (VOIDmode, scratch, eqv));
1885 emit_insn (gen_adddi3 (dest, tp, scratch));
1886 return dest;
1887
1888 case TLS_MODEL_LOCAL_EXEC:
1889 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, x), UNSPEC_TPREL);
1890 eqv = gen_rtx_CONST (Pmode, eqv);
1891 tp = gen_reg_rtx (Pmode);
1892
1893 emit_insn (gen_load_tp (tp));
1894 if (alpha_tls_size == 32)
1895 {
1896 insn = gen_rtx_HIGH (Pmode, eqv);
1897 insn = gen_rtx_PLUS (Pmode, tp, insn);
1898 tp = gen_reg_rtx (Pmode);
1899 emit_insn (gen_rtx_SET (VOIDmode, tp, insn));
1900 }
1901 return gen_rtx_LO_SUM (Pmode, tp, eqv);
1902 }
1903
1904 if (local_symbolic_operand (x, Pmode))
1905 {
1906 if (small_symbolic_operand (x, Pmode))
1907 return x;
1908 else
1909 {
1910 if (!no_new_pseudos)
1911 scratch = gen_reg_rtx (Pmode);
1912 emit_insn (gen_rtx_SET (VOIDmode, scratch,
1913 gen_rtx_HIGH (Pmode, x)));
1914 return gen_rtx_LO_SUM (Pmode, scratch, x);
1915 }
1916 }
1917 }
1918
1919 return NULL;
1920
1921 split_addend:
1922 {
1923 HOST_WIDE_INT low, high;
1924
1925 low = ((addend & 0xffff) ^ 0x8000) - 0x8000;
1926 addend -= low;
1927 high = ((addend & 0xffffffff) ^ 0x80000000) - 0x80000000;
1928 addend -= high;
1929
1930 if (addend)
1931 x = expand_simple_binop (Pmode, PLUS, x, GEN_INT (addend),
1932 (no_new_pseudos ? scratch : NULL_RTX),
1933 1, OPTAB_LIB_WIDEN);
1934 if (high)
1935 x = expand_simple_binop (Pmode, PLUS, x, GEN_INT (high),
1936 (no_new_pseudos ? scratch : NULL_RTX),
1937 1, OPTAB_LIB_WIDEN);
1938
1939 return plus_constant (x, low);
1940 }
1941 }
1942
1943 /* We do not allow indirect calls to be optimized into sibling calls, nor
1944 can we allow a call to a function with a different GP to be optimized
1945 into a sibcall. */
1946
1947 static bool
1948 alpha_function_ok_for_sibcall (tree decl, tree exp ATTRIBUTE_UNUSED)
1949 {
1950 /* Can't do indirect tail calls, since we don't know if the target
1951 uses the same GP. */
1952 if (!decl)
1953 return false;
1954
1955 /* Otherwise, we can make a tail call if the target function shares
1956 the same GP. */
1957 return decl_has_samegp (decl);
1958 }
1959
1960 /* For TARGET_EXPLICIT_RELOCS, we don't obfuscate a SYMBOL_REF to a
1961 small symbolic operand until after reload. At which point we need
1962 to replace (mem (symbol_ref)) with (mem (lo_sum $29 symbol_ref))
1963 so that sched2 has the proper dependency information. */
1964
1965 static int
1966 some_small_symbolic_operand_1 (rtx *px, void *data ATTRIBUTE_UNUSED)
1967 {
1968 rtx x = *px;
1969
1970 /* Don't re-split. */
1971 if (GET_CODE (x) == LO_SUM)
1972 return -1;
1973
1974 return small_symbolic_operand (x, Pmode) != 0;
1975 }
1976
1977 int
1978 some_small_symbolic_operand (rtx x, enum machine_mode mode ATTRIBUTE_UNUSED)
1979 {
1980 return for_each_rtx (&x, some_small_symbolic_operand_1, NULL);
1981 }
1982
1983 static int
1984 split_small_symbolic_operand_1 (rtx *px, void *data ATTRIBUTE_UNUSED)
1985 {
1986 rtx x = *px;
1987
1988 /* Don't re-split. */
1989 if (GET_CODE (x) == LO_SUM)
1990 return -1;
1991
1992 if (small_symbolic_operand (x, Pmode))
1993 {
1994 x = gen_rtx_LO_SUM (Pmode, pic_offset_table_rtx, x);
1995 *px = x;
1996 return -1;
1997 }
1998
1999 return 0;
2000 }
2001
2002 rtx
2003 split_small_symbolic_operand (rtx x)
2004 {
2005 x = copy_insn (x);
2006 for_each_rtx (&x, split_small_symbolic_operand_1, NULL);
2007 return x;
2008 }
2009
2010 /* Indicate that INSN cannot be duplicated. This is true for any insn
2011 that we've marked with gpdisp relocs, since those have to stay in
2012 1-1 correspondence with one another.
2013
2014 Technically we could copy them if we could set up a mapping from one
2015 sequence number to another, across the set of insns to be duplicated.
2016 This seems overly complicated and error-prone since interblock motion
2017 from sched-ebb could move one of the pair of insns to a different block.
2018
2019 Also cannot allow jsr insns to be duplicated. If they throw exceptions,
2020 then they'll be in a different block from their ldgp. Which could lead
2021 the bb reorder code to think that it would be ok to copy just the block
2022 containing the call and branch to the block containing the ldgp. */
2023
2024 static bool
2025 alpha_cannot_copy_insn_p (rtx insn)
2026 {
2027 if (!reload_completed || !TARGET_EXPLICIT_RELOCS)
2028 return false;
2029 if (recog_memoized (insn) >= 0)
2030 return get_attr_cannot_copy (insn);
2031 else
2032 return false;
2033 }
2034
2035
2036 /* Try a machine-dependent way of reloading an illegitimate address
2037 operand. If we find one, push the reload and return the new rtx. */
2038
2039 rtx
2040 alpha_legitimize_reload_address (rtx x,
2041 enum machine_mode mode ATTRIBUTE_UNUSED,
2042 int opnum, int type,
2043 int ind_levels ATTRIBUTE_UNUSED)
2044 {
2045 /* We must recognize output that we have already generated ourselves. */
2046 if (GET_CODE (x) == PLUS
2047 && GET_CODE (XEXP (x, 0)) == PLUS
2048 && GET_CODE (XEXP (XEXP (x, 0), 0)) == REG
2049 && GET_CODE (XEXP (XEXP (x, 0), 1)) == CONST_INT
2050 && GET_CODE (XEXP (x, 1)) == CONST_INT)
2051 {
2052 push_reload (XEXP (x, 0), NULL_RTX, &XEXP (x, 0), NULL,
2053 BASE_REG_CLASS, GET_MODE (x), VOIDmode, 0, 0,
2054 opnum, type);
2055 return x;
2056 }
2057
2058 /* We wish to handle large displacements off a base register by
2059 splitting the addend across an ldah and the mem insn. This
2060 cuts number of extra insns needed from 3 to 1. */
2061 if (GET_CODE (x) == PLUS
2062 && GET_CODE (XEXP (x, 0)) == REG
2063 && REGNO (XEXP (x, 0)) < FIRST_PSEUDO_REGISTER
2064 && REGNO_OK_FOR_BASE_P (REGNO (XEXP (x, 0)))
2065 && GET_CODE (XEXP (x, 1)) == CONST_INT)
2066 {
2067 HOST_WIDE_INT val = INTVAL (XEXP (x, 1));
2068 HOST_WIDE_INT low = ((val & 0xffff) ^ 0x8000) - 0x8000;
2069 HOST_WIDE_INT high
2070 = (((val - low) & 0xffffffff) ^ 0x80000000) - 0x80000000;
2071
2072 /* Check for 32-bit overflow. */
2073 if (high + low != val)
2074 return NULL_RTX;
2075
2076 /* Reload the high part into a base reg; leave the low part
2077 in the mem directly. */
2078 x = gen_rtx_PLUS (GET_MODE (x),
2079 gen_rtx_PLUS (GET_MODE (x), XEXP (x, 0),
2080 GEN_INT (high)),
2081 GEN_INT (low));
2082
2083 push_reload (XEXP (x, 0), NULL_RTX, &XEXP (x, 0), NULL,
2084 BASE_REG_CLASS, GET_MODE (x), VOIDmode, 0, 0,
2085 opnum, type);
2086 return x;
2087 }
2088
2089 return NULL_RTX;
2090 }
2091 \f
2092 /* Compute a (partial) cost for rtx X. Return true if the complete
2093 cost has been computed, and false if subexpressions should be
2094 scanned. In either case, *TOTAL contains the cost result. */
2095
2096 static bool
2097 alpha_rtx_costs (rtx x, int code, int outer_code, int *total)
2098 {
2099 enum machine_mode mode = GET_MODE (x);
2100 bool float_mode_p = FLOAT_MODE_P (mode);
2101
2102 switch (code)
2103 {
2104 /* If this is an 8-bit constant, return zero since it can be used
2105 nearly anywhere with no cost. If it is a valid operand for an
2106 ADD or AND, likewise return 0 if we know it will be used in that
2107 context. Otherwise, return 2 since it might be used there later.
2108 All other constants take at least two insns. */
2109 case CONST_INT:
2110 if (INTVAL (x) >= 0 && INTVAL (x) < 256)
2111 {
2112 *total = 0;
2113 return true;
2114 }
2115 /* FALLTHRU */
2116
2117 case CONST_DOUBLE:
2118 if (x == CONST0_RTX (mode))
2119 *total = 0;
2120 else if ((outer_code == PLUS && add_operand (x, VOIDmode))
2121 || (outer_code == AND && and_operand (x, VOIDmode)))
2122 *total = 0;
2123 else if (add_operand (x, VOIDmode) || and_operand (x, VOIDmode))
2124 *total = 2;
2125 else
2126 *total = COSTS_N_INSNS (2);
2127 return true;
2128
2129 case CONST:
2130 case SYMBOL_REF:
2131 case LABEL_REF:
2132 if (TARGET_EXPLICIT_RELOCS && small_symbolic_operand (x, VOIDmode))
2133 *total = COSTS_N_INSNS (outer_code != MEM);
2134 else if (TARGET_EXPLICIT_RELOCS && local_symbolic_operand (x, VOIDmode))
2135 *total = COSTS_N_INSNS (1 + (outer_code != MEM));
2136 else if (tls_symbolic_operand_type (x))
2137 /* Estimate of cost for call_pal rduniq. */
2138 *total = COSTS_N_INSNS (15);
2139 else
2140 /* Otherwise we do a load from the GOT. */
2141 *total = COSTS_N_INSNS (alpha_memory_latency);
2142 return true;
2143
2144 case PLUS:
2145 case MINUS:
2146 if (float_mode_p)
2147 *total = alpha_rtx_cost_data[alpha_cpu].fp_add;
2148 else if (GET_CODE (XEXP (x, 0)) == MULT
2149 && const48_operand (XEXP (XEXP (x, 0), 1), VOIDmode))
2150 {
2151 *total = (rtx_cost (XEXP (XEXP (x, 0), 0), outer_code)
2152 + rtx_cost (XEXP (x, 1), outer_code) + 2);
2153 return true;
2154 }
2155 return false;
2156
2157 case MULT:
2158 if (float_mode_p)
2159 *total = alpha_rtx_cost_data[alpha_cpu].fp_mult;
2160 else if (mode == DImode)
2161 *total = alpha_rtx_cost_data[alpha_cpu].int_mult_di;
2162 else
2163 *total = alpha_rtx_cost_data[alpha_cpu].int_mult_si;
2164 return false;
2165
2166 case ASHIFT:
2167 if (GET_CODE (XEXP (x, 1)) == CONST_INT
2168 && INTVAL (XEXP (x, 1)) <= 3)
2169 {
2170 *total = COSTS_N_INSNS (1);
2171 return false;
2172 }
2173 /* FALLTHRU */
2174
2175 case ASHIFTRT:
2176 case LSHIFTRT:
2177 *total = alpha_rtx_cost_data[alpha_cpu].int_shift;
2178 return false;
2179
2180 case IF_THEN_ELSE:
2181 if (float_mode_p)
2182 *total = alpha_rtx_cost_data[alpha_cpu].fp_add;
2183 else
2184 *total = alpha_rtx_cost_data[alpha_cpu].int_cmov;
2185 return false;
2186
2187 case DIV:
2188 case UDIV:
2189 case MOD:
2190 case UMOD:
2191 if (!float_mode_p)
2192 *total = COSTS_N_INSNS (70); /* ??? */
2193 else if (mode == SFmode)
2194 *total = alpha_rtx_cost_data[alpha_cpu].fp_div_sf;
2195 else
2196 *total = alpha_rtx_cost_data[alpha_cpu].fp_div_df;
2197 return false;
2198
2199 case MEM:
2200 *total = COSTS_N_INSNS (alpha_memory_latency);
2201 return true;
2202
2203 case NEG:
2204 if (! float_mode_p)
2205 {
2206 *total = COSTS_N_INSNS (1);
2207 return false;
2208 }
2209 /* FALLTHRU */
2210
2211 case ABS:
2212 if (! float_mode_p)
2213 {
2214 *total = COSTS_N_INSNS (1) + alpha_rtx_cost_data[alpha_cpu].int_cmov;
2215 return false;
2216 }
2217 /* FALLTHRU */
2218
2219 case FLOAT:
2220 case UNSIGNED_FLOAT:
2221 case FIX:
2222 case UNSIGNED_FIX:
2223 case FLOAT_EXTEND:
2224 case FLOAT_TRUNCATE:
2225 *total = alpha_rtx_cost_data[alpha_cpu].fp_add;
2226 return false;
2227
2228 default:
2229 return false;
2230 }
2231 }
2232 \f
2233 /* REF is an alignable memory location. Place an aligned SImode
2234 reference into *PALIGNED_MEM and the number of bits to shift into
2235 *PBITNUM. SCRATCH is a free register for use in reloading out
2236 of range stack slots. */
2237
2238 void
2239 get_aligned_mem (rtx ref, rtx *paligned_mem, rtx *pbitnum)
2240 {
2241 rtx base;
2242 HOST_WIDE_INT offset = 0;
2243
2244 if (GET_CODE (ref) != MEM)
2245 abort ();
2246
2247 if (reload_in_progress
2248 && ! memory_address_p (GET_MODE (ref), XEXP (ref, 0)))
2249 {
2250 base = find_replacement (&XEXP (ref, 0));
2251
2252 if (! memory_address_p (GET_MODE (ref), base))
2253 abort ();
2254 }
2255 else
2256 {
2257 base = XEXP (ref, 0);
2258 }
2259
2260 if (GET_CODE (base) == PLUS)
2261 offset += INTVAL (XEXP (base, 1)), base = XEXP (base, 0);
2262
2263 *paligned_mem
2264 = widen_memory_access (ref, SImode, (offset & ~3) - offset);
2265
2266 if (WORDS_BIG_ENDIAN)
2267 *pbitnum = GEN_INT (32 - (GET_MODE_BITSIZE (GET_MODE (ref))
2268 + (offset & 3) * 8));
2269 else
2270 *pbitnum = GEN_INT ((offset & 3) * 8);
2271 }
2272
2273 /* Similar, but just get the address. Handle the two reload cases.
2274 Add EXTRA_OFFSET to the address we return. */
2275
2276 rtx
2277 get_unaligned_address (rtx ref, int extra_offset)
2278 {
2279 rtx base;
2280 HOST_WIDE_INT offset = 0;
2281
2282 if (GET_CODE (ref) != MEM)
2283 abort ();
2284
2285 if (reload_in_progress
2286 && ! memory_address_p (GET_MODE (ref), XEXP (ref, 0)))
2287 {
2288 base = find_replacement (&XEXP (ref, 0));
2289
2290 if (! memory_address_p (GET_MODE (ref), base))
2291 abort ();
2292 }
2293 else
2294 {
2295 base = XEXP (ref, 0);
2296 }
2297
2298 if (GET_CODE (base) == PLUS)
2299 offset += INTVAL (XEXP (base, 1)), base = XEXP (base, 0);
2300
2301 return plus_constant (base, offset + extra_offset);
2302 }
2303
2304 /* On the Alpha, all (non-symbolic) constants except zero go into
2305 a floating-point register via memory. Note that we cannot
2306 return anything that is not a subset of CLASS, and that some
2307 symbolic constants cannot be dropped to memory. */
2308
2309 enum reg_class
2310 alpha_preferred_reload_class(rtx x, enum reg_class class)
2311 {
2312 /* Zero is present in any register class. */
2313 if (x == CONST0_RTX (GET_MODE (x)))
2314 return class;
2315
2316 /* These sorts of constants we can easily drop to memory. */
2317 if (GET_CODE (x) == CONST_INT || GET_CODE (x) == CONST_DOUBLE)
2318 {
2319 if (class == FLOAT_REGS)
2320 return NO_REGS;
2321 if (class == ALL_REGS)
2322 return GENERAL_REGS;
2323 return class;
2324 }
2325
2326 /* All other kinds of constants should not (and in the case of HIGH
2327 cannot) be dropped to memory -- instead we use a GENERAL_REGS
2328 secondary reload. */
2329 if (CONSTANT_P (x))
2330 return (class == ALL_REGS ? GENERAL_REGS : class);
2331
2332 return class;
2333 }
2334
2335 /* Loading and storing HImode or QImode values to and from memory
2336 usually requires a scratch register. The exceptions are loading
2337 QImode and HImode from an aligned address to a general register
2338 unless byte instructions are permitted.
2339
2340 We also cannot load an unaligned address or a paradoxical SUBREG
2341 into an FP register.
2342
2343 We also cannot do integral arithmetic into FP regs, as might result
2344 from register elimination into a DImode fp register. */
2345
2346 enum reg_class
2347 secondary_reload_class (enum reg_class class, enum machine_mode mode,
2348 rtx x, int in)
2349 {
2350 if ((mode == QImode || mode == HImode) && ! TARGET_BWX)
2351 {
2352 if (GET_CODE (x) == MEM
2353 || (GET_CODE (x) == REG && REGNO (x) >= FIRST_PSEUDO_REGISTER)
2354 || (GET_CODE (x) == SUBREG
2355 && (GET_CODE (SUBREG_REG (x)) == MEM
2356 || (GET_CODE (SUBREG_REG (x)) == REG
2357 && REGNO (SUBREG_REG (x)) >= FIRST_PSEUDO_REGISTER))))
2358 {
2359 if (!in || !aligned_memory_operand(x, mode))
2360 return GENERAL_REGS;
2361 }
2362 }
2363
2364 if (class == FLOAT_REGS)
2365 {
2366 if (GET_CODE (x) == MEM && GET_CODE (XEXP (x, 0)) == AND)
2367 return GENERAL_REGS;
2368
2369 if (GET_CODE (x) == SUBREG
2370 && (GET_MODE_SIZE (GET_MODE (x))
2371 > GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
2372 return GENERAL_REGS;
2373
2374 if (in && INTEGRAL_MODE_P (mode)
2375 && ! (memory_operand (x, mode) || x == const0_rtx))
2376 return GENERAL_REGS;
2377 }
2378
2379 return NO_REGS;
2380 }
2381 \f
2382 /* Subfunction of the following function. Update the flags of any MEM
2383 found in part of X. */
2384
2385 static void
2386 alpha_set_memflags_1 (rtx x, int in_struct_p, int volatile_p, int unchanging_p)
2387 {
2388 int i;
2389
2390 switch (GET_CODE (x))
2391 {
2392 case SEQUENCE:
2393 abort ();
2394
2395 case PARALLEL:
2396 for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
2397 alpha_set_memflags_1 (XVECEXP (x, 0, i), in_struct_p, volatile_p,
2398 unchanging_p);
2399 break;
2400
2401 case INSN:
2402 alpha_set_memflags_1 (PATTERN (x), in_struct_p, volatile_p,
2403 unchanging_p);
2404 break;
2405
2406 case SET:
2407 alpha_set_memflags_1 (SET_DEST (x), in_struct_p, volatile_p,
2408 unchanging_p);
2409 alpha_set_memflags_1 (SET_SRC (x), in_struct_p, volatile_p,
2410 unchanging_p);
2411 break;
2412
2413 case MEM:
2414 MEM_IN_STRUCT_P (x) = in_struct_p;
2415 MEM_VOLATILE_P (x) = volatile_p;
2416 RTX_UNCHANGING_P (x) = unchanging_p;
2417 /* Sadly, we cannot use alias sets because the extra aliasing
2418 produced by the AND interferes. Given that two-byte quantities
2419 are the only thing we would be able to differentiate anyway,
2420 there does not seem to be any point in convoluting the early
2421 out of the alias check. */
2422 break;
2423
2424 default:
2425 break;
2426 }
2427 }
2428
2429 /* Given INSN, which is an INSN list or the PATTERN of a single insn
2430 generated to perform a memory operation, look for any MEMs in either
2431 a SET_DEST or a SET_SRC and copy the in-struct, unchanging, and
2432 volatile flags from REF into each of the MEMs found. If REF is not
2433 a MEM, don't do anything. */
2434
2435 void
2436 alpha_set_memflags (rtx insn, rtx ref)
2437 {
2438 int in_struct_p, volatile_p, unchanging_p;
2439
2440 if (GET_CODE (ref) != MEM)
2441 return;
2442
2443 in_struct_p = MEM_IN_STRUCT_P (ref);
2444 volatile_p = MEM_VOLATILE_P (ref);
2445 unchanging_p = RTX_UNCHANGING_P (ref);
2446
2447 /* This is only called from alpha.md, after having had something
2448 generated from one of the insn patterns. So if everything is
2449 zero, the pattern is already up-to-date. */
2450 if (! in_struct_p && ! volatile_p && ! unchanging_p)
2451 return;
2452
2453 alpha_set_memflags_1 (insn, in_struct_p, volatile_p, unchanging_p);
2454 }
2455 \f
2456 /* Internal routine for alpha_emit_set_const to check for N or below insns. */
2457
2458 static rtx
2459 alpha_emit_set_const_1 (rtx target, enum machine_mode mode,
2460 HOST_WIDE_INT c, int n)
2461 {
2462 HOST_WIDE_INT new;
2463 int i, bits;
2464 /* Use a pseudo if highly optimizing and still generating RTL. */
2465 rtx subtarget
2466 = (flag_expensive_optimizations && !no_new_pseudos ? 0 : target);
2467 rtx temp, insn;
2468
2469 /* If this is a sign-extended 32-bit constant, we can do this in at most
2470 three insns, so do it if we have enough insns left. We always have
2471 a sign-extended 32-bit constant when compiling on a narrow machine. */
2472
2473 if (HOST_BITS_PER_WIDE_INT != 64
2474 || c >> 31 == -1 || c >> 31 == 0)
2475 {
2476 HOST_WIDE_INT low = ((c & 0xffff) ^ 0x8000) - 0x8000;
2477 HOST_WIDE_INT tmp1 = c - low;
2478 HOST_WIDE_INT high = (((tmp1 >> 16) & 0xffff) ^ 0x8000) - 0x8000;
2479 HOST_WIDE_INT extra = 0;
2480
2481 /* If HIGH will be interpreted as negative but the constant is
2482 positive, we must adjust it to do two ldha insns. */
2483
2484 if ((high & 0x8000) != 0 && c >= 0)
2485 {
2486 extra = 0x4000;
2487 tmp1 -= 0x40000000;
2488 high = ((tmp1 >> 16) & 0xffff) - 2 * ((tmp1 >> 16) & 0x8000);
2489 }
2490
2491 if (c == low || (low == 0 && extra == 0))
2492 {
2493 /* We used to use copy_to_suggested_reg (GEN_INT (c), target, mode)
2494 but that meant that we can't handle INT_MIN on 32-bit machines
2495 (like NT/Alpha), because we recurse indefinitely through
2496 emit_move_insn to gen_movdi. So instead, since we know exactly
2497 what we want, create it explicitly. */
2498
2499 if (target == NULL)
2500 target = gen_reg_rtx (mode);
2501 emit_insn (gen_rtx_SET (VOIDmode, target, GEN_INT (c)));
2502 return target;
2503 }
2504 else if (n >= 2 + (extra != 0))
2505 {
2506 if (no_new_pseudos)
2507 {
2508 emit_insn (gen_rtx_SET (VOIDmode, target, GEN_INT (high << 16)));
2509 temp = target;
2510 }
2511 else
2512 temp = copy_to_suggested_reg (GEN_INT (high << 16),
2513 subtarget, mode);
2514
2515 /* As of 2002-02-23, addsi3 is only available when not optimizing.
2516 This means that if we go through expand_binop, we'll try to
2517 generate extensions, etc, which will require new pseudos, which
2518 will fail during some split phases. The SImode add patterns
2519 still exist, but are not named. So build the insns by hand. */
2520
2521 if (extra != 0)
2522 {
2523 if (! subtarget)
2524 subtarget = gen_reg_rtx (mode);
2525 insn = gen_rtx_PLUS (mode, temp, GEN_INT (extra << 16));
2526 insn = gen_rtx_SET (VOIDmode, subtarget, insn);
2527 emit_insn (insn);
2528 temp = subtarget;
2529 }
2530
2531 if (target == NULL)
2532 target = gen_reg_rtx (mode);
2533 insn = gen_rtx_PLUS (mode, temp, GEN_INT (low));
2534 insn = gen_rtx_SET (VOIDmode, target, insn);
2535 emit_insn (insn);
2536 return target;
2537 }
2538 }
2539
2540 /* If we couldn't do it that way, try some other methods. But if we have
2541 no instructions left, don't bother. Likewise, if this is SImode and
2542 we can't make pseudos, we can't do anything since the expand_binop
2543 and expand_unop calls will widen and try to make pseudos. */
2544
2545 if (n == 1 || (mode == SImode && no_new_pseudos))
2546 return 0;
2547
2548 /* Next, see if we can load a related constant and then shift and possibly
2549 negate it to get the constant we want. Try this once each increasing
2550 numbers of insns. */
2551
2552 for (i = 1; i < n; i++)
2553 {
2554 /* First, see if minus some low bits, we've an easy load of
2555 high bits. */
2556
2557 new = ((c & 0xffff) ^ 0x8000) - 0x8000;
2558 if (new != 0
2559 && (temp = alpha_emit_set_const (subtarget, mode, c - new, i)) != 0)
2560 return expand_binop (mode, add_optab, temp, GEN_INT (new),
2561 target, 0, OPTAB_WIDEN);
2562
2563 /* Next try complementing. */
2564 if ((temp = alpha_emit_set_const (subtarget, mode, ~ c, i)) != 0)
2565 return expand_unop (mode, one_cmpl_optab, temp, target, 0);
2566
2567 /* Next try to form a constant and do a left shift. We can do this
2568 if some low-order bits are zero; the exact_log2 call below tells
2569 us that information. The bits we are shifting out could be any
2570 value, but here we'll just try the 0- and sign-extended forms of
2571 the constant. To try to increase the chance of having the same
2572 constant in more than one insn, start at the highest number of
2573 bits to shift, but try all possibilities in case a ZAPNOT will
2574 be useful. */
2575
2576 if ((bits = exact_log2 (c & - c)) > 0)
2577 for (; bits > 0; bits--)
2578 if ((temp = (alpha_emit_set_const
2579 (subtarget, mode, c >> bits, i))) != 0
2580 || ((temp = (alpha_emit_set_const
2581 (subtarget, mode,
2582 ((unsigned HOST_WIDE_INT) c) >> bits, i)))
2583 != 0))
2584 return expand_binop (mode, ashl_optab, temp, GEN_INT (bits),
2585 target, 0, OPTAB_WIDEN);
2586
2587 /* Now try high-order zero bits. Here we try the shifted-in bits as
2588 all zero and all ones. Be careful to avoid shifting outside the
2589 mode and to avoid shifting outside the host wide int size. */
2590 /* On narrow hosts, don't shift a 1 into the high bit, since we'll
2591 confuse the recursive call and set all of the high 32 bits. */
2592
2593 if ((bits = (MIN (HOST_BITS_PER_WIDE_INT, GET_MODE_SIZE (mode) * 8)
2594 - floor_log2 (c) - 1 - (HOST_BITS_PER_WIDE_INT < 64))) > 0)
2595 for (; bits > 0; bits--)
2596 if ((temp = alpha_emit_set_const (subtarget, mode,
2597 c << bits, i)) != 0
2598 || ((temp = (alpha_emit_set_const
2599 (subtarget, mode,
2600 ((c << bits) | (((HOST_WIDE_INT) 1 << bits) - 1)),
2601 i)))
2602 != 0))
2603 return expand_binop (mode, lshr_optab, temp, GEN_INT (bits),
2604 target, 1, OPTAB_WIDEN);
2605
2606 /* Now try high-order 1 bits. We get that with a sign-extension.
2607 But one bit isn't enough here. Be careful to avoid shifting outside
2608 the mode and to avoid shifting outside the host wide int size. */
2609
2610 if ((bits = (MIN (HOST_BITS_PER_WIDE_INT, GET_MODE_SIZE (mode) * 8)
2611 - floor_log2 (~ c) - 2)) > 0)
2612 for (; bits > 0; bits--)
2613 if ((temp = alpha_emit_set_const (subtarget, mode,
2614 c << bits, i)) != 0
2615 || ((temp = (alpha_emit_set_const
2616 (subtarget, mode,
2617 ((c << bits) | (((HOST_WIDE_INT) 1 << bits) - 1)),
2618 i)))
2619 != 0))
2620 return expand_binop (mode, ashr_optab, temp, GEN_INT (bits),
2621 target, 0, OPTAB_WIDEN);
2622 }
2623
2624 #if HOST_BITS_PER_WIDE_INT == 64
2625 /* Finally, see if can load a value into the target that is the same as the
2626 constant except that all bytes that are 0 are changed to be 0xff. If we
2627 can, then we can do a ZAPNOT to obtain the desired constant. */
2628
2629 new = c;
2630 for (i = 0; i < 64; i += 8)
2631 if ((new & ((HOST_WIDE_INT) 0xff << i)) == 0)
2632 new |= (HOST_WIDE_INT) 0xff << i;
2633
2634 /* We are only called for SImode and DImode. If this is SImode, ensure that
2635 we are sign extended to a full word. */
2636
2637 if (mode == SImode)
2638 new = ((new & 0xffffffff) ^ 0x80000000) - 0x80000000;
2639
2640 if (new != c && new != -1
2641 && (temp = alpha_emit_set_const (subtarget, mode, new, n - 1)) != 0)
2642 return expand_binop (mode, and_optab, temp, GEN_INT (c | ~ new),
2643 target, 0, OPTAB_WIDEN);
2644 #endif
2645
2646 return 0;
2647 }
2648
2649 /* Try to output insns to set TARGET equal to the constant C if it can be
2650 done in less than N insns. Do all computations in MODE. Returns the place
2651 where the output has been placed if it can be done and the insns have been
2652 emitted. If it would take more than N insns, zero is returned and no
2653 insns and emitted. */
2654
2655 rtx
2656 alpha_emit_set_const (rtx target, enum machine_mode mode,
2657 HOST_WIDE_INT c, int n)
2658 {
2659 rtx result = 0;
2660 rtx orig_target = target;
2661 int i;
2662
2663 /* If we can't make any pseudos, TARGET is an SImode hard register, we
2664 can't load this constant in one insn, do this in DImode. */
2665 if (no_new_pseudos && mode == SImode
2666 && GET_CODE (target) == REG && REGNO (target) < FIRST_PSEUDO_REGISTER
2667 && (result = alpha_emit_set_const_1 (target, mode, c, 1)) == 0)
2668 {
2669 target = gen_lowpart (DImode, target);
2670 mode = DImode;
2671 }
2672
2673 /* Try 1 insn, then 2, then up to N. */
2674 for (i = 1; i <= n; i++)
2675 {
2676 result = alpha_emit_set_const_1 (target, mode, c, i);
2677 if (result)
2678 {
2679 rtx insn = get_last_insn ();
2680 rtx set = single_set (insn);
2681 if (! CONSTANT_P (SET_SRC (set)))
2682 set_unique_reg_note (get_last_insn (), REG_EQUAL, GEN_INT (c));
2683 break;
2684 }
2685 }
2686
2687 /* Allow for the case where we changed the mode of TARGET. */
2688 if (result == target)
2689 result = orig_target;
2690
2691 return result;
2692 }
2693
2694 /* Having failed to find a 3 insn sequence in alpha_emit_set_const,
2695 fall back to a straight forward decomposition. We do this to avoid
2696 exponential run times encountered when looking for longer sequences
2697 with alpha_emit_set_const. */
2698
2699 rtx
2700 alpha_emit_set_long_const (rtx target, HOST_WIDE_INT c1, HOST_WIDE_INT c2)
2701 {
2702 HOST_WIDE_INT d1, d2, d3, d4;
2703
2704 /* Decompose the entire word */
2705 #if HOST_BITS_PER_WIDE_INT >= 64
2706 if (c2 != -(c1 < 0))
2707 abort ();
2708 d1 = ((c1 & 0xffff) ^ 0x8000) - 0x8000;
2709 c1 -= d1;
2710 d2 = ((c1 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2711 c1 = (c1 - d2) >> 32;
2712 d3 = ((c1 & 0xffff) ^ 0x8000) - 0x8000;
2713 c1 -= d3;
2714 d4 = ((c1 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2715 if (c1 != d4)
2716 abort ();
2717 #else
2718 d1 = ((c1 & 0xffff) ^ 0x8000) - 0x8000;
2719 c1 -= d1;
2720 d2 = ((c1 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2721 if (c1 != d2)
2722 abort ();
2723 c2 += (d2 < 0);
2724 d3 = ((c2 & 0xffff) ^ 0x8000) - 0x8000;
2725 c2 -= d3;
2726 d4 = ((c2 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2727 if (c2 != d4)
2728 abort ();
2729 #endif
2730
2731 /* Construct the high word */
2732 if (d4)
2733 {
2734 emit_move_insn (target, GEN_INT (d4));
2735 if (d3)
2736 emit_move_insn (target, gen_rtx_PLUS (DImode, target, GEN_INT (d3)));
2737 }
2738 else
2739 emit_move_insn (target, GEN_INT (d3));
2740
2741 /* Shift it into place */
2742 emit_move_insn (target, gen_rtx_ASHIFT (DImode, target, GEN_INT (32)));
2743
2744 /* Add in the low bits. */
2745 if (d2)
2746 emit_move_insn (target, gen_rtx_PLUS (DImode, target, GEN_INT (d2)));
2747 if (d1)
2748 emit_move_insn (target, gen_rtx_PLUS (DImode, target, GEN_INT (d1)));
2749
2750 return target;
2751 }
2752
2753 /* Expand a move instruction; return true if all work is done.
2754 We don't handle non-bwx subword loads here. */
2755
2756 bool
2757 alpha_expand_mov (enum machine_mode mode, rtx *operands)
2758 {
2759 /* If the output is not a register, the input must be. */
2760 if (GET_CODE (operands[0]) == MEM
2761 && ! reg_or_0_operand (operands[1], mode))
2762 operands[1] = force_reg (mode, operands[1]);
2763
2764 /* Allow legitimize_address to perform some simplifications. */
2765 if (mode == Pmode && symbolic_operand (operands[1], mode))
2766 {
2767 rtx tmp;
2768
2769 /* With RTL inlining, at -O3, rtl is generated, stored, then actually
2770 compiled at the end of compilation. In the meantime, someone can
2771 re-encode-section-info on some symbol changing it e.g. from global
2772 to local-not-small. If this happens, we'd have emitted a plain
2773 load rather than a high+losum load and not recognize the insn.
2774
2775 So if rtl inlining is in effect, we delay the global/not-global
2776 decision until rest_of_compilation by wrapping it in an
2777 UNSPEC_SYMBOL. */
2778 if (TARGET_EXPLICIT_RELOCS && flag_inline_functions
2779 && rtx_equal_function_value_matters
2780 && global_symbolic_operand (operands[1], mode))
2781 {
2782 emit_insn (gen_movdi_er_maybe_g (operands[0], operands[1]));
2783 return true;
2784 }
2785
2786 tmp = alpha_legitimize_address (operands[1], operands[0], mode);
2787 if (tmp)
2788 {
2789 if (tmp == operands[0])
2790 return true;
2791 operands[1] = tmp;
2792 return false;
2793 }
2794 }
2795
2796 /* Early out for non-constants and valid constants. */
2797 if (! CONSTANT_P (operands[1]) || input_operand (operands[1], mode))
2798 return false;
2799
2800 /* Split large integers. */
2801 if (GET_CODE (operands[1]) == CONST_INT
2802 || GET_CODE (operands[1]) == CONST_DOUBLE)
2803 {
2804 HOST_WIDE_INT i0, i1;
2805 rtx temp = NULL_RTX;
2806
2807 if (GET_CODE (operands[1]) == CONST_INT)
2808 {
2809 i0 = INTVAL (operands[1]);
2810 i1 = -(i0 < 0);
2811 }
2812 else if (HOST_BITS_PER_WIDE_INT >= 64)
2813 {
2814 i0 = CONST_DOUBLE_LOW (operands[1]);
2815 i1 = -(i0 < 0);
2816 }
2817 else
2818 {
2819 i0 = CONST_DOUBLE_LOW (operands[1]);
2820 i1 = CONST_DOUBLE_HIGH (operands[1]);
2821 }
2822
2823 if (HOST_BITS_PER_WIDE_INT >= 64 || i1 == -(i0 < 0))
2824 temp = alpha_emit_set_const (operands[0], mode, i0, 3);
2825
2826 if (!temp && TARGET_BUILD_CONSTANTS)
2827 temp = alpha_emit_set_long_const (operands[0], i0, i1);
2828
2829 if (temp)
2830 {
2831 if (rtx_equal_p (operands[0], temp))
2832 return true;
2833 operands[1] = temp;
2834 return false;
2835 }
2836 }
2837
2838 /* Otherwise we've nothing left but to drop the thing to memory. */
2839 operands[1] = force_const_mem (mode, operands[1]);
2840 if (reload_in_progress)
2841 {
2842 emit_move_insn (operands[0], XEXP (operands[1], 0));
2843 operands[1] = copy_rtx (operands[1]);
2844 XEXP (operands[1], 0) = operands[0];
2845 }
2846 else
2847 operands[1] = validize_mem (operands[1]);
2848 return false;
2849 }
2850
2851 /* Expand a non-bwx QImode or HImode move instruction;
2852 return true if all work is done. */
2853
2854 bool
2855 alpha_expand_mov_nobwx (enum machine_mode mode, rtx *operands)
2856 {
2857 /* If the output is not a register, the input must be. */
2858 if (GET_CODE (operands[0]) == MEM)
2859 operands[1] = force_reg (mode, operands[1]);
2860
2861 /* Handle four memory cases, unaligned and aligned for either the input
2862 or the output. The only case where we can be called during reload is
2863 for aligned loads; all other cases require temporaries. */
2864
2865 if (GET_CODE (operands[1]) == MEM
2866 || (GET_CODE (operands[1]) == SUBREG
2867 && GET_CODE (SUBREG_REG (operands[1])) == MEM)
2868 || (reload_in_progress && GET_CODE (operands[1]) == REG
2869 && REGNO (operands[1]) >= FIRST_PSEUDO_REGISTER)
2870 || (reload_in_progress && GET_CODE (operands[1]) == SUBREG
2871 && GET_CODE (SUBREG_REG (operands[1])) == REG
2872 && REGNO (SUBREG_REG (operands[1])) >= FIRST_PSEUDO_REGISTER))
2873 {
2874 if (aligned_memory_operand (operands[1], mode))
2875 {
2876 if (reload_in_progress)
2877 {
2878 emit_insn ((mode == QImode
2879 ? gen_reload_inqi_help
2880 : gen_reload_inhi_help)
2881 (operands[0], operands[1],
2882 gen_rtx_REG (SImode, REGNO (operands[0]))));
2883 }
2884 else
2885 {
2886 rtx aligned_mem, bitnum;
2887 rtx scratch = gen_reg_rtx (SImode);
2888 rtx subtarget;
2889 bool copyout;
2890
2891 get_aligned_mem (operands[1], &aligned_mem, &bitnum);
2892
2893 subtarget = operands[0];
2894 if (GET_CODE (subtarget) == REG)
2895 subtarget = gen_lowpart (DImode, subtarget), copyout = false;
2896 else
2897 subtarget = gen_reg_rtx (DImode), copyout = true;
2898
2899 emit_insn ((mode == QImode
2900 ? gen_aligned_loadqi
2901 : gen_aligned_loadhi)
2902 (subtarget, aligned_mem, bitnum, scratch));
2903
2904 if (copyout)
2905 emit_move_insn (operands[0], gen_lowpart (mode, subtarget));
2906 }
2907 }
2908 else
2909 {
2910 /* Don't pass these as parameters since that makes the generated
2911 code depend on parameter evaluation order which will cause
2912 bootstrap failures. */
2913
2914 rtx temp1, temp2, seq, subtarget;
2915 bool copyout;
2916
2917 temp1 = gen_reg_rtx (DImode);
2918 temp2 = gen_reg_rtx (DImode);
2919
2920 subtarget = operands[0];
2921 if (GET_CODE (subtarget) == REG)
2922 subtarget = gen_lowpart (DImode, subtarget), copyout = false;
2923 else
2924 subtarget = gen_reg_rtx (DImode), copyout = true;
2925
2926 seq = ((mode == QImode
2927 ? gen_unaligned_loadqi
2928 : gen_unaligned_loadhi)
2929 (subtarget, get_unaligned_address (operands[1], 0),
2930 temp1, temp2));
2931 alpha_set_memflags (seq, operands[1]);
2932 emit_insn (seq);
2933
2934 if (copyout)
2935 emit_move_insn (operands[0], gen_lowpart (mode, subtarget));
2936 }
2937 return true;
2938 }
2939
2940 if (GET_CODE (operands[0]) == MEM
2941 || (GET_CODE (operands[0]) == SUBREG
2942 && GET_CODE (SUBREG_REG (operands[0])) == MEM)
2943 || (reload_in_progress && GET_CODE (operands[0]) == REG
2944 && REGNO (operands[0]) >= FIRST_PSEUDO_REGISTER)
2945 || (reload_in_progress && GET_CODE (operands[0]) == SUBREG
2946 && GET_CODE (SUBREG_REG (operands[0])) == REG
2947 && REGNO (operands[0]) >= FIRST_PSEUDO_REGISTER))
2948 {
2949 if (aligned_memory_operand (operands[0], mode))
2950 {
2951 rtx aligned_mem, bitnum;
2952 rtx temp1 = gen_reg_rtx (SImode);
2953 rtx temp2 = gen_reg_rtx (SImode);
2954
2955 get_aligned_mem (operands[0], &aligned_mem, &bitnum);
2956
2957 emit_insn (gen_aligned_store (aligned_mem, operands[1], bitnum,
2958 temp1, temp2));
2959 }
2960 else
2961 {
2962 rtx temp1 = gen_reg_rtx (DImode);
2963 rtx temp2 = gen_reg_rtx (DImode);
2964 rtx temp3 = gen_reg_rtx (DImode);
2965 rtx seq = ((mode == QImode
2966 ? gen_unaligned_storeqi
2967 : gen_unaligned_storehi)
2968 (get_unaligned_address (operands[0], 0),
2969 operands[1], temp1, temp2, temp3));
2970
2971 alpha_set_memflags (seq, operands[0]);
2972 emit_insn (seq);
2973 }
2974 return true;
2975 }
2976
2977 return false;
2978 }
2979
2980 /* Generate an unsigned DImode to FP conversion. This is the same code
2981 optabs would emit if we didn't have TFmode patterns.
2982
2983 For SFmode, this is the only construction I've found that can pass
2984 gcc.c-torture/execute/ieee/rbug.c. No scenario that uses DFmode
2985 intermediates will work, because you'll get intermediate rounding
2986 that ruins the end result. Some of this could be fixed by turning
2987 on round-to-positive-infinity, but that requires diddling the fpsr,
2988 which kills performance. I tried turning this around and converting
2989 to a negative number, so that I could turn on /m, but either I did
2990 it wrong or there's something else cause I wound up with the exact
2991 same single-bit error. There is a branch-less form of this same code:
2992
2993 srl $16,1,$1
2994 and $16,1,$2
2995 cmplt $16,0,$3
2996 or $1,$2,$2
2997 cmovge $16,$16,$2
2998 itoft $3,$f10
2999 itoft $2,$f11
3000 cvtqs $f11,$f11
3001 adds $f11,$f11,$f0
3002 fcmoveq $f10,$f11,$f0
3003
3004 I'm not using it because it's the same number of instructions as
3005 this branch-full form, and it has more serialized long latency
3006 instructions on the critical path.
3007
3008 For DFmode, we can avoid rounding errors by breaking up the word
3009 into two pieces, converting them separately, and adding them back:
3010
3011 LC0: .long 0,0x5f800000
3012
3013 itoft $16,$f11
3014 lda $2,LC0
3015 cmplt $16,0,$1
3016 cpyse $f11,$f31,$f10
3017 cpyse $f31,$f11,$f11
3018 s4addq $1,$2,$1
3019 lds $f12,0($1)
3020 cvtqt $f10,$f10
3021 cvtqt $f11,$f11
3022 addt $f12,$f10,$f0
3023 addt $f0,$f11,$f0
3024
3025 This doesn't seem to be a clear-cut win over the optabs form.
3026 It probably all depends on the distribution of numbers being
3027 converted -- in the optabs form, all but high-bit-set has a
3028 much lower minimum execution time. */
3029
3030 void
3031 alpha_emit_floatuns (rtx operands[2])
3032 {
3033 rtx neglab, donelab, i0, i1, f0, in, out;
3034 enum machine_mode mode;
3035
3036 out = operands[0];
3037 in = force_reg (DImode, operands[1]);
3038 mode = GET_MODE (out);
3039 neglab = gen_label_rtx ();
3040 donelab = gen_label_rtx ();
3041 i0 = gen_reg_rtx (DImode);
3042 i1 = gen_reg_rtx (DImode);
3043 f0 = gen_reg_rtx (mode);
3044
3045 emit_cmp_and_jump_insns (in, const0_rtx, LT, const0_rtx, DImode, 0, neglab);
3046
3047 emit_insn (gen_rtx_SET (VOIDmode, out, gen_rtx_FLOAT (mode, in)));
3048 emit_jump_insn (gen_jump (donelab));
3049 emit_barrier ();
3050
3051 emit_label (neglab);
3052
3053 emit_insn (gen_lshrdi3 (i0, in, const1_rtx));
3054 emit_insn (gen_anddi3 (i1, in, const1_rtx));
3055 emit_insn (gen_iordi3 (i0, i0, i1));
3056 emit_insn (gen_rtx_SET (VOIDmode, f0, gen_rtx_FLOAT (mode, i0)));
3057 emit_insn (gen_rtx_SET (VOIDmode, out, gen_rtx_PLUS (mode, f0, f0)));
3058
3059 emit_label (donelab);
3060 }
3061
3062 /* Generate the comparison for a conditional branch. */
3063
3064 rtx
3065 alpha_emit_conditional_branch (enum rtx_code code)
3066 {
3067 enum rtx_code cmp_code, branch_code;
3068 enum machine_mode cmp_mode, branch_mode = VOIDmode;
3069 rtx op0 = alpha_compare.op0, op1 = alpha_compare.op1;
3070 rtx tem;
3071
3072 if (alpha_compare.fp_p && GET_MODE (op0) == TFmode)
3073 {
3074 if (! TARGET_HAS_XFLOATING_LIBS)
3075 abort ();
3076
3077 /* X_floating library comparison functions return
3078 -1 unordered
3079 0 false
3080 1 true
3081 Convert the compare against the raw return value. */
3082
3083 switch (code)
3084 {
3085 case UNORDERED:
3086 cmp_code = EQ;
3087 code = LT;
3088 break;
3089 case ORDERED:
3090 cmp_code = EQ;
3091 code = GE;
3092 break;
3093 case NE:
3094 cmp_code = NE;
3095 code = NE;
3096 break;
3097 default:
3098 cmp_code = code;
3099 code = GT;
3100 break;
3101 }
3102
3103 op0 = alpha_emit_xfloating_compare (cmp_code, op0, op1);
3104 op1 = const0_rtx;
3105 alpha_compare.fp_p = 0;
3106 }
3107
3108 /* The general case: fold the comparison code to the types of compares
3109 that we have, choosing the branch as necessary. */
3110 switch (code)
3111 {
3112 case EQ: case LE: case LT: case LEU: case LTU:
3113 case UNORDERED:
3114 /* We have these compares: */
3115 cmp_code = code, branch_code = NE;
3116 break;
3117
3118 case NE:
3119 case ORDERED:
3120 /* These must be reversed. */
3121 cmp_code = reverse_condition (code), branch_code = EQ;
3122 break;
3123
3124 case GE: case GT: case GEU: case GTU:
3125 /* For FP, we swap them, for INT, we reverse them. */
3126 if (alpha_compare.fp_p)
3127 {
3128 cmp_code = swap_condition (code);
3129 branch_code = NE;
3130 tem = op0, op0 = op1, op1 = tem;
3131 }
3132 else
3133 {
3134 cmp_code = reverse_condition (code);
3135 branch_code = EQ;
3136 }
3137 break;
3138
3139 default:
3140 abort ();
3141 }
3142
3143 if (alpha_compare.fp_p)
3144 {
3145 cmp_mode = DFmode;
3146 if (flag_unsafe_math_optimizations)
3147 {
3148 /* When we are not as concerned about non-finite values, and we
3149 are comparing against zero, we can branch directly. */
3150 if (op1 == CONST0_RTX (DFmode))
3151 cmp_code = NIL, branch_code = code;
3152 else if (op0 == CONST0_RTX (DFmode))
3153 {
3154 /* Undo the swap we probably did just above. */
3155 tem = op0, op0 = op1, op1 = tem;
3156 branch_code = swap_condition (cmp_code);
3157 cmp_code = NIL;
3158 }
3159 }
3160 else
3161 {
3162 /* ??? We mark the branch mode to be CCmode to prevent the
3163 compare and branch from being combined, since the compare
3164 insn follows IEEE rules that the branch does not. */
3165 branch_mode = CCmode;
3166 }
3167 }
3168 else
3169 {
3170 cmp_mode = DImode;
3171
3172 /* The following optimizations are only for signed compares. */
3173 if (code != LEU && code != LTU && code != GEU && code != GTU)
3174 {
3175 /* Whee. Compare and branch against 0 directly. */
3176 if (op1 == const0_rtx)
3177 cmp_code = NIL, branch_code = code;
3178
3179 /* If the constants doesn't fit into an immediate, but can
3180 be generated by lda/ldah, we adjust the argument and
3181 compare against zero, so we can use beq/bne directly. */
3182 else if (GET_CODE (op1) == CONST_INT && (code == EQ || code == NE))
3183 {
3184 HOST_WIDE_INT v = INTVAL (op1), n = -v;
3185
3186 if (! CONST_OK_FOR_LETTER_P (v, 'I')
3187 && (CONST_OK_FOR_LETTER_P (n, 'K')
3188 || CONST_OK_FOR_LETTER_P (n, 'L')))
3189 {
3190 cmp_code = PLUS, branch_code = code;
3191 op1 = GEN_INT (n);
3192 }
3193 }
3194 }
3195
3196 if (!reg_or_0_operand (op0, DImode))
3197 op0 = force_reg (DImode, op0);
3198 if (cmp_code != PLUS && !reg_or_8bit_operand (op1, DImode))
3199 op1 = force_reg (DImode, op1);
3200 }
3201
3202 /* Emit an initial compare instruction, if necessary. */
3203 tem = op0;
3204 if (cmp_code != NIL)
3205 {
3206 tem = gen_reg_rtx (cmp_mode);
3207 emit_move_insn (tem, gen_rtx_fmt_ee (cmp_code, cmp_mode, op0, op1));
3208 }
3209
3210 /* Zero the operands. */
3211 memset (&alpha_compare, 0, sizeof (alpha_compare));
3212
3213 /* Return the branch comparison. */
3214 return gen_rtx_fmt_ee (branch_code, branch_mode, tem, CONST0_RTX (cmp_mode));
3215 }
3216
3217 /* Certain simplifications can be done to make invalid setcc operations
3218 valid. Return the final comparison, or NULL if we can't work. */
3219
3220 rtx
3221 alpha_emit_setcc (enum rtx_code code)
3222 {
3223 enum rtx_code cmp_code;
3224 rtx op0 = alpha_compare.op0, op1 = alpha_compare.op1;
3225 int fp_p = alpha_compare.fp_p;
3226 rtx tmp;
3227
3228 /* Zero the operands. */
3229 memset (&alpha_compare, 0, sizeof (alpha_compare));
3230
3231 if (fp_p && GET_MODE (op0) == TFmode)
3232 {
3233 if (! TARGET_HAS_XFLOATING_LIBS)
3234 abort ();
3235
3236 /* X_floating library comparison functions return
3237 -1 unordered
3238 0 false
3239 1 true
3240 Convert the compare against the raw return value. */
3241
3242 if (code == UNORDERED || code == ORDERED)
3243 cmp_code = EQ;
3244 else
3245 cmp_code = code;
3246
3247 op0 = alpha_emit_xfloating_compare (cmp_code, op0, op1);
3248 op1 = const0_rtx;
3249 fp_p = 0;
3250
3251 if (code == UNORDERED)
3252 code = LT;
3253 else if (code == ORDERED)
3254 code = GE;
3255 else
3256 code = GT;
3257 }
3258
3259 if (fp_p && !TARGET_FIX)
3260 return NULL_RTX;
3261
3262 /* The general case: fold the comparison code to the types of compares
3263 that we have, choosing the branch as necessary. */
3264
3265 cmp_code = NIL;
3266 switch (code)
3267 {
3268 case EQ: case LE: case LT: case LEU: case LTU:
3269 case UNORDERED:
3270 /* We have these compares. */
3271 if (fp_p)
3272 cmp_code = code, code = NE;
3273 break;
3274
3275 case NE:
3276 if (!fp_p && op1 == const0_rtx)
3277 break;
3278 /* FALLTHRU */
3279
3280 case ORDERED:
3281 cmp_code = reverse_condition (code);
3282 code = EQ;
3283 break;
3284
3285 case GE: case GT: case GEU: case GTU:
3286 /* These normally need swapping, but for integer zero we have
3287 special patterns that recognize swapped operands. */
3288 if (!fp_p && op1 == const0_rtx)
3289 break;
3290 code = swap_condition (code);
3291 if (fp_p)
3292 cmp_code = code, code = NE;
3293 tmp = op0, op0 = op1, op1 = tmp;
3294 break;
3295
3296 default:
3297 abort ();
3298 }
3299
3300 if (!fp_p)
3301 {
3302 if (!register_operand (op0, DImode))
3303 op0 = force_reg (DImode, op0);
3304 if (!reg_or_8bit_operand (op1, DImode))
3305 op1 = force_reg (DImode, op1);
3306 }
3307
3308 /* Emit an initial compare instruction, if necessary. */
3309 if (cmp_code != NIL)
3310 {
3311 enum machine_mode mode = fp_p ? DFmode : DImode;
3312
3313 tmp = gen_reg_rtx (mode);
3314 emit_insn (gen_rtx_SET (VOIDmode, tmp,
3315 gen_rtx_fmt_ee (cmp_code, mode, op0, op1)));
3316
3317 op0 = fp_p ? gen_lowpart (DImode, tmp) : tmp;
3318 op1 = const0_rtx;
3319 }
3320
3321 /* Return the setcc comparison. */
3322 return gen_rtx_fmt_ee (code, DImode, op0, op1);
3323 }
3324
3325
3326 /* Rewrite a comparison against zero CMP of the form
3327 (CODE (cc0) (const_int 0)) so it can be written validly in
3328 a conditional move (if_then_else CMP ...).
3329 If both of the operands that set cc0 are nonzero we must emit
3330 an insn to perform the compare (it can't be done within
3331 the conditional move). */
3332
3333 rtx
3334 alpha_emit_conditional_move (rtx cmp, enum machine_mode mode)
3335 {
3336 enum rtx_code code = GET_CODE (cmp);
3337 enum rtx_code cmov_code = NE;
3338 rtx op0 = alpha_compare.op0;
3339 rtx op1 = alpha_compare.op1;
3340 int fp_p = alpha_compare.fp_p;
3341 enum machine_mode cmp_mode
3342 = (GET_MODE (op0) == VOIDmode ? DImode : GET_MODE (op0));
3343 enum machine_mode cmp_op_mode = fp_p ? DFmode : DImode;
3344 enum machine_mode cmov_mode = VOIDmode;
3345 int local_fast_math = flag_unsafe_math_optimizations;
3346 rtx tem;
3347
3348 /* Zero the operands. */
3349 memset (&alpha_compare, 0, sizeof (alpha_compare));
3350
3351 if (fp_p != FLOAT_MODE_P (mode))
3352 {
3353 enum rtx_code cmp_code;
3354
3355 if (! TARGET_FIX)
3356 return 0;
3357
3358 /* If we have fp<->int register move instructions, do a cmov by
3359 performing the comparison in fp registers, and move the
3360 zero/nonzero value to integer registers, where we can then
3361 use a normal cmov, or vice-versa. */
3362
3363 switch (code)
3364 {
3365 case EQ: case LE: case LT: case LEU: case LTU:
3366 /* We have these compares. */
3367 cmp_code = code, code = NE;
3368 break;
3369
3370 case NE:
3371 /* This must be reversed. */
3372 cmp_code = EQ, code = EQ;
3373 break;
3374
3375 case GE: case GT: case GEU: case GTU:
3376 /* These normally need swapping, but for integer zero we have
3377 special patterns that recognize swapped operands. */
3378 if (!fp_p && op1 == const0_rtx)
3379 cmp_code = code, code = NE;
3380 else
3381 {
3382 cmp_code = swap_condition (code);
3383 code = NE;
3384 tem = op0, op0 = op1, op1 = tem;
3385 }
3386 break;
3387
3388 default:
3389 abort ();
3390 }
3391
3392 tem = gen_reg_rtx (cmp_op_mode);
3393 emit_insn (gen_rtx_SET (VOIDmode, tem,
3394 gen_rtx_fmt_ee (cmp_code, cmp_op_mode,
3395 op0, op1)));
3396
3397 cmp_mode = cmp_op_mode = fp_p ? DImode : DFmode;
3398 op0 = gen_lowpart (cmp_op_mode, tem);
3399 op1 = CONST0_RTX (cmp_op_mode);
3400 fp_p = !fp_p;
3401 local_fast_math = 1;
3402 }
3403
3404 /* We may be able to use a conditional move directly.
3405 This avoids emitting spurious compares. */
3406 if (signed_comparison_operator (cmp, VOIDmode)
3407 && (!fp_p || local_fast_math)
3408 && (op0 == CONST0_RTX (cmp_mode) || op1 == CONST0_RTX (cmp_mode)))
3409 return gen_rtx_fmt_ee (code, VOIDmode, op0, op1);
3410
3411 /* We can't put the comparison inside the conditional move;
3412 emit a compare instruction and put that inside the
3413 conditional move. Make sure we emit only comparisons we have;
3414 swap or reverse as necessary. */
3415
3416 if (no_new_pseudos)
3417 return NULL_RTX;
3418
3419 switch (code)
3420 {
3421 case EQ: case LE: case LT: case LEU: case LTU:
3422 /* We have these compares: */
3423 break;
3424
3425 case NE:
3426 /* This must be reversed. */
3427 code = reverse_condition (code);
3428 cmov_code = EQ;
3429 break;
3430
3431 case GE: case GT: case GEU: case GTU:
3432 /* These must be swapped. */
3433 if (op1 != CONST0_RTX (cmp_mode))
3434 {
3435 code = swap_condition (code);
3436 tem = op0, op0 = op1, op1 = tem;
3437 }
3438 break;
3439
3440 default:
3441 abort ();
3442 }
3443
3444 if (!fp_p)
3445 {
3446 if (!reg_or_0_operand (op0, DImode))
3447 op0 = force_reg (DImode, op0);
3448 if (!reg_or_8bit_operand (op1, DImode))
3449 op1 = force_reg (DImode, op1);
3450 }
3451
3452 /* ??? We mark the branch mode to be CCmode to prevent the compare
3453 and cmov from being combined, since the compare insn follows IEEE
3454 rules that the cmov does not. */
3455 if (fp_p && !local_fast_math)
3456 cmov_mode = CCmode;
3457
3458 tem = gen_reg_rtx (cmp_op_mode);
3459 emit_move_insn (tem, gen_rtx_fmt_ee (code, cmp_op_mode, op0, op1));
3460 return gen_rtx_fmt_ee (cmov_code, cmov_mode, tem, CONST0_RTX (cmp_op_mode));
3461 }
3462
3463 /* Simplify a conditional move of two constants into a setcc with
3464 arithmetic. This is done with a splitter since combine would
3465 just undo the work if done during code generation. It also catches
3466 cases we wouldn't have before cse. */
3467
3468 int
3469 alpha_split_conditional_move (enum rtx_code code, rtx dest, rtx cond,
3470 rtx t_rtx, rtx f_rtx)
3471 {
3472 HOST_WIDE_INT t, f, diff;
3473 enum machine_mode mode;
3474 rtx target, subtarget, tmp;
3475
3476 mode = GET_MODE (dest);
3477 t = INTVAL (t_rtx);
3478 f = INTVAL (f_rtx);
3479 diff = t - f;
3480
3481 if (((code == NE || code == EQ) && diff < 0)
3482 || (code == GE || code == GT))
3483 {
3484 code = reverse_condition (code);
3485 diff = t, t = f, f = diff;
3486 diff = t - f;
3487 }
3488
3489 subtarget = target = dest;
3490 if (mode != DImode)
3491 {
3492 target = gen_lowpart (DImode, dest);
3493 if (! no_new_pseudos)
3494 subtarget = gen_reg_rtx (DImode);
3495 else
3496 subtarget = target;
3497 }
3498 /* Below, we must be careful to use copy_rtx on target and subtarget
3499 in intermediate insns, as they may be a subreg rtx, which may not
3500 be shared. */
3501
3502 if (f == 0 && exact_log2 (diff) > 0
3503 /* On EV6, we've got enough shifters to make non-arithmetic shifts
3504 viable over a longer latency cmove. On EV5, the E0 slot is a
3505 scarce resource, and on EV4 shift has the same latency as a cmove. */
3506 && (diff <= 8 || alpha_cpu == PROCESSOR_EV6))
3507 {
3508 tmp = gen_rtx_fmt_ee (code, DImode, cond, const0_rtx);
3509 emit_insn (gen_rtx_SET (VOIDmode, copy_rtx (subtarget), tmp));
3510
3511 tmp = gen_rtx_ASHIFT (DImode, copy_rtx (subtarget),
3512 GEN_INT (exact_log2 (t)));
3513 emit_insn (gen_rtx_SET (VOIDmode, target, tmp));
3514 }
3515 else if (f == 0 && t == -1)
3516 {
3517 tmp = gen_rtx_fmt_ee (code, DImode, cond, const0_rtx);
3518 emit_insn (gen_rtx_SET (VOIDmode, copy_rtx (subtarget), tmp));
3519
3520 emit_insn (gen_negdi2 (target, copy_rtx (subtarget)));
3521 }
3522 else if (diff == 1 || diff == 4 || diff == 8)
3523 {
3524 rtx add_op;
3525
3526 tmp = gen_rtx_fmt_ee (code, DImode, cond, const0_rtx);
3527 emit_insn (gen_rtx_SET (VOIDmode, copy_rtx (subtarget), tmp));
3528
3529 if (diff == 1)
3530 emit_insn (gen_adddi3 (target, copy_rtx (subtarget), GEN_INT (f)));
3531 else
3532 {
3533 add_op = GEN_INT (f);
3534 if (sext_add_operand (add_op, mode))
3535 {
3536 tmp = gen_rtx_MULT (DImode, copy_rtx (subtarget),
3537 GEN_INT (diff));
3538 tmp = gen_rtx_PLUS (DImode, tmp, add_op);
3539 emit_insn (gen_rtx_SET (VOIDmode, target, tmp));
3540 }
3541 else
3542 return 0;
3543 }
3544 }
3545 else
3546 return 0;
3547
3548 return 1;
3549 }
3550 \f
3551 /* Look up the function X_floating library function name for the
3552 given operation. */
3553
3554 static const char *
3555 alpha_lookup_xfloating_lib_func (enum rtx_code code)
3556 {
3557 struct xfloating_op
3558 {
3559 const enum rtx_code code;
3560 const char *const func;
3561 };
3562
3563 static const struct xfloating_op vms_xfloating_ops[] =
3564 {
3565 { PLUS, "OTS$ADD_X" },
3566 { MINUS, "OTS$SUB_X" },
3567 { MULT, "OTS$MUL_X" },
3568 { DIV, "OTS$DIV_X" },
3569 { EQ, "OTS$EQL_X" },
3570 { NE, "OTS$NEQ_X" },
3571 { LT, "OTS$LSS_X" },
3572 { LE, "OTS$LEQ_X" },
3573 { GT, "OTS$GTR_X" },
3574 { GE, "OTS$GEQ_X" },
3575 { FIX, "OTS$CVTXQ" },
3576 { FLOAT, "OTS$CVTQX" },
3577 { UNSIGNED_FLOAT, "OTS$CVTQUX" },
3578 { FLOAT_EXTEND, "OTS$CVT_FLOAT_T_X" },
3579 { FLOAT_TRUNCATE, "OTS$CVT_FLOAT_X_T" },
3580 };
3581
3582 static const struct xfloating_op osf_xfloating_ops[] =
3583 {
3584 { PLUS, "_OtsAddX" },
3585 { MINUS, "_OtsSubX" },
3586 { MULT, "_OtsMulX" },
3587 { DIV, "_OtsDivX" },
3588 { EQ, "_OtsEqlX" },
3589 { NE, "_OtsNeqX" },
3590 { LT, "_OtsLssX" },
3591 { LE, "_OtsLeqX" },
3592 { GT, "_OtsGtrX" },
3593 { GE, "_OtsGeqX" },
3594 { FIX, "_OtsCvtXQ" },
3595 { FLOAT, "_OtsCvtQX" },
3596 { UNSIGNED_FLOAT, "_OtsCvtQUX" },
3597 { FLOAT_EXTEND, "_OtsConvertFloatTX" },
3598 { FLOAT_TRUNCATE, "_OtsConvertFloatXT" },
3599 };
3600
3601 const struct xfloating_op *ops;
3602 const long n = ARRAY_SIZE (osf_xfloating_ops);
3603 long i;
3604
3605 /* How irritating. Nothing to key off for the table. Hardcode
3606 knowledge of the G_floating routines. */
3607 if (TARGET_FLOAT_VAX)
3608 {
3609 if (TARGET_ABI_OPEN_VMS)
3610 {
3611 if (code == FLOAT_EXTEND)
3612 return "OTS$CVT_FLOAT_G_X";
3613 if (code == FLOAT_TRUNCATE)
3614 return "OTS$CVT_FLOAT_X_G";
3615 }
3616 else
3617 {
3618 if (code == FLOAT_EXTEND)
3619 return "_OtsConvertFloatGX";
3620 if (code == FLOAT_TRUNCATE)
3621 return "_OtsConvertFloatXG";
3622 }
3623 }
3624
3625 if (TARGET_ABI_OPEN_VMS)
3626 ops = vms_xfloating_ops;
3627 else
3628 ops = osf_xfloating_ops;
3629
3630 for (i = 0; i < n; ++i)
3631 if (ops[i].code == code)
3632 return ops[i].func;
3633
3634 abort();
3635 }
3636
3637 /* Most X_floating operations take the rounding mode as an argument.
3638 Compute that here. */
3639
3640 static int
3641 alpha_compute_xfloating_mode_arg (enum rtx_code code,
3642 enum alpha_fp_rounding_mode round)
3643 {
3644 int mode;
3645
3646 switch (round)
3647 {
3648 case ALPHA_FPRM_NORM:
3649 mode = 2;
3650 break;
3651 case ALPHA_FPRM_MINF:
3652 mode = 1;
3653 break;
3654 case ALPHA_FPRM_CHOP:
3655 mode = 0;
3656 break;
3657 case ALPHA_FPRM_DYN:
3658 mode = 4;
3659 break;
3660 default:
3661 abort ();
3662
3663 /* XXX For reference, round to +inf is mode = 3. */
3664 }
3665
3666 if (code == FLOAT_TRUNCATE && alpha_fptm == ALPHA_FPTM_N)
3667 mode |= 0x10000;
3668
3669 return mode;
3670 }
3671
3672 /* Emit an X_floating library function call.
3673
3674 Note that these functions do not follow normal calling conventions:
3675 TFmode arguments are passed in two integer registers (as opposed to
3676 indirect); TFmode return values appear in R16+R17.
3677
3678 FUNC is the function name to call.
3679 TARGET is where the output belongs.
3680 OPERANDS are the inputs.
3681 NOPERANDS is the count of inputs.
3682 EQUIV is the expression equivalent for the function.
3683 */
3684
3685 static void
3686 alpha_emit_xfloating_libcall (const char *func, rtx target, rtx operands[],
3687 int noperands, rtx equiv)
3688 {
3689 rtx usage = NULL_RTX, tmp, reg;
3690 int regno = 16, i;
3691
3692 start_sequence ();
3693
3694 for (i = 0; i < noperands; ++i)
3695 {
3696 switch (GET_MODE (operands[i]))
3697 {
3698 case TFmode:
3699 reg = gen_rtx_REG (TFmode, regno);
3700 regno += 2;
3701 break;
3702
3703 case DFmode:
3704 reg = gen_rtx_REG (DFmode, regno + 32);
3705 regno += 1;
3706 break;
3707
3708 case VOIDmode:
3709 if (GET_CODE (operands[i]) != CONST_INT)
3710 abort ();
3711 /* FALLTHRU */
3712 case DImode:
3713 reg = gen_rtx_REG (DImode, regno);
3714 regno += 1;
3715 break;
3716
3717 default:
3718 abort ();
3719 }
3720
3721 emit_move_insn (reg, operands[i]);
3722 usage = alloc_EXPR_LIST (0, gen_rtx_USE (VOIDmode, reg), usage);
3723 }
3724
3725 switch (GET_MODE (target))
3726 {
3727 case TFmode:
3728 reg = gen_rtx_REG (TFmode, 16);
3729 break;
3730 case DFmode:
3731 reg = gen_rtx_REG (DFmode, 32);
3732 break;
3733 case DImode:
3734 reg = gen_rtx_REG (DImode, 0);
3735 break;
3736 default:
3737 abort ();
3738 }
3739
3740 tmp = gen_rtx_MEM (QImode, init_one_libfunc (func));
3741 tmp = emit_call_insn (GEN_CALL_VALUE (reg, tmp, const0_rtx,
3742 const0_rtx, const0_rtx));
3743 CALL_INSN_FUNCTION_USAGE (tmp) = usage;
3744
3745 tmp = get_insns ();
3746 end_sequence ();
3747
3748 emit_libcall_block (tmp, target, reg, equiv);
3749 }
3750
3751 /* Emit an X_floating library function call for arithmetic (+,-,*,/). */
3752
3753 void
3754 alpha_emit_xfloating_arith (enum rtx_code code, rtx operands[])
3755 {
3756 const char *func;
3757 int mode;
3758 rtx out_operands[3];
3759
3760 func = alpha_lookup_xfloating_lib_func (code);
3761 mode = alpha_compute_xfloating_mode_arg (code, alpha_fprm);
3762
3763 out_operands[0] = operands[1];
3764 out_operands[1] = operands[2];
3765 out_operands[2] = GEN_INT (mode);
3766 alpha_emit_xfloating_libcall (func, operands[0], out_operands, 3,
3767 gen_rtx_fmt_ee (code, TFmode, operands[1],
3768 operands[2]));
3769 }
3770
3771 /* Emit an X_floating library function call for a comparison. */
3772
3773 static rtx
3774 alpha_emit_xfloating_compare (enum rtx_code code, rtx op0, rtx op1)
3775 {
3776 const char *func;
3777 rtx out, operands[2];
3778
3779 func = alpha_lookup_xfloating_lib_func (code);
3780
3781 operands[0] = op0;
3782 operands[1] = op1;
3783 out = gen_reg_rtx (DImode);
3784
3785 /* ??? Strange mode for equiv because what's actually returned
3786 is -1,0,1, not a proper boolean value. */
3787 alpha_emit_xfloating_libcall (func, out, operands, 2,
3788 gen_rtx_fmt_ee (code, CCmode, op0, op1));
3789
3790 return out;
3791 }
3792
3793 /* Emit an X_floating library function call for a conversion. */
3794
3795 void
3796 alpha_emit_xfloating_cvt (enum rtx_code orig_code, rtx operands[])
3797 {
3798 int noperands = 1, mode;
3799 rtx out_operands[2];
3800 const char *func;
3801 enum rtx_code code = orig_code;
3802
3803 if (code == UNSIGNED_FIX)
3804 code = FIX;
3805
3806 func = alpha_lookup_xfloating_lib_func (code);
3807
3808 out_operands[0] = operands[1];
3809
3810 switch (code)
3811 {
3812 case FIX:
3813 mode = alpha_compute_xfloating_mode_arg (code, ALPHA_FPRM_CHOP);
3814 out_operands[1] = GEN_INT (mode);
3815 noperands = 2;
3816 break;
3817 case FLOAT_TRUNCATE:
3818 mode = alpha_compute_xfloating_mode_arg (code, alpha_fprm);
3819 out_operands[1] = GEN_INT (mode);
3820 noperands = 2;
3821 break;
3822 default:
3823 break;
3824 }
3825
3826 alpha_emit_xfloating_libcall (func, operands[0], out_operands, noperands,
3827 gen_rtx_fmt_e (orig_code,
3828 GET_MODE (operands[0]),
3829 operands[1]));
3830 }
3831
3832 /* Split a TFmode OP[1] into DImode OP[2,3] and likewise for
3833 OP[0] into OP[0,1]. Naturally, output operand ordering is
3834 little-endian. */
3835
3836 void
3837 alpha_split_tfmode_pair (rtx operands[4])
3838 {
3839 if (GET_CODE (operands[1]) == REG)
3840 {
3841 operands[3] = gen_rtx_REG (DImode, REGNO (operands[1]) + 1);
3842 operands[2] = gen_rtx_REG (DImode, REGNO (operands[1]));
3843 }
3844 else if (GET_CODE (operands[1]) == MEM)
3845 {
3846 operands[3] = adjust_address (operands[1], DImode, 8);
3847 operands[2] = adjust_address (operands[1], DImode, 0);
3848 }
3849 else if (operands[1] == CONST0_RTX (TFmode))
3850 operands[2] = operands[3] = const0_rtx;
3851 else
3852 abort ();
3853
3854 if (GET_CODE (operands[0]) == REG)
3855 {
3856 operands[1] = gen_rtx_REG (DImode, REGNO (operands[0]) + 1);
3857 operands[0] = gen_rtx_REG (DImode, REGNO (operands[0]));
3858 }
3859 else if (GET_CODE (operands[0]) == MEM)
3860 {
3861 operands[1] = adjust_address (operands[0], DImode, 8);
3862 operands[0] = adjust_address (operands[0], DImode, 0);
3863 }
3864 else
3865 abort ();
3866 }
3867
3868 /* Implement negtf2 or abstf2. Op0 is destination, op1 is source,
3869 op2 is a register containing the sign bit, operation is the
3870 logical operation to be performed. */
3871
3872 void
3873 alpha_split_tfmode_frobsign (rtx operands[3], rtx (*operation) (rtx, rtx, rtx))
3874 {
3875 rtx high_bit = operands[2];
3876 rtx scratch;
3877 int move;
3878
3879 alpha_split_tfmode_pair (operands);
3880
3881 /* Detect three flavors of operand overlap. */
3882 move = 1;
3883 if (rtx_equal_p (operands[0], operands[2]))
3884 move = 0;
3885 else if (rtx_equal_p (operands[1], operands[2]))
3886 {
3887 if (rtx_equal_p (operands[0], high_bit))
3888 move = 2;
3889 else
3890 move = -1;
3891 }
3892
3893 if (move < 0)
3894 emit_move_insn (operands[0], operands[2]);
3895
3896 /* ??? If the destination overlaps both source tf and high_bit, then
3897 assume source tf is dead in its entirety and use the other half
3898 for a scratch register. Otherwise "scratch" is just the proper
3899 destination register. */
3900 scratch = operands[move < 2 ? 1 : 3];
3901
3902 emit_insn ((*operation) (scratch, high_bit, operands[3]));
3903
3904 if (move > 0)
3905 {
3906 emit_move_insn (operands[0], operands[2]);
3907 if (move > 1)
3908 emit_move_insn (operands[1], scratch);
3909 }
3910 }
3911 \f
3912 /* Use ext[wlq][lh] as the Architecture Handbook describes for extracting
3913 unaligned data:
3914
3915 unsigned: signed:
3916 word: ldq_u r1,X(r11) ldq_u r1,X(r11)
3917 ldq_u r2,X+1(r11) ldq_u r2,X+1(r11)
3918 lda r3,X(r11) lda r3,X+2(r11)
3919 extwl r1,r3,r1 extql r1,r3,r1
3920 extwh r2,r3,r2 extqh r2,r3,r2
3921 or r1.r2.r1 or r1,r2,r1
3922 sra r1,48,r1
3923
3924 long: ldq_u r1,X(r11) ldq_u r1,X(r11)
3925 ldq_u r2,X+3(r11) ldq_u r2,X+3(r11)
3926 lda r3,X(r11) lda r3,X(r11)
3927 extll r1,r3,r1 extll r1,r3,r1
3928 extlh r2,r3,r2 extlh r2,r3,r2
3929 or r1.r2.r1 addl r1,r2,r1
3930
3931 quad: ldq_u r1,X(r11)
3932 ldq_u r2,X+7(r11)
3933 lda r3,X(r11)
3934 extql r1,r3,r1
3935 extqh r2,r3,r2
3936 or r1.r2.r1
3937 */
3938
3939 void
3940 alpha_expand_unaligned_load (rtx tgt, rtx mem, HOST_WIDE_INT size,
3941 HOST_WIDE_INT ofs, int sign)
3942 {
3943 rtx meml, memh, addr, extl, exth, tmp, mema;
3944 enum machine_mode mode;
3945
3946 meml = gen_reg_rtx (DImode);
3947 memh = gen_reg_rtx (DImode);
3948 addr = gen_reg_rtx (DImode);
3949 extl = gen_reg_rtx (DImode);
3950 exth = gen_reg_rtx (DImode);
3951
3952 mema = XEXP (mem, 0);
3953 if (GET_CODE (mema) == LO_SUM)
3954 mema = force_reg (Pmode, mema);
3955
3956 /* AND addresses cannot be in any alias set, since they may implicitly
3957 alias surrounding code. Ideally we'd have some alias set that
3958 covered all types except those with alignment 8 or higher. */
3959
3960 tmp = change_address (mem, DImode,
3961 gen_rtx_AND (DImode,
3962 plus_constant (mema, ofs),
3963 GEN_INT (-8)));
3964 set_mem_alias_set (tmp, 0);
3965 emit_move_insn (meml, tmp);
3966
3967 tmp = change_address (mem, DImode,
3968 gen_rtx_AND (DImode,
3969 plus_constant (mema, ofs + size - 1),
3970 GEN_INT (-8)));
3971 set_mem_alias_set (tmp, 0);
3972 emit_move_insn (memh, tmp);
3973
3974 if (WORDS_BIG_ENDIAN && sign && (size == 2 || size == 4))
3975 {
3976 emit_move_insn (addr, plus_constant (mema, -1));
3977
3978 emit_insn (gen_extqh_be (extl, meml, addr));
3979 emit_insn (gen_extxl_be (exth, memh, GEN_INT (64), addr));
3980
3981 addr = expand_binop (DImode, ior_optab, extl, exth, tgt, 1, OPTAB_WIDEN);
3982 addr = expand_binop (DImode, ashr_optab, addr, GEN_INT (64 - size*8),
3983 addr, 1, OPTAB_WIDEN);
3984 }
3985 else if (sign && size == 2)
3986 {
3987 emit_move_insn (addr, plus_constant (mema, ofs+2));
3988
3989 emit_insn (gen_extxl_le (extl, meml, GEN_INT (64), addr));
3990 emit_insn (gen_extqh_le (exth, memh, addr));
3991
3992 /* We must use tgt here for the target. Alpha-vms port fails if we use
3993 addr for the target, because addr is marked as a pointer and combine
3994 knows that pointers are always sign-extended 32 bit values. */
3995 addr = expand_binop (DImode, ior_optab, extl, exth, tgt, 1, OPTAB_WIDEN);
3996 addr = expand_binop (DImode, ashr_optab, addr, GEN_INT (48),
3997 addr, 1, OPTAB_WIDEN);
3998 }
3999 else
4000 {
4001 if (WORDS_BIG_ENDIAN)
4002 {
4003 emit_move_insn (addr, plus_constant (mema, ofs+size-1));
4004 switch ((int) size)
4005 {
4006 case 2:
4007 emit_insn (gen_extwh_be (extl, meml, addr));
4008 mode = HImode;
4009 break;
4010
4011 case 4:
4012 emit_insn (gen_extlh_be (extl, meml, addr));
4013 mode = SImode;
4014 break;
4015
4016 case 8:
4017 emit_insn (gen_extqh_be (extl, meml, addr));
4018 mode = DImode;
4019 break;
4020
4021 default:
4022 abort ();
4023 }
4024 emit_insn (gen_extxl_be (exth, memh, GEN_INT (size*8), addr));
4025 }
4026 else
4027 {
4028 emit_move_insn (addr, plus_constant (mema, ofs));
4029 emit_insn (gen_extxl_le (extl, meml, GEN_INT (size*8), addr));
4030 switch ((int) size)
4031 {
4032 case 2:
4033 emit_insn (gen_extwh_le (exth, memh, addr));
4034 mode = HImode;
4035 break;
4036
4037 case 4:
4038 emit_insn (gen_extlh_le (exth, memh, addr));
4039 mode = SImode;
4040 break;
4041
4042 case 8:
4043 emit_insn (gen_extqh_le (exth, memh, addr));
4044 mode = DImode;
4045 break;
4046
4047 default:
4048 abort();
4049 }
4050 }
4051
4052 addr = expand_binop (mode, ior_optab, gen_lowpart (mode, extl),
4053 gen_lowpart (mode, exth), gen_lowpart (mode, tgt),
4054 sign, OPTAB_WIDEN);
4055 }
4056
4057 if (addr != tgt)
4058 emit_move_insn (tgt, gen_lowpart(GET_MODE (tgt), addr));
4059 }
4060
4061 /* Similarly, use ins and msk instructions to perform unaligned stores. */
4062
4063 void
4064 alpha_expand_unaligned_store (rtx dst, rtx src,
4065 HOST_WIDE_INT size, HOST_WIDE_INT ofs)
4066 {
4067 rtx dstl, dsth, addr, insl, insh, meml, memh, dsta;
4068
4069 dstl = gen_reg_rtx (DImode);
4070 dsth = gen_reg_rtx (DImode);
4071 insl = gen_reg_rtx (DImode);
4072 insh = gen_reg_rtx (DImode);
4073
4074 dsta = XEXP (dst, 0);
4075 if (GET_CODE (dsta) == LO_SUM)
4076 dsta = force_reg (Pmode, dsta);
4077
4078 /* AND addresses cannot be in any alias set, since they may implicitly
4079 alias surrounding code. Ideally we'd have some alias set that
4080 covered all types except those with alignment 8 or higher. */
4081
4082 meml = change_address (dst, DImode,
4083 gen_rtx_AND (DImode,
4084 plus_constant (dsta, ofs),
4085 GEN_INT (-8)));
4086 set_mem_alias_set (meml, 0);
4087
4088 memh = change_address (dst, DImode,
4089 gen_rtx_AND (DImode,
4090 plus_constant (dsta, ofs + size - 1),
4091 GEN_INT (-8)));
4092 set_mem_alias_set (memh, 0);
4093
4094 emit_move_insn (dsth, memh);
4095 emit_move_insn (dstl, meml);
4096 if (WORDS_BIG_ENDIAN)
4097 {
4098 addr = copy_addr_to_reg (plus_constant (dsta, ofs+size-1));
4099
4100 if (src != const0_rtx)
4101 {
4102 switch ((int) size)
4103 {
4104 case 2:
4105 emit_insn (gen_inswl_be (insh, gen_lowpart (HImode,src), addr));
4106 break;
4107 case 4:
4108 emit_insn (gen_insll_be (insh, gen_lowpart (SImode,src), addr));
4109 break;
4110 case 8:
4111 emit_insn (gen_insql_be (insh, gen_lowpart (DImode,src), addr));
4112 break;
4113 }
4114 emit_insn (gen_insxh (insl, gen_lowpart (DImode, src),
4115 GEN_INT (size*8), addr));
4116 }
4117
4118 switch ((int) size)
4119 {
4120 case 2:
4121 emit_insn (gen_mskxl_be (dsth, dsth, GEN_INT (0xffff), addr));
4122 break;
4123 case 4:
4124 {
4125 rtx msk = immed_double_const (0xffffffff, 0, DImode);
4126 emit_insn (gen_mskxl_be (dsth, dsth, msk, addr));
4127 break;
4128 }
4129 case 8:
4130 emit_insn (gen_mskxl_be (dsth, dsth, constm1_rtx, addr));
4131 break;
4132 }
4133
4134 emit_insn (gen_mskxh (dstl, dstl, GEN_INT (size*8), addr));
4135 }
4136 else
4137 {
4138 addr = copy_addr_to_reg (plus_constant (dsta, ofs));
4139
4140 if (src != const0_rtx)
4141 {
4142 emit_insn (gen_insxh (insh, gen_lowpart (DImode, src),
4143 GEN_INT (size*8), addr));
4144
4145 switch ((int) size)
4146 {
4147 case 2:
4148 emit_insn (gen_inswl_le (insl, gen_lowpart (HImode, src), addr));
4149 break;
4150 case 4:
4151 emit_insn (gen_insll_le (insl, gen_lowpart (SImode, src), addr));
4152 break;
4153 case 8:
4154 emit_insn (gen_insql_le (insl, src, addr));
4155 break;
4156 }
4157 }
4158
4159 emit_insn (gen_mskxh (dsth, dsth, GEN_INT (size*8), addr));
4160
4161 switch ((int) size)
4162 {
4163 case 2:
4164 emit_insn (gen_mskxl_le (dstl, dstl, GEN_INT (0xffff), addr));
4165 break;
4166 case 4:
4167 {
4168 rtx msk = immed_double_const (0xffffffff, 0, DImode);
4169 emit_insn (gen_mskxl_le (dstl, dstl, msk, addr));
4170 break;
4171 }
4172 case 8:
4173 emit_insn (gen_mskxl_le (dstl, dstl, constm1_rtx, addr));
4174 break;
4175 }
4176 }
4177
4178 if (src != const0_rtx)
4179 {
4180 dsth = expand_binop (DImode, ior_optab, insh, dsth, dsth, 0, OPTAB_WIDEN);
4181 dstl = expand_binop (DImode, ior_optab, insl, dstl, dstl, 0, OPTAB_WIDEN);
4182 }
4183
4184 if (WORDS_BIG_ENDIAN)
4185 {
4186 emit_move_insn (meml, dstl);
4187 emit_move_insn (memh, dsth);
4188 }
4189 else
4190 {
4191 /* Must store high before low for degenerate case of aligned. */
4192 emit_move_insn (memh, dsth);
4193 emit_move_insn (meml, dstl);
4194 }
4195 }
4196
4197 /* The block move code tries to maximize speed by separating loads and
4198 stores at the expense of register pressure: we load all of the data
4199 before we store it back out. There are two secondary effects worth
4200 mentioning, that this speeds copying to/from aligned and unaligned
4201 buffers, and that it makes the code significantly easier to write. */
4202
4203 #define MAX_MOVE_WORDS 8
4204
4205 /* Load an integral number of consecutive unaligned quadwords. */
4206
4207 static void
4208 alpha_expand_unaligned_load_words (rtx *out_regs, rtx smem,
4209 HOST_WIDE_INT words, HOST_WIDE_INT ofs)
4210 {
4211 rtx const im8 = GEN_INT (-8);
4212 rtx const i64 = GEN_INT (64);
4213 rtx ext_tmps[MAX_MOVE_WORDS], data_regs[MAX_MOVE_WORDS+1];
4214 rtx sreg, areg, tmp, smema;
4215 HOST_WIDE_INT i;
4216
4217 smema = XEXP (smem, 0);
4218 if (GET_CODE (smema) == LO_SUM)
4219 smema = force_reg (Pmode, smema);
4220
4221 /* Generate all the tmp registers we need. */
4222 for (i = 0; i < words; ++i)
4223 {
4224 data_regs[i] = out_regs[i];
4225 ext_tmps[i] = gen_reg_rtx (DImode);
4226 }
4227 data_regs[words] = gen_reg_rtx (DImode);
4228
4229 if (ofs != 0)
4230 smem = adjust_address (smem, GET_MODE (smem), ofs);
4231
4232 /* Load up all of the source data. */
4233 for (i = 0; i < words; ++i)
4234 {
4235 tmp = change_address (smem, DImode,
4236 gen_rtx_AND (DImode,
4237 plus_constant (smema, 8*i),
4238 im8));
4239 set_mem_alias_set (tmp, 0);
4240 emit_move_insn (data_regs[i], tmp);
4241 }
4242
4243 tmp = change_address (smem, DImode,
4244 gen_rtx_AND (DImode,
4245 plus_constant (smema, 8*words - 1),
4246 im8));
4247 set_mem_alias_set (tmp, 0);
4248 emit_move_insn (data_regs[words], tmp);
4249
4250 /* Extract the half-word fragments. Unfortunately DEC decided to make
4251 extxh with offset zero a noop instead of zeroing the register, so
4252 we must take care of that edge condition ourselves with cmov. */
4253
4254 sreg = copy_addr_to_reg (smema);
4255 areg = expand_binop (DImode, and_optab, sreg, GEN_INT (7), NULL,
4256 1, OPTAB_WIDEN);
4257 if (WORDS_BIG_ENDIAN)
4258 emit_move_insn (sreg, plus_constant (sreg, 7));
4259 for (i = 0; i < words; ++i)
4260 {
4261 if (WORDS_BIG_ENDIAN)
4262 {
4263 emit_insn (gen_extqh_be (data_regs[i], data_regs[i], sreg));
4264 emit_insn (gen_extxl_be (ext_tmps[i], data_regs[i+1], i64, sreg));
4265 }
4266 else
4267 {
4268 emit_insn (gen_extxl_le (data_regs[i], data_regs[i], i64, sreg));
4269 emit_insn (gen_extqh_le (ext_tmps[i], data_regs[i+1], sreg));
4270 }
4271 emit_insn (gen_rtx_SET (VOIDmode, ext_tmps[i],
4272 gen_rtx_IF_THEN_ELSE (DImode,
4273 gen_rtx_EQ (DImode, areg,
4274 const0_rtx),
4275 const0_rtx, ext_tmps[i])));
4276 }
4277
4278 /* Merge the half-words into whole words. */
4279 for (i = 0; i < words; ++i)
4280 {
4281 out_regs[i] = expand_binop (DImode, ior_optab, data_regs[i],
4282 ext_tmps[i], data_regs[i], 1, OPTAB_WIDEN);
4283 }
4284 }
4285
4286 /* Store an integral number of consecutive unaligned quadwords. DATA_REGS
4287 may be NULL to store zeros. */
4288
4289 static void
4290 alpha_expand_unaligned_store_words (rtx *data_regs, rtx dmem,
4291 HOST_WIDE_INT words, HOST_WIDE_INT ofs)
4292 {
4293 rtx const im8 = GEN_INT (-8);
4294 rtx const i64 = GEN_INT (64);
4295 rtx ins_tmps[MAX_MOVE_WORDS];
4296 rtx st_tmp_1, st_tmp_2, dreg;
4297 rtx st_addr_1, st_addr_2, dmema;
4298 HOST_WIDE_INT i;
4299
4300 dmema = XEXP (dmem, 0);
4301 if (GET_CODE (dmema) == LO_SUM)
4302 dmema = force_reg (Pmode, dmema);
4303
4304 /* Generate all the tmp registers we need. */
4305 if (data_regs != NULL)
4306 for (i = 0; i < words; ++i)
4307 ins_tmps[i] = gen_reg_rtx(DImode);
4308 st_tmp_1 = gen_reg_rtx(DImode);
4309 st_tmp_2 = gen_reg_rtx(DImode);
4310
4311 if (ofs != 0)
4312 dmem = adjust_address (dmem, GET_MODE (dmem), ofs);
4313
4314 st_addr_2 = change_address (dmem, DImode,
4315 gen_rtx_AND (DImode,
4316 plus_constant (dmema, words*8 - 1),
4317 im8));
4318 set_mem_alias_set (st_addr_2, 0);
4319
4320 st_addr_1 = change_address (dmem, DImode,
4321 gen_rtx_AND (DImode, dmema, im8));
4322 set_mem_alias_set (st_addr_1, 0);
4323
4324 /* Load up the destination end bits. */
4325 emit_move_insn (st_tmp_2, st_addr_2);
4326 emit_move_insn (st_tmp_1, st_addr_1);
4327
4328 /* Shift the input data into place. */
4329 dreg = copy_addr_to_reg (dmema);
4330 if (WORDS_BIG_ENDIAN)
4331 emit_move_insn (dreg, plus_constant (dreg, 7));
4332 if (data_regs != NULL)
4333 {
4334 for (i = words-1; i >= 0; --i)
4335 {
4336 if (WORDS_BIG_ENDIAN)
4337 {
4338 emit_insn (gen_insql_be (ins_tmps[i], data_regs[i], dreg));
4339 emit_insn (gen_insxh (data_regs[i], data_regs[i], i64, dreg));
4340 }
4341 else
4342 {
4343 emit_insn (gen_insxh (ins_tmps[i], data_regs[i], i64, dreg));
4344 emit_insn (gen_insql_le (data_regs[i], data_regs[i], dreg));
4345 }
4346 }
4347 for (i = words-1; i > 0; --i)
4348 {
4349 ins_tmps[i-1] = expand_binop (DImode, ior_optab, data_regs[i],
4350 ins_tmps[i-1], ins_tmps[i-1], 1,
4351 OPTAB_WIDEN);
4352 }
4353 }
4354
4355 /* Split and merge the ends with the destination data. */
4356 if (WORDS_BIG_ENDIAN)
4357 {
4358 emit_insn (gen_mskxl_be (st_tmp_2, st_tmp_2, constm1_rtx, dreg));
4359 emit_insn (gen_mskxh (st_tmp_1, st_tmp_1, i64, dreg));
4360 }
4361 else
4362 {
4363 emit_insn (gen_mskxh (st_tmp_2, st_tmp_2, i64, dreg));
4364 emit_insn (gen_mskxl_le (st_tmp_1, st_tmp_1, constm1_rtx, dreg));
4365 }
4366
4367 if (data_regs != NULL)
4368 {
4369 st_tmp_2 = expand_binop (DImode, ior_optab, st_tmp_2, ins_tmps[words-1],
4370 st_tmp_2, 1, OPTAB_WIDEN);
4371 st_tmp_1 = expand_binop (DImode, ior_optab, st_tmp_1, data_regs[0],
4372 st_tmp_1, 1, OPTAB_WIDEN);
4373 }
4374
4375 /* Store it all. */
4376 if (WORDS_BIG_ENDIAN)
4377 emit_move_insn (st_addr_1, st_tmp_1);
4378 else
4379 emit_move_insn (st_addr_2, st_tmp_2);
4380 for (i = words-1; i > 0; --i)
4381 {
4382 rtx tmp = change_address (dmem, DImode,
4383 gen_rtx_AND (DImode,
4384 plus_constant(dmema,
4385 WORDS_BIG_ENDIAN ? i*8-1 : i*8),
4386 im8));
4387 set_mem_alias_set (tmp, 0);
4388 emit_move_insn (tmp, data_regs ? ins_tmps[i-1] : const0_rtx);
4389 }
4390 if (WORDS_BIG_ENDIAN)
4391 emit_move_insn (st_addr_2, st_tmp_2);
4392 else
4393 emit_move_insn (st_addr_1, st_tmp_1);
4394 }
4395
4396
4397 /* Expand string/block move operations.
4398
4399 operands[0] is the pointer to the destination.
4400 operands[1] is the pointer to the source.
4401 operands[2] is the number of bytes to move.
4402 operands[3] is the alignment. */
4403
4404 int
4405 alpha_expand_block_move (rtx operands[])
4406 {
4407 rtx bytes_rtx = operands[2];
4408 rtx align_rtx = operands[3];
4409 HOST_WIDE_INT orig_bytes = INTVAL (bytes_rtx);
4410 HOST_WIDE_INT bytes = orig_bytes;
4411 HOST_WIDE_INT src_align = INTVAL (align_rtx) * BITS_PER_UNIT;
4412 HOST_WIDE_INT dst_align = src_align;
4413 rtx orig_src = operands[1];
4414 rtx orig_dst = operands[0];
4415 rtx data_regs[2 * MAX_MOVE_WORDS + 16];
4416 rtx tmp;
4417 unsigned int i, words, ofs, nregs = 0;
4418
4419 if (orig_bytes <= 0)
4420 return 1;
4421 else if (orig_bytes > MAX_MOVE_WORDS * UNITS_PER_WORD)
4422 return 0;
4423
4424 /* Look for additional alignment information from recorded register info. */
4425
4426 tmp = XEXP (orig_src, 0);
4427 if (GET_CODE (tmp) == REG)
4428 src_align = MAX (src_align, REGNO_POINTER_ALIGN (REGNO (tmp)));
4429 else if (GET_CODE (tmp) == PLUS
4430 && GET_CODE (XEXP (tmp, 0)) == REG
4431 && GET_CODE (XEXP (tmp, 1)) == CONST_INT)
4432 {
4433 unsigned HOST_WIDE_INT c = INTVAL (XEXP (tmp, 1));
4434 unsigned int a = REGNO_POINTER_ALIGN (REGNO (XEXP (tmp, 0)));
4435
4436 if (a > src_align)
4437 {
4438 if (a >= 64 && c % 8 == 0)
4439 src_align = 64;
4440 else if (a >= 32 && c % 4 == 0)
4441 src_align = 32;
4442 else if (a >= 16 && c % 2 == 0)
4443 src_align = 16;
4444 }
4445 }
4446
4447 tmp = XEXP (orig_dst, 0);
4448 if (GET_CODE (tmp) == REG)
4449 dst_align = MAX (dst_align, REGNO_POINTER_ALIGN (REGNO (tmp)));
4450 else if (GET_CODE (tmp) == PLUS
4451 && GET_CODE (XEXP (tmp, 0)) == REG
4452 && GET_CODE (XEXP (tmp, 1)) == CONST_INT)
4453 {
4454 unsigned HOST_WIDE_INT c = INTVAL (XEXP (tmp, 1));
4455 unsigned int a = REGNO_POINTER_ALIGN (REGNO (XEXP (tmp, 0)));
4456
4457 if (a > dst_align)
4458 {
4459 if (a >= 64 && c % 8 == 0)
4460 dst_align = 64;
4461 else if (a >= 32 && c % 4 == 0)
4462 dst_align = 32;
4463 else if (a >= 16 && c % 2 == 0)
4464 dst_align = 16;
4465 }
4466 }
4467
4468 /* Load the entire block into registers. */
4469 if (GET_CODE (XEXP (orig_src, 0)) == ADDRESSOF)
4470 {
4471 enum machine_mode mode;
4472
4473 tmp = XEXP (XEXP (orig_src, 0), 0);
4474
4475 /* Don't use the existing register if we're reading more than
4476 is held in the register. Nor if there is not a mode that
4477 handles the exact size. */
4478 mode = mode_for_size (bytes * BITS_PER_UNIT, MODE_INT, 1);
4479 if (GET_CODE (tmp) == REG
4480 && mode != BLKmode
4481 && GET_MODE_SIZE (GET_MODE (tmp)) >= bytes)
4482 {
4483 if (mode == TImode)
4484 {
4485 data_regs[nregs] = gen_lowpart (DImode, tmp);
4486 data_regs[nregs + 1] = gen_highpart (DImode, tmp);
4487 nregs += 2;
4488 }
4489 else
4490 data_regs[nregs++] = gen_lowpart (mode, tmp);
4491
4492 goto src_done;
4493 }
4494
4495 /* No appropriate mode; fall back on memory. */
4496 orig_src = replace_equiv_address (orig_src,
4497 copy_addr_to_reg (XEXP (orig_src, 0)));
4498 src_align = GET_MODE_BITSIZE (GET_MODE (tmp));
4499 }
4500
4501 ofs = 0;
4502 if (src_align >= 64 && bytes >= 8)
4503 {
4504 words = bytes / 8;
4505
4506 for (i = 0; i < words; ++i)
4507 data_regs[nregs + i] = gen_reg_rtx (DImode);
4508
4509 for (i = 0; i < words; ++i)
4510 emit_move_insn (data_regs[nregs + i],
4511 adjust_address (orig_src, DImode, ofs + i * 8));
4512
4513 nregs += words;
4514 bytes -= words * 8;
4515 ofs += words * 8;
4516 }
4517
4518 if (src_align >= 32 && bytes >= 4)
4519 {
4520 words = bytes / 4;
4521
4522 for (i = 0; i < words; ++i)
4523 data_regs[nregs + i] = gen_reg_rtx (SImode);
4524
4525 for (i = 0; i < words; ++i)
4526 emit_move_insn (data_regs[nregs + i],
4527 adjust_address (orig_src, SImode, ofs + i * 4));
4528
4529 nregs += words;
4530 bytes -= words * 4;
4531 ofs += words * 4;
4532 }
4533
4534 if (bytes >= 8)
4535 {
4536 words = bytes / 8;
4537
4538 for (i = 0; i < words+1; ++i)
4539 data_regs[nregs + i] = gen_reg_rtx (DImode);
4540
4541 alpha_expand_unaligned_load_words (data_regs + nregs, orig_src,
4542 words, ofs);
4543
4544 nregs += words;
4545 bytes -= words * 8;
4546 ofs += words * 8;
4547 }
4548
4549 if (! TARGET_BWX && bytes >= 4)
4550 {
4551 data_regs[nregs++] = tmp = gen_reg_rtx (SImode);
4552 alpha_expand_unaligned_load (tmp, orig_src, 4, ofs, 0);
4553 bytes -= 4;
4554 ofs += 4;
4555 }
4556
4557 if (bytes >= 2)
4558 {
4559 if (src_align >= 16)
4560 {
4561 do {
4562 data_regs[nregs++] = tmp = gen_reg_rtx (HImode);
4563 emit_move_insn (tmp, adjust_address (orig_src, HImode, ofs));
4564 bytes -= 2;
4565 ofs += 2;
4566 } while (bytes >= 2);
4567 }
4568 else if (! TARGET_BWX)
4569 {
4570 data_regs[nregs++] = tmp = gen_reg_rtx (HImode);
4571 alpha_expand_unaligned_load (tmp, orig_src, 2, ofs, 0);
4572 bytes -= 2;
4573 ofs += 2;
4574 }
4575 }
4576
4577 while (bytes > 0)
4578 {
4579 data_regs[nregs++] = tmp = gen_reg_rtx (QImode);
4580 emit_move_insn (tmp, adjust_address (orig_src, QImode, ofs));
4581 bytes -= 1;
4582 ofs += 1;
4583 }
4584
4585 src_done:
4586
4587 if (nregs > ARRAY_SIZE (data_regs))
4588 abort ();
4589
4590 /* Now save it back out again. */
4591
4592 i = 0, ofs = 0;
4593
4594 if (GET_CODE (XEXP (orig_dst, 0)) == ADDRESSOF)
4595 {
4596 enum machine_mode mode;
4597 tmp = XEXP (XEXP (orig_dst, 0), 0);
4598
4599 mode = mode_for_size (orig_bytes * BITS_PER_UNIT, MODE_INT, 1);
4600 if (GET_CODE (tmp) == REG && GET_MODE (tmp) == mode)
4601 {
4602 if (nregs == 1)
4603 {
4604 emit_move_insn (tmp, data_regs[0]);
4605 i = 1;
4606 goto dst_done;
4607 }
4608
4609 else if (nregs == 2 && mode == TImode)
4610 {
4611 /* Undo the subregging done above when copying between
4612 two TImode registers. */
4613 if (GET_CODE (data_regs[0]) == SUBREG
4614 && GET_MODE (SUBREG_REG (data_regs[0])) == TImode)
4615 emit_move_insn (tmp, SUBREG_REG (data_regs[0]));
4616 else
4617 {
4618 rtx seq;
4619
4620 start_sequence ();
4621 emit_move_insn (gen_lowpart (DImode, tmp), data_regs[0]);
4622 emit_move_insn (gen_highpart (DImode, tmp), data_regs[1]);
4623 seq = get_insns ();
4624 end_sequence ();
4625
4626 emit_no_conflict_block (seq, tmp, data_regs[0],
4627 data_regs[1], NULL_RTX);
4628 }
4629
4630 i = 2;
4631 goto dst_done;
4632 }
4633 }
4634
4635 /* ??? If nregs > 1, consider reconstructing the word in regs. */
4636 /* ??? Optimize mode < dst_mode with strict_low_part. */
4637
4638 /* No appropriate mode; fall back on memory. We can speed things
4639 up by recognizing extra alignment information. */
4640 orig_dst = replace_equiv_address (orig_dst,
4641 copy_addr_to_reg (XEXP (orig_dst, 0)));
4642 dst_align = GET_MODE_BITSIZE (GET_MODE (tmp));
4643 }
4644
4645 /* Write out the data in whatever chunks reading the source allowed. */
4646 if (dst_align >= 64)
4647 {
4648 while (i < nregs && GET_MODE (data_regs[i]) == DImode)
4649 {
4650 emit_move_insn (adjust_address (orig_dst, DImode, ofs),
4651 data_regs[i]);
4652 ofs += 8;
4653 i++;
4654 }
4655 }
4656
4657 if (dst_align >= 32)
4658 {
4659 /* If the source has remaining DImode regs, write them out in
4660 two pieces. */
4661 while (i < nregs && GET_MODE (data_regs[i]) == DImode)
4662 {
4663 tmp = expand_binop (DImode, lshr_optab, data_regs[i], GEN_INT (32),
4664 NULL_RTX, 1, OPTAB_WIDEN);
4665
4666 emit_move_insn (adjust_address (orig_dst, SImode, ofs),
4667 gen_lowpart (SImode, data_regs[i]));
4668 emit_move_insn (adjust_address (orig_dst, SImode, ofs + 4),
4669 gen_lowpart (SImode, tmp));
4670 ofs += 8;
4671 i++;
4672 }
4673
4674 while (i < nregs && GET_MODE (data_regs[i]) == SImode)
4675 {
4676 emit_move_insn (adjust_address (orig_dst, SImode, ofs),
4677 data_regs[i]);
4678 ofs += 4;
4679 i++;
4680 }
4681 }
4682
4683 if (i < nregs && GET_MODE (data_regs[i]) == DImode)
4684 {
4685 /* Write out a remaining block of words using unaligned methods. */
4686
4687 for (words = 1; i + words < nregs; words++)
4688 if (GET_MODE (data_regs[i + words]) != DImode)
4689 break;
4690
4691 if (words == 1)
4692 alpha_expand_unaligned_store (orig_dst, data_regs[i], 8, ofs);
4693 else
4694 alpha_expand_unaligned_store_words (data_regs + i, orig_dst,
4695 words, ofs);
4696
4697 i += words;
4698 ofs += words * 8;
4699 }
4700
4701 /* Due to the above, this won't be aligned. */
4702 /* ??? If we have more than one of these, consider constructing full
4703 words in registers and using alpha_expand_unaligned_store_words. */
4704 while (i < nregs && GET_MODE (data_regs[i]) == SImode)
4705 {
4706 alpha_expand_unaligned_store (orig_dst, data_regs[i], 4, ofs);
4707 ofs += 4;
4708 i++;
4709 }
4710
4711 if (dst_align >= 16)
4712 while (i < nregs && GET_MODE (data_regs[i]) == HImode)
4713 {
4714 emit_move_insn (adjust_address (orig_dst, HImode, ofs), data_regs[i]);
4715 i++;
4716 ofs += 2;
4717 }
4718 else
4719 while (i < nregs && GET_MODE (data_regs[i]) == HImode)
4720 {
4721 alpha_expand_unaligned_store (orig_dst, data_regs[i], 2, ofs);
4722 i++;
4723 ofs += 2;
4724 }
4725
4726 while (i < nregs && GET_MODE (data_regs[i]) == QImode)
4727 {
4728 emit_move_insn (adjust_address (orig_dst, QImode, ofs), data_regs[i]);
4729 i++;
4730 ofs += 1;
4731 }
4732
4733 dst_done:
4734
4735 if (i != nregs)
4736 abort ();
4737
4738 return 1;
4739 }
4740
4741 int
4742 alpha_expand_block_clear (rtx operands[])
4743 {
4744 rtx bytes_rtx = operands[1];
4745 rtx align_rtx = operands[2];
4746 HOST_WIDE_INT orig_bytes = INTVAL (bytes_rtx);
4747 HOST_WIDE_INT bytes = orig_bytes;
4748 HOST_WIDE_INT align = INTVAL (align_rtx) * BITS_PER_UNIT;
4749 HOST_WIDE_INT alignofs = 0;
4750 rtx orig_dst = operands[0];
4751 rtx tmp;
4752 int i, words, ofs = 0;
4753
4754 if (orig_bytes <= 0)
4755 return 1;
4756 if (orig_bytes > MAX_MOVE_WORDS * UNITS_PER_WORD)
4757 return 0;
4758
4759 /* Look for stricter alignment. */
4760 tmp = XEXP (orig_dst, 0);
4761 if (GET_CODE (tmp) == REG)
4762 align = MAX (align, REGNO_POINTER_ALIGN (REGNO (tmp)));
4763 else if (GET_CODE (tmp) == PLUS
4764 && GET_CODE (XEXP (tmp, 0)) == REG
4765 && GET_CODE (XEXP (tmp, 1)) == CONST_INT)
4766 {
4767 HOST_WIDE_INT c = INTVAL (XEXP (tmp, 1));
4768 int a = REGNO_POINTER_ALIGN (REGNO (XEXP (tmp, 0)));
4769
4770 if (a > align)
4771 {
4772 if (a >= 64)
4773 align = a, alignofs = 8 - c % 8;
4774 else if (a >= 32)
4775 align = a, alignofs = 4 - c % 4;
4776 else if (a >= 16)
4777 align = a, alignofs = 2 - c % 2;
4778 }
4779 }
4780 else if (GET_CODE (tmp) == ADDRESSOF)
4781 {
4782 enum machine_mode mode;
4783
4784 mode = mode_for_size (bytes * BITS_PER_UNIT, MODE_INT, 1);
4785 if (GET_MODE (XEXP (tmp, 0)) == mode)
4786 {
4787 emit_move_insn (XEXP (tmp, 0), const0_rtx);
4788 return 1;
4789 }
4790
4791 /* No appropriate mode; fall back on memory. */
4792 orig_dst = replace_equiv_address (orig_dst, copy_addr_to_reg (tmp));
4793 align = GET_MODE_BITSIZE (GET_MODE (XEXP (tmp, 0)));
4794 }
4795
4796 /* Handle an unaligned prefix first. */
4797
4798 if (alignofs > 0)
4799 {
4800 #if HOST_BITS_PER_WIDE_INT >= 64
4801 /* Given that alignofs is bounded by align, the only time BWX could
4802 generate three stores is for a 7 byte fill. Prefer two individual
4803 stores over a load/mask/store sequence. */
4804 if ((!TARGET_BWX || alignofs == 7)
4805 && align >= 32
4806 && !(alignofs == 4 && bytes >= 4))
4807 {
4808 enum machine_mode mode = (align >= 64 ? DImode : SImode);
4809 int inv_alignofs = (align >= 64 ? 8 : 4) - alignofs;
4810 rtx mem, tmp;
4811 HOST_WIDE_INT mask;
4812
4813 mem = adjust_address (orig_dst, mode, ofs - inv_alignofs);
4814 set_mem_alias_set (mem, 0);
4815
4816 mask = ~(~(HOST_WIDE_INT)0 << (inv_alignofs * 8));
4817 if (bytes < alignofs)
4818 {
4819 mask |= ~(HOST_WIDE_INT)0 << ((inv_alignofs + bytes) * 8);
4820 ofs += bytes;
4821 bytes = 0;
4822 }
4823 else
4824 {
4825 bytes -= alignofs;
4826 ofs += alignofs;
4827 }
4828 alignofs = 0;
4829
4830 tmp = expand_binop (mode, and_optab, mem, GEN_INT (mask),
4831 NULL_RTX, 1, OPTAB_WIDEN);
4832
4833 emit_move_insn (mem, tmp);
4834 }
4835 #endif
4836
4837 if (TARGET_BWX && (alignofs & 1) && bytes >= 1)
4838 {
4839 emit_move_insn (adjust_address (orig_dst, QImode, ofs), const0_rtx);
4840 bytes -= 1;
4841 ofs += 1;
4842 alignofs -= 1;
4843 }
4844 if (TARGET_BWX && align >= 16 && (alignofs & 3) == 2 && bytes >= 2)
4845 {
4846 emit_move_insn (adjust_address (orig_dst, HImode, ofs), const0_rtx);
4847 bytes -= 2;
4848 ofs += 2;
4849 alignofs -= 2;
4850 }
4851 if (alignofs == 4 && bytes >= 4)
4852 {
4853 emit_move_insn (adjust_address (orig_dst, SImode, ofs), const0_rtx);
4854 bytes -= 4;
4855 ofs += 4;
4856 alignofs = 0;
4857 }
4858
4859 /* If we've not used the extra lead alignment information by now,
4860 we won't be able to. Downgrade align to match what's left over. */
4861 if (alignofs > 0)
4862 {
4863 alignofs = alignofs & -alignofs;
4864 align = MIN (align, alignofs * BITS_PER_UNIT);
4865 }
4866 }
4867
4868 /* Handle a block of contiguous long-words. */
4869
4870 if (align >= 64 && bytes >= 8)
4871 {
4872 words = bytes / 8;
4873
4874 for (i = 0; i < words; ++i)
4875 emit_move_insn (adjust_address (orig_dst, DImode, ofs + i * 8),
4876 const0_rtx);
4877
4878 bytes -= words * 8;
4879 ofs += words * 8;
4880 }
4881
4882 /* If the block is large and appropriately aligned, emit a single
4883 store followed by a sequence of stq_u insns. */
4884
4885 if (align >= 32 && bytes > 16)
4886 {
4887 rtx orig_dsta;
4888
4889 emit_move_insn (adjust_address (orig_dst, SImode, ofs), const0_rtx);
4890 bytes -= 4;
4891 ofs += 4;
4892
4893 orig_dsta = XEXP (orig_dst, 0);
4894 if (GET_CODE (orig_dsta) == LO_SUM)
4895 orig_dsta = force_reg (Pmode, orig_dsta);
4896
4897 words = bytes / 8;
4898 for (i = 0; i < words; ++i)
4899 {
4900 rtx mem
4901 = change_address (orig_dst, DImode,
4902 gen_rtx_AND (DImode,
4903 plus_constant (orig_dsta, ofs + i*8),
4904 GEN_INT (-8)));
4905 set_mem_alias_set (mem, 0);
4906 emit_move_insn (mem, const0_rtx);
4907 }
4908
4909 /* Depending on the alignment, the first stq_u may have overlapped
4910 with the initial stl, which means that the last stq_u didn't
4911 write as much as it would appear. Leave those questionable bytes
4912 unaccounted for. */
4913 bytes -= words * 8 - 4;
4914 ofs += words * 8 - 4;
4915 }
4916
4917 /* Handle a smaller block of aligned words. */
4918
4919 if ((align >= 64 && bytes == 4)
4920 || (align == 32 && bytes >= 4))
4921 {
4922 words = bytes / 4;
4923
4924 for (i = 0; i < words; ++i)
4925 emit_move_insn (adjust_address (orig_dst, SImode, ofs + i * 4),
4926 const0_rtx);
4927
4928 bytes -= words * 4;
4929 ofs += words * 4;
4930 }
4931
4932 /* An unaligned block uses stq_u stores for as many as possible. */
4933
4934 if (bytes >= 8)
4935 {
4936 words = bytes / 8;
4937
4938 alpha_expand_unaligned_store_words (NULL, orig_dst, words, ofs);
4939
4940 bytes -= words * 8;
4941 ofs += words * 8;
4942 }
4943
4944 /* Next clean up any trailing pieces. */
4945
4946 #if HOST_BITS_PER_WIDE_INT >= 64
4947 /* Count the number of bits in BYTES for which aligned stores could
4948 be emitted. */
4949 words = 0;
4950 for (i = (TARGET_BWX ? 1 : 4); i * BITS_PER_UNIT <= align ; i <<= 1)
4951 if (bytes & i)
4952 words += 1;
4953
4954 /* If we have appropriate alignment (and it wouldn't take too many
4955 instructions otherwise), mask out the bytes we need. */
4956 if (TARGET_BWX ? words > 2 : bytes > 0)
4957 {
4958 if (align >= 64)
4959 {
4960 rtx mem, tmp;
4961 HOST_WIDE_INT mask;
4962
4963 mem = adjust_address (orig_dst, DImode, ofs);
4964 set_mem_alias_set (mem, 0);
4965
4966 mask = ~(HOST_WIDE_INT)0 << (bytes * 8);
4967
4968 tmp = expand_binop (DImode, and_optab, mem, GEN_INT (mask),
4969 NULL_RTX, 1, OPTAB_WIDEN);
4970
4971 emit_move_insn (mem, tmp);
4972 return 1;
4973 }
4974 else if (align >= 32 && bytes < 4)
4975 {
4976 rtx mem, tmp;
4977 HOST_WIDE_INT mask;
4978
4979 mem = adjust_address (orig_dst, SImode, ofs);
4980 set_mem_alias_set (mem, 0);
4981
4982 mask = ~(HOST_WIDE_INT)0 << (bytes * 8);
4983
4984 tmp = expand_binop (SImode, and_optab, mem, GEN_INT (mask),
4985 NULL_RTX, 1, OPTAB_WIDEN);
4986
4987 emit_move_insn (mem, tmp);
4988 return 1;
4989 }
4990 }
4991 #endif
4992
4993 if (!TARGET_BWX && bytes >= 4)
4994 {
4995 alpha_expand_unaligned_store (orig_dst, const0_rtx, 4, ofs);
4996 bytes -= 4;
4997 ofs += 4;
4998 }
4999
5000 if (bytes >= 2)
5001 {
5002 if (align >= 16)
5003 {
5004 do {
5005 emit_move_insn (adjust_address (orig_dst, HImode, ofs),
5006 const0_rtx);
5007 bytes -= 2;
5008 ofs += 2;
5009 } while (bytes >= 2);
5010 }
5011 else if (! TARGET_BWX)
5012 {
5013 alpha_expand_unaligned_store (orig_dst, const0_rtx, 2, ofs);
5014 bytes -= 2;
5015 ofs += 2;
5016 }
5017 }
5018
5019 while (bytes > 0)
5020 {
5021 emit_move_insn (adjust_address (orig_dst, QImode, ofs), const0_rtx);
5022 bytes -= 1;
5023 ofs += 1;
5024 }
5025
5026 return 1;
5027 }
5028
5029 /* Returns a mask so that zap(x, value) == x & mask. */
5030
5031 rtx
5032 alpha_expand_zap_mask (HOST_WIDE_INT value)
5033 {
5034 rtx result;
5035 int i;
5036
5037 if (HOST_BITS_PER_WIDE_INT >= 64)
5038 {
5039 HOST_WIDE_INT mask = 0;
5040
5041 for (i = 7; i >= 0; --i)
5042 {
5043 mask <<= 8;
5044 if (!((value >> i) & 1))
5045 mask |= 0xff;
5046 }
5047
5048 result = gen_int_mode (mask, DImode);
5049 }
5050 else if (HOST_BITS_PER_WIDE_INT == 32)
5051 {
5052 HOST_WIDE_INT mask_lo = 0, mask_hi = 0;
5053
5054 for (i = 7; i >= 4; --i)
5055 {
5056 mask_hi <<= 8;
5057 if (!((value >> i) & 1))
5058 mask_hi |= 0xff;
5059 }
5060
5061 for (i = 3; i >= 0; --i)
5062 {
5063 mask_lo <<= 8;
5064 if (!((value >> i) & 1))
5065 mask_lo |= 0xff;
5066 }
5067
5068 result = immed_double_const (mask_lo, mask_hi, DImode);
5069 }
5070 else
5071 abort ();
5072
5073 return result;
5074 }
5075
5076 void
5077 alpha_expand_builtin_vector_binop (rtx (*gen) (rtx, rtx, rtx),
5078 enum machine_mode mode,
5079 rtx op0, rtx op1, rtx op2)
5080 {
5081 op0 = gen_lowpart (mode, op0);
5082
5083 if (op1 == const0_rtx)
5084 op1 = CONST0_RTX (mode);
5085 else
5086 op1 = gen_lowpart (mode, op1);
5087
5088 if (op2 == const0_rtx)
5089 op2 = CONST0_RTX (mode);
5090 else
5091 op2 = gen_lowpart (mode, op2);
5092
5093 emit_insn ((*gen) (op0, op1, op2));
5094 }
5095 \f
5096 /* Adjust the cost of a scheduling dependency. Return the new cost of
5097 a dependency LINK or INSN on DEP_INSN. COST is the current cost. */
5098
5099 static int
5100 alpha_adjust_cost (rtx insn, rtx link, rtx dep_insn, int cost)
5101 {
5102 enum attr_type insn_type, dep_insn_type;
5103
5104 /* If the dependence is an anti-dependence, there is no cost. For an
5105 output dependence, there is sometimes a cost, but it doesn't seem
5106 worth handling those few cases. */
5107 if (REG_NOTE_KIND (link) != 0)
5108 return cost;
5109
5110 /* If we can't recognize the insns, we can't really do anything. */
5111 if (recog_memoized (insn) < 0 || recog_memoized (dep_insn) < 0)
5112 return cost;
5113
5114 insn_type = get_attr_type (insn);
5115 dep_insn_type = get_attr_type (dep_insn);
5116
5117 /* Bring in the user-defined memory latency. */
5118 if (dep_insn_type == TYPE_ILD
5119 || dep_insn_type == TYPE_FLD
5120 || dep_insn_type == TYPE_LDSYM)
5121 cost += alpha_memory_latency-1;
5122
5123 /* Everything else handled in DFA bypasses now. */
5124
5125 return cost;
5126 }
5127
5128 /* The number of instructions that can be issued per cycle. */
5129
5130 static int
5131 alpha_issue_rate (void)
5132 {
5133 return (alpha_cpu == PROCESSOR_EV4 ? 2 : 4);
5134 }
5135
5136 static int
5137 alpha_use_dfa_pipeline_interface (void)
5138 {
5139 return true;
5140 }
5141
5142 /* How many alternative schedules to try. This should be as wide as the
5143 scheduling freedom in the DFA, but no wider. Making this value too
5144 large results extra work for the scheduler.
5145
5146 For EV4, loads can be issued to either IB0 or IB1, thus we have 2
5147 alternative schedules. For EV5, we can choose between E0/E1 and
5148 FA/FM. For EV6, an arithmetic insn can be issued to U0/U1/L0/L1. */
5149
5150 static int
5151 alpha_multipass_dfa_lookahead (void)
5152 {
5153 return (alpha_cpu == PROCESSOR_EV6 ? 4 : 2);
5154 }
5155 \f
5156 /* Machine-specific function data. */
5157
5158 struct machine_function GTY(())
5159 {
5160 /* For unicosmk. */
5161 /* List of call information words for calls from this function. */
5162 struct rtx_def *first_ciw;
5163 struct rtx_def *last_ciw;
5164 int ciw_count;
5165
5166 /* List of deferred case vectors. */
5167 struct rtx_def *addr_list;
5168
5169 /* For OSF. */
5170 const char *some_ld_name;
5171 };
5172
5173 /* How to allocate a 'struct machine_function'. */
5174
5175 static struct machine_function *
5176 alpha_init_machine_status (void)
5177 {
5178 return ((struct machine_function *)
5179 ggc_alloc_cleared (sizeof (struct machine_function)));
5180 }
5181
5182 /* Functions to save and restore alpha_return_addr_rtx. */
5183
5184 /* Start the ball rolling with RETURN_ADDR_RTX. */
5185
5186 rtx
5187 alpha_return_addr (int count, rtx frame ATTRIBUTE_UNUSED)
5188 {
5189 if (count != 0)
5190 return const0_rtx;
5191
5192 return get_hard_reg_initial_val (Pmode, REG_RA);
5193 }
5194
5195 /* Return or create a pseudo containing the gp value for the current
5196 function. Needed only if TARGET_LD_BUGGY_LDGP. */
5197
5198 rtx
5199 alpha_gp_save_rtx (void)
5200 {
5201 rtx r = get_hard_reg_initial_val (DImode, 29);
5202 if (GET_CODE (r) != MEM)
5203 r = gen_mem_addressof (r, NULL_TREE, /*rescan=*/true);
5204 return r;
5205 }
5206
5207 static int
5208 alpha_ra_ever_killed (void)
5209 {
5210 rtx top;
5211
5212 if (!has_hard_reg_initial_val (Pmode, REG_RA))
5213 return regs_ever_live[REG_RA];
5214
5215 push_topmost_sequence ();
5216 top = get_insns ();
5217 pop_topmost_sequence ();
5218
5219 return reg_set_between_p (gen_rtx_REG (Pmode, REG_RA), top, NULL_RTX);
5220 }
5221
5222 \f
5223 /* Return the trap mode suffix applicable to the current
5224 instruction, or NULL. */
5225
5226 static const char *
5227 get_trap_mode_suffix (void)
5228 {
5229 enum attr_trap_suffix s = get_attr_trap_suffix (current_output_insn);
5230
5231 switch (s)
5232 {
5233 case TRAP_SUFFIX_NONE:
5234 return NULL;
5235
5236 case TRAP_SUFFIX_SU:
5237 if (alpha_fptm >= ALPHA_FPTM_SU)
5238 return "su";
5239 return NULL;
5240
5241 case TRAP_SUFFIX_SUI:
5242 if (alpha_fptm >= ALPHA_FPTM_SUI)
5243 return "sui";
5244 return NULL;
5245
5246 case TRAP_SUFFIX_V_SV:
5247 switch (alpha_fptm)
5248 {
5249 case ALPHA_FPTM_N:
5250 return NULL;
5251 case ALPHA_FPTM_U:
5252 return "v";
5253 case ALPHA_FPTM_SU:
5254 case ALPHA_FPTM_SUI:
5255 return "sv";
5256 }
5257 break;
5258
5259 case TRAP_SUFFIX_V_SV_SVI:
5260 switch (alpha_fptm)
5261 {
5262 case ALPHA_FPTM_N:
5263 return NULL;
5264 case ALPHA_FPTM_U:
5265 return "v";
5266 case ALPHA_FPTM_SU:
5267 return "sv";
5268 case ALPHA_FPTM_SUI:
5269 return "svi";
5270 }
5271 break;
5272
5273 case TRAP_SUFFIX_U_SU_SUI:
5274 switch (alpha_fptm)
5275 {
5276 case ALPHA_FPTM_N:
5277 return NULL;
5278 case ALPHA_FPTM_U:
5279 return "u";
5280 case ALPHA_FPTM_SU:
5281 return "su";
5282 case ALPHA_FPTM_SUI:
5283 return "sui";
5284 }
5285 break;
5286 }
5287 abort ();
5288 }
5289
5290 /* Return the rounding mode suffix applicable to the current
5291 instruction, or NULL. */
5292
5293 static const char *
5294 get_round_mode_suffix (void)
5295 {
5296 enum attr_round_suffix s = get_attr_round_suffix (current_output_insn);
5297
5298 switch (s)
5299 {
5300 case ROUND_SUFFIX_NONE:
5301 return NULL;
5302 case ROUND_SUFFIX_NORMAL:
5303 switch (alpha_fprm)
5304 {
5305 case ALPHA_FPRM_NORM:
5306 return NULL;
5307 case ALPHA_FPRM_MINF:
5308 return "m";
5309 case ALPHA_FPRM_CHOP:
5310 return "c";
5311 case ALPHA_FPRM_DYN:
5312 return "d";
5313 }
5314 break;
5315
5316 case ROUND_SUFFIX_C:
5317 return "c";
5318 }
5319 abort ();
5320 }
5321
5322 /* Locate some local-dynamic symbol still in use by this function
5323 so that we can print its name in some movdi_er_tlsldm pattern. */
5324
5325 static int
5326 get_some_local_dynamic_name_1 (rtx *px, void *data ATTRIBUTE_UNUSED)
5327 {
5328 rtx x = *px;
5329
5330 if (GET_CODE (x) == SYMBOL_REF
5331 && SYMBOL_REF_TLS_MODEL (x) == TLS_MODEL_LOCAL_DYNAMIC)
5332 {
5333 cfun->machine->some_ld_name = XSTR (x, 0);
5334 return 1;
5335 }
5336
5337 return 0;
5338 }
5339
5340 static const char *
5341 get_some_local_dynamic_name (void)
5342 {
5343 rtx insn;
5344
5345 if (cfun->machine->some_ld_name)
5346 return cfun->machine->some_ld_name;
5347
5348 for (insn = get_insns (); insn ; insn = NEXT_INSN (insn))
5349 if (INSN_P (insn)
5350 && for_each_rtx (&PATTERN (insn), get_some_local_dynamic_name_1, 0))
5351 return cfun->machine->some_ld_name;
5352
5353 abort ();
5354 }
5355
5356 /* Print an operand. Recognize special options, documented below. */
5357
5358 void
5359 print_operand (FILE *file, rtx x, int code)
5360 {
5361 int i;
5362
5363 switch (code)
5364 {
5365 case '~':
5366 /* Print the assembler name of the current function. */
5367 assemble_name (file, alpha_fnname);
5368 break;
5369
5370 case '&':
5371 assemble_name (file, get_some_local_dynamic_name ());
5372 break;
5373
5374 case '/':
5375 {
5376 const char *trap = get_trap_mode_suffix ();
5377 const char *round = get_round_mode_suffix ();
5378
5379 if (trap || round)
5380 fprintf (file, (TARGET_AS_SLASH_BEFORE_SUFFIX ? "/%s%s" : "%s%s"),
5381 (trap ? trap : ""), (round ? round : ""));
5382 break;
5383 }
5384
5385 case ',':
5386 /* Generates single precision instruction suffix. */
5387 fputc ((TARGET_FLOAT_VAX ? 'f' : 's'), file);
5388 break;
5389
5390 case '-':
5391 /* Generates double precision instruction suffix. */
5392 fputc ((TARGET_FLOAT_VAX ? 'g' : 't'), file);
5393 break;
5394
5395 case '+':
5396 /* Generates a nop after a noreturn call at the very end of the
5397 function. */
5398 if (next_real_insn (current_output_insn) == 0)
5399 fprintf (file, "\n\tnop");
5400 break;
5401
5402 case '#':
5403 if (alpha_this_literal_sequence_number == 0)
5404 alpha_this_literal_sequence_number = alpha_next_sequence_number++;
5405 fprintf (file, "%d", alpha_this_literal_sequence_number);
5406 break;
5407
5408 case '*':
5409 if (alpha_this_gpdisp_sequence_number == 0)
5410 alpha_this_gpdisp_sequence_number = alpha_next_sequence_number++;
5411 fprintf (file, "%d", alpha_this_gpdisp_sequence_number);
5412 break;
5413
5414 case 'H':
5415 if (GET_CODE (x) == HIGH)
5416 output_addr_const (file, XEXP (x, 0));
5417 else
5418 output_operand_lossage ("invalid %%H value");
5419 break;
5420
5421 case 'J':
5422 {
5423 const char *lituse;
5424
5425 if (GET_CODE (x) == UNSPEC && XINT (x, 1) == UNSPEC_TLSGD_CALL)
5426 {
5427 x = XVECEXP (x, 0, 0);
5428 lituse = "lituse_tlsgd";
5429 }
5430 else if (GET_CODE (x) == UNSPEC && XINT (x, 1) == UNSPEC_TLSLDM_CALL)
5431 {
5432 x = XVECEXP (x, 0, 0);
5433 lituse = "lituse_tlsldm";
5434 }
5435 else if (GET_CODE (x) == CONST_INT)
5436 lituse = "lituse_jsr";
5437 else
5438 {
5439 output_operand_lossage ("invalid %%J value");
5440 break;
5441 }
5442
5443 if (x != const0_rtx)
5444 fprintf (file, "\t\t!%s!%d", lituse, (int) INTVAL (x));
5445 }
5446 break;
5447
5448 case 'r':
5449 /* If this operand is the constant zero, write it as "$31". */
5450 if (GET_CODE (x) == REG)
5451 fprintf (file, "%s", reg_names[REGNO (x)]);
5452 else if (x == CONST0_RTX (GET_MODE (x)))
5453 fprintf (file, "$31");
5454 else
5455 output_operand_lossage ("invalid %%r value");
5456 break;
5457
5458 case 'R':
5459 /* Similar, but for floating-point. */
5460 if (GET_CODE (x) == REG)
5461 fprintf (file, "%s", reg_names[REGNO (x)]);
5462 else if (x == CONST0_RTX (GET_MODE (x)))
5463 fprintf (file, "$f31");
5464 else
5465 output_operand_lossage ("invalid %%R value");
5466 break;
5467
5468 case 'N':
5469 /* Write the 1's complement of a constant. */
5470 if (GET_CODE (x) != CONST_INT)
5471 output_operand_lossage ("invalid %%N value");
5472
5473 fprintf (file, HOST_WIDE_INT_PRINT_DEC, ~ INTVAL (x));
5474 break;
5475
5476 case 'P':
5477 /* Write 1 << C, for a constant C. */
5478 if (GET_CODE (x) != CONST_INT)
5479 output_operand_lossage ("invalid %%P value");
5480
5481 fprintf (file, HOST_WIDE_INT_PRINT_DEC, (HOST_WIDE_INT) 1 << INTVAL (x));
5482 break;
5483
5484 case 'h':
5485 /* Write the high-order 16 bits of a constant, sign-extended. */
5486 if (GET_CODE (x) != CONST_INT)
5487 output_operand_lossage ("invalid %%h value");
5488
5489 fprintf (file, HOST_WIDE_INT_PRINT_DEC, INTVAL (x) >> 16);
5490 break;
5491
5492 case 'L':
5493 /* Write the low-order 16 bits of a constant, sign-extended. */
5494 if (GET_CODE (x) != CONST_INT)
5495 output_operand_lossage ("invalid %%L value");
5496
5497 fprintf (file, HOST_WIDE_INT_PRINT_DEC,
5498 (INTVAL (x) & 0xffff) - 2 * (INTVAL (x) & 0x8000));
5499 break;
5500
5501 case 'm':
5502 /* Write mask for ZAP insn. */
5503 if (GET_CODE (x) == CONST_DOUBLE)
5504 {
5505 HOST_WIDE_INT mask = 0;
5506 HOST_WIDE_INT value;
5507
5508 value = CONST_DOUBLE_LOW (x);
5509 for (i = 0; i < HOST_BITS_PER_WIDE_INT / HOST_BITS_PER_CHAR;
5510 i++, value >>= 8)
5511 if (value & 0xff)
5512 mask |= (1 << i);
5513
5514 value = CONST_DOUBLE_HIGH (x);
5515 for (i = 0; i < HOST_BITS_PER_WIDE_INT / HOST_BITS_PER_CHAR;
5516 i++, value >>= 8)
5517 if (value & 0xff)
5518 mask |= (1 << (i + sizeof (int)));
5519
5520 fprintf (file, HOST_WIDE_INT_PRINT_DEC, mask & 0xff);
5521 }
5522
5523 else if (GET_CODE (x) == CONST_INT)
5524 {
5525 HOST_WIDE_INT mask = 0, value = INTVAL (x);
5526
5527 for (i = 0; i < 8; i++, value >>= 8)
5528 if (value & 0xff)
5529 mask |= (1 << i);
5530
5531 fprintf (file, HOST_WIDE_INT_PRINT_DEC, mask);
5532 }
5533 else
5534 output_operand_lossage ("invalid %%m value");
5535 break;
5536
5537 case 'M':
5538 /* 'b', 'w', 'l', or 'q' as the value of the constant. */
5539 if (GET_CODE (x) != CONST_INT
5540 || (INTVAL (x) != 8 && INTVAL (x) != 16
5541 && INTVAL (x) != 32 && INTVAL (x) != 64))
5542 output_operand_lossage ("invalid %%M value");
5543
5544 fprintf (file, "%s",
5545 (INTVAL (x) == 8 ? "b"
5546 : INTVAL (x) == 16 ? "w"
5547 : INTVAL (x) == 32 ? "l"
5548 : "q"));
5549 break;
5550
5551 case 'U':
5552 /* Similar, except do it from the mask. */
5553 if (GET_CODE (x) == CONST_INT)
5554 {
5555 HOST_WIDE_INT value = INTVAL (x);
5556
5557 if (value == 0xff)
5558 {
5559 fputc ('b', file);
5560 break;
5561 }
5562 if (value == 0xffff)
5563 {
5564 fputc ('w', file);
5565 break;
5566 }
5567 if (value == 0xffffffff)
5568 {
5569 fputc ('l', file);
5570 break;
5571 }
5572 if (value == -1)
5573 {
5574 fputc ('q', file);
5575 break;
5576 }
5577 }
5578 else if (HOST_BITS_PER_WIDE_INT == 32
5579 && GET_CODE (x) == CONST_DOUBLE
5580 && CONST_DOUBLE_LOW (x) == 0xffffffff
5581 && CONST_DOUBLE_HIGH (x) == 0)
5582 {
5583 fputc ('l', file);
5584 break;
5585 }
5586 output_operand_lossage ("invalid %%U value");
5587 break;
5588
5589 case 's':
5590 /* Write the constant value divided by 8 for little-endian mode or
5591 (56 - value) / 8 for big-endian mode. */
5592
5593 if (GET_CODE (x) != CONST_INT
5594 || (unsigned HOST_WIDE_INT) INTVAL (x) >= (WORDS_BIG_ENDIAN
5595 ? 56
5596 : 64)
5597 || (INTVAL (x) & 7) != 0)
5598 output_operand_lossage ("invalid %%s value");
5599
5600 fprintf (file, HOST_WIDE_INT_PRINT_DEC,
5601 WORDS_BIG_ENDIAN
5602 ? (56 - INTVAL (x)) / 8
5603 : INTVAL (x) / 8);
5604 break;
5605
5606 case 'S':
5607 /* Same, except compute (64 - c) / 8 */
5608
5609 if (GET_CODE (x) != CONST_INT
5610 && (unsigned HOST_WIDE_INT) INTVAL (x) >= 64
5611 && (INTVAL (x) & 7) != 8)
5612 output_operand_lossage ("invalid %%s value");
5613
5614 fprintf (file, HOST_WIDE_INT_PRINT_DEC, (64 - INTVAL (x)) / 8);
5615 break;
5616
5617 case 't':
5618 {
5619 /* On Unicos/Mk systems: use a DEX expression if the symbol
5620 clashes with a register name. */
5621 int dex = unicosmk_need_dex (x);
5622 if (dex)
5623 fprintf (file, "DEX(%d)", dex);
5624 else
5625 output_addr_const (file, x);
5626 }
5627 break;
5628
5629 case 'C': case 'D': case 'c': case 'd':
5630 /* Write out comparison name. */
5631 {
5632 enum rtx_code c = GET_CODE (x);
5633
5634 if (GET_RTX_CLASS (c) != '<')
5635 output_operand_lossage ("invalid %%C value");
5636
5637 else if (code == 'D')
5638 c = reverse_condition (c);
5639 else if (code == 'c')
5640 c = swap_condition (c);
5641 else if (code == 'd')
5642 c = swap_condition (reverse_condition (c));
5643
5644 if (c == LEU)
5645 fprintf (file, "ule");
5646 else if (c == LTU)
5647 fprintf (file, "ult");
5648 else if (c == UNORDERED)
5649 fprintf (file, "un");
5650 else
5651 fprintf (file, "%s", GET_RTX_NAME (c));
5652 }
5653 break;
5654
5655 case 'E':
5656 /* Write the divide or modulus operator. */
5657 switch (GET_CODE (x))
5658 {
5659 case DIV:
5660 fprintf (file, "div%s", GET_MODE (x) == SImode ? "l" : "q");
5661 break;
5662 case UDIV:
5663 fprintf (file, "div%su", GET_MODE (x) == SImode ? "l" : "q");
5664 break;
5665 case MOD:
5666 fprintf (file, "rem%s", GET_MODE (x) == SImode ? "l" : "q");
5667 break;
5668 case UMOD:
5669 fprintf (file, "rem%su", GET_MODE (x) == SImode ? "l" : "q");
5670 break;
5671 default:
5672 output_operand_lossage ("invalid %%E value");
5673 break;
5674 }
5675 break;
5676
5677 case 'A':
5678 /* Write "_u" for unaligned access. */
5679 if (GET_CODE (x) == MEM && GET_CODE (XEXP (x, 0)) == AND)
5680 fprintf (file, "_u");
5681 break;
5682
5683 case 0:
5684 if (GET_CODE (x) == REG)
5685 fprintf (file, "%s", reg_names[REGNO (x)]);
5686 else if (GET_CODE (x) == MEM)
5687 output_address (XEXP (x, 0));
5688 else if (GET_CODE (x) == CONST && GET_CODE (XEXP (x, 0)) == UNSPEC)
5689 {
5690 switch (XINT (XEXP (x, 0), 1))
5691 {
5692 case UNSPEC_DTPREL:
5693 case UNSPEC_TPREL:
5694 output_addr_const (file, XVECEXP (XEXP (x, 0), 0, 0));
5695 break;
5696 default:
5697 output_operand_lossage ("unknown relocation unspec");
5698 break;
5699 }
5700 }
5701 else
5702 output_addr_const (file, x);
5703 break;
5704
5705 default:
5706 output_operand_lossage ("invalid %%xn code");
5707 }
5708 }
5709
5710 void
5711 print_operand_address (FILE *file, rtx addr)
5712 {
5713 int basereg = 31;
5714 HOST_WIDE_INT offset = 0;
5715
5716 if (GET_CODE (addr) == AND)
5717 addr = XEXP (addr, 0);
5718
5719 if (GET_CODE (addr) == PLUS
5720 && GET_CODE (XEXP (addr, 1)) == CONST_INT)
5721 {
5722 offset = INTVAL (XEXP (addr, 1));
5723 addr = XEXP (addr, 0);
5724 }
5725
5726 if (GET_CODE (addr) == LO_SUM)
5727 {
5728 const char *reloc16, *reloclo;
5729 rtx op1 = XEXP (addr, 1);
5730
5731 if (GET_CODE (op1) == CONST && GET_CODE (XEXP (op1, 0)) == UNSPEC)
5732 {
5733 op1 = XEXP (op1, 0);
5734 switch (XINT (op1, 1))
5735 {
5736 case UNSPEC_DTPREL:
5737 reloc16 = NULL;
5738 reloclo = (alpha_tls_size == 16 ? "dtprel" : "dtprello");
5739 break;
5740 case UNSPEC_TPREL:
5741 reloc16 = NULL;
5742 reloclo = (alpha_tls_size == 16 ? "tprel" : "tprello");
5743 break;
5744 default:
5745 output_operand_lossage ("unknown relocation unspec");
5746 return;
5747 }
5748
5749 output_addr_const (file, XVECEXP (op1, 0, 0));
5750 }
5751 else
5752 {
5753 reloc16 = "gprel";
5754 reloclo = "gprellow";
5755 output_addr_const (file, op1);
5756 }
5757
5758 if (offset)
5759 fprintf (file, "+" HOST_WIDE_INT_PRINT_DEC, offset);
5760
5761 addr = XEXP (addr, 0);
5762 if (GET_CODE (addr) == REG)
5763 basereg = REGNO (addr);
5764 else if (GET_CODE (addr) == SUBREG
5765 && GET_CODE (SUBREG_REG (addr)) == REG)
5766 basereg = subreg_regno (addr);
5767 else
5768 abort ();
5769
5770 fprintf (file, "($%d)\t\t!%s", basereg,
5771 (basereg == 29 ? reloc16 : reloclo));
5772 return;
5773 }
5774
5775 if (GET_CODE (addr) == REG)
5776 basereg = REGNO (addr);
5777 else if (GET_CODE (addr) == SUBREG
5778 && GET_CODE (SUBREG_REG (addr)) == REG)
5779 basereg = subreg_regno (addr);
5780 else if (GET_CODE (addr) == CONST_INT)
5781 offset = INTVAL (addr);
5782
5783 #if TARGET_ABI_OPEN_VMS
5784 else if (GET_CODE (addr) == SYMBOL_REF)
5785 {
5786 fprintf (file, "%s", XSTR (addr, 0));
5787 return;
5788 }
5789 else if (GET_CODE (addr) == CONST
5790 && GET_CODE (XEXP (addr, 0)) == PLUS
5791 && GET_CODE (XEXP (XEXP (addr, 0), 0)) == SYMBOL_REF)
5792 {
5793 fprintf (file, "%s+" HOST_WIDE_INT_PRINT_DEC,
5794 XSTR (XEXP (XEXP (addr, 0), 0), 0),
5795 INTVAL (XEXP (XEXP (addr, 0), 1)));
5796 return;
5797 }
5798 #endif
5799
5800 else
5801 abort ();
5802
5803 fprintf (file, HOST_WIDE_INT_PRINT_DEC "($%d)", offset, basereg);
5804 }
5805 \f
5806 /* Emit RTL insns to initialize the variable parts of a trampoline at
5807 TRAMP. FNADDR is an RTX for the address of the function's pure
5808 code. CXT is an RTX for the static chain value for the function.
5809
5810 The three offset parameters are for the individual template's
5811 layout. A JMPOFS < 0 indicates that the trampoline does not
5812 contain instructions at all.
5813
5814 We assume here that a function will be called many more times than
5815 its address is taken (e.g., it might be passed to qsort), so we
5816 take the trouble to initialize the "hint" field in the JMP insn.
5817 Note that the hint field is PC (new) + 4 * bits 13:0. */
5818
5819 void
5820 alpha_initialize_trampoline (rtx tramp, rtx fnaddr, rtx cxt,
5821 int fnofs, int cxtofs, int jmpofs)
5822 {
5823 rtx temp, temp1, addr;
5824 /* VMS really uses DImode pointers in memory at this point. */
5825 enum machine_mode mode = TARGET_ABI_OPEN_VMS ? Pmode : ptr_mode;
5826
5827 #ifdef POINTERS_EXTEND_UNSIGNED
5828 fnaddr = convert_memory_address (mode, fnaddr);
5829 cxt = convert_memory_address (mode, cxt);
5830 #endif
5831
5832 /* Store function address and CXT. */
5833 addr = memory_address (mode, plus_constant (tramp, fnofs));
5834 emit_move_insn (gen_rtx_MEM (mode, addr), fnaddr);
5835 addr = memory_address (mode, plus_constant (tramp, cxtofs));
5836 emit_move_insn (gen_rtx_MEM (mode, addr), cxt);
5837
5838 /* This has been disabled since the hint only has a 32k range, and in
5839 no existing OS is the stack within 32k of the text segment. */
5840 if (0 && jmpofs >= 0)
5841 {
5842 /* Compute hint value. */
5843 temp = force_operand (plus_constant (tramp, jmpofs+4), NULL_RTX);
5844 temp = expand_binop (DImode, sub_optab, fnaddr, temp, temp, 1,
5845 OPTAB_WIDEN);
5846 temp = expand_shift (RSHIFT_EXPR, Pmode, temp,
5847 build_int_2 (2, 0), NULL_RTX, 1);
5848 temp = expand_and (SImode, gen_lowpart (SImode, temp),
5849 GEN_INT (0x3fff), 0);
5850
5851 /* Merge in the hint. */
5852 addr = memory_address (SImode, plus_constant (tramp, jmpofs));
5853 temp1 = force_reg (SImode, gen_rtx_MEM (SImode, addr));
5854 temp1 = expand_and (SImode, temp1, GEN_INT (0xffffc000), NULL_RTX);
5855 temp1 = expand_binop (SImode, ior_optab, temp1, temp, temp1, 1,
5856 OPTAB_WIDEN);
5857 emit_move_insn (gen_rtx_MEM (SImode, addr), temp1);
5858 }
5859
5860 #ifdef TRANSFER_FROM_TRAMPOLINE
5861 emit_library_call (init_one_libfunc ("__enable_execute_stack"),
5862 0, VOIDmode, 1, tramp, Pmode);
5863 #endif
5864
5865 if (jmpofs >= 0)
5866 emit_insn (gen_imb ());
5867 }
5868 \f
5869 /* Determine where to put an argument to a function.
5870 Value is zero to push the argument on the stack,
5871 or a hard register in which to store the argument.
5872
5873 MODE is the argument's machine mode.
5874 TYPE is the data type of the argument (as a tree).
5875 This is null for libcalls where that information may
5876 not be available.
5877 CUM is a variable of type CUMULATIVE_ARGS which gives info about
5878 the preceding args and about the function being called.
5879 NAMED is nonzero if this argument is a named parameter
5880 (otherwise it is an extra parameter matching an ellipsis).
5881
5882 On Alpha the first 6 words of args are normally in registers
5883 and the rest are pushed. */
5884
5885 rtx
5886 function_arg (CUMULATIVE_ARGS cum, enum machine_mode mode, tree type,
5887 int named ATTRIBUTE_UNUSED)
5888 {
5889 int basereg;
5890 int num_args;
5891
5892 /* Don't get confused and pass small structures in FP registers. */
5893 if (type && AGGREGATE_TYPE_P (type))
5894 basereg = 16;
5895 else
5896 {
5897 #ifdef ENABLE_CHECKING
5898 /* With SPLIT_COMPLEX_ARGS, we shouldn't see any raw complex
5899 values here. */
5900 if (COMPLEX_MODE_P (mode))
5901 abort ();
5902 #endif
5903
5904 /* Set up defaults for FP operands passed in FP registers, and
5905 integral operands passed in integer registers. */
5906 if (TARGET_FPREGS && GET_MODE_CLASS (mode) == MODE_FLOAT)
5907 basereg = 32 + 16;
5908 else
5909 basereg = 16;
5910 }
5911
5912 /* ??? Irritatingly, the definition of CUMULATIVE_ARGS is different for
5913 the three platforms, so we can't avoid conditional compilation. */
5914 #if TARGET_ABI_OPEN_VMS
5915 {
5916 if (mode == VOIDmode)
5917 return alpha_arg_info_reg_val (cum);
5918
5919 num_args = cum.num_args;
5920 if (num_args >= 6 || MUST_PASS_IN_STACK (mode, type))
5921 return NULL_RTX;
5922 }
5923 #elif TARGET_ABI_UNICOSMK
5924 {
5925 int size;
5926
5927 /* If this is the last argument, generate the call info word (CIW). */
5928 /* ??? We don't include the caller's line number in the CIW because
5929 I don't know how to determine it if debug infos are turned off. */
5930 if (mode == VOIDmode)
5931 {
5932 int i;
5933 HOST_WIDE_INT lo;
5934 HOST_WIDE_INT hi;
5935 rtx ciw;
5936
5937 lo = 0;
5938
5939 for (i = 0; i < cum.num_reg_words && i < 5; i++)
5940 if (cum.reg_args_type[i])
5941 lo |= (1 << (7 - i));
5942
5943 if (cum.num_reg_words == 6 && cum.reg_args_type[5])
5944 lo |= 7;
5945 else
5946 lo |= cum.num_reg_words;
5947
5948 #if HOST_BITS_PER_WIDE_INT == 32
5949 hi = (cum.num_args << 20) | cum.num_arg_words;
5950 #else
5951 lo = lo | ((HOST_WIDE_INT) cum.num_args << 52)
5952 | ((HOST_WIDE_INT) cum.num_arg_words << 32);
5953 hi = 0;
5954 #endif
5955 ciw = immed_double_const (lo, hi, DImode);
5956
5957 return gen_rtx_UNSPEC (DImode, gen_rtvec (1, ciw),
5958 UNSPEC_UMK_LOAD_CIW);
5959 }
5960
5961 size = ALPHA_ARG_SIZE (mode, type, named);
5962 num_args = cum.num_reg_words;
5963 if (MUST_PASS_IN_STACK (mode, type)
5964 || cum.num_reg_words + size > 6 || cum.force_stack)
5965 return NULL_RTX;
5966 else if (type && TYPE_MODE (type) == BLKmode)
5967 {
5968 rtx reg1, reg2;
5969
5970 reg1 = gen_rtx_REG (DImode, num_args + 16);
5971 reg1 = gen_rtx_EXPR_LIST (DImode, reg1, const0_rtx);
5972
5973 /* The argument fits in two registers. Note that we still need to
5974 reserve a register for empty structures. */
5975 if (size == 0)
5976 return NULL_RTX;
5977 else if (size == 1)
5978 return gen_rtx_PARALLEL (mode, gen_rtvec (1, reg1));
5979 else
5980 {
5981 reg2 = gen_rtx_REG (DImode, num_args + 17);
5982 reg2 = gen_rtx_EXPR_LIST (DImode, reg2, GEN_INT (8));
5983 return gen_rtx_PARALLEL (mode, gen_rtvec (2, reg1, reg2));
5984 }
5985 }
5986 }
5987 #elif TARGET_ABI_OSF
5988 {
5989 if (cum >= 6)
5990 return NULL_RTX;
5991 num_args = cum;
5992
5993 /* VOID is passed as a special flag for "last argument". */
5994 if (type == void_type_node)
5995 basereg = 16;
5996 else if (MUST_PASS_IN_STACK (mode, type))
5997 return NULL_RTX;
5998 else if (FUNCTION_ARG_PASS_BY_REFERENCE (cum, mode, type, named))
5999 basereg = 16;
6000 }
6001 #else
6002 #error Unhandled ABI
6003 #endif
6004
6005 return gen_rtx_REG (mode, num_args + basereg);
6006 }
6007
6008 /* Return true if TYPE must be returned in memory, instead of in registers. */
6009
6010 static bool
6011 alpha_return_in_memory (tree type, tree fndecl ATTRIBUTE_UNUSED)
6012 {
6013 enum machine_mode mode = VOIDmode;
6014 int size;
6015
6016 if (type)
6017 {
6018 mode = TYPE_MODE (type);
6019
6020 /* All aggregates are returned in memory. */
6021 if (AGGREGATE_TYPE_P (type))
6022 return true;
6023 }
6024
6025 size = GET_MODE_SIZE (mode);
6026 switch (GET_MODE_CLASS (mode))
6027 {
6028 case MODE_VECTOR_FLOAT:
6029 /* Pass all float vectors in memory, like an aggregate. */
6030 return true;
6031
6032 case MODE_COMPLEX_FLOAT:
6033 /* We judge complex floats on the size of their element,
6034 not the size of the whole type. */
6035 size = GET_MODE_UNIT_SIZE (mode);
6036 break;
6037
6038 case MODE_INT:
6039 case MODE_FLOAT:
6040 case MODE_COMPLEX_INT:
6041 case MODE_VECTOR_INT:
6042 break;
6043
6044 default:
6045 /* ??? We get called on all sorts of random stuff from
6046 aggregate_value_p. We can't abort, but it's not clear
6047 what's safe to return. Pretend it's a struct I guess. */
6048 return true;
6049 }
6050
6051 /* Otherwise types must fit in one register. */
6052 return size > UNITS_PER_WORD;
6053 }
6054
6055 /* Define how to find the value returned by a function. VALTYPE is the
6056 data type of the value (as a tree). If the precise function being
6057 called is known, FUNC is its FUNCTION_DECL; otherwise, FUNC is 0.
6058 MODE is set instead of VALTYPE for libcalls.
6059
6060 On Alpha the value is found in $0 for integer functions and
6061 $f0 for floating-point functions. */
6062
6063 rtx
6064 function_value (tree valtype, tree func ATTRIBUTE_UNUSED,
6065 enum machine_mode mode)
6066 {
6067 unsigned int regnum;
6068 enum mode_class class;
6069
6070 #ifdef ENABLE_CHECKING
6071 if (valtype && alpha_return_in_memory (valtype, func))
6072 abort ();
6073 #endif
6074
6075 if (valtype)
6076 mode = TYPE_MODE (valtype);
6077
6078 class = GET_MODE_CLASS (mode);
6079 switch (class)
6080 {
6081 case MODE_INT:
6082 /* Do the same thing as PROMOTE_MODE. */
6083 mode = DImode;
6084 /* FALLTHRU */
6085
6086 case MODE_COMPLEX_INT:
6087 case MODE_VECTOR_INT:
6088 regnum = 0;
6089 break;
6090
6091 case MODE_FLOAT:
6092 regnum = 32;
6093 break;
6094
6095 case MODE_COMPLEX_FLOAT:
6096 {
6097 enum machine_mode cmode = GET_MODE_INNER (mode);
6098
6099 return gen_rtx_PARALLEL
6100 (VOIDmode,
6101 gen_rtvec (2,
6102 gen_rtx_EXPR_LIST (VOIDmode, gen_rtx_REG (cmode, 32),
6103 GEN_INT (0)),
6104 gen_rtx_EXPR_LIST (VOIDmode, gen_rtx_REG (cmode, 33),
6105 GEN_INT (GET_MODE_SIZE (cmode)))));
6106 }
6107
6108 default:
6109 abort ();
6110 }
6111
6112 return gen_rtx_REG (mode, regnum);
6113 }
6114
6115 static tree
6116 alpha_build_builtin_va_list (void)
6117 {
6118 tree base, ofs, space, record, type_decl;
6119
6120 if (TARGET_ABI_OPEN_VMS || TARGET_ABI_UNICOSMK)
6121 return ptr_type_node;
6122
6123 record = (*lang_hooks.types.make_type) (RECORD_TYPE);
6124 type_decl = build_decl (TYPE_DECL, get_identifier ("__va_list_tag"), record);
6125 TREE_CHAIN (record) = type_decl;
6126 TYPE_NAME (record) = type_decl;
6127
6128 /* C++? SET_IS_AGGR_TYPE (record, 1); */
6129
6130 /* Dummy field to prevent alignment warnings. */
6131 space = build_decl (FIELD_DECL, NULL_TREE, integer_type_node);
6132 DECL_FIELD_CONTEXT (space) = record;
6133 DECL_ARTIFICIAL (space) = 1;
6134 DECL_IGNORED_P (space) = 1;
6135
6136 ofs = build_decl (FIELD_DECL, get_identifier ("__offset"),
6137 integer_type_node);
6138 DECL_FIELD_CONTEXT (ofs) = record;
6139 TREE_CHAIN (ofs) = space;
6140
6141 base = build_decl (FIELD_DECL, get_identifier ("__base"),
6142 ptr_type_node);
6143 DECL_FIELD_CONTEXT (base) = record;
6144 TREE_CHAIN (base) = ofs;
6145
6146 TYPE_FIELDS (record) = base;
6147 layout_type (record);
6148
6149 return record;
6150 }
6151
6152 /* Perform any needed actions needed for a function that is receiving a
6153 variable number of arguments. */
6154
6155 static void
6156 alpha_setup_incoming_varargs (CUMULATIVE_ARGS *pcum,
6157 enum machine_mode mode ATTRIBUTE_UNUSED,
6158 tree type ATTRIBUTE_UNUSED,
6159 int *pretend_size, int no_rtl)
6160 {
6161 #if TARGET_ABI_UNICOSMK
6162 /* On Unicos/Mk, the standard subroutine __T3E_MISMATCH stores all register
6163 arguments on the stack. Unfortunately, it doesn't always store the first
6164 one (i.e. the one that arrives in $16 or $f16). This is not a problem
6165 with stdargs as we always have at least one named argument there. */
6166 int num_reg_words = pcum->num_reg_words;
6167 if (num_reg_words < 6)
6168 {
6169 if (!no_rtl)
6170 {
6171 emit_insn (gen_umk_mismatch_args (GEN_INT (num_reg_words + 1)));
6172 emit_insn (gen_arg_home_umk ());
6173 }
6174 *pretend_size = 0;
6175 }
6176 #elif TARGET_ABI_OPEN_VMS
6177 /* For VMS, we allocate space for all 6 arg registers plus a count.
6178
6179 However, if NO registers need to be saved, don't allocate any space.
6180 This is not only because we won't need the space, but because AP
6181 includes the current_pretend_args_size and we don't want to mess up
6182 any ap-relative addresses already made. */
6183 if (pcum->num_args < 6)
6184 {
6185 if (!no_rtl)
6186 {
6187 emit_move_insn (gen_rtx_REG (DImode, 1), virtual_incoming_args_rtx);
6188 emit_insn (gen_arg_home ());
6189 }
6190 *pretend_size = 7 * UNITS_PER_WORD;
6191 }
6192 #else
6193 /* On OSF/1 and friends, we allocate space for all 12 arg registers, but
6194 only push those that are remaining. However, if NO registers need to
6195 be saved, don't allocate any space. This is not only because we won't
6196 need the space, but because AP includes the current_pretend_args_size
6197 and we don't want to mess up any ap-relative addresses already made.
6198
6199 If we are not to use the floating-point registers, save the integer
6200 registers where we would put the floating-point registers. This is
6201 not the most efficient way to implement varargs with just one register
6202 class, but it isn't worth doing anything more efficient in this rare
6203 case. */
6204 CUMULATIVE_ARGS cum = *pcum;
6205
6206 if (cum >= 6)
6207 return;
6208
6209 if (!no_rtl)
6210 {
6211 int set = get_varargs_alias_set ();
6212 rtx tmp;
6213
6214 tmp = gen_rtx_MEM (BLKmode,
6215 plus_constant (virtual_incoming_args_rtx,
6216 (cum + 6) * UNITS_PER_WORD));
6217 set_mem_alias_set (tmp, set);
6218 move_block_from_reg (16 + cum, tmp, 6 - cum);
6219
6220 tmp = gen_rtx_MEM (BLKmode,
6221 plus_constant (virtual_incoming_args_rtx,
6222 cum * UNITS_PER_WORD));
6223 set_mem_alias_set (tmp, set);
6224 move_block_from_reg (16 + (TARGET_FPREGS ? 32 : 0) + cum, tmp,
6225 6 - cum);
6226 }
6227 *pretend_size = 12 * UNITS_PER_WORD;
6228 #endif
6229 }
6230
6231 void
6232 alpha_va_start (tree valist, rtx nextarg ATTRIBUTE_UNUSED)
6233 {
6234 HOST_WIDE_INT offset;
6235 tree t, offset_field, base_field;
6236
6237 if (TREE_CODE (TREE_TYPE (valist)) == ERROR_MARK)
6238 return;
6239
6240 if (TARGET_ABI_UNICOSMK)
6241 std_expand_builtin_va_start (valist, nextarg);
6242
6243 /* For Unix, TARGET_SETUP_INCOMING_VARARGS moves the starting address base
6244 up by 48, storing fp arg registers in the first 48 bytes, and the
6245 integer arg registers in the next 48 bytes. This is only done,
6246 however, if any integer registers need to be stored.
6247
6248 If no integer registers need be stored, then we must subtract 48
6249 in order to account for the integer arg registers which are counted
6250 in argsize above, but which are not actually stored on the stack.
6251 Must further be careful here about structures straddling the last
6252 integer argument register; that futzes with pretend_args_size,
6253 which changes the meaning of AP. */
6254
6255 if (NUM_ARGS <= 6)
6256 offset = TARGET_ABI_OPEN_VMS ? UNITS_PER_WORD : 6 * UNITS_PER_WORD;
6257 else
6258 offset = -6 * UNITS_PER_WORD + current_function_pretend_args_size;
6259
6260 if (TARGET_ABI_OPEN_VMS)
6261 {
6262 nextarg = plus_constant (nextarg, offset);
6263 nextarg = plus_constant (nextarg, NUM_ARGS * UNITS_PER_WORD);
6264 t = build (MODIFY_EXPR, TREE_TYPE (valist), valist,
6265 make_tree (ptr_type_node, nextarg));
6266 TREE_SIDE_EFFECTS (t) = 1;
6267
6268 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6269 }
6270 else
6271 {
6272 base_field = TYPE_FIELDS (TREE_TYPE (valist));
6273 offset_field = TREE_CHAIN (base_field);
6274
6275 base_field = build (COMPONENT_REF, TREE_TYPE (base_field),
6276 valist, base_field);
6277 offset_field = build (COMPONENT_REF, TREE_TYPE (offset_field),
6278 valist, offset_field);
6279
6280 t = make_tree (ptr_type_node, virtual_incoming_args_rtx);
6281 t = build (PLUS_EXPR, ptr_type_node, t, build_int_2 (offset, 0));
6282 t = build (MODIFY_EXPR, TREE_TYPE (base_field), base_field, t);
6283 TREE_SIDE_EFFECTS (t) = 1;
6284 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6285
6286 t = build_int_2 (NUM_ARGS * UNITS_PER_WORD, 0);
6287 t = build (MODIFY_EXPR, TREE_TYPE (offset_field), offset_field, t);
6288 TREE_SIDE_EFFECTS (t) = 1;
6289 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6290 }
6291 }
6292
6293 rtx
6294 alpha_va_arg (tree valist, tree type)
6295 {
6296 rtx addr;
6297 tree t, type_size, rounded_size;
6298 tree offset_field, base_field, addr_tree, addend;
6299 tree wide_type, wide_ofs;
6300 int indirect = 0;
6301
6302 if (TARGET_ABI_OPEN_VMS || TARGET_ABI_UNICOSMK)
6303 return std_expand_builtin_va_arg (valist, type);
6304
6305 if (type == error_mark_node
6306 || (type_size = TYPE_SIZE_UNIT (TYPE_MAIN_VARIANT (type))) == NULL
6307 || TREE_OVERFLOW (type_size))
6308 rounded_size = size_zero_node;
6309 else
6310 rounded_size = fold (build (MULT_EXPR, sizetype,
6311 fold (build (TRUNC_DIV_EXPR, sizetype,
6312 fold (build (PLUS_EXPR, sizetype,
6313 type_size,
6314 size_int (7))),
6315 size_int (8))),
6316 size_int (8)));
6317
6318 base_field = TYPE_FIELDS (TREE_TYPE (valist));
6319 offset_field = TREE_CHAIN (base_field);
6320
6321 base_field = build (COMPONENT_REF, TREE_TYPE (base_field),
6322 valist, base_field);
6323 offset_field = build (COMPONENT_REF, TREE_TYPE (offset_field),
6324 valist, offset_field);
6325
6326 /* If the type could not be passed in registers, skip the block
6327 reserved for the registers. */
6328 if (MUST_PASS_IN_STACK (TYPE_MODE (type), type))
6329 {
6330 t = build (MODIFY_EXPR, TREE_TYPE (offset_field), offset_field,
6331 build (MAX_EXPR, TREE_TYPE (offset_field),
6332 offset_field, build_int_2 (6*8, 0)));
6333 TREE_SIDE_EFFECTS (t) = 1;
6334 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6335 }
6336
6337 wide_type = make_signed_type (64);
6338 wide_ofs = save_expr (build1 (CONVERT_EXPR, wide_type, offset_field));
6339
6340 addend = wide_ofs;
6341
6342 if (TYPE_MODE (type) == TFmode || TYPE_MODE (type) == TCmode)
6343 {
6344 indirect = 1;
6345 rounded_size = size_int (UNITS_PER_WORD);
6346 }
6347 else if (TREE_CODE (type) == COMPLEX_TYPE)
6348 {
6349 rtx real_part, imag_part, value, tmp;
6350
6351 real_part = alpha_va_arg (valist, TREE_TYPE (type));
6352 imag_part = alpha_va_arg (valist, TREE_TYPE (type));
6353
6354 /* ??? Most irritatingly, we're not returning the value here,
6355 but the address. Since real_part and imag_part are not
6356 necessarily contiguous, we must copy to local storage. */
6357
6358 real_part = gen_rtx_MEM (TYPE_MODE (TREE_TYPE (type)), real_part);
6359 imag_part = gen_rtx_MEM (TYPE_MODE (TREE_TYPE (type)), imag_part);
6360 value = gen_rtx_CONCAT (TYPE_MODE (type), real_part, imag_part);
6361
6362 tmp = assign_temp (type, 0, 1, 0);
6363 emit_move_insn (tmp, value);
6364
6365 return XEXP (tmp, 0);
6366 }
6367 else if (TREE_CODE (type) == REAL_TYPE)
6368 {
6369 tree fpaddend, cond;
6370
6371 fpaddend = fold (build (PLUS_EXPR, TREE_TYPE (addend),
6372 addend, build_int_2 (-6*8, 0)));
6373
6374 cond = fold (build (LT_EXPR, integer_type_node,
6375 wide_ofs, build_int_2 (6*8, 0)));
6376
6377 addend = fold (build (COND_EXPR, TREE_TYPE (addend), cond,
6378 fpaddend, addend));
6379 }
6380
6381 addr_tree = build (PLUS_EXPR, TREE_TYPE (base_field),
6382 base_field, addend);
6383
6384 addr = expand_expr (addr_tree, NULL_RTX, Pmode, EXPAND_NORMAL);
6385 addr = copy_to_reg (addr);
6386
6387 t = build (MODIFY_EXPR, TREE_TYPE (offset_field), offset_field,
6388 build (PLUS_EXPR, TREE_TYPE (offset_field),
6389 offset_field, rounded_size));
6390 TREE_SIDE_EFFECTS (t) = 1;
6391 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6392
6393 if (indirect)
6394 {
6395 addr = force_reg (Pmode, addr);
6396 addr = gen_rtx_MEM (Pmode, addr);
6397 }
6398
6399 return addr;
6400 }
6401 \f
6402 /* Builtins. */
6403
6404 enum alpha_builtin
6405 {
6406 ALPHA_BUILTIN_CMPBGE,
6407 ALPHA_BUILTIN_EXTBL,
6408 ALPHA_BUILTIN_EXTWL,
6409 ALPHA_BUILTIN_EXTLL,
6410 ALPHA_BUILTIN_EXTQL,
6411 ALPHA_BUILTIN_EXTWH,
6412 ALPHA_BUILTIN_EXTLH,
6413 ALPHA_BUILTIN_EXTQH,
6414 ALPHA_BUILTIN_INSBL,
6415 ALPHA_BUILTIN_INSWL,
6416 ALPHA_BUILTIN_INSLL,
6417 ALPHA_BUILTIN_INSQL,
6418 ALPHA_BUILTIN_INSWH,
6419 ALPHA_BUILTIN_INSLH,
6420 ALPHA_BUILTIN_INSQH,
6421 ALPHA_BUILTIN_MSKBL,
6422 ALPHA_BUILTIN_MSKWL,
6423 ALPHA_BUILTIN_MSKLL,
6424 ALPHA_BUILTIN_MSKQL,
6425 ALPHA_BUILTIN_MSKWH,
6426 ALPHA_BUILTIN_MSKLH,
6427 ALPHA_BUILTIN_MSKQH,
6428 ALPHA_BUILTIN_UMULH,
6429 ALPHA_BUILTIN_ZAP,
6430 ALPHA_BUILTIN_ZAPNOT,
6431 ALPHA_BUILTIN_AMASK,
6432 ALPHA_BUILTIN_IMPLVER,
6433 ALPHA_BUILTIN_RPCC,
6434 ALPHA_BUILTIN_THREAD_POINTER,
6435 ALPHA_BUILTIN_SET_THREAD_POINTER,
6436
6437 /* TARGET_MAX */
6438 ALPHA_BUILTIN_MINUB8,
6439 ALPHA_BUILTIN_MINSB8,
6440 ALPHA_BUILTIN_MINUW4,
6441 ALPHA_BUILTIN_MINSW4,
6442 ALPHA_BUILTIN_MAXUB8,
6443 ALPHA_BUILTIN_MAXSB8,
6444 ALPHA_BUILTIN_MAXUW4,
6445 ALPHA_BUILTIN_MAXSW4,
6446 ALPHA_BUILTIN_PERR,
6447 ALPHA_BUILTIN_PKLB,
6448 ALPHA_BUILTIN_PKWB,
6449 ALPHA_BUILTIN_UNPKBL,
6450 ALPHA_BUILTIN_UNPKBW,
6451
6452 /* TARGET_CIX */
6453 ALPHA_BUILTIN_CTTZ,
6454 ALPHA_BUILTIN_CTLZ,
6455 ALPHA_BUILTIN_CTPOP,
6456
6457 ALPHA_BUILTIN_max
6458 };
6459
6460 static unsigned int const code_for_builtin[ALPHA_BUILTIN_max] = {
6461 CODE_FOR_builtin_cmpbge,
6462 CODE_FOR_builtin_extbl,
6463 CODE_FOR_builtin_extwl,
6464 CODE_FOR_builtin_extll,
6465 CODE_FOR_builtin_extql,
6466 CODE_FOR_builtin_extwh,
6467 CODE_FOR_builtin_extlh,
6468 CODE_FOR_builtin_extqh,
6469 CODE_FOR_builtin_insbl,
6470 CODE_FOR_builtin_inswl,
6471 CODE_FOR_builtin_insll,
6472 CODE_FOR_builtin_insql,
6473 CODE_FOR_builtin_inswh,
6474 CODE_FOR_builtin_inslh,
6475 CODE_FOR_builtin_insqh,
6476 CODE_FOR_builtin_mskbl,
6477 CODE_FOR_builtin_mskwl,
6478 CODE_FOR_builtin_mskll,
6479 CODE_FOR_builtin_mskql,
6480 CODE_FOR_builtin_mskwh,
6481 CODE_FOR_builtin_msklh,
6482 CODE_FOR_builtin_mskqh,
6483 CODE_FOR_umuldi3_highpart,
6484 CODE_FOR_builtin_zap,
6485 CODE_FOR_builtin_zapnot,
6486 CODE_FOR_builtin_amask,
6487 CODE_FOR_builtin_implver,
6488 CODE_FOR_builtin_rpcc,
6489 CODE_FOR_load_tp,
6490 CODE_FOR_set_tp,
6491
6492 /* TARGET_MAX */
6493 CODE_FOR_builtin_minub8,
6494 CODE_FOR_builtin_minsb8,
6495 CODE_FOR_builtin_minuw4,
6496 CODE_FOR_builtin_minsw4,
6497 CODE_FOR_builtin_maxub8,
6498 CODE_FOR_builtin_maxsb8,
6499 CODE_FOR_builtin_maxuw4,
6500 CODE_FOR_builtin_maxsw4,
6501 CODE_FOR_builtin_perr,
6502 CODE_FOR_builtin_pklb,
6503 CODE_FOR_builtin_pkwb,
6504 CODE_FOR_builtin_unpkbl,
6505 CODE_FOR_builtin_unpkbw,
6506
6507 /* TARGET_CIX */
6508 CODE_FOR_builtin_cttz,
6509 CODE_FOR_builtin_ctlz,
6510 CODE_FOR_builtin_ctpop
6511 };
6512
6513 struct alpha_builtin_def
6514 {
6515 const char *name;
6516 enum alpha_builtin code;
6517 unsigned int target_mask;
6518 };
6519
6520 static struct alpha_builtin_def const zero_arg_builtins[] = {
6521 { "__builtin_alpha_implver", ALPHA_BUILTIN_IMPLVER, 0 },
6522 { "__builtin_alpha_rpcc", ALPHA_BUILTIN_RPCC, 0 }
6523 };
6524
6525 static struct alpha_builtin_def const one_arg_builtins[] = {
6526 { "__builtin_alpha_amask", ALPHA_BUILTIN_AMASK, 0 },
6527 { "__builtin_alpha_pklb", ALPHA_BUILTIN_PKLB, MASK_MAX },
6528 { "__builtin_alpha_pkwb", ALPHA_BUILTIN_PKWB, MASK_MAX },
6529 { "__builtin_alpha_unpkbl", ALPHA_BUILTIN_UNPKBL, MASK_MAX },
6530 { "__builtin_alpha_unpkbw", ALPHA_BUILTIN_UNPKBW, MASK_MAX },
6531 { "__builtin_alpha_cttz", ALPHA_BUILTIN_CTTZ, MASK_CIX },
6532 { "__builtin_alpha_ctlz", ALPHA_BUILTIN_CTLZ, MASK_CIX },
6533 { "__builtin_alpha_ctpop", ALPHA_BUILTIN_CTPOP, MASK_CIX }
6534 };
6535
6536 static struct alpha_builtin_def const two_arg_builtins[] = {
6537 { "__builtin_alpha_cmpbge", ALPHA_BUILTIN_CMPBGE, 0 },
6538 { "__builtin_alpha_extbl", ALPHA_BUILTIN_EXTBL, 0 },
6539 { "__builtin_alpha_extwl", ALPHA_BUILTIN_EXTWL, 0 },
6540 { "__builtin_alpha_extll", ALPHA_BUILTIN_EXTLL, 0 },
6541 { "__builtin_alpha_extql", ALPHA_BUILTIN_EXTQL, 0 },
6542 { "__builtin_alpha_extwh", ALPHA_BUILTIN_EXTWH, 0 },
6543 { "__builtin_alpha_extlh", ALPHA_BUILTIN_EXTLH, 0 },
6544 { "__builtin_alpha_extqh", ALPHA_BUILTIN_EXTQH, 0 },
6545 { "__builtin_alpha_insbl", ALPHA_BUILTIN_INSBL, 0 },
6546 { "__builtin_alpha_inswl", ALPHA_BUILTIN_INSWL, 0 },
6547 { "__builtin_alpha_insll", ALPHA_BUILTIN_INSLL, 0 },
6548 { "__builtin_alpha_insql", ALPHA_BUILTIN_INSQL, 0 },
6549 { "__builtin_alpha_inswh", ALPHA_BUILTIN_INSWH, 0 },
6550 { "__builtin_alpha_inslh", ALPHA_BUILTIN_INSLH, 0 },
6551 { "__builtin_alpha_insqh", ALPHA_BUILTIN_INSQH, 0 },
6552 { "__builtin_alpha_mskbl", ALPHA_BUILTIN_MSKBL, 0 },
6553 { "__builtin_alpha_mskwl", ALPHA_BUILTIN_MSKWL, 0 },
6554 { "__builtin_alpha_mskll", ALPHA_BUILTIN_MSKLL, 0 },
6555 { "__builtin_alpha_mskql", ALPHA_BUILTIN_MSKQL, 0 },
6556 { "__builtin_alpha_mskwh", ALPHA_BUILTIN_MSKWH, 0 },
6557 { "__builtin_alpha_msklh", ALPHA_BUILTIN_MSKLH, 0 },
6558 { "__builtin_alpha_mskqh", ALPHA_BUILTIN_MSKQH, 0 },
6559 { "__builtin_alpha_umulh", ALPHA_BUILTIN_UMULH, 0 },
6560 { "__builtin_alpha_zap", ALPHA_BUILTIN_ZAP, 0 },
6561 { "__builtin_alpha_zapnot", ALPHA_BUILTIN_ZAPNOT, 0 },
6562 { "__builtin_alpha_minub8", ALPHA_BUILTIN_MINUB8, MASK_MAX },
6563 { "__builtin_alpha_minsb8", ALPHA_BUILTIN_MINSB8, MASK_MAX },
6564 { "__builtin_alpha_minuw4", ALPHA_BUILTIN_MINUW4, MASK_MAX },
6565 { "__builtin_alpha_minsw4", ALPHA_BUILTIN_MINSW4, MASK_MAX },
6566 { "__builtin_alpha_maxub8", ALPHA_BUILTIN_MAXUB8, MASK_MAX },
6567 { "__builtin_alpha_maxsb8", ALPHA_BUILTIN_MAXSB8, MASK_MAX },
6568 { "__builtin_alpha_maxuw4", ALPHA_BUILTIN_MAXUW4, MASK_MAX },
6569 { "__builtin_alpha_maxsw4", ALPHA_BUILTIN_MAXSW4, MASK_MAX },
6570 { "__builtin_alpha_perr", ALPHA_BUILTIN_PERR, MASK_MAX }
6571 };
6572
6573 static void
6574 alpha_init_builtins (void)
6575 {
6576 const struct alpha_builtin_def *p;
6577 tree ftype;
6578 size_t i;
6579
6580 ftype = build_function_type (long_integer_type_node, void_list_node);
6581
6582 p = zero_arg_builtins;
6583 for (i = 0; i < ARRAY_SIZE (zero_arg_builtins); ++i, ++p)
6584 if ((target_flags & p->target_mask) == p->target_mask)
6585 builtin_function (p->name, ftype, p->code, BUILT_IN_MD,
6586 NULL, NULL_TREE);
6587
6588 ftype = build_function_type_list (long_integer_type_node,
6589 long_integer_type_node, NULL_TREE);
6590
6591 p = one_arg_builtins;
6592 for (i = 0; i < ARRAY_SIZE (one_arg_builtins); ++i, ++p)
6593 if ((target_flags & p->target_mask) == p->target_mask)
6594 builtin_function (p->name, ftype, p->code, BUILT_IN_MD,
6595 NULL, NULL_TREE);
6596
6597 ftype = build_function_type_list (long_integer_type_node,
6598 long_integer_type_node,
6599 long_integer_type_node, NULL_TREE);
6600
6601 p = two_arg_builtins;
6602 for (i = 0; i < ARRAY_SIZE (two_arg_builtins); ++i, ++p)
6603 if ((target_flags & p->target_mask) == p->target_mask)
6604 builtin_function (p->name, ftype, p->code, BUILT_IN_MD,
6605 NULL, NULL_TREE);
6606
6607 ftype = build_function_type (ptr_type_node, void_list_node);
6608 builtin_function ("__builtin_thread_pointer", ftype,
6609 ALPHA_BUILTIN_THREAD_POINTER, BUILT_IN_MD,
6610 NULL, NULL_TREE);
6611
6612 ftype = build_function_type_list (void_type_node, ptr_type_node, NULL_TREE);
6613 builtin_function ("__builtin_set_thread_pointer", ftype,
6614 ALPHA_BUILTIN_SET_THREAD_POINTER, BUILT_IN_MD,
6615 NULL, NULL_TREE);
6616 }
6617
6618 /* Expand an expression EXP that calls a built-in function,
6619 with result going to TARGET if that's convenient
6620 (and in mode MODE if that's convenient).
6621 SUBTARGET may be used as the target for computing one of EXP's operands.
6622 IGNORE is nonzero if the value is to be ignored. */
6623
6624 static rtx
6625 alpha_expand_builtin (tree exp, rtx target,
6626 rtx subtarget ATTRIBUTE_UNUSED,
6627 enum machine_mode mode ATTRIBUTE_UNUSED,
6628 int ignore ATTRIBUTE_UNUSED)
6629 {
6630 #define MAX_ARGS 2
6631
6632 tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0);
6633 unsigned int fcode = DECL_FUNCTION_CODE (fndecl);
6634 tree arglist = TREE_OPERAND (exp, 1);
6635 enum insn_code icode;
6636 rtx op[MAX_ARGS], pat;
6637 int arity;
6638 bool nonvoid;
6639
6640 if (fcode >= ALPHA_BUILTIN_max)
6641 internal_error ("bad builtin fcode");
6642 icode = code_for_builtin[fcode];
6643 if (icode == 0)
6644 internal_error ("bad builtin fcode");
6645
6646 nonvoid = TREE_TYPE (TREE_TYPE (fndecl)) != void_type_node;
6647
6648 for (arglist = TREE_OPERAND (exp, 1), arity = 0;
6649 arglist;
6650 arglist = TREE_CHAIN (arglist), arity++)
6651 {
6652 const struct insn_operand_data *insn_op;
6653
6654 tree arg = TREE_VALUE (arglist);
6655 if (arg == error_mark_node)
6656 return NULL_RTX;
6657 if (arity > MAX_ARGS)
6658 return NULL_RTX;
6659
6660 insn_op = &insn_data[icode].operand[arity + nonvoid];
6661
6662 op[arity] = expand_expr (arg, NULL_RTX, insn_op->mode, 0);
6663
6664 if (!(*insn_op->predicate) (op[arity], insn_op->mode))
6665 op[arity] = copy_to_mode_reg (insn_op->mode, op[arity]);
6666 }
6667
6668 if (nonvoid)
6669 {
6670 enum machine_mode tmode = insn_data[icode].operand[0].mode;
6671 if (!target
6672 || GET_MODE (target) != tmode
6673 || !(*insn_data[icode].operand[0].predicate) (target, tmode))
6674 target = gen_reg_rtx (tmode);
6675 }
6676
6677 switch (arity)
6678 {
6679 case 0:
6680 pat = GEN_FCN (icode) (target);
6681 break;
6682 case 1:
6683 if (nonvoid)
6684 pat = GEN_FCN (icode) (target, op[0]);
6685 else
6686 pat = GEN_FCN (icode) (op[0]);
6687 break;
6688 case 2:
6689 pat = GEN_FCN (icode) (target, op[0], op[1]);
6690 break;
6691 default:
6692 abort ();
6693 }
6694 if (!pat)
6695 return NULL_RTX;
6696 emit_insn (pat);
6697
6698 if (nonvoid)
6699 return target;
6700 else
6701 return const0_rtx;
6702 }
6703 \f
6704 /* This page contains routines that are used to determine what the function
6705 prologue and epilogue code will do and write them out. */
6706
6707 /* Compute the size of the save area in the stack. */
6708
6709 /* These variables are used for communication between the following functions.
6710 They indicate various things about the current function being compiled
6711 that are used to tell what kind of prologue, epilogue and procedure
6712 descriptor to generate. */
6713
6714 /* Nonzero if we need a stack procedure. */
6715 enum alpha_procedure_types {PT_NULL = 0, PT_REGISTER = 1, PT_STACK = 2};
6716 static enum alpha_procedure_types alpha_procedure_type;
6717
6718 /* Register number (either FP or SP) that is used to unwind the frame. */
6719 static int vms_unwind_regno;
6720
6721 /* Register number used to save FP. We need not have one for RA since
6722 we don't modify it for register procedures. This is only defined
6723 for register frame procedures. */
6724 static int vms_save_fp_regno;
6725
6726 /* Register number used to reference objects off our PV. */
6727 static int vms_base_regno;
6728
6729 /* Compute register masks for saved registers. */
6730
6731 static void
6732 alpha_sa_mask (unsigned long *imaskP, unsigned long *fmaskP)
6733 {
6734 unsigned long imask = 0;
6735 unsigned long fmask = 0;
6736 unsigned int i;
6737
6738 /* Irritatingly, there are two kinds of thunks -- those created with
6739 TARGET_ASM_OUTPUT_MI_THUNK and those with DECL_THUNK_P that go
6740 through the regular part of the compiler. In the
6741 TARGET_ASM_OUTPUT_MI_THUNK case we don't have valid register life
6742 info, but assemble_start_function wants to output .frame and
6743 .mask directives. */
6744 if (current_function_is_thunk && !no_new_pseudos)
6745 {
6746 *imaskP = 0;
6747 *fmaskP = 0;
6748 return;
6749 }
6750
6751 if (TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_STACK)
6752 imask |= (1UL << HARD_FRAME_POINTER_REGNUM);
6753
6754 /* One for every register we have to save. */
6755 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
6756 if (! fixed_regs[i] && ! call_used_regs[i]
6757 && regs_ever_live[i] && i != REG_RA
6758 && (!TARGET_ABI_UNICOSMK || i != HARD_FRAME_POINTER_REGNUM))
6759 {
6760 if (i < 32)
6761 imask |= (1UL << i);
6762 else
6763 fmask |= (1UL << (i - 32));
6764 }
6765
6766 /* We need to restore these for the handler. */
6767 if (current_function_calls_eh_return)
6768 {
6769 for (i = 0; ; ++i)
6770 {
6771 unsigned regno = EH_RETURN_DATA_REGNO (i);
6772 if (regno == INVALID_REGNUM)
6773 break;
6774 imask |= 1UL << regno;
6775 }
6776
6777 /* Glibc likes to use $31 as an unwind stopper for crt0. To
6778 avoid hackery in unwind-dw2.c, we need to actively store a
6779 zero in the prologue of _Unwind_RaiseException et al. */
6780 imask |= 1UL << 31;
6781 }
6782
6783 /* If any register spilled, then spill the return address also. */
6784 /* ??? This is required by the Digital stack unwind specification
6785 and isn't needed if we're doing Dwarf2 unwinding. */
6786 if (imask || fmask || alpha_ra_ever_killed ())
6787 imask |= (1UL << REG_RA);
6788
6789 *imaskP = imask;
6790 *fmaskP = fmask;
6791 }
6792
6793 int
6794 alpha_sa_size (void)
6795 {
6796 unsigned long mask[2];
6797 int sa_size = 0;
6798 int i, j;
6799
6800 alpha_sa_mask (&mask[0], &mask[1]);
6801
6802 if (TARGET_ABI_UNICOSMK)
6803 {
6804 if (mask[0] || mask[1])
6805 sa_size = 14;
6806 }
6807 else
6808 {
6809 for (j = 0; j < 2; ++j)
6810 for (i = 0; i < 32; ++i)
6811 if ((mask[j] >> i) & 1)
6812 sa_size++;
6813 }
6814
6815 if (TARGET_ABI_UNICOSMK)
6816 {
6817 /* We might not need to generate a frame if we don't make any calls
6818 (including calls to __T3E_MISMATCH if this is a vararg function),
6819 don't have any local variables which require stack slots, don't
6820 use alloca and have not determined that we need a frame for other
6821 reasons. */
6822
6823 alpha_procedure_type
6824 = (sa_size || get_frame_size() != 0
6825 || current_function_outgoing_args_size
6826 || current_function_stdarg || current_function_calls_alloca
6827 || frame_pointer_needed)
6828 ? PT_STACK : PT_REGISTER;
6829
6830 /* Always reserve space for saving callee-saved registers if we
6831 need a frame as required by the calling convention. */
6832 if (alpha_procedure_type == PT_STACK)
6833 sa_size = 14;
6834 }
6835 else if (TARGET_ABI_OPEN_VMS)
6836 {
6837 /* Start by assuming we can use a register procedure if we don't
6838 make any calls (REG_RA not used) or need to save any
6839 registers and a stack procedure if we do. */
6840 if ((mask[0] >> REG_RA) & 1)
6841 alpha_procedure_type = PT_STACK;
6842 else if (get_frame_size() != 0)
6843 alpha_procedure_type = PT_REGISTER;
6844 else
6845 alpha_procedure_type = PT_NULL;
6846
6847 /* Don't reserve space for saving FP & RA yet. Do that later after we've
6848 made the final decision on stack procedure vs register procedure. */
6849 if (alpha_procedure_type == PT_STACK)
6850 sa_size -= 2;
6851
6852 /* Decide whether to refer to objects off our PV via FP or PV.
6853 If we need FP for something else or if we receive a nonlocal
6854 goto (which expects PV to contain the value), we must use PV.
6855 Otherwise, start by assuming we can use FP. */
6856
6857 vms_base_regno
6858 = (frame_pointer_needed
6859 || current_function_has_nonlocal_label
6860 || alpha_procedure_type == PT_STACK
6861 || current_function_outgoing_args_size)
6862 ? REG_PV : HARD_FRAME_POINTER_REGNUM;
6863
6864 /* If we want to copy PV into FP, we need to find some register
6865 in which to save FP. */
6866
6867 vms_save_fp_regno = -1;
6868 if (vms_base_regno == HARD_FRAME_POINTER_REGNUM)
6869 for (i = 0; i < 32; i++)
6870 if (! fixed_regs[i] && call_used_regs[i] && ! regs_ever_live[i])
6871 vms_save_fp_regno = i;
6872
6873 if (vms_save_fp_regno == -1 && alpha_procedure_type == PT_REGISTER)
6874 vms_base_regno = REG_PV, alpha_procedure_type = PT_STACK;
6875 else if (alpha_procedure_type == PT_NULL)
6876 vms_base_regno = REG_PV;
6877
6878 /* Stack unwinding should be done via FP unless we use it for PV. */
6879 vms_unwind_regno = (vms_base_regno == REG_PV
6880 ? HARD_FRAME_POINTER_REGNUM : STACK_POINTER_REGNUM);
6881
6882 /* If this is a stack procedure, allow space for saving FP and RA. */
6883 if (alpha_procedure_type == PT_STACK)
6884 sa_size += 2;
6885 }
6886 else
6887 {
6888 /* Our size must be even (multiple of 16 bytes). */
6889 if (sa_size & 1)
6890 sa_size++;
6891 }
6892
6893 return sa_size * 8;
6894 }
6895
6896 /* Define the offset between two registers, one to be eliminated,
6897 and the other its replacement, at the start of a routine. */
6898
6899 HOST_WIDE_INT
6900 alpha_initial_elimination_offset (unsigned int from,
6901 unsigned int to ATTRIBUTE_UNUSED)
6902 {
6903 HOST_WIDE_INT ret;
6904
6905 ret = alpha_sa_size ();
6906 ret += ALPHA_ROUND (current_function_outgoing_args_size);
6907
6908 if (from == FRAME_POINTER_REGNUM)
6909 ;
6910 else if (from == ARG_POINTER_REGNUM)
6911 ret += (ALPHA_ROUND (get_frame_size ()
6912 + current_function_pretend_args_size)
6913 - current_function_pretend_args_size);
6914 else
6915 abort ();
6916
6917 return ret;
6918 }
6919
6920 int
6921 alpha_pv_save_size (void)
6922 {
6923 alpha_sa_size ();
6924 return alpha_procedure_type == PT_STACK ? 8 : 0;
6925 }
6926
6927 int
6928 alpha_using_fp (void)
6929 {
6930 alpha_sa_size ();
6931 return vms_unwind_regno == HARD_FRAME_POINTER_REGNUM;
6932 }
6933
6934 #if TARGET_ABI_OPEN_VMS
6935
6936 const struct attribute_spec vms_attribute_table[] =
6937 {
6938 /* { name, min_len, max_len, decl_req, type_req, fn_type_req, handler } */
6939 { "overlaid", 0, 0, true, false, false, NULL },
6940 { "global", 0, 0, true, false, false, NULL },
6941 { "initialize", 0, 0, true, false, false, NULL },
6942 { NULL, 0, 0, false, false, false, NULL }
6943 };
6944
6945 #endif
6946
6947 static int
6948 find_lo_sum_using_gp (rtx *px, void *data ATTRIBUTE_UNUSED)
6949 {
6950 return GET_CODE (*px) == LO_SUM && XEXP (*px, 0) == pic_offset_table_rtx;
6951 }
6952
6953 int
6954 alpha_find_lo_sum_using_gp (rtx insn)
6955 {
6956 return for_each_rtx (&PATTERN (insn), find_lo_sum_using_gp, NULL) > 0;
6957 }
6958
6959 static int
6960 alpha_does_function_need_gp (void)
6961 {
6962 rtx insn;
6963
6964 /* The GP being variable is an OSF abi thing. */
6965 if (! TARGET_ABI_OSF)
6966 return 0;
6967
6968 /* We need the gp to load the address of __mcount. */
6969 if (TARGET_PROFILING_NEEDS_GP && current_function_profile)
6970 return 1;
6971
6972 /* The code emitted by alpha_output_mi_thunk_osf uses the gp. */
6973 if (current_function_is_thunk)
6974 return 1;
6975
6976 /* The nonlocal receiver pattern assumes that the gp is valid for
6977 the nested function. Reasonable because it's almost always set
6978 correctly already. For the cases where that's wrong, make sure
6979 the nested function loads its gp on entry. */
6980 if (current_function_has_nonlocal_goto)
6981 return 1;
6982
6983 /* If we need a GP (we have a LDSYM insn or a CALL_INSN), load it first.
6984 Even if we are a static function, we still need to do this in case
6985 our address is taken and passed to something like qsort. */
6986
6987 push_topmost_sequence ();
6988 insn = get_insns ();
6989 pop_topmost_sequence ();
6990
6991 for (; insn; insn = NEXT_INSN (insn))
6992 if (INSN_P (insn)
6993 && GET_CODE (PATTERN (insn)) != USE
6994 && GET_CODE (PATTERN (insn)) != CLOBBER
6995 && get_attr_usegp (insn))
6996 return 1;
6997
6998 return 0;
6999 }
7000
7001 \f
7002 /* Helper function to set RTX_FRAME_RELATED_P on instructions, including
7003 sequences. */
7004
7005 static rtx
7006 set_frame_related_p (void)
7007 {
7008 rtx seq = get_insns ();
7009 rtx insn;
7010
7011 end_sequence ();
7012
7013 if (!seq)
7014 return NULL_RTX;
7015
7016 if (INSN_P (seq))
7017 {
7018 insn = seq;
7019 while (insn != NULL_RTX)
7020 {
7021 RTX_FRAME_RELATED_P (insn) = 1;
7022 insn = NEXT_INSN (insn);
7023 }
7024 seq = emit_insn (seq);
7025 }
7026 else
7027 {
7028 seq = emit_insn (seq);
7029 RTX_FRAME_RELATED_P (seq) = 1;
7030 }
7031 return seq;
7032 }
7033
7034 #define FRP(exp) (start_sequence (), exp, set_frame_related_p ())
7035
7036 /* Write function prologue. */
7037
7038 /* On vms we have two kinds of functions:
7039
7040 - stack frame (PROC_STACK)
7041 these are 'normal' functions with local vars and which are
7042 calling other functions
7043 - register frame (PROC_REGISTER)
7044 keeps all data in registers, needs no stack
7045
7046 We must pass this to the assembler so it can generate the
7047 proper pdsc (procedure descriptor)
7048 This is done with the '.pdesc' command.
7049
7050 On not-vms, we don't really differentiate between the two, as we can
7051 simply allocate stack without saving registers. */
7052
7053 void
7054 alpha_expand_prologue (void)
7055 {
7056 /* Registers to save. */
7057 unsigned long imask = 0;
7058 unsigned long fmask = 0;
7059 /* Stack space needed for pushing registers clobbered by us. */
7060 HOST_WIDE_INT sa_size;
7061 /* Complete stack size needed. */
7062 HOST_WIDE_INT frame_size;
7063 /* Offset from base reg to register save area. */
7064 HOST_WIDE_INT reg_offset;
7065 rtx sa_reg, mem;
7066 int i;
7067
7068 sa_size = alpha_sa_size ();
7069
7070 frame_size = get_frame_size ();
7071 if (TARGET_ABI_OPEN_VMS)
7072 frame_size = ALPHA_ROUND (sa_size
7073 + (alpha_procedure_type == PT_STACK ? 8 : 0)
7074 + frame_size
7075 + current_function_pretend_args_size);
7076 else if (TARGET_ABI_UNICOSMK)
7077 /* We have to allocate space for the DSIB if we generate a frame. */
7078 frame_size = ALPHA_ROUND (sa_size
7079 + (alpha_procedure_type == PT_STACK ? 48 : 0))
7080 + ALPHA_ROUND (frame_size
7081 + current_function_outgoing_args_size);
7082 else
7083 frame_size = (ALPHA_ROUND (current_function_outgoing_args_size)
7084 + sa_size
7085 + ALPHA_ROUND (frame_size
7086 + current_function_pretend_args_size));
7087
7088 if (TARGET_ABI_OPEN_VMS)
7089 reg_offset = 8;
7090 else
7091 reg_offset = ALPHA_ROUND (current_function_outgoing_args_size);
7092
7093 alpha_sa_mask (&imask, &fmask);
7094
7095 /* Emit an insn to reload GP, if needed. */
7096 if (TARGET_ABI_OSF)
7097 {
7098 alpha_function_needs_gp = alpha_does_function_need_gp ();
7099 if (alpha_function_needs_gp)
7100 emit_insn (gen_prologue_ldgp ());
7101 }
7102
7103 /* TARGET_PROFILING_NEEDS_GP actually implies that we need to insert
7104 the call to mcount ourselves, rather than having the linker do it
7105 magically in response to -pg. Since _mcount has special linkage,
7106 don't represent the call as a call. */
7107 if (TARGET_PROFILING_NEEDS_GP && current_function_profile)
7108 emit_insn (gen_prologue_mcount ());
7109
7110 if (TARGET_ABI_UNICOSMK)
7111 unicosmk_gen_dsib (&imask);
7112
7113 /* Adjust the stack by the frame size. If the frame size is > 4096
7114 bytes, we need to be sure we probe somewhere in the first and last
7115 4096 bytes (we can probably get away without the latter test) and
7116 every 8192 bytes in between. If the frame size is > 32768, we
7117 do this in a loop. Otherwise, we generate the explicit probe
7118 instructions.
7119
7120 Note that we are only allowed to adjust sp once in the prologue. */
7121
7122 if (frame_size <= 32768)
7123 {
7124 if (frame_size > 4096)
7125 {
7126 int probed = 4096;
7127
7128 do
7129 emit_insn (gen_probe_stack (GEN_INT (TARGET_ABI_UNICOSMK
7130 ? -probed + 64
7131 : -probed)));
7132 while ((probed += 8192) < frame_size);
7133
7134 /* We only have to do this probe if we aren't saving registers. */
7135 if (sa_size == 0 && probed + 4096 < frame_size)
7136 emit_insn (gen_probe_stack (GEN_INT (-frame_size)));
7137 }
7138
7139 if (frame_size != 0)
7140 FRP (emit_insn (gen_adddi3 (stack_pointer_rtx, stack_pointer_rtx,
7141 GEN_INT (TARGET_ABI_UNICOSMK
7142 ? -frame_size + 64
7143 : -frame_size))));
7144 }
7145 else
7146 {
7147 /* Here we generate code to set R22 to SP + 4096 and set R23 to the
7148 number of 8192 byte blocks to probe. We then probe each block
7149 in the loop and then set SP to the proper location. If the
7150 amount remaining is > 4096, we have to do one more probe if we
7151 are not saving any registers. */
7152
7153 HOST_WIDE_INT blocks = (frame_size + 4096) / 8192;
7154 HOST_WIDE_INT leftover = frame_size + 4096 - blocks * 8192;
7155 rtx ptr = gen_rtx_REG (DImode, 22);
7156 rtx count = gen_rtx_REG (DImode, 23);
7157 rtx seq;
7158
7159 emit_move_insn (count, GEN_INT (blocks));
7160 emit_insn (gen_adddi3 (ptr, stack_pointer_rtx,
7161 GEN_INT (TARGET_ABI_UNICOSMK ? 4096 - 64 : 4096)));
7162
7163 /* Because of the difficulty in emitting a new basic block this
7164 late in the compilation, generate the loop as a single insn. */
7165 emit_insn (gen_prologue_stack_probe_loop (count, ptr));
7166
7167 if (leftover > 4096 && sa_size == 0)
7168 {
7169 rtx last = gen_rtx_MEM (DImode, plus_constant (ptr, -leftover));
7170 MEM_VOLATILE_P (last) = 1;
7171 emit_move_insn (last, const0_rtx);
7172 }
7173
7174 if (TARGET_ABI_WINDOWS_NT)
7175 {
7176 /* For NT stack unwind (done by 'reverse execution'), it's
7177 not OK to take the result of a loop, even though the value
7178 is already in ptr, so we reload it via a single operation
7179 and subtract it to sp.
7180
7181 Yes, that's correct -- we have to reload the whole constant
7182 into a temporary via ldah+lda then subtract from sp. */
7183
7184 HOST_WIDE_INT lo, hi;
7185 lo = ((frame_size & 0xffff) ^ 0x8000) - 0x8000;
7186 hi = frame_size - lo;
7187
7188 emit_move_insn (ptr, GEN_INT (hi));
7189 emit_insn (gen_adddi3 (ptr, ptr, GEN_INT (lo)));
7190 seq = emit_insn (gen_subdi3 (stack_pointer_rtx, stack_pointer_rtx,
7191 ptr));
7192 }
7193 else
7194 {
7195 seq = emit_insn (gen_adddi3 (stack_pointer_rtx, ptr,
7196 GEN_INT (-leftover)));
7197 }
7198
7199 /* This alternative is special, because the DWARF code cannot
7200 possibly intuit through the loop above. So we invent this
7201 note it looks at instead. */
7202 RTX_FRAME_RELATED_P (seq) = 1;
7203 REG_NOTES (seq)
7204 = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR,
7205 gen_rtx_SET (VOIDmode, stack_pointer_rtx,
7206 gen_rtx_PLUS (Pmode, stack_pointer_rtx,
7207 GEN_INT (TARGET_ABI_UNICOSMK
7208 ? -frame_size + 64
7209 : -frame_size))),
7210 REG_NOTES (seq));
7211 }
7212
7213 if (!TARGET_ABI_UNICOSMK)
7214 {
7215 /* Cope with very large offsets to the register save area. */
7216 sa_reg = stack_pointer_rtx;
7217 if (reg_offset + sa_size > 0x8000)
7218 {
7219 int low = ((reg_offset & 0xffff) ^ 0x8000) - 0x8000;
7220 HOST_WIDE_INT bias;
7221
7222 if (low + sa_size <= 0x8000)
7223 bias = reg_offset - low, reg_offset = low;
7224 else
7225 bias = reg_offset, reg_offset = 0;
7226
7227 sa_reg = gen_rtx_REG (DImode, 24);
7228 FRP (emit_insn (gen_adddi3 (sa_reg, stack_pointer_rtx,
7229 GEN_INT (bias))));
7230 }
7231
7232 /* Save regs in stack order. Beginning with VMS PV. */
7233 if (TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_STACK)
7234 {
7235 mem = gen_rtx_MEM (DImode, stack_pointer_rtx);
7236 set_mem_alias_set (mem, alpha_sr_alias_set);
7237 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, REG_PV)));
7238 }
7239
7240 /* Save register RA next. */
7241 if (imask & (1UL << REG_RA))
7242 {
7243 mem = gen_rtx_MEM (DImode, plus_constant (sa_reg, reg_offset));
7244 set_mem_alias_set (mem, alpha_sr_alias_set);
7245 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, REG_RA)));
7246 imask &= ~(1UL << REG_RA);
7247 reg_offset += 8;
7248 }
7249
7250 /* Now save any other registers required to be saved. */
7251 for (i = 0; i < 31; i++)
7252 if (imask & (1UL << i))
7253 {
7254 mem = gen_rtx_MEM (DImode, plus_constant (sa_reg, reg_offset));
7255 set_mem_alias_set (mem, alpha_sr_alias_set);
7256 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, i)));
7257 reg_offset += 8;
7258 }
7259
7260 /* Store a zero if requested for unwinding. */
7261 if (imask & (1UL << 31))
7262 {
7263 rtx insn, t;
7264
7265 mem = gen_rtx_MEM (DImode, plus_constant (sa_reg, reg_offset));
7266 set_mem_alias_set (mem, alpha_sr_alias_set);
7267 insn = emit_move_insn (mem, const0_rtx);
7268
7269 RTX_FRAME_RELATED_P (insn) = 1;
7270 t = gen_rtx_REG (Pmode, 31);
7271 t = gen_rtx_SET (VOIDmode, mem, t);
7272 t = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, t, REG_NOTES (insn));
7273 REG_NOTES (insn) = t;
7274
7275 reg_offset += 8;
7276 }
7277
7278 for (i = 0; i < 31; i++)
7279 if (fmask & (1UL << i))
7280 {
7281 mem = gen_rtx_MEM (DFmode, plus_constant (sa_reg, reg_offset));
7282 set_mem_alias_set (mem, alpha_sr_alias_set);
7283 FRP (emit_move_insn (mem, gen_rtx_REG (DFmode, i+32)));
7284 reg_offset += 8;
7285 }
7286 }
7287 else if (TARGET_ABI_UNICOSMK && alpha_procedure_type == PT_STACK)
7288 {
7289 /* The standard frame on the T3E includes space for saving registers.
7290 We just have to use it. We don't have to save the return address and
7291 the old frame pointer here - they are saved in the DSIB. */
7292
7293 reg_offset = -56;
7294 for (i = 9; i < 15; i++)
7295 if (imask & (1UL << i))
7296 {
7297 mem = gen_rtx_MEM (DImode, plus_constant(hard_frame_pointer_rtx,
7298 reg_offset));
7299 set_mem_alias_set (mem, alpha_sr_alias_set);
7300 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, i)));
7301 reg_offset -= 8;
7302 }
7303 for (i = 2; i < 10; i++)
7304 if (fmask & (1UL << i))
7305 {
7306 mem = gen_rtx_MEM (DFmode, plus_constant (hard_frame_pointer_rtx,
7307 reg_offset));
7308 set_mem_alias_set (mem, alpha_sr_alias_set);
7309 FRP (emit_move_insn (mem, gen_rtx_REG (DFmode, i+32)));
7310 reg_offset -= 8;
7311 }
7312 }
7313
7314 if (TARGET_ABI_OPEN_VMS)
7315 {
7316 if (alpha_procedure_type == PT_REGISTER)
7317 /* Register frame procedures save the fp.
7318 ?? Ought to have a dwarf2 save for this. */
7319 emit_move_insn (gen_rtx_REG (DImode, vms_save_fp_regno),
7320 hard_frame_pointer_rtx);
7321
7322 if (alpha_procedure_type != PT_NULL && vms_base_regno != REG_PV)
7323 emit_insn (gen_force_movdi (gen_rtx_REG (DImode, vms_base_regno),
7324 gen_rtx_REG (DImode, REG_PV)));
7325
7326 if (alpha_procedure_type != PT_NULL
7327 && vms_unwind_regno == HARD_FRAME_POINTER_REGNUM)
7328 FRP (emit_move_insn (hard_frame_pointer_rtx, stack_pointer_rtx));
7329
7330 /* If we have to allocate space for outgoing args, do it now. */
7331 if (current_function_outgoing_args_size != 0)
7332 {
7333 rtx seq
7334 = emit_move_insn (stack_pointer_rtx,
7335 plus_constant
7336 (hard_frame_pointer_rtx,
7337 - (ALPHA_ROUND
7338 (current_function_outgoing_args_size))));
7339
7340 /* Only set FRAME_RELATED_P on the stack adjustment we just emitted
7341 if ! frame_pointer_needed. Setting the bit will change the CFA
7342 computation rule to use sp again, which would be wrong if we had
7343 frame_pointer_needed, as this means sp might move unpredictably
7344 later on.
7345
7346 Also, note that
7347 frame_pointer_needed
7348 => vms_unwind_regno == HARD_FRAME_POINTER_REGNUM
7349 and
7350 current_function_outgoing_args_size != 0
7351 => alpha_procedure_type != PT_NULL,
7352
7353 so when we are not setting the bit here, we are guaranteed to
7354 have emitted an FRP frame pointer update just before. */
7355 RTX_FRAME_RELATED_P (seq) = ! frame_pointer_needed;
7356 }
7357 }
7358 else if (!TARGET_ABI_UNICOSMK)
7359 {
7360 /* If we need a frame pointer, set it from the stack pointer. */
7361 if (frame_pointer_needed)
7362 {
7363 if (TARGET_CAN_FAULT_IN_PROLOGUE)
7364 FRP (emit_move_insn (hard_frame_pointer_rtx, stack_pointer_rtx));
7365 else
7366 /* This must always be the last instruction in the
7367 prologue, thus we emit a special move + clobber. */
7368 FRP (emit_insn (gen_init_fp (hard_frame_pointer_rtx,
7369 stack_pointer_rtx, sa_reg)));
7370 }
7371 }
7372
7373 /* The ABIs for VMS and OSF/1 say that while we can schedule insns into
7374 the prologue, for exception handling reasons, we cannot do this for
7375 any insn that might fault. We could prevent this for mems with a
7376 (clobber:BLK (scratch)), but this doesn't work for fp insns. So we
7377 have to prevent all such scheduling with a blockage.
7378
7379 Linux, on the other hand, never bothered to implement OSF/1's
7380 exception handling, and so doesn't care about such things. Anyone
7381 planning to use dwarf2 frame-unwind info can also omit the blockage. */
7382
7383 if (! TARGET_CAN_FAULT_IN_PROLOGUE)
7384 emit_insn (gen_blockage ());
7385 }
7386
7387 /* Output the textual info surrounding the prologue. */
7388
7389 void
7390 alpha_start_function (FILE *file, const char *fnname,
7391 tree decl ATTRIBUTE_UNUSED)
7392 {
7393 unsigned long imask = 0;
7394 unsigned long fmask = 0;
7395 /* Stack space needed for pushing registers clobbered by us. */
7396 HOST_WIDE_INT sa_size;
7397 /* Complete stack size needed. */
7398 unsigned HOST_WIDE_INT frame_size;
7399 /* Offset from base reg to register save area. */
7400 HOST_WIDE_INT reg_offset;
7401 char *entry_label = (char *) alloca (strlen (fnname) + 6);
7402 int i;
7403
7404 /* Don't emit an extern directive for functions defined in the same file. */
7405 if (TARGET_ABI_UNICOSMK)
7406 {
7407 tree name_tree;
7408 name_tree = get_identifier (fnname);
7409 TREE_ASM_WRITTEN (name_tree) = 1;
7410 }
7411
7412 alpha_fnname = fnname;
7413 sa_size = alpha_sa_size ();
7414
7415 frame_size = get_frame_size ();
7416 if (TARGET_ABI_OPEN_VMS)
7417 frame_size = ALPHA_ROUND (sa_size
7418 + (alpha_procedure_type == PT_STACK ? 8 : 0)
7419 + frame_size
7420 + current_function_pretend_args_size);
7421 else if (TARGET_ABI_UNICOSMK)
7422 frame_size = ALPHA_ROUND (sa_size
7423 + (alpha_procedure_type == PT_STACK ? 48 : 0))
7424 + ALPHA_ROUND (frame_size
7425 + current_function_outgoing_args_size);
7426 else
7427 frame_size = (ALPHA_ROUND (current_function_outgoing_args_size)
7428 + sa_size
7429 + ALPHA_ROUND (frame_size
7430 + current_function_pretend_args_size));
7431
7432 if (TARGET_ABI_OPEN_VMS)
7433 reg_offset = 8;
7434 else
7435 reg_offset = ALPHA_ROUND (current_function_outgoing_args_size);
7436
7437 alpha_sa_mask (&imask, &fmask);
7438
7439 /* Ecoff can handle multiple .file directives, so put out file and lineno.
7440 We have to do that before the .ent directive as we cannot switch
7441 files within procedures with native ecoff because line numbers are
7442 linked to procedure descriptors.
7443 Outputting the lineno helps debugging of one line functions as they
7444 would otherwise get no line number at all. Please note that we would
7445 like to put out last_linenum from final.c, but it is not accessible. */
7446
7447 if (write_symbols == SDB_DEBUG)
7448 {
7449 #ifdef ASM_OUTPUT_SOURCE_FILENAME
7450 ASM_OUTPUT_SOURCE_FILENAME (file,
7451 DECL_SOURCE_FILE (current_function_decl));
7452 #endif
7453 #ifdef ASM_OUTPUT_SOURCE_LINE
7454 if (debug_info_level != DINFO_LEVEL_TERSE)
7455 ASM_OUTPUT_SOURCE_LINE (file,
7456 DECL_SOURCE_LINE (current_function_decl), 0);
7457 #endif
7458 }
7459
7460 /* Issue function start and label. */
7461 if (TARGET_ABI_OPEN_VMS
7462 || (!TARGET_ABI_UNICOSMK && !flag_inhibit_size_directive))
7463 {
7464 fputs ("\t.ent ", file);
7465 assemble_name (file, fnname);
7466 putc ('\n', file);
7467
7468 /* If the function needs GP, we'll write the "..ng" label there.
7469 Otherwise, do it here. */
7470 if (TARGET_ABI_OSF
7471 && ! alpha_function_needs_gp
7472 && ! current_function_is_thunk)
7473 {
7474 putc ('$', file);
7475 assemble_name (file, fnname);
7476 fputs ("..ng:\n", file);
7477 }
7478 }
7479
7480 strcpy (entry_label, fnname);
7481 if (TARGET_ABI_OPEN_VMS)
7482 strcat (entry_label, "..en");
7483
7484 /* For public functions, the label must be globalized by appending an
7485 additional colon. */
7486 if (TARGET_ABI_UNICOSMK && TREE_PUBLIC (decl))
7487 strcat (entry_label, ":");
7488
7489 ASM_OUTPUT_LABEL (file, entry_label);
7490 inside_function = TRUE;
7491
7492 if (TARGET_ABI_OPEN_VMS)
7493 fprintf (file, "\t.base $%d\n", vms_base_regno);
7494
7495 if (!TARGET_ABI_OPEN_VMS && !TARGET_ABI_UNICOSMK && TARGET_IEEE_CONFORMANT
7496 && !flag_inhibit_size_directive)
7497 {
7498 /* Set flags in procedure descriptor to request IEEE-conformant
7499 math-library routines. The value we set it to is PDSC_EXC_IEEE
7500 (/usr/include/pdsc.h). */
7501 fputs ("\t.eflag 48\n", file);
7502 }
7503
7504 /* Set up offsets to alpha virtual arg/local debugging pointer. */
7505 alpha_auto_offset = -frame_size + current_function_pretend_args_size;
7506 alpha_arg_offset = -frame_size + 48;
7507
7508 /* Describe our frame. If the frame size is larger than an integer,
7509 print it as zero to avoid an assembler error. We won't be
7510 properly describing such a frame, but that's the best we can do. */
7511 if (TARGET_ABI_UNICOSMK)
7512 ;
7513 else if (TARGET_ABI_OPEN_VMS)
7514 fprintf (file, "\t.frame $%d," HOST_WIDE_INT_PRINT_DEC ",$26,"
7515 HOST_WIDE_INT_PRINT_DEC "\n",
7516 vms_unwind_regno,
7517 frame_size >= (1UL << 31) ? 0 : frame_size,
7518 reg_offset);
7519 else if (!flag_inhibit_size_directive)
7520 fprintf (file, "\t.frame $%d," HOST_WIDE_INT_PRINT_DEC ",$26,%d\n",
7521 (frame_pointer_needed
7522 ? HARD_FRAME_POINTER_REGNUM : STACK_POINTER_REGNUM),
7523 frame_size >= (1UL << 31) ? 0 : frame_size,
7524 current_function_pretend_args_size);
7525
7526 /* Describe which registers were spilled. */
7527 if (TARGET_ABI_UNICOSMK)
7528 ;
7529 else if (TARGET_ABI_OPEN_VMS)
7530 {
7531 if (imask)
7532 /* ??? Does VMS care if mask contains ra? The old code didn't
7533 set it, so I don't here. */
7534 fprintf (file, "\t.mask 0x%lx,0\n", imask & ~(1UL << REG_RA));
7535 if (fmask)
7536 fprintf (file, "\t.fmask 0x%lx,0\n", fmask);
7537 if (alpha_procedure_type == PT_REGISTER)
7538 fprintf (file, "\t.fp_save $%d\n", vms_save_fp_regno);
7539 }
7540 else if (!flag_inhibit_size_directive)
7541 {
7542 if (imask)
7543 {
7544 fprintf (file, "\t.mask 0x%lx," HOST_WIDE_INT_PRINT_DEC "\n", imask,
7545 frame_size >= (1UL << 31) ? 0 : reg_offset - frame_size);
7546
7547 for (i = 0; i < 32; ++i)
7548 if (imask & (1UL << i))
7549 reg_offset += 8;
7550 }
7551
7552 if (fmask)
7553 fprintf (file, "\t.fmask 0x%lx," HOST_WIDE_INT_PRINT_DEC "\n", fmask,
7554 frame_size >= (1UL << 31) ? 0 : reg_offset - frame_size);
7555 }
7556
7557 #if TARGET_ABI_OPEN_VMS
7558 /* Ifdef'ed cause link_section are only available then. */
7559 readonly_data_section ();
7560 fprintf (file, "\t.align 3\n");
7561 assemble_name (file, fnname); fputs ("..na:\n", file);
7562 fputs ("\t.ascii \"", file);
7563 assemble_name (file, fnname);
7564 fputs ("\\0\"\n", file);
7565 alpha_need_linkage (fnname, 1);
7566 text_section ();
7567 #endif
7568 }
7569
7570 /* Emit the .prologue note at the scheduled end of the prologue. */
7571
7572 static void
7573 alpha_output_function_end_prologue (FILE *file)
7574 {
7575 if (TARGET_ABI_UNICOSMK)
7576 ;
7577 else if (TARGET_ABI_OPEN_VMS)
7578 fputs ("\t.prologue\n", file);
7579 else if (TARGET_ABI_WINDOWS_NT)
7580 fputs ("\t.prologue 0\n", file);
7581 else if (!flag_inhibit_size_directive)
7582 fprintf (file, "\t.prologue %d\n",
7583 alpha_function_needs_gp || current_function_is_thunk);
7584 }
7585
7586 /* Write function epilogue. */
7587
7588 /* ??? At some point we will want to support full unwind, and so will
7589 need to mark the epilogue as well. At the moment, we just confuse
7590 dwarf2out. */
7591 #undef FRP
7592 #define FRP(exp) exp
7593
7594 void
7595 alpha_expand_epilogue (void)
7596 {
7597 /* Registers to save. */
7598 unsigned long imask = 0;
7599 unsigned long fmask = 0;
7600 /* Stack space needed for pushing registers clobbered by us. */
7601 HOST_WIDE_INT sa_size;
7602 /* Complete stack size needed. */
7603 HOST_WIDE_INT frame_size;
7604 /* Offset from base reg to register save area. */
7605 HOST_WIDE_INT reg_offset;
7606 int fp_is_frame_pointer, fp_offset;
7607 rtx sa_reg, sa_reg_exp = NULL;
7608 rtx sp_adj1, sp_adj2, mem;
7609 rtx eh_ofs;
7610 int i;
7611
7612 sa_size = alpha_sa_size ();
7613
7614 frame_size = get_frame_size ();
7615 if (TARGET_ABI_OPEN_VMS)
7616 frame_size = ALPHA_ROUND (sa_size
7617 + (alpha_procedure_type == PT_STACK ? 8 : 0)
7618 + frame_size
7619 + current_function_pretend_args_size);
7620 else if (TARGET_ABI_UNICOSMK)
7621 frame_size = ALPHA_ROUND (sa_size
7622 + (alpha_procedure_type == PT_STACK ? 48 : 0))
7623 + ALPHA_ROUND (frame_size
7624 + current_function_outgoing_args_size);
7625 else
7626 frame_size = (ALPHA_ROUND (current_function_outgoing_args_size)
7627 + sa_size
7628 + ALPHA_ROUND (frame_size
7629 + current_function_pretend_args_size));
7630
7631 if (TARGET_ABI_OPEN_VMS)
7632 {
7633 if (alpha_procedure_type == PT_STACK)
7634 reg_offset = 8;
7635 else
7636 reg_offset = 0;
7637 }
7638 else
7639 reg_offset = ALPHA_ROUND (current_function_outgoing_args_size);
7640
7641 alpha_sa_mask (&imask, &fmask);
7642
7643 fp_is_frame_pointer
7644 = ((TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_STACK)
7645 || (!TARGET_ABI_OPEN_VMS && frame_pointer_needed));
7646 fp_offset = 0;
7647 sa_reg = stack_pointer_rtx;
7648
7649 if (current_function_calls_eh_return)
7650 eh_ofs = EH_RETURN_STACKADJ_RTX;
7651 else
7652 eh_ofs = NULL_RTX;
7653
7654 if (!TARGET_ABI_UNICOSMK && sa_size)
7655 {
7656 /* If we have a frame pointer, restore SP from it. */
7657 if ((TARGET_ABI_OPEN_VMS
7658 && vms_unwind_regno == HARD_FRAME_POINTER_REGNUM)
7659 || (!TARGET_ABI_OPEN_VMS && frame_pointer_needed))
7660 FRP (emit_move_insn (stack_pointer_rtx, hard_frame_pointer_rtx));
7661
7662 /* Cope with very large offsets to the register save area. */
7663 if (reg_offset + sa_size > 0x8000)
7664 {
7665 int low = ((reg_offset & 0xffff) ^ 0x8000) - 0x8000;
7666 HOST_WIDE_INT bias;
7667
7668 if (low + sa_size <= 0x8000)
7669 bias = reg_offset - low, reg_offset = low;
7670 else
7671 bias = reg_offset, reg_offset = 0;
7672
7673 sa_reg = gen_rtx_REG (DImode, 22);
7674 sa_reg_exp = plus_constant (stack_pointer_rtx, bias);
7675
7676 FRP (emit_move_insn (sa_reg, sa_reg_exp));
7677 }
7678
7679 /* Restore registers in order, excepting a true frame pointer. */
7680
7681 mem = gen_rtx_MEM (DImode, plus_constant (sa_reg, reg_offset));
7682 if (! eh_ofs)
7683 set_mem_alias_set (mem, alpha_sr_alias_set);
7684 FRP (emit_move_insn (gen_rtx_REG (DImode, REG_RA), mem));
7685
7686 reg_offset += 8;
7687 imask &= ~(1UL << REG_RA);
7688
7689 for (i = 0; i < 31; ++i)
7690 if (imask & (1UL << i))
7691 {
7692 if (i == HARD_FRAME_POINTER_REGNUM && fp_is_frame_pointer)
7693 fp_offset = reg_offset;
7694 else
7695 {
7696 mem = gen_rtx_MEM (DImode, plus_constant(sa_reg, reg_offset));
7697 set_mem_alias_set (mem, alpha_sr_alias_set);
7698 FRP (emit_move_insn (gen_rtx_REG (DImode, i), mem));
7699 }
7700 reg_offset += 8;
7701 }
7702
7703 if (imask & (1UL << 31))
7704 reg_offset += 8;
7705
7706 for (i = 0; i < 31; ++i)
7707 if (fmask & (1UL << i))
7708 {
7709 mem = gen_rtx_MEM (DFmode, plus_constant(sa_reg, reg_offset));
7710 set_mem_alias_set (mem, alpha_sr_alias_set);
7711 FRP (emit_move_insn (gen_rtx_REG (DFmode, i+32), mem));
7712 reg_offset += 8;
7713 }
7714 }
7715 else if (TARGET_ABI_UNICOSMK && alpha_procedure_type == PT_STACK)
7716 {
7717 /* Restore callee-saved general-purpose registers. */
7718
7719 reg_offset = -56;
7720
7721 for (i = 9; i < 15; i++)
7722 if (imask & (1UL << i))
7723 {
7724 mem = gen_rtx_MEM (DImode, plus_constant(hard_frame_pointer_rtx,
7725 reg_offset));
7726 set_mem_alias_set (mem, alpha_sr_alias_set);
7727 FRP (emit_move_insn (gen_rtx_REG (DImode, i), mem));
7728 reg_offset -= 8;
7729 }
7730
7731 for (i = 2; i < 10; i++)
7732 if (fmask & (1UL << i))
7733 {
7734 mem = gen_rtx_MEM (DFmode, plus_constant(hard_frame_pointer_rtx,
7735 reg_offset));
7736 set_mem_alias_set (mem, alpha_sr_alias_set);
7737 FRP (emit_move_insn (gen_rtx_REG (DFmode, i+32), mem));
7738 reg_offset -= 8;
7739 }
7740
7741 /* Restore the return address from the DSIB. */
7742
7743 mem = gen_rtx_MEM (DImode, plus_constant(hard_frame_pointer_rtx, -8));
7744 set_mem_alias_set (mem, alpha_sr_alias_set);
7745 FRP (emit_move_insn (gen_rtx_REG (DImode, REG_RA), mem));
7746 }
7747
7748 if (frame_size || eh_ofs)
7749 {
7750 sp_adj1 = stack_pointer_rtx;
7751
7752 if (eh_ofs)
7753 {
7754 sp_adj1 = gen_rtx_REG (DImode, 23);
7755 emit_move_insn (sp_adj1,
7756 gen_rtx_PLUS (Pmode, stack_pointer_rtx, eh_ofs));
7757 }
7758
7759 /* If the stack size is large, begin computation into a temporary
7760 register so as not to interfere with a potential fp restore,
7761 which must be consecutive with an SP restore. */
7762 if (frame_size < 32768
7763 && ! (TARGET_ABI_UNICOSMK && current_function_calls_alloca))
7764 sp_adj2 = GEN_INT (frame_size);
7765 else if (TARGET_ABI_UNICOSMK)
7766 {
7767 sp_adj1 = gen_rtx_REG (DImode, 23);
7768 FRP (emit_move_insn (sp_adj1, hard_frame_pointer_rtx));
7769 sp_adj2 = const0_rtx;
7770 }
7771 else if (frame_size < 0x40007fffL)
7772 {
7773 int low = ((frame_size & 0xffff) ^ 0x8000) - 0x8000;
7774
7775 sp_adj2 = plus_constant (sp_adj1, frame_size - low);
7776 if (sa_reg_exp && rtx_equal_p (sa_reg_exp, sp_adj2))
7777 sp_adj1 = sa_reg;
7778 else
7779 {
7780 sp_adj1 = gen_rtx_REG (DImode, 23);
7781 FRP (emit_move_insn (sp_adj1, sp_adj2));
7782 }
7783 sp_adj2 = GEN_INT (low);
7784 }
7785 else
7786 {
7787 rtx tmp = gen_rtx_REG (DImode, 23);
7788 FRP (sp_adj2 = alpha_emit_set_const (tmp, DImode, frame_size, 3));
7789 if (!sp_adj2)
7790 {
7791 /* We can't drop new things to memory this late, afaik,
7792 so build it up by pieces. */
7793 FRP (sp_adj2 = alpha_emit_set_long_const (tmp, frame_size,
7794 -(frame_size < 0)));
7795 if (!sp_adj2)
7796 abort ();
7797 }
7798 }
7799
7800 /* From now on, things must be in order. So emit blockages. */
7801
7802 /* Restore the frame pointer. */
7803 if (TARGET_ABI_UNICOSMK)
7804 {
7805 emit_insn (gen_blockage ());
7806 mem = gen_rtx_MEM (DImode,
7807 plus_constant (hard_frame_pointer_rtx, -16));
7808 set_mem_alias_set (mem, alpha_sr_alias_set);
7809 FRP (emit_move_insn (hard_frame_pointer_rtx, mem));
7810 }
7811 else if (fp_is_frame_pointer)
7812 {
7813 emit_insn (gen_blockage ());
7814 mem = gen_rtx_MEM (DImode, plus_constant (sa_reg, fp_offset));
7815 set_mem_alias_set (mem, alpha_sr_alias_set);
7816 FRP (emit_move_insn (hard_frame_pointer_rtx, mem));
7817 }
7818 else if (TARGET_ABI_OPEN_VMS)
7819 {
7820 emit_insn (gen_blockage ());
7821 FRP (emit_move_insn (hard_frame_pointer_rtx,
7822 gen_rtx_REG (DImode, vms_save_fp_regno)));
7823 }
7824
7825 /* Restore the stack pointer. */
7826 emit_insn (gen_blockage ());
7827 if (sp_adj2 == const0_rtx)
7828 FRP (emit_move_insn (stack_pointer_rtx, sp_adj1));
7829 else
7830 FRP (emit_move_insn (stack_pointer_rtx,
7831 gen_rtx_PLUS (DImode, sp_adj1, sp_adj2)));
7832 }
7833 else
7834 {
7835 if (TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_REGISTER)
7836 {
7837 emit_insn (gen_blockage ());
7838 FRP (emit_move_insn (hard_frame_pointer_rtx,
7839 gen_rtx_REG (DImode, vms_save_fp_regno)));
7840 }
7841 else if (TARGET_ABI_UNICOSMK && alpha_procedure_type != PT_STACK)
7842 {
7843 /* Decrement the frame pointer if the function does not have a
7844 frame. */
7845
7846 emit_insn (gen_blockage ());
7847 FRP (emit_insn (gen_adddi3 (hard_frame_pointer_rtx,
7848 hard_frame_pointer_rtx, GEN_INT (-1))));
7849 }
7850 }
7851 }
7852 \f
7853 /* Output the rest of the textual info surrounding the epilogue. */
7854
7855 void
7856 alpha_end_function (FILE *file, const char *fnname, tree decl ATTRIBUTE_UNUSED)
7857 {
7858 /* End the function. */
7859 if (!TARGET_ABI_UNICOSMK && !flag_inhibit_size_directive)
7860 {
7861 fputs ("\t.end ", file);
7862 assemble_name (file, fnname);
7863 putc ('\n', file);
7864 }
7865 inside_function = FALSE;
7866
7867 #if TARGET_ABI_OPEN_VMS
7868 alpha_write_linkage (file, fnname, decl);
7869 #endif
7870
7871 /* Output jump tables and the static subroutine information block. */
7872 if (TARGET_ABI_UNICOSMK)
7873 {
7874 unicosmk_output_ssib (file, fnname);
7875 unicosmk_output_deferred_case_vectors (file);
7876 }
7877 }
7878
7879 #if TARGET_ABI_OSF
7880 /* Emit a tail call to FUNCTION after adjusting THIS by DELTA.
7881
7882 In order to avoid the hordes of differences between generated code
7883 with and without TARGET_EXPLICIT_RELOCS, and to avoid duplicating
7884 lots of code loading up large constants, generate rtl and emit it
7885 instead of going straight to text.
7886
7887 Not sure why this idea hasn't been explored before... */
7888
7889 static void
7890 alpha_output_mi_thunk_osf (FILE *file, tree thunk_fndecl ATTRIBUTE_UNUSED,
7891 HOST_WIDE_INT delta, HOST_WIDE_INT vcall_offset,
7892 tree function)
7893 {
7894 HOST_WIDE_INT hi, lo;
7895 rtx this, insn, funexp;
7896
7897 /* We always require a valid GP. */
7898 emit_insn (gen_prologue_ldgp ());
7899 emit_note (NOTE_INSN_PROLOGUE_END);
7900
7901 /* Find the "this" pointer. If the function returns a structure,
7902 the structure return pointer is in $16. */
7903 if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function))
7904 this = gen_rtx_REG (Pmode, 17);
7905 else
7906 this = gen_rtx_REG (Pmode, 16);
7907
7908 /* Add DELTA. When possible we use ldah+lda. Otherwise load the
7909 entire constant for the add. */
7910 lo = ((delta & 0xffff) ^ 0x8000) - 0x8000;
7911 hi = (((delta - lo) & 0xffffffff) ^ 0x80000000) - 0x80000000;
7912 if (hi + lo == delta)
7913 {
7914 if (hi)
7915 emit_insn (gen_adddi3 (this, this, GEN_INT (hi)));
7916 if (lo)
7917 emit_insn (gen_adddi3 (this, this, GEN_INT (lo)));
7918 }
7919 else
7920 {
7921 rtx tmp = alpha_emit_set_long_const (gen_rtx_REG (Pmode, 0),
7922 delta, -(delta < 0));
7923 emit_insn (gen_adddi3 (this, this, tmp));
7924 }
7925
7926 /* Add a delta stored in the vtable at VCALL_OFFSET. */
7927 if (vcall_offset)
7928 {
7929 rtx tmp, tmp2;
7930
7931 tmp = gen_rtx_REG (Pmode, 0);
7932 emit_move_insn (tmp, gen_rtx_MEM (Pmode, this));
7933
7934 lo = ((vcall_offset & 0xffff) ^ 0x8000) - 0x8000;
7935 hi = (((vcall_offset - lo) & 0xffffffff) ^ 0x80000000) - 0x80000000;
7936 if (hi + lo == vcall_offset)
7937 {
7938 if (hi)
7939 emit_insn (gen_adddi3 (tmp, tmp, GEN_INT (hi)));
7940 }
7941 else
7942 {
7943 tmp2 = alpha_emit_set_long_const (gen_rtx_REG (Pmode, 1),
7944 vcall_offset, -(vcall_offset < 0));
7945 emit_insn (gen_adddi3 (tmp, tmp, tmp2));
7946 lo = 0;
7947 }
7948 if (lo)
7949 tmp2 = gen_rtx_PLUS (Pmode, tmp, GEN_INT (lo));
7950 else
7951 tmp2 = tmp;
7952 emit_move_insn (tmp, gen_rtx_MEM (Pmode, tmp2));
7953
7954 emit_insn (gen_adddi3 (this, this, tmp));
7955 }
7956
7957 /* Generate a tail call to the target function. */
7958 if (! TREE_USED (function))
7959 {
7960 assemble_external (function);
7961 TREE_USED (function) = 1;
7962 }
7963 funexp = XEXP (DECL_RTL (function), 0);
7964 funexp = gen_rtx_MEM (FUNCTION_MODE, funexp);
7965 insn = emit_call_insn (gen_sibcall (funexp, const0_rtx));
7966 SIBLING_CALL_P (insn) = 1;
7967
7968 /* Run just enough of rest_of_compilation to get the insns emitted.
7969 There's not really enough bulk here to make other passes such as
7970 instruction scheduling worth while. Note that use_thunk calls
7971 assemble_start_function and assemble_end_function. */
7972 insn = get_insns ();
7973 insn_locators_initialize ();
7974 shorten_branches (insn);
7975 final_start_function (insn, file, 1);
7976 final (insn, file, 1, 0);
7977 final_end_function ();
7978 }
7979 #endif /* TARGET_ABI_OSF */
7980 \f
7981 /* Debugging support. */
7982
7983 #include "gstab.h"
7984
7985 /* Count the number of sdb related labels are generated (to find block
7986 start and end boundaries). */
7987
7988 int sdb_label_count = 0;
7989
7990 /* Next label # for each statement. */
7991
7992 static int sym_lineno = 0;
7993
7994 /* Count the number of .file directives, so that .loc is up to date. */
7995
7996 static int num_source_filenames = 0;
7997
7998 /* Name of the file containing the current function. */
7999
8000 static const char *current_function_file = "";
8001
8002 /* Offsets to alpha virtual arg/local debugging pointers. */
8003
8004 long alpha_arg_offset;
8005 long alpha_auto_offset;
8006 \f
8007 /* Emit a new filename to a stream. */
8008
8009 void
8010 alpha_output_filename (FILE *stream, const char *name)
8011 {
8012 static int first_time = TRUE;
8013 char ltext_label_name[100];
8014
8015 if (first_time)
8016 {
8017 first_time = FALSE;
8018 ++num_source_filenames;
8019 current_function_file = name;
8020 fprintf (stream, "\t.file\t%d ", num_source_filenames);
8021 output_quoted_string (stream, name);
8022 fprintf (stream, "\n");
8023 if (!TARGET_GAS && write_symbols == DBX_DEBUG)
8024 fprintf (stream, "\t#@stabs\n");
8025 }
8026
8027 else if (write_symbols == DBX_DEBUG)
8028 {
8029 ASM_GENERATE_INTERNAL_LABEL (ltext_label_name, "Ltext", 0);
8030 fprintf (stream, "%s", ASM_STABS_OP);
8031 output_quoted_string (stream, name);
8032 fprintf (stream, ",%d,0,0,%s\n", N_SOL, &ltext_label_name[1]);
8033 }
8034
8035 else if (name != current_function_file
8036 && strcmp (name, current_function_file) != 0)
8037 {
8038 if (inside_function && ! TARGET_GAS)
8039 fprintf (stream, "\t#.file\t%d ", num_source_filenames);
8040 else
8041 {
8042 ++num_source_filenames;
8043 current_function_file = name;
8044 fprintf (stream, "\t.file\t%d ", num_source_filenames);
8045 }
8046
8047 output_quoted_string (stream, name);
8048 fprintf (stream, "\n");
8049 }
8050 }
8051
8052 /* Emit a linenumber to a stream. */
8053
8054 void
8055 alpha_output_lineno (FILE *stream, int line)
8056 {
8057 if (write_symbols == DBX_DEBUG)
8058 {
8059 /* mips-tfile doesn't understand .stabd directives. */
8060 ++sym_lineno;
8061 fprintf (stream, "$LM%d:\n%s%d,0,%d,$LM%d\n",
8062 sym_lineno, ASM_STABN_OP, N_SLINE, line, sym_lineno);
8063 }
8064 else
8065 fprintf (stream, "\n\t.loc\t%d %d\n", num_source_filenames, line);
8066 }
8067 \f
8068 /* Structure to show the current status of registers and memory. */
8069
8070 struct shadow_summary
8071 {
8072 struct {
8073 unsigned int i : 31; /* Mask of int regs */
8074 unsigned int fp : 31; /* Mask of fp regs */
8075 unsigned int mem : 1; /* mem == imem | fpmem */
8076 } used, defd;
8077 };
8078
8079 /* Summary the effects of expression X on the machine. Update SUM, a pointer
8080 to the summary structure. SET is nonzero if the insn is setting the
8081 object, otherwise zero. */
8082
8083 static void
8084 summarize_insn (rtx x, struct shadow_summary *sum, int set)
8085 {
8086 const char *format_ptr;
8087 int i, j;
8088
8089 if (x == 0)
8090 return;
8091
8092 switch (GET_CODE (x))
8093 {
8094 /* ??? Note that this case would be incorrect if the Alpha had a
8095 ZERO_EXTRACT in SET_DEST. */
8096 case SET:
8097 summarize_insn (SET_SRC (x), sum, 0);
8098 summarize_insn (SET_DEST (x), sum, 1);
8099 break;
8100
8101 case CLOBBER:
8102 summarize_insn (XEXP (x, 0), sum, 1);
8103 break;
8104
8105 case USE:
8106 summarize_insn (XEXP (x, 0), sum, 0);
8107 break;
8108
8109 case ASM_OPERANDS:
8110 for (i = ASM_OPERANDS_INPUT_LENGTH (x) - 1; i >= 0; i--)
8111 summarize_insn (ASM_OPERANDS_INPUT (x, i), sum, 0);
8112 break;
8113
8114 case PARALLEL:
8115 for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
8116 summarize_insn (XVECEXP (x, 0, i), sum, 0);
8117 break;
8118
8119 case SUBREG:
8120 summarize_insn (SUBREG_REG (x), sum, 0);
8121 break;
8122
8123 case REG:
8124 {
8125 int regno = REGNO (x);
8126 unsigned long mask = ((unsigned long) 1) << (regno % 32);
8127
8128 if (regno == 31 || regno == 63)
8129 break;
8130
8131 if (set)
8132 {
8133 if (regno < 32)
8134 sum->defd.i |= mask;
8135 else
8136 sum->defd.fp |= mask;
8137 }
8138 else
8139 {
8140 if (regno < 32)
8141 sum->used.i |= mask;
8142 else
8143 sum->used.fp |= mask;
8144 }
8145 }
8146 break;
8147
8148 case MEM:
8149 if (set)
8150 sum->defd.mem = 1;
8151 else
8152 sum->used.mem = 1;
8153
8154 /* Find the regs used in memory address computation: */
8155 summarize_insn (XEXP (x, 0), sum, 0);
8156 break;
8157
8158 case CONST_INT: case CONST_DOUBLE:
8159 case SYMBOL_REF: case LABEL_REF: case CONST:
8160 case SCRATCH: case ASM_INPUT:
8161 break;
8162
8163 /* Handle common unary and binary ops for efficiency. */
8164 case COMPARE: case PLUS: case MINUS: case MULT: case DIV:
8165 case MOD: case UDIV: case UMOD: case AND: case IOR:
8166 case XOR: case ASHIFT: case ROTATE: case ASHIFTRT: case LSHIFTRT:
8167 case ROTATERT: case SMIN: case SMAX: case UMIN: case UMAX:
8168 case NE: case EQ: case GE: case GT: case LE:
8169 case LT: case GEU: case GTU: case LEU: case LTU:
8170 summarize_insn (XEXP (x, 0), sum, 0);
8171 summarize_insn (XEXP (x, 1), sum, 0);
8172 break;
8173
8174 case NEG: case NOT: case SIGN_EXTEND: case ZERO_EXTEND:
8175 case TRUNCATE: case FLOAT_EXTEND: case FLOAT_TRUNCATE: case FLOAT:
8176 case FIX: case UNSIGNED_FLOAT: case UNSIGNED_FIX: case ABS:
8177 case SQRT: case FFS:
8178 summarize_insn (XEXP (x, 0), sum, 0);
8179 break;
8180
8181 default:
8182 format_ptr = GET_RTX_FORMAT (GET_CODE (x));
8183 for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
8184 switch (format_ptr[i])
8185 {
8186 case 'e':
8187 summarize_insn (XEXP (x, i), sum, 0);
8188 break;
8189
8190 case 'E':
8191 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
8192 summarize_insn (XVECEXP (x, i, j), sum, 0);
8193 break;
8194
8195 case 'i':
8196 break;
8197
8198 default:
8199 abort ();
8200 }
8201 }
8202 }
8203
8204 /* Ensure a sufficient number of `trapb' insns are in the code when
8205 the user requests code with a trap precision of functions or
8206 instructions.
8207
8208 In naive mode, when the user requests a trap-precision of
8209 "instruction", a trapb is needed after every instruction that may
8210 generate a trap. This ensures that the code is resumption safe but
8211 it is also slow.
8212
8213 When optimizations are turned on, we delay issuing a trapb as long
8214 as possible. In this context, a trap shadow is the sequence of
8215 instructions that starts with a (potentially) trap generating
8216 instruction and extends to the next trapb or call_pal instruction
8217 (but GCC never generates call_pal by itself). We can delay (and
8218 therefore sometimes omit) a trapb subject to the following
8219 conditions:
8220
8221 (a) On entry to the trap shadow, if any Alpha register or memory
8222 location contains a value that is used as an operand value by some
8223 instruction in the trap shadow (live on entry), then no instruction
8224 in the trap shadow may modify the register or memory location.
8225
8226 (b) Within the trap shadow, the computation of the base register
8227 for a memory load or store instruction may not involve using the
8228 result of an instruction that might generate an UNPREDICTABLE
8229 result.
8230
8231 (c) Within the trap shadow, no register may be used more than once
8232 as a destination register. (This is to make life easier for the
8233 trap-handler.)
8234
8235 (d) The trap shadow may not include any branch instructions. */
8236
8237 static void
8238 alpha_handle_trap_shadows (void)
8239 {
8240 struct shadow_summary shadow;
8241 int trap_pending, exception_nesting;
8242 rtx i, n;
8243
8244 trap_pending = 0;
8245 exception_nesting = 0;
8246 shadow.used.i = 0;
8247 shadow.used.fp = 0;
8248 shadow.used.mem = 0;
8249 shadow.defd = shadow.used;
8250
8251 for (i = get_insns (); i ; i = NEXT_INSN (i))
8252 {
8253 if (GET_CODE (i) == NOTE)
8254 {
8255 switch (NOTE_LINE_NUMBER (i))
8256 {
8257 case NOTE_INSN_EH_REGION_BEG:
8258 exception_nesting++;
8259 if (trap_pending)
8260 goto close_shadow;
8261 break;
8262
8263 case NOTE_INSN_EH_REGION_END:
8264 exception_nesting--;
8265 if (trap_pending)
8266 goto close_shadow;
8267 break;
8268
8269 case NOTE_INSN_EPILOGUE_BEG:
8270 if (trap_pending && alpha_tp >= ALPHA_TP_FUNC)
8271 goto close_shadow;
8272 break;
8273 }
8274 }
8275 else if (trap_pending)
8276 {
8277 if (alpha_tp == ALPHA_TP_FUNC)
8278 {
8279 if (GET_CODE (i) == JUMP_INSN
8280 && GET_CODE (PATTERN (i)) == RETURN)
8281 goto close_shadow;
8282 }
8283 else if (alpha_tp == ALPHA_TP_INSN)
8284 {
8285 if (optimize > 0)
8286 {
8287 struct shadow_summary sum;
8288
8289 sum.used.i = 0;
8290 sum.used.fp = 0;
8291 sum.used.mem = 0;
8292 sum.defd = sum.used;
8293
8294 switch (GET_CODE (i))
8295 {
8296 case INSN:
8297 /* Annoyingly, get_attr_trap will abort on these. */
8298 if (GET_CODE (PATTERN (i)) == USE
8299 || GET_CODE (PATTERN (i)) == CLOBBER)
8300 break;
8301
8302 summarize_insn (PATTERN (i), &sum, 0);
8303
8304 if ((sum.defd.i & shadow.defd.i)
8305 || (sum.defd.fp & shadow.defd.fp))
8306 {
8307 /* (c) would be violated */
8308 goto close_shadow;
8309 }
8310
8311 /* Combine shadow with summary of current insn: */
8312 shadow.used.i |= sum.used.i;
8313 shadow.used.fp |= sum.used.fp;
8314 shadow.used.mem |= sum.used.mem;
8315 shadow.defd.i |= sum.defd.i;
8316 shadow.defd.fp |= sum.defd.fp;
8317 shadow.defd.mem |= sum.defd.mem;
8318
8319 if ((sum.defd.i & shadow.used.i)
8320 || (sum.defd.fp & shadow.used.fp)
8321 || (sum.defd.mem & shadow.used.mem))
8322 {
8323 /* (a) would be violated (also takes care of (b)) */
8324 if (get_attr_trap (i) == TRAP_YES
8325 && ((sum.defd.i & sum.used.i)
8326 || (sum.defd.fp & sum.used.fp)))
8327 abort ();
8328
8329 goto close_shadow;
8330 }
8331 break;
8332
8333 case JUMP_INSN:
8334 case CALL_INSN:
8335 case CODE_LABEL:
8336 goto close_shadow;
8337
8338 default:
8339 abort ();
8340 }
8341 }
8342 else
8343 {
8344 close_shadow:
8345 n = emit_insn_before (gen_trapb (), i);
8346 PUT_MODE (n, TImode);
8347 PUT_MODE (i, TImode);
8348 trap_pending = 0;
8349 shadow.used.i = 0;
8350 shadow.used.fp = 0;
8351 shadow.used.mem = 0;
8352 shadow.defd = shadow.used;
8353 }
8354 }
8355 }
8356
8357 if ((exception_nesting > 0 || alpha_tp >= ALPHA_TP_FUNC)
8358 && GET_CODE (i) == INSN
8359 && GET_CODE (PATTERN (i)) != USE
8360 && GET_CODE (PATTERN (i)) != CLOBBER
8361 && get_attr_trap (i) == TRAP_YES)
8362 {
8363 if (optimize && !trap_pending)
8364 summarize_insn (PATTERN (i), &shadow, 0);
8365 trap_pending = 1;
8366 }
8367 }
8368 }
8369 \f
8370 /* Alpha can only issue instruction groups simultaneously if they are
8371 suitably aligned. This is very processor-specific. */
8372
8373 enum alphaev4_pipe {
8374 EV4_STOP = 0,
8375 EV4_IB0 = 1,
8376 EV4_IB1 = 2,
8377 EV4_IBX = 4
8378 };
8379
8380 enum alphaev5_pipe {
8381 EV5_STOP = 0,
8382 EV5_NONE = 1,
8383 EV5_E01 = 2,
8384 EV5_E0 = 4,
8385 EV5_E1 = 8,
8386 EV5_FAM = 16,
8387 EV5_FA = 32,
8388 EV5_FM = 64
8389 };
8390
8391 static enum alphaev4_pipe
8392 alphaev4_insn_pipe (rtx insn)
8393 {
8394 if (recog_memoized (insn) < 0)
8395 return EV4_STOP;
8396 if (get_attr_length (insn) != 4)
8397 return EV4_STOP;
8398
8399 switch (get_attr_type (insn))
8400 {
8401 case TYPE_ILD:
8402 case TYPE_FLD:
8403 return EV4_IBX;
8404
8405 case TYPE_LDSYM:
8406 case TYPE_IADD:
8407 case TYPE_ILOG:
8408 case TYPE_ICMOV:
8409 case TYPE_ICMP:
8410 case TYPE_IST:
8411 case TYPE_FST:
8412 case TYPE_SHIFT:
8413 case TYPE_IMUL:
8414 case TYPE_FBR:
8415 return EV4_IB0;
8416
8417 case TYPE_MISC:
8418 case TYPE_IBR:
8419 case TYPE_JSR:
8420 case TYPE_CALLPAL:
8421 case TYPE_FCPYS:
8422 case TYPE_FCMOV:
8423 case TYPE_FADD:
8424 case TYPE_FDIV:
8425 case TYPE_FMUL:
8426 return EV4_IB1;
8427
8428 default:
8429 abort ();
8430 }
8431 }
8432
8433 static enum alphaev5_pipe
8434 alphaev5_insn_pipe (rtx insn)
8435 {
8436 if (recog_memoized (insn) < 0)
8437 return EV5_STOP;
8438 if (get_attr_length (insn) != 4)
8439 return EV5_STOP;
8440
8441 switch (get_attr_type (insn))
8442 {
8443 case TYPE_ILD:
8444 case TYPE_FLD:
8445 case TYPE_LDSYM:
8446 case TYPE_IADD:
8447 case TYPE_ILOG:
8448 case TYPE_ICMOV:
8449 case TYPE_ICMP:
8450 return EV5_E01;
8451
8452 case TYPE_IST:
8453 case TYPE_FST:
8454 case TYPE_SHIFT:
8455 case TYPE_IMUL:
8456 case TYPE_MISC:
8457 case TYPE_MVI:
8458 return EV5_E0;
8459
8460 case TYPE_IBR:
8461 case TYPE_JSR:
8462 case TYPE_CALLPAL:
8463 return EV5_E1;
8464
8465 case TYPE_FCPYS:
8466 return EV5_FAM;
8467
8468 case TYPE_FBR:
8469 case TYPE_FCMOV:
8470 case TYPE_FADD:
8471 case TYPE_FDIV:
8472 return EV5_FA;
8473
8474 case TYPE_FMUL:
8475 return EV5_FM;
8476
8477 default:
8478 abort();
8479 }
8480 }
8481
8482 /* IN_USE is a mask of the slots currently filled within the insn group.
8483 The mask bits come from alphaev4_pipe above. If EV4_IBX is set, then
8484 the insn in EV4_IB0 can be swapped by the hardware into EV4_IB1.
8485
8486 LEN is, of course, the length of the group in bytes. */
8487
8488 static rtx
8489 alphaev4_next_group (rtx insn, int *pin_use, int *plen)
8490 {
8491 int len, in_use;
8492
8493 len = in_use = 0;
8494
8495 if (! INSN_P (insn)
8496 || GET_CODE (PATTERN (insn)) == CLOBBER
8497 || GET_CODE (PATTERN (insn)) == USE)
8498 goto next_and_done;
8499
8500 while (1)
8501 {
8502 enum alphaev4_pipe pipe;
8503
8504 pipe = alphaev4_insn_pipe (insn);
8505 switch (pipe)
8506 {
8507 case EV4_STOP:
8508 /* Force complex instructions to start new groups. */
8509 if (in_use)
8510 goto done;
8511
8512 /* If this is a completely unrecognized insn, its an asm.
8513 We don't know how long it is, so record length as -1 to
8514 signal a needed realignment. */
8515 if (recog_memoized (insn) < 0)
8516 len = -1;
8517 else
8518 len = get_attr_length (insn);
8519 goto next_and_done;
8520
8521 case EV4_IBX:
8522 if (in_use & EV4_IB0)
8523 {
8524 if (in_use & EV4_IB1)
8525 goto done;
8526 in_use |= EV4_IB1;
8527 }
8528 else
8529 in_use |= EV4_IB0 | EV4_IBX;
8530 break;
8531
8532 case EV4_IB0:
8533 if (in_use & EV4_IB0)
8534 {
8535 if (!(in_use & EV4_IBX) || (in_use & EV4_IB1))
8536 goto done;
8537 in_use |= EV4_IB1;
8538 }
8539 in_use |= EV4_IB0;
8540 break;
8541
8542 case EV4_IB1:
8543 if (in_use & EV4_IB1)
8544 goto done;
8545 in_use |= EV4_IB1;
8546 break;
8547
8548 default:
8549 abort();
8550 }
8551 len += 4;
8552
8553 /* Haifa doesn't do well scheduling branches. */
8554 if (GET_CODE (insn) == JUMP_INSN)
8555 goto next_and_done;
8556
8557 next:
8558 insn = next_nonnote_insn (insn);
8559
8560 if (!insn || ! INSN_P (insn))
8561 goto done;
8562
8563 /* Let Haifa tell us where it thinks insn group boundaries are. */
8564 if (GET_MODE (insn) == TImode)
8565 goto done;
8566
8567 if (GET_CODE (insn) == CLOBBER || GET_CODE (insn) == USE)
8568 goto next;
8569 }
8570
8571 next_and_done:
8572 insn = next_nonnote_insn (insn);
8573
8574 done:
8575 *plen = len;
8576 *pin_use = in_use;
8577 return insn;
8578 }
8579
8580 /* IN_USE is a mask of the slots currently filled within the insn group.
8581 The mask bits come from alphaev5_pipe above. If EV5_E01 is set, then
8582 the insn in EV5_E0 can be swapped by the hardware into EV5_E1.
8583
8584 LEN is, of course, the length of the group in bytes. */
8585
8586 static rtx
8587 alphaev5_next_group (rtx insn, int *pin_use, int *plen)
8588 {
8589 int len, in_use;
8590
8591 len = in_use = 0;
8592
8593 if (! INSN_P (insn)
8594 || GET_CODE (PATTERN (insn)) == CLOBBER
8595 || GET_CODE (PATTERN (insn)) == USE)
8596 goto next_and_done;
8597
8598 while (1)
8599 {
8600 enum alphaev5_pipe pipe;
8601
8602 pipe = alphaev5_insn_pipe (insn);
8603 switch (pipe)
8604 {
8605 case EV5_STOP:
8606 /* Force complex instructions to start new groups. */
8607 if (in_use)
8608 goto done;
8609
8610 /* If this is a completely unrecognized insn, its an asm.
8611 We don't know how long it is, so record length as -1 to
8612 signal a needed realignment. */
8613 if (recog_memoized (insn) < 0)
8614 len = -1;
8615 else
8616 len = get_attr_length (insn);
8617 goto next_and_done;
8618
8619 /* ??? Most of the places below, we would like to abort, as
8620 it would indicate an error either in Haifa, or in the
8621 scheduling description. Unfortunately, Haifa never
8622 schedules the last instruction of the BB, so we don't
8623 have an accurate TI bit to go off. */
8624 case EV5_E01:
8625 if (in_use & EV5_E0)
8626 {
8627 if (in_use & EV5_E1)
8628 goto done;
8629 in_use |= EV5_E1;
8630 }
8631 else
8632 in_use |= EV5_E0 | EV5_E01;
8633 break;
8634
8635 case EV5_E0:
8636 if (in_use & EV5_E0)
8637 {
8638 if (!(in_use & EV5_E01) || (in_use & EV5_E1))
8639 goto done;
8640 in_use |= EV5_E1;
8641 }
8642 in_use |= EV5_E0;
8643 break;
8644
8645 case EV5_E1:
8646 if (in_use & EV5_E1)
8647 goto done;
8648 in_use |= EV5_E1;
8649 break;
8650
8651 case EV5_FAM:
8652 if (in_use & EV5_FA)
8653 {
8654 if (in_use & EV5_FM)
8655 goto done;
8656 in_use |= EV5_FM;
8657 }
8658 else
8659 in_use |= EV5_FA | EV5_FAM;
8660 break;
8661
8662 case EV5_FA:
8663 if (in_use & EV5_FA)
8664 goto done;
8665 in_use |= EV5_FA;
8666 break;
8667
8668 case EV5_FM:
8669 if (in_use & EV5_FM)
8670 goto done;
8671 in_use |= EV5_FM;
8672 break;
8673
8674 case EV5_NONE:
8675 break;
8676
8677 default:
8678 abort();
8679 }
8680 len += 4;
8681
8682 /* Haifa doesn't do well scheduling branches. */
8683 /* ??? If this is predicted not-taken, slotting continues, except
8684 that no more IBR, FBR, or JSR insns may be slotted. */
8685 if (GET_CODE (insn) == JUMP_INSN)
8686 goto next_and_done;
8687
8688 next:
8689 insn = next_nonnote_insn (insn);
8690
8691 if (!insn || ! INSN_P (insn))
8692 goto done;
8693
8694 /* Let Haifa tell us where it thinks insn group boundaries are. */
8695 if (GET_MODE (insn) == TImode)
8696 goto done;
8697
8698 if (GET_CODE (insn) == CLOBBER || GET_CODE (insn) == USE)
8699 goto next;
8700 }
8701
8702 next_and_done:
8703 insn = next_nonnote_insn (insn);
8704
8705 done:
8706 *plen = len;
8707 *pin_use = in_use;
8708 return insn;
8709 }
8710
8711 static rtx
8712 alphaev4_next_nop (int *pin_use)
8713 {
8714 int in_use = *pin_use;
8715 rtx nop;
8716
8717 if (!(in_use & EV4_IB0))
8718 {
8719 in_use |= EV4_IB0;
8720 nop = gen_nop ();
8721 }
8722 else if ((in_use & (EV4_IBX|EV4_IB1)) == EV4_IBX)
8723 {
8724 in_use |= EV4_IB1;
8725 nop = gen_nop ();
8726 }
8727 else if (TARGET_FP && !(in_use & EV4_IB1))
8728 {
8729 in_use |= EV4_IB1;
8730 nop = gen_fnop ();
8731 }
8732 else
8733 nop = gen_unop ();
8734
8735 *pin_use = in_use;
8736 return nop;
8737 }
8738
8739 static rtx
8740 alphaev5_next_nop (int *pin_use)
8741 {
8742 int in_use = *pin_use;
8743 rtx nop;
8744
8745 if (!(in_use & EV5_E1))
8746 {
8747 in_use |= EV5_E1;
8748 nop = gen_nop ();
8749 }
8750 else if (TARGET_FP && !(in_use & EV5_FA))
8751 {
8752 in_use |= EV5_FA;
8753 nop = gen_fnop ();
8754 }
8755 else if (TARGET_FP && !(in_use & EV5_FM))
8756 {
8757 in_use |= EV5_FM;
8758 nop = gen_fnop ();
8759 }
8760 else
8761 nop = gen_unop ();
8762
8763 *pin_use = in_use;
8764 return nop;
8765 }
8766
8767 /* The instruction group alignment main loop. */
8768
8769 static void
8770 alpha_align_insns (unsigned int max_align,
8771 rtx (*next_group) (rtx, int *, int *),
8772 rtx (*next_nop) (int *))
8773 {
8774 /* ALIGN is the known alignment for the insn group. */
8775 unsigned int align;
8776 /* OFS is the offset of the current insn in the insn group. */
8777 int ofs;
8778 int prev_in_use, in_use, len;
8779 rtx i, next;
8780
8781 /* Let shorten branches care for assigning alignments to code labels. */
8782 shorten_branches (get_insns ());
8783
8784 if (align_functions < 4)
8785 align = 4;
8786 else if ((unsigned int) align_functions < max_align)
8787 align = align_functions;
8788 else
8789 align = max_align;
8790
8791 ofs = prev_in_use = 0;
8792 i = get_insns ();
8793 if (GET_CODE (i) == NOTE)
8794 i = next_nonnote_insn (i);
8795
8796 while (i)
8797 {
8798 next = (*next_group) (i, &in_use, &len);
8799
8800 /* When we see a label, resync alignment etc. */
8801 if (GET_CODE (i) == CODE_LABEL)
8802 {
8803 unsigned int new_align = 1 << label_to_alignment (i);
8804
8805 if (new_align >= align)
8806 {
8807 align = new_align < max_align ? new_align : max_align;
8808 ofs = 0;
8809 }
8810
8811 else if (ofs & (new_align-1))
8812 ofs = (ofs | (new_align-1)) + 1;
8813 if (len != 0)
8814 abort();
8815 }
8816
8817 /* Handle complex instructions special. */
8818 else if (in_use == 0)
8819 {
8820 /* Asms will have length < 0. This is a signal that we have
8821 lost alignment knowledge. Assume, however, that the asm
8822 will not mis-align instructions. */
8823 if (len < 0)
8824 {
8825 ofs = 0;
8826 align = 4;
8827 len = 0;
8828 }
8829 }
8830
8831 /* If the known alignment is smaller than the recognized insn group,
8832 realign the output. */
8833 else if ((int) align < len)
8834 {
8835 unsigned int new_log_align = len > 8 ? 4 : 3;
8836 rtx prev, where;
8837
8838 where = prev = prev_nonnote_insn (i);
8839 if (!where || GET_CODE (where) != CODE_LABEL)
8840 where = i;
8841
8842 /* Can't realign between a call and its gp reload. */
8843 if (! (TARGET_EXPLICIT_RELOCS
8844 && prev && GET_CODE (prev) == CALL_INSN))
8845 {
8846 emit_insn_before (gen_realign (GEN_INT (new_log_align)), where);
8847 align = 1 << new_log_align;
8848 ofs = 0;
8849 }
8850 }
8851
8852 /* If the group won't fit in the same INT16 as the previous,
8853 we need to add padding to keep the group together. Rather
8854 than simply leaving the insn filling to the assembler, we
8855 can make use of the knowledge of what sorts of instructions
8856 were issued in the previous group to make sure that all of
8857 the added nops are really free. */
8858 else if (ofs + len > (int) align)
8859 {
8860 int nop_count = (align - ofs) / 4;
8861 rtx where;
8862
8863 /* Insert nops before labels, branches, and calls to truly merge
8864 the execution of the nops with the previous instruction group. */
8865 where = prev_nonnote_insn (i);
8866 if (where)
8867 {
8868 if (GET_CODE (where) == CODE_LABEL)
8869 {
8870 rtx where2 = prev_nonnote_insn (where);
8871 if (where2 && GET_CODE (where2) == JUMP_INSN)
8872 where = where2;
8873 }
8874 else if (GET_CODE (where) == INSN)
8875 where = i;
8876 }
8877 else
8878 where = i;
8879
8880 do
8881 emit_insn_before ((*next_nop)(&prev_in_use), where);
8882 while (--nop_count);
8883 ofs = 0;
8884 }
8885
8886 ofs = (ofs + len) & (align - 1);
8887 prev_in_use = in_use;
8888 i = next;
8889 }
8890 }
8891 \f
8892 /* Machine dependent reorg pass. */
8893
8894 static void
8895 alpha_reorg (void)
8896 {
8897 if (alpha_tp != ALPHA_TP_PROG || flag_exceptions)
8898 alpha_handle_trap_shadows ();
8899
8900 /* Due to the number of extra trapb insns, don't bother fixing up
8901 alignment when trap precision is instruction. Moreover, we can
8902 only do our job when sched2 is run. */
8903 if (optimize && !optimize_size
8904 && alpha_tp != ALPHA_TP_INSN
8905 && flag_schedule_insns_after_reload)
8906 {
8907 if (alpha_cpu == PROCESSOR_EV4)
8908 alpha_align_insns (8, alphaev4_next_group, alphaev4_next_nop);
8909 else if (alpha_cpu == PROCESSOR_EV5)
8910 alpha_align_insns (16, alphaev5_next_group, alphaev5_next_nop);
8911 }
8912 }
8913 \f
8914 #if !TARGET_ABI_UNICOSMK
8915
8916 #ifdef HAVE_STAMP_H
8917 #include <stamp.h>
8918 #endif
8919
8920 static void
8921 alpha_file_start (void)
8922 {
8923 #ifdef OBJECT_FORMAT_ELF
8924 /* If emitting dwarf2 debug information, we cannot generate a .file
8925 directive to start the file, as it will conflict with dwarf2out
8926 file numbers. So it's only useful when emitting mdebug output. */
8927 targetm.file_start_file_directive = (write_symbols == DBX_DEBUG);
8928 #endif
8929
8930 default_file_start ();
8931 #ifdef MS_STAMP
8932 fprintf (asm_out_file, "\t.verstamp %d %d\n", MS_STAMP, LS_STAMP);
8933 #endif
8934
8935 fputs ("\t.set noreorder\n", asm_out_file);
8936 fputs ("\t.set volatile\n", asm_out_file);
8937 if (!TARGET_ABI_OPEN_VMS)
8938 fputs ("\t.set noat\n", asm_out_file);
8939 if (TARGET_EXPLICIT_RELOCS)
8940 fputs ("\t.set nomacro\n", asm_out_file);
8941 if (TARGET_SUPPORT_ARCH | TARGET_BWX | TARGET_MAX | TARGET_FIX | TARGET_CIX)
8942 fprintf (asm_out_file,
8943 "\t.arch %s\n",
8944 TARGET_CPU_EV6 ? "ev6"
8945 : (TARGET_CPU_EV5
8946 ? (TARGET_MAX ? "pca56" : TARGET_BWX ? "ev56" : "ev5")
8947 : "ev4"));
8948 }
8949 #endif
8950
8951 #ifdef OBJECT_FORMAT_ELF
8952
8953 /* Switch to the section to which we should output X. The only thing
8954 special we do here is to honor small data. */
8955
8956 static void
8957 alpha_elf_select_rtx_section (enum machine_mode mode, rtx x,
8958 unsigned HOST_WIDE_INT align)
8959 {
8960 if (TARGET_SMALL_DATA && GET_MODE_SIZE (mode) <= g_switch_value)
8961 /* ??? Consider using mergeable sdata sections. */
8962 sdata_section ();
8963 else
8964 default_elf_select_rtx_section (mode, x, align);
8965 }
8966
8967 #endif /* OBJECT_FORMAT_ELF */
8968 \f
8969 /* Structure to collect function names for final output in link section. */
8970 /* Note that items marked with GTY can't be ifdef'ed out. */
8971
8972 enum links_kind {KIND_UNUSED, KIND_LOCAL, KIND_EXTERN};
8973 enum reloc_kind {KIND_LINKAGE, KIND_CODEADDR};
8974
8975 struct alpha_links GTY(())
8976 {
8977 int num;
8978 rtx linkage;
8979 enum links_kind lkind;
8980 enum reloc_kind rkind;
8981 };
8982
8983 struct alpha_funcs GTY(())
8984 {
8985 int num;
8986 splay_tree GTY ((param1_is (char *), param2_is (struct alpha_links *)))
8987 links;
8988 };
8989
8990 static GTY ((param1_is (char *), param2_is (struct alpha_links *)))
8991 splay_tree alpha_links_tree;
8992 static GTY ((param1_is (tree), param2_is (struct alpha_funcs *)))
8993 splay_tree alpha_funcs_tree;
8994
8995 static GTY(()) int alpha_funcs_num;
8996
8997 #if TARGET_ABI_OPEN_VMS
8998
8999 /* Return the VMS argument type corresponding to MODE. */
9000
9001 enum avms_arg_type
9002 alpha_arg_type (enum machine_mode mode)
9003 {
9004 switch (mode)
9005 {
9006 case SFmode:
9007 return TARGET_FLOAT_VAX ? FF : FS;
9008 case DFmode:
9009 return TARGET_FLOAT_VAX ? FD : FT;
9010 default:
9011 return I64;
9012 }
9013 }
9014
9015 /* Return an rtx for an integer representing the VMS Argument Information
9016 register value. */
9017
9018 rtx
9019 alpha_arg_info_reg_val (CUMULATIVE_ARGS cum)
9020 {
9021 unsigned HOST_WIDE_INT regval = cum.num_args;
9022 int i;
9023
9024 for (i = 0; i < 6; i++)
9025 regval |= ((int) cum.atypes[i]) << (i * 3 + 8);
9026
9027 return GEN_INT (regval);
9028 }
9029 \f
9030 /* Make (or fake) .linkage entry for function call.
9031
9032 IS_LOCAL is 0 if name is used in call, 1 if name is used in definition.
9033
9034 Return an SYMBOL_REF rtx for the linkage. */
9035
9036 rtx
9037 alpha_need_linkage (const char *name, int is_local)
9038 {
9039 splay_tree_node node;
9040 struct alpha_links *al;
9041
9042 if (name[0] == '*')
9043 name++;
9044
9045 if (is_local)
9046 {
9047 struct alpha_funcs *cfaf;
9048
9049 if (!alpha_funcs_tree)
9050 alpha_funcs_tree = splay_tree_new_ggc ((splay_tree_compare_fn)
9051 splay_tree_compare_pointers);
9052
9053 cfaf = (struct alpha_funcs *) ggc_alloc (sizeof (struct alpha_funcs));
9054
9055 cfaf->links = 0;
9056 cfaf->num = ++alpha_funcs_num;
9057
9058 splay_tree_insert (alpha_funcs_tree,
9059 (splay_tree_key) current_function_decl,
9060 (splay_tree_value) cfaf);
9061 }
9062
9063 if (alpha_links_tree)
9064 {
9065 /* Is this name already defined? */
9066
9067 node = splay_tree_lookup (alpha_links_tree, (splay_tree_key) name);
9068 if (node)
9069 {
9070 al = (struct alpha_links *) node->value;
9071 if (is_local)
9072 {
9073 /* Defined here but external assumed. */
9074 if (al->lkind == KIND_EXTERN)
9075 al->lkind = KIND_LOCAL;
9076 }
9077 else
9078 {
9079 /* Used here but unused assumed. */
9080 if (al->lkind == KIND_UNUSED)
9081 al->lkind = KIND_LOCAL;
9082 }
9083 return al->linkage;
9084 }
9085 }
9086 else
9087 alpha_links_tree = splay_tree_new_ggc ((splay_tree_compare_fn) strcmp);
9088
9089 al = (struct alpha_links *) ggc_alloc (sizeof (struct alpha_links));
9090 name = ggc_strdup (name);
9091
9092 /* Assume external if no definition. */
9093 al->lkind = (is_local ? KIND_UNUSED : KIND_EXTERN);
9094
9095 /* Ensure we have an IDENTIFIER so assemble_name can mark it used. */
9096 get_identifier (name);
9097
9098 /* Construct a SYMBOL_REF for us to call. */
9099 {
9100 size_t name_len = strlen (name);
9101 char *linksym = alloca (name_len + 6);
9102 linksym[0] = '$';
9103 memcpy (linksym + 1, name, name_len);
9104 memcpy (linksym + 1 + name_len, "..lk", 5);
9105 al->linkage = gen_rtx_SYMBOL_REF (Pmode,
9106 ggc_alloc_string (linksym, name_len + 5));
9107 }
9108
9109 splay_tree_insert (alpha_links_tree, (splay_tree_key) name,
9110 (splay_tree_value) al);
9111
9112 return al->linkage;
9113 }
9114
9115 rtx
9116 alpha_use_linkage (rtx linkage, tree cfundecl, int lflag, int rflag)
9117 {
9118 splay_tree_node cfunnode;
9119 struct alpha_funcs *cfaf;
9120 struct alpha_links *al;
9121 const char *name = XSTR (linkage, 0);
9122
9123 cfaf = (struct alpha_funcs *) 0;
9124 al = (struct alpha_links *) 0;
9125
9126 cfunnode = splay_tree_lookup (alpha_funcs_tree, (splay_tree_key) cfundecl);
9127 cfaf = (struct alpha_funcs *) cfunnode->value;
9128
9129 if (cfaf->links)
9130 {
9131 splay_tree_node lnode;
9132
9133 /* Is this name already defined? */
9134
9135 lnode = splay_tree_lookup (cfaf->links, (splay_tree_key) name);
9136 if (lnode)
9137 al = (struct alpha_links *) lnode->value;
9138 }
9139 else
9140 cfaf->links = splay_tree_new_ggc ((splay_tree_compare_fn) strcmp);
9141
9142 if (!al)
9143 {
9144 size_t name_len;
9145 size_t buflen;
9146 char buf [512];
9147 char *linksym;
9148 splay_tree_node node = 0;
9149 struct alpha_links *anl;
9150
9151 if (name[0] == '*')
9152 name++;
9153
9154 name_len = strlen (name);
9155
9156 al = (struct alpha_links *) ggc_alloc (sizeof (struct alpha_links));
9157 al->num = cfaf->num;
9158
9159 node = splay_tree_lookup (alpha_links_tree, (splay_tree_key) name);
9160 if (node)
9161 {
9162 anl = (struct alpha_links *) node->value;
9163 al->lkind = anl->lkind;
9164 }
9165
9166 sprintf (buf, "$%d..%s..lk", cfaf->num, name);
9167 buflen = strlen (buf);
9168 linksym = alloca (buflen + 1);
9169 memcpy (linksym, buf, buflen + 1);
9170
9171 al->linkage = gen_rtx_SYMBOL_REF
9172 (Pmode, ggc_alloc_string (linksym, buflen + 1));
9173
9174 splay_tree_insert (cfaf->links, (splay_tree_key) name,
9175 (splay_tree_value) al);
9176 }
9177
9178 if (rflag)
9179 al->rkind = KIND_CODEADDR;
9180 else
9181 al->rkind = KIND_LINKAGE;
9182
9183 if (lflag)
9184 return gen_rtx_MEM (Pmode, plus_constant (al->linkage, 8));
9185 else
9186 return al->linkage;
9187 }
9188
9189 static int
9190 alpha_write_one_linkage (splay_tree_node node, void *data)
9191 {
9192 const char *const name = (const char *) node->key;
9193 struct alpha_links *link = (struct alpha_links *) node->value;
9194 FILE *stream = (FILE *) data;
9195
9196 fprintf (stream, "$%d..%s..lk:\n", link->num, name);
9197 if (link->rkind == KIND_CODEADDR)
9198 {
9199 if (link->lkind == KIND_LOCAL)
9200 {
9201 /* Local and used */
9202 fprintf (stream, "\t.quad %s..en\n", name);
9203 }
9204 else
9205 {
9206 /* External and used, request code address. */
9207 fprintf (stream, "\t.code_address %s\n", name);
9208 }
9209 }
9210 else
9211 {
9212 if (link->lkind == KIND_LOCAL)
9213 {
9214 /* Local and used, build linkage pair. */
9215 fprintf (stream, "\t.quad %s..en\n", name);
9216 fprintf (stream, "\t.quad %s\n", name);
9217 }
9218 else
9219 {
9220 /* External and used, request linkage pair. */
9221 fprintf (stream, "\t.linkage %s\n", name);
9222 }
9223 }
9224
9225 return 0;
9226 }
9227
9228 static void
9229 alpha_write_linkage (FILE *stream, const char *funname, tree fundecl)
9230 {
9231 splay_tree_node node;
9232 struct alpha_funcs *func;
9233
9234 link_section ();
9235 fprintf (stream, "\t.align 3\n");
9236 node = splay_tree_lookup (alpha_funcs_tree, (splay_tree_key) fundecl);
9237 func = (struct alpha_funcs *) node->value;
9238
9239 fputs ("\t.name ", stream);
9240 assemble_name (stream, funname);
9241 fputs ("..na\n", stream);
9242 ASM_OUTPUT_LABEL (stream, funname);
9243 fprintf (stream, "\t.pdesc ");
9244 assemble_name (stream, funname);
9245 fprintf (stream, "..en,%s\n",
9246 alpha_procedure_type == PT_STACK ? "stack"
9247 : alpha_procedure_type == PT_REGISTER ? "reg" : "null");
9248
9249 if (func->links)
9250 {
9251 splay_tree_foreach (func->links, alpha_write_one_linkage, stream);
9252 /* splay_tree_delete (func->links); */
9253 }
9254 }
9255
9256 /* Given a decl, a section name, and whether the decl initializer
9257 has relocs, choose attributes for the section. */
9258
9259 #define SECTION_VMS_OVERLAY SECTION_FORGET
9260 #define SECTION_VMS_GLOBAL SECTION_MACH_DEP
9261 #define SECTION_VMS_INITIALIZE (SECTION_VMS_GLOBAL << 1)
9262
9263 static unsigned int
9264 vms_section_type_flags (tree decl, const char *name, int reloc)
9265 {
9266 unsigned int flags = default_section_type_flags (decl, name, reloc);
9267
9268 if (decl && DECL_ATTRIBUTES (decl)
9269 && lookup_attribute ("overlaid", DECL_ATTRIBUTES (decl)))
9270 flags |= SECTION_VMS_OVERLAY;
9271 if (decl && DECL_ATTRIBUTES (decl)
9272 && lookup_attribute ("global", DECL_ATTRIBUTES (decl)))
9273 flags |= SECTION_VMS_GLOBAL;
9274 if (decl && DECL_ATTRIBUTES (decl)
9275 && lookup_attribute ("initialize", DECL_ATTRIBUTES (decl)))
9276 flags |= SECTION_VMS_INITIALIZE;
9277
9278 return flags;
9279 }
9280
9281 /* Switch to an arbitrary section NAME with attributes as specified
9282 by FLAGS. ALIGN specifies any known alignment requirements for
9283 the section; 0 if the default should be used. */
9284
9285 static void
9286 vms_asm_named_section (const char *name, unsigned int flags)
9287 {
9288 fputc ('\n', asm_out_file);
9289 fprintf (asm_out_file, ".section\t%s", name);
9290
9291 if (flags & SECTION_VMS_OVERLAY)
9292 fprintf (asm_out_file, ",OVR");
9293 if (flags & SECTION_VMS_GLOBAL)
9294 fprintf (asm_out_file, ",GBL");
9295 if (flags & SECTION_VMS_INITIALIZE)
9296 fprintf (asm_out_file, ",NOMOD");
9297 if (flags & SECTION_DEBUG)
9298 fprintf (asm_out_file, ",NOWRT");
9299
9300 fputc ('\n', asm_out_file);
9301 }
9302
9303 /* Record an element in the table of global constructors. SYMBOL is
9304 a SYMBOL_REF of the function to be called; PRIORITY is a number
9305 between 0 and MAX_INIT_PRIORITY.
9306
9307 Differs from default_ctors_section_asm_out_constructor in that the
9308 width of the .ctors entry is always 64 bits, rather than the 32 bits
9309 used by a normal pointer. */
9310
9311 static void
9312 vms_asm_out_constructor (rtx symbol, int priority ATTRIBUTE_UNUSED)
9313 {
9314 ctors_section ();
9315 assemble_align (BITS_PER_WORD);
9316 assemble_integer (symbol, UNITS_PER_WORD, BITS_PER_WORD, 1);
9317 }
9318
9319 static void
9320 vms_asm_out_destructor (rtx symbol, int priority ATTRIBUTE_UNUSED)
9321 {
9322 dtors_section ();
9323 assemble_align (BITS_PER_WORD);
9324 assemble_integer (symbol, UNITS_PER_WORD, BITS_PER_WORD, 1);
9325 }
9326 #else
9327
9328 rtx
9329 alpha_need_linkage (const char *name ATTRIBUTE_UNUSED,
9330 int is_local ATTRIBUTE_UNUSED)
9331 {
9332 return NULL_RTX;
9333 }
9334
9335 rtx
9336 alpha_use_linkage (rtx linkage ATTRIBUTE_UNUSED,
9337 tree cfundecl ATTRIBUTE_UNUSED,
9338 int lflag ATTRIBUTE_UNUSED,
9339 int rflag ATTRIBUTE_UNUSED)
9340 {
9341 return NULL_RTX;
9342 }
9343
9344 #endif /* TARGET_ABI_OPEN_VMS */
9345 \f
9346 #if TARGET_ABI_UNICOSMK
9347
9348 /* Define the offset between two registers, one to be eliminated, and the
9349 other its replacement, at the start of a routine. */
9350
9351 int
9352 unicosmk_initial_elimination_offset (int from, int to)
9353 {
9354 int fixed_size;
9355
9356 fixed_size = alpha_sa_size();
9357 if (fixed_size != 0)
9358 fixed_size += 48;
9359
9360 if (from == FRAME_POINTER_REGNUM && to == HARD_FRAME_POINTER_REGNUM)
9361 return -fixed_size;
9362 else if (from == ARG_POINTER_REGNUM && to == HARD_FRAME_POINTER_REGNUM)
9363 return 0;
9364 else if (from == FRAME_POINTER_REGNUM && to == STACK_POINTER_REGNUM)
9365 return (ALPHA_ROUND (current_function_outgoing_args_size)
9366 + ALPHA_ROUND (get_frame_size()));
9367 else if (from == ARG_POINTER_REGNUM && to == STACK_POINTER_REGNUM)
9368 return (ALPHA_ROUND (fixed_size)
9369 + ALPHA_ROUND (get_frame_size()
9370 + current_function_outgoing_args_size));
9371 else
9372 abort ();
9373 }
9374
9375 /* Output the module name for .ident and .end directives. We have to strip
9376 directories and add make sure that the module name starts with a letter
9377 or '$'. */
9378
9379 static void
9380 unicosmk_output_module_name (FILE *file)
9381 {
9382 const char *name = lbasename (main_input_filename);
9383 unsigned len = strlen (name);
9384 char *clean_name = alloca (len + 2);
9385 char *ptr = clean_name;
9386
9387 /* CAM only accepts module names that start with a letter or '$'. We
9388 prefix the module name with a '$' if necessary. */
9389
9390 if (!ISALPHA (*name))
9391 *ptr++ = '$';
9392 memcpy (ptr, name, len + 1);
9393 clean_symbol_name (clean_name);
9394 fputs (clean_name, file);
9395 }
9396
9397 /* Output the definition of a common variable. */
9398
9399 void
9400 unicosmk_output_common (FILE *file, const char *name, int size, int align)
9401 {
9402 tree name_tree;
9403 printf ("T3E__: common %s\n", name);
9404
9405 common_section ();
9406 fputs("\t.endp\n\n\t.psect ", file);
9407 assemble_name(file, name);
9408 fprintf(file, ",%d,common\n", floor_log2 (align / BITS_PER_UNIT));
9409 fprintf(file, "\t.byte\t0:%d\n", size);
9410
9411 /* Mark the symbol as defined in this module. */
9412 name_tree = get_identifier (name);
9413 TREE_ASM_WRITTEN (name_tree) = 1;
9414 }
9415
9416 #define SECTION_PUBLIC SECTION_MACH_DEP
9417 #define SECTION_MAIN (SECTION_PUBLIC << 1)
9418 static int current_section_align;
9419
9420 static unsigned int
9421 unicosmk_section_type_flags (tree decl, const char *name,
9422 int reloc ATTRIBUTE_UNUSED)
9423 {
9424 unsigned int flags = default_section_type_flags (decl, name, reloc);
9425
9426 if (!decl)
9427 return flags;
9428
9429 if (TREE_CODE (decl) == FUNCTION_DECL)
9430 {
9431 current_section_align = floor_log2 (FUNCTION_BOUNDARY / BITS_PER_UNIT);
9432 if (align_functions_log > current_section_align)
9433 current_section_align = align_functions_log;
9434
9435 if (! strcmp (IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)), "main"))
9436 flags |= SECTION_MAIN;
9437 }
9438 else
9439 current_section_align = floor_log2 (DECL_ALIGN (decl) / BITS_PER_UNIT);
9440
9441 if (TREE_PUBLIC (decl))
9442 flags |= SECTION_PUBLIC;
9443
9444 return flags;
9445 }
9446
9447 /* Generate a section name for decl and associate it with the
9448 declaration. */
9449
9450 static void
9451 unicosmk_unique_section (tree decl, int reloc ATTRIBUTE_UNUSED)
9452 {
9453 const char *name;
9454 int len;
9455
9456 if (!decl)
9457 abort ();
9458
9459 name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl));
9460 name = default_strip_name_encoding (name);
9461 len = strlen (name);
9462
9463 if (TREE_CODE (decl) == FUNCTION_DECL)
9464 {
9465 char *string;
9466
9467 /* It is essential that we prefix the section name here because
9468 otherwise the section names generated for constructors and
9469 destructors confuse collect2. */
9470
9471 string = alloca (len + 6);
9472 sprintf (string, "code@%s", name);
9473 DECL_SECTION_NAME (decl) = build_string (len + 5, string);
9474 }
9475 else if (TREE_PUBLIC (decl))
9476 DECL_SECTION_NAME (decl) = build_string (len, name);
9477 else
9478 {
9479 char *string;
9480
9481 string = alloca (len + 6);
9482 sprintf (string, "data@%s", name);
9483 DECL_SECTION_NAME (decl) = build_string (len + 5, string);
9484 }
9485 }
9486
9487 /* Switch to an arbitrary section NAME with attributes as specified
9488 by FLAGS. ALIGN specifies any known alignment requirements for
9489 the section; 0 if the default should be used. */
9490
9491 static void
9492 unicosmk_asm_named_section (const char *name, unsigned int flags)
9493 {
9494 const char *kind;
9495
9496 /* Close the previous section. */
9497
9498 fputs ("\t.endp\n\n", asm_out_file);
9499
9500 /* Find out what kind of section we are opening. */
9501
9502 if (flags & SECTION_MAIN)
9503 fputs ("\t.start\tmain\n", asm_out_file);
9504
9505 if (flags & SECTION_CODE)
9506 kind = "code";
9507 else if (flags & SECTION_PUBLIC)
9508 kind = "common";
9509 else
9510 kind = "data";
9511
9512 if (current_section_align != 0)
9513 fprintf (asm_out_file, "\t.psect\t%s,%d,%s\n", name,
9514 current_section_align, kind);
9515 else
9516 fprintf (asm_out_file, "\t.psect\t%s,%s\n", name, kind);
9517 }
9518
9519 static void
9520 unicosmk_insert_attributes (tree decl, tree *attr_ptr ATTRIBUTE_UNUSED)
9521 {
9522 if (DECL_P (decl)
9523 && (TREE_PUBLIC (decl) || TREE_CODE (decl) == FUNCTION_DECL))
9524 unicosmk_unique_section (decl, 0);
9525 }
9526
9527 /* Output an alignment directive. We have to use the macro 'gcc@code@align'
9528 in code sections because .align fill unused space with zeroes. */
9529
9530 void
9531 unicosmk_output_align (FILE *file, int align)
9532 {
9533 if (inside_function)
9534 fprintf (file, "\tgcc@code@align\t%d\n", align);
9535 else
9536 fprintf (file, "\t.align\t%d\n", align);
9537 }
9538
9539 /* Add a case vector to the current function's list of deferred case
9540 vectors. Case vectors have to be put into a separate section because CAM
9541 does not allow data definitions in code sections. */
9542
9543 void
9544 unicosmk_defer_case_vector (rtx lab, rtx vec)
9545 {
9546 struct machine_function *machine = cfun->machine;
9547
9548 vec = gen_rtx_EXPR_LIST (VOIDmode, lab, vec);
9549 machine->addr_list = gen_rtx_EXPR_LIST (VOIDmode, vec,
9550 machine->addr_list);
9551 }
9552
9553 /* Output a case vector. */
9554
9555 static void
9556 unicosmk_output_addr_vec (FILE *file, rtx vec)
9557 {
9558 rtx lab = XEXP (vec, 0);
9559 rtx body = XEXP (vec, 1);
9560 int vlen = XVECLEN (body, 0);
9561 int idx;
9562
9563 (*targetm.asm_out.internal_label) (file, "L", CODE_LABEL_NUMBER (lab));
9564
9565 for (idx = 0; idx < vlen; idx++)
9566 {
9567 ASM_OUTPUT_ADDR_VEC_ELT
9568 (file, CODE_LABEL_NUMBER (XEXP (XVECEXP (body, 0, idx), 0)));
9569 }
9570 }
9571
9572 /* Output current function's deferred case vectors. */
9573
9574 static void
9575 unicosmk_output_deferred_case_vectors (FILE *file)
9576 {
9577 struct machine_function *machine = cfun->machine;
9578 rtx t;
9579
9580 if (machine->addr_list == NULL_RTX)
9581 return;
9582
9583 data_section ();
9584 for (t = machine->addr_list; t; t = XEXP (t, 1))
9585 unicosmk_output_addr_vec (file, XEXP (t, 0));
9586 }
9587
9588 /* Generate the name of the SSIB section for the current function. */
9589
9590 #define SSIB_PREFIX "__SSIB_"
9591 #define SSIB_PREFIX_LEN 7
9592
9593 static const char *
9594 unicosmk_ssib_name (void)
9595 {
9596 /* This is ok since CAM won't be able to deal with names longer than that
9597 anyway. */
9598
9599 static char name[256];
9600
9601 rtx x;
9602 const char *fnname;
9603 int len;
9604
9605 x = DECL_RTL (cfun->decl);
9606 if (GET_CODE (x) != MEM)
9607 abort ();
9608 x = XEXP (x, 0);
9609 if (GET_CODE (x) != SYMBOL_REF)
9610 abort ();
9611 fnname = XSTR (x, 0);
9612
9613 len = strlen (fnname);
9614 if (len + SSIB_PREFIX_LEN > 255)
9615 len = 255 - SSIB_PREFIX_LEN;
9616
9617 strcpy (name, SSIB_PREFIX);
9618 strncpy (name + SSIB_PREFIX_LEN, fnname, len);
9619 name[len + SSIB_PREFIX_LEN] = 0;
9620
9621 return name;
9622 }
9623
9624 /* Set up the dynamic subprogram information block (DSIB) and update the
9625 frame pointer register ($15) for subroutines which have a frame. If the
9626 subroutine doesn't have a frame, simply increment $15. */
9627
9628 static void
9629 unicosmk_gen_dsib (unsigned long *imaskP)
9630 {
9631 if (alpha_procedure_type == PT_STACK)
9632 {
9633 const char *ssib_name;
9634 rtx mem;
9635
9636 /* Allocate 64 bytes for the DSIB. */
9637
9638 FRP (emit_insn (gen_adddi3 (stack_pointer_rtx, stack_pointer_rtx,
9639 GEN_INT (-64))));
9640 emit_insn (gen_blockage ());
9641
9642 /* Save the return address. */
9643
9644 mem = gen_rtx_MEM (DImode, plus_constant (stack_pointer_rtx, 56));
9645 set_mem_alias_set (mem, alpha_sr_alias_set);
9646 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, REG_RA)));
9647 (*imaskP) &= ~(1UL << REG_RA);
9648
9649 /* Save the old frame pointer. */
9650
9651 mem = gen_rtx_MEM (DImode, plus_constant (stack_pointer_rtx, 48));
9652 set_mem_alias_set (mem, alpha_sr_alias_set);
9653 FRP (emit_move_insn (mem, hard_frame_pointer_rtx));
9654 (*imaskP) &= ~(1UL << HARD_FRAME_POINTER_REGNUM);
9655
9656 emit_insn (gen_blockage ());
9657
9658 /* Store the SSIB pointer. */
9659
9660 ssib_name = ggc_strdup (unicosmk_ssib_name ());
9661 mem = gen_rtx_MEM (DImode, plus_constant (stack_pointer_rtx, 32));
9662 set_mem_alias_set (mem, alpha_sr_alias_set);
9663
9664 FRP (emit_move_insn (gen_rtx_REG (DImode, 5),
9665 gen_rtx_SYMBOL_REF (Pmode, ssib_name)));
9666 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, 5)));
9667
9668 /* Save the CIW index. */
9669
9670 mem = gen_rtx_MEM (DImode, plus_constant (stack_pointer_rtx, 24));
9671 set_mem_alias_set (mem, alpha_sr_alias_set);
9672 FRP (emit_move_insn (mem, gen_rtx_REG (DImode, 25)));
9673
9674 emit_insn (gen_blockage ());
9675
9676 /* Set the new frame pointer. */
9677
9678 FRP (emit_insn (gen_adddi3 (hard_frame_pointer_rtx,
9679 stack_pointer_rtx, GEN_INT (64))));
9680
9681 }
9682 else
9683 {
9684 /* Increment the frame pointer register to indicate that we do not
9685 have a frame. */
9686
9687 FRP (emit_insn (gen_adddi3 (hard_frame_pointer_rtx,
9688 hard_frame_pointer_rtx, GEN_INT (1))));
9689 }
9690 }
9691
9692 /* Output the static subroutine information block for the current
9693 function. */
9694
9695 static void
9696 unicosmk_output_ssib (FILE *file, const char *fnname)
9697 {
9698 int len;
9699 int i;
9700 rtx x;
9701 rtx ciw;
9702 struct machine_function *machine = cfun->machine;
9703
9704 ssib_section ();
9705 fprintf (file, "\t.endp\n\n\t.psect\t%s%s,data\n", user_label_prefix,
9706 unicosmk_ssib_name ());
9707
9708 /* Some required stuff and the function name length. */
9709
9710 len = strlen (fnname);
9711 fprintf (file, "\t.quad\t^X20008%2.2X28\n", len);
9712
9713 /* Saved registers
9714 ??? We don't do that yet. */
9715
9716 fputs ("\t.quad\t0\n", file);
9717
9718 /* Function address. */
9719
9720 fputs ("\t.quad\t", file);
9721 assemble_name (file, fnname);
9722 putc ('\n', file);
9723
9724 fputs ("\t.quad\t0\n", file);
9725 fputs ("\t.quad\t0\n", file);
9726
9727 /* Function name.
9728 ??? We do it the same way Cray CC does it but this could be
9729 simplified. */
9730
9731 for( i = 0; i < len; i++ )
9732 fprintf (file, "\t.byte\t%d\n", (int)(fnname[i]));
9733 if( (len % 8) == 0 )
9734 fputs ("\t.quad\t0\n", file);
9735 else
9736 fprintf (file, "\t.bits\t%d : 0\n", (8 - (len % 8))*8);
9737
9738 /* All call information words used in the function. */
9739
9740 for (x = machine->first_ciw; x; x = XEXP (x, 1))
9741 {
9742 ciw = XEXP (x, 0);
9743 #if HOST_BITS_PER_WIDE_INT == 32
9744 fprintf (file, "\t.quad\t" HOST_WIDE_INT_PRINT_DOUBLE_HEX "\n",
9745 CONST_DOUBLE_HIGH (ciw), CONST_DOUBLE_LOW (ciw));
9746 #else
9747 fprintf (file, "\t.quad\t" HOST_WIDE_INT_PRINT_HEX "\n", INTVAL (ciw));
9748 #endif
9749 }
9750 }
9751
9752 /* Add a call information word (CIW) to the list of the current function's
9753 CIWs and return its index.
9754
9755 X is a CONST_INT or CONST_DOUBLE representing the CIW. */
9756
9757 rtx
9758 unicosmk_add_call_info_word (rtx x)
9759 {
9760 rtx node;
9761 struct machine_function *machine = cfun->machine;
9762
9763 node = gen_rtx_EXPR_LIST (VOIDmode, x, NULL_RTX);
9764 if (machine->first_ciw == NULL_RTX)
9765 machine->first_ciw = node;
9766 else
9767 XEXP (machine->last_ciw, 1) = node;
9768
9769 machine->last_ciw = node;
9770 ++machine->ciw_count;
9771
9772 return GEN_INT (machine->ciw_count
9773 + strlen (current_function_name ())/8 + 5);
9774 }
9775
9776 static char unicosmk_section_buf[100];
9777
9778 char *
9779 unicosmk_text_section (void)
9780 {
9781 static int count = 0;
9782 sprintf (unicosmk_section_buf, "\t.endp\n\n\t.psect\tgcc@text___%d,code",
9783 count++);
9784 return unicosmk_section_buf;
9785 }
9786
9787 char *
9788 unicosmk_data_section (void)
9789 {
9790 static int count = 1;
9791 sprintf (unicosmk_section_buf, "\t.endp\n\n\t.psect\tgcc@data___%d,data",
9792 count++);
9793 return unicosmk_section_buf;
9794 }
9795
9796 /* The Cray assembler doesn't accept extern declarations for symbols which
9797 are defined in the same file. We have to keep track of all global
9798 symbols which are referenced and/or defined in a source file and output
9799 extern declarations for those which are referenced but not defined at
9800 the end of file. */
9801
9802 /* List of identifiers for which an extern declaration might have to be
9803 emitted. */
9804 /* FIXME: needs to use GC, so it can be saved and restored for PCH. */
9805
9806 struct unicosmk_extern_list
9807 {
9808 struct unicosmk_extern_list *next;
9809 const char *name;
9810 };
9811
9812 static struct unicosmk_extern_list *unicosmk_extern_head = 0;
9813
9814 /* Output extern declarations which are required for every asm file. */
9815
9816 static void
9817 unicosmk_output_default_externs (FILE *file)
9818 {
9819 static const char *const externs[] =
9820 { "__T3E_MISMATCH" };
9821
9822 int i;
9823 int n;
9824
9825 n = ARRAY_SIZE (externs);
9826
9827 for (i = 0; i < n; i++)
9828 fprintf (file, "\t.extern\t%s\n", externs[i]);
9829 }
9830
9831 /* Output extern declarations for global symbols which are have been
9832 referenced but not defined. */
9833
9834 static void
9835 unicosmk_output_externs (FILE *file)
9836 {
9837 struct unicosmk_extern_list *p;
9838 const char *real_name;
9839 int len;
9840 tree name_tree;
9841
9842 len = strlen (user_label_prefix);
9843 for (p = unicosmk_extern_head; p != 0; p = p->next)
9844 {
9845 /* We have to strip the encoding and possibly remove user_label_prefix
9846 from the identifier in order to handle -fleading-underscore and
9847 explicit asm names correctly (cf. gcc.dg/asm-names-1.c). */
9848 real_name = default_strip_name_encoding (p->name);
9849 if (len && p->name[0] == '*'
9850 && !memcmp (real_name, user_label_prefix, len))
9851 real_name += len;
9852
9853 name_tree = get_identifier (real_name);
9854 if (! TREE_ASM_WRITTEN (name_tree))
9855 {
9856 TREE_ASM_WRITTEN (name_tree) = 1;
9857 fputs ("\t.extern\t", file);
9858 assemble_name (file, p->name);
9859 putc ('\n', file);
9860 }
9861 }
9862 }
9863
9864 /* Record an extern. */
9865
9866 void
9867 unicosmk_add_extern (const char *name)
9868 {
9869 struct unicosmk_extern_list *p;
9870
9871 p = (struct unicosmk_extern_list *)
9872 xmalloc (sizeof (struct unicosmk_extern_list));
9873 p->next = unicosmk_extern_head;
9874 p->name = name;
9875 unicosmk_extern_head = p;
9876 }
9877
9878 /* The Cray assembler generates incorrect code if identifiers which
9879 conflict with register names are used as instruction operands. We have
9880 to replace such identifiers with DEX expressions. */
9881
9882 /* Structure to collect identifiers which have been replaced by DEX
9883 expressions. */
9884 /* FIXME: needs to use GC, so it can be saved and restored for PCH. */
9885
9886 struct unicosmk_dex {
9887 struct unicosmk_dex *next;
9888 const char *name;
9889 };
9890
9891 /* List of identifiers which have been replaced by DEX expressions. The DEX
9892 number is determined by the position in the list. */
9893
9894 static struct unicosmk_dex *unicosmk_dex_list = NULL;
9895
9896 /* The number of elements in the DEX list. */
9897
9898 static int unicosmk_dex_count = 0;
9899
9900 /* Check if NAME must be replaced by a DEX expression. */
9901
9902 static int
9903 unicosmk_special_name (const char *name)
9904 {
9905 if (name[0] == '*')
9906 ++name;
9907
9908 if (name[0] == '$')
9909 ++name;
9910
9911 if (name[0] != 'r' && name[0] != 'f' && name[0] != 'R' && name[0] != 'F')
9912 return 0;
9913
9914 switch (name[1])
9915 {
9916 case '1': case '2':
9917 return (name[2] == '\0' || (ISDIGIT (name[2]) && name[3] == '\0'));
9918
9919 case '3':
9920 return (name[2] == '\0'
9921 || ((name[2] == '0' || name[2] == '1') && name[3] == '\0'));
9922
9923 default:
9924 return (ISDIGIT (name[1]) && name[2] == '\0');
9925 }
9926 }
9927
9928 /* Return the DEX number if X must be replaced by a DEX expression and 0
9929 otherwise. */
9930
9931 static int
9932 unicosmk_need_dex (rtx x)
9933 {
9934 struct unicosmk_dex *dex;
9935 const char *name;
9936 int i;
9937
9938 if (GET_CODE (x) != SYMBOL_REF)
9939 return 0;
9940
9941 name = XSTR (x,0);
9942 if (! unicosmk_special_name (name))
9943 return 0;
9944
9945 i = unicosmk_dex_count;
9946 for (dex = unicosmk_dex_list; dex; dex = dex->next)
9947 {
9948 if (! strcmp (name, dex->name))
9949 return i;
9950 --i;
9951 }
9952
9953 dex = (struct unicosmk_dex *) xmalloc (sizeof (struct unicosmk_dex));
9954 dex->name = name;
9955 dex->next = unicosmk_dex_list;
9956 unicosmk_dex_list = dex;
9957
9958 ++unicosmk_dex_count;
9959 return unicosmk_dex_count;
9960 }
9961
9962 /* Output the DEX definitions for this file. */
9963
9964 static void
9965 unicosmk_output_dex (FILE *file)
9966 {
9967 struct unicosmk_dex *dex;
9968 int i;
9969
9970 if (unicosmk_dex_list == NULL)
9971 return;
9972
9973 fprintf (file, "\t.dexstart\n");
9974
9975 i = unicosmk_dex_count;
9976 for (dex = unicosmk_dex_list; dex; dex = dex->next)
9977 {
9978 fprintf (file, "\tDEX (%d) = ", i);
9979 assemble_name (file, dex->name);
9980 putc ('\n', file);
9981 --i;
9982 }
9983
9984 fprintf (file, "\t.dexend\n");
9985 }
9986
9987 /* Output text that to appear at the beginning of an assembler file. */
9988
9989 static void
9990 unicosmk_file_start (void)
9991 {
9992 int i;
9993
9994 fputs ("\t.ident\t", asm_out_file);
9995 unicosmk_output_module_name (asm_out_file);
9996 fputs ("\n\n", asm_out_file);
9997
9998 /* The Unicos/Mk assembler uses different register names. Instead of trying
9999 to support them, we simply use micro definitions. */
10000
10001 /* CAM has different register names: rN for the integer register N and fN
10002 for the floating-point register N. Instead of trying to use these in
10003 alpha.md, we define the symbols $N and $fN to refer to the appropriate
10004 register. */
10005
10006 for (i = 0; i < 32; ++i)
10007 fprintf (asm_out_file, "$%d <- r%d\n", i, i);
10008
10009 for (i = 0; i < 32; ++i)
10010 fprintf (asm_out_file, "$f%d <- f%d\n", i, i);
10011
10012 putc ('\n', asm_out_file);
10013
10014 /* The .align directive fill unused space with zeroes which does not work
10015 in code sections. We define the macro 'gcc@code@align' which uses nops
10016 instead. Note that it assumes that code sections always have the
10017 biggest possible alignment since . refers to the current offset from
10018 the beginning of the section. */
10019
10020 fputs ("\t.macro gcc@code@align n\n", asm_out_file);
10021 fputs ("gcc@n@bytes = 1 << n\n", asm_out_file);
10022 fputs ("gcc@here = . % gcc@n@bytes\n", asm_out_file);
10023 fputs ("\t.if ne, gcc@here, 0\n", asm_out_file);
10024 fputs ("\t.repeat (gcc@n@bytes - gcc@here) / 4\n", asm_out_file);
10025 fputs ("\tbis r31,r31,r31\n", asm_out_file);
10026 fputs ("\t.endr\n", asm_out_file);
10027 fputs ("\t.endif\n", asm_out_file);
10028 fputs ("\t.endm gcc@code@align\n\n", asm_out_file);
10029
10030 /* Output extern declarations which should always be visible. */
10031 unicosmk_output_default_externs (asm_out_file);
10032
10033 /* Open a dummy section. We always need to be inside a section for the
10034 section-switching code to work correctly.
10035 ??? This should be a module id or something like that. I still have to
10036 figure out what the rules for those are. */
10037 fputs ("\n\t.psect\t$SG00000,data\n", asm_out_file);
10038 }
10039
10040 /* Output text to appear at the end of an assembler file. This includes all
10041 pending extern declarations and DEX expressions. */
10042
10043 static void
10044 unicosmk_file_end (void)
10045 {
10046 fputs ("\t.endp\n\n", asm_out_file);
10047
10048 /* Output all pending externs. */
10049
10050 unicosmk_output_externs (asm_out_file);
10051
10052 /* Output dex definitions used for functions whose names conflict with
10053 register names. */
10054
10055 unicosmk_output_dex (asm_out_file);
10056
10057 fputs ("\t.end\t", asm_out_file);
10058 unicosmk_output_module_name (asm_out_file);
10059 putc ('\n', asm_out_file);
10060 }
10061
10062 #else
10063
10064 static void
10065 unicosmk_output_deferred_case_vectors (FILE *file ATTRIBUTE_UNUSED)
10066 {}
10067
10068 static void
10069 unicosmk_gen_dsib (unsigned long *imaskP ATTRIBUTE_UNUSED)
10070 {}
10071
10072 static void
10073 unicosmk_output_ssib (FILE * file ATTRIBUTE_UNUSED,
10074 const char * fnname ATTRIBUTE_UNUSED)
10075 {}
10076
10077 rtx
10078 unicosmk_add_call_info_word (rtx x ATTRIBUTE_UNUSED)
10079 {
10080 return NULL_RTX;
10081 }
10082
10083 static int
10084 unicosmk_need_dex (rtx x ATTRIBUTE_UNUSED)
10085 {
10086 return 0;
10087 }
10088
10089 #endif /* TARGET_ABI_UNICOSMK */
10090
10091 static void
10092 alpha_init_libfuncs (void)
10093 {
10094 if (TARGET_ABI_UNICOSMK)
10095 {
10096 /* Prevent gcc from generating calls to __divsi3. */
10097 set_optab_libfunc (sdiv_optab, SImode, 0);
10098 set_optab_libfunc (udiv_optab, SImode, 0);
10099
10100 /* Use the functions provided by the system library
10101 for DImode integer division. */
10102 set_optab_libfunc (sdiv_optab, DImode, "$sldiv");
10103 set_optab_libfunc (udiv_optab, DImode, "$uldiv");
10104 }
10105 else if (TARGET_ABI_OPEN_VMS)
10106 {
10107 /* Use the VMS runtime library functions for division and
10108 remainder. */
10109 set_optab_libfunc (sdiv_optab, SImode, "OTS$DIV_I");
10110 set_optab_libfunc (sdiv_optab, DImode, "OTS$DIV_L");
10111 set_optab_libfunc (udiv_optab, SImode, "OTS$DIV_UI");
10112 set_optab_libfunc (udiv_optab, DImode, "OTS$DIV_UL");
10113 set_optab_libfunc (smod_optab, SImode, "OTS$REM_I");
10114 set_optab_libfunc (smod_optab, DImode, "OTS$REM_L");
10115 set_optab_libfunc (umod_optab, SImode, "OTS$REM_UI");
10116 set_optab_libfunc (umod_optab, DImode, "OTS$REM_UL");
10117 }
10118 }
10119
10120 \f
10121 /* Initialize the GCC target structure. */
10122 #if TARGET_ABI_OPEN_VMS
10123 # undef TARGET_ATTRIBUTE_TABLE
10124 # define TARGET_ATTRIBUTE_TABLE vms_attribute_table
10125 # undef TARGET_SECTION_TYPE_FLAGS
10126 # define TARGET_SECTION_TYPE_FLAGS vms_section_type_flags
10127 #endif
10128
10129 #undef TARGET_IN_SMALL_DATA_P
10130 #define TARGET_IN_SMALL_DATA_P alpha_in_small_data_p
10131
10132 #if TARGET_ABI_UNICOSMK
10133 # undef TARGET_INSERT_ATTRIBUTES
10134 # define TARGET_INSERT_ATTRIBUTES unicosmk_insert_attributes
10135 # undef TARGET_SECTION_TYPE_FLAGS
10136 # define TARGET_SECTION_TYPE_FLAGS unicosmk_section_type_flags
10137 # undef TARGET_ASM_UNIQUE_SECTION
10138 # define TARGET_ASM_UNIQUE_SECTION unicosmk_unique_section
10139 # undef TARGET_ASM_GLOBALIZE_LABEL
10140 # define TARGET_ASM_GLOBALIZE_LABEL hook_void_FILEptr_constcharptr
10141 #endif
10142
10143 #undef TARGET_ASM_ALIGNED_HI_OP
10144 #define TARGET_ASM_ALIGNED_HI_OP "\t.word\t"
10145 #undef TARGET_ASM_ALIGNED_DI_OP
10146 #define TARGET_ASM_ALIGNED_DI_OP "\t.quad\t"
10147
10148 /* Default unaligned ops are provided for ELF systems. To get unaligned
10149 data for non-ELF systems, we have to turn off auto alignment. */
10150 #ifndef OBJECT_FORMAT_ELF
10151 #undef TARGET_ASM_UNALIGNED_HI_OP
10152 #define TARGET_ASM_UNALIGNED_HI_OP "\t.align 0\n\t.word\t"
10153 #undef TARGET_ASM_UNALIGNED_SI_OP
10154 #define TARGET_ASM_UNALIGNED_SI_OP "\t.align 0\n\t.long\t"
10155 #undef TARGET_ASM_UNALIGNED_DI_OP
10156 #define TARGET_ASM_UNALIGNED_DI_OP "\t.align 0\n\t.quad\t"
10157 #endif
10158
10159 #ifdef OBJECT_FORMAT_ELF
10160 #undef TARGET_ASM_SELECT_RTX_SECTION
10161 #define TARGET_ASM_SELECT_RTX_SECTION alpha_elf_select_rtx_section
10162 #endif
10163
10164 #undef TARGET_ASM_FUNCTION_END_PROLOGUE
10165 #define TARGET_ASM_FUNCTION_END_PROLOGUE alpha_output_function_end_prologue
10166
10167 #undef TARGET_INIT_LIBFUNCS
10168 #define TARGET_INIT_LIBFUNCS alpha_init_libfuncs
10169
10170 #if TARGET_ABI_UNICOSMK
10171 #undef TARGET_ASM_FILE_START
10172 #define TARGET_ASM_FILE_START unicosmk_file_start
10173 #undef TARGET_ASM_FILE_END
10174 #define TARGET_ASM_FILE_END unicosmk_file_end
10175 #else
10176 #undef TARGET_ASM_FILE_START
10177 #define TARGET_ASM_FILE_START alpha_file_start
10178 #undef TARGET_ASM_FILE_START_FILE_DIRECTIVE
10179 #define TARGET_ASM_FILE_START_FILE_DIRECTIVE true
10180 #endif
10181
10182 #undef TARGET_SCHED_ADJUST_COST
10183 #define TARGET_SCHED_ADJUST_COST alpha_adjust_cost
10184 #undef TARGET_SCHED_ISSUE_RATE
10185 #define TARGET_SCHED_ISSUE_RATE alpha_issue_rate
10186 #undef TARGET_SCHED_USE_DFA_PIPELINE_INTERFACE
10187 #define TARGET_SCHED_USE_DFA_PIPELINE_INTERFACE \
10188 alpha_use_dfa_pipeline_interface
10189 #undef TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD
10190 #define TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD \
10191 alpha_multipass_dfa_lookahead
10192
10193 #undef TARGET_HAVE_TLS
10194 #define TARGET_HAVE_TLS HAVE_AS_TLS
10195
10196 #undef TARGET_INIT_BUILTINS
10197 #define TARGET_INIT_BUILTINS alpha_init_builtins
10198 #undef TARGET_EXPAND_BUILTIN
10199 #define TARGET_EXPAND_BUILTIN alpha_expand_builtin
10200
10201 #undef TARGET_FUNCTION_OK_FOR_SIBCALL
10202 #define TARGET_FUNCTION_OK_FOR_SIBCALL alpha_function_ok_for_sibcall
10203 #undef TARGET_CANNOT_COPY_INSN_P
10204 #define TARGET_CANNOT_COPY_INSN_P alpha_cannot_copy_insn_p
10205
10206 #if TARGET_ABI_OSF
10207 #undef TARGET_ASM_OUTPUT_MI_THUNK
10208 #define TARGET_ASM_OUTPUT_MI_THUNK alpha_output_mi_thunk_osf
10209 #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
10210 #define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_tree_hwi_hwi_tree_true
10211 #endif
10212
10213 #undef TARGET_RTX_COSTS
10214 #define TARGET_RTX_COSTS alpha_rtx_costs
10215 #undef TARGET_ADDRESS_COST
10216 #define TARGET_ADDRESS_COST hook_int_rtx_0
10217
10218 #undef TARGET_MACHINE_DEPENDENT_REORG
10219 #define TARGET_MACHINE_DEPENDENT_REORG alpha_reorg
10220
10221 #undef TARGET_PROMOTE_FUNCTION_ARGS
10222 #define TARGET_PROMOTE_FUNCTION_ARGS hook_bool_tree_true
10223 #undef TARGET_PROMOTE_FUNCTION_RETURN
10224 #define TARGET_PROMOTE_FUNCTION_RETURN hook_bool_tree_true
10225 #undef TARGET_PROMOTE_PROTOTYPES
10226 #define TARGET_PROMOTE_PROTOTYPES hook_bool_tree_false
10227 #undef TARGET_STRUCT_VALUE_RTX
10228 #define TARGET_STRUCT_VALUE_RTX hook_rtx_tree_int_null
10229 #undef TARGET_RETURN_IN_MEMORY
10230 #define TARGET_RETURN_IN_MEMORY alpha_return_in_memory
10231 #undef TARGET_SETUP_INCOMING_VARARGS
10232 #define TARGET_SETUP_INCOMING_VARARGS alpha_setup_incoming_varargs
10233 #undef TARGET_STRICT_ARGUMENT_NAMING
10234 #define TARGET_STRICT_ARGUMENT_NAMING hook_bool_CUMULATIVE_ARGS_true
10235 #undef TARGET_PRETEND_OUTGOING_VARARGS_NAMED
10236 #define TARGET_PRETEND_OUTGOING_VARARGS_NAMED hook_bool_CUMULATIVE_ARGS_true
10237
10238 #undef TARGET_BUILD_BUILTIN_VA_LIST
10239 #define TARGET_BUILD_BUILTIN_VA_LIST alpha_build_builtin_va_list
10240
10241 struct gcc_target targetm = TARGET_INITIALIZER;
10242
10243 \f
10244 #include "gt-alpha.h"
10245