re PR target/65697 (__atomic memory barriers not strong enough for __sync builtins)
[gcc.git] / gcc / calls.c
1 /* Convert function calls to rtl insns, for GNU C compiler.
2 Copyright (C) 1989-2015 Free Software Foundation, Inc.
3
4 This file is part of GCC.
5
6 GCC is free software; you can redistribute it and/or modify it under
7 the terms of the GNU General Public License as published by the Free
8 Software Foundation; either version 3, or (at your option) any later
9 version.
10
11 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
12 WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
14 for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with GCC; see the file COPYING3. If not see
18 <http://www.gnu.org/licenses/>. */
19
20 #include "config.h"
21 #include "system.h"
22 #include "coretypes.h"
23 #include "tm.h"
24 #include "hard-reg-set.h"
25 #include "rtl.h"
26 #include "alias.h"
27 #include "symtab.h"
28 #include "tree.h"
29 #include "fold-const.h"
30 #include "stor-layout.h"
31 #include "varasm.h"
32 #include "stringpool.h"
33 #include "attribs.h"
34 #include "predict.h"
35 #include "function.h"
36 #include "basic-block.h"
37 #include "tree-ssa-alias.h"
38 #include "internal-fn.h"
39 #include "gimple-expr.h"
40 #include "gimple.h"
41 #include "flags.h"
42 #include "insn-config.h"
43 #include "expmed.h"
44 #include "dojump.h"
45 #include "explow.h"
46 #include "calls.h"
47 #include "emit-rtl.h"
48 #include "stmt.h"
49 #include "expr.h"
50 #include "insn-codes.h"
51 #include "optabs.h"
52 #include "libfuncs.h"
53 #include "regs.h"
54 #include "diagnostic-core.h"
55 #include "output.h"
56 #include "tm_p.h"
57 #include "timevar.h"
58 #include "sbitmap.h"
59 #include "bitmap.h"
60 #include "langhooks.h"
61 #include "target.h"
62 #include "cgraph.h"
63 #include "except.h"
64 #include "dbgcnt.h"
65 #include "rtl-iter.h"
66 #include "tree-chkp.h"
67 #include "rtl-chkp.h"
68
69
70 /* Like PREFERRED_STACK_BOUNDARY but in units of bytes, not bits. */
71 #define STACK_BYTES (PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT)
72
73 /* Data structure and subroutines used within expand_call. */
74
75 struct arg_data
76 {
77 /* Tree node for this argument. */
78 tree tree_value;
79 /* Mode for value; TYPE_MODE unless promoted. */
80 machine_mode mode;
81 /* Current RTL value for argument, or 0 if it isn't precomputed. */
82 rtx value;
83 /* Initially-compute RTL value for argument; only for const functions. */
84 rtx initial_value;
85 /* Register to pass this argument in, 0 if passed on stack, or an
86 PARALLEL if the arg is to be copied into multiple non-contiguous
87 registers. */
88 rtx reg;
89 /* Register to pass this argument in when generating tail call sequence.
90 This is not the same register as for normal calls on machines with
91 register windows. */
92 rtx tail_call_reg;
93 /* If REG is a PARALLEL, this is a copy of VALUE pulled into the correct
94 form for emit_group_move. */
95 rtx parallel_value;
96 /* If value is passed in neither reg nor stack, this field holds a number
97 of a special slot to be used. */
98 rtx special_slot;
99 /* For pointer bounds hold an index of parm bounds are bound to. -1 if
100 there is no such pointer. */
101 int pointer_arg;
102 /* If pointer_arg refers a structure, then pointer_offset holds an offset
103 of a pointer in this structure. */
104 int pointer_offset;
105 /* If REG was promoted from the actual mode of the argument expression,
106 indicates whether the promotion is sign- or zero-extended. */
107 int unsignedp;
108 /* Number of bytes to put in registers. 0 means put the whole arg
109 in registers. Also 0 if not passed in registers. */
110 int partial;
111 /* Nonzero if argument must be passed on stack.
112 Note that some arguments may be passed on the stack
113 even though pass_on_stack is zero, just because FUNCTION_ARG says so.
114 pass_on_stack identifies arguments that *cannot* go in registers. */
115 int pass_on_stack;
116 /* Some fields packaged up for locate_and_pad_parm. */
117 struct locate_and_pad_arg_data locate;
118 /* Location on the stack at which parameter should be stored. The store
119 has already been done if STACK == VALUE. */
120 rtx stack;
121 /* Location on the stack of the start of this argument slot. This can
122 differ from STACK if this arg pads downward. This location is known
123 to be aligned to TARGET_FUNCTION_ARG_BOUNDARY. */
124 rtx stack_slot;
125 /* Place that this stack area has been saved, if needed. */
126 rtx save_area;
127 /* If an argument's alignment does not permit direct copying into registers,
128 copy in smaller-sized pieces into pseudos. These are stored in a
129 block pointed to by this field. The next field says how many
130 word-sized pseudos we made. */
131 rtx *aligned_regs;
132 int n_aligned_regs;
133 };
134
135 /* A vector of one char per byte of stack space. A byte if nonzero if
136 the corresponding stack location has been used.
137 This vector is used to prevent a function call within an argument from
138 clobbering any stack already set up. */
139 static char *stack_usage_map;
140
141 /* Size of STACK_USAGE_MAP. */
142 static int highest_outgoing_arg_in_use;
143
144 /* A bitmap of virtual-incoming stack space. Bit is set if the corresponding
145 stack location's tail call argument has been already stored into the stack.
146 This bitmap is used to prevent sibling call optimization if function tries
147 to use parent's incoming argument slots when they have been already
148 overwritten with tail call arguments. */
149 static sbitmap stored_args_map;
150
151 /* stack_arg_under_construction is nonzero when an argument may be
152 initialized with a constructor call (including a C function that
153 returns a BLKmode struct) and expand_call must take special action
154 to make sure the object being constructed does not overlap the
155 argument list for the constructor call. */
156 static int stack_arg_under_construction;
157
158 static void emit_call_1 (rtx, tree, tree, tree, HOST_WIDE_INT, HOST_WIDE_INT,
159 HOST_WIDE_INT, rtx, rtx, int, rtx, int,
160 cumulative_args_t);
161 static void precompute_register_parameters (int, struct arg_data *, int *);
162 static void store_bounds (struct arg_data *, struct arg_data *);
163 static int store_one_arg (struct arg_data *, rtx, int, int, int);
164 static void store_unaligned_arguments_into_pseudos (struct arg_data *, int);
165 static int finalize_must_preallocate (int, int, struct arg_data *,
166 struct args_size *);
167 static void precompute_arguments (int, struct arg_data *);
168 static int compute_argument_block_size (int, struct args_size *, tree, tree, int);
169 static void initialize_argument_information (int, struct arg_data *,
170 struct args_size *, int,
171 tree, tree,
172 tree, tree, cumulative_args_t, int,
173 rtx *, int *, int *, int *,
174 bool *, bool);
175 static void compute_argument_addresses (struct arg_data *, rtx, int);
176 static rtx rtx_for_function_call (tree, tree);
177 static void load_register_parameters (struct arg_data *, int, rtx *, int,
178 int, int *);
179 static rtx emit_library_call_value_1 (int, rtx, rtx, enum libcall_type,
180 machine_mode, int, va_list);
181 static int special_function_p (const_tree, int);
182 static int check_sibcall_argument_overlap_1 (rtx);
183 static int check_sibcall_argument_overlap (rtx_insn *, struct arg_data *, int);
184
185 static int combine_pending_stack_adjustment_and_call (int, struct args_size *,
186 unsigned int);
187 static tree split_complex_types (tree);
188
189 #ifdef REG_PARM_STACK_SPACE
190 static rtx save_fixed_argument_area (int, rtx, int *, int *);
191 static void restore_fixed_argument_area (rtx, rtx, int, int);
192 #endif
193 \f
194 /* Force FUNEXP into a form suitable for the address of a CALL,
195 and return that as an rtx. Also load the static chain register
196 if FNDECL is a nested function.
197
198 CALL_FUSAGE points to a variable holding the prospective
199 CALL_INSN_FUNCTION_USAGE information. */
200
201 rtx
202 prepare_call_address (tree fndecl_or_type, rtx funexp, rtx static_chain_value,
203 rtx *call_fusage, int reg_parm_seen, int sibcallp)
204 {
205 /* Make a valid memory address and copy constants through pseudo-regs,
206 but not for a constant address if -fno-function-cse. */
207 if (GET_CODE (funexp) != SYMBOL_REF)
208 /* If we are using registers for parameters, force the
209 function address into a register now. */
210 funexp = ((reg_parm_seen
211 && targetm.small_register_classes_for_mode_p (FUNCTION_MODE))
212 ? force_not_mem (memory_address (FUNCTION_MODE, funexp))
213 : memory_address (FUNCTION_MODE, funexp));
214 else if (flag_pic
215 && fndecl_or_type
216 && TREE_CODE (fndecl_or_type) == FUNCTION_DECL
217 && (!flag_plt
218 || lookup_attribute ("noplt", DECL_ATTRIBUTES (fndecl_or_type)))
219 && !targetm.binds_local_p (fndecl_or_type))
220 {
221 /* This is done only for PIC code. There is no easy interface to force the
222 function address into GOT for non-PIC case. non-PIC case needs to be
223 handled specially by the backend. */
224 funexp = force_reg (Pmode, funexp);
225 }
226 else if (! sibcallp)
227 {
228 if (!NO_FUNCTION_CSE && optimize && ! flag_no_function_cse)
229 funexp = force_reg (Pmode, funexp);
230 }
231
232 if (static_chain_value != 0
233 && (TREE_CODE (fndecl_or_type) != FUNCTION_DECL
234 || DECL_STATIC_CHAIN (fndecl_or_type)))
235 {
236 rtx chain;
237
238 chain = targetm.calls.static_chain (fndecl_or_type, false);
239 static_chain_value = convert_memory_address (Pmode, static_chain_value);
240
241 emit_move_insn (chain, static_chain_value);
242 if (REG_P (chain))
243 use_reg (call_fusage, chain);
244 }
245
246 return funexp;
247 }
248
249 /* Generate instructions to call function FUNEXP,
250 and optionally pop the results.
251 The CALL_INSN is the first insn generated.
252
253 FNDECL is the declaration node of the function. This is given to the
254 hook TARGET_RETURN_POPS_ARGS to determine whether this function pops
255 its own args.
256
257 FUNTYPE is the data type of the function. This is given to the hook
258 TARGET_RETURN_POPS_ARGS to determine whether this function pops its
259 own args. We used to allow an identifier for library functions, but
260 that doesn't work when the return type is an aggregate type and the
261 calling convention says that the pointer to this aggregate is to be
262 popped by the callee.
263
264 STACK_SIZE is the number of bytes of arguments on the stack,
265 ROUNDED_STACK_SIZE is that number rounded up to
266 PREFERRED_STACK_BOUNDARY; zero if the size is variable. This is
267 both to put into the call insn and to generate explicit popping
268 code if necessary.
269
270 STRUCT_VALUE_SIZE is the number of bytes wanted in a structure value.
271 It is zero if this call doesn't want a structure value.
272
273 NEXT_ARG_REG is the rtx that results from executing
274 targetm.calls.function_arg (&args_so_far, VOIDmode, void_type_node, true)
275 just after all the args have had their registers assigned.
276 This could be whatever you like, but normally it is the first
277 arg-register beyond those used for args in this call,
278 or 0 if all the arg-registers are used in this call.
279 It is passed on to `gen_call' so you can put this info in the call insn.
280
281 VALREG is a hard register in which a value is returned,
282 or 0 if the call does not return a value.
283
284 OLD_INHIBIT_DEFER_POP is the value that `inhibit_defer_pop' had before
285 the args to this call were processed.
286 We restore `inhibit_defer_pop' to that value.
287
288 CALL_FUSAGE is either empty or an EXPR_LIST of USE expressions that
289 denote registers used by the called function. */
290
291 static void
292 emit_call_1 (rtx funexp, tree fntree ATTRIBUTE_UNUSED, tree fndecl ATTRIBUTE_UNUSED,
293 tree funtype ATTRIBUTE_UNUSED,
294 HOST_WIDE_INT stack_size ATTRIBUTE_UNUSED,
295 HOST_WIDE_INT rounded_stack_size,
296 HOST_WIDE_INT struct_value_size ATTRIBUTE_UNUSED,
297 rtx next_arg_reg ATTRIBUTE_UNUSED, rtx valreg,
298 int old_inhibit_defer_pop, rtx call_fusage, int ecf_flags,
299 cumulative_args_t args_so_far ATTRIBUTE_UNUSED)
300 {
301 rtx rounded_stack_size_rtx = GEN_INT (rounded_stack_size);
302 rtx call, funmem;
303 int already_popped = 0;
304 HOST_WIDE_INT n_popped
305 = targetm.calls.return_pops_args (fndecl, funtype, stack_size);
306
307 #ifdef CALL_POPS_ARGS
308 n_popped += CALL_POPS_ARGS (*get_cumulative_args (args_so_far));
309 #endif
310
311 /* Ensure address is valid. SYMBOL_REF is already valid, so no need,
312 and we don't want to load it into a register as an optimization,
313 because prepare_call_address already did it if it should be done. */
314 if (GET_CODE (funexp) != SYMBOL_REF)
315 funexp = memory_address (FUNCTION_MODE, funexp);
316
317 funmem = gen_rtx_MEM (FUNCTION_MODE, funexp);
318 if (fndecl && TREE_CODE (fndecl) == FUNCTION_DECL)
319 {
320 tree t = fndecl;
321
322 /* Although a built-in FUNCTION_DECL and its non-__builtin
323 counterpart compare equal and get a shared mem_attrs, they
324 produce different dump output in compare-debug compilations,
325 if an entry gets garbage collected in one compilation, then
326 adds a different (but equivalent) entry, while the other
327 doesn't run the garbage collector at the same spot and then
328 shares the mem_attr with the equivalent entry. */
329 if (DECL_BUILT_IN_CLASS (t) == BUILT_IN_NORMAL)
330 {
331 tree t2 = builtin_decl_explicit (DECL_FUNCTION_CODE (t));
332 if (t2)
333 t = t2;
334 }
335
336 set_mem_expr (funmem, t);
337 }
338 else if (fntree)
339 set_mem_expr (funmem, build_simple_mem_ref (CALL_EXPR_FN (fntree)));
340
341 #if defined (HAVE_sibcall_pop) && defined (HAVE_sibcall_value_pop)
342 if ((ecf_flags & ECF_SIBCALL)
343 && HAVE_sibcall_pop && HAVE_sibcall_value_pop
344 && (n_popped > 0 || stack_size == 0))
345 {
346 rtx n_pop = GEN_INT (n_popped);
347 rtx pat;
348
349 /* If this subroutine pops its own args, record that in the call insn
350 if possible, for the sake of frame pointer elimination. */
351
352 if (valreg)
353 pat = GEN_SIBCALL_VALUE_POP (valreg, funmem, rounded_stack_size_rtx,
354 next_arg_reg, n_pop);
355 else
356 pat = GEN_SIBCALL_POP (funmem, rounded_stack_size_rtx, next_arg_reg,
357 n_pop);
358
359 emit_call_insn (pat);
360 already_popped = 1;
361 }
362 else
363 #endif
364
365 #if defined (HAVE_call_pop) && defined (HAVE_call_value_pop)
366 /* If the target has "call" or "call_value" insns, then prefer them
367 if no arguments are actually popped. If the target does not have
368 "call" or "call_value" insns, then we must use the popping versions
369 even if the call has no arguments to pop. */
370 #if defined (HAVE_call) && defined (HAVE_call_value)
371 if (HAVE_call && HAVE_call_value && HAVE_call_pop && HAVE_call_value_pop
372 && n_popped > 0)
373 #else
374 if (HAVE_call_pop && HAVE_call_value_pop)
375 #endif
376 {
377 rtx n_pop = GEN_INT (n_popped);
378 rtx pat;
379
380 /* If this subroutine pops its own args, record that in the call insn
381 if possible, for the sake of frame pointer elimination. */
382
383 if (valreg)
384 pat = GEN_CALL_VALUE_POP (valreg, funmem, rounded_stack_size_rtx,
385 next_arg_reg, n_pop);
386 else
387 pat = GEN_CALL_POP (funmem, rounded_stack_size_rtx, next_arg_reg,
388 n_pop);
389
390 emit_call_insn (pat);
391 already_popped = 1;
392 }
393 else
394 #endif
395
396 #if defined (HAVE_sibcall) && defined (HAVE_sibcall_value)
397 if ((ecf_flags & ECF_SIBCALL)
398 && HAVE_sibcall && HAVE_sibcall_value)
399 {
400 if (valreg)
401 emit_call_insn (GEN_SIBCALL_VALUE (valreg, funmem,
402 rounded_stack_size_rtx,
403 next_arg_reg, NULL_RTX));
404 else
405 emit_call_insn (GEN_SIBCALL (funmem, rounded_stack_size_rtx,
406 next_arg_reg,
407 GEN_INT (struct_value_size)));
408 }
409 else
410 #endif
411
412 #if defined (HAVE_call) && defined (HAVE_call_value)
413 if (HAVE_call && HAVE_call_value)
414 {
415 if (valreg)
416 emit_call_insn (GEN_CALL_VALUE (valreg, funmem, rounded_stack_size_rtx,
417 next_arg_reg, NULL_RTX));
418 else
419 emit_call_insn (GEN_CALL (funmem, rounded_stack_size_rtx, next_arg_reg,
420 GEN_INT (struct_value_size)));
421 }
422 else
423 #endif
424 gcc_unreachable ();
425
426 /* Find the call we just emitted. */
427 rtx_call_insn *call_insn = last_call_insn ();
428
429 /* Some target create a fresh MEM instead of reusing the one provided
430 above. Set its MEM_EXPR. */
431 call = get_call_rtx_from (call_insn);
432 if (call
433 && MEM_EXPR (XEXP (call, 0)) == NULL_TREE
434 && MEM_EXPR (funmem) != NULL_TREE)
435 set_mem_expr (XEXP (call, 0), MEM_EXPR (funmem));
436
437 /* Mark instrumented calls. */
438 if (call && fntree)
439 CALL_EXPR_WITH_BOUNDS_P (call) = CALL_WITH_BOUNDS_P (fntree);
440
441 /* Put the register usage information there. */
442 add_function_usage_to (call_insn, call_fusage);
443
444 /* If this is a const call, then set the insn's unchanging bit. */
445 if (ecf_flags & ECF_CONST)
446 RTL_CONST_CALL_P (call_insn) = 1;
447
448 /* If this is a pure call, then set the insn's unchanging bit. */
449 if (ecf_flags & ECF_PURE)
450 RTL_PURE_CALL_P (call_insn) = 1;
451
452 /* If this is a const call, then set the insn's unchanging bit. */
453 if (ecf_flags & ECF_LOOPING_CONST_OR_PURE)
454 RTL_LOOPING_CONST_OR_PURE_CALL_P (call_insn) = 1;
455
456 /* Create a nothrow REG_EH_REGION note, if needed. */
457 make_reg_eh_region_note (call_insn, ecf_flags, 0);
458
459 if (ecf_flags & ECF_NORETURN)
460 add_reg_note (call_insn, REG_NORETURN, const0_rtx);
461
462 if (ecf_flags & ECF_RETURNS_TWICE)
463 {
464 add_reg_note (call_insn, REG_SETJMP, const0_rtx);
465 cfun->calls_setjmp = 1;
466 }
467
468 SIBLING_CALL_P (call_insn) = ((ecf_flags & ECF_SIBCALL) != 0);
469
470 /* Restore this now, so that we do defer pops for this call's args
471 if the context of the call as a whole permits. */
472 inhibit_defer_pop = old_inhibit_defer_pop;
473
474 if (n_popped > 0)
475 {
476 if (!already_popped)
477 CALL_INSN_FUNCTION_USAGE (call_insn)
478 = gen_rtx_EXPR_LIST (VOIDmode,
479 gen_rtx_CLOBBER (VOIDmode, stack_pointer_rtx),
480 CALL_INSN_FUNCTION_USAGE (call_insn));
481 rounded_stack_size -= n_popped;
482 rounded_stack_size_rtx = GEN_INT (rounded_stack_size);
483 stack_pointer_delta -= n_popped;
484
485 add_reg_note (call_insn, REG_ARGS_SIZE, GEN_INT (stack_pointer_delta));
486
487 /* If popup is needed, stack realign must use DRAP */
488 if (SUPPORTS_STACK_ALIGNMENT)
489 crtl->need_drap = true;
490 }
491 /* For noreturn calls when not accumulating outgoing args force
492 REG_ARGS_SIZE note to prevent crossjumping of calls with different
493 args sizes. */
494 else if (!ACCUMULATE_OUTGOING_ARGS && (ecf_flags & ECF_NORETURN) != 0)
495 add_reg_note (call_insn, REG_ARGS_SIZE, GEN_INT (stack_pointer_delta));
496
497 if (!ACCUMULATE_OUTGOING_ARGS)
498 {
499 /* If returning from the subroutine does not automatically pop the args,
500 we need an instruction to pop them sooner or later.
501 Perhaps do it now; perhaps just record how much space to pop later.
502
503 If returning from the subroutine does pop the args, indicate that the
504 stack pointer will be changed. */
505
506 if (rounded_stack_size != 0)
507 {
508 if (ecf_flags & ECF_NORETURN)
509 /* Just pretend we did the pop. */
510 stack_pointer_delta -= rounded_stack_size;
511 else if (flag_defer_pop && inhibit_defer_pop == 0
512 && ! (ecf_flags & (ECF_CONST | ECF_PURE)))
513 pending_stack_adjust += rounded_stack_size;
514 else
515 adjust_stack (rounded_stack_size_rtx);
516 }
517 }
518 /* When we accumulate outgoing args, we must avoid any stack manipulations.
519 Restore the stack pointer to its original value now. Usually
520 ACCUMULATE_OUTGOING_ARGS targets don't get here, but there are exceptions.
521 On i386 ACCUMULATE_OUTGOING_ARGS can be enabled on demand, and
522 popping variants of functions exist as well.
523
524 ??? We may optimize similar to defer_pop above, but it is
525 probably not worthwhile.
526
527 ??? It will be worthwhile to enable combine_stack_adjustments even for
528 such machines. */
529 else if (n_popped)
530 anti_adjust_stack (GEN_INT (n_popped));
531 }
532
533 /* Determine if the function identified by NAME and FNDECL is one with
534 special properties we wish to know about.
535
536 For example, if the function might return more than one time (setjmp), then
537 set RETURNS_TWICE to a nonzero value.
538
539 Similarly set NORETURN if the function is in the longjmp family.
540
541 Set MAY_BE_ALLOCA for any memory allocation function that might allocate
542 space from the stack such as alloca. */
543
544 static int
545 special_function_p (const_tree fndecl, int flags)
546 {
547 tree name_decl = DECL_NAME (fndecl);
548
549 /* For instrumentation clones we want to derive flags
550 from the original name. */
551 if (cgraph_node::get (fndecl)
552 && cgraph_node::get (fndecl)->instrumentation_clone)
553 name_decl = DECL_NAME (cgraph_node::get (fndecl)->orig_decl);
554
555 if (fndecl && name_decl
556 && IDENTIFIER_LENGTH (name_decl) <= 17
557 /* Exclude functions not at the file scope, or not `extern',
558 since they are not the magic functions we would otherwise
559 think they are.
560 FIXME: this should be handled with attributes, not with this
561 hacky imitation of DECL_ASSEMBLER_NAME. It's (also) wrong
562 because you can declare fork() inside a function if you
563 wish. */
564 && (DECL_CONTEXT (fndecl) == NULL_TREE
565 || TREE_CODE (DECL_CONTEXT (fndecl)) == TRANSLATION_UNIT_DECL)
566 && TREE_PUBLIC (fndecl))
567 {
568 const char *name = IDENTIFIER_POINTER (name_decl);
569 const char *tname = name;
570
571 /* We assume that alloca will always be called by name. It
572 makes no sense to pass it as a pointer-to-function to
573 anything that does not understand its behavior. */
574 if (((IDENTIFIER_LENGTH (name_decl) == 6
575 && name[0] == 'a'
576 && ! strcmp (name, "alloca"))
577 || (IDENTIFIER_LENGTH (name_decl) == 16
578 && name[0] == '_'
579 && ! strcmp (name, "__builtin_alloca"))))
580 flags |= ECF_MAY_BE_ALLOCA;
581
582 /* Disregard prefix _, __, __x or __builtin_. */
583 if (name[0] == '_')
584 {
585 if (name[1] == '_'
586 && name[2] == 'b'
587 && !strncmp (name + 3, "uiltin_", 7))
588 tname += 10;
589 else if (name[1] == '_' && name[2] == 'x')
590 tname += 3;
591 else if (name[1] == '_')
592 tname += 2;
593 else
594 tname += 1;
595 }
596
597 if (tname[0] == 's')
598 {
599 if ((tname[1] == 'e'
600 && (! strcmp (tname, "setjmp")
601 || ! strcmp (tname, "setjmp_syscall")))
602 || (tname[1] == 'i'
603 && ! strcmp (tname, "sigsetjmp"))
604 || (tname[1] == 'a'
605 && ! strcmp (tname, "savectx")))
606 flags |= ECF_RETURNS_TWICE | ECF_LEAF;
607
608 if (tname[1] == 'i'
609 && ! strcmp (tname, "siglongjmp"))
610 flags |= ECF_NORETURN;
611 }
612 else if ((tname[0] == 'q' && tname[1] == 's'
613 && ! strcmp (tname, "qsetjmp"))
614 || (tname[0] == 'v' && tname[1] == 'f'
615 && ! strcmp (tname, "vfork"))
616 || (tname[0] == 'g' && tname[1] == 'e'
617 && !strcmp (tname, "getcontext")))
618 flags |= ECF_RETURNS_TWICE | ECF_LEAF;
619
620 else if (tname[0] == 'l' && tname[1] == 'o'
621 && ! strcmp (tname, "longjmp"))
622 flags |= ECF_NORETURN;
623 }
624
625 return flags;
626 }
627
628 /* Similar to special_function_p; return a set of ERF_ flags for the
629 function FNDECL. */
630 static int
631 decl_return_flags (tree fndecl)
632 {
633 tree attr;
634 tree type = TREE_TYPE (fndecl);
635 if (!type)
636 return 0;
637
638 attr = lookup_attribute ("fn spec", TYPE_ATTRIBUTES (type));
639 if (!attr)
640 return 0;
641
642 attr = TREE_VALUE (TREE_VALUE (attr));
643 if (!attr || TREE_STRING_LENGTH (attr) < 1)
644 return 0;
645
646 switch (TREE_STRING_POINTER (attr)[0])
647 {
648 case '1':
649 case '2':
650 case '3':
651 case '4':
652 return ERF_RETURNS_ARG | (TREE_STRING_POINTER (attr)[0] - '1');
653
654 case 'm':
655 return ERF_NOALIAS;
656
657 case '.':
658 default:
659 return 0;
660 }
661 }
662
663 /* Return nonzero when FNDECL represents a call to setjmp. */
664
665 int
666 setjmp_call_p (const_tree fndecl)
667 {
668 if (DECL_IS_RETURNS_TWICE (fndecl))
669 return ECF_RETURNS_TWICE;
670 return special_function_p (fndecl, 0) & ECF_RETURNS_TWICE;
671 }
672
673
674 /* Return true if STMT is an alloca call. */
675
676 bool
677 gimple_alloca_call_p (const_gimple stmt)
678 {
679 tree fndecl;
680
681 if (!is_gimple_call (stmt))
682 return false;
683
684 fndecl = gimple_call_fndecl (stmt);
685 if (fndecl && (special_function_p (fndecl, 0) & ECF_MAY_BE_ALLOCA))
686 return true;
687
688 return false;
689 }
690
691 /* Return true when exp contains alloca call. */
692
693 bool
694 alloca_call_p (const_tree exp)
695 {
696 tree fndecl;
697 if (TREE_CODE (exp) == CALL_EXPR
698 && (fndecl = get_callee_fndecl (exp))
699 && (special_function_p (fndecl, 0) & ECF_MAY_BE_ALLOCA))
700 return true;
701 return false;
702 }
703
704 /* Return TRUE if FNDECL is either a TM builtin or a TM cloned
705 function. Return FALSE otherwise. */
706
707 static bool
708 is_tm_builtin (const_tree fndecl)
709 {
710 if (fndecl == NULL)
711 return false;
712
713 if (decl_is_tm_clone (fndecl))
714 return true;
715
716 if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_NORMAL)
717 {
718 switch (DECL_FUNCTION_CODE (fndecl))
719 {
720 case BUILT_IN_TM_COMMIT:
721 case BUILT_IN_TM_COMMIT_EH:
722 case BUILT_IN_TM_ABORT:
723 case BUILT_IN_TM_IRREVOCABLE:
724 case BUILT_IN_TM_GETTMCLONE_IRR:
725 case BUILT_IN_TM_MEMCPY:
726 case BUILT_IN_TM_MEMMOVE:
727 case BUILT_IN_TM_MEMSET:
728 CASE_BUILT_IN_TM_STORE (1):
729 CASE_BUILT_IN_TM_STORE (2):
730 CASE_BUILT_IN_TM_STORE (4):
731 CASE_BUILT_IN_TM_STORE (8):
732 CASE_BUILT_IN_TM_STORE (FLOAT):
733 CASE_BUILT_IN_TM_STORE (DOUBLE):
734 CASE_BUILT_IN_TM_STORE (LDOUBLE):
735 CASE_BUILT_IN_TM_STORE (M64):
736 CASE_BUILT_IN_TM_STORE (M128):
737 CASE_BUILT_IN_TM_STORE (M256):
738 CASE_BUILT_IN_TM_LOAD (1):
739 CASE_BUILT_IN_TM_LOAD (2):
740 CASE_BUILT_IN_TM_LOAD (4):
741 CASE_BUILT_IN_TM_LOAD (8):
742 CASE_BUILT_IN_TM_LOAD (FLOAT):
743 CASE_BUILT_IN_TM_LOAD (DOUBLE):
744 CASE_BUILT_IN_TM_LOAD (LDOUBLE):
745 CASE_BUILT_IN_TM_LOAD (M64):
746 CASE_BUILT_IN_TM_LOAD (M128):
747 CASE_BUILT_IN_TM_LOAD (M256):
748 case BUILT_IN_TM_LOG:
749 case BUILT_IN_TM_LOG_1:
750 case BUILT_IN_TM_LOG_2:
751 case BUILT_IN_TM_LOG_4:
752 case BUILT_IN_TM_LOG_8:
753 case BUILT_IN_TM_LOG_FLOAT:
754 case BUILT_IN_TM_LOG_DOUBLE:
755 case BUILT_IN_TM_LOG_LDOUBLE:
756 case BUILT_IN_TM_LOG_M64:
757 case BUILT_IN_TM_LOG_M128:
758 case BUILT_IN_TM_LOG_M256:
759 return true;
760 default:
761 break;
762 }
763 }
764 return false;
765 }
766
767 /* Detect flags (function attributes) from the function decl or type node. */
768
769 int
770 flags_from_decl_or_type (const_tree exp)
771 {
772 int flags = 0;
773
774 if (DECL_P (exp))
775 {
776 /* The function exp may have the `malloc' attribute. */
777 if (DECL_IS_MALLOC (exp))
778 flags |= ECF_MALLOC;
779
780 /* The function exp may have the `returns_twice' attribute. */
781 if (DECL_IS_RETURNS_TWICE (exp))
782 flags |= ECF_RETURNS_TWICE;
783
784 /* Process the pure and const attributes. */
785 if (TREE_READONLY (exp))
786 flags |= ECF_CONST;
787 if (DECL_PURE_P (exp))
788 flags |= ECF_PURE;
789 if (DECL_LOOPING_CONST_OR_PURE_P (exp))
790 flags |= ECF_LOOPING_CONST_OR_PURE;
791
792 if (DECL_IS_NOVOPS (exp))
793 flags |= ECF_NOVOPS;
794 if (lookup_attribute ("leaf", DECL_ATTRIBUTES (exp)))
795 flags |= ECF_LEAF;
796
797 if (TREE_NOTHROW (exp))
798 flags |= ECF_NOTHROW;
799
800 if (flag_tm)
801 {
802 if (is_tm_builtin (exp))
803 flags |= ECF_TM_BUILTIN;
804 else if ((flags & (ECF_CONST|ECF_NOVOPS)) != 0
805 || lookup_attribute ("transaction_pure",
806 TYPE_ATTRIBUTES (TREE_TYPE (exp))))
807 flags |= ECF_TM_PURE;
808 }
809
810 flags = special_function_p (exp, flags);
811 }
812 else if (TYPE_P (exp))
813 {
814 if (TYPE_READONLY (exp))
815 flags |= ECF_CONST;
816
817 if (flag_tm
818 && ((flags & ECF_CONST) != 0
819 || lookup_attribute ("transaction_pure", TYPE_ATTRIBUTES (exp))))
820 flags |= ECF_TM_PURE;
821 }
822 else
823 gcc_unreachable ();
824
825 if (TREE_THIS_VOLATILE (exp))
826 {
827 flags |= ECF_NORETURN;
828 if (flags & (ECF_CONST|ECF_PURE))
829 flags |= ECF_LOOPING_CONST_OR_PURE;
830 }
831
832 return flags;
833 }
834
835 /* Detect flags from a CALL_EXPR. */
836
837 int
838 call_expr_flags (const_tree t)
839 {
840 int flags;
841 tree decl = get_callee_fndecl (t);
842
843 if (decl)
844 flags = flags_from_decl_or_type (decl);
845 else if (CALL_EXPR_FN (t) == NULL_TREE)
846 flags = internal_fn_flags (CALL_EXPR_IFN (t));
847 else
848 {
849 t = TREE_TYPE (CALL_EXPR_FN (t));
850 if (t && TREE_CODE (t) == POINTER_TYPE)
851 flags = flags_from_decl_or_type (TREE_TYPE (t));
852 else
853 flags = 0;
854 }
855
856 return flags;
857 }
858
859 /* Return true if TYPE should be passed by invisible reference. */
860
861 bool
862 pass_by_reference (CUMULATIVE_ARGS *ca, machine_mode mode,
863 tree type, bool named_arg)
864 {
865 if (type)
866 {
867 /* If this type contains non-trivial constructors, then it is
868 forbidden for the middle-end to create any new copies. */
869 if (TREE_ADDRESSABLE (type))
870 return true;
871
872 /* GCC post 3.4 passes *all* variable sized types by reference. */
873 if (!TYPE_SIZE (type) || TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)
874 return true;
875
876 /* If a record type should be passed the same as its first (and only)
877 member, use the type and mode of that member. */
878 if (TREE_CODE (type) == RECORD_TYPE && TYPE_TRANSPARENT_AGGR (type))
879 {
880 type = TREE_TYPE (first_field (type));
881 mode = TYPE_MODE (type);
882 }
883 }
884
885 return targetm.calls.pass_by_reference (pack_cumulative_args (ca), mode,
886 type, named_arg);
887 }
888
889 /* Return true if TYPE, which is passed by reference, should be callee
890 copied instead of caller copied. */
891
892 bool
893 reference_callee_copied (CUMULATIVE_ARGS *ca, machine_mode mode,
894 tree type, bool named_arg)
895 {
896 if (type && TREE_ADDRESSABLE (type))
897 return false;
898 return targetm.calls.callee_copies (pack_cumulative_args (ca), mode, type,
899 named_arg);
900 }
901
902
903 /* Precompute all register parameters as described by ARGS, storing values
904 into fields within the ARGS array.
905
906 NUM_ACTUALS indicates the total number elements in the ARGS array.
907
908 Set REG_PARM_SEEN if we encounter a register parameter. */
909
910 static void
911 precompute_register_parameters (int num_actuals, struct arg_data *args,
912 int *reg_parm_seen)
913 {
914 int i;
915
916 *reg_parm_seen = 0;
917
918 for (i = 0; i < num_actuals; i++)
919 if (args[i].reg != 0 && ! args[i].pass_on_stack)
920 {
921 *reg_parm_seen = 1;
922
923 if (args[i].value == 0)
924 {
925 push_temp_slots ();
926 args[i].value = expand_normal (args[i].tree_value);
927 preserve_temp_slots (args[i].value);
928 pop_temp_slots ();
929 }
930
931 /* If we are to promote the function arg to a wider mode,
932 do it now. */
933
934 if (args[i].mode != TYPE_MODE (TREE_TYPE (args[i].tree_value)))
935 args[i].value
936 = convert_modes (args[i].mode,
937 TYPE_MODE (TREE_TYPE (args[i].tree_value)),
938 args[i].value, args[i].unsignedp);
939
940 /* If the value is a non-legitimate constant, force it into a
941 pseudo now. TLS symbols sometimes need a call to resolve. */
942 if (CONSTANT_P (args[i].value)
943 && !targetm.legitimate_constant_p (args[i].mode, args[i].value))
944 args[i].value = force_reg (args[i].mode, args[i].value);
945
946 /* If we're going to have to load the value by parts, pull the
947 parts into pseudos. The part extraction process can involve
948 non-trivial computation. */
949 if (GET_CODE (args[i].reg) == PARALLEL)
950 {
951 tree type = TREE_TYPE (args[i].tree_value);
952 args[i].parallel_value
953 = emit_group_load_into_temps (args[i].reg, args[i].value,
954 type, int_size_in_bytes (type));
955 }
956
957 /* If the value is expensive, and we are inside an appropriately
958 short loop, put the value into a pseudo and then put the pseudo
959 into the hard reg.
960
961 For small register classes, also do this if this call uses
962 register parameters. This is to avoid reload conflicts while
963 loading the parameters registers. */
964
965 else if ((! (REG_P (args[i].value)
966 || (GET_CODE (args[i].value) == SUBREG
967 && REG_P (SUBREG_REG (args[i].value)))))
968 && args[i].mode != BLKmode
969 && set_src_cost (args[i].value, optimize_insn_for_speed_p ())
970 > COSTS_N_INSNS (1)
971 && ((*reg_parm_seen
972 && targetm.small_register_classes_for_mode_p (args[i].mode))
973 || optimize))
974 args[i].value = copy_to_mode_reg (args[i].mode, args[i].value);
975 }
976 }
977
978 #ifdef REG_PARM_STACK_SPACE
979
980 /* The argument list is the property of the called routine and it
981 may clobber it. If the fixed area has been used for previous
982 parameters, we must save and restore it. */
983
984 static rtx
985 save_fixed_argument_area (int reg_parm_stack_space, rtx argblock, int *low_to_save, int *high_to_save)
986 {
987 int low;
988 int high;
989
990 /* Compute the boundary of the area that needs to be saved, if any. */
991 high = reg_parm_stack_space;
992 if (ARGS_GROW_DOWNWARD)
993 high += 1;
994
995 if (high > highest_outgoing_arg_in_use)
996 high = highest_outgoing_arg_in_use;
997
998 for (low = 0; low < high; low++)
999 if (stack_usage_map[low] != 0)
1000 {
1001 int num_to_save;
1002 machine_mode save_mode;
1003 int delta;
1004 rtx addr;
1005 rtx stack_area;
1006 rtx save_area;
1007
1008 while (stack_usage_map[--high] == 0)
1009 ;
1010
1011 *low_to_save = low;
1012 *high_to_save = high;
1013
1014 num_to_save = high - low + 1;
1015 save_mode = mode_for_size (num_to_save * BITS_PER_UNIT, MODE_INT, 1);
1016
1017 /* If we don't have the required alignment, must do this
1018 in BLKmode. */
1019 if ((low & (MIN (GET_MODE_SIZE (save_mode),
1020 BIGGEST_ALIGNMENT / UNITS_PER_WORD) - 1)))
1021 save_mode = BLKmode;
1022
1023 if (ARGS_GROW_DOWNWARD)
1024 delta = -high;
1025 else
1026 delta = low;
1027
1028 addr = plus_constant (Pmode, argblock, delta);
1029 stack_area = gen_rtx_MEM (save_mode, memory_address (save_mode, addr));
1030
1031 set_mem_align (stack_area, PARM_BOUNDARY);
1032 if (save_mode == BLKmode)
1033 {
1034 save_area = assign_stack_temp (BLKmode, num_to_save);
1035 emit_block_move (validize_mem (save_area), stack_area,
1036 GEN_INT (num_to_save), BLOCK_OP_CALL_PARM);
1037 }
1038 else
1039 {
1040 save_area = gen_reg_rtx (save_mode);
1041 emit_move_insn (save_area, stack_area);
1042 }
1043
1044 return save_area;
1045 }
1046
1047 return NULL_RTX;
1048 }
1049
1050 static void
1051 restore_fixed_argument_area (rtx save_area, rtx argblock, int high_to_save, int low_to_save)
1052 {
1053 machine_mode save_mode = GET_MODE (save_area);
1054 int delta;
1055 rtx addr, stack_area;
1056
1057 if (ARGS_GROW_DOWNWARD)
1058 delta = -high_to_save;
1059 else
1060 delta = low_to_save;
1061
1062 addr = plus_constant (Pmode, argblock, delta);
1063 stack_area = gen_rtx_MEM (save_mode, memory_address (save_mode, addr));
1064 set_mem_align (stack_area, PARM_BOUNDARY);
1065
1066 if (save_mode != BLKmode)
1067 emit_move_insn (stack_area, save_area);
1068 else
1069 emit_block_move (stack_area, validize_mem (save_area),
1070 GEN_INT (high_to_save - low_to_save + 1),
1071 BLOCK_OP_CALL_PARM);
1072 }
1073 #endif /* REG_PARM_STACK_SPACE */
1074
1075 /* If any elements in ARGS refer to parameters that are to be passed in
1076 registers, but not in memory, and whose alignment does not permit a
1077 direct copy into registers. Copy the values into a group of pseudos
1078 which we will later copy into the appropriate hard registers.
1079
1080 Pseudos for each unaligned argument will be stored into the array
1081 args[argnum].aligned_regs. The caller is responsible for deallocating
1082 the aligned_regs array if it is nonzero. */
1083
1084 static void
1085 store_unaligned_arguments_into_pseudos (struct arg_data *args, int num_actuals)
1086 {
1087 int i, j;
1088
1089 for (i = 0; i < num_actuals; i++)
1090 if (args[i].reg != 0 && ! args[i].pass_on_stack
1091 && GET_CODE (args[i].reg) != PARALLEL
1092 && args[i].mode == BLKmode
1093 && MEM_P (args[i].value)
1094 && (MEM_ALIGN (args[i].value)
1095 < (unsigned int) MIN (BIGGEST_ALIGNMENT, BITS_PER_WORD)))
1096 {
1097 int bytes = int_size_in_bytes (TREE_TYPE (args[i].tree_value));
1098 int endian_correction = 0;
1099
1100 if (args[i].partial)
1101 {
1102 gcc_assert (args[i].partial % UNITS_PER_WORD == 0);
1103 args[i].n_aligned_regs = args[i].partial / UNITS_PER_WORD;
1104 }
1105 else
1106 {
1107 args[i].n_aligned_regs
1108 = (bytes + UNITS_PER_WORD - 1) / UNITS_PER_WORD;
1109 }
1110
1111 args[i].aligned_regs = XNEWVEC (rtx, args[i].n_aligned_regs);
1112
1113 /* Structures smaller than a word are normally aligned to the
1114 least significant byte. On a BYTES_BIG_ENDIAN machine,
1115 this means we must skip the empty high order bytes when
1116 calculating the bit offset. */
1117 if (bytes < UNITS_PER_WORD
1118 #ifdef BLOCK_REG_PADDING
1119 && (BLOCK_REG_PADDING (args[i].mode,
1120 TREE_TYPE (args[i].tree_value), 1)
1121 == downward)
1122 #else
1123 && BYTES_BIG_ENDIAN
1124 #endif
1125 )
1126 endian_correction = BITS_PER_WORD - bytes * BITS_PER_UNIT;
1127
1128 for (j = 0; j < args[i].n_aligned_regs; j++)
1129 {
1130 rtx reg = gen_reg_rtx (word_mode);
1131 rtx word = operand_subword_force (args[i].value, j, BLKmode);
1132 int bitsize = MIN (bytes * BITS_PER_UNIT, BITS_PER_WORD);
1133
1134 args[i].aligned_regs[j] = reg;
1135 word = extract_bit_field (word, bitsize, 0, 1, NULL_RTX,
1136 word_mode, word_mode);
1137
1138 /* There is no need to restrict this code to loading items
1139 in TYPE_ALIGN sized hunks. The bitfield instructions can
1140 load up entire word sized registers efficiently.
1141
1142 ??? This may not be needed anymore.
1143 We use to emit a clobber here but that doesn't let later
1144 passes optimize the instructions we emit. By storing 0 into
1145 the register later passes know the first AND to zero out the
1146 bitfield being set in the register is unnecessary. The store
1147 of 0 will be deleted as will at least the first AND. */
1148
1149 emit_move_insn (reg, const0_rtx);
1150
1151 bytes -= bitsize / BITS_PER_UNIT;
1152 store_bit_field (reg, bitsize, endian_correction, 0, 0,
1153 word_mode, word);
1154 }
1155 }
1156 }
1157
1158 /* Fill in ARGS_SIZE and ARGS array based on the parameters found in
1159 CALL_EXPR EXP.
1160
1161 NUM_ACTUALS is the total number of parameters.
1162
1163 N_NAMED_ARGS is the total number of named arguments.
1164
1165 STRUCT_VALUE_ADDR_VALUE is the implicit argument for a struct return
1166 value, or null.
1167
1168 FNDECL is the tree code for the target of this call (if known)
1169
1170 ARGS_SO_FAR holds state needed by the target to know where to place
1171 the next argument.
1172
1173 REG_PARM_STACK_SPACE is the number of bytes of stack space reserved
1174 for arguments which are passed in registers.
1175
1176 OLD_STACK_LEVEL is a pointer to an rtx which olds the old stack level
1177 and may be modified by this routine.
1178
1179 OLD_PENDING_ADJ, MUST_PREALLOCATE and FLAGS are pointers to integer
1180 flags which may may be modified by this routine.
1181
1182 MAY_TAILCALL is cleared if we encounter an invisible pass-by-reference
1183 that requires allocation of stack space.
1184
1185 CALL_FROM_THUNK_P is true if this call is the jump from a thunk to
1186 the thunked-to function. */
1187
1188 static void
1189 initialize_argument_information (int num_actuals ATTRIBUTE_UNUSED,
1190 struct arg_data *args,
1191 struct args_size *args_size,
1192 int n_named_args ATTRIBUTE_UNUSED,
1193 tree exp, tree struct_value_addr_value,
1194 tree fndecl, tree fntype,
1195 cumulative_args_t args_so_far,
1196 int reg_parm_stack_space,
1197 rtx *old_stack_level, int *old_pending_adj,
1198 int *must_preallocate, int *ecf_flags,
1199 bool *may_tailcall, bool call_from_thunk_p)
1200 {
1201 CUMULATIVE_ARGS *args_so_far_pnt = get_cumulative_args (args_so_far);
1202 location_t loc = EXPR_LOCATION (exp);
1203
1204 /* Count arg position in order args appear. */
1205 int argpos;
1206
1207 int i;
1208
1209 args_size->constant = 0;
1210 args_size->var = 0;
1211
1212 bitmap_obstack_initialize (NULL);
1213
1214 /* In this loop, we consider args in the order they are written.
1215 We fill up ARGS from the back. */
1216
1217 i = num_actuals - 1;
1218 {
1219 int j = i, ptr_arg = -1;
1220 call_expr_arg_iterator iter;
1221 tree arg;
1222 bitmap slots = NULL;
1223
1224 if (struct_value_addr_value)
1225 {
1226 args[j].tree_value = struct_value_addr_value;
1227 j--;
1228
1229 /* If we pass structure address then we need to
1230 create bounds for it. Since created bounds is
1231 a call statement, we expand it right here to avoid
1232 fixing all other places where it may be expanded. */
1233 if (CALL_WITH_BOUNDS_P (exp))
1234 {
1235 args[j].value = gen_reg_rtx (targetm.chkp_bound_mode ());
1236 args[j].tree_value
1237 = chkp_make_bounds_for_struct_addr (struct_value_addr_value);
1238 expand_expr_real (args[j].tree_value, args[j].value, VOIDmode,
1239 EXPAND_NORMAL, 0, false);
1240 args[j].pointer_arg = j + 1;
1241 j--;
1242 }
1243 }
1244 FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
1245 {
1246 tree argtype = TREE_TYPE (arg);
1247
1248 /* Remember last param with pointer and associate it
1249 with following pointer bounds. */
1250 if (CALL_WITH_BOUNDS_P (exp)
1251 && chkp_type_has_pointer (argtype))
1252 {
1253 if (slots)
1254 BITMAP_FREE (slots);
1255 ptr_arg = j;
1256 if (!BOUNDED_TYPE_P (argtype))
1257 {
1258 slots = BITMAP_ALLOC (NULL);
1259 chkp_find_bound_slots (argtype, slots);
1260 }
1261 }
1262 else if (POINTER_BOUNDS_TYPE_P (argtype))
1263 {
1264 /* We expect bounds in instrumented calls only.
1265 Otherwise it is a sign we lost flag due to some optimization
1266 and may emit call args incorrectly. */
1267 gcc_assert (CALL_WITH_BOUNDS_P (exp));
1268
1269 /* For structures look for the next available pointer. */
1270 if (ptr_arg != -1 && slots)
1271 {
1272 unsigned bnd_no = bitmap_first_set_bit (slots);
1273 args[j].pointer_offset =
1274 bnd_no * POINTER_SIZE / BITS_PER_UNIT;
1275
1276 bitmap_clear_bit (slots, bnd_no);
1277
1278 /* Check we have no more pointers in the structure. */
1279 if (bitmap_empty_p (slots))
1280 BITMAP_FREE (slots);
1281 }
1282 args[j].pointer_arg = ptr_arg;
1283
1284 /* Check we covered all pointers in the previous
1285 non bounds arg. */
1286 if (!slots)
1287 ptr_arg = -1;
1288 }
1289 else
1290 ptr_arg = -1;
1291
1292 if (targetm.calls.split_complex_arg
1293 && argtype
1294 && TREE_CODE (argtype) == COMPLEX_TYPE
1295 && targetm.calls.split_complex_arg (argtype))
1296 {
1297 tree subtype = TREE_TYPE (argtype);
1298 args[j].tree_value = build1 (REALPART_EXPR, subtype, arg);
1299 j--;
1300 args[j].tree_value = build1 (IMAGPART_EXPR, subtype, arg);
1301 }
1302 else
1303 args[j].tree_value = arg;
1304 j--;
1305 }
1306
1307 if (slots)
1308 BITMAP_FREE (slots);
1309 }
1310
1311 bitmap_obstack_release (NULL);
1312
1313 /* I counts args in order (to be) pushed; ARGPOS counts in order written. */
1314 for (argpos = 0; argpos < num_actuals; i--, argpos++)
1315 {
1316 tree type = TREE_TYPE (args[i].tree_value);
1317 int unsignedp;
1318 machine_mode mode;
1319
1320 /* Replace erroneous argument with constant zero. */
1321 if (type == error_mark_node || !COMPLETE_TYPE_P (type))
1322 args[i].tree_value = integer_zero_node, type = integer_type_node;
1323
1324 /* If TYPE is a transparent union or record, pass things the way
1325 we would pass the first field of the union or record. We have
1326 already verified that the modes are the same. */
1327 if ((TREE_CODE (type) == UNION_TYPE || TREE_CODE (type) == RECORD_TYPE)
1328 && TYPE_TRANSPARENT_AGGR (type))
1329 type = TREE_TYPE (first_field (type));
1330
1331 /* Decide where to pass this arg.
1332
1333 args[i].reg is nonzero if all or part is passed in registers.
1334
1335 args[i].partial is nonzero if part but not all is passed in registers,
1336 and the exact value says how many bytes are passed in registers.
1337
1338 args[i].pass_on_stack is nonzero if the argument must at least be
1339 computed on the stack. It may then be loaded back into registers
1340 if args[i].reg is nonzero.
1341
1342 These decisions are driven by the FUNCTION_... macros and must agree
1343 with those made by function.c. */
1344
1345 /* See if this argument should be passed by invisible reference. */
1346 if (pass_by_reference (args_so_far_pnt, TYPE_MODE (type),
1347 type, argpos < n_named_args))
1348 {
1349 bool callee_copies;
1350 tree base = NULL_TREE;
1351
1352 callee_copies
1353 = reference_callee_copied (args_so_far_pnt, TYPE_MODE (type),
1354 type, argpos < n_named_args);
1355
1356 /* If we're compiling a thunk, pass through invisible references
1357 instead of making a copy. */
1358 if (call_from_thunk_p
1359 || (callee_copies
1360 && !TREE_ADDRESSABLE (type)
1361 && (base = get_base_address (args[i].tree_value))
1362 && TREE_CODE (base) != SSA_NAME
1363 && (!DECL_P (base) || MEM_P (DECL_RTL (base)))))
1364 {
1365 /* We may have turned the parameter value into an SSA name.
1366 Go back to the original parameter so we can take the
1367 address. */
1368 if (TREE_CODE (args[i].tree_value) == SSA_NAME)
1369 {
1370 gcc_assert (SSA_NAME_IS_DEFAULT_DEF (args[i].tree_value));
1371 args[i].tree_value = SSA_NAME_VAR (args[i].tree_value);
1372 gcc_assert (TREE_CODE (args[i].tree_value) == PARM_DECL);
1373 }
1374 /* Argument setup code may have copied the value to register. We
1375 revert that optimization now because the tail call code must
1376 use the original location. */
1377 if (TREE_CODE (args[i].tree_value) == PARM_DECL
1378 && !MEM_P (DECL_RTL (args[i].tree_value))
1379 && DECL_INCOMING_RTL (args[i].tree_value)
1380 && MEM_P (DECL_INCOMING_RTL (args[i].tree_value)))
1381 set_decl_rtl (args[i].tree_value,
1382 DECL_INCOMING_RTL (args[i].tree_value));
1383
1384 mark_addressable (args[i].tree_value);
1385
1386 /* We can't use sibcalls if a callee-copied argument is
1387 stored in the current function's frame. */
1388 if (!call_from_thunk_p && DECL_P (base) && !TREE_STATIC (base))
1389 *may_tailcall = false;
1390
1391 args[i].tree_value = build_fold_addr_expr_loc (loc,
1392 args[i].tree_value);
1393 type = TREE_TYPE (args[i].tree_value);
1394
1395 if (*ecf_flags & ECF_CONST)
1396 *ecf_flags &= ~(ECF_CONST | ECF_LOOPING_CONST_OR_PURE);
1397 }
1398 else
1399 {
1400 /* We make a copy of the object and pass the address to the
1401 function being called. */
1402 rtx copy;
1403
1404 if (!COMPLETE_TYPE_P (type)
1405 || TREE_CODE (TYPE_SIZE_UNIT (type)) != INTEGER_CST
1406 || (flag_stack_check == GENERIC_STACK_CHECK
1407 && compare_tree_int (TYPE_SIZE_UNIT (type),
1408 STACK_CHECK_MAX_VAR_SIZE) > 0))
1409 {
1410 /* This is a variable-sized object. Make space on the stack
1411 for it. */
1412 rtx size_rtx = expr_size (args[i].tree_value);
1413
1414 if (*old_stack_level == 0)
1415 {
1416 emit_stack_save (SAVE_BLOCK, old_stack_level);
1417 *old_pending_adj = pending_stack_adjust;
1418 pending_stack_adjust = 0;
1419 }
1420
1421 /* We can pass TRUE as the 4th argument because we just
1422 saved the stack pointer and will restore it right after
1423 the call. */
1424 copy = allocate_dynamic_stack_space (size_rtx,
1425 TYPE_ALIGN (type),
1426 TYPE_ALIGN (type),
1427 true);
1428 copy = gen_rtx_MEM (BLKmode, copy);
1429 set_mem_attributes (copy, type, 1);
1430 }
1431 else
1432 copy = assign_temp (type, 1, 0);
1433
1434 store_expr (args[i].tree_value, copy, 0, false);
1435
1436 /* Just change the const function to pure and then let
1437 the next test clear the pure based on
1438 callee_copies. */
1439 if (*ecf_flags & ECF_CONST)
1440 {
1441 *ecf_flags &= ~ECF_CONST;
1442 *ecf_flags |= ECF_PURE;
1443 }
1444
1445 if (!callee_copies && *ecf_flags & ECF_PURE)
1446 *ecf_flags &= ~(ECF_PURE | ECF_LOOPING_CONST_OR_PURE);
1447
1448 args[i].tree_value
1449 = build_fold_addr_expr_loc (loc, make_tree (type, copy));
1450 type = TREE_TYPE (args[i].tree_value);
1451 *may_tailcall = false;
1452 }
1453 }
1454
1455 unsignedp = TYPE_UNSIGNED (type);
1456 mode = promote_function_mode (type, TYPE_MODE (type), &unsignedp,
1457 fndecl ? TREE_TYPE (fndecl) : fntype, 0);
1458
1459 args[i].unsignedp = unsignedp;
1460 args[i].mode = mode;
1461
1462 args[i].reg = targetm.calls.function_arg (args_so_far, mode, type,
1463 argpos < n_named_args);
1464
1465 if (args[i].reg && CONST_INT_P (args[i].reg))
1466 {
1467 args[i].special_slot = args[i].reg;
1468 args[i].reg = NULL;
1469 }
1470
1471 /* If this is a sibling call and the machine has register windows, the
1472 register window has to be unwinded before calling the routine, so
1473 arguments have to go into the incoming registers. */
1474 if (targetm.calls.function_incoming_arg != targetm.calls.function_arg)
1475 args[i].tail_call_reg
1476 = targetm.calls.function_incoming_arg (args_so_far, mode, type,
1477 argpos < n_named_args);
1478 else
1479 args[i].tail_call_reg = args[i].reg;
1480
1481 if (args[i].reg)
1482 args[i].partial
1483 = targetm.calls.arg_partial_bytes (args_so_far, mode, type,
1484 argpos < n_named_args);
1485
1486 args[i].pass_on_stack = targetm.calls.must_pass_in_stack (mode, type);
1487
1488 /* If FUNCTION_ARG returned a (parallel [(expr_list (nil) ...) ...]),
1489 it means that we are to pass this arg in the register(s) designated
1490 by the PARALLEL, but also to pass it in the stack. */
1491 if (args[i].reg && GET_CODE (args[i].reg) == PARALLEL
1492 && XEXP (XVECEXP (args[i].reg, 0, 0), 0) == 0)
1493 args[i].pass_on_stack = 1;
1494
1495 /* If this is an addressable type, we must preallocate the stack
1496 since we must evaluate the object into its final location.
1497
1498 If this is to be passed in both registers and the stack, it is simpler
1499 to preallocate. */
1500 if (TREE_ADDRESSABLE (type)
1501 || (args[i].pass_on_stack && args[i].reg != 0))
1502 *must_preallocate = 1;
1503
1504 /* No stack allocation and padding for bounds. */
1505 if (POINTER_BOUNDS_P (args[i].tree_value))
1506 ;
1507 /* Compute the stack-size of this argument. */
1508 else if (args[i].reg == 0 || args[i].partial != 0
1509 || reg_parm_stack_space > 0
1510 || args[i].pass_on_stack)
1511 locate_and_pad_parm (mode, type,
1512 #ifdef STACK_PARMS_IN_REG_PARM_AREA
1513 1,
1514 #else
1515 args[i].reg != 0,
1516 #endif
1517 reg_parm_stack_space,
1518 args[i].pass_on_stack ? 0 : args[i].partial,
1519 fndecl, args_size, &args[i].locate);
1520 #ifdef BLOCK_REG_PADDING
1521 else
1522 /* The argument is passed entirely in registers. See at which
1523 end it should be padded. */
1524 args[i].locate.where_pad =
1525 BLOCK_REG_PADDING (mode, type,
1526 int_size_in_bytes (type) <= UNITS_PER_WORD);
1527 #endif
1528
1529 /* Update ARGS_SIZE, the total stack space for args so far. */
1530
1531 args_size->constant += args[i].locate.size.constant;
1532 if (args[i].locate.size.var)
1533 ADD_PARM_SIZE (*args_size, args[i].locate.size.var);
1534
1535 /* Increment ARGS_SO_FAR, which has info about which arg-registers
1536 have been used, etc. */
1537
1538 targetm.calls.function_arg_advance (args_so_far, TYPE_MODE (type),
1539 type, argpos < n_named_args);
1540 }
1541 }
1542
1543 /* Update ARGS_SIZE to contain the total size for the argument block.
1544 Return the original constant component of the argument block's size.
1545
1546 REG_PARM_STACK_SPACE holds the number of bytes of stack space reserved
1547 for arguments passed in registers. */
1548
1549 static int
1550 compute_argument_block_size (int reg_parm_stack_space,
1551 struct args_size *args_size,
1552 tree fndecl ATTRIBUTE_UNUSED,
1553 tree fntype ATTRIBUTE_UNUSED,
1554 int preferred_stack_boundary ATTRIBUTE_UNUSED)
1555 {
1556 int unadjusted_args_size = args_size->constant;
1557
1558 /* For accumulate outgoing args mode we don't need to align, since the frame
1559 will be already aligned. Align to STACK_BOUNDARY in order to prevent
1560 backends from generating misaligned frame sizes. */
1561 if (ACCUMULATE_OUTGOING_ARGS && preferred_stack_boundary > STACK_BOUNDARY)
1562 preferred_stack_boundary = STACK_BOUNDARY;
1563
1564 /* Compute the actual size of the argument block required. The variable
1565 and constant sizes must be combined, the size may have to be rounded,
1566 and there may be a minimum required size. */
1567
1568 if (args_size->var)
1569 {
1570 args_size->var = ARGS_SIZE_TREE (*args_size);
1571 args_size->constant = 0;
1572
1573 preferred_stack_boundary /= BITS_PER_UNIT;
1574 if (preferred_stack_boundary > 1)
1575 {
1576 /* We don't handle this case yet. To handle it correctly we have
1577 to add the delta, round and subtract the delta.
1578 Currently no machine description requires this support. */
1579 gcc_assert (!(stack_pointer_delta & (preferred_stack_boundary - 1)));
1580 args_size->var = round_up (args_size->var, preferred_stack_boundary);
1581 }
1582
1583 if (reg_parm_stack_space > 0)
1584 {
1585 args_size->var
1586 = size_binop (MAX_EXPR, args_size->var,
1587 ssize_int (reg_parm_stack_space));
1588
1589 /* The area corresponding to register parameters is not to count in
1590 the size of the block we need. So make the adjustment. */
1591 if (! OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl))))
1592 args_size->var
1593 = size_binop (MINUS_EXPR, args_size->var,
1594 ssize_int (reg_parm_stack_space));
1595 }
1596 }
1597 else
1598 {
1599 preferred_stack_boundary /= BITS_PER_UNIT;
1600 if (preferred_stack_boundary < 1)
1601 preferred_stack_boundary = 1;
1602 args_size->constant = (((args_size->constant
1603 + stack_pointer_delta
1604 + preferred_stack_boundary - 1)
1605 / preferred_stack_boundary
1606 * preferred_stack_boundary)
1607 - stack_pointer_delta);
1608
1609 args_size->constant = MAX (args_size->constant,
1610 reg_parm_stack_space);
1611
1612 if (! OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl))))
1613 args_size->constant -= reg_parm_stack_space;
1614 }
1615 return unadjusted_args_size;
1616 }
1617
1618 /* Precompute parameters as needed for a function call.
1619
1620 FLAGS is mask of ECF_* constants.
1621
1622 NUM_ACTUALS is the number of arguments.
1623
1624 ARGS is an array containing information for each argument; this
1625 routine fills in the INITIAL_VALUE and VALUE fields for each
1626 precomputed argument. */
1627
1628 static void
1629 precompute_arguments (int num_actuals, struct arg_data *args)
1630 {
1631 int i;
1632
1633 /* If this is a libcall, then precompute all arguments so that we do not
1634 get extraneous instructions emitted as part of the libcall sequence. */
1635
1636 /* If we preallocated the stack space, and some arguments must be passed
1637 on the stack, then we must precompute any parameter which contains a
1638 function call which will store arguments on the stack.
1639 Otherwise, evaluating the parameter may clobber previous parameters
1640 which have already been stored into the stack. (we have code to avoid
1641 such case by saving the outgoing stack arguments, but it results in
1642 worse code) */
1643 if (!ACCUMULATE_OUTGOING_ARGS)
1644 return;
1645
1646 for (i = 0; i < num_actuals; i++)
1647 {
1648 tree type;
1649 machine_mode mode;
1650
1651 if (TREE_CODE (args[i].tree_value) != CALL_EXPR)
1652 continue;
1653
1654 /* If this is an addressable type, we cannot pre-evaluate it. */
1655 type = TREE_TYPE (args[i].tree_value);
1656 gcc_assert (!TREE_ADDRESSABLE (type));
1657
1658 args[i].initial_value = args[i].value
1659 = expand_normal (args[i].tree_value);
1660
1661 mode = TYPE_MODE (type);
1662 if (mode != args[i].mode)
1663 {
1664 int unsignedp = args[i].unsignedp;
1665 args[i].value
1666 = convert_modes (args[i].mode, mode,
1667 args[i].value, args[i].unsignedp);
1668
1669 /* CSE will replace this only if it contains args[i].value
1670 pseudo, so convert it down to the declared mode using
1671 a SUBREG. */
1672 if (REG_P (args[i].value)
1673 && GET_MODE_CLASS (args[i].mode) == MODE_INT
1674 && promote_mode (type, mode, &unsignedp) != args[i].mode)
1675 {
1676 args[i].initial_value
1677 = gen_lowpart_SUBREG (mode, args[i].value);
1678 SUBREG_PROMOTED_VAR_P (args[i].initial_value) = 1;
1679 SUBREG_PROMOTED_SET (args[i].initial_value, args[i].unsignedp);
1680 }
1681 }
1682 }
1683 }
1684
1685 /* Given the current state of MUST_PREALLOCATE and information about
1686 arguments to a function call in NUM_ACTUALS, ARGS and ARGS_SIZE,
1687 compute and return the final value for MUST_PREALLOCATE. */
1688
1689 static int
1690 finalize_must_preallocate (int must_preallocate, int num_actuals,
1691 struct arg_data *args, struct args_size *args_size)
1692 {
1693 /* See if we have or want to preallocate stack space.
1694
1695 If we would have to push a partially-in-regs parm
1696 before other stack parms, preallocate stack space instead.
1697
1698 If the size of some parm is not a multiple of the required stack
1699 alignment, we must preallocate.
1700
1701 If the total size of arguments that would otherwise create a copy in
1702 a temporary (such as a CALL) is more than half the total argument list
1703 size, preallocation is faster.
1704
1705 Another reason to preallocate is if we have a machine (like the m88k)
1706 where stack alignment is required to be maintained between every
1707 pair of insns, not just when the call is made. However, we assume here
1708 that such machines either do not have push insns (and hence preallocation
1709 would occur anyway) or the problem is taken care of with
1710 PUSH_ROUNDING. */
1711
1712 if (! must_preallocate)
1713 {
1714 int partial_seen = 0;
1715 int copy_to_evaluate_size = 0;
1716 int i;
1717
1718 for (i = 0; i < num_actuals && ! must_preallocate; i++)
1719 {
1720 if (args[i].partial > 0 && ! args[i].pass_on_stack)
1721 partial_seen = 1;
1722 else if (partial_seen && args[i].reg == 0)
1723 must_preallocate = 1;
1724 /* We preallocate in case there are bounds passed
1725 in the bounds table to have precomputed address
1726 for bounds association. */
1727 else if (POINTER_BOUNDS_P (args[i].tree_value)
1728 && !args[i].reg)
1729 must_preallocate = 1;
1730
1731 if (TYPE_MODE (TREE_TYPE (args[i].tree_value)) == BLKmode
1732 && (TREE_CODE (args[i].tree_value) == CALL_EXPR
1733 || TREE_CODE (args[i].tree_value) == TARGET_EXPR
1734 || TREE_CODE (args[i].tree_value) == COND_EXPR
1735 || TREE_ADDRESSABLE (TREE_TYPE (args[i].tree_value))))
1736 copy_to_evaluate_size
1737 += int_size_in_bytes (TREE_TYPE (args[i].tree_value));
1738 }
1739
1740 if (copy_to_evaluate_size * 2 >= args_size->constant
1741 && args_size->constant > 0)
1742 must_preallocate = 1;
1743 }
1744 return must_preallocate;
1745 }
1746
1747 /* If we preallocated stack space, compute the address of each argument
1748 and store it into the ARGS array.
1749
1750 We need not ensure it is a valid memory address here; it will be
1751 validized when it is used.
1752
1753 ARGBLOCK is an rtx for the address of the outgoing arguments. */
1754
1755 static void
1756 compute_argument_addresses (struct arg_data *args, rtx argblock, int num_actuals)
1757 {
1758 if (argblock)
1759 {
1760 rtx arg_reg = argblock;
1761 int i, arg_offset = 0;
1762
1763 if (GET_CODE (argblock) == PLUS)
1764 arg_reg = XEXP (argblock, 0), arg_offset = INTVAL (XEXP (argblock, 1));
1765
1766 for (i = 0; i < num_actuals; i++)
1767 {
1768 rtx offset = ARGS_SIZE_RTX (args[i].locate.offset);
1769 rtx slot_offset = ARGS_SIZE_RTX (args[i].locate.slot_offset);
1770 rtx addr;
1771 unsigned int align, boundary;
1772 unsigned int units_on_stack = 0;
1773 machine_mode partial_mode = VOIDmode;
1774
1775 /* Skip this parm if it will not be passed on the stack. */
1776 if (! args[i].pass_on_stack
1777 && args[i].reg != 0
1778 && args[i].partial == 0)
1779 continue;
1780
1781 /* Pointer Bounds are never passed on the stack. */
1782 if (POINTER_BOUNDS_P (args[i].tree_value))
1783 continue;
1784
1785 if (CONST_INT_P (offset))
1786 addr = plus_constant (Pmode, arg_reg, INTVAL (offset));
1787 else
1788 addr = gen_rtx_PLUS (Pmode, arg_reg, offset);
1789
1790 addr = plus_constant (Pmode, addr, arg_offset);
1791
1792 if (args[i].partial != 0)
1793 {
1794 /* Only part of the parameter is being passed on the stack.
1795 Generate a simple memory reference of the correct size. */
1796 units_on_stack = args[i].locate.size.constant;
1797 partial_mode = mode_for_size (units_on_stack * BITS_PER_UNIT,
1798 MODE_INT, 1);
1799 args[i].stack = gen_rtx_MEM (partial_mode, addr);
1800 set_mem_size (args[i].stack, units_on_stack);
1801 }
1802 else
1803 {
1804 args[i].stack = gen_rtx_MEM (args[i].mode, addr);
1805 set_mem_attributes (args[i].stack,
1806 TREE_TYPE (args[i].tree_value), 1);
1807 }
1808 align = BITS_PER_UNIT;
1809 boundary = args[i].locate.boundary;
1810 if (args[i].locate.where_pad != downward)
1811 align = boundary;
1812 else if (CONST_INT_P (offset))
1813 {
1814 align = INTVAL (offset) * BITS_PER_UNIT | boundary;
1815 align = align & -align;
1816 }
1817 set_mem_align (args[i].stack, align);
1818
1819 if (CONST_INT_P (slot_offset))
1820 addr = plus_constant (Pmode, arg_reg, INTVAL (slot_offset));
1821 else
1822 addr = gen_rtx_PLUS (Pmode, arg_reg, slot_offset);
1823
1824 addr = plus_constant (Pmode, addr, arg_offset);
1825
1826 if (args[i].partial != 0)
1827 {
1828 /* Only part of the parameter is being passed on the stack.
1829 Generate a simple memory reference of the correct size.
1830 */
1831 args[i].stack_slot = gen_rtx_MEM (partial_mode, addr);
1832 set_mem_size (args[i].stack_slot, units_on_stack);
1833 }
1834 else
1835 {
1836 args[i].stack_slot = gen_rtx_MEM (args[i].mode, addr);
1837 set_mem_attributes (args[i].stack_slot,
1838 TREE_TYPE (args[i].tree_value), 1);
1839 }
1840 set_mem_align (args[i].stack_slot, args[i].locate.boundary);
1841
1842 /* Function incoming arguments may overlap with sibling call
1843 outgoing arguments and we cannot allow reordering of reads
1844 from function arguments with stores to outgoing arguments
1845 of sibling calls. */
1846 set_mem_alias_set (args[i].stack, 0);
1847 set_mem_alias_set (args[i].stack_slot, 0);
1848 }
1849 }
1850 }
1851
1852 /* Given a FNDECL and EXP, return an rtx suitable for use as a target address
1853 in a call instruction.
1854
1855 FNDECL is the tree node for the target function. For an indirect call
1856 FNDECL will be NULL_TREE.
1857
1858 ADDR is the operand 0 of CALL_EXPR for this call. */
1859
1860 static rtx
1861 rtx_for_function_call (tree fndecl, tree addr)
1862 {
1863 rtx funexp;
1864
1865 /* Get the function to call, in the form of RTL. */
1866 if (fndecl)
1867 {
1868 if (!TREE_USED (fndecl) && fndecl != current_function_decl)
1869 TREE_USED (fndecl) = 1;
1870
1871 /* Get a SYMBOL_REF rtx for the function address. */
1872 funexp = XEXP (DECL_RTL (fndecl), 0);
1873 }
1874 else
1875 /* Generate an rtx (probably a pseudo-register) for the address. */
1876 {
1877 push_temp_slots ();
1878 funexp = expand_normal (addr);
1879 pop_temp_slots (); /* FUNEXP can't be BLKmode. */
1880 }
1881 return funexp;
1882 }
1883
1884 /* Internal state for internal_arg_pointer_based_exp and its helpers. */
1885 static struct
1886 {
1887 /* Last insn that has been scanned by internal_arg_pointer_based_exp_scan,
1888 or NULL_RTX if none has been scanned yet. */
1889 rtx_insn *scan_start;
1890 /* Vector indexed by REGNO - FIRST_PSEUDO_REGISTER, recording if a pseudo is
1891 based on crtl->args.internal_arg_pointer. The element is NULL_RTX if the
1892 pseudo isn't based on it, a CONST_INT offset if the pseudo is based on it
1893 with fixed offset, or PC if this is with variable or unknown offset. */
1894 vec<rtx> cache;
1895 } internal_arg_pointer_exp_state;
1896
1897 static rtx internal_arg_pointer_based_exp (const_rtx, bool);
1898
1899 /* Helper function for internal_arg_pointer_based_exp. Scan insns in
1900 the tail call sequence, starting with first insn that hasn't been
1901 scanned yet, and note for each pseudo on the LHS whether it is based
1902 on crtl->args.internal_arg_pointer or not, and what offset from that
1903 that pointer it has. */
1904
1905 static void
1906 internal_arg_pointer_based_exp_scan (void)
1907 {
1908 rtx_insn *insn, *scan_start = internal_arg_pointer_exp_state.scan_start;
1909
1910 if (scan_start == NULL_RTX)
1911 insn = get_insns ();
1912 else
1913 insn = NEXT_INSN (scan_start);
1914
1915 while (insn)
1916 {
1917 rtx set = single_set (insn);
1918 if (set && REG_P (SET_DEST (set)) && !HARD_REGISTER_P (SET_DEST (set)))
1919 {
1920 rtx val = NULL_RTX;
1921 unsigned int idx = REGNO (SET_DEST (set)) - FIRST_PSEUDO_REGISTER;
1922 /* Punt on pseudos set multiple times. */
1923 if (idx < internal_arg_pointer_exp_state.cache.length ()
1924 && (internal_arg_pointer_exp_state.cache[idx]
1925 != NULL_RTX))
1926 val = pc_rtx;
1927 else
1928 val = internal_arg_pointer_based_exp (SET_SRC (set), false);
1929 if (val != NULL_RTX)
1930 {
1931 if (idx >= internal_arg_pointer_exp_state.cache.length ())
1932 internal_arg_pointer_exp_state.cache
1933 .safe_grow_cleared (idx + 1);
1934 internal_arg_pointer_exp_state.cache[idx] = val;
1935 }
1936 }
1937 if (NEXT_INSN (insn) == NULL_RTX)
1938 scan_start = insn;
1939 insn = NEXT_INSN (insn);
1940 }
1941
1942 internal_arg_pointer_exp_state.scan_start = scan_start;
1943 }
1944
1945 /* Compute whether RTL is based on crtl->args.internal_arg_pointer. Return
1946 NULL_RTX if RTL isn't based on it, a CONST_INT offset if RTL is based on
1947 it with fixed offset, or PC if this is with variable or unknown offset.
1948 TOPLEVEL is true if the function is invoked at the topmost level. */
1949
1950 static rtx
1951 internal_arg_pointer_based_exp (const_rtx rtl, bool toplevel)
1952 {
1953 if (CONSTANT_P (rtl))
1954 return NULL_RTX;
1955
1956 if (rtl == crtl->args.internal_arg_pointer)
1957 return const0_rtx;
1958
1959 if (REG_P (rtl) && HARD_REGISTER_P (rtl))
1960 return NULL_RTX;
1961
1962 if (GET_CODE (rtl) == PLUS && CONST_INT_P (XEXP (rtl, 1)))
1963 {
1964 rtx val = internal_arg_pointer_based_exp (XEXP (rtl, 0), toplevel);
1965 if (val == NULL_RTX || val == pc_rtx)
1966 return val;
1967 return plus_constant (Pmode, val, INTVAL (XEXP (rtl, 1)));
1968 }
1969
1970 /* When called at the topmost level, scan pseudo assignments in between the
1971 last scanned instruction in the tail call sequence and the latest insn
1972 in that sequence. */
1973 if (toplevel)
1974 internal_arg_pointer_based_exp_scan ();
1975
1976 if (REG_P (rtl))
1977 {
1978 unsigned int idx = REGNO (rtl) - FIRST_PSEUDO_REGISTER;
1979 if (idx < internal_arg_pointer_exp_state.cache.length ())
1980 return internal_arg_pointer_exp_state.cache[idx];
1981
1982 return NULL_RTX;
1983 }
1984
1985 subrtx_iterator::array_type array;
1986 FOR_EACH_SUBRTX (iter, array, rtl, NONCONST)
1987 {
1988 const_rtx x = *iter;
1989 if (REG_P (x) && internal_arg_pointer_based_exp (x, false) != NULL_RTX)
1990 return pc_rtx;
1991 if (MEM_P (x))
1992 iter.skip_subrtxes ();
1993 }
1994
1995 return NULL_RTX;
1996 }
1997
1998 /* Return true if and only if SIZE storage units (usually bytes)
1999 starting from address ADDR overlap with already clobbered argument
2000 area. This function is used to determine if we should give up a
2001 sibcall. */
2002
2003 static bool
2004 mem_overlaps_already_clobbered_arg_p (rtx addr, unsigned HOST_WIDE_INT size)
2005 {
2006 HOST_WIDE_INT i;
2007 rtx val;
2008
2009 if (bitmap_empty_p (stored_args_map))
2010 return false;
2011 val = internal_arg_pointer_based_exp (addr, true);
2012 if (val == NULL_RTX)
2013 return false;
2014 else if (val == pc_rtx)
2015 return true;
2016 else
2017 i = INTVAL (val);
2018
2019 if (STACK_GROWS_DOWNWARD)
2020 i -= crtl->args.pretend_args_size;
2021 else
2022 i += crtl->args.pretend_args_size;
2023
2024
2025 if (ARGS_GROW_DOWNWARD)
2026 i = -i - size;
2027
2028 if (size > 0)
2029 {
2030 unsigned HOST_WIDE_INT k;
2031
2032 for (k = 0; k < size; k++)
2033 if (i + k < SBITMAP_SIZE (stored_args_map)
2034 && bitmap_bit_p (stored_args_map, i + k))
2035 return true;
2036 }
2037
2038 return false;
2039 }
2040
2041 /* Do the register loads required for any wholly-register parms or any
2042 parms which are passed both on the stack and in a register. Their
2043 expressions were already evaluated.
2044
2045 Mark all register-parms as living through the call, putting these USE
2046 insns in the CALL_INSN_FUNCTION_USAGE field.
2047
2048 When IS_SIBCALL, perform the check_sibcall_argument_overlap
2049 checking, setting *SIBCALL_FAILURE if appropriate. */
2050
2051 static void
2052 load_register_parameters (struct arg_data *args, int num_actuals,
2053 rtx *call_fusage, int flags, int is_sibcall,
2054 int *sibcall_failure)
2055 {
2056 int i, j;
2057
2058 for (i = 0; i < num_actuals; i++)
2059 {
2060 rtx reg = ((flags & ECF_SIBCALL)
2061 ? args[i].tail_call_reg : args[i].reg);
2062 if (reg)
2063 {
2064 int partial = args[i].partial;
2065 int nregs;
2066 int size = 0;
2067 rtx_insn *before_arg = get_last_insn ();
2068 /* Set non-negative if we must move a word at a time, even if
2069 just one word (e.g, partial == 4 && mode == DFmode). Set
2070 to -1 if we just use a normal move insn. This value can be
2071 zero if the argument is a zero size structure. */
2072 nregs = -1;
2073 if (GET_CODE (reg) == PARALLEL)
2074 ;
2075 else if (partial)
2076 {
2077 gcc_assert (partial % UNITS_PER_WORD == 0);
2078 nregs = partial / UNITS_PER_WORD;
2079 }
2080 else if (TYPE_MODE (TREE_TYPE (args[i].tree_value)) == BLKmode)
2081 {
2082 size = int_size_in_bytes (TREE_TYPE (args[i].tree_value));
2083 nregs = (size + (UNITS_PER_WORD - 1)) / UNITS_PER_WORD;
2084 }
2085 else
2086 size = GET_MODE_SIZE (args[i].mode);
2087
2088 /* Handle calls that pass values in multiple non-contiguous
2089 locations. The Irix 6 ABI has examples of this. */
2090
2091 if (GET_CODE (reg) == PARALLEL)
2092 emit_group_move (reg, args[i].parallel_value);
2093
2094 /* If simple case, just do move. If normal partial, store_one_arg
2095 has already loaded the register for us. In all other cases,
2096 load the register(s) from memory. */
2097
2098 else if (nregs == -1)
2099 {
2100 emit_move_insn (reg, args[i].value);
2101 #ifdef BLOCK_REG_PADDING
2102 /* Handle case where we have a value that needs shifting
2103 up to the msb. eg. a QImode value and we're padding
2104 upward on a BYTES_BIG_ENDIAN machine. */
2105 if (size < UNITS_PER_WORD
2106 && (args[i].locate.where_pad
2107 == (BYTES_BIG_ENDIAN ? upward : downward)))
2108 {
2109 rtx x;
2110 int shift = (UNITS_PER_WORD - size) * BITS_PER_UNIT;
2111
2112 /* Assigning REG here rather than a temp makes CALL_FUSAGE
2113 report the whole reg as used. Strictly speaking, the
2114 call only uses SIZE bytes at the msb end, but it doesn't
2115 seem worth generating rtl to say that. */
2116 reg = gen_rtx_REG (word_mode, REGNO (reg));
2117 x = expand_shift (LSHIFT_EXPR, word_mode, reg, shift, reg, 1);
2118 if (x != reg)
2119 emit_move_insn (reg, x);
2120 }
2121 #endif
2122 }
2123
2124 /* If we have pre-computed the values to put in the registers in
2125 the case of non-aligned structures, copy them in now. */
2126
2127 else if (args[i].n_aligned_regs != 0)
2128 for (j = 0; j < args[i].n_aligned_regs; j++)
2129 emit_move_insn (gen_rtx_REG (word_mode, REGNO (reg) + j),
2130 args[i].aligned_regs[j]);
2131
2132 else if (partial == 0 || args[i].pass_on_stack)
2133 {
2134 rtx mem = validize_mem (copy_rtx (args[i].value));
2135
2136 /* Check for overlap with already clobbered argument area,
2137 providing that this has non-zero size. */
2138 if (is_sibcall
2139 && (size == 0
2140 || mem_overlaps_already_clobbered_arg_p
2141 (XEXP (args[i].value, 0), size)))
2142 *sibcall_failure = 1;
2143
2144 if (size % UNITS_PER_WORD == 0
2145 || MEM_ALIGN (mem) % BITS_PER_WORD == 0)
2146 move_block_to_reg (REGNO (reg), mem, nregs, args[i].mode);
2147 else
2148 {
2149 if (nregs > 1)
2150 move_block_to_reg (REGNO (reg), mem, nregs - 1,
2151 args[i].mode);
2152 rtx dest = gen_rtx_REG (word_mode, REGNO (reg) + nregs - 1);
2153 unsigned int bitoff = (nregs - 1) * BITS_PER_WORD;
2154 unsigned int bitsize = size * BITS_PER_UNIT - bitoff;
2155 rtx x = extract_bit_field (mem, bitsize, bitoff, 1,
2156 dest, word_mode, word_mode);
2157 if (BYTES_BIG_ENDIAN)
2158 x = expand_shift (LSHIFT_EXPR, word_mode, x,
2159 BITS_PER_WORD - bitsize, dest, 1);
2160 if (x != dest)
2161 emit_move_insn (dest, x);
2162 }
2163
2164 /* Handle a BLKmode that needs shifting. */
2165 if (nregs == 1 && size < UNITS_PER_WORD
2166 #ifdef BLOCK_REG_PADDING
2167 && args[i].locate.where_pad == downward
2168 #else
2169 && BYTES_BIG_ENDIAN
2170 #endif
2171 )
2172 {
2173 rtx dest = gen_rtx_REG (word_mode, REGNO (reg));
2174 int shift = (UNITS_PER_WORD - size) * BITS_PER_UNIT;
2175 enum tree_code dir = (BYTES_BIG_ENDIAN
2176 ? RSHIFT_EXPR : LSHIFT_EXPR);
2177 rtx x;
2178
2179 x = expand_shift (dir, word_mode, dest, shift, dest, 1);
2180 if (x != dest)
2181 emit_move_insn (dest, x);
2182 }
2183 }
2184
2185 /* When a parameter is a block, and perhaps in other cases, it is
2186 possible that it did a load from an argument slot that was
2187 already clobbered. */
2188 if (is_sibcall
2189 && check_sibcall_argument_overlap (before_arg, &args[i], 0))
2190 *sibcall_failure = 1;
2191
2192 /* Handle calls that pass values in multiple non-contiguous
2193 locations. The Irix 6 ABI has examples of this. */
2194 if (GET_CODE (reg) == PARALLEL)
2195 use_group_regs (call_fusage, reg);
2196 else if (nregs == -1)
2197 use_reg_mode (call_fusage, reg,
2198 TYPE_MODE (TREE_TYPE (args[i].tree_value)));
2199 else if (nregs > 0)
2200 use_regs (call_fusage, REGNO (reg), nregs);
2201 }
2202 }
2203 }
2204
2205 /* We need to pop PENDING_STACK_ADJUST bytes. But, if the arguments
2206 wouldn't fill up an even multiple of PREFERRED_UNIT_STACK_BOUNDARY
2207 bytes, then we would need to push some additional bytes to pad the
2208 arguments. So, we compute an adjust to the stack pointer for an
2209 amount that will leave the stack under-aligned by UNADJUSTED_ARGS_SIZE
2210 bytes. Then, when the arguments are pushed the stack will be perfectly
2211 aligned. ARGS_SIZE->CONSTANT is set to the number of bytes that should
2212 be popped after the call. Returns the adjustment. */
2213
2214 static int
2215 combine_pending_stack_adjustment_and_call (int unadjusted_args_size,
2216 struct args_size *args_size,
2217 unsigned int preferred_unit_stack_boundary)
2218 {
2219 /* The number of bytes to pop so that the stack will be
2220 under-aligned by UNADJUSTED_ARGS_SIZE bytes. */
2221 HOST_WIDE_INT adjustment;
2222 /* The alignment of the stack after the arguments are pushed, if we
2223 just pushed the arguments without adjust the stack here. */
2224 unsigned HOST_WIDE_INT unadjusted_alignment;
2225
2226 unadjusted_alignment
2227 = ((stack_pointer_delta + unadjusted_args_size)
2228 % preferred_unit_stack_boundary);
2229
2230 /* We want to get rid of as many of the PENDING_STACK_ADJUST bytes
2231 as possible -- leaving just enough left to cancel out the
2232 UNADJUSTED_ALIGNMENT. In other words, we want to ensure that the
2233 PENDING_STACK_ADJUST is non-negative, and congruent to
2234 -UNADJUSTED_ALIGNMENT modulo the PREFERRED_UNIT_STACK_BOUNDARY. */
2235
2236 /* Begin by trying to pop all the bytes. */
2237 unadjusted_alignment
2238 = (unadjusted_alignment
2239 - (pending_stack_adjust % preferred_unit_stack_boundary));
2240 adjustment = pending_stack_adjust;
2241 /* Push enough additional bytes that the stack will be aligned
2242 after the arguments are pushed. */
2243 if (preferred_unit_stack_boundary > 1)
2244 {
2245 if (unadjusted_alignment > 0)
2246 adjustment -= preferred_unit_stack_boundary - unadjusted_alignment;
2247 else
2248 adjustment += unadjusted_alignment;
2249 }
2250
2251 /* Now, sets ARGS_SIZE->CONSTANT so that we pop the right number of
2252 bytes after the call. The right number is the entire
2253 PENDING_STACK_ADJUST less our ADJUSTMENT plus the amount required
2254 by the arguments in the first place. */
2255 args_size->constant
2256 = pending_stack_adjust - adjustment + unadjusted_args_size;
2257
2258 return adjustment;
2259 }
2260
2261 /* Scan X expression if it does not dereference any argument slots
2262 we already clobbered by tail call arguments (as noted in stored_args_map
2263 bitmap).
2264 Return nonzero if X expression dereferences such argument slots,
2265 zero otherwise. */
2266
2267 static int
2268 check_sibcall_argument_overlap_1 (rtx x)
2269 {
2270 RTX_CODE code;
2271 int i, j;
2272 const char *fmt;
2273
2274 if (x == NULL_RTX)
2275 return 0;
2276
2277 code = GET_CODE (x);
2278
2279 /* We need not check the operands of the CALL expression itself. */
2280 if (code == CALL)
2281 return 0;
2282
2283 if (code == MEM)
2284 return mem_overlaps_already_clobbered_arg_p (XEXP (x, 0),
2285 GET_MODE_SIZE (GET_MODE (x)));
2286
2287 /* Scan all subexpressions. */
2288 fmt = GET_RTX_FORMAT (code);
2289 for (i = 0; i < GET_RTX_LENGTH (code); i++, fmt++)
2290 {
2291 if (*fmt == 'e')
2292 {
2293 if (check_sibcall_argument_overlap_1 (XEXP (x, i)))
2294 return 1;
2295 }
2296 else if (*fmt == 'E')
2297 {
2298 for (j = 0; j < XVECLEN (x, i); j++)
2299 if (check_sibcall_argument_overlap_1 (XVECEXP (x, i, j)))
2300 return 1;
2301 }
2302 }
2303 return 0;
2304 }
2305
2306 /* Scan sequence after INSN if it does not dereference any argument slots
2307 we already clobbered by tail call arguments (as noted in stored_args_map
2308 bitmap). If MARK_STORED_ARGS_MAP, add stack slots for ARG to
2309 stored_args_map bitmap afterwards (when ARG is a register MARK_STORED_ARGS_MAP
2310 should be 0). Return nonzero if sequence after INSN dereferences such argument
2311 slots, zero otherwise. */
2312
2313 static int
2314 check_sibcall_argument_overlap (rtx_insn *insn, struct arg_data *arg,
2315 int mark_stored_args_map)
2316 {
2317 int low, high;
2318
2319 if (insn == NULL_RTX)
2320 insn = get_insns ();
2321 else
2322 insn = NEXT_INSN (insn);
2323
2324 for (; insn; insn = NEXT_INSN (insn))
2325 if (INSN_P (insn)
2326 && check_sibcall_argument_overlap_1 (PATTERN (insn)))
2327 break;
2328
2329 if (mark_stored_args_map)
2330 {
2331 if (ARGS_GROW_DOWNWARD)
2332 low = -arg->locate.slot_offset.constant - arg->locate.size.constant;
2333 else
2334 low = arg->locate.slot_offset.constant;
2335
2336 for (high = low + arg->locate.size.constant; low < high; low++)
2337 bitmap_set_bit (stored_args_map, low);
2338 }
2339 return insn != NULL_RTX;
2340 }
2341
2342 /* Given that a function returns a value of mode MODE at the most
2343 significant end of hard register VALUE, shift VALUE left or right
2344 as specified by LEFT_P. Return true if some action was needed. */
2345
2346 bool
2347 shift_return_value (machine_mode mode, bool left_p, rtx value)
2348 {
2349 HOST_WIDE_INT shift;
2350
2351 gcc_assert (REG_P (value) && HARD_REGISTER_P (value));
2352 shift = GET_MODE_BITSIZE (GET_MODE (value)) - GET_MODE_BITSIZE (mode);
2353 if (shift == 0)
2354 return false;
2355
2356 /* Use ashr rather than lshr for right shifts. This is for the benefit
2357 of the MIPS port, which requires SImode values to be sign-extended
2358 when stored in 64-bit registers. */
2359 if (!force_expand_binop (GET_MODE (value), left_p ? ashl_optab : ashr_optab,
2360 value, GEN_INT (shift), value, 1, OPTAB_WIDEN))
2361 gcc_unreachable ();
2362 return true;
2363 }
2364
2365 /* If X is a likely-spilled register value, copy it to a pseudo
2366 register and return that register. Return X otherwise. */
2367
2368 static rtx
2369 avoid_likely_spilled_reg (rtx x)
2370 {
2371 rtx new_rtx;
2372
2373 if (REG_P (x)
2374 && HARD_REGISTER_P (x)
2375 && targetm.class_likely_spilled_p (REGNO_REG_CLASS (REGNO (x))))
2376 {
2377 /* Make sure that we generate a REG rather than a CONCAT.
2378 Moves into CONCATs can need nontrivial instructions,
2379 and the whole point of this function is to avoid
2380 using the hard register directly in such a situation. */
2381 generating_concat_p = 0;
2382 new_rtx = gen_reg_rtx (GET_MODE (x));
2383 generating_concat_p = 1;
2384 emit_move_insn (new_rtx, x);
2385 return new_rtx;
2386 }
2387 return x;
2388 }
2389
2390 /* Generate all the code for a CALL_EXPR exp
2391 and return an rtx for its value.
2392 Store the value in TARGET (specified as an rtx) if convenient.
2393 If the value is stored in TARGET then TARGET is returned.
2394 If IGNORE is nonzero, then we ignore the value of the function call. */
2395
2396 rtx
2397 expand_call (tree exp, rtx target, int ignore)
2398 {
2399 /* Nonzero if we are currently expanding a call. */
2400 static int currently_expanding_call = 0;
2401
2402 /* RTX for the function to be called. */
2403 rtx funexp;
2404 /* Sequence of insns to perform a normal "call". */
2405 rtx_insn *normal_call_insns = NULL;
2406 /* Sequence of insns to perform a tail "call". */
2407 rtx_insn *tail_call_insns = NULL;
2408 /* Data type of the function. */
2409 tree funtype;
2410 tree type_arg_types;
2411 tree rettype;
2412 /* Declaration of the function being called,
2413 or 0 if the function is computed (not known by name). */
2414 tree fndecl = 0;
2415 /* The type of the function being called. */
2416 tree fntype;
2417 bool try_tail_call = CALL_EXPR_TAILCALL (exp);
2418 int pass;
2419
2420 /* Register in which non-BLKmode value will be returned,
2421 or 0 if no value or if value is BLKmode. */
2422 rtx valreg;
2423 /* Register(s) in which bounds are returned. */
2424 rtx valbnd = NULL;
2425 /* Address where we should return a BLKmode value;
2426 0 if value not BLKmode. */
2427 rtx structure_value_addr = 0;
2428 /* Nonzero if that address is being passed by treating it as
2429 an extra, implicit first parameter. Otherwise,
2430 it is passed by being copied directly into struct_value_rtx. */
2431 int structure_value_addr_parm = 0;
2432 /* Holds the value of implicit argument for the struct value. */
2433 tree structure_value_addr_value = NULL_TREE;
2434 /* Size of aggregate value wanted, or zero if none wanted
2435 or if we are using the non-reentrant PCC calling convention
2436 or expecting the value in registers. */
2437 HOST_WIDE_INT struct_value_size = 0;
2438 /* Nonzero if called function returns an aggregate in memory PCC style,
2439 by returning the address of where to find it. */
2440 int pcc_struct_value = 0;
2441 rtx struct_value = 0;
2442
2443 /* Number of actual parameters in this call, including struct value addr. */
2444 int num_actuals;
2445 /* Number of named args. Args after this are anonymous ones
2446 and they must all go on the stack. */
2447 int n_named_args;
2448 /* Number of complex actual arguments that need to be split. */
2449 int num_complex_actuals = 0;
2450
2451 /* Vector of information about each argument.
2452 Arguments are numbered in the order they will be pushed,
2453 not the order they are written. */
2454 struct arg_data *args;
2455
2456 /* Total size in bytes of all the stack-parms scanned so far. */
2457 struct args_size args_size;
2458 struct args_size adjusted_args_size;
2459 /* Size of arguments before any adjustments (such as rounding). */
2460 int unadjusted_args_size;
2461 /* Data on reg parms scanned so far. */
2462 CUMULATIVE_ARGS args_so_far_v;
2463 cumulative_args_t args_so_far;
2464 /* Nonzero if a reg parm has been scanned. */
2465 int reg_parm_seen;
2466 /* Nonzero if this is an indirect function call. */
2467
2468 /* Nonzero if we must avoid push-insns in the args for this call.
2469 If stack space is allocated for register parameters, but not by the
2470 caller, then it is preallocated in the fixed part of the stack frame.
2471 So the entire argument block must then be preallocated (i.e., we
2472 ignore PUSH_ROUNDING in that case). */
2473
2474 int must_preallocate = !PUSH_ARGS;
2475
2476 /* Size of the stack reserved for parameter registers. */
2477 int reg_parm_stack_space = 0;
2478
2479 /* Address of space preallocated for stack parms
2480 (on machines that lack push insns), or 0 if space not preallocated. */
2481 rtx argblock = 0;
2482
2483 /* Mask of ECF_ and ERF_ flags. */
2484 int flags = 0;
2485 int return_flags = 0;
2486 #ifdef REG_PARM_STACK_SPACE
2487 /* Define the boundary of the register parm stack space that needs to be
2488 saved, if any. */
2489 int low_to_save, high_to_save;
2490 rtx save_area = 0; /* Place that it is saved */
2491 #endif
2492
2493 int initial_highest_arg_in_use = highest_outgoing_arg_in_use;
2494 char *initial_stack_usage_map = stack_usage_map;
2495 char *stack_usage_map_buf = NULL;
2496
2497 int old_stack_allocated;
2498
2499 /* State variables to track stack modifications. */
2500 rtx old_stack_level = 0;
2501 int old_stack_arg_under_construction = 0;
2502 int old_pending_adj = 0;
2503 int old_inhibit_defer_pop = inhibit_defer_pop;
2504
2505 /* Some stack pointer alterations we make are performed via
2506 allocate_dynamic_stack_space. This modifies the stack_pointer_delta,
2507 which we then also need to save/restore along the way. */
2508 int old_stack_pointer_delta = 0;
2509
2510 rtx call_fusage;
2511 tree addr = CALL_EXPR_FN (exp);
2512 int i;
2513 /* The alignment of the stack, in bits. */
2514 unsigned HOST_WIDE_INT preferred_stack_boundary;
2515 /* The alignment of the stack, in bytes. */
2516 unsigned HOST_WIDE_INT preferred_unit_stack_boundary;
2517 /* The static chain value to use for this call. */
2518 rtx static_chain_value;
2519 /* See if this is "nothrow" function call. */
2520 if (TREE_NOTHROW (exp))
2521 flags |= ECF_NOTHROW;
2522
2523 /* See if we can find a DECL-node for the actual function, and get the
2524 function attributes (flags) from the function decl or type node. */
2525 fndecl = get_callee_fndecl (exp);
2526 if (fndecl)
2527 {
2528 fntype = TREE_TYPE (fndecl);
2529 flags |= flags_from_decl_or_type (fndecl);
2530 return_flags |= decl_return_flags (fndecl);
2531 }
2532 else
2533 {
2534 fntype = TREE_TYPE (TREE_TYPE (addr));
2535 flags |= flags_from_decl_or_type (fntype);
2536 }
2537 rettype = TREE_TYPE (exp);
2538
2539 struct_value = targetm.calls.struct_value_rtx (fntype, 0);
2540
2541 /* Warn if this value is an aggregate type,
2542 regardless of which calling convention we are using for it. */
2543 if (AGGREGATE_TYPE_P (rettype))
2544 warning (OPT_Waggregate_return, "function call has aggregate value");
2545
2546 /* If the result of a non looping pure or const function call is
2547 ignored (or void), and none of its arguments are volatile, we can
2548 avoid expanding the call and just evaluate the arguments for
2549 side-effects. */
2550 if ((flags & (ECF_CONST | ECF_PURE))
2551 && (!(flags & ECF_LOOPING_CONST_OR_PURE))
2552 && (ignore || target == const0_rtx
2553 || TYPE_MODE (rettype) == VOIDmode))
2554 {
2555 bool volatilep = false;
2556 tree arg;
2557 call_expr_arg_iterator iter;
2558
2559 FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
2560 if (TREE_THIS_VOLATILE (arg))
2561 {
2562 volatilep = true;
2563 break;
2564 }
2565
2566 if (! volatilep)
2567 {
2568 FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
2569 expand_expr (arg, const0_rtx, VOIDmode, EXPAND_NORMAL);
2570 return const0_rtx;
2571 }
2572 }
2573
2574 #ifdef REG_PARM_STACK_SPACE
2575 reg_parm_stack_space = REG_PARM_STACK_SPACE (!fndecl ? fntype : fndecl);
2576 #endif
2577
2578 if (! OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl)))
2579 && reg_parm_stack_space > 0 && PUSH_ARGS)
2580 must_preallocate = 1;
2581
2582 /* Set up a place to return a structure. */
2583
2584 /* Cater to broken compilers. */
2585 if (aggregate_value_p (exp, fntype))
2586 {
2587 /* This call returns a big structure. */
2588 flags &= ~(ECF_CONST | ECF_PURE | ECF_LOOPING_CONST_OR_PURE);
2589
2590 #ifdef PCC_STATIC_STRUCT_RETURN
2591 {
2592 pcc_struct_value = 1;
2593 }
2594 #else /* not PCC_STATIC_STRUCT_RETURN */
2595 {
2596 struct_value_size = int_size_in_bytes (rettype);
2597
2598 /* Even if it is semantically safe to use the target as the return
2599 slot, it may be not sufficiently aligned for the return type. */
2600 if (CALL_EXPR_RETURN_SLOT_OPT (exp)
2601 && target
2602 && MEM_P (target)
2603 && !(MEM_ALIGN (target) < TYPE_ALIGN (rettype)
2604 && SLOW_UNALIGNED_ACCESS (TYPE_MODE (rettype),
2605 MEM_ALIGN (target))))
2606 structure_value_addr = XEXP (target, 0);
2607 else
2608 {
2609 /* For variable-sized objects, we must be called with a target
2610 specified. If we were to allocate space on the stack here,
2611 we would have no way of knowing when to free it. */
2612 rtx d = assign_temp (rettype, 1, 1);
2613 structure_value_addr = XEXP (d, 0);
2614 target = 0;
2615 }
2616 }
2617 #endif /* not PCC_STATIC_STRUCT_RETURN */
2618 }
2619
2620 /* Figure out the amount to which the stack should be aligned. */
2621 preferred_stack_boundary = PREFERRED_STACK_BOUNDARY;
2622 if (fndecl)
2623 {
2624 struct cgraph_rtl_info *i = cgraph_node::rtl_info (fndecl);
2625 /* Without automatic stack alignment, we can't increase preferred
2626 stack boundary. With automatic stack alignment, it is
2627 unnecessary since unless we can guarantee that all callers will
2628 align the outgoing stack properly, callee has to align its
2629 stack anyway. */
2630 if (i
2631 && i->preferred_incoming_stack_boundary
2632 && i->preferred_incoming_stack_boundary < preferred_stack_boundary)
2633 preferred_stack_boundary = i->preferred_incoming_stack_boundary;
2634 }
2635
2636 /* Operand 0 is a pointer-to-function; get the type of the function. */
2637 funtype = TREE_TYPE (addr);
2638 gcc_assert (POINTER_TYPE_P (funtype));
2639 funtype = TREE_TYPE (funtype);
2640
2641 /* Count whether there are actual complex arguments that need to be split
2642 into their real and imaginary parts. Munge the type_arg_types
2643 appropriately here as well. */
2644 if (targetm.calls.split_complex_arg)
2645 {
2646 call_expr_arg_iterator iter;
2647 tree arg;
2648 FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
2649 {
2650 tree type = TREE_TYPE (arg);
2651 if (type && TREE_CODE (type) == COMPLEX_TYPE
2652 && targetm.calls.split_complex_arg (type))
2653 num_complex_actuals++;
2654 }
2655 type_arg_types = split_complex_types (TYPE_ARG_TYPES (funtype));
2656 }
2657 else
2658 type_arg_types = TYPE_ARG_TYPES (funtype);
2659
2660 if (flags & ECF_MAY_BE_ALLOCA)
2661 cfun->calls_alloca = 1;
2662
2663 /* If struct_value_rtx is 0, it means pass the address
2664 as if it were an extra parameter. Put the argument expression
2665 in structure_value_addr_value. */
2666 if (structure_value_addr && struct_value == 0)
2667 {
2668 /* If structure_value_addr is a REG other than
2669 virtual_outgoing_args_rtx, we can use always use it. If it
2670 is not a REG, we must always copy it into a register.
2671 If it is virtual_outgoing_args_rtx, we must copy it to another
2672 register in some cases. */
2673 rtx temp = (!REG_P (structure_value_addr)
2674 || (ACCUMULATE_OUTGOING_ARGS
2675 && stack_arg_under_construction
2676 && structure_value_addr == virtual_outgoing_args_rtx)
2677 ? copy_addr_to_reg (convert_memory_address
2678 (Pmode, structure_value_addr))
2679 : structure_value_addr);
2680
2681 structure_value_addr_value =
2682 make_tree (build_pointer_type (TREE_TYPE (funtype)), temp);
2683 structure_value_addr_parm = CALL_WITH_BOUNDS_P (exp) ? 2 : 1;
2684 }
2685
2686 /* Count the arguments and set NUM_ACTUALS. */
2687 num_actuals =
2688 call_expr_nargs (exp) + num_complex_actuals + structure_value_addr_parm;
2689
2690 /* Compute number of named args.
2691 First, do a raw count of the args for INIT_CUMULATIVE_ARGS. */
2692
2693 if (type_arg_types != 0)
2694 n_named_args
2695 = (list_length (type_arg_types)
2696 /* Count the struct value address, if it is passed as a parm. */
2697 + structure_value_addr_parm);
2698 else
2699 /* If we know nothing, treat all args as named. */
2700 n_named_args = num_actuals;
2701
2702 /* Start updating where the next arg would go.
2703
2704 On some machines (such as the PA) indirect calls have a different
2705 calling convention than normal calls. The fourth argument in
2706 INIT_CUMULATIVE_ARGS tells the backend if this is an indirect call
2707 or not. */
2708 INIT_CUMULATIVE_ARGS (args_so_far_v, funtype, NULL_RTX, fndecl, n_named_args);
2709 args_so_far = pack_cumulative_args (&args_so_far_v);
2710
2711 /* Now possibly adjust the number of named args.
2712 Normally, don't include the last named arg if anonymous args follow.
2713 We do include the last named arg if
2714 targetm.calls.strict_argument_naming() returns nonzero.
2715 (If no anonymous args follow, the result of list_length is actually
2716 one too large. This is harmless.)
2717
2718 If targetm.calls.pretend_outgoing_varargs_named() returns
2719 nonzero, and targetm.calls.strict_argument_naming() returns zero,
2720 this machine will be able to place unnamed args that were passed
2721 in registers into the stack. So treat all args as named. This
2722 allows the insns emitting for a specific argument list to be
2723 independent of the function declaration.
2724
2725 If targetm.calls.pretend_outgoing_varargs_named() returns zero,
2726 we do not have any reliable way to pass unnamed args in
2727 registers, so we must force them into memory. */
2728
2729 if (type_arg_types != 0
2730 && targetm.calls.strict_argument_naming (args_so_far))
2731 ;
2732 else if (type_arg_types != 0
2733 && ! targetm.calls.pretend_outgoing_varargs_named (args_so_far))
2734 /* Don't include the last named arg. */
2735 --n_named_args;
2736 else
2737 /* Treat all args as named. */
2738 n_named_args = num_actuals;
2739
2740 /* Make a vector to hold all the information about each arg. */
2741 args = XALLOCAVEC (struct arg_data, num_actuals);
2742 memset (args, 0, num_actuals * sizeof (struct arg_data));
2743
2744 /* Build up entries in the ARGS array, compute the size of the
2745 arguments into ARGS_SIZE, etc. */
2746 initialize_argument_information (num_actuals, args, &args_size,
2747 n_named_args, exp,
2748 structure_value_addr_value, fndecl, fntype,
2749 args_so_far, reg_parm_stack_space,
2750 &old_stack_level, &old_pending_adj,
2751 &must_preallocate, &flags,
2752 &try_tail_call, CALL_FROM_THUNK_P (exp));
2753
2754 if (args_size.var)
2755 must_preallocate = 1;
2756
2757 /* Now make final decision about preallocating stack space. */
2758 must_preallocate = finalize_must_preallocate (must_preallocate,
2759 num_actuals, args,
2760 &args_size);
2761
2762 /* If the structure value address will reference the stack pointer, we
2763 must stabilize it. We don't need to do this if we know that we are
2764 not going to adjust the stack pointer in processing this call. */
2765
2766 if (structure_value_addr
2767 && (reg_mentioned_p (virtual_stack_dynamic_rtx, structure_value_addr)
2768 || reg_mentioned_p (virtual_outgoing_args_rtx,
2769 structure_value_addr))
2770 && (args_size.var
2771 || (!ACCUMULATE_OUTGOING_ARGS && args_size.constant)))
2772 structure_value_addr = copy_to_reg (structure_value_addr);
2773
2774 /* Tail calls can make things harder to debug, and we've traditionally
2775 pushed these optimizations into -O2. Don't try if we're already
2776 expanding a call, as that means we're an argument. Don't try if
2777 there's cleanups, as we know there's code to follow the call. */
2778
2779 if (currently_expanding_call++ != 0
2780 || !flag_optimize_sibling_calls
2781 || args_size.var
2782 || dbg_cnt (tail_call) == false)
2783 try_tail_call = 0;
2784
2785 /* Rest of purposes for tail call optimizations to fail. */
2786 if (
2787 #ifdef HAVE_sibcall_epilogue
2788 !HAVE_sibcall_epilogue
2789 #else
2790 1
2791 #endif
2792 || !try_tail_call
2793 /* Doing sibling call optimization needs some work, since
2794 structure_value_addr can be allocated on the stack.
2795 It does not seem worth the effort since few optimizable
2796 sibling calls will return a structure. */
2797 || structure_value_addr != NULL_RTX
2798 #ifdef REG_PARM_STACK_SPACE
2799 /* If outgoing reg parm stack space changes, we can not do sibcall. */
2800 || (OUTGOING_REG_PARM_STACK_SPACE (funtype)
2801 != OUTGOING_REG_PARM_STACK_SPACE (TREE_TYPE (current_function_decl)))
2802 || (reg_parm_stack_space != REG_PARM_STACK_SPACE (current_function_decl))
2803 #endif
2804 /* Check whether the target is able to optimize the call
2805 into a sibcall. */
2806 || !targetm.function_ok_for_sibcall (fndecl, exp)
2807 /* Functions that do not return exactly once may not be sibcall
2808 optimized. */
2809 || (flags & (ECF_RETURNS_TWICE | ECF_NORETURN))
2810 || TYPE_VOLATILE (TREE_TYPE (TREE_TYPE (addr)))
2811 /* If the called function is nested in the current one, it might access
2812 some of the caller's arguments, but could clobber them beforehand if
2813 the argument areas are shared. */
2814 || (fndecl && decl_function_context (fndecl) == current_function_decl)
2815 /* If this function requires more stack slots than the current
2816 function, we cannot change it into a sibling call.
2817 crtl->args.pretend_args_size is not part of the
2818 stack allocated by our caller. */
2819 || args_size.constant > (crtl->args.size
2820 - crtl->args.pretend_args_size)
2821 /* If the callee pops its own arguments, then it must pop exactly
2822 the same number of arguments as the current function. */
2823 || (targetm.calls.return_pops_args (fndecl, funtype, args_size.constant)
2824 != targetm.calls.return_pops_args (current_function_decl,
2825 TREE_TYPE (current_function_decl),
2826 crtl->args.size))
2827 || !lang_hooks.decls.ok_for_sibcall (fndecl))
2828 try_tail_call = 0;
2829
2830 /* Check if caller and callee disagree in promotion of function
2831 return value. */
2832 if (try_tail_call)
2833 {
2834 machine_mode caller_mode, caller_promoted_mode;
2835 machine_mode callee_mode, callee_promoted_mode;
2836 int caller_unsignedp, callee_unsignedp;
2837 tree caller_res = DECL_RESULT (current_function_decl);
2838
2839 caller_unsignedp = TYPE_UNSIGNED (TREE_TYPE (caller_res));
2840 caller_mode = DECL_MODE (caller_res);
2841 callee_unsignedp = TYPE_UNSIGNED (TREE_TYPE (funtype));
2842 callee_mode = TYPE_MODE (TREE_TYPE (funtype));
2843 caller_promoted_mode
2844 = promote_function_mode (TREE_TYPE (caller_res), caller_mode,
2845 &caller_unsignedp,
2846 TREE_TYPE (current_function_decl), 1);
2847 callee_promoted_mode
2848 = promote_function_mode (TREE_TYPE (funtype), callee_mode,
2849 &callee_unsignedp,
2850 funtype, 1);
2851 if (caller_mode != VOIDmode
2852 && (caller_promoted_mode != callee_promoted_mode
2853 || ((caller_mode != caller_promoted_mode
2854 || callee_mode != callee_promoted_mode)
2855 && (caller_unsignedp != callee_unsignedp
2856 || GET_MODE_BITSIZE (caller_mode)
2857 < GET_MODE_BITSIZE (callee_mode)))))
2858 try_tail_call = 0;
2859 }
2860
2861 /* Ensure current function's preferred stack boundary is at least
2862 what we need. Stack alignment may also increase preferred stack
2863 boundary. */
2864 if (crtl->preferred_stack_boundary < preferred_stack_boundary)
2865 crtl->preferred_stack_boundary = preferred_stack_boundary;
2866 else
2867 preferred_stack_boundary = crtl->preferred_stack_boundary;
2868
2869 preferred_unit_stack_boundary = preferred_stack_boundary / BITS_PER_UNIT;
2870
2871 /* We want to make two insn chains; one for a sibling call, the other
2872 for a normal call. We will select one of the two chains after
2873 initial RTL generation is complete. */
2874 for (pass = try_tail_call ? 0 : 1; pass < 2; pass++)
2875 {
2876 int sibcall_failure = 0;
2877 /* We want to emit any pending stack adjustments before the tail
2878 recursion "call". That way we know any adjustment after the tail
2879 recursion call can be ignored if we indeed use the tail
2880 call expansion. */
2881 saved_pending_stack_adjust save;
2882 rtx_insn *insns, *before_call, *after_args;
2883 rtx next_arg_reg;
2884
2885 if (pass == 0)
2886 {
2887 /* State variables we need to save and restore between
2888 iterations. */
2889 save_pending_stack_adjust (&save);
2890 }
2891 if (pass)
2892 flags &= ~ECF_SIBCALL;
2893 else
2894 flags |= ECF_SIBCALL;
2895
2896 /* Other state variables that we must reinitialize each time
2897 through the loop (that are not initialized by the loop itself). */
2898 argblock = 0;
2899 call_fusage = 0;
2900
2901 /* Start a new sequence for the normal call case.
2902
2903 From this point on, if the sibling call fails, we want to set
2904 sibcall_failure instead of continuing the loop. */
2905 start_sequence ();
2906
2907 /* Don't let pending stack adjusts add up to too much.
2908 Also, do all pending adjustments now if there is any chance
2909 this might be a call to alloca or if we are expanding a sibling
2910 call sequence.
2911 Also do the adjustments before a throwing call, otherwise
2912 exception handling can fail; PR 19225. */
2913 if (pending_stack_adjust >= 32
2914 || (pending_stack_adjust > 0
2915 && (flags & ECF_MAY_BE_ALLOCA))
2916 || (pending_stack_adjust > 0
2917 && flag_exceptions && !(flags & ECF_NOTHROW))
2918 || pass == 0)
2919 do_pending_stack_adjust ();
2920
2921 /* Precompute any arguments as needed. */
2922 if (pass)
2923 precompute_arguments (num_actuals, args);
2924
2925 /* Now we are about to start emitting insns that can be deleted
2926 if a libcall is deleted. */
2927 if (pass && (flags & ECF_MALLOC))
2928 start_sequence ();
2929
2930 if (pass == 0 && crtl->stack_protect_guard)
2931 stack_protect_epilogue ();
2932
2933 adjusted_args_size = args_size;
2934 /* Compute the actual size of the argument block required. The variable
2935 and constant sizes must be combined, the size may have to be rounded,
2936 and there may be a minimum required size. When generating a sibcall
2937 pattern, do not round up, since we'll be re-using whatever space our
2938 caller provided. */
2939 unadjusted_args_size
2940 = compute_argument_block_size (reg_parm_stack_space,
2941 &adjusted_args_size,
2942 fndecl, fntype,
2943 (pass == 0 ? 0
2944 : preferred_stack_boundary));
2945
2946 old_stack_allocated = stack_pointer_delta - pending_stack_adjust;
2947
2948 /* The argument block when performing a sibling call is the
2949 incoming argument block. */
2950 if (pass == 0)
2951 {
2952 argblock = crtl->args.internal_arg_pointer;
2953 if (STACK_GROWS_DOWNWARD)
2954 argblock
2955 = plus_constant (Pmode, argblock, crtl->args.pretend_args_size);
2956 else
2957 argblock
2958 = plus_constant (Pmode, argblock, -crtl->args.pretend_args_size);
2959
2960 stored_args_map = sbitmap_alloc (args_size.constant);
2961 bitmap_clear (stored_args_map);
2962 }
2963
2964 /* If we have no actual push instructions, or shouldn't use them,
2965 make space for all args right now. */
2966 else if (adjusted_args_size.var != 0)
2967 {
2968 if (old_stack_level == 0)
2969 {
2970 emit_stack_save (SAVE_BLOCK, &old_stack_level);
2971 old_stack_pointer_delta = stack_pointer_delta;
2972 old_pending_adj = pending_stack_adjust;
2973 pending_stack_adjust = 0;
2974 /* stack_arg_under_construction says whether a stack arg is
2975 being constructed at the old stack level. Pushing the stack
2976 gets a clean outgoing argument block. */
2977 old_stack_arg_under_construction = stack_arg_under_construction;
2978 stack_arg_under_construction = 0;
2979 }
2980 argblock = push_block (ARGS_SIZE_RTX (adjusted_args_size), 0, 0);
2981 if (flag_stack_usage_info)
2982 current_function_has_unbounded_dynamic_stack_size = 1;
2983 }
2984 else
2985 {
2986 /* Note that we must go through the motions of allocating an argument
2987 block even if the size is zero because we may be storing args
2988 in the area reserved for register arguments, which may be part of
2989 the stack frame. */
2990
2991 int needed = adjusted_args_size.constant;
2992
2993 /* Store the maximum argument space used. It will be pushed by
2994 the prologue (if ACCUMULATE_OUTGOING_ARGS, or stack overflow
2995 checking). */
2996
2997 if (needed > crtl->outgoing_args_size)
2998 crtl->outgoing_args_size = needed;
2999
3000 if (must_preallocate)
3001 {
3002 if (ACCUMULATE_OUTGOING_ARGS)
3003 {
3004 /* Since the stack pointer will never be pushed, it is
3005 possible for the evaluation of a parm to clobber
3006 something we have already written to the stack.
3007 Since most function calls on RISC machines do not use
3008 the stack, this is uncommon, but must work correctly.
3009
3010 Therefore, we save any area of the stack that was already
3011 written and that we are using. Here we set up to do this
3012 by making a new stack usage map from the old one. The
3013 actual save will be done by store_one_arg.
3014
3015 Another approach might be to try to reorder the argument
3016 evaluations to avoid this conflicting stack usage. */
3017
3018 /* Since we will be writing into the entire argument area,
3019 the map must be allocated for its entire size, not just
3020 the part that is the responsibility of the caller. */
3021 if (! OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl))))
3022 needed += reg_parm_stack_space;
3023
3024 if (ARGS_GROW_DOWNWARD)
3025 highest_outgoing_arg_in_use
3026 = MAX (initial_highest_arg_in_use, needed + 1);
3027 else
3028 highest_outgoing_arg_in_use
3029 = MAX (initial_highest_arg_in_use, needed);
3030
3031 free (stack_usage_map_buf);
3032 stack_usage_map_buf = XNEWVEC (char, highest_outgoing_arg_in_use);
3033 stack_usage_map = stack_usage_map_buf;
3034
3035 if (initial_highest_arg_in_use)
3036 memcpy (stack_usage_map, initial_stack_usage_map,
3037 initial_highest_arg_in_use);
3038
3039 if (initial_highest_arg_in_use != highest_outgoing_arg_in_use)
3040 memset (&stack_usage_map[initial_highest_arg_in_use], 0,
3041 (highest_outgoing_arg_in_use
3042 - initial_highest_arg_in_use));
3043 needed = 0;
3044
3045 /* The address of the outgoing argument list must not be
3046 copied to a register here, because argblock would be left
3047 pointing to the wrong place after the call to
3048 allocate_dynamic_stack_space below. */
3049
3050 argblock = virtual_outgoing_args_rtx;
3051 }
3052 else
3053 {
3054 if (inhibit_defer_pop == 0)
3055 {
3056 /* Try to reuse some or all of the pending_stack_adjust
3057 to get this space. */
3058 needed
3059 = (combine_pending_stack_adjustment_and_call
3060 (unadjusted_args_size,
3061 &adjusted_args_size,
3062 preferred_unit_stack_boundary));
3063
3064 /* combine_pending_stack_adjustment_and_call computes
3065 an adjustment before the arguments are allocated.
3066 Account for them and see whether or not the stack
3067 needs to go up or down. */
3068 needed = unadjusted_args_size - needed;
3069
3070 if (needed < 0)
3071 {
3072 /* We're releasing stack space. */
3073 /* ??? We can avoid any adjustment at all if we're
3074 already aligned. FIXME. */
3075 pending_stack_adjust = -needed;
3076 do_pending_stack_adjust ();
3077 needed = 0;
3078 }
3079 else
3080 /* We need to allocate space. We'll do that in
3081 push_block below. */
3082 pending_stack_adjust = 0;
3083 }
3084
3085 /* Special case this because overhead of `push_block' in
3086 this case is non-trivial. */
3087 if (needed == 0)
3088 argblock = virtual_outgoing_args_rtx;
3089 else
3090 {
3091 argblock = push_block (GEN_INT (needed), 0, 0);
3092 if (ARGS_GROW_DOWNWARD)
3093 argblock = plus_constant (Pmode, argblock, needed);
3094 }
3095
3096 /* We only really need to call `copy_to_reg' in the case
3097 where push insns are going to be used to pass ARGBLOCK
3098 to a function call in ARGS. In that case, the stack
3099 pointer changes value from the allocation point to the
3100 call point, and hence the value of
3101 VIRTUAL_OUTGOING_ARGS_RTX changes as well. But might
3102 as well always do it. */
3103 argblock = copy_to_reg (argblock);
3104 }
3105 }
3106 }
3107
3108 if (ACCUMULATE_OUTGOING_ARGS)
3109 {
3110 /* The save/restore code in store_one_arg handles all
3111 cases except one: a constructor call (including a C
3112 function returning a BLKmode struct) to initialize
3113 an argument. */
3114 if (stack_arg_under_construction)
3115 {
3116 rtx push_size
3117 = GEN_INT (adjusted_args_size.constant
3118 + (OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype
3119 : TREE_TYPE (fndecl))) ? 0
3120 : reg_parm_stack_space));
3121 if (old_stack_level == 0)
3122 {
3123 emit_stack_save (SAVE_BLOCK, &old_stack_level);
3124 old_stack_pointer_delta = stack_pointer_delta;
3125 old_pending_adj = pending_stack_adjust;
3126 pending_stack_adjust = 0;
3127 /* stack_arg_under_construction says whether a stack
3128 arg is being constructed at the old stack level.
3129 Pushing the stack gets a clean outgoing argument
3130 block. */
3131 old_stack_arg_under_construction
3132 = stack_arg_under_construction;
3133 stack_arg_under_construction = 0;
3134 /* Make a new map for the new argument list. */
3135 free (stack_usage_map_buf);
3136 stack_usage_map_buf = XCNEWVEC (char, highest_outgoing_arg_in_use);
3137 stack_usage_map = stack_usage_map_buf;
3138 highest_outgoing_arg_in_use = 0;
3139 }
3140 /* We can pass TRUE as the 4th argument because we just
3141 saved the stack pointer and will restore it right after
3142 the call. */
3143 allocate_dynamic_stack_space (push_size, 0,
3144 BIGGEST_ALIGNMENT, true);
3145 }
3146
3147 /* If argument evaluation might modify the stack pointer,
3148 copy the address of the argument list to a register. */
3149 for (i = 0; i < num_actuals; i++)
3150 if (args[i].pass_on_stack)
3151 {
3152 argblock = copy_addr_to_reg (argblock);
3153 break;
3154 }
3155 }
3156
3157 compute_argument_addresses (args, argblock, num_actuals);
3158
3159 /* Perform stack alignment before the first push (the last arg). */
3160 if (argblock == 0
3161 && adjusted_args_size.constant > reg_parm_stack_space
3162 && adjusted_args_size.constant != unadjusted_args_size)
3163 {
3164 /* When the stack adjustment is pending, we get better code
3165 by combining the adjustments. */
3166 if (pending_stack_adjust
3167 && ! inhibit_defer_pop)
3168 {
3169 pending_stack_adjust
3170 = (combine_pending_stack_adjustment_and_call
3171 (unadjusted_args_size,
3172 &adjusted_args_size,
3173 preferred_unit_stack_boundary));
3174 do_pending_stack_adjust ();
3175 }
3176 else if (argblock == 0)
3177 anti_adjust_stack (GEN_INT (adjusted_args_size.constant
3178 - unadjusted_args_size));
3179 }
3180 /* Now that the stack is properly aligned, pops can't safely
3181 be deferred during the evaluation of the arguments. */
3182 NO_DEFER_POP;
3183
3184 /* Record the maximum pushed stack space size. We need to delay
3185 doing it this far to take into account the optimization done
3186 by combine_pending_stack_adjustment_and_call. */
3187 if (flag_stack_usage_info
3188 && !ACCUMULATE_OUTGOING_ARGS
3189 && pass
3190 && adjusted_args_size.var == 0)
3191 {
3192 int pushed = adjusted_args_size.constant + pending_stack_adjust;
3193 if (pushed > current_function_pushed_stack_size)
3194 current_function_pushed_stack_size = pushed;
3195 }
3196
3197 funexp = rtx_for_function_call (fndecl, addr);
3198
3199 /* Precompute all register parameters. It isn't safe to compute anything
3200 once we have started filling any specific hard regs. */
3201 precompute_register_parameters (num_actuals, args, &reg_parm_seen);
3202
3203 if (CALL_EXPR_STATIC_CHAIN (exp))
3204 static_chain_value = expand_normal (CALL_EXPR_STATIC_CHAIN (exp));
3205 else
3206 static_chain_value = 0;
3207
3208 #ifdef REG_PARM_STACK_SPACE
3209 /* Save the fixed argument area if it's part of the caller's frame and
3210 is clobbered by argument setup for this call. */
3211 if (ACCUMULATE_OUTGOING_ARGS && pass)
3212 save_area = save_fixed_argument_area (reg_parm_stack_space, argblock,
3213 &low_to_save, &high_to_save);
3214 #endif
3215
3216 /* Now store (and compute if necessary) all non-register parms.
3217 These come before register parms, since they can require block-moves,
3218 which could clobber the registers used for register parms.
3219 Parms which have partial registers are not stored here,
3220 but we do preallocate space here if they want that. */
3221
3222 for (i = 0; i < num_actuals; i++)
3223 {
3224 /* Delay bounds until all other args are stored. */
3225 if (POINTER_BOUNDS_P (args[i].tree_value))
3226 continue;
3227 else if (args[i].reg == 0 || args[i].pass_on_stack)
3228 {
3229 rtx_insn *before_arg = get_last_insn ();
3230
3231 /* We don't allow passing huge (> 2^30 B) arguments
3232 by value. It would cause an overflow later on. */
3233 if (adjusted_args_size.constant
3234 >= (1 << (HOST_BITS_PER_INT - 2)))
3235 {
3236 sorry ("passing too large argument on stack");
3237 continue;
3238 }
3239
3240 if (store_one_arg (&args[i], argblock, flags,
3241 adjusted_args_size.var != 0,
3242 reg_parm_stack_space)
3243 || (pass == 0
3244 && check_sibcall_argument_overlap (before_arg,
3245 &args[i], 1)))
3246 sibcall_failure = 1;
3247 }
3248
3249 if (args[i].stack)
3250 call_fusage
3251 = gen_rtx_EXPR_LIST (TYPE_MODE (TREE_TYPE (args[i].tree_value)),
3252 gen_rtx_USE (VOIDmode, args[i].stack),
3253 call_fusage);
3254 }
3255
3256 /* If we have a parm that is passed in registers but not in memory
3257 and whose alignment does not permit a direct copy into registers,
3258 make a group of pseudos that correspond to each register that we
3259 will later fill. */
3260 if (STRICT_ALIGNMENT)
3261 store_unaligned_arguments_into_pseudos (args, num_actuals);
3262
3263 /* Now store any partially-in-registers parm.
3264 This is the last place a block-move can happen. */
3265 if (reg_parm_seen)
3266 for (i = 0; i < num_actuals; i++)
3267 if (args[i].partial != 0 && ! args[i].pass_on_stack)
3268 {
3269 rtx_insn *before_arg = get_last_insn ();
3270
3271 /* On targets with weird calling conventions (e.g. PA) it's
3272 hard to ensure that all cases of argument overlap between
3273 stack and registers work. Play it safe and bail out. */
3274 if (ARGS_GROW_DOWNWARD && !STACK_GROWS_DOWNWARD)
3275 {
3276 sibcall_failure = 1;
3277 break;
3278 }
3279
3280 if (store_one_arg (&args[i], argblock, flags,
3281 adjusted_args_size.var != 0,
3282 reg_parm_stack_space)
3283 || (pass == 0
3284 && check_sibcall_argument_overlap (before_arg,
3285 &args[i], 1)))
3286 sibcall_failure = 1;
3287 }
3288
3289 bool any_regs = false;
3290 for (i = 0; i < num_actuals; i++)
3291 if (args[i].reg != NULL_RTX)
3292 {
3293 any_regs = true;
3294 targetm.calls.call_args (args[i].reg, funtype);
3295 }
3296 if (!any_regs)
3297 targetm.calls.call_args (pc_rtx, funtype);
3298
3299 /* Figure out the register where the value, if any, will come back. */
3300 valreg = 0;
3301 valbnd = 0;
3302 if (TYPE_MODE (rettype) != VOIDmode
3303 && ! structure_value_addr)
3304 {
3305 if (pcc_struct_value)
3306 {
3307 valreg = hard_function_value (build_pointer_type (rettype),
3308 fndecl, NULL, (pass == 0));
3309 if (CALL_WITH_BOUNDS_P (exp))
3310 valbnd = targetm.calls.
3311 chkp_function_value_bounds (build_pointer_type (rettype),
3312 fndecl, (pass == 0));
3313 }
3314 else
3315 {
3316 valreg = hard_function_value (rettype, fndecl, fntype,
3317 (pass == 0));
3318 if (CALL_WITH_BOUNDS_P (exp))
3319 valbnd = targetm.calls.chkp_function_value_bounds (rettype,
3320 fndecl,
3321 (pass == 0));
3322 }
3323
3324 /* If VALREG is a PARALLEL whose first member has a zero
3325 offset, use that. This is for targets such as m68k that
3326 return the same value in multiple places. */
3327 if (GET_CODE (valreg) == PARALLEL)
3328 {
3329 rtx elem = XVECEXP (valreg, 0, 0);
3330 rtx where = XEXP (elem, 0);
3331 rtx offset = XEXP (elem, 1);
3332 if (offset == const0_rtx
3333 && GET_MODE (where) == GET_MODE (valreg))
3334 valreg = where;
3335 }
3336 }
3337
3338 /* Store all bounds not passed in registers. */
3339 for (i = 0; i < num_actuals; i++)
3340 {
3341 if (POINTER_BOUNDS_P (args[i].tree_value)
3342 && !args[i].reg)
3343 store_bounds (&args[i],
3344 args[i].pointer_arg == -1
3345 ? NULL
3346 : &args[args[i].pointer_arg]);
3347 }
3348
3349 /* If register arguments require space on the stack and stack space
3350 was not preallocated, allocate stack space here for arguments
3351 passed in registers. */
3352 if (OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl)))
3353 && !ACCUMULATE_OUTGOING_ARGS
3354 && must_preallocate == 0 && reg_parm_stack_space > 0)
3355 anti_adjust_stack (GEN_INT (reg_parm_stack_space));
3356
3357 /* Pass the function the address in which to return a
3358 structure value. */
3359 if (pass != 0 && structure_value_addr && ! structure_value_addr_parm)
3360 {
3361 structure_value_addr
3362 = convert_memory_address (Pmode, structure_value_addr);
3363 emit_move_insn (struct_value,
3364 force_reg (Pmode,
3365 force_operand (structure_value_addr,
3366 NULL_RTX)));
3367
3368 if (REG_P (struct_value))
3369 use_reg (&call_fusage, struct_value);
3370 }
3371
3372 after_args = get_last_insn ();
3373 funexp = prepare_call_address (fndecl ? fndecl : fntype, funexp,
3374 static_chain_value, &call_fusage,
3375 reg_parm_seen, pass == 0);
3376
3377 load_register_parameters (args, num_actuals, &call_fusage, flags,
3378 pass == 0, &sibcall_failure);
3379
3380 /* Save a pointer to the last insn before the call, so that we can
3381 later safely search backwards to find the CALL_INSN. */
3382 before_call = get_last_insn ();
3383
3384 /* Set up next argument register. For sibling calls on machines
3385 with register windows this should be the incoming register. */
3386 if (pass == 0)
3387 next_arg_reg = targetm.calls.function_incoming_arg (args_so_far,
3388 VOIDmode,
3389 void_type_node,
3390 true);
3391 else
3392 next_arg_reg = targetm.calls.function_arg (args_so_far,
3393 VOIDmode, void_type_node,
3394 true);
3395
3396 if (pass == 1 && (return_flags & ERF_RETURNS_ARG))
3397 {
3398 int arg_nr = return_flags & ERF_RETURN_ARG_MASK;
3399 arg_nr = num_actuals - arg_nr - 1;
3400 if (arg_nr >= 0
3401 && arg_nr < num_actuals
3402 && args[arg_nr].reg
3403 && valreg
3404 && REG_P (valreg)
3405 && GET_MODE (args[arg_nr].reg) == GET_MODE (valreg))
3406 call_fusage
3407 = gen_rtx_EXPR_LIST (TYPE_MODE (TREE_TYPE (args[arg_nr].tree_value)),
3408 gen_rtx_SET (valreg, args[arg_nr].reg),
3409 call_fusage);
3410 }
3411 /* All arguments and registers used for the call must be set up by
3412 now! */
3413
3414 /* Stack must be properly aligned now. */
3415 gcc_assert (!pass
3416 || !(stack_pointer_delta % preferred_unit_stack_boundary));
3417
3418 /* Generate the actual call instruction. */
3419 emit_call_1 (funexp, exp, fndecl, funtype, unadjusted_args_size,
3420 adjusted_args_size.constant, struct_value_size,
3421 next_arg_reg, valreg, old_inhibit_defer_pop, call_fusage,
3422 flags, args_so_far);
3423
3424 if (flag_ipa_ra)
3425 {
3426 rtx_call_insn *last;
3427 rtx datum = NULL_RTX;
3428 if (fndecl != NULL_TREE)
3429 {
3430 datum = XEXP (DECL_RTL (fndecl), 0);
3431 gcc_assert (datum != NULL_RTX
3432 && GET_CODE (datum) == SYMBOL_REF);
3433 }
3434 last = last_call_insn ();
3435 add_reg_note (last, REG_CALL_DECL, datum);
3436 }
3437
3438 /* If the call setup or the call itself overlaps with anything
3439 of the argument setup we probably clobbered our call address.
3440 In that case we can't do sibcalls. */
3441 if (pass == 0
3442 && check_sibcall_argument_overlap (after_args, 0, 0))
3443 sibcall_failure = 1;
3444
3445 /* If a non-BLKmode value is returned at the most significant end
3446 of a register, shift the register right by the appropriate amount
3447 and update VALREG accordingly. BLKmode values are handled by the
3448 group load/store machinery below. */
3449 if (!structure_value_addr
3450 && !pcc_struct_value
3451 && TYPE_MODE (rettype) != VOIDmode
3452 && TYPE_MODE (rettype) != BLKmode
3453 && REG_P (valreg)
3454 && targetm.calls.return_in_msb (rettype))
3455 {
3456 if (shift_return_value (TYPE_MODE (rettype), false, valreg))
3457 sibcall_failure = 1;
3458 valreg = gen_rtx_REG (TYPE_MODE (rettype), REGNO (valreg));
3459 }
3460
3461 if (pass && (flags & ECF_MALLOC))
3462 {
3463 rtx temp = gen_reg_rtx (GET_MODE (valreg));
3464 rtx_insn *last, *insns;
3465
3466 /* The return value from a malloc-like function is a pointer. */
3467 if (TREE_CODE (rettype) == POINTER_TYPE)
3468 mark_reg_pointer (temp, MALLOC_ABI_ALIGNMENT);
3469
3470 emit_move_insn (temp, valreg);
3471
3472 /* The return value from a malloc-like function can not alias
3473 anything else. */
3474 last = get_last_insn ();
3475 add_reg_note (last, REG_NOALIAS, temp);
3476
3477 /* Write out the sequence. */
3478 insns = get_insns ();
3479 end_sequence ();
3480 emit_insn (insns);
3481 valreg = temp;
3482 }
3483
3484 /* For calls to `setjmp', etc., inform
3485 function.c:setjmp_warnings that it should complain if
3486 nonvolatile values are live. For functions that cannot
3487 return, inform flow that control does not fall through. */
3488
3489 if ((flags & ECF_NORETURN) || pass == 0)
3490 {
3491 /* The barrier must be emitted
3492 immediately after the CALL_INSN. Some ports emit more
3493 than just a CALL_INSN above, so we must search for it here. */
3494
3495 rtx_insn *last = get_last_insn ();
3496 while (!CALL_P (last))
3497 {
3498 last = PREV_INSN (last);
3499 /* There was no CALL_INSN? */
3500 gcc_assert (last != before_call);
3501 }
3502
3503 emit_barrier_after (last);
3504
3505 /* Stack adjustments after a noreturn call are dead code.
3506 However when NO_DEFER_POP is in effect, we must preserve
3507 stack_pointer_delta. */
3508 if (inhibit_defer_pop == 0)
3509 {
3510 stack_pointer_delta = old_stack_allocated;
3511 pending_stack_adjust = 0;
3512 }
3513 }
3514
3515 /* If value type not void, return an rtx for the value. */
3516
3517 if (TYPE_MODE (rettype) == VOIDmode
3518 || ignore)
3519 target = const0_rtx;
3520 else if (structure_value_addr)
3521 {
3522 if (target == 0 || !MEM_P (target))
3523 {
3524 target
3525 = gen_rtx_MEM (TYPE_MODE (rettype),
3526 memory_address (TYPE_MODE (rettype),
3527 structure_value_addr));
3528 set_mem_attributes (target, rettype, 1);
3529 }
3530 }
3531 else if (pcc_struct_value)
3532 {
3533 /* This is the special C++ case where we need to
3534 know what the true target was. We take care to
3535 never use this value more than once in one expression. */
3536 target = gen_rtx_MEM (TYPE_MODE (rettype),
3537 copy_to_reg (valreg));
3538 set_mem_attributes (target, rettype, 1);
3539 }
3540 /* Handle calls that return values in multiple non-contiguous locations.
3541 The Irix 6 ABI has examples of this. */
3542 else if (GET_CODE (valreg) == PARALLEL)
3543 {
3544 if (target == 0)
3545 target = emit_group_move_into_temps (valreg);
3546 else if (rtx_equal_p (target, valreg))
3547 ;
3548 else if (GET_CODE (target) == PARALLEL)
3549 /* Handle the result of a emit_group_move_into_temps
3550 call in the previous pass. */
3551 emit_group_move (target, valreg);
3552 else
3553 emit_group_store (target, valreg, rettype,
3554 int_size_in_bytes (rettype));
3555 }
3556 else if (target
3557 && GET_MODE (target) == TYPE_MODE (rettype)
3558 && GET_MODE (target) == GET_MODE (valreg))
3559 {
3560 bool may_overlap = false;
3561
3562 /* We have to copy a return value in a CLASS_LIKELY_SPILLED hard
3563 reg to a plain register. */
3564 if (!REG_P (target) || HARD_REGISTER_P (target))
3565 valreg = avoid_likely_spilled_reg (valreg);
3566
3567 /* If TARGET is a MEM in the argument area, and we have
3568 saved part of the argument area, then we can't store
3569 directly into TARGET as it may get overwritten when we
3570 restore the argument save area below. Don't work too
3571 hard though and simply force TARGET to a register if it
3572 is a MEM; the optimizer is quite likely to sort it out. */
3573 if (ACCUMULATE_OUTGOING_ARGS && pass && MEM_P (target))
3574 for (i = 0; i < num_actuals; i++)
3575 if (args[i].save_area)
3576 {
3577 may_overlap = true;
3578 break;
3579 }
3580
3581 if (may_overlap)
3582 target = copy_to_reg (valreg);
3583 else
3584 {
3585 /* TARGET and VALREG cannot be equal at this point
3586 because the latter would not have
3587 REG_FUNCTION_VALUE_P true, while the former would if
3588 it were referring to the same register.
3589
3590 If they refer to the same register, this move will be
3591 a no-op, except when function inlining is being
3592 done. */
3593 emit_move_insn (target, valreg);
3594
3595 /* If we are setting a MEM, this code must be executed.
3596 Since it is emitted after the call insn, sibcall
3597 optimization cannot be performed in that case. */
3598 if (MEM_P (target))
3599 sibcall_failure = 1;
3600 }
3601 }
3602 else
3603 target = copy_to_reg (avoid_likely_spilled_reg (valreg));
3604
3605 /* If we promoted this return value, make the proper SUBREG.
3606 TARGET might be const0_rtx here, so be careful. */
3607 if (REG_P (target)
3608 && TYPE_MODE (rettype) != BLKmode
3609 && GET_MODE (target) != TYPE_MODE (rettype))
3610 {
3611 tree type = rettype;
3612 int unsignedp = TYPE_UNSIGNED (type);
3613 int offset = 0;
3614 machine_mode pmode;
3615
3616 /* Ensure we promote as expected, and get the new unsignedness. */
3617 pmode = promote_function_mode (type, TYPE_MODE (type), &unsignedp,
3618 funtype, 1);
3619 gcc_assert (GET_MODE (target) == pmode);
3620
3621 if ((WORDS_BIG_ENDIAN || BYTES_BIG_ENDIAN)
3622 && (GET_MODE_SIZE (GET_MODE (target))
3623 > GET_MODE_SIZE (TYPE_MODE (type))))
3624 {
3625 offset = GET_MODE_SIZE (GET_MODE (target))
3626 - GET_MODE_SIZE (TYPE_MODE (type));
3627 if (! BYTES_BIG_ENDIAN)
3628 offset = (offset / UNITS_PER_WORD) * UNITS_PER_WORD;
3629 else if (! WORDS_BIG_ENDIAN)
3630 offset %= UNITS_PER_WORD;
3631 }
3632
3633 target = gen_rtx_SUBREG (TYPE_MODE (type), target, offset);
3634 SUBREG_PROMOTED_VAR_P (target) = 1;
3635 SUBREG_PROMOTED_SET (target, unsignedp);
3636 }
3637
3638 /* If size of args is variable or this was a constructor call for a stack
3639 argument, restore saved stack-pointer value. */
3640
3641 if (old_stack_level)
3642 {
3643 rtx_insn *prev = get_last_insn ();
3644
3645 emit_stack_restore (SAVE_BLOCK, old_stack_level);
3646 stack_pointer_delta = old_stack_pointer_delta;
3647
3648 fixup_args_size_notes (prev, get_last_insn (), stack_pointer_delta);
3649
3650 pending_stack_adjust = old_pending_adj;
3651 old_stack_allocated = stack_pointer_delta - pending_stack_adjust;
3652 stack_arg_under_construction = old_stack_arg_under_construction;
3653 highest_outgoing_arg_in_use = initial_highest_arg_in_use;
3654 stack_usage_map = initial_stack_usage_map;
3655 sibcall_failure = 1;
3656 }
3657 else if (ACCUMULATE_OUTGOING_ARGS && pass)
3658 {
3659 #ifdef REG_PARM_STACK_SPACE
3660 if (save_area)
3661 restore_fixed_argument_area (save_area, argblock,
3662 high_to_save, low_to_save);
3663 #endif
3664
3665 /* If we saved any argument areas, restore them. */
3666 for (i = 0; i < num_actuals; i++)
3667 if (args[i].save_area)
3668 {
3669 machine_mode save_mode = GET_MODE (args[i].save_area);
3670 rtx stack_area
3671 = gen_rtx_MEM (save_mode,
3672 memory_address (save_mode,
3673 XEXP (args[i].stack_slot, 0)));
3674
3675 if (save_mode != BLKmode)
3676 emit_move_insn (stack_area, args[i].save_area);
3677 else
3678 emit_block_move (stack_area, args[i].save_area,
3679 GEN_INT (args[i].locate.size.constant),
3680 BLOCK_OP_CALL_PARM);
3681 }
3682
3683 highest_outgoing_arg_in_use = initial_highest_arg_in_use;
3684 stack_usage_map = initial_stack_usage_map;
3685 }
3686
3687 /* If this was alloca, record the new stack level. */
3688 if (flags & ECF_MAY_BE_ALLOCA)
3689 record_new_stack_level ();
3690
3691 /* Free up storage we no longer need. */
3692 for (i = 0; i < num_actuals; ++i)
3693 free (args[i].aligned_regs);
3694
3695 targetm.calls.end_call_args ();
3696
3697 insns = get_insns ();
3698 end_sequence ();
3699
3700 if (pass == 0)
3701 {
3702 tail_call_insns = insns;
3703
3704 /* Restore the pending stack adjustment now that we have
3705 finished generating the sibling call sequence. */
3706
3707 restore_pending_stack_adjust (&save);
3708
3709 /* Prepare arg structure for next iteration. */
3710 for (i = 0; i < num_actuals; i++)
3711 {
3712 args[i].value = 0;
3713 args[i].aligned_regs = 0;
3714 args[i].stack = 0;
3715 }
3716
3717 sbitmap_free (stored_args_map);
3718 internal_arg_pointer_exp_state.scan_start = NULL;
3719 internal_arg_pointer_exp_state.cache.release ();
3720 }
3721 else
3722 {
3723 normal_call_insns = insns;
3724
3725 /* Verify that we've deallocated all the stack we used. */
3726 gcc_assert ((flags & ECF_NORETURN)
3727 || (old_stack_allocated
3728 == stack_pointer_delta - pending_stack_adjust));
3729 }
3730
3731 /* If something prevents making this a sibling call,
3732 zero out the sequence. */
3733 if (sibcall_failure)
3734 tail_call_insns = NULL;
3735 else
3736 break;
3737 }
3738
3739 /* If tail call production succeeded, we need to remove REG_EQUIV notes on
3740 arguments too, as argument area is now clobbered by the call. */
3741 if (tail_call_insns)
3742 {
3743 emit_insn (tail_call_insns);
3744 crtl->tail_call_emit = true;
3745 }
3746 else
3747 emit_insn (normal_call_insns);
3748
3749 currently_expanding_call--;
3750
3751 free (stack_usage_map_buf);
3752
3753 /* Join result with returned bounds so caller may use them if needed. */
3754 target = chkp_join_splitted_slot (target, valbnd);
3755
3756 return target;
3757 }
3758
3759 /* A sibling call sequence invalidates any REG_EQUIV notes made for
3760 this function's incoming arguments.
3761
3762 At the start of RTL generation we know the only REG_EQUIV notes
3763 in the rtl chain are those for incoming arguments, so we can look
3764 for REG_EQUIV notes between the start of the function and the
3765 NOTE_INSN_FUNCTION_BEG.
3766
3767 This is (slight) overkill. We could keep track of the highest
3768 argument we clobber and be more selective in removing notes, but it
3769 does not seem to be worth the effort. */
3770
3771 void
3772 fixup_tail_calls (void)
3773 {
3774 rtx_insn *insn;
3775
3776 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
3777 {
3778 rtx note;
3779
3780 /* There are never REG_EQUIV notes for the incoming arguments
3781 after the NOTE_INSN_FUNCTION_BEG note, so stop if we see it. */
3782 if (NOTE_P (insn)
3783 && NOTE_KIND (insn) == NOTE_INSN_FUNCTION_BEG)
3784 break;
3785
3786 note = find_reg_note (insn, REG_EQUIV, 0);
3787 if (note)
3788 remove_note (insn, note);
3789 note = find_reg_note (insn, REG_EQUIV, 0);
3790 gcc_assert (!note);
3791 }
3792 }
3793
3794 /* Traverse a list of TYPES and expand all complex types into their
3795 components. */
3796 static tree
3797 split_complex_types (tree types)
3798 {
3799 tree p;
3800
3801 /* Before allocating memory, check for the common case of no complex. */
3802 for (p = types; p; p = TREE_CHAIN (p))
3803 {
3804 tree type = TREE_VALUE (p);
3805 if (TREE_CODE (type) == COMPLEX_TYPE
3806 && targetm.calls.split_complex_arg (type))
3807 goto found;
3808 }
3809 return types;
3810
3811 found:
3812 types = copy_list (types);
3813
3814 for (p = types; p; p = TREE_CHAIN (p))
3815 {
3816 tree complex_type = TREE_VALUE (p);
3817
3818 if (TREE_CODE (complex_type) == COMPLEX_TYPE
3819 && targetm.calls.split_complex_arg (complex_type))
3820 {
3821 tree next, imag;
3822
3823 /* Rewrite complex type with component type. */
3824 TREE_VALUE (p) = TREE_TYPE (complex_type);
3825 next = TREE_CHAIN (p);
3826
3827 /* Add another component type for the imaginary part. */
3828 imag = build_tree_list (NULL_TREE, TREE_VALUE (p));
3829 TREE_CHAIN (p) = imag;
3830 TREE_CHAIN (imag) = next;
3831
3832 /* Skip the newly created node. */
3833 p = TREE_CHAIN (p);
3834 }
3835 }
3836
3837 return types;
3838 }
3839 \f
3840 /* Output a library call to function FUN (a SYMBOL_REF rtx).
3841 The RETVAL parameter specifies whether return value needs to be saved, other
3842 parameters are documented in the emit_library_call function below. */
3843
3844 static rtx
3845 emit_library_call_value_1 (int retval, rtx orgfun, rtx value,
3846 enum libcall_type fn_type,
3847 machine_mode outmode, int nargs, va_list p)
3848 {
3849 /* Total size in bytes of all the stack-parms scanned so far. */
3850 struct args_size args_size;
3851 /* Size of arguments before any adjustments (such as rounding). */
3852 struct args_size original_args_size;
3853 int argnum;
3854 rtx fun;
3855 /* Todo, choose the correct decl type of orgfun. Sadly this information
3856 isn't present here, so we default to native calling abi here. */
3857 tree fndecl ATTRIBUTE_UNUSED = NULL_TREE; /* library calls default to host calling abi ? */
3858 tree fntype ATTRIBUTE_UNUSED = NULL_TREE; /* library calls default to host calling abi ? */
3859 int count;
3860 rtx argblock = 0;
3861 CUMULATIVE_ARGS args_so_far_v;
3862 cumulative_args_t args_so_far;
3863 struct arg
3864 {
3865 rtx value;
3866 machine_mode mode;
3867 rtx reg;
3868 int partial;
3869 struct locate_and_pad_arg_data locate;
3870 rtx save_area;
3871 };
3872 struct arg *argvec;
3873 int old_inhibit_defer_pop = inhibit_defer_pop;
3874 rtx call_fusage = 0;
3875 rtx mem_value = 0;
3876 rtx valreg;
3877 int pcc_struct_value = 0;
3878 int struct_value_size = 0;
3879 int flags;
3880 int reg_parm_stack_space = 0;
3881 int needed;
3882 rtx_insn *before_call;
3883 bool have_push_fusage;
3884 tree tfom; /* type_for_mode (outmode, 0) */
3885
3886 #ifdef REG_PARM_STACK_SPACE
3887 /* Define the boundary of the register parm stack space that needs to be
3888 save, if any. */
3889 int low_to_save = 0, high_to_save = 0;
3890 rtx save_area = 0; /* Place that it is saved. */
3891 #endif
3892
3893 /* Size of the stack reserved for parameter registers. */
3894 int initial_highest_arg_in_use = highest_outgoing_arg_in_use;
3895 char *initial_stack_usage_map = stack_usage_map;
3896 char *stack_usage_map_buf = NULL;
3897
3898 rtx struct_value = targetm.calls.struct_value_rtx (0, 0);
3899
3900 #ifdef REG_PARM_STACK_SPACE
3901 reg_parm_stack_space = REG_PARM_STACK_SPACE ((tree) 0);
3902 #endif
3903
3904 /* By default, library functions can not throw. */
3905 flags = ECF_NOTHROW;
3906
3907 switch (fn_type)
3908 {
3909 case LCT_NORMAL:
3910 break;
3911 case LCT_CONST:
3912 flags |= ECF_CONST;
3913 break;
3914 case LCT_PURE:
3915 flags |= ECF_PURE;
3916 break;
3917 case LCT_NORETURN:
3918 flags |= ECF_NORETURN;
3919 break;
3920 case LCT_THROW:
3921 flags = ECF_NORETURN;
3922 break;
3923 case LCT_RETURNS_TWICE:
3924 flags = ECF_RETURNS_TWICE;
3925 break;
3926 }
3927 fun = orgfun;
3928
3929 /* Ensure current function's preferred stack boundary is at least
3930 what we need. */
3931 if (crtl->preferred_stack_boundary < PREFERRED_STACK_BOUNDARY)
3932 crtl->preferred_stack_boundary = PREFERRED_STACK_BOUNDARY;
3933
3934 /* If this kind of value comes back in memory,
3935 decide where in memory it should come back. */
3936 if (outmode != VOIDmode)
3937 {
3938 tfom = lang_hooks.types.type_for_mode (outmode, 0);
3939 if (aggregate_value_p (tfom, 0))
3940 {
3941 #ifdef PCC_STATIC_STRUCT_RETURN
3942 rtx pointer_reg
3943 = hard_function_value (build_pointer_type (tfom), 0, 0, 0);
3944 mem_value = gen_rtx_MEM (outmode, pointer_reg);
3945 pcc_struct_value = 1;
3946 if (value == 0)
3947 value = gen_reg_rtx (outmode);
3948 #else /* not PCC_STATIC_STRUCT_RETURN */
3949 struct_value_size = GET_MODE_SIZE (outmode);
3950 if (value != 0 && MEM_P (value))
3951 mem_value = value;
3952 else
3953 mem_value = assign_temp (tfom, 1, 1);
3954 #endif
3955 /* This call returns a big structure. */
3956 flags &= ~(ECF_CONST | ECF_PURE | ECF_LOOPING_CONST_OR_PURE);
3957 }
3958 }
3959 else
3960 tfom = void_type_node;
3961
3962 /* ??? Unfinished: must pass the memory address as an argument. */
3963
3964 /* Copy all the libcall-arguments out of the varargs data
3965 and into a vector ARGVEC.
3966
3967 Compute how to pass each argument. We only support a very small subset
3968 of the full argument passing conventions to limit complexity here since
3969 library functions shouldn't have many args. */
3970
3971 argvec = XALLOCAVEC (struct arg, nargs + 1);
3972 memset (argvec, 0, (nargs + 1) * sizeof (struct arg));
3973
3974 #ifdef INIT_CUMULATIVE_LIBCALL_ARGS
3975 INIT_CUMULATIVE_LIBCALL_ARGS (args_so_far_v, outmode, fun);
3976 #else
3977 INIT_CUMULATIVE_ARGS (args_so_far_v, NULL_TREE, fun, 0, nargs);
3978 #endif
3979 args_so_far = pack_cumulative_args (&args_so_far_v);
3980
3981 args_size.constant = 0;
3982 args_size.var = 0;
3983
3984 count = 0;
3985
3986 push_temp_slots ();
3987
3988 /* If there's a structure value address to be passed,
3989 either pass it in the special place, or pass it as an extra argument. */
3990 if (mem_value && struct_value == 0 && ! pcc_struct_value)
3991 {
3992 rtx addr = XEXP (mem_value, 0);
3993
3994 nargs++;
3995
3996 /* Make sure it is a reasonable operand for a move or push insn. */
3997 if (!REG_P (addr) && !MEM_P (addr)
3998 && !(CONSTANT_P (addr)
3999 && targetm.legitimate_constant_p (Pmode, addr)))
4000 addr = force_operand (addr, NULL_RTX);
4001
4002 argvec[count].value = addr;
4003 argvec[count].mode = Pmode;
4004 argvec[count].partial = 0;
4005
4006 argvec[count].reg = targetm.calls.function_arg (args_so_far,
4007 Pmode, NULL_TREE, true);
4008 gcc_assert (targetm.calls.arg_partial_bytes (args_so_far, Pmode,
4009 NULL_TREE, 1) == 0);
4010
4011 locate_and_pad_parm (Pmode, NULL_TREE,
4012 #ifdef STACK_PARMS_IN_REG_PARM_AREA
4013 1,
4014 #else
4015 argvec[count].reg != 0,
4016 #endif
4017 reg_parm_stack_space, 0,
4018 NULL_TREE, &args_size, &argvec[count].locate);
4019
4020 if (argvec[count].reg == 0 || argvec[count].partial != 0
4021 || reg_parm_stack_space > 0)
4022 args_size.constant += argvec[count].locate.size.constant;
4023
4024 targetm.calls.function_arg_advance (args_so_far, Pmode, (tree) 0, true);
4025
4026 count++;
4027 }
4028
4029 for (; count < nargs; count++)
4030 {
4031 rtx val = va_arg (p, rtx);
4032 machine_mode mode = (machine_mode) va_arg (p, int);
4033 int unsigned_p = 0;
4034
4035 /* We cannot convert the arg value to the mode the library wants here;
4036 must do it earlier where we know the signedness of the arg. */
4037 gcc_assert (mode != BLKmode
4038 && (GET_MODE (val) == mode || GET_MODE (val) == VOIDmode));
4039
4040 /* Make sure it is a reasonable operand for a move or push insn. */
4041 if (!REG_P (val) && !MEM_P (val)
4042 && !(CONSTANT_P (val) && targetm.legitimate_constant_p (mode, val)))
4043 val = force_operand (val, NULL_RTX);
4044
4045 if (pass_by_reference (&args_so_far_v, mode, NULL_TREE, 1))
4046 {
4047 rtx slot;
4048 int must_copy
4049 = !reference_callee_copied (&args_so_far_v, mode, NULL_TREE, 1);
4050
4051 /* If this was a CONST function, it is now PURE since it now
4052 reads memory. */
4053 if (flags & ECF_CONST)
4054 {
4055 flags &= ~ECF_CONST;
4056 flags |= ECF_PURE;
4057 }
4058
4059 if (MEM_P (val) && !must_copy)
4060 {
4061 tree val_expr = MEM_EXPR (val);
4062 if (val_expr)
4063 mark_addressable (val_expr);
4064 slot = val;
4065 }
4066 else
4067 {
4068 slot = assign_temp (lang_hooks.types.type_for_mode (mode, 0),
4069 1, 1);
4070 emit_move_insn (slot, val);
4071 }
4072
4073 call_fusage = gen_rtx_EXPR_LIST (VOIDmode,
4074 gen_rtx_USE (VOIDmode, slot),
4075 call_fusage);
4076 if (must_copy)
4077 call_fusage = gen_rtx_EXPR_LIST (VOIDmode,
4078 gen_rtx_CLOBBER (VOIDmode,
4079 slot),
4080 call_fusage);
4081
4082 mode = Pmode;
4083 val = force_operand (XEXP (slot, 0), NULL_RTX);
4084 }
4085
4086 mode = promote_function_mode (NULL_TREE, mode, &unsigned_p, NULL_TREE, 0);
4087 argvec[count].mode = mode;
4088 argvec[count].value = convert_modes (mode, GET_MODE (val), val, unsigned_p);
4089 argvec[count].reg = targetm.calls.function_arg (args_so_far, mode,
4090 NULL_TREE, true);
4091
4092 argvec[count].partial
4093 = targetm.calls.arg_partial_bytes (args_so_far, mode, NULL_TREE, 1);
4094
4095 if (argvec[count].reg == 0
4096 || argvec[count].partial != 0
4097 || reg_parm_stack_space > 0)
4098 {
4099 locate_and_pad_parm (mode, NULL_TREE,
4100 #ifdef STACK_PARMS_IN_REG_PARM_AREA
4101 1,
4102 #else
4103 argvec[count].reg != 0,
4104 #endif
4105 reg_parm_stack_space, argvec[count].partial,
4106 NULL_TREE, &args_size, &argvec[count].locate);
4107 args_size.constant += argvec[count].locate.size.constant;
4108 gcc_assert (!argvec[count].locate.size.var);
4109 }
4110 #ifdef BLOCK_REG_PADDING
4111 else
4112 /* The argument is passed entirely in registers. See at which
4113 end it should be padded. */
4114 argvec[count].locate.where_pad =
4115 BLOCK_REG_PADDING (mode, NULL_TREE,
4116 GET_MODE_SIZE (mode) <= UNITS_PER_WORD);
4117 #endif
4118
4119 targetm.calls.function_arg_advance (args_so_far, mode, (tree) 0, true);
4120 }
4121
4122 /* If this machine requires an external definition for library
4123 functions, write one out. */
4124 assemble_external_libcall (fun);
4125
4126 original_args_size = args_size;
4127 args_size.constant = (((args_size.constant
4128 + stack_pointer_delta
4129 + STACK_BYTES - 1)
4130 / STACK_BYTES
4131 * STACK_BYTES)
4132 - stack_pointer_delta);
4133
4134 args_size.constant = MAX (args_size.constant,
4135 reg_parm_stack_space);
4136
4137 if (! OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl))))
4138 args_size.constant -= reg_parm_stack_space;
4139
4140 if (args_size.constant > crtl->outgoing_args_size)
4141 crtl->outgoing_args_size = args_size.constant;
4142
4143 if (flag_stack_usage_info && !ACCUMULATE_OUTGOING_ARGS)
4144 {
4145 int pushed = args_size.constant + pending_stack_adjust;
4146 if (pushed > current_function_pushed_stack_size)
4147 current_function_pushed_stack_size = pushed;
4148 }
4149
4150 if (ACCUMULATE_OUTGOING_ARGS)
4151 {
4152 /* Since the stack pointer will never be pushed, it is possible for
4153 the evaluation of a parm to clobber something we have already
4154 written to the stack. Since most function calls on RISC machines
4155 do not use the stack, this is uncommon, but must work correctly.
4156
4157 Therefore, we save any area of the stack that was already written
4158 and that we are using. Here we set up to do this by making a new
4159 stack usage map from the old one.
4160
4161 Another approach might be to try to reorder the argument
4162 evaluations to avoid this conflicting stack usage. */
4163
4164 needed = args_size.constant;
4165
4166 /* Since we will be writing into the entire argument area, the
4167 map must be allocated for its entire size, not just the part that
4168 is the responsibility of the caller. */
4169 if (! OUTGOING_REG_PARM_STACK_SPACE ((!fndecl ? fntype : TREE_TYPE (fndecl))))
4170 needed += reg_parm_stack_space;
4171
4172 if (ARGS_GROW_DOWNWARD)
4173 highest_outgoing_arg_in_use = MAX (initial_highest_arg_in_use,
4174 needed + 1);
4175 else
4176 highest_outgoing_arg_in_use = MAX (initial_highest_arg_in_use, needed);
4177
4178 stack_usage_map_buf = XNEWVEC (char, highest_outgoing_arg_in_use);
4179 stack_usage_map = stack_usage_map_buf;
4180
4181 if (initial_highest_arg_in_use)
4182 memcpy (stack_usage_map, initial_stack_usage_map,
4183 initial_highest_arg_in_use);
4184
4185 if (initial_highest_arg_in_use != highest_outgoing_arg_in_use)
4186 memset (&stack_usage_map[initial_highest_arg_in_use], 0,
4187 highest_outgoing_arg_in_use - initial_highest_arg_in_use);
4188 needed = 0;
4189
4190 /* We must be careful to use virtual regs before they're instantiated,
4191 and real regs afterwards. Loop optimization, for example, can create
4192 new libcalls after we've instantiated the virtual regs, and if we
4193 use virtuals anyway, they won't match the rtl patterns. */
4194
4195 if (virtuals_instantiated)
4196 argblock = plus_constant (Pmode, stack_pointer_rtx,
4197 STACK_POINTER_OFFSET);
4198 else
4199 argblock = virtual_outgoing_args_rtx;
4200 }
4201 else
4202 {
4203 if (!PUSH_ARGS)
4204 argblock = push_block (GEN_INT (args_size.constant), 0, 0);
4205 }
4206
4207 /* We push args individually in reverse order, perform stack alignment
4208 before the first push (the last arg). */
4209 if (argblock == 0)
4210 anti_adjust_stack (GEN_INT (args_size.constant
4211 - original_args_size.constant));
4212
4213 argnum = nargs - 1;
4214
4215 #ifdef REG_PARM_STACK_SPACE
4216 if (ACCUMULATE_OUTGOING_ARGS)
4217 {
4218 /* The argument list is the property of the called routine and it
4219 may clobber it. If the fixed area has been used for previous
4220 parameters, we must save and restore it. */
4221 save_area = save_fixed_argument_area (reg_parm_stack_space, argblock,
4222 &low_to_save, &high_to_save);
4223 }
4224 #endif
4225
4226 /* When expanding a normal call, args are stored in push order,
4227 which is the reverse of what we have here. */
4228 bool any_regs = false;
4229 for (int i = nargs; i-- > 0; )
4230 if (argvec[i].reg != NULL_RTX)
4231 {
4232 targetm.calls.call_args (argvec[i].reg, NULL_TREE);
4233 any_regs = true;
4234 }
4235 if (!any_regs)
4236 targetm.calls.call_args (pc_rtx, NULL_TREE);
4237
4238 /* Push the args that need to be pushed. */
4239
4240 have_push_fusage = false;
4241
4242 /* ARGNUM indexes the ARGVEC array in the order in which the arguments
4243 are to be pushed. */
4244 for (count = 0; count < nargs; count++, argnum--)
4245 {
4246 machine_mode mode = argvec[argnum].mode;
4247 rtx val = argvec[argnum].value;
4248 rtx reg = argvec[argnum].reg;
4249 int partial = argvec[argnum].partial;
4250 unsigned int parm_align = argvec[argnum].locate.boundary;
4251 int lower_bound = 0, upper_bound = 0, i;
4252
4253 if (! (reg != 0 && partial == 0))
4254 {
4255 rtx use;
4256
4257 if (ACCUMULATE_OUTGOING_ARGS)
4258 {
4259 /* If this is being stored into a pre-allocated, fixed-size,
4260 stack area, save any previous data at that location. */
4261
4262 if (ARGS_GROW_DOWNWARD)
4263 {
4264 /* stack_slot is negative, but we want to index stack_usage_map
4265 with positive values. */
4266 upper_bound = -argvec[argnum].locate.slot_offset.constant + 1;
4267 lower_bound = upper_bound - argvec[argnum].locate.size.constant;
4268 }
4269 else
4270 {
4271 lower_bound = argvec[argnum].locate.slot_offset.constant;
4272 upper_bound = lower_bound + argvec[argnum].locate.size.constant;
4273 }
4274
4275 i = lower_bound;
4276 /* Don't worry about things in the fixed argument area;
4277 it has already been saved. */
4278 if (i < reg_parm_stack_space)
4279 i = reg_parm_stack_space;
4280 while (i < upper_bound && stack_usage_map[i] == 0)
4281 i++;
4282
4283 if (i < upper_bound)
4284 {
4285 /* We need to make a save area. */
4286 unsigned int size
4287 = argvec[argnum].locate.size.constant * BITS_PER_UNIT;
4288 machine_mode save_mode
4289 = mode_for_size (size, MODE_INT, 1);
4290 rtx adr
4291 = plus_constant (Pmode, argblock,
4292 argvec[argnum].locate.offset.constant);
4293 rtx stack_area
4294 = gen_rtx_MEM (save_mode, memory_address (save_mode, adr));
4295
4296 if (save_mode == BLKmode)
4297 {
4298 argvec[argnum].save_area
4299 = assign_stack_temp (BLKmode,
4300 argvec[argnum].locate.size.constant
4301 );
4302
4303 emit_block_move (validize_mem
4304 (copy_rtx (argvec[argnum].save_area)),
4305 stack_area,
4306 GEN_INT (argvec[argnum].locate.size.constant),
4307 BLOCK_OP_CALL_PARM);
4308 }
4309 else
4310 {
4311 argvec[argnum].save_area = gen_reg_rtx (save_mode);
4312
4313 emit_move_insn (argvec[argnum].save_area, stack_area);
4314 }
4315 }
4316 }
4317
4318 emit_push_insn (val, mode, NULL_TREE, NULL_RTX, parm_align,
4319 partial, reg, 0, argblock,
4320 GEN_INT (argvec[argnum].locate.offset.constant),
4321 reg_parm_stack_space,
4322 ARGS_SIZE_RTX (argvec[argnum].locate.alignment_pad), false);
4323
4324 /* Now mark the segment we just used. */
4325 if (ACCUMULATE_OUTGOING_ARGS)
4326 for (i = lower_bound; i < upper_bound; i++)
4327 stack_usage_map[i] = 1;
4328
4329 NO_DEFER_POP;
4330
4331 /* Indicate argument access so that alias.c knows that these
4332 values are live. */
4333 if (argblock)
4334 use = plus_constant (Pmode, argblock,
4335 argvec[argnum].locate.offset.constant);
4336 else if (have_push_fusage)
4337 continue;
4338 else
4339 {
4340 /* When arguments are pushed, trying to tell alias.c where
4341 exactly this argument is won't work, because the
4342 auto-increment causes confusion. So we merely indicate
4343 that we access something with a known mode somewhere on
4344 the stack. */
4345 use = gen_rtx_PLUS (Pmode, stack_pointer_rtx,
4346 gen_rtx_SCRATCH (Pmode));
4347 have_push_fusage = true;
4348 }
4349 use = gen_rtx_MEM (argvec[argnum].mode, use);
4350 use = gen_rtx_USE (VOIDmode, use);
4351 call_fusage = gen_rtx_EXPR_LIST (VOIDmode, use, call_fusage);
4352 }
4353 }
4354
4355 argnum = nargs - 1;
4356
4357 fun = prepare_call_address (NULL, fun, NULL, &call_fusage, 0, 0);
4358
4359 /* Now load any reg parms into their regs. */
4360
4361 /* ARGNUM indexes the ARGVEC array in the order in which the arguments
4362 are to be pushed. */
4363 for (count = 0; count < nargs; count++, argnum--)
4364 {
4365 machine_mode mode = argvec[argnum].mode;
4366 rtx val = argvec[argnum].value;
4367 rtx reg = argvec[argnum].reg;
4368 int partial = argvec[argnum].partial;
4369 #ifdef BLOCK_REG_PADDING
4370 int size = 0;
4371 #endif
4372
4373 /* Handle calls that pass values in multiple non-contiguous
4374 locations. The PA64 has examples of this for library calls. */
4375 if (reg != 0 && GET_CODE (reg) == PARALLEL)
4376 emit_group_load (reg, val, NULL_TREE, GET_MODE_SIZE (mode));
4377 else if (reg != 0 && partial == 0)
4378 {
4379 emit_move_insn (reg, val);
4380 #ifdef BLOCK_REG_PADDING
4381 size = GET_MODE_SIZE (argvec[argnum].mode);
4382
4383 /* Copied from load_register_parameters. */
4384
4385 /* Handle case where we have a value that needs shifting
4386 up to the msb. eg. a QImode value and we're padding
4387 upward on a BYTES_BIG_ENDIAN machine. */
4388 if (size < UNITS_PER_WORD
4389 && (argvec[argnum].locate.where_pad
4390 == (BYTES_BIG_ENDIAN ? upward : downward)))
4391 {
4392 rtx x;
4393 int shift = (UNITS_PER_WORD - size) * BITS_PER_UNIT;
4394
4395 /* Assigning REG here rather than a temp makes CALL_FUSAGE
4396 report the whole reg as used. Strictly speaking, the
4397 call only uses SIZE bytes at the msb end, but it doesn't
4398 seem worth generating rtl to say that. */
4399 reg = gen_rtx_REG (word_mode, REGNO (reg));
4400 x = expand_shift (LSHIFT_EXPR, word_mode, reg, shift, reg, 1);
4401 if (x != reg)
4402 emit_move_insn (reg, x);
4403 }
4404 #endif
4405 }
4406
4407 NO_DEFER_POP;
4408 }
4409
4410 /* Any regs containing parms remain in use through the call. */
4411 for (count = 0; count < nargs; count++)
4412 {
4413 rtx reg = argvec[count].reg;
4414 if (reg != 0 && GET_CODE (reg) == PARALLEL)
4415 use_group_regs (&call_fusage, reg);
4416 else if (reg != 0)
4417 {
4418 int partial = argvec[count].partial;
4419 if (partial)
4420 {
4421 int nregs;
4422 gcc_assert (partial % UNITS_PER_WORD == 0);
4423 nregs = partial / UNITS_PER_WORD;
4424 use_regs (&call_fusage, REGNO (reg), nregs);
4425 }
4426 else
4427 use_reg (&call_fusage, reg);
4428 }
4429 }
4430
4431 /* Pass the function the address in which to return a structure value. */
4432 if (mem_value != 0 && struct_value != 0 && ! pcc_struct_value)
4433 {
4434 emit_move_insn (struct_value,
4435 force_reg (Pmode,
4436 force_operand (XEXP (mem_value, 0),
4437 NULL_RTX)));
4438 if (REG_P (struct_value))
4439 use_reg (&call_fusage, struct_value);
4440 }
4441
4442 /* Don't allow popping to be deferred, since then
4443 cse'ing of library calls could delete a call and leave the pop. */
4444 NO_DEFER_POP;
4445 valreg = (mem_value == 0 && outmode != VOIDmode
4446 ? hard_libcall_value (outmode, orgfun) : NULL_RTX);
4447
4448 /* Stack must be properly aligned now. */
4449 gcc_assert (!(stack_pointer_delta
4450 & (PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT - 1)));
4451
4452 before_call = get_last_insn ();
4453
4454 /* We pass the old value of inhibit_defer_pop + 1 to emit_call_1, which
4455 will set inhibit_defer_pop to that value. */
4456 /* The return type is needed to decide how many bytes the function pops.
4457 Signedness plays no role in that, so for simplicity, we pretend it's
4458 always signed. We also assume that the list of arguments passed has
4459 no impact, so we pretend it is unknown. */
4460
4461 emit_call_1 (fun, NULL,
4462 get_identifier (XSTR (orgfun, 0)),
4463 build_function_type (tfom, NULL_TREE),
4464 original_args_size.constant, args_size.constant,
4465 struct_value_size,
4466 targetm.calls.function_arg (args_so_far,
4467 VOIDmode, void_type_node, true),
4468 valreg,
4469 old_inhibit_defer_pop + 1, call_fusage, flags, args_so_far);
4470
4471 if (flag_ipa_ra)
4472 {
4473 rtx datum = orgfun;
4474 gcc_assert (GET_CODE (datum) == SYMBOL_REF);
4475 rtx_call_insn *last = last_call_insn ();
4476 add_reg_note (last, REG_CALL_DECL, datum);
4477 }
4478
4479 /* Right-shift returned value if necessary. */
4480 if (!pcc_struct_value
4481 && TYPE_MODE (tfom) != BLKmode
4482 && targetm.calls.return_in_msb (tfom))
4483 {
4484 shift_return_value (TYPE_MODE (tfom), false, valreg);
4485 valreg = gen_rtx_REG (TYPE_MODE (tfom), REGNO (valreg));
4486 }
4487
4488 targetm.calls.end_call_args ();
4489
4490 /* For calls to `setjmp', etc., inform function.c:setjmp_warnings
4491 that it should complain if nonvolatile values are live. For
4492 functions that cannot return, inform flow that control does not
4493 fall through. */
4494 if (flags & ECF_NORETURN)
4495 {
4496 /* The barrier note must be emitted
4497 immediately after the CALL_INSN. Some ports emit more than
4498 just a CALL_INSN above, so we must search for it here. */
4499 rtx_insn *last = get_last_insn ();
4500 while (!CALL_P (last))
4501 {
4502 last = PREV_INSN (last);
4503 /* There was no CALL_INSN? */
4504 gcc_assert (last != before_call);
4505 }
4506
4507 emit_barrier_after (last);
4508 }
4509
4510 /* Consider that "regular" libcalls, i.e. all of them except for LCT_THROW
4511 and LCT_RETURNS_TWICE, cannot perform non-local gotos. */
4512 if (flags & ECF_NOTHROW)
4513 {
4514 rtx_insn *last = get_last_insn ();
4515 while (!CALL_P (last))
4516 {
4517 last = PREV_INSN (last);
4518 /* There was no CALL_INSN? */
4519 gcc_assert (last != before_call);
4520 }
4521
4522 make_reg_eh_region_note_nothrow_nononlocal (last);
4523 }
4524
4525 /* Now restore inhibit_defer_pop to its actual original value. */
4526 OK_DEFER_POP;
4527
4528 pop_temp_slots ();
4529
4530 /* Copy the value to the right place. */
4531 if (outmode != VOIDmode && retval)
4532 {
4533 if (mem_value)
4534 {
4535 if (value == 0)
4536 value = mem_value;
4537 if (value != mem_value)
4538 emit_move_insn (value, mem_value);
4539 }
4540 else if (GET_CODE (valreg) == PARALLEL)
4541 {
4542 if (value == 0)
4543 value = gen_reg_rtx (outmode);
4544 emit_group_store (value, valreg, NULL_TREE, GET_MODE_SIZE (outmode));
4545 }
4546 else
4547 {
4548 /* Convert to the proper mode if a promotion has been active. */
4549 if (GET_MODE (valreg) != outmode)
4550 {
4551 int unsignedp = TYPE_UNSIGNED (tfom);
4552
4553 gcc_assert (promote_function_mode (tfom, outmode, &unsignedp,
4554 fndecl ? TREE_TYPE (fndecl) : fntype, 1)
4555 == GET_MODE (valreg));
4556 valreg = convert_modes (outmode, GET_MODE (valreg), valreg, 0);
4557 }
4558
4559 if (value != 0)
4560 emit_move_insn (value, valreg);
4561 else
4562 value = valreg;
4563 }
4564 }
4565
4566 if (ACCUMULATE_OUTGOING_ARGS)
4567 {
4568 #ifdef REG_PARM_STACK_SPACE
4569 if (save_area)
4570 restore_fixed_argument_area (save_area, argblock,
4571 high_to_save, low_to_save);
4572 #endif
4573
4574 /* If we saved any argument areas, restore them. */
4575 for (count = 0; count < nargs; count++)
4576 if (argvec[count].save_area)
4577 {
4578 machine_mode save_mode = GET_MODE (argvec[count].save_area);
4579 rtx adr = plus_constant (Pmode, argblock,
4580 argvec[count].locate.offset.constant);
4581 rtx stack_area = gen_rtx_MEM (save_mode,
4582 memory_address (save_mode, adr));
4583
4584 if (save_mode == BLKmode)
4585 emit_block_move (stack_area,
4586 validize_mem
4587 (copy_rtx (argvec[count].save_area)),
4588 GEN_INT (argvec[count].locate.size.constant),
4589 BLOCK_OP_CALL_PARM);
4590 else
4591 emit_move_insn (stack_area, argvec[count].save_area);
4592 }
4593
4594 highest_outgoing_arg_in_use = initial_highest_arg_in_use;
4595 stack_usage_map = initial_stack_usage_map;
4596 }
4597
4598 free (stack_usage_map_buf);
4599
4600 return value;
4601
4602 }
4603 \f
4604 /* Output a library call to function FUN (a SYMBOL_REF rtx)
4605 (emitting the queue unless NO_QUEUE is nonzero),
4606 for a value of mode OUTMODE,
4607 with NARGS different arguments, passed as alternating rtx values
4608 and machine_modes to convert them to.
4609
4610 FN_TYPE should be LCT_NORMAL for `normal' calls, LCT_CONST for
4611 `const' calls, LCT_PURE for `pure' calls, or other LCT_ value for
4612 other types of library calls. */
4613
4614 void
4615 emit_library_call (rtx orgfun, enum libcall_type fn_type,
4616 machine_mode outmode, int nargs, ...)
4617 {
4618 va_list p;
4619
4620 va_start (p, nargs);
4621 emit_library_call_value_1 (0, orgfun, NULL_RTX, fn_type, outmode, nargs, p);
4622 va_end (p);
4623 }
4624 \f
4625 /* Like emit_library_call except that an extra argument, VALUE,
4626 comes second and says where to store the result.
4627 (If VALUE is zero, this function chooses a convenient way
4628 to return the value.
4629
4630 This function returns an rtx for where the value is to be found.
4631 If VALUE is nonzero, VALUE is returned. */
4632
4633 rtx
4634 emit_library_call_value (rtx orgfun, rtx value,
4635 enum libcall_type fn_type,
4636 machine_mode outmode, int nargs, ...)
4637 {
4638 rtx result;
4639 va_list p;
4640
4641 va_start (p, nargs);
4642 result = emit_library_call_value_1 (1, orgfun, value, fn_type, outmode,
4643 nargs, p);
4644 va_end (p);
4645
4646 return result;
4647 }
4648 \f
4649
4650 /* Store pointer bounds argument ARG into Bounds Table entry
4651 associated with PARM. */
4652 static void
4653 store_bounds (struct arg_data *arg, struct arg_data *parm)
4654 {
4655 rtx slot = NULL, ptr = NULL, addr = NULL;
4656
4657 /* We may pass bounds not associated with any pointer. */
4658 if (!parm)
4659 {
4660 gcc_assert (arg->special_slot);
4661 slot = arg->special_slot;
4662 ptr = const0_rtx;
4663 }
4664 /* Find pointer associated with bounds and where it is
4665 passed. */
4666 else
4667 {
4668 if (!parm->reg)
4669 {
4670 gcc_assert (!arg->special_slot);
4671
4672 addr = adjust_address (parm->stack, Pmode, arg->pointer_offset);
4673 }
4674 else if (REG_P (parm->reg))
4675 {
4676 gcc_assert (arg->special_slot);
4677 slot = arg->special_slot;
4678
4679 if (MEM_P (parm->value))
4680 addr = adjust_address (parm->value, Pmode, arg->pointer_offset);
4681 else if (REG_P (parm->value))
4682 ptr = gen_rtx_SUBREG (Pmode, parm->value, arg->pointer_offset);
4683 else
4684 {
4685 gcc_assert (!arg->pointer_offset);
4686 ptr = parm->value;
4687 }
4688 }
4689 else
4690 {
4691 gcc_assert (GET_CODE (parm->reg) == PARALLEL);
4692
4693 gcc_assert (arg->special_slot);
4694 slot = arg->special_slot;
4695
4696 if (parm->parallel_value)
4697 ptr = chkp_get_value_with_offs (parm->parallel_value,
4698 GEN_INT (arg->pointer_offset));
4699 else
4700 gcc_unreachable ();
4701 }
4702 }
4703
4704 /* Expand bounds. */
4705 if (!arg->value)
4706 arg->value = expand_normal (arg->tree_value);
4707
4708 targetm.calls.store_bounds_for_arg (ptr, addr, arg->value, slot);
4709 }
4710
4711 /* Store a single argument for a function call
4712 into the register or memory area where it must be passed.
4713 *ARG describes the argument value and where to pass it.
4714
4715 ARGBLOCK is the address of the stack-block for all the arguments,
4716 or 0 on a machine where arguments are pushed individually.
4717
4718 MAY_BE_ALLOCA nonzero says this could be a call to `alloca'
4719 so must be careful about how the stack is used.
4720
4721 VARIABLE_SIZE nonzero says that this was a variable-sized outgoing
4722 argument stack. This is used if ACCUMULATE_OUTGOING_ARGS to indicate
4723 that we need not worry about saving and restoring the stack.
4724
4725 FNDECL is the declaration of the function we are calling.
4726
4727 Return nonzero if this arg should cause sibcall failure,
4728 zero otherwise. */
4729
4730 static int
4731 store_one_arg (struct arg_data *arg, rtx argblock, int flags,
4732 int variable_size ATTRIBUTE_UNUSED, int reg_parm_stack_space)
4733 {
4734 tree pval = arg->tree_value;
4735 rtx reg = 0;
4736 int partial = 0;
4737 int used = 0;
4738 int i, lower_bound = 0, upper_bound = 0;
4739 int sibcall_failure = 0;
4740
4741 if (TREE_CODE (pval) == ERROR_MARK)
4742 return 1;
4743
4744 /* Push a new temporary level for any temporaries we make for
4745 this argument. */
4746 push_temp_slots ();
4747
4748 if (ACCUMULATE_OUTGOING_ARGS && !(flags & ECF_SIBCALL))
4749 {
4750 /* If this is being stored into a pre-allocated, fixed-size, stack area,
4751 save any previous data at that location. */
4752 if (argblock && ! variable_size && arg->stack)
4753 {
4754 if (ARGS_GROW_DOWNWARD)
4755 {
4756 /* stack_slot is negative, but we want to index stack_usage_map
4757 with positive values. */
4758 if (GET_CODE (XEXP (arg->stack_slot, 0)) == PLUS)
4759 upper_bound = -INTVAL (XEXP (XEXP (arg->stack_slot, 0), 1)) + 1;
4760 else
4761 upper_bound = 0;
4762
4763 lower_bound = upper_bound - arg->locate.size.constant;
4764 }
4765 else
4766 {
4767 if (GET_CODE (XEXP (arg->stack_slot, 0)) == PLUS)
4768 lower_bound = INTVAL (XEXP (XEXP (arg->stack_slot, 0), 1));
4769 else
4770 lower_bound = 0;
4771
4772 upper_bound = lower_bound + arg->locate.size.constant;
4773 }
4774
4775 i = lower_bound;
4776 /* Don't worry about things in the fixed argument area;
4777 it has already been saved. */
4778 if (i < reg_parm_stack_space)
4779 i = reg_parm_stack_space;
4780 while (i < upper_bound && stack_usage_map[i] == 0)
4781 i++;
4782
4783 if (i < upper_bound)
4784 {
4785 /* We need to make a save area. */
4786 unsigned int size = arg->locate.size.constant * BITS_PER_UNIT;
4787 machine_mode save_mode = mode_for_size (size, MODE_INT, 1);
4788 rtx adr = memory_address (save_mode, XEXP (arg->stack_slot, 0));
4789 rtx stack_area = gen_rtx_MEM (save_mode, adr);
4790
4791 if (save_mode == BLKmode)
4792 {
4793 arg->save_area
4794 = assign_temp (TREE_TYPE (arg->tree_value), 1, 1);
4795 preserve_temp_slots (arg->save_area);
4796 emit_block_move (validize_mem (copy_rtx (arg->save_area)),
4797 stack_area,
4798 GEN_INT (arg->locate.size.constant),
4799 BLOCK_OP_CALL_PARM);
4800 }
4801 else
4802 {
4803 arg->save_area = gen_reg_rtx (save_mode);
4804 emit_move_insn (arg->save_area, stack_area);
4805 }
4806 }
4807 }
4808 }
4809
4810 /* If this isn't going to be placed on both the stack and in registers,
4811 set up the register and number of words. */
4812 if (! arg->pass_on_stack)
4813 {
4814 if (flags & ECF_SIBCALL)
4815 reg = arg->tail_call_reg;
4816 else
4817 reg = arg->reg;
4818 partial = arg->partial;
4819 }
4820
4821 /* Being passed entirely in a register. We shouldn't be called in
4822 this case. */
4823 gcc_assert (reg == 0 || partial != 0);
4824
4825 /* If this arg needs special alignment, don't load the registers
4826 here. */
4827 if (arg->n_aligned_regs != 0)
4828 reg = 0;
4829
4830 /* If this is being passed partially in a register, we can't evaluate
4831 it directly into its stack slot. Otherwise, we can. */
4832 if (arg->value == 0)
4833 {
4834 /* stack_arg_under_construction is nonzero if a function argument is
4835 being evaluated directly into the outgoing argument list and
4836 expand_call must take special action to preserve the argument list
4837 if it is called recursively.
4838
4839 For scalar function arguments stack_usage_map is sufficient to
4840 determine which stack slots must be saved and restored. Scalar
4841 arguments in general have pass_on_stack == 0.
4842
4843 If this argument is initialized by a function which takes the
4844 address of the argument (a C++ constructor or a C function
4845 returning a BLKmode structure), then stack_usage_map is
4846 insufficient and expand_call must push the stack around the
4847 function call. Such arguments have pass_on_stack == 1.
4848
4849 Note that it is always safe to set stack_arg_under_construction,
4850 but this generates suboptimal code if set when not needed. */
4851
4852 if (arg->pass_on_stack)
4853 stack_arg_under_construction++;
4854
4855 arg->value = expand_expr (pval,
4856 (partial
4857 || TYPE_MODE (TREE_TYPE (pval)) != arg->mode)
4858 ? NULL_RTX : arg->stack,
4859 VOIDmode, EXPAND_STACK_PARM);
4860
4861 /* If we are promoting object (or for any other reason) the mode
4862 doesn't agree, convert the mode. */
4863
4864 if (arg->mode != TYPE_MODE (TREE_TYPE (pval)))
4865 arg->value = convert_modes (arg->mode, TYPE_MODE (TREE_TYPE (pval)),
4866 arg->value, arg->unsignedp);
4867
4868 if (arg->pass_on_stack)
4869 stack_arg_under_construction--;
4870 }
4871
4872 /* Check for overlap with already clobbered argument area. */
4873 if ((flags & ECF_SIBCALL)
4874 && MEM_P (arg->value)
4875 && mem_overlaps_already_clobbered_arg_p (XEXP (arg->value, 0),
4876 arg->locate.size.constant))
4877 sibcall_failure = 1;
4878
4879 /* Don't allow anything left on stack from computation
4880 of argument to alloca. */
4881 if (flags & ECF_MAY_BE_ALLOCA)
4882 do_pending_stack_adjust ();
4883
4884 if (arg->value == arg->stack)
4885 /* If the value is already in the stack slot, we are done. */
4886 ;
4887 else if (arg->mode != BLKmode)
4888 {
4889 int size;
4890 unsigned int parm_align;
4891
4892 /* Argument is a scalar, not entirely passed in registers.
4893 (If part is passed in registers, arg->partial says how much
4894 and emit_push_insn will take care of putting it there.)
4895
4896 Push it, and if its size is less than the
4897 amount of space allocated to it,
4898 also bump stack pointer by the additional space.
4899 Note that in C the default argument promotions
4900 will prevent such mismatches. */
4901
4902 size = GET_MODE_SIZE (arg->mode);
4903 /* Compute how much space the push instruction will push.
4904 On many machines, pushing a byte will advance the stack
4905 pointer by a halfword. */
4906 #ifdef PUSH_ROUNDING
4907 size = PUSH_ROUNDING (size);
4908 #endif
4909 used = size;
4910
4911 /* Compute how much space the argument should get:
4912 round up to a multiple of the alignment for arguments. */
4913 if (none != FUNCTION_ARG_PADDING (arg->mode, TREE_TYPE (pval)))
4914 used = (((size + PARM_BOUNDARY / BITS_PER_UNIT - 1)
4915 / (PARM_BOUNDARY / BITS_PER_UNIT))
4916 * (PARM_BOUNDARY / BITS_PER_UNIT));
4917
4918 /* Compute the alignment of the pushed argument. */
4919 parm_align = arg->locate.boundary;
4920 if (FUNCTION_ARG_PADDING (arg->mode, TREE_TYPE (pval)) == downward)
4921 {
4922 int pad = used - size;
4923 if (pad)
4924 {
4925 unsigned int pad_align = (pad & -pad) * BITS_PER_UNIT;
4926 parm_align = MIN (parm_align, pad_align);
4927 }
4928 }
4929
4930 /* This isn't already where we want it on the stack, so put it there.
4931 This can either be done with push or copy insns. */
4932 if (!emit_push_insn (arg->value, arg->mode, TREE_TYPE (pval), NULL_RTX,
4933 parm_align, partial, reg, used - size, argblock,
4934 ARGS_SIZE_RTX (arg->locate.offset), reg_parm_stack_space,
4935 ARGS_SIZE_RTX (arg->locate.alignment_pad), true))
4936 sibcall_failure = 1;
4937
4938 /* Unless this is a partially-in-register argument, the argument is now
4939 in the stack. */
4940 if (partial == 0)
4941 arg->value = arg->stack;
4942 }
4943 else
4944 {
4945 /* BLKmode, at least partly to be pushed. */
4946
4947 unsigned int parm_align;
4948 int excess;
4949 rtx size_rtx;
4950
4951 /* Pushing a nonscalar.
4952 If part is passed in registers, PARTIAL says how much
4953 and emit_push_insn will take care of putting it there. */
4954
4955 /* Round its size up to a multiple
4956 of the allocation unit for arguments. */
4957
4958 if (arg->locate.size.var != 0)
4959 {
4960 excess = 0;
4961 size_rtx = ARGS_SIZE_RTX (arg->locate.size);
4962 }
4963 else
4964 {
4965 /* PUSH_ROUNDING has no effect on us, because emit_push_insn
4966 for BLKmode is careful to avoid it. */
4967 excess = (arg->locate.size.constant
4968 - int_size_in_bytes (TREE_TYPE (pval))
4969 + partial);
4970 size_rtx = expand_expr (size_in_bytes (TREE_TYPE (pval)),
4971 NULL_RTX, TYPE_MODE (sizetype),
4972 EXPAND_NORMAL);
4973 }
4974
4975 parm_align = arg->locate.boundary;
4976
4977 /* When an argument is padded down, the block is aligned to
4978 PARM_BOUNDARY, but the actual argument isn't. */
4979 if (FUNCTION_ARG_PADDING (arg->mode, TREE_TYPE (pval)) == downward)
4980 {
4981 if (arg->locate.size.var)
4982 parm_align = BITS_PER_UNIT;
4983 else if (excess)
4984 {
4985 unsigned int excess_align = (excess & -excess) * BITS_PER_UNIT;
4986 parm_align = MIN (parm_align, excess_align);
4987 }
4988 }
4989
4990 if ((flags & ECF_SIBCALL) && MEM_P (arg->value))
4991 {
4992 /* emit_push_insn might not work properly if arg->value and
4993 argblock + arg->locate.offset areas overlap. */
4994 rtx x = arg->value;
4995 int i = 0;
4996
4997 if (XEXP (x, 0) == crtl->args.internal_arg_pointer
4998 || (GET_CODE (XEXP (x, 0)) == PLUS
4999 && XEXP (XEXP (x, 0), 0) ==
5000 crtl->args.internal_arg_pointer
5001 && CONST_INT_P (XEXP (XEXP (x, 0), 1))))
5002 {
5003 if (XEXP (x, 0) != crtl->args.internal_arg_pointer)
5004 i = INTVAL (XEXP (XEXP (x, 0), 1));
5005
5006 /* expand_call should ensure this. */
5007 gcc_assert (!arg->locate.offset.var
5008 && arg->locate.size.var == 0
5009 && CONST_INT_P (size_rtx));
5010
5011 if (arg->locate.offset.constant > i)
5012 {
5013 if (arg->locate.offset.constant < i + INTVAL (size_rtx))
5014 sibcall_failure = 1;
5015 }
5016 else if (arg->locate.offset.constant < i)
5017 {
5018 /* Use arg->locate.size.constant instead of size_rtx
5019 because we only care about the part of the argument
5020 on the stack. */
5021 if (i < (arg->locate.offset.constant
5022 + arg->locate.size.constant))
5023 sibcall_failure = 1;
5024 }
5025 else
5026 {
5027 /* Even though they appear to be at the same location,
5028 if part of the outgoing argument is in registers,
5029 they aren't really at the same location. Check for
5030 this by making sure that the incoming size is the
5031 same as the outgoing size. */
5032 if (arg->locate.size.constant != INTVAL (size_rtx))
5033 sibcall_failure = 1;
5034 }
5035 }
5036 }
5037
5038 emit_push_insn (arg->value, arg->mode, TREE_TYPE (pval), size_rtx,
5039 parm_align, partial, reg, excess, argblock,
5040 ARGS_SIZE_RTX (arg->locate.offset), reg_parm_stack_space,
5041 ARGS_SIZE_RTX (arg->locate.alignment_pad), false);
5042
5043 /* Unless this is a partially-in-register argument, the argument is now
5044 in the stack.
5045
5046 ??? Unlike the case above, in which we want the actual
5047 address of the data, so that we can load it directly into a
5048 register, here we want the address of the stack slot, so that
5049 it's properly aligned for word-by-word copying or something
5050 like that. It's not clear that this is always correct. */
5051 if (partial == 0)
5052 arg->value = arg->stack_slot;
5053 }
5054
5055 if (arg->reg && GET_CODE (arg->reg) == PARALLEL)
5056 {
5057 tree type = TREE_TYPE (arg->tree_value);
5058 arg->parallel_value
5059 = emit_group_load_into_temps (arg->reg, arg->value, type,
5060 int_size_in_bytes (type));
5061 }
5062
5063 /* Mark all slots this store used. */
5064 if (ACCUMULATE_OUTGOING_ARGS && !(flags & ECF_SIBCALL)
5065 && argblock && ! variable_size && arg->stack)
5066 for (i = lower_bound; i < upper_bound; i++)
5067 stack_usage_map[i] = 1;
5068
5069 /* Once we have pushed something, pops can't safely
5070 be deferred during the rest of the arguments. */
5071 NO_DEFER_POP;
5072
5073 /* Free any temporary slots made in processing this argument. */
5074 pop_temp_slots ();
5075
5076 return sibcall_failure;
5077 }
5078
5079 /* Nonzero if we do not know how to pass TYPE solely in registers. */
5080
5081 bool
5082 must_pass_in_stack_var_size (machine_mode mode ATTRIBUTE_UNUSED,
5083 const_tree type)
5084 {
5085 if (!type)
5086 return false;
5087
5088 /* If the type has variable size... */
5089 if (TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)
5090 return true;
5091
5092 /* If the type is marked as addressable (it is required
5093 to be constructed into the stack)... */
5094 if (TREE_ADDRESSABLE (type))
5095 return true;
5096
5097 return false;
5098 }
5099
5100 /* Another version of the TARGET_MUST_PASS_IN_STACK hook. This one
5101 takes trailing padding of a structure into account. */
5102 /* ??? Should be able to merge these two by examining BLOCK_REG_PADDING. */
5103
5104 bool
5105 must_pass_in_stack_var_size_or_pad (machine_mode mode, const_tree type)
5106 {
5107 if (!type)
5108 return false;
5109
5110 /* If the type has variable size... */
5111 if (TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)
5112 return true;
5113
5114 /* If the type is marked as addressable (it is required
5115 to be constructed into the stack)... */
5116 if (TREE_ADDRESSABLE (type))
5117 return true;
5118
5119 /* If the padding and mode of the type is such that a copy into
5120 a register would put it into the wrong part of the register. */
5121 if (mode == BLKmode
5122 && int_size_in_bytes (type) % (PARM_BOUNDARY / BITS_PER_UNIT)
5123 && (FUNCTION_ARG_PADDING (mode, type)
5124 == (BYTES_BIG_ENDIAN ? upward : downward)))
5125 return true;
5126
5127 return false;
5128 }