(c-decl.c, [...]): Replace DECL_SAVED_INSNS with DECL_STRUCT_FUNCTION.
[gcc.git] / gcc / ada / utils.c
1 /****************************************************************************
2 * *
3 * GNAT COMPILER COMPONENTS *
4 * *
5 * U T I L S *
6 * *
7 * C Implementation File *
8 * *
9 * Copyright (C) 1992-2004, Free Software Foundation, Inc. *
10 * *
11 * GNAT is free software; you can redistribute it and/or modify it under *
12 * terms of the GNU General Public License as published by the Free Soft- *
13 * ware Foundation; either version 2, or (at your option) any later ver- *
14 * sion. GNAT is distributed in the hope that it will be useful, but WITH- *
15 * OUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY *
16 * or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License *
17 * for more details. You should have received a copy of the GNU General *
18 * Public License distributed with GNAT; see file COPYING. If not, write *
19 * to the Free Software Foundation, 59 Temple Place - Suite 330, Boston, *
20 * MA 02111-1307, USA. *
21 * *
22 * GNAT was originally developed by the GNAT team at New York University. *
23 * Extensive contributions were provided by Ada Core Technologies Inc. *
24 * *
25 ****************************************************************************/
26
27 #include "config.h"
28 #include "system.h"
29 #include "coretypes.h"
30 #include "tm.h"
31 #include "tree.h"
32 #include "flags.h"
33 #include "defaults.h"
34 #include "toplev.h"
35 #include "output.h"
36 #include "ggc.h"
37 #include "debug.h"
38 #include "convert.h"
39 #include "target.h"
40
41 #include "ada.h"
42 #include "types.h"
43 #include "atree.h"
44 #include "elists.h"
45 #include "namet.h"
46 #include "nlists.h"
47 #include "stringt.h"
48 #include "uintp.h"
49 #include "fe.h"
50 #include "sinfo.h"
51 #include "einfo.h"
52 #include "ada-tree.h"
53 #include "gigi.h"
54
55 #ifndef MAX_FIXED_MODE_SIZE
56 #define MAX_FIXED_MODE_SIZE GET_MODE_BITSIZE (DImode)
57 #endif
58
59 #ifndef MAX_BITS_PER_WORD
60 #define MAX_BITS_PER_WORD BITS_PER_WORD
61 #endif
62
63 /* If nonzero, pretend we are allocating at global level. */
64 int force_global;
65
66 /* Tree nodes for the various types and decls we create. */
67 tree gnat_std_decls[(int) ADT_LAST];
68
69 /* Functions to call for each of the possible raise reasons. */
70 tree gnat_raise_decls[(int) LAST_REASON_CODE + 1];
71
72 /* Associates a GNAT tree node to a GCC tree node. It is used in
73 `save_gnu_tree', `get_gnu_tree' and `present_gnu_tree'. See documentation
74 of `save_gnu_tree' for more info. */
75 static GTY((length ("max_gnat_nodes"))) tree *associate_gnat_to_gnu;
76
77 /* This listhead is used to record any global objects that need elaboration.
78 TREE_PURPOSE is the variable to be elaborated and TREE_VALUE is the
79 initial value to assign. */
80
81 static GTY(()) tree pending_elaborations;
82
83 /* This stack allows us to momentarily switch to generating elaboration
84 lists for an inner context. */
85
86 struct e_stack GTY(()) {
87 struct e_stack *next;
88 tree elab_list;
89 };
90 static GTY(()) struct e_stack *elist_stack;
91
92 /* This variable keeps a table for types for each precision so that we only
93 allocate each of them once. Signed and unsigned types are kept separate.
94
95 Note that these types are only used when fold-const requests something
96 special. Perhaps we should NOT share these types; we'll see how it
97 goes later. */
98 static GTY(()) tree signed_and_unsigned_types[2 * MAX_BITS_PER_WORD + 1][2];
99
100 /* Likewise for float types, but record these by mode. */
101 static GTY(()) tree float_types[NUM_MACHINE_MODES];
102
103 /* For each binding contour we allocate a binding_level structure which records
104 the entities defined or declared in that contour. Contours include:
105
106 the global one
107 one for each subprogram definition
108 one for each compound statement (declare block)
109
110 Binding contours are used to create GCC tree BLOCK nodes. */
111
112 struct binding_level GTY(())
113 {
114 /* A chain of ..._DECL nodes for all variables, constants, functions,
115 parameters and type declarations. These ..._DECL nodes are chained
116 through the TREE_CHAIN field. Note that these ..._DECL nodes are stored
117 in the reverse of the order supplied to be compatible with the
118 back-end. */
119 tree names;
120 /* For each level (except the global one), a chain of BLOCK nodes for all
121 the levels that were entered and exited one level down from this one. */
122 tree blocks;
123 /* The BLOCK node for this level, if one has been preallocated.
124 If 0, the BLOCK is allocated (if needed) when the level is popped. */
125 tree this_block;
126 /* The binding level containing this one (the enclosing binding level). */
127 struct binding_level *level_chain;
128 };
129
130 /* The binding level currently in effect. */
131 static GTY(()) struct binding_level *current_binding_level;
132
133 /* A chain of binding_level structures awaiting reuse. */
134 static GTY((deletable (""))) struct binding_level *free_binding_level;
135
136 /* The outermost binding level. This binding level is created when the
137 compiler is started and it will exist through the entire compilation. */
138 static struct binding_level *global_binding_level;
139
140 /* Binding level structures are initialized by copying this one. */
141 static struct binding_level clear_binding_level = {NULL, NULL, NULL, NULL};
142
143 struct language_function GTY(())
144 {
145 int unused;
146 };
147
148 static tree merge_sizes (tree, tree, tree, int, int);
149 static tree compute_related_constant (tree, tree);
150 static tree split_plus (tree, tree *);
151 static int value_zerop (tree);
152 static tree float_type_for_precision (int, enum machine_mode);
153 static tree convert_to_fat_pointer (tree, tree);
154 static tree convert_to_thin_pointer (tree, tree);
155 static tree make_descriptor_field (const char *,tree, tree, tree);
156 static int value_factor_p (tree, int);
157 static int potential_alignment_gap (tree, tree, tree);
158 \f
159 /* Initialize the association of GNAT nodes to GCC trees. */
160
161 void
162 init_gnat_to_gnu (void)
163 {
164 associate_gnat_to_gnu
165 = (tree *) ggc_alloc_cleared (max_gnat_nodes * sizeof (tree));
166
167 pending_elaborations = build_tree_list (NULL_TREE, NULL_TREE);
168 }
169
170 /* GNAT_ENTITY is a GNAT tree node for an entity. GNU_DECL is the GCC tree
171 which is to be associated with GNAT_ENTITY. Such GCC tree node is always
172 a ..._DECL node. If NO_CHECK is nonzero, the latter check is suppressed.
173
174 If GNU_DECL is zero, a previous association is to be reset. */
175
176 void
177 save_gnu_tree (Entity_Id gnat_entity, tree gnu_decl, int no_check)
178 {
179 /* Check that GNAT_ENTITY is not already defined and that it is being set
180 to something which is a decl. Raise gigi 401 if not. Usually, this
181 means GNAT_ENTITY is defined twice, but occasionally is due to some
182 Gigi problem. */
183 if (gnu_decl
184 && (associate_gnat_to_gnu[gnat_entity - First_Node_Id]
185 || (! no_check && ! DECL_P (gnu_decl))))
186 gigi_abort (401);
187
188 associate_gnat_to_gnu[gnat_entity - First_Node_Id] = gnu_decl;
189 }
190
191 /* GNAT_ENTITY is a GNAT tree node for a defining identifier.
192 Return the ..._DECL node that was associated with it. If there is no tree
193 node associated with GNAT_ENTITY, abort.
194
195 In some cases, such as delayed elaboration or expressions that need to
196 be elaborated only once, GNAT_ENTITY is really not an entity. */
197
198 tree
199 get_gnu_tree (Entity_Id gnat_entity)
200 {
201 if (! associate_gnat_to_gnu[gnat_entity - First_Node_Id])
202 gigi_abort (402);
203
204 return associate_gnat_to_gnu[gnat_entity - First_Node_Id];
205 }
206
207 /* Return nonzero if a GCC tree has been associated with GNAT_ENTITY. */
208
209 int
210 present_gnu_tree (Entity_Id gnat_entity)
211 {
212 return (associate_gnat_to_gnu[gnat_entity - First_Node_Id] != NULL_TREE);
213 }
214
215 \f
216 /* Return non-zero if we are currently in the global binding level. */
217
218 int
219 global_bindings_p (void)
220 {
221 return (force_global != 0 || current_binding_level == global_binding_level
222 ? -1 : 0);
223 }
224
225 /* Return the list of declarations in the current level. Note that this list
226 is in reverse order (it has to be so for back-end compatibility). */
227
228 tree
229 getdecls (void)
230 {
231 return current_binding_level->names;
232 }
233
234 /* Nonzero if the current level needs to have a BLOCK made. */
235
236 int
237 kept_level_p (void)
238 {
239 return (current_binding_level->names != 0);
240 }
241
242 /* Enter a new binding level. The input parameter is ignored, but has to be
243 specified for back-end compatibility. */
244
245 void
246 pushlevel (int ignore ATTRIBUTE_UNUSED)
247 {
248 struct binding_level *newlevel = NULL;
249
250 /* Reuse a struct for this binding level, if there is one. */
251 if (free_binding_level)
252 {
253 newlevel = free_binding_level;
254 free_binding_level = free_binding_level->level_chain;
255 }
256 else
257 newlevel
258 = (struct binding_level *) ggc_alloc (sizeof (struct binding_level));
259
260 *newlevel = clear_binding_level;
261
262 /* Add this level to the front of the chain (stack) of levels that are
263 active. */
264 newlevel->level_chain = current_binding_level;
265 current_binding_level = newlevel;
266 }
267
268 /* Exit a binding level.
269 Pop the level off, and restore the state of the identifier-decl mappings
270 that were in effect when this level was entered.
271
272 If KEEP is nonzero, this level had explicit declarations, so
273 and create a "block" (a BLOCK node) for the level
274 to record its declarations and subblocks for symbol table output.
275
276 If FUNCTIONBODY is nonzero, this level is the body of a function,
277 so create a block as if KEEP were set and also clear out all
278 label names.
279
280 If REVERSE is nonzero, reverse the order of decls before putting
281 them into the BLOCK. */
282
283 tree
284 poplevel (int keep, int reverse, int functionbody)
285 {
286 /* Points to a GCC BLOCK tree node. This is the BLOCK node construted for the
287 binding level that we are about to exit and which is returned by this
288 routine. */
289 tree block = NULL_TREE;
290 tree decl_chain;
291 tree decl_node;
292 tree subblock_chain = current_binding_level->blocks;
293 tree subblock_node;
294 int block_previously_created;
295
296 /* Reverse the list of XXXX_DECL nodes if desired. Note that the ..._DECL
297 nodes chained through the `names' field of current_binding_level are in
298 reverse order except for PARM_DECL node, which are explicitly stored in
299 the right order. */
300 current_binding_level->names
301 = decl_chain = (reverse) ? nreverse (current_binding_level->names)
302 : current_binding_level->names;
303
304 /* Output any nested inline functions within this block which must be
305 compiled because their address is needed. */
306 for (decl_node = decl_chain; decl_node; decl_node = TREE_CHAIN (decl_node))
307 if (TREE_CODE (decl_node) == FUNCTION_DECL
308 && ! TREE_ASM_WRITTEN (decl_node) && TREE_ADDRESSABLE (decl_node)
309 && DECL_INITIAL (decl_node) != 0)
310 {
311 push_function_context ();
312 output_inline_function (decl_node);
313 pop_function_context ();
314 }
315
316 block = 0;
317 block_previously_created = (current_binding_level->this_block != 0);
318 if (block_previously_created)
319 block = current_binding_level->this_block;
320 else if (keep || functionbody)
321 block = make_node (BLOCK);
322 if (block != 0)
323 {
324 BLOCK_VARS (block) = keep ? decl_chain : 0;
325 BLOCK_SUBBLOCKS (block) = subblock_chain;
326 }
327
328 /* Record the BLOCK node just built as the subblock its enclosing scope. */
329 for (subblock_node = subblock_chain; subblock_node;
330 subblock_node = TREE_CHAIN (subblock_node))
331 BLOCK_SUPERCONTEXT (subblock_node) = block;
332
333 /* Clear out the meanings of the local variables of this level. */
334
335 for (subblock_node = decl_chain; subblock_node;
336 subblock_node = TREE_CHAIN (subblock_node))
337 if (DECL_NAME (subblock_node) != 0)
338 /* If the identifier was used or addressed via a local extern decl,
339 don't forget that fact. */
340 if (DECL_EXTERNAL (subblock_node))
341 {
342 if (TREE_USED (subblock_node))
343 TREE_USED (DECL_NAME (subblock_node)) = 1;
344 if (TREE_ADDRESSABLE (subblock_node))
345 TREE_ADDRESSABLE (DECL_ASSEMBLER_NAME (subblock_node)) = 1;
346 }
347
348 {
349 /* Pop the current level, and free the structure for reuse. */
350 struct binding_level *level = current_binding_level;
351 current_binding_level = current_binding_level->level_chain;
352 level->level_chain = free_binding_level;
353 free_binding_level = level;
354 }
355
356 if (functionbody)
357 {
358 /* This is the top level block of a function. The ..._DECL chain stored
359 in BLOCK_VARS are the function's parameters (PARM_DECL nodes). Don't
360 leave them in the BLOCK because they are found in the FUNCTION_DECL
361 instead. */
362 DECL_INITIAL (current_function_decl) = block;
363 BLOCK_VARS (block) = 0;
364 }
365 else if (block)
366 {
367 if (!block_previously_created)
368 current_binding_level->blocks
369 = chainon (current_binding_level->blocks, block);
370 }
371
372 /* If we did not make a block for the level just exited, any blocks made for
373 inner levels (since they cannot be recorded as subblocks in that level)
374 must be carried forward so they will later become subblocks of something
375 else. */
376 else if (subblock_chain)
377 current_binding_level->blocks
378 = chainon (current_binding_level->blocks, subblock_chain);
379 if (block)
380 TREE_USED (block) = 1;
381
382 return block;
383 }
384 \f
385 /* Insert BLOCK at the end of the list of subblocks of the
386 current binding level. This is used when a BIND_EXPR is expanded,
387 to handle the BLOCK node inside the BIND_EXPR. */
388
389 void
390 insert_block (tree block)
391 {
392 TREE_USED (block) = 1;
393 current_binding_level->blocks
394 = chainon (current_binding_level->blocks, block);
395 }
396
397 /* Set the BLOCK node for the innermost scope
398 (the one we are currently in). */
399
400 void
401 set_block (tree block)
402 {
403 current_binding_level->this_block = block;
404 current_binding_level->names = chainon (current_binding_level->names,
405 BLOCK_VARS (block));
406 current_binding_level->blocks = chainon (current_binding_level->blocks,
407 BLOCK_SUBBLOCKS (block));
408 }
409
410 /* Records a ..._DECL node DECL as belonging to the current lexical scope.
411 Returns the ..._DECL node. */
412
413 tree
414 pushdecl (tree decl)
415 {
416 struct binding_level *b;
417
418 /* If at top level, there is no context. But PARM_DECLs always go in the
419 level of its function. */
420 if (global_bindings_p () && TREE_CODE (decl) != PARM_DECL)
421 {
422 b = global_binding_level;
423 DECL_CONTEXT (decl) = 0;
424 }
425 else
426 {
427 b = current_binding_level;
428 DECL_CONTEXT (decl) = current_function_decl;
429 }
430
431 /* Put the declaration on the list. The list of declarations is in reverse
432 order. The list will be reversed later if necessary. This needs to be
433 this way for compatibility with the back-end.
434
435 Don't put TYPE_DECLs for UNCONSTRAINED_ARRAY_TYPE into the list. They
436 will cause trouble with the debugger and aren't needed anyway. */
437 if (TREE_CODE (decl) != TYPE_DECL
438 || TREE_CODE (TREE_TYPE (decl)) != UNCONSTRAINED_ARRAY_TYPE)
439 {
440 TREE_CHAIN (decl) = b->names;
441 b->names = decl;
442 }
443
444 /* For the declaration of a type, set its name if it either is not already
445 set, was set to an IDENTIFIER_NODE, indicating an internal name,
446 or if the previous type name was not derived from a source name.
447 We'd rather have the type named with a real name and all the pointer
448 types to the same object have the same POINTER_TYPE node. Code in this
449 function in c-decl.c makes a copy of the type node here, but that may
450 cause us trouble with incomplete types, so let's not try it (at least
451 for now). */
452
453 if (TREE_CODE (decl) == TYPE_DECL
454 && DECL_NAME (decl) != 0
455 && (TYPE_NAME (TREE_TYPE (decl)) == 0
456 || TREE_CODE (TYPE_NAME (TREE_TYPE (decl))) == IDENTIFIER_NODE
457 || (TREE_CODE (TYPE_NAME (TREE_TYPE (decl))) == TYPE_DECL
458 && DECL_ARTIFICIAL (TYPE_NAME (TREE_TYPE (decl)))
459 && ! DECL_ARTIFICIAL (decl))))
460 TYPE_NAME (TREE_TYPE (decl)) = decl;
461
462 return decl;
463 }
464 \f
465 /* Do little here. Set up the standard declarations later after the
466 front end has been run. */
467
468 void
469 gnat_init_decl_processing (void)
470 {
471 input_line = 0;
472
473 /* Make the binding_level structure for global names. */
474 current_function_decl = 0;
475 current_binding_level = 0;
476 free_binding_level = 0;
477 pushlevel (0);
478 global_binding_level = current_binding_level;
479
480 build_common_tree_nodes (0);
481
482 /* In Ada, we use a signed type for SIZETYPE. Use the signed type
483 corresponding to the size of Pmode. In most cases when ptr_mode and
484 Pmode differ, C will use the width of ptr_mode as sizetype. But we get
485 far better code using the width of Pmode. Make this here since we need
486 this before we can expand the GNAT types. */
487 set_sizetype (gnat_type_for_size (GET_MODE_BITSIZE (Pmode), 0));
488 build_common_tree_nodes_2 (0);
489
490 pushdecl (build_decl (TYPE_DECL, get_identifier (SIZE_TYPE), sizetype));
491
492 /* We need to make the integer type before doing anything else.
493 We stitch this in to the appropriate GNAT type later. */
494 pushdecl (build_decl (TYPE_DECL, get_identifier ("integer"),
495 integer_type_node));
496 pushdecl (build_decl (TYPE_DECL, get_identifier ("unsigned char"),
497 char_type_node));
498
499 ptr_void_type_node = build_pointer_type (void_type_node);
500
501 }
502
503 /* Create the predefined scalar types such as `integer_type_node' needed
504 in the gcc back-end and initialize the global binding level. */
505
506 void
507 init_gigi_decls (tree long_long_float_type, tree exception_type)
508 {
509 tree endlink, decl;
510 unsigned int i;
511
512 /* Set the types that GCC and Gigi use from the front end. We would like
513 to do this for char_type_node, but it needs to correspond to the C
514 char type. */
515 if (TREE_CODE (TREE_TYPE (long_long_float_type)) == INTEGER_TYPE)
516 {
517 /* In this case, the builtin floating point types are VAX float,
518 so make up a type for use. */
519 longest_float_type_node = make_node (REAL_TYPE);
520 TYPE_PRECISION (longest_float_type_node) = LONG_DOUBLE_TYPE_SIZE;
521 layout_type (longest_float_type_node);
522 pushdecl (build_decl (TYPE_DECL, get_identifier ("longest float type"),
523 longest_float_type_node));
524 }
525 else
526 longest_float_type_node = TREE_TYPE (long_long_float_type);
527
528 except_type_node = TREE_TYPE (exception_type);
529
530 unsigned_type_node = gnat_type_for_size (INT_TYPE_SIZE, 1);
531 pushdecl (build_decl (TYPE_DECL, get_identifier ("unsigned int"),
532 unsigned_type_node));
533
534 void_type_decl_node
535 = pushdecl (build_decl (TYPE_DECL, get_identifier ("void"),
536 void_type_node));
537
538 void_ftype = build_function_type (void_type_node, NULL_TREE);
539 ptr_void_ftype = build_pointer_type (void_ftype);
540
541 /* Now declare runtime functions. */
542 endlink = tree_cons (NULL_TREE, void_type_node, NULL_TREE);
543
544 /* malloc is a function declaration tree for a function to allocate
545 memory. */
546 malloc_decl = create_subprog_decl (get_identifier ("__gnat_malloc"),
547 NULL_TREE,
548 build_function_type (ptr_void_type_node,
549 tree_cons (NULL_TREE,
550 sizetype,
551 endlink)),
552 NULL_TREE, 0, 1, 1, 0);
553
554 /* free is a function declaration tree for a function to free memory. */
555 free_decl
556 = create_subprog_decl (get_identifier ("__gnat_free"), NULL_TREE,
557 build_function_type (void_type_node,
558 tree_cons (NULL_TREE,
559 ptr_void_type_node,
560 endlink)),
561 NULL_TREE, 0, 1, 1, 0);
562
563 /* Make the types and functions used for exception processing. */
564 jmpbuf_type
565 = build_array_type (gnat_type_for_mode (Pmode, 0),
566 build_index_type (build_int_2 (5, 0)));
567 pushdecl (build_decl (TYPE_DECL, get_identifier ("JMPBUF_T"), jmpbuf_type));
568 jmpbuf_ptr_type = build_pointer_type (jmpbuf_type);
569
570 /* Functions to get and set the jumpbuf pointer for the current thread. */
571 get_jmpbuf_decl
572 = create_subprog_decl
573 (get_identifier ("system__soft_links__get_jmpbuf_address_soft"),
574 NULL_TREE, build_function_type (jmpbuf_ptr_type, NULL_TREE),
575 NULL_TREE, 0, 1, 1, 0);
576
577 set_jmpbuf_decl
578 = create_subprog_decl
579 (get_identifier ("system__soft_links__set_jmpbuf_address_soft"),
580 NULL_TREE,
581 build_function_type (void_type_node,
582 tree_cons (NULL_TREE, jmpbuf_ptr_type, endlink)),
583 NULL_TREE, 0, 1, 1, 0);
584
585 /* Function to get the current exception. */
586 get_excptr_decl
587 = create_subprog_decl
588 (get_identifier ("system__soft_links__get_gnat_exception"),
589 NULL_TREE,
590 build_function_type (build_pointer_type (except_type_node), NULL_TREE),
591 NULL_TREE, 0, 1, 1, 0);
592
593 /* Functions that raise exceptions. */
594 raise_nodefer_decl
595 = create_subprog_decl
596 (get_identifier ("__gnat_raise_nodefer_with_msg"), NULL_TREE,
597 build_function_type (void_type_node,
598 tree_cons (NULL_TREE,
599 build_pointer_type (except_type_node),
600 endlink)),
601 NULL_TREE, 0, 1, 1, 0);
602
603 /* Hooks to call when entering/leaving an exception handler. */
604 begin_handler_decl
605 = create_subprog_decl (get_identifier ("__gnat_begin_handler"), NULL_TREE,
606 build_function_type (void_type_node,
607 tree_cons (NULL_TREE,
608 ptr_void_type_node,
609 endlink)),
610 NULL_TREE, 0, 1, 1, 0);
611
612 end_handler_decl
613 = create_subprog_decl (get_identifier ("__gnat_end_handler"), NULL_TREE,
614 build_function_type (void_type_node,
615 tree_cons (NULL_TREE,
616 ptr_void_type_node,
617 endlink)),
618 NULL_TREE, 0, 1, 1, 0);
619
620 /* If in no exception handlers mode, all raise statements are redirected to
621 __gnat_last_chance_handler. No need to redefine raise_nodefer_decl, since
622 this procedure will never be called in this mode. */
623 if (No_Exception_Handlers_Set ())
624 {
625 decl
626 = create_subprog_decl
627 (get_identifier ("__gnat_last_chance_handler"), NULL_TREE,
628 build_function_type (void_type_node,
629 tree_cons (NULL_TREE,
630 build_pointer_type (char_type_node),
631 tree_cons (NULL_TREE,
632 integer_type_node,
633 endlink))),
634 NULL_TREE, 0, 1, 1, 0);
635
636 for (i = 0; i < ARRAY_SIZE (gnat_raise_decls); i++)
637 gnat_raise_decls[i] = decl;
638 }
639 else
640 /* Otherwise, make one decl for each exception reason. */
641 for (i = 0; i < ARRAY_SIZE (gnat_raise_decls); i++)
642 {
643 char name[17];
644
645 sprintf (name, "__gnat_rcheck_%.2d", i);
646 gnat_raise_decls[i]
647 = create_subprog_decl
648 (get_identifier (name), NULL_TREE,
649 build_function_type (void_type_node,
650 tree_cons (NULL_TREE,
651 build_pointer_type
652 (char_type_node),
653 tree_cons (NULL_TREE,
654 integer_type_node,
655 endlink))),
656 NULL_TREE, 0, 1, 1, 0);
657 }
658
659 /* Indicate that these never return. */
660 TREE_THIS_VOLATILE (raise_nodefer_decl) = 1;
661 TREE_SIDE_EFFECTS (raise_nodefer_decl) = 1;
662 TREE_TYPE (raise_nodefer_decl)
663 = build_qualified_type (TREE_TYPE (raise_nodefer_decl),
664 TYPE_QUAL_VOLATILE);
665
666 for (i = 0; i < ARRAY_SIZE (gnat_raise_decls); i++)
667 {
668 TREE_THIS_VOLATILE (gnat_raise_decls[i]) = 1;
669 TREE_SIDE_EFFECTS (gnat_raise_decls[i]) = 1;
670 TREE_TYPE (gnat_raise_decls[i])
671 = build_qualified_type (TREE_TYPE (gnat_raise_decls[i]),
672 TYPE_QUAL_VOLATILE);
673 }
674
675 /* setjmp returns an integer and has one operand, which is a pointer to
676 a jmpbuf. */
677 setjmp_decl
678 = create_subprog_decl
679 (get_identifier ("__builtin_setjmp"), NULL_TREE,
680 build_function_type (integer_type_node,
681 tree_cons (NULL_TREE, jmpbuf_ptr_type, endlink)),
682 NULL_TREE, 0, 1, 1, 0);
683
684 DECL_BUILT_IN_CLASS (setjmp_decl) = BUILT_IN_NORMAL;
685 DECL_FUNCTION_CODE (setjmp_decl) = BUILT_IN_SETJMP;
686
687 main_identifier_node = get_identifier ("main");
688 }
689 \f
690 /* Given a record type (RECORD_TYPE) and a chain of FIELD_DECL
691 nodes (FIELDLIST), finish constructing the record or union type.
692 If HAS_REP is nonzero, this record has a rep clause; don't call
693 layout_type but merely set the size and alignment ourselves.
694 If DEFER_DEBUG is nonzero, do not call the debugging routines
695 on this type; it will be done later. */
696
697 void
698 finish_record_type (tree record_type,
699 tree fieldlist,
700 int has_rep,
701 int defer_debug)
702 {
703 enum tree_code code = TREE_CODE (record_type);
704 tree ada_size = bitsize_zero_node;
705 tree size = bitsize_zero_node;
706 tree size_unit = size_zero_node;
707 int var_size = 0;
708 tree field;
709
710 TYPE_FIELDS (record_type) = fieldlist;
711
712 if (TYPE_NAME (record_type) != 0
713 && TREE_CODE (TYPE_NAME (record_type)) == TYPE_DECL)
714 TYPE_STUB_DECL (record_type) = TYPE_NAME (record_type);
715 else
716 TYPE_STUB_DECL (record_type)
717 = pushdecl (build_decl (TYPE_DECL, TYPE_NAME (record_type),
718 record_type));
719
720 /* We don't need both the typedef name and the record name output in
721 the debugging information, since they are the same. */
722 DECL_ARTIFICIAL (TYPE_STUB_DECL (record_type)) = 1;
723
724 /* Globally initialize the record first. If this is a rep'ed record,
725 that just means some initializations; otherwise, layout the record. */
726
727 if (has_rep)
728 {
729 TYPE_ALIGN (record_type) = MAX (BITS_PER_UNIT, TYPE_ALIGN (record_type));
730 TYPE_MODE (record_type) = BLKmode;
731 if (TYPE_SIZE (record_type) == 0)
732 {
733 TYPE_SIZE (record_type) = bitsize_zero_node;
734 TYPE_SIZE_UNIT (record_type) = size_zero_node;
735 }
736 /* For all-repped records with a size specified, lay the QUAL_UNION_TYPE
737 out just like a UNION_TYPE, since the size will be fixed. */
738 else if (code == QUAL_UNION_TYPE)
739 code = UNION_TYPE;
740 }
741 else
742 {
743 /* Ensure there isn't a size already set. There can be in an error
744 case where there is a rep clause but all fields have errors and
745 no longer have a position. */
746 TYPE_SIZE (record_type) = 0;
747 layout_type (record_type);
748 }
749
750 /* At this point, the position and size of each field is known. It was
751 either set before entry by a rep clause, or by laying out the type above.
752
753 We now run a pass over the fields (in reverse order for QUAL_UNION_TYPEs)
754 to compute the Ada size; the GCC size and alignment (for rep'ed records
755 that are not padding types); and the mode (for rep'ed records). We also
756 clear the DECL_BIT_FIELD indication for the cases we know have not been
757 handled yet, and adjust DECL_NONADDRESSABLE_P accordingly. */
758
759 if (code == QUAL_UNION_TYPE)
760 fieldlist = nreverse (fieldlist);
761
762 for (field = fieldlist; field; field = TREE_CHAIN (field))
763 {
764 tree pos = bit_position (field);
765
766 tree type = TREE_TYPE (field);
767 tree this_size = DECL_SIZE (field);
768 tree this_size_unit = DECL_SIZE_UNIT (field);
769 tree this_ada_size = DECL_SIZE (field);
770
771 /* We need to make an XVE/XVU record if any field has variable size,
772 whether or not the record does. For example, if we have an union,
773 it may be that all fields, rounded up to the alignment, have the
774 same size, in which case we'll use that size. But the debug
775 output routines (except Dwarf2) won't be able to output the fields,
776 so we need to make the special record. */
777 if (TREE_CODE (this_size) != INTEGER_CST)
778 var_size = 1;
779
780 if ((TREE_CODE (type) == RECORD_TYPE || TREE_CODE (type) == UNION_TYPE
781 || TREE_CODE (type) == QUAL_UNION_TYPE)
782 && ! TYPE_IS_FAT_POINTER_P (type)
783 && ! TYPE_CONTAINS_TEMPLATE_P (type)
784 && TYPE_ADA_SIZE (type) != 0)
785 this_ada_size = TYPE_ADA_SIZE (type);
786
787 /* Clear DECL_BIT_FIELD for the cases layout_decl does not handle. */
788 if (DECL_BIT_FIELD (field) && !STRICT_ALIGNMENT
789 && value_factor_p (pos, BITS_PER_UNIT)
790 && operand_equal_p (this_size, TYPE_SIZE (type), 0))
791 DECL_BIT_FIELD (field) = 0;
792
793 /* If we still have DECL_BIT_FIELD set at this point, we know the field
794 is technically not addressable. */
795 DECL_NONADDRESSABLE_P (field) |= DECL_BIT_FIELD (field);
796
797 if (has_rep && ! DECL_BIT_FIELD (field))
798 TYPE_ALIGN (record_type)
799 = MAX (TYPE_ALIGN (record_type), DECL_ALIGN (field));
800
801 switch (code)
802 {
803 case UNION_TYPE:
804 ada_size = size_binop (MAX_EXPR, ada_size, this_ada_size);
805 size = size_binop (MAX_EXPR, size, this_size);
806 size_unit = size_binop (MAX_EXPR, size_unit, this_size_unit);
807 break;
808
809 case QUAL_UNION_TYPE:
810 ada_size
811 = fold (build (COND_EXPR, bitsizetype, DECL_QUALIFIER (field),
812 this_ada_size, ada_size));
813 size = fold (build (COND_EXPR, bitsizetype, DECL_QUALIFIER (field),
814 this_size, size));
815 size_unit = fold (build (COND_EXPR, sizetype, DECL_QUALIFIER (field),
816 this_size_unit, size_unit));
817 break;
818
819 case RECORD_TYPE:
820 /* Since we know here that all fields are sorted in order of
821 increasing bit position, the size of the record is one
822 higher than the ending bit of the last field processed
823 unless we have a rep clause, since in that case we might
824 have a field outside a QUAL_UNION_TYPE that has a higher ending
825 position. So use a MAX in that case. Also, if this field is a
826 QUAL_UNION_TYPE, we need to take into account the previous size in
827 the case of empty variants. */
828 ada_size
829 = merge_sizes (ada_size, pos, this_ada_size,
830 TREE_CODE (type) == QUAL_UNION_TYPE, has_rep);
831 size = merge_sizes (size, pos, this_size,
832 TREE_CODE (type) == QUAL_UNION_TYPE, has_rep);
833 size_unit
834 = merge_sizes (size_unit, byte_position (field), this_size_unit,
835 TREE_CODE (type) == QUAL_UNION_TYPE, has_rep);
836 break;
837
838 default:
839 abort ();
840 }
841 }
842
843 if (code == QUAL_UNION_TYPE)
844 nreverse (fieldlist);
845
846 /* If this is a padding record, we never want to make the size smaller than
847 what was specified in it, if any. */
848 if (TREE_CODE (record_type) == RECORD_TYPE
849 && TYPE_IS_PADDING_P (record_type) && TYPE_SIZE (record_type) != 0)
850 {
851 size = TYPE_SIZE (record_type);
852 size_unit = TYPE_SIZE_UNIT (record_type);
853 }
854
855 /* Now set any of the values we've just computed that apply. */
856 if (! TYPE_IS_FAT_POINTER_P (record_type)
857 && ! TYPE_CONTAINS_TEMPLATE_P (record_type))
858 SET_TYPE_ADA_SIZE (record_type, ada_size);
859
860 if (has_rep)
861 {
862 if (! (TREE_CODE (record_type) == RECORD_TYPE
863 && TYPE_IS_PADDING_P (record_type)
864 && CONTAINS_PLACEHOLDER_P (size)))
865 {
866 TYPE_SIZE (record_type) = round_up (size, TYPE_ALIGN (record_type));
867 TYPE_SIZE_UNIT (record_type)
868 = round_up (size_unit,
869 TYPE_ALIGN (record_type) / BITS_PER_UNIT);
870 }
871
872 compute_record_mode (record_type);
873 }
874
875 if (! defer_debug)
876 {
877 /* If this record is of variable size, rename it so that the
878 debugger knows it is and make a new, parallel, record
879 that tells the debugger how the record is laid out. See
880 exp_dbug.ads. But don't do this for records that are padding
881 since they confuse GDB. */
882 if (var_size
883 && ! (TREE_CODE (record_type) == RECORD_TYPE
884 && TYPE_IS_PADDING_P (record_type)))
885 {
886 tree new_record_type
887 = make_node (TREE_CODE (record_type) == QUAL_UNION_TYPE
888 ? UNION_TYPE : TREE_CODE (record_type));
889 tree orig_id = DECL_NAME (TYPE_STUB_DECL (record_type));
890 tree new_id
891 = concat_id_with_name (orig_id,
892 TREE_CODE (record_type) == QUAL_UNION_TYPE
893 ? "XVU" : "XVE");
894 tree last_pos = bitsize_zero_node;
895 tree old_field;
896 tree prev_old_field = 0;
897
898 TYPE_NAME (new_record_type) = new_id;
899 TYPE_ALIGN (new_record_type) = BIGGEST_ALIGNMENT;
900 TYPE_STUB_DECL (new_record_type)
901 = pushdecl (build_decl (TYPE_DECL, new_id, new_record_type));
902 DECL_ARTIFICIAL (TYPE_STUB_DECL (new_record_type)) = 1;
903 DECL_IGNORED_P (TYPE_STUB_DECL (new_record_type))
904 = DECL_IGNORED_P (TYPE_STUB_DECL (record_type));
905 TYPE_SIZE (new_record_type) = size_int (TYPE_ALIGN (record_type));
906
907 /* Now scan all the fields, replacing each field with a new
908 field corresponding to the new encoding. */
909 for (old_field = TYPE_FIELDS (record_type); old_field != 0;
910 old_field = TREE_CHAIN (old_field))
911 {
912 tree field_type = TREE_TYPE (old_field);
913 tree field_name = DECL_NAME (old_field);
914 tree new_field;
915 tree curpos = bit_position (old_field);
916 int var = 0;
917 unsigned int align = 0;
918 tree pos;
919
920 /* See how the position was modified from the last position.
921
922 There are two basic cases we support: a value was added
923 to the last position or the last position was rounded to
924 a boundary and they something was added. Check for the
925 first case first. If not, see if there is any evidence
926 of rounding. If so, round the last position and try
927 again.
928
929 If this is a union, the position can be taken as zero. */
930
931 if (TREE_CODE (new_record_type) == UNION_TYPE)
932 pos = bitsize_zero_node, align = 0;
933 else
934 pos = compute_related_constant (curpos, last_pos);
935
936 if (pos == 0 && TREE_CODE (curpos) == MULT_EXPR
937 && TREE_CODE (TREE_OPERAND (curpos, 1)) == INTEGER_CST)
938 {
939 align = TREE_INT_CST_LOW (TREE_OPERAND (curpos, 1));
940 pos = compute_related_constant (curpos,
941 round_up (last_pos, align));
942 }
943 else if (pos == 0 && TREE_CODE (curpos) == PLUS_EXPR
944 && TREE_CODE (TREE_OPERAND (curpos, 1)) == INTEGER_CST
945 && TREE_CODE (TREE_OPERAND (curpos, 0)) == MULT_EXPR
946 && host_integerp (TREE_OPERAND
947 (TREE_OPERAND (curpos, 0), 1),
948 1))
949 {
950 align
951 = tree_low_cst
952 (TREE_OPERAND (TREE_OPERAND (curpos, 0), 1), 1);
953 pos = compute_related_constant (curpos,
954 round_up (last_pos, align));
955 }
956 else if (potential_alignment_gap (prev_old_field, old_field,
957 pos))
958 {
959 align = TYPE_ALIGN (field_type);
960 pos = compute_related_constant (curpos,
961 round_up (last_pos, align));
962 }
963
964 /* If we can't compute a position, set it to zero.
965
966 ??? We really should abort here, but it's too much work
967 to get this correct for all cases. */
968
969 if (pos == 0)
970 pos = bitsize_zero_node;
971
972 /* See if this type is variable-size and make a new type
973 and indicate the indirection if so. */
974 if (TREE_CODE (DECL_SIZE (old_field)) != INTEGER_CST)
975 {
976 field_type = build_pointer_type (field_type);
977 var = 1;
978 }
979
980 /* Make a new field name, if necessary. */
981 if (var || align != 0)
982 {
983 char suffix[6];
984
985 if (align != 0)
986 sprintf (suffix, "XV%c%u", var ? 'L' : 'A',
987 align / BITS_PER_UNIT);
988 else
989 strcpy (suffix, "XVL");
990
991 field_name = concat_id_with_name (field_name, suffix);
992 }
993
994 new_field = create_field_decl (field_name, field_type,
995 new_record_type, 0,
996 DECL_SIZE (old_field), pos, 0);
997 TREE_CHAIN (new_field) = TYPE_FIELDS (new_record_type);
998 TYPE_FIELDS (new_record_type) = new_field;
999
1000 /* If old_field is a QUAL_UNION_TYPE, take its size as being
1001 zero. The only time it's not the last field of the record
1002 is when there are other components at fixed positions after
1003 it (meaning there was a rep clause for every field) and we
1004 want to be able to encode them. */
1005 last_pos = size_binop (PLUS_EXPR, bit_position (old_field),
1006 (TREE_CODE (TREE_TYPE (old_field))
1007 == QUAL_UNION_TYPE)
1008 ? bitsize_zero_node
1009 : DECL_SIZE (old_field));
1010 prev_old_field = old_field;
1011 }
1012
1013 TYPE_FIELDS (new_record_type)
1014 = nreverse (TYPE_FIELDS (new_record_type));
1015
1016 rest_of_type_compilation (new_record_type, global_bindings_p ());
1017 }
1018
1019 rest_of_type_compilation (record_type, global_bindings_p ());
1020 }
1021 }
1022
1023 /* Utility function of above to merge LAST_SIZE, the previous size of a record
1024 with FIRST_BIT and SIZE that describe a field. SPECIAL is nonzero
1025 if this represents a QUAL_UNION_TYPE in which case we must look for
1026 COND_EXPRs and replace a value of zero with the old size. If HAS_REP
1027 is nonzero, we must take the MAX of the end position of this field
1028 with LAST_SIZE. In all other cases, we use FIRST_BIT plus SIZE.
1029
1030 We return an expression for the size. */
1031
1032 static tree
1033 merge_sizes (tree last_size,
1034 tree first_bit,
1035 tree size,
1036 int special,
1037 int has_rep)
1038 {
1039 tree type = TREE_TYPE (last_size);
1040 tree new;
1041
1042 if (! special || TREE_CODE (size) != COND_EXPR)
1043 {
1044 new = size_binop (PLUS_EXPR, first_bit, size);
1045 if (has_rep)
1046 new = size_binop (MAX_EXPR, last_size, new);
1047 }
1048
1049 else
1050 new = fold (build (COND_EXPR, type, TREE_OPERAND (size, 0),
1051 integer_zerop (TREE_OPERAND (size, 1))
1052 ? last_size : merge_sizes (last_size, first_bit,
1053 TREE_OPERAND (size, 1),
1054 1, has_rep),
1055 integer_zerop (TREE_OPERAND (size, 2))
1056 ? last_size : merge_sizes (last_size, first_bit,
1057 TREE_OPERAND (size, 2),
1058 1, has_rep)));
1059
1060 /* We don't need any NON_VALUE_EXPRs and they can confuse us (especially
1061 when fed through substitute_in_expr) into thinking that a constant
1062 size is not constant. */
1063 while (TREE_CODE (new) == NON_LVALUE_EXPR)
1064 new = TREE_OPERAND (new, 0);
1065
1066 return new;
1067 }
1068
1069 /* Utility function of above to see if OP0 and OP1, both of SIZETYPE, are
1070 related by the addition of a constant. Return that constant if so. */
1071
1072 static tree
1073 compute_related_constant (tree op0, tree op1)
1074 {
1075 tree op0_var, op1_var;
1076 tree op0_con = split_plus (op0, &op0_var);
1077 tree op1_con = split_plus (op1, &op1_var);
1078 tree result = size_binop (MINUS_EXPR, op0_con, op1_con);
1079
1080 if (operand_equal_p (op0_var, op1_var, 0))
1081 return result;
1082 else if (operand_equal_p (op0, size_binop (PLUS_EXPR, op1_var, result), 0))
1083 return result;
1084 else
1085 return 0;
1086 }
1087
1088 /* Utility function of above to split a tree OP which may be a sum, into a
1089 constant part, which is returned, and a variable part, which is stored
1090 in *PVAR. *PVAR may be bitsize_zero_node. All operations must be of
1091 bitsizetype. */
1092
1093 static tree
1094 split_plus (tree in, tree *pvar)
1095 {
1096 /* Strip NOPS in order to ease the tree traversal and maximize the
1097 potential for constant or plus/minus discovery. We need to be careful
1098 to always return and set *pvar to bitsizetype trees, but it's worth
1099 the effort. */
1100 STRIP_NOPS (in);
1101
1102 *pvar = convert (bitsizetype, in);
1103
1104 if (TREE_CODE (in) == INTEGER_CST)
1105 {
1106 *pvar = bitsize_zero_node;
1107 return convert (bitsizetype, in);
1108 }
1109 else if (TREE_CODE (in) == PLUS_EXPR || TREE_CODE (in) == MINUS_EXPR)
1110 {
1111 tree lhs_var, rhs_var;
1112 tree lhs_con = split_plus (TREE_OPERAND (in, 0), &lhs_var);
1113 tree rhs_con = split_plus (TREE_OPERAND (in, 1), &rhs_var);
1114
1115 if (lhs_var == TREE_OPERAND (in, 0)
1116 && rhs_var == TREE_OPERAND (in, 1))
1117 return bitsize_zero_node;
1118
1119 *pvar = size_binop (TREE_CODE (in), lhs_var, rhs_var);
1120 return size_binop (TREE_CODE (in), lhs_con, rhs_con);
1121 }
1122 else
1123 return bitsize_zero_node;
1124 }
1125 \f
1126 /* Return a FUNCTION_TYPE node. RETURN_TYPE is the type returned by the
1127 subprogram. If it is void_type_node, then we are dealing with a procedure,
1128 otherwise we are dealing with a function. PARAM_DECL_LIST is a list of
1129 PARM_DECL nodes that are the subprogram arguments. CICO_LIST is the
1130 copy-in/copy-out list to be stored into TYPE_CICO_LIST.
1131 RETURNS_UNCONSTRAINED is nonzero if the function returns an unconstrained
1132 object. RETURNS_BY_REF is nonzero if the function returns by reference.
1133 RETURNS_WITH_DSP is nonzero if the function is to return with a
1134 depressed stack pointer. */
1135
1136 tree
1137 create_subprog_type (tree return_type,
1138 tree param_decl_list,
1139 tree cico_list,
1140 int returns_unconstrained,
1141 int returns_by_ref,
1142 int returns_with_dsp)
1143 {
1144 /* A chain of TREE_LIST nodes whose TREE_VALUEs are the data type nodes of
1145 the subprogram formal parameters. This list is generated by traversing the
1146 input list of PARM_DECL nodes. */
1147 tree param_type_list = NULL;
1148 tree param_decl;
1149 tree type;
1150
1151 for (param_decl = param_decl_list; param_decl;
1152 param_decl = TREE_CHAIN (param_decl))
1153 param_type_list = tree_cons (NULL_TREE, TREE_TYPE (param_decl),
1154 param_type_list);
1155
1156 /* The list of the function parameter types has to be terminated by the void
1157 type to signal to the back-end that we are not dealing with a variable
1158 parameter subprogram, but that the subprogram has a fixed number of
1159 parameters. */
1160 param_type_list = tree_cons (NULL_TREE, void_type_node, param_type_list);
1161
1162 /* The list of argument types has been created in reverse
1163 so nreverse it. */
1164 param_type_list = nreverse (param_type_list);
1165
1166 type = build_function_type (return_type, param_type_list);
1167
1168 /* TYPE may have been shared since GCC hashes types. If it has a CICO_LIST
1169 or the new type should, make a copy of TYPE. Likewise for
1170 RETURNS_UNCONSTRAINED and RETURNS_BY_REF. */
1171 if (TYPE_CI_CO_LIST (type) != 0 || cico_list != 0
1172 || TYPE_RETURNS_UNCONSTRAINED_P (type) != returns_unconstrained
1173 || TYPE_RETURNS_BY_REF_P (type) != returns_by_ref)
1174 type = copy_type (type);
1175
1176 SET_TYPE_CI_CO_LIST (type, cico_list);
1177 TYPE_RETURNS_UNCONSTRAINED_P (type) = returns_unconstrained;
1178 TYPE_RETURNS_STACK_DEPRESSED (type) = returns_with_dsp;
1179 TYPE_RETURNS_BY_REF_P (type) = returns_by_ref;
1180 return type;
1181 }
1182 \f
1183 /* Return a copy of TYPE but safe to modify in any way. */
1184
1185 tree
1186 copy_type (tree type)
1187 {
1188 tree new = copy_node (type);
1189
1190 /* copy_node clears this field instead of copying it, because it is
1191 aliased with TREE_CHAIN. */
1192 TYPE_STUB_DECL (new) = TYPE_STUB_DECL (type);
1193
1194 TYPE_POINTER_TO (new) = 0;
1195 TYPE_REFERENCE_TO (new) = 0;
1196 TYPE_MAIN_VARIANT (new) = new;
1197 TYPE_NEXT_VARIANT (new) = 0;
1198
1199 return new;
1200 }
1201 \f
1202 /* Return an INTEGER_TYPE of SIZETYPE with range MIN to MAX and whose
1203 TYPE_INDEX_TYPE is INDEX. */
1204
1205 tree
1206 create_index_type (tree min, tree max, tree index)
1207 {
1208 /* First build a type for the desired range. */
1209 tree type = build_index_2_type (min, max);
1210
1211 /* If this type has the TYPE_INDEX_TYPE we want, return it. Otherwise, if it
1212 doesn't have TYPE_INDEX_TYPE set, set it to INDEX. If TYPE_INDEX_TYPE
1213 is set, but not to INDEX, make a copy of this type with the requested
1214 index type. Note that we have no way of sharing these types, but that's
1215 only a small hole. */
1216 if (TYPE_INDEX_TYPE (type) == index)
1217 return type;
1218 else if (TYPE_INDEX_TYPE (type) != 0)
1219 type = copy_type (type);
1220
1221 SET_TYPE_INDEX_TYPE (type, index);
1222 return type;
1223 }
1224 \f
1225 /* Return a TYPE_DECL node. TYPE_NAME gives the name of the type (a character
1226 string) and TYPE is a ..._TYPE node giving its data type.
1227 ARTIFICIAL_P is nonzero if this is a declaration that was generated
1228 by the compiler. DEBUG_INFO_P is nonzero if we need to write debugging
1229 information about this type. */
1230
1231 tree
1232 create_type_decl (tree type_name,
1233 tree type,
1234 struct attrib *attr_list,
1235 int artificial_p,
1236 int debug_info_p)
1237 {
1238 tree type_decl = build_decl (TYPE_DECL, type_name, type);
1239 enum tree_code code = TREE_CODE (type);
1240
1241 DECL_ARTIFICIAL (type_decl) = artificial_p;
1242 pushdecl (type_decl);
1243 process_attributes (type_decl, attr_list);
1244
1245 /* Pass type declaration information to the debugger unless this is an
1246 UNCONSTRAINED_ARRAY_TYPE, which the debugger does not support,
1247 and ENUMERAL_TYPE or RECORD_TYPE which is handled separately,
1248 a dummy type, which will be completed later, or a type for which
1249 debugging information was not requested. */
1250 if (code == UNCONSTRAINED_ARRAY_TYPE || TYPE_IS_DUMMY_P (type)
1251 || ! debug_info_p)
1252 DECL_IGNORED_P (type_decl) = 1;
1253 else if (code != ENUMERAL_TYPE && code != RECORD_TYPE
1254 && ! ((code == POINTER_TYPE || code == REFERENCE_TYPE)
1255 && TYPE_IS_DUMMY_P (TREE_TYPE (type))))
1256 rest_of_decl_compilation (type_decl, NULL, global_bindings_p (), 0);
1257
1258 return type_decl;
1259 }
1260
1261 /* Returns a GCC VAR_DECL node. VAR_NAME gives the name of the variable.
1262 ASM_NAME is its assembler name (if provided). TYPE is its data type
1263 (a GCC ..._TYPE node). VAR_INIT is the GCC tree for an optional initial
1264 expression; NULL_TREE if none.
1265
1266 CONST_FLAG is nonzero if this variable is constant.
1267
1268 PUBLIC_FLAG is nonzero if this definition is to be made visible outside of
1269 the current compilation unit. This flag should be set when processing the
1270 variable definitions in a package specification. EXTERN_FLAG is nonzero
1271 when processing an external variable declaration (as opposed to a
1272 definition: no storage is to be allocated for the variable here).
1273
1274 STATIC_FLAG is only relevant when not at top level. In that case
1275 it indicates whether to always allocate storage to the variable. */
1276
1277 tree
1278 create_var_decl (tree var_name,
1279 tree asm_name,
1280 tree type,
1281 tree var_init,
1282 int const_flag,
1283 int public_flag,
1284 int extern_flag,
1285 int static_flag,
1286 struct attrib *attr_list)
1287 {
1288 int init_const
1289 = (var_init == 0
1290 ? 0
1291 : (TYPE_MAIN_VARIANT (type) == TYPE_MAIN_VARIANT (TREE_TYPE (var_init))
1292 && (global_bindings_p () || static_flag
1293 ? 0 != initializer_constant_valid_p (var_init,
1294 TREE_TYPE (var_init))
1295 : TREE_CONSTANT (var_init))));
1296 tree var_decl
1297 = build_decl ((const_flag && init_const
1298 /* Only make a CONST_DECL for sufficiently-small objects.
1299 We consider complex double "sufficiently-small" */
1300 && TYPE_SIZE (type) != 0
1301 && host_integerp (TYPE_SIZE_UNIT (type), 1)
1302 && 0 >= compare_tree_int (TYPE_SIZE_UNIT (type),
1303 GET_MODE_SIZE (DCmode)))
1304 ? CONST_DECL : VAR_DECL, var_name, type);
1305 tree assign_init = 0;
1306
1307 /* If this is external, throw away any initializations unless this is a
1308 CONST_DECL (meaning we have a constant); they will be done elsewhere. If
1309 we are defining a global here, leave a constant initialization and save
1310 any variable elaborations for the elaboration routine. Otherwise, if
1311 the initializing expression is not the same as TYPE, generate the
1312 initialization with an assignment statement, since it knows how
1313 to do the required adjustents. If we are just annotating types,
1314 throw away the initialization if it isn't a constant. */
1315
1316 if ((extern_flag && TREE_CODE (var_decl) != CONST_DECL)
1317 || (type_annotate_only && var_init != 0 && ! TREE_CONSTANT (var_init)))
1318 var_init = 0;
1319
1320 if (global_bindings_p () && var_init != 0 && ! init_const)
1321 {
1322 add_pending_elaborations (var_decl, var_init);
1323 var_init = 0;
1324 }
1325
1326 else if (var_init != 0
1327 && ((TYPE_MAIN_VARIANT (TREE_TYPE (var_init))
1328 != TYPE_MAIN_VARIANT (type))
1329 || (static_flag && ! init_const)))
1330 assign_init = var_init, var_init = 0;
1331
1332 DECL_COMMON (var_decl) = !flag_no_common;
1333 DECL_INITIAL (var_decl) = var_init;
1334 TREE_READONLY (var_decl) = const_flag;
1335 DECL_EXTERNAL (var_decl) = extern_flag;
1336 TREE_PUBLIC (var_decl) = public_flag || extern_flag;
1337 TREE_CONSTANT (var_decl) = TREE_CODE (var_decl) == CONST_DECL;
1338 TREE_THIS_VOLATILE (var_decl) = TREE_SIDE_EFFECTS (var_decl)
1339 = TYPE_VOLATILE (type);
1340
1341 /* At the global binding level we need to allocate static storage for the
1342 variable if and only if its not external. If we are not at the top level
1343 we allocate automatic storage unless requested not to. */
1344 TREE_STATIC (var_decl) = global_bindings_p () ? !extern_flag : static_flag;
1345
1346 if (asm_name != 0)
1347 SET_DECL_ASSEMBLER_NAME (var_decl, asm_name);
1348
1349 process_attributes (var_decl, attr_list);
1350
1351 /* Add this decl to the current binding level and generate any
1352 needed code and RTL. */
1353 var_decl = pushdecl (var_decl);
1354 expand_decl (var_decl);
1355
1356 if (DECL_CONTEXT (var_decl) != 0)
1357 expand_decl_init (var_decl);
1358
1359 /* If this is volatile, force it into memory. */
1360 if (TREE_SIDE_EFFECTS (var_decl))
1361 gnat_mark_addressable (var_decl);
1362
1363 if (TREE_CODE (var_decl) != CONST_DECL)
1364 rest_of_decl_compilation (var_decl, 0, global_bindings_p (), 0);
1365
1366 if (assign_init != 0)
1367 {
1368 /* If VAR_DECL has a padded type, convert it to the unpadded
1369 type so the assignment is done properly. */
1370 tree lhs = var_decl;
1371
1372 if (TREE_CODE (TREE_TYPE (lhs)) == RECORD_TYPE
1373 && TYPE_IS_PADDING_P (TREE_TYPE (lhs)))
1374 lhs = convert (TREE_TYPE (TYPE_FIELDS (TREE_TYPE (lhs))), lhs);
1375
1376 expand_expr_stmt (build_binary_op (MODIFY_EXPR, NULL_TREE, lhs,
1377 assign_init));
1378 }
1379
1380 return var_decl;
1381 }
1382 \f
1383 /* Returns a FIELD_DECL node. FIELD_NAME the field name, FIELD_TYPE is its
1384 type, and RECORD_TYPE is the type of the parent. PACKED is nonzero if
1385 this field is in a record type with a "pragma pack". If SIZE is nonzero
1386 it is the specified size for this field. If POS is nonzero, it is the bit
1387 position. If ADDRESSABLE is nonzero, it means we are allowed to take
1388 the address of this field for aliasing purposes. */
1389
1390 tree
1391 create_field_decl (tree field_name,
1392 tree field_type,
1393 tree record_type,
1394 int packed,
1395 tree size,
1396 tree pos,
1397 int addressable)
1398 {
1399 tree field_decl = build_decl (FIELD_DECL, field_name, field_type);
1400
1401 DECL_CONTEXT (field_decl) = record_type;
1402 TREE_READONLY (field_decl) = TREE_READONLY (field_type);
1403
1404 /* If FIELD_TYPE is BLKmode, we must ensure this is aligned to at least a
1405 byte boundary since GCC cannot handle less-aligned BLKmode bitfields. */
1406 if (packed && TYPE_MODE (field_type) == BLKmode)
1407 DECL_ALIGN (field_decl) = BITS_PER_UNIT;
1408
1409 /* If a size is specified, use it. Otherwise, if the record type is packed
1410 compute a size to use, which may differ from the object's natural size.
1411 We always set a size in this case to trigger the checks for bitfield
1412 creation below, which is typically required when no position has been
1413 specified. */
1414 if (size != 0)
1415 size = convert (bitsizetype, size);
1416 else if (packed == 1)
1417 {
1418 size = rm_size (field_type);
1419
1420 /* For a constant size larger than MAX_FIXED_MODE_SIZE, round up to
1421 byte. */
1422 if (TREE_CODE (size) == INTEGER_CST
1423 && compare_tree_int (size, MAX_FIXED_MODE_SIZE) > 0)
1424 size = round_up (size, BITS_PER_UNIT);
1425 }
1426
1427 /* Make a bitfield if a size is specified for two reasons: first if the size
1428 differs from the natural size. Second, if the alignment is insufficient.
1429 There are a number of ways the latter can be true.
1430
1431 We never make a bitfield if the type of the field has a nonconstant size,
1432 or if it is claimed to be addressable, because no such entity requiring
1433 bitfield operations should reach here.
1434
1435 We do *preventively* make a bitfield when there might be the need for it
1436 but we don't have all the necessary information to decide, as is the case
1437 of a field with no specified position in a packed record.
1438
1439 We also don't look at STRICT_ALIGNMENT here, and rely on later processing
1440 in layout_decl or finish_record_type to clear the bit_field indication if
1441 it is in fact not needed. */
1442 if (size != 0 && TREE_CODE (size) == INTEGER_CST
1443 && TREE_CODE (TYPE_SIZE (field_type)) == INTEGER_CST
1444 && ! addressable
1445 && (! operand_equal_p (TYPE_SIZE (field_type), size, 0)
1446 || (pos != 0
1447 && ! value_zerop (size_binop (TRUNC_MOD_EXPR, pos,
1448 bitsize_int (TYPE_ALIGN
1449 (field_type)))))
1450 || packed
1451 || (TYPE_ALIGN (record_type) != 0
1452 && TYPE_ALIGN (record_type) < TYPE_ALIGN (field_type))))
1453 {
1454 DECL_BIT_FIELD (field_decl) = 1;
1455 DECL_SIZE (field_decl) = size;
1456 if (! packed && pos == 0)
1457 DECL_ALIGN (field_decl)
1458 = (TYPE_ALIGN (record_type) != 0
1459 ? MIN (TYPE_ALIGN (record_type), TYPE_ALIGN (field_type))
1460 : TYPE_ALIGN (field_type));
1461 }
1462
1463 DECL_PACKED (field_decl) = pos != 0 ? DECL_BIT_FIELD (field_decl) : packed;
1464 DECL_ALIGN (field_decl)
1465 = MAX (DECL_ALIGN (field_decl),
1466 DECL_BIT_FIELD (field_decl) ? 1
1467 : packed && TYPE_MODE (field_type) != BLKmode ? BITS_PER_UNIT
1468 : TYPE_ALIGN (field_type));
1469
1470 if (pos != 0)
1471 {
1472 /* We need to pass in the alignment the DECL is known to have.
1473 This is the lowest-order bit set in POS, but no more than
1474 the alignment of the record, if one is specified. Note
1475 that an alignment of 0 is taken as infinite. */
1476 unsigned int known_align;
1477
1478 if (host_integerp (pos, 1))
1479 known_align = tree_low_cst (pos, 1) & - tree_low_cst (pos, 1);
1480 else
1481 known_align = BITS_PER_UNIT;
1482
1483 if (TYPE_ALIGN (record_type)
1484 && (known_align == 0 || known_align > TYPE_ALIGN (record_type)))
1485 known_align = TYPE_ALIGN (record_type);
1486
1487 layout_decl (field_decl, known_align);
1488 SET_DECL_OFFSET_ALIGN (field_decl,
1489 host_integerp (pos, 1) ? BIGGEST_ALIGNMENT
1490 : BITS_PER_UNIT);
1491 pos_from_bit (&DECL_FIELD_OFFSET (field_decl),
1492 &DECL_FIELD_BIT_OFFSET (field_decl),
1493 DECL_OFFSET_ALIGN (field_decl), pos);
1494
1495 DECL_HAS_REP_P (field_decl) = 1;
1496 }
1497
1498 /* If the field type is passed by reference, we will have pointers to the
1499 field, so it is addressable. */
1500 if (must_pass_by_ref (field_type) || default_pass_by_ref (field_type))
1501 addressable = 1;
1502
1503 /* ??? For now, we say that any field of aggregate type is addressable
1504 because the front end may take 'Reference of it. */
1505 if (AGGREGATE_TYPE_P (field_type))
1506 addressable = 1;
1507
1508 /* Mark the decl as nonaddressable if it is indicated so semantically,
1509 meaning we won't ever attempt to take the address of the field.
1510
1511 It may also be "technically" nonaddressable, meaning that even if we
1512 attempt to take the field's address we will actually get the address of a
1513 copy. This is the case for true bitfields, but the DECL_BIT_FIELD value
1514 we have at this point is not accurate enough, so we don't account for
1515 this here and let finish_record_type decide. */
1516 DECL_NONADDRESSABLE_P (field_decl) = ! addressable;
1517
1518 return field_decl;
1519 }
1520
1521 /* Subroutine of previous function: return nonzero if EXP, ignoring any side
1522 effects, has the value of zero. */
1523
1524 static int
1525 value_zerop (tree exp)
1526 {
1527 if (TREE_CODE (exp) == COMPOUND_EXPR)
1528 return value_zerop (TREE_OPERAND (exp, 1));
1529
1530 return integer_zerop (exp);
1531 }
1532 \f
1533 /* Returns a PARM_DECL node. PARAM_NAME is the name of the parameter,
1534 PARAM_TYPE is its type. READONLY is nonzero if the parameter is
1535 readonly (either an IN parameter or an address of a pass-by-ref
1536 parameter). */
1537
1538 tree
1539 create_param_decl (tree param_name, tree param_type, int readonly)
1540 {
1541 tree param_decl = build_decl (PARM_DECL, param_name, param_type);
1542
1543 /* Honor targetm.calls.promote_prototypes(), as not doing so can
1544 lead to various ABI violations. */
1545 if (targetm.calls.promote_prototypes (param_type)
1546 && (TREE_CODE (param_type) == INTEGER_TYPE
1547 || TREE_CODE (param_type) == ENUMERAL_TYPE)
1548 && TYPE_PRECISION (param_type) < TYPE_PRECISION (integer_type_node))
1549 {
1550 /* We have to be careful about biased types here. Make a subtype
1551 of integer_type_node with the proper biasing. */
1552 if (TREE_CODE (param_type) == INTEGER_TYPE
1553 && TYPE_BIASED_REPRESENTATION_P (param_type))
1554 {
1555 param_type
1556 = copy_type (build_range_type (integer_type_node,
1557 TYPE_MIN_VALUE (param_type),
1558 TYPE_MAX_VALUE (param_type)));
1559
1560 TYPE_BIASED_REPRESENTATION_P (param_type) = 1;
1561 }
1562 else
1563 param_type = integer_type_node;
1564 }
1565
1566 DECL_ARG_TYPE (param_decl) = param_type;
1567 DECL_ARG_TYPE_AS_WRITTEN (param_decl) = param_type;
1568 TREE_READONLY (param_decl) = readonly;
1569 return param_decl;
1570 }
1571 \f
1572 /* Given a DECL and ATTR_LIST, process the listed attributes. */
1573
1574 void
1575 process_attributes (tree decl, struct attrib *attr_list)
1576 {
1577 for (; attr_list; attr_list = attr_list->next)
1578 switch (attr_list->type)
1579 {
1580 case ATTR_MACHINE_ATTRIBUTE:
1581 decl_attributes (&decl, tree_cons (attr_list->name, attr_list->arg,
1582 NULL_TREE),
1583 ATTR_FLAG_TYPE_IN_PLACE);
1584 break;
1585
1586 case ATTR_LINK_ALIAS:
1587 TREE_STATIC (decl) = 1;
1588 assemble_alias (decl, attr_list->name);
1589 break;
1590
1591 case ATTR_WEAK_EXTERNAL:
1592 if (SUPPORTS_WEAK)
1593 declare_weak (decl);
1594 else
1595 post_error ("?weak declarations not supported on this target",
1596 attr_list->error_point);
1597 break;
1598
1599 case ATTR_LINK_SECTION:
1600 if (targetm.have_named_sections)
1601 {
1602 DECL_SECTION_NAME (decl)
1603 = build_string (IDENTIFIER_LENGTH (attr_list->name),
1604 IDENTIFIER_POINTER (attr_list->name));
1605 DECL_COMMON (decl) = 0;
1606 }
1607 else
1608 post_error ("?section attributes are not supported for this target",
1609 attr_list->error_point);
1610 break;
1611 }
1612 }
1613 \f
1614 /* Add some pending elaborations on the list. */
1615
1616 void
1617 add_pending_elaborations (tree var_decl, tree var_init)
1618 {
1619 if (var_init != 0)
1620 Check_Elaboration_Code_Allowed (error_gnat_node);
1621
1622 pending_elaborations
1623 = chainon (pending_elaborations, build_tree_list (var_decl, var_init));
1624 }
1625
1626 /* Obtain any pending elaborations and clear the old list. */
1627
1628 tree
1629 get_pending_elaborations (void)
1630 {
1631 /* Each thing added to the list went on the end; we want it on the
1632 beginning. */
1633 tree result = TREE_CHAIN (pending_elaborations);
1634
1635 TREE_CHAIN (pending_elaborations) = 0;
1636 return result;
1637 }
1638
1639 /* Return true if VALUE is a multiple of FACTOR. FACTOR must be a power
1640 of 2. */
1641
1642 static int
1643 value_factor_p (tree value, int factor)
1644 {
1645 if (host_integerp (value, 1))
1646 return tree_low_cst (value, 1) % factor == 0;
1647
1648 if (TREE_CODE (value) == MULT_EXPR)
1649 return (value_factor_p (TREE_OPERAND (value, 0), factor)
1650 || value_factor_p (TREE_OPERAND (value, 1), factor));
1651
1652 return 0;
1653 }
1654
1655 /* Given 2 consecutive field decls PREV_FIELD and CURR_FIELD, return true
1656 unless we can prove these 2 fields are laid out in such a way that no gap
1657 exist between the end of PREV_FIELD and the begining of CURR_FIELD. OFFSET
1658 is the distance in bits between the end of PREV_FIELD and the starting
1659 position of CURR_FIELD. It is ignored if null. */
1660
1661 static int
1662 potential_alignment_gap (tree prev_field, tree curr_field, tree offset)
1663 {
1664 /* If this is the first field of the record, there cannot be any gap */
1665 if (!prev_field)
1666 return 0;
1667
1668 /* If the previous field is a union type, then return False: The only
1669 time when such a field is not the last field of the record is when
1670 there are other components at fixed positions after it (meaning there
1671 was a rep clause for every field), in which case we don't want the
1672 alignment constraint to override them. */
1673 if (TREE_CODE (TREE_TYPE (prev_field)) == QUAL_UNION_TYPE)
1674 return 0;
1675
1676 /* If the distance between the end of prev_field and the begining of
1677 curr_field is constant, then there is a gap if the value of this
1678 constant is not null. */
1679 if (offset && host_integerp (offset, 1))
1680 return (!integer_zerop (offset));
1681
1682 /* If the size and position of the previous field are constant,
1683 then check the sum of this size and position. There will be a gap
1684 iff it is not multiple of the current field alignment. */
1685 if (host_integerp (DECL_SIZE (prev_field), 1)
1686 && host_integerp (bit_position (prev_field), 1))
1687 return ((tree_low_cst (bit_position (prev_field), 1)
1688 + tree_low_cst (DECL_SIZE (prev_field), 1))
1689 % DECL_ALIGN (curr_field) != 0);
1690
1691 /* If both the position and size of the previous field are multiples
1692 of the current field alignment, there can not be any gap. */
1693 if (value_factor_p (bit_position (prev_field), DECL_ALIGN (curr_field))
1694 && value_factor_p (DECL_SIZE (prev_field), DECL_ALIGN (curr_field)))
1695 return 0;
1696
1697 /* Fallback, return that there may be a potential gap */
1698 return 1;
1699 }
1700
1701 /* Return nonzero if there are pending elaborations. */
1702
1703 int
1704 pending_elaborations_p (void)
1705 {
1706 return TREE_CHAIN (pending_elaborations) != 0;
1707 }
1708
1709 /* Save a copy of the current pending elaboration list and make a new
1710 one. */
1711
1712 void
1713 push_pending_elaborations (void)
1714 {
1715 struct e_stack *p = (struct e_stack *) ggc_alloc (sizeof (struct e_stack));
1716
1717 p->next = elist_stack;
1718 p->elab_list = pending_elaborations;
1719 elist_stack = p;
1720 pending_elaborations = build_tree_list (NULL_TREE, NULL_TREE);
1721 }
1722
1723 /* Pop the stack of pending elaborations. */
1724
1725 void
1726 pop_pending_elaborations (void)
1727 {
1728 struct e_stack *p = elist_stack;
1729
1730 pending_elaborations = p->elab_list;
1731 elist_stack = p->next;
1732 }
1733
1734 /* Return the current position in pending_elaborations so we can insert
1735 elaborations after that point. */
1736
1737 tree
1738 get_elaboration_location (void)
1739 {
1740 return tree_last (pending_elaborations);
1741 }
1742
1743 /* Insert the current elaborations after ELAB, which is in some elaboration
1744 list. */
1745
1746 void
1747 insert_elaboration_list (tree elab)
1748 {
1749 tree next = TREE_CHAIN (elab);
1750
1751 if (TREE_CHAIN (pending_elaborations))
1752 {
1753 TREE_CHAIN (elab) = TREE_CHAIN (pending_elaborations);
1754 TREE_CHAIN (tree_last (pending_elaborations)) = next;
1755 TREE_CHAIN (pending_elaborations) = 0;
1756 }
1757 }
1758
1759 /* Returns a LABEL_DECL node for LABEL_NAME. */
1760
1761 tree
1762 create_label_decl (tree label_name)
1763 {
1764 tree label_decl = build_decl (LABEL_DECL, label_name, void_type_node);
1765
1766 DECL_CONTEXT (label_decl) = current_function_decl;
1767 DECL_MODE (label_decl) = VOIDmode;
1768 DECL_SOURCE_LOCATION (label_decl) = input_location;
1769
1770 return label_decl;
1771 }
1772 \f
1773 /* Returns a FUNCTION_DECL node. SUBPROG_NAME is the name of the subprogram,
1774 ASM_NAME is its assembler name, SUBPROG_TYPE is its type (a FUNCTION_TYPE
1775 node), PARAM_DECL_LIST is the list of the subprogram arguments (a list of
1776 PARM_DECL nodes chained through the TREE_CHAIN field).
1777
1778 INLINE_FLAG, PUBLIC_FLAG, EXTERN_FLAG, and ATTR_LIST are used to set the
1779 appropriate fields in the FUNCTION_DECL. */
1780
1781 tree
1782 create_subprog_decl (tree subprog_name,
1783 tree asm_name,
1784 tree subprog_type,
1785 tree param_decl_list,
1786 int inline_flag,
1787 int public_flag,
1788 int extern_flag,
1789 struct attrib *attr_list)
1790 {
1791 tree return_type = TREE_TYPE (subprog_type);
1792 tree subprog_decl = build_decl (FUNCTION_DECL, subprog_name, subprog_type);
1793
1794 /* If this is a function nested inside an inlined external function, it
1795 means we aren't going to compile the outer function unless it is
1796 actually inlined, so do the same for us. */
1797 if (current_function_decl != 0 && DECL_INLINE (current_function_decl)
1798 && DECL_EXTERNAL (current_function_decl))
1799 extern_flag = 1;
1800
1801 DECL_EXTERNAL (subprog_decl) = extern_flag;
1802 TREE_PUBLIC (subprog_decl) = public_flag;
1803 DECL_INLINE (subprog_decl) = inline_flag;
1804 TREE_READONLY (subprog_decl) = TYPE_READONLY (subprog_type);
1805 TREE_THIS_VOLATILE (subprog_decl) = TYPE_VOLATILE (subprog_type);
1806 TREE_SIDE_EFFECTS (subprog_decl) = TYPE_VOLATILE (subprog_type);
1807 DECL_ARGUMENTS (subprog_decl) = param_decl_list;
1808 DECL_RESULT (subprog_decl) = build_decl (RESULT_DECL, 0, return_type);
1809
1810 if (asm_name != 0)
1811 SET_DECL_ASSEMBLER_NAME (subprog_decl, asm_name);
1812
1813 process_attributes (subprog_decl, attr_list);
1814
1815 /* Add this decl to the current binding level. */
1816 subprog_decl = pushdecl (subprog_decl);
1817
1818 /* Output the assembler code and/or RTL for the declaration. */
1819 rest_of_decl_compilation (subprog_decl, 0, global_bindings_p (), 0);
1820
1821 return subprog_decl;
1822 }
1823 \f
1824 /* Count how deep we are into nested functions. This is because
1825 we shouldn't call the backend function context routines unless we
1826 are in a nested function. */
1827
1828 static int function_nesting_depth;
1829
1830 /* Set up the framework for generating code for SUBPROG_DECL, a subprogram
1831 body. This routine needs to be invoked before processing the declarations
1832 appearing in the subprogram. */
1833
1834 void
1835 begin_subprog_body (tree subprog_decl)
1836 {
1837 tree param_decl_list;
1838 tree param_decl;
1839 tree next_param;
1840
1841 if (function_nesting_depth++ != 0)
1842 push_function_context ();
1843
1844 announce_function (subprog_decl);
1845
1846 /* Make this field nonzero so further routines know that this is not
1847 tentative. error_mark_node is replaced below (in poplevel) with the
1848 adequate BLOCK. */
1849 DECL_INITIAL (subprog_decl) = error_mark_node;
1850
1851 /* This function exists in static storage. This does not mean `static' in
1852 the C sense! */
1853 TREE_STATIC (subprog_decl) = 1;
1854
1855 /* Enter a new binding level. */
1856 current_function_decl = subprog_decl;
1857 pushlevel (0);
1858
1859 /* Push all the PARM_DECL nodes onto the current scope (i.e. the scope of the
1860 subprogram body) so that they can be recognized as local variables in the
1861 subprogram.
1862
1863 The list of PARM_DECL nodes is stored in the right order in
1864 DECL_ARGUMENTS. Since ..._DECL nodes get stored in the reverse order in
1865 which they are transmitted to `pushdecl' we need to reverse the list of
1866 PARM_DECLs if we want it to be stored in the right order. The reason why
1867 we want to make sure the PARM_DECLs are stored in the correct order is
1868 that this list will be retrieved in a few lines with a call to `getdecl'
1869 to store it back into the DECL_ARGUMENTS field. */
1870 param_decl_list = nreverse (DECL_ARGUMENTS (subprog_decl));
1871
1872 for (param_decl = param_decl_list; param_decl; param_decl = next_param)
1873 {
1874 next_param = TREE_CHAIN (param_decl);
1875 TREE_CHAIN (param_decl) = NULL;
1876 pushdecl (param_decl);
1877 }
1878
1879 /* Store back the PARM_DECL nodes. They appear in the right order. */
1880 DECL_ARGUMENTS (subprog_decl) = getdecls ();
1881
1882 init_function_start (subprog_decl);
1883 expand_function_start (subprog_decl, 0);
1884
1885 /* If this function is `main', emit a call to `__main'
1886 to run global initializers, etc. */
1887 if (DECL_ASSEMBLER_NAME (subprog_decl) != 0
1888 && MAIN_NAME_P (DECL_ASSEMBLER_NAME (subprog_decl))
1889 && DECL_CONTEXT (subprog_decl) == NULL_TREE)
1890 expand_main_function ();
1891 }
1892
1893 /* Finish the definition of the current subprogram and compile it all the way
1894 to assembler language output. */
1895
1896 void
1897 end_subprog_body (void)
1898 {
1899 tree decl;
1900 tree cico_list;
1901
1902 poplevel (1, 0, 1);
1903 BLOCK_SUPERCONTEXT (DECL_INITIAL (current_function_decl))
1904 = current_function_decl;
1905
1906 /* Mark the RESULT_DECL as being in this subprogram. */
1907 DECL_CONTEXT (DECL_RESULT (current_function_decl)) = current_function_decl;
1908
1909 expand_function_end ();
1910
1911 /* If this is a nested function, push a new GC context. That will keep
1912 local variables on the stack from being collected while we're doing
1913 the compilation of this function. */
1914 if (function_nesting_depth > 1)
1915 ggc_push_context ();
1916
1917 /* If we're only annotating types, don't actually compile this
1918 function. */
1919 if (!type_annotate_only)
1920 rest_of_compilation (current_function_decl);
1921
1922 if (function_nesting_depth > 1)
1923 ggc_pop_context ();
1924
1925 /* Throw away any VAR_DECLs we made for OUT parameters; they must
1926 not be seen when we call this function and will be in
1927 unallocated memory anyway. */
1928 for (cico_list = TYPE_CI_CO_LIST (TREE_TYPE (current_function_decl));
1929 cico_list != 0; cico_list = TREE_CHAIN (cico_list))
1930 TREE_VALUE (cico_list) = 0;
1931
1932 if (DECL_STRUCT_FUNCTION (current_function_decl) == 0)
1933 {
1934 /* Throw away DECL_RTL in any PARM_DECLs unless this function
1935 was saved for inline, in which case the DECL_RTLs are in
1936 preserved memory. */
1937 for (decl = DECL_ARGUMENTS (current_function_decl);
1938 decl != 0; decl = TREE_CHAIN (decl))
1939 {
1940 SET_DECL_RTL (decl, 0);
1941 DECL_INCOMING_RTL (decl) = 0;
1942 }
1943
1944 /* Similarly, discard DECL_RTL of the return value. */
1945 SET_DECL_RTL (DECL_RESULT (current_function_decl), 0);
1946
1947 /* But DECL_INITIAL must remain nonzero so we know this
1948 was an actual function definition unless toplev.c decided not
1949 to inline it. */
1950 if (DECL_INITIAL (current_function_decl) != 0)
1951 DECL_INITIAL (current_function_decl) = error_mark_node;
1952
1953 DECL_ARGUMENTS (current_function_decl) = 0;
1954 }
1955
1956 /* If we are not at the bottom of the function nesting stack, pop up to
1957 the containing function. Otherwise show we aren't in any function. */
1958 if (--function_nesting_depth != 0)
1959 pop_function_context ();
1960 else
1961 current_function_decl = 0;
1962 }
1963 \f
1964 /* Return a definition for a builtin function named NAME and whose data type
1965 is TYPE. TYPE should be a function type with argument types.
1966 FUNCTION_CODE tells later passes how to compile calls to this function.
1967 See tree.h for its possible values.
1968
1969 If LIBRARY_NAME is nonzero, use that for DECL_ASSEMBLER_NAME,
1970 the name to be called if we can't opencode the function. If
1971 ATTRS is nonzero, use that for the function attribute list. */
1972
1973 tree
1974 builtin_function (const char *name,
1975 tree type,
1976 int function_code,
1977 enum built_in_class class,
1978 const char *library_name,
1979 tree attrs)
1980 {
1981 tree decl = build_decl (FUNCTION_DECL, get_identifier (name), type);
1982
1983 DECL_EXTERNAL (decl) = 1;
1984 TREE_PUBLIC (decl) = 1;
1985 if (library_name)
1986 SET_DECL_ASSEMBLER_NAME (decl, get_identifier (library_name));
1987
1988 pushdecl (decl);
1989 DECL_BUILT_IN_CLASS (decl) = class;
1990 DECL_FUNCTION_CODE (decl) = function_code;
1991 if (attrs)
1992 decl_attributes (&decl, attrs, ATTR_FLAG_BUILT_IN);
1993 return decl;
1994 }
1995
1996 /* Return an integer type with the number of bits of precision given by
1997 PRECISION. UNSIGNEDP is nonzero if the type is unsigned; otherwise
1998 it is a signed type. */
1999
2000 tree
2001 gnat_type_for_size (unsigned precision, int unsignedp)
2002 {
2003 tree t;
2004 char type_name[20];
2005
2006 if (precision <= 2 * MAX_BITS_PER_WORD
2007 && signed_and_unsigned_types[precision][unsignedp] != 0)
2008 return signed_and_unsigned_types[precision][unsignedp];
2009
2010 if (unsignedp)
2011 t = make_unsigned_type (precision);
2012 else
2013 t = make_signed_type (precision);
2014
2015 if (precision <= 2 * MAX_BITS_PER_WORD)
2016 signed_and_unsigned_types[precision][unsignedp] = t;
2017
2018 if (TYPE_NAME (t) == 0)
2019 {
2020 sprintf (type_name, "%sSIGNED_%d", unsignedp ? "UN" : "", precision);
2021 TYPE_NAME (t) = get_identifier (type_name);
2022 }
2023
2024 return t;
2025 }
2026
2027 /* Likewise for floating-point types. */
2028
2029 static tree
2030 float_type_for_precision (int precision, enum machine_mode mode)
2031 {
2032 tree t;
2033 char type_name[20];
2034
2035 if (float_types[(int) mode] != 0)
2036 return float_types[(int) mode];
2037
2038 float_types[(int) mode] = t = make_node (REAL_TYPE);
2039 TYPE_PRECISION (t) = precision;
2040 layout_type (t);
2041
2042 if (TYPE_MODE (t) != mode)
2043 gigi_abort (414);
2044
2045 if (TYPE_NAME (t) == 0)
2046 {
2047 sprintf (type_name, "FLOAT_%d", precision);
2048 TYPE_NAME (t) = get_identifier (type_name);
2049 }
2050
2051 return t;
2052 }
2053
2054 /* Return a data type that has machine mode MODE. UNSIGNEDP selects
2055 an unsigned type; otherwise a signed type is returned. */
2056
2057 tree
2058 gnat_type_for_mode (enum machine_mode mode, int unsignedp)
2059 {
2060 if (GET_MODE_CLASS (mode) == MODE_FLOAT)
2061 return float_type_for_precision (GET_MODE_PRECISION (mode), mode);
2062 else
2063 return gnat_type_for_size (GET_MODE_BITSIZE (mode), unsignedp);
2064 }
2065
2066 /* Return the unsigned version of a TYPE_NODE, a scalar type. */
2067
2068 tree
2069 gnat_unsigned_type (tree type_node)
2070 {
2071 tree type = gnat_type_for_size (TYPE_PRECISION (type_node), 1);
2072
2073 if (TREE_CODE (type_node) == INTEGER_TYPE && TYPE_MODULAR_P (type_node))
2074 {
2075 type = copy_node (type);
2076 TREE_TYPE (type) = type_node;
2077 }
2078 else if (TREE_TYPE (type_node) != 0
2079 && TREE_CODE (TREE_TYPE (type_node)) == INTEGER_TYPE
2080 && TYPE_MODULAR_P (TREE_TYPE (type_node)))
2081 {
2082 type = copy_node (type);
2083 TREE_TYPE (type) = TREE_TYPE (type_node);
2084 }
2085
2086 return type;
2087 }
2088
2089 /* Return the signed version of a TYPE_NODE, a scalar type. */
2090
2091 tree
2092 gnat_signed_type (tree type_node)
2093 {
2094 tree type = gnat_type_for_size (TYPE_PRECISION (type_node), 0);
2095
2096 if (TREE_CODE (type_node) == INTEGER_TYPE && TYPE_MODULAR_P (type_node))
2097 {
2098 type = copy_node (type);
2099 TREE_TYPE (type) = type_node;
2100 }
2101 else if (TREE_TYPE (type_node) != 0
2102 && TREE_CODE (TREE_TYPE (type_node)) == INTEGER_TYPE
2103 && TYPE_MODULAR_P (TREE_TYPE (type_node)))
2104 {
2105 type = copy_node (type);
2106 TREE_TYPE (type) = TREE_TYPE (type_node);
2107 }
2108
2109 return type;
2110 }
2111
2112 /* Return a type the same as TYPE except unsigned or signed according to
2113 UNSIGNEDP. */
2114
2115 tree
2116 gnat_signed_or_unsigned_type (int unsignedp, tree type)
2117 {
2118 if (! INTEGRAL_TYPE_P (type) || TREE_UNSIGNED (type) == unsignedp)
2119 return type;
2120 else
2121 return gnat_type_for_size (TYPE_PRECISION (type), unsignedp);
2122 }
2123 \f
2124 /* EXP is an expression for the size of an object. If this size contains
2125 discriminant references, replace them with the maximum (if MAX_P) or
2126 minimum (if ! MAX_P) possible value of the discriminant. */
2127
2128 tree
2129 max_size (tree exp, int max_p)
2130 {
2131 enum tree_code code = TREE_CODE (exp);
2132 tree type = TREE_TYPE (exp);
2133
2134 switch (TREE_CODE_CLASS (code))
2135 {
2136 case 'd':
2137 case 'c':
2138 return exp;
2139
2140 case 'x':
2141 if (code == TREE_LIST)
2142 return tree_cons (TREE_PURPOSE (exp),
2143 max_size (TREE_VALUE (exp), max_p),
2144 TREE_CHAIN (exp) != 0
2145 ? max_size (TREE_CHAIN (exp), max_p) : 0);
2146 break;
2147
2148 case 'r':
2149 /* If this contains a PLACEHOLDER_EXPR, it is the thing we want to
2150 modify. Otherwise, we treat it like a variable. */
2151 if (! CONTAINS_PLACEHOLDER_P (exp))
2152 return exp;
2153
2154 type = TREE_TYPE (TREE_OPERAND (exp, 1));
2155 return
2156 max_size (max_p ? TYPE_MAX_VALUE (type) : TYPE_MIN_VALUE (type), 1);
2157
2158 case '<':
2159 return max_p ? size_one_node : size_zero_node;
2160
2161 case '1':
2162 case '2':
2163 case 'e':
2164 switch (TREE_CODE_LENGTH (code))
2165 {
2166 case 1:
2167 if (code == NON_LVALUE_EXPR)
2168 return max_size (TREE_OPERAND (exp, 0), max_p);
2169 else
2170 return
2171 fold (build1 (code, type,
2172 max_size (TREE_OPERAND (exp, 0),
2173 code == NEGATE_EXPR ? ! max_p : max_p)));
2174
2175 case 2:
2176 if (code == RTL_EXPR)
2177 gigi_abort (407);
2178 else if (code == COMPOUND_EXPR)
2179 return max_size (TREE_OPERAND (exp, 1), max_p);
2180 else if (code == WITH_RECORD_EXPR)
2181 return exp;
2182
2183 {
2184 tree lhs = max_size (TREE_OPERAND (exp, 0), max_p);
2185 tree rhs = max_size (TREE_OPERAND (exp, 1),
2186 code == MINUS_EXPR ? ! max_p : max_p);
2187
2188 /* Special-case wanting the maximum value of a MIN_EXPR.
2189 In that case, if one side overflows, return the other.
2190 sizetype is signed, but we know sizes are non-negative.
2191 Likewise, handle a MINUS_EXPR or PLUS_EXPR with the LHS
2192 overflowing or the maximum possible value and the RHS
2193 a variable. */
2194 if (max_p && code == MIN_EXPR && TREE_OVERFLOW (rhs))
2195 return lhs;
2196 else if (max_p && code == MIN_EXPR && TREE_OVERFLOW (lhs))
2197 return rhs;
2198 else if ((code == MINUS_EXPR || code == PLUS_EXPR)
2199 && ((TREE_CONSTANT (lhs) && TREE_OVERFLOW (lhs))
2200 || operand_equal_p (lhs, TYPE_MAX_VALUE (type), 0))
2201 && ! TREE_CONSTANT (rhs))
2202 return lhs;
2203 else
2204 return fold (build (code, type, lhs, rhs));
2205 }
2206
2207 case 3:
2208 if (code == SAVE_EXPR)
2209 return exp;
2210 else if (code == COND_EXPR)
2211 return fold (build (MAX_EXPR, type,
2212 max_size (TREE_OPERAND (exp, 1), max_p),
2213 max_size (TREE_OPERAND (exp, 2), max_p)));
2214 else if (code == CALL_EXPR && TREE_OPERAND (exp, 1) != 0)
2215 return build (CALL_EXPR, type, TREE_OPERAND (exp, 0),
2216 max_size (TREE_OPERAND (exp, 1), max_p));
2217 }
2218 }
2219
2220 gigi_abort (408);
2221 }
2222 \f
2223 /* Build a template of type TEMPLATE_TYPE from the array bounds of ARRAY_TYPE.
2224 EXPR is an expression that we can use to locate any PLACEHOLDER_EXPRs.
2225 Return a constructor for the template. */
2226
2227 tree
2228 build_template (tree template_type, tree array_type, tree expr)
2229 {
2230 tree template_elts = NULL_TREE;
2231 tree bound_list = NULL_TREE;
2232 tree field;
2233
2234 if (TREE_CODE (array_type) == RECORD_TYPE
2235 && (TYPE_IS_PADDING_P (array_type)
2236 || TYPE_LEFT_JUSTIFIED_MODULAR_P (array_type)))
2237 array_type = TREE_TYPE (TYPE_FIELDS (array_type));
2238
2239 if (TREE_CODE (array_type) == ARRAY_TYPE
2240 || (TREE_CODE (array_type) == INTEGER_TYPE
2241 && TYPE_HAS_ACTUAL_BOUNDS_P (array_type)))
2242 bound_list = TYPE_ACTUAL_BOUNDS (array_type);
2243
2244 /* First make the list for a CONSTRUCTOR for the template. Go down the
2245 field list of the template instead of the type chain because this
2246 array might be an Ada array of arrays and we can't tell where the
2247 nested arrays stop being the underlying object. */
2248
2249 for (field = TYPE_FIELDS (template_type); field;
2250 (bound_list != 0
2251 ? (bound_list = TREE_CHAIN (bound_list))
2252 : (array_type = TREE_TYPE (array_type))),
2253 field = TREE_CHAIN (TREE_CHAIN (field)))
2254 {
2255 tree bounds, min, max;
2256
2257 /* If we have a bound list, get the bounds from there. Likewise
2258 for an ARRAY_TYPE. Otherwise, if expr is a PARM_DECL with
2259 DECL_BY_COMPONENT_PTR_P, use the bounds of the field in the template.
2260 This will give us a maximum range. */
2261 if (bound_list != 0)
2262 bounds = TREE_VALUE (bound_list);
2263 else if (TREE_CODE (array_type) == ARRAY_TYPE)
2264 bounds = TYPE_INDEX_TYPE (TYPE_DOMAIN (array_type));
2265 else if (expr != 0 && TREE_CODE (expr) == PARM_DECL
2266 && DECL_BY_COMPONENT_PTR_P (expr))
2267 bounds = TREE_TYPE (field);
2268 else
2269 gigi_abort (411);
2270
2271 min = convert (TREE_TYPE (TREE_CHAIN (field)), TYPE_MIN_VALUE (bounds));
2272 max = convert (TREE_TYPE (field), TYPE_MAX_VALUE (bounds));
2273
2274 /* If either MIN or MAX involve a PLACEHOLDER_EXPR, we must
2275 surround them with a WITH_RECORD_EXPR giving EXPR as the
2276 OBJECT. */
2277 if (CONTAINS_PLACEHOLDER_P (min))
2278 min = build (WITH_RECORD_EXPR, TREE_TYPE (min), min, expr);
2279 if (CONTAINS_PLACEHOLDER_P (max))
2280 max = build (WITH_RECORD_EXPR, TREE_TYPE (max), max, expr);
2281
2282 template_elts = tree_cons (TREE_CHAIN (field), max,
2283 tree_cons (field, min, template_elts));
2284 }
2285
2286 return gnat_build_constructor (template_type, nreverse (template_elts));
2287 }
2288 \f
2289 /* Build a VMS descriptor from a Mechanism_Type, which must specify
2290 a descriptor type, and the GCC type of an object. Each FIELD_DECL
2291 in the type contains in its DECL_INITIAL the expression to use when
2292 a constructor is made for the type. GNAT_ENTITY is a gnat node used
2293 to print out an error message if the mechanism cannot be applied to
2294 an object of that type and also for the name. */
2295
2296 tree
2297 build_vms_descriptor (tree type, Mechanism_Type mech, Entity_Id gnat_entity)
2298 {
2299 tree record_type = make_node (RECORD_TYPE);
2300 tree field_list = 0;
2301 int class;
2302 int dtype = 0;
2303 tree inner_type;
2304 int ndim;
2305 int i;
2306 tree *idx_arr;
2307 tree tem;
2308
2309 /* If TYPE is an unconstrained array, use the underlying array type. */
2310 if (TREE_CODE (type) == UNCONSTRAINED_ARRAY_TYPE)
2311 type = TREE_TYPE (TREE_TYPE (TYPE_FIELDS (TREE_TYPE (type))));
2312
2313 /* If this is an array, compute the number of dimensions in the array,
2314 get the index types, and point to the inner type. */
2315 if (TREE_CODE (type) != ARRAY_TYPE)
2316 ndim = 0;
2317 else
2318 for (ndim = 1, inner_type = type;
2319 TREE_CODE (TREE_TYPE (inner_type)) == ARRAY_TYPE
2320 && TYPE_MULTI_ARRAY_P (TREE_TYPE (inner_type));
2321 ndim++, inner_type = TREE_TYPE (inner_type))
2322 ;
2323
2324 idx_arr = (tree *) alloca (ndim * sizeof (tree));
2325
2326 if (mech != By_Descriptor_NCA
2327 && TREE_CODE (type) == ARRAY_TYPE && TYPE_CONVENTION_FORTRAN_P (type))
2328 for (i = ndim - 1, inner_type = type;
2329 i >= 0;
2330 i--, inner_type = TREE_TYPE (inner_type))
2331 idx_arr[i] = TYPE_DOMAIN (inner_type);
2332 else
2333 for (i = 0, inner_type = type;
2334 i < ndim;
2335 i++, inner_type = TREE_TYPE (inner_type))
2336 idx_arr[i] = TYPE_DOMAIN (inner_type);
2337
2338 /* Now get the DTYPE value. */
2339 switch (TREE_CODE (type))
2340 {
2341 case INTEGER_TYPE:
2342 case ENUMERAL_TYPE:
2343 if (TYPE_VAX_FLOATING_POINT_P (type))
2344 switch (tree_low_cst (TYPE_DIGITS_VALUE (type), 1))
2345 {
2346 case 6:
2347 dtype = 10;
2348 break;
2349 case 9:
2350 dtype = 11;
2351 break;
2352 case 15:
2353 dtype = 27;
2354 break;
2355 }
2356 else
2357 switch (GET_MODE_BITSIZE (TYPE_MODE (type)))
2358 {
2359 case 8:
2360 dtype = TREE_UNSIGNED (type) ? 2 : 6;
2361 break;
2362 case 16:
2363 dtype = TREE_UNSIGNED (type) ? 3 : 7;
2364 break;
2365 case 32:
2366 dtype = TREE_UNSIGNED (type) ? 4 : 8;
2367 break;
2368 case 64:
2369 dtype = TREE_UNSIGNED (type) ? 5 : 9;
2370 break;
2371 case 128:
2372 dtype = TREE_UNSIGNED (type) ? 25 : 26;
2373 break;
2374 }
2375 break;
2376
2377 case REAL_TYPE:
2378 dtype = GET_MODE_BITSIZE (TYPE_MODE (type)) == 32 ? 52 : 53;
2379 break;
2380
2381 case COMPLEX_TYPE:
2382 if (TREE_CODE (TREE_TYPE (type)) == INTEGER_TYPE
2383 && TYPE_VAX_FLOATING_POINT_P (type))
2384 switch (tree_low_cst (TYPE_DIGITS_VALUE (type), 1))
2385 {
2386 case 6:
2387 dtype = 12;
2388 break;
2389 case 9:
2390 dtype = 13;
2391 break;
2392 case 15:
2393 dtype = 29;
2394 }
2395 else
2396 dtype = GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (type))) == 32 ? 54: 55;
2397 break;
2398
2399 case ARRAY_TYPE:
2400 dtype = 14;
2401 break;
2402
2403 default:
2404 break;
2405 }
2406
2407 /* Get the CLASS value. */
2408 switch (mech)
2409 {
2410 case By_Descriptor_A:
2411 class = 4;
2412 break;
2413 case By_Descriptor_NCA:
2414 class = 10;
2415 break;
2416 case By_Descriptor_SB:
2417 class = 15;
2418 break;
2419 default:
2420 class = 1;
2421 }
2422
2423 /* Make the type for a descriptor for VMS. The first four fields
2424 are the same for all types. */
2425
2426 field_list
2427 = chainon (field_list,
2428 make_descriptor_field
2429 ("LENGTH", gnat_type_for_size (16, 1), record_type,
2430 size_in_bytes (mech == By_Descriptor_A ? inner_type : type)));
2431
2432 field_list = chainon (field_list,
2433 make_descriptor_field ("DTYPE",
2434 gnat_type_for_size (8, 1),
2435 record_type, size_int (dtype)));
2436 field_list = chainon (field_list,
2437 make_descriptor_field ("CLASS",
2438 gnat_type_for_size (8, 1),
2439 record_type, size_int (class)));
2440
2441 field_list
2442 = chainon (field_list,
2443 make_descriptor_field ("POINTER",
2444 build_pointer_type (type),
2445 record_type,
2446 build1 (ADDR_EXPR,
2447 build_pointer_type (type),
2448 build (PLACEHOLDER_EXPR,
2449 type))));
2450
2451 switch (mech)
2452 {
2453 case By_Descriptor:
2454 case By_Descriptor_S:
2455 break;
2456
2457 case By_Descriptor_SB:
2458 field_list
2459 = chainon (field_list,
2460 make_descriptor_field
2461 ("SB_L1", gnat_type_for_size (32, 1), record_type,
2462 TREE_CODE (type) == ARRAY_TYPE
2463 ? TYPE_MIN_VALUE (TYPE_DOMAIN (type)) : size_zero_node));
2464 field_list
2465 = chainon (field_list,
2466 make_descriptor_field
2467 ("SB_L2", gnat_type_for_size (32, 1), record_type,
2468 TREE_CODE (type) == ARRAY_TYPE
2469 ? TYPE_MAX_VALUE (TYPE_DOMAIN (type)) : size_zero_node));
2470 break;
2471
2472 case By_Descriptor_A:
2473 case By_Descriptor_NCA:
2474 field_list = chainon (field_list,
2475 make_descriptor_field ("SCALE",
2476 gnat_type_for_size (8, 1),
2477 record_type,
2478 size_zero_node));
2479
2480 field_list = chainon (field_list,
2481 make_descriptor_field ("DIGITS",
2482 gnat_type_for_size (8, 1),
2483 record_type,
2484 size_zero_node));
2485
2486 field_list
2487 = chainon (field_list,
2488 make_descriptor_field
2489 ("AFLAGS", gnat_type_for_size (8, 1), record_type,
2490 size_int (mech == By_Descriptor_NCA
2491 ? 0
2492 /* Set FL_COLUMN, FL_COEFF, and FL_BOUNDS. */
2493 : (TREE_CODE (type) == ARRAY_TYPE
2494 && TYPE_CONVENTION_FORTRAN_P (type)
2495 ? 224 : 192))));
2496
2497 field_list = chainon (field_list,
2498 make_descriptor_field ("DIMCT",
2499 gnat_type_for_size (8, 1),
2500 record_type,
2501 size_int (ndim)));
2502
2503 field_list = chainon (field_list,
2504 make_descriptor_field ("ARSIZE",
2505 gnat_type_for_size (32, 1),
2506 record_type,
2507 size_in_bytes (type)));
2508
2509 /* Now build a pointer to the 0,0,0... element. */
2510 tem = build (PLACEHOLDER_EXPR, type);
2511 for (i = 0, inner_type = type; i < ndim;
2512 i++, inner_type = TREE_TYPE (inner_type))
2513 tem = build (ARRAY_REF, TREE_TYPE (inner_type), tem,
2514 convert (TYPE_DOMAIN (inner_type), size_zero_node));
2515
2516 field_list
2517 = chainon (field_list,
2518 make_descriptor_field
2519 ("A0", build_pointer_type (inner_type), record_type,
2520 build1 (ADDR_EXPR, build_pointer_type (inner_type), tem)));
2521
2522 /* Next come the addressing coefficients. */
2523 tem = size_int (1);
2524 for (i = 0; i < ndim; i++)
2525 {
2526 char fname[3];
2527 tree idx_length
2528 = size_binop (MULT_EXPR, tem,
2529 size_binop (PLUS_EXPR,
2530 size_binop (MINUS_EXPR,
2531 TYPE_MAX_VALUE (idx_arr[i]),
2532 TYPE_MIN_VALUE (idx_arr[i])),
2533 size_int (1)));
2534
2535 fname[0] = (mech == By_Descriptor_NCA ? 'S' : 'M');
2536 fname[1] = '0' + i, fname[2] = 0;
2537 field_list
2538 = chainon (field_list,
2539 make_descriptor_field (fname,
2540 gnat_type_for_size (32, 1),
2541 record_type, idx_length));
2542
2543 if (mech == By_Descriptor_NCA)
2544 tem = idx_length;
2545 }
2546
2547 /* Finally here are the bounds. */
2548 for (i = 0; i < ndim; i++)
2549 {
2550 char fname[3];
2551
2552 fname[0] = 'L', fname[1] = '0' + i, fname[2] = 0;
2553 field_list
2554 = chainon (field_list,
2555 make_descriptor_field
2556 (fname, gnat_type_for_size (32, 1), record_type,
2557 TYPE_MIN_VALUE (idx_arr[i])));
2558
2559 fname[0] = 'U';
2560 field_list
2561 = chainon (field_list,
2562 make_descriptor_field
2563 (fname, gnat_type_for_size (32, 1), record_type,
2564 TYPE_MAX_VALUE (idx_arr[i])));
2565 }
2566 break;
2567
2568 default:
2569 post_error ("unsupported descriptor type for &", gnat_entity);
2570 }
2571
2572 finish_record_type (record_type, field_list, 0, 1);
2573 pushdecl (build_decl (TYPE_DECL, create_concat_name (gnat_entity, "DESC"),
2574 record_type));
2575
2576 return record_type;
2577 }
2578
2579 /* Utility routine for above code to make a field. */
2580
2581 static tree
2582 make_descriptor_field (const char *name, tree type,
2583 tree rec_type, tree initial)
2584 {
2585 tree field
2586 = create_field_decl (get_identifier (name), type, rec_type, 0, 0, 0, 0);
2587
2588 DECL_INITIAL (field) = initial;
2589 return field;
2590 }
2591 \f
2592 /* Build a type to be used to represent an aliased object whose nominal
2593 type is an unconstrained array. This consists of a RECORD_TYPE containing
2594 a field of TEMPLATE_TYPE and a field of OBJECT_TYPE, which is an
2595 ARRAY_TYPE. If ARRAY_TYPE is that of the unconstrained array, this
2596 is used to represent an arbitrary unconstrained object. Use NAME
2597 as the name of the record. */
2598
2599 tree
2600 build_unc_object_type (tree template_type, tree object_type, tree name)
2601 {
2602 tree type = make_node (RECORD_TYPE);
2603 tree template_field = create_field_decl (get_identifier ("BOUNDS"),
2604 template_type, type, 0, 0, 0, 1);
2605 tree array_field = create_field_decl (get_identifier ("ARRAY"), object_type,
2606 type, 0, 0, 0, 1);
2607
2608 TYPE_NAME (type) = name;
2609 TYPE_CONTAINS_TEMPLATE_P (type) = 1;
2610 finish_record_type (type,
2611 chainon (chainon (NULL_TREE, template_field),
2612 array_field),
2613 0, 0);
2614
2615 return type;
2616 }
2617 \f
2618 /* Update anything previously pointing to OLD_TYPE to point to NEW_TYPE. In
2619 the normal case this is just two adjustments, but we have more to do
2620 if NEW is an UNCONSTRAINED_ARRAY_TYPE. */
2621
2622 void
2623 update_pointer_to (tree old_type, tree new_type)
2624 {
2625 tree ptr = TYPE_POINTER_TO (old_type);
2626 tree ref = TYPE_REFERENCE_TO (old_type);
2627 tree type;
2628
2629 /* If this is the main variant, process all the other variants first. */
2630 if (TYPE_MAIN_VARIANT (old_type) == old_type)
2631 for (type = TYPE_NEXT_VARIANT (old_type); type != 0;
2632 type = TYPE_NEXT_VARIANT (type))
2633 update_pointer_to (type, new_type);
2634
2635 /* If no pointer or reference, we are done. */
2636 if (ptr == 0 && ref == 0)
2637 return;
2638
2639 /* Merge the old type qualifiers in the new type.
2640
2641 Each old variant has qualifiers for specific reasons, and the new
2642 designated type as well. Each set of qualifiers represents useful
2643 information grabbed at some point, and merging the two simply unifies
2644 these inputs into the final type description.
2645
2646 Consider for instance a volatile type frozen after an access to constant
2647 type designating it. After the designated type freeze, we get here with a
2648 volatile new_type and a dummy old_type with a readonly variant, created
2649 when the access type was processed. We shall make a volatile and readonly
2650 designated type, because that's what it really is.
2651
2652 We might also get here for a non-dummy old_type variant with different
2653 qualifiers than the new_type ones, for instance in some cases of pointers
2654 to private record type elaboration (see the comments around the call to
2655 this routine from gnat_to_gnu_entity/E_Access_Type). We have to merge the
2656 qualifiers in thoses cases too, to avoid accidentally discarding the
2657 initial set, and will often end up with old_type == new_type then. */
2658 new_type = build_qualified_type (new_type,
2659 TYPE_QUALS (old_type)
2660 | TYPE_QUALS (new_type));
2661
2662 /* If the new type and the old one are identical, there is nothing to
2663 update. */
2664 if (old_type == new_type)
2665 return;
2666
2667 /* Otherwise, first handle the simple case. */
2668 if (TREE_CODE (new_type) != UNCONSTRAINED_ARRAY_TYPE)
2669 {
2670 if (ptr != 0)
2671 TREE_TYPE (ptr) = new_type;
2672 TYPE_POINTER_TO (new_type) = ptr;
2673
2674 if (ref != 0)
2675 TREE_TYPE (ref) = new_type;
2676 TYPE_REFERENCE_TO (new_type) = ref;
2677
2678 if (ptr != 0 && TYPE_NAME (ptr) != 0
2679 && TREE_CODE (TYPE_NAME (ptr)) == TYPE_DECL
2680 && TREE_CODE (new_type) != ENUMERAL_TYPE)
2681 rest_of_decl_compilation (TYPE_NAME (ptr), NULL,
2682 global_bindings_p (), 0);
2683 if (ref != 0 && TYPE_NAME (ref) != 0
2684 && TREE_CODE (TYPE_NAME (ref)) == TYPE_DECL
2685 && TREE_CODE (new_type) != ENUMERAL_TYPE)
2686 rest_of_decl_compilation (TYPE_NAME (ref), NULL,
2687 global_bindings_p (), 0);
2688 }
2689
2690 /* Now deal with the unconstrained array case. In this case the "pointer"
2691 is actually a RECORD_TYPE where the types of both fields are
2692 pointers to void. In that case, copy the field list from the
2693 old type to the new one and update the fields' context. */
2694 else if (TREE_CODE (ptr) != RECORD_TYPE || ! TYPE_IS_FAT_POINTER_P (ptr))
2695 gigi_abort (412);
2696
2697 else
2698 {
2699 tree new_obj_rec = TYPE_OBJECT_RECORD_TYPE (new_type);
2700 tree ptr_temp_type;
2701 tree new_ref;
2702 tree var;
2703
2704 TYPE_FIELDS (ptr) = TYPE_FIELDS (TYPE_POINTER_TO (new_type));
2705 DECL_CONTEXT (TYPE_FIELDS (ptr)) = ptr;
2706 DECL_CONTEXT (TREE_CHAIN (TYPE_FIELDS (ptr))) = ptr;
2707
2708 /* Rework the PLACEHOLDER_EXPR inside the reference to the
2709 template bounds.
2710
2711 ??? This is now the only use of gnat_substitute_in_type, which
2712 is now a very "heavy" routine to do this, so it should be replaced
2713 at some point. */
2714 ptr_temp_type = TREE_TYPE (TREE_CHAIN (TYPE_FIELDS (ptr)));
2715 new_ref = build (COMPONENT_REF, ptr_temp_type,
2716 build (PLACEHOLDER_EXPR, ptr),
2717 TREE_CHAIN (TYPE_FIELDS (ptr)));
2718
2719 update_pointer_to
2720 (TREE_TYPE (TREE_TYPE (TYPE_FIELDS (ptr))),
2721 gnat_substitute_in_type (TREE_TYPE (TREE_TYPE (TYPE_FIELDS (ptr))),
2722 TREE_CHAIN (TYPE_FIELDS (ptr)), new_ref));
2723
2724 for (var = TYPE_MAIN_VARIANT (ptr); var; var = TYPE_NEXT_VARIANT (var))
2725 SET_TYPE_UNCONSTRAINED_ARRAY (var, new_type);
2726
2727 TYPE_POINTER_TO (new_type) = TYPE_REFERENCE_TO (new_type)
2728 = TREE_TYPE (new_type) = ptr;
2729
2730 /* Now handle updating the allocation record, what the thin pointer
2731 points to. Update all pointers from the old record into the new
2732 one, update the types of the fields, and recompute the size. */
2733
2734 update_pointer_to (TYPE_OBJECT_RECORD_TYPE (old_type), new_obj_rec);
2735
2736 TREE_TYPE (TYPE_FIELDS (new_obj_rec)) = TREE_TYPE (ptr_temp_type);
2737 TREE_TYPE (TREE_CHAIN (TYPE_FIELDS (new_obj_rec)))
2738 = TREE_TYPE (TREE_TYPE (TYPE_FIELDS (ptr)));
2739 DECL_SIZE (TREE_CHAIN (TYPE_FIELDS (new_obj_rec)))
2740 = TYPE_SIZE (TREE_TYPE (TREE_TYPE (TYPE_FIELDS (ptr))));
2741 DECL_SIZE_UNIT (TREE_CHAIN (TYPE_FIELDS (new_obj_rec)))
2742 = TYPE_SIZE_UNIT (TREE_TYPE (TREE_TYPE (TYPE_FIELDS (ptr))));
2743
2744 TYPE_SIZE (new_obj_rec)
2745 = size_binop (PLUS_EXPR,
2746 DECL_SIZE (TYPE_FIELDS (new_obj_rec)),
2747 DECL_SIZE (TREE_CHAIN (TYPE_FIELDS (new_obj_rec))));
2748 TYPE_SIZE_UNIT (new_obj_rec)
2749 = size_binop (PLUS_EXPR,
2750 DECL_SIZE_UNIT (TYPE_FIELDS (new_obj_rec)),
2751 DECL_SIZE_UNIT (TREE_CHAIN (TYPE_FIELDS (new_obj_rec))));
2752 rest_of_type_compilation (ptr, global_bindings_p ());
2753 }
2754 }
2755 \f
2756 /* Convert a pointer to a constrained array into a pointer to a fat
2757 pointer. This involves making or finding a template. */
2758
2759 static tree
2760 convert_to_fat_pointer (tree type, tree expr)
2761 {
2762 tree template_type = TREE_TYPE (TREE_TYPE (TREE_CHAIN (TYPE_FIELDS (type))));
2763 tree template, template_addr;
2764 tree etype = TREE_TYPE (expr);
2765
2766 /* If EXPR is a constant of zero, we make a fat pointer that has a null
2767 pointer to the template and array. */
2768 if (integer_zerop (expr))
2769 return
2770 gnat_build_constructor
2771 (type,
2772 tree_cons (TYPE_FIELDS (type),
2773 convert (TREE_TYPE (TYPE_FIELDS (type)), expr),
2774 tree_cons (TREE_CHAIN (TYPE_FIELDS (type)),
2775 convert (build_pointer_type (template_type),
2776 expr),
2777 NULL_TREE)));
2778
2779 /* If EXPR is a thin pointer, make the template and data from the record. */
2780
2781 else if (TYPE_THIN_POINTER_P (etype))
2782 {
2783 tree fields = TYPE_FIELDS (TREE_TYPE (etype));
2784
2785 expr = save_expr (expr);
2786 if (TREE_CODE (expr) == ADDR_EXPR)
2787 expr = TREE_OPERAND (expr, 0);
2788 else
2789 expr = build1 (INDIRECT_REF, TREE_TYPE (etype), expr);
2790
2791 template = build_component_ref (expr, NULL_TREE, fields, 0);
2792 expr = build_unary_op (ADDR_EXPR, NULL_TREE,
2793 build_component_ref (expr, NULL_TREE,
2794 TREE_CHAIN (fields), 0));
2795 }
2796 else
2797 /* Otherwise, build the constructor for the template. */
2798 template = build_template (template_type, TREE_TYPE (etype), expr);
2799
2800 template_addr = build_unary_op (ADDR_EXPR, NULL_TREE, template);
2801
2802 /* The result is a CONSTRUCTOR for the fat pointer.
2803
2804 If expr is an argument of a foreign convention subprogram, the type it
2805 points to is directly the component type. In this case, the expression
2806 type may not match the corresponding FIELD_DECL type at this point, so we
2807 call "convert" here to fix that up if necessary. This type consistency is
2808 required, for instance because it ensures that possible later folding of
2809 component_refs against this constructor always yields something of the
2810 same type as the initial reference.
2811
2812 Note that the call to "build_template" above is still fine, because it
2813 will only refer to the provided template_type in this case. */
2814 return
2815 gnat_build_constructor
2816 (type, tree_cons (TYPE_FIELDS (type),
2817 convert (TREE_TYPE (TYPE_FIELDS (type)), expr),
2818 tree_cons (TREE_CHAIN (TYPE_FIELDS (type)),
2819 template_addr, NULL_TREE)));
2820 }
2821 \f
2822 /* Convert to a thin pointer type, TYPE. The only thing we know how to convert
2823 is something that is a fat pointer, so convert to it first if it EXPR
2824 is not already a fat pointer. */
2825
2826 static tree
2827 convert_to_thin_pointer (tree type, tree expr)
2828 {
2829 if (! TYPE_FAT_POINTER_P (TREE_TYPE (expr)))
2830 expr
2831 = convert_to_fat_pointer
2832 (TREE_TYPE (TYPE_UNCONSTRAINED_ARRAY (TREE_TYPE (type))), expr);
2833
2834 /* We get the pointer to the data and use a NOP_EXPR to make it the
2835 proper GCC type. */
2836 expr
2837 = build_component_ref (expr, NULL_TREE, TYPE_FIELDS (TREE_TYPE (expr)), 0);
2838 expr = build1 (NOP_EXPR, type, expr);
2839
2840 return expr;
2841 }
2842 \f
2843 /* Create an expression whose value is that of EXPR,
2844 converted to type TYPE. The TREE_TYPE of the value
2845 is always TYPE. This function implements all reasonable
2846 conversions; callers should filter out those that are
2847 not permitted by the language being compiled. */
2848
2849 tree
2850 convert (tree type, tree expr)
2851 {
2852 enum tree_code code = TREE_CODE (type);
2853 tree etype = TREE_TYPE (expr);
2854 enum tree_code ecode = TREE_CODE (etype);
2855 tree tem;
2856
2857 /* If EXPR is already the right type, we are done. */
2858 if (type == etype)
2859 return expr;
2860 /* If we're converting between two aggregate types that have the same main
2861 variant, just make a NOP_EXPR. */
2862 else if (AGGREGATE_TYPE_P (type)
2863 && TYPE_MAIN_VARIANT (type) == TYPE_MAIN_VARIANT (etype))
2864 return build1 (NOP_EXPR, type, expr);
2865 /* If EXPR is a WITH_RECORD_EXPR, do the conversion inside and then make a
2866 new one. */
2867 else if (TREE_CODE (expr) == WITH_RECORD_EXPR)
2868 return build (WITH_RECORD_EXPR, type,
2869 convert (type, TREE_OPERAND (expr, 0)),
2870 TREE_OPERAND (expr, 1));
2871
2872 /* If the input type has padding, remove it by doing a component reference
2873 to the field. If the output type has padding, make a constructor
2874 to build the record. If both input and output have padding and are
2875 of variable size, do this as an unchecked conversion. */
2876 else if (ecode == RECORD_TYPE && code == RECORD_TYPE
2877 && TYPE_IS_PADDING_P (type) && TYPE_IS_PADDING_P (etype)
2878 && (! TREE_CONSTANT (TYPE_SIZE (type))
2879 || ! TREE_CONSTANT (TYPE_SIZE (etype))))
2880 ;
2881 else if (ecode == RECORD_TYPE && TYPE_IS_PADDING_P (etype))
2882 {
2883 /* If we have just converted to this padded type, just get
2884 the inner expression. */
2885 if (TREE_CODE (expr) == CONSTRUCTOR
2886 && CONSTRUCTOR_ELTS (expr) != 0
2887 && TREE_PURPOSE (CONSTRUCTOR_ELTS (expr)) == TYPE_FIELDS (etype))
2888 return TREE_VALUE (CONSTRUCTOR_ELTS (expr));
2889 else
2890 return convert (type, build_component_ref (expr, NULL_TREE,
2891 TYPE_FIELDS (etype), 0));
2892 }
2893 else if (code == RECORD_TYPE && TYPE_IS_PADDING_P (type))
2894 {
2895 /* If we previously converted from another type and our type is
2896 of variable size, remove the conversion to avoid the need for
2897 variable-size temporaries. */
2898 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR
2899 && ! TREE_CONSTANT (TYPE_SIZE (type)))
2900 expr = TREE_OPERAND (expr, 0);
2901
2902 /* If we are just removing the padding from expr, convert the original
2903 object if we have variable size. That will avoid the need
2904 for some variable-size temporaries. */
2905 if (TREE_CODE (expr) == COMPONENT_REF
2906 && TREE_CODE (TREE_TYPE (TREE_OPERAND (expr, 0))) == RECORD_TYPE
2907 && TYPE_IS_PADDING_P (TREE_TYPE (TREE_OPERAND (expr, 0)))
2908 && ! TREE_CONSTANT (TYPE_SIZE (type)))
2909 return convert (type, TREE_OPERAND (expr, 0));
2910
2911 /* If the result type is a padded type with a self-referentially-sized
2912 field and the expression type is a record, do this as an
2913 unchecked converstion. */
2914 else if (TREE_CODE (etype) == RECORD_TYPE
2915 && CONTAINS_PLACEHOLDER_P (DECL_SIZE (TYPE_FIELDS (type))))
2916 return unchecked_convert (type, expr, 0);
2917
2918 else
2919 return
2920 gnat_build_constructor (type,
2921 tree_cons (TYPE_FIELDS (type),
2922 convert (TREE_TYPE
2923 (TYPE_FIELDS (type)),
2924 expr),
2925 NULL_TREE));
2926 }
2927
2928 /* If the input is a biased type, adjust first. */
2929 if (ecode == INTEGER_TYPE && TYPE_BIASED_REPRESENTATION_P (etype))
2930 return convert (type, fold (build (PLUS_EXPR, TREE_TYPE (etype),
2931 fold (build1 (GNAT_NOP_EXPR,
2932 TREE_TYPE (etype), expr)),
2933 TYPE_MIN_VALUE (etype))));
2934
2935 /* If the input is a left-justified modular type, we need to extract
2936 the actual object before converting it to any other type with the
2937 exception of an unconstrained array. */
2938 if (ecode == RECORD_TYPE && TYPE_LEFT_JUSTIFIED_MODULAR_P (etype)
2939 && code != UNCONSTRAINED_ARRAY_TYPE)
2940 return convert (type, build_component_ref (expr, NULL_TREE,
2941 TYPE_FIELDS (etype), 0));
2942
2943 /* If converting to a type that contains a template, convert to the data
2944 type and then build the template. */
2945 if (code == RECORD_TYPE && TYPE_CONTAINS_TEMPLATE_P (type))
2946 {
2947 tree obj_type = TREE_TYPE (TREE_CHAIN (TYPE_FIELDS (type)));
2948
2949 /* If the source already has a template, get a reference to the
2950 associated array only, as we are going to rebuild a template
2951 for the target type anyway. */
2952 expr = maybe_unconstrained_array (expr);
2953
2954 return
2955 gnat_build_constructor
2956 (type,
2957 tree_cons (TYPE_FIELDS (type),
2958 build_template (TREE_TYPE (TYPE_FIELDS (type)),
2959 obj_type, NULL_TREE),
2960 tree_cons (TREE_CHAIN (TYPE_FIELDS (type)),
2961 convert (obj_type, expr), NULL_TREE)));
2962 }
2963
2964 /* There are some special cases of expressions that we process
2965 specially. */
2966 switch (TREE_CODE (expr))
2967 {
2968 case ERROR_MARK:
2969 return expr;
2970
2971 case TRANSFORM_EXPR:
2972 case NULL_EXPR:
2973 /* Just set its type here. For TRANSFORM_EXPR, we will do the actual
2974 conversion in gnat_expand_expr. NULL_EXPR does not represent
2975 and actual value, so no conversion is needed. */
2976 expr = copy_node (expr);
2977 TREE_TYPE (expr) = type;
2978 return expr;
2979
2980 case STRING_CST:
2981 case CONSTRUCTOR:
2982 /* If we are converting a STRING_CST to another constrained array type,
2983 just make a new one in the proper type. Likewise for a
2984 CONSTRUCTOR. */
2985 if (code == ecode && AGGREGATE_TYPE_P (etype)
2986 && ! (TREE_CODE (TYPE_SIZE (etype)) == INTEGER_CST
2987 && TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST))
2988 {
2989 expr = copy_node (expr);
2990 TREE_TYPE (expr) = type;
2991 return expr;
2992 }
2993 break;
2994
2995 case COMPONENT_REF:
2996 /* If we are converting between two aggregate types of the same
2997 kind, size, mode, and alignment, just make a new COMPONENT_REF.
2998 This avoid unneeded conversions which makes reference computations
2999 more complex. */
3000 if (code == ecode && TYPE_MODE (type) == TYPE_MODE (etype)
3001 && AGGREGATE_TYPE_P (type) && AGGREGATE_TYPE_P (etype)
3002 && TYPE_ALIGN (type) == TYPE_ALIGN (etype)
3003 && operand_equal_p (TYPE_SIZE (type), TYPE_SIZE (etype), 0))
3004 return build (COMPONENT_REF, type, TREE_OPERAND (expr, 0),
3005 TREE_OPERAND (expr, 1));
3006
3007 break;
3008
3009 case UNCONSTRAINED_ARRAY_REF:
3010 /* Convert this to the type of the inner array by getting the address of
3011 the array from the template. */
3012 expr = build_unary_op (INDIRECT_REF, NULL_TREE,
3013 build_component_ref (TREE_OPERAND (expr, 0),
3014 get_identifier ("P_ARRAY"),
3015 NULL_TREE, 0));
3016 etype = TREE_TYPE (expr);
3017 ecode = TREE_CODE (etype);
3018 break;
3019
3020 case VIEW_CONVERT_EXPR:
3021 if (AGGREGATE_TYPE_P (type) && AGGREGATE_TYPE_P (etype)
3022 && ! TYPE_FAT_POINTER_P (type) && ! TYPE_FAT_POINTER_P (etype))
3023 return convert (type, TREE_OPERAND (expr, 0));
3024 break;
3025
3026 case INDIRECT_REF:
3027 /* If both types are record types, just convert the pointer and
3028 make a new INDIRECT_REF.
3029
3030 ??? Disable this for now since it causes problems with the
3031 code in build_binary_op for MODIFY_EXPR which wants to
3032 strip off conversions. But that code really is a mess and
3033 we need to do this a much better way some time. */
3034 if (0
3035 && (TREE_CODE (type) == RECORD_TYPE
3036 || TREE_CODE (type) == UNION_TYPE)
3037 && (TREE_CODE (etype) == RECORD_TYPE
3038 || TREE_CODE (etype) == UNION_TYPE)
3039 && ! TYPE_FAT_POINTER_P (type) && ! TYPE_FAT_POINTER_P (etype))
3040 return build_unary_op (INDIRECT_REF, NULL_TREE,
3041 convert (build_pointer_type (type),
3042 TREE_OPERAND (expr, 0)));
3043 break;
3044
3045 default:
3046 break;
3047 }
3048
3049 /* Check for converting to a pointer to an unconstrained array. */
3050 if (TYPE_FAT_POINTER_P (type) && ! TYPE_FAT_POINTER_P (etype))
3051 return convert_to_fat_pointer (type, expr);
3052
3053 if (TYPE_MAIN_VARIANT (type) == TYPE_MAIN_VARIANT (etype)
3054 || (code == INTEGER_CST && ecode == INTEGER_CST
3055 && (type == TREE_TYPE (etype) || etype == TREE_TYPE (type))))
3056 return fold (build1 (NOP_EXPR, type, expr));
3057
3058 switch (code)
3059 {
3060 case VOID_TYPE:
3061 return build1 (CONVERT_EXPR, type, expr);
3062
3063 case INTEGER_TYPE:
3064 if (TYPE_HAS_ACTUAL_BOUNDS_P (type)
3065 && (ecode == ARRAY_TYPE || ecode == UNCONSTRAINED_ARRAY_TYPE
3066 || (ecode == RECORD_TYPE && TYPE_CONTAINS_TEMPLATE_P (etype))))
3067 return unchecked_convert (type, expr, 0);
3068 else if (TYPE_BIASED_REPRESENTATION_P (type))
3069 return fold (build1 (CONVERT_EXPR, type,
3070 fold (build (MINUS_EXPR, TREE_TYPE (type),
3071 convert (TREE_TYPE (type), expr),
3072 TYPE_MIN_VALUE (type)))));
3073
3074 /* ... fall through ... */
3075
3076 case ENUMERAL_TYPE:
3077 return fold (convert_to_integer (type, expr));
3078
3079 case POINTER_TYPE:
3080 case REFERENCE_TYPE:
3081 /* If converting between two pointers to records denoting
3082 both a template and type, adjust if needed to account
3083 for any differing offsets, since one might be negative. */
3084 if (TYPE_THIN_POINTER_P (etype) && TYPE_THIN_POINTER_P (type))
3085 {
3086 tree bit_diff
3087 = size_diffop (bit_position (TYPE_FIELDS (TREE_TYPE (etype))),
3088 bit_position (TYPE_FIELDS (TREE_TYPE (type))));
3089 tree byte_diff = size_binop (CEIL_DIV_EXPR, bit_diff,
3090 sbitsize_int (BITS_PER_UNIT));
3091
3092 expr = build1 (NOP_EXPR, type, expr);
3093 TREE_CONSTANT (expr) = TREE_CONSTANT (TREE_OPERAND (expr, 0));
3094 if (integer_zerop (byte_diff))
3095 return expr;
3096
3097 return build_binary_op (PLUS_EXPR, type, expr,
3098 fold (convert_to_pointer (type, byte_diff)));
3099 }
3100
3101 /* If converting to a thin pointer, handle specially. */
3102 if (TYPE_THIN_POINTER_P (type)
3103 && TYPE_UNCONSTRAINED_ARRAY (TREE_TYPE (type)) != 0)
3104 return convert_to_thin_pointer (type, expr);
3105
3106 /* If converting fat pointer to normal pointer, get the pointer to the
3107 array and then convert it. */
3108 else if (TYPE_FAT_POINTER_P (etype))
3109 expr = build_component_ref (expr, get_identifier ("P_ARRAY"),
3110 NULL_TREE, 0);
3111
3112 return fold (convert_to_pointer (type, expr));
3113
3114 case REAL_TYPE:
3115 return fold (convert_to_real (type, expr));
3116
3117 case RECORD_TYPE:
3118 if (TYPE_LEFT_JUSTIFIED_MODULAR_P (type) && ! AGGREGATE_TYPE_P (etype))
3119 return
3120 gnat_build_constructor
3121 (type, tree_cons (TYPE_FIELDS (type),
3122 convert (TREE_TYPE (TYPE_FIELDS (type)), expr),
3123 NULL_TREE));
3124
3125 /* ... fall through ... */
3126
3127 case ARRAY_TYPE:
3128 /* In these cases, assume the front-end has validated the conversion.
3129 If the conversion is valid, it will be a bit-wise conversion, so
3130 it can be viewed as an unchecked conversion. */
3131 return unchecked_convert (type, expr, 0);
3132
3133 case UNION_TYPE:
3134 /* Just validate that the type is indeed that of a field
3135 of the type. Then make the simple conversion. */
3136 for (tem = TYPE_FIELDS (type); tem; tem = TREE_CHAIN (tem))
3137 {
3138 if (TREE_TYPE (tem) == etype)
3139 return build1 (CONVERT_EXPR, type, expr);
3140 else if (TREE_CODE (TREE_TYPE (tem)) == RECORD_TYPE
3141 && (TYPE_LEFT_JUSTIFIED_MODULAR_P (TREE_TYPE (tem))
3142 || TYPE_IS_PADDING_P (TREE_TYPE (tem)))
3143 && TREE_TYPE (TYPE_FIELDS (TREE_TYPE (tem))) == etype)
3144 return build1 (CONVERT_EXPR, type,
3145 convert (TREE_TYPE (tem), expr));
3146 }
3147
3148 gigi_abort (413);
3149
3150 case UNCONSTRAINED_ARRAY_TYPE:
3151 /* If EXPR is a constrained array, take its address, convert it to a
3152 fat pointer, and then dereference it. Likewise if EXPR is a
3153 record containing both a template and a constrained array.
3154 Note that a record representing a left justified modular type
3155 always represents a packed constrained array. */
3156 if (ecode == ARRAY_TYPE
3157 || (ecode == INTEGER_TYPE && TYPE_HAS_ACTUAL_BOUNDS_P (etype))
3158 || (ecode == RECORD_TYPE && TYPE_CONTAINS_TEMPLATE_P (etype))
3159 || (ecode == RECORD_TYPE && TYPE_LEFT_JUSTIFIED_MODULAR_P (etype)))
3160 return
3161 build_unary_op
3162 (INDIRECT_REF, NULL_TREE,
3163 convert_to_fat_pointer (TREE_TYPE (type),
3164 build_unary_op (ADDR_EXPR,
3165 NULL_TREE, expr)));
3166
3167 /* Do something very similar for converting one unconstrained
3168 array to another. */
3169 else if (ecode == UNCONSTRAINED_ARRAY_TYPE)
3170 return
3171 build_unary_op (INDIRECT_REF, NULL_TREE,
3172 convert (TREE_TYPE (type),
3173 build_unary_op (ADDR_EXPR,
3174 NULL_TREE, expr)));
3175 else
3176 gigi_abort (409);
3177
3178 case COMPLEX_TYPE:
3179 return fold (convert_to_complex (type, expr));
3180
3181 default:
3182 gigi_abort (410);
3183 }
3184 }
3185 \f
3186 /* Remove all conversions that are done in EXP. This includes converting
3187 from a padded type or to a left-justified modular type. If TRUE_ADDRESS
3188 is nonzero, always return the address of the containing object even if
3189 the address is not bit-aligned. */
3190
3191 tree
3192 remove_conversions (tree exp, int true_address)
3193 {
3194 switch (TREE_CODE (exp))
3195 {
3196 case CONSTRUCTOR:
3197 if (true_address
3198 && TREE_CODE (TREE_TYPE (exp)) == RECORD_TYPE
3199 && TYPE_LEFT_JUSTIFIED_MODULAR_P (TREE_TYPE (exp)))
3200 return remove_conversions (TREE_VALUE (CONSTRUCTOR_ELTS (exp)), 1);
3201 break;
3202
3203 case COMPONENT_REF:
3204 if (TREE_CODE (TREE_TYPE (TREE_OPERAND (exp, 0))) == RECORD_TYPE
3205 && TYPE_IS_PADDING_P (TREE_TYPE (TREE_OPERAND (exp, 0))))
3206 return remove_conversions (TREE_OPERAND (exp, 0), true_address);
3207 break;
3208
3209 case VIEW_CONVERT_EXPR: case NON_LVALUE_EXPR:
3210 case NOP_EXPR: case CONVERT_EXPR: case GNAT_NOP_EXPR:
3211 return remove_conversions (TREE_OPERAND (exp, 0), true_address);
3212
3213 default:
3214 break;
3215 }
3216
3217 return exp;
3218 }
3219 \f
3220 /* If EXP's type is an UNCONSTRAINED_ARRAY_TYPE, return an expression that
3221 refers to the underlying array. If its type has TYPE_CONTAINS_TEMPLATE_P,
3222 likewise return an expression pointing to the underlying array. */
3223
3224 tree
3225 maybe_unconstrained_array (tree exp)
3226 {
3227 enum tree_code code = TREE_CODE (exp);
3228 tree new;
3229
3230 switch (TREE_CODE (TREE_TYPE (exp)))
3231 {
3232 case UNCONSTRAINED_ARRAY_TYPE:
3233 if (code == UNCONSTRAINED_ARRAY_REF)
3234 {
3235 new
3236 = build_unary_op (INDIRECT_REF, NULL_TREE,
3237 build_component_ref (TREE_OPERAND (exp, 0),
3238 get_identifier ("P_ARRAY"),
3239 NULL_TREE, 0));
3240 TREE_READONLY (new) = TREE_STATIC (new) = TREE_READONLY (exp);
3241 return new;
3242 }
3243
3244 else if (code == NULL_EXPR)
3245 return build1 (NULL_EXPR,
3246 TREE_TYPE (TREE_TYPE (TYPE_FIELDS
3247 (TREE_TYPE (TREE_TYPE (exp))))),
3248 TREE_OPERAND (exp, 0));
3249
3250 else if (code == WITH_RECORD_EXPR
3251 && (TREE_OPERAND (exp, 0)
3252 != (new = maybe_unconstrained_array
3253 (TREE_OPERAND (exp, 0)))))
3254 return build (WITH_RECORD_EXPR, TREE_TYPE (new), new,
3255 TREE_OPERAND (exp, 1));
3256
3257 case RECORD_TYPE:
3258 /* If this is a padded type, convert to the unpadded type and see if
3259 it contains a template. */
3260 if (TYPE_IS_PADDING_P (TREE_TYPE (exp)))
3261 {
3262 new = convert (TREE_TYPE (TYPE_FIELDS (TREE_TYPE (exp))), exp);
3263 if (TREE_CODE (TREE_TYPE (new)) == RECORD_TYPE
3264 && TYPE_CONTAINS_TEMPLATE_P (TREE_TYPE (new)))
3265 return
3266 build_component_ref (new, NULL_TREE,
3267 TREE_CHAIN (TYPE_FIELDS (TREE_TYPE (new))),
3268 0);
3269 }
3270 else if (TYPE_CONTAINS_TEMPLATE_P (TREE_TYPE (exp)))
3271 return
3272 build_component_ref (exp, NULL_TREE,
3273 TREE_CHAIN (TYPE_FIELDS (TREE_TYPE (exp))), 0);
3274 break;
3275
3276 default:
3277 break;
3278 }
3279
3280 return exp;
3281 }
3282 \f
3283 /* Return an expression that does an unchecked converstion of EXPR to TYPE.
3284 If NOTRUNC_P is set, truncation operations should be suppressed. */
3285
3286 tree
3287 unchecked_convert (tree type, tree expr, int notrunc_p)
3288 {
3289 tree etype = TREE_TYPE (expr);
3290
3291 /* If the expression is already the right type, we are done. */
3292 if (etype == type)
3293 return expr;
3294
3295 /* If EXPR is a WITH_RECORD_EXPR, do the conversion inside and then make a
3296 new one. */
3297 if (TREE_CODE (expr) == WITH_RECORD_EXPR)
3298 return build (WITH_RECORD_EXPR, type,
3299 unchecked_convert (type, TREE_OPERAND (expr, 0), notrunc_p),
3300 TREE_OPERAND (expr, 1));
3301
3302 /* If both types types are integral just do a normal conversion.
3303 Likewise for a conversion to an unconstrained array. */
3304 if ((((INTEGRAL_TYPE_P (type)
3305 && ! (TREE_CODE (type) == INTEGER_TYPE
3306 && TYPE_VAX_FLOATING_POINT_P (type)))
3307 || (POINTER_TYPE_P (type) && ! TYPE_THIN_POINTER_P (type))
3308 || (TREE_CODE (type) == RECORD_TYPE
3309 && TYPE_LEFT_JUSTIFIED_MODULAR_P (type)))
3310 && ((INTEGRAL_TYPE_P (etype)
3311 && ! (TREE_CODE (etype) == INTEGER_TYPE
3312 && TYPE_VAX_FLOATING_POINT_P (etype)))
3313 || (POINTER_TYPE_P (etype) && ! TYPE_THIN_POINTER_P (etype))
3314 || (TREE_CODE (etype) == RECORD_TYPE
3315 && TYPE_LEFT_JUSTIFIED_MODULAR_P (etype))))
3316 || TREE_CODE (type) == UNCONSTRAINED_ARRAY_TYPE)
3317 {
3318 tree rtype = type;
3319
3320 if (TREE_CODE (etype) == INTEGER_TYPE
3321 && TYPE_BIASED_REPRESENTATION_P (etype))
3322 {
3323 tree ntype = copy_type (etype);
3324
3325 TYPE_BIASED_REPRESENTATION_P (ntype) = 0;
3326 TYPE_MAIN_VARIANT (ntype) = ntype;
3327 expr = build1 (GNAT_NOP_EXPR, ntype, expr);
3328 }
3329
3330 if (TREE_CODE (type) == INTEGER_TYPE
3331 && TYPE_BIASED_REPRESENTATION_P (type))
3332 {
3333 rtype = copy_type (type);
3334 TYPE_BIASED_REPRESENTATION_P (rtype) = 0;
3335 TYPE_MAIN_VARIANT (rtype) = rtype;
3336 }
3337
3338 expr = convert (rtype, expr);
3339 if (type != rtype)
3340 expr = build1 (GNAT_NOP_EXPR, type, expr);
3341 }
3342
3343 /* If we are converting TO an integral type whose precision is not the
3344 same as its size, first unchecked convert to a record that contains
3345 an object of the output type. Then extract the field. */
3346 else if (INTEGRAL_TYPE_P (type) && TYPE_RM_SIZE (type) != 0
3347 && 0 != compare_tree_int (TYPE_RM_SIZE (type),
3348 GET_MODE_BITSIZE (TYPE_MODE (type))))
3349 {
3350 tree rec_type = make_node (RECORD_TYPE);
3351 tree field = create_field_decl (get_identifier ("OBJ"), type,
3352 rec_type, 1, 0, 0, 0);
3353
3354 TYPE_FIELDS (rec_type) = field;
3355 layout_type (rec_type);
3356
3357 expr = unchecked_convert (rec_type, expr, notrunc_p);
3358 expr = build_component_ref (expr, NULL_TREE, field, 0);
3359 }
3360
3361 /* Similarly for integral input type whose precision is not equal to its
3362 size. */
3363 else if (INTEGRAL_TYPE_P (etype) && TYPE_RM_SIZE (etype) != 0
3364 && 0 != compare_tree_int (TYPE_RM_SIZE (etype),
3365 GET_MODE_BITSIZE (TYPE_MODE (etype))))
3366 {
3367 tree rec_type = make_node (RECORD_TYPE);
3368 tree field
3369 = create_field_decl (get_identifier ("OBJ"), etype, rec_type,
3370 1, 0, 0, 0);
3371
3372 TYPE_FIELDS (rec_type) = field;
3373 layout_type (rec_type);
3374
3375 expr = gnat_build_constructor (rec_type, build_tree_list (field, expr));
3376 expr = unchecked_convert (type, expr, notrunc_p);
3377 }
3378
3379 /* We have a special case when we are converting between two
3380 unconstrained array types. In that case, take the address,
3381 convert the fat pointer types, and dereference. */
3382 else if (TREE_CODE (etype) == UNCONSTRAINED_ARRAY_TYPE
3383 && TREE_CODE (type) == UNCONSTRAINED_ARRAY_TYPE)
3384 expr = build_unary_op (INDIRECT_REF, NULL_TREE,
3385 build1 (VIEW_CONVERT_EXPR, TREE_TYPE (type),
3386 build_unary_op (ADDR_EXPR, NULL_TREE,
3387 expr)));
3388 else
3389 {
3390 expr = maybe_unconstrained_array (expr);
3391 etype = TREE_TYPE (expr);
3392 expr = build1 (VIEW_CONVERT_EXPR, type, expr);
3393 }
3394
3395 /* If the result is an integral type whose size is not equal to
3396 the size of the underlying machine type, sign- or zero-extend
3397 the result. We need not do this in the case where the input is
3398 an integral type of the same precision and signedness or if the output
3399 is a biased type or if both the input and output are unsigned. */
3400 if (! notrunc_p
3401 && INTEGRAL_TYPE_P (type) && TYPE_RM_SIZE (type) != 0
3402 && ! (TREE_CODE (type) == INTEGER_TYPE
3403 && TYPE_BIASED_REPRESENTATION_P (type))
3404 && 0 != compare_tree_int (TYPE_RM_SIZE (type),
3405 GET_MODE_BITSIZE (TYPE_MODE (type)))
3406 && ! (INTEGRAL_TYPE_P (etype)
3407 && TREE_UNSIGNED (type) == TREE_UNSIGNED (etype)
3408 && operand_equal_p (TYPE_RM_SIZE (type),
3409 (TYPE_RM_SIZE (etype) != 0
3410 ? TYPE_RM_SIZE (etype) : TYPE_SIZE (etype)),
3411 0))
3412 && ! (TREE_UNSIGNED (type) && TREE_UNSIGNED (etype)))
3413 {
3414 tree base_type = gnat_type_for_mode (TYPE_MODE (type),
3415 TREE_UNSIGNED (type));
3416 tree shift_expr
3417 = convert (base_type,
3418 size_binop (MINUS_EXPR,
3419 bitsize_int
3420 (GET_MODE_BITSIZE (TYPE_MODE (type))),
3421 TYPE_RM_SIZE (type)));
3422 expr
3423 = convert (type,
3424 build_binary_op (RSHIFT_EXPR, base_type,
3425 build_binary_op (LSHIFT_EXPR, base_type,
3426 convert (base_type, expr),
3427 shift_expr),
3428 shift_expr));
3429 }
3430
3431 /* An unchecked conversion should never raise Constraint_Error. The code
3432 below assumes that GCC's conversion routines overflow the same way that
3433 the underlying hardware does. This is probably true. In the rare case
3434 when it is false, we can rely on the fact that such conversions are
3435 erroneous anyway. */
3436 if (TREE_CODE (expr) == INTEGER_CST)
3437 TREE_OVERFLOW (expr) = TREE_CONSTANT_OVERFLOW (expr) = 0;
3438
3439 /* If the sizes of the types differ and this is an VIEW_CONVERT_EXPR,
3440 show no longer constant. */
3441 if (TREE_CODE (expr) == VIEW_CONVERT_EXPR
3442 && ! operand_equal_p (TYPE_SIZE_UNIT (type), TYPE_SIZE_UNIT (etype), 1))
3443 TREE_CONSTANT (expr) = 0;
3444
3445 return expr;
3446 }
3447
3448 #include "gt-ada-utils.h"
3449 #include "gtype-ada.h"