epiphany.opt (mfp-iarith): New option.
[gcc.git] / gcc / gimplify.c
1 /* Tree lowering pass. This pass converts the GENERIC functions-as-trees
2 tree representation into the GIMPLE form.
3 Copyright (C) 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011,
4 2012 Free Software Foundation, Inc.
5 Major work done by Sebastian Pop <s.pop@laposte.net>,
6 Diego Novillo <dnovillo@redhat.com> and Jason Merrill <jason@redhat.com>.
7
8 This file is part of GCC.
9
10 GCC is free software; you can redistribute it and/or modify it under
11 the terms of the GNU General Public License as published by the Free
12 Software Foundation; either version 3, or (at your option) any later
13 version.
14
15 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
16 WARRANTY; without even the implied warranty of MERCHANTABILITY or
17 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 for more details.
19
20 You should have received a copy of the GNU General Public License
21 along with GCC; see the file COPYING3. If not see
22 <http://www.gnu.org/licenses/>. */
23
24 #include "config.h"
25 #include "system.h"
26 #include "coretypes.h"
27 #include "tm.h"
28 #include "tree.h"
29 #include "gimple.h"
30 #include "tree-iterator.h"
31 #include "tree-inline.h"
32 #include "tree-pretty-print.h"
33 #include "langhooks.h"
34 #include "tree-flow.h"
35 #include "cgraph.h"
36 #include "timevar.h"
37 #include "hashtab.h"
38 #include "flags.h"
39 #include "function.h"
40 #include "ggc.h"
41 #include "diagnostic-core.h"
42 #include "target.h"
43 #include "pointer-set.h"
44 #include "splay-tree.h"
45 #include "vec.h"
46 #include "gimple.h"
47
48 #include "langhooks-def.h" /* FIXME: for lhd_set_decl_assembler_name */
49 #include "tree-pass.h" /* FIXME: only for PROP_gimple_any */
50
51 enum gimplify_omp_var_data
52 {
53 GOVD_SEEN = 1,
54 GOVD_EXPLICIT = 2,
55 GOVD_SHARED = 4,
56 GOVD_PRIVATE = 8,
57 GOVD_FIRSTPRIVATE = 16,
58 GOVD_LASTPRIVATE = 32,
59 GOVD_REDUCTION = 64,
60 GOVD_LOCAL = 128,
61 GOVD_DEBUG_PRIVATE = 256,
62 GOVD_PRIVATE_OUTER_REF = 512,
63 GOVD_DATA_SHARE_CLASS = (GOVD_SHARED | GOVD_PRIVATE | GOVD_FIRSTPRIVATE
64 | GOVD_LASTPRIVATE | GOVD_REDUCTION | GOVD_LOCAL)
65 };
66
67
68 enum omp_region_type
69 {
70 ORT_WORKSHARE = 0,
71 ORT_PARALLEL = 2,
72 ORT_COMBINED_PARALLEL = 3,
73 ORT_TASK = 4,
74 ORT_UNTIED_TASK = 5
75 };
76
77 struct gimplify_omp_ctx
78 {
79 struct gimplify_omp_ctx *outer_context;
80 splay_tree variables;
81 struct pointer_set_t *privatized_types;
82 location_t location;
83 enum omp_clause_default_kind default_kind;
84 enum omp_region_type region_type;
85 };
86
87 static struct gimplify_ctx *gimplify_ctxp;
88 static struct gimplify_omp_ctx *gimplify_omp_ctxp;
89
90
91 /* Formal (expression) temporary table handling: multiple occurrences of
92 the same scalar expression are evaluated into the same temporary. */
93
94 typedef struct gimple_temp_hash_elt
95 {
96 tree val; /* Key */
97 tree temp; /* Value */
98 } elt_t;
99
100 /* Forward declaration. */
101 static enum gimplify_status gimplify_compound_expr (tree *, gimple_seq *, bool);
102
103 /* Mark X addressable. Unlike the langhook we expect X to be in gimple
104 form and we don't do any syntax checking. */
105
106 void
107 mark_addressable (tree x)
108 {
109 while (handled_component_p (x))
110 x = TREE_OPERAND (x, 0);
111 if (TREE_CODE (x) == MEM_REF
112 && TREE_CODE (TREE_OPERAND (x, 0)) == ADDR_EXPR)
113 x = TREE_OPERAND (TREE_OPERAND (x, 0), 0);
114 if (TREE_CODE (x) != VAR_DECL
115 && TREE_CODE (x) != PARM_DECL
116 && TREE_CODE (x) != RESULT_DECL)
117 return;
118 TREE_ADDRESSABLE (x) = 1;
119
120 /* Also mark the artificial SSA_NAME that points to the partition of X. */
121 if (TREE_CODE (x) == VAR_DECL
122 && !DECL_EXTERNAL (x)
123 && !TREE_STATIC (x)
124 && cfun->gimple_df != NULL
125 && cfun->gimple_df->decls_to_pointers != NULL)
126 {
127 void *namep
128 = pointer_map_contains (cfun->gimple_df->decls_to_pointers, x);
129 if (namep)
130 TREE_ADDRESSABLE (*(tree *)namep) = 1;
131 }
132 }
133
134 /* Return a hash value for a formal temporary table entry. */
135
136 static hashval_t
137 gimple_tree_hash (const void *p)
138 {
139 tree t = ((const elt_t *) p)->val;
140 return iterative_hash_expr (t, 0);
141 }
142
143 /* Compare two formal temporary table entries. */
144
145 static int
146 gimple_tree_eq (const void *p1, const void *p2)
147 {
148 tree t1 = ((const elt_t *) p1)->val;
149 tree t2 = ((const elt_t *) p2)->val;
150 enum tree_code code = TREE_CODE (t1);
151
152 if (TREE_CODE (t2) != code
153 || TREE_TYPE (t1) != TREE_TYPE (t2))
154 return 0;
155
156 if (!operand_equal_p (t1, t2, 0))
157 return 0;
158
159 #ifdef ENABLE_CHECKING
160 /* Only allow them to compare equal if they also hash equal; otherwise
161 results are nondeterminate, and we fail bootstrap comparison. */
162 gcc_assert (gimple_tree_hash (p1) == gimple_tree_hash (p2));
163 #endif
164
165 return 1;
166 }
167
168 /* Link gimple statement GS to the end of the sequence *SEQ_P. If
169 *SEQ_P is NULL, a new sequence is allocated. This function is
170 similar to gimple_seq_add_stmt, but does not scan the operands.
171 During gimplification, we need to manipulate statement sequences
172 before the def/use vectors have been constructed. */
173
174 void
175 gimple_seq_add_stmt_without_update (gimple_seq *seq_p, gimple gs)
176 {
177 gimple_stmt_iterator si;
178
179 if (gs == NULL)
180 return;
181
182 si = gsi_last (*seq_p);
183 gsi_insert_after_without_update (&si, gs, GSI_NEW_STMT);
184 }
185
186 /* Shorter alias name for the above function for use in gimplify.c
187 only. */
188
189 static inline void
190 gimplify_seq_add_stmt (gimple_seq *seq_p, gimple gs)
191 {
192 gimple_seq_add_stmt_without_update (seq_p, gs);
193 }
194
195 /* Append sequence SRC to the end of sequence *DST_P. If *DST_P is
196 NULL, a new sequence is allocated. This function is
197 similar to gimple_seq_add_seq, but does not scan the operands.
198 During gimplification, we need to manipulate statement sequences
199 before the def/use vectors have been constructed. */
200
201 static void
202 gimplify_seq_add_seq (gimple_seq *dst_p, gimple_seq src)
203 {
204 gimple_stmt_iterator si;
205
206 if (src == NULL)
207 return;
208
209 si = gsi_last (*dst_p);
210 gsi_insert_seq_after_without_update (&si, src, GSI_NEW_STMT);
211 }
212
213 /* Set up a context for the gimplifier. */
214
215 void
216 push_gimplify_context (struct gimplify_ctx *c)
217 {
218 memset (c, '\0', sizeof (*c));
219 c->prev_context = gimplify_ctxp;
220 gimplify_ctxp = c;
221 }
222
223 /* Tear down a context for the gimplifier. If BODY is non-null, then
224 put the temporaries into the outer BIND_EXPR. Otherwise, put them
225 in the local_decls.
226
227 BODY is not a sequence, but the first tuple in a sequence. */
228
229 void
230 pop_gimplify_context (gimple body)
231 {
232 struct gimplify_ctx *c = gimplify_ctxp;
233
234 gcc_assert (c
235 && (!c->bind_expr_stack.exists ()
236 || c->bind_expr_stack.is_empty ()));
237 c->bind_expr_stack.release ();
238 gimplify_ctxp = c->prev_context;
239
240 if (body)
241 declare_vars (c->temps, body, false);
242 else
243 record_vars (c->temps);
244
245 if (c->temp_htab)
246 htab_delete (c->temp_htab);
247 }
248
249 /* Push a GIMPLE_BIND tuple onto the stack of bindings. */
250
251 static void
252 gimple_push_bind_expr (gimple gimple_bind)
253 {
254 gimplify_ctxp->bind_expr_stack.reserve (8);
255 gimplify_ctxp->bind_expr_stack.safe_push (gimple_bind);
256 }
257
258 /* Pop the first element off the stack of bindings. */
259
260 static void
261 gimple_pop_bind_expr (void)
262 {
263 gimplify_ctxp->bind_expr_stack.pop ();
264 }
265
266 /* Return the first element of the stack of bindings. */
267
268 gimple
269 gimple_current_bind_expr (void)
270 {
271 return gimplify_ctxp->bind_expr_stack.last ();
272 }
273
274 /* Return the stack of bindings created during gimplification. */
275
276 vec<gimple>
277 gimple_bind_expr_stack (void)
278 {
279 return gimplify_ctxp->bind_expr_stack;
280 }
281
282 /* Return true iff there is a COND_EXPR between us and the innermost
283 CLEANUP_POINT_EXPR. This info is used by gimple_push_cleanup. */
284
285 static bool
286 gimple_conditional_context (void)
287 {
288 return gimplify_ctxp->conditions > 0;
289 }
290
291 /* Note that we've entered a COND_EXPR. */
292
293 static void
294 gimple_push_condition (void)
295 {
296 #ifdef ENABLE_GIMPLE_CHECKING
297 if (gimplify_ctxp->conditions == 0)
298 gcc_assert (gimple_seq_empty_p (gimplify_ctxp->conditional_cleanups));
299 #endif
300 ++(gimplify_ctxp->conditions);
301 }
302
303 /* Note that we've left a COND_EXPR. If we're back at unconditional scope
304 now, add any conditional cleanups we've seen to the prequeue. */
305
306 static void
307 gimple_pop_condition (gimple_seq *pre_p)
308 {
309 int conds = --(gimplify_ctxp->conditions);
310
311 gcc_assert (conds >= 0);
312 if (conds == 0)
313 {
314 gimplify_seq_add_seq (pre_p, gimplify_ctxp->conditional_cleanups);
315 gimplify_ctxp->conditional_cleanups = NULL;
316 }
317 }
318
319 /* A stable comparison routine for use with splay trees and DECLs. */
320
321 static int
322 splay_tree_compare_decl_uid (splay_tree_key xa, splay_tree_key xb)
323 {
324 tree a = (tree) xa;
325 tree b = (tree) xb;
326
327 return DECL_UID (a) - DECL_UID (b);
328 }
329
330 /* Create a new omp construct that deals with variable remapping. */
331
332 static struct gimplify_omp_ctx *
333 new_omp_context (enum omp_region_type region_type)
334 {
335 struct gimplify_omp_ctx *c;
336
337 c = XCNEW (struct gimplify_omp_ctx);
338 c->outer_context = gimplify_omp_ctxp;
339 c->variables = splay_tree_new (splay_tree_compare_decl_uid, 0, 0);
340 c->privatized_types = pointer_set_create ();
341 c->location = input_location;
342 c->region_type = region_type;
343 if ((region_type & ORT_TASK) == 0)
344 c->default_kind = OMP_CLAUSE_DEFAULT_SHARED;
345 else
346 c->default_kind = OMP_CLAUSE_DEFAULT_UNSPECIFIED;
347
348 return c;
349 }
350
351 /* Destroy an omp construct that deals with variable remapping. */
352
353 static void
354 delete_omp_context (struct gimplify_omp_ctx *c)
355 {
356 splay_tree_delete (c->variables);
357 pointer_set_destroy (c->privatized_types);
358 XDELETE (c);
359 }
360
361 static void omp_add_variable (struct gimplify_omp_ctx *, tree, unsigned int);
362 static bool omp_notice_variable (struct gimplify_omp_ctx *, tree, bool);
363
364 /* Both gimplify the statement T and append it to *SEQ_P. This function
365 behaves exactly as gimplify_stmt, but you don't have to pass T as a
366 reference. */
367
368 void
369 gimplify_and_add (tree t, gimple_seq *seq_p)
370 {
371 gimplify_stmt (&t, seq_p);
372 }
373
374 /* Gimplify statement T into sequence *SEQ_P, and return the first
375 tuple in the sequence of generated tuples for this statement.
376 Return NULL if gimplifying T produced no tuples. */
377
378 static gimple
379 gimplify_and_return_first (tree t, gimple_seq *seq_p)
380 {
381 gimple_stmt_iterator last = gsi_last (*seq_p);
382
383 gimplify_and_add (t, seq_p);
384
385 if (!gsi_end_p (last))
386 {
387 gsi_next (&last);
388 return gsi_stmt (last);
389 }
390 else
391 return gimple_seq_first_stmt (*seq_p);
392 }
393
394 /* Strip off a legitimate source ending from the input string NAME of
395 length LEN. Rather than having to know the names used by all of
396 our front ends, we strip off an ending of a period followed by
397 up to five characters. (Java uses ".class".) */
398
399 static inline void
400 remove_suffix (char *name, int len)
401 {
402 int i;
403
404 for (i = 2; i < 8 && len > i; i++)
405 {
406 if (name[len - i] == '.')
407 {
408 name[len - i] = '\0';
409 break;
410 }
411 }
412 }
413
414 /* Create a new temporary name with PREFIX. Return an identifier. */
415
416 static GTY(()) unsigned int tmp_var_id_num;
417
418 tree
419 create_tmp_var_name (const char *prefix)
420 {
421 char *tmp_name;
422
423 if (prefix)
424 {
425 char *preftmp = ASTRDUP (prefix);
426
427 remove_suffix (preftmp, strlen (preftmp));
428 clean_symbol_name (preftmp);
429
430 prefix = preftmp;
431 }
432
433 ASM_FORMAT_PRIVATE_NAME (tmp_name, prefix ? prefix : "T", tmp_var_id_num++);
434 return get_identifier (tmp_name);
435 }
436
437 /* Create a new temporary variable declaration of type TYPE.
438 Do NOT push it into the current binding. */
439
440 tree
441 create_tmp_var_raw (tree type, const char *prefix)
442 {
443 tree tmp_var;
444
445 tmp_var = build_decl (input_location,
446 VAR_DECL, prefix ? create_tmp_var_name (prefix) : NULL,
447 type);
448
449 /* The variable was declared by the compiler. */
450 DECL_ARTIFICIAL (tmp_var) = 1;
451 /* And we don't want debug info for it. */
452 DECL_IGNORED_P (tmp_var) = 1;
453
454 /* Make the variable writable. */
455 TREE_READONLY (tmp_var) = 0;
456
457 DECL_EXTERNAL (tmp_var) = 0;
458 TREE_STATIC (tmp_var) = 0;
459 TREE_USED (tmp_var) = 1;
460
461 return tmp_var;
462 }
463
464 /* Create a new temporary variable declaration of type TYPE. DO push the
465 variable into the current binding. Further, assume that this is called
466 only from gimplification or optimization, at which point the creation of
467 certain types are bugs. */
468
469 tree
470 create_tmp_var (tree type, const char *prefix)
471 {
472 tree tmp_var;
473
474 /* We don't allow types that are addressable (meaning we can't make copies),
475 or incomplete. We also used to reject every variable size objects here,
476 but now support those for which a constant upper bound can be obtained.
477 The processing for variable sizes is performed in gimple_add_tmp_var,
478 point at which it really matters and possibly reached via paths not going
479 through this function, e.g. after direct calls to create_tmp_var_raw. */
480 gcc_assert (!TREE_ADDRESSABLE (type) && COMPLETE_TYPE_P (type));
481
482 tmp_var = create_tmp_var_raw (type, prefix);
483 gimple_add_tmp_var (tmp_var);
484 return tmp_var;
485 }
486
487 /* Create a new temporary variable declaration of type TYPE by calling
488 create_tmp_var and if TYPE is a vector or a complex number, mark the new
489 temporary as gimple register. */
490
491 tree
492 create_tmp_reg (tree type, const char *prefix)
493 {
494 tree tmp;
495
496 tmp = create_tmp_var (type, prefix);
497 if (TREE_CODE (type) == COMPLEX_TYPE
498 || TREE_CODE (type) == VECTOR_TYPE)
499 DECL_GIMPLE_REG_P (tmp) = 1;
500
501 return tmp;
502 }
503
504 /* Returns true iff T is a valid RHS for an assignment to a renamed
505 user -- or front-end generated artificial -- variable. */
506
507 static bool
508 is_gimple_reg_rhs (tree t)
509 {
510 return get_gimple_rhs_class (TREE_CODE (t)) != GIMPLE_INVALID_RHS;
511 }
512
513 /* Returns true iff T is a valid RHS for an assignment to an un-renamed
514 LHS, or for a call argument. */
515
516 static bool
517 is_gimple_mem_rhs (tree t)
518 {
519 /* If we're dealing with a renamable type, either source or dest must be
520 a renamed variable. */
521 if (is_gimple_reg_type (TREE_TYPE (t)))
522 return is_gimple_val (t);
523 else
524 return is_gimple_val (t) || is_gimple_lvalue (t);
525 }
526
527 /* Return true if T is a CALL_EXPR or an expression that can be
528 assigned to a temporary. Note that this predicate should only be
529 used during gimplification. See the rationale for this in
530 gimplify_modify_expr. */
531
532 static bool
533 is_gimple_reg_rhs_or_call (tree t)
534 {
535 return (get_gimple_rhs_class (TREE_CODE (t)) != GIMPLE_INVALID_RHS
536 || TREE_CODE (t) == CALL_EXPR);
537 }
538
539 /* Return true if T is a valid memory RHS or a CALL_EXPR. Note that
540 this predicate should only be used during gimplification. See the
541 rationale for this in gimplify_modify_expr. */
542
543 static bool
544 is_gimple_mem_rhs_or_call (tree t)
545 {
546 /* If we're dealing with a renamable type, either source or dest must be
547 a renamed variable. */
548 if (is_gimple_reg_type (TREE_TYPE (t)))
549 return is_gimple_val (t);
550 else
551 return (is_gimple_val (t) || is_gimple_lvalue (t)
552 || TREE_CODE (t) == CALL_EXPR);
553 }
554
555 /* Create a temporary with a name derived from VAL. Subroutine of
556 lookup_tmp_var; nobody else should call this function. */
557
558 static inline tree
559 create_tmp_from_val (tree val, bool is_formal)
560 {
561 /* Drop all qualifiers and address-space information from the value type. */
562 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (val));
563 tree var = create_tmp_var (type, get_name (val));
564 if (is_formal
565 && (TREE_CODE (TREE_TYPE (var)) == COMPLEX_TYPE
566 || TREE_CODE (TREE_TYPE (var)) == VECTOR_TYPE))
567 DECL_GIMPLE_REG_P (var) = 1;
568 return var;
569 }
570
571 /* Create a temporary to hold the value of VAL. If IS_FORMAL, try to reuse
572 an existing expression temporary. */
573
574 static tree
575 lookup_tmp_var (tree val, bool is_formal)
576 {
577 tree ret;
578
579 /* If not optimizing, never really reuse a temporary. local-alloc
580 won't allocate any variable that is used in more than one basic
581 block, which means it will go into memory, causing much extra
582 work in reload and final and poorer code generation, outweighing
583 the extra memory allocation here. */
584 if (!optimize || !is_formal || TREE_SIDE_EFFECTS (val))
585 ret = create_tmp_from_val (val, is_formal);
586 else
587 {
588 elt_t elt, *elt_p;
589 void **slot;
590
591 elt.val = val;
592 if (gimplify_ctxp->temp_htab == NULL)
593 gimplify_ctxp->temp_htab
594 = htab_create (1000, gimple_tree_hash, gimple_tree_eq, free);
595 slot = htab_find_slot (gimplify_ctxp->temp_htab, (void *)&elt, INSERT);
596 if (*slot == NULL)
597 {
598 elt_p = XNEW (elt_t);
599 elt_p->val = val;
600 elt_p->temp = ret = create_tmp_from_val (val, is_formal);
601 *slot = (void *) elt_p;
602 }
603 else
604 {
605 elt_p = (elt_t *) *slot;
606 ret = elt_p->temp;
607 }
608 }
609
610 return ret;
611 }
612
613 /* Helper for get_formal_tmp_var and get_initialized_tmp_var. */
614
615 static tree
616 internal_get_tmp_var (tree val, gimple_seq *pre_p, gimple_seq *post_p,
617 bool is_formal)
618 {
619 tree t, mod;
620
621 /* Notice that we explicitly allow VAL to be a CALL_EXPR so that we
622 can create an INIT_EXPR and convert it into a GIMPLE_CALL below. */
623 gimplify_expr (&val, pre_p, post_p, is_gimple_reg_rhs_or_call,
624 fb_rvalue);
625
626 if (gimplify_ctxp->into_ssa
627 && is_gimple_reg_type (TREE_TYPE (val)))
628 t = make_ssa_name (TYPE_MAIN_VARIANT (TREE_TYPE (val)), NULL);
629 else
630 t = lookup_tmp_var (val, is_formal);
631
632 mod = build2 (INIT_EXPR, TREE_TYPE (t), t, unshare_expr (val));
633
634 SET_EXPR_LOCATION (mod, EXPR_LOC_OR_HERE (val));
635
636 /* gimplify_modify_expr might want to reduce this further. */
637 gimplify_and_add (mod, pre_p);
638 ggc_free (mod);
639
640 return t;
641 }
642
643 /* Return a formal temporary variable initialized with VAL. PRE_P is as
644 in gimplify_expr. Only use this function if:
645
646 1) The value of the unfactored expression represented by VAL will not
647 change between the initialization and use of the temporary, and
648 2) The temporary will not be otherwise modified.
649
650 For instance, #1 means that this is inappropriate for SAVE_EXPR temps,
651 and #2 means it is inappropriate for && temps.
652
653 For other cases, use get_initialized_tmp_var instead. */
654
655 tree
656 get_formal_tmp_var (tree val, gimple_seq *pre_p)
657 {
658 return internal_get_tmp_var (val, pre_p, NULL, true);
659 }
660
661 /* Return a temporary variable initialized with VAL. PRE_P and POST_P
662 are as in gimplify_expr. */
663
664 tree
665 get_initialized_tmp_var (tree val, gimple_seq *pre_p, gimple_seq *post_p)
666 {
667 return internal_get_tmp_var (val, pre_p, post_p, false);
668 }
669
670 /* Declare all the variables in VARS in SCOPE. If DEBUG_INFO is true,
671 generate debug info for them; otherwise don't. */
672
673 void
674 declare_vars (tree vars, gimple scope, bool debug_info)
675 {
676 tree last = vars;
677 if (last)
678 {
679 tree temps, block;
680
681 gcc_assert (gimple_code (scope) == GIMPLE_BIND);
682
683 temps = nreverse (last);
684
685 block = gimple_bind_block (scope);
686 gcc_assert (!block || TREE_CODE (block) == BLOCK);
687 if (!block || !debug_info)
688 {
689 DECL_CHAIN (last) = gimple_bind_vars (scope);
690 gimple_bind_set_vars (scope, temps);
691 }
692 else
693 {
694 /* We need to attach the nodes both to the BIND_EXPR and to its
695 associated BLOCK for debugging purposes. The key point here
696 is that the BLOCK_VARS of the BIND_EXPR_BLOCK of a BIND_EXPR
697 is a subchain of the BIND_EXPR_VARS of the BIND_EXPR. */
698 if (BLOCK_VARS (block))
699 BLOCK_VARS (block) = chainon (BLOCK_VARS (block), temps);
700 else
701 {
702 gimple_bind_set_vars (scope,
703 chainon (gimple_bind_vars (scope), temps));
704 BLOCK_VARS (block) = temps;
705 }
706 }
707 }
708 }
709
710 /* For VAR a VAR_DECL of variable size, try to find a constant upper bound
711 for the size and adjust DECL_SIZE/DECL_SIZE_UNIT accordingly. Abort if
712 no such upper bound can be obtained. */
713
714 static void
715 force_constant_size (tree var)
716 {
717 /* The only attempt we make is by querying the maximum size of objects
718 of the variable's type. */
719
720 HOST_WIDE_INT max_size;
721
722 gcc_assert (TREE_CODE (var) == VAR_DECL);
723
724 max_size = max_int_size_in_bytes (TREE_TYPE (var));
725
726 gcc_assert (max_size >= 0);
727
728 DECL_SIZE_UNIT (var)
729 = build_int_cst (TREE_TYPE (DECL_SIZE_UNIT (var)), max_size);
730 DECL_SIZE (var)
731 = build_int_cst (TREE_TYPE (DECL_SIZE (var)), max_size * BITS_PER_UNIT);
732 }
733
734 /* Push the temporary variable TMP into the current binding. */
735
736 void
737 gimple_add_tmp_var (tree tmp)
738 {
739 gcc_assert (!DECL_CHAIN (tmp) && !DECL_SEEN_IN_BIND_EXPR_P (tmp));
740
741 /* Later processing assumes that the object size is constant, which might
742 not be true at this point. Force the use of a constant upper bound in
743 this case. */
744 if (!host_integerp (DECL_SIZE_UNIT (tmp), 1))
745 force_constant_size (tmp);
746
747 DECL_CONTEXT (tmp) = current_function_decl;
748 DECL_SEEN_IN_BIND_EXPR_P (tmp) = 1;
749
750 if (gimplify_ctxp)
751 {
752 DECL_CHAIN (tmp) = gimplify_ctxp->temps;
753 gimplify_ctxp->temps = tmp;
754
755 /* Mark temporaries local within the nearest enclosing parallel. */
756 if (gimplify_omp_ctxp)
757 {
758 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
759 while (ctx && ctx->region_type == ORT_WORKSHARE)
760 ctx = ctx->outer_context;
761 if (ctx)
762 omp_add_variable (ctx, tmp, GOVD_LOCAL | GOVD_SEEN);
763 }
764 }
765 else if (cfun)
766 record_vars (tmp);
767 else
768 {
769 gimple_seq body_seq;
770
771 /* This case is for nested functions. We need to expose the locals
772 they create. */
773 body_seq = gimple_body (current_function_decl);
774 declare_vars (tmp, gimple_seq_first_stmt (body_seq), false);
775 }
776 }
777
778 /* Determine whether to assign a location to the statement GS. */
779
780 static bool
781 should_carry_location_p (gimple gs)
782 {
783 /* Don't emit a line note for a label. We particularly don't want to
784 emit one for the break label, since it doesn't actually correspond
785 to the beginning of the loop/switch. */
786 if (gimple_code (gs) == GIMPLE_LABEL)
787 return false;
788
789 return true;
790 }
791
792 /* Return true if a location should not be emitted for this statement
793 by annotate_one_with_location. */
794
795 static inline bool
796 gimple_do_not_emit_location_p (gimple g)
797 {
798 return gimple_plf (g, GF_PLF_1);
799 }
800
801 /* Mark statement G so a location will not be emitted by
802 annotate_one_with_location. */
803
804 static inline void
805 gimple_set_do_not_emit_location (gimple g)
806 {
807 /* The PLF flags are initialized to 0 when a new tuple is created,
808 so no need to initialize it anywhere. */
809 gimple_set_plf (g, GF_PLF_1, true);
810 }
811
812 /* Set the location for gimple statement GS to LOCATION. */
813
814 static void
815 annotate_one_with_location (gimple gs, location_t location)
816 {
817 if (!gimple_has_location (gs)
818 && !gimple_do_not_emit_location_p (gs)
819 && should_carry_location_p (gs))
820 gimple_set_location (gs, location);
821 }
822
823 /* Set LOCATION for all the statements after iterator GSI in sequence
824 SEQ. If GSI is pointing to the end of the sequence, start with the
825 first statement in SEQ. */
826
827 static void
828 annotate_all_with_location_after (gimple_seq seq, gimple_stmt_iterator gsi,
829 location_t location)
830 {
831 if (gsi_end_p (gsi))
832 gsi = gsi_start (seq);
833 else
834 gsi_next (&gsi);
835
836 for (; !gsi_end_p (gsi); gsi_next (&gsi))
837 annotate_one_with_location (gsi_stmt (gsi), location);
838 }
839
840 /* Set the location for all the statements in a sequence STMT_P to LOCATION. */
841
842 void
843 annotate_all_with_location (gimple_seq stmt_p, location_t location)
844 {
845 gimple_stmt_iterator i;
846
847 if (gimple_seq_empty_p (stmt_p))
848 return;
849
850 for (i = gsi_start (stmt_p); !gsi_end_p (i); gsi_next (&i))
851 {
852 gimple gs = gsi_stmt (i);
853 annotate_one_with_location (gs, location);
854 }
855 }
856 \f
857 /* This page contains routines to unshare tree nodes, i.e. to duplicate tree
858 nodes that are referenced more than once in GENERIC functions. This is
859 necessary because gimplification (translation into GIMPLE) is performed
860 by modifying tree nodes in-place, so gimplication of a shared node in a
861 first context could generate an invalid GIMPLE form in a second context.
862
863 This is achieved with a simple mark/copy/unmark algorithm that walks the
864 GENERIC representation top-down, marks nodes with TREE_VISITED the first
865 time it encounters them, duplicates them if they already have TREE_VISITED
866 set, and finally removes the TREE_VISITED marks it has set.
867
868 The algorithm works only at the function level, i.e. it generates a GENERIC
869 representation of a function with no nodes shared within the function when
870 passed a GENERIC function (except for nodes that are allowed to be shared).
871
872 At the global level, it is also necessary to unshare tree nodes that are
873 referenced in more than one function, for the same aforementioned reason.
874 This requires some cooperation from the front-end. There are 2 strategies:
875
876 1. Manual unsharing. The front-end needs to call unshare_expr on every
877 expression that might end up being shared across functions.
878
879 2. Deep unsharing. This is an extension of regular unsharing. Instead
880 of calling unshare_expr on expressions that might be shared across
881 functions, the front-end pre-marks them with TREE_VISITED. This will
882 ensure that they are unshared on the first reference within functions
883 when the regular unsharing algorithm runs. The counterpart is that
884 this algorithm must look deeper than for manual unsharing, which is
885 specified by LANG_HOOKS_DEEP_UNSHARING.
886
887 If there are only few specific cases of node sharing across functions, it is
888 probably easier for a front-end to unshare the expressions manually. On the
889 contrary, if the expressions generated at the global level are as widespread
890 as expressions generated within functions, deep unsharing is very likely the
891 way to go. */
892
893 /* Similar to copy_tree_r but do not copy SAVE_EXPR or TARGET_EXPR nodes.
894 These nodes model computations that must be done once. If we were to
895 unshare something like SAVE_EXPR(i++), the gimplification process would
896 create wrong code. However, if DATA is non-null, it must hold a pointer
897 set that is used to unshare the subtrees of these nodes. */
898
899 static tree
900 mostly_copy_tree_r (tree *tp, int *walk_subtrees, void *data)
901 {
902 tree t = *tp;
903 enum tree_code code = TREE_CODE (t);
904
905 /* Do not copy SAVE_EXPR, TARGET_EXPR or BIND_EXPR nodes themselves, but
906 copy their subtrees if we can make sure to do it only once. */
907 if (code == SAVE_EXPR || code == TARGET_EXPR || code == BIND_EXPR)
908 {
909 if (data && !pointer_set_insert ((struct pointer_set_t *)data, t))
910 ;
911 else
912 *walk_subtrees = 0;
913 }
914
915 /* Stop at types, decls, constants like copy_tree_r. */
916 else if (TREE_CODE_CLASS (code) == tcc_type
917 || TREE_CODE_CLASS (code) == tcc_declaration
918 || TREE_CODE_CLASS (code) == tcc_constant
919 /* We can't do anything sensible with a BLOCK used as an
920 expression, but we also can't just die when we see it
921 because of non-expression uses. So we avert our eyes
922 and cross our fingers. Silly Java. */
923 || code == BLOCK)
924 *walk_subtrees = 0;
925
926 /* Cope with the statement expression extension. */
927 else if (code == STATEMENT_LIST)
928 ;
929
930 /* Leave the bulk of the work to copy_tree_r itself. */
931 else
932 copy_tree_r (tp, walk_subtrees, NULL);
933
934 return NULL_TREE;
935 }
936
937 /* Callback for walk_tree to unshare most of the shared trees rooted at *TP.
938 If *TP has been visited already, then *TP is deeply copied by calling
939 mostly_copy_tree_r. DATA is passed to mostly_copy_tree_r unmodified. */
940
941 static tree
942 copy_if_shared_r (tree *tp, int *walk_subtrees, void *data)
943 {
944 tree t = *tp;
945 enum tree_code code = TREE_CODE (t);
946
947 /* Skip types, decls, and constants. But we do want to look at their
948 types and the bounds of types. Mark them as visited so we properly
949 unmark their subtrees on the unmark pass. If we've already seen them,
950 don't look down further. */
951 if (TREE_CODE_CLASS (code) == tcc_type
952 || TREE_CODE_CLASS (code) == tcc_declaration
953 || TREE_CODE_CLASS (code) == tcc_constant)
954 {
955 if (TREE_VISITED (t))
956 *walk_subtrees = 0;
957 else
958 TREE_VISITED (t) = 1;
959 }
960
961 /* If this node has been visited already, unshare it and don't look
962 any deeper. */
963 else if (TREE_VISITED (t))
964 {
965 walk_tree (tp, mostly_copy_tree_r, data, NULL);
966 *walk_subtrees = 0;
967 }
968
969 /* Otherwise, mark the node as visited and keep looking. */
970 else
971 TREE_VISITED (t) = 1;
972
973 return NULL_TREE;
974 }
975
976 /* Unshare most of the shared trees rooted at *TP. DATA is passed to the
977 copy_if_shared_r callback unmodified. */
978
979 static inline void
980 copy_if_shared (tree *tp, void *data)
981 {
982 walk_tree (tp, copy_if_shared_r, data, NULL);
983 }
984
985 /* Unshare all the trees in the body of FNDECL, as well as in the bodies of
986 any nested functions. */
987
988 static void
989 unshare_body (tree fndecl)
990 {
991 struct cgraph_node *cgn = cgraph_get_node (fndecl);
992 /* If the language requires deep unsharing, we need a pointer set to make
993 sure we don't repeatedly unshare subtrees of unshareable nodes. */
994 struct pointer_set_t *visited
995 = lang_hooks.deep_unsharing ? pointer_set_create () : NULL;
996
997 copy_if_shared (&DECL_SAVED_TREE (fndecl), visited);
998 copy_if_shared (&DECL_SIZE (DECL_RESULT (fndecl)), visited);
999 copy_if_shared (&DECL_SIZE_UNIT (DECL_RESULT (fndecl)), visited);
1000
1001 if (visited)
1002 pointer_set_destroy (visited);
1003
1004 if (cgn)
1005 for (cgn = cgn->nested; cgn; cgn = cgn->next_nested)
1006 unshare_body (cgn->symbol.decl);
1007 }
1008
1009 /* Callback for walk_tree to unmark the visited trees rooted at *TP.
1010 Subtrees are walked until the first unvisited node is encountered. */
1011
1012 static tree
1013 unmark_visited_r (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED)
1014 {
1015 tree t = *tp;
1016
1017 /* If this node has been visited, unmark it and keep looking. */
1018 if (TREE_VISITED (t))
1019 TREE_VISITED (t) = 0;
1020
1021 /* Otherwise, don't look any deeper. */
1022 else
1023 *walk_subtrees = 0;
1024
1025 return NULL_TREE;
1026 }
1027
1028 /* Unmark the visited trees rooted at *TP. */
1029
1030 static inline void
1031 unmark_visited (tree *tp)
1032 {
1033 walk_tree (tp, unmark_visited_r, NULL, NULL);
1034 }
1035
1036 /* Likewise, but mark all trees as not visited. */
1037
1038 static void
1039 unvisit_body (tree fndecl)
1040 {
1041 struct cgraph_node *cgn = cgraph_get_node (fndecl);
1042
1043 unmark_visited (&DECL_SAVED_TREE (fndecl));
1044 unmark_visited (&DECL_SIZE (DECL_RESULT (fndecl)));
1045 unmark_visited (&DECL_SIZE_UNIT (DECL_RESULT (fndecl)));
1046
1047 if (cgn)
1048 for (cgn = cgn->nested; cgn; cgn = cgn->next_nested)
1049 unvisit_body (cgn->symbol.decl);
1050 }
1051
1052 /* Unconditionally make an unshared copy of EXPR. This is used when using
1053 stored expressions which span multiple functions, such as BINFO_VTABLE,
1054 as the normal unsharing process can't tell that they're shared. */
1055
1056 tree
1057 unshare_expr (tree expr)
1058 {
1059 walk_tree (&expr, mostly_copy_tree_r, NULL, NULL);
1060 return expr;
1061 }
1062 \f
1063 /* WRAPPER is a code such as BIND_EXPR or CLEANUP_POINT_EXPR which can both
1064 contain statements and have a value. Assign its value to a temporary
1065 and give it void_type_node. Return the temporary, or NULL_TREE if
1066 WRAPPER was already void. */
1067
1068 tree
1069 voidify_wrapper_expr (tree wrapper, tree temp)
1070 {
1071 tree type = TREE_TYPE (wrapper);
1072 if (type && !VOID_TYPE_P (type))
1073 {
1074 tree *p;
1075
1076 /* Set p to point to the body of the wrapper. Loop until we find
1077 something that isn't a wrapper. */
1078 for (p = &wrapper; p && *p; )
1079 {
1080 switch (TREE_CODE (*p))
1081 {
1082 case BIND_EXPR:
1083 TREE_SIDE_EFFECTS (*p) = 1;
1084 TREE_TYPE (*p) = void_type_node;
1085 /* For a BIND_EXPR, the body is operand 1. */
1086 p = &BIND_EXPR_BODY (*p);
1087 break;
1088
1089 case CLEANUP_POINT_EXPR:
1090 case TRY_FINALLY_EXPR:
1091 case TRY_CATCH_EXPR:
1092 TREE_SIDE_EFFECTS (*p) = 1;
1093 TREE_TYPE (*p) = void_type_node;
1094 p = &TREE_OPERAND (*p, 0);
1095 break;
1096
1097 case STATEMENT_LIST:
1098 {
1099 tree_stmt_iterator i = tsi_last (*p);
1100 TREE_SIDE_EFFECTS (*p) = 1;
1101 TREE_TYPE (*p) = void_type_node;
1102 p = tsi_end_p (i) ? NULL : tsi_stmt_ptr (i);
1103 }
1104 break;
1105
1106 case COMPOUND_EXPR:
1107 /* Advance to the last statement. Set all container types to
1108 void. */
1109 for (; TREE_CODE (*p) == COMPOUND_EXPR; p = &TREE_OPERAND (*p, 1))
1110 {
1111 TREE_SIDE_EFFECTS (*p) = 1;
1112 TREE_TYPE (*p) = void_type_node;
1113 }
1114 break;
1115
1116 case TRANSACTION_EXPR:
1117 TREE_SIDE_EFFECTS (*p) = 1;
1118 TREE_TYPE (*p) = void_type_node;
1119 p = &TRANSACTION_EXPR_BODY (*p);
1120 break;
1121
1122 default:
1123 /* Assume that any tree upon which voidify_wrapper_expr is
1124 directly called is a wrapper, and that its body is op0. */
1125 if (p == &wrapper)
1126 {
1127 TREE_SIDE_EFFECTS (*p) = 1;
1128 TREE_TYPE (*p) = void_type_node;
1129 p = &TREE_OPERAND (*p, 0);
1130 break;
1131 }
1132 goto out;
1133 }
1134 }
1135
1136 out:
1137 if (p == NULL || IS_EMPTY_STMT (*p))
1138 temp = NULL_TREE;
1139 else if (temp)
1140 {
1141 /* The wrapper is on the RHS of an assignment that we're pushing
1142 down. */
1143 gcc_assert (TREE_CODE (temp) == INIT_EXPR
1144 || TREE_CODE (temp) == MODIFY_EXPR);
1145 TREE_OPERAND (temp, 1) = *p;
1146 *p = temp;
1147 }
1148 else
1149 {
1150 temp = create_tmp_var (type, "retval");
1151 *p = build2 (INIT_EXPR, type, temp, *p);
1152 }
1153
1154 return temp;
1155 }
1156
1157 return NULL_TREE;
1158 }
1159
1160 /* Prepare calls to builtins to SAVE and RESTORE the stack as well as
1161 a temporary through which they communicate. */
1162
1163 static void
1164 build_stack_save_restore (gimple *save, gimple *restore)
1165 {
1166 tree tmp_var;
1167
1168 *save = gimple_build_call (builtin_decl_implicit (BUILT_IN_STACK_SAVE), 0);
1169 tmp_var = create_tmp_var (ptr_type_node, "saved_stack");
1170 gimple_call_set_lhs (*save, tmp_var);
1171
1172 *restore
1173 = gimple_build_call (builtin_decl_implicit (BUILT_IN_STACK_RESTORE),
1174 1, tmp_var);
1175 }
1176
1177 /* Gimplify a BIND_EXPR. Just voidify and recurse. */
1178
1179 static enum gimplify_status
1180 gimplify_bind_expr (tree *expr_p, gimple_seq *pre_p)
1181 {
1182 tree bind_expr = *expr_p;
1183 bool old_save_stack = gimplify_ctxp->save_stack;
1184 tree t;
1185 gimple gimple_bind;
1186 gimple_seq body, cleanup;
1187 gimple stack_save;
1188
1189 tree temp = voidify_wrapper_expr (bind_expr, NULL);
1190
1191 /* Mark variables seen in this bind expr. */
1192 for (t = BIND_EXPR_VARS (bind_expr); t ; t = DECL_CHAIN (t))
1193 {
1194 if (TREE_CODE (t) == VAR_DECL)
1195 {
1196 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
1197
1198 /* Mark variable as local. */
1199 if (ctx && !DECL_EXTERNAL (t)
1200 && (! DECL_SEEN_IN_BIND_EXPR_P (t)
1201 || splay_tree_lookup (ctx->variables,
1202 (splay_tree_key) t) == NULL))
1203 omp_add_variable (gimplify_omp_ctxp, t, GOVD_LOCAL | GOVD_SEEN);
1204
1205 DECL_SEEN_IN_BIND_EXPR_P (t) = 1;
1206
1207 if (DECL_HARD_REGISTER (t) && !is_global_var (t) && cfun)
1208 cfun->has_local_explicit_reg_vars = true;
1209 }
1210
1211 /* Preliminarily mark non-addressed complex variables as eligible
1212 for promotion to gimple registers. We'll transform their uses
1213 as we find them. */
1214 if ((TREE_CODE (TREE_TYPE (t)) == COMPLEX_TYPE
1215 || TREE_CODE (TREE_TYPE (t)) == VECTOR_TYPE)
1216 && !TREE_THIS_VOLATILE (t)
1217 && (TREE_CODE (t) == VAR_DECL && !DECL_HARD_REGISTER (t))
1218 && !needs_to_live_in_memory (t))
1219 DECL_GIMPLE_REG_P (t) = 1;
1220 }
1221
1222 gimple_bind = gimple_build_bind (BIND_EXPR_VARS (bind_expr), NULL,
1223 BIND_EXPR_BLOCK (bind_expr));
1224 gimple_push_bind_expr (gimple_bind);
1225
1226 gimplify_ctxp->save_stack = false;
1227
1228 /* Gimplify the body into the GIMPLE_BIND tuple's body. */
1229 body = NULL;
1230 gimplify_stmt (&BIND_EXPR_BODY (bind_expr), &body);
1231 gimple_bind_set_body (gimple_bind, body);
1232
1233 cleanup = NULL;
1234 stack_save = NULL;
1235 if (gimplify_ctxp->save_stack)
1236 {
1237 gimple stack_restore;
1238
1239 /* Save stack on entry and restore it on exit. Add a try_finally
1240 block to achieve this. Note that mudflap depends on the
1241 format of the emitted code: see mx_register_decls(). */
1242 build_stack_save_restore (&stack_save, &stack_restore);
1243
1244 gimplify_seq_add_stmt (&cleanup, stack_restore);
1245 }
1246
1247 /* Add clobbers for all variables that go out of scope. */
1248 for (t = BIND_EXPR_VARS (bind_expr); t ; t = DECL_CHAIN (t))
1249 {
1250 if (TREE_CODE (t) == VAR_DECL
1251 && !is_global_var (t)
1252 && DECL_CONTEXT (t) == current_function_decl
1253 && !DECL_HARD_REGISTER (t)
1254 && !TREE_THIS_VOLATILE (t)
1255 && !DECL_HAS_VALUE_EXPR_P (t)
1256 /* Only care for variables that have to be in memory. Others
1257 will be rewritten into SSA names, hence moved to the top-level. */
1258 && !is_gimple_reg (t)
1259 && flag_stack_reuse != SR_NONE)
1260 {
1261 tree clobber = build_constructor (TREE_TYPE (t),
1262 NULL);
1263 TREE_THIS_VOLATILE (clobber) = 1;
1264 gimplify_seq_add_stmt (&cleanup, gimple_build_assign (t, clobber));
1265 }
1266 }
1267
1268 if (cleanup)
1269 {
1270 gimple gs;
1271 gimple_seq new_body;
1272
1273 new_body = NULL;
1274 gs = gimple_build_try (gimple_bind_body (gimple_bind), cleanup,
1275 GIMPLE_TRY_FINALLY);
1276
1277 if (stack_save)
1278 gimplify_seq_add_stmt (&new_body, stack_save);
1279 gimplify_seq_add_stmt (&new_body, gs);
1280 gimple_bind_set_body (gimple_bind, new_body);
1281 }
1282
1283 gimplify_ctxp->save_stack = old_save_stack;
1284 gimple_pop_bind_expr ();
1285
1286 gimplify_seq_add_stmt (pre_p, gimple_bind);
1287
1288 if (temp)
1289 {
1290 *expr_p = temp;
1291 return GS_OK;
1292 }
1293
1294 *expr_p = NULL_TREE;
1295 return GS_ALL_DONE;
1296 }
1297
1298 /* Gimplify a RETURN_EXPR. If the expression to be returned is not a
1299 GIMPLE value, it is assigned to a new temporary and the statement is
1300 re-written to return the temporary.
1301
1302 PRE_P points to the sequence where side effects that must happen before
1303 STMT should be stored. */
1304
1305 static enum gimplify_status
1306 gimplify_return_expr (tree stmt, gimple_seq *pre_p)
1307 {
1308 gimple ret;
1309 tree ret_expr = TREE_OPERAND (stmt, 0);
1310 tree result_decl, result;
1311
1312 if (ret_expr == error_mark_node)
1313 return GS_ERROR;
1314
1315 if (!ret_expr
1316 || TREE_CODE (ret_expr) == RESULT_DECL
1317 || ret_expr == error_mark_node)
1318 {
1319 gimple ret = gimple_build_return (ret_expr);
1320 gimple_set_no_warning (ret, TREE_NO_WARNING (stmt));
1321 gimplify_seq_add_stmt (pre_p, ret);
1322 return GS_ALL_DONE;
1323 }
1324
1325 if (VOID_TYPE_P (TREE_TYPE (TREE_TYPE (current_function_decl))))
1326 result_decl = NULL_TREE;
1327 else
1328 {
1329 result_decl = TREE_OPERAND (ret_expr, 0);
1330
1331 /* See through a return by reference. */
1332 if (TREE_CODE (result_decl) == INDIRECT_REF)
1333 result_decl = TREE_OPERAND (result_decl, 0);
1334
1335 gcc_assert ((TREE_CODE (ret_expr) == MODIFY_EXPR
1336 || TREE_CODE (ret_expr) == INIT_EXPR)
1337 && TREE_CODE (result_decl) == RESULT_DECL);
1338 }
1339
1340 /* If aggregate_value_p is true, then we can return the bare RESULT_DECL.
1341 Recall that aggregate_value_p is FALSE for any aggregate type that is
1342 returned in registers. If we're returning values in registers, then
1343 we don't want to extend the lifetime of the RESULT_DECL, particularly
1344 across another call. In addition, for those aggregates for which
1345 hard_function_value generates a PARALLEL, we'll die during normal
1346 expansion of structure assignments; there's special code in expand_return
1347 to handle this case that does not exist in expand_expr. */
1348 if (!result_decl)
1349 result = NULL_TREE;
1350 else if (aggregate_value_p (result_decl, TREE_TYPE (current_function_decl)))
1351 {
1352 if (TREE_CODE (DECL_SIZE (result_decl)) != INTEGER_CST)
1353 {
1354 if (!TYPE_SIZES_GIMPLIFIED (TREE_TYPE (result_decl)))
1355 gimplify_type_sizes (TREE_TYPE (result_decl), pre_p);
1356 /* Note that we don't use gimplify_vla_decl because the RESULT_DECL
1357 should be effectively allocated by the caller, i.e. all calls to
1358 this function must be subject to the Return Slot Optimization. */
1359 gimplify_one_sizepos (&DECL_SIZE (result_decl), pre_p);
1360 gimplify_one_sizepos (&DECL_SIZE_UNIT (result_decl), pre_p);
1361 }
1362 result = result_decl;
1363 }
1364 else if (gimplify_ctxp->return_temp)
1365 result = gimplify_ctxp->return_temp;
1366 else
1367 {
1368 result = create_tmp_reg (TREE_TYPE (result_decl), NULL);
1369
1370 /* ??? With complex control flow (usually involving abnormal edges),
1371 we can wind up warning about an uninitialized value for this. Due
1372 to how this variable is constructed and initialized, this is never
1373 true. Give up and never warn. */
1374 TREE_NO_WARNING (result) = 1;
1375
1376 gimplify_ctxp->return_temp = result;
1377 }
1378
1379 /* Smash the lhs of the MODIFY_EXPR to the temporary we plan to use.
1380 Then gimplify the whole thing. */
1381 if (result != result_decl)
1382 TREE_OPERAND (ret_expr, 0) = result;
1383
1384 gimplify_and_add (TREE_OPERAND (stmt, 0), pre_p);
1385
1386 ret = gimple_build_return (result);
1387 gimple_set_no_warning (ret, TREE_NO_WARNING (stmt));
1388 gimplify_seq_add_stmt (pre_p, ret);
1389
1390 return GS_ALL_DONE;
1391 }
1392
1393 /* Gimplify a variable-length array DECL. */
1394
1395 static void
1396 gimplify_vla_decl (tree decl, gimple_seq *seq_p)
1397 {
1398 /* This is a variable-sized decl. Simplify its size and mark it
1399 for deferred expansion. Note that mudflap depends on the format
1400 of the emitted code: see mx_register_decls(). */
1401 tree t, addr, ptr_type;
1402
1403 gimplify_one_sizepos (&DECL_SIZE (decl), seq_p);
1404 gimplify_one_sizepos (&DECL_SIZE_UNIT (decl), seq_p);
1405
1406 /* All occurrences of this decl in final gimplified code will be
1407 replaced by indirection. Setting DECL_VALUE_EXPR does two
1408 things: First, it lets the rest of the gimplifier know what
1409 replacement to use. Second, it lets the debug info know
1410 where to find the value. */
1411 ptr_type = build_pointer_type (TREE_TYPE (decl));
1412 addr = create_tmp_var (ptr_type, get_name (decl));
1413 DECL_IGNORED_P (addr) = 0;
1414 t = build_fold_indirect_ref (addr);
1415 TREE_THIS_NOTRAP (t) = 1;
1416 SET_DECL_VALUE_EXPR (decl, t);
1417 DECL_HAS_VALUE_EXPR_P (decl) = 1;
1418
1419 t = builtin_decl_explicit (BUILT_IN_ALLOCA_WITH_ALIGN);
1420 t = build_call_expr (t, 2, DECL_SIZE_UNIT (decl),
1421 size_int (DECL_ALIGN (decl)));
1422 /* The call has been built for a variable-sized object. */
1423 CALL_ALLOCA_FOR_VAR_P (t) = 1;
1424 t = fold_convert (ptr_type, t);
1425 t = build2 (MODIFY_EXPR, TREE_TYPE (addr), addr, t);
1426
1427 gimplify_and_add (t, seq_p);
1428
1429 /* Indicate that we need to restore the stack level when the
1430 enclosing BIND_EXPR is exited. */
1431 gimplify_ctxp->save_stack = true;
1432 }
1433
1434 /* Gimplify a DECL_EXPR node *STMT_P by making any necessary allocation
1435 and initialization explicit. */
1436
1437 static enum gimplify_status
1438 gimplify_decl_expr (tree *stmt_p, gimple_seq *seq_p)
1439 {
1440 tree stmt = *stmt_p;
1441 tree decl = DECL_EXPR_DECL (stmt);
1442
1443 *stmt_p = NULL_TREE;
1444
1445 if (TREE_TYPE (decl) == error_mark_node)
1446 return GS_ERROR;
1447
1448 if ((TREE_CODE (decl) == TYPE_DECL
1449 || TREE_CODE (decl) == VAR_DECL)
1450 && !TYPE_SIZES_GIMPLIFIED (TREE_TYPE (decl)))
1451 gimplify_type_sizes (TREE_TYPE (decl), seq_p);
1452
1453 /* ??? DECL_ORIGINAL_TYPE is streamed for LTO so it needs to be gimplified
1454 in case its size expressions contain problematic nodes like CALL_EXPR. */
1455 if (TREE_CODE (decl) == TYPE_DECL
1456 && DECL_ORIGINAL_TYPE (decl)
1457 && !TYPE_SIZES_GIMPLIFIED (DECL_ORIGINAL_TYPE (decl)))
1458 gimplify_type_sizes (DECL_ORIGINAL_TYPE (decl), seq_p);
1459
1460 if (TREE_CODE (decl) == VAR_DECL && !DECL_EXTERNAL (decl))
1461 {
1462 tree init = DECL_INITIAL (decl);
1463
1464 if (TREE_CODE (DECL_SIZE_UNIT (decl)) != INTEGER_CST
1465 || (!TREE_STATIC (decl)
1466 && flag_stack_check == GENERIC_STACK_CHECK
1467 && compare_tree_int (DECL_SIZE_UNIT (decl),
1468 STACK_CHECK_MAX_VAR_SIZE) > 0))
1469 gimplify_vla_decl (decl, seq_p);
1470
1471 /* Some front ends do not explicitly declare all anonymous
1472 artificial variables. We compensate here by declaring the
1473 variables, though it would be better if the front ends would
1474 explicitly declare them. */
1475 if (!DECL_SEEN_IN_BIND_EXPR_P (decl)
1476 && DECL_ARTIFICIAL (decl) && DECL_NAME (decl) == NULL_TREE)
1477 gimple_add_tmp_var (decl);
1478
1479 if (init && init != error_mark_node)
1480 {
1481 if (!TREE_STATIC (decl))
1482 {
1483 DECL_INITIAL (decl) = NULL_TREE;
1484 init = build2 (INIT_EXPR, void_type_node, decl, init);
1485 gimplify_and_add (init, seq_p);
1486 ggc_free (init);
1487 }
1488 else
1489 /* We must still examine initializers for static variables
1490 as they may contain a label address. */
1491 walk_tree (&init, force_labels_r, NULL, NULL);
1492 }
1493 }
1494
1495 return GS_ALL_DONE;
1496 }
1497
1498 /* Gimplify a LOOP_EXPR. Normally this just involves gimplifying the body
1499 and replacing the LOOP_EXPR with goto, but if the loop contains an
1500 EXIT_EXPR, we need to append a label for it to jump to. */
1501
1502 static enum gimplify_status
1503 gimplify_loop_expr (tree *expr_p, gimple_seq *pre_p)
1504 {
1505 tree saved_label = gimplify_ctxp->exit_label;
1506 tree start_label = create_artificial_label (UNKNOWN_LOCATION);
1507
1508 gimplify_seq_add_stmt (pre_p, gimple_build_label (start_label));
1509
1510 gimplify_ctxp->exit_label = NULL_TREE;
1511
1512 gimplify_and_add (LOOP_EXPR_BODY (*expr_p), pre_p);
1513
1514 gimplify_seq_add_stmt (pre_p, gimple_build_goto (start_label));
1515
1516 if (gimplify_ctxp->exit_label)
1517 gimplify_seq_add_stmt (pre_p,
1518 gimple_build_label (gimplify_ctxp->exit_label));
1519
1520 gimplify_ctxp->exit_label = saved_label;
1521
1522 *expr_p = NULL;
1523 return GS_ALL_DONE;
1524 }
1525
1526 /* Gimplify a statement list onto a sequence. These may be created either
1527 by an enlightened front-end, or by shortcut_cond_expr. */
1528
1529 static enum gimplify_status
1530 gimplify_statement_list (tree *expr_p, gimple_seq *pre_p)
1531 {
1532 tree temp = voidify_wrapper_expr (*expr_p, NULL);
1533
1534 tree_stmt_iterator i = tsi_start (*expr_p);
1535
1536 while (!tsi_end_p (i))
1537 {
1538 gimplify_stmt (tsi_stmt_ptr (i), pre_p);
1539 tsi_delink (&i);
1540 }
1541
1542 if (temp)
1543 {
1544 *expr_p = temp;
1545 return GS_OK;
1546 }
1547
1548 return GS_ALL_DONE;
1549 }
1550 \f
1551 /* Compare two case labels. Because the front end should already have
1552 made sure that case ranges do not overlap, it is enough to only compare
1553 the CASE_LOW values of each case label. */
1554
1555 static int
1556 compare_case_labels (const void *p1, const void *p2)
1557 {
1558 const_tree const case1 = *(const_tree const*)p1;
1559 const_tree const case2 = *(const_tree const*)p2;
1560
1561 /* The 'default' case label always goes first. */
1562 if (!CASE_LOW (case1))
1563 return -1;
1564 else if (!CASE_LOW (case2))
1565 return 1;
1566 else
1567 return tree_int_cst_compare (CASE_LOW (case1), CASE_LOW (case2));
1568 }
1569
1570 /* Sort the case labels in LABEL_VEC in place in ascending order. */
1571
1572 void
1573 sort_case_labels (vec<tree> label_vec)
1574 {
1575 label_vec.qsort (compare_case_labels);
1576 }
1577 \f
1578 /* Prepare a vector of case labels to be used in a GIMPLE_SWITCH statement.
1579
1580 LABELS is a vector that contains all case labels to look at.
1581
1582 INDEX_TYPE is the type of the switch index expression. Case labels
1583 in LABELS are discarded if their values are not in the value range
1584 covered by INDEX_TYPE. The remaining case label values are folded
1585 to INDEX_TYPE.
1586
1587 If a default case exists in LABELS, it is removed from LABELS and
1588 returned in DEFAULT_CASEP. If no default case exists, but the
1589 case labels already cover the whole range of INDEX_TYPE, a default
1590 case is returned pointing to one of the existing case labels.
1591 Otherwise DEFAULT_CASEP is set to NULL_TREE.
1592
1593 DEFAULT_CASEP may be NULL, in which case the above comment doesn't
1594 apply and no action is taken regardless of whether a default case is
1595 found or not. */
1596
1597 void
1598 preprocess_case_label_vec_for_gimple (vec<tree> labels,
1599 tree index_type,
1600 tree *default_casep)
1601 {
1602 tree min_value, max_value;
1603 tree default_case = NULL_TREE;
1604 size_t i, len;
1605
1606 i = 0;
1607 min_value = TYPE_MIN_VALUE (index_type);
1608 max_value = TYPE_MAX_VALUE (index_type);
1609 while (i < labels.length ())
1610 {
1611 tree elt = labels[i];
1612 tree low = CASE_LOW (elt);
1613 tree high = CASE_HIGH (elt);
1614 bool remove_element = FALSE;
1615
1616 if (low)
1617 {
1618 gcc_checking_assert (TREE_CODE (low) == INTEGER_CST);
1619 gcc_checking_assert (!high || TREE_CODE (high) == INTEGER_CST);
1620
1621 /* This is a non-default case label, i.e. it has a value.
1622
1623 See if the case label is reachable within the range of
1624 the index type. Remove out-of-range case values. Turn
1625 case ranges into a canonical form (high > low strictly)
1626 and convert the case label values to the index type.
1627
1628 NB: The type of gimple_switch_index() may be the promoted
1629 type, but the case labels retain the original type. */
1630
1631 if (high)
1632 {
1633 /* This is a case range. Discard empty ranges.
1634 If the bounds or the range are equal, turn this
1635 into a simple (one-value) case. */
1636 int cmp = tree_int_cst_compare (high, low);
1637 if (cmp < 0)
1638 remove_element = TRUE;
1639 else if (cmp == 0)
1640 high = NULL_TREE;
1641 }
1642
1643 if (! high)
1644 {
1645 /* If the simple case value is unreachable, ignore it. */
1646 if ((TREE_CODE (min_value) == INTEGER_CST
1647 && tree_int_cst_compare (low, min_value) < 0)
1648 || (TREE_CODE (max_value) == INTEGER_CST
1649 && tree_int_cst_compare (low, max_value) > 0))
1650 remove_element = TRUE;
1651 else
1652 low = fold_convert (index_type, low);
1653 }
1654 else
1655 {
1656 /* If the entire case range is unreachable, ignore it. */
1657 if ((TREE_CODE (min_value) == INTEGER_CST
1658 && tree_int_cst_compare (high, min_value) < 0)
1659 || (TREE_CODE (max_value) == INTEGER_CST
1660 && tree_int_cst_compare (low, max_value) > 0))
1661 remove_element = TRUE;
1662 else
1663 {
1664 /* If the lower bound is less than the index type's
1665 minimum value, truncate the range bounds. */
1666 if (TREE_CODE (min_value) == INTEGER_CST
1667 && tree_int_cst_compare (low, min_value) < 0)
1668 low = min_value;
1669 low = fold_convert (index_type, low);
1670
1671 /* If the upper bound is greater than the index type's
1672 maximum value, truncate the range bounds. */
1673 if (TREE_CODE (max_value) == INTEGER_CST
1674 && tree_int_cst_compare (high, max_value) > 0)
1675 high = max_value;
1676 high = fold_convert (index_type, high);
1677
1678 /* We may have folded a case range to a one-value case. */
1679 if (tree_int_cst_equal (low, high))
1680 high = NULL_TREE;
1681 }
1682 }
1683
1684 CASE_LOW (elt) = low;
1685 CASE_HIGH (elt) = high;
1686 }
1687 else
1688 {
1689 gcc_assert (!default_case);
1690 default_case = elt;
1691 /* The default case must be passed separately to the
1692 gimple_build_switch routine. But if DEFAULT_CASEP
1693 is NULL, we do not remove the default case (it would
1694 be completely lost). */
1695 if (default_casep)
1696 remove_element = TRUE;
1697 }
1698
1699 if (remove_element)
1700 labels.ordered_remove (i);
1701 else
1702 i++;
1703 }
1704 len = i;
1705
1706 if (!labels.is_empty ())
1707 sort_case_labels (labels);
1708
1709 if (default_casep && !default_case)
1710 {
1711 /* If the switch has no default label, add one, so that we jump
1712 around the switch body. If the labels already cover the whole
1713 range of the switch index_type, add the default label pointing
1714 to one of the existing labels. */
1715 if (len
1716 && TYPE_MIN_VALUE (index_type)
1717 && TYPE_MAX_VALUE (index_type)
1718 && tree_int_cst_equal (CASE_LOW (labels[0]),
1719 TYPE_MIN_VALUE (index_type)))
1720 {
1721 tree low, high = CASE_HIGH (labels[len - 1]);
1722 if (!high)
1723 high = CASE_LOW (labels[len - 1]);
1724 if (tree_int_cst_equal (high, TYPE_MAX_VALUE (index_type)))
1725 {
1726 for (i = 1; i < len; i++)
1727 {
1728 high = CASE_LOW (labels[i]);
1729 low = CASE_HIGH (labels[i - 1]);
1730 if (!low)
1731 low = CASE_LOW (labels[i - 1]);
1732 if ((TREE_INT_CST_LOW (low) + 1
1733 != TREE_INT_CST_LOW (high))
1734 || (TREE_INT_CST_HIGH (low)
1735 + (TREE_INT_CST_LOW (high) == 0)
1736 != TREE_INT_CST_HIGH (high)))
1737 break;
1738 }
1739 if (i == len)
1740 {
1741 tree label = CASE_LABEL (labels[0]);
1742 default_case = build_case_label (NULL_TREE, NULL_TREE,
1743 label);
1744 }
1745 }
1746 }
1747 }
1748
1749 if (default_casep)
1750 *default_casep = default_case;
1751 }
1752 \f
1753 /* Gimplify a SWITCH_EXPR, and collect the vector of labels it can
1754 branch to. */
1755
1756 static enum gimplify_status
1757 gimplify_switch_expr (tree *expr_p, gimple_seq *pre_p)
1758 {
1759 tree switch_expr = *expr_p;
1760 gimple_seq switch_body_seq = NULL;
1761 enum gimplify_status ret;
1762 tree index_type = TREE_TYPE (switch_expr);
1763 if (index_type == NULL_TREE)
1764 index_type = TREE_TYPE (SWITCH_COND (switch_expr));
1765
1766 ret = gimplify_expr (&SWITCH_COND (switch_expr), pre_p, NULL, is_gimple_val,
1767 fb_rvalue);
1768 if (ret == GS_ERROR || ret == GS_UNHANDLED)
1769 return ret;
1770
1771 if (SWITCH_BODY (switch_expr))
1772 {
1773 vec<tree> labels;
1774 vec<tree> saved_labels;
1775 tree default_case = NULL_TREE;
1776 gimple gimple_switch;
1777
1778 /* If someone can be bothered to fill in the labels, they can
1779 be bothered to null out the body too. */
1780 gcc_assert (!SWITCH_LABELS (switch_expr));
1781
1782 /* Save old labels, get new ones from body, then restore the old
1783 labels. Save all the things from the switch body to append after. */
1784 saved_labels = gimplify_ctxp->case_labels;
1785 gimplify_ctxp->case_labels.create (8);
1786
1787 gimplify_stmt (&SWITCH_BODY (switch_expr), &switch_body_seq);
1788 labels = gimplify_ctxp->case_labels;
1789 gimplify_ctxp->case_labels = saved_labels;
1790
1791 preprocess_case_label_vec_for_gimple (labels, index_type,
1792 &default_case);
1793
1794 if (!default_case)
1795 {
1796 gimple new_default;
1797
1798 default_case
1799 = build_case_label (NULL_TREE, NULL_TREE,
1800 create_artificial_label (UNKNOWN_LOCATION));
1801 new_default = gimple_build_label (CASE_LABEL (default_case));
1802 gimplify_seq_add_stmt (&switch_body_seq, new_default);
1803 }
1804
1805 gimple_switch = gimple_build_switch (SWITCH_COND (switch_expr),
1806 default_case, labels);
1807 gimplify_seq_add_stmt (pre_p, gimple_switch);
1808 gimplify_seq_add_seq (pre_p, switch_body_seq);
1809 labels.release ();
1810 }
1811 else
1812 gcc_assert (SWITCH_LABELS (switch_expr));
1813
1814 return GS_ALL_DONE;
1815 }
1816
1817 /* Gimplify the CASE_LABEL_EXPR pointed to by EXPR_P. */
1818
1819 static enum gimplify_status
1820 gimplify_case_label_expr (tree *expr_p, gimple_seq *pre_p)
1821 {
1822 struct gimplify_ctx *ctxp;
1823 gimple gimple_label;
1824
1825 /* Invalid OpenMP programs can play Duff's Device type games with
1826 #pragma omp parallel. At least in the C front end, we don't
1827 detect such invalid branches until after gimplification. */
1828 for (ctxp = gimplify_ctxp; ; ctxp = ctxp->prev_context)
1829 if (ctxp->case_labels.exists ())
1830 break;
1831
1832 gimple_label = gimple_build_label (CASE_LABEL (*expr_p));
1833 ctxp->case_labels.safe_push (*expr_p);
1834 gimplify_seq_add_stmt (pre_p, gimple_label);
1835
1836 return GS_ALL_DONE;
1837 }
1838
1839 /* Build a GOTO to the LABEL_DECL pointed to by LABEL_P, building it first
1840 if necessary. */
1841
1842 tree
1843 build_and_jump (tree *label_p)
1844 {
1845 if (label_p == NULL)
1846 /* If there's nowhere to jump, just fall through. */
1847 return NULL_TREE;
1848
1849 if (*label_p == NULL_TREE)
1850 {
1851 tree label = create_artificial_label (UNKNOWN_LOCATION);
1852 *label_p = label;
1853 }
1854
1855 return build1 (GOTO_EXPR, void_type_node, *label_p);
1856 }
1857
1858 /* Gimplify an EXIT_EXPR by converting to a GOTO_EXPR inside a COND_EXPR.
1859 This also involves building a label to jump to and communicating it to
1860 gimplify_loop_expr through gimplify_ctxp->exit_label. */
1861
1862 static enum gimplify_status
1863 gimplify_exit_expr (tree *expr_p)
1864 {
1865 tree cond = TREE_OPERAND (*expr_p, 0);
1866 tree expr;
1867
1868 expr = build_and_jump (&gimplify_ctxp->exit_label);
1869 expr = build3 (COND_EXPR, void_type_node, cond, expr, NULL_TREE);
1870 *expr_p = expr;
1871
1872 return GS_OK;
1873 }
1874
1875 /* A helper function to be called via walk_tree. Mark all labels under *TP
1876 as being forced. To be called for DECL_INITIAL of static variables. */
1877
1878 tree
1879 force_labels_r (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED)
1880 {
1881 if (TYPE_P (*tp))
1882 *walk_subtrees = 0;
1883 if (TREE_CODE (*tp) == LABEL_DECL)
1884 FORCED_LABEL (*tp) = 1;
1885
1886 return NULL_TREE;
1887 }
1888
1889 /* *EXPR_P is a COMPONENT_REF being used as an rvalue. If its type is
1890 different from its canonical type, wrap the whole thing inside a
1891 NOP_EXPR and force the type of the COMPONENT_REF to be the canonical
1892 type.
1893
1894 The canonical type of a COMPONENT_REF is the type of the field being
1895 referenced--unless the field is a bit-field which can be read directly
1896 in a smaller mode, in which case the canonical type is the
1897 sign-appropriate type corresponding to that mode. */
1898
1899 static void
1900 canonicalize_component_ref (tree *expr_p)
1901 {
1902 tree expr = *expr_p;
1903 tree type;
1904
1905 gcc_assert (TREE_CODE (expr) == COMPONENT_REF);
1906
1907 if (INTEGRAL_TYPE_P (TREE_TYPE (expr)))
1908 type = TREE_TYPE (get_unwidened (expr, NULL_TREE));
1909 else
1910 type = TREE_TYPE (TREE_OPERAND (expr, 1));
1911
1912 /* One could argue that all the stuff below is not necessary for
1913 the non-bitfield case and declare it a FE error if type
1914 adjustment would be needed. */
1915 if (TREE_TYPE (expr) != type)
1916 {
1917 #ifdef ENABLE_TYPES_CHECKING
1918 tree old_type = TREE_TYPE (expr);
1919 #endif
1920 int type_quals;
1921
1922 /* We need to preserve qualifiers and propagate them from
1923 operand 0. */
1924 type_quals = TYPE_QUALS (type)
1925 | TYPE_QUALS (TREE_TYPE (TREE_OPERAND (expr, 0)));
1926 if (TYPE_QUALS (type) != type_quals)
1927 type = build_qualified_type (TYPE_MAIN_VARIANT (type), type_quals);
1928
1929 /* Set the type of the COMPONENT_REF to the underlying type. */
1930 TREE_TYPE (expr) = type;
1931
1932 #ifdef ENABLE_TYPES_CHECKING
1933 /* It is now a FE error, if the conversion from the canonical
1934 type to the original expression type is not useless. */
1935 gcc_assert (useless_type_conversion_p (old_type, type));
1936 #endif
1937 }
1938 }
1939
1940 /* If a NOP conversion is changing a pointer to array of foo to a pointer
1941 to foo, embed that change in the ADDR_EXPR by converting
1942 T array[U];
1943 (T *)&array
1944 ==>
1945 &array[L]
1946 where L is the lower bound. For simplicity, only do this for constant
1947 lower bound.
1948 The constraint is that the type of &array[L] is trivially convertible
1949 to T *. */
1950
1951 static void
1952 canonicalize_addr_expr (tree *expr_p)
1953 {
1954 tree expr = *expr_p;
1955 tree addr_expr = TREE_OPERAND (expr, 0);
1956 tree datype, ddatype, pddatype;
1957
1958 /* We simplify only conversions from an ADDR_EXPR to a pointer type. */
1959 if (!POINTER_TYPE_P (TREE_TYPE (expr))
1960 || TREE_CODE (addr_expr) != ADDR_EXPR)
1961 return;
1962
1963 /* The addr_expr type should be a pointer to an array. */
1964 datype = TREE_TYPE (TREE_TYPE (addr_expr));
1965 if (TREE_CODE (datype) != ARRAY_TYPE)
1966 return;
1967
1968 /* The pointer to element type shall be trivially convertible to
1969 the expression pointer type. */
1970 ddatype = TREE_TYPE (datype);
1971 pddatype = build_pointer_type (ddatype);
1972 if (!useless_type_conversion_p (TYPE_MAIN_VARIANT (TREE_TYPE (expr)),
1973 pddatype))
1974 return;
1975
1976 /* The lower bound and element sizes must be constant. */
1977 if (!TYPE_SIZE_UNIT (ddatype)
1978 || TREE_CODE (TYPE_SIZE_UNIT (ddatype)) != INTEGER_CST
1979 || !TYPE_DOMAIN (datype) || !TYPE_MIN_VALUE (TYPE_DOMAIN (datype))
1980 || TREE_CODE (TYPE_MIN_VALUE (TYPE_DOMAIN (datype))) != INTEGER_CST)
1981 return;
1982
1983 /* All checks succeeded. Build a new node to merge the cast. */
1984 *expr_p = build4 (ARRAY_REF, ddatype, TREE_OPERAND (addr_expr, 0),
1985 TYPE_MIN_VALUE (TYPE_DOMAIN (datype)),
1986 NULL_TREE, NULL_TREE);
1987 *expr_p = build1 (ADDR_EXPR, pddatype, *expr_p);
1988
1989 /* We can have stripped a required restrict qualifier above. */
1990 if (!useless_type_conversion_p (TREE_TYPE (expr), TREE_TYPE (*expr_p)))
1991 *expr_p = fold_convert (TREE_TYPE (expr), *expr_p);
1992 }
1993
1994 /* *EXPR_P is a NOP_EXPR or CONVERT_EXPR. Remove it and/or other conversions
1995 underneath as appropriate. */
1996
1997 static enum gimplify_status
1998 gimplify_conversion (tree *expr_p)
1999 {
2000 location_t loc = EXPR_LOCATION (*expr_p);
2001 gcc_assert (CONVERT_EXPR_P (*expr_p));
2002
2003 /* Then strip away all but the outermost conversion. */
2004 STRIP_SIGN_NOPS (TREE_OPERAND (*expr_p, 0));
2005
2006 /* And remove the outermost conversion if it's useless. */
2007 if (tree_ssa_useless_type_conversion (*expr_p))
2008 *expr_p = TREE_OPERAND (*expr_p, 0);
2009
2010 /* If we still have a conversion at the toplevel,
2011 then canonicalize some constructs. */
2012 if (CONVERT_EXPR_P (*expr_p))
2013 {
2014 tree sub = TREE_OPERAND (*expr_p, 0);
2015
2016 /* If a NOP conversion is changing the type of a COMPONENT_REF
2017 expression, then canonicalize its type now in order to expose more
2018 redundant conversions. */
2019 if (TREE_CODE (sub) == COMPONENT_REF)
2020 canonicalize_component_ref (&TREE_OPERAND (*expr_p, 0));
2021
2022 /* If a NOP conversion is changing a pointer to array of foo
2023 to a pointer to foo, embed that change in the ADDR_EXPR. */
2024 else if (TREE_CODE (sub) == ADDR_EXPR)
2025 canonicalize_addr_expr (expr_p);
2026 }
2027
2028 /* If we have a conversion to a non-register type force the
2029 use of a VIEW_CONVERT_EXPR instead. */
2030 if (CONVERT_EXPR_P (*expr_p) && !is_gimple_reg_type (TREE_TYPE (*expr_p)))
2031 *expr_p = fold_build1_loc (loc, VIEW_CONVERT_EXPR, TREE_TYPE (*expr_p),
2032 TREE_OPERAND (*expr_p, 0));
2033
2034 return GS_OK;
2035 }
2036
2037 /* Nonlocal VLAs seen in the current function. */
2038 static struct pointer_set_t *nonlocal_vlas;
2039
2040 /* Gimplify a VAR_DECL or PARM_DECL. Return GS_OK if we expanded a
2041 DECL_VALUE_EXPR, and it's worth re-examining things. */
2042
2043 static enum gimplify_status
2044 gimplify_var_or_parm_decl (tree *expr_p)
2045 {
2046 tree decl = *expr_p;
2047
2048 /* ??? If this is a local variable, and it has not been seen in any
2049 outer BIND_EXPR, then it's probably the result of a duplicate
2050 declaration, for which we've already issued an error. It would
2051 be really nice if the front end wouldn't leak these at all.
2052 Currently the only known culprit is C++ destructors, as seen
2053 in g++.old-deja/g++.jason/binding.C. */
2054 if (TREE_CODE (decl) == VAR_DECL
2055 && !DECL_SEEN_IN_BIND_EXPR_P (decl)
2056 && !TREE_STATIC (decl) && !DECL_EXTERNAL (decl)
2057 && decl_function_context (decl) == current_function_decl)
2058 {
2059 gcc_assert (seen_error ());
2060 return GS_ERROR;
2061 }
2062
2063 /* When within an OpenMP context, notice uses of variables. */
2064 if (gimplify_omp_ctxp && omp_notice_variable (gimplify_omp_ctxp, decl, true))
2065 return GS_ALL_DONE;
2066
2067 /* If the decl is an alias for another expression, substitute it now. */
2068 if (DECL_HAS_VALUE_EXPR_P (decl))
2069 {
2070 tree value_expr = DECL_VALUE_EXPR (decl);
2071
2072 /* For referenced nonlocal VLAs add a decl for debugging purposes
2073 to the current function. */
2074 if (TREE_CODE (decl) == VAR_DECL
2075 && TREE_CODE (DECL_SIZE_UNIT (decl)) != INTEGER_CST
2076 && nonlocal_vlas != NULL
2077 && TREE_CODE (value_expr) == INDIRECT_REF
2078 && TREE_CODE (TREE_OPERAND (value_expr, 0)) == VAR_DECL
2079 && decl_function_context (decl) != current_function_decl)
2080 {
2081 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
2082 while (ctx && ctx->region_type == ORT_WORKSHARE)
2083 ctx = ctx->outer_context;
2084 if (!ctx && !pointer_set_insert (nonlocal_vlas, decl))
2085 {
2086 tree copy = copy_node (decl), block;
2087
2088 lang_hooks.dup_lang_specific_decl (copy);
2089 SET_DECL_RTL (copy, 0);
2090 TREE_USED (copy) = 1;
2091 block = DECL_INITIAL (current_function_decl);
2092 DECL_CHAIN (copy) = BLOCK_VARS (block);
2093 BLOCK_VARS (block) = copy;
2094 SET_DECL_VALUE_EXPR (copy, unshare_expr (value_expr));
2095 DECL_HAS_VALUE_EXPR_P (copy) = 1;
2096 }
2097 }
2098
2099 *expr_p = unshare_expr (value_expr);
2100 return GS_OK;
2101 }
2102
2103 return GS_ALL_DONE;
2104 }
2105
2106 /* Gimplify the COMPONENT_REF, ARRAY_REF, REALPART_EXPR or IMAGPART_EXPR
2107 node *EXPR_P.
2108
2109 compound_lval
2110 : min_lval '[' val ']'
2111 | min_lval '.' ID
2112 | compound_lval '[' val ']'
2113 | compound_lval '.' ID
2114
2115 This is not part of the original SIMPLE definition, which separates
2116 array and member references, but it seems reasonable to handle them
2117 together. Also, this way we don't run into problems with union
2118 aliasing; gcc requires that for accesses through a union to alias, the
2119 union reference must be explicit, which was not always the case when we
2120 were splitting up array and member refs.
2121
2122 PRE_P points to the sequence where side effects that must happen before
2123 *EXPR_P should be stored.
2124
2125 POST_P points to the sequence where side effects that must happen after
2126 *EXPR_P should be stored. */
2127
2128 static enum gimplify_status
2129 gimplify_compound_lval (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
2130 fallback_t fallback)
2131 {
2132 tree *p;
2133 vec<tree> expr_stack;
2134 enum gimplify_status ret = GS_ALL_DONE, tret;
2135 int i;
2136 location_t loc = EXPR_LOCATION (*expr_p);
2137 tree expr = *expr_p;
2138
2139 /* Create a stack of the subexpressions so later we can walk them in
2140 order from inner to outer. */
2141 expr_stack.create (10);
2142
2143 /* We can handle anything that get_inner_reference can deal with. */
2144 for (p = expr_p; ; p = &TREE_OPERAND (*p, 0))
2145 {
2146 restart:
2147 /* Fold INDIRECT_REFs now to turn them into ARRAY_REFs. */
2148 if (TREE_CODE (*p) == INDIRECT_REF)
2149 *p = fold_indirect_ref_loc (loc, *p);
2150
2151 if (handled_component_p (*p))
2152 ;
2153 /* Expand DECL_VALUE_EXPR now. In some cases that may expose
2154 additional COMPONENT_REFs. */
2155 else if ((TREE_CODE (*p) == VAR_DECL || TREE_CODE (*p) == PARM_DECL)
2156 && gimplify_var_or_parm_decl (p) == GS_OK)
2157 goto restart;
2158 else
2159 break;
2160
2161 expr_stack.safe_push (*p);
2162 }
2163
2164 gcc_assert (expr_stack.length ());
2165
2166 /* Now EXPR_STACK is a stack of pointers to all the refs we've
2167 walked through and P points to the innermost expression.
2168
2169 Java requires that we elaborated nodes in source order. That
2170 means we must gimplify the inner expression followed by each of
2171 the indices, in order. But we can't gimplify the inner
2172 expression until we deal with any variable bounds, sizes, or
2173 positions in order to deal with PLACEHOLDER_EXPRs.
2174
2175 So we do this in three steps. First we deal with the annotations
2176 for any variables in the components, then we gimplify the base,
2177 then we gimplify any indices, from left to right. */
2178 for (i = expr_stack.length () - 1; i >= 0; i--)
2179 {
2180 tree t = expr_stack[i];
2181
2182 if (TREE_CODE (t) == ARRAY_REF || TREE_CODE (t) == ARRAY_RANGE_REF)
2183 {
2184 /* Gimplify the low bound and element type size and put them into
2185 the ARRAY_REF. If these values are set, they have already been
2186 gimplified. */
2187 if (TREE_OPERAND (t, 2) == NULL_TREE)
2188 {
2189 tree low = unshare_expr (array_ref_low_bound (t));
2190 if (!is_gimple_min_invariant (low))
2191 {
2192 TREE_OPERAND (t, 2) = low;
2193 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p,
2194 post_p, is_gimple_reg,
2195 fb_rvalue);
2196 ret = MIN (ret, tret);
2197 }
2198 }
2199 else
2200 {
2201 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p, post_p,
2202 is_gimple_reg, fb_rvalue);
2203 ret = MIN (ret, tret);
2204 }
2205
2206 if (TREE_OPERAND (t, 3) == NULL_TREE)
2207 {
2208 tree elmt_type = TREE_TYPE (TREE_TYPE (TREE_OPERAND (t, 0)));
2209 tree elmt_size = unshare_expr (array_ref_element_size (t));
2210 tree factor = size_int (TYPE_ALIGN_UNIT (elmt_type));
2211
2212 /* Divide the element size by the alignment of the element
2213 type (above). */
2214 elmt_size
2215 = size_binop_loc (loc, EXACT_DIV_EXPR, elmt_size, factor);
2216
2217 if (!is_gimple_min_invariant (elmt_size))
2218 {
2219 TREE_OPERAND (t, 3) = elmt_size;
2220 tret = gimplify_expr (&TREE_OPERAND (t, 3), pre_p,
2221 post_p, is_gimple_reg,
2222 fb_rvalue);
2223 ret = MIN (ret, tret);
2224 }
2225 }
2226 else
2227 {
2228 tret = gimplify_expr (&TREE_OPERAND (t, 3), pre_p, post_p,
2229 is_gimple_reg, fb_rvalue);
2230 ret = MIN (ret, tret);
2231 }
2232 }
2233 else if (TREE_CODE (t) == COMPONENT_REF)
2234 {
2235 /* Set the field offset into T and gimplify it. */
2236 if (TREE_OPERAND (t, 2) == NULL_TREE)
2237 {
2238 tree offset = unshare_expr (component_ref_field_offset (t));
2239 tree field = TREE_OPERAND (t, 1);
2240 tree factor
2241 = size_int (DECL_OFFSET_ALIGN (field) / BITS_PER_UNIT);
2242
2243 /* Divide the offset by its alignment. */
2244 offset = size_binop_loc (loc, EXACT_DIV_EXPR, offset, factor);
2245
2246 if (!is_gimple_min_invariant (offset))
2247 {
2248 TREE_OPERAND (t, 2) = offset;
2249 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p,
2250 post_p, is_gimple_reg,
2251 fb_rvalue);
2252 ret = MIN (ret, tret);
2253 }
2254 }
2255 else
2256 {
2257 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p, post_p,
2258 is_gimple_reg, fb_rvalue);
2259 ret = MIN (ret, tret);
2260 }
2261 }
2262 }
2263
2264 /* Step 2 is to gimplify the base expression. Make sure lvalue is set
2265 so as to match the min_lval predicate. Failure to do so may result
2266 in the creation of large aggregate temporaries. */
2267 tret = gimplify_expr (p, pre_p, post_p, is_gimple_min_lval,
2268 fallback | fb_lvalue);
2269 ret = MIN (ret, tret);
2270
2271 /* And finally, the indices and operands of ARRAY_REF. During this
2272 loop we also remove any useless conversions. */
2273 for (; expr_stack.length () > 0; )
2274 {
2275 tree t = expr_stack.pop ();
2276
2277 if (TREE_CODE (t) == ARRAY_REF || TREE_CODE (t) == ARRAY_RANGE_REF)
2278 {
2279 /* Gimplify the dimension. */
2280 if (!is_gimple_min_invariant (TREE_OPERAND (t, 1)))
2281 {
2282 tret = gimplify_expr (&TREE_OPERAND (t, 1), pre_p, post_p,
2283 is_gimple_val, fb_rvalue);
2284 ret = MIN (ret, tret);
2285 }
2286 }
2287
2288 STRIP_USELESS_TYPE_CONVERSION (TREE_OPERAND (t, 0));
2289
2290 /* The innermost expression P may have originally had
2291 TREE_SIDE_EFFECTS set which would have caused all the outer
2292 expressions in *EXPR_P leading to P to also have had
2293 TREE_SIDE_EFFECTS set. */
2294 recalculate_side_effects (t);
2295 }
2296
2297 /* If the outermost expression is a COMPONENT_REF, canonicalize its type. */
2298 if ((fallback & fb_rvalue) && TREE_CODE (*expr_p) == COMPONENT_REF)
2299 {
2300 canonicalize_component_ref (expr_p);
2301 }
2302
2303 expr_stack.release ();
2304
2305 gcc_assert (*expr_p == expr || ret != GS_ALL_DONE);
2306
2307 return ret;
2308 }
2309
2310 /* Gimplify the self modifying expression pointed to by EXPR_P
2311 (++, --, +=, -=).
2312
2313 PRE_P points to the list where side effects that must happen before
2314 *EXPR_P should be stored.
2315
2316 POST_P points to the list where side effects that must happen after
2317 *EXPR_P should be stored.
2318
2319 WANT_VALUE is nonzero iff we want to use the value of this expression
2320 in another expression.
2321
2322 ARITH_TYPE is the type the computation should be performed in. */
2323
2324 enum gimplify_status
2325 gimplify_self_mod_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
2326 bool want_value, tree arith_type)
2327 {
2328 enum tree_code code;
2329 tree lhs, lvalue, rhs, t1;
2330 gimple_seq post = NULL, *orig_post_p = post_p;
2331 bool postfix;
2332 enum tree_code arith_code;
2333 enum gimplify_status ret;
2334 location_t loc = EXPR_LOCATION (*expr_p);
2335
2336 code = TREE_CODE (*expr_p);
2337
2338 gcc_assert (code == POSTINCREMENT_EXPR || code == POSTDECREMENT_EXPR
2339 || code == PREINCREMENT_EXPR || code == PREDECREMENT_EXPR);
2340
2341 /* Prefix or postfix? */
2342 if (code == POSTINCREMENT_EXPR || code == POSTDECREMENT_EXPR)
2343 /* Faster to treat as prefix if result is not used. */
2344 postfix = want_value;
2345 else
2346 postfix = false;
2347
2348 /* For postfix, make sure the inner expression's post side effects
2349 are executed after side effects from this expression. */
2350 if (postfix)
2351 post_p = &post;
2352
2353 /* Add or subtract? */
2354 if (code == PREINCREMENT_EXPR || code == POSTINCREMENT_EXPR)
2355 arith_code = PLUS_EXPR;
2356 else
2357 arith_code = MINUS_EXPR;
2358
2359 /* Gimplify the LHS into a GIMPLE lvalue. */
2360 lvalue = TREE_OPERAND (*expr_p, 0);
2361 ret = gimplify_expr (&lvalue, pre_p, post_p, is_gimple_lvalue, fb_lvalue);
2362 if (ret == GS_ERROR)
2363 return ret;
2364
2365 /* Extract the operands to the arithmetic operation. */
2366 lhs = lvalue;
2367 rhs = TREE_OPERAND (*expr_p, 1);
2368
2369 /* For postfix operator, we evaluate the LHS to an rvalue and then use
2370 that as the result value and in the postqueue operation. We also
2371 make sure to make lvalue a minimal lval, see
2372 gcc.c-torture/execute/20040313-1.c for an example where this matters. */
2373 if (postfix)
2374 {
2375 if (!is_gimple_min_lval (lvalue))
2376 {
2377 mark_addressable (lvalue);
2378 lvalue = build_fold_addr_expr_loc (input_location, lvalue);
2379 gimplify_expr (&lvalue, pre_p, post_p, is_gimple_val, fb_rvalue);
2380 lvalue = build_fold_indirect_ref_loc (input_location, lvalue);
2381 }
2382 ret = gimplify_expr (&lhs, pre_p, post_p, is_gimple_val, fb_rvalue);
2383 if (ret == GS_ERROR)
2384 return ret;
2385 }
2386
2387 if (postfix)
2388 lhs = get_initialized_tmp_var (lhs, pre_p, NULL);
2389
2390 /* For POINTERs increment, use POINTER_PLUS_EXPR. */
2391 if (POINTER_TYPE_P (TREE_TYPE (lhs)))
2392 {
2393 rhs = convert_to_ptrofftype_loc (loc, rhs);
2394 if (arith_code == MINUS_EXPR)
2395 rhs = fold_build1_loc (loc, NEGATE_EXPR, TREE_TYPE (rhs), rhs);
2396 t1 = fold_build2 (POINTER_PLUS_EXPR, TREE_TYPE (*expr_p), lhs, rhs);
2397 }
2398 else
2399 t1 = fold_convert (TREE_TYPE (*expr_p),
2400 fold_build2 (arith_code, arith_type,
2401 fold_convert (arith_type, lhs),
2402 fold_convert (arith_type, rhs)));
2403
2404 if (postfix)
2405 {
2406 gimplify_assign (lvalue, t1, pre_p);
2407 gimplify_seq_add_seq (orig_post_p, post);
2408 *expr_p = lhs;
2409 return GS_ALL_DONE;
2410 }
2411 else
2412 {
2413 *expr_p = build2 (MODIFY_EXPR, TREE_TYPE (lvalue), lvalue, t1);
2414 return GS_OK;
2415 }
2416 }
2417
2418 /* If *EXPR_P has a variable sized type, wrap it in a WITH_SIZE_EXPR. */
2419
2420 static void
2421 maybe_with_size_expr (tree *expr_p)
2422 {
2423 tree expr = *expr_p;
2424 tree type = TREE_TYPE (expr);
2425 tree size;
2426
2427 /* If we've already wrapped this or the type is error_mark_node, we can't do
2428 anything. */
2429 if (TREE_CODE (expr) == WITH_SIZE_EXPR
2430 || type == error_mark_node)
2431 return;
2432
2433 /* If the size isn't known or is a constant, we have nothing to do. */
2434 size = TYPE_SIZE_UNIT (type);
2435 if (!size || TREE_CODE (size) == INTEGER_CST)
2436 return;
2437
2438 /* Otherwise, make a WITH_SIZE_EXPR. */
2439 size = unshare_expr (size);
2440 size = SUBSTITUTE_PLACEHOLDER_IN_EXPR (size, expr);
2441 *expr_p = build2 (WITH_SIZE_EXPR, type, expr, size);
2442 }
2443
2444 /* Helper for gimplify_call_expr. Gimplify a single argument *ARG_P
2445 Store any side-effects in PRE_P. CALL_LOCATION is the location of
2446 the CALL_EXPR. */
2447
2448 static enum gimplify_status
2449 gimplify_arg (tree *arg_p, gimple_seq *pre_p, location_t call_location)
2450 {
2451 bool (*test) (tree);
2452 fallback_t fb;
2453
2454 /* In general, we allow lvalues for function arguments to avoid
2455 extra overhead of copying large aggregates out of even larger
2456 aggregates into temporaries only to copy the temporaries to
2457 the argument list. Make optimizers happy by pulling out to
2458 temporaries those types that fit in registers. */
2459 if (is_gimple_reg_type (TREE_TYPE (*arg_p)))
2460 test = is_gimple_val, fb = fb_rvalue;
2461 else
2462 {
2463 test = is_gimple_lvalue, fb = fb_either;
2464 /* Also strip a TARGET_EXPR that would force an extra copy. */
2465 if (TREE_CODE (*arg_p) == TARGET_EXPR)
2466 {
2467 tree init = TARGET_EXPR_INITIAL (*arg_p);
2468 if (init
2469 && !VOID_TYPE_P (TREE_TYPE (init)))
2470 *arg_p = init;
2471 }
2472 }
2473
2474 /* If this is a variable sized type, we must remember the size. */
2475 maybe_with_size_expr (arg_p);
2476
2477 /* FIXME diagnostics: This will mess up gcc.dg/Warray-bounds.c. */
2478 /* Make sure arguments have the same location as the function call
2479 itself. */
2480 protected_set_expr_location (*arg_p, call_location);
2481
2482 /* There is a sequence point before a function call. Side effects in
2483 the argument list must occur before the actual call. So, when
2484 gimplifying arguments, force gimplify_expr to use an internal
2485 post queue which is then appended to the end of PRE_P. */
2486 return gimplify_expr (arg_p, pre_p, NULL, test, fb);
2487 }
2488
2489 /* Gimplify the CALL_EXPR node *EXPR_P into the GIMPLE sequence PRE_P.
2490 WANT_VALUE is true if the result of the call is desired. */
2491
2492 static enum gimplify_status
2493 gimplify_call_expr (tree *expr_p, gimple_seq *pre_p, bool want_value)
2494 {
2495 tree fndecl, parms, p, fnptrtype;
2496 enum gimplify_status ret;
2497 int i, nargs;
2498 gimple call;
2499 bool builtin_va_start_p = FALSE;
2500 location_t loc = EXPR_LOCATION (*expr_p);
2501
2502 gcc_assert (TREE_CODE (*expr_p) == CALL_EXPR);
2503
2504 /* For reliable diagnostics during inlining, it is necessary that
2505 every call_expr be annotated with file and line. */
2506 if (! EXPR_HAS_LOCATION (*expr_p))
2507 SET_EXPR_LOCATION (*expr_p, input_location);
2508
2509 /* This may be a call to a builtin function.
2510
2511 Builtin function calls may be transformed into different
2512 (and more efficient) builtin function calls under certain
2513 circumstances. Unfortunately, gimplification can muck things
2514 up enough that the builtin expanders are not aware that certain
2515 transformations are still valid.
2516
2517 So we attempt transformation/gimplification of the call before
2518 we gimplify the CALL_EXPR. At this time we do not manage to
2519 transform all calls in the same manner as the expanders do, but
2520 we do transform most of them. */
2521 fndecl = get_callee_fndecl (*expr_p);
2522 if (fndecl
2523 && DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_NORMAL)
2524 switch (DECL_FUNCTION_CODE (fndecl))
2525 {
2526 case BUILT_IN_VA_START:
2527 {
2528 builtin_va_start_p = TRUE;
2529 if (call_expr_nargs (*expr_p) < 2)
2530 {
2531 error ("too few arguments to function %<va_start%>");
2532 *expr_p = build_empty_stmt (EXPR_LOCATION (*expr_p));
2533 return GS_OK;
2534 }
2535
2536 if (fold_builtin_next_arg (*expr_p, true))
2537 {
2538 *expr_p = build_empty_stmt (EXPR_LOCATION (*expr_p));
2539 return GS_OK;
2540 }
2541 break;
2542 }
2543 case BUILT_IN_LINE:
2544 {
2545 expanded_location loc = expand_location (EXPR_LOCATION (*expr_p));
2546 *expr_p = build_int_cst (TREE_TYPE (*expr_p), loc.line);
2547 return GS_OK;
2548 }
2549 case BUILT_IN_FILE:
2550 {
2551 expanded_location loc = expand_location (EXPR_LOCATION (*expr_p));
2552 *expr_p = build_string_literal (strlen (loc.file) + 1, loc.file);
2553 return GS_OK;
2554 }
2555 case BUILT_IN_FUNCTION:
2556 {
2557 const char *function;
2558 function = IDENTIFIER_POINTER (DECL_NAME (current_function_decl));
2559 *expr_p = build_string_literal (strlen (function) + 1, function);
2560 return GS_OK;
2561 }
2562 default:
2563 ;
2564 }
2565 if (fndecl && DECL_BUILT_IN (fndecl))
2566 {
2567 tree new_tree = fold_call_expr (input_location, *expr_p, !want_value);
2568 if (new_tree && new_tree != *expr_p)
2569 {
2570 /* There was a transformation of this call which computes the
2571 same value, but in a more efficient way. Return and try
2572 again. */
2573 *expr_p = new_tree;
2574 return GS_OK;
2575 }
2576 }
2577
2578 /* Remember the original function pointer type. */
2579 fnptrtype = TREE_TYPE (CALL_EXPR_FN (*expr_p));
2580
2581 /* There is a sequence point before the call, so any side effects in
2582 the calling expression must occur before the actual call. Force
2583 gimplify_expr to use an internal post queue. */
2584 ret = gimplify_expr (&CALL_EXPR_FN (*expr_p), pre_p, NULL,
2585 is_gimple_call_addr, fb_rvalue);
2586
2587 nargs = call_expr_nargs (*expr_p);
2588
2589 /* Get argument types for verification. */
2590 fndecl = get_callee_fndecl (*expr_p);
2591 parms = NULL_TREE;
2592 if (fndecl)
2593 parms = TYPE_ARG_TYPES (TREE_TYPE (fndecl));
2594 else if (POINTER_TYPE_P (TREE_TYPE (CALL_EXPR_FN (*expr_p))))
2595 parms = TYPE_ARG_TYPES (TREE_TYPE (TREE_TYPE (CALL_EXPR_FN (*expr_p))));
2596
2597 if (fndecl && DECL_ARGUMENTS (fndecl))
2598 p = DECL_ARGUMENTS (fndecl);
2599 else if (parms)
2600 p = parms;
2601 else
2602 p = NULL_TREE;
2603 for (i = 0; i < nargs && p; i++, p = TREE_CHAIN (p))
2604 ;
2605
2606 /* If the last argument is __builtin_va_arg_pack () and it is not
2607 passed as a named argument, decrease the number of CALL_EXPR
2608 arguments and set instead the CALL_EXPR_VA_ARG_PACK flag. */
2609 if (!p
2610 && i < nargs
2611 && TREE_CODE (CALL_EXPR_ARG (*expr_p, nargs - 1)) == CALL_EXPR)
2612 {
2613 tree last_arg = CALL_EXPR_ARG (*expr_p, nargs - 1);
2614 tree last_arg_fndecl = get_callee_fndecl (last_arg);
2615
2616 if (last_arg_fndecl
2617 && TREE_CODE (last_arg_fndecl) == FUNCTION_DECL
2618 && DECL_BUILT_IN_CLASS (last_arg_fndecl) == BUILT_IN_NORMAL
2619 && DECL_FUNCTION_CODE (last_arg_fndecl) == BUILT_IN_VA_ARG_PACK)
2620 {
2621 tree call = *expr_p;
2622
2623 --nargs;
2624 *expr_p = build_call_array_loc (loc, TREE_TYPE (call),
2625 CALL_EXPR_FN (call),
2626 nargs, CALL_EXPR_ARGP (call));
2627
2628 /* Copy all CALL_EXPR flags, location and block, except
2629 CALL_EXPR_VA_ARG_PACK flag. */
2630 CALL_EXPR_STATIC_CHAIN (*expr_p) = CALL_EXPR_STATIC_CHAIN (call);
2631 CALL_EXPR_TAILCALL (*expr_p) = CALL_EXPR_TAILCALL (call);
2632 CALL_EXPR_RETURN_SLOT_OPT (*expr_p)
2633 = CALL_EXPR_RETURN_SLOT_OPT (call);
2634 CALL_FROM_THUNK_P (*expr_p) = CALL_FROM_THUNK_P (call);
2635 SET_EXPR_LOCATION (*expr_p, EXPR_LOCATION (call));
2636
2637 /* Set CALL_EXPR_VA_ARG_PACK. */
2638 CALL_EXPR_VA_ARG_PACK (*expr_p) = 1;
2639 }
2640 }
2641
2642 /* Finally, gimplify the function arguments. */
2643 if (nargs > 0)
2644 {
2645 for (i = (PUSH_ARGS_REVERSED ? nargs - 1 : 0);
2646 PUSH_ARGS_REVERSED ? i >= 0 : i < nargs;
2647 PUSH_ARGS_REVERSED ? i-- : i++)
2648 {
2649 enum gimplify_status t;
2650
2651 /* Avoid gimplifying the second argument to va_start, which needs to
2652 be the plain PARM_DECL. */
2653 if ((i != 1) || !builtin_va_start_p)
2654 {
2655 t = gimplify_arg (&CALL_EXPR_ARG (*expr_p, i), pre_p,
2656 EXPR_LOCATION (*expr_p));
2657
2658 if (t == GS_ERROR)
2659 ret = GS_ERROR;
2660 }
2661 }
2662 }
2663
2664 /* Verify the function result. */
2665 if (want_value && fndecl
2666 && VOID_TYPE_P (TREE_TYPE (TREE_TYPE (fnptrtype))))
2667 {
2668 error_at (loc, "using result of function returning %<void%>");
2669 ret = GS_ERROR;
2670 }
2671
2672 /* Try this again in case gimplification exposed something. */
2673 if (ret != GS_ERROR)
2674 {
2675 tree new_tree = fold_call_expr (input_location, *expr_p, !want_value);
2676
2677 if (new_tree && new_tree != *expr_p)
2678 {
2679 /* There was a transformation of this call which computes the
2680 same value, but in a more efficient way. Return and try
2681 again. */
2682 *expr_p = new_tree;
2683 return GS_OK;
2684 }
2685 }
2686 else
2687 {
2688 *expr_p = error_mark_node;
2689 return GS_ERROR;
2690 }
2691
2692 /* If the function is "const" or "pure", then clear TREE_SIDE_EFFECTS on its
2693 decl. This allows us to eliminate redundant or useless
2694 calls to "const" functions. */
2695 if (TREE_CODE (*expr_p) == CALL_EXPR)
2696 {
2697 int flags = call_expr_flags (*expr_p);
2698 if (flags & (ECF_CONST | ECF_PURE)
2699 /* An infinite loop is considered a side effect. */
2700 && !(flags & (ECF_LOOPING_CONST_OR_PURE)))
2701 TREE_SIDE_EFFECTS (*expr_p) = 0;
2702 }
2703
2704 /* If the value is not needed by the caller, emit a new GIMPLE_CALL
2705 and clear *EXPR_P. Otherwise, leave *EXPR_P in its gimplified
2706 form and delegate the creation of a GIMPLE_CALL to
2707 gimplify_modify_expr. This is always possible because when
2708 WANT_VALUE is true, the caller wants the result of this call into
2709 a temporary, which means that we will emit an INIT_EXPR in
2710 internal_get_tmp_var which will then be handled by
2711 gimplify_modify_expr. */
2712 if (!want_value)
2713 {
2714 /* The CALL_EXPR in *EXPR_P is already in GIMPLE form, so all we
2715 have to do is replicate it as a GIMPLE_CALL tuple. */
2716 gimple_stmt_iterator gsi;
2717 call = gimple_build_call_from_tree (*expr_p);
2718 gimple_call_set_fntype (call, TREE_TYPE (fnptrtype));
2719 gimplify_seq_add_stmt (pre_p, call);
2720 gsi = gsi_last (*pre_p);
2721 fold_stmt (&gsi);
2722 *expr_p = NULL_TREE;
2723 }
2724 else
2725 /* Remember the original function type. */
2726 CALL_EXPR_FN (*expr_p) = build1 (NOP_EXPR, fnptrtype,
2727 CALL_EXPR_FN (*expr_p));
2728
2729 return ret;
2730 }
2731
2732 /* Handle shortcut semantics in the predicate operand of a COND_EXPR by
2733 rewriting it into multiple COND_EXPRs, and possibly GOTO_EXPRs.
2734
2735 TRUE_LABEL_P and FALSE_LABEL_P point to the labels to jump to if the
2736 condition is true or false, respectively. If null, we should generate
2737 our own to skip over the evaluation of this specific expression.
2738
2739 LOCUS is the source location of the COND_EXPR.
2740
2741 This function is the tree equivalent of do_jump.
2742
2743 shortcut_cond_r should only be called by shortcut_cond_expr. */
2744
2745 static tree
2746 shortcut_cond_r (tree pred, tree *true_label_p, tree *false_label_p,
2747 location_t locus)
2748 {
2749 tree local_label = NULL_TREE;
2750 tree t, expr = NULL;
2751
2752 /* OK, it's not a simple case; we need to pull apart the COND_EXPR to
2753 retain the shortcut semantics. Just insert the gotos here;
2754 shortcut_cond_expr will append the real blocks later. */
2755 if (TREE_CODE (pred) == TRUTH_ANDIF_EXPR)
2756 {
2757 location_t new_locus;
2758
2759 /* Turn if (a && b) into
2760
2761 if (a); else goto no;
2762 if (b) goto yes; else goto no;
2763 (no:) */
2764
2765 if (false_label_p == NULL)
2766 false_label_p = &local_label;
2767
2768 /* Keep the original source location on the first 'if'. */
2769 t = shortcut_cond_r (TREE_OPERAND (pred, 0), NULL, false_label_p, locus);
2770 append_to_statement_list (t, &expr);
2771
2772 /* Set the source location of the && on the second 'if'. */
2773 new_locus = EXPR_HAS_LOCATION (pred) ? EXPR_LOCATION (pred) : locus;
2774 t = shortcut_cond_r (TREE_OPERAND (pred, 1), true_label_p, false_label_p,
2775 new_locus);
2776 append_to_statement_list (t, &expr);
2777 }
2778 else if (TREE_CODE (pred) == TRUTH_ORIF_EXPR)
2779 {
2780 location_t new_locus;
2781
2782 /* Turn if (a || b) into
2783
2784 if (a) goto yes;
2785 if (b) goto yes; else goto no;
2786 (yes:) */
2787
2788 if (true_label_p == NULL)
2789 true_label_p = &local_label;
2790
2791 /* Keep the original source location on the first 'if'. */
2792 t = shortcut_cond_r (TREE_OPERAND (pred, 0), true_label_p, NULL, locus);
2793 append_to_statement_list (t, &expr);
2794
2795 /* Set the source location of the || on the second 'if'. */
2796 new_locus = EXPR_HAS_LOCATION (pred) ? EXPR_LOCATION (pred) : locus;
2797 t = shortcut_cond_r (TREE_OPERAND (pred, 1), true_label_p, false_label_p,
2798 new_locus);
2799 append_to_statement_list (t, &expr);
2800 }
2801 else if (TREE_CODE (pred) == COND_EXPR
2802 && !VOID_TYPE_P (TREE_TYPE (TREE_OPERAND (pred, 1)))
2803 && !VOID_TYPE_P (TREE_TYPE (TREE_OPERAND (pred, 2))))
2804 {
2805 location_t new_locus;
2806
2807 /* As long as we're messing with gotos, turn if (a ? b : c) into
2808 if (a)
2809 if (b) goto yes; else goto no;
2810 else
2811 if (c) goto yes; else goto no;
2812
2813 Don't do this if one of the arms has void type, which can happen
2814 in C++ when the arm is throw. */
2815
2816 /* Keep the original source location on the first 'if'. Set the source
2817 location of the ? on the second 'if'. */
2818 new_locus = EXPR_HAS_LOCATION (pred) ? EXPR_LOCATION (pred) : locus;
2819 expr = build3 (COND_EXPR, void_type_node, TREE_OPERAND (pred, 0),
2820 shortcut_cond_r (TREE_OPERAND (pred, 1), true_label_p,
2821 false_label_p, locus),
2822 shortcut_cond_r (TREE_OPERAND (pred, 2), true_label_p,
2823 false_label_p, new_locus));
2824 }
2825 else
2826 {
2827 expr = build3 (COND_EXPR, void_type_node, pred,
2828 build_and_jump (true_label_p),
2829 build_and_jump (false_label_p));
2830 SET_EXPR_LOCATION (expr, locus);
2831 }
2832
2833 if (local_label)
2834 {
2835 t = build1 (LABEL_EXPR, void_type_node, local_label);
2836 append_to_statement_list (t, &expr);
2837 }
2838
2839 return expr;
2840 }
2841
2842 /* Given a conditional expression EXPR with short-circuit boolean
2843 predicates using TRUTH_ANDIF_EXPR or TRUTH_ORIF_EXPR, break the
2844 predicate apart into the equivalent sequence of conditionals. */
2845
2846 static tree
2847 shortcut_cond_expr (tree expr)
2848 {
2849 tree pred = TREE_OPERAND (expr, 0);
2850 tree then_ = TREE_OPERAND (expr, 1);
2851 tree else_ = TREE_OPERAND (expr, 2);
2852 tree true_label, false_label, end_label, t;
2853 tree *true_label_p;
2854 tree *false_label_p;
2855 bool emit_end, emit_false, jump_over_else;
2856 bool then_se = then_ && TREE_SIDE_EFFECTS (then_);
2857 bool else_se = else_ && TREE_SIDE_EFFECTS (else_);
2858
2859 /* First do simple transformations. */
2860 if (!else_se)
2861 {
2862 /* If there is no 'else', turn
2863 if (a && b) then c
2864 into
2865 if (a) if (b) then c. */
2866 while (TREE_CODE (pred) == TRUTH_ANDIF_EXPR)
2867 {
2868 /* Keep the original source location on the first 'if'. */
2869 location_t locus = EXPR_LOC_OR_HERE (expr);
2870 TREE_OPERAND (expr, 0) = TREE_OPERAND (pred, 1);
2871 /* Set the source location of the && on the second 'if'. */
2872 if (EXPR_HAS_LOCATION (pred))
2873 SET_EXPR_LOCATION (expr, EXPR_LOCATION (pred));
2874 then_ = shortcut_cond_expr (expr);
2875 then_se = then_ && TREE_SIDE_EFFECTS (then_);
2876 pred = TREE_OPERAND (pred, 0);
2877 expr = build3 (COND_EXPR, void_type_node, pred, then_, NULL_TREE);
2878 SET_EXPR_LOCATION (expr, locus);
2879 }
2880 }
2881
2882 if (!then_se)
2883 {
2884 /* If there is no 'then', turn
2885 if (a || b); else d
2886 into
2887 if (a); else if (b); else d. */
2888 while (TREE_CODE (pred) == TRUTH_ORIF_EXPR)
2889 {
2890 /* Keep the original source location on the first 'if'. */
2891 location_t locus = EXPR_LOC_OR_HERE (expr);
2892 TREE_OPERAND (expr, 0) = TREE_OPERAND (pred, 1);
2893 /* Set the source location of the || on the second 'if'. */
2894 if (EXPR_HAS_LOCATION (pred))
2895 SET_EXPR_LOCATION (expr, EXPR_LOCATION (pred));
2896 else_ = shortcut_cond_expr (expr);
2897 else_se = else_ && TREE_SIDE_EFFECTS (else_);
2898 pred = TREE_OPERAND (pred, 0);
2899 expr = build3 (COND_EXPR, void_type_node, pred, NULL_TREE, else_);
2900 SET_EXPR_LOCATION (expr, locus);
2901 }
2902 }
2903
2904 /* If we're done, great. */
2905 if (TREE_CODE (pred) != TRUTH_ANDIF_EXPR
2906 && TREE_CODE (pred) != TRUTH_ORIF_EXPR)
2907 return expr;
2908
2909 /* Otherwise we need to mess with gotos. Change
2910 if (a) c; else d;
2911 to
2912 if (a); else goto no;
2913 c; goto end;
2914 no: d; end:
2915 and recursively gimplify the condition. */
2916
2917 true_label = false_label = end_label = NULL_TREE;
2918
2919 /* If our arms just jump somewhere, hijack those labels so we don't
2920 generate jumps to jumps. */
2921
2922 if (then_
2923 && TREE_CODE (then_) == GOTO_EXPR
2924 && TREE_CODE (GOTO_DESTINATION (then_)) == LABEL_DECL)
2925 {
2926 true_label = GOTO_DESTINATION (then_);
2927 then_ = NULL;
2928 then_se = false;
2929 }
2930
2931 if (else_
2932 && TREE_CODE (else_) == GOTO_EXPR
2933 && TREE_CODE (GOTO_DESTINATION (else_)) == LABEL_DECL)
2934 {
2935 false_label = GOTO_DESTINATION (else_);
2936 else_ = NULL;
2937 else_se = false;
2938 }
2939
2940 /* If we aren't hijacking a label for the 'then' branch, it falls through. */
2941 if (true_label)
2942 true_label_p = &true_label;
2943 else
2944 true_label_p = NULL;
2945
2946 /* The 'else' branch also needs a label if it contains interesting code. */
2947 if (false_label || else_se)
2948 false_label_p = &false_label;
2949 else
2950 false_label_p = NULL;
2951
2952 /* If there was nothing else in our arms, just forward the label(s). */
2953 if (!then_se && !else_se)
2954 return shortcut_cond_r (pred, true_label_p, false_label_p,
2955 EXPR_LOC_OR_HERE (expr));
2956
2957 /* If our last subexpression already has a terminal label, reuse it. */
2958 if (else_se)
2959 t = expr_last (else_);
2960 else if (then_se)
2961 t = expr_last (then_);
2962 else
2963 t = NULL;
2964 if (t && TREE_CODE (t) == LABEL_EXPR)
2965 end_label = LABEL_EXPR_LABEL (t);
2966
2967 /* If we don't care about jumping to the 'else' branch, jump to the end
2968 if the condition is false. */
2969 if (!false_label_p)
2970 false_label_p = &end_label;
2971
2972 /* We only want to emit these labels if we aren't hijacking them. */
2973 emit_end = (end_label == NULL_TREE);
2974 emit_false = (false_label == NULL_TREE);
2975
2976 /* We only emit the jump over the else clause if we have to--if the
2977 then clause may fall through. Otherwise we can wind up with a
2978 useless jump and a useless label at the end of gimplified code,
2979 which will cause us to think that this conditional as a whole
2980 falls through even if it doesn't. If we then inline a function
2981 which ends with such a condition, that can cause us to issue an
2982 inappropriate warning about control reaching the end of a
2983 non-void function. */
2984 jump_over_else = block_may_fallthru (then_);
2985
2986 pred = shortcut_cond_r (pred, true_label_p, false_label_p,
2987 EXPR_LOC_OR_HERE (expr));
2988
2989 expr = NULL;
2990 append_to_statement_list (pred, &expr);
2991
2992 append_to_statement_list (then_, &expr);
2993 if (else_se)
2994 {
2995 if (jump_over_else)
2996 {
2997 tree last = expr_last (expr);
2998 t = build_and_jump (&end_label);
2999 if (EXPR_HAS_LOCATION (last))
3000 SET_EXPR_LOCATION (t, EXPR_LOCATION (last));
3001 append_to_statement_list (t, &expr);
3002 }
3003 if (emit_false)
3004 {
3005 t = build1 (LABEL_EXPR, void_type_node, false_label);
3006 append_to_statement_list (t, &expr);
3007 }
3008 append_to_statement_list (else_, &expr);
3009 }
3010 if (emit_end && end_label)
3011 {
3012 t = build1 (LABEL_EXPR, void_type_node, end_label);
3013 append_to_statement_list (t, &expr);
3014 }
3015
3016 return expr;
3017 }
3018
3019 /* EXPR is used in a boolean context; make sure it has BOOLEAN_TYPE. */
3020
3021 tree
3022 gimple_boolify (tree expr)
3023 {
3024 tree type = TREE_TYPE (expr);
3025 location_t loc = EXPR_LOCATION (expr);
3026
3027 if (TREE_CODE (expr) == NE_EXPR
3028 && TREE_CODE (TREE_OPERAND (expr, 0)) == CALL_EXPR
3029 && integer_zerop (TREE_OPERAND (expr, 1)))
3030 {
3031 tree call = TREE_OPERAND (expr, 0);
3032 tree fn = get_callee_fndecl (call);
3033
3034 /* For __builtin_expect ((long) (x), y) recurse into x as well
3035 if x is truth_value_p. */
3036 if (fn
3037 && DECL_BUILT_IN_CLASS (fn) == BUILT_IN_NORMAL
3038 && DECL_FUNCTION_CODE (fn) == BUILT_IN_EXPECT
3039 && call_expr_nargs (call) == 2)
3040 {
3041 tree arg = CALL_EXPR_ARG (call, 0);
3042 if (arg)
3043 {
3044 if (TREE_CODE (arg) == NOP_EXPR
3045 && TREE_TYPE (arg) == TREE_TYPE (call))
3046 arg = TREE_OPERAND (arg, 0);
3047 if (truth_value_p (TREE_CODE (arg)))
3048 {
3049 arg = gimple_boolify (arg);
3050 CALL_EXPR_ARG (call, 0)
3051 = fold_convert_loc (loc, TREE_TYPE (call), arg);
3052 }
3053 }
3054 }
3055 }
3056
3057 switch (TREE_CODE (expr))
3058 {
3059 case TRUTH_AND_EXPR:
3060 case TRUTH_OR_EXPR:
3061 case TRUTH_XOR_EXPR:
3062 case TRUTH_ANDIF_EXPR:
3063 case TRUTH_ORIF_EXPR:
3064 /* Also boolify the arguments of truth exprs. */
3065 TREE_OPERAND (expr, 1) = gimple_boolify (TREE_OPERAND (expr, 1));
3066 /* FALLTHRU */
3067
3068 case TRUTH_NOT_EXPR:
3069 TREE_OPERAND (expr, 0) = gimple_boolify (TREE_OPERAND (expr, 0));
3070
3071 /* These expressions always produce boolean results. */
3072 if (TREE_CODE (type) != BOOLEAN_TYPE)
3073 TREE_TYPE (expr) = boolean_type_node;
3074 return expr;
3075
3076 default:
3077 if (COMPARISON_CLASS_P (expr))
3078 {
3079 /* There expressions always prduce boolean results. */
3080 if (TREE_CODE (type) != BOOLEAN_TYPE)
3081 TREE_TYPE (expr) = boolean_type_node;
3082 return expr;
3083 }
3084 /* Other expressions that get here must have boolean values, but
3085 might need to be converted to the appropriate mode. */
3086 if (TREE_CODE (type) == BOOLEAN_TYPE)
3087 return expr;
3088 return fold_convert_loc (loc, boolean_type_node, expr);
3089 }
3090 }
3091
3092 /* Given a conditional expression *EXPR_P without side effects, gimplify
3093 its operands. New statements are inserted to PRE_P. */
3094
3095 static enum gimplify_status
3096 gimplify_pure_cond_expr (tree *expr_p, gimple_seq *pre_p)
3097 {
3098 tree expr = *expr_p, cond;
3099 enum gimplify_status ret, tret;
3100 enum tree_code code;
3101
3102 cond = gimple_boolify (COND_EXPR_COND (expr));
3103
3104 /* We need to handle && and || specially, as their gimplification
3105 creates pure cond_expr, thus leading to an infinite cycle otherwise. */
3106 code = TREE_CODE (cond);
3107 if (code == TRUTH_ANDIF_EXPR)
3108 TREE_SET_CODE (cond, TRUTH_AND_EXPR);
3109 else if (code == TRUTH_ORIF_EXPR)
3110 TREE_SET_CODE (cond, TRUTH_OR_EXPR);
3111 ret = gimplify_expr (&cond, pre_p, NULL, is_gimple_condexpr, fb_rvalue);
3112 COND_EXPR_COND (*expr_p) = cond;
3113
3114 tret = gimplify_expr (&COND_EXPR_THEN (expr), pre_p, NULL,
3115 is_gimple_val, fb_rvalue);
3116 ret = MIN (ret, tret);
3117 tret = gimplify_expr (&COND_EXPR_ELSE (expr), pre_p, NULL,
3118 is_gimple_val, fb_rvalue);
3119
3120 return MIN (ret, tret);
3121 }
3122
3123 /* Return true if evaluating EXPR could trap.
3124 EXPR is GENERIC, while tree_could_trap_p can be called
3125 only on GIMPLE. */
3126
3127 static bool
3128 generic_expr_could_trap_p (tree expr)
3129 {
3130 unsigned i, n;
3131
3132 if (!expr || is_gimple_val (expr))
3133 return false;
3134
3135 if (!EXPR_P (expr) || tree_could_trap_p (expr))
3136 return true;
3137
3138 n = TREE_OPERAND_LENGTH (expr);
3139 for (i = 0; i < n; i++)
3140 if (generic_expr_could_trap_p (TREE_OPERAND (expr, i)))
3141 return true;
3142
3143 return false;
3144 }
3145
3146 /* Convert the conditional expression pointed to by EXPR_P '(p) ? a : b;'
3147 into
3148
3149 if (p) if (p)
3150 t1 = a; a;
3151 else or else
3152 t1 = b; b;
3153 t1;
3154
3155 The second form is used when *EXPR_P is of type void.
3156
3157 PRE_P points to the list where side effects that must happen before
3158 *EXPR_P should be stored. */
3159
3160 static enum gimplify_status
3161 gimplify_cond_expr (tree *expr_p, gimple_seq *pre_p, fallback_t fallback)
3162 {
3163 tree expr = *expr_p;
3164 tree type = TREE_TYPE (expr);
3165 location_t loc = EXPR_LOCATION (expr);
3166 tree tmp, arm1, arm2;
3167 enum gimplify_status ret;
3168 tree label_true, label_false, label_cont;
3169 bool have_then_clause_p, have_else_clause_p;
3170 gimple gimple_cond;
3171 enum tree_code pred_code;
3172 gimple_seq seq = NULL;
3173
3174 /* If this COND_EXPR has a value, copy the values into a temporary within
3175 the arms. */
3176 if (!VOID_TYPE_P (type))
3177 {
3178 tree then_ = TREE_OPERAND (expr, 1), else_ = TREE_OPERAND (expr, 2);
3179 tree result;
3180
3181 /* If either an rvalue is ok or we do not require an lvalue, create the
3182 temporary. But we cannot do that if the type is addressable. */
3183 if (((fallback & fb_rvalue) || !(fallback & fb_lvalue))
3184 && !TREE_ADDRESSABLE (type))
3185 {
3186 if (gimplify_ctxp->allow_rhs_cond_expr
3187 /* If either branch has side effects or could trap, it can't be
3188 evaluated unconditionally. */
3189 && !TREE_SIDE_EFFECTS (then_)
3190 && !generic_expr_could_trap_p (then_)
3191 && !TREE_SIDE_EFFECTS (else_)
3192 && !generic_expr_could_trap_p (else_))
3193 return gimplify_pure_cond_expr (expr_p, pre_p);
3194
3195 tmp = create_tmp_var (type, "iftmp");
3196 result = tmp;
3197 }
3198
3199 /* Otherwise, only create and copy references to the values. */
3200 else
3201 {
3202 type = build_pointer_type (type);
3203
3204 if (!VOID_TYPE_P (TREE_TYPE (then_)))
3205 then_ = build_fold_addr_expr_loc (loc, then_);
3206
3207 if (!VOID_TYPE_P (TREE_TYPE (else_)))
3208 else_ = build_fold_addr_expr_loc (loc, else_);
3209
3210 expr
3211 = build3 (COND_EXPR, type, TREE_OPERAND (expr, 0), then_, else_);
3212
3213 tmp = create_tmp_var (type, "iftmp");
3214 result = build_simple_mem_ref_loc (loc, tmp);
3215 }
3216
3217 /* Build the new then clause, `tmp = then_;'. But don't build the
3218 assignment if the value is void; in C++ it can be if it's a throw. */
3219 if (!VOID_TYPE_P (TREE_TYPE (then_)))
3220 TREE_OPERAND (expr, 1) = build2 (MODIFY_EXPR, type, tmp, then_);
3221
3222 /* Similarly, build the new else clause, `tmp = else_;'. */
3223 if (!VOID_TYPE_P (TREE_TYPE (else_)))
3224 TREE_OPERAND (expr, 2) = build2 (MODIFY_EXPR, type, tmp, else_);
3225
3226 TREE_TYPE (expr) = void_type_node;
3227 recalculate_side_effects (expr);
3228
3229 /* Move the COND_EXPR to the prequeue. */
3230 gimplify_stmt (&expr, pre_p);
3231
3232 *expr_p = result;
3233 return GS_ALL_DONE;
3234 }
3235
3236 /* Remove any COMPOUND_EXPR so the following cases will be caught. */
3237 STRIP_TYPE_NOPS (TREE_OPERAND (expr, 0));
3238 if (TREE_CODE (TREE_OPERAND (expr, 0)) == COMPOUND_EXPR)
3239 gimplify_compound_expr (&TREE_OPERAND (expr, 0), pre_p, true);
3240
3241 /* Make sure the condition has BOOLEAN_TYPE. */
3242 TREE_OPERAND (expr, 0) = gimple_boolify (TREE_OPERAND (expr, 0));
3243
3244 /* Break apart && and || conditions. */
3245 if (TREE_CODE (TREE_OPERAND (expr, 0)) == TRUTH_ANDIF_EXPR
3246 || TREE_CODE (TREE_OPERAND (expr, 0)) == TRUTH_ORIF_EXPR)
3247 {
3248 expr = shortcut_cond_expr (expr);
3249
3250 if (expr != *expr_p)
3251 {
3252 *expr_p = expr;
3253
3254 /* We can't rely on gimplify_expr to re-gimplify the expanded
3255 form properly, as cleanups might cause the target labels to be
3256 wrapped in a TRY_FINALLY_EXPR. To prevent that, we need to
3257 set up a conditional context. */
3258 gimple_push_condition ();
3259 gimplify_stmt (expr_p, &seq);
3260 gimple_pop_condition (pre_p);
3261 gimple_seq_add_seq (pre_p, seq);
3262
3263 return GS_ALL_DONE;
3264 }
3265 }
3266
3267 /* Now do the normal gimplification. */
3268
3269 /* Gimplify condition. */
3270 ret = gimplify_expr (&TREE_OPERAND (expr, 0), pre_p, NULL, is_gimple_condexpr,
3271 fb_rvalue);
3272 if (ret == GS_ERROR)
3273 return GS_ERROR;
3274 gcc_assert (TREE_OPERAND (expr, 0) != NULL_TREE);
3275
3276 gimple_push_condition ();
3277
3278 have_then_clause_p = have_else_clause_p = false;
3279 if (TREE_OPERAND (expr, 1) != NULL
3280 && TREE_CODE (TREE_OPERAND (expr, 1)) == GOTO_EXPR
3281 && TREE_CODE (GOTO_DESTINATION (TREE_OPERAND (expr, 1))) == LABEL_DECL
3282 && (DECL_CONTEXT (GOTO_DESTINATION (TREE_OPERAND (expr, 1)))
3283 == current_function_decl)
3284 /* For -O0 avoid this optimization if the COND_EXPR and GOTO_EXPR
3285 have different locations, otherwise we end up with incorrect
3286 location information on the branches. */
3287 && (optimize
3288 || !EXPR_HAS_LOCATION (expr)
3289 || !EXPR_HAS_LOCATION (TREE_OPERAND (expr, 1))
3290 || EXPR_LOCATION (expr) == EXPR_LOCATION (TREE_OPERAND (expr, 1))))
3291 {
3292 label_true = GOTO_DESTINATION (TREE_OPERAND (expr, 1));
3293 have_then_clause_p = true;
3294 }
3295 else
3296 label_true = create_artificial_label (UNKNOWN_LOCATION);
3297 if (TREE_OPERAND (expr, 2) != NULL
3298 && TREE_CODE (TREE_OPERAND (expr, 2)) == GOTO_EXPR
3299 && TREE_CODE (GOTO_DESTINATION (TREE_OPERAND (expr, 2))) == LABEL_DECL
3300 && (DECL_CONTEXT (GOTO_DESTINATION (TREE_OPERAND (expr, 2)))
3301 == current_function_decl)
3302 /* For -O0 avoid this optimization if the COND_EXPR and GOTO_EXPR
3303 have different locations, otherwise we end up with incorrect
3304 location information on the branches. */
3305 && (optimize
3306 || !EXPR_HAS_LOCATION (expr)
3307 || !EXPR_HAS_LOCATION (TREE_OPERAND (expr, 2))
3308 || EXPR_LOCATION (expr) == EXPR_LOCATION (TREE_OPERAND (expr, 2))))
3309 {
3310 label_false = GOTO_DESTINATION (TREE_OPERAND (expr, 2));
3311 have_else_clause_p = true;
3312 }
3313 else
3314 label_false = create_artificial_label (UNKNOWN_LOCATION);
3315
3316 gimple_cond_get_ops_from_tree (COND_EXPR_COND (expr), &pred_code, &arm1,
3317 &arm2);
3318
3319 gimple_cond = gimple_build_cond (pred_code, arm1, arm2, label_true,
3320 label_false);
3321
3322 gimplify_seq_add_stmt (&seq, gimple_cond);
3323 label_cont = NULL_TREE;
3324 if (!have_then_clause_p)
3325 {
3326 /* For if (...) {} else { code; } put label_true after
3327 the else block. */
3328 if (TREE_OPERAND (expr, 1) == NULL_TREE
3329 && !have_else_clause_p
3330 && TREE_OPERAND (expr, 2) != NULL_TREE)
3331 label_cont = label_true;
3332 else
3333 {
3334 gimplify_seq_add_stmt (&seq, gimple_build_label (label_true));
3335 have_then_clause_p = gimplify_stmt (&TREE_OPERAND (expr, 1), &seq);
3336 /* For if (...) { code; } else {} or
3337 if (...) { code; } else goto label; or
3338 if (...) { code; return; } else { ... }
3339 label_cont isn't needed. */
3340 if (!have_else_clause_p
3341 && TREE_OPERAND (expr, 2) != NULL_TREE
3342 && gimple_seq_may_fallthru (seq))
3343 {
3344 gimple g;
3345 label_cont = create_artificial_label (UNKNOWN_LOCATION);
3346
3347 g = gimple_build_goto (label_cont);
3348
3349 /* GIMPLE_COND's are very low level; they have embedded
3350 gotos. This particular embedded goto should not be marked
3351 with the location of the original COND_EXPR, as it would
3352 correspond to the COND_EXPR's condition, not the ELSE or the
3353 THEN arms. To avoid marking it with the wrong location, flag
3354 it as "no location". */
3355 gimple_set_do_not_emit_location (g);
3356
3357 gimplify_seq_add_stmt (&seq, g);
3358 }
3359 }
3360 }
3361 if (!have_else_clause_p)
3362 {
3363 gimplify_seq_add_stmt (&seq, gimple_build_label (label_false));
3364 have_else_clause_p = gimplify_stmt (&TREE_OPERAND (expr, 2), &seq);
3365 }
3366 if (label_cont)
3367 gimplify_seq_add_stmt (&seq, gimple_build_label (label_cont));
3368
3369 gimple_pop_condition (pre_p);
3370 gimple_seq_add_seq (pre_p, seq);
3371
3372 if (ret == GS_ERROR)
3373 ; /* Do nothing. */
3374 else if (have_then_clause_p || have_else_clause_p)
3375 ret = GS_ALL_DONE;
3376 else
3377 {
3378 /* Both arms are empty; replace the COND_EXPR with its predicate. */
3379 expr = TREE_OPERAND (expr, 0);
3380 gimplify_stmt (&expr, pre_p);
3381 }
3382
3383 *expr_p = NULL;
3384 return ret;
3385 }
3386
3387 /* Prepare the node pointed to by EXPR_P, an is_gimple_addressable expression,
3388 to be marked addressable.
3389
3390 We cannot rely on such an expression being directly markable if a temporary
3391 has been created by the gimplification. In this case, we create another
3392 temporary and initialize it with a copy, which will become a store after we
3393 mark it addressable. This can happen if the front-end passed us something
3394 that it could not mark addressable yet, like a Fortran pass-by-reference
3395 parameter (int) floatvar. */
3396
3397 static void
3398 prepare_gimple_addressable (tree *expr_p, gimple_seq *seq_p)
3399 {
3400 while (handled_component_p (*expr_p))
3401 expr_p = &TREE_OPERAND (*expr_p, 0);
3402 if (is_gimple_reg (*expr_p))
3403 *expr_p = get_initialized_tmp_var (*expr_p, seq_p, NULL);
3404 }
3405
3406 /* A subroutine of gimplify_modify_expr. Replace a MODIFY_EXPR with
3407 a call to __builtin_memcpy. */
3408
3409 static enum gimplify_status
3410 gimplify_modify_expr_to_memcpy (tree *expr_p, tree size, bool want_value,
3411 gimple_seq *seq_p)
3412 {
3413 tree t, to, to_ptr, from, from_ptr;
3414 gimple gs;
3415 location_t loc = EXPR_LOCATION (*expr_p);
3416
3417 to = TREE_OPERAND (*expr_p, 0);
3418 from = TREE_OPERAND (*expr_p, 1);
3419
3420 /* Mark the RHS addressable. Beware that it may not be possible to do so
3421 directly if a temporary has been created by the gimplification. */
3422 prepare_gimple_addressable (&from, seq_p);
3423
3424 mark_addressable (from);
3425 from_ptr = build_fold_addr_expr_loc (loc, from);
3426 gimplify_arg (&from_ptr, seq_p, loc);
3427
3428 mark_addressable (to);
3429 to_ptr = build_fold_addr_expr_loc (loc, to);
3430 gimplify_arg (&to_ptr, seq_p, loc);
3431
3432 t = builtin_decl_implicit (BUILT_IN_MEMCPY);
3433
3434 gs = gimple_build_call (t, 3, to_ptr, from_ptr, size);
3435
3436 if (want_value)
3437 {
3438 /* tmp = memcpy() */
3439 t = create_tmp_var (TREE_TYPE (to_ptr), NULL);
3440 gimple_call_set_lhs (gs, t);
3441 gimplify_seq_add_stmt (seq_p, gs);
3442
3443 *expr_p = build_simple_mem_ref (t);
3444 return GS_ALL_DONE;
3445 }
3446
3447 gimplify_seq_add_stmt (seq_p, gs);
3448 *expr_p = NULL;
3449 return GS_ALL_DONE;
3450 }
3451
3452 /* A subroutine of gimplify_modify_expr. Replace a MODIFY_EXPR with
3453 a call to __builtin_memset. In this case we know that the RHS is
3454 a CONSTRUCTOR with an empty element list. */
3455
3456 static enum gimplify_status
3457 gimplify_modify_expr_to_memset (tree *expr_p, tree size, bool want_value,
3458 gimple_seq *seq_p)
3459 {
3460 tree t, from, to, to_ptr;
3461 gimple gs;
3462 location_t loc = EXPR_LOCATION (*expr_p);
3463
3464 /* Assert our assumptions, to abort instead of producing wrong code
3465 silently if they are not met. Beware that the RHS CONSTRUCTOR might
3466 not be immediately exposed. */
3467 from = TREE_OPERAND (*expr_p, 1);
3468 if (TREE_CODE (from) == WITH_SIZE_EXPR)
3469 from = TREE_OPERAND (from, 0);
3470
3471 gcc_assert (TREE_CODE (from) == CONSTRUCTOR
3472 && vec_safe_is_empty (CONSTRUCTOR_ELTS (from)));
3473
3474 /* Now proceed. */
3475 to = TREE_OPERAND (*expr_p, 0);
3476
3477 to_ptr = build_fold_addr_expr_loc (loc, to);
3478 gimplify_arg (&to_ptr, seq_p, loc);
3479 t = builtin_decl_implicit (BUILT_IN_MEMSET);
3480
3481 gs = gimple_build_call (t, 3, to_ptr, integer_zero_node, size);
3482
3483 if (want_value)
3484 {
3485 /* tmp = memset() */
3486 t = create_tmp_var (TREE_TYPE (to_ptr), NULL);
3487 gimple_call_set_lhs (gs, t);
3488 gimplify_seq_add_stmt (seq_p, gs);
3489
3490 *expr_p = build1 (INDIRECT_REF, TREE_TYPE (to), t);
3491 return GS_ALL_DONE;
3492 }
3493
3494 gimplify_seq_add_stmt (seq_p, gs);
3495 *expr_p = NULL;
3496 return GS_ALL_DONE;
3497 }
3498
3499 /* A subroutine of gimplify_init_ctor_preeval. Called via walk_tree,
3500 determine, cautiously, if a CONSTRUCTOR overlaps the lhs of an
3501 assignment. Return non-null if we detect a potential overlap. */
3502
3503 struct gimplify_init_ctor_preeval_data
3504 {
3505 /* The base decl of the lhs object. May be NULL, in which case we
3506 have to assume the lhs is indirect. */
3507 tree lhs_base_decl;
3508
3509 /* The alias set of the lhs object. */
3510 alias_set_type lhs_alias_set;
3511 };
3512
3513 static tree
3514 gimplify_init_ctor_preeval_1 (tree *tp, int *walk_subtrees, void *xdata)
3515 {
3516 struct gimplify_init_ctor_preeval_data *data
3517 = (struct gimplify_init_ctor_preeval_data *) xdata;
3518 tree t = *tp;
3519
3520 /* If we find the base object, obviously we have overlap. */
3521 if (data->lhs_base_decl == t)
3522 return t;
3523
3524 /* If the constructor component is indirect, determine if we have a
3525 potential overlap with the lhs. The only bits of information we
3526 have to go on at this point are addressability and alias sets. */
3527 if ((INDIRECT_REF_P (t)
3528 || TREE_CODE (t) == MEM_REF)
3529 && (!data->lhs_base_decl || TREE_ADDRESSABLE (data->lhs_base_decl))
3530 && alias_sets_conflict_p (data->lhs_alias_set, get_alias_set (t)))
3531 return t;
3532
3533 /* If the constructor component is a call, determine if it can hide a
3534 potential overlap with the lhs through an INDIRECT_REF like above.
3535 ??? Ugh - this is completely broken. In fact this whole analysis
3536 doesn't look conservative. */
3537 if (TREE_CODE (t) == CALL_EXPR)
3538 {
3539 tree type, fntype = TREE_TYPE (TREE_TYPE (CALL_EXPR_FN (t)));
3540
3541 for (type = TYPE_ARG_TYPES (fntype); type; type = TREE_CHAIN (type))
3542 if (POINTER_TYPE_P (TREE_VALUE (type))
3543 && (!data->lhs_base_decl || TREE_ADDRESSABLE (data->lhs_base_decl))
3544 && alias_sets_conflict_p (data->lhs_alias_set,
3545 get_alias_set
3546 (TREE_TYPE (TREE_VALUE (type)))))
3547 return t;
3548 }
3549
3550 if (IS_TYPE_OR_DECL_P (t))
3551 *walk_subtrees = 0;
3552 return NULL;
3553 }
3554
3555 /* A subroutine of gimplify_init_constructor. Pre-evaluate EXPR,
3556 force values that overlap with the lhs (as described by *DATA)
3557 into temporaries. */
3558
3559 static void
3560 gimplify_init_ctor_preeval (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
3561 struct gimplify_init_ctor_preeval_data *data)
3562 {
3563 enum gimplify_status one;
3564
3565 /* If the value is constant, then there's nothing to pre-evaluate. */
3566 if (TREE_CONSTANT (*expr_p))
3567 {
3568 /* Ensure it does not have side effects, it might contain a reference to
3569 the object we're initializing. */
3570 gcc_assert (!TREE_SIDE_EFFECTS (*expr_p));
3571 return;
3572 }
3573
3574 /* If the type has non-trivial constructors, we can't pre-evaluate. */
3575 if (TREE_ADDRESSABLE (TREE_TYPE (*expr_p)))
3576 return;
3577
3578 /* Recurse for nested constructors. */
3579 if (TREE_CODE (*expr_p) == CONSTRUCTOR)
3580 {
3581 unsigned HOST_WIDE_INT ix;
3582 constructor_elt *ce;
3583 vec<constructor_elt, va_gc> *v = CONSTRUCTOR_ELTS (*expr_p);
3584
3585 FOR_EACH_VEC_SAFE_ELT (v, ix, ce)
3586 gimplify_init_ctor_preeval (&ce->value, pre_p, post_p, data);
3587
3588 return;
3589 }
3590
3591 /* If this is a variable sized type, we must remember the size. */
3592 maybe_with_size_expr (expr_p);
3593
3594 /* Gimplify the constructor element to something appropriate for the rhs
3595 of a MODIFY_EXPR. Given that we know the LHS is an aggregate, we know
3596 the gimplifier will consider this a store to memory. Doing this
3597 gimplification now means that we won't have to deal with complicated
3598 language-specific trees, nor trees like SAVE_EXPR that can induce
3599 exponential search behavior. */
3600 one = gimplify_expr (expr_p, pre_p, post_p, is_gimple_mem_rhs, fb_rvalue);
3601 if (one == GS_ERROR)
3602 {
3603 *expr_p = NULL;
3604 return;
3605 }
3606
3607 /* If we gimplified to a bare decl, we can be sure that it doesn't overlap
3608 with the lhs, since "a = { .x=a }" doesn't make sense. This will
3609 always be true for all scalars, since is_gimple_mem_rhs insists on a
3610 temporary variable for them. */
3611 if (DECL_P (*expr_p))
3612 return;
3613
3614 /* If this is of variable size, we have no choice but to assume it doesn't
3615 overlap since we can't make a temporary for it. */
3616 if (TREE_CODE (TYPE_SIZE (TREE_TYPE (*expr_p))) != INTEGER_CST)
3617 return;
3618
3619 /* Otherwise, we must search for overlap ... */
3620 if (!walk_tree (expr_p, gimplify_init_ctor_preeval_1, data, NULL))
3621 return;
3622
3623 /* ... and if found, force the value into a temporary. */
3624 *expr_p = get_formal_tmp_var (*expr_p, pre_p);
3625 }
3626
3627 /* A subroutine of gimplify_init_ctor_eval. Create a loop for
3628 a RANGE_EXPR in a CONSTRUCTOR for an array.
3629
3630 var = lower;
3631 loop_entry:
3632 object[var] = value;
3633 if (var == upper)
3634 goto loop_exit;
3635 var = var + 1;
3636 goto loop_entry;
3637 loop_exit:
3638
3639 We increment var _after_ the loop exit check because we might otherwise
3640 fail if upper == TYPE_MAX_VALUE (type for upper).
3641
3642 Note that we never have to deal with SAVE_EXPRs here, because this has
3643 already been taken care of for us, in gimplify_init_ctor_preeval(). */
3644
3645 static void gimplify_init_ctor_eval (tree, vec<constructor_elt, va_gc> *,
3646 gimple_seq *, bool);
3647
3648 static void
3649 gimplify_init_ctor_eval_range (tree object, tree lower, tree upper,
3650 tree value, tree array_elt_type,
3651 gimple_seq *pre_p, bool cleared)
3652 {
3653 tree loop_entry_label, loop_exit_label, fall_thru_label;
3654 tree var, var_type, cref, tmp;
3655
3656 loop_entry_label = create_artificial_label (UNKNOWN_LOCATION);
3657 loop_exit_label = create_artificial_label (UNKNOWN_LOCATION);
3658 fall_thru_label = create_artificial_label (UNKNOWN_LOCATION);
3659
3660 /* Create and initialize the index variable. */
3661 var_type = TREE_TYPE (upper);
3662 var = create_tmp_var (var_type, NULL);
3663 gimplify_seq_add_stmt (pre_p, gimple_build_assign (var, lower));
3664
3665 /* Add the loop entry label. */
3666 gimplify_seq_add_stmt (pre_p, gimple_build_label (loop_entry_label));
3667
3668 /* Build the reference. */
3669 cref = build4 (ARRAY_REF, array_elt_type, unshare_expr (object),
3670 var, NULL_TREE, NULL_TREE);
3671
3672 /* If we are a constructor, just call gimplify_init_ctor_eval to do
3673 the store. Otherwise just assign value to the reference. */
3674
3675 if (TREE_CODE (value) == CONSTRUCTOR)
3676 /* NB we might have to call ourself recursively through
3677 gimplify_init_ctor_eval if the value is a constructor. */
3678 gimplify_init_ctor_eval (cref, CONSTRUCTOR_ELTS (value),
3679 pre_p, cleared);
3680 else
3681 gimplify_seq_add_stmt (pre_p, gimple_build_assign (cref, value));
3682
3683 /* We exit the loop when the index var is equal to the upper bound. */
3684 gimplify_seq_add_stmt (pre_p,
3685 gimple_build_cond (EQ_EXPR, var, upper,
3686 loop_exit_label, fall_thru_label));
3687
3688 gimplify_seq_add_stmt (pre_p, gimple_build_label (fall_thru_label));
3689
3690 /* Otherwise, increment the index var... */
3691 tmp = build2 (PLUS_EXPR, var_type, var,
3692 fold_convert (var_type, integer_one_node));
3693 gimplify_seq_add_stmt (pre_p, gimple_build_assign (var, tmp));
3694
3695 /* ...and jump back to the loop entry. */
3696 gimplify_seq_add_stmt (pre_p, gimple_build_goto (loop_entry_label));
3697
3698 /* Add the loop exit label. */
3699 gimplify_seq_add_stmt (pre_p, gimple_build_label (loop_exit_label));
3700 }
3701
3702 /* Return true if FDECL is accessing a field that is zero sized. */
3703
3704 static bool
3705 zero_sized_field_decl (const_tree fdecl)
3706 {
3707 if (TREE_CODE (fdecl) == FIELD_DECL && DECL_SIZE (fdecl)
3708 && integer_zerop (DECL_SIZE (fdecl)))
3709 return true;
3710 return false;
3711 }
3712
3713 /* Return true if TYPE is zero sized. */
3714
3715 static bool
3716 zero_sized_type (const_tree type)
3717 {
3718 if (AGGREGATE_TYPE_P (type) && TYPE_SIZE (type)
3719 && integer_zerop (TYPE_SIZE (type)))
3720 return true;
3721 return false;
3722 }
3723
3724 /* A subroutine of gimplify_init_constructor. Generate individual
3725 MODIFY_EXPRs for a CONSTRUCTOR. OBJECT is the LHS against which the
3726 assignments should happen. ELTS is the CONSTRUCTOR_ELTS of the
3727 CONSTRUCTOR. CLEARED is true if the entire LHS object has been
3728 zeroed first. */
3729
3730 static void
3731 gimplify_init_ctor_eval (tree object, vec<constructor_elt, va_gc> *elts,
3732 gimple_seq *pre_p, bool cleared)
3733 {
3734 tree array_elt_type = NULL;
3735 unsigned HOST_WIDE_INT ix;
3736 tree purpose, value;
3737
3738 if (TREE_CODE (TREE_TYPE (object)) == ARRAY_TYPE)
3739 array_elt_type = TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (object)));
3740
3741 FOR_EACH_CONSTRUCTOR_ELT (elts, ix, purpose, value)
3742 {
3743 tree cref;
3744
3745 /* NULL values are created above for gimplification errors. */
3746 if (value == NULL)
3747 continue;
3748
3749 if (cleared && initializer_zerop (value))
3750 continue;
3751
3752 /* ??? Here's to hoping the front end fills in all of the indices,
3753 so we don't have to figure out what's missing ourselves. */
3754 gcc_assert (purpose);
3755
3756 /* Skip zero-sized fields, unless value has side-effects. This can
3757 happen with calls to functions returning a zero-sized type, which
3758 we shouldn't discard. As a number of downstream passes don't
3759 expect sets of zero-sized fields, we rely on the gimplification of
3760 the MODIFY_EXPR we make below to drop the assignment statement. */
3761 if (! TREE_SIDE_EFFECTS (value) && zero_sized_field_decl (purpose))
3762 continue;
3763
3764 /* If we have a RANGE_EXPR, we have to build a loop to assign the
3765 whole range. */
3766 if (TREE_CODE (purpose) == RANGE_EXPR)
3767 {
3768 tree lower = TREE_OPERAND (purpose, 0);
3769 tree upper = TREE_OPERAND (purpose, 1);
3770
3771 /* If the lower bound is equal to upper, just treat it as if
3772 upper was the index. */
3773 if (simple_cst_equal (lower, upper))
3774 purpose = upper;
3775 else
3776 {
3777 gimplify_init_ctor_eval_range (object, lower, upper, value,
3778 array_elt_type, pre_p, cleared);
3779 continue;
3780 }
3781 }
3782
3783 if (array_elt_type)
3784 {
3785 /* Do not use bitsizetype for ARRAY_REF indices. */
3786 if (TYPE_DOMAIN (TREE_TYPE (object)))
3787 purpose
3788 = fold_convert (TREE_TYPE (TYPE_DOMAIN (TREE_TYPE (object))),
3789 purpose);
3790 cref = build4 (ARRAY_REF, array_elt_type, unshare_expr (object),
3791 purpose, NULL_TREE, NULL_TREE);
3792 }
3793 else
3794 {
3795 gcc_assert (TREE_CODE (purpose) == FIELD_DECL);
3796 cref = build3 (COMPONENT_REF, TREE_TYPE (purpose),
3797 unshare_expr (object), purpose, NULL_TREE);
3798 }
3799
3800 if (TREE_CODE (value) == CONSTRUCTOR
3801 && TREE_CODE (TREE_TYPE (value)) != VECTOR_TYPE)
3802 gimplify_init_ctor_eval (cref, CONSTRUCTOR_ELTS (value),
3803 pre_p, cleared);
3804 else
3805 {
3806 tree init = build2 (INIT_EXPR, TREE_TYPE (cref), cref, value);
3807 gimplify_and_add (init, pre_p);
3808 ggc_free (init);
3809 }
3810 }
3811 }
3812
3813 /* Return the appropriate RHS predicate for this LHS. */
3814
3815 gimple_predicate
3816 rhs_predicate_for (tree lhs)
3817 {
3818 if (is_gimple_reg (lhs))
3819 return is_gimple_reg_rhs_or_call;
3820 else
3821 return is_gimple_mem_rhs_or_call;
3822 }
3823
3824 /* Gimplify a C99 compound literal expression. This just means adding
3825 the DECL_EXPR before the current statement and using its anonymous
3826 decl instead. */
3827
3828 static enum gimplify_status
3829 gimplify_compound_literal_expr (tree *expr_p, gimple_seq *pre_p,
3830 bool (*gimple_test_f) (tree),
3831 fallback_t fallback)
3832 {
3833 tree decl_s = COMPOUND_LITERAL_EXPR_DECL_EXPR (*expr_p);
3834 tree decl = DECL_EXPR_DECL (decl_s);
3835 tree init = DECL_INITIAL (decl);
3836 /* Mark the decl as addressable if the compound literal
3837 expression is addressable now, otherwise it is marked too late
3838 after we gimplify the initialization expression. */
3839 if (TREE_ADDRESSABLE (*expr_p))
3840 TREE_ADDRESSABLE (decl) = 1;
3841 /* Otherwise, if we don't need an lvalue and have a literal directly
3842 substitute it. Check if it matches the gimple predicate, as
3843 otherwise we'd generate a new temporary, and we can as well just
3844 use the decl we already have. */
3845 else if (!TREE_ADDRESSABLE (decl)
3846 && init
3847 && (fallback & fb_lvalue) == 0
3848 && gimple_test_f (init))
3849 {
3850 *expr_p = init;
3851 return GS_OK;
3852 }
3853
3854 /* Preliminarily mark non-addressed complex variables as eligible
3855 for promotion to gimple registers. We'll transform their uses
3856 as we find them. */
3857 if ((TREE_CODE (TREE_TYPE (decl)) == COMPLEX_TYPE
3858 || TREE_CODE (TREE_TYPE (decl)) == VECTOR_TYPE)
3859 && !TREE_THIS_VOLATILE (decl)
3860 && !needs_to_live_in_memory (decl))
3861 DECL_GIMPLE_REG_P (decl) = 1;
3862
3863 /* If the decl is not addressable, then it is being used in some
3864 expression or on the right hand side of a statement, and it can
3865 be put into a readonly data section. */
3866 if (!TREE_ADDRESSABLE (decl) && (fallback & fb_lvalue) == 0)
3867 TREE_READONLY (decl) = 1;
3868
3869 /* This decl isn't mentioned in the enclosing block, so add it to the
3870 list of temps. FIXME it seems a bit of a kludge to say that
3871 anonymous artificial vars aren't pushed, but everything else is. */
3872 if (DECL_NAME (decl) == NULL_TREE && !DECL_SEEN_IN_BIND_EXPR_P (decl))
3873 gimple_add_tmp_var (decl);
3874
3875 gimplify_and_add (decl_s, pre_p);
3876 *expr_p = decl;
3877 return GS_OK;
3878 }
3879
3880 /* Optimize embedded COMPOUND_LITERAL_EXPRs within a CONSTRUCTOR,
3881 return a new CONSTRUCTOR if something changed. */
3882
3883 static tree
3884 optimize_compound_literals_in_ctor (tree orig_ctor)
3885 {
3886 tree ctor = orig_ctor;
3887 vec<constructor_elt, va_gc> *elts = CONSTRUCTOR_ELTS (ctor);
3888 unsigned int idx, num = vec_safe_length (elts);
3889
3890 for (idx = 0; idx < num; idx++)
3891 {
3892 tree value = (*elts)[idx].value;
3893 tree newval = value;
3894 if (TREE_CODE (value) == CONSTRUCTOR)
3895 newval = optimize_compound_literals_in_ctor (value);
3896 else if (TREE_CODE (value) == COMPOUND_LITERAL_EXPR)
3897 {
3898 tree decl_s = COMPOUND_LITERAL_EXPR_DECL_EXPR (value);
3899 tree decl = DECL_EXPR_DECL (decl_s);
3900 tree init = DECL_INITIAL (decl);
3901
3902 if (!TREE_ADDRESSABLE (value)
3903 && !TREE_ADDRESSABLE (decl)
3904 && init
3905 && TREE_CODE (init) == CONSTRUCTOR)
3906 newval = optimize_compound_literals_in_ctor (init);
3907 }
3908 if (newval == value)
3909 continue;
3910
3911 if (ctor == orig_ctor)
3912 {
3913 ctor = copy_node (orig_ctor);
3914 CONSTRUCTOR_ELTS (ctor) = vec_safe_copy (elts);
3915 elts = CONSTRUCTOR_ELTS (ctor);
3916 }
3917 (*elts)[idx].value = newval;
3918 }
3919 return ctor;
3920 }
3921
3922 /* A subroutine of gimplify_modify_expr. Break out elements of a
3923 CONSTRUCTOR used as an initializer into separate MODIFY_EXPRs.
3924
3925 Note that we still need to clear any elements that don't have explicit
3926 initializers, so if not all elements are initialized we keep the
3927 original MODIFY_EXPR, we just remove all of the constructor elements.
3928
3929 If NOTIFY_TEMP_CREATION is true, do not gimplify, just return
3930 GS_ERROR if we would have to create a temporary when gimplifying
3931 this constructor. Otherwise, return GS_OK.
3932
3933 If NOTIFY_TEMP_CREATION is false, just do the gimplification. */
3934
3935 static enum gimplify_status
3936 gimplify_init_constructor (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
3937 bool want_value, bool notify_temp_creation)
3938 {
3939 tree object, ctor, type;
3940 enum gimplify_status ret;
3941 vec<constructor_elt, va_gc> *elts;
3942
3943 gcc_assert (TREE_CODE (TREE_OPERAND (*expr_p, 1)) == CONSTRUCTOR);
3944
3945 if (!notify_temp_creation)
3946 {
3947 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
3948 is_gimple_lvalue, fb_lvalue);
3949 if (ret == GS_ERROR)
3950 return ret;
3951 }
3952
3953 object = TREE_OPERAND (*expr_p, 0);
3954 ctor = TREE_OPERAND (*expr_p, 1) =
3955 optimize_compound_literals_in_ctor (TREE_OPERAND (*expr_p, 1));
3956 type = TREE_TYPE (ctor);
3957 elts = CONSTRUCTOR_ELTS (ctor);
3958 ret = GS_ALL_DONE;
3959
3960 switch (TREE_CODE (type))
3961 {
3962 case RECORD_TYPE:
3963 case UNION_TYPE:
3964 case QUAL_UNION_TYPE:
3965 case ARRAY_TYPE:
3966 {
3967 struct gimplify_init_ctor_preeval_data preeval_data;
3968 HOST_WIDE_INT num_ctor_elements, num_nonzero_elements;
3969 bool cleared, complete_p, valid_const_initializer;
3970
3971 /* Aggregate types must lower constructors to initialization of
3972 individual elements. The exception is that a CONSTRUCTOR node
3973 with no elements indicates zero-initialization of the whole. */
3974 if (vec_safe_is_empty (elts))
3975 {
3976 if (notify_temp_creation)
3977 return GS_OK;
3978 break;
3979 }
3980
3981 /* Fetch information about the constructor to direct later processing.
3982 We might want to make static versions of it in various cases, and
3983 can only do so if it known to be a valid constant initializer. */
3984 valid_const_initializer
3985 = categorize_ctor_elements (ctor, &num_nonzero_elements,
3986 &num_ctor_elements, &complete_p);
3987
3988 /* If a const aggregate variable is being initialized, then it
3989 should never be a lose to promote the variable to be static. */
3990 if (valid_const_initializer
3991 && num_nonzero_elements > 1
3992 && TREE_READONLY (object)
3993 && TREE_CODE (object) == VAR_DECL
3994 && (flag_merge_constants >= 2 || !TREE_ADDRESSABLE (object)))
3995 {
3996 if (notify_temp_creation)
3997 return GS_ERROR;
3998 DECL_INITIAL (object) = ctor;
3999 TREE_STATIC (object) = 1;
4000 if (!DECL_NAME (object))
4001 DECL_NAME (object) = create_tmp_var_name ("C");
4002 walk_tree (&DECL_INITIAL (object), force_labels_r, NULL, NULL);
4003
4004 /* ??? C++ doesn't automatically append a .<number> to the
4005 assembler name, and even when it does, it looks at FE private
4006 data structures to figure out what that number should be,
4007 which are not set for this variable. I suppose this is
4008 important for local statics for inline functions, which aren't
4009 "local" in the object file sense. So in order to get a unique
4010 TU-local symbol, we must invoke the lhd version now. */
4011 lhd_set_decl_assembler_name (object);
4012
4013 *expr_p = NULL_TREE;
4014 break;
4015 }
4016
4017 /* If there are "lots" of initialized elements, even discounting
4018 those that are not address constants (and thus *must* be
4019 computed at runtime), then partition the constructor into
4020 constant and non-constant parts. Block copy the constant
4021 parts in, then generate code for the non-constant parts. */
4022 /* TODO. There's code in cp/typeck.c to do this. */
4023
4024 if (int_size_in_bytes (TREE_TYPE (ctor)) < 0)
4025 /* store_constructor will ignore the clearing of variable-sized
4026 objects. Initializers for such objects must explicitly set
4027 every field that needs to be set. */
4028 cleared = false;
4029 else if (!complete_p)
4030 /* If the constructor isn't complete, clear the whole object
4031 beforehand.
4032
4033 ??? This ought not to be needed. For any element not present
4034 in the initializer, we should simply set them to zero. Except
4035 we'd need to *find* the elements that are not present, and that
4036 requires trickery to avoid quadratic compile-time behavior in
4037 large cases or excessive memory use in small cases. */
4038 cleared = true;
4039 else if (num_ctor_elements - num_nonzero_elements
4040 > CLEAR_RATIO (optimize_function_for_speed_p (cfun))
4041 && num_nonzero_elements < num_ctor_elements / 4)
4042 /* If there are "lots" of zeros, it's more efficient to clear
4043 the memory and then set the nonzero elements. */
4044 cleared = true;
4045 else
4046 cleared = false;
4047
4048 /* If there are "lots" of initialized elements, and all of them
4049 are valid address constants, then the entire initializer can
4050 be dropped to memory, and then memcpy'd out. Don't do this
4051 for sparse arrays, though, as it's more efficient to follow
4052 the standard CONSTRUCTOR behavior of memset followed by
4053 individual element initialization. Also don't do this for small
4054 all-zero initializers (which aren't big enough to merit
4055 clearing), and don't try to make bitwise copies of
4056 TREE_ADDRESSABLE types. */
4057 if (valid_const_initializer
4058 && !(cleared || num_nonzero_elements == 0)
4059 && !TREE_ADDRESSABLE (type))
4060 {
4061 HOST_WIDE_INT size = int_size_in_bytes (type);
4062 unsigned int align;
4063
4064 /* ??? We can still get unbounded array types, at least
4065 from the C++ front end. This seems wrong, but attempt
4066 to work around it for now. */
4067 if (size < 0)
4068 {
4069 size = int_size_in_bytes (TREE_TYPE (object));
4070 if (size >= 0)
4071 TREE_TYPE (ctor) = type = TREE_TYPE (object);
4072 }
4073
4074 /* Find the maximum alignment we can assume for the object. */
4075 /* ??? Make use of DECL_OFFSET_ALIGN. */
4076 if (DECL_P (object))
4077 align = DECL_ALIGN (object);
4078 else
4079 align = TYPE_ALIGN (type);
4080
4081 /* Do a block move either if the size is so small as to make
4082 each individual move a sub-unit move on average, or if it
4083 is so large as to make individual moves inefficient. */
4084 if (size > 0
4085 && num_nonzero_elements > 1
4086 && (size < num_nonzero_elements
4087 || !can_move_by_pieces (size, align)))
4088 {
4089 if (notify_temp_creation)
4090 return GS_ERROR;
4091
4092 walk_tree (&ctor, force_labels_r, NULL, NULL);
4093 ctor = tree_output_constant_def (ctor);
4094 if (!useless_type_conversion_p (type, TREE_TYPE (ctor)))
4095 ctor = build1 (VIEW_CONVERT_EXPR, type, ctor);
4096 TREE_OPERAND (*expr_p, 1) = ctor;
4097
4098 /* This is no longer an assignment of a CONSTRUCTOR, but
4099 we still may have processing to do on the LHS. So
4100 pretend we didn't do anything here to let that happen. */
4101 return GS_UNHANDLED;
4102 }
4103 }
4104
4105 /* If the target is volatile, we have non-zero elements and more than
4106 one field to assign, initialize the target from a temporary. */
4107 if (TREE_THIS_VOLATILE (object)
4108 && !TREE_ADDRESSABLE (type)
4109 && num_nonzero_elements > 0
4110 && vec_safe_length (elts) > 1)
4111 {
4112 tree temp = create_tmp_var (TYPE_MAIN_VARIANT (type), NULL);
4113 TREE_OPERAND (*expr_p, 0) = temp;
4114 *expr_p = build2 (COMPOUND_EXPR, TREE_TYPE (*expr_p),
4115 *expr_p,
4116 build2 (MODIFY_EXPR, void_type_node,
4117 object, temp));
4118 return GS_OK;
4119 }
4120
4121 if (notify_temp_creation)
4122 return GS_OK;
4123
4124 /* If there are nonzero elements and if needed, pre-evaluate to capture
4125 elements overlapping with the lhs into temporaries. We must do this
4126 before clearing to fetch the values before they are zeroed-out. */
4127 if (num_nonzero_elements > 0 && TREE_CODE (*expr_p) != INIT_EXPR)
4128 {
4129 preeval_data.lhs_base_decl = get_base_address (object);
4130 if (!DECL_P (preeval_data.lhs_base_decl))
4131 preeval_data.lhs_base_decl = NULL;
4132 preeval_data.lhs_alias_set = get_alias_set (object);
4133
4134 gimplify_init_ctor_preeval (&TREE_OPERAND (*expr_p, 1),
4135 pre_p, post_p, &preeval_data);
4136 }
4137
4138 if (cleared)
4139 {
4140 /* Zap the CONSTRUCTOR element list, which simplifies this case.
4141 Note that we still have to gimplify, in order to handle the
4142 case of variable sized types. Avoid shared tree structures. */
4143 CONSTRUCTOR_ELTS (ctor) = NULL;
4144 TREE_SIDE_EFFECTS (ctor) = 0;
4145 object = unshare_expr (object);
4146 gimplify_stmt (expr_p, pre_p);
4147 }
4148
4149 /* If we have not block cleared the object, or if there are nonzero
4150 elements in the constructor, add assignments to the individual
4151 scalar fields of the object. */
4152 if (!cleared || num_nonzero_elements > 0)
4153 gimplify_init_ctor_eval (object, elts, pre_p, cleared);
4154
4155 *expr_p = NULL_TREE;
4156 }
4157 break;
4158
4159 case COMPLEX_TYPE:
4160 {
4161 tree r, i;
4162
4163 if (notify_temp_creation)
4164 return GS_OK;
4165
4166 /* Extract the real and imaginary parts out of the ctor. */
4167 gcc_assert (elts->length () == 2);
4168 r = (*elts)[0].value;
4169 i = (*elts)[1].value;
4170 if (r == NULL || i == NULL)
4171 {
4172 tree zero = build_zero_cst (TREE_TYPE (type));
4173 if (r == NULL)
4174 r = zero;
4175 if (i == NULL)
4176 i = zero;
4177 }
4178
4179 /* Complex types have either COMPLEX_CST or COMPLEX_EXPR to
4180 represent creation of a complex value. */
4181 if (TREE_CONSTANT (r) && TREE_CONSTANT (i))
4182 {
4183 ctor = build_complex (type, r, i);
4184 TREE_OPERAND (*expr_p, 1) = ctor;
4185 }
4186 else
4187 {
4188 ctor = build2 (COMPLEX_EXPR, type, r, i);
4189 TREE_OPERAND (*expr_p, 1) = ctor;
4190 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 1),
4191 pre_p,
4192 post_p,
4193 rhs_predicate_for (TREE_OPERAND (*expr_p, 0)),
4194 fb_rvalue);
4195 }
4196 }
4197 break;
4198
4199 case VECTOR_TYPE:
4200 {
4201 unsigned HOST_WIDE_INT ix;
4202 constructor_elt *ce;
4203
4204 if (notify_temp_creation)
4205 return GS_OK;
4206
4207 /* Go ahead and simplify constant constructors to VECTOR_CST. */
4208 if (TREE_CONSTANT (ctor))
4209 {
4210 bool constant_p = true;
4211 tree value;
4212
4213 /* Even when ctor is constant, it might contain non-*_CST
4214 elements, such as addresses or trapping values like
4215 1.0/0.0 - 1.0/0.0. Such expressions don't belong
4216 in VECTOR_CST nodes. */
4217 FOR_EACH_CONSTRUCTOR_VALUE (elts, ix, value)
4218 if (!CONSTANT_CLASS_P (value))
4219 {
4220 constant_p = false;
4221 break;
4222 }
4223
4224 if (constant_p)
4225 {
4226 TREE_OPERAND (*expr_p, 1) = build_vector_from_ctor (type, elts);
4227 break;
4228 }
4229
4230 /* Don't reduce an initializer constant even if we can't
4231 make a VECTOR_CST. It won't do anything for us, and it'll
4232 prevent us from representing it as a single constant. */
4233 if (initializer_constant_valid_p (ctor, type))
4234 break;
4235
4236 TREE_CONSTANT (ctor) = 0;
4237 }
4238
4239 /* Vector types use CONSTRUCTOR all the way through gimple
4240 compilation as a general initializer. */
4241 FOR_EACH_VEC_SAFE_ELT (elts, ix, ce)
4242 {
4243 enum gimplify_status tret;
4244 tret = gimplify_expr (&ce->value, pre_p, post_p, is_gimple_val,
4245 fb_rvalue);
4246 if (tret == GS_ERROR)
4247 ret = GS_ERROR;
4248 }
4249 if (!is_gimple_reg (TREE_OPERAND (*expr_p, 0)))
4250 TREE_OPERAND (*expr_p, 1) = get_formal_tmp_var (ctor, pre_p);
4251 }
4252 break;
4253
4254 default:
4255 /* So how did we get a CONSTRUCTOR for a scalar type? */
4256 gcc_unreachable ();
4257 }
4258
4259 if (ret == GS_ERROR)
4260 return GS_ERROR;
4261 else if (want_value)
4262 {
4263 *expr_p = object;
4264 return GS_OK;
4265 }
4266 else
4267 {
4268 /* If we have gimplified both sides of the initializer but have
4269 not emitted an assignment, do so now. */
4270 if (*expr_p)
4271 {
4272 tree lhs = TREE_OPERAND (*expr_p, 0);
4273 tree rhs = TREE_OPERAND (*expr_p, 1);
4274 gimple init = gimple_build_assign (lhs, rhs);
4275 gimplify_seq_add_stmt (pre_p, init);
4276 *expr_p = NULL;
4277 }
4278
4279 return GS_ALL_DONE;
4280 }
4281 }
4282
4283 /* Given a pointer value OP0, return a simplified version of an
4284 indirection through OP0, or NULL_TREE if no simplification is
4285 possible. Note that the resulting type may be different from
4286 the type pointed to in the sense that it is still compatible
4287 from the langhooks point of view. */
4288
4289 tree
4290 gimple_fold_indirect_ref (tree t)
4291 {
4292 tree ptype = TREE_TYPE (t), type = TREE_TYPE (ptype);
4293 tree sub = t;
4294 tree subtype;
4295
4296 STRIP_NOPS (sub);
4297 subtype = TREE_TYPE (sub);
4298 if (!POINTER_TYPE_P (subtype))
4299 return NULL_TREE;
4300
4301 if (TREE_CODE (sub) == ADDR_EXPR)
4302 {
4303 tree op = TREE_OPERAND (sub, 0);
4304 tree optype = TREE_TYPE (op);
4305 /* *&p => p */
4306 if (useless_type_conversion_p (type, optype))
4307 return op;
4308
4309 /* *(foo *)&fooarray => fooarray[0] */
4310 if (TREE_CODE (optype) == ARRAY_TYPE
4311 && TREE_CODE (TYPE_SIZE (TREE_TYPE (optype))) == INTEGER_CST
4312 && useless_type_conversion_p (type, TREE_TYPE (optype)))
4313 {
4314 tree type_domain = TYPE_DOMAIN (optype);
4315 tree min_val = size_zero_node;
4316 if (type_domain && TYPE_MIN_VALUE (type_domain))
4317 min_val = TYPE_MIN_VALUE (type_domain);
4318 if (TREE_CODE (min_val) == INTEGER_CST)
4319 return build4 (ARRAY_REF, type, op, min_val, NULL_TREE, NULL_TREE);
4320 }
4321 /* *(foo *)&complexfoo => __real__ complexfoo */
4322 else if (TREE_CODE (optype) == COMPLEX_TYPE
4323 && useless_type_conversion_p (type, TREE_TYPE (optype)))
4324 return fold_build1 (REALPART_EXPR, type, op);
4325 /* *(foo *)&vectorfoo => BIT_FIELD_REF<vectorfoo,...> */
4326 else if (TREE_CODE (optype) == VECTOR_TYPE
4327 && useless_type_conversion_p (type, TREE_TYPE (optype)))
4328 {
4329 tree part_width = TYPE_SIZE (type);
4330 tree index = bitsize_int (0);
4331 return fold_build3 (BIT_FIELD_REF, type, op, part_width, index);
4332 }
4333 }
4334
4335 /* *(p + CST) -> ... */
4336 if (TREE_CODE (sub) == POINTER_PLUS_EXPR
4337 && TREE_CODE (TREE_OPERAND (sub, 1)) == INTEGER_CST)
4338 {
4339 tree addr = TREE_OPERAND (sub, 0);
4340 tree off = TREE_OPERAND (sub, 1);
4341 tree addrtype;
4342
4343 STRIP_NOPS (addr);
4344 addrtype = TREE_TYPE (addr);
4345
4346 /* ((foo*)&vectorfoo)[1] -> BIT_FIELD_REF<vectorfoo,...> */
4347 if (TREE_CODE (addr) == ADDR_EXPR
4348 && TREE_CODE (TREE_TYPE (addrtype)) == VECTOR_TYPE
4349 && useless_type_conversion_p (type, TREE_TYPE (TREE_TYPE (addrtype)))
4350 && host_integerp (off, 1))
4351 {
4352 unsigned HOST_WIDE_INT offset = tree_low_cst (off, 1);
4353 tree part_width = TYPE_SIZE (type);
4354 unsigned HOST_WIDE_INT part_widthi
4355 = tree_low_cst (part_width, 0) / BITS_PER_UNIT;
4356 unsigned HOST_WIDE_INT indexi = offset * BITS_PER_UNIT;
4357 tree index = bitsize_int (indexi);
4358 if (offset / part_widthi
4359 <= TYPE_VECTOR_SUBPARTS (TREE_TYPE (addrtype)))
4360 return fold_build3 (BIT_FIELD_REF, type, TREE_OPERAND (addr, 0),
4361 part_width, index);
4362 }
4363
4364 /* ((foo*)&complexfoo)[1] -> __imag__ complexfoo */
4365 if (TREE_CODE (addr) == ADDR_EXPR
4366 && TREE_CODE (TREE_TYPE (addrtype)) == COMPLEX_TYPE
4367 && useless_type_conversion_p (type, TREE_TYPE (TREE_TYPE (addrtype))))
4368 {
4369 tree size = TYPE_SIZE_UNIT (type);
4370 if (tree_int_cst_equal (size, off))
4371 return fold_build1 (IMAGPART_EXPR, type, TREE_OPERAND (addr, 0));
4372 }
4373
4374 /* *(p + CST) -> MEM_REF <p, CST>. */
4375 if (TREE_CODE (addr) != ADDR_EXPR
4376 || DECL_P (TREE_OPERAND (addr, 0)))
4377 return fold_build2 (MEM_REF, type,
4378 addr,
4379 build_int_cst_wide (ptype,
4380 TREE_INT_CST_LOW (off),
4381 TREE_INT_CST_HIGH (off)));
4382 }
4383
4384 /* *(foo *)fooarrptr => (*fooarrptr)[0] */
4385 if (TREE_CODE (TREE_TYPE (subtype)) == ARRAY_TYPE
4386 && TREE_CODE (TYPE_SIZE (TREE_TYPE (TREE_TYPE (subtype)))) == INTEGER_CST
4387 && useless_type_conversion_p (type, TREE_TYPE (TREE_TYPE (subtype))))
4388 {
4389 tree type_domain;
4390 tree min_val = size_zero_node;
4391 tree osub = sub;
4392 sub = gimple_fold_indirect_ref (sub);
4393 if (! sub)
4394 sub = build1 (INDIRECT_REF, TREE_TYPE (subtype), osub);
4395 type_domain = TYPE_DOMAIN (TREE_TYPE (sub));
4396 if (type_domain && TYPE_MIN_VALUE (type_domain))
4397 min_val = TYPE_MIN_VALUE (type_domain);
4398 if (TREE_CODE (min_val) == INTEGER_CST)
4399 return build4 (ARRAY_REF, type, sub, min_val, NULL_TREE, NULL_TREE);
4400 }
4401
4402 return NULL_TREE;
4403 }
4404
4405 /* Given a pointer value OP0, return a simplified version of an
4406 indirection through OP0, or NULL_TREE if no simplification is
4407 possible. This may only be applied to a rhs of an expression.
4408 Note that the resulting type may be different from the type pointed
4409 to in the sense that it is still compatible from the langhooks
4410 point of view. */
4411
4412 static tree
4413 gimple_fold_indirect_ref_rhs (tree t)
4414 {
4415 return gimple_fold_indirect_ref (t);
4416 }
4417
4418 /* Subroutine of gimplify_modify_expr to do simplifications of
4419 MODIFY_EXPRs based on the code of the RHS. We loop for as long as
4420 something changes. */
4421
4422 static enum gimplify_status
4423 gimplify_modify_expr_rhs (tree *expr_p, tree *from_p, tree *to_p,
4424 gimple_seq *pre_p, gimple_seq *post_p,
4425 bool want_value)
4426 {
4427 enum gimplify_status ret = GS_UNHANDLED;
4428 bool changed;
4429
4430 do
4431 {
4432 changed = false;
4433 switch (TREE_CODE (*from_p))
4434 {
4435 case VAR_DECL:
4436 /* If we're assigning from a read-only variable initialized with
4437 a constructor, do the direct assignment from the constructor,
4438 but only if neither source nor target are volatile since this
4439 latter assignment might end up being done on a per-field basis. */
4440 if (DECL_INITIAL (*from_p)
4441 && TREE_READONLY (*from_p)
4442 && !TREE_THIS_VOLATILE (*from_p)
4443 && !TREE_THIS_VOLATILE (*to_p)
4444 && TREE_CODE (DECL_INITIAL (*from_p)) == CONSTRUCTOR)
4445 {
4446 tree old_from = *from_p;
4447 enum gimplify_status subret;
4448
4449 /* Move the constructor into the RHS. */
4450 *from_p = unshare_expr (DECL_INITIAL (*from_p));
4451
4452 /* Let's see if gimplify_init_constructor will need to put
4453 it in memory. */
4454 subret = gimplify_init_constructor (expr_p, NULL, NULL,
4455 false, true);
4456 if (subret == GS_ERROR)
4457 {
4458 /* If so, revert the change. */
4459 *from_p = old_from;
4460 }
4461 else
4462 {
4463 ret = GS_OK;
4464 changed = true;
4465 }
4466 }
4467 break;
4468 case INDIRECT_REF:
4469 {
4470 /* If we have code like
4471
4472 *(const A*)(A*)&x
4473
4474 where the type of "x" is a (possibly cv-qualified variant
4475 of "A"), treat the entire expression as identical to "x".
4476 This kind of code arises in C++ when an object is bound
4477 to a const reference, and if "x" is a TARGET_EXPR we want
4478 to take advantage of the optimization below. */
4479 bool volatile_p = TREE_THIS_VOLATILE (*from_p);
4480 tree t = gimple_fold_indirect_ref_rhs (TREE_OPERAND (*from_p, 0));
4481 if (t)
4482 {
4483 if (TREE_THIS_VOLATILE (t) != volatile_p)
4484 {
4485 if (TREE_CODE_CLASS (TREE_CODE (t)) == tcc_declaration)
4486 t = build_simple_mem_ref_loc (EXPR_LOCATION (*from_p),
4487 build_fold_addr_expr (t));
4488 if (REFERENCE_CLASS_P (t))
4489 TREE_THIS_VOLATILE (t) = volatile_p;
4490 }
4491 *from_p = t;
4492 ret = GS_OK;
4493 changed = true;
4494 }
4495 break;
4496 }
4497
4498 case TARGET_EXPR:
4499 {
4500 /* If we are initializing something from a TARGET_EXPR, strip the
4501 TARGET_EXPR and initialize it directly, if possible. This can't
4502 be done if the initializer is void, since that implies that the
4503 temporary is set in some non-trivial way.
4504
4505 ??? What about code that pulls out the temp and uses it
4506 elsewhere? I think that such code never uses the TARGET_EXPR as
4507 an initializer. If I'm wrong, we'll die because the temp won't
4508 have any RTL. In that case, I guess we'll need to replace
4509 references somehow. */
4510 tree init = TARGET_EXPR_INITIAL (*from_p);
4511
4512 if (init
4513 && !VOID_TYPE_P (TREE_TYPE (init)))
4514 {
4515 *from_p = init;
4516 ret = GS_OK;
4517 changed = true;
4518 }
4519 }
4520 break;
4521
4522 case COMPOUND_EXPR:
4523 /* Remove any COMPOUND_EXPR in the RHS so the following cases will be
4524 caught. */
4525 gimplify_compound_expr (from_p, pre_p, true);
4526 ret = GS_OK;
4527 changed = true;
4528 break;
4529
4530 case CONSTRUCTOR:
4531 /* If we already made some changes, let the front end have a
4532 crack at this before we break it down. */
4533 if (ret != GS_UNHANDLED)
4534 break;
4535 /* If we're initializing from a CONSTRUCTOR, break this into
4536 individual MODIFY_EXPRs. */
4537 return gimplify_init_constructor (expr_p, pre_p, post_p, want_value,
4538 false);
4539
4540 case COND_EXPR:
4541 /* If we're assigning to a non-register type, push the assignment
4542 down into the branches. This is mandatory for ADDRESSABLE types,
4543 since we cannot generate temporaries for such, but it saves a
4544 copy in other cases as well. */
4545 if (!is_gimple_reg_type (TREE_TYPE (*from_p)))
4546 {
4547 /* This code should mirror the code in gimplify_cond_expr. */
4548 enum tree_code code = TREE_CODE (*expr_p);
4549 tree cond = *from_p;
4550 tree result = *to_p;
4551
4552 ret = gimplify_expr (&result, pre_p, post_p,
4553 is_gimple_lvalue, fb_lvalue);
4554 if (ret != GS_ERROR)
4555 ret = GS_OK;
4556
4557 if (TREE_TYPE (TREE_OPERAND (cond, 1)) != void_type_node)
4558 TREE_OPERAND (cond, 1)
4559 = build2 (code, void_type_node, result,
4560 TREE_OPERAND (cond, 1));
4561 if (TREE_TYPE (TREE_OPERAND (cond, 2)) != void_type_node)
4562 TREE_OPERAND (cond, 2)
4563 = build2 (code, void_type_node, unshare_expr (result),
4564 TREE_OPERAND (cond, 2));
4565
4566 TREE_TYPE (cond) = void_type_node;
4567 recalculate_side_effects (cond);
4568
4569 if (want_value)
4570 {
4571 gimplify_and_add (cond, pre_p);
4572 *expr_p = unshare_expr (result);
4573 }
4574 else
4575 *expr_p = cond;
4576 return ret;
4577 }
4578 break;
4579
4580 case CALL_EXPR:
4581 /* For calls that return in memory, give *to_p as the CALL_EXPR's
4582 return slot so that we don't generate a temporary. */
4583 if (!CALL_EXPR_RETURN_SLOT_OPT (*from_p)
4584 && aggregate_value_p (*from_p, *from_p))
4585 {
4586 bool use_target;
4587
4588 if (!(rhs_predicate_for (*to_p))(*from_p))
4589 /* If we need a temporary, *to_p isn't accurate. */
4590 use_target = false;
4591 /* It's OK to use the return slot directly unless it's an NRV. */
4592 else if (TREE_CODE (*to_p) == RESULT_DECL
4593 && DECL_NAME (*to_p) == NULL_TREE
4594 && needs_to_live_in_memory (*to_p))
4595 use_target = true;
4596 else if (is_gimple_reg_type (TREE_TYPE (*to_p))
4597 || (DECL_P (*to_p) && DECL_REGISTER (*to_p)))
4598 /* Don't force regs into memory. */
4599 use_target = false;
4600 else if (TREE_CODE (*expr_p) == INIT_EXPR)
4601 /* It's OK to use the target directly if it's being
4602 initialized. */
4603 use_target = true;
4604 else if (variably_modified_type_p (TREE_TYPE (*to_p), NULL_TREE))
4605 /* Always use the target and thus RSO for variable-sized types.
4606 GIMPLE cannot deal with a variable-sized assignment
4607 embedded in a call statement. */
4608 use_target = true;
4609 else if (TREE_CODE (*to_p) != SSA_NAME
4610 && (!is_gimple_variable (*to_p)
4611 || needs_to_live_in_memory (*to_p)))
4612 /* Don't use the original target if it's already addressable;
4613 if its address escapes, and the called function uses the
4614 NRV optimization, a conforming program could see *to_p
4615 change before the called function returns; see c++/19317.
4616 When optimizing, the return_slot pass marks more functions
4617 as safe after we have escape info. */
4618 use_target = false;
4619 else
4620 use_target = true;
4621
4622 if (use_target)
4623 {
4624 CALL_EXPR_RETURN_SLOT_OPT (*from_p) = 1;
4625 mark_addressable (*to_p);
4626 }
4627 }
4628 break;
4629
4630 case WITH_SIZE_EXPR:
4631 /* Likewise for calls that return an aggregate of non-constant size,
4632 since we would not be able to generate a temporary at all. */
4633 if (TREE_CODE (TREE_OPERAND (*from_p, 0)) == CALL_EXPR)
4634 {
4635 *from_p = TREE_OPERAND (*from_p, 0);
4636 /* We don't change ret in this case because the
4637 WITH_SIZE_EXPR might have been added in
4638 gimplify_modify_expr, so returning GS_OK would lead to an
4639 infinite loop. */
4640 changed = true;
4641 }
4642 break;
4643
4644 /* If we're initializing from a container, push the initialization
4645 inside it. */
4646 case CLEANUP_POINT_EXPR:
4647 case BIND_EXPR:
4648 case STATEMENT_LIST:
4649 {
4650 tree wrap = *from_p;
4651 tree t;
4652
4653 ret = gimplify_expr (to_p, pre_p, post_p, is_gimple_min_lval,
4654 fb_lvalue);
4655 if (ret != GS_ERROR)
4656 ret = GS_OK;
4657
4658 t = voidify_wrapper_expr (wrap, *expr_p);
4659 gcc_assert (t == *expr_p);
4660
4661 if (want_value)
4662 {
4663 gimplify_and_add (wrap, pre_p);
4664 *expr_p = unshare_expr (*to_p);
4665 }
4666 else
4667 *expr_p = wrap;
4668 return GS_OK;
4669 }
4670
4671 case COMPOUND_LITERAL_EXPR:
4672 {
4673 tree complit = TREE_OPERAND (*expr_p, 1);
4674 tree decl_s = COMPOUND_LITERAL_EXPR_DECL_EXPR (complit);
4675 tree decl = DECL_EXPR_DECL (decl_s);
4676 tree init = DECL_INITIAL (decl);
4677
4678 /* struct T x = (struct T) { 0, 1, 2 } can be optimized
4679 into struct T x = { 0, 1, 2 } if the address of the
4680 compound literal has never been taken. */
4681 if (!TREE_ADDRESSABLE (complit)
4682 && !TREE_ADDRESSABLE (decl)
4683 && init)
4684 {
4685 *expr_p = copy_node (*expr_p);
4686 TREE_OPERAND (*expr_p, 1) = init;
4687 return GS_OK;
4688 }
4689 }
4690
4691 default:
4692 break;
4693 }
4694 }
4695 while (changed);
4696
4697 return ret;
4698 }
4699
4700
4701 /* Return true if T looks like a valid GIMPLE statement. */
4702
4703 static bool
4704 is_gimple_stmt (tree t)
4705 {
4706 const enum tree_code code = TREE_CODE (t);
4707
4708 switch (code)
4709 {
4710 case NOP_EXPR:
4711 /* The only valid NOP_EXPR is the empty statement. */
4712 return IS_EMPTY_STMT (t);
4713
4714 case BIND_EXPR:
4715 case COND_EXPR:
4716 /* These are only valid if they're void. */
4717 return TREE_TYPE (t) == NULL || VOID_TYPE_P (TREE_TYPE (t));
4718
4719 case SWITCH_EXPR:
4720 case GOTO_EXPR:
4721 case RETURN_EXPR:
4722 case LABEL_EXPR:
4723 case CASE_LABEL_EXPR:
4724 case TRY_CATCH_EXPR:
4725 case TRY_FINALLY_EXPR:
4726 case EH_FILTER_EXPR:
4727 case CATCH_EXPR:
4728 case ASM_EXPR:
4729 case STATEMENT_LIST:
4730 case OMP_PARALLEL:
4731 case OMP_FOR:
4732 case OMP_SECTIONS:
4733 case OMP_SECTION:
4734 case OMP_SINGLE:
4735 case OMP_MASTER:
4736 case OMP_ORDERED:
4737 case OMP_CRITICAL:
4738 case OMP_TASK:
4739 /* These are always void. */
4740 return true;
4741
4742 case CALL_EXPR:
4743 case MODIFY_EXPR:
4744 case PREDICT_EXPR:
4745 /* These are valid regardless of their type. */
4746 return true;
4747
4748 default:
4749 return false;
4750 }
4751 }
4752
4753
4754 /* Promote partial stores to COMPLEX variables to total stores. *EXPR_P is
4755 a MODIFY_EXPR with a lhs of a REAL/IMAGPART_EXPR of a variable with
4756 DECL_GIMPLE_REG_P set.
4757
4758 IMPORTANT NOTE: This promotion is performed by introducing a load of the
4759 other, unmodified part of the complex object just before the total store.
4760 As a consequence, if the object is still uninitialized, an undefined value
4761 will be loaded into a register, which may result in a spurious exception
4762 if the register is floating-point and the value happens to be a signaling
4763 NaN for example. Then the fully-fledged complex operations lowering pass
4764 followed by a DCE pass are necessary in order to fix things up. */
4765
4766 static enum gimplify_status
4767 gimplify_modify_expr_complex_part (tree *expr_p, gimple_seq *pre_p,
4768 bool want_value)
4769 {
4770 enum tree_code code, ocode;
4771 tree lhs, rhs, new_rhs, other, realpart, imagpart;
4772
4773 lhs = TREE_OPERAND (*expr_p, 0);
4774 rhs = TREE_OPERAND (*expr_p, 1);
4775 code = TREE_CODE (lhs);
4776 lhs = TREE_OPERAND (lhs, 0);
4777
4778 ocode = code == REALPART_EXPR ? IMAGPART_EXPR : REALPART_EXPR;
4779 other = build1 (ocode, TREE_TYPE (rhs), lhs);
4780 TREE_NO_WARNING (other) = 1;
4781 other = get_formal_tmp_var (other, pre_p);
4782
4783 realpart = code == REALPART_EXPR ? rhs : other;
4784 imagpart = code == REALPART_EXPR ? other : rhs;
4785
4786 if (TREE_CONSTANT (realpart) && TREE_CONSTANT (imagpart))
4787 new_rhs = build_complex (TREE_TYPE (lhs), realpart, imagpart);
4788 else
4789 new_rhs = build2 (COMPLEX_EXPR, TREE_TYPE (lhs), realpart, imagpart);
4790
4791 gimplify_seq_add_stmt (pre_p, gimple_build_assign (lhs, new_rhs));
4792 *expr_p = (want_value) ? rhs : NULL_TREE;
4793
4794 return GS_ALL_DONE;
4795 }
4796
4797 /* Gimplify the MODIFY_EXPR node pointed to by EXPR_P.
4798
4799 modify_expr
4800 : varname '=' rhs
4801 | '*' ID '=' rhs
4802
4803 PRE_P points to the list where side effects that must happen before
4804 *EXPR_P should be stored.
4805
4806 POST_P points to the list where side effects that must happen after
4807 *EXPR_P should be stored.
4808
4809 WANT_VALUE is nonzero iff we want to use the value of this expression
4810 in another expression. */
4811
4812 static enum gimplify_status
4813 gimplify_modify_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
4814 bool want_value)
4815 {
4816 tree *from_p = &TREE_OPERAND (*expr_p, 1);
4817 tree *to_p = &TREE_OPERAND (*expr_p, 0);
4818 enum gimplify_status ret = GS_UNHANDLED;
4819 gimple assign;
4820 location_t loc = EXPR_LOCATION (*expr_p);
4821 gimple_stmt_iterator gsi;
4822
4823 gcc_assert (TREE_CODE (*expr_p) == MODIFY_EXPR
4824 || TREE_CODE (*expr_p) == INIT_EXPR);
4825
4826 /* Trying to simplify a clobber using normal logic doesn't work,
4827 so handle it here. */
4828 if (TREE_CLOBBER_P (*from_p))
4829 {
4830 gcc_assert (!want_value && TREE_CODE (*to_p) == VAR_DECL);
4831 gimplify_seq_add_stmt (pre_p, gimple_build_assign (*to_p, *from_p));
4832 *expr_p = NULL;
4833 return GS_ALL_DONE;
4834 }
4835
4836 /* Insert pointer conversions required by the middle-end that are not
4837 required by the frontend. This fixes middle-end type checking for
4838 for example gcc.dg/redecl-6.c. */
4839 if (POINTER_TYPE_P (TREE_TYPE (*to_p)))
4840 {
4841 STRIP_USELESS_TYPE_CONVERSION (*from_p);
4842 if (!useless_type_conversion_p (TREE_TYPE (*to_p), TREE_TYPE (*from_p)))
4843 *from_p = fold_convert_loc (loc, TREE_TYPE (*to_p), *from_p);
4844 }
4845
4846 /* See if any simplifications can be done based on what the RHS is. */
4847 ret = gimplify_modify_expr_rhs (expr_p, from_p, to_p, pre_p, post_p,
4848 want_value);
4849 if (ret != GS_UNHANDLED)
4850 return ret;
4851
4852 /* For zero sized types only gimplify the left hand side and right hand
4853 side as statements and throw away the assignment. Do this after
4854 gimplify_modify_expr_rhs so we handle TARGET_EXPRs of addressable
4855 types properly. */
4856 if (zero_sized_type (TREE_TYPE (*from_p)) && !want_value)
4857 {
4858 gimplify_stmt (from_p, pre_p);
4859 gimplify_stmt (to_p, pre_p);
4860 *expr_p = NULL_TREE;
4861 return GS_ALL_DONE;
4862 }
4863
4864 /* If the value being copied is of variable width, compute the length
4865 of the copy into a WITH_SIZE_EXPR. Note that we need to do this
4866 before gimplifying any of the operands so that we can resolve any
4867 PLACEHOLDER_EXPRs in the size. Also note that the RTL expander uses
4868 the size of the expression to be copied, not of the destination, so
4869 that is what we must do here. */
4870 maybe_with_size_expr (from_p);
4871
4872 ret = gimplify_expr (to_p, pre_p, post_p, is_gimple_lvalue, fb_lvalue);
4873 if (ret == GS_ERROR)
4874 return ret;
4875
4876 /* As a special case, we have to temporarily allow for assignments
4877 with a CALL_EXPR on the RHS. Since in GIMPLE a function call is
4878 a toplevel statement, when gimplifying the GENERIC expression
4879 MODIFY_EXPR <a, CALL_EXPR <foo>>, we cannot create the tuple
4880 GIMPLE_ASSIGN <a, GIMPLE_CALL <foo>>.
4881
4882 Instead, we need to create the tuple GIMPLE_CALL <a, foo>. To
4883 prevent gimplify_expr from trying to create a new temporary for
4884 foo's LHS, we tell it that it should only gimplify until it
4885 reaches the CALL_EXPR. On return from gimplify_expr, the newly
4886 created GIMPLE_CALL <foo> will be the last statement in *PRE_P
4887 and all we need to do here is set 'a' to be its LHS. */
4888 ret = gimplify_expr (from_p, pre_p, post_p, rhs_predicate_for (*to_p),
4889 fb_rvalue);
4890 if (ret == GS_ERROR)
4891 return ret;
4892
4893 /* Now see if the above changed *from_p to something we handle specially. */
4894 ret = gimplify_modify_expr_rhs (expr_p, from_p, to_p, pre_p, post_p,
4895 want_value);
4896 if (ret != GS_UNHANDLED)
4897 return ret;
4898
4899 /* If we've got a variable sized assignment between two lvalues (i.e. does
4900 not involve a call), then we can make things a bit more straightforward
4901 by converting the assignment to memcpy or memset. */
4902 if (TREE_CODE (*from_p) == WITH_SIZE_EXPR)
4903 {
4904 tree from = TREE_OPERAND (*from_p, 0);
4905 tree size = TREE_OPERAND (*from_p, 1);
4906
4907 if (TREE_CODE (from) == CONSTRUCTOR)
4908 return gimplify_modify_expr_to_memset (expr_p, size, want_value, pre_p);
4909
4910 if (is_gimple_addressable (from))
4911 {
4912 *from_p = from;
4913 return gimplify_modify_expr_to_memcpy (expr_p, size, want_value,
4914 pre_p);
4915 }
4916 }
4917
4918 /* Transform partial stores to non-addressable complex variables into
4919 total stores. This allows us to use real instead of virtual operands
4920 for these variables, which improves optimization. */
4921 if ((TREE_CODE (*to_p) == REALPART_EXPR
4922 || TREE_CODE (*to_p) == IMAGPART_EXPR)
4923 && is_gimple_reg (TREE_OPERAND (*to_p, 0)))
4924 return gimplify_modify_expr_complex_part (expr_p, pre_p, want_value);
4925
4926 /* Try to alleviate the effects of the gimplification creating artificial
4927 temporaries (see for example is_gimple_reg_rhs) on the debug info. */
4928 if (!gimplify_ctxp->into_ssa
4929 && TREE_CODE (*from_p) == VAR_DECL
4930 && DECL_IGNORED_P (*from_p)
4931 && DECL_P (*to_p)
4932 && !DECL_IGNORED_P (*to_p))
4933 {
4934 if (!DECL_NAME (*from_p) && DECL_NAME (*to_p))
4935 DECL_NAME (*from_p)
4936 = create_tmp_var_name (IDENTIFIER_POINTER (DECL_NAME (*to_p)));
4937 DECL_DEBUG_EXPR_IS_FROM (*from_p) = 1;
4938 SET_DECL_DEBUG_EXPR (*from_p, *to_p);
4939 }
4940
4941 if (want_value && TREE_THIS_VOLATILE (*to_p))
4942 *from_p = get_initialized_tmp_var (*from_p, pre_p, post_p);
4943
4944 if (TREE_CODE (*from_p) == CALL_EXPR)
4945 {
4946 /* Since the RHS is a CALL_EXPR, we need to create a GIMPLE_CALL
4947 instead of a GIMPLE_ASSIGN. */
4948 tree fnptrtype = TREE_TYPE (CALL_EXPR_FN (*from_p));
4949 CALL_EXPR_FN (*from_p) = TREE_OPERAND (CALL_EXPR_FN (*from_p), 0);
4950 STRIP_USELESS_TYPE_CONVERSION (CALL_EXPR_FN (*from_p));
4951 assign = gimple_build_call_from_tree (*from_p);
4952 gimple_call_set_fntype (assign, TREE_TYPE (fnptrtype));
4953 if (!gimple_call_noreturn_p (assign))
4954 gimple_call_set_lhs (assign, *to_p);
4955 }
4956 else
4957 {
4958 assign = gimple_build_assign (*to_p, *from_p);
4959 gimple_set_location (assign, EXPR_LOCATION (*expr_p));
4960 }
4961
4962 if (gimplify_ctxp->into_ssa && is_gimple_reg (*to_p))
4963 {
4964 /* We should have got an SSA name from the start. */
4965 gcc_assert (TREE_CODE (*to_p) == SSA_NAME);
4966 }
4967
4968 gimplify_seq_add_stmt (pre_p, assign);
4969 gsi = gsi_last (*pre_p);
4970 fold_stmt (&gsi);
4971
4972 if (want_value)
4973 {
4974 *expr_p = TREE_THIS_VOLATILE (*to_p) ? *from_p : unshare_expr (*to_p);
4975 return GS_OK;
4976 }
4977 else
4978 *expr_p = NULL;
4979
4980 return GS_ALL_DONE;
4981 }
4982
4983 /* Gimplify a comparison between two variable-sized objects. Do this
4984 with a call to BUILT_IN_MEMCMP. */
4985
4986 static enum gimplify_status
4987 gimplify_variable_sized_compare (tree *expr_p)
4988 {
4989 location_t loc = EXPR_LOCATION (*expr_p);
4990 tree op0 = TREE_OPERAND (*expr_p, 0);
4991 tree op1 = TREE_OPERAND (*expr_p, 1);
4992 tree t, arg, dest, src, expr;
4993
4994 arg = TYPE_SIZE_UNIT (TREE_TYPE (op0));
4995 arg = unshare_expr (arg);
4996 arg = SUBSTITUTE_PLACEHOLDER_IN_EXPR (arg, op0);
4997 src = build_fold_addr_expr_loc (loc, op1);
4998 dest = build_fold_addr_expr_loc (loc, op0);
4999 t = builtin_decl_implicit (BUILT_IN_MEMCMP);
5000 t = build_call_expr_loc (loc, t, 3, dest, src, arg);
5001
5002 expr
5003 = build2 (TREE_CODE (*expr_p), TREE_TYPE (*expr_p), t, integer_zero_node);
5004 SET_EXPR_LOCATION (expr, loc);
5005 *expr_p = expr;
5006
5007 return GS_OK;
5008 }
5009
5010 /* Gimplify a comparison between two aggregate objects of integral scalar
5011 mode as a comparison between the bitwise equivalent scalar values. */
5012
5013 static enum gimplify_status
5014 gimplify_scalar_mode_aggregate_compare (tree *expr_p)
5015 {
5016 location_t loc = EXPR_LOCATION (*expr_p);
5017 tree op0 = TREE_OPERAND (*expr_p, 0);
5018 tree op1 = TREE_OPERAND (*expr_p, 1);
5019
5020 tree type = TREE_TYPE (op0);
5021 tree scalar_type = lang_hooks.types.type_for_mode (TYPE_MODE (type), 1);
5022
5023 op0 = fold_build1_loc (loc, VIEW_CONVERT_EXPR, scalar_type, op0);
5024 op1 = fold_build1_loc (loc, VIEW_CONVERT_EXPR, scalar_type, op1);
5025
5026 *expr_p
5027 = fold_build2_loc (loc, TREE_CODE (*expr_p), TREE_TYPE (*expr_p), op0, op1);
5028
5029 return GS_OK;
5030 }
5031
5032 /* Gimplify an expression sequence. This function gimplifies each
5033 expression and rewrites the original expression with the last
5034 expression of the sequence in GIMPLE form.
5035
5036 PRE_P points to the list where the side effects for all the
5037 expressions in the sequence will be emitted.
5038
5039 WANT_VALUE is true when the result of the last COMPOUND_EXPR is used. */
5040
5041 static enum gimplify_status
5042 gimplify_compound_expr (tree *expr_p, gimple_seq *pre_p, bool want_value)
5043 {
5044 tree t = *expr_p;
5045
5046 do
5047 {
5048 tree *sub_p = &TREE_OPERAND (t, 0);
5049
5050 if (TREE_CODE (*sub_p) == COMPOUND_EXPR)
5051 gimplify_compound_expr (sub_p, pre_p, false);
5052 else
5053 gimplify_stmt (sub_p, pre_p);
5054
5055 t = TREE_OPERAND (t, 1);
5056 }
5057 while (TREE_CODE (t) == COMPOUND_EXPR);
5058
5059 *expr_p = t;
5060 if (want_value)
5061 return GS_OK;
5062 else
5063 {
5064 gimplify_stmt (expr_p, pre_p);
5065 return GS_ALL_DONE;
5066 }
5067 }
5068
5069 /* Gimplify a SAVE_EXPR node. EXPR_P points to the expression to
5070 gimplify. After gimplification, EXPR_P will point to a new temporary
5071 that holds the original value of the SAVE_EXPR node.
5072
5073 PRE_P points to the list where side effects that must happen before
5074 *EXPR_P should be stored. */
5075
5076 static enum gimplify_status
5077 gimplify_save_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
5078 {
5079 enum gimplify_status ret = GS_ALL_DONE;
5080 tree val;
5081
5082 gcc_assert (TREE_CODE (*expr_p) == SAVE_EXPR);
5083 val = TREE_OPERAND (*expr_p, 0);
5084
5085 /* If the SAVE_EXPR has not been resolved, then evaluate it once. */
5086 if (!SAVE_EXPR_RESOLVED_P (*expr_p))
5087 {
5088 /* The operand may be a void-valued expression such as SAVE_EXPRs
5089 generated by the Java frontend for class initialization. It is
5090 being executed only for its side-effects. */
5091 if (TREE_TYPE (val) == void_type_node)
5092 {
5093 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
5094 is_gimple_stmt, fb_none);
5095 val = NULL;
5096 }
5097 else
5098 val = get_initialized_tmp_var (val, pre_p, post_p);
5099
5100 TREE_OPERAND (*expr_p, 0) = val;
5101 SAVE_EXPR_RESOLVED_P (*expr_p) = 1;
5102 }
5103
5104 *expr_p = val;
5105
5106 return ret;
5107 }
5108
5109 /* Rewrite the ADDR_EXPR node pointed to by EXPR_P
5110
5111 unary_expr
5112 : ...
5113 | '&' varname
5114 ...
5115
5116 PRE_P points to the list where side effects that must happen before
5117 *EXPR_P should be stored.
5118
5119 POST_P points to the list where side effects that must happen after
5120 *EXPR_P should be stored. */
5121
5122 static enum gimplify_status
5123 gimplify_addr_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
5124 {
5125 tree expr = *expr_p;
5126 tree op0 = TREE_OPERAND (expr, 0);
5127 enum gimplify_status ret;
5128 location_t loc = EXPR_LOCATION (*expr_p);
5129
5130 switch (TREE_CODE (op0))
5131 {
5132 case INDIRECT_REF:
5133 do_indirect_ref:
5134 /* Check if we are dealing with an expression of the form '&*ptr'.
5135 While the front end folds away '&*ptr' into 'ptr', these
5136 expressions may be generated internally by the compiler (e.g.,
5137 builtins like __builtin_va_end). */
5138 /* Caution: the silent array decomposition semantics we allow for
5139 ADDR_EXPR means we can't always discard the pair. */
5140 /* Gimplification of the ADDR_EXPR operand may drop
5141 cv-qualification conversions, so make sure we add them if
5142 needed. */
5143 {
5144 tree op00 = TREE_OPERAND (op0, 0);
5145 tree t_expr = TREE_TYPE (expr);
5146 tree t_op00 = TREE_TYPE (op00);
5147
5148 if (!useless_type_conversion_p (t_expr, t_op00))
5149 op00 = fold_convert_loc (loc, TREE_TYPE (expr), op00);
5150 *expr_p = op00;
5151 ret = GS_OK;
5152 }
5153 break;
5154
5155 case VIEW_CONVERT_EXPR:
5156 /* Take the address of our operand and then convert it to the type of
5157 this ADDR_EXPR.
5158
5159 ??? The interactions of VIEW_CONVERT_EXPR and aliasing is not at
5160 all clear. The impact of this transformation is even less clear. */
5161
5162 /* If the operand is a useless conversion, look through it. Doing so
5163 guarantees that the ADDR_EXPR and its operand will remain of the
5164 same type. */
5165 if (tree_ssa_useless_type_conversion (TREE_OPERAND (op0, 0)))
5166 op0 = TREE_OPERAND (op0, 0);
5167
5168 *expr_p = fold_convert_loc (loc, TREE_TYPE (expr),
5169 build_fold_addr_expr_loc (loc,
5170 TREE_OPERAND (op0, 0)));
5171 ret = GS_OK;
5172 break;
5173
5174 default:
5175 /* We use fb_either here because the C frontend sometimes takes
5176 the address of a call that returns a struct; see
5177 gcc.dg/c99-array-lval-1.c. The gimplifier will correctly make
5178 the implied temporary explicit. */
5179
5180 /* Make the operand addressable. */
5181 ret = gimplify_expr (&TREE_OPERAND (expr, 0), pre_p, post_p,
5182 is_gimple_addressable, fb_either);
5183 if (ret == GS_ERROR)
5184 break;
5185
5186 /* Then mark it. Beware that it may not be possible to do so directly
5187 if a temporary has been created by the gimplification. */
5188 prepare_gimple_addressable (&TREE_OPERAND (expr, 0), pre_p);
5189
5190 op0 = TREE_OPERAND (expr, 0);
5191
5192 /* For various reasons, the gimplification of the expression
5193 may have made a new INDIRECT_REF. */
5194 if (TREE_CODE (op0) == INDIRECT_REF)
5195 goto do_indirect_ref;
5196
5197 mark_addressable (TREE_OPERAND (expr, 0));
5198
5199 /* The FEs may end up building ADDR_EXPRs early on a decl with
5200 an incomplete type. Re-build ADDR_EXPRs in canonical form
5201 here. */
5202 if (!types_compatible_p (TREE_TYPE (op0), TREE_TYPE (TREE_TYPE (expr))))
5203 *expr_p = build_fold_addr_expr (op0);
5204
5205 /* Make sure TREE_CONSTANT and TREE_SIDE_EFFECTS are set properly. */
5206 recompute_tree_invariant_for_addr_expr (*expr_p);
5207
5208 /* If we re-built the ADDR_EXPR add a conversion to the original type
5209 if required. */
5210 if (!useless_type_conversion_p (TREE_TYPE (expr), TREE_TYPE (*expr_p)))
5211 *expr_p = fold_convert (TREE_TYPE (expr), *expr_p);
5212
5213 break;
5214 }
5215
5216 return ret;
5217 }
5218
5219 /* Gimplify the operands of an ASM_EXPR. Input operands should be a gimple
5220 value; output operands should be a gimple lvalue. */
5221
5222 static enum gimplify_status
5223 gimplify_asm_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
5224 {
5225 tree expr;
5226 int noutputs;
5227 const char **oconstraints;
5228 int i;
5229 tree link;
5230 const char *constraint;
5231 bool allows_mem, allows_reg, is_inout;
5232 enum gimplify_status ret, tret;
5233 gimple stmt;
5234 vec<tree, va_gc> *inputs;
5235 vec<tree, va_gc> *outputs;
5236 vec<tree, va_gc> *clobbers;
5237 vec<tree, va_gc> *labels;
5238 tree link_next;
5239
5240 expr = *expr_p;
5241 noutputs = list_length (ASM_OUTPUTS (expr));
5242 oconstraints = (const char **) alloca ((noutputs) * sizeof (const char *));
5243
5244 inputs = NULL;
5245 outputs = NULL;
5246 clobbers = NULL;
5247 labels = NULL;
5248
5249 ret = GS_ALL_DONE;
5250 link_next = NULL_TREE;
5251 for (i = 0, link = ASM_OUTPUTS (expr); link; ++i, link = link_next)
5252 {
5253 bool ok;
5254 size_t constraint_len;
5255
5256 link_next = TREE_CHAIN (link);
5257
5258 oconstraints[i]
5259 = constraint
5260 = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (link)));
5261 constraint_len = strlen (constraint);
5262 if (constraint_len == 0)
5263 continue;
5264
5265 ok = parse_output_constraint (&constraint, i, 0, 0,
5266 &allows_mem, &allows_reg, &is_inout);
5267 if (!ok)
5268 {
5269 ret = GS_ERROR;
5270 is_inout = false;
5271 }
5272
5273 if (!allows_reg && allows_mem)
5274 mark_addressable (TREE_VALUE (link));
5275
5276 tret = gimplify_expr (&TREE_VALUE (link), pre_p, post_p,
5277 is_inout ? is_gimple_min_lval : is_gimple_lvalue,
5278 fb_lvalue | fb_mayfail);
5279 if (tret == GS_ERROR)
5280 {
5281 error ("invalid lvalue in asm output %d", i);
5282 ret = tret;
5283 }
5284
5285 vec_safe_push (outputs, link);
5286 TREE_CHAIN (link) = NULL_TREE;
5287
5288 if (is_inout)
5289 {
5290 /* An input/output operand. To give the optimizers more
5291 flexibility, split it into separate input and output
5292 operands. */
5293 tree input;
5294 char buf[10];
5295
5296 /* Turn the in/out constraint into an output constraint. */
5297 char *p = xstrdup (constraint);
5298 p[0] = '=';
5299 TREE_VALUE (TREE_PURPOSE (link)) = build_string (constraint_len, p);
5300
5301 /* And add a matching input constraint. */
5302 if (allows_reg)
5303 {
5304 sprintf (buf, "%d", i);
5305
5306 /* If there are multiple alternatives in the constraint,
5307 handle each of them individually. Those that allow register
5308 will be replaced with operand number, the others will stay
5309 unchanged. */
5310 if (strchr (p, ',') != NULL)
5311 {
5312 size_t len = 0, buflen = strlen (buf);
5313 char *beg, *end, *str, *dst;
5314
5315 for (beg = p + 1;;)
5316 {
5317 end = strchr (beg, ',');
5318 if (end == NULL)
5319 end = strchr (beg, '\0');
5320 if ((size_t) (end - beg) < buflen)
5321 len += buflen + 1;
5322 else
5323 len += end - beg + 1;
5324 if (*end)
5325 beg = end + 1;
5326 else
5327 break;
5328 }
5329
5330 str = (char *) alloca (len);
5331 for (beg = p + 1, dst = str;;)
5332 {
5333 const char *tem;
5334 bool mem_p, reg_p, inout_p;
5335
5336 end = strchr (beg, ',');
5337 if (end)
5338 *end = '\0';
5339 beg[-1] = '=';
5340 tem = beg - 1;
5341 parse_output_constraint (&tem, i, 0, 0,
5342 &mem_p, &reg_p, &inout_p);
5343 if (dst != str)
5344 *dst++ = ',';
5345 if (reg_p)
5346 {
5347 memcpy (dst, buf, buflen);
5348 dst += buflen;
5349 }
5350 else
5351 {
5352 if (end)
5353 len = end - beg;
5354 else
5355 len = strlen (beg);
5356 memcpy (dst, beg, len);
5357 dst += len;
5358 }
5359 if (end)
5360 beg = end + 1;
5361 else
5362 break;
5363 }
5364 *dst = '\0';
5365 input = build_string (dst - str, str);
5366 }
5367 else
5368 input = build_string (strlen (buf), buf);
5369 }
5370 else
5371 input = build_string (constraint_len - 1, constraint + 1);
5372
5373 free (p);
5374
5375 input = build_tree_list (build_tree_list (NULL_TREE, input),
5376 unshare_expr (TREE_VALUE (link)));
5377 ASM_INPUTS (expr) = chainon (ASM_INPUTS (expr), input);
5378 }
5379 }
5380
5381 link_next = NULL_TREE;
5382 for (link = ASM_INPUTS (expr); link; ++i, link = link_next)
5383 {
5384 link_next = TREE_CHAIN (link);
5385 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (link)));
5386 parse_input_constraint (&constraint, 0, 0, noutputs, 0,
5387 oconstraints, &allows_mem, &allows_reg);
5388
5389 /* If we can't make copies, we can only accept memory. */
5390 if (TREE_ADDRESSABLE (TREE_TYPE (TREE_VALUE (link))))
5391 {
5392 if (allows_mem)
5393 allows_reg = 0;
5394 else
5395 {
5396 error ("impossible constraint in %<asm%>");
5397 error ("non-memory input %d must stay in memory", i);
5398 return GS_ERROR;
5399 }
5400 }
5401
5402 /* If the operand is a memory input, it should be an lvalue. */
5403 if (!allows_reg && allows_mem)
5404 {
5405 tree inputv = TREE_VALUE (link);
5406 STRIP_NOPS (inputv);
5407 if (TREE_CODE (inputv) == PREDECREMENT_EXPR
5408 || TREE_CODE (inputv) == PREINCREMENT_EXPR
5409 || TREE_CODE (inputv) == POSTDECREMENT_EXPR
5410 || TREE_CODE (inputv) == POSTINCREMENT_EXPR)
5411 TREE_VALUE (link) = error_mark_node;
5412 tret = gimplify_expr (&TREE_VALUE (link), pre_p, post_p,
5413 is_gimple_lvalue, fb_lvalue | fb_mayfail);
5414 mark_addressable (TREE_VALUE (link));
5415 if (tret == GS_ERROR)
5416 {
5417 if (EXPR_HAS_LOCATION (TREE_VALUE (link)))
5418 input_location = EXPR_LOCATION (TREE_VALUE (link));
5419 error ("memory input %d is not directly addressable", i);
5420 ret = tret;
5421 }
5422 }
5423 else
5424 {
5425 tret = gimplify_expr (&TREE_VALUE (link), pre_p, post_p,
5426 is_gimple_asm_val, fb_rvalue);
5427 if (tret == GS_ERROR)
5428 ret = tret;
5429 }
5430
5431 TREE_CHAIN (link) = NULL_TREE;
5432 vec_safe_push (inputs, link);
5433 }
5434
5435 for (link = ASM_CLOBBERS (expr); link; ++i, link = TREE_CHAIN (link))
5436 vec_safe_push (clobbers, link);
5437
5438 for (link = ASM_LABELS (expr); link; ++i, link = TREE_CHAIN (link))
5439 vec_safe_push (labels, link);
5440
5441 /* Do not add ASMs with errors to the gimple IL stream. */
5442 if (ret != GS_ERROR)
5443 {
5444 stmt = gimple_build_asm_vec (TREE_STRING_POINTER (ASM_STRING (expr)),
5445 inputs, outputs, clobbers, labels);
5446
5447 gimple_asm_set_volatile (stmt, ASM_VOLATILE_P (expr));
5448 gimple_asm_set_input (stmt, ASM_INPUT_P (expr));
5449
5450 gimplify_seq_add_stmt (pre_p, stmt);
5451 }
5452
5453 return ret;
5454 }
5455
5456 /* Gimplify a CLEANUP_POINT_EXPR. Currently this works by adding
5457 GIMPLE_WITH_CLEANUP_EXPRs to the prequeue as we encounter cleanups while
5458 gimplifying the body, and converting them to TRY_FINALLY_EXPRs when we
5459 return to this function.
5460
5461 FIXME should we complexify the prequeue handling instead? Or use flags
5462 for all the cleanups and let the optimizer tighten them up? The current
5463 code seems pretty fragile; it will break on a cleanup within any
5464 non-conditional nesting. But any such nesting would be broken, anyway;
5465 we can't write a TRY_FINALLY_EXPR that starts inside a nesting construct
5466 and continues out of it. We can do that at the RTL level, though, so
5467 having an optimizer to tighten up try/finally regions would be a Good
5468 Thing. */
5469
5470 static enum gimplify_status
5471 gimplify_cleanup_point_expr (tree *expr_p, gimple_seq *pre_p)
5472 {
5473 gimple_stmt_iterator iter;
5474 gimple_seq body_sequence = NULL;
5475
5476 tree temp = voidify_wrapper_expr (*expr_p, NULL);
5477
5478 /* We only care about the number of conditions between the innermost
5479 CLEANUP_POINT_EXPR and the cleanup. So save and reset the count and
5480 any cleanups collected outside the CLEANUP_POINT_EXPR. */
5481 int old_conds = gimplify_ctxp->conditions;
5482 gimple_seq old_cleanups = gimplify_ctxp->conditional_cleanups;
5483 bool old_in_cleanup_point_expr = gimplify_ctxp->in_cleanup_point_expr;
5484 gimplify_ctxp->conditions = 0;
5485 gimplify_ctxp->conditional_cleanups = NULL;
5486 gimplify_ctxp->in_cleanup_point_expr = true;
5487
5488 gimplify_stmt (&TREE_OPERAND (*expr_p, 0), &body_sequence);
5489
5490 gimplify_ctxp->conditions = old_conds;
5491 gimplify_ctxp->conditional_cleanups = old_cleanups;
5492 gimplify_ctxp->in_cleanup_point_expr = old_in_cleanup_point_expr;
5493
5494 for (iter = gsi_start (body_sequence); !gsi_end_p (iter); )
5495 {
5496 gimple wce = gsi_stmt (iter);
5497
5498 if (gimple_code (wce) == GIMPLE_WITH_CLEANUP_EXPR)
5499 {
5500 if (gsi_one_before_end_p (iter))
5501 {
5502 /* Note that gsi_insert_seq_before and gsi_remove do not
5503 scan operands, unlike some other sequence mutators. */
5504 if (!gimple_wce_cleanup_eh_only (wce))
5505 gsi_insert_seq_before_without_update (&iter,
5506 gimple_wce_cleanup (wce),
5507 GSI_SAME_STMT);
5508 gsi_remove (&iter, true);
5509 break;
5510 }
5511 else
5512 {
5513 gimple gtry;
5514 gimple_seq seq;
5515 enum gimple_try_flags kind;
5516
5517 if (gimple_wce_cleanup_eh_only (wce))
5518 kind = GIMPLE_TRY_CATCH;
5519 else
5520 kind = GIMPLE_TRY_FINALLY;
5521 seq = gsi_split_seq_after (iter);
5522
5523 gtry = gimple_build_try (seq, gimple_wce_cleanup (wce), kind);
5524 /* Do not use gsi_replace here, as it may scan operands.
5525 We want to do a simple structural modification only. */
5526 gsi_set_stmt (&iter, gtry);
5527 iter = gsi_start (gtry->gimple_try.eval);
5528 }
5529 }
5530 else
5531 gsi_next (&iter);
5532 }
5533
5534 gimplify_seq_add_seq (pre_p, body_sequence);
5535 if (temp)
5536 {
5537 *expr_p = temp;
5538 return GS_OK;
5539 }
5540 else
5541 {
5542 *expr_p = NULL;
5543 return GS_ALL_DONE;
5544 }
5545 }
5546
5547 /* Insert a cleanup marker for gimplify_cleanup_point_expr. CLEANUP
5548 is the cleanup action required. EH_ONLY is true if the cleanup should
5549 only be executed if an exception is thrown, not on normal exit. */
5550
5551 static void
5552 gimple_push_cleanup (tree var, tree cleanup, bool eh_only, gimple_seq *pre_p)
5553 {
5554 gimple wce;
5555 gimple_seq cleanup_stmts = NULL;
5556
5557 /* Errors can result in improperly nested cleanups. Which results in
5558 confusion when trying to resolve the GIMPLE_WITH_CLEANUP_EXPR. */
5559 if (seen_error ())
5560 return;
5561
5562 if (gimple_conditional_context ())
5563 {
5564 /* If we're in a conditional context, this is more complex. We only
5565 want to run the cleanup if we actually ran the initialization that
5566 necessitates it, but we want to run it after the end of the
5567 conditional context. So we wrap the try/finally around the
5568 condition and use a flag to determine whether or not to actually
5569 run the destructor. Thus
5570
5571 test ? f(A()) : 0
5572
5573 becomes (approximately)
5574
5575 flag = 0;
5576 try {
5577 if (test) { A::A(temp); flag = 1; val = f(temp); }
5578 else { val = 0; }
5579 } finally {
5580 if (flag) A::~A(temp);
5581 }
5582 val
5583 */
5584 tree flag = create_tmp_var (boolean_type_node, "cleanup");
5585 gimple ffalse = gimple_build_assign (flag, boolean_false_node);
5586 gimple ftrue = gimple_build_assign (flag, boolean_true_node);
5587
5588 cleanup = build3 (COND_EXPR, void_type_node, flag, cleanup, NULL);
5589 gimplify_stmt (&cleanup, &cleanup_stmts);
5590 wce = gimple_build_wce (cleanup_stmts);
5591
5592 gimplify_seq_add_stmt (&gimplify_ctxp->conditional_cleanups, ffalse);
5593 gimplify_seq_add_stmt (&gimplify_ctxp->conditional_cleanups, wce);
5594 gimplify_seq_add_stmt (pre_p, ftrue);
5595
5596 /* Because of this manipulation, and the EH edges that jump
5597 threading cannot redirect, the temporary (VAR) will appear
5598 to be used uninitialized. Don't warn. */
5599 TREE_NO_WARNING (var) = 1;
5600 }
5601 else
5602 {
5603 gimplify_stmt (&cleanup, &cleanup_stmts);
5604 wce = gimple_build_wce (cleanup_stmts);
5605 gimple_wce_set_cleanup_eh_only (wce, eh_only);
5606 gimplify_seq_add_stmt (pre_p, wce);
5607 }
5608 }
5609
5610 /* Gimplify a TARGET_EXPR which doesn't appear on the rhs of an INIT_EXPR. */
5611
5612 static enum gimplify_status
5613 gimplify_target_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
5614 {
5615 tree targ = *expr_p;
5616 tree temp = TARGET_EXPR_SLOT (targ);
5617 tree init = TARGET_EXPR_INITIAL (targ);
5618 enum gimplify_status ret;
5619
5620 if (init)
5621 {
5622 tree cleanup = NULL_TREE;
5623
5624 /* TARGET_EXPR temps aren't part of the enclosing block, so add it
5625 to the temps list. Handle also variable length TARGET_EXPRs. */
5626 if (TREE_CODE (DECL_SIZE (temp)) != INTEGER_CST)
5627 {
5628 if (!TYPE_SIZES_GIMPLIFIED (TREE_TYPE (temp)))
5629 gimplify_type_sizes (TREE_TYPE (temp), pre_p);
5630 gimplify_vla_decl (temp, pre_p);
5631 }
5632 else
5633 gimple_add_tmp_var (temp);
5634
5635 /* If TARGET_EXPR_INITIAL is void, then the mere evaluation of the
5636 expression is supposed to initialize the slot. */
5637 if (VOID_TYPE_P (TREE_TYPE (init)))
5638 ret = gimplify_expr (&init, pre_p, post_p, is_gimple_stmt, fb_none);
5639 else
5640 {
5641 tree init_expr = build2 (INIT_EXPR, void_type_node, temp, init);
5642 init = init_expr;
5643 ret = gimplify_expr (&init, pre_p, post_p, is_gimple_stmt, fb_none);
5644 init = NULL;
5645 ggc_free (init_expr);
5646 }
5647 if (ret == GS_ERROR)
5648 {
5649 /* PR c++/28266 Make sure this is expanded only once. */
5650 TARGET_EXPR_INITIAL (targ) = NULL_TREE;
5651 return GS_ERROR;
5652 }
5653 if (init)
5654 gimplify_and_add (init, pre_p);
5655
5656 /* If needed, push the cleanup for the temp. */
5657 if (TARGET_EXPR_CLEANUP (targ))
5658 {
5659 if (CLEANUP_EH_ONLY (targ))
5660 gimple_push_cleanup (temp, TARGET_EXPR_CLEANUP (targ),
5661 CLEANUP_EH_ONLY (targ), pre_p);
5662 else
5663 cleanup = TARGET_EXPR_CLEANUP (targ);
5664 }
5665
5666 /* Add a clobber for the temporary going out of scope, like
5667 gimplify_bind_expr. */
5668 if (gimplify_ctxp->in_cleanup_point_expr
5669 && needs_to_live_in_memory (temp)
5670 && flag_stack_reuse == SR_ALL)
5671 {
5672 tree clobber = build_constructor (TREE_TYPE (temp),
5673 NULL);
5674 TREE_THIS_VOLATILE (clobber) = true;
5675 clobber = build2 (MODIFY_EXPR, TREE_TYPE (temp), temp, clobber);
5676 if (cleanup)
5677 cleanup = build2 (COMPOUND_EXPR, void_type_node, cleanup,
5678 clobber);
5679 else
5680 cleanup = clobber;
5681 }
5682
5683 if (cleanup)
5684 gimple_push_cleanup (temp, cleanup, false, pre_p);
5685
5686 /* Only expand this once. */
5687 TREE_OPERAND (targ, 3) = init;
5688 TARGET_EXPR_INITIAL (targ) = NULL_TREE;
5689 }
5690 else
5691 /* We should have expanded this before. */
5692 gcc_assert (DECL_SEEN_IN_BIND_EXPR_P (temp));
5693
5694 *expr_p = temp;
5695 return GS_OK;
5696 }
5697
5698 /* Gimplification of expression trees. */
5699
5700 /* Gimplify an expression which appears at statement context. The
5701 corresponding GIMPLE statements are added to *SEQ_P. If *SEQ_P is
5702 NULL, a new sequence is allocated.
5703
5704 Return true if we actually added a statement to the queue. */
5705
5706 bool
5707 gimplify_stmt (tree *stmt_p, gimple_seq *seq_p)
5708 {
5709 gimple_seq_node last;
5710
5711 last = gimple_seq_last (*seq_p);
5712 gimplify_expr (stmt_p, seq_p, NULL, is_gimple_stmt, fb_none);
5713 return last != gimple_seq_last (*seq_p);
5714 }
5715
5716 /* Add FIRSTPRIVATE entries for DECL in the OpenMP the surrounding parallels
5717 to CTX. If entries already exist, force them to be some flavor of private.
5718 If there is no enclosing parallel, do nothing. */
5719
5720 void
5721 omp_firstprivatize_variable (struct gimplify_omp_ctx *ctx, tree decl)
5722 {
5723 splay_tree_node n;
5724
5725 if (decl == NULL || !DECL_P (decl))
5726 return;
5727
5728 do
5729 {
5730 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5731 if (n != NULL)
5732 {
5733 if (n->value & GOVD_SHARED)
5734 n->value = GOVD_FIRSTPRIVATE | (n->value & GOVD_SEEN);
5735 else
5736 return;
5737 }
5738 else if (ctx->region_type != ORT_WORKSHARE)
5739 omp_add_variable (ctx, decl, GOVD_FIRSTPRIVATE);
5740
5741 ctx = ctx->outer_context;
5742 }
5743 while (ctx);
5744 }
5745
5746 /* Similarly for each of the type sizes of TYPE. */
5747
5748 static void
5749 omp_firstprivatize_type_sizes (struct gimplify_omp_ctx *ctx, tree type)
5750 {
5751 if (type == NULL || type == error_mark_node)
5752 return;
5753 type = TYPE_MAIN_VARIANT (type);
5754
5755 if (pointer_set_insert (ctx->privatized_types, type))
5756 return;
5757
5758 switch (TREE_CODE (type))
5759 {
5760 case INTEGER_TYPE:
5761 case ENUMERAL_TYPE:
5762 case BOOLEAN_TYPE:
5763 case REAL_TYPE:
5764 case FIXED_POINT_TYPE:
5765 omp_firstprivatize_variable (ctx, TYPE_MIN_VALUE (type));
5766 omp_firstprivatize_variable (ctx, TYPE_MAX_VALUE (type));
5767 break;
5768
5769 case ARRAY_TYPE:
5770 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (type));
5771 omp_firstprivatize_type_sizes (ctx, TYPE_DOMAIN (type));
5772 break;
5773
5774 case RECORD_TYPE:
5775 case UNION_TYPE:
5776 case QUAL_UNION_TYPE:
5777 {
5778 tree field;
5779 for (field = TYPE_FIELDS (type); field; field = DECL_CHAIN (field))
5780 if (TREE_CODE (field) == FIELD_DECL)
5781 {
5782 omp_firstprivatize_variable (ctx, DECL_FIELD_OFFSET (field));
5783 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (field));
5784 }
5785 }
5786 break;
5787
5788 case POINTER_TYPE:
5789 case REFERENCE_TYPE:
5790 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (type));
5791 break;
5792
5793 default:
5794 break;
5795 }
5796
5797 omp_firstprivatize_variable (ctx, TYPE_SIZE (type));
5798 omp_firstprivatize_variable (ctx, TYPE_SIZE_UNIT (type));
5799 lang_hooks.types.omp_firstprivatize_type_sizes (ctx, type);
5800 }
5801
5802 /* Add an entry for DECL in the OpenMP context CTX with FLAGS. */
5803
5804 static void
5805 omp_add_variable (struct gimplify_omp_ctx *ctx, tree decl, unsigned int flags)
5806 {
5807 splay_tree_node n;
5808 unsigned int nflags;
5809 tree t;
5810
5811 if (error_operand_p (decl))
5812 return;
5813
5814 /* Never elide decls whose type has TREE_ADDRESSABLE set. This means
5815 there are constructors involved somewhere. */
5816 if (TREE_ADDRESSABLE (TREE_TYPE (decl))
5817 || TYPE_NEEDS_CONSTRUCTING (TREE_TYPE (decl)))
5818 flags |= GOVD_SEEN;
5819
5820 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5821 if (n != NULL)
5822 {
5823 /* We shouldn't be re-adding the decl with the same data
5824 sharing class. */
5825 gcc_assert ((n->value & GOVD_DATA_SHARE_CLASS & flags) == 0);
5826 /* The only combination of data sharing classes we should see is
5827 FIRSTPRIVATE and LASTPRIVATE. */
5828 nflags = n->value | flags;
5829 gcc_assert ((nflags & GOVD_DATA_SHARE_CLASS)
5830 == (GOVD_FIRSTPRIVATE | GOVD_LASTPRIVATE));
5831 n->value = nflags;
5832 return;
5833 }
5834
5835 /* When adding a variable-sized variable, we have to handle all sorts
5836 of additional bits of data: the pointer replacement variable, and
5837 the parameters of the type. */
5838 if (DECL_SIZE (decl) && TREE_CODE (DECL_SIZE (decl)) != INTEGER_CST)
5839 {
5840 /* Add the pointer replacement variable as PRIVATE if the variable
5841 replacement is private, else FIRSTPRIVATE since we'll need the
5842 address of the original variable either for SHARED, or for the
5843 copy into or out of the context. */
5844 if (!(flags & GOVD_LOCAL))
5845 {
5846 nflags = flags & GOVD_PRIVATE ? GOVD_PRIVATE : GOVD_FIRSTPRIVATE;
5847 nflags |= flags & GOVD_SEEN;
5848 t = DECL_VALUE_EXPR (decl);
5849 gcc_assert (TREE_CODE (t) == INDIRECT_REF);
5850 t = TREE_OPERAND (t, 0);
5851 gcc_assert (DECL_P (t));
5852 omp_add_variable (ctx, t, nflags);
5853 }
5854
5855 /* Add all of the variable and type parameters (which should have
5856 been gimplified to a formal temporary) as FIRSTPRIVATE. */
5857 omp_firstprivatize_variable (ctx, DECL_SIZE_UNIT (decl));
5858 omp_firstprivatize_variable (ctx, DECL_SIZE (decl));
5859 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (decl));
5860
5861 /* The variable-sized variable itself is never SHARED, only some form
5862 of PRIVATE. The sharing would take place via the pointer variable
5863 which we remapped above. */
5864 if (flags & GOVD_SHARED)
5865 flags = GOVD_PRIVATE | GOVD_DEBUG_PRIVATE
5866 | (flags & (GOVD_SEEN | GOVD_EXPLICIT));
5867
5868 /* We're going to make use of the TYPE_SIZE_UNIT at least in the
5869 alloca statement we generate for the variable, so make sure it
5870 is available. This isn't automatically needed for the SHARED
5871 case, since we won't be allocating local storage then.
5872 For local variables TYPE_SIZE_UNIT might not be gimplified yet,
5873 in this case omp_notice_variable will be called later
5874 on when it is gimplified. */
5875 else if (! (flags & GOVD_LOCAL)
5876 && DECL_P (TYPE_SIZE_UNIT (TREE_TYPE (decl))))
5877 omp_notice_variable (ctx, TYPE_SIZE_UNIT (TREE_TYPE (decl)), true);
5878 }
5879 else if (lang_hooks.decls.omp_privatize_by_reference (decl))
5880 {
5881 gcc_assert ((flags & GOVD_LOCAL) == 0);
5882 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (decl));
5883
5884 /* Similar to the direct variable sized case above, we'll need the
5885 size of references being privatized. */
5886 if ((flags & GOVD_SHARED) == 0)
5887 {
5888 t = TYPE_SIZE_UNIT (TREE_TYPE (TREE_TYPE (decl)));
5889 if (TREE_CODE (t) != INTEGER_CST)
5890 omp_notice_variable (ctx, t, true);
5891 }
5892 }
5893
5894 splay_tree_insert (ctx->variables, (splay_tree_key)decl, flags);
5895 }
5896
5897 /* Notice a threadprivate variable DECL used in OpenMP context CTX.
5898 This just prints out diagnostics about threadprivate variable uses
5899 in untied tasks. If DECL2 is non-NULL, prevent this warning
5900 on that variable. */
5901
5902 static bool
5903 omp_notice_threadprivate_variable (struct gimplify_omp_ctx *ctx, tree decl,
5904 tree decl2)
5905 {
5906 splay_tree_node n;
5907
5908 if (ctx->region_type != ORT_UNTIED_TASK)
5909 return false;
5910 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5911 if (n == NULL)
5912 {
5913 error ("threadprivate variable %qE used in untied task",
5914 DECL_NAME (decl));
5915 error_at (ctx->location, "enclosing task");
5916 splay_tree_insert (ctx->variables, (splay_tree_key)decl, 0);
5917 }
5918 if (decl2)
5919 splay_tree_insert (ctx->variables, (splay_tree_key)decl2, 0);
5920 return false;
5921 }
5922
5923 /* Record the fact that DECL was used within the OpenMP context CTX.
5924 IN_CODE is true when real code uses DECL, and false when we should
5925 merely emit default(none) errors. Return true if DECL is going to
5926 be remapped and thus DECL shouldn't be gimplified into its
5927 DECL_VALUE_EXPR (if any). */
5928
5929 static bool
5930 omp_notice_variable (struct gimplify_omp_ctx *ctx, tree decl, bool in_code)
5931 {
5932 splay_tree_node n;
5933 unsigned flags = in_code ? GOVD_SEEN : 0;
5934 bool ret = false, shared;
5935
5936 if (error_operand_p (decl))
5937 return false;
5938
5939 /* Threadprivate variables are predetermined. */
5940 if (is_global_var (decl))
5941 {
5942 if (DECL_THREAD_LOCAL_P (decl))
5943 return omp_notice_threadprivate_variable (ctx, decl, NULL_TREE);
5944
5945 if (DECL_HAS_VALUE_EXPR_P (decl))
5946 {
5947 tree value = get_base_address (DECL_VALUE_EXPR (decl));
5948
5949 if (value && DECL_P (value) && DECL_THREAD_LOCAL_P (value))
5950 return omp_notice_threadprivate_variable (ctx, decl, value);
5951 }
5952 }
5953
5954 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5955 if (n == NULL)
5956 {
5957 enum omp_clause_default_kind default_kind, kind;
5958 struct gimplify_omp_ctx *octx;
5959
5960 if (ctx->region_type == ORT_WORKSHARE)
5961 goto do_outer;
5962
5963 /* ??? Some compiler-generated variables (like SAVE_EXPRs) could be
5964 remapped firstprivate instead of shared. To some extent this is
5965 addressed in omp_firstprivatize_type_sizes, but not effectively. */
5966 default_kind = ctx->default_kind;
5967 kind = lang_hooks.decls.omp_predetermined_sharing (decl);
5968 if (kind != OMP_CLAUSE_DEFAULT_UNSPECIFIED)
5969 default_kind = kind;
5970
5971 switch (default_kind)
5972 {
5973 case OMP_CLAUSE_DEFAULT_NONE:
5974 error ("%qE not specified in enclosing parallel",
5975 DECL_NAME (lang_hooks.decls.omp_report_decl (decl)));
5976 if ((ctx->region_type & ORT_TASK) != 0)
5977 error_at (ctx->location, "enclosing task");
5978 else
5979 error_at (ctx->location, "enclosing parallel");
5980 /* FALLTHRU */
5981 case OMP_CLAUSE_DEFAULT_SHARED:
5982 flags |= GOVD_SHARED;
5983 break;
5984 case OMP_CLAUSE_DEFAULT_PRIVATE:
5985 flags |= GOVD_PRIVATE;
5986 break;
5987 case OMP_CLAUSE_DEFAULT_FIRSTPRIVATE:
5988 flags |= GOVD_FIRSTPRIVATE;
5989 break;
5990 case OMP_CLAUSE_DEFAULT_UNSPECIFIED:
5991 /* decl will be either GOVD_FIRSTPRIVATE or GOVD_SHARED. */
5992 gcc_assert ((ctx->region_type & ORT_TASK) != 0);
5993 if (ctx->outer_context)
5994 omp_notice_variable (ctx->outer_context, decl, in_code);
5995 for (octx = ctx->outer_context; octx; octx = octx->outer_context)
5996 {
5997 splay_tree_node n2;
5998
5999 n2 = splay_tree_lookup (octx->variables, (splay_tree_key) decl);
6000 if (n2 && (n2->value & GOVD_DATA_SHARE_CLASS) != GOVD_SHARED)
6001 {
6002 flags |= GOVD_FIRSTPRIVATE;
6003 break;
6004 }
6005 if ((octx->region_type & ORT_PARALLEL) != 0)
6006 break;
6007 }
6008 if (flags & GOVD_FIRSTPRIVATE)
6009 break;
6010 if (octx == NULL
6011 && (TREE_CODE (decl) == PARM_DECL
6012 || (!is_global_var (decl)
6013 && DECL_CONTEXT (decl) == current_function_decl)))
6014 {
6015 flags |= GOVD_FIRSTPRIVATE;
6016 break;
6017 }
6018 flags |= GOVD_SHARED;
6019 break;
6020 default:
6021 gcc_unreachable ();
6022 }
6023
6024 if ((flags & GOVD_PRIVATE)
6025 && lang_hooks.decls.omp_private_outer_ref (decl))
6026 flags |= GOVD_PRIVATE_OUTER_REF;
6027
6028 omp_add_variable (ctx, decl, flags);
6029
6030 shared = (flags & GOVD_SHARED) != 0;
6031 ret = lang_hooks.decls.omp_disregard_value_expr (decl, shared);
6032 goto do_outer;
6033 }
6034
6035 if ((n->value & (GOVD_SEEN | GOVD_LOCAL)) == 0
6036 && (flags & (GOVD_SEEN | GOVD_LOCAL)) == GOVD_SEEN
6037 && DECL_SIZE (decl)
6038 && TREE_CODE (DECL_SIZE (decl)) != INTEGER_CST)
6039 {
6040 splay_tree_node n2;
6041 tree t = DECL_VALUE_EXPR (decl);
6042 gcc_assert (TREE_CODE (t) == INDIRECT_REF);
6043 t = TREE_OPERAND (t, 0);
6044 gcc_assert (DECL_P (t));
6045 n2 = splay_tree_lookup (ctx->variables, (splay_tree_key) t);
6046 n2->value |= GOVD_SEEN;
6047 }
6048
6049 shared = ((flags | n->value) & GOVD_SHARED) != 0;
6050 ret = lang_hooks.decls.omp_disregard_value_expr (decl, shared);
6051
6052 /* If nothing changed, there's nothing left to do. */
6053 if ((n->value & flags) == flags)
6054 return ret;
6055 flags |= n->value;
6056 n->value = flags;
6057
6058 do_outer:
6059 /* If the variable is private in the current context, then we don't
6060 need to propagate anything to an outer context. */
6061 if ((flags & GOVD_PRIVATE) && !(flags & GOVD_PRIVATE_OUTER_REF))
6062 return ret;
6063 if (ctx->outer_context
6064 && omp_notice_variable (ctx->outer_context, decl, in_code))
6065 return true;
6066 return ret;
6067 }
6068
6069 /* Verify that DECL is private within CTX. If there's specific information
6070 to the contrary in the innermost scope, generate an error. */
6071
6072 static bool
6073 omp_is_private (struct gimplify_omp_ctx *ctx, tree decl)
6074 {
6075 splay_tree_node n;
6076
6077 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
6078 if (n != NULL)
6079 {
6080 if (n->value & GOVD_SHARED)
6081 {
6082 if (ctx == gimplify_omp_ctxp)
6083 {
6084 error ("iteration variable %qE should be private",
6085 DECL_NAME (decl));
6086 n->value = GOVD_PRIVATE;
6087 return true;
6088 }
6089 else
6090 return false;
6091 }
6092 else if ((n->value & GOVD_EXPLICIT) != 0
6093 && (ctx == gimplify_omp_ctxp
6094 || (ctx->region_type == ORT_COMBINED_PARALLEL
6095 && gimplify_omp_ctxp->outer_context == ctx)))
6096 {
6097 if ((n->value & GOVD_FIRSTPRIVATE) != 0)
6098 error ("iteration variable %qE should not be firstprivate",
6099 DECL_NAME (decl));
6100 else if ((n->value & GOVD_REDUCTION) != 0)
6101 error ("iteration variable %qE should not be reduction",
6102 DECL_NAME (decl));
6103 }
6104 return (ctx == gimplify_omp_ctxp
6105 || (ctx->region_type == ORT_COMBINED_PARALLEL
6106 && gimplify_omp_ctxp->outer_context == ctx));
6107 }
6108
6109 if (ctx->region_type != ORT_WORKSHARE)
6110 return false;
6111 else if (ctx->outer_context)
6112 return omp_is_private (ctx->outer_context, decl);
6113 return false;
6114 }
6115
6116 /* Return true if DECL is private within a parallel region
6117 that binds to the current construct's context or in parallel
6118 region's REDUCTION clause. */
6119
6120 static bool
6121 omp_check_private (struct gimplify_omp_ctx *ctx, tree decl)
6122 {
6123 splay_tree_node n;
6124
6125 do
6126 {
6127 ctx = ctx->outer_context;
6128 if (ctx == NULL)
6129 return !(is_global_var (decl)
6130 /* References might be private, but might be shared too. */
6131 || lang_hooks.decls.omp_privatize_by_reference (decl));
6132
6133 n = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
6134 if (n != NULL)
6135 return (n->value & GOVD_SHARED) == 0;
6136 }
6137 while (ctx->region_type == ORT_WORKSHARE);
6138 return false;
6139 }
6140
6141 /* Scan the OpenMP clauses in *LIST_P, installing mappings into a new
6142 and previous omp contexts. */
6143
6144 static void
6145 gimplify_scan_omp_clauses (tree *list_p, gimple_seq *pre_p,
6146 enum omp_region_type region_type)
6147 {
6148 struct gimplify_omp_ctx *ctx, *outer_ctx;
6149 struct gimplify_ctx gctx;
6150 tree c;
6151
6152 ctx = new_omp_context (region_type);
6153 outer_ctx = ctx->outer_context;
6154
6155 while ((c = *list_p) != NULL)
6156 {
6157 bool remove = false;
6158 bool notice_outer = true;
6159 const char *check_non_private = NULL;
6160 unsigned int flags;
6161 tree decl;
6162
6163 switch (OMP_CLAUSE_CODE (c))
6164 {
6165 case OMP_CLAUSE_PRIVATE:
6166 flags = GOVD_PRIVATE | GOVD_EXPLICIT;
6167 if (lang_hooks.decls.omp_private_outer_ref (OMP_CLAUSE_DECL (c)))
6168 {
6169 flags |= GOVD_PRIVATE_OUTER_REF;
6170 OMP_CLAUSE_PRIVATE_OUTER_REF (c) = 1;
6171 }
6172 else
6173 notice_outer = false;
6174 goto do_add;
6175 case OMP_CLAUSE_SHARED:
6176 flags = GOVD_SHARED | GOVD_EXPLICIT;
6177 goto do_add;
6178 case OMP_CLAUSE_FIRSTPRIVATE:
6179 flags = GOVD_FIRSTPRIVATE | GOVD_EXPLICIT;
6180 check_non_private = "firstprivate";
6181 goto do_add;
6182 case OMP_CLAUSE_LASTPRIVATE:
6183 flags = GOVD_LASTPRIVATE | GOVD_SEEN | GOVD_EXPLICIT;
6184 check_non_private = "lastprivate";
6185 goto do_add;
6186 case OMP_CLAUSE_REDUCTION:
6187 flags = GOVD_REDUCTION | GOVD_SEEN | GOVD_EXPLICIT;
6188 check_non_private = "reduction";
6189 goto do_add;
6190
6191 do_add:
6192 decl = OMP_CLAUSE_DECL (c);
6193 if (error_operand_p (decl))
6194 {
6195 remove = true;
6196 break;
6197 }
6198 omp_add_variable (ctx, decl, flags);
6199 if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_REDUCTION
6200 && OMP_CLAUSE_REDUCTION_PLACEHOLDER (c))
6201 {
6202 omp_add_variable (ctx, OMP_CLAUSE_REDUCTION_PLACEHOLDER (c),
6203 GOVD_LOCAL | GOVD_SEEN);
6204 gimplify_omp_ctxp = ctx;
6205 push_gimplify_context (&gctx);
6206
6207 OMP_CLAUSE_REDUCTION_GIMPLE_INIT (c) = NULL;
6208 OMP_CLAUSE_REDUCTION_GIMPLE_MERGE (c) = NULL;
6209
6210 gimplify_and_add (OMP_CLAUSE_REDUCTION_INIT (c),
6211 &OMP_CLAUSE_REDUCTION_GIMPLE_INIT (c));
6212 pop_gimplify_context
6213 (gimple_seq_first_stmt (OMP_CLAUSE_REDUCTION_GIMPLE_INIT (c)));
6214 push_gimplify_context (&gctx);
6215 gimplify_and_add (OMP_CLAUSE_REDUCTION_MERGE (c),
6216 &OMP_CLAUSE_REDUCTION_GIMPLE_MERGE (c));
6217 pop_gimplify_context
6218 (gimple_seq_first_stmt (OMP_CLAUSE_REDUCTION_GIMPLE_MERGE (c)));
6219 OMP_CLAUSE_REDUCTION_INIT (c) = NULL_TREE;
6220 OMP_CLAUSE_REDUCTION_MERGE (c) = NULL_TREE;
6221
6222 gimplify_omp_ctxp = outer_ctx;
6223 }
6224 else if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_LASTPRIVATE
6225 && OMP_CLAUSE_LASTPRIVATE_STMT (c))
6226 {
6227 gimplify_omp_ctxp = ctx;
6228 push_gimplify_context (&gctx);
6229 if (TREE_CODE (OMP_CLAUSE_LASTPRIVATE_STMT (c)) != BIND_EXPR)
6230 {
6231 tree bind = build3 (BIND_EXPR, void_type_node, NULL,
6232 NULL, NULL);
6233 TREE_SIDE_EFFECTS (bind) = 1;
6234 BIND_EXPR_BODY (bind) = OMP_CLAUSE_LASTPRIVATE_STMT (c);
6235 OMP_CLAUSE_LASTPRIVATE_STMT (c) = bind;
6236 }
6237 gimplify_and_add (OMP_CLAUSE_LASTPRIVATE_STMT (c),
6238 &OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c));
6239 pop_gimplify_context
6240 (gimple_seq_first_stmt (OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c)));
6241 OMP_CLAUSE_LASTPRIVATE_STMT (c) = NULL_TREE;
6242
6243 gimplify_omp_ctxp = outer_ctx;
6244 }
6245 if (notice_outer)
6246 goto do_notice;
6247 break;
6248
6249 case OMP_CLAUSE_COPYIN:
6250 case OMP_CLAUSE_COPYPRIVATE:
6251 decl = OMP_CLAUSE_DECL (c);
6252 if (error_operand_p (decl))
6253 {
6254 remove = true;
6255 break;
6256 }
6257 do_notice:
6258 if (outer_ctx)
6259 omp_notice_variable (outer_ctx, decl, true);
6260 if (check_non_private
6261 && region_type == ORT_WORKSHARE
6262 && omp_check_private (ctx, decl))
6263 {
6264 error ("%s variable %qE is private in outer context",
6265 check_non_private, DECL_NAME (decl));
6266 remove = true;
6267 }
6268 break;
6269
6270 case OMP_CLAUSE_FINAL:
6271 case OMP_CLAUSE_IF:
6272 OMP_CLAUSE_OPERAND (c, 0)
6273 = gimple_boolify (OMP_CLAUSE_OPERAND (c, 0));
6274 /* Fall through. */
6275
6276 case OMP_CLAUSE_SCHEDULE:
6277 case OMP_CLAUSE_NUM_THREADS:
6278 if (gimplify_expr (&OMP_CLAUSE_OPERAND (c, 0), pre_p, NULL,
6279 is_gimple_val, fb_rvalue) == GS_ERROR)
6280 remove = true;
6281 break;
6282
6283 case OMP_CLAUSE_NOWAIT:
6284 case OMP_CLAUSE_ORDERED:
6285 case OMP_CLAUSE_UNTIED:
6286 case OMP_CLAUSE_COLLAPSE:
6287 case OMP_CLAUSE_MERGEABLE:
6288 break;
6289
6290 case OMP_CLAUSE_DEFAULT:
6291 ctx->default_kind = OMP_CLAUSE_DEFAULT_KIND (c);
6292 break;
6293
6294 default:
6295 gcc_unreachable ();
6296 }
6297
6298 if (remove)
6299 *list_p = OMP_CLAUSE_CHAIN (c);
6300 else
6301 list_p = &OMP_CLAUSE_CHAIN (c);
6302 }
6303
6304 gimplify_omp_ctxp = ctx;
6305 }
6306
6307 /* For all variables that were not actually used within the context,
6308 remove PRIVATE, SHARED, and FIRSTPRIVATE clauses. */
6309
6310 static int
6311 gimplify_adjust_omp_clauses_1 (splay_tree_node n, void *data)
6312 {
6313 tree *list_p = (tree *) data;
6314 tree decl = (tree) n->key;
6315 unsigned flags = n->value;
6316 enum omp_clause_code code;
6317 tree clause;
6318 bool private_debug;
6319
6320 if (flags & (GOVD_EXPLICIT | GOVD_LOCAL))
6321 return 0;
6322 if ((flags & GOVD_SEEN) == 0)
6323 return 0;
6324 if (flags & GOVD_DEBUG_PRIVATE)
6325 {
6326 gcc_assert ((flags & GOVD_DATA_SHARE_CLASS) == GOVD_PRIVATE);
6327 private_debug = true;
6328 }
6329 else
6330 private_debug
6331 = lang_hooks.decls.omp_private_debug_clause (decl,
6332 !!(flags & GOVD_SHARED));
6333 if (private_debug)
6334 code = OMP_CLAUSE_PRIVATE;
6335 else if (flags & GOVD_SHARED)
6336 {
6337 if (is_global_var (decl))
6338 {
6339 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp->outer_context;
6340 while (ctx != NULL)
6341 {
6342 splay_tree_node on
6343 = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
6344 if (on && (on->value & (GOVD_FIRSTPRIVATE | GOVD_LASTPRIVATE
6345 | GOVD_PRIVATE | GOVD_REDUCTION)) != 0)
6346 break;
6347 ctx = ctx->outer_context;
6348 }
6349 if (ctx == NULL)
6350 return 0;
6351 }
6352 code = OMP_CLAUSE_SHARED;
6353 }
6354 else if (flags & GOVD_PRIVATE)
6355 code = OMP_CLAUSE_PRIVATE;
6356 else if (flags & GOVD_FIRSTPRIVATE)
6357 code = OMP_CLAUSE_FIRSTPRIVATE;
6358 else
6359 gcc_unreachable ();
6360
6361 clause = build_omp_clause (input_location, code);
6362 OMP_CLAUSE_DECL (clause) = decl;
6363 OMP_CLAUSE_CHAIN (clause) = *list_p;
6364 if (private_debug)
6365 OMP_CLAUSE_PRIVATE_DEBUG (clause) = 1;
6366 else if (code == OMP_CLAUSE_PRIVATE && (flags & GOVD_PRIVATE_OUTER_REF))
6367 OMP_CLAUSE_PRIVATE_OUTER_REF (clause) = 1;
6368 *list_p = clause;
6369 lang_hooks.decls.omp_finish_clause (clause);
6370
6371 return 0;
6372 }
6373
6374 static void
6375 gimplify_adjust_omp_clauses (tree *list_p)
6376 {
6377 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
6378 tree c, decl;
6379
6380 while ((c = *list_p) != NULL)
6381 {
6382 splay_tree_node n;
6383 bool remove = false;
6384
6385 switch (OMP_CLAUSE_CODE (c))
6386 {
6387 case OMP_CLAUSE_PRIVATE:
6388 case OMP_CLAUSE_SHARED:
6389 case OMP_CLAUSE_FIRSTPRIVATE:
6390 decl = OMP_CLAUSE_DECL (c);
6391 n = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
6392 remove = !(n->value & GOVD_SEEN);
6393 if (! remove)
6394 {
6395 bool shared = OMP_CLAUSE_CODE (c) == OMP_CLAUSE_SHARED;
6396 if ((n->value & GOVD_DEBUG_PRIVATE)
6397 || lang_hooks.decls.omp_private_debug_clause (decl, shared))
6398 {
6399 gcc_assert ((n->value & GOVD_DEBUG_PRIVATE) == 0
6400 || ((n->value & GOVD_DATA_SHARE_CLASS)
6401 == GOVD_PRIVATE));
6402 OMP_CLAUSE_SET_CODE (c, OMP_CLAUSE_PRIVATE);
6403 OMP_CLAUSE_PRIVATE_DEBUG (c) = 1;
6404 }
6405 }
6406 break;
6407
6408 case OMP_CLAUSE_LASTPRIVATE:
6409 /* Make sure OMP_CLAUSE_LASTPRIVATE_FIRSTPRIVATE is set to
6410 accurately reflect the presence of a FIRSTPRIVATE clause. */
6411 decl = OMP_CLAUSE_DECL (c);
6412 n = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
6413 OMP_CLAUSE_LASTPRIVATE_FIRSTPRIVATE (c)
6414 = (n->value & GOVD_FIRSTPRIVATE) != 0;
6415 break;
6416
6417 case OMP_CLAUSE_REDUCTION:
6418 case OMP_CLAUSE_COPYIN:
6419 case OMP_CLAUSE_COPYPRIVATE:
6420 case OMP_CLAUSE_IF:
6421 case OMP_CLAUSE_NUM_THREADS:
6422 case OMP_CLAUSE_SCHEDULE:
6423 case OMP_CLAUSE_NOWAIT:
6424 case OMP_CLAUSE_ORDERED:
6425 case OMP_CLAUSE_DEFAULT:
6426 case OMP_CLAUSE_UNTIED:
6427 case OMP_CLAUSE_COLLAPSE:
6428 case OMP_CLAUSE_FINAL:
6429 case OMP_CLAUSE_MERGEABLE:
6430 break;
6431
6432 default:
6433 gcc_unreachable ();
6434 }
6435
6436 if (remove)
6437 *list_p = OMP_CLAUSE_CHAIN (c);
6438 else
6439 list_p = &OMP_CLAUSE_CHAIN (c);
6440 }
6441
6442 /* Add in any implicit data sharing. */
6443 splay_tree_foreach (ctx->variables, gimplify_adjust_omp_clauses_1, list_p);
6444
6445 gimplify_omp_ctxp = ctx->outer_context;
6446 delete_omp_context (ctx);
6447 }
6448
6449 /* Gimplify the contents of an OMP_PARALLEL statement. This involves
6450 gimplification of the body, as well as scanning the body for used
6451 variables. We need to do this scan now, because variable-sized
6452 decls will be decomposed during gimplification. */
6453
6454 static void
6455 gimplify_omp_parallel (tree *expr_p, gimple_seq *pre_p)
6456 {
6457 tree expr = *expr_p;
6458 gimple g;
6459 gimple_seq body = NULL;
6460 struct gimplify_ctx gctx;
6461
6462 gimplify_scan_omp_clauses (&OMP_PARALLEL_CLAUSES (expr), pre_p,
6463 OMP_PARALLEL_COMBINED (expr)
6464 ? ORT_COMBINED_PARALLEL
6465 : ORT_PARALLEL);
6466
6467 push_gimplify_context (&gctx);
6468
6469 g = gimplify_and_return_first (OMP_PARALLEL_BODY (expr), &body);
6470 if (gimple_code (g) == GIMPLE_BIND)
6471 pop_gimplify_context (g);
6472 else
6473 pop_gimplify_context (NULL);
6474
6475 gimplify_adjust_omp_clauses (&OMP_PARALLEL_CLAUSES (expr));
6476
6477 g = gimple_build_omp_parallel (body,
6478 OMP_PARALLEL_CLAUSES (expr),
6479 NULL_TREE, NULL_TREE);
6480 if (OMP_PARALLEL_COMBINED (expr))
6481 gimple_omp_set_subcode (g, GF_OMP_PARALLEL_COMBINED);
6482 gimplify_seq_add_stmt (pre_p, g);
6483 *expr_p = NULL_TREE;
6484 }
6485
6486 /* Gimplify the contents of an OMP_TASK statement. This involves
6487 gimplification of the body, as well as scanning the body for used
6488 variables. We need to do this scan now, because variable-sized
6489 decls will be decomposed during gimplification. */
6490
6491 static void
6492 gimplify_omp_task (tree *expr_p, gimple_seq *pre_p)
6493 {
6494 tree expr = *expr_p;
6495 gimple g;
6496 gimple_seq body = NULL;
6497 struct gimplify_ctx gctx;
6498
6499 gimplify_scan_omp_clauses (&OMP_TASK_CLAUSES (expr), pre_p,
6500 find_omp_clause (OMP_TASK_CLAUSES (expr),
6501 OMP_CLAUSE_UNTIED)
6502 ? ORT_UNTIED_TASK : ORT_TASK);
6503
6504 push_gimplify_context (&gctx);
6505
6506 g = gimplify_and_return_first (OMP_TASK_BODY (expr), &body);
6507 if (gimple_code (g) == GIMPLE_BIND)
6508 pop_gimplify_context (g);
6509 else
6510 pop_gimplify_context (NULL);
6511
6512 gimplify_adjust_omp_clauses (&OMP_TASK_CLAUSES (expr));
6513
6514 g = gimple_build_omp_task (body,
6515 OMP_TASK_CLAUSES (expr),
6516 NULL_TREE, NULL_TREE,
6517 NULL_TREE, NULL_TREE, NULL_TREE);
6518 gimplify_seq_add_stmt (pre_p, g);
6519 *expr_p = NULL_TREE;
6520 }
6521
6522 /* Gimplify the gross structure of an OMP_FOR statement. */
6523
6524 static enum gimplify_status
6525 gimplify_omp_for (tree *expr_p, gimple_seq *pre_p)
6526 {
6527 tree for_stmt, decl, var, t;
6528 enum gimplify_status ret = GS_ALL_DONE;
6529 enum gimplify_status tret;
6530 gimple gfor;
6531 gimple_seq for_body, for_pre_body;
6532 int i;
6533
6534 for_stmt = *expr_p;
6535
6536 gimplify_scan_omp_clauses (&OMP_FOR_CLAUSES (for_stmt), pre_p,
6537 ORT_WORKSHARE);
6538
6539 /* Handle OMP_FOR_INIT. */
6540 for_pre_body = NULL;
6541 gimplify_and_add (OMP_FOR_PRE_BODY (for_stmt), &for_pre_body);
6542 OMP_FOR_PRE_BODY (for_stmt) = NULL_TREE;
6543
6544 for_body = NULL;
6545 gcc_assert (TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt))
6546 == TREE_VEC_LENGTH (OMP_FOR_COND (for_stmt)));
6547 gcc_assert (TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt))
6548 == TREE_VEC_LENGTH (OMP_FOR_INCR (for_stmt)));
6549 for (i = 0; i < TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)); i++)
6550 {
6551 t = TREE_VEC_ELT (OMP_FOR_INIT (for_stmt), i);
6552 gcc_assert (TREE_CODE (t) == MODIFY_EXPR);
6553 decl = TREE_OPERAND (t, 0);
6554 gcc_assert (DECL_P (decl));
6555 gcc_assert (INTEGRAL_TYPE_P (TREE_TYPE (decl))
6556 || POINTER_TYPE_P (TREE_TYPE (decl)));
6557
6558 /* Make sure the iteration variable is private. */
6559 if (omp_is_private (gimplify_omp_ctxp, decl))
6560 omp_notice_variable (gimplify_omp_ctxp, decl, true);
6561 else
6562 omp_add_variable (gimplify_omp_ctxp, decl, GOVD_PRIVATE | GOVD_SEEN);
6563
6564 /* If DECL is not a gimple register, create a temporary variable to act
6565 as an iteration counter. This is valid, since DECL cannot be
6566 modified in the body of the loop. */
6567 if (!is_gimple_reg (decl))
6568 {
6569 var = create_tmp_var (TREE_TYPE (decl), get_name (decl));
6570 TREE_OPERAND (t, 0) = var;
6571
6572 gimplify_seq_add_stmt (&for_body, gimple_build_assign (decl, var));
6573
6574 omp_add_variable (gimplify_omp_ctxp, var, GOVD_PRIVATE | GOVD_SEEN);
6575 }
6576 else
6577 var = decl;
6578
6579 tret = gimplify_expr (&TREE_OPERAND (t, 1), &for_pre_body, NULL,
6580 is_gimple_val, fb_rvalue);
6581 ret = MIN (ret, tret);
6582 if (ret == GS_ERROR)
6583 return ret;
6584
6585 /* Handle OMP_FOR_COND. */
6586 t = TREE_VEC_ELT (OMP_FOR_COND (for_stmt), i);
6587 gcc_assert (COMPARISON_CLASS_P (t));
6588 gcc_assert (TREE_OPERAND (t, 0) == decl);
6589
6590 tret = gimplify_expr (&TREE_OPERAND (t, 1), &for_pre_body, NULL,
6591 is_gimple_val, fb_rvalue);
6592 ret = MIN (ret, tret);
6593
6594 /* Handle OMP_FOR_INCR. */
6595 t = TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i);
6596 switch (TREE_CODE (t))
6597 {
6598 case PREINCREMENT_EXPR:
6599 case POSTINCREMENT_EXPR:
6600 t = build_int_cst (TREE_TYPE (decl), 1);
6601 t = build2 (PLUS_EXPR, TREE_TYPE (decl), var, t);
6602 t = build2 (MODIFY_EXPR, TREE_TYPE (var), var, t);
6603 TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i) = t;
6604 break;
6605
6606 case PREDECREMENT_EXPR:
6607 case POSTDECREMENT_EXPR:
6608 t = build_int_cst (TREE_TYPE (decl), -1);
6609 t = build2 (PLUS_EXPR, TREE_TYPE (decl), var, t);
6610 t = build2 (MODIFY_EXPR, TREE_TYPE (var), var, t);
6611 TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i) = t;
6612 break;
6613
6614 case MODIFY_EXPR:
6615 gcc_assert (TREE_OPERAND (t, 0) == decl);
6616 TREE_OPERAND (t, 0) = var;
6617
6618 t = TREE_OPERAND (t, 1);
6619 switch (TREE_CODE (t))
6620 {
6621 case PLUS_EXPR:
6622 if (TREE_OPERAND (t, 1) == decl)
6623 {
6624 TREE_OPERAND (t, 1) = TREE_OPERAND (t, 0);
6625 TREE_OPERAND (t, 0) = var;
6626 break;
6627 }
6628
6629 /* Fallthru. */
6630 case MINUS_EXPR:
6631 case POINTER_PLUS_EXPR:
6632 gcc_assert (TREE_OPERAND (t, 0) == decl);
6633 TREE_OPERAND (t, 0) = var;
6634 break;
6635 default:
6636 gcc_unreachable ();
6637 }
6638
6639 tret = gimplify_expr (&TREE_OPERAND (t, 1), &for_pre_body, NULL,
6640 is_gimple_val, fb_rvalue);
6641 ret = MIN (ret, tret);
6642 break;
6643
6644 default:
6645 gcc_unreachable ();
6646 }
6647
6648 if (var != decl || TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)) > 1)
6649 {
6650 tree c;
6651 for (c = OMP_FOR_CLAUSES (for_stmt); c ; c = OMP_CLAUSE_CHAIN (c))
6652 if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_LASTPRIVATE
6653 && OMP_CLAUSE_DECL (c) == decl
6654 && OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c) == NULL)
6655 {
6656 t = TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i);
6657 gcc_assert (TREE_CODE (t) == MODIFY_EXPR);
6658 gcc_assert (TREE_OPERAND (t, 0) == var);
6659 t = TREE_OPERAND (t, 1);
6660 gcc_assert (TREE_CODE (t) == PLUS_EXPR
6661 || TREE_CODE (t) == MINUS_EXPR
6662 || TREE_CODE (t) == POINTER_PLUS_EXPR);
6663 gcc_assert (TREE_OPERAND (t, 0) == var);
6664 t = build2 (TREE_CODE (t), TREE_TYPE (decl), decl,
6665 TREE_OPERAND (t, 1));
6666 gimplify_assign (decl, t,
6667 &OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c));
6668 }
6669 }
6670 }
6671
6672 gimplify_and_add (OMP_FOR_BODY (for_stmt), &for_body);
6673
6674 gimplify_adjust_omp_clauses (&OMP_FOR_CLAUSES (for_stmt));
6675
6676 gfor = gimple_build_omp_for (for_body, OMP_FOR_CLAUSES (for_stmt),
6677 TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)),
6678 for_pre_body);
6679
6680 for (i = 0; i < TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)); i++)
6681 {
6682 t = TREE_VEC_ELT (OMP_FOR_INIT (for_stmt), i);
6683 gimple_omp_for_set_index (gfor, i, TREE_OPERAND (t, 0));
6684 gimple_omp_for_set_initial (gfor, i, TREE_OPERAND (t, 1));
6685 t = TREE_VEC_ELT (OMP_FOR_COND (for_stmt), i);
6686 gimple_omp_for_set_cond (gfor, i, TREE_CODE (t));
6687 gimple_omp_for_set_final (gfor, i, TREE_OPERAND (t, 1));
6688 t = TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i);
6689 gimple_omp_for_set_incr (gfor, i, TREE_OPERAND (t, 1));
6690 }
6691
6692 gimplify_seq_add_stmt (pre_p, gfor);
6693 return ret == GS_ALL_DONE ? GS_ALL_DONE : GS_ERROR;
6694 }
6695
6696 /* Gimplify the gross structure of other OpenMP worksharing constructs.
6697 In particular, OMP_SECTIONS and OMP_SINGLE. */
6698
6699 static void
6700 gimplify_omp_workshare (tree *expr_p, gimple_seq *pre_p)
6701 {
6702 tree expr = *expr_p;
6703 gimple stmt;
6704 gimple_seq body = NULL;
6705
6706 gimplify_scan_omp_clauses (&OMP_CLAUSES (expr), pre_p, ORT_WORKSHARE);
6707 gimplify_and_add (OMP_BODY (expr), &body);
6708 gimplify_adjust_omp_clauses (&OMP_CLAUSES (expr));
6709
6710 if (TREE_CODE (expr) == OMP_SECTIONS)
6711 stmt = gimple_build_omp_sections (body, OMP_CLAUSES (expr));
6712 else if (TREE_CODE (expr) == OMP_SINGLE)
6713 stmt = gimple_build_omp_single (body, OMP_CLAUSES (expr));
6714 else
6715 gcc_unreachable ();
6716
6717 gimplify_seq_add_stmt (pre_p, stmt);
6718 }
6719
6720 /* A subroutine of gimplify_omp_atomic. The front end is supposed to have
6721 stabilized the lhs of the atomic operation as *ADDR. Return true if
6722 EXPR is this stabilized form. */
6723
6724 static bool
6725 goa_lhs_expr_p (tree expr, tree addr)
6726 {
6727 /* Also include casts to other type variants. The C front end is fond
6728 of adding these for e.g. volatile variables. This is like
6729 STRIP_TYPE_NOPS but includes the main variant lookup. */
6730 STRIP_USELESS_TYPE_CONVERSION (expr);
6731
6732 if (TREE_CODE (expr) == INDIRECT_REF)
6733 {
6734 expr = TREE_OPERAND (expr, 0);
6735 while (expr != addr
6736 && (CONVERT_EXPR_P (expr)
6737 || TREE_CODE (expr) == NON_LVALUE_EXPR)
6738 && TREE_CODE (expr) == TREE_CODE (addr)
6739 && types_compatible_p (TREE_TYPE (expr), TREE_TYPE (addr)))
6740 {
6741 expr = TREE_OPERAND (expr, 0);
6742 addr = TREE_OPERAND (addr, 0);
6743 }
6744 if (expr == addr)
6745 return true;
6746 return (TREE_CODE (addr) == ADDR_EXPR
6747 && TREE_CODE (expr) == ADDR_EXPR
6748 && TREE_OPERAND (addr, 0) == TREE_OPERAND (expr, 0));
6749 }
6750 if (TREE_CODE (addr) == ADDR_EXPR && expr == TREE_OPERAND (addr, 0))
6751 return true;
6752 return false;
6753 }
6754
6755 /* Walk *EXPR_P and replace appearances of *LHS_ADDR with LHS_VAR. If an
6756 expression does not involve the lhs, evaluate it into a temporary.
6757 Return 1 if the lhs appeared as a subexpression, 0 if it did not,
6758 or -1 if an error was encountered. */
6759
6760 static int
6761 goa_stabilize_expr (tree *expr_p, gimple_seq *pre_p, tree lhs_addr,
6762 tree lhs_var)
6763 {
6764 tree expr = *expr_p;
6765 int saw_lhs;
6766
6767 if (goa_lhs_expr_p (expr, lhs_addr))
6768 {
6769 *expr_p = lhs_var;
6770 return 1;
6771 }
6772 if (is_gimple_val (expr))
6773 return 0;
6774
6775 saw_lhs = 0;
6776 switch (TREE_CODE_CLASS (TREE_CODE (expr)))
6777 {
6778 case tcc_binary:
6779 case tcc_comparison:
6780 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 1), pre_p, lhs_addr,
6781 lhs_var);
6782 case tcc_unary:
6783 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 0), pre_p, lhs_addr,
6784 lhs_var);
6785 break;
6786 case tcc_expression:
6787 switch (TREE_CODE (expr))
6788 {
6789 case TRUTH_ANDIF_EXPR:
6790 case TRUTH_ORIF_EXPR:
6791 case TRUTH_AND_EXPR:
6792 case TRUTH_OR_EXPR:
6793 case TRUTH_XOR_EXPR:
6794 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 1), pre_p,
6795 lhs_addr, lhs_var);
6796 case TRUTH_NOT_EXPR:
6797 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 0), pre_p,
6798 lhs_addr, lhs_var);
6799 break;
6800 case COMPOUND_EXPR:
6801 /* Break out any preevaluations from cp_build_modify_expr. */
6802 for (; TREE_CODE (expr) == COMPOUND_EXPR;
6803 expr = TREE_OPERAND (expr, 1))
6804 gimplify_stmt (&TREE_OPERAND (expr, 0), pre_p);
6805 *expr_p = expr;
6806 return goa_stabilize_expr (expr_p, pre_p, lhs_addr, lhs_var);
6807 default:
6808 break;
6809 }
6810 break;
6811 default:
6812 break;
6813 }
6814
6815 if (saw_lhs == 0)
6816 {
6817 enum gimplify_status gs;
6818 gs = gimplify_expr (expr_p, pre_p, NULL, is_gimple_val, fb_rvalue);
6819 if (gs != GS_ALL_DONE)
6820 saw_lhs = -1;
6821 }
6822
6823 return saw_lhs;
6824 }
6825
6826 /* Gimplify an OMP_ATOMIC statement. */
6827
6828 static enum gimplify_status
6829 gimplify_omp_atomic (tree *expr_p, gimple_seq *pre_p)
6830 {
6831 tree addr = TREE_OPERAND (*expr_p, 0);
6832 tree rhs = TREE_CODE (*expr_p) == OMP_ATOMIC_READ
6833 ? NULL : TREE_OPERAND (*expr_p, 1);
6834 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (addr)));
6835 tree tmp_load;
6836 gimple loadstmt, storestmt;
6837
6838 tmp_load = create_tmp_reg (type, NULL);
6839 if (rhs && goa_stabilize_expr (&rhs, pre_p, addr, tmp_load) < 0)
6840 return GS_ERROR;
6841
6842 if (gimplify_expr (&addr, pre_p, NULL, is_gimple_val, fb_rvalue)
6843 != GS_ALL_DONE)
6844 return GS_ERROR;
6845
6846 loadstmt = gimple_build_omp_atomic_load (tmp_load, addr);
6847 gimplify_seq_add_stmt (pre_p, loadstmt);
6848 if (rhs && gimplify_expr (&rhs, pre_p, NULL, is_gimple_val, fb_rvalue)
6849 != GS_ALL_DONE)
6850 return GS_ERROR;
6851
6852 if (TREE_CODE (*expr_p) == OMP_ATOMIC_READ)
6853 rhs = tmp_load;
6854 storestmt = gimple_build_omp_atomic_store (rhs);
6855 gimplify_seq_add_stmt (pre_p, storestmt);
6856 switch (TREE_CODE (*expr_p))
6857 {
6858 case OMP_ATOMIC_READ:
6859 case OMP_ATOMIC_CAPTURE_OLD:
6860 *expr_p = tmp_load;
6861 gimple_omp_atomic_set_need_value (loadstmt);
6862 break;
6863 case OMP_ATOMIC_CAPTURE_NEW:
6864 *expr_p = rhs;
6865 gimple_omp_atomic_set_need_value (storestmt);
6866 break;
6867 default:
6868 *expr_p = NULL;
6869 break;
6870 }
6871
6872 return GS_ALL_DONE;
6873 }
6874
6875 /* Gimplify a TRANSACTION_EXPR. This involves gimplification of the
6876 body, and adding some EH bits. */
6877
6878 static enum gimplify_status
6879 gimplify_transaction (tree *expr_p, gimple_seq *pre_p)
6880 {
6881 tree expr = *expr_p, temp, tbody = TRANSACTION_EXPR_BODY (expr);
6882 gimple g;
6883 gimple_seq body = NULL;
6884 struct gimplify_ctx gctx;
6885 int subcode = 0;
6886
6887 /* Wrap the transaction body in a BIND_EXPR so we have a context
6888 where to put decls for OpenMP. */
6889 if (TREE_CODE (tbody) != BIND_EXPR)
6890 {
6891 tree bind = build3 (BIND_EXPR, void_type_node, NULL, tbody, NULL);
6892 TREE_SIDE_EFFECTS (bind) = 1;
6893 SET_EXPR_LOCATION (bind, EXPR_LOCATION (tbody));
6894 TRANSACTION_EXPR_BODY (expr) = bind;
6895 }
6896
6897 push_gimplify_context (&gctx);
6898 temp = voidify_wrapper_expr (*expr_p, NULL);
6899
6900 g = gimplify_and_return_first (TRANSACTION_EXPR_BODY (expr), &body);
6901 pop_gimplify_context (g);
6902
6903 g = gimple_build_transaction (body, NULL);
6904 if (TRANSACTION_EXPR_OUTER (expr))
6905 subcode = GTMA_IS_OUTER;
6906 else if (TRANSACTION_EXPR_RELAXED (expr))
6907 subcode = GTMA_IS_RELAXED;
6908 gimple_transaction_set_subcode (g, subcode);
6909
6910 gimplify_seq_add_stmt (pre_p, g);
6911
6912 if (temp)
6913 {
6914 *expr_p = temp;
6915 return GS_OK;
6916 }
6917
6918 *expr_p = NULL_TREE;
6919 return GS_ALL_DONE;
6920 }
6921
6922 /* Convert the GENERIC expression tree *EXPR_P to GIMPLE. If the
6923 expression produces a value to be used as an operand inside a GIMPLE
6924 statement, the value will be stored back in *EXPR_P. This value will
6925 be a tree of class tcc_declaration, tcc_constant, tcc_reference or
6926 an SSA_NAME. The corresponding sequence of GIMPLE statements is
6927 emitted in PRE_P and POST_P.
6928
6929 Additionally, this process may overwrite parts of the input
6930 expression during gimplification. Ideally, it should be
6931 possible to do non-destructive gimplification.
6932
6933 EXPR_P points to the GENERIC expression to convert to GIMPLE. If
6934 the expression needs to evaluate to a value to be used as
6935 an operand in a GIMPLE statement, this value will be stored in
6936 *EXPR_P on exit. This happens when the caller specifies one
6937 of fb_lvalue or fb_rvalue fallback flags.
6938
6939 PRE_P will contain the sequence of GIMPLE statements corresponding
6940 to the evaluation of EXPR and all the side-effects that must
6941 be executed before the main expression. On exit, the last
6942 statement of PRE_P is the core statement being gimplified. For
6943 instance, when gimplifying 'if (++a)' the last statement in
6944 PRE_P will be 'if (t.1)' where t.1 is the result of
6945 pre-incrementing 'a'.
6946
6947 POST_P will contain the sequence of GIMPLE statements corresponding
6948 to the evaluation of all the side-effects that must be executed
6949 after the main expression. If this is NULL, the post
6950 side-effects are stored at the end of PRE_P.
6951
6952 The reason why the output is split in two is to handle post
6953 side-effects explicitly. In some cases, an expression may have
6954 inner and outer post side-effects which need to be emitted in
6955 an order different from the one given by the recursive
6956 traversal. For instance, for the expression (*p--)++ the post
6957 side-effects of '--' must actually occur *after* the post
6958 side-effects of '++'. However, gimplification will first visit
6959 the inner expression, so if a separate POST sequence was not
6960 used, the resulting sequence would be:
6961
6962 1 t.1 = *p
6963 2 p = p - 1
6964 3 t.2 = t.1 + 1
6965 4 *p = t.2
6966
6967 However, the post-decrement operation in line #2 must not be
6968 evaluated until after the store to *p at line #4, so the
6969 correct sequence should be:
6970
6971 1 t.1 = *p
6972 2 t.2 = t.1 + 1
6973 3 *p = t.2
6974 4 p = p - 1
6975
6976 So, by specifying a separate post queue, it is possible
6977 to emit the post side-effects in the correct order.
6978 If POST_P is NULL, an internal queue will be used. Before
6979 returning to the caller, the sequence POST_P is appended to
6980 the main output sequence PRE_P.
6981
6982 GIMPLE_TEST_F points to a function that takes a tree T and
6983 returns nonzero if T is in the GIMPLE form requested by the
6984 caller. The GIMPLE predicates are in gimple.c.
6985
6986 FALLBACK tells the function what sort of a temporary we want if
6987 gimplification cannot produce an expression that complies with
6988 GIMPLE_TEST_F.
6989
6990 fb_none means that no temporary should be generated
6991 fb_rvalue means that an rvalue is OK to generate
6992 fb_lvalue means that an lvalue is OK to generate
6993 fb_either means that either is OK, but an lvalue is preferable.
6994 fb_mayfail means that gimplification may fail (in which case
6995 GS_ERROR will be returned)
6996
6997 The return value is either GS_ERROR or GS_ALL_DONE, since this
6998 function iterates until EXPR is completely gimplified or an error
6999 occurs. */
7000
7001 enum gimplify_status
7002 gimplify_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
7003 bool (*gimple_test_f) (tree), fallback_t fallback)
7004 {
7005 tree tmp;
7006 gimple_seq internal_pre = NULL;
7007 gimple_seq internal_post = NULL;
7008 tree save_expr;
7009 bool is_statement;
7010 location_t saved_location;
7011 enum gimplify_status ret;
7012 gimple_stmt_iterator pre_last_gsi, post_last_gsi;
7013
7014 save_expr = *expr_p;
7015 if (save_expr == NULL_TREE)
7016 return GS_ALL_DONE;
7017
7018 /* If we are gimplifying a top-level statement, PRE_P must be valid. */
7019 is_statement = gimple_test_f == is_gimple_stmt;
7020 if (is_statement)
7021 gcc_assert (pre_p);
7022
7023 /* Consistency checks. */
7024 if (gimple_test_f == is_gimple_reg)
7025 gcc_assert (fallback & (fb_rvalue | fb_lvalue));
7026 else if (gimple_test_f == is_gimple_val
7027 || gimple_test_f == is_gimple_call_addr
7028 || gimple_test_f == is_gimple_condexpr
7029 || gimple_test_f == is_gimple_mem_rhs
7030 || gimple_test_f == is_gimple_mem_rhs_or_call
7031 || gimple_test_f == is_gimple_reg_rhs
7032 || gimple_test_f == is_gimple_reg_rhs_or_call
7033 || gimple_test_f == is_gimple_asm_val
7034 || gimple_test_f == is_gimple_mem_ref_addr)
7035 gcc_assert (fallback & fb_rvalue);
7036 else if (gimple_test_f == is_gimple_min_lval
7037 || gimple_test_f == is_gimple_lvalue)
7038 gcc_assert (fallback & fb_lvalue);
7039 else if (gimple_test_f == is_gimple_addressable)
7040 gcc_assert (fallback & fb_either);
7041 else if (gimple_test_f == is_gimple_stmt)
7042 gcc_assert (fallback == fb_none);
7043 else
7044 {
7045 /* We should have recognized the GIMPLE_TEST_F predicate to
7046 know what kind of fallback to use in case a temporary is
7047 needed to hold the value or address of *EXPR_P. */
7048 gcc_unreachable ();
7049 }
7050
7051 /* We used to check the predicate here and return immediately if it
7052 succeeds. This is wrong; the design is for gimplification to be
7053 idempotent, and for the predicates to only test for valid forms, not
7054 whether they are fully simplified. */
7055 if (pre_p == NULL)
7056 pre_p = &internal_pre;
7057
7058 if (post_p == NULL)
7059 post_p = &internal_post;
7060
7061 /* Remember the last statements added to PRE_P and POST_P. Every
7062 new statement added by the gimplification helpers needs to be
7063 annotated with location information. To centralize the
7064 responsibility, we remember the last statement that had been
7065 added to both queues before gimplifying *EXPR_P. If
7066 gimplification produces new statements in PRE_P and POST_P, those
7067 statements will be annotated with the same location information
7068 as *EXPR_P. */
7069 pre_last_gsi = gsi_last (*pre_p);
7070 post_last_gsi = gsi_last (*post_p);
7071
7072 saved_location = input_location;
7073 if (save_expr != error_mark_node
7074 && EXPR_HAS_LOCATION (*expr_p))
7075 input_location = EXPR_LOCATION (*expr_p);
7076
7077 /* Loop over the specific gimplifiers until the toplevel node
7078 remains the same. */
7079 do
7080 {
7081 /* Strip away as many useless type conversions as possible
7082 at the toplevel. */
7083 STRIP_USELESS_TYPE_CONVERSION (*expr_p);
7084
7085 /* Remember the expr. */
7086 save_expr = *expr_p;
7087
7088 /* Die, die, die, my darling. */
7089 if (save_expr == error_mark_node
7090 || (TREE_TYPE (save_expr)
7091 && TREE_TYPE (save_expr) == error_mark_node))
7092 {
7093 ret = GS_ERROR;
7094 break;
7095 }
7096
7097 /* Do any language-specific gimplification. */
7098 ret = ((enum gimplify_status)
7099 lang_hooks.gimplify_expr (expr_p, pre_p, post_p));
7100 if (ret == GS_OK)
7101 {
7102 if (*expr_p == NULL_TREE)
7103 break;
7104 if (*expr_p != save_expr)
7105 continue;
7106 }
7107 else if (ret != GS_UNHANDLED)
7108 break;
7109
7110 /* Make sure that all the cases set 'ret' appropriately. */
7111 ret = GS_UNHANDLED;
7112 switch (TREE_CODE (*expr_p))
7113 {
7114 /* First deal with the special cases. */
7115
7116 case POSTINCREMENT_EXPR:
7117 case POSTDECREMENT_EXPR:
7118 case PREINCREMENT_EXPR:
7119 case PREDECREMENT_EXPR:
7120 ret = gimplify_self_mod_expr (expr_p, pre_p, post_p,
7121 fallback != fb_none,
7122 TREE_TYPE (*expr_p));
7123 break;
7124
7125 case ARRAY_REF:
7126 case ARRAY_RANGE_REF:
7127 case REALPART_EXPR:
7128 case IMAGPART_EXPR:
7129 case COMPONENT_REF:
7130 case VIEW_CONVERT_EXPR:
7131 ret = gimplify_compound_lval (expr_p, pre_p, post_p,
7132 fallback ? fallback : fb_rvalue);
7133 break;
7134
7135 case COND_EXPR:
7136 ret = gimplify_cond_expr (expr_p, pre_p, fallback);
7137
7138 /* C99 code may assign to an array in a structure value of a
7139 conditional expression, and this has undefined behavior
7140 only on execution, so create a temporary if an lvalue is
7141 required. */
7142 if (fallback == fb_lvalue)
7143 {
7144 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
7145 mark_addressable (*expr_p);
7146 ret = GS_OK;
7147 }
7148 break;
7149
7150 case CALL_EXPR:
7151 ret = gimplify_call_expr (expr_p, pre_p, fallback != fb_none);
7152
7153 /* C99 code may assign to an array in a structure returned
7154 from a function, and this has undefined behavior only on
7155 execution, so create a temporary if an lvalue is
7156 required. */
7157 if (fallback == fb_lvalue)
7158 {
7159 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
7160 mark_addressable (*expr_p);
7161 ret = GS_OK;
7162 }
7163 break;
7164
7165 case TREE_LIST:
7166 gcc_unreachable ();
7167
7168 case COMPOUND_EXPR:
7169 ret = gimplify_compound_expr (expr_p, pre_p, fallback != fb_none);
7170 break;
7171
7172 case COMPOUND_LITERAL_EXPR:
7173 ret = gimplify_compound_literal_expr (expr_p, pre_p,
7174 gimple_test_f, fallback);
7175 break;
7176
7177 case MODIFY_EXPR:
7178 case INIT_EXPR:
7179 ret = gimplify_modify_expr (expr_p, pre_p, post_p,
7180 fallback != fb_none);
7181 break;
7182
7183 case TRUTH_ANDIF_EXPR:
7184 case TRUTH_ORIF_EXPR:
7185 {
7186 /* Preserve the original type of the expression and the
7187 source location of the outer expression. */
7188 tree org_type = TREE_TYPE (*expr_p);
7189 *expr_p = gimple_boolify (*expr_p);
7190 *expr_p = build3_loc (input_location, COND_EXPR,
7191 org_type, *expr_p,
7192 fold_convert_loc
7193 (input_location,
7194 org_type, boolean_true_node),
7195 fold_convert_loc
7196 (input_location,
7197 org_type, boolean_false_node));
7198 ret = GS_OK;
7199 break;
7200 }
7201
7202 case TRUTH_NOT_EXPR:
7203 {
7204 tree type = TREE_TYPE (*expr_p);
7205 /* The parsers are careful to generate TRUTH_NOT_EXPR
7206 only with operands that are always zero or one.
7207 We do not fold here but handle the only interesting case
7208 manually, as fold may re-introduce the TRUTH_NOT_EXPR. */
7209 *expr_p = gimple_boolify (*expr_p);
7210 if (TYPE_PRECISION (TREE_TYPE (*expr_p)) == 1)
7211 *expr_p = build1_loc (input_location, BIT_NOT_EXPR,
7212 TREE_TYPE (*expr_p),
7213 TREE_OPERAND (*expr_p, 0));
7214 else
7215 *expr_p = build2_loc (input_location, BIT_XOR_EXPR,
7216 TREE_TYPE (*expr_p),
7217 TREE_OPERAND (*expr_p, 0),
7218 build_int_cst (TREE_TYPE (*expr_p), 1));
7219 if (!useless_type_conversion_p (type, TREE_TYPE (*expr_p)))
7220 *expr_p = fold_convert_loc (input_location, type, *expr_p);
7221 ret = GS_OK;
7222 break;
7223 }
7224
7225 case ADDR_EXPR:
7226 ret = gimplify_addr_expr (expr_p, pre_p, post_p);
7227 break;
7228
7229 case VA_ARG_EXPR:
7230 ret = gimplify_va_arg_expr (expr_p, pre_p, post_p);
7231 break;
7232
7233 CASE_CONVERT:
7234 if (IS_EMPTY_STMT (*expr_p))
7235 {
7236 ret = GS_ALL_DONE;
7237 break;
7238 }
7239
7240 if (VOID_TYPE_P (TREE_TYPE (*expr_p))
7241 || fallback == fb_none)
7242 {
7243 /* Just strip a conversion to void (or in void context) and
7244 try again. */
7245 *expr_p = TREE_OPERAND (*expr_p, 0);
7246 ret = GS_OK;
7247 break;
7248 }
7249
7250 ret = gimplify_conversion (expr_p);
7251 if (ret == GS_ERROR)
7252 break;
7253 if (*expr_p != save_expr)
7254 break;
7255 /* FALLTHRU */
7256
7257 case FIX_TRUNC_EXPR:
7258 /* unary_expr: ... | '(' cast ')' val | ... */
7259 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7260 is_gimple_val, fb_rvalue);
7261 recalculate_side_effects (*expr_p);
7262 break;
7263
7264 case INDIRECT_REF:
7265 {
7266 bool volatilep = TREE_THIS_VOLATILE (*expr_p);
7267 bool notrap = TREE_THIS_NOTRAP (*expr_p);
7268 tree saved_ptr_type = TREE_TYPE (TREE_OPERAND (*expr_p, 0));
7269
7270 *expr_p = fold_indirect_ref_loc (input_location, *expr_p);
7271 if (*expr_p != save_expr)
7272 {
7273 ret = GS_OK;
7274 break;
7275 }
7276
7277 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7278 is_gimple_reg, fb_rvalue);
7279 if (ret == GS_ERROR)
7280 break;
7281
7282 recalculate_side_effects (*expr_p);
7283 *expr_p = fold_build2_loc (input_location, MEM_REF,
7284 TREE_TYPE (*expr_p),
7285 TREE_OPERAND (*expr_p, 0),
7286 build_int_cst (saved_ptr_type, 0));
7287 TREE_THIS_VOLATILE (*expr_p) = volatilep;
7288 TREE_THIS_NOTRAP (*expr_p) = notrap;
7289 ret = GS_OK;
7290 break;
7291 }
7292
7293 /* We arrive here through the various re-gimplifcation paths. */
7294 case MEM_REF:
7295 /* First try re-folding the whole thing. */
7296 tmp = fold_binary (MEM_REF, TREE_TYPE (*expr_p),
7297 TREE_OPERAND (*expr_p, 0),
7298 TREE_OPERAND (*expr_p, 1));
7299 if (tmp)
7300 {
7301 *expr_p = tmp;
7302 recalculate_side_effects (*expr_p);
7303 ret = GS_OK;
7304 break;
7305 }
7306 /* Avoid re-gimplifying the address operand if it is already
7307 in suitable form. Re-gimplifying would mark the address
7308 operand addressable. Always gimplify when not in SSA form
7309 as we still may have to gimplify decls with value-exprs. */
7310 if (!gimplify_ctxp || !gimplify_ctxp->into_ssa
7311 || !is_gimple_mem_ref_addr (TREE_OPERAND (*expr_p, 0)))
7312 {
7313 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7314 is_gimple_mem_ref_addr, fb_rvalue);
7315 if (ret == GS_ERROR)
7316 break;
7317 }
7318 recalculate_side_effects (*expr_p);
7319 ret = GS_ALL_DONE;
7320 break;
7321
7322 /* Constants need not be gimplified. */
7323 case INTEGER_CST:
7324 case REAL_CST:
7325 case FIXED_CST:
7326 case STRING_CST:
7327 case COMPLEX_CST:
7328 case VECTOR_CST:
7329 ret = GS_ALL_DONE;
7330 break;
7331
7332 case CONST_DECL:
7333 /* If we require an lvalue, such as for ADDR_EXPR, retain the
7334 CONST_DECL node. Otherwise the decl is replaceable by its
7335 value. */
7336 /* ??? Should be == fb_lvalue, but ADDR_EXPR passes fb_either. */
7337 if (fallback & fb_lvalue)
7338 ret = GS_ALL_DONE;
7339 else
7340 {
7341 *expr_p = DECL_INITIAL (*expr_p);
7342 ret = GS_OK;
7343 }
7344 break;
7345
7346 case DECL_EXPR:
7347 ret = gimplify_decl_expr (expr_p, pre_p);
7348 break;
7349
7350 case BIND_EXPR:
7351 ret = gimplify_bind_expr (expr_p, pre_p);
7352 break;
7353
7354 case LOOP_EXPR:
7355 ret = gimplify_loop_expr (expr_p, pre_p);
7356 break;
7357
7358 case SWITCH_EXPR:
7359 ret = gimplify_switch_expr (expr_p, pre_p);
7360 break;
7361
7362 case EXIT_EXPR:
7363 ret = gimplify_exit_expr (expr_p);
7364 break;
7365
7366 case GOTO_EXPR:
7367 /* If the target is not LABEL, then it is a computed jump
7368 and the target needs to be gimplified. */
7369 if (TREE_CODE (GOTO_DESTINATION (*expr_p)) != LABEL_DECL)
7370 {
7371 ret = gimplify_expr (&GOTO_DESTINATION (*expr_p), pre_p,
7372 NULL, is_gimple_val, fb_rvalue);
7373 if (ret == GS_ERROR)
7374 break;
7375 }
7376 gimplify_seq_add_stmt (pre_p,
7377 gimple_build_goto (GOTO_DESTINATION (*expr_p)));
7378 ret = GS_ALL_DONE;
7379 break;
7380
7381 case PREDICT_EXPR:
7382 gimplify_seq_add_stmt (pre_p,
7383 gimple_build_predict (PREDICT_EXPR_PREDICTOR (*expr_p),
7384 PREDICT_EXPR_OUTCOME (*expr_p)));
7385 ret = GS_ALL_DONE;
7386 break;
7387
7388 case LABEL_EXPR:
7389 ret = GS_ALL_DONE;
7390 gcc_assert (decl_function_context (LABEL_EXPR_LABEL (*expr_p))
7391 == current_function_decl);
7392 gimplify_seq_add_stmt (pre_p,
7393 gimple_build_label (LABEL_EXPR_LABEL (*expr_p)));
7394 break;
7395
7396 case CASE_LABEL_EXPR:
7397 ret = gimplify_case_label_expr (expr_p, pre_p);
7398 break;
7399
7400 case RETURN_EXPR:
7401 ret = gimplify_return_expr (*expr_p, pre_p);
7402 break;
7403
7404 case CONSTRUCTOR:
7405 /* Don't reduce this in place; let gimplify_init_constructor work its
7406 magic. Buf if we're just elaborating this for side effects, just
7407 gimplify any element that has side-effects. */
7408 if (fallback == fb_none)
7409 {
7410 unsigned HOST_WIDE_INT ix;
7411 tree val;
7412 tree temp = NULL_TREE;
7413 FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (*expr_p), ix, val)
7414 if (TREE_SIDE_EFFECTS (val))
7415 append_to_statement_list (val, &temp);
7416
7417 *expr_p = temp;
7418 ret = temp ? GS_OK : GS_ALL_DONE;
7419 }
7420 /* C99 code may assign to an array in a constructed
7421 structure or union, and this has undefined behavior only
7422 on execution, so create a temporary if an lvalue is
7423 required. */
7424 else if (fallback == fb_lvalue)
7425 {
7426 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
7427 mark_addressable (*expr_p);
7428 ret = GS_OK;
7429 }
7430 else
7431 ret = GS_ALL_DONE;
7432 break;
7433
7434 /* The following are special cases that are not handled by the
7435 original GIMPLE grammar. */
7436
7437 /* SAVE_EXPR nodes are converted into a GIMPLE identifier and
7438 eliminated. */
7439 case SAVE_EXPR:
7440 ret = gimplify_save_expr (expr_p, pre_p, post_p);
7441 break;
7442
7443 case BIT_FIELD_REF:
7444 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7445 post_p, is_gimple_lvalue, fb_either);
7446 recalculate_side_effects (*expr_p);
7447 break;
7448
7449 case TARGET_MEM_REF:
7450 {
7451 enum gimplify_status r0 = GS_ALL_DONE, r1 = GS_ALL_DONE;
7452
7453 if (TMR_BASE (*expr_p))
7454 r0 = gimplify_expr (&TMR_BASE (*expr_p), pre_p,
7455 post_p, is_gimple_mem_ref_addr, fb_either);
7456 if (TMR_INDEX (*expr_p))
7457 r1 = gimplify_expr (&TMR_INDEX (*expr_p), pre_p,
7458 post_p, is_gimple_val, fb_rvalue);
7459 if (TMR_INDEX2 (*expr_p))
7460 r1 = gimplify_expr (&TMR_INDEX2 (*expr_p), pre_p,
7461 post_p, is_gimple_val, fb_rvalue);
7462 /* TMR_STEP and TMR_OFFSET are always integer constants. */
7463 ret = MIN (r0, r1);
7464 }
7465 break;
7466
7467 case NON_LVALUE_EXPR:
7468 /* This should have been stripped above. */
7469 gcc_unreachable ();
7470
7471 case ASM_EXPR:
7472 ret = gimplify_asm_expr (expr_p, pre_p, post_p);
7473 break;
7474
7475 case TRY_FINALLY_EXPR:
7476 case TRY_CATCH_EXPR:
7477 {
7478 gimple_seq eval, cleanup;
7479 gimple try_;
7480
7481 /* Calls to destructors are generated automatically in FINALLY/CATCH
7482 block. They should have location as UNKNOWN_LOCATION. However,
7483 gimplify_call_expr will reset these call stmts to input_location
7484 if it finds stmt's location is unknown. To prevent resetting for
7485 destructors, we set the input_location to unknown.
7486 Note that this only affects the destructor calls in FINALLY/CATCH
7487 block, and will automatically reset to its original value by the
7488 end of gimplify_expr. */
7489 input_location = UNKNOWN_LOCATION;
7490 eval = cleanup = NULL;
7491 gimplify_and_add (TREE_OPERAND (*expr_p, 0), &eval);
7492 gimplify_and_add (TREE_OPERAND (*expr_p, 1), &cleanup);
7493 /* Don't create bogus GIMPLE_TRY with empty cleanup. */
7494 if (gimple_seq_empty_p (cleanup))
7495 {
7496 gimple_seq_add_seq (pre_p, eval);
7497 ret = GS_ALL_DONE;
7498 break;
7499 }
7500 try_ = gimple_build_try (eval, cleanup,
7501 TREE_CODE (*expr_p) == TRY_FINALLY_EXPR
7502 ? GIMPLE_TRY_FINALLY
7503 : GIMPLE_TRY_CATCH);
7504 if (LOCATION_LOCUS (saved_location) != UNKNOWN_LOCATION)
7505 gimple_set_location (try_, saved_location);
7506 else
7507 gimple_set_location (try_, EXPR_LOCATION (save_expr));
7508 if (TREE_CODE (*expr_p) == TRY_CATCH_EXPR)
7509 gimple_try_set_catch_is_cleanup (try_,
7510 TRY_CATCH_IS_CLEANUP (*expr_p));
7511 gimplify_seq_add_stmt (pre_p, try_);
7512 ret = GS_ALL_DONE;
7513 break;
7514 }
7515
7516 case CLEANUP_POINT_EXPR:
7517 ret = gimplify_cleanup_point_expr (expr_p, pre_p);
7518 break;
7519
7520 case TARGET_EXPR:
7521 ret = gimplify_target_expr (expr_p, pre_p, post_p);
7522 break;
7523
7524 case CATCH_EXPR:
7525 {
7526 gimple c;
7527 gimple_seq handler = NULL;
7528 gimplify_and_add (CATCH_BODY (*expr_p), &handler);
7529 c = gimple_build_catch (CATCH_TYPES (*expr_p), handler);
7530 gimplify_seq_add_stmt (pre_p, c);
7531 ret = GS_ALL_DONE;
7532 break;
7533 }
7534
7535 case EH_FILTER_EXPR:
7536 {
7537 gimple ehf;
7538 gimple_seq failure = NULL;
7539
7540 gimplify_and_add (EH_FILTER_FAILURE (*expr_p), &failure);
7541 ehf = gimple_build_eh_filter (EH_FILTER_TYPES (*expr_p), failure);
7542 gimple_set_no_warning (ehf, TREE_NO_WARNING (*expr_p));
7543 gimplify_seq_add_stmt (pre_p, ehf);
7544 ret = GS_ALL_DONE;
7545 break;
7546 }
7547
7548 case OBJ_TYPE_REF:
7549 {
7550 enum gimplify_status r0, r1;
7551 r0 = gimplify_expr (&OBJ_TYPE_REF_OBJECT (*expr_p), pre_p,
7552 post_p, is_gimple_val, fb_rvalue);
7553 r1 = gimplify_expr (&OBJ_TYPE_REF_EXPR (*expr_p), pre_p,
7554 post_p, is_gimple_val, fb_rvalue);
7555 TREE_SIDE_EFFECTS (*expr_p) = 0;
7556 ret = MIN (r0, r1);
7557 }
7558 break;
7559
7560 case LABEL_DECL:
7561 /* We get here when taking the address of a label. We mark
7562 the label as "forced"; meaning it can never be removed and
7563 it is a potential target for any computed goto. */
7564 FORCED_LABEL (*expr_p) = 1;
7565 ret = GS_ALL_DONE;
7566 break;
7567
7568 case STATEMENT_LIST:
7569 ret = gimplify_statement_list (expr_p, pre_p);
7570 break;
7571
7572 case WITH_SIZE_EXPR:
7573 {
7574 gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7575 post_p == &internal_post ? NULL : post_p,
7576 gimple_test_f, fallback);
7577 gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p, post_p,
7578 is_gimple_val, fb_rvalue);
7579 ret = GS_ALL_DONE;
7580 }
7581 break;
7582
7583 case VAR_DECL:
7584 case PARM_DECL:
7585 ret = gimplify_var_or_parm_decl (expr_p);
7586 break;
7587
7588 case RESULT_DECL:
7589 /* When within an OpenMP context, notice uses of variables. */
7590 if (gimplify_omp_ctxp)
7591 omp_notice_variable (gimplify_omp_ctxp, *expr_p, true);
7592 ret = GS_ALL_DONE;
7593 break;
7594
7595 case SSA_NAME:
7596 /* Allow callbacks into the gimplifier during optimization. */
7597 ret = GS_ALL_DONE;
7598 break;
7599
7600 case OMP_PARALLEL:
7601 gimplify_omp_parallel (expr_p, pre_p);
7602 ret = GS_ALL_DONE;
7603 break;
7604
7605 case OMP_TASK:
7606 gimplify_omp_task (expr_p, pre_p);
7607 ret = GS_ALL_DONE;
7608 break;
7609
7610 case OMP_FOR:
7611 ret = gimplify_omp_for (expr_p, pre_p);
7612 break;
7613
7614 case OMP_SECTIONS:
7615 case OMP_SINGLE:
7616 gimplify_omp_workshare (expr_p, pre_p);
7617 ret = GS_ALL_DONE;
7618 break;
7619
7620 case OMP_SECTION:
7621 case OMP_MASTER:
7622 case OMP_ORDERED:
7623 case OMP_CRITICAL:
7624 {
7625 gimple_seq body = NULL;
7626 gimple g;
7627
7628 gimplify_and_add (OMP_BODY (*expr_p), &body);
7629 switch (TREE_CODE (*expr_p))
7630 {
7631 case OMP_SECTION:
7632 g = gimple_build_omp_section (body);
7633 break;
7634 case OMP_MASTER:
7635 g = gimple_build_omp_master (body);
7636 break;
7637 case OMP_ORDERED:
7638 g = gimple_build_omp_ordered (body);
7639 break;
7640 case OMP_CRITICAL:
7641 g = gimple_build_omp_critical (body,
7642 OMP_CRITICAL_NAME (*expr_p));
7643 break;
7644 default:
7645 gcc_unreachable ();
7646 }
7647 gimplify_seq_add_stmt (pre_p, g);
7648 ret = GS_ALL_DONE;
7649 break;
7650 }
7651
7652 case OMP_ATOMIC:
7653 case OMP_ATOMIC_READ:
7654 case OMP_ATOMIC_CAPTURE_OLD:
7655 case OMP_ATOMIC_CAPTURE_NEW:
7656 ret = gimplify_omp_atomic (expr_p, pre_p);
7657 break;
7658
7659 case TRANSACTION_EXPR:
7660 ret = gimplify_transaction (expr_p, pre_p);
7661 break;
7662
7663 case TRUTH_AND_EXPR:
7664 case TRUTH_OR_EXPR:
7665 case TRUTH_XOR_EXPR:
7666 {
7667 tree orig_type = TREE_TYPE (*expr_p);
7668 tree new_type, xop0, xop1;
7669 *expr_p = gimple_boolify (*expr_p);
7670 new_type = TREE_TYPE (*expr_p);
7671 if (!useless_type_conversion_p (orig_type, new_type))
7672 {
7673 *expr_p = fold_convert_loc (input_location, orig_type, *expr_p);
7674 ret = GS_OK;
7675 break;
7676 }
7677
7678 /* Boolified binary truth expressions are semantically equivalent
7679 to bitwise binary expressions. Canonicalize them to the
7680 bitwise variant. */
7681 switch (TREE_CODE (*expr_p))
7682 {
7683 case TRUTH_AND_EXPR:
7684 TREE_SET_CODE (*expr_p, BIT_AND_EXPR);
7685 break;
7686 case TRUTH_OR_EXPR:
7687 TREE_SET_CODE (*expr_p, BIT_IOR_EXPR);
7688 break;
7689 case TRUTH_XOR_EXPR:
7690 TREE_SET_CODE (*expr_p, BIT_XOR_EXPR);
7691 break;
7692 default:
7693 break;
7694 }
7695 /* Now make sure that operands have compatible type to
7696 expression's new_type. */
7697 xop0 = TREE_OPERAND (*expr_p, 0);
7698 xop1 = TREE_OPERAND (*expr_p, 1);
7699 if (!useless_type_conversion_p (new_type, TREE_TYPE (xop0)))
7700 TREE_OPERAND (*expr_p, 0) = fold_convert_loc (input_location,
7701 new_type,
7702 xop0);
7703 if (!useless_type_conversion_p (new_type, TREE_TYPE (xop1)))
7704 TREE_OPERAND (*expr_p, 1) = fold_convert_loc (input_location,
7705 new_type,
7706 xop1);
7707 /* Continue classified as tcc_binary. */
7708 goto expr_2;
7709 }
7710
7711 case FMA_EXPR:
7712 case VEC_COND_EXPR:
7713 case VEC_PERM_EXPR:
7714 /* Classified as tcc_expression. */
7715 goto expr_3;
7716
7717 case POINTER_PLUS_EXPR:
7718 {
7719 enum gimplify_status r0, r1;
7720 r0 = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7721 post_p, is_gimple_val, fb_rvalue);
7722 r1 = gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p,
7723 post_p, is_gimple_val, fb_rvalue);
7724 recalculate_side_effects (*expr_p);
7725 ret = MIN (r0, r1);
7726 /* Convert &X + CST to invariant &MEM[&X, CST]. Do this
7727 after gimplifying operands - this is similar to how
7728 it would be folding all gimplified stmts on creation
7729 to have them canonicalized, which is what we eventually
7730 should do anyway. */
7731 if (TREE_CODE (TREE_OPERAND (*expr_p, 1)) == INTEGER_CST
7732 && is_gimple_min_invariant (TREE_OPERAND (*expr_p, 0)))
7733 {
7734 *expr_p = build_fold_addr_expr_with_type_loc
7735 (input_location,
7736 fold_build2 (MEM_REF, TREE_TYPE (TREE_TYPE (*expr_p)),
7737 TREE_OPERAND (*expr_p, 0),
7738 fold_convert (ptr_type_node,
7739 TREE_OPERAND (*expr_p, 1))),
7740 TREE_TYPE (*expr_p));
7741 ret = MIN (ret, GS_OK);
7742 }
7743 break;
7744 }
7745
7746 default:
7747 switch (TREE_CODE_CLASS (TREE_CODE (*expr_p)))
7748 {
7749 case tcc_comparison:
7750 /* Handle comparison of objects of non scalar mode aggregates
7751 with a call to memcmp. It would be nice to only have to do
7752 this for variable-sized objects, but then we'd have to allow
7753 the same nest of reference nodes we allow for MODIFY_EXPR and
7754 that's too complex.
7755
7756 Compare scalar mode aggregates as scalar mode values. Using
7757 memcmp for them would be very inefficient at best, and is
7758 plain wrong if bitfields are involved. */
7759 {
7760 tree type = TREE_TYPE (TREE_OPERAND (*expr_p, 1));
7761
7762 /* Vector comparisons need no boolification. */
7763 if (TREE_CODE (type) == VECTOR_TYPE)
7764 goto expr_2;
7765 else if (!AGGREGATE_TYPE_P (type))
7766 {
7767 tree org_type = TREE_TYPE (*expr_p);
7768 *expr_p = gimple_boolify (*expr_p);
7769 if (!useless_type_conversion_p (org_type,
7770 TREE_TYPE (*expr_p)))
7771 {
7772 *expr_p = fold_convert_loc (input_location,
7773 org_type, *expr_p);
7774 ret = GS_OK;
7775 }
7776 else
7777 goto expr_2;
7778 }
7779 else if (TYPE_MODE (type) != BLKmode)
7780 ret = gimplify_scalar_mode_aggregate_compare (expr_p);
7781 else
7782 ret = gimplify_variable_sized_compare (expr_p);
7783
7784 break;
7785 }
7786
7787 /* If *EXPR_P does not need to be special-cased, handle it
7788 according to its class. */
7789 case tcc_unary:
7790 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7791 post_p, is_gimple_val, fb_rvalue);
7792 break;
7793
7794 case tcc_binary:
7795 expr_2:
7796 {
7797 enum gimplify_status r0, r1;
7798
7799 r0 = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7800 post_p, is_gimple_val, fb_rvalue);
7801 r1 = gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p,
7802 post_p, is_gimple_val, fb_rvalue);
7803
7804 ret = MIN (r0, r1);
7805 break;
7806 }
7807
7808 expr_3:
7809 {
7810 enum gimplify_status r0, r1, r2;
7811
7812 r0 = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7813 post_p, is_gimple_val, fb_rvalue);
7814 r1 = gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p,
7815 post_p, is_gimple_val, fb_rvalue);
7816 r2 = gimplify_expr (&TREE_OPERAND (*expr_p, 2), pre_p,
7817 post_p, is_gimple_val, fb_rvalue);
7818
7819 ret = MIN (MIN (r0, r1), r2);
7820 break;
7821 }
7822
7823 case tcc_declaration:
7824 case tcc_constant:
7825 ret = GS_ALL_DONE;
7826 goto dont_recalculate;
7827
7828 default:
7829 gcc_unreachable ();
7830 }
7831
7832 recalculate_side_effects (*expr_p);
7833
7834 dont_recalculate:
7835 break;
7836 }
7837
7838 gcc_assert (*expr_p || ret != GS_OK);
7839 }
7840 while (ret == GS_OK);
7841
7842 /* If we encountered an error_mark somewhere nested inside, either
7843 stub out the statement or propagate the error back out. */
7844 if (ret == GS_ERROR)
7845 {
7846 if (is_statement)
7847 *expr_p = NULL;
7848 goto out;
7849 }
7850
7851 /* This was only valid as a return value from the langhook, which
7852 we handled. Make sure it doesn't escape from any other context. */
7853 gcc_assert (ret != GS_UNHANDLED);
7854
7855 if (fallback == fb_none && *expr_p && !is_gimple_stmt (*expr_p))
7856 {
7857 /* We aren't looking for a value, and we don't have a valid
7858 statement. If it doesn't have side-effects, throw it away. */
7859 if (!TREE_SIDE_EFFECTS (*expr_p))
7860 *expr_p = NULL;
7861 else if (!TREE_THIS_VOLATILE (*expr_p))
7862 {
7863 /* This is probably a _REF that contains something nested that
7864 has side effects. Recurse through the operands to find it. */
7865 enum tree_code code = TREE_CODE (*expr_p);
7866
7867 switch (code)
7868 {
7869 case COMPONENT_REF:
7870 case REALPART_EXPR:
7871 case IMAGPART_EXPR:
7872 case VIEW_CONVERT_EXPR:
7873 gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7874 gimple_test_f, fallback);
7875 break;
7876
7877 case ARRAY_REF:
7878 case ARRAY_RANGE_REF:
7879 gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7880 gimple_test_f, fallback);
7881 gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p, post_p,
7882 gimple_test_f, fallback);
7883 break;
7884
7885 default:
7886 /* Anything else with side-effects must be converted to
7887 a valid statement before we get here. */
7888 gcc_unreachable ();
7889 }
7890
7891 *expr_p = NULL;
7892 }
7893 else if (COMPLETE_TYPE_P (TREE_TYPE (*expr_p))
7894 && TYPE_MODE (TREE_TYPE (*expr_p)) != BLKmode)
7895 {
7896 /* Historically, the compiler has treated a bare reference
7897 to a non-BLKmode volatile lvalue as forcing a load. */
7898 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (*expr_p));
7899
7900 /* Normally, we do not want to create a temporary for a
7901 TREE_ADDRESSABLE type because such a type should not be
7902 copied by bitwise-assignment. However, we make an
7903 exception here, as all we are doing here is ensuring that
7904 we read the bytes that make up the type. We use
7905 create_tmp_var_raw because create_tmp_var will abort when
7906 given a TREE_ADDRESSABLE type. */
7907 tree tmp = create_tmp_var_raw (type, "vol");
7908 gimple_add_tmp_var (tmp);
7909 gimplify_assign (tmp, *expr_p, pre_p);
7910 *expr_p = NULL;
7911 }
7912 else
7913 /* We can't do anything useful with a volatile reference to
7914 an incomplete type, so just throw it away. Likewise for
7915 a BLKmode type, since any implicit inner load should
7916 already have been turned into an explicit one by the
7917 gimplification process. */
7918 *expr_p = NULL;
7919 }
7920
7921 /* If we are gimplifying at the statement level, we're done. Tack
7922 everything together and return. */
7923 if (fallback == fb_none || is_statement)
7924 {
7925 /* Since *EXPR_P has been converted into a GIMPLE tuple, clear
7926 it out for GC to reclaim it. */
7927 *expr_p = NULL_TREE;
7928
7929 if (!gimple_seq_empty_p (internal_pre)
7930 || !gimple_seq_empty_p (internal_post))
7931 {
7932 gimplify_seq_add_seq (&internal_pre, internal_post);
7933 gimplify_seq_add_seq (pre_p, internal_pre);
7934 }
7935
7936 /* The result of gimplifying *EXPR_P is going to be the last few
7937 statements in *PRE_P and *POST_P. Add location information
7938 to all the statements that were added by the gimplification
7939 helpers. */
7940 if (!gimple_seq_empty_p (*pre_p))
7941 annotate_all_with_location_after (*pre_p, pre_last_gsi, input_location);
7942
7943 if (!gimple_seq_empty_p (*post_p))
7944 annotate_all_with_location_after (*post_p, post_last_gsi,
7945 input_location);
7946
7947 goto out;
7948 }
7949
7950 #ifdef ENABLE_GIMPLE_CHECKING
7951 if (*expr_p)
7952 {
7953 enum tree_code code = TREE_CODE (*expr_p);
7954 /* These expressions should already be in gimple IR form. */
7955 gcc_assert (code != MODIFY_EXPR
7956 && code != ASM_EXPR
7957 && code != BIND_EXPR
7958 && code != CATCH_EXPR
7959 && (code != COND_EXPR || gimplify_ctxp->allow_rhs_cond_expr)
7960 && code != EH_FILTER_EXPR
7961 && code != GOTO_EXPR
7962 && code != LABEL_EXPR
7963 && code != LOOP_EXPR
7964 && code != SWITCH_EXPR
7965 && code != TRY_FINALLY_EXPR
7966 && code != OMP_CRITICAL
7967 && code != OMP_FOR
7968 && code != OMP_MASTER
7969 && code != OMP_ORDERED
7970 && code != OMP_PARALLEL
7971 && code != OMP_SECTIONS
7972 && code != OMP_SECTION
7973 && code != OMP_SINGLE);
7974 }
7975 #endif
7976
7977 /* Otherwise we're gimplifying a subexpression, so the resulting
7978 value is interesting. If it's a valid operand that matches
7979 GIMPLE_TEST_F, we're done. Unless we are handling some
7980 post-effects internally; if that's the case, we need to copy into
7981 a temporary before adding the post-effects to POST_P. */
7982 if (gimple_seq_empty_p (internal_post) && (*gimple_test_f) (*expr_p))
7983 goto out;
7984
7985 /* Otherwise, we need to create a new temporary for the gimplified
7986 expression. */
7987
7988 /* We can't return an lvalue if we have an internal postqueue. The
7989 object the lvalue refers to would (probably) be modified by the
7990 postqueue; we need to copy the value out first, which means an
7991 rvalue. */
7992 if ((fallback & fb_lvalue)
7993 && gimple_seq_empty_p (internal_post)
7994 && is_gimple_addressable (*expr_p))
7995 {
7996 /* An lvalue will do. Take the address of the expression, store it
7997 in a temporary, and replace the expression with an INDIRECT_REF of
7998 that temporary. */
7999 tmp = build_fold_addr_expr_loc (input_location, *expr_p);
8000 gimplify_expr (&tmp, pre_p, post_p, is_gimple_reg, fb_rvalue);
8001 *expr_p = build_simple_mem_ref (tmp);
8002 }
8003 else if ((fallback & fb_rvalue) && is_gimple_reg_rhs_or_call (*expr_p))
8004 {
8005 /* An rvalue will do. Assign the gimplified expression into a
8006 new temporary TMP and replace the original expression with
8007 TMP. First, make sure that the expression has a type so that
8008 it can be assigned into a temporary. */
8009 gcc_assert (!VOID_TYPE_P (TREE_TYPE (*expr_p)));
8010 *expr_p = get_formal_tmp_var (*expr_p, pre_p);
8011 }
8012 else
8013 {
8014 #ifdef ENABLE_GIMPLE_CHECKING
8015 if (!(fallback & fb_mayfail))
8016 {
8017 fprintf (stderr, "gimplification failed:\n");
8018 print_generic_expr (stderr, *expr_p, 0);
8019 debug_tree (*expr_p);
8020 internal_error ("gimplification failed");
8021 }
8022 #endif
8023 gcc_assert (fallback & fb_mayfail);
8024
8025 /* If this is an asm statement, and the user asked for the
8026 impossible, don't die. Fail and let gimplify_asm_expr
8027 issue an error. */
8028 ret = GS_ERROR;
8029 goto out;
8030 }
8031
8032 /* Make sure the temporary matches our predicate. */
8033 gcc_assert ((*gimple_test_f) (*expr_p));
8034
8035 if (!gimple_seq_empty_p (internal_post))
8036 {
8037 annotate_all_with_location (internal_post, input_location);
8038 gimplify_seq_add_seq (pre_p, internal_post);
8039 }
8040
8041 out:
8042 input_location = saved_location;
8043 return ret;
8044 }
8045
8046 /* Look through TYPE for variable-sized objects and gimplify each such
8047 size that we find. Add to LIST_P any statements generated. */
8048
8049 void
8050 gimplify_type_sizes (tree type, gimple_seq *list_p)
8051 {
8052 tree field, t;
8053
8054 if (type == NULL || type == error_mark_node)
8055 return;
8056
8057 /* We first do the main variant, then copy into any other variants. */
8058 type = TYPE_MAIN_VARIANT (type);
8059
8060 /* Avoid infinite recursion. */
8061 if (TYPE_SIZES_GIMPLIFIED (type))
8062 return;
8063
8064 TYPE_SIZES_GIMPLIFIED (type) = 1;
8065
8066 switch (TREE_CODE (type))
8067 {
8068 case INTEGER_TYPE:
8069 case ENUMERAL_TYPE:
8070 case BOOLEAN_TYPE:
8071 case REAL_TYPE:
8072 case FIXED_POINT_TYPE:
8073 gimplify_one_sizepos (&TYPE_MIN_VALUE (type), list_p);
8074 gimplify_one_sizepos (&TYPE_MAX_VALUE (type), list_p);
8075
8076 for (t = TYPE_NEXT_VARIANT (type); t; t = TYPE_NEXT_VARIANT (t))
8077 {
8078 TYPE_MIN_VALUE (t) = TYPE_MIN_VALUE (type);
8079 TYPE_MAX_VALUE (t) = TYPE_MAX_VALUE (type);
8080 }
8081 break;
8082
8083 case ARRAY_TYPE:
8084 /* These types may not have declarations, so handle them here. */
8085 gimplify_type_sizes (TREE_TYPE (type), list_p);
8086 gimplify_type_sizes (TYPE_DOMAIN (type), list_p);
8087 /* Ensure VLA bounds aren't removed, for -O0 they should be variables
8088 with assigned stack slots, for -O1+ -g they should be tracked
8089 by VTA. */
8090 if (!(TYPE_NAME (type)
8091 && TREE_CODE (TYPE_NAME (type)) == TYPE_DECL
8092 && DECL_IGNORED_P (TYPE_NAME (type)))
8093 && TYPE_DOMAIN (type)
8094 && INTEGRAL_TYPE_P (TYPE_DOMAIN (type)))
8095 {
8096 t = TYPE_MIN_VALUE (TYPE_DOMAIN (type));
8097 if (t && TREE_CODE (t) == VAR_DECL && DECL_ARTIFICIAL (t))
8098 DECL_IGNORED_P (t) = 0;
8099 t = TYPE_MAX_VALUE (TYPE_DOMAIN (type));
8100 if (t && TREE_CODE (t) == VAR_DECL && DECL_ARTIFICIAL (t))
8101 DECL_IGNORED_P (t) = 0;
8102 }
8103 break;
8104
8105 case RECORD_TYPE:
8106 case UNION_TYPE:
8107 case QUAL_UNION_TYPE:
8108 for (field = TYPE_FIELDS (type); field; field = DECL_CHAIN (field))
8109 if (TREE_CODE (field) == FIELD_DECL)
8110 {
8111 gimplify_one_sizepos (&DECL_FIELD_OFFSET (field), list_p);
8112 gimplify_one_sizepos (&DECL_SIZE (field), list_p);
8113 gimplify_one_sizepos (&DECL_SIZE_UNIT (field), list_p);
8114 gimplify_type_sizes (TREE_TYPE (field), list_p);
8115 }
8116 break;
8117
8118 case POINTER_TYPE:
8119 case REFERENCE_TYPE:
8120 /* We used to recurse on the pointed-to type here, which turned out to
8121 be incorrect because its definition might refer to variables not
8122 yet initialized at this point if a forward declaration is involved.
8123
8124 It was actually useful for anonymous pointed-to types to ensure
8125 that the sizes evaluation dominates every possible later use of the
8126 values. Restricting to such types here would be safe since there
8127 is no possible forward declaration around, but would introduce an
8128 undesirable middle-end semantic to anonymity. We then defer to
8129 front-ends the responsibility of ensuring that the sizes are
8130 evaluated both early and late enough, e.g. by attaching artificial
8131 type declarations to the tree. */
8132 break;
8133
8134 default:
8135 break;
8136 }
8137
8138 gimplify_one_sizepos (&TYPE_SIZE (type), list_p);
8139 gimplify_one_sizepos (&TYPE_SIZE_UNIT (type), list_p);
8140
8141 for (t = TYPE_NEXT_VARIANT (type); t; t = TYPE_NEXT_VARIANT (t))
8142 {
8143 TYPE_SIZE (t) = TYPE_SIZE (type);
8144 TYPE_SIZE_UNIT (t) = TYPE_SIZE_UNIT (type);
8145 TYPE_SIZES_GIMPLIFIED (t) = 1;
8146 }
8147 }
8148
8149 /* A subroutine of gimplify_type_sizes to make sure that *EXPR_P,
8150 a size or position, has had all of its SAVE_EXPRs evaluated.
8151 We add any required statements to *STMT_P. */
8152
8153 void
8154 gimplify_one_sizepos (tree *expr_p, gimple_seq *stmt_p)
8155 {
8156 tree expr = *expr_p;
8157
8158 /* We don't do anything if the value isn't there, is constant, or contains
8159 A PLACEHOLDER_EXPR. We also don't want to do anything if it's already
8160 a VAR_DECL. If it's a VAR_DECL from another function, the gimplifier
8161 will want to replace it with a new variable, but that will cause problems
8162 if this type is from outside the function. It's OK to have that here. */
8163 if (is_gimple_sizepos (expr))
8164 return;
8165
8166 *expr_p = unshare_expr (expr);
8167
8168 gimplify_expr (expr_p, stmt_p, NULL, is_gimple_val, fb_rvalue);
8169 }
8170
8171 /* Gimplify the body of statements of FNDECL and return a GIMPLE_BIND node
8172 containing the sequence of corresponding GIMPLE statements. If DO_PARMS
8173 is true, also gimplify the parameters. */
8174
8175 gimple
8176 gimplify_body (tree fndecl, bool do_parms)
8177 {
8178 location_t saved_location = input_location;
8179 gimple_seq parm_stmts, seq;
8180 gimple outer_bind;
8181 struct gimplify_ctx gctx;
8182 struct cgraph_node *cgn;
8183
8184 timevar_push (TV_TREE_GIMPLIFY);
8185
8186 /* Initialize for optimize_insn_for_s{ize,peed}_p possibly called during
8187 gimplification. */
8188 default_rtl_profile ();
8189
8190 gcc_assert (gimplify_ctxp == NULL);
8191 push_gimplify_context (&gctx);
8192
8193 /* Unshare most shared trees in the body and in that of any nested functions.
8194 It would seem we don't have to do this for nested functions because
8195 they are supposed to be output and then the outer function gimplified
8196 first, but the g++ front end doesn't always do it that way. */
8197 unshare_body (fndecl);
8198 unvisit_body (fndecl);
8199
8200 cgn = cgraph_get_node (fndecl);
8201 if (cgn && cgn->origin)
8202 nonlocal_vlas = pointer_set_create ();
8203
8204 /* Make sure input_location isn't set to something weird. */
8205 input_location = DECL_SOURCE_LOCATION (fndecl);
8206
8207 /* Resolve callee-copies. This has to be done before processing
8208 the body so that DECL_VALUE_EXPR gets processed correctly. */
8209 parm_stmts = do_parms ? gimplify_parameters () : NULL;
8210
8211 /* Gimplify the function's body. */
8212 seq = NULL;
8213 gimplify_stmt (&DECL_SAVED_TREE (fndecl), &seq);
8214 outer_bind = gimple_seq_first_stmt (seq);
8215 if (!outer_bind)
8216 {
8217 outer_bind = gimple_build_nop ();
8218 gimplify_seq_add_stmt (&seq, outer_bind);
8219 }
8220
8221 /* The body must contain exactly one statement, a GIMPLE_BIND. If this is
8222 not the case, wrap everything in a GIMPLE_BIND to make it so. */
8223 if (gimple_code (outer_bind) == GIMPLE_BIND
8224 && gimple_seq_first (seq) == gimple_seq_last (seq))
8225 ;
8226 else
8227 outer_bind = gimple_build_bind (NULL_TREE, seq, NULL);
8228
8229 DECL_SAVED_TREE (fndecl) = NULL_TREE;
8230
8231 /* If we had callee-copies statements, insert them at the beginning
8232 of the function and clear DECL_VALUE_EXPR_P on the parameters. */
8233 if (!gimple_seq_empty_p (parm_stmts))
8234 {
8235 tree parm;
8236
8237 gimplify_seq_add_seq (&parm_stmts, gimple_bind_body (outer_bind));
8238 gimple_bind_set_body (outer_bind, parm_stmts);
8239
8240 for (parm = DECL_ARGUMENTS (current_function_decl);
8241 parm; parm = DECL_CHAIN (parm))
8242 if (DECL_HAS_VALUE_EXPR_P (parm))
8243 {
8244 DECL_HAS_VALUE_EXPR_P (parm) = 0;
8245 DECL_IGNORED_P (parm) = 0;
8246 }
8247 }
8248
8249 if (nonlocal_vlas)
8250 {
8251 pointer_set_destroy (nonlocal_vlas);
8252 nonlocal_vlas = NULL;
8253 }
8254
8255 pop_gimplify_context (outer_bind);
8256 gcc_assert (gimplify_ctxp == NULL);
8257
8258 #ifdef ENABLE_CHECKING
8259 if (!seen_error ())
8260 verify_gimple_in_seq (gimple_bind_body (outer_bind));
8261 #endif
8262
8263 timevar_pop (TV_TREE_GIMPLIFY);
8264 input_location = saved_location;
8265
8266 return outer_bind;
8267 }
8268
8269 typedef char *char_p; /* For DEF_VEC_P. */
8270
8271 /* Return whether we should exclude FNDECL from instrumentation. */
8272
8273 static bool
8274 flag_instrument_functions_exclude_p (tree fndecl)
8275 {
8276 vec<char_p> *v;
8277
8278 v = (vec<char_p> *) flag_instrument_functions_exclude_functions;
8279 if (v && v->length () > 0)
8280 {
8281 const char *name;
8282 int i;
8283 char *s;
8284
8285 name = lang_hooks.decl_printable_name (fndecl, 0);
8286 FOR_EACH_VEC_ELT (*v, i, s)
8287 if (strstr (name, s) != NULL)
8288 return true;
8289 }
8290
8291 v = (vec<char_p> *) flag_instrument_functions_exclude_files;
8292 if (v && v->length () > 0)
8293 {
8294 const char *name;
8295 int i;
8296 char *s;
8297
8298 name = DECL_SOURCE_FILE (fndecl);
8299 FOR_EACH_VEC_ELT (*v, i, s)
8300 if (strstr (name, s) != NULL)
8301 return true;
8302 }
8303
8304 return false;
8305 }
8306
8307 /* Entry point to the gimplification pass. FNDECL is the FUNCTION_DECL
8308 node for the function we want to gimplify.
8309
8310 Return the sequence of GIMPLE statements corresponding to the body
8311 of FNDECL. */
8312
8313 void
8314 gimplify_function_tree (tree fndecl)
8315 {
8316 tree parm, ret;
8317 gimple_seq seq;
8318 gimple bind;
8319
8320 gcc_assert (!gimple_body (fndecl));
8321
8322 if (DECL_STRUCT_FUNCTION (fndecl))
8323 push_cfun (DECL_STRUCT_FUNCTION (fndecl));
8324 else
8325 push_struct_function (fndecl);
8326
8327 for (parm = DECL_ARGUMENTS (fndecl); parm ; parm = DECL_CHAIN (parm))
8328 {
8329 /* Preliminarily mark non-addressed complex variables as eligible
8330 for promotion to gimple registers. We'll transform their uses
8331 as we find them. */
8332 if ((TREE_CODE (TREE_TYPE (parm)) == COMPLEX_TYPE
8333 || TREE_CODE (TREE_TYPE (parm)) == VECTOR_TYPE)
8334 && !TREE_THIS_VOLATILE (parm)
8335 && !needs_to_live_in_memory (parm))
8336 DECL_GIMPLE_REG_P (parm) = 1;
8337 }
8338
8339 ret = DECL_RESULT (fndecl);
8340 if ((TREE_CODE (TREE_TYPE (ret)) == COMPLEX_TYPE
8341 || TREE_CODE (TREE_TYPE (ret)) == VECTOR_TYPE)
8342 && !needs_to_live_in_memory (ret))
8343 DECL_GIMPLE_REG_P (ret) = 1;
8344
8345 bind = gimplify_body (fndecl, true);
8346
8347 /* The tree body of the function is no longer needed, replace it
8348 with the new GIMPLE body. */
8349 seq = NULL;
8350 gimple_seq_add_stmt (&seq, bind);
8351 gimple_set_body (fndecl, seq);
8352
8353 /* If we're instrumenting function entry/exit, then prepend the call to
8354 the entry hook and wrap the whole function in a TRY_FINALLY_EXPR to
8355 catch the exit hook. */
8356 /* ??? Add some way to ignore exceptions for this TFE. */
8357 if (flag_instrument_function_entry_exit
8358 && !DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (fndecl)
8359 && !flag_instrument_functions_exclude_p (fndecl))
8360 {
8361 tree x;
8362 gimple new_bind;
8363 gimple tf;
8364 gimple_seq cleanup = NULL, body = NULL;
8365 tree tmp_var;
8366 gimple call;
8367
8368 x = builtin_decl_implicit (BUILT_IN_RETURN_ADDRESS);
8369 call = gimple_build_call (x, 1, integer_zero_node);
8370 tmp_var = create_tmp_var (ptr_type_node, "return_addr");
8371 gimple_call_set_lhs (call, tmp_var);
8372 gimplify_seq_add_stmt (&cleanup, call);
8373 x = builtin_decl_implicit (BUILT_IN_PROFILE_FUNC_EXIT);
8374 call = gimple_build_call (x, 2,
8375 build_fold_addr_expr (current_function_decl),
8376 tmp_var);
8377 gimplify_seq_add_stmt (&cleanup, call);
8378 tf = gimple_build_try (seq, cleanup, GIMPLE_TRY_FINALLY);
8379
8380 x = builtin_decl_implicit (BUILT_IN_RETURN_ADDRESS);
8381 call = gimple_build_call (x, 1, integer_zero_node);
8382 tmp_var = create_tmp_var (ptr_type_node, "return_addr");
8383 gimple_call_set_lhs (call, tmp_var);
8384 gimplify_seq_add_stmt (&body, call);
8385 x = builtin_decl_implicit (BUILT_IN_PROFILE_FUNC_ENTER);
8386 call = gimple_build_call (x, 2,
8387 build_fold_addr_expr (current_function_decl),
8388 tmp_var);
8389 gimplify_seq_add_stmt (&body, call);
8390 gimplify_seq_add_stmt (&body, tf);
8391 new_bind = gimple_build_bind (NULL, body, gimple_bind_block (bind));
8392 /* Clear the block for BIND, since it is no longer directly inside
8393 the function, but within a try block. */
8394 gimple_bind_set_block (bind, NULL);
8395
8396 /* Replace the current function body with the body
8397 wrapped in the try/finally TF. */
8398 seq = NULL;
8399 gimple_seq_add_stmt (&seq, new_bind);
8400 gimple_set_body (fndecl, seq);
8401 }
8402
8403 DECL_SAVED_TREE (fndecl) = NULL_TREE;
8404 cfun->curr_properties = PROP_gimple_any;
8405
8406 pop_cfun ();
8407 }
8408
8409 /* Some transformations like inlining may invalidate the GIMPLE form
8410 for operands. This function traverses all the operands in STMT and
8411 gimplifies anything that is not a valid gimple operand. Any new
8412 GIMPLE statements are inserted before *GSI_P. */
8413
8414 void
8415 gimple_regimplify_operands (gimple stmt, gimple_stmt_iterator *gsi_p)
8416 {
8417 size_t i, num_ops;
8418 tree lhs;
8419 gimple_seq pre = NULL;
8420 gimple post_stmt = NULL;
8421 struct gimplify_ctx gctx;
8422
8423 push_gimplify_context (&gctx);
8424 gimplify_ctxp->into_ssa = gimple_in_ssa_p (cfun);
8425
8426 switch (gimple_code (stmt))
8427 {
8428 case GIMPLE_COND:
8429 gimplify_expr (gimple_cond_lhs_ptr (stmt), &pre, NULL,
8430 is_gimple_val, fb_rvalue);
8431 gimplify_expr (gimple_cond_rhs_ptr (stmt), &pre, NULL,
8432 is_gimple_val, fb_rvalue);
8433 break;
8434 case GIMPLE_SWITCH:
8435 gimplify_expr (gimple_switch_index_ptr (stmt), &pre, NULL,
8436 is_gimple_val, fb_rvalue);
8437 break;
8438 case GIMPLE_OMP_ATOMIC_LOAD:
8439 gimplify_expr (gimple_omp_atomic_load_rhs_ptr (stmt), &pre, NULL,
8440 is_gimple_val, fb_rvalue);
8441 break;
8442 case GIMPLE_ASM:
8443 {
8444 size_t i, noutputs = gimple_asm_noutputs (stmt);
8445 const char *constraint, **oconstraints;
8446 bool allows_mem, allows_reg, is_inout;
8447
8448 oconstraints
8449 = (const char **) alloca ((noutputs) * sizeof (const char *));
8450 for (i = 0; i < noutputs; i++)
8451 {
8452 tree op = gimple_asm_output_op (stmt, i);
8453 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
8454 oconstraints[i] = constraint;
8455 parse_output_constraint (&constraint, i, 0, 0, &allows_mem,
8456 &allows_reg, &is_inout);
8457 gimplify_expr (&TREE_VALUE (op), &pre, NULL,
8458 is_inout ? is_gimple_min_lval : is_gimple_lvalue,
8459 fb_lvalue | fb_mayfail);
8460 }
8461 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
8462 {
8463 tree op = gimple_asm_input_op (stmt, i);
8464 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
8465 parse_input_constraint (&constraint, 0, 0, noutputs, 0,
8466 oconstraints, &allows_mem, &allows_reg);
8467 if (TREE_ADDRESSABLE (TREE_TYPE (TREE_VALUE (op))) && allows_mem)
8468 allows_reg = 0;
8469 if (!allows_reg && allows_mem)
8470 gimplify_expr (&TREE_VALUE (op), &pre, NULL,
8471 is_gimple_lvalue, fb_lvalue | fb_mayfail);
8472 else
8473 gimplify_expr (&TREE_VALUE (op), &pre, NULL,
8474 is_gimple_asm_val, fb_rvalue);
8475 }
8476 }
8477 break;
8478 default:
8479 /* NOTE: We start gimplifying operands from last to first to
8480 make sure that side-effects on the RHS of calls, assignments
8481 and ASMs are executed before the LHS. The ordering is not
8482 important for other statements. */
8483 num_ops = gimple_num_ops (stmt);
8484 for (i = num_ops; i > 0; i--)
8485 {
8486 tree op = gimple_op (stmt, i - 1);
8487 if (op == NULL_TREE)
8488 continue;
8489 if (i == 1 && (is_gimple_call (stmt) || is_gimple_assign (stmt)))
8490 gimplify_expr (&op, &pre, NULL, is_gimple_lvalue, fb_lvalue);
8491 else if (i == 2
8492 && is_gimple_assign (stmt)
8493 && num_ops == 2
8494 && get_gimple_rhs_class (gimple_expr_code (stmt))
8495 == GIMPLE_SINGLE_RHS)
8496 gimplify_expr (&op, &pre, NULL,
8497 rhs_predicate_for (gimple_assign_lhs (stmt)),
8498 fb_rvalue);
8499 else if (i == 2 && is_gimple_call (stmt))
8500 {
8501 if (TREE_CODE (op) == FUNCTION_DECL)
8502 continue;
8503 gimplify_expr (&op, &pre, NULL, is_gimple_call_addr, fb_rvalue);
8504 }
8505 else
8506 gimplify_expr (&op, &pre, NULL, is_gimple_val, fb_rvalue);
8507 gimple_set_op (stmt, i - 1, op);
8508 }
8509
8510 lhs = gimple_get_lhs (stmt);
8511 /* If the LHS changed it in a way that requires a simple RHS,
8512 create temporary. */
8513 if (lhs && !is_gimple_reg (lhs))
8514 {
8515 bool need_temp = false;
8516
8517 if (is_gimple_assign (stmt)
8518 && num_ops == 2
8519 && get_gimple_rhs_class (gimple_expr_code (stmt))
8520 == GIMPLE_SINGLE_RHS)
8521 gimplify_expr (gimple_assign_rhs1_ptr (stmt), &pre, NULL,
8522 rhs_predicate_for (gimple_assign_lhs (stmt)),
8523 fb_rvalue);
8524 else if (is_gimple_reg (lhs))
8525 {
8526 if (is_gimple_reg_type (TREE_TYPE (lhs)))
8527 {
8528 if (is_gimple_call (stmt))
8529 {
8530 i = gimple_call_flags (stmt);
8531 if ((i & ECF_LOOPING_CONST_OR_PURE)
8532 || !(i & (ECF_CONST | ECF_PURE)))
8533 need_temp = true;
8534 }
8535 if (stmt_can_throw_internal (stmt))
8536 need_temp = true;
8537 }
8538 }
8539 else
8540 {
8541 if (is_gimple_reg_type (TREE_TYPE (lhs)))
8542 need_temp = true;
8543 else if (TYPE_MODE (TREE_TYPE (lhs)) != BLKmode)
8544 {
8545 if (is_gimple_call (stmt))
8546 {
8547 tree fndecl = gimple_call_fndecl (stmt);
8548
8549 if (!aggregate_value_p (TREE_TYPE (lhs), fndecl)
8550 && !(fndecl && DECL_RESULT (fndecl)
8551 && DECL_BY_REFERENCE (DECL_RESULT (fndecl))))
8552 need_temp = true;
8553 }
8554 else
8555 need_temp = true;
8556 }
8557 }
8558 if (need_temp)
8559 {
8560 tree temp = create_tmp_reg (TREE_TYPE (lhs), NULL);
8561 if (gimple_in_ssa_p (cfun))
8562 temp = make_ssa_name (temp, NULL);
8563 gimple_set_lhs (stmt, temp);
8564 post_stmt = gimple_build_assign (lhs, temp);
8565 if (TREE_CODE (lhs) == SSA_NAME)
8566 SSA_NAME_DEF_STMT (lhs) = post_stmt;
8567 }
8568 }
8569 break;
8570 }
8571
8572 if (!gimple_seq_empty_p (pre))
8573 gsi_insert_seq_before (gsi_p, pre, GSI_SAME_STMT);
8574 if (post_stmt)
8575 gsi_insert_after (gsi_p, post_stmt, GSI_NEW_STMT);
8576
8577 pop_gimplify_context (NULL);
8578 }
8579
8580 /* Expand EXPR to list of gimple statements STMTS. GIMPLE_TEST_F specifies
8581 the predicate that will hold for the result. If VAR is not NULL, make the
8582 base variable of the final destination be VAR if suitable. */
8583
8584 tree
8585 force_gimple_operand_1 (tree expr, gimple_seq *stmts,
8586 gimple_predicate gimple_test_f, tree var)
8587 {
8588 enum gimplify_status ret;
8589 struct gimplify_ctx gctx;
8590
8591 *stmts = NULL;
8592
8593 /* gimple_test_f might be more strict than is_gimple_val, make
8594 sure we pass both. Just checking gimple_test_f doesn't work
8595 because most gimple predicates do not work recursively. */
8596 if (is_gimple_val (expr)
8597 && (*gimple_test_f) (expr))
8598 return expr;
8599
8600 push_gimplify_context (&gctx);
8601 gimplify_ctxp->into_ssa = gimple_in_ssa_p (cfun);
8602 gimplify_ctxp->allow_rhs_cond_expr = true;
8603
8604 if (var)
8605 {
8606 if (gimplify_ctxp->into_ssa
8607 && is_gimple_reg (var))
8608 var = make_ssa_name (var, NULL);
8609 expr = build2 (MODIFY_EXPR, TREE_TYPE (var), var, expr);
8610 }
8611
8612 if (TREE_CODE (expr) != MODIFY_EXPR
8613 && TREE_TYPE (expr) == void_type_node)
8614 {
8615 gimplify_and_add (expr, stmts);
8616 expr = NULL_TREE;
8617 }
8618 else
8619 {
8620 ret = gimplify_expr (&expr, stmts, NULL, gimple_test_f, fb_rvalue);
8621 gcc_assert (ret != GS_ERROR);
8622 }
8623
8624 pop_gimplify_context (NULL);
8625
8626 return expr;
8627 }
8628
8629 /* Expand EXPR to list of gimple statements STMTS. If SIMPLE is true,
8630 force the result to be either ssa_name or an invariant, otherwise
8631 just force it to be a rhs expression. If VAR is not NULL, make the
8632 base variable of the final destination be VAR if suitable. */
8633
8634 tree
8635 force_gimple_operand (tree expr, gimple_seq *stmts, bool simple, tree var)
8636 {
8637 return force_gimple_operand_1 (expr, stmts,
8638 simple ? is_gimple_val : is_gimple_reg_rhs,
8639 var);
8640 }
8641
8642 /* Invoke force_gimple_operand_1 for EXPR with parameters GIMPLE_TEST_F
8643 and VAR. If some statements are produced, emits them at GSI.
8644 If BEFORE is true. the statements are appended before GSI, otherwise
8645 they are appended after it. M specifies the way GSI moves after
8646 insertion (GSI_SAME_STMT or GSI_CONTINUE_LINKING are the usual values). */
8647
8648 tree
8649 force_gimple_operand_gsi_1 (gimple_stmt_iterator *gsi, tree expr,
8650 gimple_predicate gimple_test_f,
8651 tree var, bool before,
8652 enum gsi_iterator_update m)
8653 {
8654 gimple_seq stmts;
8655
8656 expr = force_gimple_operand_1 (expr, &stmts, gimple_test_f, var);
8657
8658 if (!gimple_seq_empty_p (stmts))
8659 {
8660 if (before)
8661 gsi_insert_seq_before (gsi, stmts, m);
8662 else
8663 gsi_insert_seq_after (gsi, stmts, m);
8664 }
8665
8666 return expr;
8667 }
8668
8669 /* Invoke force_gimple_operand_1 for EXPR with parameter VAR.
8670 If SIMPLE is true, force the result to be either ssa_name or an invariant,
8671 otherwise just force it to be a rhs expression. If some statements are
8672 produced, emits them at GSI. If BEFORE is true, the statements are
8673 appended before GSI, otherwise they are appended after it. M specifies
8674 the way GSI moves after insertion (GSI_SAME_STMT or GSI_CONTINUE_LINKING
8675 are the usual values). */
8676
8677 tree
8678 force_gimple_operand_gsi (gimple_stmt_iterator *gsi, tree expr,
8679 bool simple_p, tree var, bool before,
8680 enum gsi_iterator_update m)
8681 {
8682 return force_gimple_operand_gsi_1 (gsi, expr,
8683 simple_p
8684 ? is_gimple_val : is_gimple_reg_rhs,
8685 var, before, m);
8686 }
8687
8688
8689 #include "gt-gimplify.h"