1 /* intrinsics.cc -- D language compiler intrinsics.
2 Copyright (C) 2006-2020 Free Software Foundation, Inc.
4 GCC is free software; you can redistribute it and/or modify
5 it under the terms of the GNU General Public License as published by
6 the Free Software Foundation; either version 3, or (at your option)
9 GCC is distributed in the hope that it will be useful,
10 but WITHOUT ANY WARRANTY; without even the implied warranty of
11 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 GNU General Public License for more details.
14 You should have received a copy of the GNU General Public License
15 along with GCC; see the file COPYING3. If not see
16 <http://www.gnu.org/licenses/>. */
20 #include "coretypes.h"
22 #include "dmd/declaration.h"
23 #include "dmd/identifier.h"
24 #include "dmd/mangle.h"
25 #include "dmd/mangle.h"
26 #include "dmd/module.h"
27 #include "dmd/template.h"
32 #include "fold-const.h"
33 #include "stringpool.h"
39 /* An internal struct used to hold information on D intrinsics. */
43 /* The DECL_INTRINSIC_CODE of this decl. */
46 /* The DECL_FUNCTION_CODE of this decl, if it directly maps to any. */
47 built_in_function built_in
;
49 /* The name of the intrinsic. */
52 /* The module where the intrinsic is located. */
55 /* The mangled signature decoration of the intrinsic. */
58 /* True if the intrinsic is only handled in CTFE. */
62 static const intrinsic_decl intrinsic_decls
[] =
64 #define DEF_D_INTRINSIC(CODE, BUILTIN, NAME, MODULE, DECO, CTFE) \
65 { INTRINSIC_ ## CODE, BUILT_IN_ ## BUILTIN, NAME, MODULE, DECO, CTFE },
67 #include "intrinsics.def"
69 #undef DEF_D_INTRINSIC
72 /* Checks if DECL is an intrinsic or run time library function that requires
73 special processing. Sets DECL_INTRINSIC_CODE so it can be identified
74 later in maybe_expand_intrinsic. */
77 maybe_set_intrinsic (FuncDeclaration
*decl
)
79 if (!decl
->ident
|| decl
->builtin
!= BUILTINunknown
)
82 /* The builtin flag is updated only if we can evaluate the intrinsic
83 at compile-time. Such as the math or bitop intrinsics. */
84 decl
->builtin
= BUILTINno
;
86 /* Check if it's a compiler intrinsic. We only require that any
87 internally recognised intrinsics are declared in a module with
88 an explicit module declaration. */
89 Module
*m
= decl
->getModule ();
94 TemplateInstance
*ti
= decl
->isInstantiated ();
95 TemplateDeclaration
*td
= ti
? ti
->tempdecl
->isTemplateDeclaration () : NULL
;
97 const char *tname
= decl
->ident
->toChars ();
98 const char *tmodule
= m
->md
->toChars ();
99 const char *tdeco
= (td
== NULL
) ? decl
->type
->deco
: NULL
;
101 /* Look through all D intrinsics. */
102 for (size_t i
= 0; i
< (int) INTRINSIC_LAST
; i
++)
104 if (!intrinsic_decls
[i
].name
)
107 if (strcmp (intrinsic_decls
[i
].name
, tname
) != 0
108 || strcmp (intrinsic_decls
[i
].module
, tmodule
) != 0)
111 /* Instantiated functions would have the wrong type deco, get it from the
112 template member instead. */
115 if (!td
|| !td
->onemember
)
118 FuncDeclaration
*fd
= td
->onemember
->isFuncDeclaration ();
123 mangleToBuffer (fd
->type
, &buf
);
124 tdeco
= buf
.extractChars ();
127 /* Matching the type deco may be a bit too strict, as it means that all
128 function attributes that end up in the signature must be kept aligned
129 between the compiler and library declaration. */
130 if (strcmp (intrinsic_decls
[i
].deco
, tdeco
) == 0)
132 intrinsic_code code
= intrinsic_decls
[i
].code
;
134 if (decl
->csym
== NULL
)
135 get_symbol_decl (decl
);
137 /* If there is no function body, then the implementation is always
138 provided by the compiler. */
140 set_decl_built_in_function (decl
->csym
, BUILT_IN_FRONTEND
, code
);
142 /* Infer whether the intrinsic can be used for CTFE, let the
143 front-end know that it can be evaluated at compile-time. */
146 case INTRINSIC_VA_ARG
:
147 case INTRINSIC_C_VA_ARG
:
148 case INTRINSIC_VASTART
:
150 case INTRINSIC_ADDSL
:
152 case INTRINSIC_ADDUL
:
154 case INTRINSIC_SUBSL
:
156 case INTRINSIC_SUBUL
:
158 case INTRINSIC_MULSL
:
160 case INTRINSIC_MULUI
:
161 case INTRINSIC_MULUL
:
163 case INTRINSIC_NEGSL
:
164 case INTRINSIC_VLOAD8
:
165 case INTRINSIC_VLOAD16
:
166 case INTRINSIC_VLOAD32
:
167 case INTRINSIC_VLOAD64
:
168 case INTRINSIC_VSTORE8
:
169 case INTRINSIC_VSTORE16
:
170 case INTRINSIC_VSTORE32
:
171 case INTRINSIC_VSTORE64
:
176 /* Check that this overload of pow() is has an equivalent
177 built-in function. It could be `int pow(int, int)'. */
178 tree rettype
= TREE_TYPE (TREE_TYPE (decl
->csym
));
179 if (mathfn_built_in (rettype
, BUILT_IN_POW
) != NULL_TREE
)
180 decl
->builtin
= BUILTINyes
;
185 decl
->builtin
= BUILTINyes
;
189 /* The intrinsic was marked as CTFE-only. */
190 if (intrinsic_decls
[i
].ctfeonly
)
191 DECL_BUILT_IN_CTFE (decl
->csym
) = 1;
193 DECL_INTRINSIC_CODE (decl
->csym
) = code
;
199 /* Construct a function call to the built-in function CODE, N is the number of
200 arguments, and the `...' parameters are the argument expressions.
201 The original call expression is held in CALLEXP. */
204 call_builtin_fn (tree callexp
, built_in_function code
, int n
, ...)
206 tree
*argarray
= XALLOCAVEC (tree
, n
);
210 for (int i
= 0; i
< n
; i
++)
211 argarray
[i
] = va_arg (ap
, tree
);
214 tree exp
= build_call_expr_loc_array (EXPR_LOCATION (callexp
),
215 builtin_decl_explicit (code
),
217 return convert (TREE_TYPE (callexp
), fold (exp
));
220 /* Expand a front-end instrinsic call to bsf(). This takes one argument,
221 the signature to which can be either:
226 This scans all bits in the given argument starting with the first,
227 returning the bit number of the first bit set. The original call
228 expression is held in CALLEXP. */
231 expand_intrinsic_bsf (tree callexp
)
233 /* The bsr() intrinsic gets turned into __builtin_ctz(arg).
234 The return value is supposed to be undefined if arg is zero. */
235 tree arg
= CALL_EXPR_ARG (callexp
, 0);
236 int argsize
= TYPE_PRECISION (TREE_TYPE (arg
));
238 /* Which variant of __builtin_ctz* should we call? */
239 built_in_function code
= (argsize
<= INT_TYPE_SIZE
) ? BUILT_IN_CTZ
240 : (argsize
<= LONG_TYPE_SIZE
) ? BUILT_IN_CTZL
241 : (argsize
<= LONG_LONG_TYPE_SIZE
) ? BUILT_IN_CTZLL
244 gcc_assert (code
!= END_BUILTINS
);
246 return call_builtin_fn (callexp
, code
, 1, arg
);
249 /* Expand a front-end instrinsic call to bsr(). This takes one argument,
250 the signature to which can be either:
255 This scans all bits in the given argument from the most significant bit
256 to the least significant, returning the bit number of the first bit set.
257 The original call expression is held in CALLEXP. */
260 expand_intrinsic_bsr (tree callexp
)
262 /* The bsr() intrinsic gets turned into (size - 1) - __builtin_clz(arg).
263 The return value is supposed to be undefined if arg is zero. */
264 tree arg
= CALL_EXPR_ARG (callexp
, 0);
265 tree type
= TREE_TYPE (arg
);
266 int argsize
= TYPE_PRECISION (type
);
268 /* Which variant of __builtin_clz* should we call? */
269 built_in_function code
= (argsize
<= INT_TYPE_SIZE
) ? BUILT_IN_CLZ
270 : (argsize
<= LONG_TYPE_SIZE
) ? BUILT_IN_CLZL
271 : (argsize
<= LONG_LONG_TYPE_SIZE
) ? BUILT_IN_CLZLL
274 gcc_assert (code
!= END_BUILTINS
);
276 tree result
= call_builtin_fn (callexp
, code
, 1, arg
);
278 /* Handle int -> long conversions. */
279 if (TREE_TYPE (result
) != type
)
280 result
= fold_convert (type
, result
);
282 result
= fold_build2 (MINUS_EXPR
, type
,
283 build_integer_cst (argsize
- 1, type
), result
);
284 return fold_convert (TREE_TYPE (callexp
), result
);
287 /* Expand a front-end intrinsic call to INTRINSIC, which is either a call to
288 bt(), btc(), btr(), or bts(). These intrinsics expect to take two arguments,
289 the signature to which is:
291 int bt (size_t* ptr, size_t bitnum);
293 All intrinsics test if a bit is set and return the result of that condition.
294 Variants of `bt' will then update that bit. `btc' compliments the bit, `bts'
295 sets the bit, and `btr' resets the bit. The original call expression is
299 expand_intrinsic_bt (intrinsic_code intrinsic
, tree callexp
)
301 tree ptr
= CALL_EXPR_ARG (callexp
, 0);
302 tree bitnum
= CALL_EXPR_ARG (callexp
, 1);
303 tree type
= TREE_TYPE (TREE_TYPE (ptr
));
305 /* size_t bitsize = sizeof(*ptr) * BITS_PER_UNIT; */
306 tree bitsize
= fold_convert (type
, TYPE_SIZE (type
));
308 /* ptr[bitnum / bitsize] */
309 ptr
= build_array_index (ptr
, fold_build2 (TRUNC_DIV_EXPR
, type
,
311 ptr
= indirect_ref (type
, ptr
);
313 /* mask = 1 << (bitnum % bitsize); */
314 bitnum
= fold_build2 (TRUNC_MOD_EXPR
, type
, bitnum
, bitsize
);
315 bitnum
= fold_build2 (LSHIFT_EXPR
, type
, size_one_node
, bitnum
);
317 /* cond = ptr[bitnum / size] & mask; */
318 tree cond
= fold_build2 (BIT_AND_EXPR
, type
, ptr
, bitnum
);
321 cond
= build_condition (TREE_TYPE (callexp
), d_truthvalue_conversion (cond
),
322 integer_minus_one_node
, integer_zero_node
);
324 /* Update the bit as needed, only testing the bit for bt(). */
334 case INTRINSIC_BTC64
:
339 case INTRINSIC_BTR64
:
340 bitnum
= fold_build1 (BIT_NOT_EXPR
, TREE_TYPE (bitnum
), bitnum
);
345 case INTRINSIC_BTS64
:
353 /* ptr[bitnum / size] op= mask; */
354 if (intrinsic
== INTRINSIC_BTR
)
355 bitnum
= fold_build1 (BIT_NOT_EXPR
, TREE_TYPE (bitnum
), bitnum
);
357 ptr
= modify_expr (ptr
, fold_build2 (code
, TREE_TYPE (ptr
), ptr
, bitnum
));
359 /* Store the condition result in a temporary, and return expressions in
360 correct order of evaluation. */
361 tree tmp
= build_local_temp (TREE_TYPE (callexp
));
362 cond
= modify_expr (tmp
, cond
);
364 return compound_expr (cond
, compound_expr (ptr
, tmp
));
367 /* Expand a front-end intrinsic call to popcnt(). This takes one argument, the
368 signature to which can be either:
370 int popcnt (uint arg);
371 int popcnt (ulong arg);
373 Calculates the number of set bits in an integer. The original call
374 expression is held in CALLEXP. */
377 expand_intrinsic_popcnt (tree callexp
)
379 tree arg
= CALL_EXPR_ARG (callexp
, 0);
380 int argsize
= TYPE_PRECISION (TREE_TYPE (arg
));
382 /* Which variant of __builtin_popcount* should we call? */
383 built_in_function code
= (argsize
<= INT_TYPE_SIZE
) ? BUILT_IN_POPCOUNT
384 : (argsize
<= LONG_TYPE_SIZE
) ? BUILT_IN_POPCOUNTL
385 : (argsize
<= LONG_LONG_TYPE_SIZE
) ? BUILT_IN_POPCOUNTLL
388 gcc_assert (code
!= END_BUILTINS
);
390 return call_builtin_fn (callexp
, code
, 1, arg
);
393 /* Expand a front-end intrinsic call to INTRINSIC, which is either a call to
394 rol() or ror(). These intrinsics expect to take one or two arguments,
395 the signature to which can be either:
397 T rol(T) (const T value, const uint count);
398 T rol(uint count, T) (const T value);
399 T ror(T) (const T value, const uint count);
400 T ror(uint count, T) (const T value);
402 This bitwise rotates VALUE left or right by COUNT bit positions. */
405 expand_intrinsic_rotate (intrinsic_code intrinsic
, tree callexp
)
407 tree type
= TREE_TYPE (callexp
);
408 tree value
= CALL_EXPR_ARG (callexp
, 0);
412 /* Get the equivalent tree code for the intrinsic. */
413 if (intrinsic
== INTRINSIC_ROL
|| intrinsic
== INTRINSIC_ROL_TIARG
)
415 else if (intrinsic
== INTRINSIC_ROR
|| intrinsic
== INTRINSIC_ROR_TIARG
)
420 /* Get the COUNT parameter. Either from the call expression arguments or the
421 template instantiation arguments. */
422 if (intrinsic
== INTRINSIC_ROL
|| intrinsic
== INTRINSIC_ROR
)
423 count
= CALL_EXPR_ARG (callexp
, 1);
426 tree callee
= CALL_EXPR_FN (callexp
);
428 if (TREE_CODE (callee
) == ADDR_EXPR
)
429 callee
= TREE_OPERAND (callee
, 0);
431 /* Retrieve from the encoded template instantation. */
432 TemplateInstance
*ti
= DECL_LANG_FRONTEND (callee
)->isInstantiated ();
433 gcc_assert (ti
&& ti
->tiargs
&& ti
->tiargs
->length
== 2);
435 Expression
*e
= isExpression ((*ti
->tiargs
)[0]);
436 gcc_assert (e
&& e
->op
== TOKint64
);
437 count
= build_expr (e
, true);
440 return fold_build2 (code
, type
, value
, count
);
443 /* Expand a front-end intrinsic call to copysign(). This takes two arguments,
444 the signature to which can be either:
446 float copysign (T to, float from);
447 double copysign (T to, double from);
448 real copysign (T to, real from);
450 This computes a value composed of TO with the sign bit of FROM. The original
451 call expression is held in CALLEXP. */
454 expand_intrinsic_copysign (tree callexp
)
456 tree to
= CALL_EXPR_ARG (callexp
, 0);
457 tree from
= CALL_EXPR_ARG (callexp
, 1);
458 tree type
= TREE_TYPE (to
);
460 /* Convert parameters to the same type. Prefer the first parameter unless it
461 is an integral type. */
462 if (INTEGRAL_TYPE_P (type
))
464 to
= fold_convert (TREE_TYPE (from
), to
);
465 type
= TREE_TYPE (to
);
468 from
= fold_convert (type
, from
);
470 /* Which variant of __builtin_copysign* should we call? */
471 tree builtin
= mathfn_built_in (type
, BUILT_IN_COPYSIGN
);
472 gcc_assert (builtin
!= NULL_TREE
);
474 return call_builtin_fn (callexp
, DECL_FUNCTION_CODE (builtin
), 2,
478 /* Expand a front-end intrinsic call to pow(). This takes two arguments, the
479 signature to which can be either:
481 float pow (float base, T exponent);
482 double pow (double base, T exponent);
483 real pow (real base, T exponent);
485 This computes the value of BASE raised to the power of EXPONENT.
486 The original call expression is held in CALLEXP. */
489 expand_intrinsic_pow (tree callexp
)
491 tree base
= CALL_EXPR_ARG (callexp
, 0);
492 tree exponent
= CALL_EXPR_ARG (callexp
, 1);
493 tree exptype
= TREE_TYPE (exponent
);
495 /* Which variant of __builtin_pow* should we call? */
496 built_in_function code
= SCALAR_FLOAT_TYPE_P (exptype
) ? BUILT_IN_POW
497 : INTEGRAL_TYPE_P (exptype
) ? BUILT_IN_POWI
499 gcc_assert (code
!= END_BUILTINS
);
501 tree builtin
= mathfn_built_in (TREE_TYPE (base
), code
);
502 gcc_assert (builtin
!= NULL_TREE
);
504 return call_builtin_fn (callexp
, DECL_FUNCTION_CODE (builtin
), 2,
508 /* Expand a front-end intrinsic call to toPrec(). This takes one argument, the
509 signature to which can be either:
511 T toPrec(T)(float f);
512 T toPrec(T)(double f);
515 This rounds the argument F to the precision of the specified floating
516 point type T. The original call expression is held in CALLEXP. */
519 expand_intrinsic_toprec (tree callexp
)
521 tree f
= CALL_EXPR_ARG (callexp
, 0);
522 tree type
= TREE_TYPE (callexp
);
524 return convert (type
, f
);
527 /* Expand a front-end intrinsic call to va_arg(). This takes either one or two
528 arguments, the signature to which can be either:
530 T va_arg(T) (ref va_list ap);
531 void va_arg(T) (va_list ap, ref T parmn);
533 This retrieves the next variadic parameter that is type T from the given
534 va_list. If also given, store the value into parmn, otherwise return it.
535 The original call expression is held in CALLEXP. */
538 expand_intrinsic_vaarg (tree callexp
)
540 tree ap
= CALL_EXPR_ARG (callexp
, 0);
541 tree parmn
= NULL_TREE
;
546 if (call_expr_nargs (callexp
) == 1)
547 type
= TREE_TYPE (callexp
);
550 parmn
= CALL_EXPR_ARG (callexp
, 1);
552 gcc_assert (TREE_CODE (parmn
) == ADDR_EXPR
);
553 parmn
= TREE_OPERAND (parmn
, 0);
554 type
= TREE_TYPE (parmn
);
557 /* (T) VA_ARG_EXP<ap>; */
558 tree exp
= build1 (VA_ARG_EXPR
, type
, ap
);
560 /* parmn = (T) VA_ARG_EXP<ap>; */
561 if (parmn
!= NULL_TREE
)
562 exp
= modify_expr (parmn
, exp
);
567 /* Expand a front-end intrinsic call to va_start(), which takes two arguments,
568 the signature to which is:
570 void va_start(T) (out va_list ap, ref T parmn);
572 This initializes the va_list type, where parmn should be the last named
573 parameter. The original call expression is held in CALLEXP. */
576 expand_intrinsic_vastart (tree callexp
)
578 tree ap
= CALL_EXPR_ARG (callexp
, 0);
579 tree parmn
= CALL_EXPR_ARG (callexp
, 1);
584 /* The va_list argument should already have its address taken. The second
585 argument, however, is inout and that needs to be fixed to prevent a
586 warning. Could be casting, so need to check type too? */
587 gcc_assert (TREE_CODE (ap
) == ADDR_EXPR
&& TREE_CODE (parmn
) == ADDR_EXPR
);
589 /* Assuming nobody tries to change the return type. */
590 parmn
= TREE_OPERAND (parmn
, 0);
592 return call_builtin_fn (callexp
, BUILT_IN_VA_START
, 2, ap
, parmn
);
595 /* Expand a front-end instrinsic call to INTRINSIC, which is either a call to
596 adds(), addu(), subs(), subu(), negs(), muls(), or mulu(). These intrinsics
597 expect to take two or three arguments, the signature to which can be either:
599 int adds (int x, int y, ref bool overflow);
600 long adds (long x, long y, ref bool overflow);
601 int negs (int x, ref bool overflow);
602 long negs (long x, ref bool overflow);
604 This performs an operation on two signed or unsigned integers, checking for
605 overflow. The overflow is sticky, meaning that a sequence of operations
606 can be done and overflow need only be checked at the end. The original call
607 expression is held in CALLEXP. */
610 expand_intrinsic_checkedint (intrinsic_code intrinsic
, tree callexp
)
612 tree type
= TREE_TYPE (callexp
);
618 /* Which variant of *_OVERFLOW should we generate? */
622 case INTRINSIC_ADDSL
:
624 case INTRINSIC_ADDUL
:
625 x
= CALL_EXPR_ARG (callexp
, 0);
626 y
= CALL_EXPR_ARG (callexp
, 1);
627 overflow
= CALL_EXPR_ARG (callexp
, 2);
628 icode
= IFN_ADD_OVERFLOW
;
632 case INTRINSIC_SUBSL
:
634 case INTRINSIC_SUBUL
:
635 x
= CALL_EXPR_ARG (callexp
, 0);
636 y
= CALL_EXPR_ARG (callexp
, 1);
637 overflow
= CALL_EXPR_ARG (callexp
, 2);
638 icode
= IFN_SUB_OVERFLOW
;
642 case INTRINSIC_MULSL
:
644 case INTRINSIC_MULUI
:
645 case INTRINSIC_MULUL
:
646 x
= CALL_EXPR_ARG (callexp
, 0);
647 y
= CALL_EXPR_ARG (callexp
, 1);
648 overflow
= CALL_EXPR_ARG (callexp
, 2);
649 icode
= IFN_MUL_OVERFLOW
;
653 case INTRINSIC_NEGSL
:
654 /* The negs() intrinsic gets turned into SUB_OVERFLOW (0, y). */
655 x
= fold_convert (type
, integer_zero_node
);
656 y
= CALL_EXPR_ARG (callexp
, 0);
657 overflow
= CALL_EXPR_ARG (callexp
, 1);
658 icode
= IFN_SUB_OVERFLOW
;
666 = build_call_expr_internal_loc (EXPR_LOCATION (callexp
), icode
,
667 build_complex_type (type
), 2, x
, y
);
669 STRIP_NOPS (overflow
);
670 overflow
= build_deref (overflow
);
672 /* Assign returned result to overflow parameter, however if overflow is
673 already true, maintain its value. */
674 type
= TREE_TYPE (overflow
);
675 result
= save_expr (result
);
677 tree exp
= fold_build2 (BIT_IOR_EXPR
, type
, overflow
,
678 fold_convert (type
, imaginary_part (result
)));
679 exp
= modify_expr (overflow
, exp
);
681 /* Return the value of result. */
682 return compound_expr (exp
, real_part (result
));
685 /* Expand a front-end instrinsic call to volatileLoad(). This takes one
686 argument, the signature to which can be either:
688 ubyte volatileLoad (ubyte* ptr);
689 ushort volatileLoad (ushort* ptr);
690 uint volatileLoad (uint* ptr);
691 ulong volatileLoad (ulong* ptr);
693 This reads a value from the memory location indicated by ptr. Calls to
694 them are be guaranteed to not be removed (such as during DCE) or reordered
695 in the same thread. The original call expression is held in CALLEXP. */
698 expand_volatile_load (tree callexp
)
700 tree ptr
= CALL_EXPR_ARG (callexp
, 0);
701 tree ptrtype
= TREE_TYPE (ptr
);
702 gcc_assert (POINTER_TYPE_P (ptrtype
));
704 /* (T) *(volatile T *) ptr; */
705 tree type
= build_qualified_type (TREE_TYPE (ptrtype
), TYPE_QUAL_VOLATILE
);
706 tree result
= indirect_ref (type
, ptr
);
707 TREE_THIS_VOLATILE (result
) = 1;
712 /* Expand a front-end instrinsic call to volatileStore(). This takes two
713 arguments, the signature to which can be either:
715 void volatileStore (ubyte* ptr, ubyte value);
716 void volatileStore (ushort* ptr, ushort value);
717 void volatileStore (uint* ptr, uint value);
718 void volatileStore (ulong* ptr, ulong value);
720 This writes a value to the memory location indicated by ptr. Calls to
721 them are be guaranteed to not be removed (such as during DCE) or reordered
722 in the same thread. The original call expression is held in CALLEXP. */
725 expand_volatile_store (tree callexp
)
727 tree ptr
= CALL_EXPR_ARG (callexp
, 0);
728 tree ptrtype
= TREE_TYPE (ptr
);
729 gcc_assert (POINTER_TYPE_P (ptrtype
));
731 /* (T) *(volatile T *) ptr; */
732 tree type
= build_qualified_type (TREE_TYPE (ptrtype
), TYPE_QUAL_VOLATILE
);
733 tree result
= indirect_ref (type
, ptr
);
734 TREE_THIS_VOLATILE (result
) = 1;
736 /* (*(volatile T *) ptr) = value; */
737 tree value
= CALL_EXPR_ARG (callexp
, 1);
738 return modify_expr (result
, value
);
741 /* If CALLEXP is for an intrinsic , expand and return inlined compiler
742 generated instructions. Most map directly to GCC builtins, others
743 require a little extra work around them. */
746 maybe_expand_intrinsic (tree callexp
)
748 tree callee
= CALL_EXPR_FN (callexp
);
750 if (TREE_CODE (callee
) == ADDR_EXPR
)
751 callee
= TREE_OPERAND (callee
, 0);
753 if (TREE_CODE (callee
) != FUNCTION_DECL
)
756 /* Don't expand CTFE-only intrinsics outside of semantic processing. */
757 if (DECL_BUILT_IN_CTFE (callee
) && !doing_semantic_analysis_p
)
760 intrinsic_code intrinsic
= DECL_INTRINSIC_CODE (callee
);
761 built_in_function code
;
769 case INTRINSIC_BSF64
:
770 return expand_intrinsic_bsf (callexp
);
773 case INTRINSIC_BSR64
:
774 return expand_intrinsic_bsr (callexp
);
779 case INTRINSIC_BTC64
:
781 case INTRINSIC_BTR64
:
783 case INTRINSIC_BTS64
:
784 return expand_intrinsic_bt (intrinsic
, callexp
);
786 case INTRINSIC_POPCNT32
:
787 case INTRINSIC_POPCNT64
:
788 return expand_intrinsic_popcnt (callexp
);
791 case INTRINSIC_ROL_TIARG
:
793 case INTRINSIC_ROR_TIARG
:
794 return expand_intrinsic_rotate (intrinsic
, callexp
);
796 case INTRINSIC_BSWAP32
:
797 case INTRINSIC_BSWAP64
:
799 case INTRINSIC_CEILF
:
800 case INTRINSIC_CEILL
:
804 case INTRINSIC_EXPM1
:
805 case INTRINSIC_FABSL
:
806 case INTRINSIC_FLOOR
:
807 case INTRINSIC_FLOORF
:
808 case INTRINSIC_FLOORL
:
809 case INTRINSIC_ISFINITE
:
810 case INTRINSIC_ISINFINITY
:
811 case INTRINSIC_ISNAN
:
813 case INTRINSIC_LOG10
:
815 case INTRINSIC_RINTL
:
816 case INTRINSIC_RNDTOLL
:
817 case INTRINSIC_ROUND
:
820 case INTRINSIC_SQRTF
:
821 case INTRINSIC_SQRTL
:
823 case INTRINSIC_TRUNC
:
824 code
= intrinsic_decls
[intrinsic
].built_in
;
825 gcc_assert (code
!= BUILT_IN_NONE
);
826 return call_builtin_fn (callexp
, code
, 1,
827 CALL_EXPR_ARG (callexp
, 0));
831 case INTRINSIC_LDEXPL
:
832 code
= intrinsic_decls
[intrinsic
].built_in
;
833 gcc_assert (code
!= BUILT_IN_NONE
);
834 return call_builtin_fn (callexp
, code
, 2,
835 CALL_EXPR_ARG (callexp
, 0),
836 CALL_EXPR_ARG (callexp
, 1));
839 code
= intrinsic_decls
[intrinsic
].built_in
;
840 gcc_assert (code
!= BUILT_IN_NONE
);
841 return call_builtin_fn (callexp
, code
, 3,
842 CALL_EXPR_ARG (callexp
, 0),
843 CALL_EXPR_ARG (callexp
, 1),
844 CALL_EXPR_ARG (callexp
, 2));
846 case INTRINSIC_COPYSIGN
:
847 case INTRINSIC_COPYSIGNI
:
848 return expand_intrinsic_copysign (callexp
);
851 return expand_intrinsic_pow (callexp
);
853 case INTRINSIC_TOPREC
:
854 case INTRINSIC_TOPRECF
:
855 case INTRINSIC_TOPRECL
:
856 return expand_intrinsic_toprec (callexp
);
858 case INTRINSIC_VA_ARG
:
859 case INTRINSIC_C_VA_ARG
:
860 return expand_intrinsic_vaarg (callexp
);
862 case INTRINSIC_VASTART
:
863 return expand_intrinsic_vastart (callexp
);
866 case INTRINSIC_ADDSL
:
868 case INTRINSIC_ADDUL
:
870 case INTRINSIC_SUBSL
:
872 case INTRINSIC_SUBUL
:
874 case INTRINSIC_MULSL
:
876 case INTRINSIC_MULUI
:
877 case INTRINSIC_MULUL
:
879 case INTRINSIC_NEGSL
:
880 return expand_intrinsic_checkedint (intrinsic
, callexp
);
882 case INTRINSIC_VLOAD8
:
883 case INTRINSIC_VLOAD16
:
884 case INTRINSIC_VLOAD32
:
885 case INTRINSIC_VLOAD64
:
886 return expand_volatile_load (callexp
);
888 case INTRINSIC_VSTORE8
:
889 case INTRINSIC_VSTORE16
:
890 case INTRINSIC_VSTORE32
:
891 case INTRINSIC_VSTORE64
:
892 return expand_volatile_store (callexp
);